1 /************************************************************************** 2 * 3 * Copyright (c) 2006-2009 VMware, Inc., Palo Alto, CA., USA 4 * All Rights Reserved. 5 * 6 * Permission is hereby granted, free of charge, to any person obtaining a 7 * copy of this software and associated documentation files (the 8 * "Software"), to deal in the Software without restriction, including 9 * without limitation the rights to use, copy, modify, merge, publish, 10 * distribute, sub license, and/or sell copies of the Software, and to 11 * permit persons to whom the Software is furnished to do so, subject to 12 * the following conditions: 13 * 14 * The above copyright notice and this permission notice (including the 15 * next paragraph) shall be included in all copies or substantial portions 16 * of the Software. 17 * 18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 20 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL 21 * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, 22 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR 23 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE 24 * USE OR OTHER DEALINGS IN THE SOFTWARE. 25 * 26 **************************************************************************/ 27 /* 28 * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com> 29 */ 30 /* $FreeBSD: head/sys/dev/drm2/ttm/ttm_bo_api.h 247835 2013-03-05 09:49:34Z kib $ */ 31 32 #ifndef _TTM_BO_API_H_ 33 #define _TTM_BO_API_H_ 34 35 #include <drm/drmP.h> 36 #include <drm/drm_hashtab.h> 37 #include <linux/kref.h> 38 #include <linux/list.h> 39 #include <linux/wait.h> 40 #include <linux/mutex.h> 41 #include <linux/reservation.h> 42 43 struct ttm_bo_device; 44 45 struct drm_mm_node; 46 47 /** 48 * struct ttm_place 49 * 50 * @fpfn: first valid page frame number to put the object 51 * @lpfn: last valid page frame number to put the object 52 * @flags: memory domain and caching flags for the object 53 * 54 * Structure indicating a possible place to put an object. 55 */ 56 struct ttm_place { 57 unsigned fpfn; 58 unsigned lpfn; 59 uint32_t flags; 60 }; 61 62 /** 63 * struct ttm_placement 64 * 65 * @num_placement: number of preferred placements 66 * @placement: preferred placements 67 * @num_busy_placement: number of preferred placements when need to evict buffer 68 * @busy_placement: preferred placements when need to evict buffer 69 * 70 * Structure indicating the placement you request for an object. 71 */ 72 struct ttm_placement { 73 unsigned num_placement; 74 const struct ttm_place *placement; 75 unsigned num_busy_placement; 76 const struct ttm_place *busy_placement; 77 }; 78 79 /** 80 * struct ttm_bus_placement 81 * 82 * @addr: mapped virtual address 83 * @base: bus base address 84 * @is_iomem: is this io memory ? 85 * @size: size in byte 86 * @offset: offset from the base address 87 * @io_reserved_vm: The VM system has a refcount in @io_reserved_count 88 * @io_reserved_count: Refcounting the numbers of callers to ttm_mem_io_reserve 89 * 90 * Structure indicating the bus placement of an object. 91 */ 92 struct ttm_bus_placement { 93 void *addr; 94 unsigned long base; 95 unsigned long size; 96 unsigned long offset; 97 bool is_iomem; 98 bool io_reserved_vm; 99 uint64_t io_reserved_count; 100 }; 101 102 103 /** 104 * struct ttm_mem_reg 105 * 106 * @mm_node: Memory manager node. 107 * @size: Requested size of memory region. 108 * @num_pages: Actual size of memory region in pages. 109 * @page_alignment: Page alignment. 110 * @placement: Placement flags. 111 * @bus: Placement on io bus accessible to the CPU 112 * 113 * Structure indicating the placement and space resources used by a 114 * buffer object. 115 */ 116 117 struct ttm_mem_reg { 118 void *mm_node; 119 unsigned long start; 120 unsigned long size; 121 unsigned long num_pages; 122 uint32_t page_alignment; 123 uint32_t mem_type; 124 uint32_t placement; 125 struct ttm_bus_placement bus; 126 }; 127 128 /** 129 * enum ttm_bo_type 130 * 131 * @ttm_bo_type_device: These are 'normal' buffers that can 132 * be mmapped by user space. Each of these bos occupy a slot in the 133 * device address space, that can be used for normal vm operations. 134 * 135 * @ttm_bo_type_kernel: These buffers are like ttm_bo_type_device buffers, 136 * but they cannot be accessed from user-space. For kernel-only use. 137 * 138 * @ttm_bo_type_sg: Buffer made from dmabuf sg table shared with another 139 * driver. 140 */ 141 142 enum ttm_bo_type { 143 ttm_bo_type_device, 144 ttm_bo_type_kernel, 145 ttm_bo_type_sg 146 }; 147 148 struct ttm_tt; 149 150 /** 151 * struct ttm_buffer_object 152 * 153 * @bdev: Pointer to the buffer object device structure. 154 * @type: The bo type. 155 * @destroy: Destruction function. If NULL, kfree is used. 156 * @num_pages: Actual number of pages. 157 * @addr_space_offset: Address space offset. 158 * @acc_size: Accounted size for this object. 159 * @kref: Reference count of this buffer object. When this refcount reaches 160 * zero, the object is put on the delayed delete list. 161 * @list_kref: List reference count of this buffer object. This member is 162 * used to avoid destruction while the buffer object is still on a list. 163 * Lru lists may keep one refcount, the delayed delete list, and kref != 0 164 * keeps one refcount. When this refcount reaches zero, 165 * the object is destroyed. 166 * @event_queue: Queue for processes waiting on buffer object status change. 167 * @mem: structure describing current placement. 168 * @persistent_swap_storage: Usually the swap storage is deleted for buffers 169 * pinned in physical memory. If this behaviour is not desired, this member 170 * holds a pointer to a persistent shmem object. 171 * @ttm: TTM structure holding system pages. 172 * @evicted: Whether the object was evicted without user-space knowing. 173 * @cpu_writes: For synchronization. Number of cpu writers. 174 * @lru: List head for the lru list. 175 * @ddestroy: List head for the delayed destroy list. 176 * @swap: List head for swap LRU list. 177 * @val_seq: Sequence of the validation holding the @reserved lock. 178 * Used to avoid starvation when many processes compete to validate the 179 * buffer. This member is protected by the bo_device::lru_lock. 180 * @seq_valid: The value of @val_seq is valid. This value is protected by 181 * the bo_device::lru_lock. 182 * @reserved: Deadlock-free lock used for synchronization state transitions. 183 * @sync_obj: Pointer to a synchronization object. 184 * @priv_flags: Flags describing buffer object internal state. 185 * @vm_rb: Rb node for the vm rb tree. 186 * @vm_node: Address space manager node. 187 * @offset: The current GPU offset, which can have different meanings 188 * depending on the memory type. For SYSTEM type memory, it should be 0. 189 * @cur_placement: Hint of current placement. 190 * 191 * Base class for TTM buffer object, that deals with data placement and CPU 192 * mappings. GPU mappings are really up to the driver, but for simpler GPUs 193 * the driver can usually use the placement offset @offset directly as the 194 * GPU virtual address. For drivers implementing multiple 195 * GPU memory manager contexts, the driver should manage the address space 196 * in these contexts separately and use these objects to get the correct 197 * placement and caching for these GPU maps. This makes it possible to use 198 * these objects for even quite elaborate memory management schemes. 199 * The destroy member, the API visibility of this object makes it possible 200 * to derive driver specific types. 201 */ 202 203 struct ttm_buffer_object { 204 /** 205 * Members constant at init. 206 */ 207 208 struct ttm_bo_global *glob; 209 struct ttm_bo_device *bdev; 210 enum ttm_bo_type type; 211 void (*destroy) (struct ttm_buffer_object *); 212 unsigned long num_pages; 213 uint64_t addr_space_offset; 214 size_t acc_size; 215 216 /** 217 * Members not needing protection. 218 */ 219 220 struct kref kref; 221 struct kref list_kref; 222 223 wait_queue_head_t event_queue; 224 225 /** 226 * Members protected by the bo::reserved lock. 227 */ 228 229 struct ttm_mem_reg mem; 230 struct vm_object *persistent_swap_storage; 231 struct ttm_tt *ttm; 232 bool evicted; 233 234 /** 235 * Members protected by the bo::reserved lock only when written to. 236 */ 237 238 atomic_t cpu_writers; 239 240 /** 241 * Members protected by the bdev::lru_lock. 242 */ 243 244 struct list_head lru; 245 struct list_head ddestroy; 246 struct list_head swap; 247 struct list_head io_reserve_lru; 248 unsigned long val_seq; 249 bool seq_valid; 250 251 /** 252 * Members protected by the bdev::lru_lock 253 * only when written to. 254 */ 255 256 atomic_t reserved; 257 258 /** 259 * Members protected by struct buffer_object_device::fence_lock 260 * In addition, setting sync_obj to anything else 261 * than NULL requires bo::reserved to be held. This allows for 262 * checking NULL while reserved but not holding the mentioned lock. 263 */ 264 265 void *sync_obj; 266 unsigned long priv_flags; 267 268 /** 269 * Members protected by the bdev::vm_lock 270 */ 271 272 RB_ENTRY(ttm_buffer_object) vm_rb; 273 struct drm_mm_node *vm_node; 274 275 276 /** 277 * Special members that are protected by the reserve lock 278 * and the bo::lock when written to. Can be read with 279 * either of these locks held. 280 */ 281 282 unsigned long offset; 283 uint32_t cur_placement; 284 285 struct sg_table *sg; 286 }; 287 288 /** 289 * struct ttm_bo_kmap_obj 290 * 291 * @virtual: The current kernel virtual address. 292 * @page: The page when kmap'ing a single page. 293 * @bo_kmap_type: Type of bo_kmap. 294 * 295 * Object describing a kernel mapping. Since a TTM bo may be located 296 * in various memory types with various caching policies, the 297 * mapping can either be an ioremap, a vmap, a kmap or part of a 298 * premapped region. 299 */ 300 301 #define TTM_BO_MAP_IOMEM_MASK 0x80 302 struct ttm_bo_kmap_obj { 303 void *virtual; 304 struct vm_page *page; 305 struct sf_buf *sf; 306 int num_pages; 307 unsigned long size; 308 enum { 309 ttm_bo_map_iomap = 1 | TTM_BO_MAP_IOMEM_MASK, 310 ttm_bo_map_vmap = 2, 311 ttm_bo_map_kmap = 3, 312 ttm_bo_map_premapped = 4 | TTM_BO_MAP_IOMEM_MASK, 313 } bo_kmap_type; 314 struct ttm_buffer_object *bo; 315 }; 316 317 /** 318 * ttm_bo_reference - reference a struct ttm_buffer_object 319 * 320 * @bo: The buffer object. 321 * 322 * Returns a refcounted pointer to a buffer object. 323 */ 324 325 static inline struct ttm_buffer_object * 326 ttm_bo_reference(struct ttm_buffer_object *bo) 327 { 328 kref_get(&bo->kref); 329 return bo; 330 } 331 332 /** 333 * ttm_bo_wait - wait for buffer idle. 334 * 335 * @bo: The buffer object. 336 * @interruptible: Use interruptible wait. 337 * @no_wait: Return immediately if buffer is busy. 338 * 339 * This function must be called with the bo::mutex held, and makes 340 * sure any previous rendering to the buffer is completed. 341 * Note: It might be necessary to block validations before the 342 * wait by reserving the buffer. 343 * Returns -EBUSY if no_wait is true and the buffer is busy. 344 * Returns -ERESTARTSYS if interrupted by a signal. 345 */ 346 extern int ttm_bo_wait(struct ttm_buffer_object *bo, bool lazy, 347 bool interruptible, bool no_wait); 348 /** 349 * ttm_bo_validate 350 * 351 * @bo: The buffer object. 352 * @placement: Proposed placement for the buffer object. 353 * @interruptible: Sleep interruptible if sleeping. 354 * @no_wait_gpu: Return immediately if the GPU is busy. 355 * 356 * Changes placement and caching policy of the buffer object 357 * according proposed placement. 358 * Returns 359 * -EINVAL on invalid proposed placement. 360 * -ENOMEM on out-of-memory condition. 361 * -EBUSY if no_wait is true and buffer busy. 362 * -ERESTARTSYS if interrupted by a signal. 363 */ 364 extern int ttm_bo_validate(struct ttm_buffer_object *bo, 365 struct ttm_placement *placement, 366 bool interruptible, 367 bool no_wait_gpu); 368 369 /** 370 * ttm_bo_unref 371 * 372 * @bo: The buffer object. 373 * 374 * Unreference and clear a pointer to a buffer object. 375 */ 376 extern void ttm_bo_unref(struct ttm_buffer_object **bo); 377 378 379 /** 380 * ttm_bo_list_ref_sub 381 * 382 * @bo: The buffer object. 383 * @count: The number of references with which to decrease @bo::list_kref; 384 * @never_free: The refcount should not reach zero with this operation. 385 * 386 * Release @count lru list references to this buffer object. 387 */ 388 extern void ttm_bo_list_ref_sub(struct ttm_buffer_object *bo, int count, 389 bool never_free); 390 391 /** 392 * ttm_bo_add_to_lru 393 * 394 * @bo: The buffer object. 395 * 396 * Add this bo to the relevant mem type lru and, if it's backed by 397 * system pages (ttms) to the swap list. 398 * This function must be called with struct ttm_bo_global::lru_lock held, and 399 * is typically called immediately prior to unreserving a bo. 400 */ 401 extern void ttm_bo_add_to_lru(struct ttm_buffer_object *bo); 402 403 /** 404 * ttm_bo_del_from_lru 405 * 406 * @bo: The buffer object. 407 * 408 * Remove this bo from all lru lists used to lookup and reserve an object. 409 * This function must be called with struct ttm_bo_global::lru_lock held, 410 * and is usually called just immediately after the bo has been reserved to 411 * avoid recursive reservation from lru lists. 412 */ 413 extern int ttm_bo_del_from_lru(struct ttm_buffer_object *bo); 414 415 416 /** 417 * ttm_bo_lock_delayed_workqueue 418 * 419 * Prevent the delayed workqueue from running. 420 * Returns 421 * True if the workqueue was queued at the time 422 */ 423 extern int ttm_bo_lock_delayed_workqueue(struct ttm_bo_device *bdev); 424 425 /** 426 * ttm_bo_unlock_delayed_workqueue 427 * 428 * Allows the delayed workqueue to run. 429 */ 430 extern void ttm_bo_unlock_delayed_workqueue(struct ttm_bo_device *bdev, 431 int resched); 432 433 /** 434 * ttm_bo_synccpu_write_grab 435 * 436 * @bo: The buffer object: 437 * @no_wait: Return immediately if buffer is busy. 438 * 439 * Synchronizes a buffer object for CPU RW access. This means 440 * command submission that affects the buffer will return -EBUSY 441 * until ttm_bo_synccpu_write_release is called. 442 * 443 * Returns 444 * -EBUSY if the buffer is busy and no_wait is true. 445 * -ERESTARTSYS if interrupted by a signal. 446 */ 447 extern int 448 ttm_bo_synccpu_write_grab(struct ttm_buffer_object *bo, bool no_wait); 449 450 /** 451 * ttm_bo_synccpu_write_release: 452 * 453 * @bo : The buffer object. 454 * 455 * Releases a synccpu lock. 456 */ 457 extern void ttm_bo_synccpu_write_release(struct ttm_buffer_object *bo); 458 459 /** 460 * ttm_bo_acc_size 461 * 462 * @bdev: Pointer to a ttm_bo_device struct. 463 * @bo_size: size of the buffer object in byte. 464 * @struct_size: size of the structure holding buffer object datas 465 * 466 * Returns size to account for a buffer object 467 */ 468 size_t ttm_bo_acc_size(struct ttm_bo_device *bdev, 469 unsigned long bo_size, 470 unsigned struct_size); 471 size_t ttm_bo_dma_acc_size(struct ttm_bo_device *bdev, 472 unsigned long bo_size, 473 unsigned struct_size); 474 475 /** 476 * ttm_bo_init 477 * 478 * @bdev: Pointer to a ttm_bo_device struct. 479 * @bo: Pointer to a ttm_buffer_object to be initialized. 480 * @size: Requested size of buffer object. 481 * @type: Requested type of buffer object. 482 * @flags: Initial placement flags. 483 * @page_alignment: Data alignment in pages. 484 * @interruptible: If needing to sleep to wait for GPU resources, 485 * sleep interruptible. 486 * @persistent_swap_storage: Usually the swap storage is deleted for buffers 487 * pinned in physical memory. If this behaviour is not desired, this member 488 * holds a pointer to a persistent shmem object. Typically, this would 489 * point to the shmem object backing a GEM object if TTM is used to back a 490 * GEM user interface. 491 * @acc_size: Accounted size for this object. 492 * @destroy: Destroy function. Use NULL for kfree(). 493 * 494 * This function initializes a pre-allocated struct ttm_buffer_object. 495 * As this object may be part of a larger structure, this function, 496 * together with the @destroy function, 497 * enables driver-specific objects derived from a ttm_buffer_object. 498 * On successful return, the object kref and list_kref are set to 1. 499 * If a failure occurs, the function will call the @destroy function, or 500 * kfree() if @destroy is NULL. Thus, after a failure, dereferencing @bo is 501 * illegal and will likely cause memory corruption. 502 * 503 * Returns 504 * -ENOMEM: Out of memory. 505 * -EINVAL: Invalid placement flags. 506 * -ERESTARTSYS: Interrupted by signal while sleeping waiting for resources. 507 */ 508 509 extern int ttm_bo_init(struct ttm_bo_device *bdev, 510 struct ttm_buffer_object *bo, 511 unsigned long size, 512 enum ttm_bo_type type, 513 struct ttm_placement *placement, 514 uint32_t page_alignment, 515 bool interrubtible, 516 struct vm_object *persistent_swap_storage, 517 size_t acc_size, 518 struct sg_table *sg, 519 void (*destroy) (struct ttm_buffer_object *)); 520 521 /** 522 * ttm_bo_create 523 * 524 * @bdev: Pointer to a ttm_bo_device struct. 525 * @size: Requested size of buffer object. 526 * @type: Requested type of buffer object. 527 * @placement: Initial placement. 528 * @page_alignment: Data alignment in pages. 529 * @interruptible: If needing to sleep while waiting for GPU resources, 530 * sleep interruptible. 531 * @persistent_swap_storage: Usually the swap storage is deleted for buffers 532 * pinned in physical memory. If this behaviour is not desired, this member 533 * holds a pointer to a persistent shmem object. Typically, this would 534 * point to the shmem object backing a GEM object if TTM is used to back a 535 * GEM user interface. 536 * @p_bo: On successful completion *p_bo points to the created object. 537 * 538 * This function allocates a ttm_buffer_object, and then calls ttm_bo_init 539 * on that object. The destroy function is set to kfree(). 540 * Returns 541 * -ENOMEM: Out of memory. 542 * -EINVAL: Invalid placement flags. 543 * -ERESTARTSYS: Interrupted by signal while waiting for resources. 544 */ 545 546 extern int ttm_bo_create(struct ttm_bo_device *bdev, 547 unsigned long size, 548 enum ttm_bo_type type, 549 struct ttm_placement *placement, 550 uint32_t page_alignment, 551 bool interruptible, 552 struct vm_object *persistent_swap_storage, 553 struct ttm_buffer_object **p_bo); 554 555 /** 556 * ttm_bo_init_mm 557 * 558 * @bdev: Pointer to a ttm_bo_device struct. 559 * @mem_type: The memory type. 560 * @p_size: size managed area in pages. 561 * 562 * Initialize a manager for a given memory type. 563 * Note: if part of driver firstopen, it must be protected from a 564 * potentially racing lastclose. 565 * Returns: 566 * -EINVAL: invalid size or memory type. 567 * -ENOMEM: Not enough memory. 568 * May also return driver-specified errors. 569 */ 570 571 extern int ttm_bo_init_mm(struct ttm_bo_device *bdev, unsigned type, 572 unsigned long p_size); 573 /** 574 * ttm_bo_clean_mm 575 * 576 * @bdev: Pointer to a ttm_bo_device struct. 577 * @mem_type: The memory type. 578 * 579 * Take down a manager for a given memory type after first walking 580 * the LRU list to evict any buffers left alive. 581 * 582 * Normally, this function is part of lastclose() or unload(), and at that 583 * point there shouldn't be any buffers left created by user-space, since 584 * there should've been removed by the file descriptor release() method. 585 * However, before this function is run, make sure to signal all sync objects, 586 * and verify that the delayed delete queue is empty. The driver must also 587 * make sure that there are no NO_EVICT buffers present in this memory type 588 * when the call is made. 589 * 590 * If this function is part of a VT switch, the caller must make sure that 591 * there are no appications currently validating buffers before this 592 * function is called. The caller can do that by first taking the 593 * struct ttm_bo_device::ttm_lock in write mode. 594 * 595 * Returns: 596 * -EINVAL: invalid or uninitialized memory type. 597 * -EBUSY: There are still buffers left in this memory type. 598 */ 599 600 extern int ttm_bo_clean_mm(struct ttm_bo_device *bdev, unsigned mem_type); 601 602 /** 603 * ttm_bo_evict_mm 604 * 605 * @bdev: Pointer to a ttm_bo_device struct. 606 * @mem_type: The memory type. 607 * 608 * Evicts all buffers on the lru list of the memory type. 609 * This is normally part of a VT switch or an 610 * out-of-memory-space-due-to-fragmentation handler. 611 * The caller must make sure that there are no other processes 612 * currently validating buffers, and can do that by taking the 613 * struct ttm_bo_device::ttm_lock in write mode. 614 * 615 * Returns: 616 * -EINVAL: Invalid or uninitialized memory type. 617 * -ERESTARTSYS: The call was interrupted by a signal while waiting to 618 * evict a buffer. 619 */ 620 621 extern int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type); 622 623 /** 624 * ttm_kmap_obj_virtual 625 * 626 * @map: A struct ttm_bo_kmap_obj returned from ttm_bo_kmap. 627 * @is_iomem: Pointer to an integer that on return indicates 1 if the 628 * virtual map is io memory, 0 if normal memory. 629 * 630 * Returns the virtual address of a buffer object area mapped by ttm_bo_kmap. 631 * If *is_iomem is 1 on return, the virtual address points to an io memory area, 632 * that should strictly be accessed by the iowriteXX() and similar functions. 633 */ 634 635 static inline void *ttm_kmap_obj_virtual(struct ttm_bo_kmap_obj *map, 636 bool *is_iomem) 637 { 638 *is_iomem = !!(map->bo_kmap_type & TTM_BO_MAP_IOMEM_MASK); 639 return map->virtual; 640 } 641 642 /** 643 * ttm_bo_kmap 644 * 645 * @bo: The buffer object. 646 * @start_page: The first page to map. 647 * @num_pages: Number of pages to map. 648 * @map: pointer to a struct ttm_bo_kmap_obj representing the map. 649 * 650 * Sets up a kernel virtual mapping, using ioremap, vmap or kmap to the 651 * data in the buffer object. The ttm_kmap_obj_virtual function can then be 652 * used to obtain a virtual address to the data. 653 * 654 * Returns 655 * -ENOMEM: Out of memory. 656 * -EINVAL: Invalid range. 657 */ 658 659 extern int ttm_bo_kmap(struct ttm_buffer_object *bo, unsigned long start_page, 660 unsigned long num_pages, struct ttm_bo_kmap_obj *map); 661 662 /** 663 * ttm_bo_kunmap 664 * 665 * @map: Object describing the map to unmap. 666 * 667 * Unmaps a kernel map set up by ttm_bo_kmap. 668 */ 669 670 extern void ttm_bo_kunmap(struct ttm_bo_kmap_obj *map); 671 672 /** 673 * ttm_fbdev_mmap - mmap fbdev memory backed by a ttm buffer object. 674 * 675 * @vma: vma as input from the fbdev mmap method. 676 * @bo: The bo backing the address space. The address space will 677 * have the same size as the bo, and start at offset 0. 678 * 679 * This function is intended to be called by the fbdev mmap method 680 * if the fbdev address space is to be backed by a bo. 681 */ 682 683 /* XXXKIB 684 extern int ttm_fbdev_mmap(struct vm_area_struct *vma, 685 struct ttm_buffer_object *bo); 686 */ 687 /** 688 * ttm_bo_mmap - mmap out of the ttm device address space. 689 * 690 * @filp: filp as input from the mmap method. 691 * @vma: vma as input from the mmap method. 692 * @bdev: Pointer to the ttm_bo_device with the address space manager. 693 * 694 * This function is intended to be called by the device mmap method. 695 * if the device address space is to be backed by the bo manager. 696 */ 697 /* XXXKIB 698 extern int ttm_bo_mmap(struct file *filp, struct vm_area_struct *vma, 699 struct ttm_bo_device *bdev); 700 */ 701 /** 702 * ttm_bo_io 703 * 704 * @bdev: Pointer to the struct ttm_bo_device. 705 * @filp: Pointer to the struct file attempting to read / write. 706 * @wbuf: User-space pointer to address of buffer to write. NULL on read. 707 * @rbuf: User-space pointer to address of buffer to read into. 708 * Null on write. 709 * @count: Number of bytes to read / write. 710 * @f_pos: Pointer to current file position. 711 * @write: 1 for read, 0 for write. 712 * 713 * This function implements read / write into ttm buffer objects, and is 714 * intended to 715 * be called from the fops::read and fops::write method. 716 * Returns: 717 * See man (2) write, man(2) read. In particular, 718 * the function may return -ERESTARTSYS if 719 * interrupted by a signal. 720 */ 721 722 extern ssize_t ttm_bo_io(struct ttm_bo_device *bdev, struct file *filp, 723 const char *wbuf, char *rbuf, 724 size_t count, off_t *f_pos, bool write); 725 726 extern void ttm_bo_swapout_all(struct ttm_bo_device *bdev); 727 728 /** 729 * ttm_bo_is_reserved - return an indication if a ttm buffer object is reserved 730 * 731 * @bo: The buffer object to check. 732 * 733 * This function returns an indication if a bo is reserved or not, and should 734 * only be used to print an error when it is not from incorrect api usage, since 735 * there's no guarantee that it is the caller that is holding the reservation. 736 */ 737 static inline bool ttm_bo_is_reserved(struct ttm_buffer_object *bo) 738 { 739 return atomic_read(&bo->reserved); 740 } 741 742 #endif 743