2 * Copyright 2012 Red Hat Inc.
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the
6 * "Software"), to deal in the Software without restriction, including
7 * without limitation the rights to use, copy, modify, merge, publish,
8 * distribute, sub license, and/or sell copies of the Software, and to
9 * permit persons to whom the Software is furnished to do so, subject to
10 * the following conditions:
12 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
13 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
14 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
15 * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM,
16 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
17 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
18 * USE OR OTHER DEALINGS IN THE SOFTWARE.
20 * The above copyright notice and this permission notice (including the
21 * next paragraph) shall be included in all copies or substantial portions
26 * Authors: Dave Airlie <airlied@redhat.com>
29 #include <drm/ttm/ttm_page_alloc.h>
31 #include "cirrus_drv.h"
33 static inline struct cirrus_device
*
34 cirrus_bdev(struct ttm_bo_device
*bd
)
36 return container_of(bd
, struct cirrus_device
, ttm
.bdev
);
39 static void cirrus_bo_ttm_destroy(struct ttm_buffer_object
*tbo
)
43 bo
= container_of(tbo
, struct cirrus_bo
, bo
);
45 drm_gem_object_release(&bo
->gem
);
49 static bool cirrus_ttm_bo_is_cirrus_bo(struct ttm_buffer_object
*bo
)
51 if (bo
->destroy
== &cirrus_bo_ttm_destroy
)
57 cirrus_bo_init_mem_type(struct ttm_bo_device
*bdev
, uint32_t type
,
58 struct ttm_mem_type_manager
*man
)
62 man
->flags
= TTM_MEMTYPE_FLAG_MAPPABLE
;
63 man
->available_caching
= TTM_PL_MASK_CACHING
;
64 man
->default_caching
= TTM_PL_FLAG_CACHED
;
67 man
->func
= &ttm_bo_manager_func
;
68 man
->flags
= TTM_MEMTYPE_FLAG_FIXED
|
69 TTM_MEMTYPE_FLAG_MAPPABLE
;
70 man
->available_caching
= TTM_PL_FLAG_UNCACHED
|
72 man
->default_caching
= TTM_PL_FLAG_WC
;
75 DRM_ERROR("Unsupported memory type %u\n", (unsigned)type
);
82 cirrus_bo_evict_flags(struct ttm_buffer_object
*bo
, struct ttm_placement
*pl
)
84 struct cirrus_bo
*cirrusbo
= cirrus_bo(bo
);
86 if (!cirrus_ttm_bo_is_cirrus_bo(bo
))
89 cirrus_ttm_placement(cirrusbo
, TTM_PL_FLAG_SYSTEM
);
90 *pl
= cirrusbo
->placement
;
93 static int cirrus_bo_verify_access(struct ttm_buffer_object
*bo
, struct file
*filp
)
95 struct cirrus_bo
*cirrusbo
= cirrus_bo(bo
);
97 return drm_vma_node_verify_access(&cirrusbo
->gem
.vma_node
,
101 static int cirrus_ttm_io_mem_reserve(struct ttm_bo_device
*bdev
,
102 struct ttm_mem_reg
*mem
)
104 struct ttm_mem_type_manager
*man
= &bdev
->man
[mem
->mem_type
];
105 struct cirrus_device
*cirrus
= cirrus_bdev(bdev
);
107 mem
->bus
.addr
= NULL
;
109 mem
->bus
.size
= mem
->num_pages
<< PAGE_SHIFT
;
111 mem
->bus
.is_iomem
= false;
112 if (!(man
->flags
& TTM_MEMTYPE_FLAG_MAPPABLE
))
114 switch (mem
->mem_type
) {
119 mem
->bus
.offset
= mem
->start
<< PAGE_SHIFT
;
120 mem
->bus
.base
= pci_resource_start(cirrus
->dev
->pdev
, 0);
121 mem
->bus
.is_iomem
= true;
130 static void cirrus_ttm_io_mem_free(struct ttm_bo_device
*bdev
, struct ttm_mem_reg
*mem
)
134 static void cirrus_ttm_backend_destroy(struct ttm_tt
*tt
)
140 static struct ttm_backend_func cirrus_tt_backend_func
= {
141 .destroy
= &cirrus_ttm_backend_destroy
,
145 static struct ttm_tt
*cirrus_ttm_tt_create(struct ttm_buffer_object
*bo
,
150 tt
= kzalloc(sizeof(struct ttm_tt
), GFP_KERNEL
);
153 tt
->func
= &cirrus_tt_backend_func
;
154 if (ttm_tt_init(tt
, bo
, page_flags
)) {
161 struct ttm_bo_driver cirrus_bo_driver
= {
162 .ttm_tt_create
= cirrus_ttm_tt_create
,
163 .init_mem_type
= cirrus_bo_init_mem_type
,
164 .eviction_valuable
= ttm_bo_eviction_valuable
,
165 .evict_flags
= cirrus_bo_evict_flags
,
167 .verify_access
= cirrus_bo_verify_access
,
168 .io_mem_reserve
= &cirrus_ttm_io_mem_reserve
,
169 .io_mem_free
= &cirrus_ttm_io_mem_free
,
172 int cirrus_mm_init(struct cirrus_device
*cirrus
)
175 struct drm_device
*dev
= cirrus
->dev
;
176 struct ttm_bo_device
*bdev
= &cirrus
->ttm
.bdev
;
178 ret
= ttm_bo_device_init(&cirrus
->ttm
.bdev
,
180 dev
->anon_inode
->i_mapping
,
183 DRM_ERROR("Error initialising bo driver; %d\n", ret
);
187 ret
= ttm_bo_init_mm(bdev
, TTM_PL_VRAM
,
188 cirrus
->mc
.vram_size
>> PAGE_SHIFT
);
190 DRM_ERROR("Failed ttm VRAM init: %d\n", ret
);
194 arch_io_reserve_memtype_wc(pci_resource_start(dev
->pdev
, 0),
195 pci_resource_len(dev
->pdev
, 0));
197 cirrus
->fb_mtrr
= arch_phys_wc_add(pci_resource_start(dev
->pdev
, 0),
198 pci_resource_len(dev
->pdev
, 0));
200 cirrus
->mm_inited
= true;
204 void cirrus_mm_fini(struct cirrus_device
*cirrus
)
206 struct drm_device
*dev
= cirrus
->dev
;
208 if (!cirrus
->mm_inited
)
211 ttm_bo_device_release(&cirrus
->ttm
.bdev
);
213 arch_phys_wc_del(cirrus
->fb_mtrr
);
215 arch_io_free_memtype_wc(pci_resource_start(dev
->pdev
, 0),
216 pci_resource_len(dev
->pdev
, 0));
219 void cirrus_ttm_placement(struct cirrus_bo
*bo
, int domain
)
223 bo
->placement
.placement
= bo
->placements
;
224 bo
->placement
.busy_placement
= bo
->placements
;
225 if (domain
& TTM_PL_FLAG_VRAM
)
226 bo
->placements
[c
++].flags
= TTM_PL_FLAG_WC
| TTM_PL_FLAG_UNCACHED
| TTM_PL_FLAG_VRAM
;
227 if (domain
& TTM_PL_FLAG_SYSTEM
)
228 bo
->placements
[c
++].flags
= TTM_PL_MASK_CACHING
| TTM_PL_FLAG_SYSTEM
;
230 bo
->placements
[c
++].flags
= TTM_PL_MASK_CACHING
| TTM_PL_FLAG_SYSTEM
;
231 bo
->placement
.num_placement
= c
;
232 bo
->placement
.num_busy_placement
= c
;
233 for (i
= 0; i
< c
; ++i
) {
234 bo
->placements
[i
].fpfn
= 0;
235 bo
->placements
[i
].lpfn
= 0;
239 int cirrus_bo_create(struct drm_device
*dev
, int size
, int align
,
240 uint32_t flags
, struct cirrus_bo
**pcirrusbo
)
242 struct cirrus_device
*cirrus
= dev
->dev_private
;
243 struct cirrus_bo
*cirrusbo
;
247 cirrusbo
= kzalloc(sizeof(struct cirrus_bo
), GFP_KERNEL
);
251 ret
= drm_gem_object_init(dev
, &cirrusbo
->gem
, size
);
257 cirrusbo
->bo
.bdev
= &cirrus
->ttm
.bdev
;
259 cirrus_ttm_placement(cirrusbo
, TTM_PL_FLAG_VRAM
| TTM_PL_FLAG_SYSTEM
);
261 acc_size
= ttm_bo_dma_acc_size(&cirrus
->ttm
.bdev
, size
,
262 sizeof(struct cirrus_bo
));
264 ret
= ttm_bo_init(&cirrus
->ttm
.bdev
, &cirrusbo
->bo
, size
,
265 ttm_bo_type_device
, &cirrusbo
->placement
,
266 align
>> PAGE_SHIFT
, false, acc_size
,
267 NULL
, NULL
, cirrus_bo_ttm_destroy
);
271 *pcirrusbo
= cirrusbo
;
275 static inline u64
cirrus_bo_gpu_offset(struct cirrus_bo
*bo
)
277 return bo
->bo
.offset
;
280 int cirrus_bo_pin(struct cirrus_bo
*bo
, u32 pl_flag
, u64
*gpu_addr
)
282 struct ttm_operation_ctx ctx
= { false, false };
288 *gpu_addr
= cirrus_bo_gpu_offset(bo
);
291 cirrus_ttm_placement(bo
, pl_flag
);
292 for (i
= 0; i
< bo
->placement
.num_placement
; i
++)
293 bo
->placements
[i
].flags
|= TTM_PL_FLAG_NO_EVICT
;
294 ret
= ttm_bo_validate(&bo
->bo
, &bo
->placement
, &ctx
);
300 *gpu_addr
= cirrus_bo_gpu_offset(bo
);
304 int cirrus_bo_push_sysram(struct cirrus_bo
*bo
)
306 struct ttm_operation_ctx ctx
= { false, false };
308 if (!bo
->pin_count
) {
309 DRM_ERROR("unpin bad %p\n", bo
);
316 if (bo
->kmap
.virtual)
317 ttm_bo_kunmap(&bo
->kmap
);
319 cirrus_ttm_placement(bo
, TTM_PL_FLAG_SYSTEM
);
320 for (i
= 0; i
< bo
->placement
.num_placement
; i
++)
321 bo
->placements
[i
].flags
|= TTM_PL_FLAG_NO_EVICT
;
323 ret
= ttm_bo_validate(&bo
->bo
, &bo
->placement
, &ctx
);
325 DRM_ERROR("pushing to VRAM failed\n");
331 int cirrus_mmap(struct file
*filp
, struct vm_area_struct
*vma
)
333 struct drm_file
*file_priv
= filp
->private_data
;
334 struct cirrus_device
*cirrus
= file_priv
->minor
->dev
->dev_private
;
336 return ttm_bo_mmap(filp
, vma
, &cirrus
->ttm
.bdev
);