|
|
@@ -195,7 +195,7 @@ bo_tile_size(struct brw_bufmgr *bufmgr, uint64_t size, uint32_t tiling) |
|
|
|
return size; |
|
|
|
|
|
|
|
/* 965+ just need multiples of page size for tiling */ |
|
|
|
return ALIGN(size, 4096); |
|
|
|
return ALIGN(size, PAGE_SIZE); |
|
|
|
} |
|
|
|
|
|
|
|
/* |
|
|
@@ -1577,12 +1577,12 @@ init_cache_buckets(struct brw_bufmgr *bufmgr) |
|
|
|
* width/height alignment and rounding of sizes to pages will |
|
|
|
* get us useful cache hit rates anyway) |
|
|
|
*/ |
|
|
|
add_bucket(bufmgr, 4096); |
|
|
|
add_bucket(bufmgr, 4096 * 2); |
|
|
|
add_bucket(bufmgr, 4096 * 3); |
|
|
|
add_bucket(bufmgr, PAGE_SIZE); |
|
|
|
add_bucket(bufmgr, PAGE_SIZE * 2); |
|
|
|
add_bucket(bufmgr, PAGE_SIZE * 3); |
|
|
|
|
|
|
|
/* Initialize the linked lists for BO reuse cache. */ |
|
|
|
for (size = 4 * 4096; size <= cache_max_size; size *= 2) { |
|
|
|
for (size = 4 * PAGE_SIZE; size <= cache_max_size; size *= 2) { |
|
|
|
add_bucket(bufmgr, size); |
|
|
|
|
|
|
|
add_bucket(bufmgr, size + size * 1 / 4); |
|
|
@@ -1728,7 +1728,7 @@ brw_bufmgr_init(struct gen_device_info *devinfo, int fd) |
|
|
|
bufmgr->initial_kflags |= EXEC_OBJECT_PINNED; |
|
|
|
|
|
|
|
util_vma_heap_init(&bufmgr->vma_allocator[BRW_MEMZONE_LOW_4G], |
|
|
|
4096, _4GB); |
|
|
|
PAGE_SIZE, _4GB); |
|
|
|
util_vma_heap_init(&bufmgr->vma_allocator[BRW_MEMZONE_OTHER], |
|
|
|
1 * _4GB, gtt_size - 1 * _4GB); |
|
|
|
} else if (devinfo->gen >= 10) { |