drm/i915/gtt: Rename i915_hw_ppgtt base member
authorChris Wilson <chris@chris-wilson.co.uk>
Tue, 5 Jun 2018 15:37:58 +0000 (16:37 +0100)
committerChris Wilson <chris@chris-wilson.co.uk>
Tue, 5 Jun 2018 20:11:20 +0000 (21:11 +0100)
In the near future, I want to subclass gen6_hw_ppgtt as it contains a
few specialised members and I wish to add more. To avoid the ugliness of
using ppgtt->base.base, rename the i915_hw_ppgtt base member
(i915_address_space) as vm, which is our common shorthand for an
i915_address_space local.

Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com>
Cc: Matthew Auld <matthew.william.auld@gmail.com>
Reviewed-by: Mika Kuoppala <mika.kuoppala@linux.intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20180605153758.18422-1-chris@chris-wilson.co.uk
32 files changed:
drivers/gpu/drm/i915/gvt/aperture_gm.c
drivers/gpu/drm/i915/gvt/gvt.h
drivers/gpu/drm/i915/i915_debugfs.c
drivers/gpu/drm/i915/i915_drv.h
drivers/gpu/drm/i915/i915_gem.c
drivers/gpu/drm/i915/i915_gem_context.c
drivers/gpu/drm/i915/i915_gem_execbuffer.c
drivers/gpu/drm/i915/i915_gem_gtt.c
drivers/gpu/drm/i915/i915_gem_gtt.h
drivers/gpu/drm/i915/i915_gem_render_state.c
drivers/gpu/drm/i915/i915_gem_shrinker.c
drivers/gpu/drm/i915/i915_gem_stolen.c
drivers/gpu/drm/i915/i915_gpu_error.c
drivers/gpu/drm/i915/i915_trace.h
drivers/gpu/drm/i915/i915_vgpu.c
drivers/gpu/drm/i915/i915_vma.c
drivers/gpu/drm/i915/intel_engine_cs.c
drivers/gpu/drm/i915/intel_guc.c
drivers/gpu/drm/i915/intel_guc_submission.c
drivers/gpu/drm/i915/intel_lrc.c
drivers/gpu/drm/i915/intel_ringbuffer.c
drivers/gpu/drm/i915/selftests/huge_pages.c
drivers/gpu/drm/i915/selftests/i915_gem_context.c
drivers/gpu/drm/i915/selftests/i915_gem_evict.c
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
drivers/gpu/drm/i915/selftests/i915_gem_object.c
drivers/gpu/drm/i915/selftests/i915_request.c
drivers/gpu/drm/i915/selftests/i915_vma.c
drivers/gpu/drm/i915/selftests/intel_hangcheck.c
drivers/gpu/drm/i915/selftests/intel_lrc.c
drivers/gpu/drm/i915/selftests/intel_workarounds.c
drivers/gpu/drm/i915/selftests/mock_gtt.c

index 7c9ec4f4f36c747464342d14c8e3b5dd4c778335..380eeb2a0e83c60c1067a6d7150bd8608662d5b1 100644 (file)
@@ -61,7 +61,7 @@ static int alloc_gm(struct intel_vgpu *vgpu, bool high_gm)
        }
 
        mutex_lock(&dev_priv->drm.struct_mutex);
-       ret = i915_gem_gtt_insert(&dev_priv->ggtt.base, node,
+       ret = i915_gem_gtt_insert(&dev_priv->ggtt.vm, node,
                                  size, I915_GTT_PAGE_SIZE,
                                  I915_COLOR_UNEVICTABLE,
                                  start, end, flags);
index 05d15a095310d41b75d6ab64aa04162799f393a1..2ff0d40281a9a8c28b8b5237da010fdfc0720ca0 100644 (file)
@@ -361,9 +361,9 @@ int intel_gvt_load_firmware(struct intel_gvt *gvt);
 #define gvt_aperture_sz(gvt)     (gvt->dev_priv->ggtt.mappable_end)
 #define gvt_aperture_pa_base(gvt) (gvt->dev_priv->ggtt.gmadr.start)
 
-#define gvt_ggtt_gm_sz(gvt)      (gvt->dev_priv->ggtt.base.total)
+#define gvt_ggtt_gm_sz(gvt)      (gvt->dev_priv->ggtt.vm.total)
 #define gvt_ggtt_sz(gvt) \
-       ((gvt->dev_priv->ggtt.base.total >> PAGE_SHIFT) << 3)
+       ((gvt->dev_priv->ggtt.vm.total >> PAGE_SHIFT) << 3)
 #define gvt_hidden_sz(gvt)       (gvt_ggtt_gm_sz(gvt) - gvt_aperture_sz(gvt))
 
 #define gvt_aperture_gmadr_base(gvt) (0)
index 15e86d34a81ce2cc68e54cffff5235a03074f8e7..698af45e229ce2e4517b3f062ff7803fb7c9f62e 100644 (file)
@@ -328,7 +328,7 @@ static int per_file_stats(int id, void *ptr, void *data)
                } else {
                        struct i915_hw_ppgtt *ppgtt = i915_vm_to_ppgtt(vma->vm);
 
-                       if (ppgtt->base.file != stats->file_priv)
+                       if (ppgtt->vm.file != stats->file_priv)
                                continue;
                }
 
@@ -508,7 +508,7 @@ static int i915_gem_object_info(struct seq_file *m, void *data)
                   dpy_count, dpy_size);
 
        seq_printf(m, "%llu [%pa] gtt total\n",
-                  ggtt->base.total, &ggtt->mappable_end);
+                  ggtt->vm.total, &ggtt->mappable_end);
        seq_printf(m, "Supported page sizes: %s\n",
                   stringify_page_sizes(INTEL_INFO(dev_priv)->page_sizes,
                                        buf, sizeof(buf)));
index 06ecac4c32534e0a573e68095fa1cc9289836ecf..a4bb30c32a527e8e706a1413e5f36e560e88281f 100644 (file)
@@ -3213,7 +3213,7 @@ struct dma_buf *i915_gem_prime_export(struct drm_device *dev,
 static inline struct i915_hw_ppgtt *
 i915_vm_to_ppgtt(struct i915_address_space *vm)
 {
-       return container_of(vm, struct i915_hw_ppgtt, base);
+       return container_of(vm, struct i915_hw_ppgtt, vm);
 }
 
 /* i915_gem_fence_reg.c */
index 23374f3dd79d386e40efe21dd7ba70f5026ecdad..86f1f9aaa119886b8a0fcec72763f4cb8df11c56 100644 (file)
@@ -65,7 +65,7 @@ insert_mappable_node(struct i915_ggtt *ggtt,
                      struct drm_mm_node *node, u32 size)
 {
        memset(node, 0, sizeof(*node));
-       return drm_mm_insert_node_in_range(&ggtt->base.mm, node,
+       return drm_mm_insert_node_in_range(&ggtt->vm.mm, node,
                                           size, 0, I915_COLOR_UNEVICTABLE,
                                           0, ggtt->mappable_end,
                                           DRM_MM_INSERT_LOW);
@@ -249,17 +249,17 @@ i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data,
        struct i915_vma *vma;
        u64 pinned;
 
-       pinned = ggtt->base.reserved;
+       pinned = ggtt->vm.reserved;
        mutex_lock(&dev->struct_mutex);
-       list_for_each_entry(vma, &ggtt->base.active_list, vm_link)
+       list_for_each_entry(vma, &ggtt->vm.active_list, vm_link)
                if (i915_vma_is_pinned(vma))
                        pinned += vma->node.size;
-       list_for_each_entry(vma, &ggtt->base.inactive_list, vm_link)
+       list_for_each_entry(vma, &ggtt->vm.inactive_list, vm_link)
                if (i915_vma_is_pinned(vma))
                        pinned += vma->node.size;
        mutex_unlock(&dev->struct_mutex);
 
-       args->aper_size = ggtt->base.total;
+       args->aper_size = ggtt->vm.total;
        args->aper_available_size = args->aper_size - pinned;
 
        return 0;
@@ -1223,9 +1223,9 @@ i915_gem_gtt_pread(struct drm_i915_gem_object *obj,
                page_length = remain < page_length ? remain : page_length;
                if (node.allocated) {
                        wmb();
-                       ggtt->base.insert_page(&ggtt->base,
-                                              i915_gem_object_get_dma_address(obj, offset >> PAGE_SHIFT),
-                                              node.start, I915_CACHE_NONE, 0);
+                       ggtt->vm.insert_page(&ggtt->vm,
+                                            i915_gem_object_get_dma_address(obj, offset >> PAGE_SHIFT),
+                                            node.start, I915_CACHE_NONE, 0);
                        wmb();
                } else {
                        page_base += offset & PAGE_MASK;
@@ -1246,8 +1246,7 @@ i915_gem_gtt_pread(struct drm_i915_gem_object *obj,
 out_unpin:
        if (node.allocated) {
                wmb();
-               ggtt->base.clear_range(&ggtt->base,
-                                      node.start, node.size);
+               ggtt->vm.clear_range(&ggtt->vm, node.start, node.size);
                remove_mappable_node(&node);
        } else {
                i915_vma_unpin(vma);
@@ -1426,9 +1425,9 @@ i915_gem_gtt_pwrite_fast(struct drm_i915_gem_object *obj,
                page_length = remain < page_length ? remain : page_length;
                if (node.allocated) {
                        wmb(); /* flush the write before we modify the GGTT */
-                       ggtt->base.insert_page(&ggtt->base,
-                                              i915_gem_object_get_dma_address(obj, offset >> PAGE_SHIFT),
-                                              node.start, I915_CACHE_NONE, 0);
+                       ggtt->vm.insert_page(&ggtt->vm,
+                                            i915_gem_object_get_dma_address(obj, offset >> PAGE_SHIFT),
+                                            node.start, I915_CACHE_NONE, 0);
                        wmb(); /* flush modifications to the GGTT (insert_page) */
                } else {
                        page_base += offset & PAGE_MASK;
@@ -1455,8 +1454,7 @@ i915_gem_gtt_pwrite_fast(struct drm_i915_gem_object *obj,
 out_unpin:
        if (node.allocated) {
                wmb();
-               ggtt->base.clear_range(&ggtt->base,
-                                      node.start, node.size);
+               ggtt->vm.clear_range(&ggtt->vm, node.start, node.size);
                remove_mappable_node(&node);
        } else {
                i915_vma_unpin(vma);
@@ -4374,7 +4372,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
                         u64 flags)
 {
        struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
-       struct i915_address_space *vm = &dev_priv->ggtt.base;
+       struct i915_address_space *vm = &dev_priv->ggtt.vm;
        struct i915_vma *vma;
        int ret;
 
index 38c6e9e4e91bfe4482a3e6f31c0234491e6ac031..b2c7ac1b074d42f54fbfae9bc64c06787b5b0226 100644 (file)
@@ -197,7 +197,7 @@ static void context_close(struct i915_gem_context *ctx)
         */
        lut_close(ctx);
        if (ctx->ppgtt)
-               i915_ppgtt_close(&ctx->ppgtt->base);
+               i915_ppgtt_close(&ctx->ppgtt->vm);
 
        ctx->file_priv = ERR_PTR(-EBADF);
        i915_gem_context_put(ctx);
@@ -249,7 +249,7 @@ static u32 default_desc_template(const struct drm_i915_private *i915,
        desc = GEN8_CTX_VALID | GEN8_CTX_PRIVILEGE;
 
        address_mode = INTEL_LEGACY_32B_CONTEXT;
-       if (ppgtt && i915_vm_is_48bit(&ppgtt->base))
+       if (ppgtt && i915_vm_is_48bit(&ppgtt->vm))
                address_mode = INTEL_LEGACY_64B_CONTEXT;
        desc |= address_mode << GEN8_CTX_ADDRESSING_MODE_SHIFT;
 
@@ -810,11 +810,11 @@ int i915_gem_context_getparam_ioctl(struct drm_device *dev, void *data,
                break;
        case I915_CONTEXT_PARAM_GTT_SIZE:
                if (ctx->ppgtt)
-                       args->value = ctx->ppgtt->base.total;
+                       args->value = ctx->ppgtt->vm.total;
                else if (to_i915(dev)->mm.aliasing_ppgtt)
-                       args->value = to_i915(dev)->mm.aliasing_ppgtt->base.total;
+                       args->value = to_i915(dev)->mm.aliasing_ppgtt->vm.total;
                else
-                       args->value = to_i915(dev)->ggtt.base.total;
+                       args->value = to_i915(dev)->ggtt.vm.total;
                break;
        case I915_CONTEXT_PARAM_NO_ERROR_CAPTURE:
                args->value = i915_gem_context_no_error_capture(ctx);
index f627a8c47c58a36f6ff92f17a4d6d672b28bc00b..eefd449502e27399b83a088dd667b08fd9c38471 100644 (file)
@@ -703,7 +703,7 @@ static int eb_select_context(struct i915_execbuffer *eb)
                return -ENOENT;
 
        eb->ctx = ctx;
-       eb->vm = ctx->ppgtt ? &ctx->ppgtt->base : &eb->i915->ggtt.base;
+       eb->vm = ctx->ppgtt ? &ctx->ppgtt->vm : &eb->i915->ggtt.vm;
 
        eb->context_flags = 0;
        if (ctx->flags & CONTEXT_NO_ZEROMAP)
@@ -943,9 +943,9 @@ static void reloc_cache_reset(struct reloc_cache *cache)
                if (cache->node.allocated) {
                        struct i915_ggtt *ggtt = cache_to_ggtt(cache);
 
-                       ggtt->base.clear_range(&ggtt->base,
-                                              cache->node.start,
-                                              cache->node.size);
+                       ggtt->vm.clear_range(&ggtt->vm,
+                                            cache->node.start,
+                                            cache->node.size);
                        drm_mm_remove_node(&cache->node);
                } else {
                        i915_vma_unpin((struct i915_vma *)cache->node.mm);
@@ -1016,7 +1016,7 @@ static void *reloc_iomap(struct drm_i915_gem_object *obj,
                if (IS_ERR(vma)) {
                        memset(&cache->node, 0, sizeof(cache->node));
                        err = drm_mm_insert_node_in_range
-                               (&ggtt->base.mm, &cache->node,
+                               (&ggtt->vm.mm, &cache->node,
                                 PAGE_SIZE, 0, I915_COLOR_UNEVICTABLE,
                                 0, ggtt->mappable_end,
                                 DRM_MM_INSERT_LOW);
@@ -1037,9 +1037,9 @@ static void *reloc_iomap(struct drm_i915_gem_object *obj,
        offset = cache->node.start;
        if (cache->node.allocated) {
                wmb();
-               ggtt->base.insert_page(&ggtt->base,
-                                      i915_gem_object_get_dma_address(obj, page),
-                                      offset, I915_CACHE_NONE, 0);
+               ggtt->vm.insert_page(&ggtt->vm,
+                                    i915_gem_object_get_dma_address(obj, page),
+                                    offset, I915_CACHE_NONE, 0);
        } else {
                offset += page << PAGE_SHIFT;
        }
index 9302ee6e717d662e8cfcbd3b2f47736a36b9fa1b..029a5f4fbd92b5dc157bf8dec5efed8f91d9654b 100644 (file)
@@ -780,7 +780,7 @@ static void gen8_initialize_pml4(struct i915_address_space *vm,
  */
 static void mark_tlbs_dirty(struct i915_hw_ppgtt *ppgtt)
 {
-       ppgtt->pd_dirty_rings = INTEL_INFO(ppgtt->base.i915)->ring_mask;
+       ppgtt->pd_dirty_rings = INTEL_INFO(ppgtt->vm.i915)->ring_mask;
 }
 
 /* Removes entries from a single page table, releasing it if it's empty.
@@ -973,7 +973,7 @@ gen8_ppgtt_insert_pte_entries(struct i915_hw_ppgtt *ppgtt,
        gen8_pte_t *vaddr;
        bool ret;
 
-       GEM_BUG_ON(idx->pdpe >= i915_pdpes_per_pdp(&ppgtt->base));
+       GEM_BUG_ON(idx->pdpe >= i915_pdpes_per_pdp(&ppgtt->vm));
        pd = pdp->page_directory[idx->pdpe];
        vaddr = kmap_atomic_px(pd->page_table[idx->pde]);
        do {
@@ -1004,7 +1004,7 @@ gen8_ppgtt_insert_pte_entries(struct i915_hw_ppgtt *ppgtt,
                                        break;
                                }
 
-                               GEM_BUG_ON(idx->pdpe >= i915_pdpes_per_pdp(&ppgtt->base));
+                               GEM_BUG_ON(idx->pdpe >= i915_pdpes_per_pdp(&ppgtt->vm));
                                pd = pdp->page_directory[idx->pdpe];
                        }
 
@@ -1233,7 +1233,7 @@ free_scratch_page:
 
 static int gen8_ppgtt_notify_vgt(struct i915_hw_ppgtt *ppgtt, bool create)
 {
-       struct i915_address_space *vm = &ppgtt->base;
+       struct i915_address_space *vm = &ppgtt->vm;
        struct drm_i915_private *dev_priv = vm->i915;
        enum vgt_g2v_type msg;
        int i;
@@ -1294,13 +1294,13 @@ static void gen8_ppgtt_cleanup_4lvl(struct i915_hw_ppgtt *ppgtt)
        int i;
 
        for (i = 0; i < GEN8_PML4ES_PER_PML4; i++) {
-               if (ppgtt->pml4.pdps[i] == ppgtt->base.scratch_pdp)
+               if (ppgtt->pml4.pdps[i] == ppgtt->vm.scratch_pdp)
                        continue;
 
-               gen8_ppgtt_cleanup_3lvl(&ppgtt->base, ppgtt->pml4.pdps[i]);
+               gen8_ppgtt_cleanup_3lvl(&ppgtt->vm, ppgtt->pml4.pdps[i]);
        }
 
-       cleanup_px(&ppgtt->base, &ppgtt->pml4);
+       cleanup_px(&ppgtt->vm, &ppgtt->pml4);
 }
 
 static void gen8_ppgtt_cleanup(struct i915_address_space *vm)
@@ -1314,7 +1314,7 @@ static void gen8_ppgtt_cleanup(struct i915_address_space *vm)
        if (use_4lvl(vm))
                gen8_ppgtt_cleanup_4lvl(ppgtt);
        else
-               gen8_ppgtt_cleanup_3lvl(&ppgtt->base, &ppgtt->pdp);
+               gen8_ppgtt_cleanup_3lvl(&ppgtt->vm, &ppgtt->pdp);
 
        gen8_free_scratch(vm);
 }
@@ -1450,7 +1450,7 @@ static void gen8_dump_pdp(struct i915_hw_ppgtt *ppgtt,
                          gen8_pte_t scratch_pte,
                          struct seq_file *m)
 {
-       struct i915_address_space *vm = &ppgtt->base;
+       struct i915_address_space *vm = &ppgtt->vm;
        struct i915_page_directory *pd;
        u32 pdpe;
 
@@ -1460,7 +1460,7 @@ static void gen8_dump_pdp(struct i915_hw_ppgtt *ppgtt,
                u64 pd_start = start;
                u32 pde;
 
-               if (pdp->page_directory[pdpe] == ppgtt->base.scratch_pd)
+               if (pdp->page_directory[pdpe] == ppgtt->vm.scratch_pd)
                        continue;
 
                seq_printf(m, "\tPDPE #%d\n", pdpe);
@@ -1468,7 +1468,7 @@ static void gen8_dump_pdp(struct i915_hw_ppgtt *ppgtt,
                        u32 pte;
                        gen8_pte_t *pt_vaddr;
 
-                       if (pd->page_table[pde] == ppgtt->base.scratch_pt)
+                       if (pd->page_table[pde] == ppgtt->vm.scratch_pt)
                                continue;
 
                        pt_vaddr = kmap_atomic_px(pt);
@@ -1501,10 +1501,10 @@ static void gen8_dump_pdp(struct i915_hw_ppgtt *ppgtt,
 
 static void gen8_dump_ppgtt(struct i915_hw_ppgtt *ppgtt, struct seq_file *m)
 {
-       struct i915_address_space *vm = &ppgtt->base;
+       struct i915_address_space *vm = &ppgtt->vm;
        const gen8_pte_t scratch_pte =
                gen8_pte_encode(vm->scratch_page.daddr, I915_CACHE_LLC);
-       u64 start = 0, length = ppgtt->base.total;
+       u64 start = 0, length = ppgtt->vm.total;
 
        if (use_4lvl(vm)) {
                u64 pml4e;
@@ -1512,7 +1512,7 @@ static void gen8_dump_ppgtt(struct i915_hw_ppgtt *ppgtt, struct seq_file *m)
                struct i915_page_directory_pointer *pdp;
 
                gen8_for_each_pml4e(pdp, pml4, start, length, pml4e) {
-                       if (pml4->pdps[pml4e] == ppgtt->base.scratch_pdp)
+                       if (pml4->pdps[pml4e] == ppgtt->vm.scratch_pdp)
                                continue;
 
                        seq_printf(m, "    PML4E #%llu\n", pml4e);
@@ -1525,10 +1525,10 @@ static void gen8_dump_ppgtt(struct i915_hw_ppgtt *ppgtt, struct seq_file *m)
 
 static int gen8_preallocate_top_level_pdp(struct i915_hw_ppgtt *ppgtt)
 {
-       struct i915_address_space *vm = &ppgtt->base;
+       struct i915_address_space *vm = &ppgtt->vm;
        struct i915_page_directory_pointer *pdp = &ppgtt->pdp;
        struct i915_page_directory *pd;
-       u64 start = 0, length = ppgtt->base.total;
+       u64 start = 0, length = ppgtt->vm.total;
        u64 from = start;
        unsigned int pdpe;
 
@@ -1564,11 +1564,11 @@ unwind:
  */
 static int gen8_ppgtt_init(struct i915_hw_ppgtt *ppgtt)
 {
-       struct i915_address_space *vm = &ppgtt->base;
+       struct i915_address_space *vm = &ppgtt->vm;
        struct drm_i915_private *dev_priv = vm->i915;
        int ret;
 
-       ppgtt->base.total = USES_FULL_48BIT_PPGTT(dev_priv) ?
+       ppgtt->vm.total = USES_FULL_48BIT_PPGTT(dev_priv) ?
                1ULL << 48 :
                1ULL << 32;
 
@@ -1576,26 +1576,26 @@ static int gen8_ppgtt_init(struct i915_hw_ppgtt *ppgtt)
         * And we are not sure about the latter so play safe for now.
         */
        if (IS_CHERRYVIEW(dev_priv) || IS_BROXTON(dev_priv))
-               ppgtt->base.pt_kmap_wc = true;
+               ppgtt->vm.pt_kmap_wc = true;
 
-       ret = gen8_init_scratch(&ppgtt->base);
+       ret = gen8_init_scratch(&ppgtt->vm);
        if (ret) {
-               ppgtt->base.total = 0;
+               ppgtt->vm.total = 0;
                return ret;
        }
 
        if (use_4lvl(vm)) {
-               ret = setup_px(&ppgtt->base, &ppgtt->pml4);
+               ret = setup_px(&ppgtt->vm, &ppgtt->pml4);
                if (ret)
                        goto free_scratch;
 
-               gen8_initialize_pml4(&ppgtt->base, &ppgtt->pml4);
+               gen8_initialize_pml4(&ppgtt->vm, &ppgtt->pml4);
 
-               ppgtt->base.allocate_va_range = gen8_ppgtt_alloc_4lvl;
-               ppgtt->base.insert_entries = gen8_ppgtt_insert_4lvl;
-               ppgtt->base.clear_range = gen8_ppgtt_clear_4lvl;
+               ppgtt->vm.allocate_va_range = gen8_ppgtt_alloc_4lvl;
+               ppgtt->vm.insert_entries = gen8_ppgtt_insert_4lvl;
+               ppgtt->vm.clear_range = gen8_ppgtt_clear_4lvl;
        } else {
-               ret = __pdp_init(&ppgtt->base, &ppgtt->pdp);
+               ret = __pdp_init(&ppgtt->vm, &ppgtt->pdp);
                if (ret)
                        goto free_scratch;
 
@@ -1607,35 +1607,35 @@ static int gen8_ppgtt_init(struct i915_hw_ppgtt *ppgtt)
                        }
                }
 
-               ppgtt->base.allocate_va_range = gen8_ppgtt_alloc_3lvl;
-               ppgtt->base.insert_entries = gen8_ppgtt_insert_3lvl;
-               ppgtt->base.clear_range = gen8_ppgtt_clear_3lvl;
+               ppgtt->vm.allocate_va_range = gen8_ppgtt_alloc_3lvl;
+               ppgtt->vm.insert_entries = gen8_ppgtt_insert_3lvl;
+               ppgtt->vm.clear_range = gen8_ppgtt_clear_3lvl;
        }
 
        if (intel_vgpu_active(dev_priv))
                gen8_ppgtt_notify_vgt(ppgtt, true);
 
-       ppgtt->base.cleanup = gen8_ppgtt_cleanup;
-       ppgtt->base.bind_vma = gen8_ppgtt_bind_vma;
-       ppgtt->base.unbind_vma = ppgtt_unbind_vma;
-       ppgtt->base.set_pages = ppgtt_set_pages;
-       ppgtt->base.clear_pages = clear_pages;
+       ppgtt->vm.cleanup = gen8_ppgtt_cleanup;
+       ppgtt->vm.bind_vma = gen8_ppgtt_bind_vma;
+       ppgtt->vm.unbind_vma = ppgtt_unbind_vma;
+       ppgtt->vm.set_pages = ppgtt_set_pages;
+       ppgtt->vm.clear_pages = clear_pages;
        ppgtt->debug_dump = gen8_dump_ppgtt;
 
        return 0;
 
 free_scratch:
-       gen8_free_scratch(&ppgtt->base);
+       gen8_free_scratch(&ppgtt->vm);
        return ret;
 }
 
 static void gen6_dump_ppgtt(struct i915_hw_ppgtt *ppgtt, struct seq_file *m)
 {
-       struct i915_address_space *vm = &ppgtt->base;
+       struct i915_address_space *vm = &ppgtt->vm;
        struct i915_page_table *unused;
        gen6_pte_t scratch_pte;
        u32 pd_entry, pte, pde;
-       u32 start = 0, length = ppgtt->base.total;
+       u32 start = 0, length = ppgtt->vm.total;
 
        scratch_pte = vm->pte_encode(vm->scratch_page.daddr,
                                     I915_CACHE_LLC, 0);
@@ -1972,8 +1972,8 @@ static void gen6_ppgtt_cleanup(struct i915_address_space *vm)
 
 static int gen6_ppgtt_allocate_page_directories(struct i915_hw_ppgtt *ppgtt)
 {
-       struct i915_address_space *vm = &ppgtt->base;
-       struct drm_i915_private *dev_priv = ppgtt->base.i915;
+       struct i915_address_space *vm = &ppgtt->vm;
+       struct drm_i915_private *dev_priv = ppgtt->vm.i915;
        struct i915_ggtt *ggtt = &dev_priv->ggtt;
        int ret;
 
@@ -1981,16 +1981,16 @@ static int gen6_ppgtt_allocate_page_directories(struct i915_hw_ppgtt *ppgtt)
         * allocator works in address space sizes, so it's multiplied by page
         * size. We allocate at the top of the GTT to avoid fragmentation.
         */
-       BUG_ON(!drm_mm_initialized(&ggtt->base.mm));
+       BUG_ON(!drm_mm_initialized(&ggtt->vm.mm));
 
        ret = gen6_init_scratch(vm);
        if (ret)
                return ret;
 
-       ret = i915_gem_gtt_insert(&ggtt->base, &ppgtt->node,
+       ret = i915_gem_gtt_insert(&ggtt->vm, &ppgtt->node,
                                  GEN6_PD_SIZE, GEN6_PD_ALIGN,
                                  I915_COLOR_UNEVICTABLE,
-                                 0, ggtt->base.total,
+                                 0, ggtt->vm.total,
                                  PIN_HIGH);
        if (ret)
                goto err_out;
@@ -2023,16 +2023,16 @@ static void gen6_scratch_va_range(struct i915_hw_ppgtt *ppgtt,
        u32 pde;
 
        gen6_for_each_pde(unused, &ppgtt->pd, start, length, pde)
-               ppgtt->pd.page_table[pde] = ppgtt->base.scratch_pt;
+               ppgtt->pd.page_table[pde] = ppgtt->vm.scratch_pt;
 }
 
 static int gen6_ppgtt_init(struct i915_hw_ppgtt *ppgtt)
 {
-       struct drm_i915_private *dev_priv = ppgtt->base.i915;
+       struct drm_i915_private *dev_priv = ppgtt->vm.i915;
        struct i915_ggtt *ggtt = &dev_priv->ggtt;
        int ret;
 
-       ppgtt->base.pte_encode = ggtt->base.pte_encode;
+       ppgtt->vm.pte_encode = ggtt->vm.pte_encode;
        if (intel_vgpu_active(dev_priv) || IS_GEN6(dev_priv))
                ppgtt->switch_mm = gen6_mm_switch;
        else if (IS_HASWELL(dev_priv))
@@ -2046,24 +2046,24 @@ static int gen6_ppgtt_init(struct i915_hw_ppgtt *ppgtt)
        if (ret)
                return ret;
 
-       ppgtt->base.total = I915_PDES * GEN6_PTES * PAGE_SIZE;
+       ppgtt->vm.total = I915_PDES * GEN6_PTES * PAGE_SIZE;
 
-       gen6_scratch_va_range(ppgtt, 0, ppgtt->base.total);
-       gen6_write_page_range(ppgtt, 0, ppgtt->base.total);
+       gen6_scratch_va_range(ppgtt, 0, ppgtt->vm.total);
+       gen6_write_page_range(ppgtt, 0, ppgtt->vm.total);
 
-       ret = gen6_alloc_va_range(&ppgtt->base, 0, ppgtt->base.total);
+       ret = gen6_alloc_va_range(&ppgtt->vm, 0, ppgtt->vm.total);
        if (ret) {
-               gen6_ppgtt_cleanup(&ppgtt->base);
+               gen6_ppgtt_cleanup(&ppgtt->vm);
                return ret;
        }
 
-       ppgtt->base.clear_range = gen6_ppgtt_clear_range;
-       ppgtt->base.insert_entries = gen6_ppgtt_insert_entries;
-       ppgtt->base.bind_vma = gen6_ppgtt_bind_vma;
-       ppgtt->base.unbind_vma = ppgtt_unbind_vma;
-       ppgtt->base.set_pages = ppgtt_set_pages;
-       ppgtt->base.clear_pages = clear_pages;
-       ppgtt->base.cleanup = gen6_ppgtt_cleanup;
+       ppgtt->vm.clear_range = gen6_ppgtt_clear_range;
+       ppgtt->vm.insert_entries = gen6_ppgtt_insert_entries;
+       ppgtt->vm.bind_vma = gen6_ppgtt_bind_vma;
+       ppgtt->vm.unbind_vma = ppgtt_unbind_vma;
+       ppgtt->vm.set_pages = ppgtt_set_pages;
+       ppgtt->vm.clear_pages = clear_pages;
+       ppgtt->vm.cleanup = gen6_ppgtt_cleanup;
        ppgtt->debug_dump = gen6_dump_ppgtt;
 
        DRM_DEBUG_DRIVER("Allocated pde space (%lldM) at GTT entry: %llx\n",
@@ -2079,8 +2079,8 @@ static int gen6_ppgtt_init(struct i915_hw_ppgtt *ppgtt)
 static int __hw_ppgtt_init(struct i915_hw_ppgtt *ppgtt,
                           struct drm_i915_private *dev_priv)
 {
-       ppgtt->base.i915 = dev_priv;
-       ppgtt->base.dma = &dev_priv->drm.pdev->dev;
+       ppgtt->vm.i915 = dev_priv;
+       ppgtt->vm.dma = &dev_priv->drm.pdev->dev;
 
        if (INTEL_GEN(dev_priv) < 8)
                return gen6_ppgtt_init(ppgtt);
@@ -2190,10 +2190,10 @@ i915_ppgtt_create(struct drm_i915_private *dev_priv,
        }
 
        kref_init(&ppgtt->ref);
-       i915_address_space_init(&ppgtt->base, dev_priv, name);
-       ppgtt->base.file = fpriv;
+       i915_address_space_init(&ppgtt->vm, dev_priv, name);
+       ppgtt->vm.file = fpriv;
 
-       trace_i915_ppgtt_create(&ppgtt->base);
+       trace_i915_ppgtt_create(&ppgtt->vm);
 
        return ppgtt;
 }
@@ -2227,16 +2227,16 @@ void i915_ppgtt_release(struct kref *kref)
        struct i915_hw_ppgtt *ppgtt =
                container_of(kref, struct i915_hw_ppgtt, ref);
 
-       trace_i915_ppgtt_release(&ppgtt->base);
+       trace_i915_ppgtt_release(&ppgtt->vm);
 
-       ppgtt_destroy_vma(&ppgtt->base);
+       ppgtt_destroy_vma(&ppgtt->vm);
 
-       GEM_BUG_ON(!list_empty(&ppgtt->base.active_list));
-       GEM_BUG_ON(!list_empty(&ppgtt->base.inactive_list));
-       GEM_BUG_ON(!list_empty(&ppgtt->base.unbound_list));
+       GEM_BUG_ON(!list_empty(&ppgtt->vm.active_list));
+       GEM_BUG_ON(!list_empty(&ppgtt->vm.inactive_list));
+       GEM_BUG_ON(!list_empty(&ppgtt->vm.unbound_list));
 
-       ppgtt->base.cleanup(&ppgtt->base);
-       i915_address_space_fini(&ppgtt->base);
+       ppgtt->vm.cleanup(&ppgtt->vm);
+       i915_address_space_fini(&ppgtt->vm);
        kfree(ppgtt);
 }
 
@@ -2332,7 +2332,7 @@ void i915_gem_suspend_gtt_mappings(struct drm_i915_private *dev_priv)
 
        i915_check_and_clear_faults(dev_priv);
 
-       ggtt->base.clear_range(&ggtt->base, 0, ggtt->base.total);
+       ggtt->vm.clear_range(&ggtt->vm, 0, ggtt->vm.total);
 
        i915_ggtt_invalidate(dev_priv);
 }
@@ -2675,16 +2675,16 @@ static int aliasing_gtt_bind_vma(struct i915_vma *vma,
                struct i915_hw_ppgtt *appgtt = i915->mm.aliasing_ppgtt;
 
                if (!(vma->flags & I915_VMA_LOCAL_BIND) &&
-                   appgtt->base.allocate_va_range) {
-                       ret = appgtt->base.allocate_va_range(&appgtt->base,
-                                                            vma->node.start,
-                                                            vma->size);
+                   appgtt->vm.allocate_va_range) {
+                       ret = appgtt->vm.allocate_va_range(&appgtt->vm,
+                                                          vma->node.start,
+                                                          vma->size);
                        if (ret)
                                return ret;
                }
 
-               appgtt->base.insert_entries(&appgtt->base, vma, cache_level,
-                                           pte_flags);
+               appgtt->vm.insert_entries(&appgtt->vm, vma, cache_level,
+                                         pte_flags);
        }
 
        if (flags & I915_VMA_GLOBAL_BIND) {
@@ -2707,7 +2707,7 @@ static void aliasing_gtt_unbind_vma(struct i915_vma *vma)
        }
 
        if (vma->flags & I915_VMA_LOCAL_BIND) {
-               struct i915_address_space *vm = &i915->mm.aliasing_ppgtt->base;
+               struct i915_address_space *vm = &i915->mm.aliasing_ppgtt->vm;
 
                vm->clear_range(vm, vma->node.start, vma->size);
        }
@@ -2774,30 +2774,30 @@ int i915_gem_init_aliasing_ppgtt(struct drm_i915_private *i915)
        if (IS_ERR(ppgtt))
                return PTR_ERR(ppgtt);
 
-       if (WARN_ON(ppgtt->base.total < ggtt->base.total)) {
+       if (WARN_ON(ppgtt->vm.total < ggtt->vm.total)) {
                err = -ENODEV;
                goto err_ppgtt;
        }
 
-       if (ppgtt->base.allocate_va_range) {
+       if (ppgtt->vm.allocate_va_range) {
                /* Note we only pre-allocate as far as the end of the global
                 * GTT. On 48b / 4-level page-tables, the difference is very,
                 * very significant! We have to preallocate as GVT/vgpu does
                 * not like the page directory disappearing.
                 */
-               err = ppgtt->base.allocate_va_range(&ppgtt->base,
-                                                   0, ggtt->base.total);
+               err = ppgtt->vm.allocate_va_range(&ppgtt->vm,
+                                                 0, ggtt->vm.total);
                if (err)
                        goto err_ppgtt;
        }
 
        i915->mm.aliasing_ppgtt = ppgtt;
 
-       GEM_BUG_ON(ggtt->base.bind_vma != ggtt_bind_vma);
-       ggtt->base.bind_vma = aliasing_gtt_bind_vma;
+       GEM_BUG_ON(ggtt->vm.bind_vma != ggtt_bind_vma);
+       ggtt->vm.bind_vma = aliasing_gtt_bind_vma;
 
-       GEM_BUG_ON(ggtt->base.unbind_vma != ggtt_unbind_vma);
-       ggtt->base.unbind_vma = aliasing_gtt_unbind_vma;
+       GEM_BUG_ON(ggtt->vm.unbind_vma != ggtt_unbind_vma);
+       ggtt->vm.unbind_vma = aliasing_gtt_unbind_vma;
 
        return 0;
 
@@ -2817,8 +2817,8 @@ void i915_gem_fini_aliasing_ppgtt(struct drm_i915_private *i915)
 
        i915_ppgtt_put(ppgtt);
 
-       ggtt->base.bind_vma = ggtt_bind_vma;
-       ggtt->base.unbind_vma = ggtt_unbind_vma;
+       ggtt->vm.bind_vma = ggtt_bind_vma;
+       ggtt->vm.unbind_vma = ggtt_unbind_vma;
 }
 
 int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
@@ -2842,7 +2842,7 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
                return ret;
 
        /* Reserve a mappable slot for our lockless error capture */
-       ret = drm_mm_insert_node_in_range(&ggtt->base.mm, &ggtt->error_capture,
+       ret = drm_mm_insert_node_in_range(&ggtt->vm.mm, &ggtt->error_capture,
                                          PAGE_SIZE, 0, I915_COLOR_UNEVICTABLE,
                                          0, ggtt->mappable_end,
                                          DRM_MM_INSERT_LOW);
@@ -2850,16 +2850,15 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
                return ret;
 
        /* Clear any non-preallocated blocks */
-       drm_mm_for_each_hole(entry, &ggtt->base.mm, hole_start, hole_end) {
+       drm_mm_for_each_hole(entry, &ggtt->vm.mm, hole_start, hole_end) {
                DRM_DEBUG_KMS("clearing unused GTT space: [%lx, %lx]\n",
                              hole_start, hole_end);
-               ggtt->base.clear_range(&ggtt->base, hole_start,
-                                      hole_end - hole_start);
+               ggtt->vm.clear_range(&ggtt->vm, hole_start,
+                                    hole_end - hole_start);
        }
 
        /* And finally clear the reserved guard page */
-       ggtt->base.clear_range(&ggtt->base,
-                              ggtt->base.total - PAGE_SIZE, PAGE_SIZE);
+       ggtt->vm.clear_range(&ggtt->vm, ggtt->vm.total - PAGE_SIZE, PAGE_SIZE);
 
        if (USES_PPGTT(dev_priv) && !USES_FULL_PPGTT(dev_priv)) {
                ret = i915_gem_init_aliasing_ppgtt(dev_priv);
@@ -2884,11 +2883,11 @@ void i915_ggtt_cleanup_hw(struct drm_i915_private *dev_priv)
        struct i915_vma *vma, *vn;
        struct pagevec *pvec;
 
-       ggtt->base.closed = true;
+       ggtt->vm.closed = true;
 
        mutex_lock(&dev_priv->drm.struct_mutex);
-       GEM_BUG_ON(!list_empty(&ggtt->base.active_list));
-       list_for_each_entry_safe(vma, vn, &ggtt->base.inactive_list, vm_link)
+       GEM_BUG_ON(!list_empty(&ggtt->vm.active_list));
+       list_for_each_entry_safe(vma, vn, &ggtt->vm.inactive_list, vm_link)
                WARN_ON(i915_vma_unbind(vma));
        mutex_unlock(&dev_priv->drm.struct_mutex);
 
@@ -2900,12 +2899,12 @@ void i915_ggtt_cleanup_hw(struct drm_i915_private *dev_priv)
        if (drm_mm_node_allocated(&ggtt->error_capture))
                drm_mm_remove_node(&ggtt->error_capture);
 
-       if (drm_mm_initialized(&ggtt->base.mm)) {
+       if (drm_mm_initialized(&ggtt->vm.mm)) {
                intel_vgt_deballoon(dev_priv);
-               i915_address_space_fini(&ggtt->base);
+               i915_address_space_fini(&ggtt->vm);
        }
 
-       ggtt->base.cleanup(&ggtt->base);
+       ggtt->vm.cleanup(&ggtt->vm);
 
        pvec = &dev_priv->mm.wc_stash;
        if (pvec->nr) {
@@ -2955,7 +2954,7 @@ static unsigned int chv_get_total_gtt_size(u16 gmch_ctrl)
 
 static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size)
 {
-       struct drm_i915_private *dev_priv = ggtt->base.i915;
+       struct drm_i915_private *dev_priv = ggtt->vm.i915;
        struct pci_dev *pdev = dev_priv->drm.pdev;
        phys_addr_t phys_addr;
        int ret;
@@ -2979,7 +2978,7 @@ static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size)
                return -ENOMEM;
        }
 
-       ret = setup_scratch_page(&ggtt->base, GFP_DMA32);
+       ret = setup_scratch_page(&ggtt->vm, GFP_DMA32);
        if (ret) {
                DRM_ERROR("Scratch setup failed\n");
                /* iounmap will also get called at remove, but meh */
@@ -3285,7 +3284,7 @@ static void setup_private_pat(struct drm_i915_private *dev_priv)
 
 static int gen8_gmch_probe(struct i915_ggtt *ggtt)
 {
-       struct drm_i915_private *dev_priv = ggtt->base.i915;
+       struct drm_i915_private *dev_priv = ggtt->vm.i915;
        struct pci_dev *pdev = dev_priv->drm.pdev;
        unsigned int size;
        u16 snb_gmch_ctl;
@@ -3309,25 +3308,25 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt)
        else
                size = gen8_get_total_gtt_size(snb_gmch_ctl);
 
-       ggtt->base.total = (size / sizeof(gen8_pte_t)) << PAGE_SHIFT;
-       ggtt->base.cleanup = gen6_gmch_remove;
-       ggtt->base.bind_vma = ggtt_bind_vma;
-       ggtt->base.unbind_vma = ggtt_unbind_vma;
-       ggtt->base.set_pages = ggtt_set_pages;
-       ggtt->base.clear_pages = clear_pages;
-       ggtt->base.insert_page = gen8_ggtt_insert_page;
-       ggtt->base.clear_range = nop_clear_range;
+       ggtt->vm.total = (size / sizeof(gen8_pte_t)) << PAGE_SHIFT;
+       ggtt->vm.cleanup = gen6_gmch_remove;
+       ggtt->vm.bind_vma = ggtt_bind_vma;
+       ggtt->vm.unbind_vma = ggtt_unbind_vma;
+       ggtt->vm.set_pages = ggtt_set_pages;
+       ggtt->vm.clear_pages = clear_pages;
+       ggtt->vm.insert_page = gen8_ggtt_insert_page;
+       ggtt->vm.clear_range = nop_clear_range;
        if (!USES_FULL_PPGTT(dev_priv) || intel_scanout_needs_vtd_wa(dev_priv))
-               ggtt->base.clear_range = gen8_ggtt_clear_range;
+               ggtt->vm.clear_range = gen8_ggtt_clear_range;
 
-       ggtt->base.insert_entries = gen8_ggtt_insert_entries;
+       ggtt->vm.insert_entries = gen8_ggtt_insert_entries;
 
        /* Serialize GTT updates with aperture access on BXT if VT-d is on. */
        if (intel_ggtt_update_needs_vtd_wa(dev_priv)) {
-               ggtt->base.insert_entries = bxt_vtd_ggtt_insert_entries__BKL;
-               ggtt->base.insert_page    = bxt_vtd_ggtt_insert_page__BKL;
-               if (ggtt->base.clear_range != nop_clear_range)
-                       ggtt->base.clear_range = bxt_vtd_ggtt_clear_range__BKL;
+               ggtt->vm.insert_entries = bxt_vtd_ggtt_insert_entries__BKL;
+               ggtt->vm.insert_page    = bxt_vtd_ggtt_insert_page__BKL;
+               if (ggtt->vm.clear_range != nop_clear_range)
+                       ggtt->vm.clear_range = bxt_vtd_ggtt_clear_range__BKL;
        }
 
        ggtt->invalidate = gen6_ggtt_invalidate;
@@ -3339,7 +3338,7 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt)
 
 static int gen6_gmch_probe(struct i915_ggtt *ggtt)
 {
-       struct drm_i915_private *dev_priv = ggtt->base.i915;
+       struct drm_i915_private *dev_priv = ggtt->vm.i915;
        struct pci_dev *pdev = dev_priv->drm.pdev;
        unsigned int size;
        u16 snb_gmch_ctl;
@@ -3366,29 +3365,29 @@ static int gen6_gmch_probe(struct i915_ggtt *ggtt)
        pci_read_config_word(pdev, SNB_GMCH_CTRL, &snb_gmch_ctl);
 
        size = gen6_get_total_gtt_size(snb_gmch_ctl);
-       ggtt->base.total = (size / sizeof(gen6_pte_t)) << PAGE_SHIFT;
+       ggtt->vm.total = (size / sizeof(gen6_pte_t)) << PAGE_SHIFT;
 
-       ggtt->base.clear_range = gen6_ggtt_clear_range;
-       ggtt->base.insert_page = gen6_ggtt_insert_page;
-       ggtt->base.insert_entries = gen6_ggtt_insert_entries;
-       ggtt->base.bind_vma = ggtt_bind_vma;
-       ggtt->base.unbind_vma = ggtt_unbind_vma;
-       ggtt->base.set_pages = ggtt_set_pages;
-       ggtt->base.clear_pages = clear_pages;
-       ggtt->base.cleanup = gen6_gmch_remove;
+       ggtt->vm.clear_range = gen6_ggtt_clear_range;
+       ggtt->vm.insert_page = gen6_ggtt_insert_page;
+       ggtt->vm.insert_entries = gen6_ggtt_insert_entries;
+       ggtt->vm.bind_vma = ggtt_bind_vma;
+       ggtt->vm.unbind_vma = ggtt_unbind_vma;
+       ggtt->vm.set_pages = ggtt_set_pages;
+       ggtt->vm.clear_pages = clear_pages;
+       ggtt->vm.cleanup = gen6_gmch_remove;
 
        ggtt->invalidate = gen6_ggtt_invalidate;
 
        if (HAS_EDRAM(dev_priv))
-               ggtt->base.pte_encode = iris_pte_encode;
+               ggtt->vm.pte_encode = iris_pte_encode;
        else if (IS_HASWELL(dev_priv))
-               ggtt->base.pte_encode = hsw_pte_encode;
+               ggtt->vm.pte_encode = hsw_pte_encode;
        else if (IS_VALLEYVIEW(dev_priv))
-               ggtt->base.pte_encode = byt_pte_encode;
+               ggtt->vm.pte_encode = byt_pte_encode;
        else if (INTEL_GEN(dev_priv) >= 7)
-               ggtt->base.pte_encode = ivb_pte_encode;
+               ggtt->vm.pte_encode = ivb_pte_encode;
        else
-               ggtt->base.pte_encode = snb_pte_encode;
+               ggtt->vm.pte_encode = snb_pte_encode;
 
        return ggtt_probe_common(ggtt, size);
 }
@@ -3400,7 +3399,7 @@ static void i915_gmch_remove(struct i915_address_space *vm)
 
 static int i915_gmch_probe(struct i915_ggtt *ggtt)
 {
-       struct drm_i915_private *dev_priv = ggtt->base.i915;
+       struct drm_i915_private *dev_priv = ggtt->vm.i915;
        phys_addr_t gmadr_base;
        int ret;
 
@@ -3410,23 +3409,21 @@ static int i915_gmch_probe(struct i915_ggtt *ggtt)
                return -EIO;
        }
 
-       intel_gtt_get(&ggtt->base.total,
-                     &gmadr_base,
-                     &ggtt->mappable_end);
+       intel_gtt_get(&ggtt->vm.total, &gmadr_base, &ggtt->mappable_end);
 
        ggtt->gmadr =
                (struct resource) DEFINE_RES_MEM(gmadr_base,
                                                 ggtt->mappable_end);
 
        ggtt->do_idle_maps = needs_idle_maps(dev_priv);
-       ggtt->base.insert_page = i915_ggtt_insert_page;
-       ggtt->base.insert_entries = i915_ggtt_insert_entries;
-       ggtt->base.clear_range = i915_ggtt_clear_range;
-       ggtt->base.bind_vma = ggtt_bind_vma;
-       ggtt->base.unbind_vma = ggtt_unbind_vma;
-       ggtt->base.set_pages = ggtt_set_pages;
-       ggtt->base.clear_pages = clear_pages;
-       ggtt->base.cleanup = i915_gmch_remove;
+       ggtt->vm.insert_page = i915_ggtt_insert_page;
+       ggtt->vm.insert_entries = i915_ggtt_insert_entries;
+       ggtt->vm.clear_range = i915_ggtt_clear_range;
+       ggtt->vm.bind_vma = ggtt_bind_vma;
+       ggtt->vm.unbind_vma = ggtt_unbind_vma;
+       ggtt->vm.set_pages = ggtt_set_pages;
+       ggtt->vm.clear_pages = clear_pages;
+       ggtt->vm.cleanup = i915_gmch_remove;
 
        ggtt->invalidate = gmch_ggtt_invalidate;
 
@@ -3445,8 +3442,8 @@ int i915_ggtt_probe_hw(struct drm_i915_private *dev_priv)
        struct i915_ggtt *ggtt = &dev_priv->ggtt;
        int ret;
 
-       ggtt->base.i915 = dev_priv;
-       ggtt->base.dma = &dev_priv->drm.pdev->dev;
+       ggtt->vm.i915 = dev_priv;
+       ggtt->vm.dma = &dev_priv->drm.pdev->dev;
 
        if (INTEL_GEN(dev_priv) <= 5)
                ret = i915_gmch_probe(ggtt);
@@ -3463,27 +3460,29 @@ int i915_ggtt_probe_hw(struct drm_i915_private *dev_priv)
         * restriction!
         */
        if (USES_GUC(dev_priv)) {
-               ggtt->base.total = min_t(u64, ggtt->base.total, GUC_GGTT_TOP);
-               ggtt->mappable_end = min_t(u64, ggtt->mappable_end, ggtt->base.total);
+               ggtt->vm.total = min_t(u64, ggtt->vm.total, GUC_GGTT_TOP);
+               ggtt->mappable_end =
+                       min_t(u64, ggtt->mappable_end, ggtt->vm.total);
        }
 
-       if ((ggtt->base.total - 1) >> 32) {
+       if ((ggtt->vm.total - 1) >> 32) {
                DRM_ERROR("We never expected a Global GTT with more than 32bits"
                          " of address space! Found %lldM!\n",
-                         ggtt->base.total >> 20);
-               ggtt->base.total = 1ULL << 32;
-               ggtt->mappable_end = min_t(u64, ggtt->mappable_end, ggtt->base.total);
+                         ggtt->vm.total >> 20);
+               ggtt->vm.total = 1ULL << 32;
+               ggtt->mappable_end =
+                       min_t(u64, ggtt->mappable_end, ggtt->vm.total);
        }
 
-       if (ggtt->mappable_end > ggtt->base.total) {
+       if (ggtt->mappable_end > ggtt->vm.total) {
                DRM_ERROR("mappable aperture extends past end of GGTT,"
                          " aperture=%pa, total=%llx\n",
-                         &ggtt->mappable_end, ggtt->base.total);
-               ggtt->mappable_end = ggtt->base.total;
+                         &ggtt->mappable_end, ggtt->vm.total);
+               ggtt->mappable_end = ggtt->vm.total;
        }
 
        /* GMADR is the PCI mmio aperture into the global GTT. */
-       DRM_DEBUG_DRIVER("GGTT size = %lluM\n", ggtt->base.total >> 20);
+       DRM_DEBUG_DRIVER("GGTT size = %lluM\n", ggtt->vm.total >> 20);
        DRM_DEBUG_DRIVER("GMADR size = %lluM\n", (u64)ggtt->mappable_end >> 20);
        DRM_DEBUG_DRIVER("DSM size = %lluM\n",
                         (u64)resource_size(&intel_graphics_stolen_res) >> 20);
@@ -3510,9 +3509,9 @@ int i915_ggtt_init_hw(struct drm_i915_private *dev_priv)
         * and beyond the end of the GTT if we do not provide a guard.
         */
        mutex_lock(&dev_priv->drm.struct_mutex);
-       i915_address_space_init(&ggtt->base, dev_priv, "[global]");
+       i915_address_space_init(&ggtt->vm, dev_priv, "[global]");
        if (!HAS_LLC(dev_priv) && !USES_PPGTT(dev_priv))
-               ggtt->base.mm.color_adjust = i915_gtt_color_adjust;
+               ggtt->vm.mm.color_adjust = i915_gtt_color_adjust;
        mutex_unlock(&dev_priv->drm.struct_mutex);
 
        if (!io_mapping_init_wc(&dev_priv->ggtt.iomap,
@@ -3535,7 +3534,7 @@ int i915_ggtt_init_hw(struct drm_i915_private *dev_priv)
        return 0;
 
 out_gtt_cleanup:
-       ggtt->base.cleanup(&ggtt->base);
+       ggtt->vm.cleanup(&ggtt->vm);
        return ret;
 }
 
@@ -3574,13 +3573,13 @@ void i915_gem_restore_gtt_mappings(struct drm_i915_private *dev_priv)
        i915_check_and_clear_faults(dev_priv);
 
        /* First fill our portion of the GTT with scratch pages */
-       ggtt->base.clear_range(&ggtt->base, 0, ggtt->base.total);
+       ggtt->vm.clear_range(&ggtt->vm, 0, ggtt->vm.total);
 
-       ggtt->base.closed = true; /* skip rewriting PTE on VMA unbind */
+       ggtt->vm.closed = true; /* skip rewriting PTE on VMA unbind */
 
        /* clflush objects bound into the GGTT and rebind them. */
-       GEM_BUG_ON(!list_empty(&ggtt->base.active_list));
-       list_for_each_entry_safe(vma, vn, &ggtt->base.inactive_list, vm_link) {
+       GEM_BUG_ON(!list_empty(&ggtt->vm.active_list));
+       list_for_each_entry_safe(vma, vn, &ggtt->vm.inactive_list, vm_link) {
                struct drm_i915_gem_object *obj = vma->obj;
 
                if (!(vma->flags & I915_VMA_GLOBAL_BIND))
@@ -3593,7 +3592,7 @@ void i915_gem_restore_gtt_mappings(struct drm_i915_private *dev_priv)
                WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false));
        }
 
-       ggtt->base.closed = false;
+       ggtt->vm.closed = false;
 
        if (INTEL_GEN(dev_priv) >= 8) {
                struct intel_ppat *ppat = &dev_priv->ppat;
@@ -3616,7 +3615,7 @@ void i915_gem_restore_gtt_mappings(struct drm_i915_private *dev_priv)
                        if (!ppgtt)
                                continue;
 
-                       gen6_write_page_range(ppgtt, 0, ppgtt->base.total);
+                       gen6_write_page_range(ppgtt, 0, ppgtt->vm.total);
                }
        }
 
@@ -3838,7 +3837,7 @@ int i915_gem_gtt_reserve(struct i915_address_space *vm,
        GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE));
        GEM_BUG_ON(!IS_ALIGNED(offset, I915_GTT_MIN_ALIGNMENT));
        GEM_BUG_ON(range_overflows(offset, size, vm->total));
-       GEM_BUG_ON(vm == &vm->i915->mm.aliasing_ppgtt->base);
+       GEM_BUG_ON(vm == &vm->i915->mm.aliasing_ppgtt->vm);
        GEM_BUG_ON(drm_mm_node_allocated(node));
 
        node->size = size;
@@ -3935,7 +3934,7 @@ int i915_gem_gtt_insert(struct i915_address_space *vm,
        GEM_BUG_ON(start >= end);
        GEM_BUG_ON(start > 0  && !IS_ALIGNED(start, I915_GTT_PAGE_SIZE));
        GEM_BUG_ON(end < U64_MAX && !IS_ALIGNED(end, I915_GTT_PAGE_SIZE));
-       GEM_BUG_ON(vm == &vm->i915->mm.aliasing_ppgtt->base);
+       GEM_BUG_ON(vm == &vm->i915->mm.aliasing_ppgtt->vm);
        GEM_BUG_ON(drm_mm_node_allocated(node));
 
        if (unlikely(range_overflows(start, size, end)))
index aec4f73574f4de684be02e0687b4da763344435e..197c2c06ecb73afb835c2fe2a0f10eee07a9895d 100644 (file)
@@ -65,7 +65,7 @@ typedef u64 gen8_pde_t;
 typedef u64 gen8_ppgtt_pdpe_t;
 typedef u64 gen8_ppgtt_pml4e_t;
 
-#define ggtt_total_entries(ggtt) ((ggtt)->base.total >> PAGE_SHIFT)
+#define ggtt_total_entries(ggtt) ((ggtt)->vm.total >> PAGE_SHIFT)
 
 /* gen6-hsw has bit 11-4 for physical addr bit 39-32 */
 #define GEN6_GTT_ADDR_ENCODE(addr)     ((addr) | (((addr) >> 28) & 0xff0))
@@ -367,7 +367,7 @@ i915_vm_has_scratch_64K(struct i915_address_space *vm)
  * the spec.
  */
 struct i915_ggtt {
-       struct i915_address_space base;
+       struct i915_address_space vm;
 
        struct io_mapping iomap;        /* Mapping to our CPU mappable region */
        struct resource gmadr;          /* GMADR resource */
@@ -385,7 +385,7 @@ struct i915_ggtt {
 };
 
 struct i915_hw_ppgtt {
-       struct i915_address_space base;
+       struct i915_address_space vm;
        struct kref ref;
        struct drm_mm_node node;
        unsigned long pd_dirty_rings;
@@ -543,7 +543,7 @@ static inline struct i915_ggtt *
 i915_vm_to_ggtt(struct i915_address_space *vm)
 {
        GEM_BUG_ON(!i915_is_ggtt(vm));
-       return container_of(vm, struct i915_ggtt, base);
+       return container_of(vm, struct i915_ggtt, vm);
 }
 
 #define INTEL_MAX_PPAT_ENTRIES 8
index 1036e8686916578accdb30f850401b4b384a6ba0..3210cedfa46c374829fa014932899a5637c610e7 100644 (file)
@@ -194,7 +194,7 @@ int i915_gem_render_state_emit(struct i915_request *rq)
        if (IS_ERR(so.obj))
                return PTR_ERR(so.obj);
 
-       so.vma = i915_vma_instance(so.obj, &engine->i915->ggtt.base, NULL);
+       so.vma = i915_vma_instance(so.obj, &engine->i915->ggtt.vm, NULL);
        if (IS_ERR(so.vma)) {
                err = PTR_ERR(so.vma);
                goto err_obj;
index 5757fb7c4b5af567d6841f128887875f03d96487..55e84e71f526d23bb34a71f1db60cd864e151e6c 100644 (file)
@@ -480,7 +480,7 @@ i915_gem_shrinker_vmap(struct notifier_block *nb, unsigned long event, void *ptr
 
        /* We also want to clear any cached iomaps as they wrap vmap */
        list_for_each_entry_safe(vma, next,
-                                &i915->ggtt.base.inactive_list, vm_link) {
+                                &i915->ggtt.vm.inactive_list, vm_link) {
                unsigned long count = vma->node.size >> PAGE_SHIFT;
                if (vma->iomap && i915_vma_unbind(vma) == 0)
                        freed_pages += count;
index ad949cc3092816cc123d8109d80151dfcc9e6ac9..79a347295e006e53de34acbdc8d5e86892ca21bd 100644 (file)
@@ -642,7 +642,7 @@ i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *dev_priv
        if (ret)
                goto err;
 
-       vma = i915_vma_instance(obj, &ggtt->base, NULL);
+       vma = i915_vma_instance(obj, &ggtt->vm, NULL);
        if (IS_ERR(vma)) {
                ret = PTR_ERR(vma);
                goto err_pages;
@@ -653,7 +653,7 @@ i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *dev_priv
         * setting up the GTT space. The actual reservation will occur
         * later.
         */
-       ret = i915_gem_gtt_reserve(&ggtt->base, &vma->node,
+       ret = i915_gem_gtt_reserve(&ggtt->vm, &vma->node,
                                   size, gtt_offset, obj->cache_level,
                                   0);
        if (ret) {
@@ -666,7 +666,7 @@ i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *dev_priv
        vma->pages = obj->mm.pages;
        vma->flags |= I915_VMA_GLOBAL_BIND;
        __i915_vma_set_map_and_fenceable(vma);
-       list_move_tail(&vma->vm_link, &ggtt->base.inactive_list);
+       list_move_tail(&vma->vm_link, &ggtt->vm.inactive_list);
 
        spin_lock(&dev_priv->mm.obj_lock);
        list_move_tail(&obj->mm.link, &dev_priv->mm.bound_list);
index 6702776303bfd0ad0c4ea963a768681baeb81660..758234d20f4ea9b464bef01321a2d64a7253f538 100644 (file)
@@ -973,8 +973,7 @@ i915_error_object_create(struct drm_i915_private *i915,
                void __iomem *s;
                int ret;
 
-               ggtt->base.insert_page(&ggtt->base, dma, slot,
-                                      I915_CACHE_NONE, 0);
+               ggtt->vm.insert_page(&ggtt->vm, dma, slot, I915_CACHE_NONE, 0);
 
                s = io_mapping_map_atomic_wc(&ggtt->iomap, slot);
                ret = compress_page(&compress, (void  __force *)s, dst);
@@ -993,7 +992,7 @@ unwind:
 
 out:
        compress_fini(&compress, dst);
-       ggtt->base.clear_range(&ggtt->base, slot, PAGE_SIZE);
+       ggtt->vm.clear_range(&ggtt->vm, slot, PAGE_SIZE);
        return dst;
 }
 
@@ -1466,7 +1465,7 @@ static void gem_record_rings(struct i915_gpu_state *error)
                        struct i915_gem_context *ctx = request->gem_context;
                        struct intel_ring *ring;
 
-                       ee->vm = ctx->ppgtt ? &ctx->ppgtt->base : &ggtt->base;
+                       ee->vm = ctx->ppgtt ? &ctx->ppgtt->vm : &ggtt->vm;
 
                        record_context(&ee->context, ctx);
 
@@ -1564,7 +1563,7 @@ static void capture_active_buffers(struct i915_gpu_state *error)
 
 static void capture_pinned_buffers(struct i915_gpu_state *error)
 {
-       struct i915_address_space *vm = &error->i915->ggtt.base;
+       struct i915_address_space *vm = &error->i915->ggtt.vm;
        struct drm_i915_error_buffer *bo;
        struct i915_vma *vma;
        int count_inactive, count_active;
index 3d5716d86e279ddf5306aeb5729837374456845c..1472f48ab2e804ca9404155b8004d18237b657ec 100644 (file)
@@ -956,7 +956,7 @@ DECLARE_EVENT_CLASS(i915_context,
                        __entry->dev = ctx->i915->drm.primary->index;
                        __entry->ctx = ctx;
                        __entry->hw_id = ctx->hw_id;
-                       __entry->vm = ctx->ppgtt ? &ctx->ppgtt->base : NULL;
+                       __entry->vm = ctx->ppgtt ? &ctx->ppgtt->vm : NULL;
        ),
 
        TP_printk("dev=%u, ctx=%p, ctx_vm=%p, hw_id=%u",
@@ -997,7 +997,7 @@ TRACE_EVENT(switch_mm,
                        __entry->class = engine->uabi_class;
                        __entry->instance = engine->instance;
                        __entry->to = to;
-                       __entry->vm = to->ppgtt? &to->ppgtt->base : NULL;
+                       __entry->vm = to->ppgtt ? &to->ppgtt->vm : NULL;
                        __entry->dev = engine->i915->drm.primary->index;
        ),
 
index 5fe9f3f3946728aa7985ab6f251f685bc97cfe6b..869cf4a3b6de75fee593c0f66c953cc1035434a6 100644 (file)
@@ -105,7 +105,7 @@ static void vgt_deballoon_space(struct i915_ggtt *ggtt,
                         node->start + node->size,
                         node->size / 1024);
 
-       ggtt->base.reserved -= node->size;
+       ggtt->vm.reserved -= node->size;
        drm_mm_remove_node(node);
 }
 
@@ -141,11 +141,11 @@ static int vgt_balloon_space(struct i915_ggtt *ggtt,
 
        DRM_INFO("balloon space: range [ 0x%lx - 0x%lx ] %lu KiB.\n",
                 start, end, size / 1024);
-       ret = i915_gem_gtt_reserve(&ggtt->base, node,
+       ret = i915_gem_gtt_reserve(&ggtt->vm, node,
                                   size, start, I915_COLOR_UNEVICTABLE,
                                   0);
        if (!ret)
-               ggtt->base.reserved += size;
+               ggtt->vm.reserved += size;
 
        return ret;
 }
@@ -197,7 +197,7 @@ static int vgt_balloon_space(struct i915_ggtt *ggtt,
 int intel_vgt_balloon(struct drm_i915_private *dev_priv)
 {
        struct i915_ggtt *ggtt = &dev_priv->ggtt;
-       unsigned long ggtt_end = ggtt->base.total;
+       unsigned long ggtt_end = ggtt->vm.total;
 
        unsigned long mappable_base, mappable_size, mappable_end;
        unsigned long unmappable_base, unmappable_size, unmappable_end;
index 10bf654cd0235daf4b0a9c02ab4a8b7f0f0e222c..912f16ffe7eea4b62f1f562fd16938d1c69fb14f 100644 (file)
@@ -85,7 +85,7 @@ vma_create(struct drm_i915_gem_object *obj,
        int i;
 
        /* The aliasing_ppgtt should never be used directly! */
-       GEM_BUG_ON(vm == &vm->i915->mm.aliasing_ppgtt->base);
+       GEM_BUG_ON(vm == &vm->i915->mm.aliasing_ppgtt->vm);
 
        vma = kmem_cache_zalloc(vm->i915->vmas, GFP_KERNEL);
        if (vma == NULL)
index 13448ea76f57bb653a383f2105dc8e00b75dc867..2ec2e60dc670711e2593994d5b58773c32733e06 100644 (file)
@@ -515,7 +515,7 @@ int intel_engine_create_scratch(struct intel_engine_cs *engine, int size)
                return PTR_ERR(obj);
        }
 
-       vma = i915_vma_instance(obj, &engine->i915->ggtt.base, NULL);
+       vma = i915_vma_instance(obj, &engine->i915->ggtt.vm, NULL);
        if (IS_ERR(vma)) {
                ret = PTR_ERR(vma);
                goto err_unref;
@@ -585,7 +585,7 @@ static int init_status_page(struct intel_engine_cs *engine)
        if (ret)
                goto err;
 
-       vma = i915_vma_instance(obj, &engine->i915->ggtt.base, NULL);
+       vma = i915_vma_instance(obj, &engine->i915->ggtt.vm, NULL);
        if (IS_ERR(vma)) {
                ret = PTR_ERR(vma);
                goto err;
index e28a996b9604a0f75db675f57bed2601848d0bd4..29fd95c1306bd1a52bf453068aa00cf3b7573be8 100644 (file)
@@ -570,7 +570,7 @@ struct i915_vma *intel_guc_allocate_vma(struct intel_guc *guc, u32 size)
        if (IS_ERR(obj))
                return ERR_CAST(obj);
 
-       vma = i915_vma_instance(obj, &dev_priv->ggtt.base, NULL);
+       vma = i915_vma_instance(obj, &dev_priv->ggtt.vm, NULL);
        if (IS_ERR(vma))
                goto err;
 
index 23e9f2023bc56201a6f405e4ad3755893f580709..f3945258fe1b53c1723de5a584caa66522aab774 100644 (file)
@@ -536,7 +536,7 @@ static void guc_add_request(struct intel_guc *guc, struct i915_request *rq)
  */
 static void flush_ggtt_writes(struct i915_vma *vma)
 {
-       struct drm_i915_private *dev_priv = to_i915(vma->obj->base.dev);
+       struct drm_i915_private *dev_priv = vma->vm->i915;
 
        if (i915_vma_is_map_and_fenceable(vma))
                POSTING_READ_FW(GUC_STATUS);
index f6f09f808f74ae0d6c34a4a0ecefc6704c21af87..0935a19aca96f7ed3d740554103d4cdeae4b5715 100644 (file)
@@ -431,7 +431,7 @@ static u64 execlists_update_context(struct i915_request *rq)
         * PML4 is allocated during ppgtt init, so this is not needed
         * in 48-bit mode.
         */
-       if (ppgtt && !i915_vm_is_48bit(&ppgtt->base))
+       if (ppgtt && !i915_vm_is_48bit(&ppgtt->vm))
                execlists_update_context_pdps(ppgtt, reg_state);
 
        return ce->lrc_desc;
@@ -1672,7 +1672,7 @@ static int lrc_setup_wa_ctx(struct intel_engine_cs *engine)
        if (IS_ERR(obj))
                return PTR_ERR(obj);
 
-       vma = i915_vma_instance(obj, &engine->i915->ggtt.base, NULL);
+       vma = i915_vma_instance(obj, &engine->i915->ggtt.vm, NULL);
        if (IS_ERR(vma)) {
                err = PTR_ERR(vma);
                goto err;
@@ -2070,7 +2070,7 @@ static int gen8_emit_bb_start(struct i915_request *rq,
         * not needed in 48-bit.*/
        if (rq->gem_context->ppgtt &&
            (intel_engine_flag(rq->engine) & rq->gem_context->ppgtt->pd_dirty_rings) &&
-           !i915_vm_is_48bit(&rq->gem_context->ppgtt->base) &&
+           !i915_vm_is_48bit(&rq->gem_context->ppgtt->vm) &&
            !intel_vgpu_active(rq->i915)) {
                ret = intel_logical_ring_emit_pdps(rq);
                if (ret)
@@ -2668,7 +2668,7 @@ static void execlists_init_reg_state(u32 *regs,
        CTX_REG(regs, CTX_PDP0_UDW, GEN8_RING_PDP_UDW(engine, 0), 0);
        CTX_REG(regs, CTX_PDP0_LDW, GEN8_RING_PDP_LDW(engine, 0), 0);
 
-       if (ppgtt && i915_vm_is_48bit(&ppgtt->base)) {
+       if (ppgtt && i915_vm_is_48bit(&ppgtt->vm)) {
                /* 64b PPGTT (48bit canonical)
                 * PDP0_DESCRIPTOR contains the base address to PML4 and
                 * other PDP Descriptors are ignored.
@@ -2774,7 +2774,7 @@ static int execlists_context_deferred_alloc(struct i915_gem_context *ctx,
                goto error_deref_obj;
        }
 
-       vma = i915_vma_instance(ctx_obj, &ctx->i915->ggtt.base, NULL);
+       vma = i915_vma_instance(ctx_obj, &ctx->i915->ggtt.vm, NULL);
        if (IS_ERR(vma)) {
                ret = PTR_ERR(vma);
                goto error_deref_obj;
index 3f0eb538df098cc0cb656b5a76064941910006e7..6496c1d00dbbac3f96f876d7f531f5188124692c 100644 (file)
@@ -1121,7 +1121,7 @@ intel_ring_create_vma(struct drm_i915_private *dev_priv, int size)
        /* mark ring buffers as read-only from GPU side by default */
        obj->gt_ro = 1;
 
-       vma = i915_vma_instance(obj, &dev_priv->ggtt.base, NULL);
+       vma = i915_vma_instance(obj, &dev_priv->ggtt.vm, NULL);
        if (IS_ERR(vma))
                goto err;
 
@@ -1300,7 +1300,7 @@ alloc_context_vma(struct intel_engine_cs *engine)
                i915_gem_object_set_cache_level(obj, I915_CACHE_L3_LLC);
        }
 
-       vma = i915_vma_instance(obj, &i915->ggtt.base, NULL);
+       vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
        if (IS_ERR(vma)) {
                err = PTR_ERR(vma);
                goto err_obj;
index 91c72911be3cb73b126e9925648746c7db214f93..7846ea4a99bc6065546694970f9fd3e7ed9a8e08 100644 (file)
@@ -338,7 +338,7 @@ fake_huge_pages_object(struct drm_i915_private *i915, u64 size, bool single)
 
 static int igt_check_page_sizes(struct i915_vma *vma)
 {
-       struct drm_i915_private *i915 = to_i915(vma->obj->base.dev);
+       struct drm_i915_private *i915 = vma->vm->i915;
        unsigned int supported = INTEL_INFO(i915)->page_sizes;
        struct drm_i915_gem_object *obj = vma->obj;
        int err = 0;
@@ -379,7 +379,7 @@ static int igt_check_page_sizes(struct i915_vma *vma)
 static int igt_mock_exhaust_device_supported_pages(void *arg)
 {
        struct i915_hw_ppgtt *ppgtt = arg;
-       struct drm_i915_private *i915 = ppgtt->base.i915;
+       struct drm_i915_private *i915 = ppgtt->vm.i915;
        unsigned int saved_mask = INTEL_INFO(i915)->page_sizes;
        struct drm_i915_gem_object *obj;
        struct i915_vma *vma;
@@ -415,7 +415,7 @@ static int igt_mock_exhaust_device_supported_pages(void *arg)
                                goto out_put;
                        }
 
-                       vma = i915_vma_instance(obj, &ppgtt->base, NULL);
+                       vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
                        if (IS_ERR(vma)) {
                                err = PTR_ERR(vma);
                                goto out_put;
@@ -458,7 +458,7 @@ out_device:
 static int igt_mock_ppgtt_misaligned_dma(void *arg)
 {
        struct i915_hw_ppgtt *ppgtt = arg;
-       struct drm_i915_private *i915 = ppgtt->base.i915;
+       struct drm_i915_private *i915 = ppgtt->vm.i915;
        unsigned long supported = INTEL_INFO(i915)->page_sizes;
        struct drm_i915_gem_object *obj;
        int bit;
@@ -500,7 +500,7 @@ static int igt_mock_ppgtt_misaligned_dma(void *arg)
                /* Force the page size for this object */
                obj->mm.page_sizes.sg = page_size;
 
-               vma = i915_vma_instance(obj, &ppgtt->base, NULL);
+               vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
                if (IS_ERR(vma)) {
                        err = PTR_ERR(vma);
                        goto out_unpin;
@@ -591,7 +591,7 @@ static void close_object_list(struct list_head *objects,
        list_for_each_entry_safe(obj, on, objects, st_link) {
                struct i915_vma *vma;
 
-               vma = i915_vma_instance(obj, &ppgtt->base, NULL);
+               vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
                if (!IS_ERR(vma))
                        i915_vma_close(vma);
 
@@ -604,8 +604,8 @@ static void close_object_list(struct list_head *objects,
 static int igt_mock_ppgtt_huge_fill(void *arg)
 {
        struct i915_hw_ppgtt *ppgtt = arg;
-       struct drm_i915_private *i915 = ppgtt->base.i915;
-       unsigned long max_pages = ppgtt->base.total >> PAGE_SHIFT;
+       struct drm_i915_private *i915 = ppgtt->vm.i915;
+       unsigned long max_pages = ppgtt->vm.total >> PAGE_SHIFT;
        unsigned long page_num;
        bool single = false;
        LIST_HEAD(objects);
@@ -641,7 +641,7 @@ static int igt_mock_ppgtt_huge_fill(void *arg)
 
                list_add(&obj->st_link, &objects);
 
-               vma = i915_vma_instance(obj, &ppgtt->base, NULL);
+               vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
                if (IS_ERR(vma)) {
                        err = PTR_ERR(vma);
                        break;
@@ -725,7 +725,7 @@ static int igt_mock_ppgtt_huge_fill(void *arg)
 static int igt_mock_ppgtt_64K(void *arg)
 {
        struct i915_hw_ppgtt *ppgtt = arg;
-       struct drm_i915_private *i915 = ppgtt->base.i915;
+       struct drm_i915_private *i915 = ppgtt->vm.i915;
        struct drm_i915_gem_object *obj;
        const struct object_info {
                unsigned int size;
@@ -819,7 +819,7 @@ static int igt_mock_ppgtt_64K(void *arg)
                         */
                        obj->mm.page_sizes.sg &= ~I915_GTT_PAGE_SIZE_2M;
 
-                       vma = i915_vma_instance(obj, &ppgtt->base, NULL);
+                       vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
                        if (IS_ERR(vma)) {
                                err = PTR_ERR(vma);
                                goto out_object_unpin;
@@ -887,8 +887,8 @@ out_object_put:
 static struct i915_vma *
 gpu_write_dw(struct i915_vma *vma, u64 offset, u32 val)
 {
-       struct drm_i915_private *i915 = to_i915(vma->obj->base.dev);
-       const int gen = INTEL_GEN(vma->vm->i915);
+       struct drm_i915_private *i915 = vma->vm->i915;
+       const int gen = INTEL_GEN(i915);
        unsigned int count = vma->size >> PAGE_SHIFT;
        struct drm_i915_gem_object *obj;
        struct i915_vma *batch;
@@ -1047,7 +1047,8 @@ static int __igt_write_huge(struct i915_gem_context *ctx,
                            u32 dword, u32 val)
 {
        struct drm_i915_private *i915 = to_i915(obj->base.dev);
-       struct i915_address_space *vm = ctx->ppgtt ? &ctx->ppgtt->base : &i915->ggtt.base;
+       struct i915_address_space *vm =
+               ctx->ppgtt ? &ctx->ppgtt->vm : &i915->ggtt.vm;
        unsigned int flags = PIN_USER | PIN_OFFSET_FIXED;
        struct i915_vma *vma;
        int err;
@@ -1100,7 +1101,8 @@ static int igt_write_huge(struct i915_gem_context *ctx,
                          struct drm_i915_gem_object *obj)
 {
        struct drm_i915_private *i915 = to_i915(obj->base.dev);
-       struct i915_address_space *vm = ctx->ppgtt ? &ctx->ppgtt->base : &i915->ggtt.base;
+       struct i915_address_space *vm =
+               ctx->ppgtt ? &ctx->ppgtt->vm : &i915->ggtt.vm;
        static struct intel_engine_cs *engines[I915_NUM_ENGINES];
        struct intel_engine_cs *engine;
        I915_RND_STATE(prng);
@@ -1439,7 +1441,7 @@ static int igt_ppgtt_pin_update(void *arg)
                if (IS_ERR(obj))
                        return PTR_ERR(obj);
 
-               vma = i915_vma_instance(obj, &ppgtt->base, NULL);
+               vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
                if (IS_ERR(vma)) {
                        err = PTR_ERR(vma);
                        goto out_put;
@@ -1493,7 +1495,7 @@ static int igt_ppgtt_pin_update(void *arg)
        if (IS_ERR(obj))
                return PTR_ERR(obj);
 
-       vma = i915_vma_instance(obj, &ppgtt->base, NULL);
+       vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
        if (IS_ERR(vma)) {
                err = PTR_ERR(vma);
                goto out_put;
@@ -1531,7 +1533,8 @@ static int igt_tmpfs_fallback(void *arg)
        struct i915_gem_context *ctx = arg;
        struct drm_i915_private *i915 = ctx->i915;
        struct vfsmount *gemfs = i915->mm.gemfs;
-       struct i915_address_space *vm = ctx->ppgtt ? &ctx->ppgtt->base : &i915->ggtt.base;
+       struct i915_address_space *vm =
+               ctx->ppgtt ? &ctx->ppgtt->vm : &i915->ggtt.vm;
        struct drm_i915_gem_object *obj;
        struct i915_vma *vma;
        u32 *vaddr;
@@ -1587,7 +1590,8 @@ static int igt_shrink_thp(void *arg)
 {
        struct i915_gem_context *ctx = arg;
        struct drm_i915_private *i915 = ctx->i915;
-       struct i915_address_space *vm = ctx->ppgtt ? &ctx->ppgtt->base : &i915->ggtt.base;
+       struct i915_address_space *vm =
+               ctx->ppgtt ? &ctx->ppgtt->vm : &i915->ggtt.vm;
        struct drm_i915_gem_object *obj;
        struct i915_vma *vma;
        unsigned int flags = PIN_USER;
@@ -1696,14 +1700,14 @@ int i915_gem_huge_page_mock_selftests(void)
                goto out_unlock;
        }
 
-       if (!i915_vm_is_48bit(&ppgtt->base)) {
+       if (!i915_vm_is_48bit(&ppgtt->vm)) {
                pr_err("failed to create 48b PPGTT\n");
                err = -EINVAL;
                goto out_close;
        }
 
        /* If we were ever hit this then it's time to mock the 64K scratch */
-       if (!i915_vm_has_scratch_64K(&ppgtt->base)) {
+       if (!i915_vm_has_scratch_64K(&ppgtt->vm)) {
                pr_err("PPGTT missing 64K scratch page\n");
                err = -EINVAL;
                goto out_close;
@@ -1712,7 +1716,7 @@ int i915_gem_huge_page_mock_selftests(void)
        err = i915_subtests(tests, ppgtt);
 
 out_close:
-       i915_ppgtt_close(&ppgtt->base);
+       i915_ppgtt_close(&ppgtt->vm);
        i915_ppgtt_put(ppgtt);
 
 out_unlock:
@@ -1758,7 +1762,7 @@ int i915_gem_huge_page_live_selftests(struct drm_i915_private *dev_priv)
        }
 
        if (ctx->ppgtt)
-               ctx->ppgtt->base.scrub_64K = true;
+               ctx->ppgtt->vm.scrub_64K = true;
 
        err = i915_subtests(tests, ctx);
 
index b39392a00a6fe110f7900cb06cf0f4618303dc0f..708e8d721448868ce6d198459b2660fe74eafd96 100644 (file)
@@ -115,7 +115,7 @@ static int gpu_fill(struct drm_i915_gem_object *obj,
 {
        struct drm_i915_private *i915 = to_i915(obj->base.dev);
        struct i915_address_space *vm =
-               ctx->ppgtt ? &ctx->ppgtt->base : &i915->ggtt.base;
+               ctx->ppgtt ? &ctx->ppgtt->vm : &i915->ggtt.vm;
        struct i915_request *rq;
        struct i915_vma *vma;
        struct i915_vma *batch;
@@ -290,7 +290,7 @@ create_test_object(struct i915_gem_context *ctx,
 {
        struct drm_i915_gem_object *obj;
        struct i915_address_space *vm =
-               ctx->ppgtt ? &ctx->ppgtt->base : &ctx->i915->ggtt.base;
+               ctx->ppgtt ? &ctx->ppgtt->vm : &ctx->i915->ggtt.vm;
        u64 size;
        int err;
 
@@ -557,7 +557,7 @@ static int fake_aliasing_ppgtt_enable(struct drm_i915_private *i915)
        list_for_each_entry(obj, &i915->mm.bound_list, mm.link) {
                struct i915_vma *vma;
 
-               vma = i915_vma_instance(obj, &i915->ggtt.base, NULL);
+               vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
                if (IS_ERR(vma))
                        continue;
 
index ab9d7bee0aae1f3fbe371dbe68274f4423b1826d..2dc72a984d450c72bc863ca217af9b9f471f2e47 100644 (file)
@@ -35,7 +35,7 @@ static int populate_ggtt(struct drm_i915_private *i915)
        u64 size;
 
        for (size = 0;
-            size + I915_GTT_PAGE_SIZE <= i915->ggtt.base.total;
+            size + I915_GTT_PAGE_SIZE <= i915->ggtt.vm.total;
             size += I915_GTT_PAGE_SIZE) {
                struct i915_vma *vma;
 
@@ -57,7 +57,7 @@ static int populate_ggtt(struct drm_i915_private *i915)
                return -EINVAL;
        }
 
-       if (list_empty(&i915->ggtt.base.inactive_list)) {
+       if (list_empty(&i915->ggtt.vm.inactive_list)) {
                pr_err("No objects on the GGTT inactive list!\n");
                return -EINVAL;
        }
@@ -69,7 +69,7 @@ static void unpin_ggtt(struct drm_i915_private *i915)
 {
        struct i915_vma *vma;
 
-       list_for_each_entry(vma, &i915->ggtt.base.inactive_list, vm_link)
+       list_for_each_entry(vma, &i915->ggtt.vm.inactive_list, vm_link)
                i915_vma_unpin(vma);
 }
 
@@ -103,7 +103,7 @@ static int igt_evict_something(void *arg)
                goto cleanup;
 
        /* Everything is pinned, nothing should happen */
-       err = i915_gem_evict_something(&ggtt->base,
+       err = i915_gem_evict_something(&ggtt->vm,
                                       I915_GTT_PAGE_SIZE, 0, 0,
                                       0, U64_MAX,
                                       0);
@@ -116,7 +116,7 @@ static int igt_evict_something(void *arg)
        unpin_ggtt(i915);
 
        /* Everything is unpinned, we should be able to evict something */
-       err = i915_gem_evict_something(&ggtt->base,
+       err = i915_gem_evict_something(&ggtt->vm,
                                       I915_GTT_PAGE_SIZE, 0, 0,
                                       0, U64_MAX,
                                       0);
@@ -181,7 +181,7 @@ static int igt_evict_for_vma(void *arg)
                goto cleanup;
 
        /* Everything is pinned, nothing should happen */
-       err = i915_gem_evict_for_node(&ggtt->base, &target, 0);
+       err = i915_gem_evict_for_node(&ggtt->vm, &target, 0);
        if (err != -ENOSPC) {
                pr_err("i915_gem_evict_for_node on a full GGTT returned err=%d\n",
                       err);
@@ -191,7 +191,7 @@ static int igt_evict_for_vma(void *arg)
        unpin_ggtt(i915);
 
        /* Everything is unpinned, we should be able to evict the node */
-       err = i915_gem_evict_for_node(&ggtt->base, &target, 0);
+       err = i915_gem_evict_for_node(&ggtt->vm, &target, 0);
        if (err) {
                pr_err("i915_gem_evict_for_node returned err=%d\n",
                       err);
@@ -229,7 +229,7 @@ static int igt_evict_for_cache_color(void *arg)
         * i915_gtt_color_adjust throughout our driver, so using a mock color
         * adjust will work just fine for our purposes.
         */
-       ggtt->base.mm.color_adjust = mock_color_adjust;
+       ggtt->vm.mm.color_adjust = mock_color_adjust;
 
        obj = i915_gem_object_create_internal(i915, I915_GTT_PAGE_SIZE);
        if (IS_ERR(obj)) {
@@ -265,7 +265,7 @@ static int igt_evict_for_cache_color(void *arg)
        i915_vma_unpin(vma);
 
        /* Remove just the second vma */
-       err = i915_gem_evict_for_node(&ggtt->base, &target, 0);
+       err = i915_gem_evict_for_node(&ggtt->vm, &target, 0);
        if (err) {
                pr_err("[0]i915_gem_evict_for_node returned err=%d\n", err);
                goto cleanup;
@@ -276,7 +276,7 @@ static int igt_evict_for_cache_color(void *arg)
         */
        target.color = I915_CACHE_L3_LLC;
 
-       err = i915_gem_evict_for_node(&ggtt->base, &target, 0);
+       err = i915_gem_evict_for_node(&ggtt->vm, &target, 0);
        if (!err) {
                pr_err("[1]i915_gem_evict_for_node returned err=%d\n", err);
                err = -EINVAL;
@@ -288,7 +288,7 @@ static int igt_evict_for_cache_color(void *arg)
 cleanup:
        unpin_ggtt(i915);
        cleanup_objects(i915);
-       ggtt->base.mm.color_adjust = NULL;
+       ggtt->vm.mm.color_adjust = NULL;
        return err;
 }
 
@@ -305,7 +305,7 @@ static int igt_evict_vm(void *arg)
                goto cleanup;
 
        /* Everything is pinned, nothing should happen */
-       err = i915_gem_evict_vm(&ggtt->base);
+       err = i915_gem_evict_vm(&ggtt->vm);
        if (err) {
                pr_err("i915_gem_evict_vm on a full GGTT returned err=%d]\n",
                       err);
@@ -314,7 +314,7 @@ static int igt_evict_vm(void *arg)
 
        unpin_ggtt(i915);
 
-       err = i915_gem_evict_vm(&ggtt->base);
+       err = i915_gem_evict_vm(&ggtt->vm);
        if (err) {
                pr_err("i915_gem_evict_vm on a full GGTT returned err=%d]\n",
                       err);
@@ -359,9 +359,9 @@ static int igt_evict_contexts(void *arg)
 
        /* Reserve a block so that we know we have enough to fit a few rq */
        memset(&hole, 0, sizeof(hole));
-       err = i915_gem_gtt_insert(&i915->ggtt.base, &hole,
+       err = i915_gem_gtt_insert(&i915->ggtt.vm, &hole,
                                  PRETEND_GGTT_SIZE, 0, I915_COLOR_UNEVICTABLE,
-                                 0, i915->ggtt.base.total,
+                                 0, i915->ggtt.vm.total,
                                  PIN_NOEVICT);
        if (err)
                goto out_locked;
@@ -377,9 +377,9 @@ static int igt_evict_contexts(void *arg)
                        goto out_locked;
                }
 
-               if (i915_gem_gtt_insert(&i915->ggtt.base, &r->node,
+               if (i915_gem_gtt_insert(&i915->ggtt.vm, &r->node,
                                        1ul << 20, 0, I915_COLOR_UNEVICTABLE,
-                                       0, i915->ggtt.base.total,
+                                       0, i915->ggtt.vm.total,
                                        PIN_NOEVICT)) {
                        kfree(r);
                        break;
index f7dc926f4ef1fa610e5415d0b2771b30fcbdbbe5..58ab5e84ceb7c789eafc0a925b0cc5adaa1050c7 100644 (file)
@@ -151,14 +151,14 @@ static int igt_ppgtt_alloc(void *arg)
        if (err)
                goto err_ppgtt;
 
-       if (!ppgtt->base.allocate_va_range)
+       if (!ppgtt->vm.allocate_va_range)
                goto err_ppgtt_cleanup;
 
        /* Check we can allocate the entire range */
        for (size = 4096;
-            size <= ppgtt->base.total;
+            size <= ppgtt->vm.total;
             size <<= 2) {
-               err = ppgtt->base.allocate_va_range(&ppgtt->base, 0, size);
+               err = ppgtt->vm.allocate_va_range(&ppgtt->vm, 0, size);
                if (err) {
                        if (err == -ENOMEM) {
                                pr_info("[1] Ran out of memory for va_range [0 + %llx] [bit %d]\n",
@@ -168,15 +168,15 @@ static int igt_ppgtt_alloc(void *arg)
                        goto err_ppgtt_cleanup;
                }
 
-               ppgtt->base.clear_range(&ppgtt->base, 0, size);
+               ppgtt->vm.clear_range(&ppgtt->vm, 0, size);
        }
 
        /* Check we can incrementally allocate the entire range */
        for (last = 0, size = 4096;
-            size <= ppgtt->base.total;
+            size <= ppgtt->vm.total;
             last = size, size <<= 2) {
-               err = ppgtt->base.allocate_va_range(&ppgtt->base,
-                                                   last, size - last);
+               err = ppgtt->vm.allocate_va_range(&ppgtt->vm,
+                                                 last, size - last);
                if (err) {
                        if (err == -ENOMEM) {
                                pr_info("[2] Ran out of memory for va_range [%llx + %llx] [bit %d]\n",
@@ -188,7 +188,7 @@ static int igt_ppgtt_alloc(void *arg)
        }
 
 err_ppgtt_cleanup:
-       ppgtt->base.cleanup(&ppgtt->base);
+       ppgtt->vm.cleanup(&ppgtt->vm);
 err_ppgtt:
        mutex_unlock(&dev_priv->drm.struct_mutex);
        kfree(ppgtt);
@@ -987,12 +987,12 @@ static int exercise_ppgtt(struct drm_i915_private *dev_priv,
                err = PTR_ERR(ppgtt);
                goto out_unlock;
        }
-       GEM_BUG_ON(offset_in_page(ppgtt->base.total));
-       GEM_BUG_ON(ppgtt->base.closed);
+       GEM_BUG_ON(offset_in_page(ppgtt->vm.total));
+       GEM_BUG_ON(ppgtt->vm.closed);
 
-       err = func(dev_priv, &ppgtt->base, 0, ppgtt->base.total, end_time);
+       err = func(dev_priv, &ppgtt->vm, 0, ppgtt->vm.total, end_time);
 
-       i915_ppgtt_close(&ppgtt->base);
+       i915_ppgtt_close(&ppgtt->vm);
        i915_ppgtt_put(ppgtt);
 out_unlock:
        mutex_unlock(&dev_priv->drm.struct_mutex);
@@ -1061,18 +1061,18 @@ static int exercise_ggtt(struct drm_i915_private *i915,
 
        mutex_lock(&i915->drm.struct_mutex);
 restart:
-       list_sort(NULL, &ggtt->base.mm.hole_stack, sort_holes);
-       drm_mm_for_each_hole(node, &ggtt->base.mm, hole_start, hole_end) {
+       list_sort(NULL, &ggtt->vm.mm.hole_stack, sort_holes);
+       drm_mm_for_each_hole(node, &ggtt->vm.mm, hole_start, hole_end) {
                if (hole_start < last)
                        continue;
 
-               if (ggtt->base.mm.color_adjust)
-                       ggtt->base.mm.color_adjust(node, 0,
-                                                  &hole_start, &hole_end);
+               if (ggtt->vm.mm.color_adjust)
+                       ggtt->vm.mm.color_adjust(node, 0,
+                                                &hole_start, &hole_end);
                if (hole_start >= hole_end)
                        continue;
 
-               err = func(i915, &ggtt->base, hole_start, hole_end, end_time);
+               err = func(i915, &ggtt->vm, hole_start, hole_end, end_time);
                if (err)
                        break;
 
@@ -1134,7 +1134,7 @@ static int igt_ggtt_page(void *arg)
                goto out_free;
 
        memset(&tmp, 0, sizeof(tmp));
-       err = drm_mm_insert_node_in_range(&ggtt->base.mm, &tmp,
+       err = drm_mm_insert_node_in_range(&ggtt->vm.mm, &tmp,
                                          count * PAGE_SIZE, 0,
                                          I915_COLOR_UNEVICTABLE,
                                          0, ggtt->mappable_end,
@@ -1147,9 +1147,9 @@ static int igt_ggtt_page(void *arg)
        for (n = 0; n < count; n++) {
                u64 offset = tmp.start + n * PAGE_SIZE;
 
-               ggtt->base.insert_page(&ggtt->base,
-                                      i915_gem_object_get_dma_address(obj, 0),
-                                      offset, I915_CACHE_NONE, 0);
+               ggtt->vm.insert_page(&ggtt->vm,
+                                    i915_gem_object_get_dma_address(obj, 0),
+                                    offset, I915_CACHE_NONE, 0);
        }
 
        order = i915_random_order(count, &prng);
@@ -1188,7 +1188,7 @@ static int igt_ggtt_page(void *arg)
 
        kfree(order);
 out_remove:
-       ggtt->base.clear_range(&ggtt->base, tmp.start, tmp.size);
+       ggtt->vm.clear_range(&ggtt->vm, tmp.start, tmp.size);
        intel_runtime_pm_put(i915);
        drm_mm_remove_node(&tmp);
 out_unpin:
@@ -1229,7 +1229,7 @@ static int exercise_mock(struct drm_i915_private *i915,
        ppgtt = ctx->ppgtt;
        GEM_BUG_ON(!ppgtt);
 
-       err = func(i915, &ppgtt->base, 0, ppgtt->base.total, end_time);
+       err = func(i915, &ppgtt->vm, 0, ppgtt->vm.total, end_time);
 
        mock_context_close(ctx);
        return err;
@@ -1270,7 +1270,7 @@ static int igt_gtt_reserve(void *arg)
 
        /* Start by filling the GGTT */
        for (total = 0;
-            total + 2*I915_GTT_PAGE_SIZE <= i915->ggtt.base.total;
+            total + 2*I915_GTT_PAGE_SIZE <= i915->ggtt.vm.total;
             total += 2*I915_GTT_PAGE_SIZE) {
                struct i915_vma *vma;
 
@@ -1288,20 +1288,20 @@ static int igt_gtt_reserve(void *arg)
 
                list_add(&obj->st_link, &objects);
 
-               vma = i915_vma_instance(obj, &i915->ggtt.base, NULL);
+               vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
                if (IS_ERR(vma)) {
                        err = PTR_ERR(vma);
                        goto out;
                }
 
-               err = i915_gem_gtt_reserve(&i915->ggtt.base, &vma->node,
+               err = i915_gem_gtt_reserve(&i915->ggtt.vm, &vma->node,
                                           obj->base.size,
                                           total,
                                           obj->cache_level,
                                           0);
                if (err) {
                        pr_err("i915_gem_gtt_reserve (pass 1) failed at %llu/%llu with err=%d\n",
-                              total, i915->ggtt.base.total, err);
+                              total, i915->ggtt.vm.total, err);
                        goto out;
                }
                track_vma_bind(vma);
@@ -1319,7 +1319,7 @@ static int igt_gtt_reserve(void *arg)
 
        /* Now we start forcing evictions */
        for (total = I915_GTT_PAGE_SIZE;
-            total + 2*I915_GTT_PAGE_SIZE <= i915->ggtt.base.total;
+            total + 2*I915_GTT_PAGE_SIZE <= i915->ggtt.vm.total;
             total += 2*I915_GTT_PAGE_SIZE) {
                struct i915_vma *vma;
 
@@ -1337,20 +1337,20 @@ static int igt_gtt_reserve(void *arg)
 
                list_add(&obj->st_link, &objects);
 
-               vma = i915_vma_instance(obj, &i915->ggtt.base, NULL);
+               vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
                if (IS_ERR(vma)) {
                        err = PTR_ERR(vma);
                        goto out;
                }
 
-               err = i915_gem_gtt_reserve(&i915->ggtt.base, &vma->node,
+               err = i915_gem_gtt_reserve(&i915->ggtt.vm, &vma->node,
                                           obj->base.size,
                                           total,
                                           obj->cache_level,
                                           0);
                if (err) {
                        pr_err("i915_gem_gtt_reserve (pass 2) failed at %llu/%llu with err=%d\n",
-                              total, i915->ggtt.base.total, err);
+                              total, i915->ggtt.vm.total, err);
                        goto out;
                }
                track_vma_bind(vma);
@@ -1371,7 +1371,7 @@ static int igt_gtt_reserve(void *arg)
                struct i915_vma *vma;
                u64 offset;
 
-               vma = i915_vma_instance(obj, &i915->ggtt.base, NULL);
+               vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
                if (IS_ERR(vma)) {
                        err = PTR_ERR(vma);
                        goto out;
@@ -1383,18 +1383,18 @@ static int igt_gtt_reserve(void *arg)
                        goto out;
                }
 
-               offset = random_offset(0, i915->ggtt.base.total,
+               offset = random_offset(0, i915->ggtt.vm.total,
                                       2*I915_GTT_PAGE_SIZE,
                                       I915_GTT_MIN_ALIGNMENT);
 
-               err = i915_gem_gtt_reserve(&i915->ggtt.base, &vma->node,
+               err = i915_gem_gtt_reserve(&i915->ggtt.vm, &vma->node,
                                           obj->base.size,
                                           offset,
                                           obj->cache_level,
                                           0);
                if (err) {
                        pr_err("i915_gem_gtt_reserve (pass 3) failed at %llu/%llu with err=%d\n",
-                              total, i915->ggtt.base.total, err);
+                              total, i915->ggtt.vm.total, err);
                        goto out;
                }
                track_vma_bind(vma);
@@ -1429,8 +1429,8 @@ static int igt_gtt_insert(void *arg)
                u64 start, end;
        } invalid_insert[] = {
                {
-                       i915->ggtt.base.total + I915_GTT_PAGE_SIZE, 0,
-                       0, i915->ggtt.base.total,
+                       i915->ggtt.vm.total + I915_GTT_PAGE_SIZE, 0,
+                       0, i915->ggtt.vm.total,
                },
                {
                        2*I915_GTT_PAGE_SIZE, 0,
@@ -1460,7 +1460,7 @@ static int igt_gtt_insert(void *arg)
 
        /* Check a couple of obviously invalid requests */
        for (ii = invalid_insert; ii->size; ii++) {
-               err = i915_gem_gtt_insert(&i915->ggtt.base, &tmp,
+               err = i915_gem_gtt_insert(&i915->ggtt.vm, &tmp,
                                          ii->size, ii->alignment,
                                          I915_COLOR_UNEVICTABLE,
                                          ii->start, ii->end,
@@ -1475,7 +1475,7 @@ static int igt_gtt_insert(void *arg)
 
        /* Start by filling the GGTT */
        for (total = 0;
-            total + I915_GTT_PAGE_SIZE <= i915->ggtt.base.total;
+            total + I915_GTT_PAGE_SIZE <= i915->ggtt.vm.total;
             total += I915_GTT_PAGE_SIZE) {
                struct i915_vma *vma;
 
@@ -1493,15 +1493,15 @@ static int igt_gtt_insert(void *arg)
 
                list_add(&obj->st_link, &objects);
 
-               vma = i915_vma_instance(obj, &i915->ggtt.base, NULL);
+               vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
                if (IS_ERR(vma)) {
                        err = PTR_ERR(vma);
                        goto out;
                }
 
-               err = i915_gem_gtt_insert(&i915->ggtt.base, &vma->node,
+               err = i915_gem_gtt_insert(&i915->ggtt.vm, &vma->node,
                                          obj->base.size, 0, obj->cache_level,
-                                         0, i915->ggtt.base.total,
+                                         0, i915->ggtt.vm.total,
                                          0);
                if (err == -ENOSPC) {
                        /* maxed out the GGTT space */
@@ -1510,7 +1510,7 @@ static int igt_gtt_insert(void *arg)
                }
                if (err) {
                        pr_err("i915_gem_gtt_insert (pass 1) failed at %llu/%llu with err=%d\n",
-                              total, i915->ggtt.base.total, err);
+                              total, i915->ggtt.vm.total, err);
                        goto out;
                }
                track_vma_bind(vma);
@@ -1522,7 +1522,7 @@ static int igt_gtt_insert(void *arg)
        list_for_each_entry(obj, &objects, st_link) {
                struct i915_vma *vma;
 
-               vma = i915_vma_instance(obj, &i915->ggtt.base, NULL);
+               vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
                if (IS_ERR(vma)) {
                        err = PTR_ERR(vma);
                        goto out;
@@ -1542,7 +1542,7 @@ static int igt_gtt_insert(void *arg)
                struct i915_vma *vma;
                u64 offset;
 
-               vma = i915_vma_instance(obj, &i915->ggtt.base, NULL);
+               vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
                if (IS_ERR(vma)) {
                        err = PTR_ERR(vma);
                        goto out;
@@ -1557,13 +1557,13 @@ static int igt_gtt_insert(void *arg)
                        goto out;
                }
 
-               err = i915_gem_gtt_insert(&i915->ggtt.base, &vma->node,
+               err = i915_gem_gtt_insert(&i915->ggtt.vm, &vma->node,
                                          obj->base.size, 0, obj->cache_level,
-                                         0, i915->ggtt.base.total,
+                                         0, i915->ggtt.vm.total,
                                          0);
                if (err) {
                        pr_err("i915_gem_gtt_insert (pass 2) failed at %llu/%llu with err=%d\n",
-                              total, i915->ggtt.base.total, err);
+                              total, i915->ggtt.vm.total, err);
                        goto out;
                }
                track_vma_bind(vma);
@@ -1579,7 +1579,7 @@ static int igt_gtt_insert(void *arg)
 
        /* And then force evictions */
        for (total = 0;
-            total + 2*I915_GTT_PAGE_SIZE <= i915->ggtt.base.total;
+            total + 2*I915_GTT_PAGE_SIZE <= i915->ggtt.vm.total;
             total += 2*I915_GTT_PAGE_SIZE) {
                struct i915_vma *vma;
 
@@ -1597,19 +1597,19 @@ static int igt_gtt_insert(void *arg)
 
                list_add(&obj->st_link, &objects);
 
-               vma = i915_vma_instance(obj, &i915->ggtt.base, NULL);
+               vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
                if (IS_ERR(vma)) {
                        err = PTR_ERR(vma);
                        goto out;
                }
 
-               err = i915_gem_gtt_insert(&i915->ggtt.base, &vma->node,
+               err = i915_gem_gtt_insert(&i915->ggtt.vm, &vma->node,
                                          obj->base.size, 0, obj->cache_level,
-                                         0, i915->ggtt.base.total,
+                                         0, i915->ggtt.vm.total,
                                          0);
                if (err) {
                        pr_err("i915_gem_gtt_insert (pass 3) failed at %llu/%llu with err=%d\n",
-                              total, i915->ggtt.base.total, err);
+                              total, i915->ggtt.vm.total, err);
                        goto out;
                }
                track_vma_bind(vma);
@@ -1669,7 +1669,7 @@ int i915_gem_gtt_live_selftests(struct drm_i915_private *i915)
                SUBTEST(igt_ggtt_page),
        };
 
-       GEM_BUG_ON(offset_in_page(i915->ggtt.base.total));
+       GEM_BUG_ON(offset_in_page(i915->ggtt.vm.total));
 
        return i915_subtests(tests, i915);
 }
index fbdb2419d418cbc385d7ca1a11e09f543d41e81c..2b2dde94526f70fdfa399c06f31f4ac93b6c9dd0 100644 (file)
@@ -113,7 +113,7 @@ static int igt_gem_huge(void *arg)
 
        obj = huge_gem_object(i915,
                              nreal * PAGE_SIZE,
-                             i915->ggtt.base.total + PAGE_SIZE);
+                             i915->ggtt.vm.total + PAGE_SIZE);
        if (IS_ERR(obj))
                return PTR_ERR(obj);
 
@@ -311,7 +311,7 @@ static int igt_partial_tiling(void *arg)
 
        obj = huge_gem_object(i915,
                              nreal << PAGE_SHIFT,
-                             (1 + next_prime_number(i915->ggtt.base.total >> PAGE_SHIFT)) << PAGE_SHIFT);
+                             (1 + next_prime_number(i915->ggtt.vm.total >> PAGE_SHIFT)) << PAGE_SHIFT);
        if (IS_ERR(obj))
                return PTR_ERR(obj);
 
@@ -440,7 +440,7 @@ static int make_obj_busy(struct drm_i915_gem_object *obj)
        struct i915_vma *vma;
        int err;
 
-       vma = i915_vma_instance(obj, &i915->ggtt.base, NULL);
+       vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
        if (IS_ERR(vma))
                return PTR_ERR(vma);
 
index 94bc2e1898a4df70c622a8a41f133fdc676272fa..a3a89aadeccb0563ec2dbfd49e0bc2517bc17f02 100644 (file)
@@ -430,7 +430,7 @@ static struct i915_vma *empty_batch(struct drm_i915_private *i915)
        if (err)
                goto err;
 
-       vma = i915_vma_instance(obj, &i915->ggtt.base, NULL);
+       vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
        if (IS_ERR(vma)) {
                err = PTR_ERR(vma);
                goto err;
@@ -555,7 +555,8 @@ out_unlock:
 static struct i915_vma *recursive_batch(struct drm_i915_private *i915)
 {
        struct i915_gem_context *ctx = i915->kernel_context;
-       struct i915_address_space *vm = ctx->ppgtt ? &ctx->ppgtt->base : &i915->ggtt.base;
+       struct i915_address_space *vm =
+               ctx->ppgtt ? &ctx->ppgtt->vm : &i915->ggtt.vm;
        struct drm_i915_gem_object *obj;
        const int gen = INTEL_GEN(i915);
        struct i915_vma *vma;
index e90f97236e505ee9bec04a2b61c8a6bcc46cc6ab..8400a8cc5cf20d49d7025e2274949673ffd423a7 100644 (file)
@@ -35,7 +35,7 @@ static bool assert_vma(struct i915_vma *vma,
 {
        bool ok = true;
 
-       if (vma->vm != &ctx->ppgtt->base) {
+       if (vma->vm != &ctx->ppgtt->vm) {
                pr_err("VMA created with wrong VM\n");
                ok = false;
        }
@@ -110,8 +110,7 @@ static int create_vmas(struct drm_i915_private *i915,
        list_for_each_entry(obj, objects, st_link) {
                for (pinned = 0; pinned <= 1; pinned++) {
                        list_for_each_entry(ctx, contexts, link) {
-                               struct i915_address_space *vm =
-                                       &ctx->ppgtt->base;
+                               struct i915_address_space *vm = &ctx->ppgtt->vm;
                                struct i915_vma *vma;
                                int err;
 
@@ -259,12 +258,12 @@ static int igt_vma_pin1(void *arg)
                VALID(0, PIN_GLOBAL | PIN_OFFSET_BIAS | 8192),
                VALID(0, PIN_GLOBAL | PIN_OFFSET_BIAS | (i915->ggtt.mappable_end - 4096)),
                VALID(0, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_BIAS | (i915->ggtt.mappable_end - 4096)),
-               VALID(0, PIN_GLOBAL | PIN_OFFSET_BIAS | (i915->ggtt.base.total - 4096)),
+               VALID(0, PIN_GLOBAL | PIN_OFFSET_BIAS | (i915->ggtt.vm.total - 4096)),
 
                VALID(0, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_FIXED | (i915->ggtt.mappable_end - 4096)),
                INVALID(0, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_FIXED | i915->ggtt.mappable_end),
-               VALID(0, PIN_GLOBAL | PIN_OFFSET_FIXED | (i915->ggtt.base.total - 4096)),
-               INVALID(0, PIN_GLOBAL | PIN_OFFSET_FIXED | i915->ggtt.base.total),
+               VALID(0, PIN_GLOBAL | PIN_OFFSET_FIXED | (i915->ggtt.vm.total - 4096)),
+               INVALID(0, PIN_GLOBAL | PIN_OFFSET_FIXED | i915->ggtt.vm.total),
                INVALID(0, PIN_GLOBAL | PIN_OFFSET_FIXED | round_down(U64_MAX, PAGE_SIZE)),
 
                VALID(4096, PIN_GLOBAL),
@@ -272,12 +271,12 @@ static int igt_vma_pin1(void *arg)
                VALID(i915->ggtt.mappable_end - 4096, PIN_GLOBAL | PIN_MAPPABLE),
                VALID(i915->ggtt.mappable_end, PIN_GLOBAL | PIN_MAPPABLE),
                NOSPACE(i915->ggtt.mappable_end + 4096, PIN_GLOBAL | PIN_MAPPABLE),
-               VALID(i915->ggtt.base.total - 4096, PIN_GLOBAL),
-               VALID(i915->ggtt.base.total, PIN_GLOBAL),
-               NOSPACE(i915->ggtt.base.total + 4096, PIN_GLOBAL),
+               VALID(i915->ggtt.vm.total - 4096, PIN_GLOBAL),
+               VALID(i915->ggtt.vm.total, PIN_GLOBAL),
+               NOSPACE(i915->ggtt.vm.total + 4096, PIN_GLOBAL),
                NOSPACE(round_down(U64_MAX, PAGE_SIZE), PIN_GLOBAL),
                INVALID(8192, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_FIXED | (i915->ggtt.mappable_end - 4096)),
-               INVALID(8192, PIN_GLOBAL | PIN_OFFSET_FIXED | (i915->ggtt.base.total - 4096)),
+               INVALID(8192, PIN_GLOBAL | PIN_OFFSET_FIXED | (i915->ggtt.vm.total - 4096)),
                INVALID(8192, PIN_GLOBAL | PIN_OFFSET_FIXED | (round_down(U64_MAX, PAGE_SIZE) - 4096)),
 
                VALID(8192, PIN_GLOBAL | PIN_OFFSET_BIAS | (i915->ggtt.mappable_end - 4096)),
@@ -289,9 +288,9 @@ static int igt_vma_pin1(void *arg)
                 * variable start, end and size.
                 */
                NOSPACE(0, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_BIAS | i915->ggtt.mappable_end),
-               NOSPACE(0, PIN_GLOBAL | PIN_OFFSET_BIAS | i915->ggtt.base.total),
+               NOSPACE(0, PIN_GLOBAL | PIN_OFFSET_BIAS | i915->ggtt.vm.total),
                NOSPACE(8192, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_BIAS | (i915->ggtt.mappable_end - 4096)),
-               NOSPACE(8192, PIN_GLOBAL | PIN_OFFSET_BIAS | (i915->ggtt.base.total - 4096)),
+               NOSPACE(8192, PIN_GLOBAL | PIN_OFFSET_BIAS | (i915->ggtt.vm.total - 4096)),
 #endif
                { },
 #undef NOSPACE
@@ -307,13 +306,13 @@ static int igt_vma_pin1(void *arg)
         * focusing on error handling of boundary conditions.
         */
 
-       GEM_BUG_ON(!drm_mm_clean(&i915->ggtt.base.mm));
+       GEM_BUG_ON(!drm_mm_clean(&i915->ggtt.vm.mm));
 
        obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
        if (IS_ERR(obj))
                return PTR_ERR(obj);
 
-       vma = checked_vma_instance(obj, &i915->ggtt.base, NULL);
+       vma = checked_vma_instance(obj, &i915->ggtt.vm, NULL);
        if (IS_ERR(vma))
                goto out;
 
@@ -405,7 +404,7 @@ static unsigned int rotated_size(const struct intel_rotation_plane_info *a,
 static int igt_vma_rotate(void *arg)
 {
        struct drm_i915_private *i915 = arg;
-       struct i915_address_space *vm = &i915->ggtt.base;
+       struct i915_address_space *vm = &i915->ggtt.vm;
        struct drm_i915_gem_object *obj;
        const struct intel_rotation_plane_info planes[] = {
                { .width = 1, .height = 1, .stride = 1 },
@@ -604,7 +603,7 @@ static bool assert_pin(struct i915_vma *vma,
 static int igt_vma_partial(void *arg)
 {
        struct drm_i915_private *i915 = arg;
-       struct i915_address_space *vm = &i915->ggtt.base;
+       struct i915_address_space *vm = &i915->ggtt.vm;
        const unsigned int npages = 1021; /* prime! */
        struct drm_i915_gem_object *obj;
        const struct phase {
index 2091e3a6a5be8d1dafd744ccdac874e7b738ec00..390a157b37c3f5e1f52f9e6e859e71e1e574e97e 100644 (file)
@@ -107,8 +107,8 @@ static int emit_recurse_batch(struct hang *h,
        struct drm_i915_private *i915 = h->i915;
        struct i915_address_space *vm =
                rq->gem_context->ppgtt ?
-               &rq->gem_context->ppgtt->base :
-               &i915->ggtt.base;
+               &rq->gem_context->ppgtt->vm :
+               &i915->ggtt.vm;
        struct i915_vma *hws, *vma;
        unsigned int flags;
        u32 *batch;
index 68cb9126b3e14ff007606848ad4f40d711657fd1..0b6da08c8caec6c53f3060a55d4a7ec4d1715dfd 100644 (file)
@@ -83,7 +83,7 @@ static int emit_recurse_batch(struct spinner *spin,
                              struct i915_request *rq,
                              u32 arbitration_command)
 {
-       struct i915_address_space *vm = &rq->gem_context->ppgtt->base;
+       struct i915_address_space *vm = &rq->gem_context->ppgtt->vm;
        struct i915_vma *hws, *vma;
        u32 *batch;
        int err;
index 17444a3abbb933d396ba4e98f74af02e503370a4..f1cfb0fb6bea42f75ad09c03d7a771a3f1e50b52 100644 (file)
@@ -33,7 +33,7 @@ read_nonprivs(struct i915_gem_context *ctx, struct intel_engine_cs *engine)
        memset(cs, 0xc5, PAGE_SIZE);
        i915_gem_object_unpin_map(result);
 
-       vma = i915_vma_instance(result, &engine->i915->ggtt.base, NULL);
+       vma = i915_vma_instance(result, &engine->i915->ggtt.vm, NULL);
        if (IS_ERR(vma)) {
                err = PTR_ERR(vma);
                goto err_obj;
index 36c112088940585c8bcf0bcdb66027f3ec9e61da..556c546f2715ba7903bf85e55257f1c2f02c2198 100644 (file)
@@ -66,25 +66,25 @@ mock_ppgtt(struct drm_i915_private *i915,
                return NULL;
 
        kref_init(&ppgtt->ref);
-       ppgtt->base.i915 = i915;
-       ppgtt->base.total = round_down(U64_MAX, PAGE_SIZE);
-       ppgtt->base.file = ERR_PTR(-ENODEV);
-
-       INIT_LIST_HEAD(&ppgtt->base.active_list);
-       INIT_LIST_HEAD(&ppgtt->base.inactive_list);
-       INIT_LIST_HEAD(&ppgtt->base.unbound_list);
-
-       INIT_LIST_HEAD(&ppgtt->base.global_link);
-       drm_mm_init(&ppgtt->base.mm, 0, ppgtt->base.total);
-
-       ppgtt->base.clear_range = nop_clear_range;
-       ppgtt->base.insert_page = mock_insert_page;
-       ppgtt->base.insert_entries = mock_insert_entries;
-       ppgtt->base.bind_vma = mock_bind_ppgtt;
-       ppgtt->base.unbind_vma = mock_unbind_ppgtt;
-       ppgtt->base.set_pages = ppgtt_set_pages;
-       ppgtt->base.clear_pages = clear_pages;
-       ppgtt->base.cleanup = mock_cleanup;
+       ppgtt->vm.i915 = i915;
+       ppgtt->vm.total = round_down(U64_MAX, PAGE_SIZE);
+       ppgtt->vm.file = ERR_PTR(-ENODEV);
+
+       INIT_LIST_HEAD(&ppgtt->vm.active_list);
+       INIT_LIST_HEAD(&ppgtt->vm.inactive_list);
+       INIT_LIST_HEAD(&ppgtt->vm.unbound_list);
+
+       INIT_LIST_HEAD(&ppgtt->vm.global_link);
+       drm_mm_init(&ppgtt->vm.mm, 0, ppgtt->vm.total);
+
+       ppgtt->vm.clear_range = nop_clear_range;
+       ppgtt->vm.insert_page = mock_insert_page;
+       ppgtt->vm.insert_entries = mock_insert_entries;
+       ppgtt->vm.bind_vma = mock_bind_ppgtt;
+       ppgtt->vm.unbind_vma = mock_unbind_ppgtt;
+       ppgtt->vm.set_pages = ppgtt_set_pages;
+       ppgtt->vm.clear_pages = clear_pages;
+       ppgtt->vm.cleanup = mock_cleanup;
 
        return ppgtt;
 }
@@ -107,27 +107,27 @@ void mock_init_ggtt(struct drm_i915_private *i915)
 
        INIT_LIST_HEAD(&i915->vm_list);
 
-       ggtt->base.i915 = i915;
+       ggtt->vm.i915 = i915;
 
        ggtt->gmadr = (struct resource) DEFINE_RES_MEM(0, 2048 * PAGE_SIZE);
        ggtt->mappable_end = resource_size(&ggtt->gmadr);
-       ggtt->base.total = 4096 * PAGE_SIZE;
-
-       ggtt->base.clear_range = nop_clear_range;
-       ggtt->base.insert_page = mock_insert_page;
-       ggtt->base.insert_entries = mock_insert_entries;
-       ggtt->base.bind_vma = mock_bind_ggtt;
-       ggtt->base.unbind_vma = mock_unbind_ggtt;
-       ggtt->base.set_pages = ggtt_set_pages;
-       ggtt->base.clear_pages = clear_pages;
-       ggtt->base.cleanup = mock_cleanup;
-
-       i915_address_space_init(&ggtt->base, i915, "global");
+       ggtt->vm.total = 4096 * PAGE_SIZE;
+
+       ggtt->vm.clear_range = nop_clear_range;
+       ggtt->vm.insert_page = mock_insert_page;
+       ggtt->vm.insert_entries = mock_insert_entries;
+       ggtt->vm.bind_vma = mock_bind_ggtt;
+       ggtt->vm.unbind_vma = mock_unbind_ggtt;
+       ggtt->vm.set_pages = ggtt_set_pages;
+       ggtt->vm.clear_pages = clear_pages;
+       ggtt->vm.cleanup = mock_cleanup;
+
+       i915_address_space_init(&ggtt->vm, i915, "global");
 }
 
 void mock_fini_ggtt(struct drm_i915_private *i915)
 {
        struct i915_ggtt *ggtt = &i915->ggtt;
 
-       i915_address_space_fini(&ggtt->base);
+       i915_address_space_fini(&ggtt->vm);
 }