1 // SPDX-License-Identifier: MIT
2 /*
3  * Copyright © 2019 Intel Corporation
4  */
5 
6 #include "i915_drv.h"
7 #include "intel_memory_region.h"
8 #include "gem/i915_gem_lmem.h"
9 #include "gem/i915_gem_region.h"
10 #include "intel_region_lmem.h"
11 
init_fake_lmem_bar(struct intel_memory_region * mem)12 static int init_fake_lmem_bar(struct intel_memory_region *mem)
13 {
14 	struct drm_i915_private *i915 = mem->i915;
15 	struct i915_ggtt *ggtt = &i915->ggtt;
16 	unsigned long n;
17 	int ret;
18 
19 	/* We want to 1:1 map the mappable aperture to our reserved region */
20 
21 	mem->fake_mappable.start = 0;
22 	mem->fake_mappable.size = resource_size(&mem->region);
23 	mem->fake_mappable.color = I915_COLOR_UNEVICTABLE;
24 
25 	ret = drm_mm_reserve_node(&ggtt->vm.mm, &mem->fake_mappable);
26 	if (ret)
27 		return ret;
28 
29 	mem->remap_addr = dma_map_resource(i915->drm.dev,
30 					   mem->region.start,
31 					   mem->fake_mappable.size,
32 					   PCI_DMA_BIDIRECTIONAL,
33 					   DMA_ATTR_FORCE_CONTIGUOUS);
34 	if (dma_mapping_error(i915->drm.dev, mem->remap_addr)) {
35 		drm_mm_remove_node(&mem->fake_mappable);
36 		return -EINVAL;
37 	}
38 
39 	for (n = 0; n < mem->fake_mappable.size >> PAGE_SHIFT; ++n) {
40 		ggtt->vm.insert_page(&ggtt->vm,
41 				     mem->remap_addr + (n << PAGE_SHIFT),
42 				     n << PAGE_SHIFT,
43 				     I915_CACHE_NONE, 0);
44 	}
45 
46 	mem->region = (struct resource)DEFINE_RES_MEM(mem->remap_addr,
47 						      mem->fake_mappable.size);
48 
49 	return 0;
50 }
51 
release_fake_lmem_bar(struct intel_memory_region * mem)52 static void release_fake_lmem_bar(struct intel_memory_region *mem)
53 {
54 	if (!drm_mm_node_allocated(&mem->fake_mappable))
55 		return;
56 
57 	drm_mm_remove_node(&mem->fake_mappable);
58 
59 	dma_unmap_resource(mem->i915->drm.dev,
60 			   mem->remap_addr,
61 			   mem->fake_mappable.size,
62 			   PCI_DMA_BIDIRECTIONAL,
63 			   DMA_ATTR_FORCE_CONTIGUOUS);
64 }
65 
66 static void
region_lmem_release(struct intel_memory_region * mem)67 region_lmem_release(struct intel_memory_region *mem)
68 {
69 	release_fake_lmem_bar(mem);
70 	io_mapping_fini(&mem->iomap);
71 	intel_memory_region_release_buddy(mem);
72 }
73 
74 static int
region_lmem_init(struct intel_memory_region * mem)75 region_lmem_init(struct intel_memory_region *mem)
76 {
77 	int ret;
78 
79 	if (mem->i915->params.fake_lmem_start) {
80 		ret = init_fake_lmem_bar(mem);
81 		GEM_BUG_ON(ret);
82 	}
83 
84 	if (!io_mapping_init_wc(&mem->iomap,
85 				mem->io_start,
86 				resource_size(&mem->region)))
87 		return -EIO;
88 
89 	ret = intel_memory_region_init_buddy(mem);
90 	if (ret)
91 		io_mapping_fini(&mem->iomap);
92 
93 	return ret;
94 }
95 
96 static const struct intel_memory_region_ops intel_region_lmem_ops = {
97 	.init = region_lmem_init,
98 	.release = region_lmem_release,
99 	.init_object = __i915_gem_lmem_object_init,
100 };
101 
102 struct intel_memory_region *
intel_gt_setup_fake_lmem(struct intel_gt * gt)103 intel_gt_setup_fake_lmem(struct intel_gt *gt)
104 {
105 	struct drm_i915_private *i915 = gt->i915;
106 	struct pci_dev *pdev = to_pci_dev(i915->drm.dev);
107 	struct intel_memory_region *mem;
108 	resource_size_t mappable_end;
109 	resource_size_t io_start;
110 	resource_size_t start;
111 
112 	if (!HAS_LMEM(i915))
113 		return ERR_PTR(-ENODEV);
114 
115 	if (!i915->params.fake_lmem_start)
116 		return ERR_PTR(-ENODEV);
117 
118 	GEM_BUG_ON(i915_ggtt_has_aperture(&i915->ggtt));
119 
120 	/* Your mappable aperture belongs to me now! */
121 	mappable_end = pci_resource_len(pdev, 2);
122 	io_start = pci_resource_start(pdev, 2);
123 	start = i915->params.fake_lmem_start;
124 
125 	mem = intel_memory_region_create(i915,
126 					 start,
127 					 mappable_end,
128 					 PAGE_SIZE,
129 					 io_start,
130 					 &intel_region_lmem_ops);
131 	if (!IS_ERR(mem)) {
132 		drm_info(&i915->drm, "Intel graphics fake LMEM: %pR\n",
133 			 &mem->region);
134 		drm_info(&i915->drm,
135 			 "Intel graphics fake LMEM IO start: %llx\n",
136 			(u64)mem->io_start);
137 		drm_info(&i915->drm, "Intel graphics fake LMEM size: %llx\n",
138 			 (u64)resource_size(&mem->region));
139 	}
140 
141 	return mem;
142 }
143 
get_legacy_lowmem_region(struct intel_uncore * uncore,u64 * start,u32 * size)144 static bool get_legacy_lowmem_region(struct intel_uncore *uncore,
145 				     u64 *start, u32 *size)
146 {
147 	if (!IS_DG1_REVID(uncore->i915, DG1_REVID_A0, DG1_REVID_B0))
148 		return false;
149 
150 	*start = 0;
151 	*size = SZ_1M;
152 
153 	drm_dbg(&uncore->i915->drm, "LMEM: reserved legacy low-memory [0x%llx-0x%llx]\n",
154 		*start, *start + *size);
155 
156 	return true;
157 }
158 
reserve_lowmem_region(struct intel_uncore * uncore,struct intel_memory_region * mem)159 static int reserve_lowmem_region(struct intel_uncore *uncore,
160 				 struct intel_memory_region *mem)
161 {
162 	u64 reserve_start;
163 	u32 reserve_size;
164 	int ret;
165 
166 	if (!get_legacy_lowmem_region(uncore, &reserve_start, &reserve_size))
167 		return 0;
168 
169 	ret = intel_memory_region_reserve(mem, reserve_start, reserve_size);
170 	if (ret)
171 		drm_err(&uncore->i915->drm, "LMEM: reserving low memory region failed\n");
172 
173 	return ret;
174 }
175 
setup_lmem(struct intel_gt * gt)176 static struct intel_memory_region *setup_lmem(struct intel_gt *gt)
177 {
178 	struct drm_i915_private *i915 = gt->i915;
179 	struct intel_uncore *uncore = gt->uncore;
180 	struct pci_dev *pdev = i915->drm.pdev;
181 	struct intel_memory_region *mem;
182 	resource_size_t io_start;
183 	resource_size_t lmem_size;
184 	int err;
185 
186 	if (!IS_DGFX(i915))
187 		return ERR_PTR(-ENODEV);
188 
189 	/* Stolen starts from GSMBASE on DG1 */
190 	lmem_size = intel_uncore_read64(uncore, GEN12_GSMBASE);
191 
192 	io_start = pci_resource_start(pdev, 2);
193 	if (GEM_WARN_ON(lmem_size > pci_resource_len(pdev, 2)))
194 		return ERR_PTR(-ENODEV);
195 
196 	mem = intel_memory_region_create(i915,
197 					 0,
198 					 lmem_size,
199 					 I915_GTT_PAGE_SIZE_4K,
200 					 io_start,
201 					 &intel_region_lmem_ops);
202 	if (IS_ERR(mem))
203 		return mem;
204 
205 	err = reserve_lowmem_region(uncore, mem);
206 	if (err)
207 		goto err_region_put;
208 
209 	drm_dbg(&i915->drm, "Local memory: %pR\n", &mem->region);
210 	drm_dbg(&i915->drm, "Local memory IO start: %pa\n",
211 		&mem->io_start);
212 	drm_info(&i915->drm, "Local memory available: %pa\n",
213 		 &lmem_size);
214 
215 	return mem;
216 
217 err_region_put:
218 	intel_memory_region_put(mem);
219 	return ERR_PTR(err);
220 }
221 
intel_gt_setup_lmem(struct intel_gt * gt)222 struct intel_memory_region *intel_gt_setup_lmem(struct intel_gt *gt)
223 {
224 	return setup_lmem(gt);
225 }
226