1 // SPDX-License-Identifier: MIT
3 * Copyright © 2019 Intel Corporation
8 #include "intel_gt_pm.h"
9 #include "intel_gt_requests.h"
10 #include "intel_mocs.h"
11 #include "intel_rc6.h"
12 #include "intel_rps.h"
13 #include "intel_uncore.h"
16 void intel_gt_init_early(struct intel_gt *gt, struct drm_i915_private *i915)
19 gt->uncore = &i915->uncore;
21 spin_lock_init(>->irq_lock);
23 INIT_LIST_HEAD(>->closed_vma);
24 spin_lock_init(>->closed_lock);
26 intel_gt_init_reset(gt);
27 intel_gt_init_requests(gt);
28 intel_gt_init_timelines(gt);
29 intel_gt_pm_init_early(gt);
31 intel_rps_init_early(>->rps);
32 intel_uc_init_early(>->uc);
35 void intel_gt_init_hw_early(struct intel_gt *gt, struct i915_ggtt *ggtt)
39 intel_gt_sanitize(gt, false);
42 static void init_unused_ring(struct intel_gt *gt, u32 base)
44 struct intel_uncore *uncore = gt->uncore;
46 intel_uncore_write(uncore, RING_CTL(base), 0);
47 intel_uncore_write(uncore, RING_HEAD(base), 0);
48 intel_uncore_write(uncore, RING_TAIL(base), 0);
49 intel_uncore_write(uncore, RING_START(base), 0);
52 static void init_unused_rings(struct intel_gt *gt)
54 struct drm_i915_private *i915 = gt->i915;
57 init_unused_ring(gt, PRB1_BASE);
58 init_unused_ring(gt, SRB0_BASE);
59 init_unused_ring(gt, SRB1_BASE);
60 init_unused_ring(gt, SRB2_BASE);
61 init_unused_ring(gt, SRB3_BASE);
62 } else if (IS_GEN(i915, 2)) {
63 init_unused_ring(gt, SRB0_BASE);
64 init_unused_ring(gt, SRB1_BASE);
65 } else if (IS_GEN(i915, 3)) {
66 init_unused_ring(gt, PRB1_BASE);
67 init_unused_ring(gt, PRB2_BASE);
71 int intel_gt_init_hw(struct intel_gt *gt)
73 struct drm_i915_private *i915 = gt->i915;
74 struct intel_uncore *uncore = gt->uncore;
77 BUG_ON(!i915->kernel_context);
78 ret = intel_gt_terminally_wedged(gt);
82 gt->last_init_time = ktime_get();
84 /* Double layer security blanket, see i915_gem_init() */
85 intel_uncore_forcewake_get(uncore, FORCEWAKE_ALL);
87 if (HAS_EDRAM(i915) && INTEL_GEN(i915) < 9)
88 intel_uncore_rmw(uncore, HSW_IDICR, 0, IDIHASHMSK(0xf));
91 intel_uncore_write(uncore,
92 MI_PREDICATE_RESULT_2,
94 LOWER_SLICE_ENABLED : LOWER_SLICE_DISABLED);
96 /* Apply the GT workarounds... */
97 intel_gt_apply_workarounds(gt);
98 /* ...and determine whether they are sticking. */
99 intel_gt_verify_workarounds(gt, "init");
101 intel_gt_init_swizzling(gt);
104 * At least 830 can leave some of the unused rings
105 * "active" (ie. head != tail) after resume which
106 * will prevent c3 entry. Makes sure all unused rings
109 init_unused_rings(gt);
111 ret = i915_ppgtt_init_hw(gt);
113 DRM_ERROR("Enabling PPGTT failed (%d)\n", ret);
117 /* We can't enable contexts until all firmware is loaded */
118 ret = intel_uc_init_hw(>->uc);
120 i915_probe_error(i915, "Enabling uc failed (%d)\n", ret);
127 intel_uncore_forcewake_put(uncore, FORCEWAKE_ALL);
131 static void rmw_set(struct intel_uncore *uncore, i915_reg_t reg, u32 set)
133 intel_uncore_rmw(uncore, reg, 0, set);
136 static void rmw_clear(struct intel_uncore *uncore, i915_reg_t reg, u32 clr)
138 intel_uncore_rmw(uncore, reg, clr, 0);
141 static void clear_register(struct intel_uncore *uncore, i915_reg_t reg)
143 intel_uncore_rmw(uncore, reg, 0, 0);
146 static void gen8_clear_engine_error_register(struct intel_engine_cs *engine)
148 GEN6_RING_FAULT_REG_RMW(engine, RING_FAULT_VALID, 0);
149 GEN6_RING_FAULT_REG_POSTING_READ(engine);
153 intel_gt_clear_error_registers(struct intel_gt *gt,
154 intel_engine_mask_t engine_mask)
156 struct drm_i915_private *i915 = gt->i915;
157 struct intel_uncore *uncore = gt->uncore;
160 if (!IS_GEN(i915, 2))
161 clear_register(uncore, PGTBL_ER);
163 if (INTEL_GEN(i915) < 4)
164 clear_register(uncore, IPEIR(RENDER_RING_BASE));
166 clear_register(uncore, IPEIR_I965);
168 clear_register(uncore, EIR);
169 eir = intel_uncore_read(uncore, EIR);
172 * some errors might have become stuck,
175 DRM_DEBUG_DRIVER("EIR stuck: 0x%08x, masking\n", eir);
176 rmw_set(uncore, EMR, eir);
177 intel_uncore_write(uncore, GEN2_IIR,
178 I915_MASTER_ERROR_INTERRUPT);
181 if (INTEL_GEN(i915) >= 12) {
182 rmw_clear(uncore, GEN12_RING_FAULT_REG, RING_FAULT_VALID);
183 intel_uncore_posting_read(uncore, GEN12_RING_FAULT_REG);
184 } else if (INTEL_GEN(i915) >= 8) {
185 rmw_clear(uncore, GEN8_RING_FAULT_REG, RING_FAULT_VALID);
186 intel_uncore_posting_read(uncore, GEN8_RING_FAULT_REG);
187 } else if (INTEL_GEN(i915) >= 6) {
188 struct intel_engine_cs *engine;
189 enum intel_engine_id id;
191 for_each_engine_masked(engine, gt, engine_mask, id)
192 gen8_clear_engine_error_register(engine);
196 static void gen6_check_faults(struct intel_gt *gt)
198 struct intel_engine_cs *engine;
199 enum intel_engine_id id;
202 for_each_engine(engine, gt, id) {
203 fault = GEN6_RING_FAULT_REG_READ(engine);
204 if (fault & RING_FAULT_VALID) {
205 DRM_DEBUG_DRIVER("Unexpected fault\n"
207 "\tAddress space: %s\n"
211 fault & RING_FAULT_GTTSEL_MASK ?
213 RING_FAULT_SRCID(fault),
214 RING_FAULT_FAULT_TYPE(fault));
219 static void gen8_check_faults(struct intel_gt *gt)
221 struct intel_uncore *uncore = gt->uncore;
222 i915_reg_t fault_reg, fault_data0_reg, fault_data1_reg;
225 if (INTEL_GEN(gt->i915) >= 12) {
226 fault_reg = GEN12_RING_FAULT_REG;
227 fault_data0_reg = GEN12_FAULT_TLB_DATA0;
228 fault_data1_reg = GEN12_FAULT_TLB_DATA1;
230 fault_reg = GEN8_RING_FAULT_REG;
231 fault_data0_reg = GEN8_FAULT_TLB_DATA0;
232 fault_data1_reg = GEN8_FAULT_TLB_DATA1;
235 fault = intel_uncore_read(uncore, fault_reg);
236 if (fault & RING_FAULT_VALID) {
237 u32 fault_data0, fault_data1;
240 fault_data0 = intel_uncore_read(uncore, fault_data0_reg);
241 fault_data1 = intel_uncore_read(uncore, fault_data1_reg);
243 fault_addr = ((u64)(fault_data1 & FAULT_VA_HIGH_BITS) << 44) |
244 ((u64)fault_data0 << 12);
246 DRM_DEBUG_DRIVER("Unexpected fault\n"
247 "\tAddr: 0x%08x_%08x\n"
248 "\tAddress space: %s\n"
252 upper_32_bits(fault_addr),
253 lower_32_bits(fault_addr),
254 fault_data1 & FAULT_GTT_SEL ? "GGTT" : "PPGTT",
255 GEN8_RING_FAULT_ENGINE_ID(fault),
256 RING_FAULT_SRCID(fault),
257 RING_FAULT_FAULT_TYPE(fault));
261 void intel_gt_check_and_clear_faults(struct intel_gt *gt)
263 struct drm_i915_private *i915 = gt->i915;
265 /* From GEN8 onwards we only have one 'All Engine Fault Register' */
266 if (INTEL_GEN(i915) >= 8)
267 gen8_check_faults(gt);
268 else if (INTEL_GEN(i915) >= 6)
269 gen6_check_faults(gt);
273 intel_gt_clear_error_registers(gt, ALL_ENGINES);
276 void intel_gt_flush_ggtt_writes(struct intel_gt *gt)
278 struct intel_uncore *uncore = gt->uncore;
279 intel_wakeref_t wakeref;
282 * No actual flushing is required for the GTT write domain for reads
283 * from the GTT domain. Writes to it "immediately" go to main memory
284 * as far as we know, so there's no chipset flush. It also doesn't
285 * land in the GPU render cache.
287 * However, we do have to enforce the order so that all writes through
288 * the GTT land before any writes to the device, such as updates to
291 * We also have to wait a bit for the writes to land from the GTT.
292 * An uncached read (i.e. mmio) seems to be ideal for the round-trip
293 * timing. This issue has only been observed when switching quickly
294 * between GTT writes and CPU reads from inside the kernel on recent hw,
295 * and it appears to only affect discrete GTT blocks (i.e. on LLC
296 * system agents we cannot reproduce this behaviour, until Cannonlake
302 if (INTEL_INFO(gt->i915)->has_coherent_ggtt)
305 intel_gt_chipset_flush(gt);
307 with_intel_runtime_pm_if_in_use(uncore->rpm, wakeref) {
310 spin_lock_irqsave(&uncore->lock, flags);
311 intel_uncore_posting_read_fw(uncore,
312 RING_HEAD(RENDER_RING_BASE));
313 spin_unlock_irqrestore(&uncore->lock, flags);
317 void intel_gt_chipset_flush(struct intel_gt *gt)
320 if (INTEL_GEN(gt->i915) < 6)
321 intel_gtt_chipset_flush();
324 void intel_gt_driver_register(struct intel_gt *gt)
326 intel_rps_driver_register(>->rps);
329 static int intel_gt_init_scratch(struct intel_gt *gt, unsigned int size)
331 struct drm_i915_private *i915 = gt->i915;
332 struct drm_i915_gem_object *obj;
333 struct i915_vma *vma;
336 obj = i915_gem_object_create_stolen(i915, size);
338 obj = i915_gem_object_create_internal(i915, size);
340 DRM_ERROR("Failed to allocate scratch page\n");
344 vma = i915_vma_instance(obj, >->ggtt->vm, NULL);
350 ret = i915_vma_pin(vma, 0, 0, PIN_GLOBAL | PIN_HIGH);
354 gt->scratch = i915_vma_make_unshrinkable(vma);
359 i915_gem_object_put(obj);
363 static void intel_gt_fini_scratch(struct intel_gt *gt)
365 i915_vma_unpin_and_release(>->scratch, 0);
368 int intel_gt_init(struct intel_gt *gt)
372 err = intel_gt_init_scratch(gt, IS_GEN(gt->i915, 2) ? SZ_256K : SZ_4K);
376 intel_gt_pm_init(gt);
381 void intel_gt_driver_remove(struct intel_gt *gt)
383 GEM_BUG_ON(gt->awake);
386 void intel_gt_driver_unregister(struct intel_gt *gt)
388 intel_rps_driver_unregister(>->rps);
391 void intel_gt_driver_release(struct intel_gt *gt)
393 intel_gt_pm_fini(gt);
394 intel_gt_fini_scratch(gt);
397 void intel_gt_driver_late_release(struct intel_gt *gt)
399 intel_uc_driver_late_release(>->uc);
400 intel_gt_fini_requests(gt);
401 intel_gt_fini_reset(gt);
402 intel_gt_fini_timelines(gt);