return xe_device_get_root_tile(xe)->primary_gt;
}
-static inline bool xe_device_guc_submission_enabled(struct xe_device *xe)
+static inline bool xe_device_uc_enabled(struct xe_device *xe)
{
return !xe->info.force_execlist;
}
u16 gt_id;
u32 return_mask = 0, prev_mask;
- if (XE_IOCTL_DBG(xe, !xe_device_guc_submission_enabled(xe) &&
+ if (XE_IOCTL_DBG(xe, !xe_device_uc_enabled(xe) &&
len > 1))
return 0;
struct xe_device *xe = gt_to_xe(q->gt);
int err;
- xe_assert(xe, !xe_device_guc_submission_enabled(xe));
+ xe_assert(xe, !xe_device_uc_enabled(xe));
drm_info(&xe->drm, "Enabling execlist submission (GuC submission disabled)\n");
struct xe_device *xe = gt_to_xe(q->gt);
unsigned long flags;
- xe_assert(xe, !xe_device_guc_submission_enabled(xe));
+ xe_assert(xe, !xe_device_uc_enabled(xe));
spin_lock_irqsave(&exl->port->lock, flags);
if (WARN_ON(exl->active_priority != XE_EXEC_QUEUE_PRIORITY_UNSET))
int xe_execlist_init(struct xe_gt *gt)
{
/* GuC submission enabled, nothing to do */
- if (xe_device_guc_submission_enabled(gt_to_xe(gt)))
+ if (xe_device_uc_enabled(gt_to_xe(gt)))
return 0;
gt->exec_queue_ops = &execlist_exec_queue_ops;
xe_gt_assert(gt, seqno > 0);
if (seqno > 0)
xe_gt_tlb_invalidation_wait(gt, seqno);
- } else if (xe_device_guc_submission_enabled(gt_to_xe(gt))) {
+ } else if (xe_device_uc_enabled(gt_to_xe(gt))) {
struct xe_device *xe = gt_to_xe(gt);
if (xe->info.platform == XE_PVC) {
int err;
/* We only support GT resets with GuC submission */
- if (!xe_device_guc_submission_enabled(gt_to_xe(gt)))
+ if (!xe_device_uc_enabled(gt_to_xe(gt)))
return -ENODEV;
xe_gt_info(gt, "reset started\n");
int err;
/* For now suspend/resume is only allowed with GuC */
- if (!xe_device_guc_submission_enabled(gt_to_xe(gt)))
+ if (!xe_device_uc_enabled(gt_to_xe(gt)))
return -ENODEV;
xe_gt_sanitize(gt);
u32 size = PAGE_ALIGN(sizeof(struct slpc_shared_data));
int ret;
- xe_gt_assert(gt, xe_device_guc_submission_enabled(xe));
+ xe_gt_assert(gt, xe_device_uc_enabled(xe));
xe_device_mem_access_get(pc_to_xe(pc));
long timeout;
int err;
- xe_assert(xe, xe_device_guc_submission_enabled(guc_to_xe(guc)));
+ xe_assert(xe, xe_device_uc_enabled(guc_to_xe(guc)));
ge = kzalloc(sizeof(*ge), GFP_KERNEL);
if (!ge)
struct xe_exec_queue *q;
unsigned long index;
- if (!xe_device_guc_submission_enabled(guc_to_xe(guc)))
+ if (!xe_device_uc_enabled(guc_to_xe(guc)))
return;
mutex_lock(&guc->submission_state.lock);
if (err)
goto err_hwsp;
- if (!xe_device_guc_submission_enabled(xe)) {
+ if (!xe_device_uc_enabled(xe)) {
hwe->exl_port = xe_execlist_port_create(xe, hwe);
if (IS_ERR(hwe->exl_port)) {
err = PTR_ERR(hwe->exl_port);
}
}
- if (xe_device_guc_submission_enabled(xe))
+ if (xe_device_uc_enabled(xe))
xe_hw_engine_enable_ring(hwe);
/* We reserve the highest BCS instance for USM */
u32 ccs_mask, bcs_mask;
u32 irqs, dmask, smask;
- if (xe_device_guc_submission_enabled(xe)) {
+ if (xe_device_uc_enabled(xe)) {
irqs = GT_RENDER_USER_INTERRUPT |
GT_RENDER_PIPECTL_NOTIFY_INTERRUPT;
} else {
int ret;
/* GuC submission not enabled, nothing to do */
- if (!xe_device_guc_submission_enabled(uc_to_xe(uc)))
+ if (!xe_device_uc_enabled(uc_to_xe(uc)))
return 0;
ret = xe_guc_init(&uc->guc);
int xe_uc_init_post_hwconfig(struct xe_uc *uc)
{
/* GuC submission not enabled, nothing to do */
- if (!xe_device_guc_submission_enabled(uc_to_xe(uc)))
+ if (!xe_device_uc_enabled(uc_to_xe(uc)))
return 0;
return xe_guc_init_post_hwconfig(&uc->guc);
int ret;
/* GuC submission not enabled, nothing to do */
- if (!xe_device_guc_submission_enabled(uc_to_xe(uc)))
+ if (!xe_device_uc_enabled(uc_to_xe(uc)))
return 0;
ret = xe_guc_min_load_for_hwconfig(&uc->guc);
int ret;
/* GuC submission not enabled, nothing to do */
- if (!xe_device_guc_submission_enabled(uc_to_xe(uc)))
+ if (!xe_device_uc_enabled(uc_to_xe(uc)))
return 0;
ret = xe_uc_sanitize_reset(uc);
int xe_uc_reset_prepare(struct xe_uc *uc)
{
/* GuC submission not enabled, nothing to do */
- if (!xe_device_guc_submission_enabled(uc_to_xe(uc)))
+ if (!xe_device_uc_enabled(uc_to_xe(uc)))
return 0;
return xe_guc_reset_prepare(&uc->guc);
int xe_uc_stop(struct xe_uc *uc)
{
/* GuC submission not enabled, nothing to do */
- if (!xe_device_guc_submission_enabled(uc_to_xe(uc)))
+ if (!xe_device_uc_enabled(uc_to_xe(uc)))
return 0;
return xe_guc_stop(&uc->guc);
int xe_uc_start(struct xe_uc *uc)
{
/* GuC submission not enabled, nothing to do */
- if (!xe_device_guc_submission_enabled(uc_to_xe(uc)))
+ if (!xe_device_uc_enabled(uc_to_xe(uc)))
return 0;
return xe_guc_start(&uc->guc);
int ret;
/* GuC submission not enabled, nothing to do */
- if (!xe_device_guc_submission_enabled(uc_to_xe(uc)))
+ if (!xe_device_uc_enabled(uc_to_xe(uc)))
return 0;
uc_reset_wait(uc);