return 0;
}
-static int query_gts(struct xe_device *xe, struct drm_xe_device_query *query)
+static int query_gt_list(struct xe_device *xe, struct drm_xe_device_query *query)
{
struct xe_gt *gt;
- size_t size = sizeof(struct drm_xe_query_gts) +
+ size_t size = sizeof(struct drm_xe_query_gt_list) +
xe->info.gt_count * sizeof(struct drm_xe_query_gt);
- struct drm_xe_query_gts __user *query_ptr =
+ struct drm_xe_query_gt_list __user *query_ptr =
u64_to_user_ptr(query->data);
- struct drm_xe_query_gts *gts;
+ struct drm_xe_query_gt_list *gt_list;
u8 id;
if (query->size == 0) {
return -EINVAL;
}
- gts = kzalloc(size, GFP_KERNEL);
- if (!gts)
+ gt_list = kzalloc(size, GFP_KERNEL);
+ if (!gt_list)
return -ENOMEM;
- gts->num_gt = xe->info.gt_count;
+ gt_list->num_gt = xe->info.gt_count;
for_each_gt(gt, xe, id) {
if (xe_gt_is_media_type(gt))
- gts->gts[id].type = XE_QUERY_GT_TYPE_MEDIA;
+ gt_list->gt_list[id].type = XE_QUERY_GT_TYPE_MEDIA;
else if (gt_to_tile(gt)->id > 0)
- gts->gts[id].type = XE_QUERY_GT_TYPE_REMOTE;
+ gt_list->gt_list[id].type = XE_QUERY_GT_TYPE_REMOTE;
else
- gts->gts[id].type = XE_QUERY_GT_TYPE_MAIN;
- gts->gts[id].gt_id = gt->info.id;
- gts->gts[id].clock_freq = gt->info.clock_freq;
+ gt_list->gt_list[id].type = XE_QUERY_GT_TYPE_MAIN;
+ gt_list->gt_list[id].gt_id = gt->info.id;
+ gt_list->gt_list[id].clock_freq = gt->info.clock_freq;
if (!IS_DGFX(xe))
- gts->gts[id].native_mem_regions = 0x1;
+ gt_list->gt_list[id].native_mem_regions = 0x1;
else
- gts->gts[id].native_mem_regions =
+ gt_list->gt_list[id].native_mem_regions =
BIT(gt_to_tile(gt)->id) << 1;
- gts->gts[id].slow_mem_regions = xe->info.mem_region_mask ^
- gts->gts[id].native_mem_regions;
+ gt_list->gt_list[id].slow_mem_regions = xe->info.mem_region_mask ^
+ gt_list->gt_list[id].native_mem_regions;
}
- if (copy_to_user(query_ptr, gts, size)) {
- kfree(gts);
+ if (copy_to_user(query_ptr, gt_list, size)) {
+ kfree(gt_list);
return -EFAULT;
}
- kfree(gts);
+ kfree(gt_list);
return 0;
}
query_engines,
query_memory_usage,
query_config,
- query_gts,
+ query_gt_list,
query_hwconfig,
query_gt_topology,
query_engine_cycles,
/**
* struct drm_xe_query_gt - describe an individual GT.
*
- * To be used with drm_xe_query_gts, which will return a list with all the
+ * To be used with drm_xe_query_gt_list, which will return a list with all the
* existing GT individual descriptions.
* Graphics Technology (GT) is a subset of a GPU/tile that is responsible for
* implementing graphics and/or media operations.
};
/**
- * struct drm_xe_query_gts - describe GTs
+ * struct drm_xe_query_gt_list - A list with GT description items.
*
* If a query is made with a struct drm_xe_device_query where .query
- * is equal to DRM_XE_DEVICE_QUERY_GTS, then the reply uses struct
- * drm_xe_query_gts in .data.
+ * is equal to DRM_XE_DEVICE_QUERY_GT_LIST, then the reply uses struct
+ * drm_xe_query_gt_list in .data.
*/
-struct drm_xe_query_gts {
- /** @num_gt: number of GTs returned in gts */
+struct drm_xe_query_gt_list {
+ /** @num_gt: number of GT items returned in gt_list */
__u32 num_gt;
/** @pad: MBZ */
__u32 pad;
- /** @gts: The GT list returned for this device */
- struct drm_xe_query_gt gts[];
+ /** @gt_list: The GT list returned for this device */
+ struct drm_xe_query_gt gt_list[];
};
/**
#define DRM_XE_DEVICE_QUERY_ENGINES 0
#define DRM_XE_DEVICE_QUERY_MEM_USAGE 1
#define DRM_XE_DEVICE_QUERY_CONFIG 2
-#define DRM_XE_DEVICE_QUERY_GTS 3
+#define DRM_XE_DEVICE_QUERY_GT_LIST 3
#define DRM_XE_DEVICE_QUERY_HWCONFIG 4
#define DRM_XE_DEVICE_QUERY_GT_TOPOLOGY 5
#define DRM_XE_DEVICE_QUERY_ENGINE_CYCLES 6