Skip to content

Commit

Permalink
drm/xe: Make xe_ggtt_node struct independent
Browse files Browse the repository at this point in the history
In some rare cases, the drm_mm node cannot be removed synchronously
due to runtime PM conditions. In this situation, the node removal will
be delegated to a workqueue that will be able to wake up the device
before removing the node.

However, in this situation, the lifetime of the xe_ggtt_node cannot
be restricted to the lifetime of the parent object. So, this patch
introduces the infrastructure so the xe_ggtt_node struct can be
allocated in advance and freed when needed.

By having the ggtt backpointer, it also ensure that the init function
is always called before any attempt to insert or reserve the node
in the GGTT.

v2: s/xe_ggtt_node_force_fini/xe_ggtt_node_fini and use it
    internaly (Brost)
v3: - Use GF_NOFS for node allocation (CI)
    - Avoid ggtt argument, now that we have it inside the node (Lucas)
    - Fix some missed fini cases (CI)
v4: - Fix SRIOV critical case where config->ggtt_region was
    lost (Michal)
    - Remove useless checks (Michal)
    - Return 0 instead of negative errno on a u32 addr. (Michal)
    - Use xe_tile_assert instead of WARN_ON whenever possible (Michal)
    - Avoid ggtt argument also on removal (missed case on v3) (Michal)
    - s/xe_ggtt_assign/xe_ggtt_node_assign for coherence, while we
    are touching it (Michal)

Cc: Matthew Auld <matthew.auld@intel.com>
Cc: Michal Wajdeczko <michal.wajdeczko@intel.com>
Cc: Matthew Brost <matthew.brost@intel.com>
Reviewed-by: Matthew Brost <matthew.brost@intel.com> #2
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
  • Loading branch information
rodrigovivi committed Aug 19, 2024
1 parent 9e08f65 commit 2ac6c7d
Show file tree
Hide file tree
Showing 12 changed files with 192 additions and 83 deletions.
4 changes: 2 additions & 2 deletions drivers/gpu/drm/xe/compat-i915-headers/i915_vma.h
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@ struct xe_bo;

struct i915_vma {
struct xe_bo *bo, *dpt;
struct xe_ggtt_node node;
struct xe_ggtt_node *node;
};

#define i915_ggtt_clear_scanout(bo) do { } while (0)
Expand All @@ -29,7 +29,7 @@ struct i915_vma {

static inline u32 i915_ggtt_offset(const struct i915_vma *vma)
{
return vma->node.base.start;
return vma->node->base.start;
}

#endif
36 changes: 26 additions & 10 deletions drivers/gpu/drm/xe/display/xe_fb_pin.c
Original file line number Diff line number Diff line change
Expand Up @@ -204,20 +204,28 @@ static int __xe_pin_fb_vma_ggtt(const struct intel_framebuffer *fb,
if (xe_bo_is_vram(bo) && ggtt->flags & XE_GGTT_FLAGS_64K)
align = max_t(u32, align, SZ_64K);

if (bo->ggtt_node.base.size && view->type == I915_GTT_VIEW_NORMAL) {
if (bo->ggtt_node && view->type == I915_GTT_VIEW_NORMAL) {
vma->node = bo->ggtt_node;
} else if (view->type == I915_GTT_VIEW_NORMAL) {
u32 x, size = bo->ttm.base.size;

ret = xe_ggtt_node_insert_locked(ggtt, &vma->node, size, align, 0);
if (ret)
vma->node = xe_ggtt_node_init(ggtt);
if (IS_ERR(vma->node)) {
ret = PTR_ERR(vma->node);
goto out_unlock;
}

ret = xe_ggtt_node_insert_locked(vma->node, size, align, 0);
if (ret) {
xe_ggtt_node_fini(vma->node);
goto out_unlock;
}

for (x = 0; x < size; x += XE_PAGE_SIZE) {
u64 pte = ggtt->pt_ops->pte_encode_bo(bo, x,
xe->pat.idx[XE_CACHE_NONE]);

ggtt->pt_ops->ggtt_set_pte(ggtt, vma->node.base.start + x, pte);
ggtt->pt_ops->ggtt_set_pte(ggtt, vma->node->base.start + x, pte);
}
} else {
u32 i, ggtt_ofs;
Expand All @@ -226,11 +234,19 @@ static int __xe_pin_fb_vma_ggtt(const struct intel_framebuffer *fb,
/* display seems to use tiles instead of bytes here, so convert it back.. */
u32 size = intel_rotation_info_size(rot_info) * XE_PAGE_SIZE;

ret = xe_ggtt_node_insert_locked(ggtt, &vma->node, size, align, 0);
if (ret)
vma->node = xe_ggtt_node_init(ggtt);
if (IS_ERR(vma->node)) {
ret = PTR_ERR(vma->node);
goto out_unlock;
}

ret = xe_ggtt_node_insert_locked(vma->node, size, align, 0);
if (ret) {
xe_ggtt_node_fini(vma->node);
goto out_unlock;
}

ggtt_ofs = vma->node.base.start;
ggtt_ofs = vma->node->base.start;

for (i = 0; i < ARRAY_SIZE(rot_info->plane); i++)
write_ggtt_rotated(bo, ggtt, &ggtt_ofs,
Expand Down Expand Up @@ -323,9 +339,9 @@ static void __xe_unpin_fb_vma(struct i915_vma *vma)

if (vma->dpt)
xe_bo_unpin_map_no_vm(vma->dpt);
else if (!xe_ggtt_node_allocated(&vma->bo->ggtt_node) ||
vma->bo->ggtt_node.base.start != vma->node.base.start)
xe_ggtt_node_remove(ggtt, &vma->node, false);
else if (!xe_ggtt_node_allocated(vma->bo->ggtt_node) ||
vma->bo->ggtt_node->base.start != vma->node->base.start)
xe_ggtt_node_remove(ggtt, vma->node, false);

ttm_bo_reserve(&vma->bo->ttm, false, false, NULL);
ttm_bo_unpin(&vma->bo->ttm);
Expand Down
2 changes: 1 addition & 1 deletion drivers/gpu/drm/xe/xe_bo.c
Original file line number Diff line number Diff line change
Expand Up @@ -1120,7 +1120,7 @@ static void xe_ttm_bo_destroy(struct ttm_buffer_object *ttm_bo)

xe_assert(xe, list_empty(&ttm_bo->base.gpuva.list));

if (bo->ggtt_node.base.size)
if (bo->ggtt_node && bo->ggtt_node->base.size)
xe_ggtt_remove_bo(bo->tile->mem.ggtt, bo);

#ifdef CONFIG_PROC_FS
Expand Down
10 changes: 7 additions & 3 deletions drivers/gpu/drm/xe/xe_bo.h
Original file line number Diff line number Diff line change
Expand Up @@ -194,9 +194,13 @@ xe_bo_main_addr(struct xe_bo *bo, size_t page_size)
static inline u32
xe_bo_ggtt_addr(struct xe_bo *bo)
{
XE_WARN_ON(bo->ggtt_node.base.size > bo->size);
XE_WARN_ON(bo->ggtt_node.base.start + bo->ggtt_node.base.size > (1ull << 32));
return bo->ggtt_node.base.start;
if (XE_WARN_ON(!bo->ggtt_node))
return 0;

xe_tile_assert(bo->tile, bo->ggtt_node->base.size > bo->size);
xe_tile_assert(bo->tile, bo->ggtt_node->base.start + bo->ggtt_node->base.size >
(1ull << 32));
return bo->ggtt_node->base.start;
}

int xe_bo_vmap(struct xe_bo *bo);
Expand Down
2 changes: 1 addition & 1 deletion drivers/gpu/drm/xe/xe_bo_types.h
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ struct xe_bo {
/** @placement: current placement for this BO */
struct ttm_placement placement;
/** @ggtt_node: GGTT node if this BO is mapped in the GGTT */
struct xe_ggtt_node ggtt_node;
struct xe_ggtt_node *ggtt_node;
/** @vmap: iosys map of this buffer */
struct iosys_map vmap;
/** @ttm_kmap: TTM bo kmap object for internal use only. Keep off. */
Expand Down
2 changes: 1 addition & 1 deletion drivers/gpu/drm/xe/xe_device_types.h
Original file line number Diff line number Diff line change
Expand Up @@ -204,7 +204,7 @@ struct xe_tile {
struct xe_memirq memirq;

/** @sriov.vf.ggtt_balloon: GGTT regions excluded from use. */
struct xe_ggtt_node ggtt_balloon[2];
struct xe_ggtt_node *ggtt_balloon[2];
} vf;
} sriov;

Expand Down
125 changes: 91 additions & 34 deletions drivers/gpu/drm/xe/xe_ggtt.c
Original file line number Diff line number Diff line change
Expand Up @@ -348,17 +348,18 @@ static void xe_ggtt_dump_node(struct xe_ggtt *ggtt,

/**
* xe_ggtt_node_insert_balloon - prevent allocation of specified GGTT addresses
* @ggtt: the &xe_ggtt where we want to make reservation
* @node: the &xe_ggtt_node to hold reserved GGTT node
* @start: the starting GGTT address of the reserved region
* @end: then end GGTT address of the reserved region
*
* It cannot be called without first having called xe_ggtt_init().
* Use xe_ggtt_node_remove_balloon() to release a reserved GGTT node.
*
* Return: 0 on success or a negative error code on failure.
*/
int xe_ggtt_node_insert_balloon(struct xe_ggtt *ggtt, struct xe_ggtt_node *node, u64 start, u64 end)
int xe_ggtt_node_insert_balloon(struct xe_ggtt_node *node, u64 start, u64 end)
{
struct xe_ggtt *ggtt = node->ggtt;
int err;

xe_tile_assert(ggtt->tile, start < end);
Expand All @@ -385,64 +386,102 @@ int xe_ggtt_node_insert_balloon(struct xe_ggtt *ggtt, struct xe_ggtt_node *node,

/**
* xe_ggtt_node_remove_balloon - release a reserved GGTT region
* @ggtt: the &xe_ggtt where reserved node belongs
* @node: the &xe_ggtt_node with reserved GGTT region
*
* See xe_ggtt_node_insert_balloon() for details.
*/
void xe_ggtt_node_remove_balloon(struct xe_ggtt *ggtt, struct xe_ggtt_node *node)
void xe_ggtt_node_remove_balloon(struct xe_ggtt_node *node)
{
if (!drm_mm_node_allocated(&node->base))
return;
goto free_node;

xe_ggtt_dump_node(ggtt, &node->base, "deballoon");
xe_ggtt_dump_node(node->ggtt, &node->base, "deballoon");

mutex_lock(&ggtt->lock);
mutex_lock(&node->ggtt->lock);
drm_mm_remove_node(&node->base);
mutex_unlock(&ggtt->lock);
mutex_unlock(&node->ggtt->lock);

free_node:
xe_ggtt_node_fini(node);
}

/**
* xe_ggtt_node_insert_locked - Locked version to insert a &xe_ggtt_node into the GGTT
* @ggtt: the &xe_ggtt where node will be inserted
* @node: the &xe_ggtt_node to be inserted
* @size: size of the node
* @align: alignment constrain of the node
* @mm_flags: flags to control the node behavior
*
* It cannot be called without first having called xe_ggtt_init() once.
* To be used in cases where ggtt->lock is already taken.
*
* Return: 0 on success or a negative error code on failure.
*/
int xe_ggtt_node_insert_locked(struct xe_ggtt *ggtt, struct xe_ggtt_node *node,
int xe_ggtt_node_insert_locked(struct xe_ggtt_node *node,
u32 size, u32 align, u32 mm_flags)
{
return drm_mm_insert_node_generic(&ggtt->mm, &node->base, size, align, 0,
mm_flags);
return drm_mm_insert_node_generic(&node->ggtt->mm, &node->base, size, align, 0, mm_flags);
}

/**
* xe_ggtt_node_insert - Insert a &xe_ggtt_node into the GGTT
* @ggtt: the &xe_ggtt where node will be inserted
* @node: the &xe_ggtt_node to be inserted
* @size: size of the node
* @align: alignment constrain of the node
*
* It cannot be called without first having called xe_ggtt_init() once.
*
* Return: 0 on success or a negative error code on failure.
*/
int xe_ggtt_node_insert(struct xe_ggtt *ggtt, struct xe_ggtt_node *node,
u32 size, u32 align)
int xe_ggtt_node_insert(struct xe_ggtt_node *node, u32 size, u32 align)
{
int ret;

mutex_lock(&ggtt->lock);
ret = xe_ggtt_node_insert_locked(ggtt, node, size,
align, DRM_MM_INSERT_HIGH);
mutex_unlock(&ggtt->lock);
mutex_lock(&node->ggtt->lock);
ret = xe_ggtt_node_insert_locked(node, size, align,
DRM_MM_INSERT_HIGH);
mutex_unlock(&node->ggtt->lock);

return ret;
}

/**
* xe_ggtt_node_init - Initialize %xe_ggtt_node struct
* @ggtt: the &xe_ggtt where the new node will later be inserted/reserved.
*
* This function will allocated the struct %xe_ggtt_node and return it's pointer.
* This struct will then be freed after the node removal upon xe_ggtt_node_remove()
* or xe_ggtt_node_deballoon().
* Having %xe_ggtt_node struct allocated doesn't mean that the node is already allocated
* in GGTT. Only the xe_ggtt_node_insert(), xe_ggtt_node_insert_locked(),
* xe_ggtt_node_balloon() will ensure the node is inserted or reserved in GGTT.
*
* Return: A pointer to %xe_ggtt_node struct on success. An ERR_PTR otherwise.
**/
struct xe_ggtt_node *xe_ggtt_node_init(struct xe_ggtt *ggtt)
{
struct xe_ggtt_node *node = kzalloc(sizeof(*node), GFP_NOFS);

if (!node)
return ERR_PTR(-ENOMEM);

node->ggtt = ggtt;
return node;
}

/**
* xe_ggtt_node_fini - Forcebly finalize %xe_ggtt_node struct
* @node: the &xe_ggtt_node to be freed
*
* If anything went wrong with either xe_ggtt_node_insert(), xe_ggtt_node_insert_locked(),
* or xe_ggtt_node_balloon(); and this @node is not going to be reused, then,
* this function needs to be called to free the %xe_ggtt_node struct
**/
void xe_ggtt_node_fini(struct xe_ggtt_node *node)
{
kfree(node);
}

/**
* xe_ggtt_node_remove - Remove a &xe_ggtt_node from the GGTT
* @ggtt: the &xe_ggtt where node will be removed
Expand All @@ -468,17 +507,20 @@ void xe_ggtt_node_remove(struct xe_ggtt *ggtt, struct xe_ggtt_node *node,
mutex_unlock(&ggtt->lock);

if (!bound)
return;
goto free_node;

if (invalidate)
xe_ggtt_invalidate(ggtt);

xe_pm_runtime_put(xe);
drm_dev_exit(idx);

free_node:
xe_ggtt_node_fini(node);
}

/**
* xe_ggtt_node_allocated - Check if node is allocated
* xe_ggtt_node_allocated - Check if node is allocated in GGTT
* @node: the &xe_ggtt_node to be inspected
*
* Return: True if allocated, False otherwise.
Expand All @@ -497,9 +539,14 @@ void xe_ggtt_map_bo(struct xe_ggtt *ggtt, struct xe_bo *bo)
{
u16 cache_mode = bo->flags & XE_BO_FLAG_NEEDS_UC ? XE_CACHE_NONE : XE_CACHE_WB;
u16 pat_index = tile_to_xe(ggtt->tile)->pat.idx[cache_mode];
u64 start = bo->ggtt_node.base.start;
u64 start;
u64 offset, pte;

if (XE_WARN_ON(!bo->ggtt_node))
return;

start = bo->ggtt_node->base.start;

for (offset = 0; offset < bo->size; offset += XE_PAGE_SIZE) {
pte = ggtt->pt_ops->pte_encode_bo(bo, offset, pat_index);
ggtt->pt_ops->ggtt_set_pte(ggtt, start + offset, pte);
Expand All @@ -515,9 +562,9 @@ static int __xe_ggtt_insert_bo_at(struct xe_ggtt *ggtt, struct xe_bo *bo,
if (xe_bo_is_vram(bo) && ggtt->flags & XE_GGTT_FLAGS_64K)
alignment = SZ_64K;

if (XE_WARN_ON(bo->ggtt_node.base.size)) {
if (XE_WARN_ON(bo->ggtt_node)) {
/* Someone's already inserted this BO in the GGTT */
xe_tile_assert(ggtt->tile, bo->ggtt_node.base.size == bo->size);
xe_tile_assert(ggtt->tile, bo->ggtt_node->base.size == bo->size);
return 0;
}

Expand All @@ -526,15 +573,26 @@ static int __xe_ggtt_insert_bo_at(struct xe_ggtt *ggtt, struct xe_bo *bo,
return err;

xe_pm_runtime_get_noresume(tile_to_xe(ggtt->tile));

bo->ggtt_node = xe_ggtt_node_init(ggtt);
if (IS_ERR(bo->ggtt_node)) {
err = PTR_ERR(bo->ggtt_node);
goto out;
}

mutex_lock(&ggtt->lock);
err = drm_mm_insert_node_in_range(&ggtt->mm, &bo->ggtt_node.base, bo->size,
err = drm_mm_insert_node_in_range(&ggtt->mm, &bo->ggtt_node->base, bo->size,
alignment, 0, start, end, 0);
if (!err)
if (err)
xe_ggtt_node_fini(bo->ggtt_node);
else
xe_ggtt_map_bo(ggtt, bo);
mutex_unlock(&ggtt->lock);

if (!err && bo->flags & XE_BO_FLAG_GGTT_INVALIDATE)
xe_ggtt_invalidate(ggtt);

out:
xe_pm_runtime_put(tile_to_xe(ggtt->tile));

return err;
Expand Down Expand Up @@ -574,13 +632,13 @@ int xe_ggtt_insert_bo(struct xe_ggtt *ggtt, struct xe_bo *bo)
*/
void xe_ggtt_remove_bo(struct xe_ggtt *ggtt, struct xe_bo *bo)
{
if (XE_WARN_ON(!bo->ggtt_node.base.size))
if (XE_WARN_ON(!bo->ggtt_node))
return;

/* This BO is not currently in the GGTT */
xe_tile_assert(ggtt->tile, bo->ggtt_node.base.size == bo->size);
xe_tile_assert(ggtt->tile, bo->ggtt_node->base.size == bo->size);

xe_ggtt_node_remove(ggtt, &bo->ggtt_node,
xe_ggtt_node_remove(ggtt, bo->ggtt_node,
bo->flags & XE_BO_FLAG_GGTT_INVALIDATE);
}

Expand Down Expand Up @@ -647,19 +705,18 @@ static void xe_ggtt_assign_locked(struct xe_ggtt *ggtt, const struct drm_mm_node

/**
* xe_ggtt_assign - assign a GGTT region to the VF
* @ggtt: the &xe_ggtt where the node belongs
* @node: the &xe_ggtt_node to update
* @vfid: the VF identifier
*
* This function is used by the PF driver to assign a GGTT region to the VF.
* In addition to PTE's VFID bits 11:2 also PRESENT bit 0 is set as on some
* platforms VFs can't modify that either.
*/
void xe_ggtt_assign(struct xe_ggtt *ggtt, const struct xe_ggtt_node *node, u16 vfid)
void xe_ggtt_assign(const struct xe_ggtt_node *node, u16 vfid)
{
mutex_lock(&ggtt->lock);
xe_ggtt_assign_locked(ggtt, &node->base, vfid);
mutex_unlock(&ggtt->lock);
mutex_lock(&node->ggtt->lock);
xe_ggtt_assign_locked(node->ggtt, &node->base, vfid);
mutex_unlock(&node->ggtt->lock);
}
#endif

Expand Down
Loading

0 comments on commit 2ac6c7d

Please sign in to comment.