drm/xe: Simplify rebar sizing

"Right sizing" the PCI BAR is not necessary.  If rebar is needed
size to the maximum available.

Preserve the force_vram_bar_size sizing.

Reviewed-by: Matthew Auld <matthew.auld@intel.com>
Signed-off-by: Michael J. Ruhl <michael.j.ruhl@intel.com>
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
This commit is contained in:
Michael J. Ruhl 2023-05-25 15:43:24 -04:00 committed by Rodrigo Vivi
parent 61f288a897
commit 7f075300a3
3 changed files with 91 additions and 70 deletions

View File

@ -1,6 +1,6 @@
/* SPDX-License-Identifier: MIT */ /* SPDX-License-Identifier: MIT */
/* /*
* Copyright © 2022 Intel Corporation * Copyright © 2022-2023 Intel Corporation
*/ */
#ifndef _XE_DEVICE_TYPES_H_ #ifndef _XE_DEVICE_TYPES_H_
@ -133,11 +133,13 @@ struct xe_device {
/** /**
* @io_size: IO size of VRAM. * @io_size: IO size of VRAM.
* *
* This represents how much of VRAM we can access via * This represents how much of VRAM the CPU can access
* the CPU through the VRAM BAR. This can be smaller * via the VRAM BAR.
* than @size, in which case only part of VRAM is CPU * On systems that do not support large BAR IO space,
* accessible (typically the first 256M). This * this can be smaller than the actual memory size, in
* configuration is known as small-bar. * which case only part of VRAM is CPU accessible
* (typically the first 256M). This configuration is
* known as small-bar.
*/ */
resource_size_t io_size; resource_size_t io_size;
/** @size: Total size of VRAM */ /** @size: Total size of VRAM */

View File

@ -1,6 +1,6 @@
/* SPDX-License-Identifier: MIT */ /* SPDX-License-Identifier: MIT */
/* /*
* Copyright © 2022 Intel Corporation * Copyright © 2022-2023 Intel Corporation
*/ */
#ifndef _XE_GT_TYPES_H_ #ifndef _XE_GT_TYPES_H_
@ -148,11 +148,11 @@ struct xe_gt {
/** /**
* @io_size: IO size of this VRAM instance * @io_size: IO size of this VRAM instance
* *
* This represents how much of this VRAM we can access * This represents how much of the VRAM the CPU can access
* via the CPU through the VRAM BAR. This can be smaller * via the VRAM BAR.
* than @size, in which case only part of VRAM is CPU * This can be smaller than the actual @size, in which
* accessible (typically the first 256M). This * case only part of VRAM is CPU accessible (typically
* configuration is known as small-bar. * the first 256M). This configuration is known as small-bar.
*/ */
resource_size_t io_size; resource_size_t io_size;
/** @size: size of VRAM. */ /** @size: size of VRAM. */

View File

@ -3,6 +3,8 @@
* Copyright © 2021-2023 Intel Corporation * Copyright © 2021-2023 Intel Corporation
*/ */
#include <linux/minmax.h>
#include "xe_mmio.h" #include "xe_mmio.h"
#include <drm/drm_managed.h> #include <drm/drm_managed.h>
@ -21,6 +23,8 @@
#define XEHP_MTCFG_ADDR XE_REG(0x101800) #define XEHP_MTCFG_ADDR XE_REG(0x101800)
#define TILE_COUNT REG_GENMASK(15, 8) #define TILE_COUNT REG_GENMASK(15, 8)
#define BAR_SIZE_SHIFT 20
static int xe_set_dma_info(struct xe_device *xe) static int xe_set_dma_info(struct xe_device *xe)
{ {
unsigned int mask_size = xe->info.dma_mask_size; unsigned int mask_size = xe->info.dma_mask_size;
@ -57,49 +61,61 @@ _resize_bar(struct xe_device *xe, int resno, resource_size_t size)
if (ret) { if (ret) {
drm_info(&xe->drm, "Failed to resize BAR%d to %dM (%pe). Consider enabling 'Resizable BAR' support in your BIOS\n", drm_info(&xe->drm, "Failed to resize BAR%d to %dM (%pe). Consider enabling 'Resizable BAR' support in your BIOS\n",
resno, 1 << bar_size, ERR_PTR(ret)); resno, 1 << bar_size, ERR_PTR(ret));
return -1; return ret;
} }
drm_info(&xe->drm, "BAR%d resized to %dM\n", resno, 1 << bar_size); drm_info(&xe->drm, "BAR%d resized to %dM\n", resno, 1 << bar_size);
return 1; return ret;
} }
static int xe_resize_vram_bar(struct xe_device *xe, resource_size_t vram_size) /*
* if force_vram_bar_size is set, attempt to set to the requested size
* else set to maximum possible size
*/
static int xe_resize_vram_bar(struct xe_device *xe)
{ {
u64 force_vram_bar_size = xe_force_vram_bar_size;
struct pci_dev *pdev = to_pci_dev(xe->drm.dev); struct pci_dev *pdev = to_pci_dev(xe->drm.dev);
struct pci_bus *root = pdev->bus; struct pci_bus *root = pdev->bus;
struct resource *root_res;
resource_size_t rebar_size;
resource_size_t current_size; resource_size_t current_size;
resource_size_t rebar_size;
struct resource *root_res;
u32 bar_size_mask;
u32 pci_cmd; u32 pci_cmd;
int i; int i;
int ret; int ret;
u64 force_vram_bar_size = xe_force_vram_bar_size;
current_size = roundup_pow_of_two(pci_resource_len(pdev, GEN12_LMEM_BAR)); /* gather some relevant info */
current_size = pci_resource_len(pdev, GEN12_LMEM_BAR);
bar_size_mask = pci_rebar_get_possible_sizes(pdev, GEN12_LMEM_BAR);
if (!bar_size_mask)
return 0;
/* set to a specific size? */
if (force_vram_bar_size) { if (force_vram_bar_size) {
u32 bar_sizes; u32 bar_size_bit;
rebar_size = force_vram_bar_size * (resource_size_t)SZ_1M; rebar_size = force_vram_bar_size * (resource_size_t)SZ_1M;
bar_sizes = pci_rebar_get_possible_sizes(pdev, GEN12_LMEM_BAR);
bar_size_bit = bar_size_mask & BIT(pci_rebar_bytes_to_size(rebar_size));
if (!bar_size_bit) {
drm_info(&xe->drm,
"Requested size: %lluMiB is not supported by rebar sizes: 0x%x. Leaving default: %lluMiB\n",
(u64)rebar_size >> 20, bar_size_mask, (u64)current_size >> 20);
return 0;
}
rebar_size = 1ULL << (__fls(bar_size_bit) + BAR_SIZE_SHIFT);
if (rebar_size == current_size) if (rebar_size == current_size)
return 0; return 0;
if (!(bar_sizes & BIT(pci_rebar_bytes_to_size(rebar_size))) ||
rebar_size >= roundup_pow_of_two(vram_size)) {
rebar_size = vram_size;
drm_info(&xe->drm,
"Given bar size is not within supported size, setting it to default: %lluMiB\n",
(u64)vram_size >> 20);
}
} else { } else {
rebar_size = current_size; rebar_size = 1ULL << (__fls(bar_size_mask) + BAR_SIZE_SHIFT);
if (rebar_size != roundup_pow_of_two(vram_size)) /* only resize if larger than current */
rebar_size = vram_size; if (rebar_size <= current_size)
else
return 0; return 0;
} }
@ -144,6 +160,31 @@ static bool xe_pci_resource_valid(struct pci_dev *pdev, int bar)
return true; return true;
} }
static int xe_determine_lmem_bar_size(struct xe_device *xe)
{
struct pci_dev *pdev = to_pci_dev(xe->drm.dev);
int err;
if (!xe_pci_resource_valid(pdev, GEN12_LMEM_BAR)) {
drm_err(&xe->drm, "pci resource is not valid\n");
return -ENXIO;
}
err = xe_resize_vram_bar(xe);
if (err)
return err;
xe->mem.vram.io_start = pci_resource_start(pdev, GEN12_LMEM_BAR);
xe->mem.vram.io_size = pci_resource_len(pdev, GEN12_LMEM_BAR);
if (!xe->mem.vram.io_size)
return -EIO;
/* set up a map to the total memory area. */
xe->mem.vram.mapping = ioremap_wc(xe->mem.vram.io_start, xe->mem.vram.io_size);
return 0;
}
/** /**
* xe_mmio_tile_vram_size() - Collect vram size and offset information * xe_mmio_tile_vram_size() - Collect vram size and offset information
* @gt: tile to get info for * @gt: tile to get info for
@ -199,59 +240,37 @@ int xe_mmio_tile_vram_size(struct xe_gt *gt, u64 *vram_size, u64 *tile_size, u64
int xe_mmio_probe_vram(struct xe_device *xe) int xe_mmio_probe_vram(struct xe_device *xe)
{ {
struct pci_dev *pdev = to_pci_dev(xe->drm.dev);
struct xe_gt *gt; struct xe_gt *gt;
u64 original_size;
u64 tile_offset; u64 tile_offset;
u64 tile_size; u64 tile_size;
u64 vram_size; u64 vram_size;
int err; int err;
u8 id; u8 id;
if (!IS_DGFX(xe)) { if (!IS_DGFX(xe))
xe->mem.vram.mapping = 0;
xe->mem.vram.size = 0;
xe->mem.vram.io_start = 0;
xe->mem.vram.io_size = 0;
for_each_gt(gt, xe, id) {
gt->mem.vram.mapping = 0;
gt->mem.vram.size = 0;
gt->mem.vram.io_start = 0;
gt->mem.vram.io_size = 0;
}
return 0; return 0;
}
if (!xe_pci_resource_valid(pdev, GEN12_LMEM_BAR)) {
drm_err(&xe->drm, "pci resource is not valid\n");
return -ENXIO;
}
/* Get the size of the gt0 vram for later accessibility comparison */
gt = xe_device_get_gt(xe, 0); gt = xe_device_get_gt(xe, 0);
original_size = pci_resource_len(pdev, GEN12_LMEM_BAR);
err = xe_mmio_tile_vram_size(gt, &vram_size, &tile_size, &tile_offset); err = xe_mmio_tile_vram_size(gt, &vram_size, &tile_size, &tile_offset);
if (err) if (err)
return err; return err;
xe_resize_vram_bar(xe, vram_size); err = xe_determine_lmem_bar_size(xe);
xe->mem.vram.io_start = pci_resource_start(pdev, GEN12_LMEM_BAR); if (err)
xe->mem.vram.io_size = min(vram_size, return err;
pci_resource_len(pdev, GEN12_LMEM_BAR));
/* small bar issues will only cover gt0 sizes */
if (xe->mem.vram.io_size < vram_size)
drm_warn(&xe->drm, "Restricting VRAM size to PCI resource size (0x%llx->0x%llx)\n",
vram_size, (u64)xe->mem.vram.io_size);
/* Limit size to available memory to account for the current memory algorithm */
xe->mem.vram.io_size = min_t(u64, xe->mem.vram.io_size, vram_size);
xe->mem.vram.size = xe->mem.vram.io_size; xe->mem.vram.size = xe->mem.vram.io_size;
if (!xe->mem.vram.size) drm_info(&xe->drm, "VISIBLE VRAM: %pa, %pa\n", &xe->mem.vram.io_start,
return -EIO; &xe->mem.vram.io_size);
if (vram_size > xe->mem.vram.io_size)
drm_warn(&xe->drm, "Restricting VRAM size to PCI resource size (%lluMiB->%lluMiB)\n",
(u64)vram_size >> 20, (u64)xe->mem.vram.io_size >> 20);
xe->mem.vram.mapping = ioremap_wc(xe->mem.vram.io_start, xe->mem.vram.io_size);
xe->mem.vram.size = min_t(u64, xe->mem.vram.size, vram_size);
drm_info(&xe->drm, "TOTAL VRAM: %pa, %pa\n", &xe->mem.vram.io_start, &xe->mem.vram.size);
/* FIXME: Assuming equally partitioned VRAM, incorrect */ /* FIXME: Assuming equally partitioned VRAM, incorrect */
if (xe->info.tile_count > 1) { if (xe->info.tile_count > 1) {