// SPDX-License-Identifier: GPL-2.0-only // Copyright(c) 2021 Intel Corporation. All rights reserved. #include #include #include #include #include #include #include #include #include "../watermark.h" #include "mock.h" static int interleave_arithmetic; #define FAKE_QTG_ID 42 #define NR_CXL_HOST_BRIDGES 2 #define NR_CXL_SINGLE_HOST 1 #define NR_CXL_RCH 1 #define NR_CXL_ROOT_PORTS 2 #define NR_CXL_SWITCH_PORTS 2 #define NR_CXL_PORT_DECODERS 8 #define NR_BRIDGES (NR_CXL_HOST_BRIDGES + NR_CXL_SINGLE_HOST + NR_CXL_RCH) static struct platform_device *cxl_acpi; static struct platform_device *cxl_host_bridge[NR_CXL_HOST_BRIDGES]; #define NR_MULTI_ROOT (NR_CXL_HOST_BRIDGES * NR_CXL_ROOT_PORTS) static struct platform_device *cxl_root_port[NR_MULTI_ROOT]; static struct platform_device *cxl_switch_uport[NR_MULTI_ROOT]; #define NR_MEM_MULTI \ (NR_CXL_HOST_BRIDGES * NR_CXL_ROOT_PORTS * NR_CXL_SWITCH_PORTS) static struct platform_device *cxl_switch_dport[NR_MEM_MULTI]; static struct platform_device *cxl_hb_single[NR_CXL_SINGLE_HOST]; static struct platform_device *cxl_root_single[NR_CXL_SINGLE_HOST]; static struct platform_device *cxl_swu_single[NR_CXL_SINGLE_HOST]; #define NR_MEM_SINGLE (NR_CXL_SINGLE_HOST * NR_CXL_SWITCH_PORTS) static struct platform_device *cxl_swd_single[NR_MEM_SINGLE]; struct platform_device *cxl_mem[NR_MEM_MULTI]; struct platform_device *cxl_mem_single[NR_MEM_SINGLE]; static struct platform_device *cxl_rch[NR_CXL_RCH]; static struct platform_device *cxl_rcd[NR_CXL_RCH]; static inline bool is_multi_bridge(struct device *dev) { int i; for (i = 0; i < ARRAY_SIZE(cxl_host_bridge); i++) if (&cxl_host_bridge[i]->dev == dev) return true; return false; } static inline bool is_single_bridge(struct device *dev) { int i; for (i = 0; i < ARRAY_SIZE(cxl_hb_single); i++) if (&cxl_hb_single[i]->dev == dev) return true; return false; } static struct acpi_device acpi0017_mock; static struct acpi_device host_bridge[NR_BRIDGES] = { [0] = { .handle = &host_bridge[0], .pnp.unique_id = "0", }, [1] = { .handle = &host_bridge[1], .pnp.unique_id = "1", }, [2] = { .handle = &host_bridge[2], .pnp.unique_id = "2", }, [3] = { .handle = &host_bridge[3], .pnp.unique_id = "3", }, }; static bool is_mock_dev(struct device *dev) { int i; for (i = 0; i < ARRAY_SIZE(cxl_mem); i++) if (dev == &cxl_mem[i]->dev) return true; for (i = 0; i < ARRAY_SIZE(cxl_mem_single); i++) if (dev == &cxl_mem_single[i]->dev) return true; for (i = 0; i < ARRAY_SIZE(cxl_rcd); i++) if (dev == &cxl_rcd[i]->dev) return true; if (dev == &cxl_acpi->dev) return true; return false; } static bool is_mock_adev(struct acpi_device *adev) { int i; if (adev == &acpi0017_mock) return true; for (i = 0; i < ARRAY_SIZE(host_bridge); i++) if (adev == &host_bridge[i]) return true; return false; } static struct { struct acpi_table_cedt cedt; struct acpi_cedt_chbs chbs[NR_BRIDGES]; struct { struct acpi_cedt_cfmws cfmws; u32 target[1]; } cfmws0; struct { struct acpi_cedt_cfmws cfmws; u32 target[2]; } cfmws1; struct { struct acpi_cedt_cfmws cfmws; u32 target[1]; } cfmws2; struct { struct acpi_cedt_cfmws cfmws; u32 target[2]; } cfmws3; struct { struct acpi_cedt_cfmws cfmws; u32 target[1]; } cfmws4; struct { struct acpi_cedt_cfmws cfmws; u32 target[1]; } cfmws5; struct { struct acpi_cedt_cfmws cfmws; u32 target[1]; } cfmws6; struct { struct acpi_cedt_cfmws cfmws; u32 target[2]; } cfmws7; struct { struct acpi_cedt_cfmws cfmws; u32 target[4]; } cfmws8; struct { struct acpi_cedt_cxims cxims; u64 xormap_list[2]; } cxims0; } __packed mock_cedt = { .cedt = { .header = { .signature = "CEDT", .length = sizeof(mock_cedt), .revision = 1, }, }, .chbs[0] = { .header = { .type = ACPI_CEDT_TYPE_CHBS, .length = sizeof(mock_cedt.chbs[0]), }, .uid = 0, .cxl_version = ACPI_CEDT_CHBS_VERSION_CXL20, }, .chbs[1] = { .header = { .type = ACPI_CEDT_TYPE_CHBS, .length = sizeof(mock_cedt.chbs[0]), }, .uid = 1, .cxl_version = ACPI_CEDT_CHBS_VERSION_CXL20, }, .chbs[2] = { .header = { .type = ACPI_CEDT_TYPE_CHBS, .length = sizeof(mock_cedt.chbs[0]), }, .uid = 2, .cxl_version = ACPI_CEDT_CHBS_VERSION_CXL20, }, .chbs[3] = { .header = { .type = ACPI_CEDT_TYPE_CHBS, .length = sizeof(mock_cedt.chbs[0]), }, .uid = 3, .cxl_version = ACPI_CEDT_CHBS_VERSION_CXL11, }, .cfmws0 = { .cfmws = { .header = { .type = ACPI_CEDT_TYPE_CFMWS, .length = sizeof(mock_cedt.cfmws0), }, .interleave_ways = 0, .granularity = 4, .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | ACPI_CEDT_CFMWS_RESTRICT_VOLATILE, .qtg_id = FAKE_QTG_ID, .window_size = SZ_256M * 4UL, }, .target = { 0 }, }, .cfmws1 = { .cfmws = { .header = { .type = ACPI_CEDT_TYPE_CFMWS, .length = sizeof(mock_cedt.cfmws1), }, .interleave_ways = 1, .granularity = 4, .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | ACPI_CEDT_CFMWS_RESTRICT_VOLATILE, .qtg_id = FAKE_QTG_ID, .window_size = SZ_256M * 8UL, }, .target = { 0, 1, }, }, .cfmws2 = { .cfmws = { .header = { .type = ACPI_CEDT_TYPE_CFMWS, .length = sizeof(mock_cedt.cfmws2), }, .interleave_ways = 0, .granularity = 4, .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | ACPI_CEDT_CFMWS_RESTRICT_PMEM, .qtg_id = FAKE_QTG_ID, .window_size = SZ_256M * 4UL, }, .target = { 0 }, }, .cfmws3 = { .cfmws = { .header = { .type = ACPI_CEDT_TYPE_CFMWS, .length = sizeof(mock_cedt.cfmws3), }, .interleave_ways = 1, .granularity = 4, .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | ACPI_CEDT_CFMWS_RESTRICT_PMEM, .qtg_id = FAKE_QTG_ID, .window_size = SZ_256M * 8UL, }, .target = { 0, 1, }, }, .cfmws4 = { .cfmws = { .header = { .type = ACPI_CEDT_TYPE_CFMWS, .length = sizeof(mock_cedt.cfmws4), }, .interleave_ways = 0, .granularity = 4, .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | ACPI_CEDT_CFMWS_RESTRICT_PMEM, .qtg_id = FAKE_QTG_ID, .window_size = SZ_256M * 4UL, }, .target = { 2 }, }, .cfmws5 = { .cfmws = { .header = { .type = ACPI_CEDT_TYPE_CFMWS, .length = sizeof(mock_cedt.cfmws5), }, .interleave_ways = 0, .granularity = 4, .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | ACPI_CEDT_CFMWS_RESTRICT_VOLATILE, .qtg_id = FAKE_QTG_ID, .window_size = SZ_256M, }, .target = { 3 }, }, /* .cfmws6,7,8 use ACPI_CEDT_CFMWS_ARITHMETIC_XOR */ .cfmws6 = { .cfmws = { .header = { .type = ACPI_CEDT_TYPE_CFMWS, .length = sizeof(mock_cedt.cfmws6), }, .interleave_arithmetic = ACPI_CEDT_CFMWS_ARITHMETIC_XOR, .interleave_ways = 0, .granularity = 4, .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | ACPI_CEDT_CFMWS_RESTRICT_PMEM, .qtg_id = FAKE_QTG_ID, .window_size = SZ_256M * 8UL, }, .target = { 0, }, }, .cfmws7 = { .cfmws = { .header = { .type = ACPI_CEDT_TYPE_CFMWS, .length = sizeof(mock_cedt.cfmws7), }, .interleave_arithmetic = ACPI_CEDT_CFMWS_ARITHMETIC_XOR, .interleave_ways = 1, .granularity = 0, .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | ACPI_CEDT_CFMWS_RESTRICT_PMEM, .qtg_id = FAKE_QTG_ID, .window_size = SZ_256M * 8UL, }, .target = { 0, 1, }, }, .cfmws8 = { .cfmws = { .header = { .type = ACPI_CEDT_TYPE_CFMWS, .length = sizeof(mock_cedt.cfmws8), }, .interleave_arithmetic = ACPI_CEDT_CFMWS_ARITHMETIC_XOR, .interleave_ways = 2, .granularity = 0, .restrictions = ACPI_CEDT_CFMWS_RESTRICT_TYPE3 | ACPI_CEDT_CFMWS_RESTRICT_PMEM, .qtg_id = FAKE_QTG_ID, .window_size = SZ_256M * 16UL, }, .target = { 0, 1, 0, 1, }, }, .cxims0 = { .cxims = { .header = { .type = ACPI_CEDT_TYPE_CXIMS, .length = sizeof(mock_cedt.cxims0), }, .hbig = 0, .nr_xormaps = 2, }, .xormap_list = { 0x404100, 0x808200, }, }, }; struct acpi_cedt_cfmws *mock_cfmws[] = { [0] = &mock_cedt.cfmws0.cfmws, [1] = &mock_cedt.cfmws1.cfmws, [2] = &mock_cedt.cfmws2.cfmws, [3] = &mock_cedt.cfmws3.cfmws, [4] = &mock_cedt.cfmws4.cfmws, [5] = &mock_cedt.cfmws5.cfmws, /* Modulo Math above, XOR Math below */ [6] = &mock_cedt.cfmws6.cfmws, [7] = &mock_cedt.cfmws7.cfmws, [8] = &mock_cedt.cfmws8.cfmws, }; static int cfmws_start; static int cfmws_end; #define CFMWS_MOD_ARRAY_START 0 #define CFMWS_MOD_ARRAY_END 5 #define CFMWS_XOR_ARRAY_START 6 #define CFMWS_XOR_ARRAY_END 8 struct acpi_cedt_cxims *mock_cxims[1] = { [0] = &mock_cedt.cxims0.cxims, }; struct cxl_mock_res { struct list_head list; struct range range; }; static LIST_HEAD(mock_res); static DEFINE_MUTEX(mock_res_lock); static struct gen_pool *cxl_mock_pool; static void depopulate_all_mock_resources(void) { struct cxl_mock_res *res, *_res; mutex_lock(&mock_res_lock); list_for_each_entry_safe(res, _res, &mock_res, list) { gen_pool_free(cxl_mock_pool, res->range.start, range_len(&res->range)); list_del(&res->list); kfree(res); } mutex_unlock(&mock_res_lock); } static struct cxl_mock_res *alloc_mock_res(resource_size_t size, int align) { struct cxl_mock_res *res = kzalloc(sizeof(*res), GFP_KERNEL); struct genpool_data_align data = { .align = align, }; unsigned long phys; INIT_LIST_HEAD(&res->list); phys = gen_pool_alloc_algo(cxl_mock_pool, size, gen_pool_first_fit_align, &data); if (!phys) return NULL; res->range = (struct range) { .start = phys, .end = phys + size - 1, }; mutex_lock(&mock_res_lock); list_add(&res->list, &mock_res); mutex_unlock(&mock_res_lock); return res; } static int populate_cedt(void) { struct cxl_mock_res *res; int i; for (i = 0; i < ARRAY_SIZE(mock_cedt.chbs); i++) { struct acpi_cedt_chbs *chbs = &mock_cedt.chbs[i]; resource_size_t size; if (chbs->cxl_version == ACPI_CEDT_CHBS_VERSION_CXL20) size = ACPI_CEDT_CHBS_LENGTH_CXL20; else size = ACPI_CEDT_CHBS_LENGTH_CXL11; res = alloc_mock_res(size, size); if (!res) return -ENOMEM; chbs->base = res->range.start; chbs->length = size; } for (i = cfmws_start; i <= cfmws_end; i++) { struct acpi_cedt_cfmws *window = mock_cfmws[i]; res = alloc_mock_res(window->window_size, SZ_256M); if (!res) return -ENOMEM; window->base_hpa = res->range.start; } return 0; } static bool is_mock_port(struct device *dev); /* * WARNING, this hack assumes the format of 'struct cxl_cfmws_context' * and 'struct cxl_chbs_context' share the property that the first * struct member is a cxl_test device being probed by the cxl_acpi * driver. */ struct cxl_cedt_context { struct device *dev; }; static int mock_acpi_table_parse_cedt(enum acpi_cedt_type id, acpi_tbl_entry_handler_arg handler_arg, void *arg) { struct cxl_cedt_context *ctx = arg; struct device *dev = ctx->dev; union acpi_subtable_headers *h; unsigned long end; int i; if (!is_mock_port(dev) && !is_mock_dev(dev)) return acpi_table_parse_cedt(id, handler_arg, arg); if (id == ACPI_CEDT_TYPE_CHBS) for (i = 0; i < ARRAY_SIZE(mock_cedt.chbs); i++) { h = (union acpi_subtable_headers *)&mock_cedt.chbs[i]; end = (unsigned long)&mock_cedt.chbs[i + 1]; handler_arg(h, arg, end); } if (id == ACPI_CEDT_TYPE_CFMWS) for (i = cfmws_start; i <= cfmws_end; i++) { h = (union acpi_subtable_headers *) mock_cfmws[i]; end = (unsigned long) h + mock_cfmws[i]->header.length; handler_arg(h, arg, end); } if (id == ACPI_CEDT_TYPE_CXIMS) for (i = 0; i < ARRAY_SIZE(mock_cxims); i++) { h = (union acpi_subtable_headers *)mock_cxims[i]; end = (unsigned long)h + mock_cxims[i]->header.length; handler_arg(h, arg, end); } return 0; } static bool is_mock_bridge(struct device *dev) { int i; for (i = 0; i < ARRAY_SIZE(cxl_host_bridge); i++) if (dev == &cxl_host_bridge[i]->dev) return true; for (i = 0; i < ARRAY_SIZE(cxl_hb_single); i++) if (dev == &cxl_hb_single[i]->dev) return true; for (i = 0; i < ARRAY_SIZE(cxl_rch); i++) if (dev == &cxl_rch[i]->dev) return true; return false; } static bool is_mock_port(struct device *dev) { int i; if (is_mock_bridge(dev)) return true; for (i = 0; i < ARRAY_SIZE(cxl_root_port); i++) if (dev == &cxl_root_port[i]->dev) return true; for (i = 0; i < ARRAY_SIZE(cxl_switch_uport); i++) if (dev == &cxl_switch_uport[i]->dev) return true; for (i = 0; i < ARRAY_SIZE(cxl_switch_dport); i++) if (dev == &cxl_switch_dport[i]->dev) return true; for (i = 0; i < ARRAY_SIZE(cxl_root_single); i++) if (dev == &cxl_root_single[i]->dev) return true; for (i = 0; i < ARRAY_SIZE(cxl_swu_single); i++) if (dev == &cxl_swu_single[i]->dev) return true; for (i = 0; i < ARRAY_SIZE(cxl_swd_single); i++) if (dev == &cxl_swd_single[i]->dev) return true; if (is_cxl_memdev(dev)) return is_mock_dev(dev->parent); return false; } static int host_bridge_index(struct acpi_device *adev) { return adev - host_bridge; } static struct acpi_device *find_host_bridge(acpi_handle handle) { int i; for (i = 0; i < ARRAY_SIZE(host_bridge); i++) if (handle == host_bridge[i].handle) return &host_bridge[i]; return NULL; } static acpi_status mock_acpi_evaluate_integer(acpi_handle handle, acpi_string pathname, struct acpi_object_list *arguments, unsigned long long *data) { struct acpi_device *adev = find_host_bridge(handle); if (!adev || strcmp(pathname, METHOD_NAME__UID) != 0) return acpi_evaluate_integer(handle, pathname, arguments, data); *data = host_bridge_index(adev); return AE_OK; } static struct pci_bus mock_pci_bus[NR_BRIDGES]; static struct acpi_pci_root mock_pci_root[ARRAY_SIZE(mock_pci_bus)] = { [0] = { .bus = &mock_pci_bus[0], }, [1] = { .bus = &mock_pci_bus[1], }, [2] = { .bus = &mock_pci_bus[2], }, [3] = { .bus = &mock_pci_bus[3], }, }; static bool is_mock_bus(struct pci_bus *bus) { int i; for (i = 0; i < ARRAY_SIZE(mock_pci_bus); i++) if (bus == &mock_pci_bus[i]) return true; return false; } static struct acpi_pci_root *mock_acpi_pci_find_root(acpi_handle handle) { struct acpi_device *adev = find_host_bridge(handle); if (!adev) return acpi_pci_find_root(handle); return &mock_pci_root[host_bridge_index(adev)]; } static struct cxl_hdm *mock_cxl_setup_hdm(struct cxl_port *port, struct cxl_endpoint_dvsec_info *info) { struct cxl_hdm *cxlhdm = devm_kzalloc(&port->dev, sizeof(*cxlhdm), GFP_KERNEL); struct device *dev = &port->dev; if (!cxlhdm) return ERR_PTR(-ENOMEM); cxlhdm->port = port; cxlhdm->interleave_mask = ~0U; cxlhdm->iw_cap_mask = ~0UL; dev_set_drvdata(dev, cxlhdm); return cxlhdm; } static int mock_cxl_add_passthrough_decoder(struct cxl_port *port) { dev_err(&port->dev, "unexpected passthrough decoder for cxl_test\n"); return -EOPNOTSUPP; } struct target_map_ctx { int *target_map; int index; int target_count; }; static int map_targets(struct device *dev, void *data) { struct platform_device *pdev = to_platform_device(dev); struct target_map_ctx *ctx = data; ctx->target_map[ctx->index++] = pdev->id; if (ctx->index > ctx->target_count) { dev_WARN_ONCE(dev, 1, "too many targets found?\n"); return -ENXIO; } return 0; } static int mock_decoder_commit(struct cxl_decoder *cxld) { struct cxl_port *port = to_cxl_port(cxld->dev.parent); int id = cxld->id; if (cxld->flags & CXL_DECODER_F_ENABLE) return 0; dev_dbg(&port->dev, "%s commit\n", dev_name(&cxld->dev)); if (cxl_num_decoders_committed(port) != id) { dev_dbg(&port->dev, "%s: out of order commit, expected decoder%d.%d\n", dev_name(&cxld->dev), port->id, cxl_num_decoders_committed(port)); return -EBUSY; } port->commit_end++; cxld->flags |= CXL_DECODER_F_ENABLE; return 0; } static void mock_decoder_reset(struct cxl_decoder *cxld) { struct cxl_port *port = to_cxl_port(cxld->dev.parent); int id = cxld->id; if ((cxld->flags & CXL_DECODER_F_ENABLE) == 0) return; dev_dbg(&port->dev, "%s reset\n", dev_name(&cxld->dev)); if (port->commit_end == id) cxl_port_commit_reap(cxld); else dev_dbg(&port->dev, "%s: out of order reset, expected decoder%d.%d\n", dev_name(&cxld->dev), port->id, port->commit_end); cxld->flags &= ~CXL_DECODER_F_ENABLE; } static void default_mock_decoder(struct cxl_decoder *cxld) { cxld->hpa_range = (struct range){ .start = 0, .end = -1, }; cxld->interleave_ways = 1; cxld->interleave_granularity = 256; cxld->target_type = CXL_DECODER_HOSTONLYMEM; cxld->commit = mock_decoder_commit; cxld->reset = mock_decoder_reset; } static int first_decoder(struct device *dev, void *data) { struct cxl_decoder *cxld; if (!is_switch_decoder(dev)) return 0; cxld = to_cxl_decoder(dev); if (cxld->id == 0) return 1; return 0; } static void mock_init_hdm_decoder(struct cxl_decoder *cxld) { struct acpi_cedt_cfmws *window = mock_cfmws[0]; struct platform_device *pdev = NULL; struct cxl_endpoint_decoder *cxled; struct cxl_switch_decoder *cxlsd; struct cxl_port *port, *iter; const int size = SZ_512M; struct cxl_memdev *cxlmd; struct cxl_dport *dport; struct device *dev; bool hb0 = false; u64 base; int i; if (is_endpoint_decoder(&cxld->dev)) { cxled = to_cxl_endpoint_decoder(&cxld->dev); cxlmd = cxled_to_memdev(cxled); WARN_ON(!dev_is_platform(cxlmd->dev.parent)); pdev = to_platform_device(cxlmd->dev.parent); /* check is endpoint is attach to host-bridge0 */ port = cxled_to_port(cxled); do { if (port->uport_dev == &cxl_host_bridge[0]->dev) { hb0 = true; break; } if (is_cxl_port(port->dev.parent)) port = to_cxl_port(port->dev.parent); else port = NULL; } while (port); port = cxled_to_port(cxled); } /* * The first decoder on the first 2 devices on the first switch * attached to host-bridge0 mock a fake / static RAM region. All * other decoders are default disabled. Given the round robin * assignment those devices are named cxl_mem.0, and cxl_mem.4. * * See 'cxl list -BMPu -m cxl_mem.0,cxl_mem.4' */ if (!hb0 || pdev->id % 4 || pdev->id > 4 || cxld->id > 0) { default_mock_decoder(cxld); return; } base = window->base_hpa; cxld->hpa_range = (struct range) { .start = base, .end = base + size - 1, }; cxld->interleave_ways = 2; eig_to_granularity(window->granularity, &cxld->interleave_granularity); cxld->target_type = CXL_DECODER_HOSTONLYMEM; cxld->flags = CXL_DECODER_F_ENABLE; cxled->state = CXL_DECODER_STATE_AUTO; port->commit_end = cxld->id; devm_cxl_dpa_reserve(cxled, 0, size / cxld->interleave_ways, 0); cxld->commit = mock_decoder_commit; cxld->reset = mock_decoder_reset; /* * Now that endpoint decoder is set up, walk up the hierarchy * and setup the switch and root port decoders targeting @cxlmd. */ iter = port; for (i = 0; i < 2; i++) { dport = iter->parent_dport; iter = dport->port; dev = device_find_child(&iter->dev, NULL, first_decoder); /* * Ancestor ports are guaranteed to be enumerated before * @port, and all ports have at least one decoder. */ if (WARN_ON(!dev)) continue; cxlsd = to_cxl_switch_decoder(dev); if (i == 0) { /* put cxl_mem.4 second in the decode order */ if (pdev->id == 4) cxlsd->target[1] = dport; else cxlsd->target[0] = dport; } else cxlsd->target[0] = dport; cxld = &cxlsd->cxld; cxld->target_type = CXL_DECODER_HOSTONLYMEM; cxld->flags = CXL_DECODER_F_ENABLE; iter->commit_end = 0; /* * Switch targets 2 endpoints, while host bridge targets * one root port */ if (i == 0) cxld->interleave_ways = 2; else cxld->interleave_ways = 1; cxld->interleave_granularity = 4096; cxld->hpa_range = (struct range) { .start = base, .end = base + size - 1, }; put_device(dev); } } static int mock_cxl_enumerate_decoders(struct cxl_hdm *cxlhdm, struct cxl_endpoint_dvsec_info *info) { struct cxl_port *port = cxlhdm->port; struct cxl_port *parent_port = to_cxl_port(port->dev.parent); int target_count, i; if (is_cxl_endpoint(port)) target_count = 0; else if (is_cxl_root(parent_port)) target_count = NR_CXL_ROOT_PORTS; else target_count = NR_CXL_SWITCH_PORTS; for (i = 0; i < NR_CXL_PORT_DECODERS; i++) { int target_map[CXL_DECODER_MAX_INTERLEAVE] = { 0 }; struct target_map_ctx ctx = { .target_map = target_map, .target_count = target_count, }; struct cxl_decoder *cxld; int rc; if (target_count) { struct cxl_switch_decoder *cxlsd; cxlsd = cxl_switch_decoder_alloc(port, target_count); if (IS_ERR(cxlsd)) { dev_warn(&port->dev, "Failed to allocate the decoder\n"); return PTR_ERR(cxlsd); } cxld = &cxlsd->cxld; } else { struct cxl_endpoint_decoder *cxled; cxled = cxl_endpoint_decoder_alloc(port); if (IS_ERR(cxled)) { dev_warn(&port->dev, "Failed to allocate the decoder\n"); return PTR_ERR(cxled); } cxld = &cxled->cxld; } mock_init_hdm_decoder(cxld); if (target_count) { rc = device_for_each_child(port->uport_dev, &ctx, map_targets); if (rc) { put_device(&cxld->dev); return rc; } } rc = cxl_decoder_add_locked(cxld, target_map); if (rc) { put_device(&cxld->dev); dev_err(&port->dev, "Failed to add decoder\n"); return rc; } rc = cxl_decoder_autoremove(&port->dev, cxld); if (rc) return rc; dev_dbg(&cxld->dev, "Added to port %s\n", dev_name(&port->dev)); } return 0; } static int mock_cxl_port_enumerate_dports(struct cxl_port *port) { struct platform_device **array; int i, array_size; if (port->depth == 1) { if (is_multi_bridge(port->uport_dev)) { array_size = ARRAY_SIZE(cxl_root_port); array = cxl_root_port; } else if (is_single_bridge(port->uport_dev)) { array_size = ARRAY_SIZE(cxl_root_single); array = cxl_root_single; } else { dev_dbg(&port->dev, "%s: unknown bridge type\n", dev_name(port->uport_dev)); return -ENXIO; } } else if (port->depth == 2) { struct cxl_port *parent = to_cxl_port(port->dev.parent); if (is_multi_bridge(parent->uport_dev)) { array_size = ARRAY_SIZE(cxl_switch_dport); array = cxl_switch_dport; } else if (is_single_bridge(parent->uport_dev)) { array_size = ARRAY_SIZE(cxl_swd_single); array = cxl_swd_single; } else { dev_dbg(&port->dev, "%s: unknown bridge type\n", dev_name(port->uport_dev)); return -ENXIO; } } else { dev_WARN_ONCE(&port->dev, 1, "unexpected depth %d\n", port->depth); return -ENXIO; } for (i = 0; i < array_size; i++) { struct platform_device *pdev = array[i]; struct cxl_dport *dport; if (pdev->dev.parent != port->uport_dev) { dev_dbg(&port->dev, "%s: mismatch parent %s\n", dev_name(port->uport_dev), dev_name(pdev->dev.parent)); continue; } dport = devm_cxl_add_dport(port, &pdev->dev, pdev->id, CXL_RESOURCE_NONE); if (IS_ERR(dport)) return PTR_ERR(dport); } return 0; } /* * Faking the cxl_dpa_perf for the memdev when appropriate. */ static void dpa_perf_setup(struct cxl_port *endpoint, struct range *range, struct cxl_dpa_perf *dpa_perf) { dpa_perf->qos_class = FAKE_QTG_ID; dpa_perf->dpa_range = *range; for (int i = 0; i < ACCESS_COORDINATE_MAX; i++) { dpa_perf->coord[i].read_latency = 500; dpa_perf->coord[i].write_latency = 500; dpa_perf->coord[i].read_bandwidth = 1000; dpa_perf->coord[i].write_bandwidth = 1000; } } static void mock_cxl_endpoint_parse_cdat(struct cxl_port *port) { struct cxl_root *cxl_root __free(put_cxl_root) = find_cxl_root(port); struct cxl_memdev *cxlmd = to_cxl_memdev(port->uport_dev); struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlds); struct access_coordinate ep_c[ACCESS_COORDINATE_MAX]; struct range pmem_range = { .start = cxlds->pmem_res.start, .end = cxlds->pmem_res.end, }; struct range ram_range = { .start = cxlds->ram_res.start, .end = cxlds->ram_res.end, }; if (!cxl_root) return; if (range_len(&ram_range)) dpa_perf_setup(port, &ram_range, &mds->ram_perf); if (range_len(&pmem_range)) dpa_perf_setup(port, &pmem_range, &mds->pmem_perf); cxl_memdev_update_perf(cxlmd); /* * This function is here to only test the topology iterator. It serves * no other purpose. */ cxl_endpoint_get_perf_coordinates(port, ep_c); } static struct cxl_mock_ops cxl_mock_ops = { .is_mock_adev = is_mock_adev, .is_mock_bridge = is_mock_bridge, .is_mock_bus = is_mock_bus, .is_mock_port = is_mock_port, .is_mock_dev = is_mock_dev, .acpi_table_parse_cedt = mock_acpi_table_parse_cedt, .acpi_evaluate_integer = mock_acpi_evaluate_integer, .acpi_pci_find_root = mock_acpi_pci_find_root, .devm_cxl_port_enumerate_dports = mock_cxl_port_enumerate_dports, .devm_cxl_setup_hdm = mock_cxl_setup_hdm, .devm_cxl_add_passthrough_decoder = mock_cxl_add_passthrough_decoder, .devm_cxl_enumerate_decoders = mock_cxl_enumerate_decoders, .cxl_endpoint_parse_cdat = mock_cxl_endpoint_parse_cdat, .list = LIST_HEAD_INIT(cxl_mock_ops.list), }; static void mock_companion(struct acpi_device *adev, struct device *dev) { device_initialize(&adev->dev); fwnode_init(&adev->fwnode, NULL); dev->fwnode = &adev->fwnode; adev->fwnode.dev = dev; } #ifndef SZ_64G #define SZ_64G (SZ_32G * 2) #endif static __init int cxl_rch_topo_init(void) { int rc, i; for (i = 0; i < ARRAY_SIZE(cxl_rch); i++) { int idx = NR_CXL_HOST_BRIDGES + NR_CXL_SINGLE_HOST + i; struct acpi_device *adev = &host_bridge[idx]; struct platform_device *pdev; pdev = platform_device_alloc("cxl_host_bridge", idx); if (!pdev) goto err_bridge; mock_companion(adev, &pdev->dev); rc = platform_device_add(pdev); if (rc) { platform_device_put(pdev); goto err_bridge; } cxl_rch[i] = pdev; mock_pci_bus[idx].bridge = &pdev->dev; rc = sysfs_create_link(&pdev->dev.kobj, &pdev->dev.kobj, "firmware_node"); if (rc) goto err_bridge; } return 0; err_bridge: for (i = ARRAY_SIZE(cxl_rch) - 1; i >= 0; i--) { struct platform_device *pdev = cxl_rch[i]; if (!pdev) continue; sysfs_remove_link(&pdev->dev.kobj, "firmware_node"); platform_device_unregister(cxl_rch[i]); } return rc; } static void cxl_rch_topo_exit(void) { int i; for (i = ARRAY_SIZE(cxl_rch) - 1; i >= 0; i--) { struct platform_device *pdev = cxl_rch[i]; if (!pdev) continue; sysfs_remove_link(&pdev->dev.kobj, "firmware_node"); platform_device_unregister(cxl_rch[i]); } } static __init int cxl_single_topo_init(void) { int i, rc; for (i = 0; i < ARRAY_SIZE(cxl_hb_single); i++) { struct acpi_device *adev = &host_bridge[NR_CXL_HOST_BRIDGES + i]; struct platform_device *pdev; pdev = platform_device_alloc("cxl_host_bridge", NR_CXL_HOST_BRIDGES + i); if (!pdev) goto err_bridge; mock_companion(adev, &pdev->dev); rc = platform_device_add(pdev); if (rc) { platform_device_put(pdev); goto err_bridge; } cxl_hb_single[i] = pdev; mock_pci_bus[i + NR_CXL_HOST_BRIDGES].bridge = &pdev->dev; rc = sysfs_create_link(&pdev->dev.kobj, &pdev->dev.kobj, "physical_node"); if (rc) goto err_bridge; } for (i = 0; i < ARRAY_SIZE(cxl_root_single); i++) { struct platform_device *bridge = cxl_hb_single[i % ARRAY_SIZE(cxl_hb_single)]; struct platform_device *pdev; pdev = platform_device_alloc("cxl_root_port", NR_MULTI_ROOT + i); if (!pdev) goto err_port; pdev->dev.parent = &bridge->dev; rc = platform_device_add(pdev); if (rc) { platform_device_put(pdev); goto err_port; } cxl_root_single[i] = pdev; } for (i = 0; i < ARRAY_SIZE(cxl_swu_single); i++) { struct platform_device *root_port = cxl_root_single[i]; struct platform_device *pdev; pdev = platform_device_alloc("cxl_switch_uport", NR_MULTI_ROOT + i); if (!pdev) goto err_uport; pdev->dev.parent = &root_port->dev; rc = platform_device_add(pdev); if (rc) { platform_device_put(pdev); goto err_uport; } cxl_swu_single[i] = pdev; } for (i = 0; i < ARRAY_SIZE(cxl_swd_single); i++) { struct platform_device *uport = cxl_swu_single[i % ARRAY_SIZE(cxl_swu_single)]; struct platform_device *pdev; pdev = platform_device_alloc("cxl_switch_dport", i + NR_MEM_MULTI); if (!pdev) goto err_dport; pdev->dev.parent = &uport->dev; rc = platform_device_add(pdev); if (rc) { platform_device_put(pdev); goto err_dport; } cxl_swd_single[i] = pdev; } return 0; err_dport: for (i = ARRAY_SIZE(cxl_swd_single) - 1; i >= 0; i--) platform_device_unregister(cxl_swd_single[i]); err_uport: for (i = ARRAY_SIZE(cxl_swu_single) - 1; i >= 0; i--) platform_device_unregister(cxl_swu_single[i]); err_port: for (i = ARRAY_SIZE(cxl_root_single) - 1; i >= 0; i--) platform_device_unregister(cxl_root_single[i]); err_bridge: for (i = ARRAY_SIZE(cxl_hb_single) - 1; i >= 0; i--) { struct platform_device *pdev = cxl_hb_single[i]; if (!pdev) continue; sysfs_remove_link(&pdev->dev.kobj, "physical_node"); platform_device_unregister(cxl_hb_single[i]); } return rc; } static void cxl_single_topo_exit(void) { int i; for (i = ARRAY_SIZE(cxl_swd_single) - 1; i >= 0; i--) platform_device_unregister(cxl_swd_single[i]); for (i = ARRAY_SIZE(cxl_swu_single) - 1; i >= 0; i--) platform_device_unregister(cxl_swu_single[i]); for (i = ARRAY_SIZE(cxl_root_single) - 1; i >= 0; i--) platform_device_unregister(cxl_root_single[i]); for (i = ARRAY_SIZE(cxl_hb_single) - 1; i >= 0; i--) { struct platform_device *pdev = cxl_hb_single[i]; if (!pdev) continue; sysfs_remove_link(&pdev->dev.kobj, "physical_node"); platform_device_unregister(cxl_hb_single[i]); } } static void cxl_mem_exit(void) { int i; for (i = ARRAY_SIZE(cxl_rcd) - 1; i >= 0; i--) platform_device_unregister(cxl_rcd[i]); for (i = ARRAY_SIZE(cxl_mem_single) - 1; i >= 0; i--) platform_device_unregister(cxl_mem_single[i]); for (i = ARRAY_SIZE(cxl_mem) - 1; i >= 0; i--) platform_device_unregister(cxl_mem[i]); } static int cxl_mem_init(void) { int i, rc; for (i = 0; i < ARRAY_SIZE(cxl_mem); i++) { struct platform_device *dport = cxl_switch_dport[i]; struct platform_device *pdev; pdev = platform_device_alloc("cxl_mem", i); if (!pdev) goto err_mem; pdev->dev.parent = &dport->dev; set_dev_node(&pdev->dev, i % 2); rc = platform_device_add(pdev); if (rc) { platform_device_put(pdev); goto err_mem; } cxl_mem[i] = pdev; } for (i = 0; i < ARRAY_SIZE(cxl_mem_single); i++) { struct platform_device *dport = cxl_swd_single[i]; struct platform_device *pdev; pdev = platform_device_alloc("cxl_mem", NR_MEM_MULTI + i); if (!pdev) goto err_single; pdev->dev.parent = &dport->dev; set_dev_node(&pdev->dev, i % 2); rc = platform_device_add(pdev); if (rc) { platform_device_put(pdev); goto err_single; } cxl_mem_single[i] = pdev; } for (i = 0; i < ARRAY_SIZE(cxl_rcd); i++) { int idx = NR_MEM_MULTI + NR_MEM_SINGLE + i; struct platform_device *rch = cxl_rch[i]; struct platform_device *pdev; pdev = platform_device_alloc("cxl_rcd", idx); if (!pdev) goto err_rcd; pdev->dev.parent = &rch->dev; set_dev_node(&pdev->dev, i % 2); rc = platform_device_add(pdev); if (rc) { platform_device_put(pdev); goto err_rcd; } cxl_rcd[i] = pdev; } return 0; err_rcd: for (i = ARRAY_SIZE(cxl_rcd) - 1; i >= 0; i--) platform_device_unregister(cxl_rcd[i]); err_single: for (i = ARRAY_SIZE(cxl_mem_single) - 1; i >= 0; i--) platform_device_unregister(cxl_mem_single[i]); err_mem: for (i = ARRAY_SIZE(cxl_mem) - 1; i >= 0; i--) platform_device_unregister(cxl_mem[i]); return rc; } static __init int cxl_test_init(void) { int rc, i; cxl_acpi_test(); cxl_core_test(); cxl_mem_test(); cxl_pmem_test(); cxl_port_test(); register_cxl_mock_ops(&cxl_mock_ops); cxl_mock_pool = gen_pool_create(ilog2(SZ_2M), NUMA_NO_NODE); if (!cxl_mock_pool) { rc = -ENOMEM; goto err_gen_pool_create; } rc = gen_pool_add(cxl_mock_pool, iomem_resource.end + 1 - SZ_64G, SZ_64G, NUMA_NO_NODE); if (rc) goto err_gen_pool_add; if (interleave_arithmetic == 1) { cfmws_start = CFMWS_XOR_ARRAY_START; cfmws_end = CFMWS_XOR_ARRAY_END; } else { cfmws_start = CFMWS_MOD_ARRAY_START; cfmws_end = CFMWS_MOD_ARRAY_END; } rc = populate_cedt(); if (rc) goto err_populate; for (i = 0; i < ARRAY_SIZE(cxl_host_bridge); i++) { struct acpi_device *adev = &host_bridge[i]; struct platform_device *pdev; pdev = platform_device_alloc("cxl_host_bridge", i); if (!pdev) goto err_bridge; mock_companion(adev, &pdev->dev); rc = platform_device_add(pdev); if (rc) { platform_device_put(pdev); goto err_bridge; } cxl_host_bridge[i] = pdev; mock_pci_bus[i].bridge = &pdev->dev; rc = sysfs_create_link(&pdev->dev.kobj, &pdev->dev.kobj, "physical_node"); if (rc) goto err_bridge; } for (i = 0; i < ARRAY_SIZE(cxl_root_port); i++) { struct platform_device *bridge = cxl_host_bridge[i % ARRAY_SIZE(cxl_host_bridge)]; struct platform_device *pdev; pdev = platform_device_alloc("cxl_root_port", i); if (!pdev) goto err_port; pdev->dev.parent = &bridge->dev; rc = platform_device_add(pdev); if (rc) { platform_device_put(pdev); goto err_port; } cxl_root_port[i] = pdev; } BUILD_BUG_ON(ARRAY_SIZE(cxl_switch_uport) != ARRAY_SIZE(cxl_root_port)); for (i = 0; i < ARRAY_SIZE(cxl_switch_uport); i++) { struct platform_device *root_port = cxl_root_port[i]; struct platform_device *pdev; pdev = platform_device_alloc("cxl_switch_uport", i); if (!pdev) goto err_uport; pdev->dev.parent = &root_port->dev; rc = platform_device_add(pdev); if (rc) { platform_device_put(pdev); goto err_uport; } cxl_switch_uport[i] = pdev; } for (i = 0; i < ARRAY_SIZE(cxl_switch_dport); i++) { struct platform_device *uport = cxl_switch_uport[i % ARRAY_SIZE(cxl_switch_uport)]; struct platform_device *pdev; pdev = platform_device_alloc("cxl_switch_dport", i); if (!pdev) goto err_dport; pdev->dev.parent = &uport->dev; rc = platform_device_add(pdev); if (rc) { platform_device_put(pdev); goto err_dport; } cxl_switch_dport[i] = pdev; } rc = cxl_single_topo_init(); if (rc) goto err_dport; rc = cxl_rch_topo_init(); if (rc) goto err_single; cxl_acpi = platform_device_alloc("cxl_acpi", 0); if (!cxl_acpi) goto err_rch; mock_companion(&acpi0017_mock, &cxl_acpi->dev); acpi0017_mock.dev.bus = &platform_bus_type; rc = platform_device_add(cxl_acpi); if (rc) goto err_root; rc = cxl_mem_init(); if (rc) goto err_root; return 0; err_root: platform_device_put(cxl_acpi); err_rch: cxl_rch_topo_exit(); err_single: cxl_single_topo_exit(); err_dport: for (i = ARRAY_SIZE(cxl_switch_dport) - 1; i >= 0; i--) platform_device_unregister(cxl_switch_dport[i]); err_uport: for (i = ARRAY_SIZE(cxl_switch_uport) - 1; i >= 0; i--) platform_device_unregister(cxl_switch_uport[i]); err_port: for (i = ARRAY_SIZE(cxl_root_port) - 1; i >= 0; i--) platform_device_unregister(cxl_root_port[i]); err_bridge: for (i = ARRAY_SIZE(cxl_host_bridge) - 1; i >= 0; i--) { struct platform_device *pdev = cxl_host_bridge[i]; if (!pdev) continue; sysfs_remove_link(&pdev->dev.kobj, "physical_node"); platform_device_unregister(cxl_host_bridge[i]); } err_populate: depopulate_all_mock_resources(); err_gen_pool_add: gen_pool_destroy(cxl_mock_pool); err_gen_pool_create: unregister_cxl_mock_ops(&cxl_mock_ops); return rc; } static __exit void cxl_test_exit(void) { int i; cxl_mem_exit(); platform_device_unregister(cxl_acpi); cxl_rch_topo_exit(); cxl_single_topo_exit(); for (i = ARRAY_SIZE(cxl_switch_dport) - 1; i >= 0; i--) platform_device_unregister(cxl_switch_dport[i]); for (i = ARRAY_SIZE(cxl_switch_uport) - 1; i >= 0; i--) platform_device_unregister(cxl_switch_uport[i]); for (i = ARRAY_SIZE(cxl_root_port) - 1; i >= 0; i--) platform_device_unregister(cxl_root_port[i]); for (i = ARRAY_SIZE(cxl_host_bridge) - 1; i >= 0; i--) { struct platform_device *pdev = cxl_host_bridge[i]; if (!pdev) continue; sysfs_remove_link(&pdev->dev.kobj, "physical_node"); platform_device_unregister(cxl_host_bridge[i]); } depopulate_all_mock_resources(); gen_pool_destroy(cxl_mock_pool); unregister_cxl_mock_ops(&cxl_mock_ops); } module_param(interleave_arithmetic, int, 0444); MODULE_PARM_DESC(interleave_arithmetic, "Modulo:0, XOR:1"); module_init(cxl_test_init); module_exit(cxl_test_exit); MODULE_LICENSE("GPL v2"); MODULE_IMPORT_NS(ACPI); MODULE_IMPORT_NS(CXL);