mmconfig-shared.c 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668
  1. /*
  2. * mmconfig-shared.c - Low-level direct PCI config space access via
  3. * MMCONFIG - common code between i386 and x86-64.
  4. *
  5. * This code does:
  6. * - known chipset handling
  7. * - ACPI decoding and validation
  8. *
  9. * Per-architecture code takes care of the mappings and accesses
  10. * themselves.
  11. */
  12. #include <linux/pci.h>
  13. #include <linux/init.h>
  14. #include <linux/acpi.h>
  15. #include <linux/sfi_acpi.h>
  16. #include <linux/bitmap.h>
  17. #include <linux/dmi.h>
  18. #include <linux/slab.h>
  19. #include <asm/e820.h>
  20. #include <asm/pci_x86.h>
  21. #include <asm/acpi.h>
  22. #define PREFIX "PCI: "
  23. /* Indicate if the mmcfg resources have been placed into the resource table. */
  24. static int __initdata pci_mmcfg_resources_inserted;
  25. LIST_HEAD(pci_mmcfg_list);
  26. static __init void pci_mmconfig_remove(struct pci_mmcfg_region *cfg)
  27. {
  28. if (cfg->res.parent)
  29. release_resource(&cfg->res);
  30. list_del(&cfg->list);
  31. kfree(cfg);
  32. }
  33. static __init void free_all_mmcfg(void)
  34. {
  35. struct pci_mmcfg_region *cfg, *tmp;
  36. pci_mmcfg_arch_free();
  37. list_for_each_entry_safe(cfg, tmp, &pci_mmcfg_list, list)
  38. pci_mmconfig_remove(cfg);
  39. }
  40. static __init void list_add_sorted(struct pci_mmcfg_region *new)
  41. {
  42. struct pci_mmcfg_region *cfg;
  43. /* keep list sorted by segment and starting bus number */
  44. list_for_each_entry(cfg, &pci_mmcfg_list, list) {
  45. if (cfg->segment > new->segment ||
  46. (cfg->segment == new->segment &&
  47. cfg->start_bus >= new->start_bus)) {
  48. list_add_tail(&new->list, &cfg->list);
  49. return;
  50. }
  51. }
  52. list_add_tail(&new->list, &pci_mmcfg_list);
  53. }
  54. static __init struct pci_mmcfg_region *pci_mmconfig_add(int segment, int start,
  55. int end, u64 addr)
  56. {
  57. struct pci_mmcfg_region *new;
  58. struct resource *res;
  59. if (addr == 0)
  60. return NULL;
  61. new = kzalloc(sizeof(*new), GFP_KERNEL);
  62. if (!new)
  63. return NULL;
  64. new->address = addr;
  65. new->segment = segment;
  66. new->start_bus = start;
  67. new->end_bus = end;
  68. list_add_sorted(new);
  69. res = &new->res;
  70. res->start = addr + PCI_MMCFG_BUS_OFFSET(start);
  71. res->end = addr + PCI_MMCFG_BUS_OFFSET(end + 1) - 1;
  72. res->flags = IORESOURCE_MEM | IORESOURCE_BUSY;
  73. snprintf(new->name, PCI_MMCFG_RESOURCE_NAME_LEN,
  74. "PCI MMCONFIG %04x [bus %02x-%02x]", segment, start, end);
  75. res->name = new->name;
  76. printk(KERN_INFO PREFIX "MMCONFIG for domain %04x [bus %02x-%02x] at "
  77. "%pR (base %#lx)\n", segment, start, end, &new->res,
  78. (unsigned long) addr);
  79. return new;
  80. }
  81. struct pci_mmcfg_region *pci_mmconfig_lookup(int segment, int bus)
  82. {
  83. struct pci_mmcfg_region *cfg;
  84. list_for_each_entry(cfg, &pci_mmcfg_list, list)
  85. if (cfg->segment == segment &&
  86. cfg->start_bus <= bus && bus <= cfg->end_bus)
  87. return cfg;
  88. return NULL;
  89. }
  90. static const char __init *pci_mmcfg_e7520(void)
  91. {
  92. u32 win;
  93. raw_pci_ops->read(0, 0, PCI_DEVFN(0, 0), 0xce, 2, &win);
  94. win = win & 0xf000;
  95. if (win == 0x0000 || win == 0xf000)
  96. return NULL;
  97. if (pci_mmconfig_add(0, 0, 255, win << 16) == NULL)
  98. return NULL;
  99. return "Intel Corporation E7520 Memory Controller Hub";
  100. }
  101. static const char __init *pci_mmcfg_intel_945(void)
  102. {
  103. u32 pciexbar, mask = 0, len = 0;
  104. raw_pci_ops->read(0, 0, PCI_DEVFN(0, 0), 0x48, 4, &pciexbar);
  105. /* Enable bit */
  106. if (!(pciexbar & 1))
  107. return NULL;
  108. /* Size bits */
  109. switch ((pciexbar >> 1) & 3) {
  110. case 0:
  111. mask = 0xf0000000U;
  112. len = 0x10000000U;
  113. break;
  114. case 1:
  115. mask = 0xf8000000U;
  116. len = 0x08000000U;
  117. break;
  118. case 2:
  119. mask = 0xfc000000U;
  120. len = 0x04000000U;
  121. break;
  122. default:
  123. return NULL;
  124. }
  125. /* Errata #2, things break when not aligned on a 256Mb boundary */
  126. /* Can only happen in 64M/128M mode */
  127. if ((pciexbar & mask) & 0x0fffffffU)
  128. return NULL;
  129. /* Don't hit the APIC registers and their friends */
  130. if ((pciexbar & mask) >= 0xf0000000U)
  131. return NULL;
  132. if (pci_mmconfig_add(0, 0, (len >> 20) - 1, pciexbar & mask) == NULL)
  133. return NULL;
  134. return "Intel Corporation 945G/GZ/P/PL Express Memory Controller Hub";
  135. }
  136. static const char __init *pci_mmcfg_amd_fam10h(void)
  137. {
  138. u32 low, high, address;
  139. u64 base, msr;
  140. int i;
  141. unsigned segnbits = 0, busnbits, end_bus;
  142. if (!(pci_probe & PCI_CHECK_ENABLE_AMD_MMCONF))
  143. return NULL;
  144. address = MSR_FAM10H_MMIO_CONF_BASE;
  145. if (rdmsr_safe(address, &low, &high))
  146. return NULL;
  147. msr = high;
  148. msr <<= 32;
  149. msr |= low;
  150. /* mmconfig is not enable */
  151. if (!(msr & FAM10H_MMIO_CONF_ENABLE))
  152. return NULL;
  153. base = msr & (FAM10H_MMIO_CONF_BASE_MASK<<FAM10H_MMIO_CONF_BASE_SHIFT);
  154. busnbits = (msr >> FAM10H_MMIO_CONF_BUSRANGE_SHIFT) &
  155. FAM10H_MMIO_CONF_BUSRANGE_MASK;
  156. /*
  157. * only handle bus 0 ?
  158. * need to skip it
  159. */
  160. if (!busnbits)
  161. return NULL;
  162. if (busnbits > 8) {
  163. segnbits = busnbits - 8;
  164. busnbits = 8;
  165. }
  166. end_bus = (1 << busnbits) - 1;
  167. for (i = 0; i < (1 << segnbits); i++)
  168. if (pci_mmconfig_add(i, 0, end_bus,
  169. base + (1<<28) * i) == NULL) {
  170. free_all_mmcfg();
  171. return NULL;
  172. }
  173. return "AMD Family 10h NB";
  174. }
  175. static bool __initdata mcp55_checked;
  176. static const char __init *pci_mmcfg_nvidia_mcp55(void)
  177. {
  178. int bus;
  179. int mcp55_mmconf_found = 0;
  180. static const u32 extcfg_regnum = 0x90;
  181. static const u32 extcfg_regsize = 4;
  182. static const u32 extcfg_enable_mask = 1<<31;
  183. static const u32 extcfg_start_mask = 0xff<<16;
  184. static const int extcfg_start_shift = 16;
  185. static const u32 extcfg_size_mask = 0x3<<28;
  186. static const int extcfg_size_shift = 28;
  187. static const int extcfg_sizebus[] = {0x100, 0x80, 0x40, 0x20};
  188. static const u32 extcfg_base_mask[] = {0x7ff8, 0x7ffc, 0x7ffe, 0x7fff};
  189. static const int extcfg_base_lshift = 25;
  190. /*
  191. * do check if amd fam10h already took over
  192. */
  193. if (!acpi_disabled || !list_empty(&pci_mmcfg_list) || mcp55_checked)
  194. return NULL;
  195. mcp55_checked = true;
  196. for (bus = 0; bus < 256; bus++) {
  197. u64 base;
  198. u32 l, extcfg;
  199. u16 vendor, device;
  200. int start, size_index, end;
  201. raw_pci_ops->read(0, bus, PCI_DEVFN(0, 0), 0, 4, &l);
  202. vendor = l & 0xffff;
  203. device = (l >> 16) & 0xffff;
  204. if (PCI_VENDOR_ID_NVIDIA != vendor || 0x0369 != device)
  205. continue;
  206. raw_pci_ops->read(0, bus, PCI_DEVFN(0, 0), extcfg_regnum,
  207. extcfg_regsize, &extcfg);
  208. if (!(extcfg & extcfg_enable_mask))
  209. continue;
  210. size_index = (extcfg & extcfg_size_mask) >> extcfg_size_shift;
  211. base = extcfg & extcfg_base_mask[size_index];
  212. /* base could > 4G */
  213. base <<= extcfg_base_lshift;
  214. start = (extcfg & extcfg_start_mask) >> extcfg_start_shift;
  215. end = start + extcfg_sizebus[size_index] - 1;
  216. if (pci_mmconfig_add(0, start, end, base) == NULL)
  217. continue;
  218. mcp55_mmconf_found++;
  219. }
  220. if (!mcp55_mmconf_found)
  221. return NULL;
  222. return "nVidia MCP55";
  223. }
  224. struct pci_mmcfg_hostbridge_probe {
  225. u32 bus;
  226. u32 devfn;
  227. u32 vendor;
  228. u32 device;
  229. const char *(*probe)(void);
  230. };
  231. static struct pci_mmcfg_hostbridge_probe pci_mmcfg_probes[] __initdata = {
  232. { 0, PCI_DEVFN(0, 0), PCI_VENDOR_ID_INTEL,
  233. PCI_DEVICE_ID_INTEL_E7520_MCH, pci_mmcfg_e7520 },
  234. { 0, PCI_DEVFN(0, 0), PCI_VENDOR_ID_INTEL,
  235. PCI_DEVICE_ID_INTEL_82945G_HB, pci_mmcfg_intel_945 },
  236. { 0, PCI_DEVFN(0x18, 0), PCI_VENDOR_ID_AMD,
  237. 0x1200, pci_mmcfg_amd_fam10h },
  238. { 0xff, PCI_DEVFN(0, 0), PCI_VENDOR_ID_AMD,
  239. 0x1200, pci_mmcfg_amd_fam10h },
  240. { 0, PCI_DEVFN(0, 0), PCI_VENDOR_ID_NVIDIA,
  241. 0x0369, pci_mmcfg_nvidia_mcp55 },
  242. };
  243. static void __init pci_mmcfg_check_end_bus_number(void)
  244. {
  245. struct pci_mmcfg_region *cfg, *cfgx;
  246. /* Fixup overlaps */
  247. list_for_each_entry(cfg, &pci_mmcfg_list, list) {
  248. if (cfg->end_bus < cfg->start_bus)
  249. cfg->end_bus = 255;
  250. /* Don't access the list head ! */
  251. if (cfg->list.next == &pci_mmcfg_list)
  252. break;
  253. cfgx = list_entry(cfg->list.next, typeof(*cfg), list);
  254. if (cfg->end_bus >= cfgx->start_bus)
  255. cfg->end_bus = cfgx->start_bus - 1;
  256. }
  257. }
  258. static int __init pci_mmcfg_check_hostbridge(void)
  259. {
  260. u32 l;
  261. u32 bus, devfn;
  262. u16 vendor, device;
  263. int i;
  264. const char *name;
  265. if (!raw_pci_ops)
  266. return 0;
  267. free_all_mmcfg();
  268. for (i = 0; i < ARRAY_SIZE(pci_mmcfg_probes); i++) {
  269. bus = pci_mmcfg_probes[i].bus;
  270. devfn = pci_mmcfg_probes[i].devfn;
  271. raw_pci_ops->read(0, bus, devfn, 0, 4, &l);
  272. vendor = l & 0xffff;
  273. device = (l >> 16) & 0xffff;
  274. name = NULL;
  275. if (pci_mmcfg_probes[i].vendor == vendor &&
  276. pci_mmcfg_probes[i].device == device)
  277. name = pci_mmcfg_probes[i].probe();
  278. if (name)
  279. printk(KERN_INFO PREFIX "%s with MMCONFIG support\n",
  280. name);
  281. }
  282. /* some end_bus_number is crazy, fix it */
  283. pci_mmcfg_check_end_bus_number();
  284. return !list_empty(&pci_mmcfg_list);
  285. }
  286. static void __init pci_mmcfg_insert_resources(void)
  287. {
  288. struct pci_mmcfg_region *cfg;
  289. list_for_each_entry(cfg, &pci_mmcfg_list, list)
  290. insert_resource(&iomem_resource, &cfg->res);
  291. /* Mark that the resources have been inserted. */
  292. pci_mmcfg_resources_inserted = 1;
  293. }
  294. static acpi_status __init check_mcfg_resource(struct acpi_resource *res,
  295. void *data)
  296. {
  297. struct resource *mcfg_res = data;
  298. struct acpi_resource_address64 address;
  299. acpi_status status;
  300. if (res->type == ACPI_RESOURCE_TYPE_FIXED_MEMORY32) {
  301. struct acpi_resource_fixed_memory32 *fixmem32 =
  302. &res->data.fixed_memory32;
  303. if (!fixmem32)
  304. return AE_OK;
  305. if ((mcfg_res->start >= fixmem32->address) &&
  306. (mcfg_res->end < (fixmem32->address +
  307. fixmem32->address_length))) {
  308. mcfg_res->flags = 1;
  309. return AE_CTRL_TERMINATE;
  310. }
  311. }
  312. if ((res->type != ACPI_RESOURCE_TYPE_ADDRESS32) &&
  313. (res->type != ACPI_RESOURCE_TYPE_ADDRESS64))
  314. return AE_OK;
  315. status = acpi_resource_to_address64(res, &address);
  316. if (ACPI_FAILURE(status) ||
  317. (address.address_length <= 0) ||
  318. (address.resource_type != ACPI_MEMORY_RANGE))
  319. return AE_OK;
  320. if ((mcfg_res->start >= address.minimum) &&
  321. (mcfg_res->end < (address.minimum + address.address_length))) {
  322. mcfg_res->flags = 1;
  323. return AE_CTRL_TERMINATE;
  324. }
  325. return AE_OK;
  326. }
  327. static acpi_status __init find_mboard_resource(acpi_handle handle, u32 lvl,
  328. void *context, void **rv)
  329. {
  330. struct resource *mcfg_res = context;
  331. acpi_walk_resources(handle, METHOD_NAME__CRS,
  332. check_mcfg_resource, context);
  333. if (mcfg_res->flags)
  334. return AE_CTRL_TERMINATE;
  335. return AE_OK;
  336. }
  337. static int __init is_acpi_reserved(u64 start, u64 end, unsigned not_used)
  338. {
  339. struct resource mcfg_res;
  340. mcfg_res.start = start;
  341. mcfg_res.end = end - 1;
  342. mcfg_res.flags = 0;
  343. acpi_get_devices("PNP0C01", find_mboard_resource, &mcfg_res, NULL);
  344. if (!mcfg_res.flags)
  345. acpi_get_devices("PNP0C02", find_mboard_resource, &mcfg_res,
  346. NULL);
  347. return mcfg_res.flags;
  348. }
  349. typedef int (*check_reserved_t)(u64 start, u64 end, unsigned type);
  350. static int __init is_mmconf_reserved(check_reserved_t is_reserved,
  351. struct pci_mmcfg_region *cfg, int with_e820)
  352. {
  353. u64 addr = cfg->res.start;
  354. u64 size = resource_size(&cfg->res);
  355. u64 old_size = size;
  356. int valid = 0, num_buses;
  357. while (!is_reserved(addr, addr + size, E820_RESERVED)) {
  358. size >>= 1;
  359. if (size < (16UL<<20))
  360. break;
  361. }
  362. if (size >= (16UL<<20) || size == old_size) {
  363. printk(KERN_INFO PREFIX "MMCONFIG at %pR reserved in %s\n",
  364. &cfg->res,
  365. with_e820 ? "E820" : "ACPI motherboard resources");
  366. valid = 1;
  367. if (old_size != size) {
  368. /* update end_bus */
  369. cfg->end_bus = cfg->start_bus + ((size>>20) - 1);
  370. num_buses = cfg->end_bus - cfg->start_bus + 1;
  371. cfg->res.end = cfg->res.start +
  372. PCI_MMCFG_BUS_OFFSET(num_buses) - 1;
  373. snprintf(cfg->name, PCI_MMCFG_RESOURCE_NAME_LEN,
  374. "PCI MMCONFIG %04x [bus %02x-%02x]",
  375. cfg->segment, cfg->start_bus, cfg->end_bus);
  376. printk(KERN_INFO PREFIX
  377. "MMCONFIG for %04x [bus%02x-%02x] "
  378. "at %pR (base %#lx) (size reduced!)\n",
  379. cfg->segment, cfg->start_bus, cfg->end_bus,
  380. &cfg->res, (unsigned long) cfg->address);
  381. }
  382. }
  383. return valid;
  384. }
  385. static void __init pci_mmcfg_reject_broken(int early)
  386. {
  387. struct pci_mmcfg_region *cfg;
  388. list_for_each_entry(cfg, &pci_mmcfg_list, list) {
  389. int valid = 0;
  390. if (!early && !acpi_disabled) {
  391. valid = is_mmconf_reserved(is_acpi_reserved, cfg, 0);
  392. if (valid)
  393. continue;
  394. else
  395. printk(KERN_ERR FW_BUG PREFIX
  396. "MMCONFIG at %pR not reserved in "
  397. "ACPI motherboard resources\n",
  398. &cfg->res);
  399. }
  400. /* Don't try to do this check unless configuration
  401. type 1 is available. how about type 2 ?*/
  402. if (raw_pci_ops)
  403. valid = is_mmconf_reserved(e820_all_mapped, cfg, 1);
  404. if (!valid)
  405. goto reject;
  406. }
  407. return;
  408. reject:
  409. printk(KERN_INFO PREFIX "not using MMCONFIG\n");
  410. free_all_mmcfg();
  411. }
  412. static int __initdata known_bridge;
  413. static int __init acpi_mcfg_check_entry(struct acpi_table_mcfg *mcfg,
  414. struct acpi_mcfg_allocation *cfg)
  415. {
  416. int year;
  417. if (cfg->address < 0xFFFFFFFF)
  418. return 0;
  419. if (!strcmp(mcfg->header.oem_id, "SGI") ||
  420. !strcmp(mcfg->header.oem_id, "SGI2"))
  421. return 0;
  422. if (mcfg->header.revision >= 1) {
  423. if (dmi_get_date(DMI_BIOS_DATE, &year, NULL, NULL) &&
  424. year >= 2010)
  425. return 0;
  426. }
  427. printk(KERN_ERR PREFIX "MCFG region for %04x [bus %02x-%02x] at %#llx "
  428. "is above 4GB, ignored\n", cfg->pci_segment,
  429. cfg->start_bus_number, cfg->end_bus_number, cfg->address);
  430. return -EINVAL;
  431. }
  432. static int __init pci_parse_mcfg(struct acpi_table_header *header)
  433. {
  434. struct acpi_table_mcfg *mcfg;
  435. struct acpi_mcfg_allocation *cfg_table, *cfg;
  436. unsigned long i;
  437. int entries;
  438. if (!header)
  439. return -EINVAL;
  440. mcfg = (struct acpi_table_mcfg *)header;
  441. /* how many config structures do we have */
  442. free_all_mmcfg();
  443. entries = 0;
  444. i = header->length - sizeof(struct acpi_table_mcfg);
  445. while (i >= sizeof(struct acpi_mcfg_allocation)) {
  446. entries++;
  447. i -= sizeof(struct acpi_mcfg_allocation);
  448. };
  449. if (entries == 0) {
  450. printk(KERN_ERR PREFIX "MMCONFIG has no entries\n");
  451. return -ENODEV;
  452. }
  453. cfg_table = (struct acpi_mcfg_allocation *) &mcfg[1];
  454. for (i = 0; i < entries; i++) {
  455. cfg = &cfg_table[i];
  456. if (acpi_mcfg_check_entry(mcfg, cfg)) {
  457. free_all_mmcfg();
  458. return -ENODEV;
  459. }
  460. if (pci_mmconfig_add(cfg->pci_segment, cfg->start_bus_number,
  461. cfg->end_bus_number, cfg->address) == NULL) {
  462. printk(KERN_WARNING PREFIX
  463. "no memory for MCFG entries\n");
  464. free_all_mmcfg();
  465. return -ENOMEM;
  466. }
  467. }
  468. return 0;
  469. }
  470. static void __init __pci_mmcfg_init(int early)
  471. {
  472. /* MMCONFIG disabled */
  473. if ((pci_probe & PCI_PROBE_MMCONF) == 0)
  474. return;
  475. /* MMCONFIG already enabled */
  476. if (!early && !(pci_probe & PCI_PROBE_MASK & ~PCI_PROBE_MMCONF))
  477. return;
  478. /* for late to exit */
  479. if (known_bridge)
  480. return;
  481. if (early) {
  482. if (pci_mmcfg_check_hostbridge())
  483. known_bridge = 1;
  484. }
  485. if (!known_bridge)
  486. acpi_sfi_table_parse(ACPI_SIG_MCFG, pci_parse_mcfg);
  487. pci_mmcfg_reject_broken(early);
  488. if (list_empty(&pci_mmcfg_list))
  489. return;
  490. if (pcibios_last_bus < 0) {
  491. const struct pci_mmcfg_region *cfg;
  492. list_for_each_entry(cfg, &pci_mmcfg_list, list) {
  493. if (cfg->segment)
  494. break;
  495. pcibios_last_bus = cfg->end_bus;
  496. }
  497. }
  498. if (pci_mmcfg_arch_init())
  499. pci_probe = (pci_probe & ~PCI_PROBE_MASK) | PCI_PROBE_MMCONF;
  500. else {
  501. /*
  502. * Signal not to attempt to insert mmcfg resources because
  503. * the architecture mmcfg setup could not initialize.
  504. */
  505. pci_mmcfg_resources_inserted = 1;
  506. }
  507. }
  508. void __init pci_mmcfg_early_init(void)
  509. {
  510. __pci_mmcfg_init(1);
  511. }
  512. void __init pci_mmcfg_late_init(void)
  513. {
  514. __pci_mmcfg_init(0);
  515. }
  516. static int __init pci_mmcfg_late_insert_resources(void)
  517. {
  518. /*
  519. * If resources are already inserted or we are not using MMCONFIG,
  520. * don't insert the resources.
  521. */
  522. if ((pci_mmcfg_resources_inserted == 1) ||
  523. (pci_probe & PCI_PROBE_MMCONF) == 0 ||
  524. list_empty(&pci_mmcfg_list))
  525. return 1;
  526. /*
  527. * Attempt to insert the mmcfg resources but not with the busy flag
  528. * marked so it won't cause request errors when __request_region is
  529. * called.
  530. */
  531. pci_mmcfg_insert_resources();
  532. return 0;
  533. }
  534. /*
  535. * Perform MMCONFIG resource insertion after PCI initialization to allow for
  536. * misprogrammed MCFG tables that state larger sizes but actually conflict
  537. * with other system resources.
  538. */
  539. late_initcall(pci_mmcfg_late_insert_resources);