quirks.c 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558
  1. /*
  2. * This file contains work-arounds for x86 and x86_64 platform bugs.
  3. */
  4. #include <linux/pci.h>
  5. #include <linux/irq.h>
  6. #include <asm/hpet.h>
  7. #if defined(CONFIG_X86_IO_APIC) && defined(CONFIG_SMP) && defined(CONFIG_PCI)
  8. static void __devinit quirk_intel_irqbalance(struct pci_dev *dev)
  9. {
  10. u8 config, rev;
  11. u16 word;
  12. /* BIOS may enable hardware IRQ balancing for
  13. * E7520/E7320/E7525(revision ID 0x9 and below)
  14. * based platforms.
  15. * Disable SW irqbalance/affinity on those platforms.
  16. */
  17. pci_read_config_byte(dev, PCI_CLASS_REVISION, &rev);
  18. if (rev > 0x9)
  19. return;
  20. /* enable access to config space*/
  21. pci_read_config_byte(dev, 0xf4, &config);
  22. pci_write_config_byte(dev, 0xf4, config|0x2);
  23. /*
  24. * read xTPR register. We may not have a pci_dev for device 8
  25. * because it might be hidden until the above write.
  26. */
  27. pci_bus_read_config_word(dev->bus, PCI_DEVFN(8, 0), 0x4c, &word);
  28. if (!(word & (1 << 13))) {
  29. dev_info(&dev->dev, "Intel E7520/7320/7525 detected; "
  30. "disabling irq balancing and affinity\n");
  31. noirqdebug_setup("");
  32. #ifdef CONFIG_PROC_FS
  33. no_irq_affinity = 1;
  34. #endif
  35. }
  36. /* put back the original value for config space*/
  37. if (!(config & 0x2))
  38. pci_write_config_byte(dev, 0xf4, config);
  39. }
  40. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7320_MCH,
  41. quirk_intel_irqbalance);
  42. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7525_MCH,
  43. quirk_intel_irqbalance);
  44. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7520_MCH,
  45. quirk_intel_irqbalance);
  46. #endif
  47. #if defined(CONFIG_HPET_TIMER)
  48. unsigned long force_hpet_address;
  49. static enum {
  50. NONE_FORCE_HPET_RESUME,
  51. OLD_ICH_FORCE_HPET_RESUME,
  52. ICH_FORCE_HPET_RESUME,
  53. VT8237_FORCE_HPET_RESUME,
  54. NVIDIA_FORCE_HPET_RESUME,
  55. ATI_FORCE_HPET_RESUME,
  56. } force_hpet_resume_type;
  57. static void __iomem *rcba_base;
  58. static void ich_force_hpet_resume(void)
  59. {
  60. u32 val;
  61. if (!force_hpet_address)
  62. return;
  63. BUG_ON(rcba_base == NULL);
  64. /* read the Function Disable register, dword mode only */
  65. val = readl(rcba_base + 0x3404);
  66. if (!(val & 0x80)) {
  67. /* HPET disabled in HPTC. Trying to enable */
  68. writel(val | 0x80, rcba_base + 0x3404);
  69. }
  70. val = readl(rcba_base + 0x3404);
  71. if (!(val & 0x80))
  72. BUG();
  73. else
  74. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  75. return;
  76. }
  77. static void ich_force_enable_hpet(struct pci_dev *dev)
  78. {
  79. u32 val;
  80. u32 uninitialized_var(rcba);
  81. int err = 0;
  82. if (hpet_address || force_hpet_address)
  83. return;
  84. pci_read_config_dword(dev, 0xF0, &rcba);
  85. rcba &= 0xFFFFC000;
  86. if (rcba == 0) {
  87. dev_printk(KERN_DEBUG, &dev->dev, "RCBA disabled; "
  88. "cannot force enable HPET\n");
  89. return;
  90. }
  91. /* use bits 31:14, 16 kB aligned */
  92. rcba_base = ioremap_nocache(rcba, 0x4000);
  93. if (rcba_base == NULL) {
  94. dev_printk(KERN_DEBUG, &dev->dev, "ioremap failed; "
  95. "cannot force enable HPET\n");
  96. return;
  97. }
  98. /* read the Function Disable register, dword mode only */
  99. val = readl(rcba_base + 0x3404);
  100. if (val & 0x80) {
  101. /* HPET is enabled in HPTC. Just not reported by BIOS */
  102. val = val & 0x3;
  103. force_hpet_address = 0xFED00000 | (val << 12);
  104. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  105. "0x%lx\n", force_hpet_address);
  106. iounmap(rcba_base);
  107. return;
  108. }
  109. /* HPET disabled in HPTC. Trying to enable */
  110. writel(val | 0x80, rcba_base + 0x3404);
  111. val = readl(rcba_base + 0x3404);
  112. if (!(val & 0x80)) {
  113. err = 1;
  114. } else {
  115. val = val & 0x3;
  116. force_hpet_address = 0xFED00000 | (val << 12);
  117. }
  118. if (err) {
  119. force_hpet_address = 0;
  120. iounmap(rcba_base);
  121. dev_printk(KERN_DEBUG, &dev->dev,
  122. "Failed to force enable HPET\n");
  123. } else {
  124. force_hpet_resume_type = ICH_FORCE_HPET_RESUME;
  125. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  126. "0x%lx\n", force_hpet_address);
  127. }
  128. }
  129. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ESB2_0,
  130. ich_force_enable_hpet);
  131. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH6_0,
  132. ich_force_enable_hpet);
  133. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH6_1,
  134. ich_force_enable_hpet);
  135. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_0,
  136. ich_force_enable_hpet);
  137. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_1,
  138. ich_force_enable_hpet);
  139. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_31,
  140. ich_force_enable_hpet);
  141. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH8_1,
  142. ich_force_enable_hpet);
  143. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH8_4,
  144. ich_force_enable_hpet);
  145. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH9_7,
  146. ich_force_enable_hpet);
  147. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, 0x3a16, /* ICH10 */
  148. ich_force_enable_hpet);
  149. static struct pci_dev *cached_dev;
  150. static void hpet_print_force_info(void)
  151. {
  152. printk(KERN_INFO "HPET not enabled in BIOS. "
  153. "You might try hpet=force boot option\n");
  154. }
  155. static void old_ich_force_hpet_resume(void)
  156. {
  157. u32 val;
  158. u32 uninitialized_var(gen_cntl);
  159. if (!force_hpet_address || !cached_dev)
  160. return;
  161. pci_read_config_dword(cached_dev, 0xD0, &gen_cntl);
  162. gen_cntl &= (~(0x7 << 15));
  163. gen_cntl |= (0x4 << 15);
  164. pci_write_config_dword(cached_dev, 0xD0, gen_cntl);
  165. pci_read_config_dword(cached_dev, 0xD0, &gen_cntl);
  166. val = gen_cntl >> 15;
  167. val &= 0x7;
  168. if (val == 0x4)
  169. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  170. else
  171. BUG();
  172. }
  173. static void old_ich_force_enable_hpet(struct pci_dev *dev)
  174. {
  175. u32 val;
  176. u32 uninitialized_var(gen_cntl);
  177. if (hpet_address || force_hpet_address)
  178. return;
  179. pci_read_config_dword(dev, 0xD0, &gen_cntl);
  180. /*
  181. * Bit 17 is HPET enable bit.
  182. * Bit 16:15 control the HPET base address.
  183. */
  184. val = gen_cntl >> 15;
  185. val &= 0x7;
  186. if (val & 0x4) {
  187. val &= 0x3;
  188. force_hpet_address = 0xFED00000 | (val << 12);
  189. dev_printk(KERN_DEBUG, &dev->dev, "HPET at 0x%lx\n",
  190. force_hpet_address);
  191. return;
  192. }
  193. /*
  194. * HPET is disabled. Trying enabling at FED00000 and check
  195. * whether it sticks
  196. */
  197. gen_cntl &= (~(0x7 << 15));
  198. gen_cntl |= (0x4 << 15);
  199. pci_write_config_dword(dev, 0xD0, gen_cntl);
  200. pci_read_config_dword(dev, 0xD0, &gen_cntl);
  201. val = gen_cntl >> 15;
  202. val &= 0x7;
  203. if (val & 0x4) {
  204. /* HPET is enabled in HPTC. Just not reported by BIOS */
  205. val &= 0x3;
  206. force_hpet_address = 0xFED00000 | (val << 12);
  207. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  208. "0x%lx\n", force_hpet_address);
  209. cached_dev = dev;
  210. force_hpet_resume_type = OLD_ICH_FORCE_HPET_RESUME;
  211. return;
  212. }
  213. dev_printk(KERN_DEBUG, &dev->dev, "Failed to force enable HPET\n");
  214. }
  215. /*
  216. * Undocumented chipset features. Make sure that the user enforced
  217. * this.
  218. */
  219. static void old_ich_force_enable_hpet_user(struct pci_dev *dev)
  220. {
  221. if (hpet_force_user)
  222. old_ich_force_enable_hpet(dev);
  223. }
  224. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ESB_1,
  225. old_ich_force_enable_hpet_user);
  226. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801CA_0,
  227. old_ich_force_enable_hpet_user);
  228. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801CA_12,
  229. old_ich_force_enable_hpet_user);
  230. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801DB_0,
  231. old_ich_force_enable_hpet_user);
  232. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801DB_12,
  233. old_ich_force_enable_hpet_user);
  234. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801EB_0,
  235. old_ich_force_enable_hpet);
  236. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801EB_12,
  237. old_ich_force_enable_hpet);
  238. static void vt8237_force_hpet_resume(void)
  239. {
  240. u32 val;
  241. if (!force_hpet_address || !cached_dev)
  242. return;
  243. val = 0xfed00000 | 0x80;
  244. pci_write_config_dword(cached_dev, 0x68, val);
  245. pci_read_config_dword(cached_dev, 0x68, &val);
  246. if (val & 0x80)
  247. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  248. else
  249. BUG();
  250. }
  251. static void vt8237_force_enable_hpet(struct pci_dev *dev)
  252. {
  253. u32 uninitialized_var(val);
  254. if (hpet_address || force_hpet_address)
  255. return;
  256. if (!hpet_force_user) {
  257. hpet_print_force_info();
  258. return;
  259. }
  260. pci_read_config_dword(dev, 0x68, &val);
  261. /*
  262. * Bit 7 is HPET enable bit.
  263. * Bit 31:10 is HPET base address (contrary to what datasheet claims)
  264. */
  265. if (val & 0x80) {
  266. force_hpet_address = (val & ~0x3ff);
  267. dev_printk(KERN_DEBUG, &dev->dev, "HPET at 0x%lx\n",
  268. force_hpet_address);
  269. return;
  270. }
  271. /*
  272. * HPET is disabled. Trying enabling at FED00000 and check
  273. * whether it sticks
  274. */
  275. val = 0xfed00000 | 0x80;
  276. pci_write_config_dword(dev, 0x68, val);
  277. pci_read_config_dword(dev, 0x68, &val);
  278. if (val & 0x80) {
  279. force_hpet_address = (val & ~0x3ff);
  280. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  281. "0x%lx\n", force_hpet_address);
  282. cached_dev = dev;
  283. force_hpet_resume_type = VT8237_FORCE_HPET_RESUME;
  284. return;
  285. }
  286. dev_printk(KERN_DEBUG, &dev->dev, "Failed to force enable HPET\n");
  287. }
  288. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8235,
  289. vt8237_force_enable_hpet);
  290. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8237,
  291. vt8237_force_enable_hpet);
  292. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_CX700,
  293. vt8237_force_enable_hpet);
  294. static void ati_force_hpet_resume(void)
  295. {
  296. pci_write_config_dword(cached_dev, 0x14, 0xfed00000);
  297. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  298. }
  299. static u32 ati_ixp4x0_rev(struct pci_dev *dev)
  300. {
  301. u32 d;
  302. u8 b;
  303. pci_read_config_byte(dev, 0xac, &b);
  304. b &= ~(1<<5);
  305. pci_write_config_byte(dev, 0xac, b);
  306. pci_read_config_dword(dev, 0x70, &d);
  307. d |= 1<<8;
  308. pci_write_config_dword(dev, 0x70, d);
  309. pci_read_config_dword(dev, 0x8, &d);
  310. d &= 0xff;
  311. dev_printk(KERN_DEBUG, &dev->dev, "SB4X0 revision 0x%x\n", d);
  312. return d;
  313. }
  314. static void ati_force_enable_hpet(struct pci_dev *dev)
  315. {
  316. u32 d, val;
  317. u8 b;
  318. if (hpet_address || force_hpet_address)
  319. return;
  320. if (!hpet_force_user) {
  321. hpet_print_force_info();
  322. return;
  323. }
  324. d = ati_ixp4x0_rev(dev);
  325. if (d < 0x82)
  326. return;
  327. /* base address */
  328. pci_write_config_dword(dev, 0x14, 0xfed00000);
  329. pci_read_config_dword(dev, 0x14, &val);
  330. /* enable interrupt */
  331. outb(0x72, 0xcd6); b = inb(0xcd7);
  332. b |= 0x1;
  333. outb(0x72, 0xcd6); outb(b, 0xcd7);
  334. outb(0x72, 0xcd6); b = inb(0xcd7);
  335. if (!(b & 0x1))
  336. return;
  337. pci_read_config_dword(dev, 0x64, &d);
  338. d |= (1<<10);
  339. pci_write_config_dword(dev, 0x64, d);
  340. pci_read_config_dword(dev, 0x64, &d);
  341. if (!(d & (1<<10)))
  342. return;
  343. force_hpet_address = val;
  344. force_hpet_resume_type = ATI_FORCE_HPET_RESUME;
  345. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at 0x%lx\n",
  346. force_hpet_address);
  347. cached_dev = dev;
  348. }
  349. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ATI, PCI_DEVICE_ID_ATI_IXP400_SMBUS,
  350. ati_force_enable_hpet);
  351. /*
  352. * Undocumented chipset feature taken from LinuxBIOS.
  353. */
  354. static void nvidia_force_hpet_resume(void)
  355. {
  356. pci_write_config_dword(cached_dev, 0x44, 0xfed00001);
  357. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  358. }
  359. static void nvidia_force_enable_hpet(struct pci_dev *dev)
  360. {
  361. u32 uninitialized_var(val);
  362. if (hpet_address || force_hpet_address)
  363. return;
  364. if (!hpet_force_user) {
  365. hpet_print_force_info();
  366. return;
  367. }
  368. pci_write_config_dword(dev, 0x44, 0xfed00001);
  369. pci_read_config_dword(dev, 0x44, &val);
  370. force_hpet_address = val & 0xfffffffe;
  371. force_hpet_resume_type = NVIDIA_FORCE_HPET_RESUME;
  372. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at 0x%lx\n",
  373. force_hpet_address);
  374. cached_dev = dev;
  375. return;
  376. }
  377. /* ISA Bridges */
  378. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0050,
  379. nvidia_force_enable_hpet);
  380. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0051,
  381. nvidia_force_enable_hpet);
  382. /* LPC bridges */
  383. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0260,
  384. nvidia_force_enable_hpet);
  385. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0360,
  386. nvidia_force_enable_hpet);
  387. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0361,
  388. nvidia_force_enable_hpet);
  389. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0362,
  390. nvidia_force_enable_hpet);
  391. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0363,
  392. nvidia_force_enable_hpet);
  393. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0364,
  394. nvidia_force_enable_hpet);
  395. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0365,
  396. nvidia_force_enable_hpet);
  397. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0366,
  398. nvidia_force_enable_hpet);
  399. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0367,
  400. nvidia_force_enable_hpet);
  401. void force_hpet_resume(void)
  402. {
  403. switch (force_hpet_resume_type) {
  404. case ICH_FORCE_HPET_RESUME:
  405. ich_force_hpet_resume();
  406. return;
  407. case OLD_ICH_FORCE_HPET_RESUME:
  408. old_ich_force_hpet_resume();
  409. return;
  410. case VT8237_FORCE_HPET_RESUME:
  411. vt8237_force_hpet_resume();
  412. return;
  413. case NVIDIA_FORCE_HPET_RESUME:
  414. nvidia_force_hpet_resume();
  415. return;
  416. case ATI_FORCE_HPET_RESUME:
  417. ati_force_hpet_resume();
  418. return;
  419. default:
  420. break;
  421. }
  422. }
  423. /*
  424. * HPET MSI on some boards (ATI SB700/SB800) has side effect on
  425. * floppy DMA. Disable HPET MSI on such platforms.
  426. * See erratum #27 (Misinterpreted MSI Requests May Result in
  427. * Corrupted LPC DMA Data) in AMD Publication #46837,
  428. * "SB700 Family Product Errata", Rev. 1.0, March 2010.
  429. */
  430. static void force_disable_hpet_msi(struct pci_dev *unused)
  431. {
  432. hpet_msi_disable = 1;
  433. }
  434. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ATI, PCI_DEVICE_ID_ATI_SBX00_SMBUS,
  435. force_disable_hpet_msi);
  436. #endif
  437. #if defined(CONFIG_PCI) && defined(CONFIG_NUMA)
  438. /* Set correct numa_node information for AMD NB functions */
  439. static void __init quirk_amd_nb_node(struct pci_dev *dev)
  440. {
  441. struct pci_dev *nb_ht;
  442. unsigned int devfn;
  443. u32 node;
  444. u32 val;
  445. devfn = PCI_DEVFN(PCI_SLOT(dev->devfn), 0);
  446. nb_ht = pci_get_slot(dev->bus, devfn);
  447. if (!nb_ht)
  448. return;
  449. pci_read_config_dword(nb_ht, 0x60, &val);
  450. node = val & 7;
  451. /*
  452. * Some hardware may return an invalid node ID,
  453. * so check it first:
  454. */
  455. if (node_online(node))
  456. set_dev_node(&dev->dev, node);
  457. pci_dev_put(nb_ht);
  458. }
  459. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB,
  460. quirk_amd_nb_node);
  461. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_ADDRMAP,
  462. quirk_amd_nb_node);
  463. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_MEMCTL,
  464. quirk_amd_nb_node);
  465. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_MISC,
  466. quirk_amd_nb_node);
  467. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_HT,
  468. quirk_amd_nb_node);
  469. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_MAP,
  470. quirk_amd_nb_node);
  471. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_DRAM,
  472. quirk_amd_nb_node);
  473. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_MISC,
  474. quirk_amd_nb_node);
  475. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_LINK,
  476. quirk_amd_nb_node);
  477. #endif