quirks.c 10 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400
  1. /*
  2. * This file contains work-arounds for x86 and x86_64 platform bugs.
  3. */
  4. #include <linux/pci.h>
  5. #include <linux/irq.h>
  6. #include <asm/hpet.h>
  7. #if defined(CONFIG_X86_IO_APIC) && defined(CONFIG_SMP) && defined(CONFIG_PCI)
  8. static void __devinit quirk_intel_irqbalance(struct pci_dev *dev)
  9. {
  10. u8 config, rev;
  11. u32 word;
  12. /* BIOS may enable hardware IRQ balancing for
  13. * E7520/E7320/E7525(revision ID 0x9 and below)
  14. * based platforms.
  15. * Disable SW irqbalance/affinity on those platforms.
  16. */
  17. pci_read_config_byte(dev, PCI_CLASS_REVISION, &rev);
  18. if (rev > 0x9)
  19. return;
  20. /* enable access to config space*/
  21. pci_read_config_byte(dev, 0xf4, &config);
  22. pci_write_config_byte(dev, 0xf4, config|0x2);
  23. /* read xTPR register */
  24. raw_pci_ops->read(0, 0, 0x40, 0x4c, 2, &word);
  25. if (!(word & (1 << 13))) {
  26. dev_info(&dev->dev, "Intel E7520/7320/7525 detected; "
  27. "disabling irq balancing and affinity\n");
  28. #ifdef CONFIG_IRQBALANCE
  29. irqbalance_disable("");
  30. #endif
  31. noirqdebug_setup("");
  32. #ifdef CONFIG_PROC_FS
  33. no_irq_affinity = 1;
  34. #endif
  35. }
  36. /* put back the original value for config space*/
  37. if (!(config & 0x2))
  38. pci_write_config_byte(dev, 0xf4, config);
  39. }
  40. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7320_MCH,
  41. quirk_intel_irqbalance);
  42. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7525_MCH,
  43. quirk_intel_irqbalance);
  44. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7520_MCH,
  45. quirk_intel_irqbalance);
  46. #endif
  47. #if defined(CONFIG_HPET_TIMER)
  48. unsigned long force_hpet_address;
  49. static enum {
  50. NONE_FORCE_HPET_RESUME,
  51. OLD_ICH_FORCE_HPET_RESUME,
  52. ICH_FORCE_HPET_RESUME,
  53. VT8237_FORCE_HPET_RESUME,
  54. NVIDIA_FORCE_HPET_RESUME,
  55. } force_hpet_resume_type;
  56. static void __iomem *rcba_base;
  57. static void ich_force_hpet_resume(void)
  58. {
  59. u32 val;
  60. if (!force_hpet_address)
  61. return;
  62. if (rcba_base == NULL)
  63. BUG();
  64. /* read the Function Disable register, dword mode only */
  65. val = readl(rcba_base + 0x3404);
  66. if (!(val & 0x80)) {
  67. /* HPET disabled in HPTC. Trying to enable */
  68. writel(val | 0x80, rcba_base + 0x3404);
  69. }
  70. val = readl(rcba_base + 0x3404);
  71. if (!(val & 0x80))
  72. BUG();
  73. else
  74. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  75. return;
  76. }
  77. static void ich_force_enable_hpet(struct pci_dev *dev)
  78. {
  79. u32 val;
  80. u32 uninitialized_var(rcba);
  81. int err = 0;
  82. if (hpet_address || force_hpet_address)
  83. return;
  84. pci_read_config_dword(dev, 0xF0, &rcba);
  85. rcba &= 0xFFFFC000;
  86. if (rcba == 0) {
  87. dev_printk(KERN_DEBUG, &dev->dev, "RCBA disabled; "
  88. "cannot force enable HPET\n");
  89. return;
  90. }
  91. /* use bits 31:14, 16 kB aligned */
  92. rcba_base = ioremap_nocache(rcba, 0x4000);
  93. if (rcba_base == NULL) {
  94. dev_printk(KERN_DEBUG, &dev->dev, "ioremap failed; "
  95. "cannot force enable HPET\n");
  96. return;
  97. }
  98. /* read the Function Disable register, dword mode only */
  99. val = readl(rcba_base + 0x3404);
  100. if (val & 0x80) {
  101. /* HPET is enabled in HPTC. Just not reported by BIOS */
  102. val = val & 0x3;
  103. force_hpet_address = 0xFED00000 | (val << 12);
  104. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  105. "0x%lx\n", force_hpet_address);
  106. iounmap(rcba_base);
  107. return;
  108. }
  109. /* HPET disabled in HPTC. Trying to enable */
  110. writel(val | 0x80, rcba_base + 0x3404);
  111. val = readl(rcba_base + 0x3404);
  112. if (!(val & 0x80)) {
  113. err = 1;
  114. } else {
  115. val = val & 0x3;
  116. force_hpet_address = 0xFED00000 | (val << 12);
  117. }
  118. if (err) {
  119. force_hpet_address = 0;
  120. iounmap(rcba_base);
  121. dev_printk(KERN_DEBUG, &dev->dev,
  122. "Failed to force enable HPET\n");
  123. } else {
  124. force_hpet_resume_type = ICH_FORCE_HPET_RESUME;
  125. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  126. "0x%lx\n", force_hpet_address);
  127. }
  128. }
  129. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ESB2_0,
  130. ich_force_enable_hpet);
  131. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH6_1,
  132. ich_force_enable_hpet);
  133. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_0,
  134. ich_force_enable_hpet);
  135. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_1,
  136. ich_force_enable_hpet);
  137. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_31,
  138. ich_force_enable_hpet);
  139. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH8_1,
  140. ich_force_enable_hpet);
  141. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH9_7,
  142. ich_force_enable_hpet);
  143. static struct pci_dev *cached_dev;
  144. static void old_ich_force_hpet_resume(void)
  145. {
  146. u32 val;
  147. u32 uninitialized_var(gen_cntl);
  148. if (!force_hpet_address || !cached_dev)
  149. return;
  150. pci_read_config_dword(cached_dev, 0xD0, &gen_cntl);
  151. gen_cntl &= (~(0x7 << 15));
  152. gen_cntl |= (0x4 << 15);
  153. pci_write_config_dword(cached_dev, 0xD0, gen_cntl);
  154. pci_read_config_dword(cached_dev, 0xD0, &gen_cntl);
  155. val = gen_cntl >> 15;
  156. val &= 0x7;
  157. if (val == 0x4)
  158. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  159. else
  160. BUG();
  161. }
  162. static void old_ich_force_enable_hpet(struct pci_dev *dev)
  163. {
  164. u32 val;
  165. u32 uninitialized_var(gen_cntl);
  166. if (hpet_address || force_hpet_address)
  167. return;
  168. pci_read_config_dword(dev, 0xD0, &gen_cntl);
  169. /*
  170. * Bit 17 is HPET enable bit.
  171. * Bit 16:15 control the HPET base address.
  172. */
  173. val = gen_cntl >> 15;
  174. val &= 0x7;
  175. if (val & 0x4) {
  176. val &= 0x3;
  177. force_hpet_address = 0xFED00000 | (val << 12);
  178. dev_printk(KERN_DEBUG, &dev->dev, "HPET at 0x%lx\n",
  179. force_hpet_address);
  180. return;
  181. }
  182. /*
  183. * HPET is disabled. Trying enabling at FED00000 and check
  184. * whether it sticks
  185. */
  186. gen_cntl &= (~(0x7 << 15));
  187. gen_cntl |= (0x4 << 15);
  188. pci_write_config_dword(dev, 0xD0, gen_cntl);
  189. pci_read_config_dword(dev, 0xD0, &gen_cntl);
  190. val = gen_cntl >> 15;
  191. val &= 0x7;
  192. if (val & 0x4) {
  193. /* HPET is enabled in HPTC. Just not reported by BIOS */
  194. val &= 0x3;
  195. force_hpet_address = 0xFED00000 | (val << 12);
  196. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  197. "0x%lx\n", force_hpet_address);
  198. cached_dev = dev;
  199. force_hpet_resume_type = OLD_ICH_FORCE_HPET_RESUME;
  200. return;
  201. }
  202. dev_printk(KERN_DEBUG, &dev->dev, "Failed to force enable HPET\n");
  203. }
  204. /*
  205. * Undocumented chipset features. Make sure that the user enforced
  206. * this.
  207. */
  208. static void old_ich_force_enable_hpet_user(struct pci_dev *dev)
  209. {
  210. if (hpet_force_user)
  211. old_ich_force_enable_hpet(dev);
  212. }
  213. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801CA_0,
  214. old_ich_force_enable_hpet_user);
  215. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801CA_12,
  216. old_ich_force_enable_hpet_user);
  217. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801DB_0,
  218. old_ich_force_enable_hpet_user);
  219. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801DB_12,
  220. old_ich_force_enable_hpet_user);
  221. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801EB_0,
  222. old_ich_force_enable_hpet);
  223. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801EB_12,
  224. old_ich_force_enable_hpet);
  225. static void vt8237_force_hpet_resume(void)
  226. {
  227. u32 val;
  228. if (!force_hpet_address || !cached_dev)
  229. return;
  230. val = 0xfed00000 | 0x80;
  231. pci_write_config_dword(cached_dev, 0x68, val);
  232. pci_read_config_dword(cached_dev, 0x68, &val);
  233. if (val & 0x80)
  234. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  235. else
  236. BUG();
  237. }
  238. static void vt8237_force_enable_hpet(struct pci_dev *dev)
  239. {
  240. u32 uninitialized_var(val);
  241. if (!hpet_force_user || hpet_address || force_hpet_address)
  242. return;
  243. pci_read_config_dword(dev, 0x68, &val);
  244. /*
  245. * Bit 7 is HPET enable bit.
  246. * Bit 31:10 is HPET base address (contrary to what datasheet claims)
  247. */
  248. if (val & 0x80) {
  249. force_hpet_address = (val & ~0x3ff);
  250. dev_printk(KERN_DEBUG, &dev->dev, "HPET at 0x%lx\n",
  251. force_hpet_address);
  252. return;
  253. }
  254. /*
  255. * HPET is disabled. Trying enabling at FED00000 and check
  256. * whether it sticks
  257. */
  258. val = 0xfed00000 | 0x80;
  259. pci_write_config_dword(dev, 0x68, val);
  260. pci_read_config_dword(dev, 0x68, &val);
  261. if (val & 0x80) {
  262. force_hpet_address = (val & ~0x3ff);
  263. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  264. "0x%lx\n", force_hpet_address);
  265. cached_dev = dev;
  266. force_hpet_resume_type = VT8237_FORCE_HPET_RESUME;
  267. return;
  268. }
  269. dev_printk(KERN_DEBUG, &dev->dev, "Failed to force enable HPET\n");
  270. }
  271. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8235,
  272. vt8237_force_enable_hpet);
  273. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8237,
  274. vt8237_force_enable_hpet);
  275. /*
  276. * Undocumented chipset feature taken from LinuxBIOS.
  277. */
  278. static void nvidia_force_hpet_resume(void)
  279. {
  280. pci_write_config_dword(cached_dev, 0x44, 0xfed00001);
  281. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  282. }
  283. static void nvidia_force_enable_hpet(struct pci_dev *dev)
  284. {
  285. u32 uninitialized_var(val);
  286. if (!hpet_force_user || hpet_address || force_hpet_address)
  287. return;
  288. pci_write_config_dword(dev, 0x44, 0xfed00001);
  289. pci_read_config_dword(dev, 0x44, &val);
  290. force_hpet_address = val & 0xfffffffe;
  291. force_hpet_resume_type = NVIDIA_FORCE_HPET_RESUME;
  292. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at 0x%lx\n",
  293. force_hpet_address);
  294. cached_dev = dev;
  295. return;
  296. }
  297. /* ISA Bridges */
  298. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0050,
  299. nvidia_force_enable_hpet);
  300. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0051,
  301. nvidia_force_enable_hpet);
  302. /* LPC bridges */
  303. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0360,
  304. nvidia_force_enable_hpet);
  305. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0361,
  306. nvidia_force_enable_hpet);
  307. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0362,
  308. nvidia_force_enable_hpet);
  309. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0363,
  310. nvidia_force_enable_hpet);
  311. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0364,
  312. nvidia_force_enable_hpet);
  313. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0365,
  314. nvidia_force_enable_hpet);
  315. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0366,
  316. nvidia_force_enable_hpet);
  317. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0367,
  318. nvidia_force_enable_hpet);
  319. void force_hpet_resume(void)
  320. {
  321. switch (force_hpet_resume_type) {
  322. case ICH_FORCE_HPET_RESUME:
  323. return ich_force_hpet_resume();
  324. case OLD_ICH_FORCE_HPET_RESUME:
  325. return old_ich_force_hpet_resume();
  326. case VT8237_FORCE_HPET_RESUME:
  327. return vt8237_force_hpet_resume();
  328. case NVIDIA_FORCE_HPET_RESUME:
  329. return nvidia_force_hpet_resume();
  330. default:
  331. break;
  332. }
  333. }
  334. #endif