qlge_dbg.c 56 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670
  1. #include "qlge.h"
  2. /* Read a NIC register from the alternate function. */
  3. static u32 ql_read_other_func_reg(struct ql_adapter *qdev,
  4. u32 reg)
  5. {
  6. u32 register_to_read;
  7. u32 reg_val;
  8. unsigned int status = 0;
  9. register_to_read = MPI_NIC_REG_BLOCK
  10. | MPI_NIC_READ
  11. | (qdev->alt_func << MPI_NIC_FUNCTION_SHIFT)
  12. | reg;
  13. status = ql_read_mpi_reg(qdev, register_to_read, &reg_val);
  14. if (status != 0)
  15. return 0xffffffff;
  16. return reg_val;
  17. }
  18. static int ql_get_ets_regs(struct ql_adapter *qdev, u32 * buf)
  19. {
  20. int status = 0;
  21. int i;
  22. for (i = 0; i < 8; i++, buf++) {
  23. ql_write32(qdev, NIC_ETS, i << 29 | 0x08000000);
  24. *buf = ql_read32(qdev, NIC_ETS);
  25. }
  26. for (i = 0; i < 2; i++, buf++) {
  27. ql_write32(qdev, CNA_ETS, i << 29 | 0x08000000);
  28. *buf = ql_read32(qdev, CNA_ETS);
  29. }
  30. return status;
  31. }
  32. static void ql_get_intr_states(struct ql_adapter *qdev, u32 * buf)
  33. {
  34. int i;
  35. for (i = 0; i < qdev->rx_ring_count; i++, buf++) {
  36. ql_write32(qdev, INTR_EN,
  37. qdev->intr_context[i].intr_read_mask);
  38. *buf = ql_read32(qdev, INTR_EN);
  39. }
  40. }
  41. static int ql_get_cam_entries(struct ql_adapter *qdev, u32 * buf)
  42. {
  43. int i, status;
  44. u32 value[3];
  45. status = ql_sem_spinlock(qdev, SEM_MAC_ADDR_MASK);
  46. if (status)
  47. return status;
  48. for (i = 0; i < 16; i++) {
  49. status = ql_get_mac_addr_reg(qdev,
  50. MAC_ADDR_TYPE_CAM_MAC, i, value);
  51. if (status) {
  52. QPRINTK(qdev, DRV, ERR,
  53. "Failed read of mac index register.\n");
  54. goto err;
  55. }
  56. *buf++ = value[0]; /* lower MAC address */
  57. *buf++ = value[1]; /* upper MAC address */
  58. *buf++ = value[2]; /* output */
  59. }
  60. for (i = 0; i < 32; i++) {
  61. status = ql_get_mac_addr_reg(qdev,
  62. MAC_ADDR_TYPE_MULTI_MAC, i, value);
  63. if (status) {
  64. QPRINTK(qdev, DRV, ERR,
  65. "Failed read of mac index register.\n");
  66. goto err;
  67. }
  68. *buf++ = value[0]; /* lower Mcast address */
  69. *buf++ = value[1]; /* upper Mcast address */
  70. }
  71. err:
  72. ql_sem_unlock(qdev, SEM_MAC_ADDR_MASK);
  73. return status;
  74. }
  75. static int ql_get_routing_entries(struct ql_adapter *qdev, u32 * buf)
  76. {
  77. int status;
  78. u32 value, i;
  79. status = ql_sem_spinlock(qdev, SEM_RT_IDX_MASK);
  80. if (status)
  81. return status;
  82. for (i = 0; i < 16; i++) {
  83. status = ql_get_routing_reg(qdev, i, &value);
  84. if (status) {
  85. QPRINTK(qdev, DRV, ERR,
  86. "Failed read of routing index register.\n");
  87. goto err;
  88. } else {
  89. *buf++ = value;
  90. }
  91. }
  92. err:
  93. ql_sem_unlock(qdev, SEM_RT_IDX_MASK);
  94. return status;
  95. }
  96. /* Read the MPI Processor shadow registers */
  97. static int ql_get_mpi_shadow_regs(struct ql_adapter *qdev, u32 * buf)
  98. {
  99. u32 i;
  100. int status;
  101. for (i = 0; i < MPI_CORE_SH_REGS_CNT; i++, buf++) {
  102. status = ql_write_mpi_reg(qdev, RISC_124,
  103. (SHADOW_OFFSET | i << SHADOW_REG_SHIFT));
  104. if (status)
  105. goto end;
  106. status = ql_read_mpi_reg(qdev, RISC_127, buf);
  107. if (status)
  108. goto end;
  109. }
  110. end:
  111. return status;
  112. }
  113. /* Read the MPI Processor core registers */
  114. static int ql_get_mpi_regs(struct ql_adapter *qdev, u32 * buf,
  115. u32 offset, u32 count)
  116. {
  117. int i, status = 0;
  118. for (i = 0; i < count; i++, buf++) {
  119. status = ql_read_mpi_reg(qdev, offset + i, buf);
  120. if (status)
  121. return status;
  122. }
  123. return status;
  124. }
  125. /* Read the ASIC probe dump */
  126. static unsigned int *ql_get_probe(struct ql_adapter *qdev, u32 clock,
  127. u32 valid, u32 *buf)
  128. {
  129. u32 module, mux_sel, probe, lo_val, hi_val;
  130. for (module = 0; module < PRB_MX_ADDR_MAX_MODS; module++) {
  131. if (!((valid >> module) & 1))
  132. continue;
  133. for (mux_sel = 0; mux_sel < PRB_MX_ADDR_MAX_MUX; mux_sel++) {
  134. probe = clock
  135. | PRB_MX_ADDR_ARE
  136. | mux_sel
  137. | (module << PRB_MX_ADDR_MOD_SEL_SHIFT);
  138. ql_write32(qdev, PRB_MX_ADDR, probe);
  139. lo_val = ql_read32(qdev, PRB_MX_DATA);
  140. if (mux_sel == 0) {
  141. *buf = probe;
  142. buf++;
  143. }
  144. probe |= PRB_MX_ADDR_UP;
  145. ql_write32(qdev, PRB_MX_ADDR, probe);
  146. hi_val = ql_read32(qdev, PRB_MX_DATA);
  147. *buf = lo_val;
  148. buf++;
  149. *buf = hi_val;
  150. buf++;
  151. }
  152. }
  153. return buf;
  154. }
  155. static int ql_get_probe_dump(struct ql_adapter *qdev, unsigned int *buf)
  156. {
  157. /* First we have to enable the probe mux */
  158. ql_write_mpi_reg(qdev, MPI_TEST_FUNC_PRB_CTL, MPI_TEST_FUNC_PRB_EN);
  159. buf = ql_get_probe(qdev, PRB_MX_ADDR_SYS_CLOCK,
  160. PRB_MX_ADDR_VALID_SYS_MOD, buf);
  161. buf = ql_get_probe(qdev, PRB_MX_ADDR_PCI_CLOCK,
  162. PRB_MX_ADDR_VALID_PCI_MOD, buf);
  163. buf = ql_get_probe(qdev, PRB_MX_ADDR_XGM_CLOCK,
  164. PRB_MX_ADDR_VALID_XGM_MOD, buf);
  165. buf = ql_get_probe(qdev, PRB_MX_ADDR_FC_CLOCK,
  166. PRB_MX_ADDR_VALID_FC_MOD, buf);
  167. return 0;
  168. }
  169. /* Read out the routing index registers */
  170. static int ql_get_routing_index_registers(struct ql_adapter *qdev, u32 *buf)
  171. {
  172. int status;
  173. u32 type, index, index_max;
  174. u32 result_index;
  175. u32 result_data;
  176. u32 val;
  177. status = ql_sem_spinlock(qdev, SEM_RT_IDX_MASK);
  178. if (status)
  179. return status;
  180. for (type = 0; type < 4; type++) {
  181. if (type < 2)
  182. index_max = 8;
  183. else
  184. index_max = 16;
  185. for (index = 0; index < index_max; index++) {
  186. val = RT_IDX_RS
  187. | (type << RT_IDX_TYPE_SHIFT)
  188. | (index << RT_IDX_IDX_SHIFT);
  189. ql_write32(qdev, RT_IDX, val);
  190. result_index = 0;
  191. while ((result_index & RT_IDX_MR) == 0)
  192. result_index = ql_read32(qdev, RT_IDX);
  193. result_data = ql_read32(qdev, RT_DATA);
  194. *buf = type;
  195. buf++;
  196. *buf = index;
  197. buf++;
  198. *buf = result_index;
  199. buf++;
  200. *buf = result_data;
  201. buf++;
  202. }
  203. }
  204. ql_sem_unlock(qdev, SEM_RT_IDX_MASK);
  205. return status;
  206. }
  207. /* Read out the MAC protocol registers */
  208. static void ql_get_mac_protocol_registers(struct ql_adapter *qdev, u32 *buf)
  209. {
  210. u32 result_index, result_data;
  211. u32 type;
  212. u32 index;
  213. u32 offset;
  214. u32 val;
  215. u32 initial_val = MAC_ADDR_RS;
  216. u32 max_index;
  217. u32 max_offset;
  218. for (type = 0; type < MAC_ADDR_TYPE_COUNT; type++) {
  219. switch (type) {
  220. case 0: /* CAM */
  221. initial_val |= MAC_ADDR_ADR;
  222. max_index = MAC_ADDR_MAX_CAM_ENTRIES;
  223. max_offset = MAC_ADDR_MAX_CAM_WCOUNT;
  224. break;
  225. case 1: /* Multicast MAC Address */
  226. max_index = MAC_ADDR_MAX_CAM_WCOUNT;
  227. max_offset = MAC_ADDR_MAX_CAM_WCOUNT;
  228. break;
  229. case 2: /* VLAN filter mask */
  230. case 3: /* MC filter mask */
  231. max_index = MAC_ADDR_MAX_CAM_WCOUNT;
  232. max_offset = MAC_ADDR_MAX_CAM_WCOUNT;
  233. break;
  234. case 4: /* FC MAC addresses */
  235. max_index = MAC_ADDR_MAX_FC_MAC_ENTRIES;
  236. max_offset = MAC_ADDR_MAX_FC_MAC_WCOUNT;
  237. break;
  238. case 5: /* Mgmt MAC addresses */
  239. max_index = MAC_ADDR_MAX_MGMT_MAC_ENTRIES;
  240. max_offset = MAC_ADDR_MAX_MGMT_MAC_WCOUNT;
  241. break;
  242. case 6: /* Mgmt VLAN addresses */
  243. max_index = MAC_ADDR_MAX_MGMT_VLAN_ENTRIES;
  244. max_offset = MAC_ADDR_MAX_MGMT_VLAN_WCOUNT;
  245. break;
  246. case 7: /* Mgmt IPv4 address */
  247. max_index = MAC_ADDR_MAX_MGMT_V4_ENTRIES;
  248. max_offset = MAC_ADDR_MAX_MGMT_V4_WCOUNT;
  249. break;
  250. case 8: /* Mgmt IPv6 address */
  251. max_index = MAC_ADDR_MAX_MGMT_V6_ENTRIES;
  252. max_offset = MAC_ADDR_MAX_MGMT_V6_WCOUNT;
  253. break;
  254. case 9: /* Mgmt TCP/UDP Dest port */
  255. max_index = MAC_ADDR_MAX_MGMT_TU_DP_ENTRIES;
  256. max_offset = MAC_ADDR_MAX_MGMT_TU_DP_WCOUNT;
  257. break;
  258. default:
  259. printk(KERN_ERR"Bad type!!! 0x%08x\n", type);
  260. max_index = 0;
  261. max_offset = 0;
  262. break;
  263. }
  264. for (index = 0; index < max_index; index++) {
  265. for (offset = 0; offset < max_offset; offset++) {
  266. val = initial_val
  267. | (type << MAC_ADDR_TYPE_SHIFT)
  268. | (index << MAC_ADDR_IDX_SHIFT)
  269. | (offset);
  270. ql_write32(qdev, MAC_ADDR_IDX, val);
  271. result_index = 0;
  272. while ((result_index & MAC_ADDR_MR) == 0) {
  273. result_index = ql_read32(qdev,
  274. MAC_ADDR_IDX);
  275. }
  276. result_data = ql_read32(qdev, MAC_ADDR_DATA);
  277. *buf = result_index;
  278. buf++;
  279. *buf = result_data;
  280. buf++;
  281. }
  282. }
  283. }
  284. }
  285. static void ql_get_sem_registers(struct ql_adapter *qdev, u32 *buf)
  286. {
  287. u32 func_num, reg, reg_val;
  288. int status;
  289. for (func_num = 0; func_num < MAX_SEMAPHORE_FUNCTIONS ; func_num++) {
  290. reg = MPI_NIC_REG_BLOCK
  291. | (func_num << MPI_NIC_FUNCTION_SHIFT)
  292. | (SEM / 4);
  293. status = ql_read_mpi_reg(qdev, reg, &reg_val);
  294. *buf = reg_val;
  295. /* if the read failed then dead fill the element. */
  296. if (!status)
  297. *buf = 0xdeadbeef;
  298. buf++;
  299. }
  300. }
  301. /* Create a coredump segment header */
  302. static void ql_build_coredump_seg_header(
  303. struct mpi_coredump_segment_header *seg_hdr,
  304. u32 seg_number, u32 seg_size, u8 *desc)
  305. {
  306. memset(seg_hdr, 0, sizeof(struct mpi_coredump_segment_header));
  307. seg_hdr->cookie = MPI_COREDUMP_COOKIE;
  308. seg_hdr->segNum = seg_number;
  309. seg_hdr->segSize = seg_size;
  310. memcpy(seg_hdr->description, desc, (sizeof(seg_hdr->description)) - 1);
  311. }
  312. /*
  313. * This function should be called when a coredump / probedump
  314. * is to be extracted from the HBA. It is assumed there is a
  315. * qdev structure that contains the base address of the register
  316. * space for this function as well as a coredump structure that
  317. * will contain the dump.
  318. */
  319. int ql_core_dump(struct ql_adapter *qdev, struct ql_mpi_coredump *mpi_coredump)
  320. {
  321. int status;
  322. int i;
  323. if (!mpi_coredump) {
  324. QPRINTK(qdev, DRV, ERR,
  325. "No memory available.\n");
  326. return -ENOMEM;
  327. }
  328. /* Try to get the spinlock, but dont worry if
  329. * it isn't available. If the firmware died it
  330. * might be holding the sem.
  331. */
  332. ql_sem_spinlock(qdev, SEM_PROC_REG_MASK);
  333. status = ql_pause_mpi_risc(qdev);
  334. if (status) {
  335. QPRINTK(qdev, DRV, ERR,
  336. "Failed RISC pause. Status = 0x%.08x\n", status);
  337. goto err;
  338. }
  339. /* Insert the global header */
  340. memset(&(mpi_coredump->mpi_global_header), 0,
  341. sizeof(struct mpi_coredump_global_header));
  342. mpi_coredump->mpi_global_header.cookie = MPI_COREDUMP_COOKIE;
  343. mpi_coredump->mpi_global_header.headerSize =
  344. sizeof(struct mpi_coredump_global_header);
  345. mpi_coredump->mpi_global_header.imageSize =
  346. sizeof(struct ql_mpi_coredump);
  347. memcpy(mpi_coredump->mpi_global_header.idString, "MPI Coredump",
  348. sizeof(mpi_coredump->mpi_global_header.idString));
  349. /* Get generic NIC reg dump */
  350. ql_build_coredump_seg_header(&mpi_coredump->nic_regs_seg_hdr,
  351. NIC1_CONTROL_SEG_NUM,
  352. sizeof(struct mpi_coredump_segment_header) +
  353. sizeof(mpi_coredump->nic_regs), "NIC1 Registers");
  354. ql_build_coredump_seg_header(&mpi_coredump->nic2_regs_seg_hdr,
  355. NIC2_CONTROL_SEG_NUM,
  356. sizeof(struct mpi_coredump_segment_header) +
  357. sizeof(mpi_coredump->nic2_regs), "NIC2 Registers");
  358. if (qdev->func & 1) {
  359. /* Odd means our function is NIC 2 */
  360. for (i = 0; i < NIC_REGS_DUMP_WORD_COUNT; i++)
  361. mpi_coredump->nic2_regs[i] =
  362. ql_read32(qdev, i * sizeof(u32));
  363. for (i = 0; i < NIC_REGS_DUMP_WORD_COUNT; i++)
  364. mpi_coredump->nic_regs[i] =
  365. ql_read_other_func_reg(qdev, (i * sizeof(u32)) / 4);
  366. } else {
  367. /* Even means our function is NIC 1 */
  368. for (i = 0; i < NIC_REGS_DUMP_WORD_COUNT; i++)
  369. mpi_coredump->nic_regs[i] =
  370. ql_read32(qdev, i * sizeof(u32));
  371. for (i = 0; i < NIC_REGS_DUMP_WORD_COUNT; i++)
  372. mpi_coredump->nic2_regs[i] =
  373. ql_read_other_func_reg(qdev, (i * sizeof(u32)) / 4);
  374. }
  375. ql_build_coredump_seg_header(&mpi_coredump->core_regs_seg_hdr,
  376. CORE_SEG_NUM,
  377. sizeof(mpi_coredump->core_regs_seg_hdr) +
  378. sizeof(mpi_coredump->mpi_core_regs) +
  379. sizeof(mpi_coredump->mpi_core_sh_regs),
  380. "Core Registers");
  381. /* Get the MPI Core Registers */
  382. status = ql_get_mpi_regs(qdev, &mpi_coredump->mpi_core_regs[0],
  383. MPI_CORE_REGS_ADDR, MPI_CORE_REGS_CNT);
  384. if (status)
  385. goto err;
  386. /* Get the 16 MPI shadow registers */
  387. status = ql_get_mpi_shadow_regs(qdev,
  388. &mpi_coredump->mpi_core_sh_regs[0]);
  389. if (status)
  390. goto err;
  391. /* Get the Test Logic Registers */
  392. ql_build_coredump_seg_header(&mpi_coredump->test_logic_regs_seg_hdr,
  393. TEST_LOGIC_SEG_NUM,
  394. sizeof(struct mpi_coredump_segment_header)
  395. + sizeof(mpi_coredump->test_logic_regs),
  396. "Test Logic Regs");
  397. status = ql_get_mpi_regs(qdev, &mpi_coredump->test_logic_regs[0],
  398. TEST_REGS_ADDR, TEST_REGS_CNT);
  399. if (status)
  400. goto err;
  401. /* Get the RMII Registers */
  402. ql_build_coredump_seg_header(&mpi_coredump->rmii_regs_seg_hdr,
  403. RMII_SEG_NUM,
  404. sizeof(struct mpi_coredump_segment_header)
  405. + sizeof(mpi_coredump->rmii_regs),
  406. "RMII Registers");
  407. status = ql_get_mpi_regs(qdev, &mpi_coredump->rmii_regs[0],
  408. RMII_REGS_ADDR, RMII_REGS_CNT);
  409. if (status)
  410. goto err;
  411. /* Get the FCMAC1 Registers */
  412. ql_build_coredump_seg_header(&mpi_coredump->fcmac1_regs_seg_hdr,
  413. FCMAC1_SEG_NUM,
  414. sizeof(struct mpi_coredump_segment_header)
  415. + sizeof(mpi_coredump->fcmac1_regs),
  416. "FCMAC1 Registers");
  417. status = ql_get_mpi_regs(qdev, &mpi_coredump->fcmac1_regs[0],
  418. FCMAC1_REGS_ADDR, FCMAC_REGS_CNT);
  419. if (status)
  420. goto err;
  421. /* Get the FCMAC2 Registers */
  422. ql_build_coredump_seg_header(&mpi_coredump->fcmac2_regs_seg_hdr,
  423. FCMAC2_SEG_NUM,
  424. sizeof(struct mpi_coredump_segment_header)
  425. + sizeof(mpi_coredump->fcmac2_regs),
  426. "FCMAC2 Registers");
  427. status = ql_get_mpi_regs(qdev, &mpi_coredump->fcmac2_regs[0],
  428. FCMAC2_REGS_ADDR, FCMAC_REGS_CNT);
  429. if (status)
  430. goto err;
  431. /* Get the FC1 MBX Registers */
  432. ql_build_coredump_seg_header(&mpi_coredump->fc1_mbx_regs_seg_hdr,
  433. FC1_MBOX_SEG_NUM,
  434. sizeof(struct mpi_coredump_segment_header)
  435. + sizeof(mpi_coredump->fc1_mbx_regs),
  436. "FC1 MBox Regs");
  437. status = ql_get_mpi_regs(qdev, &mpi_coredump->fc1_mbx_regs[0],
  438. FC1_MBX_REGS_ADDR, FC_MBX_REGS_CNT);
  439. if (status)
  440. goto err;
  441. /* Get the IDE Registers */
  442. ql_build_coredump_seg_header(&mpi_coredump->ide_regs_seg_hdr,
  443. IDE_SEG_NUM,
  444. sizeof(struct mpi_coredump_segment_header)
  445. + sizeof(mpi_coredump->ide_regs),
  446. "IDE Registers");
  447. status = ql_get_mpi_regs(qdev, &mpi_coredump->ide_regs[0],
  448. IDE_REGS_ADDR, IDE_REGS_CNT);
  449. if (status)
  450. goto err;
  451. /* Get the NIC1 MBX Registers */
  452. ql_build_coredump_seg_header(&mpi_coredump->nic1_mbx_regs_seg_hdr,
  453. NIC1_MBOX_SEG_NUM,
  454. sizeof(struct mpi_coredump_segment_header)
  455. + sizeof(mpi_coredump->nic1_mbx_regs),
  456. "NIC1 MBox Regs");
  457. status = ql_get_mpi_regs(qdev, &mpi_coredump->nic1_mbx_regs[0],
  458. NIC1_MBX_REGS_ADDR, NIC_MBX_REGS_CNT);
  459. if (status)
  460. goto err;
  461. /* Get the SMBus Registers */
  462. ql_build_coredump_seg_header(&mpi_coredump->smbus_regs_seg_hdr,
  463. SMBUS_SEG_NUM,
  464. sizeof(struct mpi_coredump_segment_header)
  465. + sizeof(mpi_coredump->smbus_regs),
  466. "SMBus Registers");
  467. status = ql_get_mpi_regs(qdev, &mpi_coredump->smbus_regs[0],
  468. SMBUS_REGS_ADDR, SMBUS_REGS_CNT);
  469. if (status)
  470. goto err;
  471. /* Get the FC2 MBX Registers */
  472. ql_build_coredump_seg_header(&mpi_coredump->fc2_mbx_regs_seg_hdr,
  473. FC2_MBOX_SEG_NUM,
  474. sizeof(struct mpi_coredump_segment_header)
  475. + sizeof(mpi_coredump->fc2_mbx_regs),
  476. "FC2 MBox Regs");
  477. status = ql_get_mpi_regs(qdev, &mpi_coredump->fc2_mbx_regs[0],
  478. FC2_MBX_REGS_ADDR, FC_MBX_REGS_CNT);
  479. if (status)
  480. goto err;
  481. /* Get the NIC2 MBX Registers */
  482. ql_build_coredump_seg_header(&mpi_coredump->nic2_mbx_regs_seg_hdr,
  483. NIC2_MBOX_SEG_NUM,
  484. sizeof(struct mpi_coredump_segment_header)
  485. + sizeof(mpi_coredump->nic2_mbx_regs),
  486. "NIC2 MBox Regs");
  487. status = ql_get_mpi_regs(qdev, &mpi_coredump->nic2_mbx_regs[0],
  488. NIC2_MBX_REGS_ADDR, NIC_MBX_REGS_CNT);
  489. if (status)
  490. goto err;
  491. /* Get the I2C Registers */
  492. ql_build_coredump_seg_header(&mpi_coredump->i2c_regs_seg_hdr,
  493. I2C_SEG_NUM,
  494. sizeof(struct mpi_coredump_segment_header)
  495. + sizeof(mpi_coredump->i2c_regs),
  496. "I2C Registers");
  497. status = ql_get_mpi_regs(qdev, &mpi_coredump->i2c_regs[0],
  498. I2C_REGS_ADDR, I2C_REGS_CNT);
  499. if (status)
  500. goto err;
  501. /* Get the MEMC Registers */
  502. ql_build_coredump_seg_header(&mpi_coredump->memc_regs_seg_hdr,
  503. MEMC_SEG_NUM,
  504. sizeof(struct mpi_coredump_segment_header)
  505. + sizeof(mpi_coredump->memc_regs),
  506. "MEMC Registers");
  507. status = ql_get_mpi_regs(qdev, &mpi_coredump->memc_regs[0],
  508. MEMC_REGS_ADDR, MEMC_REGS_CNT);
  509. if (status)
  510. goto err;
  511. /* Get the PBus Registers */
  512. ql_build_coredump_seg_header(&mpi_coredump->pbus_regs_seg_hdr,
  513. PBUS_SEG_NUM,
  514. sizeof(struct mpi_coredump_segment_header)
  515. + sizeof(mpi_coredump->pbus_regs),
  516. "PBUS Registers");
  517. status = ql_get_mpi_regs(qdev, &mpi_coredump->pbus_regs[0],
  518. PBUS_REGS_ADDR, PBUS_REGS_CNT);
  519. if (status)
  520. goto err;
  521. /* Get the MDE Registers */
  522. ql_build_coredump_seg_header(&mpi_coredump->mde_regs_seg_hdr,
  523. MDE_SEG_NUM,
  524. sizeof(struct mpi_coredump_segment_header)
  525. + sizeof(mpi_coredump->mde_regs),
  526. "MDE Registers");
  527. status = ql_get_mpi_regs(qdev, &mpi_coredump->mde_regs[0],
  528. MDE_REGS_ADDR, MDE_REGS_CNT);
  529. if (status)
  530. goto err;
  531. ql_build_coredump_seg_header(&mpi_coredump->misc_nic_seg_hdr,
  532. MISC_NIC_INFO_SEG_NUM,
  533. sizeof(struct mpi_coredump_segment_header)
  534. + sizeof(mpi_coredump->misc_nic_info),
  535. "MISC NIC INFO");
  536. mpi_coredump->misc_nic_info.rx_ring_count = qdev->rx_ring_count;
  537. mpi_coredump->misc_nic_info.tx_ring_count = qdev->tx_ring_count;
  538. mpi_coredump->misc_nic_info.intr_count = qdev->intr_count;
  539. mpi_coredump->misc_nic_info.function = qdev->func;
  540. /* Segment 31 */
  541. /* Get indexed register values. */
  542. ql_build_coredump_seg_header(&mpi_coredump->intr_states_seg_hdr,
  543. INTR_STATES_SEG_NUM,
  544. sizeof(struct mpi_coredump_segment_header)
  545. + sizeof(mpi_coredump->intr_states),
  546. "INTR States");
  547. ql_get_intr_states(qdev, &mpi_coredump->intr_states[0]);
  548. ql_build_coredump_seg_header(&mpi_coredump->cam_entries_seg_hdr,
  549. CAM_ENTRIES_SEG_NUM,
  550. sizeof(struct mpi_coredump_segment_header)
  551. + sizeof(mpi_coredump->cam_entries),
  552. "CAM Entries");
  553. status = ql_get_cam_entries(qdev, &mpi_coredump->cam_entries[0]);
  554. if (status)
  555. goto err;
  556. ql_build_coredump_seg_header(&mpi_coredump->nic_routing_words_seg_hdr,
  557. ROUTING_WORDS_SEG_NUM,
  558. sizeof(struct mpi_coredump_segment_header)
  559. + sizeof(mpi_coredump->nic_routing_words),
  560. "Routing Words");
  561. status = ql_get_routing_entries(qdev,
  562. &mpi_coredump->nic_routing_words[0]);
  563. if (status)
  564. goto err;
  565. /* Segment 34 (Rev C. step 23) */
  566. ql_build_coredump_seg_header(&mpi_coredump->ets_seg_hdr,
  567. ETS_SEG_NUM,
  568. sizeof(struct mpi_coredump_segment_header)
  569. + sizeof(mpi_coredump->ets),
  570. "ETS Registers");
  571. status = ql_get_ets_regs(qdev, &mpi_coredump->ets[0]);
  572. if (status)
  573. goto err;
  574. ql_build_coredump_seg_header(&mpi_coredump->probe_dump_seg_hdr,
  575. PROBE_DUMP_SEG_NUM,
  576. sizeof(struct mpi_coredump_segment_header)
  577. + sizeof(mpi_coredump->probe_dump),
  578. "Probe Dump");
  579. ql_get_probe_dump(qdev, &mpi_coredump->probe_dump[0]);
  580. ql_build_coredump_seg_header(&mpi_coredump->routing_reg_seg_hdr,
  581. ROUTING_INDEX_SEG_NUM,
  582. sizeof(struct mpi_coredump_segment_header)
  583. + sizeof(mpi_coredump->routing_regs),
  584. "Routing Regs");
  585. status = ql_get_routing_index_registers(qdev,
  586. &mpi_coredump->routing_regs[0]);
  587. if (status)
  588. goto err;
  589. ql_build_coredump_seg_header(&mpi_coredump->mac_prot_reg_seg_hdr,
  590. MAC_PROTOCOL_SEG_NUM,
  591. sizeof(struct mpi_coredump_segment_header)
  592. + sizeof(mpi_coredump->mac_prot_regs),
  593. "MAC Prot Regs");
  594. ql_get_mac_protocol_registers(qdev, &mpi_coredump->mac_prot_regs[0]);
  595. /* Get the semaphore registers for all 5 functions */
  596. ql_build_coredump_seg_header(&mpi_coredump->sem_regs_seg_hdr,
  597. SEM_REGS_SEG_NUM,
  598. sizeof(struct mpi_coredump_segment_header) +
  599. sizeof(mpi_coredump->sem_regs), "Sem Registers");
  600. ql_get_sem_registers(qdev, &mpi_coredump->sem_regs[0]);
  601. /* Prevent the mpi restarting while we dump the memory.*/
  602. ql_write_mpi_reg(qdev, MPI_TEST_FUNC_RST_STS, MPI_TEST_FUNC_RST_FRC);
  603. /* clear the pause */
  604. status = ql_unpause_mpi_risc(qdev);
  605. if (status) {
  606. QPRINTK(qdev, DRV, ERR,
  607. "Failed RISC unpause. Status = 0x%.08x\n", status);
  608. goto err;
  609. }
  610. /* Reset the RISC so we can dump RAM */
  611. status = ql_hard_reset_mpi_risc(qdev);
  612. if (status) {
  613. QPRINTK(qdev, DRV, ERR,
  614. "Failed RISC reset. Status = 0x%.08x\n", status);
  615. goto err;
  616. }
  617. ql_build_coredump_seg_header(&mpi_coredump->code_ram_seg_hdr,
  618. WCS_RAM_SEG_NUM,
  619. sizeof(struct mpi_coredump_segment_header)
  620. + sizeof(mpi_coredump->code_ram),
  621. "WCS RAM");
  622. status = ql_dump_risc_ram_area(qdev, &mpi_coredump->code_ram[0],
  623. CODE_RAM_ADDR, CODE_RAM_CNT);
  624. if (status) {
  625. QPRINTK(qdev, DRV, ERR,
  626. "Failed Dump of CODE RAM. Status = 0x%.08x\n", status);
  627. goto err;
  628. }
  629. /* Insert the segment header */
  630. ql_build_coredump_seg_header(&mpi_coredump->memc_ram_seg_hdr,
  631. MEMC_RAM_SEG_NUM,
  632. sizeof(struct mpi_coredump_segment_header)
  633. + sizeof(mpi_coredump->memc_ram),
  634. "MEMC RAM");
  635. status = ql_dump_risc_ram_area(qdev, &mpi_coredump->memc_ram[0],
  636. MEMC_RAM_ADDR, MEMC_RAM_CNT);
  637. if (status) {
  638. QPRINTK(qdev, DRV, ERR,
  639. "Failed Dump of MEMC RAM. Status = 0x%.08x\n", status);
  640. goto err;
  641. }
  642. err:
  643. ql_sem_unlock(qdev, SEM_PROC_REG_MASK); /* does flush too */
  644. return status;
  645. }
  646. void ql_gen_reg_dump(struct ql_adapter *qdev,
  647. struct ql_reg_dump *mpi_coredump)
  648. {
  649. int i, status;
  650. memset(&(mpi_coredump->mpi_global_header), 0,
  651. sizeof(struct mpi_coredump_global_header));
  652. mpi_coredump->mpi_global_header.cookie = MPI_COREDUMP_COOKIE;
  653. mpi_coredump->mpi_global_header.headerSize =
  654. sizeof(struct mpi_coredump_global_header);
  655. mpi_coredump->mpi_global_header.imageSize =
  656. sizeof(struct ql_reg_dump);
  657. memcpy(mpi_coredump->mpi_global_header.idString, "MPI Coredump",
  658. sizeof(mpi_coredump->mpi_global_header.idString));
  659. /* segment 16 */
  660. ql_build_coredump_seg_header(&mpi_coredump->misc_nic_seg_hdr,
  661. MISC_NIC_INFO_SEG_NUM,
  662. sizeof(struct mpi_coredump_segment_header)
  663. + sizeof(mpi_coredump->misc_nic_info),
  664. "MISC NIC INFO");
  665. mpi_coredump->misc_nic_info.rx_ring_count = qdev->rx_ring_count;
  666. mpi_coredump->misc_nic_info.tx_ring_count = qdev->tx_ring_count;
  667. mpi_coredump->misc_nic_info.intr_count = qdev->intr_count;
  668. mpi_coredump->misc_nic_info.function = qdev->func;
  669. /* Segment 16, Rev C. Step 18 */
  670. ql_build_coredump_seg_header(&mpi_coredump->nic_regs_seg_hdr,
  671. NIC1_CONTROL_SEG_NUM,
  672. sizeof(struct mpi_coredump_segment_header)
  673. + sizeof(mpi_coredump->nic_regs),
  674. "NIC Registers");
  675. /* Get generic reg dump */
  676. for (i = 0; i < 64; i++)
  677. mpi_coredump->nic_regs[i] = ql_read32(qdev, i * sizeof(u32));
  678. /* Segment 31 */
  679. /* Get indexed register values. */
  680. ql_build_coredump_seg_header(&mpi_coredump->intr_states_seg_hdr,
  681. INTR_STATES_SEG_NUM,
  682. sizeof(struct mpi_coredump_segment_header)
  683. + sizeof(mpi_coredump->intr_states),
  684. "INTR States");
  685. ql_get_intr_states(qdev, &mpi_coredump->intr_states[0]);
  686. ql_build_coredump_seg_header(&mpi_coredump->cam_entries_seg_hdr,
  687. CAM_ENTRIES_SEG_NUM,
  688. sizeof(struct mpi_coredump_segment_header)
  689. + sizeof(mpi_coredump->cam_entries),
  690. "CAM Entries");
  691. status = ql_get_cam_entries(qdev, &mpi_coredump->cam_entries[0]);
  692. if (status)
  693. return;
  694. ql_build_coredump_seg_header(&mpi_coredump->nic_routing_words_seg_hdr,
  695. ROUTING_WORDS_SEG_NUM,
  696. sizeof(struct mpi_coredump_segment_header)
  697. + sizeof(mpi_coredump->nic_routing_words),
  698. "Routing Words");
  699. status = ql_get_routing_entries(qdev,
  700. &mpi_coredump->nic_routing_words[0]);
  701. if (status)
  702. return;
  703. /* Segment 34 (Rev C. step 23) */
  704. ql_build_coredump_seg_header(&mpi_coredump->ets_seg_hdr,
  705. ETS_SEG_NUM,
  706. sizeof(struct mpi_coredump_segment_header)
  707. + sizeof(mpi_coredump->ets),
  708. "ETS Registers");
  709. status = ql_get_ets_regs(qdev, &mpi_coredump->ets[0]);
  710. if (status)
  711. return;
  712. }
  713. /* Coredump to messages log file using separate worker thread */
  714. void ql_mpi_core_to_log(struct work_struct *work)
  715. {
  716. struct ql_adapter *qdev =
  717. container_of(work, struct ql_adapter, mpi_core_to_log.work);
  718. u32 *tmp, count;
  719. int i;
  720. count = sizeof(struct ql_mpi_coredump) / sizeof(u32);
  721. tmp = (u32 *)qdev->mpi_coredump;
  722. QPRINTK(qdev, DRV, DEBUG, "Core is dumping to log file!\n");
  723. for (i = 0; i < count; i += 8) {
  724. printk(KERN_ERR "%.08x: %.08x %.08x %.08x %.08x %.08x "
  725. "%.08x %.08x %.08x \n", i,
  726. tmp[i + 0],
  727. tmp[i + 1],
  728. tmp[i + 2],
  729. tmp[i + 3],
  730. tmp[i + 4],
  731. tmp[i + 5],
  732. tmp[i + 6],
  733. tmp[i + 7]);
  734. msleep(5);
  735. }
  736. }
  737. #ifdef QL_REG_DUMP
  738. static void ql_dump_intr_states(struct ql_adapter *qdev)
  739. {
  740. int i;
  741. u32 value;
  742. for (i = 0; i < qdev->intr_count; i++) {
  743. ql_write32(qdev, INTR_EN, qdev->intr_context[i].intr_read_mask);
  744. value = ql_read32(qdev, INTR_EN);
  745. printk(KERN_ERR PFX
  746. "%s: Interrupt %d is %s.\n",
  747. qdev->ndev->name, i,
  748. (value & INTR_EN_EN ? "enabled" : "disabled"));
  749. }
  750. }
  751. void ql_dump_xgmac_control_regs(struct ql_adapter *qdev)
  752. {
  753. u32 data;
  754. if (ql_sem_spinlock(qdev, qdev->xg_sem_mask)) {
  755. printk(KERN_ERR "%s: Couldn't get xgmac sem.\n", __func__);
  756. return;
  757. }
  758. ql_read_xgmac_reg(qdev, PAUSE_SRC_LO, &data);
  759. printk(KERN_ERR PFX "%s: PAUSE_SRC_LO = 0x%.08x.\n", qdev->ndev->name,
  760. data);
  761. ql_read_xgmac_reg(qdev, PAUSE_SRC_HI, &data);
  762. printk(KERN_ERR PFX "%s: PAUSE_SRC_HI = 0x%.08x.\n", qdev->ndev->name,
  763. data);
  764. ql_read_xgmac_reg(qdev, GLOBAL_CFG, &data);
  765. printk(KERN_ERR PFX "%s: GLOBAL_CFG = 0x%.08x.\n", qdev->ndev->name,
  766. data);
  767. ql_read_xgmac_reg(qdev, TX_CFG, &data);
  768. printk(KERN_ERR PFX "%s: TX_CFG = 0x%.08x.\n", qdev->ndev->name, data);
  769. ql_read_xgmac_reg(qdev, RX_CFG, &data);
  770. printk(KERN_ERR PFX "%s: RX_CFG = 0x%.08x.\n", qdev->ndev->name, data);
  771. ql_read_xgmac_reg(qdev, FLOW_CTL, &data);
  772. printk(KERN_ERR PFX "%s: FLOW_CTL = 0x%.08x.\n", qdev->ndev->name,
  773. data);
  774. ql_read_xgmac_reg(qdev, PAUSE_OPCODE, &data);
  775. printk(KERN_ERR PFX "%s: PAUSE_OPCODE = 0x%.08x.\n", qdev->ndev->name,
  776. data);
  777. ql_read_xgmac_reg(qdev, PAUSE_TIMER, &data);
  778. printk(KERN_ERR PFX "%s: PAUSE_TIMER = 0x%.08x.\n", qdev->ndev->name,
  779. data);
  780. ql_read_xgmac_reg(qdev, PAUSE_FRM_DEST_LO, &data);
  781. printk(KERN_ERR PFX "%s: PAUSE_FRM_DEST_LO = 0x%.08x.\n",
  782. qdev->ndev->name, data);
  783. ql_read_xgmac_reg(qdev, PAUSE_FRM_DEST_HI, &data);
  784. printk(KERN_ERR PFX "%s: PAUSE_FRM_DEST_HI = 0x%.08x.\n",
  785. qdev->ndev->name, data);
  786. ql_read_xgmac_reg(qdev, MAC_TX_PARAMS, &data);
  787. printk(KERN_ERR PFX "%s: MAC_TX_PARAMS = 0x%.08x.\n", qdev->ndev->name,
  788. data);
  789. ql_read_xgmac_reg(qdev, MAC_RX_PARAMS, &data);
  790. printk(KERN_ERR PFX "%s: MAC_RX_PARAMS = 0x%.08x.\n", qdev->ndev->name,
  791. data);
  792. ql_read_xgmac_reg(qdev, MAC_SYS_INT, &data);
  793. printk(KERN_ERR PFX "%s: MAC_SYS_INT = 0x%.08x.\n", qdev->ndev->name,
  794. data);
  795. ql_read_xgmac_reg(qdev, MAC_SYS_INT_MASK, &data);
  796. printk(KERN_ERR PFX "%s: MAC_SYS_INT_MASK = 0x%.08x.\n",
  797. qdev->ndev->name, data);
  798. ql_read_xgmac_reg(qdev, MAC_MGMT_INT, &data);
  799. printk(KERN_ERR PFX "%s: MAC_MGMT_INT = 0x%.08x.\n", qdev->ndev->name,
  800. data);
  801. ql_read_xgmac_reg(qdev, MAC_MGMT_IN_MASK, &data);
  802. printk(KERN_ERR PFX "%s: MAC_MGMT_IN_MASK = 0x%.08x.\n",
  803. qdev->ndev->name, data);
  804. ql_read_xgmac_reg(qdev, EXT_ARB_MODE, &data);
  805. printk(KERN_ERR PFX "%s: EXT_ARB_MODE = 0x%.08x.\n", qdev->ndev->name,
  806. data);
  807. ql_sem_unlock(qdev, qdev->xg_sem_mask);
  808. }
  809. static void ql_dump_ets_regs(struct ql_adapter *qdev)
  810. {
  811. }
  812. static void ql_dump_cam_entries(struct ql_adapter *qdev)
  813. {
  814. int i;
  815. u32 value[3];
  816. i = ql_sem_spinlock(qdev, SEM_MAC_ADDR_MASK);
  817. if (i)
  818. return;
  819. for (i = 0; i < 4; i++) {
  820. if (ql_get_mac_addr_reg(qdev, MAC_ADDR_TYPE_CAM_MAC, i, value)) {
  821. printk(KERN_ERR PFX
  822. "%s: Failed read of mac index register.\n",
  823. __func__);
  824. return;
  825. } else {
  826. if (value[0])
  827. printk(KERN_ERR PFX
  828. "%s: CAM index %d CAM Lookup Lower = 0x%.08x:%.08x, Output = 0x%.08x.\n",
  829. qdev->ndev->name, i, value[1], value[0],
  830. value[2]);
  831. }
  832. }
  833. for (i = 0; i < 32; i++) {
  834. if (ql_get_mac_addr_reg
  835. (qdev, MAC_ADDR_TYPE_MULTI_MAC, i, value)) {
  836. printk(KERN_ERR PFX
  837. "%s: Failed read of mac index register.\n",
  838. __func__);
  839. return;
  840. } else {
  841. if (value[0])
  842. printk(KERN_ERR PFX
  843. "%s: MCAST index %d CAM Lookup Lower = 0x%.08x:%.08x.\n",
  844. qdev->ndev->name, i, value[1], value[0]);
  845. }
  846. }
  847. ql_sem_unlock(qdev, SEM_MAC_ADDR_MASK);
  848. }
  849. void ql_dump_routing_entries(struct ql_adapter *qdev)
  850. {
  851. int i;
  852. u32 value;
  853. i = ql_sem_spinlock(qdev, SEM_RT_IDX_MASK);
  854. if (i)
  855. return;
  856. for (i = 0; i < 16; i++) {
  857. value = 0;
  858. if (ql_get_routing_reg(qdev, i, &value)) {
  859. printk(KERN_ERR PFX
  860. "%s: Failed read of routing index register.\n",
  861. __func__);
  862. return;
  863. } else {
  864. if (value)
  865. printk(KERN_ERR PFX
  866. "%s: Routing Mask %d = 0x%.08x.\n",
  867. qdev->ndev->name, i, value);
  868. }
  869. }
  870. ql_sem_unlock(qdev, SEM_RT_IDX_MASK);
  871. }
  872. void ql_dump_regs(struct ql_adapter *qdev)
  873. {
  874. printk(KERN_ERR PFX "reg dump for function #%d.\n", qdev->func);
  875. printk(KERN_ERR PFX "SYS = 0x%x.\n",
  876. ql_read32(qdev, SYS));
  877. printk(KERN_ERR PFX "RST_FO = 0x%x.\n",
  878. ql_read32(qdev, RST_FO));
  879. printk(KERN_ERR PFX "FSC = 0x%x.\n",
  880. ql_read32(qdev, FSC));
  881. printk(KERN_ERR PFX "CSR = 0x%x.\n",
  882. ql_read32(qdev, CSR));
  883. printk(KERN_ERR PFX "ICB_RID = 0x%x.\n",
  884. ql_read32(qdev, ICB_RID));
  885. printk(KERN_ERR PFX "ICB_L = 0x%x.\n",
  886. ql_read32(qdev, ICB_L));
  887. printk(KERN_ERR PFX "ICB_H = 0x%x.\n",
  888. ql_read32(qdev, ICB_H));
  889. printk(KERN_ERR PFX "CFG = 0x%x.\n",
  890. ql_read32(qdev, CFG));
  891. printk(KERN_ERR PFX "BIOS_ADDR = 0x%x.\n",
  892. ql_read32(qdev, BIOS_ADDR));
  893. printk(KERN_ERR PFX "STS = 0x%x.\n",
  894. ql_read32(qdev, STS));
  895. printk(KERN_ERR PFX "INTR_EN = 0x%x.\n",
  896. ql_read32(qdev, INTR_EN));
  897. printk(KERN_ERR PFX "INTR_MASK = 0x%x.\n",
  898. ql_read32(qdev, INTR_MASK));
  899. printk(KERN_ERR PFX "ISR1 = 0x%x.\n",
  900. ql_read32(qdev, ISR1));
  901. printk(KERN_ERR PFX "ISR2 = 0x%x.\n",
  902. ql_read32(qdev, ISR2));
  903. printk(KERN_ERR PFX "ISR3 = 0x%x.\n",
  904. ql_read32(qdev, ISR3));
  905. printk(KERN_ERR PFX "ISR4 = 0x%x.\n",
  906. ql_read32(qdev, ISR4));
  907. printk(KERN_ERR PFX "REV_ID = 0x%x.\n",
  908. ql_read32(qdev, REV_ID));
  909. printk(KERN_ERR PFX "FRC_ECC_ERR = 0x%x.\n",
  910. ql_read32(qdev, FRC_ECC_ERR));
  911. printk(KERN_ERR PFX "ERR_STS = 0x%x.\n",
  912. ql_read32(qdev, ERR_STS));
  913. printk(KERN_ERR PFX "RAM_DBG_ADDR = 0x%x.\n",
  914. ql_read32(qdev, RAM_DBG_ADDR));
  915. printk(KERN_ERR PFX "RAM_DBG_DATA = 0x%x.\n",
  916. ql_read32(qdev, RAM_DBG_DATA));
  917. printk(KERN_ERR PFX "ECC_ERR_CNT = 0x%x.\n",
  918. ql_read32(qdev, ECC_ERR_CNT));
  919. printk(KERN_ERR PFX "SEM = 0x%x.\n",
  920. ql_read32(qdev, SEM));
  921. printk(KERN_ERR PFX "GPIO_1 = 0x%x.\n",
  922. ql_read32(qdev, GPIO_1));
  923. printk(KERN_ERR PFX "GPIO_2 = 0x%x.\n",
  924. ql_read32(qdev, GPIO_2));
  925. printk(KERN_ERR PFX "GPIO_3 = 0x%x.\n",
  926. ql_read32(qdev, GPIO_3));
  927. printk(KERN_ERR PFX "XGMAC_ADDR = 0x%x.\n",
  928. ql_read32(qdev, XGMAC_ADDR));
  929. printk(KERN_ERR PFX "XGMAC_DATA = 0x%x.\n",
  930. ql_read32(qdev, XGMAC_DATA));
  931. printk(KERN_ERR PFX "NIC_ETS = 0x%x.\n",
  932. ql_read32(qdev, NIC_ETS));
  933. printk(KERN_ERR PFX "CNA_ETS = 0x%x.\n",
  934. ql_read32(qdev, CNA_ETS));
  935. printk(KERN_ERR PFX "FLASH_ADDR = 0x%x.\n",
  936. ql_read32(qdev, FLASH_ADDR));
  937. printk(KERN_ERR PFX "FLASH_DATA = 0x%x.\n",
  938. ql_read32(qdev, FLASH_DATA));
  939. printk(KERN_ERR PFX "CQ_STOP = 0x%x.\n",
  940. ql_read32(qdev, CQ_STOP));
  941. printk(KERN_ERR PFX "PAGE_TBL_RID = 0x%x.\n",
  942. ql_read32(qdev, PAGE_TBL_RID));
  943. printk(KERN_ERR PFX "WQ_PAGE_TBL_LO = 0x%x.\n",
  944. ql_read32(qdev, WQ_PAGE_TBL_LO));
  945. printk(KERN_ERR PFX "WQ_PAGE_TBL_HI = 0x%x.\n",
  946. ql_read32(qdev, WQ_PAGE_TBL_HI));
  947. printk(KERN_ERR PFX "CQ_PAGE_TBL_LO = 0x%x.\n",
  948. ql_read32(qdev, CQ_PAGE_TBL_LO));
  949. printk(KERN_ERR PFX "CQ_PAGE_TBL_HI = 0x%x.\n",
  950. ql_read32(qdev, CQ_PAGE_TBL_HI));
  951. printk(KERN_ERR PFX "COS_DFLT_CQ1 = 0x%x.\n",
  952. ql_read32(qdev, COS_DFLT_CQ1));
  953. printk(KERN_ERR PFX "COS_DFLT_CQ2 = 0x%x.\n",
  954. ql_read32(qdev, COS_DFLT_CQ2));
  955. printk(KERN_ERR PFX "SPLT_HDR = 0x%x.\n",
  956. ql_read32(qdev, SPLT_HDR));
  957. printk(KERN_ERR PFX "FC_PAUSE_THRES = 0x%x.\n",
  958. ql_read32(qdev, FC_PAUSE_THRES));
  959. printk(KERN_ERR PFX "NIC_PAUSE_THRES = 0x%x.\n",
  960. ql_read32(qdev, NIC_PAUSE_THRES));
  961. printk(KERN_ERR PFX "FC_ETHERTYPE = 0x%x.\n",
  962. ql_read32(qdev, FC_ETHERTYPE));
  963. printk(KERN_ERR PFX "FC_RCV_CFG = 0x%x.\n",
  964. ql_read32(qdev, FC_RCV_CFG));
  965. printk(KERN_ERR PFX "NIC_RCV_CFG = 0x%x.\n",
  966. ql_read32(qdev, NIC_RCV_CFG));
  967. printk(KERN_ERR PFX "FC_COS_TAGS = 0x%x.\n",
  968. ql_read32(qdev, FC_COS_TAGS));
  969. printk(KERN_ERR PFX "NIC_COS_TAGS = 0x%x.\n",
  970. ql_read32(qdev, NIC_COS_TAGS));
  971. printk(KERN_ERR PFX "MGMT_RCV_CFG = 0x%x.\n",
  972. ql_read32(qdev, MGMT_RCV_CFG));
  973. printk(KERN_ERR PFX "XG_SERDES_ADDR = 0x%x.\n",
  974. ql_read32(qdev, XG_SERDES_ADDR));
  975. printk(KERN_ERR PFX "XG_SERDES_DATA = 0x%x.\n",
  976. ql_read32(qdev, XG_SERDES_DATA));
  977. printk(KERN_ERR PFX "PRB_MX_ADDR = 0x%x.\n",
  978. ql_read32(qdev, PRB_MX_ADDR));
  979. printk(KERN_ERR PFX "PRB_MX_DATA = 0x%x.\n",
  980. ql_read32(qdev, PRB_MX_DATA));
  981. ql_dump_intr_states(qdev);
  982. ql_dump_xgmac_control_regs(qdev);
  983. ql_dump_ets_regs(qdev);
  984. ql_dump_cam_entries(qdev);
  985. ql_dump_routing_entries(qdev);
  986. }
  987. #endif
  988. #ifdef QL_STAT_DUMP
  989. void ql_dump_stat(struct ql_adapter *qdev)
  990. {
  991. printk(KERN_ERR "%s: Enter.\n", __func__);
  992. printk(KERN_ERR "tx_pkts = %ld\n",
  993. (unsigned long)qdev->nic_stats.tx_pkts);
  994. printk(KERN_ERR "tx_bytes = %ld\n",
  995. (unsigned long)qdev->nic_stats.tx_bytes);
  996. printk(KERN_ERR "tx_mcast_pkts = %ld.\n",
  997. (unsigned long)qdev->nic_stats.tx_mcast_pkts);
  998. printk(KERN_ERR "tx_bcast_pkts = %ld.\n",
  999. (unsigned long)qdev->nic_stats.tx_bcast_pkts);
  1000. printk(KERN_ERR "tx_ucast_pkts = %ld.\n",
  1001. (unsigned long)qdev->nic_stats.tx_ucast_pkts);
  1002. printk(KERN_ERR "tx_ctl_pkts = %ld.\n",
  1003. (unsigned long)qdev->nic_stats.tx_ctl_pkts);
  1004. printk(KERN_ERR "tx_pause_pkts = %ld.\n",
  1005. (unsigned long)qdev->nic_stats.tx_pause_pkts);
  1006. printk(KERN_ERR "tx_64_pkt = %ld.\n",
  1007. (unsigned long)qdev->nic_stats.tx_64_pkt);
  1008. printk(KERN_ERR "tx_65_to_127_pkt = %ld.\n",
  1009. (unsigned long)qdev->nic_stats.tx_65_to_127_pkt);
  1010. printk(KERN_ERR "tx_128_to_255_pkt = %ld.\n",
  1011. (unsigned long)qdev->nic_stats.tx_128_to_255_pkt);
  1012. printk(KERN_ERR "tx_256_511_pkt = %ld.\n",
  1013. (unsigned long)qdev->nic_stats.tx_256_511_pkt);
  1014. printk(KERN_ERR "tx_512_to_1023_pkt = %ld.\n",
  1015. (unsigned long)qdev->nic_stats.tx_512_to_1023_pkt);
  1016. printk(KERN_ERR "tx_1024_to_1518_pkt = %ld.\n",
  1017. (unsigned long)qdev->nic_stats.tx_1024_to_1518_pkt);
  1018. printk(KERN_ERR "tx_1519_to_max_pkt = %ld.\n",
  1019. (unsigned long)qdev->nic_stats.tx_1519_to_max_pkt);
  1020. printk(KERN_ERR "tx_undersize_pkt = %ld.\n",
  1021. (unsigned long)qdev->nic_stats.tx_undersize_pkt);
  1022. printk(KERN_ERR "tx_oversize_pkt = %ld.\n",
  1023. (unsigned long)qdev->nic_stats.tx_oversize_pkt);
  1024. printk(KERN_ERR "rx_bytes = %ld.\n",
  1025. (unsigned long)qdev->nic_stats.rx_bytes);
  1026. printk(KERN_ERR "rx_bytes_ok = %ld.\n",
  1027. (unsigned long)qdev->nic_stats.rx_bytes_ok);
  1028. printk(KERN_ERR "rx_pkts = %ld.\n",
  1029. (unsigned long)qdev->nic_stats.rx_pkts);
  1030. printk(KERN_ERR "rx_pkts_ok = %ld.\n",
  1031. (unsigned long)qdev->nic_stats.rx_pkts_ok);
  1032. printk(KERN_ERR "rx_bcast_pkts = %ld.\n",
  1033. (unsigned long)qdev->nic_stats.rx_bcast_pkts);
  1034. printk(KERN_ERR "rx_mcast_pkts = %ld.\n",
  1035. (unsigned long)qdev->nic_stats.rx_mcast_pkts);
  1036. printk(KERN_ERR "rx_ucast_pkts = %ld.\n",
  1037. (unsigned long)qdev->nic_stats.rx_ucast_pkts);
  1038. printk(KERN_ERR "rx_undersize_pkts = %ld.\n",
  1039. (unsigned long)qdev->nic_stats.rx_undersize_pkts);
  1040. printk(KERN_ERR "rx_oversize_pkts = %ld.\n",
  1041. (unsigned long)qdev->nic_stats.rx_oversize_pkts);
  1042. printk(KERN_ERR "rx_jabber_pkts = %ld.\n",
  1043. (unsigned long)qdev->nic_stats.rx_jabber_pkts);
  1044. printk(KERN_ERR "rx_undersize_fcerr_pkts = %ld.\n",
  1045. (unsigned long)qdev->nic_stats.rx_undersize_fcerr_pkts);
  1046. printk(KERN_ERR "rx_drop_events = %ld.\n",
  1047. (unsigned long)qdev->nic_stats.rx_drop_events);
  1048. printk(KERN_ERR "rx_fcerr_pkts = %ld.\n",
  1049. (unsigned long)qdev->nic_stats.rx_fcerr_pkts);
  1050. printk(KERN_ERR "rx_align_err = %ld.\n",
  1051. (unsigned long)qdev->nic_stats.rx_align_err);
  1052. printk(KERN_ERR "rx_symbol_err = %ld.\n",
  1053. (unsigned long)qdev->nic_stats.rx_symbol_err);
  1054. printk(KERN_ERR "rx_mac_err = %ld.\n",
  1055. (unsigned long)qdev->nic_stats.rx_mac_err);
  1056. printk(KERN_ERR "rx_ctl_pkts = %ld.\n",
  1057. (unsigned long)qdev->nic_stats.rx_ctl_pkts);
  1058. printk(KERN_ERR "rx_pause_pkts = %ld.\n",
  1059. (unsigned long)qdev->nic_stats.rx_pause_pkts);
  1060. printk(KERN_ERR "rx_64_pkts = %ld.\n",
  1061. (unsigned long)qdev->nic_stats.rx_64_pkts);
  1062. printk(KERN_ERR "rx_65_to_127_pkts = %ld.\n",
  1063. (unsigned long)qdev->nic_stats.rx_65_to_127_pkts);
  1064. printk(KERN_ERR "rx_128_255_pkts = %ld.\n",
  1065. (unsigned long)qdev->nic_stats.rx_128_255_pkts);
  1066. printk(KERN_ERR "rx_256_511_pkts = %ld.\n",
  1067. (unsigned long)qdev->nic_stats.rx_256_511_pkts);
  1068. printk(KERN_ERR "rx_512_to_1023_pkts = %ld.\n",
  1069. (unsigned long)qdev->nic_stats.rx_512_to_1023_pkts);
  1070. printk(KERN_ERR "rx_1024_to_1518_pkts = %ld.\n",
  1071. (unsigned long)qdev->nic_stats.rx_1024_to_1518_pkts);
  1072. printk(KERN_ERR "rx_1519_to_max_pkts = %ld.\n",
  1073. (unsigned long)qdev->nic_stats.rx_1519_to_max_pkts);
  1074. printk(KERN_ERR "rx_len_err_pkts = %ld.\n",
  1075. (unsigned long)qdev->nic_stats.rx_len_err_pkts);
  1076. };
  1077. #endif
  1078. #ifdef QL_DEV_DUMP
  1079. void ql_dump_qdev(struct ql_adapter *qdev)
  1080. {
  1081. int i;
  1082. printk(KERN_ERR PFX "qdev->flags = %lx.\n",
  1083. qdev->flags);
  1084. printk(KERN_ERR PFX "qdev->vlgrp = %p.\n",
  1085. qdev->vlgrp);
  1086. printk(KERN_ERR PFX "qdev->pdev = %p.\n",
  1087. qdev->pdev);
  1088. printk(KERN_ERR PFX "qdev->ndev = %p.\n",
  1089. qdev->ndev);
  1090. printk(KERN_ERR PFX "qdev->chip_rev_id = %d.\n",
  1091. qdev->chip_rev_id);
  1092. printk(KERN_ERR PFX "qdev->reg_base = %p.\n",
  1093. qdev->reg_base);
  1094. printk(KERN_ERR PFX "qdev->doorbell_area = %p.\n",
  1095. qdev->doorbell_area);
  1096. printk(KERN_ERR PFX "qdev->doorbell_area_size = %d.\n",
  1097. qdev->doorbell_area_size);
  1098. printk(KERN_ERR PFX "msg_enable = %x.\n",
  1099. qdev->msg_enable);
  1100. printk(KERN_ERR PFX "qdev->rx_ring_shadow_reg_area = %p.\n",
  1101. qdev->rx_ring_shadow_reg_area);
  1102. printk(KERN_ERR PFX "qdev->rx_ring_shadow_reg_dma = %llx.\n",
  1103. (unsigned long long) qdev->rx_ring_shadow_reg_dma);
  1104. printk(KERN_ERR PFX "qdev->tx_ring_shadow_reg_area = %p.\n",
  1105. qdev->tx_ring_shadow_reg_area);
  1106. printk(KERN_ERR PFX "qdev->tx_ring_shadow_reg_dma = %llx.\n",
  1107. (unsigned long long) qdev->tx_ring_shadow_reg_dma);
  1108. printk(KERN_ERR PFX "qdev->intr_count = %d.\n",
  1109. qdev->intr_count);
  1110. if (qdev->msi_x_entry)
  1111. for (i = 0; i < qdev->intr_count; i++) {
  1112. printk(KERN_ERR PFX
  1113. "msi_x_entry.[%d]vector = %d.\n", i,
  1114. qdev->msi_x_entry[i].vector);
  1115. printk(KERN_ERR PFX
  1116. "msi_x_entry.[%d]entry = %d.\n", i,
  1117. qdev->msi_x_entry[i].entry);
  1118. }
  1119. for (i = 0; i < qdev->intr_count; i++) {
  1120. printk(KERN_ERR PFX
  1121. "intr_context[%d].qdev = %p.\n", i,
  1122. qdev->intr_context[i].qdev);
  1123. printk(KERN_ERR PFX
  1124. "intr_context[%d].intr = %d.\n", i,
  1125. qdev->intr_context[i].intr);
  1126. printk(KERN_ERR PFX
  1127. "intr_context[%d].hooked = %d.\n", i,
  1128. qdev->intr_context[i].hooked);
  1129. printk(KERN_ERR PFX
  1130. "intr_context[%d].intr_en_mask = 0x%08x.\n", i,
  1131. qdev->intr_context[i].intr_en_mask);
  1132. printk(KERN_ERR PFX
  1133. "intr_context[%d].intr_dis_mask = 0x%08x.\n", i,
  1134. qdev->intr_context[i].intr_dis_mask);
  1135. printk(KERN_ERR PFX
  1136. "intr_context[%d].intr_read_mask = 0x%08x.\n", i,
  1137. qdev->intr_context[i].intr_read_mask);
  1138. }
  1139. printk(KERN_ERR PFX "qdev->tx_ring_count = %d.\n", qdev->tx_ring_count);
  1140. printk(KERN_ERR PFX "qdev->rx_ring_count = %d.\n", qdev->rx_ring_count);
  1141. printk(KERN_ERR PFX "qdev->ring_mem_size = %d.\n", qdev->ring_mem_size);
  1142. printk(KERN_ERR PFX "qdev->ring_mem = %p.\n", qdev->ring_mem);
  1143. printk(KERN_ERR PFX "qdev->intr_count = %d.\n", qdev->intr_count);
  1144. printk(KERN_ERR PFX "qdev->tx_ring = %p.\n",
  1145. qdev->tx_ring);
  1146. printk(KERN_ERR PFX "qdev->rss_ring_count = %d.\n",
  1147. qdev->rss_ring_count);
  1148. printk(KERN_ERR PFX "qdev->rx_ring = %p.\n", qdev->rx_ring);
  1149. printk(KERN_ERR PFX "qdev->default_rx_queue = %d.\n",
  1150. qdev->default_rx_queue);
  1151. printk(KERN_ERR PFX "qdev->xg_sem_mask = 0x%08x.\n",
  1152. qdev->xg_sem_mask);
  1153. printk(KERN_ERR PFX "qdev->port_link_up = 0x%08x.\n",
  1154. qdev->port_link_up);
  1155. printk(KERN_ERR PFX "qdev->port_init = 0x%08x.\n",
  1156. qdev->port_init);
  1157. }
  1158. #endif
  1159. #ifdef QL_CB_DUMP
  1160. void ql_dump_wqicb(struct wqicb *wqicb)
  1161. {
  1162. printk(KERN_ERR PFX "Dumping wqicb stuff...\n");
  1163. printk(KERN_ERR PFX "wqicb->len = 0x%x.\n", le16_to_cpu(wqicb->len));
  1164. printk(KERN_ERR PFX "wqicb->flags = %x.\n", le16_to_cpu(wqicb->flags));
  1165. printk(KERN_ERR PFX "wqicb->cq_id_rss = %d.\n",
  1166. le16_to_cpu(wqicb->cq_id_rss));
  1167. printk(KERN_ERR PFX "wqicb->rid = 0x%x.\n", le16_to_cpu(wqicb->rid));
  1168. printk(KERN_ERR PFX "wqicb->wq_addr = 0x%llx.\n",
  1169. (unsigned long long) le64_to_cpu(wqicb->addr));
  1170. printk(KERN_ERR PFX "wqicb->wq_cnsmr_idx_addr = 0x%llx.\n",
  1171. (unsigned long long) le64_to_cpu(wqicb->cnsmr_idx_addr));
  1172. }
  1173. void ql_dump_tx_ring(struct tx_ring *tx_ring)
  1174. {
  1175. if (tx_ring == NULL)
  1176. return;
  1177. printk(KERN_ERR PFX
  1178. "===================== Dumping tx_ring %d ===============.\n",
  1179. tx_ring->wq_id);
  1180. printk(KERN_ERR PFX "tx_ring->base = %p.\n", tx_ring->wq_base);
  1181. printk(KERN_ERR PFX "tx_ring->base_dma = 0x%llx.\n",
  1182. (unsigned long long) tx_ring->wq_base_dma);
  1183. printk(KERN_ERR PFX
  1184. "tx_ring->cnsmr_idx_sh_reg, addr = 0x%p, value = %d.\n",
  1185. tx_ring->cnsmr_idx_sh_reg,
  1186. tx_ring->cnsmr_idx_sh_reg
  1187. ? ql_read_sh_reg(tx_ring->cnsmr_idx_sh_reg) : 0);
  1188. printk(KERN_ERR PFX "tx_ring->size = %d.\n", tx_ring->wq_size);
  1189. printk(KERN_ERR PFX "tx_ring->len = %d.\n", tx_ring->wq_len);
  1190. printk(KERN_ERR PFX "tx_ring->prod_idx_db_reg = %p.\n",
  1191. tx_ring->prod_idx_db_reg);
  1192. printk(KERN_ERR PFX "tx_ring->valid_db_reg = %p.\n",
  1193. tx_ring->valid_db_reg);
  1194. printk(KERN_ERR PFX "tx_ring->prod_idx = %d.\n", tx_ring->prod_idx);
  1195. printk(KERN_ERR PFX "tx_ring->cq_id = %d.\n", tx_ring->cq_id);
  1196. printk(KERN_ERR PFX "tx_ring->wq_id = %d.\n", tx_ring->wq_id);
  1197. printk(KERN_ERR PFX "tx_ring->q = %p.\n", tx_ring->q);
  1198. printk(KERN_ERR PFX "tx_ring->tx_count = %d.\n",
  1199. atomic_read(&tx_ring->tx_count));
  1200. }
  1201. void ql_dump_ricb(struct ricb *ricb)
  1202. {
  1203. int i;
  1204. printk(KERN_ERR PFX
  1205. "===================== Dumping ricb ===============.\n");
  1206. printk(KERN_ERR PFX "Dumping ricb stuff...\n");
  1207. printk(KERN_ERR PFX "ricb->base_cq = %d.\n", ricb->base_cq & 0x1f);
  1208. printk(KERN_ERR PFX "ricb->flags = %s%s%s%s%s%s%s%s%s.\n",
  1209. ricb->base_cq & RSS_L4K ? "RSS_L4K " : "",
  1210. ricb->flags & RSS_L6K ? "RSS_L6K " : "",
  1211. ricb->flags & RSS_LI ? "RSS_LI " : "",
  1212. ricb->flags & RSS_LB ? "RSS_LB " : "",
  1213. ricb->flags & RSS_LM ? "RSS_LM " : "",
  1214. ricb->flags & RSS_RI4 ? "RSS_RI4 " : "",
  1215. ricb->flags & RSS_RT4 ? "RSS_RT4 " : "",
  1216. ricb->flags & RSS_RI6 ? "RSS_RI6 " : "",
  1217. ricb->flags & RSS_RT6 ? "RSS_RT6 " : "");
  1218. printk(KERN_ERR PFX "ricb->mask = 0x%.04x.\n", le16_to_cpu(ricb->mask));
  1219. for (i = 0; i < 16; i++)
  1220. printk(KERN_ERR PFX "ricb->hash_cq_id[%d] = 0x%.08x.\n", i,
  1221. le32_to_cpu(ricb->hash_cq_id[i]));
  1222. for (i = 0; i < 10; i++)
  1223. printk(KERN_ERR PFX "ricb->ipv6_hash_key[%d] = 0x%.08x.\n", i,
  1224. le32_to_cpu(ricb->ipv6_hash_key[i]));
  1225. for (i = 0; i < 4; i++)
  1226. printk(KERN_ERR PFX "ricb->ipv4_hash_key[%d] = 0x%.08x.\n", i,
  1227. le32_to_cpu(ricb->ipv4_hash_key[i]));
  1228. }
  1229. void ql_dump_cqicb(struct cqicb *cqicb)
  1230. {
  1231. printk(KERN_ERR PFX "Dumping cqicb stuff...\n");
  1232. printk(KERN_ERR PFX "cqicb->msix_vect = %d.\n", cqicb->msix_vect);
  1233. printk(KERN_ERR PFX "cqicb->flags = %x.\n", cqicb->flags);
  1234. printk(KERN_ERR PFX "cqicb->len = %d.\n", le16_to_cpu(cqicb->len));
  1235. printk(KERN_ERR PFX "cqicb->addr = 0x%llx.\n",
  1236. (unsigned long long) le64_to_cpu(cqicb->addr));
  1237. printk(KERN_ERR PFX "cqicb->prod_idx_addr = 0x%llx.\n",
  1238. (unsigned long long) le64_to_cpu(cqicb->prod_idx_addr));
  1239. printk(KERN_ERR PFX "cqicb->pkt_delay = 0x%.04x.\n",
  1240. le16_to_cpu(cqicb->pkt_delay));
  1241. printk(KERN_ERR PFX "cqicb->irq_delay = 0x%.04x.\n",
  1242. le16_to_cpu(cqicb->irq_delay));
  1243. printk(KERN_ERR PFX "cqicb->lbq_addr = 0x%llx.\n",
  1244. (unsigned long long) le64_to_cpu(cqicb->lbq_addr));
  1245. printk(KERN_ERR PFX "cqicb->lbq_buf_size = 0x%.04x.\n",
  1246. le16_to_cpu(cqicb->lbq_buf_size));
  1247. printk(KERN_ERR PFX "cqicb->lbq_len = 0x%.04x.\n",
  1248. le16_to_cpu(cqicb->lbq_len));
  1249. printk(KERN_ERR PFX "cqicb->sbq_addr = 0x%llx.\n",
  1250. (unsigned long long) le64_to_cpu(cqicb->sbq_addr));
  1251. printk(KERN_ERR PFX "cqicb->sbq_buf_size = 0x%.04x.\n",
  1252. le16_to_cpu(cqicb->sbq_buf_size));
  1253. printk(KERN_ERR PFX "cqicb->sbq_len = 0x%.04x.\n",
  1254. le16_to_cpu(cqicb->sbq_len));
  1255. }
  1256. void ql_dump_rx_ring(struct rx_ring *rx_ring)
  1257. {
  1258. if (rx_ring == NULL)
  1259. return;
  1260. printk(KERN_ERR PFX
  1261. "===================== Dumping rx_ring %d ===============.\n",
  1262. rx_ring->cq_id);
  1263. printk(KERN_ERR PFX "Dumping rx_ring %d, type = %s%s%s.\n",
  1264. rx_ring->cq_id, rx_ring->type == DEFAULT_Q ? "DEFAULT" : "",
  1265. rx_ring->type == TX_Q ? "OUTBOUND COMPLETIONS" : "",
  1266. rx_ring->type == RX_Q ? "INBOUND_COMPLETIONS" : "");
  1267. printk(KERN_ERR PFX "rx_ring->cqicb = %p.\n", &rx_ring->cqicb);
  1268. printk(KERN_ERR PFX "rx_ring->cq_base = %p.\n", rx_ring->cq_base);
  1269. printk(KERN_ERR PFX "rx_ring->cq_base_dma = %llx.\n",
  1270. (unsigned long long) rx_ring->cq_base_dma);
  1271. printk(KERN_ERR PFX "rx_ring->cq_size = %d.\n", rx_ring->cq_size);
  1272. printk(KERN_ERR PFX "rx_ring->cq_len = %d.\n", rx_ring->cq_len);
  1273. printk(KERN_ERR PFX
  1274. "rx_ring->prod_idx_sh_reg, addr = 0x%p, value = %d.\n",
  1275. rx_ring->prod_idx_sh_reg,
  1276. rx_ring->prod_idx_sh_reg
  1277. ? ql_read_sh_reg(rx_ring->prod_idx_sh_reg) : 0);
  1278. printk(KERN_ERR PFX "rx_ring->prod_idx_sh_reg_dma = %llx.\n",
  1279. (unsigned long long) rx_ring->prod_idx_sh_reg_dma);
  1280. printk(KERN_ERR PFX "rx_ring->cnsmr_idx_db_reg = %p.\n",
  1281. rx_ring->cnsmr_idx_db_reg);
  1282. printk(KERN_ERR PFX "rx_ring->cnsmr_idx = %d.\n", rx_ring->cnsmr_idx);
  1283. printk(KERN_ERR PFX "rx_ring->curr_entry = %p.\n", rx_ring->curr_entry);
  1284. printk(KERN_ERR PFX "rx_ring->valid_db_reg = %p.\n",
  1285. rx_ring->valid_db_reg);
  1286. printk(KERN_ERR PFX "rx_ring->lbq_base = %p.\n", rx_ring->lbq_base);
  1287. printk(KERN_ERR PFX "rx_ring->lbq_base_dma = %llx.\n",
  1288. (unsigned long long) rx_ring->lbq_base_dma);
  1289. printk(KERN_ERR PFX "rx_ring->lbq_base_indirect = %p.\n",
  1290. rx_ring->lbq_base_indirect);
  1291. printk(KERN_ERR PFX "rx_ring->lbq_base_indirect_dma = %llx.\n",
  1292. (unsigned long long) rx_ring->lbq_base_indirect_dma);
  1293. printk(KERN_ERR PFX "rx_ring->lbq = %p.\n", rx_ring->lbq);
  1294. printk(KERN_ERR PFX "rx_ring->lbq_len = %d.\n", rx_ring->lbq_len);
  1295. printk(KERN_ERR PFX "rx_ring->lbq_size = %d.\n", rx_ring->lbq_size);
  1296. printk(KERN_ERR PFX "rx_ring->lbq_prod_idx_db_reg = %p.\n",
  1297. rx_ring->lbq_prod_idx_db_reg);
  1298. printk(KERN_ERR PFX "rx_ring->lbq_prod_idx = %d.\n",
  1299. rx_ring->lbq_prod_idx);
  1300. printk(KERN_ERR PFX "rx_ring->lbq_curr_idx = %d.\n",
  1301. rx_ring->lbq_curr_idx);
  1302. printk(KERN_ERR PFX "rx_ring->lbq_clean_idx = %d.\n",
  1303. rx_ring->lbq_clean_idx);
  1304. printk(KERN_ERR PFX "rx_ring->lbq_free_cnt = %d.\n",
  1305. rx_ring->lbq_free_cnt);
  1306. printk(KERN_ERR PFX "rx_ring->lbq_buf_size = %d.\n",
  1307. rx_ring->lbq_buf_size);
  1308. printk(KERN_ERR PFX "rx_ring->sbq_base = %p.\n", rx_ring->sbq_base);
  1309. printk(KERN_ERR PFX "rx_ring->sbq_base_dma = %llx.\n",
  1310. (unsigned long long) rx_ring->sbq_base_dma);
  1311. printk(KERN_ERR PFX "rx_ring->sbq_base_indirect = %p.\n",
  1312. rx_ring->sbq_base_indirect);
  1313. printk(KERN_ERR PFX "rx_ring->sbq_base_indirect_dma = %llx.\n",
  1314. (unsigned long long) rx_ring->sbq_base_indirect_dma);
  1315. printk(KERN_ERR PFX "rx_ring->sbq = %p.\n", rx_ring->sbq);
  1316. printk(KERN_ERR PFX "rx_ring->sbq_len = %d.\n", rx_ring->sbq_len);
  1317. printk(KERN_ERR PFX "rx_ring->sbq_size = %d.\n", rx_ring->sbq_size);
  1318. printk(KERN_ERR PFX "rx_ring->sbq_prod_idx_db_reg addr = %p.\n",
  1319. rx_ring->sbq_prod_idx_db_reg);
  1320. printk(KERN_ERR PFX "rx_ring->sbq_prod_idx = %d.\n",
  1321. rx_ring->sbq_prod_idx);
  1322. printk(KERN_ERR PFX "rx_ring->sbq_curr_idx = %d.\n",
  1323. rx_ring->sbq_curr_idx);
  1324. printk(KERN_ERR PFX "rx_ring->sbq_clean_idx = %d.\n",
  1325. rx_ring->sbq_clean_idx);
  1326. printk(KERN_ERR PFX "rx_ring->sbq_free_cnt = %d.\n",
  1327. rx_ring->sbq_free_cnt);
  1328. printk(KERN_ERR PFX "rx_ring->sbq_buf_size = %d.\n",
  1329. rx_ring->sbq_buf_size);
  1330. printk(KERN_ERR PFX "rx_ring->cq_id = %d.\n", rx_ring->cq_id);
  1331. printk(KERN_ERR PFX "rx_ring->irq = %d.\n", rx_ring->irq);
  1332. printk(KERN_ERR PFX "rx_ring->cpu = %d.\n", rx_ring->cpu);
  1333. printk(KERN_ERR PFX "rx_ring->qdev = %p.\n", rx_ring->qdev);
  1334. }
  1335. void ql_dump_hw_cb(struct ql_adapter *qdev, int size, u32 bit, u16 q_id)
  1336. {
  1337. void *ptr;
  1338. printk(KERN_ERR PFX "%s: Enter.\n", __func__);
  1339. ptr = kmalloc(size, GFP_ATOMIC);
  1340. if (ptr == NULL) {
  1341. printk(KERN_ERR PFX "%s: Couldn't allocate a buffer.\n",
  1342. __func__);
  1343. return;
  1344. }
  1345. if (ql_write_cfg(qdev, ptr, size, bit, q_id)) {
  1346. printk(KERN_ERR "%s: Failed to upload control block!\n",
  1347. __func__);
  1348. goto fail_it;
  1349. }
  1350. switch (bit) {
  1351. case CFG_DRQ:
  1352. ql_dump_wqicb((struct wqicb *)ptr);
  1353. break;
  1354. case CFG_DCQ:
  1355. ql_dump_cqicb((struct cqicb *)ptr);
  1356. break;
  1357. case CFG_DR:
  1358. ql_dump_ricb((struct ricb *)ptr);
  1359. break;
  1360. default:
  1361. printk(KERN_ERR PFX "%s: Invalid bit value = %x.\n",
  1362. __func__, bit);
  1363. break;
  1364. }
  1365. fail_it:
  1366. kfree(ptr);
  1367. }
  1368. #endif
  1369. #ifdef QL_OB_DUMP
  1370. void ql_dump_tx_desc(struct tx_buf_desc *tbd)
  1371. {
  1372. printk(KERN_ERR PFX "tbd->addr = 0x%llx\n",
  1373. le64_to_cpu((u64) tbd->addr));
  1374. printk(KERN_ERR PFX "tbd->len = %d\n",
  1375. le32_to_cpu(tbd->len & TX_DESC_LEN_MASK));
  1376. printk(KERN_ERR PFX "tbd->flags = %s %s\n",
  1377. tbd->len & TX_DESC_C ? "C" : ".",
  1378. tbd->len & TX_DESC_E ? "E" : ".");
  1379. tbd++;
  1380. printk(KERN_ERR PFX "tbd->addr = 0x%llx\n",
  1381. le64_to_cpu((u64) tbd->addr));
  1382. printk(KERN_ERR PFX "tbd->len = %d\n",
  1383. le32_to_cpu(tbd->len & TX_DESC_LEN_MASK));
  1384. printk(KERN_ERR PFX "tbd->flags = %s %s\n",
  1385. tbd->len & TX_DESC_C ? "C" : ".",
  1386. tbd->len & TX_DESC_E ? "E" : ".");
  1387. tbd++;
  1388. printk(KERN_ERR PFX "tbd->addr = 0x%llx\n",
  1389. le64_to_cpu((u64) tbd->addr));
  1390. printk(KERN_ERR PFX "tbd->len = %d\n",
  1391. le32_to_cpu(tbd->len & TX_DESC_LEN_MASK));
  1392. printk(KERN_ERR PFX "tbd->flags = %s %s\n",
  1393. tbd->len & TX_DESC_C ? "C" : ".",
  1394. tbd->len & TX_DESC_E ? "E" : ".");
  1395. }
  1396. void ql_dump_ob_mac_iocb(struct ob_mac_iocb_req *ob_mac_iocb)
  1397. {
  1398. struct ob_mac_tso_iocb_req *ob_mac_tso_iocb =
  1399. (struct ob_mac_tso_iocb_req *)ob_mac_iocb;
  1400. struct tx_buf_desc *tbd;
  1401. u16 frame_len;
  1402. printk(KERN_ERR PFX "%s\n", __func__);
  1403. printk(KERN_ERR PFX "opcode = %s\n",
  1404. (ob_mac_iocb->opcode == OPCODE_OB_MAC_IOCB) ? "MAC" : "TSO");
  1405. printk(KERN_ERR PFX "flags1 = %s %s %s %s %s\n",
  1406. ob_mac_tso_iocb->flags1 & OB_MAC_TSO_IOCB_OI ? "OI" : "",
  1407. ob_mac_tso_iocb->flags1 & OB_MAC_TSO_IOCB_I ? "I" : "",
  1408. ob_mac_tso_iocb->flags1 & OB_MAC_TSO_IOCB_D ? "D" : "",
  1409. ob_mac_tso_iocb->flags1 & OB_MAC_TSO_IOCB_IP4 ? "IP4" : "",
  1410. ob_mac_tso_iocb->flags1 & OB_MAC_TSO_IOCB_IP6 ? "IP6" : "");
  1411. printk(KERN_ERR PFX "flags2 = %s %s %s\n",
  1412. ob_mac_tso_iocb->flags2 & OB_MAC_TSO_IOCB_LSO ? "LSO" : "",
  1413. ob_mac_tso_iocb->flags2 & OB_MAC_TSO_IOCB_UC ? "UC" : "",
  1414. ob_mac_tso_iocb->flags2 & OB_MAC_TSO_IOCB_TC ? "TC" : "");
  1415. printk(KERN_ERR PFX "flags3 = %s %s %s \n",
  1416. ob_mac_tso_iocb->flags3 & OB_MAC_TSO_IOCB_IC ? "IC" : "",
  1417. ob_mac_tso_iocb->flags3 & OB_MAC_TSO_IOCB_DFP ? "DFP" : "",
  1418. ob_mac_tso_iocb->flags3 & OB_MAC_TSO_IOCB_V ? "V" : "");
  1419. printk(KERN_ERR PFX "tid = %x\n", ob_mac_iocb->tid);
  1420. printk(KERN_ERR PFX "txq_idx = %d\n", ob_mac_iocb->txq_idx);
  1421. printk(KERN_ERR PFX "vlan_tci = %x\n", ob_mac_tso_iocb->vlan_tci);
  1422. if (ob_mac_iocb->opcode == OPCODE_OB_MAC_TSO_IOCB) {
  1423. printk(KERN_ERR PFX "frame_len = %d\n",
  1424. le32_to_cpu(ob_mac_tso_iocb->frame_len));
  1425. printk(KERN_ERR PFX "mss = %d\n",
  1426. le16_to_cpu(ob_mac_tso_iocb->mss));
  1427. printk(KERN_ERR PFX "prot_hdr_len = %d\n",
  1428. le16_to_cpu(ob_mac_tso_iocb->total_hdrs_len));
  1429. printk(KERN_ERR PFX "hdr_offset = 0x%.04x\n",
  1430. le16_to_cpu(ob_mac_tso_iocb->net_trans_offset));
  1431. frame_len = le32_to_cpu(ob_mac_tso_iocb->frame_len);
  1432. } else {
  1433. printk(KERN_ERR PFX "frame_len = %d\n",
  1434. le16_to_cpu(ob_mac_iocb->frame_len));
  1435. frame_len = le16_to_cpu(ob_mac_iocb->frame_len);
  1436. }
  1437. tbd = &ob_mac_iocb->tbd[0];
  1438. ql_dump_tx_desc(tbd);
  1439. }
  1440. void ql_dump_ob_mac_rsp(struct ob_mac_iocb_rsp *ob_mac_rsp)
  1441. {
  1442. printk(KERN_ERR PFX "%s\n", __func__);
  1443. printk(KERN_ERR PFX "opcode = %d\n", ob_mac_rsp->opcode);
  1444. printk(KERN_ERR PFX "flags = %s %s %s %s %s %s %s\n",
  1445. ob_mac_rsp->flags1 & OB_MAC_IOCB_RSP_OI ? "OI" : ".",
  1446. ob_mac_rsp->flags1 & OB_MAC_IOCB_RSP_I ? "I" : ".",
  1447. ob_mac_rsp->flags1 & OB_MAC_IOCB_RSP_E ? "E" : ".",
  1448. ob_mac_rsp->flags1 & OB_MAC_IOCB_RSP_S ? "S" : ".",
  1449. ob_mac_rsp->flags1 & OB_MAC_IOCB_RSP_L ? "L" : ".",
  1450. ob_mac_rsp->flags1 & OB_MAC_IOCB_RSP_P ? "P" : ".",
  1451. ob_mac_rsp->flags2 & OB_MAC_IOCB_RSP_B ? "B" : ".");
  1452. printk(KERN_ERR PFX "tid = %x\n", ob_mac_rsp->tid);
  1453. }
  1454. #endif
  1455. #ifdef QL_IB_DUMP
  1456. void ql_dump_ib_mac_rsp(struct ib_mac_iocb_rsp *ib_mac_rsp)
  1457. {
  1458. printk(KERN_ERR PFX "%s\n", __func__);
  1459. printk(KERN_ERR PFX "opcode = 0x%x\n", ib_mac_rsp->opcode);
  1460. printk(KERN_ERR PFX "flags1 = %s%s%s%s%s%s\n",
  1461. ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_OI ? "OI " : "",
  1462. ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_I ? "I " : "",
  1463. ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_TE ? "TE " : "",
  1464. ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_NU ? "NU " : "",
  1465. ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_IE ? "IE " : "",
  1466. ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_B ? "B " : "");
  1467. if (ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_M_MASK)
  1468. printk(KERN_ERR PFX "%s%s%s Multicast.\n",
  1469. (ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_M_MASK) ==
  1470. IB_MAC_IOCB_RSP_M_HASH ? "Hash" : "",
  1471. (ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_M_MASK) ==
  1472. IB_MAC_IOCB_RSP_M_REG ? "Registered" : "",
  1473. (ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_M_MASK) ==
  1474. IB_MAC_IOCB_RSP_M_PROM ? "Promiscuous" : "");
  1475. printk(KERN_ERR PFX "flags2 = %s%s%s%s%s\n",
  1476. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_P) ? "P " : "",
  1477. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_V) ? "V " : "",
  1478. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_U) ? "U " : "",
  1479. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_T) ? "T " : "",
  1480. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_FO) ? "FO " : "");
  1481. if (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK)
  1482. printk(KERN_ERR PFX "%s%s%s%s%s error.\n",
  1483. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK) ==
  1484. IB_MAC_IOCB_RSP_ERR_OVERSIZE ? "oversize" : "",
  1485. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK) ==
  1486. IB_MAC_IOCB_RSP_ERR_UNDERSIZE ? "undersize" : "",
  1487. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK) ==
  1488. IB_MAC_IOCB_RSP_ERR_PREAMBLE ? "preamble" : "",
  1489. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK) ==
  1490. IB_MAC_IOCB_RSP_ERR_FRAME_LEN ? "frame length" : "",
  1491. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK) ==
  1492. IB_MAC_IOCB_RSP_ERR_CRC ? "CRC" : "");
  1493. printk(KERN_ERR PFX "flags3 = %s%s.\n",
  1494. ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_DS ? "DS " : "",
  1495. ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_DL ? "DL " : "");
  1496. if (ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_RSS_MASK)
  1497. printk(KERN_ERR PFX "RSS flags = %s%s%s%s.\n",
  1498. ((ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_RSS_MASK) ==
  1499. IB_MAC_IOCB_RSP_M_IPV4) ? "IPv4 RSS" : "",
  1500. ((ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_RSS_MASK) ==
  1501. IB_MAC_IOCB_RSP_M_IPV6) ? "IPv6 RSS " : "",
  1502. ((ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_RSS_MASK) ==
  1503. IB_MAC_IOCB_RSP_M_TCP_V4) ? "TCP/IPv4 RSS" : "",
  1504. ((ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_RSS_MASK) ==
  1505. IB_MAC_IOCB_RSP_M_TCP_V6) ? "TCP/IPv6 RSS" : "");
  1506. printk(KERN_ERR PFX "data_len = %d\n",
  1507. le32_to_cpu(ib_mac_rsp->data_len));
  1508. printk(KERN_ERR PFX "data_addr = 0x%llx\n",
  1509. (unsigned long long) le64_to_cpu(ib_mac_rsp->data_addr));
  1510. if (ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_RSS_MASK)
  1511. printk(KERN_ERR PFX "rss = %x\n",
  1512. le32_to_cpu(ib_mac_rsp->rss));
  1513. if (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_V)
  1514. printk(KERN_ERR PFX "vlan_id = %x\n",
  1515. le16_to_cpu(ib_mac_rsp->vlan_id));
  1516. printk(KERN_ERR PFX "flags4 = %s%s%s.\n",
  1517. ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HV ? "HV " : "",
  1518. ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HS ? "HS " : "",
  1519. ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HL ? "HL " : "");
  1520. if (ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HV) {
  1521. printk(KERN_ERR PFX "hdr length = %d.\n",
  1522. le32_to_cpu(ib_mac_rsp->hdr_len));
  1523. printk(KERN_ERR PFX "hdr addr = 0x%llx.\n",
  1524. (unsigned long long) le64_to_cpu(ib_mac_rsp->hdr_addr));
  1525. }
  1526. }
  1527. #endif
  1528. #ifdef QL_ALL_DUMP
  1529. void ql_dump_all(struct ql_adapter *qdev)
  1530. {
  1531. int i;
  1532. QL_DUMP_REGS(qdev);
  1533. QL_DUMP_QDEV(qdev);
  1534. for (i = 0; i < qdev->tx_ring_count; i++) {
  1535. QL_DUMP_TX_RING(&qdev->tx_ring[i]);
  1536. QL_DUMP_WQICB((struct wqicb *)&qdev->tx_ring[i]);
  1537. }
  1538. for (i = 0; i < qdev->rx_ring_count; i++) {
  1539. QL_DUMP_RX_RING(&qdev->rx_ring[i]);
  1540. QL_DUMP_CQICB((struct cqicb *)&qdev->rx_ring[i]);
  1541. }
  1542. }
  1543. #endif