ql4_83xx.c 44 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638
  1. /*
  2. * QLogic iSCSI HBA Driver
  3. * Copyright (c) 2003-2012 QLogic Corporation
  4. *
  5. * See LICENSE.qla4xxx for copyright and licensing details.
  6. */
  7. #include <linux/ratelimit.h>
  8. #include "ql4_def.h"
  9. #include "ql4_version.h"
  10. #include "ql4_glbl.h"
  11. #include "ql4_dbg.h"
  12. #include "ql4_inline.h"
  13. uint32_t qla4_83xx_rd_reg(struct scsi_qla_host *ha, ulong addr)
  14. {
  15. return readl((void __iomem *)(ha->nx_pcibase + addr));
  16. }
  17. void qla4_83xx_wr_reg(struct scsi_qla_host *ha, ulong addr, uint32_t val)
  18. {
  19. writel(val, (void __iomem *)(ha->nx_pcibase + addr));
  20. }
  21. static int qla4_83xx_set_win_base(struct scsi_qla_host *ha, uint32_t addr)
  22. {
  23. uint32_t val;
  24. int ret_val = QLA_SUCCESS;
  25. qla4_83xx_wr_reg(ha, QLA83XX_CRB_WIN_FUNC(ha->func_num), addr);
  26. val = qla4_83xx_rd_reg(ha, QLA83XX_CRB_WIN_FUNC(ha->func_num));
  27. if (val != addr) {
  28. ql4_printk(KERN_ERR, ha, "%s: Failed to set register window : addr written 0x%x, read 0x%x!\n",
  29. __func__, addr, val);
  30. ret_val = QLA_ERROR;
  31. }
  32. return ret_val;
  33. }
  34. int qla4_83xx_rd_reg_indirect(struct scsi_qla_host *ha, uint32_t addr,
  35. uint32_t *data)
  36. {
  37. int ret_val;
  38. ret_val = qla4_83xx_set_win_base(ha, addr);
  39. if (ret_val == QLA_SUCCESS)
  40. *data = qla4_83xx_rd_reg(ha, QLA83XX_WILDCARD);
  41. else
  42. ql4_printk(KERN_ERR, ha, "%s: failed read of addr 0x%x!\n",
  43. __func__, addr);
  44. return ret_val;
  45. }
  46. int qla4_83xx_wr_reg_indirect(struct scsi_qla_host *ha, uint32_t addr,
  47. uint32_t data)
  48. {
  49. int ret_val;
  50. ret_val = qla4_83xx_set_win_base(ha, addr);
  51. if (ret_val == QLA_SUCCESS)
  52. qla4_83xx_wr_reg(ha, QLA83XX_WILDCARD, data);
  53. else
  54. ql4_printk(KERN_ERR, ha, "%s: failed wrt to addr 0x%x, data 0x%x\n",
  55. __func__, addr, data);
  56. return ret_val;
  57. }
  58. static int qla4_83xx_flash_lock(struct scsi_qla_host *ha)
  59. {
  60. int lock_owner;
  61. int timeout = 0;
  62. uint32_t lock_status = 0;
  63. int ret_val = QLA_SUCCESS;
  64. while (lock_status == 0) {
  65. lock_status = qla4_83xx_rd_reg(ha, QLA83XX_FLASH_LOCK);
  66. if (lock_status)
  67. break;
  68. if (++timeout >= QLA83XX_FLASH_LOCK_TIMEOUT / 20) {
  69. lock_owner = qla4_83xx_rd_reg(ha,
  70. QLA83XX_FLASH_LOCK_ID);
  71. ql4_printk(KERN_ERR, ha, "%s: flash lock by func %d failed, held by func %d\n",
  72. __func__, ha->func_num, lock_owner);
  73. ret_val = QLA_ERROR;
  74. break;
  75. }
  76. msleep(20);
  77. }
  78. qla4_83xx_wr_reg(ha, QLA83XX_FLASH_LOCK_ID, ha->func_num);
  79. return ret_val;
  80. }
  81. static void qla4_83xx_flash_unlock(struct scsi_qla_host *ha)
  82. {
  83. /* Reading FLASH_UNLOCK register unlocks the Flash */
  84. qla4_83xx_wr_reg(ha, QLA83XX_FLASH_LOCK_ID, 0xFF);
  85. qla4_83xx_rd_reg(ha, QLA83XX_FLASH_UNLOCK);
  86. }
  87. int qla4_83xx_flash_read_u32(struct scsi_qla_host *ha, uint32_t flash_addr,
  88. uint8_t *p_data, int u32_word_count)
  89. {
  90. int i;
  91. uint32_t u32_word;
  92. uint32_t addr = flash_addr;
  93. int ret_val = QLA_SUCCESS;
  94. ret_val = qla4_83xx_flash_lock(ha);
  95. if (ret_val == QLA_ERROR)
  96. goto exit_lock_error;
  97. if (addr & 0x03) {
  98. ql4_printk(KERN_ERR, ha, "%s: Illegal addr = 0x%x\n",
  99. __func__, addr);
  100. ret_val = QLA_ERROR;
  101. goto exit_flash_read;
  102. }
  103. for (i = 0; i < u32_word_count; i++) {
  104. ret_val = qla4_83xx_wr_reg_indirect(ha,
  105. QLA83XX_FLASH_DIRECT_WINDOW,
  106. (addr & 0xFFFF0000));
  107. if (ret_val == QLA_ERROR) {
  108. ql4_printk(KERN_ERR, ha, "%s: failed to write addr 0x%x to FLASH_DIRECT_WINDOW\n!",
  109. __func__, addr);
  110. goto exit_flash_read;
  111. }
  112. ret_val = qla4_83xx_rd_reg_indirect(ha,
  113. QLA83XX_FLASH_DIRECT_DATA(addr),
  114. &u32_word);
  115. if (ret_val == QLA_ERROR) {
  116. ql4_printk(KERN_ERR, ha, "%s: failed to read addr 0x%x!\n",
  117. __func__, addr);
  118. goto exit_flash_read;
  119. }
  120. *(__le32 *)p_data = le32_to_cpu(u32_word);
  121. p_data = p_data + 4;
  122. addr = addr + 4;
  123. }
  124. exit_flash_read:
  125. qla4_83xx_flash_unlock(ha);
  126. exit_lock_error:
  127. return ret_val;
  128. }
  129. int qla4_83xx_lockless_flash_read_u32(struct scsi_qla_host *ha,
  130. uint32_t flash_addr, uint8_t *p_data,
  131. int u32_word_count)
  132. {
  133. uint32_t i;
  134. uint32_t u32_word;
  135. uint32_t flash_offset;
  136. uint32_t addr = flash_addr;
  137. int ret_val = QLA_SUCCESS;
  138. flash_offset = addr & (QLA83XX_FLASH_SECTOR_SIZE - 1);
  139. if (addr & 0x3) {
  140. ql4_printk(KERN_ERR, ha, "%s: Illegal addr = 0x%x\n",
  141. __func__, addr);
  142. ret_val = QLA_ERROR;
  143. goto exit_lockless_read;
  144. }
  145. ret_val = qla4_83xx_wr_reg_indirect(ha, QLA83XX_FLASH_DIRECT_WINDOW,
  146. addr);
  147. if (ret_val == QLA_ERROR) {
  148. ql4_printk(KERN_ERR, ha, "%s: failed to write addr 0x%x to FLASH_DIRECT_WINDOW!\n",
  149. __func__, addr);
  150. goto exit_lockless_read;
  151. }
  152. /* Check if data is spread across multiple sectors */
  153. if ((flash_offset + (u32_word_count * sizeof(uint32_t))) >
  154. (QLA83XX_FLASH_SECTOR_SIZE - 1)) {
  155. /* Multi sector read */
  156. for (i = 0; i < u32_word_count; i++) {
  157. ret_val = qla4_83xx_rd_reg_indirect(ha,
  158. QLA83XX_FLASH_DIRECT_DATA(addr),
  159. &u32_word);
  160. if (ret_val == QLA_ERROR) {
  161. ql4_printk(KERN_ERR, ha, "%s: failed to read addr 0x%x!\n",
  162. __func__, addr);
  163. goto exit_lockless_read;
  164. }
  165. *(__le32 *)p_data = le32_to_cpu(u32_word);
  166. p_data = p_data + 4;
  167. addr = addr + 4;
  168. flash_offset = flash_offset + 4;
  169. if (flash_offset > (QLA83XX_FLASH_SECTOR_SIZE - 1)) {
  170. /* This write is needed once for each sector */
  171. ret_val = qla4_83xx_wr_reg_indirect(ha,
  172. QLA83XX_FLASH_DIRECT_WINDOW,
  173. addr);
  174. if (ret_val == QLA_ERROR) {
  175. ql4_printk(KERN_ERR, ha, "%s: failed to write addr 0x%x to FLASH_DIRECT_WINDOW!\n",
  176. __func__, addr);
  177. goto exit_lockless_read;
  178. }
  179. flash_offset = 0;
  180. }
  181. }
  182. } else {
  183. /* Single sector read */
  184. for (i = 0; i < u32_word_count; i++) {
  185. ret_val = qla4_83xx_rd_reg_indirect(ha,
  186. QLA83XX_FLASH_DIRECT_DATA(addr),
  187. &u32_word);
  188. if (ret_val == QLA_ERROR) {
  189. ql4_printk(KERN_ERR, ha, "%s: failed to read addr 0x%x!\n",
  190. __func__, addr);
  191. goto exit_lockless_read;
  192. }
  193. *(__le32 *)p_data = le32_to_cpu(u32_word);
  194. p_data = p_data + 4;
  195. addr = addr + 4;
  196. }
  197. }
  198. exit_lockless_read:
  199. return ret_val;
  200. }
  201. void qla4_83xx_rom_lock_recovery(struct scsi_qla_host *ha)
  202. {
  203. if (qla4_83xx_flash_lock(ha))
  204. ql4_printk(KERN_INFO, ha, "%s: Resetting rom lock\n", __func__);
  205. /*
  206. * We got the lock, or someone else is holding the lock
  207. * since we are restting, forcefully unlock
  208. */
  209. qla4_83xx_flash_unlock(ha);
  210. }
  211. /**
  212. * qla4_83xx_ms_mem_write_128b - Writes data to MS/off-chip memory
  213. * @ha: Pointer to adapter structure
  214. * @addr: Flash address to write to
  215. * @data: Data to be written
  216. * @count: word_count to be written
  217. *
  218. * Return: On success return QLA_SUCCESS
  219. * On error return QLA_ERROR
  220. **/
  221. static int qla4_83xx_ms_mem_write_128b(struct scsi_qla_host *ha, uint64_t addr,
  222. uint32_t *data, uint32_t count)
  223. {
  224. int i, j;
  225. uint32_t agt_ctrl;
  226. unsigned long flags;
  227. int ret_val = QLA_SUCCESS;
  228. /* Only 128-bit aligned access */
  229. if (addr & 0xF) {
  230. ret_val = QLA_ERROR;
  231. goto exit_ms_mem_write;
  232. }
  233. write_lock_irqsave(&ha->hw_lock, flags);
  234. /* Write address */
  235. ret_val = qla4_83xx_wr_reg_indirect(ha, MD_MIU_TEST_AGT_ADDR_HI, 0);
  236. if (ret_val == QLA_ERROR) {
  237. ql4_printk(KERN_ERR, ha, "%s: write to AGT_ADDR_HI failed\n",
  238. __func__);
  239. goto exit_ms_mem_write_unlock;
  240. }
  241. for (i = 0; i < count; i++, addr += 16) {
  242. if (!((QLA8XXX_ADDR_IN_RANGE(addr, QLA8XXX_ADDR_QDR_NET,
  243. QLA8XXX_ADDR_QDR_NET_MAX)) ||
  244. (QLA8XXX_ADDR_IN_RANGE(addr, QLA8XXX_ADDR_DDR_NET,
  245. QLA8XXX_ADDR_DDR_NET_MAX)))) {
  246. ret_val = QLA_ERROR;
  247. goto exit_ms_mem_write_unlock;
  248. }
  249. ret_val = qla4_83xx_wr_reg_indirect(ha, MD_MIU_TEST_AGT_ADDR_LO,
  250. addr);
  251. /* Write data */
  252. ret_val |= qla4_83xx_wr_reg_indirect(ha,
  253. MD_MIU_TEST_AGT_WRDATA_LO,
  254. *data++);
  255. ret_val |= qla4_83xx_wr_reg_indirect(ha,
  256. MD_MIU_TEST_AGT_WRDATA_HI,
  257. *data++);
  258. ret_val |= qla4_83xx_wr_reg_indirect(ha,
  259. MD_MIU_TEST_AGT_WRDATA_ULO,
  260. *data++);
  261. ret_val |= qla4_83xx_wr_reg_indirect(ha,
  262. MD_MIU_TEST_AGT_WRDATA_UHI,
  263. *data++);
  264. if (ret_val == QLA_ERROR) {
  265. ql4_printk(KERN_ERR, ha, "%s: write to AGT_WRDATA failed\n",
  266. __func__);
  267. goto exit_ms_mem_write_unlock;
  268. }
  269. /* Check write status */
  270. ret_val = qla4_83xx_wr_reg_indirect(ha, MD_MIU_TEST_AGT_CTRL,
  271. MIU_TA_CTL_WRITE_ENABLE);
  272. ret_val |= qla4_83xx_wr_reg_indirect(ha, MD_MIU_TEST_AGT_CTRL,
  273. MIU_TA_CTL_WRITE_START);
  274. if (ret_val == QLA_ERROR) {
  275. ql4_printk(KERN_ERR, ha, "%s: write to AGT_CTRL failed\n",
  276. __func__);
  277. goto exit_ms_mem_write_unlock;
  278. }
  279. for (j = 0; j < MAX_CTL_CHECK; j++) {
  280. ret_val = qla4_83xx_rd_reg_indirect(ha,
  281. MD_MIU_TEST_AGT_CTRL,
  282. &agt_ctrl);
  283. if (ret_val == QLA_ERROR) {
  284. ql4_printk(KERN_ERR, ha, "%s: failed to read MD_MIU_TEST_AGT_CTRL\n",
  285. __func__);
  286. goto exit_ms_mem_write_unlock;
  287. }
  288. if ((agt_ctrl & MIU_TA_CTL_BUSY) == 0)
  289. break;
  290. }
  291. /* Status check failed */
  292. if (j >= MAX_CTL_CHECK) {
  293. printk_ratelimited(KERN_ERR "%s: MS memory write failed!\n",
  294. __func__);
  295. ret_val = QLA_ERROR;
  296. goto exit_ms_mem_write_unlock;
  297. }
  298. }
  299. exit_ms_mem_write_unlock:
  300. write_unlock_irqrestore(&ha->hw_lock, flags);
  301. exit_ms_mem_write:
  302. return ret_val;
  303. }
  304. #define INTENT_TO_RECOVER 0x01
  305. #define PROCEED_TO_RECOVER 0x02
  306. static int qla4_83xx_lock_recovery(struct scsi_qla_host *ha)
  307. {
  308. uint32_t lock = 0, lockid;
  309. int ret_val = QLA_ERROR;
  310. lockid = ha->isp_ops->rd_reg_direct(ha, QLA83XX_DRV_LOCKRECOVERY);
  311. /* Check for other Recovery in progress, go wait */
  312. if ((lockid & 0x3) != 0)
  313. goto exit_lock_recovery;
  314. /* Intent to Recover */
  315. ha->isp_ops->wr_reg_direct(ha, QLA83XX_DRV_LOCKRECOVERY,
  316. (ha->func_num << 2) | INTENT_TO_RECOVER);
  317. msleep(200);
  318. /* Check Intent to Recover is advertised */
  319. lockid = ha->isp_ops->rd_reg_direct(ha, QLA83XX_DRV_LOCKRECOVERY);
  320. if ((lockid & 0x3C) != (ha->func_num << 2))
  321. goto exit_lock_recovery;
  322. ql4_printk(KERN_INFO, ha, "%s: IDC Lock recovery initiated for func %d\n",
  323. __func__, ha->func_num);
  324. /* Proceed to Recover */
  325. ha->isp_ops->wr_reg_direct(ha, QLA83XX_DRV_LOCKRECOVERY,
  326. (ha->func_num << 2) | PROCEED_TO_RECOVER);
  327. /* Force Unlock */
  328. ha->isp_ops->wr_reg_direct(ha, QLA83XX_DRV_LOCK_ID, 0xFF);
  329. ha->isp_ops->rd_reg_direct(ha, QLA83XX_DRV_UNLOCK);
  330. /* Clear bits 0-5 in IDC_RECOVERY register*/
  331. ha->isp_ops->wr_reg_direct(ha, QLA83XX_DRV_LOCKRECOVERY, 0);
  332. /* Get lock */
  333. lock = ha->isp_ops->rd_reg_direct(ha, QLA83XX_DRV_LOCK);
  334. if (lock) {
  335. lockid = ha->isp_ops->rd_reg_direct(ha, QLA83XX_DRV_LOCK_ID);
  336. lockid = ((lockid + (1 << 8)) & ~0xFF) | ha->func_num;
  337. ha->isp_ops->wr_reg_direct(ha, QLA83XX_DRV_LOCK_ID, lockid);
  338. ret_val = QLA_SUCCESS;
  339. }
  340. exit_lock_recovery:
  341. return ret_val;
  342. }
  343. #define QLA83XX_DRV_LOCK_MSLEEP 200
  344. int qla4_83xx_drv_lock(struct scsi_qla_host *ha)
  345. {
  346. int timeout = 0;
  347. uint32_t status = 0;
  348. int ret_val = QLA_SUCCESS;
  349. uint32_t first_owner = 0;
  350. uint32_t tmo_owner = 0;
  351. uint32_t lock_id;
  352. uint32_t func_num;
  353. uint32_t lock_cnt;
  354. while (status == 0) {
  355. status = qla4_83xx_rd_reg(ha, QLA83XX_DRV_LOCK);
  356. if (status) {
  357. /* Increment Counter (8-31) and update func_num (0-7) on
  358. * getting a successful lock */
  359. lock_id = qla4_83xx_rd_reg(ha, QLA83XX_DRV_LOCK_ID);
  360. lock_id = ((lock_id + (1 << 8)) & ~0xFF) | ha->func_num;
  361. qla4_83xx_wr_reg(ha, QLA83XX_DRV_LOCK_ID, lock_id);
  362. break;
  363. }
  364. if (timeout == 0)
  365. /* Save counter + ID of function holding the lock for
  366. * first failure */
  367. first_owner = ha->isp_ops->rd_reg_direct(ha,
  368. QLA83XX_DRV_LOCK_ID);
  369. if (++timeout >=
  370. (QLA83XX_DRV_LOCK_TIMEOUT / QLA83XX_DRV_LOCK_MSLEEP)) {
  371. tmo_owner = qla4_83xx_rd_reg(ha, QLA83XX_DRV_LOCK_ID);
  372. func_num = tmo_owner & 0xFF;
  373. lock_cnt = tmo_owner >> 8;
  374. ql4_printk(KERN_INFO, ha, "%s: Lock by func %d failed after 2s, lock held by func %d, lock count %d, first_owner %d\n",
  375. __func__, ha->func_num, func_num, lock_cnt,
  376. (first_owner & 0xFF));
  377. if (first_owner != tmo_owner) {
  378. /* Some other driver got lock, OR same driver
  379. * got lock again (counter value changed), when
  380. * we were waiting for lock.
  381. * Retry for another 2 sec */
  382. ql4_printk(KERN_INFO, ha, "%s: IDC lock failed for func %d\n",
  383. __func__, ha->func_num);
  384. timeout = 0;
  385. } else {
  386. /* Same driver holding lock > 2sec.
  387. * Force Recovery */
  388. ret_val = qla4_83xx_lock_recovery(ha);
  389. if (ret_val == QLA_SUCCESS) {
  390. /* Recovered and got lock */
  391. ql4_printk(KERN_INFO, ha, "%s: IDC lock Recovery by %d successful\n",
  392. __func__, ha->func_num);
  393. break;
  394. }
  395. /* Recovery Failed, some other function
  396. * has the lock, wait for 2secs and retry */
  397. ql4_printk(KERN_INFO, ha, "%s: IDC lock Recovery by %d failed, Retrying timout\n",
  398. __func__, ha->func_num);
  399. timeout = 0;
  400. }
  401. }
  402. msleep(QLA83XX_DRV_LOCK_MSLEEP);
  403. }
  404. return ret_val;
  405. }
  406. void qla4_83xx_drv_unlock(struct scsi_qla_host *ha)
  407. {
  408. int id;
  409. id = qla4_83xx_rd_reg(ha, QLA83XX_DRV_LOCK_ID);
  410. if ((id & 0xFF) != ha->func_num) {
  411. ql4_printk(KERN_ERR, ha, "%s: IDC Unlock by %d failed, lock owner is %d\n",
  412. __func__, ha->func_num, (id & 0xFF));
  413. return;
  414. }
  415. /* Keep lock counter value, update the ha->func_num to 0xFF */
  416. qla4_83xx_wr_reg(ha, QLA83XX_DRV_LOCK_ID, (id | 0xFF));
  417. qla4_83xx_rd_reg(ha, QLA83XX_DRV_UNLOCK);
  418. }
  419. void qla4_83xx_set_idc_dontreset(struct scsi_qla_host *ha)
  420. {
  421. uint32_t idc_ctrl;
  422. idc_ctrl = qla4_83xx_rd_reg(ha, QLA83XX_IDC_DRV_CTRL);
  423. idc_ctrl |= DONTRESET_BIT0;
  424. qla4_83xx_wr_reg(ha, QLA83XX_IDC_DRV_CTRL, idc_ctrl);
  425. DEBUG2(ql4_printk(KERN_INFO, ha, "%s: idc_ctrl = %d\n", __func__,
  426. idc_ctrl));
  427. }
  428. void qla4_83xx_clear_idc_dontreset(struct scsi_qla_host *ha)
  429. {
  430. uint32_t idc_ctrl;
  431. idc_ctrl = qla4_83xx_rd_reg(ha, QLA83XX_IDC_DRV_CTRL);
  432. idc_ctrl &= ~DONTRESET_BIT0;
  433. qla4_83xx_wr_reg(ha, QLA83XX_IDC_DRV_CTRL, idc_ctrl);
  434. DEBUG2(ql4_printk(KERN_INFO, ha, "%s: idc_ctrl = %d\n", __func__,
  435. idc_ctrl));
  436. }
  437. int qla4_83xx_idc_dontreset(struct scsi_qla_host *ha)
  438. {
  439. uint32_t idc_ctrl;
  440. idc_ctrl = qla4_83xx_rd_reg(ha, QLA83XX_IDC_DRV_CTRL);
  441. return idc_ctrl & DONTRESET_BIT0;
  442. }
  443. /*-------------------------IDC State Machine ---------------------*/
  444. enum {
  445. UNKNOWN_CLASS = 0,
  446. NIC_CLASS,
  447. FCOE_CLASS,
  448. ISCSI_CLASS
  449. };
  450. struct device_info {
  451. int func_num;
  452. int device_type;
  453. int port_num;
  454. };
  455. int qla4_83xx_can_perform_reset(struct scsi_qla_host *ha)
  456. {
  457. uint32_t drv_active;
  458. uint32_t dev_part, dev_part1, dev_part2;
  459. int i;
  460. struct device_info device_map[16];
  461. int func_nibble;
  462. int nibble;
  463. int nic_present = 0;
  464. int iscsi_present = 0;
  465. int iscsi_func_low = 0;
  466. /* Use the dev_partition register to determine the PCI function number
  467. * and then check drv_active register to see which driver is loaded */
  468. dev_part1 = qla4_83xx_rd_reg(ha,
  469. ha->reg_tbl[QLA8XXX_CRB_DEV_PART_INFO]);
  470. dev_part2 = qla4_83xx_rd_reg(ha, QLA83XX_CRB_DEV_PART_INFO2);
  471. drv_active = qla4_83xx_rd_reg(ha, ha->reg_tbl[QLA8XXX_CRB_DRV_ACTIVE]);
  472. /* Each function has 4 bits in dev_partition Info register,
  473. * Lower 2 bits - device type, Upper 2 bits - physical port number */
  474. dev_part = dev_part1;
  475. for (i = nibble = 0; i <= 15; i++, nibble++) {
  476. func_nibble = dev_part & (0xF << (nibble * 4));
  477. func_nibble >>= (nibble * 4);
  478. device_map[i].func_num = i;
  479. device_map[i].device_type = func_nibble & 0x3;
  480. device_map[i].port_num = func_nibble & 0xC;
  481. if (device_map[i].device_type == NIC_CLASS) {
  482. if (drv_active & (1 << device_map[i].func_num)) {
  483. nic_present++;
  484. break;
  485. }
  486. } else if (device_map[i].device_type == ISCSI_CLASS) {
  487. if (drv_active & (1 << device_map[i].func_num)) {
  488. if (!iscsi_present ||
  489. (iscsi_present &&
  490. (iscsi_func_low > device_map[i].func_num)))
  491. iscsi_func_low = device_map[i].func_num;
  492. iscsi_present++;
  493. }
  494. }
  495. /* For function_num[8..15] get info from dev_part2 register */
  496. if (nibble == 7) {
  497. nibble = 0;
  498. dev_part = dev_part2;
  499. }
  500. }
  501. /* NIC, iSCSI and FCOE are the Reset owners based on order, NIC gets
  502. * precedence over iSCSI and FCOE and iSCSI over FCOE, based on drivers
  503. * present. */
  504. if (!nic_present && (ha->func_num == iscsi_func_low)) {
  505. DEBUG2(ql4_printk(KERN_INFO, ha,
  506. "%s: can reset - NIC not present and lower iSCSI function is %d\n",
  507. __func__, ha->func_num));
  508. return 1;
  509. }
  510. return 0;
  511. }
  512. /**
  513. * qla4_83xx_need_reset_handler - Code to start reset sequence
  514. * @ha: pointer to adapter structure
  515. *
  516. * Note: IDC lock must be held upon entry
  517. **/
  518. void qla4_83xx_need_reset_handler(struct scsi_qla_host *ha)
  519. {
  520. uint32_t dev_state, drv_state, drv_active;
  521. unsigned long reset_timeout, dev_init_timeout;
  522. ql4_printk(KERN_INFO, ha, "%s: Performing ISP error recovery\n",
  523. __func__);
  524. if (!test_bit(AF_8XXX_RST_OWNER, &ha->flags)) {
  525. DEBUG2(ql4_printk(KERN_INFO, ha, "%s: reset acknowledged\n",
  526. __func__));
  527. qla4_8xxx_set_rst_ready(ha);
  528. /* Non-reset owners ACK Reset and wait for device INIT state
  529. * as part of Reset Recovery by Reset Owner */
  530. dev_init_timeout = jiffies + (ha->nx_dev_init_timeout * HZ);
  531. do {
  532. if (time_after_eq(jiffies, dev_init_timeout)) {
  533. ql4_printk(KERN_INFO, ha, "%s: Non Reset owner dev init timeout\n",
  534. __func__);
  535. break;
  536. }
  537. ha->isp_ops->idc_unlock(ha);
  538. msleep(1000);
  539. ha->isp_ops->idc_lock(ha);
  540. dev_state = qla4_8xxx_rd_direct(ha,
  541. QLA8XXX_CRB_DEV_STATE);
  542. } while (dev_state == QLA8XXX_DEV_NEED_RESET);
  543. } else {
  544. qla4_8xxx_set_rst_ready(ha);
  545. reset_timeout = jiffies + (ha->nx_reset_timeout * HZ);
  546. drv_state = qla4_8xxx_rd_direct(ha, QLA8XXX_CRB_DRV_STATE);
  547. drv_active = qla4_8xxx_rd_direct(ha, QLA8XXX_CRB_DRV_ACTIVE);
  548. ql4_printk(KERN_INFO, ha, "%s: drv_state = 0x%x, drv_active = 0x%x\n",
  549. __func__, drv_state, drv_active);
  550. while (drv_state != drv_active) {
  551. if (time_after_eq(jiffies, reset_timeout)) {
  552. ql4_printk(KERN_INFO, ha, "%s: %s: RESET TIMEOUT! drv_state: 0x%08x, drv_active: 0x%08x\n",
  553. __func__, DRIVER_NAME, drv_state,
  554. drv_active);
  555. break;
  556. }
  557. ha->isp_ops->idc_unlock(ha);
  558. msleep(1000);
  559. ha->isp_ops->idc_lock(ha);
  560. drv_state = qla4_8xxx_rd_direct(ha,
  561. QLA8XXX_CRB_DRV_STATE);
  562. drv_active = qla4_8xxx_rd_direct(ha,
  563. QLA8XXX_CRB_DRV_ACTIVE);
  564. }
  565. if (drv_state != drv_active) {
  566. ql4_printk(KERN_INFO, ha, "%s: Reset_owner turning off drv_active of non-acking function 0x%x\n",
  567. __func__, (drv_active ^ drv_state));
  568. drv_active = drv_active & drv_state;
  569. qla4_8xxx_wr_direct(ha, QLA8XXX_CRB_DRV_ACTIVE,
  570. drv_active);
  571. }
  572. clear_bit(AF_8XXX_RST_OWNER, &ha->flags);
  573. /* Start Reset Recovery */
  574. qla4_8xxx_device_bootstrap(ha);
  575. }
  576. }
  577. void qla4_83xx_get_idc_param(struct scsi_qla_host *ha)
  578. {
  579. uint32_t idc_params, ret_val;
  580. ret_val = qla4_83xx_flash_read_u32(ha, QLA83XX_IDC_PARAM_ADDR,
  581. (uint8_t *)&idc_params, 1);
  582. if (ret_val == QLA_SUCCESS) {
  583. ha->nx_dev_init_timeout = idc_params & 0xFFFF;
  584. ha->nx_reset_timeout = (idc_params >> 16) & 0xFFFF;
  585. } else {
  586. ha->nx_dev_init_timeout = ROM_DEV_INIT_TIMEOUT;
  587. ha->nx_reset_timeout = ROM_DRV_RESET_ACK_TIMEOUT;
  588. }
  589. DEBUG2(ql4_printk(KERN_DEBUG, ha,
  590. "%s: ha->nx_dev_init_timeout = %d, ha->nx_reset_timeout = %d\n",
  591. __func__, ha->nx_dev_init_timeout,
  592. ha->nx_reset_timeout));
  593. }
  594. /*-------------------------Reset Sequence Functions-----------------------*/
  595. static void qla4_83xx_dump_reset_seq_hdr(struct scsi_qla_host *ha)
  596. {
  597. uint8_t *phdr;
  598. if (!ha->reset_tmplt.buff) {
  599. ql4_printk(KERN_ERR, ha, "%s: Error: Invalid reset_seq_template\n",
  600. __func__);
  601. return;
  602. }
  603. phdr = ha->reset_tmplt.buff;
  604. DEBUG2(ql4_printk(KERN_INFO, ha,
  605. "Reset Template: 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X\n",
  606. *phdr, *(phdr+1), *(phdr+2), *(phdr+3), *(phdr+4),
  607. *(phdr+5), *(phdr+6), *(phdr+7), *(phdr + 8),
  608. *(phdr+9), *(phdr+10), *(phdr+11), *(phdr+12),
  609. *(phdr+13), *(phdr+14), *(phdr+15)));
  610. }
  611. static int qla4_83xx_copy_bootloader(struct scsi_qla_host *ha)
  612. {
  613. uint8_t *p_cache;
  614. uint32_t src, count, size;
  615. uint64_t dest;
  616. int ret_val = QLA_SUCCESS;
  617. src = QLA83XX_BOOTLOADER_FLASH_ADDR;
  618. dest = qla4_83xx_rd_reg(ha, QLA83XX_BOOTLOADER_ADDR);
  619. size = qla4_83xx_rd_reg(ha, QLA83XX_BOOTLOADER_SIZE);
  620. /* 128 bit alignment check */
  621. if (size & 0xF)
  622. size = (size + 16) & ~0xF;
  623. /* 16 byte count */
  624. count = size/16;
  625. p_cache = vmalloc(size);
  626. if (p_cache == NULL) {
  627. ql4_printk(KERN_ERR, ha, "%s: Failed to allocate memory for boot loader cache\n",
  628. __func__);
  629. ret_val = QLA_ERROR;
  630. goto exit_copy_bootloader;
  631. }
  632. ret_val = qla4_83xx_lockless_flash_read_u32(ha, src, p_cache,
  633. size / sizeof(uint32_t));
  634. if (ret_val == QLA_ERROR) {
  635. ql4_printk(KERN_ERR, ha, "%s: Error reading firmware from flash\n",
  636. __func__);
  637. goto exit_copy_error;
  638. }
  639. DEBUG2(ql4_printk(KERN_INFO, ha, "%s: Read firmware from flash\n",
  640. __func__));
  641. /* 128 bit/16 byte write to MS memory */
  642. ret_val = qla4_83xx_ms_mem_write_128b(ha, dest, (uint32_t *)p_cache,
  643. count);
  644. if (ret_val == QLA_ERROR) {
  645. ql4_printk(KERN_ERR, ha, "%s: Error writing firmware to MS\n",
  646. __func__);
  647. goto exit_copy_error;
  648. }
  649. DEBUG2(ql4_printk(KERN_INFO, ha, "%s: Wrote firmware size %d to MS\n",
  650. __func__, size));
  651. exit_copy_error:
  652. vfree(p_cache);
  653. exit_copy_bootloader:
  654. return ret_val;
  655. }
  656. static int qla4_83xx_check_cmd_peg_status(struct scsi_qla_host *ha)
  657. {
  658. uint32_t val, ret_val = QLA_ERROR;
  659. int retries = CRB_CMDPEG_CHECK_RETRY_COUNT;
  660. do {
  661. val = qla4_83xx_rd_reg(ha, QLA83XX_CMDPEG_STATE);
  662. if (val == PHAN_INITIALIZE_COMPLETE) {
  663. DEBUG2(ql4_printk(KERN_INFO, ha,
  664. "%s: Command Peg initialization complete. State=0x%x\n",
  665. __func__, val));
  666. ret_val = QLA_SUCCESS;
  667. break;
  668. }
  669. msleep(CRB_CMDPEG_CHECK_DELAY);
  670. } while (--retries);
  671. return ret_val;
  672. }
  673. /**
  674. * qla4_83xx_poll_reg - Poll the given CRB addr for duration msecs till
  675. * value read ANDed with test_mask is equal to test_result.
  676. *
  677. * @ha : Pointer to adapter structure
  678. * @addr : CRB register address
  679. * @duration : Poll for total of "duration" msecs
  680. * @test_mask : Mask value read with "test_mask"
  681. * @test_result : Compare (value&test_mask) with test_result.
  682. **/
  683. static int qla4_83xx_poll_reg(struct scsi_qla_host *ha, uint32_t addr,
  684. int duration, uint32_t test_mask,
  685. uint32_t test_result)
  686. {
  687. uint32_t value;
  688. uint8_t retries;
  689. int ret_val = QLA_SUCCESS;
  690. ret_val = qla4_83xx_rd_reg_indirect(ha, addr, &value);
  691. if (ret_val == QLA_ERROR)
  692. goto exit_poll_reg;
  693. retries = duration / 10;
  694. do {
  695. if ((value & test_mask) != test_result) {
  696. msleep(duration / 10);
  697. ret_val = qla4_83xx_rd_reg_indirect(ha, addr, &value);
  698. if (ret_val == QLA_ERROR)
  699. goto exit_poll_reg;
  700. ret_val = QLA_ERROR;
  701. } else {
  702. ret_val = QLA_SUCCESS;
  703. break;
  704. }
  705. } while (retries--);
  706. exit_poll_reg:
  707. if (ret_val == QLA_ERROR) {
  708. ha->reset_tmplt.seq_error++;
  709. ql4_printk(KERN_ERR, ha, "%s: Poll Failed: 0x%08x 0x%08x 0x%08x\n",
  710. __func__, value, test_mask, test_result);
  711. }
  712. return ret_val;
  713. }
  714. static int qla4_83xx_reset_seq_checksum_test(struct scsi_qla_host *ha)
  715. {
  716. uint32_t sum = 0;
  717. uint16_t *buff = (uint16_t *)ha->reset_tmplt.buff;
  718. int u16_count = ha->reset_tmplt.hdr->size / sizeof(uint16_t);
  719. int ret_val;
  720. while (u16_count-- > 0)
  721. sum += *buff++;
  722. while (sum >> 16)
  723. sum = (sum & 0xFFFF) + (sum >> 16);
  724. /* checksum of 0 indicates a valid template */
  725. if (~sum) {
  726. ret_val = QLA_SUCCESS;
  727. } else {
  728. ql4_printk(KERN_ERR, ha, "%s: Reset seq checksum failed\n",
  729. __func__);
  730. ret_val = QLA_ERROR;
  731. }
  732. return ret_val;
  733. }
  734. /**
  735. * qla4_83xx_read_reset_template - Read Reset Template from Flash
  736. * @ha: Pointer to adapter structure
  737. **/
  738. void qla4_83xx_read_reset_template(struct scsi_qla_host *ha)
  739. {
  740. uint8_t *p_buff;
  741. uint32_t addr, tmplt_hdr_def_size, tmplt_hdr_size;
  742. uint32_t ret_val;
  743. ha->reset_tmplt.seq_error = 0;
  744. ha->reset_tmplt.buff = vmalloc(QLA83XX_RESTART_TEMPLATE_SIZE);
  745. if (ha->reset_tmplt.buff == NULL) {
  746. ql4_printk(KERN_ERR, ha, "%s: Failed to allocate reset template resources\n",
  747. __func__);
  748. goto exit_read_reset_template;
  749. }
  750. p_buff = ha->reset_tmplt.buff;
  751. addr = QLA83XX_RESET_TEMPLATE_ADDR;
  752. tmplt_hdr_def_size = sizeof(struct qla4_83xx_reset_template_hdr) /
  753. sizeof(uint32_t);
  754. DEBUG2(ql4_printk(KERN_INFO, ha,
  755. "%s: Read template hdr size %d from Flash\n",
  756. __func__, tmplt_hdr_def_size));
  757. /* Copy template header from flash */
  758. ret_val = qla4_83xx_flash_read_u32(ha, addr, p_buff,
  759. tmplt_hdr_def_size);
  760. if (ret_val != QLA_SUCCESS) {
  761. ql4_printk(KERN_ERR, ha, "%s: Failed to read reset template\n",
  762. __func__);
  763. goto exit_read_template_error;
  764. }
  765. ha->reset_tmplt.hdr =
  766. (struct qla4_83xx_reset_template_hdr *)ha->reset_tmplt.buff;
  767. /* Validate the template header size and signature */
  768. tmplt_hdr_size = ha->reset_tmplt.hdr->hdr_size/sizeof(uint32_t);
  769. if ((tmplt_hdr_size != tmplt_hdr_def_size) ||
  770. (ha->reset_tmplt.hdr->signature != RESET_TMPLT_HDR_SIGNATURE)) {
  771. ql4_printk(KERN_ERR, ha, "%s: Template Header size %d is invalid, tmplt_hdr_def_size %d\n",
  772. __func__, tmplt_hdr_size, tmplt_hdr_def_size);
  773. goto exit_read_template_error;
  774. }
  775. addr = QLA83XX_RESET_TEMPLATE_ADDR + ha->reset_tmplt.hdr->hdr_size;
  776. p_buff = ha->reset_tmplt.buff + ha->reset_tmplt.hdr->hdr_size;
  777. tmplt_hdr_def_size = (ha->reset_tmplt.hdr->size -
  778. ha->reset_tmplt.hdr->hdr_size) / sizeof(uint32_t);
  779. DEBUG2(ql4_printk(KERN_INFO, ha,
  780. "%s: Read rest of the template size %d\n",
  781. __func__, ha->reset_tmplt.hdr->size));
  782. /* Copy rest of the template */
  783. ret_val = qla4_83xx_flash_read_u32(ha, addr, p_buff,
  784. tmplt_hdr_def_size);
  785. if (ret_val != QLA_SUCCESS) {
  786. ql4_printk(KERN_ERR, ha, "%s: Failed to read reset tempelate\n",
  787. __func__);
  788. goto exit_read_template_error;
  789. }
  790. /* Integrity check */
  791. if (qla4_83xx_reset_seq_checksum_test(ha)) {
  792. ql4_printk(KERN_ERR, ha, "%s: Reset Seq checksum failed!\n",
  793. __func__);
  794. goto exit_read_template_error;
  795. }
  796. DEBUG2(ql4_printk(KERN_INFO, ha,
  797. "%s: Reset Seq checksum passed, Get stop, start and init seq offsets\n",
  798. __func__));
  799. /* Get STOP, START, INIT sequence offsets */
  800. ha->reset_tmplt.init_offset = ha->reset_tmplt.buff +
  801. ha->reset_tmplt.hdr->init_seq_offset;
  802. ha->reset_tmplt.start_offset = ha->reset_tmplt.buff +
  803. ha->reset_tmplt.hdr->start_seq_offset;
  804. ha->reset_tmplt.stop_offset = ha->reset_tmplt.buff +
  805. ha->reset_tmplt.hdr->hdr_size;
  806. qla4_83xx_dump_reset_seq_hdr(ha);
  807. goto exit_read_reset_template;
  808. exit_read_template_error:
  809. vfree(ha->reset_tmplt.buff);
  810. exit_read_reset_template:
  811. return;
  812. }
  813. /**
  814. * qla4_83xx_read_write_crb_reg - Read from raddr and write value to waddr.
  815. *
  816. * @ha : Pointer to adapter structure
  817. * @raddr : CRB address to read from
  818. * @waddr : CRB address to write to
  819. **/
  820. static void qla4_83xx_read_write_crb_reg(struct scsi_qla_host *ha,
  821. uint32_t raddr, uint32_t waddr)
  822. {
  823. uint32_t value;
  824. qla4_83xx_rd_reg_indirect(ha, raddr, &value);
  825. qla4_83xx_wr_reg_indirect(ha, waddr, value);
  826. }
  827. /**
  828. * qla4_83xx_rmw_crb_reg - Read Modify Write crb register
  829. *
  830. * This function read value from raddr, AND with test_mask,
  831. * Shift Left,Right/OR/XOR with values RMW header and write value to waddr.
  832. *
  833. * @ha : Pointer to adapter structure
  834. * @raddr : CRB address to read from
  835. * @waddr : CRB address to write to
  836. * @p_rmw_hdr : header with shift/or/xor values.
  837. **/
  838. static void qla4_83xx_rmw_crb_reg(struct scsi_qla_host *ha, uint32_t raddr,
  839. uint32_t waddr,
  840. struct qla4_83xx_rmw *p_rmw_hdr)
  841. {
  842. uint32_t value;
  843. if (p_rmw_hdr->index_a)
  844. value = ha->reset_tmplt.array[p_rmw_hdr->index_a];
  845. else
  846. qla4_83xx_rd_reg_indirect(ha, raddr, &value);
  847. value &= p_rmw_hdr->test_mask;
  848. value <<= p_rmw_hdr->shl;
  849. value >>= p_rmw_hdr->shr;
  850. value |= p_rmw_hdr->or_value;
  851. value ^= p_rmw_hdr->xor_value;
  852. qla4_83xx_wr_reg_indirect(ha, waddr, value);
  853. return;
  854. }
  855. static void qla4_83xx_write_list(struct scsi_qla_host *ha,
  856. struct qla4_83xx_reset_entry_hdr *p_hdr)
  857. {
  858. struct qla4_83xx_entry *p_entry;
  859. uint32_t i;
  860. p_entry = (struct qla4_83xx_entry *)
  861. ((char *)p_hdr + sizeof(struct qla4_83xx_reset_entry_hdr));
  862. for (i = 0; i < p_hdr->count; i++, p_entry++) {
  863. qla4_83xx_wr_reg_indirect(ha, p_entry->arg1, p_entry->arg2);
  864. if (p_hdr->delay)
  865. udelay((uint32_t)(p_hdr->delay));
  866. }
  867. }
  868. static void qla4_83xx_read_write_list(struct scsi_qla_host *ha,
  869. struct qla4_83xx_reset_entry_hdr *p_hdr)
  870. {
  871. struct qla4_83xx_entry *p_entry;
  872. uint32_t i;
  873. p_entry = (struct qla4_83xx_entry *)
  874. ((char *)p_hdr + sizeof(struct qla4_83xx_reset_entry_hdr));
  875. for (i = 0; i < p_hdr->count; i++, p_entry++) {
  876. qla4_83xx_read_write_crb_reg(ha, p_entry->arg1, p_entry->arg2);
  877. if (p_hdr->delay)
  878. udelay((uint32_t)(p_hdr->delay));
  879. }
  880. }
  881. static void qla4_83xx_poll_list(struct scsi_qla_host *ha,
  882. struct qla4_83xx_reset_entry_hdr *p_hdr)
  883. {
  884. long delay;
  885. struct qla4_83xx_entry *p_entry;
  886. struct qla4_83xx_poll *p_poll;
  887. uint32_t i;
  888. uint32_t value;
  889. p_poll = (struct qla4_83xx_poll *)
  890. ((char *)p_hdr + sizeof(struct qla4_83xx_reset_entry_hdr));
  891. /* Entries start after 8 byte qla4_83xx_poll, poll header contains
  892. * the test_mask, test_value. */
  893. p_entry = (struct qla4_83xx_entry *)((char *)p_poll +
  894. sizeof(struct qla4_83xx_poll));
  895. delay = (long)p_hdr->delay;
  896. if (!delay) {
  897. for (i = 0; i < p_hdr->count; i++, p_entry++) {
  898. qla4_83xx_poll_reg(ha, p_entry->arg1, delay,
  899. p_poll->test_mask,
  900. p_poll->test_value);
  901. }
  902. } else {
  903. for (i = 0; i < p_hdr->count; i++, p_entry++) {
  904. if (qla4_83xx_poll_reg(ha, p_entry->arg1, delay,
  905. p_poll->test_mask,
  906. p_poll->test_value)) {
  907. qla4_83xx_rd_reg_indirect(ha, p_entry->arg1,
  908. &value);
  909. qla4_83xx_rd_reg_indirect(ha, p_entry->arg2,
  910. &value);
  911. }
  912. }
  913. }
  914. }
  915. static void qla4_83xx_poll_write_list(struct scsi_qla_host *ha,
  916. struct qla4_83xx_reset_entry_hdr *p_hdr)
  917. {
  918. long delay;
  919. struct qla4_83xx_quad_entry *p_entry;
  920. struct qla4_83xx_poll *p_poll;
  921. uint32_t i;
  922. p_poll = (struct qla4_83xx_poll *)
  923. ((char *)p_hdr + sizeof(struct qla4_83xx_reset_entry_hdr));
  924. p_entry = (struct qla4_83xx_quad_entry *)
  925. ((char *)p_poll + sizeof(struct qla4_83xx_poll));
  926. delay = (long)p_hdr->delay;
  927. for (i = 0; i < p_hdr->count; i++, p_entry++) {
  928. qla4_83xx_wr_reg_indirect(ha, p_entry->dr_addr,
  929. p_entry->dr_value);
  930. qla4_83xx_wr_reg_indirect(ha, p_entry->ar_addr,
  931. p_entry->ar_value);
  932. if (delay) {
  933. if (qla4_83xx_poll_reg(ha, p_entry->ar_addr, delay,
  934. p_poll->test_mask,
  935. p_poll->test_value)) {
  936. DEBUG2(ql4_printk(KERN_INFO, ha,
  937. "%s: Timeout Error: poll list, item_num %d, entry_num %d\n",
  938. __func__, i,
  939. ha->reset_tmplt.seq_index));
  940. }
  941. }
  942. }
  943. }
  944. static void qla4_83xx_read_modify_write(struct scsi_qla_host *ha,
  945. struct qla4_83xx_reset_entry_hdr *p_hdr)
  946. {
  947. struct qla4_83xx_entry *p_entry;
  948. struct qla4_83xx_rmw *p_rmw_hdr;
  949. uint32_t i;
  950. p_rmw_hdr = (struct qla4_83xx_rmw *)
  951. ((char *)p_hdr + sizeof(struct qla4_83xx_reset_entry_hdr));
  952. p_entry = (struct qla4_83xx_entry *)
  953. ((char *)p_rmw_hdr + sizeof(struct qla4_83xx_rmw));
  954. for (i = 0; i < p_hdr->count; i++, p_entry++) {
  955. qla4_83xx_rmw_crb_reg(ha, p_entry->arg1, p_entry->arg2,
  956. p_rmw_hdr);
  957. if (p_hdr->delay)
  958. udelay((uint32_t)(p_hdr->delay));
  959. }
  960. }
  961. static void qla4_83xx_pause(struct scsi_qla_host *ha,
  962. struct qla4_83xx_reset_entry_hdr *p_hdr)
  963. {
  964. if (p_hdr->delay)
  965. mdelay((uint32_t)((long)p_hdr->delay));
  966. }
  967. static void qla4_83xx_poll_read_list(struct scsi_qla_host *ha,
  968. struct qla4_83xx_reset_entry_hdr *p_hdr)
  969. {
  970. long delay;
  971. int index;
  972. struct qla4_83xx_quad_entry *p_entry;
  973. struct qla4_83xx_poll *p_poll;
  974. uint32_t i;
  975. uint32_t value;
  976. p_poll = (struct qla4_83xx_poll *)
  977. ((char *)p_hdr + sizeof(struct qla4_83xx_reset_entry_hdr));
  978. p_entry = (struct qla4_83xx_quad_entry *)
  979. ((char *)p_poll + sizeof(struct qla4_83xx_poll));
  980. delay = (long)p_hdr->delay;
  981. for (i = 0; i < p_hdr->count; i++, p_entry++) {
  982. qla4_83xx_wr_reg_indirect(ha, p_entry->ar_addr,
  983. p_entry->ar_value);
  984. if (delay) {
  985. if (qla4_83xx_poll_reg(ha, p_entry->ar_addr, delay,
  986. p_poll->test_mask,
  987. p_poll->test_value)) {
  988. DEBUG2(ql4_printk(KERN_INFO, ha,
  989. "%s: Timeout Error: poll list, Item_num %d, entry_num %d\n",
  990. __func__, i,
  991. ha->reset_tmplt.seq_index));
  992. } else {
  993. index = ha->reset_tmplt.array_index;
  994. qla4_83xx_rd_reg_indirect(ha, p_entry->dr_addr,
  995. &value);
  996. ha->reset_tmplt.array[index++] = value;
  997. if (index == QLA83XX_MAX_RESET_SEQ_ENTRIES)
  998. ha->reset_tmplt.array_index = 1;
  999. }
  1000. }
  1001. }
  1002. }
  1003. static void qla4_83xx_seq_end(struct scsi_qla_host *ha,
  1004. struct qla4_83xx_reset_entry_hdr *p_hdr)
  1005. {
  1006. ha->reset_tmplt.seq_end = 1;
  1007. }
  1008. static void qla4_83xx_template_end(struct scsi_qla_host *ha,
  1009. struct qla4_83xx_reset_entry_hdr *p_hdr)
  1010. {
  1011. ha->reset_tmplt.template_end = 1;
  1012. if (ha->reset_tmplt.seq_error == 0) {
  1013. DEBUG2(ql4_printk(KERN_INFO, ha,
  1014. "%s: Reset sequence completed SUCCESSFULLY.\n",
  1015. __func__));
  1016. } else {
  1017. ql4_printk(KERN_ERR, ha, "%s: Reset sequence completed with some timeout errors.\n",
  1018. __func__);
  1019. }
  1020. }
  1021. /**
  1022. * qla4_83xx_process_reset_template - Process reset template.
  1023. *
  1024. * Process all entries in reset template till entry with SEQ_END opcode,
  1025. * which indicates end of the reset template processing. Each entry has a
  1026. * Reset Entry header, entry opcode/command, with size of the entry, number
  1027. * of entries in sub-sequence and delay in microsecs or timeout in millisecs.
  1028. *
  1029. * @ha : Pointer to adapter structure
  1030. * @p_buff : Common reset entry header.
  1031. **/
  1032. static void qla4_83xx_process_reset_template(struct scsi_qla_host *ha,
  1033. char *p_buff)
  1034. {
  1035. int index, entries;
  1036. struct qla4_83xx_reset_entry_hdr *p_hdr;
  1037. char *p_entry = p_buff;
  1038. ha->reset_tmplt.seq_end = 0;
  1039. ha->reset_tmplt.template_end = 0;
  1040. entries = ha->reset_tmplt.hdr->entries;
  1041. index = ha->reset_tmplt.seq_index;
  1042. for (; (!ha->reset_tmplt.seq_end) && (index < entries); index++) {
  1043. p_hdr = (struct qla4_83xx_reset_entry_hdr *)p_entry;
  1044. switch (p_hdr->cmd) {
  1045. case OPCODE_NOP:
  1046. break;
  1047. case OPCODE_WRITE_LIST:
  1048. qla4_83xx_write_list(ha, p_hdr);
  1049. break;
  1050. case OPCODE_READ_WRITE_LIST:
  1051. qla4_83xx_read_write_list(ha, p_hdr);
  1052. break;
  1053. case OPCODE_POLL_LIST:
  1054. qla4_83xx_poll_list(ha, p_hdr);
  1055. break;
  1056. case OPCODE_POLL_WRITE_LIST:
  1057. qla4_83xx_poll_write_list(ha, p_hdr);
  1058. break;
  1059. case OPCODE_READ_MODIFY_WRITE:
  1060. qla4_83xx_read_modify_write(ha, p_hdr);
  1061. break;
  1062. case OPCODE_SEQ_PAUSE:
  1063. qla4_83xx_pause(ha, p_hdr);
  1064. break;
  1065. case OPCODE_SEQ_END:
  1066. qla4_83xx_seq_end(ha, p_hdr);
  1067. break;
  1068. case OPCODE_TMPL_END:
  1069. qla4_83xx_template_end(ha, p_hdr);
  1070. break;
  1071. case OPCODE_POLL_READ_LIST:
  1072. qla4_83xx_poll_read_list(ha, p_hdr);
  1073. break;
  1074. default:
  1075. ql4_printk(KERN_ERR, ha, "%s: Unknown command ==> 0x%04x on entry = %d\n",
  1076. __func__, p_hdr->cmd, index);
  1077. break;
  1078. }
  1079. /* Set pointer to next entry in the sequence. */
  1080. p_entry += p_hdr->size;
  1081. }
  1082. ha->reset_tmplt.seq_index = index;
  1083. }
  1084. static void qla4_83xx_process_stop_seq(struct scsi_qla_host *ha)
  1085. {
  1086. ha->reset_tmplt.seq_index = 0;
  1087. qla4_83xx_process_reset_template(ha, ha->reset_tmplt.stop_offset);
  1088. if (ha->reset_tmplt.seq_end != 1)
  1089. ql4_printk(KERN_ERR, ha, "%s: Abrupt STOP Sub-Sequence end.\n",
  1090. __func__);
  1091. }
  1092. static void qla4_83xx_process_start_seq(struct scsi_qla_host *ha)
  1093. {
  1094. qla4_83xx_process_reset_template(ha, ha->reset_tmplt.start_offset);
  1095. if (ha->reset_tmplt.template_end != 1)
  1096. ql4_printk(KERN_ERR, ha, "%s: Abrupt START Sub-Sequence end.\n",
  1097. __func__);
  1098. }
  1099. static void qla4_83xx_process_init_seq(struct scsi_qla_host *ha)
  1100. {
  1101. qla4_83xx_process_reset_template(ha, ha->reset_tmplt.init_offset);
  1102. if (ha->reset_tmplt.seq_end != 1)
  1103. ql4_printk(KERN_ERR, ha, "%s: Abrupt INIT Sub-Sequence end.\n",
  1104. __func__);
  1105. }
  1106. static int qla4_83xx_restart(struct scsi_qla_host *ha)
  1107. {
  1108. int ret_val = QLA_SUCCESS;
  1109. qla4_83xx_process_stop_seq(ha);
  1110. /* Collect minidump*/
  1111. if (!test_and_clear_bit(AF_83XX_NO_FW_DUMP, &ha->flags))
  1112. qla4_8xxx_get_minidump(ha);
  1113. qla4_83xx_process_init_seq(ha);
  1114. if (qla4_83xx_copy_bootloader(ha)) {
  1115. ql4_printk(KERN_ERR, ha, "%s: Copy bootloader, firmware restart failed!\n",
  1116. __func__);
  1117. ret_val = QLA_ERROR;
  1118. goto exit_restart;
  1119. }
  1120. qla4_83xx_wr_reg(ha, QLA83XX_FW_IMAGE_VALID, QLA83XX_BOOT_FROM_FLASH);
  1121. qla4_83xx_process_start_seq(ha);
  1122. exit_restart:
  1123. return ret_val;
  1124. }
  1125. int qla4_83xx_start_firmware(struct scsi_qla_host *ha)
  1126. {
  1127. int ret_val = QLA_SUCCESS;
  1128. ret_val = qla4_83xx_restart(ha);
  1129. if (ret_val == QLA_ERROR) {
  1130. ql4_printk(KERN_ERR, ha, "%s: Restart error\n", __func__);
  1131. goto exit_start_fw;
  1132. } else {
  1133. DEBUG2(ql4_printk(KERN_INFO, ha, "%s: Restart done\n",
  1134. __func__));
  1135. }
  1136. ret_val = qla4_83xx_check_cmd_peg_status(ha);
  1137. if (ret_val == QLA_ERROR)
  1138. ql4_printk(KERN_ERR, ha, "%s: Peg not initialized\n",
  1139. __func__);
  1140. exit_start_fw:
  1141. return ret_val;
  1142. }
  1143. /*----------------------Interrupt Related functions ---------------------*/
  1144. static void qla4_83xx_disable_iocb_intrs(struct scsi_qla_host *ha)
  1145. {
  1146. if (test_and_clear_bit(AF_83XX_IOCB_INTR_ON, &ha->flags))
  1147. qla4_8xxx_intr_disable(ha);
  1148. }
  1149. static void qla4_83xx_disable_mbox_intrs(struct scsi_qla_host *ha)
  1150. {
  1151. uint32_t mb_int, ret;
  1152. if (test_and_clear_bit(AF_83XX_MBOX_INTR_ON, &ha->flags)) {
  1153. ret = readl(&ha->qla4_83xx_reg->mbox_int);
  1154. mb_int = ret & ~INT_ENABLE_FW_MB;
  1155. writel(mb_int, &ha->qla4_83xx_reg->mbox_int);
  1156. writel(1, &ha->qla4_83xx_reg->leg_int_mask);
  1157. }
  1158. }
  1159. void qla4_83xx_disable_intrs(struct scsi_qla_host *ha)
  1160. {
  1161. qla4_83xx_disable_mbox_intrs(ha);
  1162. qla4_83xx_disable_iocb_intrs(ha);
  1163. }
  1164. static void qla4_83xx_enable_iocb_intrs(struct scsi_qla_host *ha)
  1165. {
  1166. if (!test_bit(AF_83XX_IOCB_INTR_ON, &ha->flags)) {
  1167. qla4_8xxx_intr_enable(ha);
  1168. set_bit(AF_83XX_IOCB_INTR_ON, &ha->flags);
  1169. }
  1170. }
  1171. void qla4_83xx_enable_mbox_intrs(struct scsi_qla_host *ha)
  1172. {
  1173. uint32_t mb_int;
  1174. if (!test_bit(AF_83XX_MBOX_INTR_ON, &ha->flags)) {
  1175. mb_int = INT_ENABLE_FW_MB;
  1176. writel(mb_int, &ha->qla4_83xx_reg->mbox_int);
  1177. writel(0, &ha->qla4_83xx_reg->leg_int_mask);
  1178. set_bit(AF_83XX_MBOX_INTR_ON, &ha->flags);
  1179. }
  1180. }
  1181. void qla4_83xx_enable_intrs(struct scsi_qla_host *ha)
  1182. {
  1183. qla4_83xx_enable_mbox_intrs(ha);
  1184. qla4_83xx_enable_iocb_intrs(ha);
  1185. }
  1186. void qla4_83xx_queue_mbox_cmd(struct scsi_qla_host *ha, uint32_t *mbx_cmd,
  1187. int incount)
  1188. {
  1189. int i;
  1190. /* Load all mailbox registers, except mailbox 0. */
  1191. for (i = 1; i < incount; i++)
  1192. writel(mbx_cmd[i], &ha->qla4_83xx_reg->mailbox_in[i]);
  1193. writel(mbx_cmd[0], &ha->qla4_83xx_reg->mailbox_in[0]);
  1194. /* Set Host Interrupt register to 1, to tell the firmware that
  1195. * a mailbox command is pending. Firmware after reading the
  1196. * mailbox command, clears the host interrupt register */
  1197. writel(HINT_MBX_INT_PENDING, &ha->qla4_83xx_reg->host_intr);
  1198. }
  1199. void qla4_83xx_process_mbox_intr(struct scsi_qla_host *ha, int outcount)
  1200. {
  1201. int intr_status;
  1202. intr_status = readl(&ha->qla4_83xx_reg->risc_intr);
  1203. if (intr_status) {
  1204. ha->mbox_status_count = outcount;
  1205. ha->isp_ops->interrupt_service_routine(ha, intr_status);
  1206. }
  1207. }
  1208. /**
  1209. * qla4_83xx_isp_reset - Resets ISP and aborts all outstanding commands.
  1210. * @ha: pointer to host adapter structure.
  1211. **/
  1212. int qla4_83xx_isp_reset(struct scsi_qla_host *ha)
  1213. {
  1214. int rval;
  1215. uint32_t dev_state;
  1216. ha->isp_ops->idc_lock(ha);
  1217. dev_state = qla4_8xxx_rd_direct(ha, QLA8XXX_CRB_DEV_STATE);
  1218. if (ql4xdontresethba)
  1219. qla4_83xx_set_idc_dontreset(ha);
  1220. if (dev_state == QLA8XXX_DEV_READY) {
  1221. /* If IDC_CTRL DONTRESETHBA_BIT0 is set dont do reset
  1222. * recovery */
  1223. if (qla4_83xx_idc_dontreset(ha) == DONTRESET_BIT0) {
  1224. ql4_printk(KERN_ERR, ha, "%s: Reset recovery disabled\n",
  1225. __func__);
  1226. rval = QLA_ERROR;
  1227. goto exit_isp_reset;
  1228. }
  1229. DEBUG2(ql4_printk(KERN_INFO, ha, "%s: HW State: NEED RESET\n",
  1230. __func__));
  1231. qla4_8xxx_wr_direct(ha, QLA8XXX_CRB_DEV_STATE,
  1232. QLA8XXX_DEV_NEED_RESET);
  1233. } else {
  1234. /* If device_state is NEED_RESET, go ahead with
  1235. * Reset,irrespective of ql4xdontresethba. This is to allow a
  1236. * non-reset-owner to force a reset. Non-reset-owner sets
  1237. * the IDC_CTRL BIT0 to prevent Reset-owner from doing a Reset
  1238. * and then forces a Reset by setting device_state to
  1239. * NEED_RESET. */
  1240. DEBUG2(ql4_printk(KERN_INFO, ha,
  1241. "%s: HW state already set to NEED_RESET\n",
  1242. __func__));
  1243. }
  1244. /* For ISP8324, Reset owner is NIC, iSCSI or FCOE based on priority
  1245. * and which drivers are present. Unlike ISP8022, the function setting
  1246. * NEED_RESET, may not be the Reset owner. */
  1247. if (qla4_83xx_can_perform_reset(ha))
  1248. set_bit(AF_8XXX_RST_OWNER, &ha->flags);
  1249. ha->isp_ops->idc_unlock(ha);
  1250. rval = qla4_8xxx_device_state_handler(ha);
  1251. ha->isp_ops->idc_lock(ha);
  1252. qla4_8xxx_clear_rst_ready(ha);
  1253. exit_isp_reset:
  1254. ha->isp_ops->idc_unlock(ha);
  1255. if (rval == QLA_SUCCESS)
  1256. clear_bit(AF_FW_RECOVERY, &ha->flags);
  1257. return rval;
  1258. }
  1259. static void qla4_83xx_dump_pause_control_regs(struct scsi_qla_host *ha)
  1260. {
  1261. u32 val = 0, val1 = 0;
  1262. int i, status = QLA_SUCCESS;
  1263. status = qla4_83xx_rd_reg_indirect(ha, QLA83XX_SRE_SHIM_CONTROL, &val);
  1264. DEBUG2(ql4_printk(KERN_INFO, ha, "SRE-Shim Ctrl:0x%x\n", val));
  1265. /* Port 0 Rx Buffer Pause Threshold Registers. */
  1266. DEBUG2(ql4_printk(KERN_INFO, ha,
  1267. "Port 0 Rx Buffer Pause Threshold Registers[TC7..TC0]:"));
  1268. for (i = 0; i < 8; i++) {
  1269. status = qla4_83xx_rd_reg_indirect(ha,
  1270. QLA83XX_PORT0_RXB_PAUSE_THRS + (i * 0x4), &val);
  1271. DEBUG2(pr_info("0x%x ", val));
  1272. }
  1273. DEBUG2(pr_info("\n"));
  1274. /* Port 1 Rx Buffer Pause Threshold Registers. */
  1275. DEBUG2(ql4_printk(KERN_INFO, ha,
  1276. "Port 1 Rx Buffer Pause Threshold Registers[TC7..TC0]:"));
  1277. for (i = 0; i < 8; i++) {
  1278. status = qla4_83xx_rd_reg_indirect(ha,
  1279. QLA83XX_PORT1_RXB_PAUSE_THRS + (i * 0x4), &val);
  1280. DEBUG2(pr_info("0x%x ", val));
  1281. }
  1282. DEBUG2(pr_info("\n"));
  1283. /* Port 0 RxB Traffic Class Max Cell Registers. */
  1284. DEBUG2(ql4_printk(KERN_INFO, ha,
  1285. "Port 0 RxB Traffic Class Max Cell Registers[3..0]:"));
  1286. for (i = 0; i < 4; i++) {
  1287. status = qla4_83xx_rd_reg_indirect(ha,
  1288. QLA83XX_PORT0_RXB_TC_MAX_CELL + (i * 0x4), &val);
  1289. DEBUG2(pr_info("0x%x ", val));
  1290. }
  1291. DEBUG2(pr_info("\n"));
  1292. /* Port 1 RxB Traffic Class Max Cell Registers. */
  1293. DEBUG2(ql4_printk(KERN_INFO, ha,
  1294. "Port 1 RxB Traffic Class Max Cell Registers[3..0]:"));
  1295. for (i = 0; i < 4; i++) {
  1296. status = qla4_83xx_rd_reg_indirect(ha,
  1297. QLA83XX_PORT1_RXB_TC_MAX_CELL + (i * 0x4), &val);
  1298. DEBUG2(pr_info("0x%x ", val));
  1299. }
  1300. DEBUG2(pr_info("\n"));
  1301. /* Port 0 RxB Rx Traffic Class Stats. */
  1302. DEBUG2(ql4_printk(KERN_INFO, ha,
  1303. "Port 0 RxB Rx Traffic Class Stats [TC7..TC0]"));
  1304. for (i = 7; i >= 0; i--) {
  1305. status = qla4_83xx_rd_reg_indirect(ha,
  1306. QLA83XX_PORT0_RXB_TC_STATS,
  1307. &val);
  1308. val &= ~(0x7 << 29); /* Reset bits 29 to 31 */
  1309. qla4_83xx_wr_reg_indirect(ha, QLA83XX_PORT0_RXB_TC_STATS,
  1310. (val | (i << 29)));
  1311. status = qla4_83xx_rd_reg_indirect(ha,
  1312. QLA83XX_PORT0_RXB_TC_STATS,
  1313. &val);
  1314. DEBUG2(pr_info("0x%x ", val));
  1315. }
  1316. DEBUG2(pr_info("\n"));
  1317. /* Port 1 RxB Rx Traffic Class Stats. */
  1318. DEBUG2(ql4_printk(KERN_INFO, ha,
  1319. "Port 1 RxB Rx Traffic Class Stats [TC7..TC0]"));
  1320. for (i = 7; i >= 0; i--) {
  1321. status = qla4_83xx_rd_reg_indirect(ha,
  1322. QLA83XX_PORT1_RXB_TC_STATS,
  1323. &val);
  1324. val &= ~(0x7 << 29); /* Reset bits 29 to 31 */
  1325. qla4_83xx_wr_reg_indirect(ha, QLA83XX_PORT1_RXB_TC_STATS,
  1326. (val | (i << 29)));
  1327. status = qla4_83xx_rd_reg_indirect(ha,
  1328. QLA83XX_PORT1_RXB_TC_STATS,
  1329. &val);
  1330. DEBUG2(pr_info("0x%x ", val));
  1331. }
  1332. DEBUG2(pr_info("\n"));
  1333. status = qla4_83xx_rd_reg_indirect(ha, QLA83XX_PORT2_IFB_PAUSE_THRS,
  1334. &val);
  1335. status = qla4_83xx_rd_reg_indirect(ha, QLA83XX_PORT3_IFB_PAUSE_THRS,
  1336. &val1);
  1337. DEBUG2(ql4_printk(KERN_INFO, ha,
  1338. "IFB-Pause Thresholds: Port 2:0x%x, Port 3:0x%x\n",
  1339. val, val1));
  1340. }
  1341. static void __qla4_83xx_disable_pause(struct scsi_qla_host *ha)
  1342. {
  1343. int i;
  1344. /* set SRE-Shim Control Register */
  1345. qla4_83xx_wr_reg_indirect(ha, QLA83XX_SRE_SHIM_CONTROL,
  1346. QLA83XX_SET_PAUSE_VAL);
  1347. for (i = 0; i < 8; i++) {
  1348. /* Port 0 Rx Buffer Pause Threshold Registers. */
  1349. qla4_83xx_wr_reg_indirect(ha,
  1350. QLA83XX_PORT0_RXB_PAUSE_THRS + (i * 0x4),
  1351. QLA83XX_SET_PAUSE_VAL);
  1352. /* Port 1 Rx Buffer Pause Threshold Registers. */
  1353. qla4_83xx_wr_reg_indirect(ha,
  1354. QLA83XX_PORT1_RXB_PAUSE_THRS + (i * 0x4),
  1355. QLA83XX_SET_PAUSE_VAL);
  1356. }
  1357. for (i = 0; i < 4; i++) {
  1358. /* Port 0 RxB Traffic Class Max Cell Registers. */
  1359. qla4_83xx_wr_reg_indirect(ha,
  1360. QLA83XX_PORT0_RXB_TC_MAX_CELL + (i * 0x4),
  1361. QLA83XX_SET_TC_MAX_CELL_VAL);
  1362. /* Port 1 RxB Traffic Class Max Cell Registers. */
  1363. qla4_83xx_wr_reg_indirect(ha,
  1364. QLA83XX_PORT1_RXB_TC_MAX_CELL + (i * 0x4),
  1365. QLA83XX_SET_TC_MAX_CELL_VAL);
  1366. }
  1367. qla4_83xx_wr_reg_indirect(ha, QLA83XX_PORT2_IFB_PAUSE_THRS,
  1368. QLA83XX_SET_PAUSE_VAL);
  1369. qla4_83xx_wr_reg_indirect(ha, QLA83XX_PORT3_IFB_PAUSE_THRS,
  1370. QLA83XX_SET_PAUSE_VAL);
  1371. ql4_printk(KERN_INFO, ha, "Disabled pause frames successfully.\n");
  1372. }
  1373. void qla4_83xx_disable_pause(struct scsi_qla_host *ha)
  1374. {
  1375. ha->isp_ops->idc_lock(ha);
  1376. qla4_83xx_dump_pause_control_regs(ha);
  1377. __qla4_83xx_disable_pause(ha);
  1378. ha->isp_ops->idc_unlock(ha);
  1379. }