1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189 |
- /*
- * linux/drivers/scsi/esas2r/esas2r_disc.c
- * esas2r device discovery routines
- *
- * Copyright (c) 2001-2013 ATTO Technology, Inc.
- * (mailto:linuxdrivers@attotech.com)
- */
- /*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
- /*
- * This program is free software; you can redistribute it and/or modify
- * it under the terms of the GNU General Public License as published by
- * the Free Software Foundation; version 2 of the License.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- * GNU General Public License for more details.
- *
- * NO WARRANTY
- * THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR
- * CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT
- * LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT,
- * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is
- * solely responsible for determining the appropriateness of using and
- * distributing the Program and assumes all risks associated with its
- * exercise of rights under this Agreement, including but not limited to
- * the risks and costs of program errors, damage to or loss of data,
- * programs or equipment, and unavailability or interruption of operations.
- *
- * DISCLAIMER OF LIABILITY
- * NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY
- * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
- * DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND
- * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR
- * TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
- * USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED
- * HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES
- *
- * You should have received a copy of the GNU General Public License
- * along with this program; if not, write to the Free Software
- * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
- */
- /*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
- #include "esas2r.h"
- /* Miscellaneous internal discovery routines */
- static void esas2r_disc_abort(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static bool esas2r_disc_continue(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static void esas2r_disc_fix_curr_requests(struct esas2r_adapter *a);
- static u32 esas2r_disc_get_phys_addr(struct esas2r_sg_context *sgc, u64 *addr);
- static bool esas2r_disc_start_request(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- /* Internal discovery routines that process the states */
- static bool esas2r_disc_block_dev_scan(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static void esas2r_disc_block_dev_scan_cb(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static bool esas2r_disc_dev_add(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static bool esas2r_disc_dev_remove(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static bool esas2r_disc_part_info(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static void esas2r_disc_part_info_cb(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static bool esas2r_disc_passthru_dev_info(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static void esas2r_disc_passthru_dev_info_cb(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static bool esas2r_disc_passthru_dev_addr(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static void esas2r_disc_passthru_dev_addr_cb(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static bool esas2r_disc_raid_grp_info(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- static void esas2r_disc_raid_grp_info_cb(struct esas2r_adapter *a,
- struct esas2r_request *rq);
- void esas2r_disc_initialize(struct esas2r_adapter *a)
- {
- struct esas2r_sas_nvram *nvr = a->nvram;
- esas2r_trace_enter();
- esas2r_lock_clear_flags(&a->flags, AF_DISC_IN_PROG);
- esas2r_lock_clear_flags(&a->flags2, AF2_DEV_SCAN);
- esas2r_lock_clear_flags(&a->flags2, AF2_DEV_CNT_OK);
- a->disc_start_time = jiffies_to_msecs(jiffies);
- a->disc_wait_time = nvr->dev_wait_time * 1000;
- a->disc_wait_cnt = nvr->dev_wait_count;
- if (a->disc_wait_cnt > ESAS2R_MAX_TARGETS)
- a->disc_wait_cnt = ESAS2R_MAX_TARGETS;
- /*
- * If we are doing chip reset or power management processing, always
- * wait for devices. use the NVRAM device count if it is greater than
- * previously discovered devices.
- */
- esas2r_hdebug("starting discovery...");
- a->general_req.interrupt_cx = NULL;
- if (a->flags & (AF_CHPRST_DETECTED | AF_POWER_MGT)) {
- if (a->prev_dev_cnt == 0) {
- /* Don't bother waiting if there is nothing to wait
- * for.
- */
- a->disc_wait_time = 0;
- } else {
- /*
- * Set the device wait count to what was previously
- * found. We don't care if the user only configured
- * a time because we know the exact count to wait for.
- * There is no need to honor the user's wishes to
- * always wait the full time.
- */
- a->disc_wait_cnt = a->prev_dev_cnt;
- /*
- * bump the minimum wait time to 15 seconds since the
- * default is 3 (system boot or the boot driver usually
- * buys us more time).
- */
- if (a->disc_wait_time < 15000)
- a->disc_wait_time = 15000;
- }
- }
- esas2r_trace("disc wait count: %d", a->disc_wait_cnt);
- esas2r_trace("disc wait time: %d", a->disc_wait_time);
- if (a->disc_wait_time == 0)
- esas2r_disc_check_complete(a);
- esas2r_trace_exit();
- }
- void esas2r_disc_start_waiting(struct esas2r_adapter *a)
- {
- unsigned long flags;
- spin_lock_irqsave(&a->mem_lock, flags);
- if (a->disc_ctx.disc_evt)
- esas2r_disc_start_port(a);
- spin_unlock_irqrestore(&a->mem_lock, flags);
- }
- void esas2r_disc_check_for_work(struct esas2r_adapter *a)
- {
- struct esas2r_request *rq = &a->general_req;
- /* service any pending interrupts first */
- esas2r_polled_interrupt(a);
- /*
- * now, interrupt processing may have queued up a discovery event. go
- * see if we have one to start. we couldn't start it in the ISR since
- * polled discovery would cause a deadlock.
- */
- esas2r_disc_start_waiting(a);
- if (rq->interrupt_cx == NULL)
- return;
- if (rq->req_stat == RS_STARTED
- && rq->timeout <= RQ_MAX_TIMEOUT) {
- /* wait for the current discovery request to complete. */
- esas2r_wait_request(a, rq);
- if (rq->req_stat == RS_TIMEOUT) {
- esas2r_disc_abort(a, rq);
- esas2r_local_reset_adapter(a);
- return;
- }
- }
- if (rq->req_stat == RS_PENDING
- || rq->req_stat == RS_STARTED)
- return;
- esas2r_disc_continue(a, rq);
- }
- void esas2r_disc_check_complete(struct esas2r_adapter *a)
- {
- unsigned long flags;
- esas2r_trace_enter();
- /* check to see if we should be waiting for devices */
- if (a->disc_wait_time) {
- u32 currtime = jiffies_to_msecs(jiffies);
- u32 time = currtime - a->disc_start_time;
- /*
- * Wait until the device wait time is exhausted or the device
- * wait count is satisfied.
- */
- if (time < a->disc_wait_time
- && (esas2r_targ_db_get_tgt_cnt(a) < a->disc_wait_cnt
- || a->disc_wait_cnt == 0)) {
- /* After three seconds of waiting, schedule a scan. */
- if (time >= 3000
- && !(esas2r_lock_set_flags(&a->flags2,
- AF2_DEV_SCAN) &
- ilog2(AF2_DEV_SCAN))) {
- spin_lock_irqsave(&a->mem_lock, flags);
- esas2r_disc_queue_event(a, DCDE_DEV_SCAN);
- spin_unlock_irqrestore(&a->mem_lock, flags);
- }
- esas2r_trace_exit();
- return;
- }
- /*
- * We are done waiting...we think. Adjust the wait time to
- * consume events after the count is met.
- */
- if (!(esas2r_lock_set_flags(&a->flags2, AF2_DEV_CNT_OK)
- & ilog2(AF2_DEV_CNT_OK)))
- a->disc_wait_time = time + 3000;
- /* If we haven't done a full scan yet, do it now. */
- if (!(esas2r_lock_set_flags(&a->flags2,
- AF2_DEV_SCAN) &
- ilog2(AF2_DEV_SCAN))) {
- spin_lock_irqsave(&a->mem_lock, flags);
- esas2r_disc_queue_event(a, DCDE_DEV_SCAN);
- spin_unlock_irqrestore(&a->mem_lock, flags);
- esas2r_trace_exit();
- return;
- }
- /*
- * Now, if there is still time left to consume events, continue
- * waiting.
- */
- if (time < a->disc_wait_time) {
- esas2r_trace_exit();
- return;
- }
- } else {
- if (!(esas2r_lock_set_flags(&a->flags2,
- AF2_DEV_SCAN) &
- ilog2(AF2_DEV_SCAN))) {
- spin_lock_irqsave(&a->mem_lock, flags);
- esas2r_disc_queue_event(a, DCDE_DEV_SCAN);
- spin_unlock_irqrestore(&a->mem_lock, flags);
- }
- }
- /* We want to stop waiting for devices. */
- a->disc_wait_time = 0;
- if ((a->flags & AF_DISC_POLLED)
- && (a->flags & AF_DISC_IN_PROG)) {
- /*
- * Polled discovery is still pending so continue the active
- * discovery until it is done. At that point, we will stop
- * polled discovery and transition to interrupt driven
- * discovery.
- */
- } else {
- /*
- * Done waiting for devices. Note that we get here immediately
- * after deferred waiting completes because that is interrupt
- * driven; i.e. There is no transition.
- */
- esas2r_disc_fix_curr_requests(a);
- esas2r_lock_clear_flags(&a->flags, AF_DISC_PENDING);
- /*
- * We have deferred target state changes until now because we
- * don't want to report any removals (due to the first arrival)
- * until the device wait time expires.
- */
- esas2r_lock_set_flags(&a->flags, AF_PORT_CHANGE);
- }
- esas2r_trace_exit();
- }
- void esas2r_disc_queue_event(struct esas2r_adapter *a, u8 disc_evt)
- {
- struct esas2r_disc_context *dc = &a->disc_ctx;
- esas2r_trace_enter();
- esas2r_trace("disc_event: %d", disc_evt);
- /* Initialize the discovery context */
- dc->disc_evt |= disc_evt;
- /*
- * Don't start discovery before or during polled discovery. if we did,
- * we would have a deadlock if we are in the ISR already.
- */
- if (!(a->flags & (AF_CHPRST_PENDING | AF_DISC_POLLED)))
- esas2r_disc_start_port(a);
- esas2r_trace_exit();
- }
- bool esas2r_disc_start_port(struct esas2r_adapter *a)
- {
- struct esas2r_request *rq = &a->general_req;
- struct esas2r_disc_context *dc = &a->disc_ctx;
- bool ret;
- esas2r_trace_enter();
- if (a->flags & AF_DISC_IN_PROG) {
- esas2r_trace_exit();
- return false;
- }
- /* If there is a discovery waiting, process it. */
- if (dc->disc_evt) {
- if ((a->flags & AF_DISC_POLLED)
- && a->disc_wait_time == 0) {
- /*
- * We are doing polled discovery, but we no longer want
- * to wait for devices. Stop polled discovery and
- * transition to interrupt driven discovery.
- */
- esas2r_trace_exit();
- return false;
- }
- } else {
- /* Discovery is complete. */
- esas2r_hdebug("disc done");
- esas2r_lock_set_flags(&a->flags, AF_PORT_CHANGE);
- esas2r_trace_exit();
- return false;
- }
- /* Handle the discovery context */
- esas2r_trace("disc_evt: %d", dc->disc_evt);
- esas2r_lock_set_flags(&a->flags, AF_DISC_IN_PROG);
- dc->flags = 0;
- if (a->flags & AF_DISC_POLLED)
- dc->flags |= DCF_POLLED;
- rq->interrupt_cx = dc;
- rq->req_stat = RS_SUCCESS;
- /* Decode the event code */
- if (dc->disc_evt & DCDE_DEV_SCAN) {
- dc->disc_evt &= ~DCDE_DEV_SCAN;
- dc->flags |= DCF_DEV_SCAN;
- dc->state = DCS_BLOCK_DEV_SCAN;
- } else if (dc->disc_evt & DCDE_DEV_CHANGE) {
- dc->disc_evt &= ~DCDE_DEV_CHANGE;
- dc->flags |= DCF_DEV_CHANGE;
- dc->state = DCS_DEV_RMV;
- }
- /* Continue interrupt driven discovery */
- if (!(a->flags & AF_DISC_POLLED))
- ret = esas2r_disc_continue(a, rq);
- else
- ret = true;
- esas2r_trace_exit();
- return ret;
- }
- static bool esas2r_disc_continue(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- bool rslt;
- /* Device discovery/removal */
- while (dc->flags & (DCF_DEV_CHANGE | DCF_DEV_SCAN)) {
- rslt = false;
- switch (dc->state) {
- case DCS_DEV_RMV:
- rslt = esas2r_disc_dev_remove(a, rq);
- break;
- case DCS_DEV_ADD:
- rslt = esas2r_disc_dev_add(a, rq);
- break;
- case DCS_BLOCK_DEV_SCAN:
- rslt = esas2r_disc_block_dev_scan(a, rq);
- break;
- case DCS_RAID_GRP_INFO:
- rslt = esas2r_disc_raid_grp_info(a, rq);
- break;
- case DCS_PART_INFO:
- rslt = esas2r_disc_part_info(a, rq);
- break;
- case DCS_PT_DEV_INFO:
- rslt = esas2r_disc_passthru_dev_info(a, rq);
- break;
- case DCS_PT_DEV_ADDR:
- rslt = esas2r_disc_passthru_dev_addr(a, rq);
- break;
- case DCS_DISC_DONE:
- dc->flags &= ~(DCF_DEV_CHANGE | DCF_DEV_SCAN);
- break;
- default:
- esas2r_bugon();
- dc->state = DCS_DISC_DONE;
- break;
- }
- if (rslt)
- return true;
- }
- /* Discovery is done...for now. */
- rq->interrupt_cx = NULL;
- if (!(a->flags & AF_DISC_PENDING))
- esas2r_disc_fix_curr_requests(a);
- esas2r_lock_clear_flags(&a->flags, AF_DISC_IN_PROG);
- /* Start the next discovery. */
- return esas2r_disc_start_port(a);
- }
- static bool esas2r_disc_start_request(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- unsigned long flags;
- /* Set the timeout to a minimum value. */
- if (rq->timeout < ESAS2R_DEFAULT_TMO)
- rq->timeout = ESAS2R_DEFAULT_TMO;
- /*
- * Override the request type to distinguish discovery requests. If we
- * end up deferring the request, esas2r_disc_local_start_request()
- * will be called to restart it.
- */
- rq->req_type = RT_DISC_REQ;
- spin_lock_irqsave(&a->queue_lock, flags);
- if (!(a->flags & (AF_CHPRST_PENDING | AF_FLASHING)))
- esas2r_disc_local_start_request(a, rq);
- else
- list_add_tail(&rq->req_list, &a->defer_list);
- spin_unlock_irqrestore(&a->queue_lock, flags);
- return true;
- }
- void esas2r_disc_local_start_request(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- esas2r_trace_enter();
- list_add_tail(&rq->req_list, &a->active_list);
- esas2r_start_vda_request(a, rq);
- esas2r_trace_exit();
- return;
- }
- static void esas2r_disc_abort(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- esas2r_trace_enter();
- /* abort the current discovery */
- dc->state = DCS_DISC_DONE;
- esas2r_trace_exit();
- }
- static bool esas2r_disc_block_dev_scan(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- bool rslt;
- esas2r_trace_enter();
- esas2r_rq_init_request(rq, a);
- esas2r_build_mgt_req(a,
- rq,
- VDAMGT_DEV_SCAN,
- 0,
- 0,
- 0,
- NULL);
- rq->comp_cb = esas2r_disc_block_dev_scan_cb;
- rq->timeout = 30000;
- rq->interrupt_cx = dc;
- rslt = esas2r_disc_start_request(a, rq);
- esas2r_trace_exit();
- return rslt;
- }
- static void esas2r_disc_block_dev_scan_cb(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- unsigned long flags;
- esas2r_trace_enter();
- spin_lock_irqsave(&a->mem_lock, flags);
- if (rq->req_stat == RS_SUCCESS)
- dc->scan_gen = rq->func_rsp.mgt_rsp.scan_generation;
- dc->state = DCS_RAID_GRP_INFO;
- dc->raid_grp_ix = 0;
- esas2r_rq_destroy_request(rq, a);
- /* continue discovery if it's interrupt driven */
- if (!(dc->flags & DCF_POLLED))
- esas2r_disc_continue(a, rq);
- spin_unlock_irqrestore(&a->mem_lock, flags);
- esas2r_trace_exit();
- }
- static bool esas2r_disc_raid_grp_info(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- bool rslt;
- struct atto_vda_grp_info *grpinfo;
- esas2r_trace_enter();
- esas2r_trace("raid_group_idx: %d", dc->raid_grp_ix);
- if (dc->raid_grp_ix >= VDA_MAX_RAID_GROUPS) {
- dc->state = DCS_DISC_DONE;
- esas2r_trace_exit();
- return false;
- }
- esas2r_rq_init_request(rq, a);
- grpinfo = &rq->vda_rsp_data->mgt_data.data.grp_info;
- memset(grpinfo, 0, sizeof(struct atto_vda_grp_info));
- esas2r_build_mgt_req(a,
- rq,
- VDAMGT_GRP_INFO,
- dc->scan_gen,
- 0,
- sizeof(struct atto_vda_grp_info),
- NULL);
- grpinfo->grp_index = dc->raid_grp_ix;
- rq->comp_cb = esas2r_disc_raid_grp_info_cb;
- rq->interrupt_cx = dc;
- rslt = esas2r_disc_start_request(a, rq);
- esas2r_trace_exit();
- return rslt;
- }
- static void esas2r_disc_raid_grp_info_cb(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- unsigned long flags;
- struct atto_vda_grp_info *grpinfo;
- esas2r_trace_enter();
- spin_lock_irqsave(&a->mem_lock, flags);
- if (rq->req_stat == RS_SCAN_GEN) {
- dc->scan_gen = rq->func_rsp.mgt_rsp.scan_generation;
- dc->raid_grp_ix = 0;
- goto done;
- }
- if (rq->req_stat == RS_SUCCESS) {
- grpinfo = &rq->vda_rsp_data->mgt_data.data.grp_info;
- if (grpinfo->status != VDA_GRP_STAT_ONLINE
- && grpinfo->status != VDA_GRP_STAT_DEGRADED) {
- /* go to the next group. */
- dc->raid_grp_ix++;
- } else {
- memcpy(&dc->raid_grp_name[0],
- &grpinfo->grp_name[0],
- sizeof(grpinfo->grp_name));
- dc->interleave = le32_to_cpu(grpinfo->interleave);
- dc->block_size = le32_to_cpu(grpinfo->block_size);
- dc->state = DCS_PART_INFO;
- dc->part_num = 0;
- }
- } else {
- if (!(rq->req_stat == RS_GRP_INVALID)) {
- esas2r_log(ESAS2R_LOG_WARN,
- "A request for RAID group info failed - "
- "returned with %x",
- rq->req_stat);
- }
- dc->dev_ix = 0;
- dc->state = DCS_PT_DEV_INFO;
- }
- done:
- esas2r_rq_destroy_request(rq, a);
- /* continue discovery if it's interrupt driven */
- if (!(dc->flags & DCF_POLLED))
- esas2r_disc_continue(a, rq);
- spin_unlock_irqrestore(&a->mem_lock, flags);
- esas2r_trace_exit();
- }
- static bool esas2r_disc_part_info(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- bool rslt;
- struct atto_vdapart_info *partinfo;
- esas2r_trace_enter();
- esas2r_trace("part_num: %d", dc->part_num);
- if (dc->part_num >= VDA_MAX_PARTITIONS) {
- dc->state = DCS_RAID_GRP_INFO;
- dc->raid_grp_ix++;
- esas2r_trace_exit();
- return false;
- }
- esas2r_rq_init_request(rq, a);
- partinfo = &rq->vda_rsp_data->mgt_data.data.part_info;
- memset(partinfo, 0, sizeof(struct atto_vdapart_info));
- esas2r_build_mgt_req(a,
- rq,
- VDAMGT_PART_INFO,
- dc->scan_gen,
- 0,
- sizeof(struct atto_vdapart_info),
- NULL);
- partinfo->part_no = dc->part_num;
- memcpy(&partinfo->grp_name[0],
- &dc->raid_grp_name[0],
- sizeof(partinfo->grp_name));
- rq->comp_cb = esas2r_disc_part_info_cb;
- rq->interrupt_cx = dc;
- rslt = esas2r_disc_start_request(a, rq);
- esas2r_trace_exit();
- return rslt;
- }
- static void esas2r_disc_part_info_cb(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- unsigned long flags;
- struct atto_vdapart_info *partinfo;
- esas2r_trace_enter();
- spin_lock_irqsave(&a->mem_lock, flags);
- if (rq->req_stat == RS_SCAN_GEN) {
- dc->scan_gen = rq->func_rsp.mgt_rsp.scan_generation;
- dc->raid_grp_ix = 0;
- dc->state = DCS_RAID_GRP_INFO;
- } else if (rq->req_stat == RS_SUCCESS) {
- partinfo = &rq->vda_rsp_data->mgt_data.data.part_info;
- dc->part_num = partinfo->part_no;
- dc->curr_virt_id = le16_to_cpu(partinfo->target_id);
- esas2r_targ_db_add_raid(a, dc);
- dc->part_num++;
- } else {
- if (!(rq->req_stat == RS_PART_LAST)) {
- esas2r_log(ESAS2R_LOG_WARN,
- "A request for RAID group partition info "
- "failed - status:%d", rq->req_stat);
- }
- dc->state = DCS_RAID_GRP_INFO;
- dc->raid_grp_ix++;
- }
- esas2r_rq_destroy_request(rq, a);
- /* continue discovery if it's interrupt driven */
- if (!(dc->flags & DCF_POLLED))
- esas2r_disc_continue(a, rq);
- spin_unlock_irqrestore(&a->mem_lock, flags);
- esas2r_trace_exit();
- }
- static bool esas2r_disc_passthru_dev_info(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- bool rslt;
- struct atto_vda_devinfo *devinfo;
- esas2r_trace_enter();
- esas2r_trace("dev_ix: %d", dc->dev_ix);
- esas2r_rq_init_request(rq, a);
- devinfo = &rq->vda_rsp_data->mgt_data.data.dev_info;
- memset(devinfo, 0, sizeof(struct atto_vda_devinfo));
- esas2r_build_mgt_req(a,
- rq,
- VDAMGT_DEV_PT_INFO,
- dc->scan_gen,
- dc->dev_ix,
- sizeof(struct atto_vda_devinfo),
- NULL);
- rq->comp_cb = esas2r_disc_passthru_dev_info_cb;
- rq->interrupt_cx = dc;
- rslt = esas2r_disc_start_request(a, rq);
- esas2r_trace_exit();
- return rslt;
- }
- static void esas2r_disc_passthru_dev_info_cb(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- unsigned long flags;
- struct atto_vda_devinfo *devinfo;
- esas2r_trace_enter();
- spin_lock_irqsave(&a->mem_lock, flags);
- if (rq->req_stat == RS_SCAN_GEN) {
- dc->scan_gen = rq->func_rsp.mgt_rsp.scan_generation;
- dc->dev_ix = 0;
- dc->state = DCS_PT_DEV_INFO;
- } else if (rq->req_stat == RS_SUCCESS) {
- devinfo = &rq->vda_rsp_data->mgt_data.data.dev_info;
- dc->dev_ix = le16_to_cpu(rq->func_rsp.mgt_rsp.dev_index);
- dc->curr_virt_id = le16_to_cpu(devinfo->target_id);
- if (le16_to_cpu(devinfo->features) & VDADEVFEAT_PHYS_ID) {
- dc->curr_phys_id =
- le16_to_cpu(devinfo->phys_target_id);
- dc->dev_addr_type = ATTO_GDA_AT_PORT;
- dc->state = DCS_PT_DEV_ADDR;
- esas2r_trace("curr_virt_id: %d", dc->curr_virt_id);
- esas2r_trace("curr_phys_id: %d", dc->curr_phys_id);
- } else {
- dc->dev_ix++;
- }
- } else {
- if (!(rq->req_stat == RS_DEV_INVALID)) {
- esas2r_log(ESAS2R_LOG_WARN,
- "A request for device information failed - "
- "status:%d", rq->req_stat);
- }
- dc->state = DCS_DISC_DONE;
- }
- esas2r_rq_destroy_request(rq, a);
- /* continue discovery if it's interrupt driven */
- if (!(dc->flags & DCF_POLLED))
- esas2r_disc_continue(a, rq);
- spin_unlock_irqrestore(&a->mem_lock, flags);
- esas2r_trace_exit();
- }
- static bool esas2r_disc_passthru_dev_addr(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- bool rslt;
- struct atto_ioctl *hi;
- struct esas2r_sg_context sgc;
- esas2r_trace_enter();
- esas2r_rq_init_request(rq, a);
- /* format the request. */
- sgc.cur_offset = NULL;
- sgc.get_phys_addr = (PGETPHYSADDR)esas2r_disc_get_phys_addr;
- sgc.length = offsetof(struct atto_ioctl, data)
- + sizeof(struct atto_hba_get_device_address);
- esas2r_sgc_init(&sgc, a, rq, rq->vrq->ioctl.sge);
- esas2r_build_ioctl_req(a, rq, sgc.length, VDA_IOCTL_HBA);
- if (!esas2r_build_sg_list(a, rq, &sgc)) {
- esas2r_rq_destroy_request(rq, a);
- esas2r_trace_exit();
- return false;
- }
- rq->comp_cb = esas2r_disc_passthru_dev_addr_cb;
- rq->interrupt_cx = dc;
- /* format the IOCTL data. */
- hi = (struct atto_ioctl *)a->disc_buffer;
- memset(a->disc_buffer, 0, ESAS2R_DISC_BUF_LEN);
- hi->version = ATTO_VER_GET_DEV_ADDR0;
- hi->function = ATTO_FUNC_GET_DEV_ADDR;
- hi->flags = HBAF_TUNNEL;
- hi->data.get_dev_addr.target_id = le32_to_cpu(dc->curr_phys_id);
- hi->data.get_dev_addr.addr_type = dc->dev_addr_type;
- /* start it up. */
- rslt = esas2r_disc_start_request(a, rq);
- esas2r_trace_exit();
- return rslt;
- }
- static void esas2r_disc_passthru_dev_addr_cb(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- struct esas2r_target *t = NULL;
- unsigned long flags;
- struct atto_ioctl *hi;
- u16 addrlen;
- esas2r_trace_enter();
- spin_lock_irqsave(&a->mem_lock, flags);
- hi = (struct atto_ioctl *)a->disc_buffer;
- if (rq->req_stat == RS_SUCCESS
- && hi->status == ATTO_STS_SUCCESS) {
- addrlen = le16_to_cpu(hi->data.get_dev_addr.addr_len);
- if (dc->dev_addr_type == ATTO_GDA_AT_PORT) {
- if (addrlen == sizeof(u64))
- memcpy(&dc->sas_addr,
- &hi->data.get_dev_addr.address[0],
- addrlen);
- else
- memset(&dc->sas_addr, 0, sizeof(dc->sas_addr));
- /* Get the unique identifier. */
- dc->dev_addr_type = ATTO_GDA_AT_UNIQUE;
- goto next_dev_addr;
- } else {
- /* Add the pass through target. */
- if (HIBYTE(addrlen) == 0) {
- t = esas2r_targ_db_add_pthru(a,
- dc,
- &hi->data.
- get_dev_addr.
- address[0],
- (u8)hi->data.
- get_dev_addr.
- addr_len);
- if (t)
- memcpy(&t->sas_addr, &dc->sas_addr,
- sizeof(t->sas_addr));
- } else {
- /* getting the back end data failed */
- esas2r_log(ESAS2R_LOG_WARN,
- "an error occurred retrieving the "
- "back end data (%s:%d)",
- __func__,
- __LINE__);
- }
- }
- } else {
- /* getting the back end data failed */
- esas2r_log(ESAS2R_LOG_WARN,
- "an error occurred retrieving the back end data - "
- "rq->req_stat:%d hi->status:%d",
- rq->req_stat, hi->status);
- }
- /* proceed to the next device. */
- if (dc->flags & DCF_DEV_SCAN) {
- dc->dev_ix++;
- dc->state = DCS_PT_DEV_INFO;
- } else if (dc->flags & DCF_DEV_CHANGE) {
- dc->curr_targ++;
- dc->state = DCS_DEV_ADD;
- } else {
- esas2r_bugon();
- }
- next_dev_addr:
- esas2r_rq_destroy_request(rq, a);
- /* continue discovery if it's interrupt driven */
- if (!(dc->flags & DCF_POLLED))
- esas2r_disc_continue(a, rq);
- spin_unlock_irqrestore(&a->mem_lock, flags);
- esas2r_trace_exit();
- }
- static u32 esas2r_disc_get_phys_addr(struct esas2r_sg_context *sgc, u64 *addr)
- {
- struct esas2r_adapter *a = sgc->adapter;
- if (sgc->length > ESAS2R_DISC_BUF_LEN)
- esas2r_bugon();
- *addr = a->uncached_phys
- + (u64)((u8 *)a->disc_buffer - a->uncached);
- return sgc->length;
- }
- static bool esas2r_disc_dev_remove(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- struct esas2r_target *t;
- struct esas2r_target *t2;
- esas2r_trace_enter();
- /* process removals. */
- for (t = a->targetdb; t < a->targetdb_end; t++) {
- if (t->new_target_state != TS_NOT_PRESENT)
- continue;
- t->new_target_state = TS_INVALID;
- /* remove the right target! */
- t2 =
- esas2r_targ_db_find_by_virt_id(a,
- esas2r_targ_get_id(t,
- a));
- if (t2)
- esas2r_targ_db_remove(a, t2);
- }
- /* removals complete. process arrivals. */
- dc->state = DCS_DEV_ADD;
- dc->curr_targ = a->targetdb;
- esas2r_trace_exit();
- return false;
- }
- static bool esas2r_disc_dev_add(struct esas2r_adapter *a,
- struct esas2r_request *rq)
- {
- struct esas2r_disc_context *dc =
- (struct esas2r_disc_context *)rq->interrupt_cx;
- struct esas2r_target *t = dc->curr_targ;
- if (t >= a->targetdb_end) {
- /* done processing state changes. */
- dc->state = DCS_DISC_DONE;
- } else if (t->new_target_state == TS_PRESENT) {
- struct atto_vda_ae_lu *luevt = &t->lu_event;
- esas2r_trace_enter();
- /* clear this now in case more events come in. */
- t->new_target_state = TS_INVALID;
- /* setup the discovery context for adding this device. */
- dc->curr_virt_id = esas2r_targ_get_id(t, a);
- if ((luevt->hdr.bylength >= offsetof(struct atto_vda_ae_lu, id)
- + sizeof(struct atto_vda_ae_lu_tgt_lun_raid))
- && !(luevt->dwevent & VDAAE_LU_PASSTHROUGH)) {
- dc->block_size = luevt->id.tgtlun_raid.dwblock_size;
- dc->interleave = luevt->id.tgtlun_raid.dwinterleave;
- } else {
- dc->block_size = 0;
- dc->interleave = 0;
- }
- /* determine the device type being added. */
- if (luevt->dwevent & VDAAE_LU_PASSTHROUGH) {
- if (luevt->dwevent & VDAAE_LU_PHYS_ID) {
- dc->state = DCS_PT_DEV_ADDR;
- dc->dev_addr_type = ATTO_GDA_AT_PORT;
- dc->curr_phys_id = luevt->wphys_target_id;
- } else {
- esas2r_log(ESAS2R_LOG_WARN,
- "luevt->dwevent does not have the "
- "VDAAE_LU_PHYS_ID bit set (%s:%d)",
- __func__, __LINE__);
- }
- } else {
- dc->raid_grp_name[0] = 0;
- esas2r_targ_db_add_raid(a, dc);
- }
- esas2r_trace("curr_virt_id: %d", dc->curr_virt_id);
- esas2r_trace("curr_phys_id: %d", dc->curr_phys_id);
- esas2r_trace("dwevent: %d", luevt->dwevent);
- esas2r_trace_exit();
- }
- if (dc->state == DCS_DEV_ADD) {
- /* go to the next device. */
- dc->curr_targ++;
- }
- return false;
- }
- /*
- * When discovery is done, find all requests on defer queue and
- * test if they need to be modified. If a target is no longer present
- * then complete the request with RS_SEL. Otherwise, update the
- * target_id since after a hibernate it can be a different value.
- * VDA does not make passthrough target IDs persistent.
- */
- static void esas2r_disc_fix_curr_requests(struct esas2r_adapter *a)
- {
- unsigned long flags;
- struct esas2r_target *t;
- struct esas2r_request *rq;
- struct list_head *element;
- /* update virt_targ_id in any outstanding esas2r_requests */
- spin_lock_irqsave(&a->queue_lock, flags);
- list_for_each(element, &a->defer_list) {
- rq = list_entry(element, struct esas2r_request, req_list);
- if (rq->vrq->scsi.function == VDA_FUNC_SCSI) {
- t = a->targetdb + rq->target_id;
- if (t->target_state == TS_PRESENT)
- rq->vrq->scsi.target_id = le16_to_cpu(
- t->virt_targ_id);
- else
- rq->req_stat = RS_SEL;
- }
- }
- spin_unlock_irqrestore(&a->queue_lock, flags);
- }
|