|
@@ -1,38 +1,41 @@
|
|
/*
|
|
/*
|
|
- mv_init.c - Marvell 88SE6440 SAS/SATA init support
|
|
|
|
|
|
+ * Marvell 88SE64xx/88SE94xx pci init
|
|
|
|
+ *
|
|
|
|
+ * Copyright 2007 Red Hat, Inc.
|
|
|
|
+ * Copyright 2008 Marvell. <kewei@marvell.com>
|
|
|
|
+ *
|
|
|
|
+ * This file is licensed under GPLv2.
|
|
|
|
+ *
|
|
|
|
+ * This program is free software; you can redistribute it and/or
|
|
|
|
+ * modify it under the terms of the GNU General Public License as
|
|
|
|
+ * published by the Free Software Foundation; version 2 of the
|
|
|
|
+ * License.
|
|
|
|
+ *
|
|
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
+ * General Public License for more details.
|
|
|
|
+ *
|
|
|
|
+ * You should have received a copy of the GNU General Public License
|
|
|
|
+ * along with this program; if not, write to the Free Software
|
|
|
|
+ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307
|
|
|
|
+ * USA
|
|
|
|
+*/
|
|
|
|
|
|
- Copyright 2007 Red Hat, Inc.
|
|
|
|
- Copyright 2008 Marvell. <kewei@marvell.com>
|
|
|
|
-
|
|
|
|
- This program is free software; you can redistribute it and/or
|
|
|
|
- modify it under the terms of the GNU General Public License as
|
|
|
|
- published by the Free Software Foundation; either version 2,
|
|
|
|
- or (at your option) any later version.
|
|
|
|
-
|
|
|
|
- This program is distributed in the hope that it will be useful,
|
|
|
|
- but WITHOUT ANY WARRANTY; without even the implied warranty
|
|
|
|
- of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
|
|
|
|
- See the GNU General Public License for more details.
|
|
|
|
-
|
|
|
|
- You should have received a copy of the GNU General Public
|
|
|
|
- License along with this program; see the file COPYING. If not,
|
|
|
|
- write to the Free Software Foundation, 675 Mass Ave, Cambridge,
|
|
|
|
- MA 02139, USA.
|
|
|
|
-
|
|
|
|
- */
|
|
|
|
|
|
|
|
#include "mv_sas.h"
|
|
#include "mv_sas.h"
|
|
-#include "mv_64xx.h"
|
|
|
|
-#include "mv_chips.h"
|
|
|
|
|
|
|
|
static struct scsi_transport_template *mvs_stt;
|
|
static struct scsi_transport_template *mvs_stt;
|
|
-
|
|
|
|
static const struct mvs_chip_info mvs_chips[] = {
|
|
static const struct mvs_chip_info mvs_chips[] = {
|
|
- [chip_6320] = { 2, 16, 9 },
|
|
|
|
- [chip_6440] = { 4, 16, 9 },
|
|
|
|
- [chip_6480] = { 8, 32, 10 },
|
|
|
|
|
|
+ [chip_6320] = { 1, 2, 0x400, 17, 16, 9, &mvs_64xx_dispatch, },
|
|
|
|
+ [chip_6440] = { 1, 4, 0x400, 17, 16, 9, &mvs_64xx_dispatch, },
|
|
|
|
+ [chip_6485] = { 1, 8, 0x800, 33, 32, 10, &mvs_64xx_dispatch, },
|
|
|
|
+ [chip_9180] = { 2, 4, 0x800, 17, 64, 9, &mvs_94xx_dispatch, },
|
|
|
|
+ [chip_9480] = { 2, 4, 0x800, 17, 64, 9, &mvs_94xx_dispatch, },
|
|
};
|
|
};
|
|
|
|
|
|
|
|
+#define SOC_SAS_NUM 2
|
|
|
|
+
|
|
static struct scsi_host_template mvs_sht = {
|
|
static struct scsi_host_template mvs_sht = {
|
|
.module = THIS_MODULE,
|
|
.module = THIS_MODULE,
|
|
.name = DRV_NAME,
|
|
.name = DRV_NAME,
|
|
@@ -53,17 +56,29 @@ static struct scsi_host_template mvs_sht = {
|
|
.use_clustering = ENABLE_CLUSTERING,
|
|
.use_clustering = ENABLE_CLUSTERING,
|
|
.eh_device_reset_handler = sas_eh_device_reset_handler,
|
|
.eh_device_reset_handler = sas_eh_device_reset_handler,
|
|
.eh_bus_reset_handler = sas_eh_bus_reset_handler,
|
|
.eh_bus_reset_handler = sas_eh_bus_reset_handler,
|
|
- .slave_alloc = sas_slave_alloc,
|
|
|
|
|
|
+ .slave_alloc = mvs_slave_alloc,
|
|
.target_destroy = sas_target_destroy,
|
|
.target_destroy = sas_target_destroy,
|
|
.ioctl = sas_ioctl,
|
|
.ioctl = sas_ioctl,
|
|
};
|
|
};
|
|
|
|
|
|
static struct sas_domain_function_template mvs_transport_ops = {
|
|
static struct sas_domain_function_template mvs_transport_ops = {
|
|
- .lldd_execute_task = mvs_task_exec,
|
|
|
|
|
|
+ .lldd_dev_found = mvs_dev_found,
|
|
|
|
+ .lldd_dev_gone = mvs_dev_gone,
|
|
|
|
+
|
|
|
|
+ .lldd_execute_task = mvs_queue_command,
|
|
.lldd_control_phy = mvs_phy_control,
|
|
.lldd_control_phy = mvs_phy_control,
|
|
- .lldd_abort_task = mvs_task_abort,
|
|
|
|
- .lldd_port_formed = mvs_port_formed,
|
|
|
|
|
|
+
|
|
|
|
+ .lldd_abort_task = mvs_abort_task,
|
|
|
|
+ .lldd_abort_task_set = mvs_abort_task_set,
|
|
|
|
+ .lldd_clear_aca = mvs_clear_aca,
|
|
|
|
+ .lldd_clear_task_set = mvs_clear_task_set,
|
|
.lldd_I_T_nexus_reset = mvs_I_T_nexus_reset,
|
|
.lldd_I_T_nexus_reset = mvs_I_T_nexus_reset,
|
|
|
|
+ .lldd_lu_reset = mvs_lu_reset,
|
|
|
|
+ .lldd_query_task = mvs_query_task,
|
|
|
|
+
|
|
|
|
+ .lldd_port_formed = mvs_port_formed,
|
|
|
|
+ .lldd_port_deformed = mvs_port_deformed,
|
|
|
|
+
|
|
};
|
|
};
|
|
|
|
|
|
static void __devinit mvs_phy_init(struct mvs_info *mvi, int phy_id)
|
|
static void __devinit mvs_phy_init(struct mvs_info *mvi, int phy_id)
|
|
@@ -71,6 +86,8 @@ static void __devinit mvs_phy_init(struct mvs_info *mvi, int phy_id)
|
|
struct mvs_phy *phy = &mvi->phy[phy_id];
|
|
struct mvs_phy *phy = &mvi->phy[phy_id];
|
|
struct asd_sas_phy *sas_phy = &phy->sas_phy;
|
|
struct asd_sas_phy *sas_phy = &phy->sas_phy;
|
|
|
|
|
|
|
|
+ phy->mvi = mvi;
|
|
|
|
+ init_timer(&phy->timer);
|
|
sas_phy->enabled = (phy_id < mvi->chip->n_phy) ? 1 : 0;
|
|
sas_phy->enabled = (phy_id < mvi->chip->n_phy) ? 1 : 0;
|
|
sas_phy->class = SAS;
|
|
sas_phy->class = SAS;
|
|
sas_phy->iproto = SAS_PROTOCOL_ALL;
|
|
sas_phy->iproto = SAS_PROTOCOL_ALL;
|
|
@@ -83,248 +100,283 @@ static void __devinit mvs_phy_init(struct mvs_info *mvi, int phy_id)
|
|
sas_phy->id = phy_id;
|
|
sas_phy->id = phy_id;
|
|
sas_phy->sas_addr = &mvi->sas_addr[0];
|
|
sas_phy->sas_addr = &mvi->sas_addr[0];
|
|
sas_phy->frame_rcvd = &phy->frame_rcvd[0];
|
|
sas_phy->frame_rcvd = &phy->frame_rcvd[0];
|
|
- sas_phy->ha = &mvi->sas;
|
|
|
|
|
|
+ sas_phy->ha = (struct sas_ha_struct *)mvi->shost->hostdata;
|
|
sas_phy->lldd_phy = phy;
|
|
sas_phy->lldd_phy = phy;
|
|
}
|
|
}
|
|
|
|
|
|
static void mvs_free(struct mvs_info *mvi)
|
|
static void mvs_free(struct mvs_info *mvi)
|
|
{
|
|
{
|
|
int i;
|
|
int i;
|
|
|
|
+ struct mvs_wq *mwq;
|
|
|
|
+ int slot_nr;
|
|
|
|
|
|
if (!mvi)
|
|
if (!mvi)
|
|
return;
|
|
return;
|
|
|
|
|
|
- for (i = 0; i < MVS_SLOTS; i++) {
|
|
|
|
- struct mvs_slot_info *slot = &mvi->slot_info[i];
|
|
|
|
|
|
+ if (mvi->flags & MVF_FLAG_SOC)
|
|
|
|
+ slot_nr = MVS_SOC_SLOTS;
|
|
|
|
+ else
|
|
|
|
+ slot_nr = MVS_SLOTS;
|
|
|
|
|
|
|
|
+ for (i = 0; i < mvi->tags_num; i++) {
|
|
|
|
+ struct mvs_slot_info *slot = &mvi->slot_info[i];
|
|
if (slot->buf)
|
|
if (slot->buf)
|
|
- dma_free_coherent(&mvi->pdev->dev, MVS_SLOT_BUF_SZ,
|
|
|
|
|
|
+ dma_free_coherent(mvi->dev, MVS_SLOT_BUF_SZ,
|
|
slot->buf, slot->buf_dma);
|
|
slot->buf, slot->buf_dma);
|
|
}
|
|
}
|
|
|
|
|
|
if (mvi->tx)
|
|
if (mvi->tx)
|
|
- dma_free_coherent(&mvi->pdev->dev,
|
|
|
|
|
|
+ dma_free_coherent(mvi->dev,
|
|
sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ,
|
|
sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ,
|
|
mvi->tx, mvi->tx_dma);
|
|
mvi->tx, mvi->tx_dma);
|
|
if (mvi->rx_fis)
|
|
if (mvi->rx_fis)
|
|
- dma_free_coherent(&mvi->pdev->dev, MVS_RX_FISL_SZ,
|
|
|
|
|
|
+ dma_free_coherent(mvi->dev, MVS_RX_FISL_SZ,
|
|
mvi->rx_fis, mvi->rx_fis_dma);
|
|
mvi->rx_fis, mvi->rx_fis_dma);
|
|
if (mvi->rx)
|
|
if (mvi->rx)
|
|
- dma_free_coherent(&mvi->pdev->dev,
|
|
|
|
|
|
+ dma_free_coherent(mvi->dev,
|
|
sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1),
|
|
sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1),
|
|
mvi->rx, mvi->rx_dma);
|
|
mvi->rx, mvi->rx_dma);
|
|
if (mvi->slot)
|
|
if (mvi->slot)
|
|
- dma_free_coherent(&mvi->pdev->dev,
|
|
|
|
- sizeof(*mvi->slot) * MVS_SLOTS,
|
|
|
|
|
|
+ dma_free_coherent(mvi->dev,
|
|
|
|
+ sizeof(*mvi->slot) * slot_nr,
|
|
mvi->slot, mvi->slot_dma);
|
|
mvi->slot, mvi->slot_dma);
|
|
-#ifdef MVS_ENABLE_PERI
|
|
|
|
- if (mvi->peri_regs)
|
|
|
|
- iounmap(mvi->peri_regs);
|
|
|
|
|
|
+#ifndef DISABLE_HOTPLUG_DMA_FIX
|
|
|
|
+ if (mvi->bulk_buffer)
|
|
|
|
+ dma_free_coherent(mvi->dev, TRASH_BUCKET_SIZE,
|
|
|
|
+ mvi->bulk_buffer, mvi->bulk_buffer_dma);
|
|
#endif
|
|
#endif
|
|
- if (mvi->regs)
|
|
|
|
- iounmap(mvi->regs);
|
|
|
|
|
|
+
|
|
|
|
+ MVS_CHIP_DISP->chip_iounmap(mvi);
|
|
if (mvi->shost)
|
|
if (mvi->shost)
|
|
scsi_host_put(mvi->shost);
|
|
scsi_host_put(mvi->shost);
|
|
- kfree(mvi->sas.sas_port);
|
|
|
|
- kfree(mvi->sas.sas_phy);
|
|
|
|
|
|
+ list_for_each_entry(mwq, &mvi->wq_list, entry)
|
|
|
|
+ cancel_delayed_work(&mwq->work_q);
|
|
kfree(mvi);
|
|
kfree(mvi);
|
|
}
|
|
}
|
|
|
|
|
|
#ifdef MVS_USE_TASKLET
|
|
#ifdef MVS_USE_TASKLET
|
|
-static void mvs_tasklet(unsigned long data)
|
|
|
|
|
|
+struct tasklet_struct mv_tasklet;
|
|
|
|
+static void mvs_tasklet(unsigned long opaque)
|
|
{
|
|
{
|
|
- struct mvs_info *mvi = (struct mvs_info *) data;
|
|
|
|
unsigned long flags;
|
|
unsigned long flags;
|
|
|
|
+ u32 stat;
|
|
|
|
+ u16 core_nr, i = 0;
|
|
|
|
|
|
- spin_lock_irqsave(&mvi->lock, flags);
|
|
|
|
|
|
+ struct mvs_info *mvi;
|
|
|
|
+ struct sas_ha_struct *sha = (struct sas_ha_struct *)opaque;
|
|
|
|
+
|
|
|
|
+ core_nr = ((struct mvs_prv_info *)sha->lldd_ha)->n_host;
|
|
|
|
+ mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[0];
|
|
|
|
+
|
|
|
|
+ if (unlikely(!mvi))
|
|
|
|
+ BUG_ON(1);
|
|
|
|
+
|
|
|
|
+ for (i = 0; i < core_nr; i++) {
|
|
|
|
+ mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[i];
|
|
|
|
+ stat = MVS_CHIP_DISP->isr_status(mvi, mvi->irq);
|
|
|
|
+ if (stat)
|
|
|
|
+ MVS_CHIP_DISP->isr(mvi, mvi->irq, stat);
|
|
|
|
+ }
|
|
|
|
|
|
-#ifdef MVS_DISABLE_MSI
|
|
|
|
- mvs_int_full(mvi);
|
|
|
|
-#else
|
|
|
|
- mvs_int_rx(mvi, true);
|
|
|
|
-#endif
|
|
|
|
- spin_unlock_irqrestore(&mvi->lock, flags);
|
|
|
|
}
|
|
}
|
|
#endif
|
|
#endif
|
|
|
|
|
|
static irqreturn_t mvs_interrupt(int irq, void *opaque)
|
|
static irqreturn_t mvs_interrupt(int irq, void *opaque)
|
|
{
|
|
{
|
|
- struct mvs_info *mvi = opaque;
|
|
|
|
- void __iomem *regs = mvi->regs;
|
|
|
|
|
|
+ u32 core_nr, i = 0;
|
|
u32 stat;
|
|
u32 stat;
|
|
|
|
+ struct mvs_info *mvi;
|
|
|
|
+ struct sas_ha_struct *sha = opaque;
|
|
|
|
|
|
- stat = mr32(GBL_INT_STAT);
|
|
|
|
|
|
+ core_nr = ((struct mvs_prv_info *)sha->lldd_ha)->n_host;
|
|
|
|
+ mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[0];
|
|
|
|
|
|
- if (stat == 0 || stat == 0xffffffff)
|
|
|
|
|
|
+ if (unlikely(!mvi))
|
|
return IRQ_NONE;
|
|
return IRQ_NONE;
|
|
|
|
|
|
- /* clear CMD_CMPLT ASAP */
|
|
|
|
- mw32_f(INT_STAT, CINT_DONE);
|
|
|
|
-
|
|
|
|
-#ifndef MVS_USE_TASKLET
|
|
|
|
- spin_lock(&mvi->lock);
|
|
|
|
-
|
|
|
|
- mvs_int_full(mvi);
|
|
|
|
|
|
+ stat = MVS_CHIP_DISP->isr_status(mvi, irq);
|
|
|
|
+ if (!stat)
|
|
|
|
+ return IRQ_NONE;
|
|
|
|
|
|
- spin_unlock(&mvi->lock);
|
|
|
|
|
|
+#ifdef MVS_USE_TASKLET
|
|
|
|
+ tasklet_schedule(&mv_tasklet);
|
|
#else
|
|
#else
|
|
- tasklet_schedule(&mvi->tasklet);
|
|
|
|
|
|
+ for (i = 0; i < core_nr; i++) {
|
|
|
|
+ mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[i];
|
|
|
|
+ MVS_CHIP_DISP->isr(mvi, irq, stat);
|
|
|
|
+ }
|
|
#endif
|
|
#endif
|
|
return IRQ_HANDLED;
|
|
return IRQ_HANDLED;
|
|
}
|
|
}
|
|
|
|
|
|
-static struct mvs_info *__devinit mvs_alloc(struct pci_dev *pdev,
|
|
|
|
- const struct pci_device_id *ent)
|
|
|
|
|
|
+static int __devinit mvs_alloc(struct mvs_info *mvi, struct Scsi_Host *shost)
|
|
{
|
|
{
|
|
- struct mvs_info *mvi;
|
|
|
|
- unsigned long res_start, res_len, res_flag;
|
|
|
|
- struct asd_sas_phy **arr_phy;
|
|
|
|
- struct asd_sas_port **arr_port;
|
|
|
|
- const struct mvs_chip_info *chip = &mvs_chips[ent->driver_data];
|
|
|
|
- int i;
|
|
|
|
|
|
+ int i, slot_nr;
|
|
|
|
|
|
- /*
|
|
|
|
- * alloc and init our per-HBA mvs_info struct
|
|
|
|
- */
|
|
|
|
-
|
|
|
|
- mvi = kzalloc(sizeof(*mvi), GFP_KERNEL);
|
|
|
|
- if (!mvi)
|
|
|
|
- return NULL;
|
|
|
|
|
|
+ if (mvi->flags & MVF_FLAG_SOC)
|
|
|
|
+ slot_nr = MVS_SOC_SLOTS;
|
|
|
|
+ else
|
|
|
|
+ slot_nr = MVS_SLOTS;
|
|
|
|
|
|
spin_lock_init(&mvi->lock);
|
|
spin_lock_init(&mvi->lock);
|
|
-#ifdef MVS_USE_TASKLET
|
|
|
|
- tasklet_init(&mvi->tasklet, mvs_tasklet, (unsigned long)mvi);
|
|
|
|
-#endif
|
|
|
|
- mvi->pdev = pdev;
|
|
|
|
- mvi->chip = chip;
|
|
|
|
-
|
|
|
|
- if (pdev->device == 0x6440 && pdev->revision == 0)
|
|
|
|
- mvi->flags |= MVF_PHY_PWR_FIX;
|
|
|
|
-
|
|
|
|
- /*
|
|
|
|
- * alloc and init SCSI, SAS glue
|
|
|
|
- */
|
|
|
|
-
|
|
|
|
- mvi->shost = scsi_host_alloc(&mvs_sht, sizeof(void *));
|
|
|
|
- if (!mvi->shost)
|
|
|
|
- goto err_out;
|
|
|
|
-
|
|
|
|
- arr_phy = kcalloc(MVS_MAX_PHYS, sizeof(void *), GFP_KERNEL);
|
|
|
|
- arr_port = kcalloc(MVS_MAX_PHYS, sizeof(void *), GFP_KERNEL);
|
|
|
|
- if (!arr_phy || !arr_port)
|
|
|
|
- goto err_out;
|
|
|
|
-
|
|
|
|
- for (i = 0; i < MVS_MAX_PHYS; i++) {
|
|
|
|
|
|
+ for (i = 0; i < mvi->chip->n_phy; i++) {
|
|
mvs_phy_init(mvi, i);
|
|
mvs_phy_init(mvi, i);
|
|
- arr_phy[i] = &mvi->phy[i].sas_phy;
|
|
|
|
- arr_port[i] = &mvi->port[i].sas_port;
|
|
|
|
- mvi->port[i].taskfileset = MVS_ID_NOT_MAPPED;
|
|
|
|
mvi->port[i].wide_port_phymap = 0;
|
|
mvi->port[i].wide_port_phymap = 0;
|
|
mvi->port[i].port_attached = 0;
|
|
mvi->port[i].port_attached = 0;
|
|
INIT_LIST_HEAD(&mvi->port[i].list);
|
|
INIT_LIST_HEAD(&mvi->port[i].list);
|
|
}
|
|
}
|
|
-
|
|
|
|
- SHOST_TO_SAS_HA(mvi->shost) = &mvi->sas;
|
|
|
|
- mvi->shost->transportt = mvs_stt;
|
|
|
|
- mvi->shost->max_id = 21;
|
|
|
|
- mvi->shost->max_lun = ~0;
|
|
|
|
- mvi->shost->max_channel = 0;
|
|
|
|
- mvi->shost->max_cmd_len = 16;
|
|
|
|
-
|
|
|
|
- mvi->sas.sas_ha_name = DRV_NAME;
|
|
|
|
- mvi->sas.dev = &pdev->dev;
|
|
|
|
- mvi->sas.lldd_module = THIS_MODULE;
|
|
|
|
- mvi->sas.sas_addr = &mvi->sas_addr[0];
|
|
|
|
- mvi->sas.sas_phy = arr_phy;
|
|
|
|
- mvi->sas.sas_port = arr_port;
|
|
|
|
- mvi->sas.num_phys = chip->n_phy;
|
|
|
|
- mvi->sas.lldd_max_execute_num = 1;
|
|
|
|
- mvi->sas.lldd_queue_size = MVS_QUEUE_SIZE;
|
|
|
|
- mvi->shost->can_queue = MVS_CAN_QUEUE;
|
|
|
|
- mvi->shost->cmd_per_lun = MVS_SLOTS / mvi->sas.num_phys;
|
|
|
|
- mvi->sas.lldd_ha = mvi;
|
|
|
|
- mvi->sas.core.shost = mvi->shost;
|
|
|
|
-
|
|
|
|
- mvs_tag_init(mvi);
|
|
|
|
-
|
|
|
|
- /*
|
|
|
|
- * ioremap main and peripheral registers
|
|
|
|
- */
|
|
|
|
-
|
|
|
|
-#ifdef MVS_ENABLE_PERI
|
|
|
|
- res_start = pci_resource_start(pdev, 2);
|
|
|
|
- res_len = pci_resource_len(pdev, 2);
|
|
|
|
- if (!res_start || !res_len)
|
|
|
|
- goto err_out;
|
|
|
|
-
|
|
|
|
- mvi->peri_regs = ioremap_nocache(res_start, res_len);
|
|
|
|
- if (!mvi->peri_regs)
|
|
|
|
- goto err_out;
|
|
|
|
-#endif
|
|
|
|
-
|
|
|
|
- res_start = pci_resource_start(pdev, 4);
|
|
|
|
- res_len = pci_resource_len(pdev, 4);
|
|
|
|
- if (!res_start || !res_len)
|
|
|
|
- goto err_out;
|
|
|
|
-
|
|
|
|
- res_flag = pci_resource_flags(pdev, 4);
|
|
|
|
- if (res_flag & IORESOURCE_CACHEABLE)
|
|
|
|
- mvi->regs = ioremap(res_start, res_len);
|
|
|
|
- else
|
|
|
|
- mvi->regs = ioremap_nocache(res_start, res_len);
|
|
|
|
-
|
|
|
|
- if (!mvi->regs)
|
|
|
|
- goto err_out;
|
|
|
|
|
|
+ for (i = 0; i < MVS_MAX_DEVICES; i++) {
|
|
|
|
+ mvi->devices[i].taskfileset = MVS_ID_NOT_MAPPED;
|
|
|
|
+ mvi->devices[i].dev_type = NO_DEVICE;
|
|
|
|
+ mvi->devices[i].device_id = i;
|
|
|
|
+ mvi->devices[i].dev_status = MVS_DEV_NORMAL;
|
|
|
|
+ }
|
|
|
|
|
|
/*
|
|
/*
|
|
* alloc and init our DMA areas
|
|
* alloc and init our DMA areas
|
|
*/
|
|
*/
|
|
-
|
|
|
|
- mvi->tx = dma_alloc_coherent(&pdev->dev,
|
|
|
|
|
|
+ mvi->tx = dma_alloc_coherent(mvi->dev,
|
|
sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ,
|
|
sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ,
|
|
&mvi->tx_dma, GFP_KERNEL);
|
|
&mvi->tx_dma, GFP_KERNEL);
|
|
if (!mvi->tx)
|
|
if (!mvi->tx)
|
|
goto err_out;
|
|
goto err_out;
|
|
memset(mvi->tx, 0, sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ);
|
|
memset(mvi->tx, 0, sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ);
|
|
-
|
|
|
|
- mvi->rx_fis = dma_alloc_coherent(&pdev->dev, MVS_RX_FISL_SZ,
|
|
|
|
|
|
+ mvi->rx_fis = dma_alloc_coherent(mvi->dev, MVS_RX_FISL_SZ,
|
|
&mvi->rx_fis_dma, GFP_KERNEL);
|
|
&mvi->rx_fis_dma, GFP_KERNEL);
|
|
if (!mvi->rx_fis)
|
|
if (!mvi->rx_fis)
|
|
goto err_out;
|
|
goto err_out;
|
|
memset(mvi->rx_fis, 0, MVS_RX_FISL_SZ);
|
|
memset(mvi->rx_fis, 0, MVS_RX_FISL_SZ);
|
|
|
|
|
|
- mvi->rx = dma_alloc_coherent(&pdev->dev,
|
|
|
|
|
|
+ mvi->rx = dma_alloc_coherent(mvi->dev,
|
|
sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1),
|
|
sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1),
|
|
&mvi->rx_dma, GFP_KERNEL);
|
|
&mvi->rx_dma, GFP_KERNEL);
|
|
if (!mvi->rx)
|
|
if (!mvi->rx)
|
|
goto err_out;
|
|
goto err_out;
|
|
memset(mvi->rx, 0, sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1));
|
|
memset(mvi->rx, 0, sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1));
|
|
-
|
|
|
|
mvi->rx[0] = cpu_to_le32(0xfff);
|
|
mvi->rx[0] = cpu_to_le32(0xfff);
|
|
mvi->rx_cons = 0xfff;
|
|
mvi->rx_cons = 0xfff;
|
|
|
|
|
|
- mvi->slot = dma_alloc_coherent(&pdev->dev,
|
|
|
|
- sizeof(*mvi->slot) * MVS_SLOTS,
|
|
|
|
|
|
+ mvi->slot = dma_alloc_coherent(mvi->dev,
|
|
|
|
+ sizeof(*mvi->slot) * slot_nr,
|
|
&mvi->slot_dma, GFP_KERNEL);
|
|
&mvi->slot_dma, GFP_KERNEL);
|
|
if (!mvi->slot)
|
|
if (!mvi->slot)
|
|
goto err_out;
|
|
goto err_out;
|
|
- memset(mvi->slot, 0, sizeof(*mvi->slot) * MVS_SLOTS);
|
|
|
|
|
|
+ memset(mvi->slot, 0, sizeof(*mvi->slot) * slot_nr);
|
|
|
|
|
|
- for (i = 0; i < MVS_SLOTS; i++) {
|
|
|
|
|
|
+#ifndef DISABLE_HOTPLUG_DMA_FIX
|
|
|
|
+ mvi->bulk_buffer = dma_alloc_coherent(mvi->dev,
|
|
|
|
+ TRASH_BUCKET_SIZE,
|
|
|
|
+ &mvi->bulk_buffer_dma, GFP_KERNEL);
|
|
|
|
+ if (!mvi->bulk_buffer)
|
|
|
|
+ goto err_out;
|
|
|
|
+#endif
|
|
|
|
+ for (i = 0; i < slot_nr; i++) {
|
|
struct mvs_slot_info *slot = &mvi->slot_info[i];
|
|
struct mvs_slot_info *slot = &mvi->slot_info[i];
|
|
|
|
|
|
- slot->buf = dma_alloc_coherent(&pdev->dev, MVS_SLOT_BUF_SZ,
|
|
|
|
|
|
+ slot->buf = dma_alloc_coherent(mvi->dev, MVS_SLOT_BUF_SZ,
|
|
&slot->buf_dma, GFP_KERNEL);
|
|
&slot->buf_dma, GFP_KERNEL);
|
|
- if (!slot->buf)
|
|
|
|
|
|
+ if (!slot->buf) {
|
|
|
|
+ printk(KERN_DEBUG"failed to allocate slot->buf.\n");
|
|
goto err_out;
|
|
goto err_out;
|
|
|
|
+ }
|
|
memset(slot->buf, 0, MVS_SLOT_BUF_SZ);
|
|
memset(slot->buf, 0, MVS_SLOT_BUF_SZ);
|
|
|
|
+ ++mvi->tags_num;
|
|
}
|
|
}
|
|
|
|
+ /* Initialize tags */
|
|
|
|
+ mvs_tag_init(mvi);
|
|
|
|
+ return 0;
|
|
|
|
+err_out:
|
|
|
|
+ return 1;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
|
|
- /* finally, read NVRAM to get our SAS address */
|
|
|
|
- if (mvs_nvram_read(mvi, NVR_SAS_ADDR, &mvi->sas_addr, 8))
|
|
|
|
|
|
+int mvs_ioremap(struct mvs_info *mvi, int bar, int bar_ex)
|
|
|
|
+{
|
|
|
|
+ unsigned long res_start, res_len, res_flag, res_flag_ex = 0;
|
|
|
|
+ struct pci_dev *pdev = mvi->pdev;
|
|
|
|
+ if (bar_ex != -1) {
|
|
|
|
+ /*
|
|
|
|
+ * ioremap main and peripheral registers
|
|
|
|
+ */
|
|
|
|
+ res_start = pci_resource_start(pdev, bar_ex);
|
|
|
|
+ res_len = pci_resource_len(pdev, bar_ex);
|
|
|
|
+ if (!res_start || !res_len)
|
|
|
|
+ goto err_out;
|
|
|
|
+
|
|
|
|
+ res_flag_ex = pci_resource_flags(pdev, bar_ex);
|
|
|
|
+ if (res_flag_ex & IORESOURCE_MEM) {
|
|
|
|
+ if (res_flag_ex & IORESOURCE_CACHEABLE)
|
|
|
|
+ mvi->regs_ex = ioremap(res_start, res_len);
|
|
|
|
+ else
|
|
|
|
+ mvi->regs_ex = ioremap_nocache(res_start,
|
|
|
|
+ res_len);
|
|
|
|
+ } else
|
|
|
|
+ mvi->regs_ex = (void *)res_start;
|
|
|
|
+ if (!mvi->regs_ex)
|
|
|
|
+ goto err_out;
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ res_start = pci_resource_start(pdev, bar);
|
|
|
|
+ res_len = pci_resource_len(pdev, bar);
|
|
|
|
+ if (!res_start || !res_len)
|
|
|
|
+ goto err_out;
|
|
|
|
+
|
|
|
|
+ res_flag = pci_resource_flags(pdev, bar);
|
|
|
|
+ if (res_flag & IORESOURCE_CACHEABLE)
|
|
|
|
+ mvi->regs = ioremap(res_start, res_len);
|
|
|
|
+ else
|
|
|
|
+ mvi->regs = ioremap_nocache(res_start, res_len);
|
|
|
|
+
|
|
|
|
+ if (!mvi->regs) {
|
|
|
|
+ if (mvi->regs_ex && (res_flag_ex & IORESOURCE_MEM))
|
|
|
|
+ iounmap(mvi->regs_ex);
|
|
|
|
+ mvi->regs_ex = NULL;
|
|
goto err_out;
|
|
goto err_out;
|
|
- return mvi;
|
|
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ return 0;
|
|
|
|
+err_out:
|
|
|
|
+ return -1;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+void mvs_iounmap(void __iomem *regs)
|
|
|
|
+{
|
|
|
|
+ iounmap(regs);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static struct mvs_info *__devinit mvs_pci_alloc(struct pci_dev *pdev,
|
|
|
|
+ const struct pci_device_id *ent,
|
|
|
|
+ struct Scsi_Host *shost, unsigned int id)
|
|
|
|
+{
|
|
|
|
+ struct mvs_info *mvi;
|
|
|
|
+ struct sas_ha_struct *sha = SHOST_TO_SAS_HA(shost);
|
|
|
|
+
|
|
|
|
+ mvi = kzalloc(sizeof(*mvi) + MVS_SLOTS * sizeof(struct mvs_slot_info),
|
|
|
|
+ GFP_KERNEL);
|
|
|
|
+ if (!mvi)
|
|
|
|
+ return NULL;
|
|
|
|
|
|
|
|
+ mvi->pdev = pdev;
|
|
|
|
+ mvi->dev = &pdev->dev;
|
|
|
|
+ mvi->chip_id = ent->driver_data;
|
|
|
|
+ mvi->chip = &mvs_chips[mvi->chip_id];
|
|
|
|
+ INIT_LIST_HEAD(&mvi->wq_list);
|
|
|
|
+ mvi->irq = pdev->irq;
|
|
|
|
+
|
|
|
|
+ ((struct mvs_prv_info *)sha->lldd_ha)->mvi[id] = mvi;
|
|
|
|
+ ((struct mvs_prv_info *)sha->lldd_ha)->n_phy = mvi->chip->n_phy;
|
|
|
|
+
|
|
|
|
+ mvi->id = id;
|
|
|
|
+ mvi->sas = sha;
|
|
|
|
+ mvi->shost = shost;
|
|
|
|
+#ifdef MVS_USE_TASKLET
|
|
|
|
+ tasklet_init(&mv_tasklet, mvs_tasklet, (unsigned long)sha);
|
|
|
|
+#endif
|
|
|
|
+
|
|
|
|
+ if (MVS_CHIP_DISP->chip_ioremap(mvi))
|
|
|
|
+ goto err_out;
|
|
|
|
+ if (!mvs_alloc(mvi, shost))
|
|
|
|
+ return mvi;
|
|
err_out:
|
|
err_out:
|
|
mvs_free(mvi);
|
|
mvs_free(mvi);
|
|
return NULL;
|
|
return NULL;
|
|
@@ -363,16 +415,111 @@ static int pci_go_64(struct pci_dev *pdev)
|
|
return rc;
|
|
return rc;
|
|
}
|
|
}
|
|
|
|
|
|
|
|
+static int __devinit mvs_prep_sas_ha_init(struct Scsi_Host *shost,
|
|
|
|
+ const struct mvs_chip_info *chip_info)
|
|
|
|
+{
|
|
|
|
+ int phy_nr, port_nr; unsigned short core_nr;
|
|
|
|
+ struct asd_sas_phy **arr_phy;
|
|
|
|
+ struct asd_sas_port **arr_port;
|
|
|
|
+ struct sas_ha_struct *sha = SHOST_TO_SAS_HA(shost);
|
|
|
|
+
|
|
|
|
+ core_nr = chip_info->n_host;
|
|
|
|
+ phy_nr = core_nr * chip_info->n_phy;
|
|
|
|
+ port_nr = phy_nr;
|
|
|
|
+
|
|
|
|
+ memset(sha, 0x00, sizeof(struct sas_ha_struct));
|
|
|
|
+ arr_phy = kcalloc(phy_nr, sizeof(void *), GFP_KERNEL);
|
|
|
|
+ arr_port = kcalloc(port_nr, sizeof(void *), GFP_KERNEL);
|
|
|
|
+ if (!arr_phy || !arr_port)
|
|
|
|
+ goto exit_free;
|
|
|
|
+
|
|
|
|
+ sha->sas_phy = arr_phy;
|
|
|
|
+ sha->sas_port = arr_port;
|
|
|
|
+
|
|
|
|
+ sha->lldd_ha = kzalloc(sizeof(struct mvs_prv_info), GFP_KERNEL);
|
|
|
|
+ if (!sha->lldd_ha)
|
|
|
|
+ goto exit_free;
|
|
|
|
+
|
|
|
|
+ ((struct mvs_prv_info *)sha->lldd_ha)->n_host = core_nr;
|
|
|
|
+
|
|
|
|
+ shost->transportt = mvs_stt;
|
|
|
|
+ shost->max_id = 128;
|
|
|
|
+ shost->max_lun = ~0;
|
|
|
|
+ shost->max_channel = 1;
|
|
|
|
+ shost->max_cmd_len = 16;
|
|
|
|
+
|
|
|
|
+ return 0;
|
|
|
|
+exit_free:
|
|
|
|
+ kfree(arr_phy);
|
|
|
|
+ kfree(arr_port);
|
|
|
|
+ return -1;
|
|
|
|
+
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void __devinit mvs_post_sas_ha_init(struct Scsi_Host *shost,
|
|
|
|
+ const struct mvs_chip_info *chip_info)
|
|
|
|
+{
|
|
|
|
+ int can_queue, i = 0, j = 0;
|
|
|
|
+ struct mvs_info *mvi = NULL;
|
|
|
|
+ struct sas_ha_struct *sha = SHOST_TO_SAS_HA(shost);
|
|
|
|
+ unsigned short nr_core = ((struct mvs_prv_info *)sha->lldd_ha)->n_host;
|
|
|
|
+
|
|
|
|
+ for (j = 0; j < nr_core; j++) {
|
|
|
|
+ mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[j];
|
|
|
|
+ for (i = 0; i < chip_info->n_phy; i++) {
|
|
|
|
+ sha->sas_phy[j * chip_info->n_phy + i] =
|
|
|
|
+ &mvi->phy[i].sas_phy;
|
|
|
|
+ sha->sas_port[j * chip_info->n_phy + i] =
|
|
|
|
+ &mvi->port[i].sas_port;
|
|
|
|
+ }
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ sha->sas_ha_name = DRV_NAME;
|
|
|
|
+ sha->dev = mvi->dev;
|
|
|
|
+ sha->lldd_module = THIS_MODULE;
|
|
|
|
+ sha->sas_addr = &mvi->sas_addr[0];
|
|
|
|
+
|
|
|
|
+ sha->num_phys = nr_core * chip_info->n_phy;
|
|
|
|
+
|
|
|
|
+ sha->lldd_max_execute_num = 1;
|
|
|
|
+
|
|
|
|
+ if (mvi->flags & MVF_FLAG_SOC)
|
|
|
|
+ can_queue = MVS_SOC_CAN_QUEUE;
|
|
|
|
+ else
|
|
|
|
+ can_queue = MVS_CAN_QUEUE;
|
|
|
|
+
|
|
|
|
+ sha->lldd_queue_size = can_queue;
|
|
|
|
+ shost->can_queue = can_queue;
|
|
|
|
+ mvi->shost->cmd_per_lun = MVS_SLOTS/sha->num_phys;
|
|
|
|
+ sha->core.shost = mvi->shost;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void mvs_init_sas_add(struct mvs_info *mvi)
|
|
|
|
+{
|
|
|
|
+ u8 i;
|
|
|
|
+ for (i = 0; i < mvi->chip->n_phy; i++) {
|
|
|
|
+ mvi->phy[i].dev_sas_addr = 0x5005043011ab0000ULL;
|
|
|
|
+ mvi->phy[i].dev_sas_addr =
|
|
|
|
+ cpu_to_be64((u64)(*(u64 *)&mvi->phy[i].dev_sas_addr));
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ memcpy(mvi->sas_addr, &mvi->phy[0].dev_sas_addr, SAS_ADDR_SIZE);
|
|
|
|
+}
|
|
|
|
+
|
|
static int __devinit mvs_pci_init(struct pci_dev *pdev,
|
|
static int __devinit mvs_pci_init(struct pci_dev *pdev,
|
|
const struct pci_device_id *ent)
|
|
const struct pci_device_id *ent)
|
|
{
|
|
{
|
|
- int rc;
|
|
|
|
|
|
+ unsigned int rc, nhost = 0;
|
|
struct mvs_info *mvi;
|
|
struct mvs_info *mvi;
|
|
irq_handler_t irq_handler = mvs_interrupt;
|
|
irq_handler_t irq_handler = mvs_interrupt;
|
|
|
|
+ struct Scsi_Host *shost = NULL;
|
|
|
|
+ const struct mvs_chip_info *chip;
|
|
|
|
|
|
|
|
+ dev_printk(KERN_INFO, &pdev->dev,
|
|
|
|
+ "mvsas: driver version %s\n", DRV_VERSION);
|
|
rc = pci_enable_device(pdev);
|
|
rc = pci_enable_device(pdev);
|
|
if (rc)
|
|
if (rc)
|
|
- return rc;
|
|
|
|
|
|
+ goto err_out_enable;
|
|
|
|
|
|
pci_set_master(pdev);
|
|
pci_set_master(pdev);
|
|
|
|
|
|
@@ -384,84 +531,110 @@ static int __devinit mvs_pci_init(struct pci_dev *pdev,
|
|
if (rc)
|
|
if (rc)
|
|
goto err_out_regions;
|
|
goto err_out_regions;
|
|
|
|
|
|
- mvi = mvs_alloc(pdev, ent);
|
|
|
|
- if (!mvi) {
|
|
|
|
|
|
+ shost = scsi_host_alloc(&mvs_sht, sizeof(void *));
|
|
|
|
+ if (!shost) {
|
|
rc = -ENOMEM;
|
|
rc = -ENOMEM;
|
|
goto err_out_regions;
|
|
goto err_out_regions;
|
|
}
|
|
}
|
|
|
|
|
|
- rc = mvs_hw_init(mvi);
|
|
|
|
- if (rc)
|
|
|
|
- goto err_out_mvi;
|
|
|
|
-
|
|
|
|
-#ifndef MVS_DISABLE_MSI
|
|
|
|
- if (!pci_enable_msi(pdev)) {
|
|
|
|
- u32 tmp;
|
|
|
|
- void __iomem *regs = mvi->regs;
|
|
|
|
- mvi->flags |= MVF_MSI;
|
|
|
|
- irq_handler = mvs_msi_interrupt;
|
|
|
|
- tmp = mr32(PCS);
|
|
|
|
- mw32(PCS, tmp | PCS_SELF_CLEAR);
|
|
|
|
|
|
+ chip = &mvs_chips[ent->driver_data];
|
|
|
|
+ SHOST_TO_SAS_HA(shost) =
|
|
|
|
+ kcalloc(1, sizeof(struct sas_ha_struct), GFP_KERNEL);
|
|
|
|
+ if (!SHOST_TO_SAS_HA(shost)) {
|
|
|
|
+ kfree(shost);
|
|
|
|
+ rc = -ENOMEM;
|
|
|
|
+ goto err_out_regions;
|
|
}
|
|
}
|
|
-#endif
|
|
|
|
|
|
|
|
- rc = request_irq(pdev->irq, irq_handler, IRQF_SHARED, DRV_NAME, mvi);
|
|
|
|
- if (rc)
|
|
|
|
- goto err_out_msi;
|
|
|
|
|
|
+ rc = mvs_prep_sas_ha_init(shost, chip);
|
|
|
|
+ if (rc) {
|
|
|
|
+ kfree(shost);
|
|
|
|
+ rc = -ENOMEM;
|
|
|
|
+ goto err_out_regions;
|
|
|
|
+ }
|
|
|
|
|
|
- rc = scsi_add_host(mvi->shost, &pdev->dev);
|
|
|
|
- if (rc)
|
|
|
|
- goto err_out_irq;
|
|
|
|
|
|
+ pci_set_drvdata(pdev, SHOST_TO_SAS_HA(shost));
|
|
|
|
|
|
- rc = sas_register_ha(&mvi->sas);
|
|
|
|
|
|
+ do {
|
|
|
|
+ mvi = mvs_pci_alloc(pdev, ent, shost, nhost);
|
|
|
|
+ if (!mvi) {
|
|
|
|
+ rc = -ENOMEM;
|
|
|
|
+ goto err_out_regions;
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ mvs_init_sas_add(mvi);
|
|
|
|
+
|
|
|
|
+ mvi->instance = nhost;
|
|
|
|
+ rc = MVS_CHIP_DISP->chip_init(mvi);
|
|
|
|
+ if (rc) {
|
|
|
|
+ mvs_free(mvi);
|
|
|
|
+ goto err_out_regions;
|
|
|
|
+ }
|
|
|
|
+ nhost++;
|
|
|
|
+ } while (nhost < chip->n_host);
|
|
|
|
+
|
|
|
|
+ mvs_post_sas_ha_init(shost, chip);
|
|
|
|
+
|
|
|
|
+ rc = scsi_add_host(shost, &pdev->dev);
|
|
if (rc)
|
|
if (rc)
|
|
goto err_out_shost;
|
|
goto err_out_shost;
|
|
|
|
|
|
- pci_set_drvdata(pdev, mvi);
|
|
|
|
-
|
|
|
|
- mvs_print_info(mvi);
|
|
|
|
|
|
+ rc = sas_register_ha(SHOST_TO_SAS_HA(shost));
|
|
|
|
+ if (rc)
|
|
|
|
+ goto err_out_shost;
|
|
|
|
+ rc = request_irq(pdev->irq, irq_handler, IRQF_SHARED,
|
|
|
|
+ DRV_NAME, SHOST_TO_SAS_HA(shost));
|
|
|
|
+ if (rc)
|
|
|
|
+ goto err_not_sas;
|
|
|
|
|
|
- mvs_hba_interrupt_enable(mvi);
|
|
|
|
|
|
+ MVS_CHIP_DISP->interrupt_enable(mvi);
|
|
|
|
|
|
scsi_scan_host(mvi->shost);
|
|
scsi_scan_host(mvi->shost);
|
|
|
|
|
|
return 0;
|
|
return 0;
|
|
|
|
|
|
|
|
+err_not_sas:
|
|
|
|
+ sas_unregister_ha(SHOST_TO_SAS_HA(shost));
|
|
err_out_shost:
|
|
err_out_shost:
|
|
scsi_remove_host(mvi->shost);
|
|
scsi_remove_host(mvi->shost);
|
|
-err_out_irq:
|
|
|
|
- free_irq(pdev->irq, mvi);
|
|
|
|
-err_out_msi:
|
|
|
|
- if (mvi->flags |= MVF_MSI)
|
|
|
|
- pci_disable_msi(pdev);
|
|
|
|
-err_out_mvi:
|
|
|
|
- mvs_free(mvi);
|
|
|
|
err_out_regions:
|
|
err_out_regions:
|
|
pci_release_regions(pdev);
|
|
pci_release_regions(pdev);
|
|
err_out_disable:
|
|
err_out_disable:
|
|
pci_disable_device(pdev);
|
|
pci_disable_device(pdev);
|
|
|
|
+err_out_enable:
|
|
return rc;
|
|
return rc;
|
|
}
|
|
}
|
|
|
|
|
|
static void __devexit mvs_pci_remove(struct pci_dev *pdev)
|
|
static void __devexit mvs_pci_remove(struct pci_dev *pdev)
|
|
{
|
|
{
|
|
- struct mvs_info *mvi = pci_get_drvdata(pdev);
|
|
|
|
|
|
+ unsigned short core_nr, i = 0;
|
|
|
|
+ struct sas_ha_struct *sha = pci_get_drvdata(pdev);
|
|
|
|
+ struct mvs_info *mvi = NULL;
|
|
|
|
|
|
- pci_set_drvdata(pdev, NULL);
|
|
|
|
|
|
+ core_nr = ((struct mvs_prv_info *)sha->lldd_ha)->n_host;
|
|
|
|
+ mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[0];
|
|
|
|
|
|
- if (mvi) {
|
|
|
|
- sas_unregister_ha(&mvi->sas);
|
|
|
|
- mvs_hba_interrupt_disable(mvi);
|
|
|
|
- sas_remove_host(mvi->shost);
|
|
|
|
- scsi_remove_host(mvi->shost);
|
|
|
|
|
|
+#ifdef MVS_USE_TASKLET
|
|
|
|
+ tasklet_kill(&mv_tasklet);
|
|
|
|
+#endif
|
|
|
|
|
|
- free_irq(pdev->irq, mvi);
|
|
|
|
- if (mvi->flags & MVF_MSI)
|
|
|
|
- pci_disable_msi(pdev);
|
|
|
|
|
|
+ pci_set_drvdata(pdev, NULL);
|
|
|
|
+ sas_unregister_ha(sha);
|
|
|
|
+ sas_remove_host(mvi->shost);
|
|
|
|
+ scsi_remove_host(mvi->shost);
|
|
|
|
+
|
|
|
|
+ MVS_CHIP_DISP->interrupt_disable(mvi);
|
|
|
|
+ free_irq(mvi->irq, sha);
|
|
|
|
+ for (i = 0; i < core_nr; i++) {
|
|
|
|
+ mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[i];
|
|
mvs_free(mvi);
|
|
mvs_free(mvi);
|
|
- pci_release_regions(pdev);
|
|
|
|
}
|
|
}
|
|
|
|
+ kfree(sha->sas_phy);
|
|
|
|
+ kfree(sha->sas_port);
|
|
|
|
+ kfree(sha);
|
|
|
|
+ pci_release_regions(pdev);
|
|
pci_disable_device(pdev);
|
|
pci_disable_device(pdev);
|
|
|
|
+ return;
|
|
}
|
|
}
|
|
|
|
|
|
static struct pci_device_id __devinitdata mvs_pci_table[] = {
|
|
static struct pci_device_id __devinitdata mvs_pci_table[] = {
|
|
@@ -474,10 +647,12 @@ static struct pci_device_id __devinitdata mvs_pci_table[] = {
|
|
.subdevice = 0x6480,
|
|
.subdevice = 0x6480,
|
|
.class = 0,
|
|
.class = 0,
|
|
.class_mask = 0,
|
|
.class_mask = 0,
|
|
- .driver_data = chip_6480,
|
|
|
|
|
|
+ .driver_data = chip_6485,
|
|
},
|
|
},
|
|
{ PCI_VDEVICE(MARVELL, 0x6440), chip_6440 },
|
|
{ PCI_VDEVICE(MARVELL, 0x6440), chip_6440 },
|
|
- { PCI_VDEVICE(MARVELL, 0x6480), chip_6480 },
|
|
|
|
|
|
+ { PCI_VDEVICE(MARVELL, 0x6485), chip_6485 },
|
|
|
|
+ { PCI_VDEVICE(MARVELL, 0x9480), chip_9480 },
|
|
|
|
+ { PCI_VDEVICE(MARVELL, 0x9180), chip_9180 },
|
|
|
|
|
|
{ } /* terminate list */
|
|
{ } /* terminate list */
|
|
};
|
|
};
|
|
@@ -489,15 +664,17 @@ static struct pci_driver mvs_pci_driver = {
|
|
.remove = __devexit_p(mvs_pci_remove),
|
|
.remove = __devexit_p(mvs_pci_remove),
|
|
};
|
|
};
|
|
|
|
|
|
|
|
+/* task handler */
|
|
|
|
+struct task_struct *mvs_th;
|
|
static int __init mvs_init(void)
|
|
static int __init mvs_init(void)
|
|
{
|
|
{
|
|
int rc;
|
|
int rc;
|
|
-
|
|
|
|
mvs_stt = sas_domain_attach_transport(&mvs_transport_ops);
|
|
mvs_stt = sas_domain_attach_transport(&mvs_transport_ops);
|
|
if (!mvs_stt)
|
|
if (!mvs_stt)
|
|
return -ENOMEM;
|
|
return -ENOMEM;
|
|
|
|
|
|
rc = pci_register_driver(&mvs_pci_driver);
|
|
rc = pci_register_driver(&mvs_pci_driver);
|
|
|
|
+
|
|
if (rc)
|
|
if (rc)
|
|
goto err_out;
|
|
goto err_out;
|
|
|
|
|
|
@@ -521,4 +698,6 @@ MODULE_AUTHOR("Jeff Garzik <jgarzik@pobox.com>");
|
|
MODULE_DESCRIPTION("Marvell 88SE6440 SAS/SATA controller driver");
|
|
MODULE_DESCRIPTION("Marvell 88SE6440 SAS/SATA controller driver");
|
|
MODULE_VERSION(DRV_VERSION);
|
|
MODULE_VERSION(DRV_VERSION);
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_LICENSE("GPL");
|
|
|
|
+#ifdef CONFIG_PCI
|
|
MODULE_DEVICE_TABLE(pci, mvs_pci_table);
|
|
MODULE_DEVICE_TABLE(pci, mvs_pci_table);
|
|
|
|
+#endif
|