|
@@ -101,242 +101,14 @@ int __ipath_layer_rcv_lid(struct ipath_devdata *dd, void *hdr)
|
|
|
return ret;
|
|
|
}
|
|
|
|
|
|
-int ipath_layer_set_linkstate(struct ipath_devdata *dd, u8 newstate)
|
|
|
+void ipath_layer_lid_changed(struct ipath_devdata *dd)
|
|
|
{
|
|
|
- u32 lstate;
|
|
|
- int ret;
|
|
|
-
|
|
|
- switch (newstate) {
|
|
|
- case IPATH_IB_LINKDOWN:
|
|
|
- ipath_set_ib_lstate(dd, INFINIPATH_IBCC_LINKINITCMD_POLL <<
|
|
|
- INFINIPATH_IBCC_LINKINITCMD_SHIFT);
|
|
|
- /* don't wait */
|
|
|
- ret = 0;
|
|
|
- goto bail;
|
|
|
-
|
|
|
- case IPATH_IB_LINKDOWN_SLEEP:
|
|
|
- ipath_set_ib_lstate(dd, INFINIPATH_IBCC_LINKINITCMD_SLEEP <<
|
|
|
- INFINIPATH_IBCC_LINKINITCMD_SHIFT);
|
|
|
- /* don't wait */
|
|
|
- ret = 0;
|
|
|
- goto bail;
|
|
|
-
|
|
|
- case IPATH_IB_LINKDOWN_DISABLE:
|
|
|
- ipath_set_ib_lstate(dd,
|
|
|
- INFINIPATH_IBCC_LINKINITCMD_DISABLE <<
|
|
|
- INFINIPATH_IBCC_LINKINITCMD_SHIFT);
|
|
|
- /* don't wait */
|
|
|
- ret = 0;
|
|
|
- goto bail;
|
|
|
-
|
|
|
- case IPATH_IB_LINKINIT:
|
|
|
- if (dd->ipath_flags & IPATH_LINKINIT) {
|
|
|
- ret = 0;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- ipath_set_ib_lstate(dd, INFINIPATH_IBCC_LINKCMD_INIT <<
|
|
|
- INFINIPATH_IBCC_LINKCMD_SHIFT);
|
|
|
- lstate = IPATH_LINKINIT;
|
|
|
- break;
|
|
|
-
|
|
|
- case IPATH_IB_LINKARM:
|
|
|
- if (dd->ipath_flags & IPATH_LINKARMED) {
|
|
|
- ret = 0;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- if (!(dd->ipath_flags &
|
|
|
- (IPATH_LINKINIT | IPATH_LINKACTIVE))) {
|
|
|
- ret = -EINVAL;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- ipath_set_ib_lstate(dd, INFINIPATH_IBCC_LINKCMD_ARMED <<
|
|
|
- INFINIPATH_IBCC_LINKCMD_SHIFT);
|
|
|
- /*
|
|
|
- * Since the port can transition to ACTIVE by receiving
|
|
|
- * a non VL 15 packet, wait for either state.
|
|
|
- */
|
|
|
- lstate = IPATH_LINKARMED | IPATH_LINKACTIVE;
|
|
|
- break;
|
|
|
-
|
|
|
- case IPATH_IB_LINKACTIVE:
|
|
|
- if (dd->ipath_flags & IPATH_LINKACTIVE) {
|
|
|
- ret = 0;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- if (!(dd->ipath_flags & IPATH_LINKARMED)) {
|
|
|
- ret = -EINVAL;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- ipath_set_ib_lstate(dd, INFINIPATH_IBCC_LINKCMD_ACTIVE <<
|
|
|
- INFINIPATH_IBCC_LINKCMD_SHIFT);
|
|
|
- lstate = IPATH_LINKACTIVE;
|
|
|
- break;
|
|
|
-
|
|
|
- default:
|
|
|
- ipath_dbg("Invalid linkstate 0x%x requested\n", newstate);
|
|
|
- ret = -EINVAL;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- ret = ipath_wait_linkstate(dd, lstate, 2000);
|
|
|
-
|
|
|
-bail:
|
|
|
- return ret;
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * ipath_layer_set_mtu - set the MTU
|
|
|
- * @dd: the infinipath device
|
|
|
- * @arg: the new MTU
|
|
|
- *
|
|
|
- * we can handle "any" incoming size, the issue here is whether we
|
|
|
- * need to restrict our outgoing size. For now, we don't do any
|
|
|
- * sanity checking on this, and we don't deal with what happens to
|
|
|
- * programs that are already running when the size changes.
|
|
|
- * NOTE: changing the MTU will usually cause the IBC to go back to
|
|
|
- * link initialize (IPATH_IBSTATE_INIT) state...
|
|
|
- */
|
|
|
-int ipath_layer_set_mtu(struct ipath_devdata *dd, u16 arg)
|
|
|
-{
|
|
|
- u32 piosize;
|
|
|
- int changed = 0;
|
|
|
- int ret;
|
|
|
-
|
|
|
- /*
|
|
|
- * mtu is IB data payload max. It's the largest power of 2 less
|
|
|
- * than piosize (or even larger, since it only really controls the
|
|
|
- * largest we can receive; we can send the max of the mtu and
|
|
|
- * piosize). We check that it's one of the valid IB sizes.
|
|
|
- */
|
|
|
- if (arg != 256 && arg != 512 && arg != 1024 && arg != 2048 &&
|
|
|
- arg != 4096) {
|
|
|
- ipath_dbg("Trying to set invalid mtu %u, failing\n", arg);
|
|
|
- ret = -EINVAL;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- if (dd->ipath_ibmtu == arg) {
|
|
|
- ret = 0; /* same as current */
|
|
|
- goto bail;
|
|
|
- }
|
|
|
-
|
|
|
- piosize = dd->ipath_ibmaxlen;
|
|
|
- dd->ipath_ibmtu = arg;
|
|
|
-
|
|
|
- if (arg >= (piosize - IPATH_PIO_MAXIBHDR)) {
|
|
|
- /* Only if it's not the initial value (or reset to it) */
|
|
|
- if (piosize != dd->ipath_init_ibmaxlen) {
|
|
|
- dd->ipath_ibmaxlen = piosize;
|
|
|
- changed = 1;
|
|
|
- }
|
|
|
- } else if ((arg + IPATH_PIO_MAXIBHDR) != dd->ipath_ibmaxlen) {
|
|
|
- piosize = arg + IPATH_PIO_MAXIBHDR;
|
|
|
- ipath_cdbg(VERBOSE, "ibmaxlen was 0x%x, setting to 0x%x "
|
|
|
- "(mtu 0x%x)\n", dd->ipath_ibmaxlen, piosize,
|
|
|
- arg);
|
|
|
- dd->ipath_ibmaxlen = piosize;
|
|
|
- changed = 1;
|
|
|
- }
|
|
|
-
|
|
|
- if (changed) {
|
|
|
- /*
|
|
|
- * set the IBC maxpktlength to the size of our pio
|
|
|
- * buffers in words
|
|
|
- */
|
|
|
- u64 ibc = dd->ipath_ibcctrl;
|
|
|
- ibc &= ~(INFINIPATH_IBCC_MAXPKTLEN_MASK <<
|
|
|
- INFINIPATH_IBCC_MAXPKTLEN_SHIFT);
|
|
|
-
|
|
|
- piosize = piosize - 2 * sizeof(u32); /* ignore pbc */
|
|
|
- dd->ipath_ibmaxlen = piosize;
|
|
|
- piosize /= sizeof(u32); /* in words */
|
|
|
- /*
|
|
|
- * for ICRC, which we only send in diag test pkt mode, and
|
|
|
- * we don't need to worry about that for mtu
|
|
|
- */
|
|
|
- piosize += 1;
|
|
|
-
|
|
|
- ibc |= piosize << INFINIPATH_IBCC_MAXPKTLEN_SHIFT;
|
|
|
- dd->ipath_ibcctrl = ibc;
|
|
|
- ipath_write_kreg(dd, dd->ipath_kregs->kr_ibcctrl,
|
|
|
- dd->ipath_ibcctrl);
|
|
|
- dd->ipath_f_tidtemplate(dd);
|
|
|
- }
|
|
|
-
|
|
|
- ret = 0;
|
|
|
-
|
|
|
-bail:
|
|
|
- return ret;
|
|
|
-}
|
|
|
-
|
|
|
-int ipath_set_lid(struct ipath_devdata *dd, u32 arg, u8 lmc)
|
|
|
-{
|
|
|
- dd->ipath_lid = arg;
|
|
|
- dd->ipath_lmc = lmc;
|
|
|
-
|
|
|
mutex_lock(&ipath_layer_mutex);
|
|
|
|
|
|
if (dd->ipath_layer.l_arg && layer_intr)
|
|
|
layer_intr(dd->ipath_layer.l_arg, IPATH_LAYER_INT_LID);
|
|
|
|
|
|
mutex_unlock(&ipath_layer_mutex);
|
|
|
-
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-int ipath_layer_set_guid(struct ipath_devdata *dd, __be64 guid)
|
|
|
-{
|
|
|
- /* XXX - need to inform anyone who cares this just happened. */
|
|
|
- dd->ipath_guid = guid;
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-__be64 ipath_layer_get_guid(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return dd->ipath_guid;
|
|
|
-}
|
|
|
-
|
|
|
-u32 ipath_layer_get_majrev(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return dd->ipath_majrev;
|
|
|
-}
|
|
|
-
|
|
|
-u32 ipath_layer_get_minrev(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return dd->ipath_minrev;
|
|
|
-}
|
|
|
-
|
|
|
-u32 ipath_layer_get_pcirev(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return dd->ipath_pcirev;
|
|
|
-}
|
|
|
-
|
|
|
-u32 ipath_layer_get_flags(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return dd->ipath_flags;
|
|
|
-}
|
|
|
-
|
|
|
-struct device *ipath_layer_get_device(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return &dd->pcidev->dev;
|
|
|
-}
|
|
|
-
|
|
|
-u16 ipath_layer_get_deviceid(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return dd->ipath_deviceid;
|
|
|
-}
|
|
|
-
|
|
|
-u32 ipath_layer_get_vendorid(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return dd->ipath_vendorid;
|
|
|
-}
|
|
|
-
|
|
|
-u64 ipath_layer_get_lastibcstat(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return dd->ipath_lastibcstat;
|
|
|
-}
|
|
|
-
|
|
|
-u32 ipath_layer_get_ibmtu(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return dd->ipath_ibmtu;
|
|
|
}
|
|
|
|
|
|
void ipath_layer_add(struct ipath_devdata *dd)
|
|
@@ -436,22 +208,6 @@ void ipath_layer_unregister(void)
|
|
|
|
|
|
EXPORT_SYMBOL_GPL(ipath_layer_unregister);
|
|
|
|
|
|
-static void __ipath_verbs_timer(unsigned long arg)
|
|
|
-{
|
|
|
- struct ipath_devdata *dd = (struct ipath_devdata *) arg;
|
|
|
-
|
|
|
- /*
|
|
|
- * If port 0 receive packet interrupts are not available, or
|
|
|
- * can be missed, poll the receive queue
|
|
|
- */
|
|
|
- if (dd->ipath_flags & IPATH_POLL_RX_INTR)
|
|
|
- ipath_kreceive(dd);
|
|
|
-
|
|
|
- /* Handle verbs layer timeouts. */
|
|
|
- ipath_ib_timer(dd->verbs_dev);
|
|
|
- mod_timer(&dd->verbs_timer, jiffies + 1);
|
|
|
-}
|
|
|
-
|
|
|
int ipath_layer_open(struct ipath_devdata *dd, u32 * pktmax)
|
|
|
{
|
|
|
int ret;
|
|
@@ -540,380 +296,6 @@ u16 ipath_layer_get_bcast(struct ipath_devdata *dd)
|
|
|
|
|
|
EXPORT_SYMBOL_GPL(ipath_layer_get_bcast);
|
|
|
|
|
|
-u32 ipath_layer_get_cr_errpkey(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return ipath_read_creg32(dd, dd->ipath_cregs->cr_errpkey);
|
|
|
-}
|
|
|
-
|
|
|
-static void update_sge(struct ipath_sge_state *ss, u32 length)
|
|
|
-{
|
|
|
- struct ipath_sge *sge = &ss->sge;
|
|
|
-
|
|
|
- sge->vaddr += length;
|
|
|
- sge->length -= length;
|
|
|
- sge->sge_length -= length;
|
|
|
- if (sge->sge_length == 0) {
|
|
|
- if (--ss->num_sge)
|
|
|
- *sge = *ss->sg_list++;
|
|
|
- } else if (sge->length == 0 && sge->mr != NULL) {
|
|
|
- if (++sge->n >= IPATH_SEGSZ) {
|
|
|
- if (++sge->m >= sge->mr->mapsz)
|
|
|
- return;
|
|
|
- sge->n = 0;
|
|
|
- }
|
|
|
- sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr;
|
|
|
- sge->length = sge->mr->map[sge->m]->segs[sge->n].length;
|
|
|
- }
|
|
|
-}
|
|
|
-
|
|
|
-#ifdef __LITTLE_ENDIAN
|
|
|
-static inline u32 get_upper_bits(u32 data, u32 shift)
|
|
|
-{
|
|
|
- return data >> shift;
|
|
|
-}
|
|
|
-
|
|
|
-static inline u32 set_upper_bits(u32 data, u32 shift)
|
|
|
-{
|
|
|
- return data << shift;
|
|
|
-}
|
|
|
-
|
|
|
-static inline u32 clear_upper_bytes(u32 data, u32 n, u32 off)
|
|
|
-{
|
|
|
- data <<= ((sizeof(u32) - n) * BITS_PER_BYTE);
|
|
|
- data >>= ((sizeof(u32) - n - off) * BITS_PER_BYTE);
|
|
|
- return data;
|
|
|
-}
|
|
|
-#else
|
|
|
-static inline u32 get_upper_bits(u32 data, u32 shift)
|
|
|
-{
|
|
|
- return data << shift;
|
|
|
-}
|
|
|
-
|
|
|
-static inline u32 set_upper_bits(u32 data, u32 shift)
|
|
|
-{
|
|
|
- return data >> shift;
|
|
|
-}
|
|
|
-
|
|
|
-static inline u32 clear_upper_bytes(u32 data, u32 n, u32 off)
|
|
|
-{
|
|
|
- data >>= ((sizeof(u32) - n) * BITS_PER_BYTE);
|
|
|
- data <<= ((sizeof(u32) - n - off) * BITS_PER_BYTE);
|
|
|
- return data;
|
|
|
-}
|
|
|
-#endif
|
|
|
-
|
|
|
-static void copy_io(u32 __iomem *piobuf, struct ipath_sge_state *ss,
|
|
|
- u32 length)
|
|
|
-{
|
|
|
- u32 extra = 0;
|
|
|
- u32 data = 0;
|
|
|
- u32 last;
|
|
|
-
|
|
|
- while (1) {
|
|
|
- u32 len = ss->sge.length;
|
|
|
- u32 off;
|
|
|
-
|
|
|
- BUG_ON(len == 0);
|
|
|
- if (len > length)
|
|
|
- len = length;
|
|
|
- if (len > ss->sge.sge_length)
|
|
|
- len = ss->sge.sge_length;
|
|
|
- /* If the source address is not aligned, try to align it. */
|
|
|
- off = (unsigned long)ss->sge.vaddr & (sizeof(u32) - 1);
|
|
|
- if (off) {
|
|
|
- u32 *addr = (u32 *)((unsigned long)ss->sge.vaddr &
|
|
|
- ~(sizeof(u32) - 1));
|
|
|
- u32 v = get_upper_bits(*addr, off * BITS_PER_BYTE);
|
|
|
- u32 y;
|
|
|
-
|
|
|
- y = sizeof(u32) - off;
|
|
|
- if (len > y)
|
|
|
- len = y;
|
|
|
- if (len + extra >= sizeof(u32)) {
|
|
|
- data |= set_upper_bits(v, extra *
|
|
|
- BITS_PER_BYTE);
|
|
|
- len = sizeof(u32) - extra;
|
|
|
- if (len == length) {
|
|
|
- last = data;
|
|
|
- break;
|
|
|
- }
|
|
|
- __raw_writel(data, piobuf);
|
|
|
- piobuf++;
|
|
|
- extra = 0;
|
|
|
- data = 0;
|
|
|
- } else {
|
|
|
- /* Clear unused upper bytes */
|
|
|
- data |= clear_upper_bytes(v, len, extra);
|
|
|
- if (len == length) {
|
|
|
- last = data;
|
|
|
- break;
|
|
|
- }
|
|
|
- extra += len;
|
|
|
- }
|
|
|
- } else if (extra) {
|
|
|
- /* Source address is aligned. */
|
|
|
- u32 *addr = (u32 *) ss->sge.vaddr;
|
|
|
- int shift = extra * BITS_PER_BYTE;
|
|
|
- int ushift = 32 - shift;
|
|
|
- u32 l = len;
|
|
|
-
|
|
|
- while (l >= sizeof(u32)) {
|
|
|
- u32 v = *addr;
|
|
|
-
|
|
|
- data |= set_upper_bits(v, shift);
|
|
|
- __raw_writel(data, piobuf);
|
|
|
- data = get_upper_bits(v, ushift);
|
|
|
- piobuf++;
|
|
|
- addr++;
|
|
|
- l -= sizeof(u32);
|
|
|
- }
|
|
|
- /*
|
|
|
- * We still have 'extra' number of bytes leftover.
|
|
|
- */
|
|
|
- if (l) {
|
|
|
- u32 v = *addr;
|
|
|
-
|
|
|
- if (l + extra >= sizeof(u32)) {
|
|
|
- data |= set_upper_bits(v, shift);
|
|
|
- len -= l + extra - sizeof(u32);
|
|
|
- if (len == length) {
|
|
|
- last = data;
|
|
|
- break;
|
|
|
- }
|
|
|
- __raw_writel(data, piobuf);
|
|
|
- piobuf++;
|
|
|
- extra = 0;
|
|
|
- data = 0;
|
|
|
- } else {
|
|
|
- /* Clear unused upper bytes */
|
|
|
- data |= clear_upper_bytes(v, l,
|
|
|
- extra);
|
|
|
- if (len == length) {
|
|
|
- last = data;
|
|
|
- break;
|
|
|
- }
|
|
|
- extra += l;
|
|
|
- }
|
|
|
- } else if (len == length) {
|
|
|
- last = data;
|
|
|
- break;
|
|
|
- }
|
|
|
- } else if (len == length) {
|
|
|
- u32 w;
|
|
|
-
|
|
|
- /*
|
|
|
- * Need to round up for the last dword in the
|
|
|
- * packet.
|
|
|
- */
|
|
|
- w = (len + 3) >> 2;
|
|
|
- __iowrite32_copy(piobuf, ss->sge.vaddr, w - 1);
|
|
|
- piobuf += w - 1;
|
|
|
- last = ((u32 *) ss->sge.vaddr)[w - 1];
|
|
|
- break;
|
|
|
- } else {
|
|
|
- u32 w = len >> 2;
|
|
|
-
|
|
|
- __iowrite32_copy(piobuf, ss->sge.vaddr, w);
|
|
|
- piobuf += w;
|
|
|
-
|
|
|
- extra = len & (sizeof(u32) - 1);
|
|
|
- if (extra) {
|
|
|
- u32 v = ((u32 *) ss->sge.vaddr)[w];
|
|
|
-
|
|
|
- /* Clear unused upper bytes */
|
|
|
- data = clear_upper_bytes(v, extra, 0);
|
|
|
- }
|
|
|
- }
|
|
|
- update_sge(ss, len);
|
|
|
- length -= len;
|
|
|
- }
|
|
|
- /* Update address before sending packet. */
|
|
|
- update_sge(ss, length);
|
|
|
- /* must flush early everything before trigger word */
|
|
|
- ipath_flush_wc();
|
|
|
- __raw_writel(last, piobuf);
|
|
|
- /* be sure trigger word is written */
|
|
|
- ipath_flush_wc();
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * ipath_verbs_send - send a packet from the verbs layer
|
|
|
- * @dd: the infinipath device
|
|
|
- * @hdrwords: the number of words in the header
|
|
|
- * @hdr: the packet header
|
|
|
- * @len: the length of the packet in bytes
|
|
|
- * @ss: the SGE to send
|
|
|
- *
|
|
|
- * This is like ipath_sma_send_pkt() in that we need to be able to send
|
|
|
- * packets after the chip is initialized (MADs) but also like
|
|
|
- * ipath_layer_send_hdr() since its used by the verbs layer.
|
|
|
- */
|
|
|
-int ipath_verbs_send(struct ipath_devdata *dd, u32 hdrwords,
|
|
|
- u32 *hdr, u32 len, struct ipath_sge_state *ss)
|
|
|
-{
|
|
|
- u32 __iomem *piobuf;
|
|
|
- u32 plen;
|
|
|
- int ret;
|
|
|
-
|
|
|
- /* +1 is for the qword padding of pbc */
|
|
|
- plen = hdrwords + ((len + 3) >> 2) + 1;
|
|
|
- if (unlikely((plen << 2) > dd->ipath_ibmaxlen)) {
|
|
|
- ipath_dbg("packet len 0x%x too long, failing\n", plen);
|
|
|
- ret = -EINVAL;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
-
|
|
|
- /* Get a PIO buffer to use. */
|
|
|
- piobuf = ipath_getpiobuf(dd, NULL);
|
|
|
- if (unlikely(piobuf == NULL)) {
|
|
|
- ret = -EBUSY;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
-
|
|
|
- /*
|
|
|
- * Write len to control qword, no flags.
|
|
|
- * We have to flush after the PBC for correctness on some cpus
|
|
|
- * or WC buffer can be written out of order.
|
|
|
- */
|
|
|
- writeq(plen, piobuf);
|
|
|
- ipath_flush_wc();
|
|
|
- piobuf += 2;
|
|
|
- if (len == 0) {
|
|
|
- /*
|
|
|
- * If there is just the header portion, must flush before
|
|
|
- * writing last word of header for correctness, and after
|
|
|
- * the last header word (trigger word).
|
|
|
- */
|
|
|
- __iowrite32_copy(piobuf, hdr, hdrwords - 1);
|
|
|
- ipath_flush_wc();
|
|
|
- __raw_writel(hdr[hdrwords - 1], piobuf + hdrwords - 1);
|
|
|
- ipath_flush_wc();
|
|
|
- ret = 0;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
-
|
|
|
- __iowrite32_copy(piobuf, hdr, hdrwords);
|
|
|
- piobuf += hdrwords;
|
|
|
-
|
|
|
- /* The common case is aligned and contained in one segment. */
|
|
|
- if (likely(ss->num_sge == 1 && len <= ss->sge.length &&
|
|
|
- !((unsigned long)ss->sge.vaddr & (sizeof(u32) - 1)))) {
|
|
|
- u32 w;
|
|
|
- u32 *addr = (u32 *) ss->sge.vaddr;
|
|
|
-
|
|
|
- /* Update address before sending packet. */
|
|
|
- update_sge(ss, len);
|
|
|
- /* Need to round up for the last dword in the packet. */
|
|
|
- w = (len + 3) >> 2;
|
|
|
- __iowrite32_copy(piobuf, addr, w - 1);
|
|
|
- /* must flush early everything before trigger word */
|
|
|
- ipath_flush_wc();
|
|
|
- __raw_writel(addr[w - 1], piobuf + w - 1);
|
|
|
- /* be sure trigger word is written */
|
|
|
- ipath_flush_wc();
|
|
|
- ret = 0;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- copy_io(piobuf, ss, len);
|
|
|
- ret = 0;
|
|
|
-
|
|
|
-bail:
|
|
|
- return ret;
|
|
|
-}
|
|
|
-
|
|
|
-int ipath_layer_snapshot_counters(struct ipath_devdata *dd, u64 *swords,
|
|
|
- u64 *rwords, u64 *spkts, u64 *rpkts,
|
|
|
- u64 *xmit_wait)
|
|
|
-{
|
|
|
- int ret;
|
|
|
-
|
|
|
- if (!(dd->ipath_flags & IPATH_INITTED)) {
|
|
|
- /* no hardware, freeze, etc. */
|
|
|
- ipath_dbg("unit %u not usable\n", dd->ipath_unit);
|
|
|
- ret = -EINVAL;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- *swords = ipath_snap_cntr(dd, dd->ipath_cregs->cr_wordsendcnt);
|
|
|
- *rwords = ipath_snap_cntr(dd, dd->ipath_cregs->cr_wordrcvcnt);
|
|
|
- *spkts = ipath_snap_cntr(dd, dd->ipath_cregs->cr_pktsendcnt);
|
|
|
- *rpkts = ipath_snap_cntr(dd, dd->ipath_cregs->cr_pktrcvcnt);
|
|
|
- *xmit_wait = ipath_snap_cntr(dd, dd->ipath_cregs->cr_sendstallcnt);
|
|
|
-
|
|
|
- ret = 0;
|
|
|
-
|
|
|
-bail:
|
|
|
- return ret;
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * ipath_layer_get_counters - get various chip counters
|
|
|
- * @dd: the infinipath device
|
|
|
- * @cntrs: counters are placed here
|
|
|
- *
|
|
|
- * Return the counters needed by recv_pma_get_portcounters().
|
|
|
- */
|
|
|
-int ipath_layer_get_counters(struct ipath_devdata *dd,
|
|
|
- struct ipath_layer_counters *cntrs)
|
|
|
-{
|
|
|
- int ret;
|
|
|
-
|
|
|
- if (!(dd->ipath_flags & IPATH_INITTED)) {
|
|
|
- /* no hardware, freeze, etc. */
|
|
|
- ipath_dbg("unit %u not usable\n", dd->ipath_unit);
|
|
|
- ret = -EINVAL;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- cntrs->symbol_error_counter =
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_ibsymbolerrcnt);
|
|
|
- cntrs->link_error_recovery_counter =
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_iblinkerrrecovcnt);
|
|
|
- /*
|
|
|
- * The link downed counter counts when the other side downs the
|
|
|
- * connection. We add in the number of times we downed the link
|
|
|
- * due to local link integrity errors to compensate.
|
|
|
- */
|
|
|
- cntrs->link_downed_counter =
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_iblinkdowncnt);
|
|
|
- cntrs->port_rcv_errors =
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_rxdroppktcnt) +
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_rcvovflcnt) +
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_portovflcnt) +
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_err_rlencnt) +
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_invalidrlencnt) +
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_erricrccnt) +
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_errvcrccnt) +
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_errlpcrccnt) +
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_badformatcnt);
|
|
|
- cntrs->port_rcv_remphys_errors =
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_rcvebpcnt);
|
|
|
- cntrs->port_xmit_discards =
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_unsupvlcnt);
|
|
|
- cntrs->port_xmit_data =
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_wordsendcnt);
|
|
|
- cntrs->port_rcv_data =
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_wordrcvcnt);
|
|
|
- cntrs->port_xmit_packets =
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_pktsendcnt);
|
|
|
- cntrs->port_rcv_packets =
|
|
|
- ipath_snap_cntr(dd, dd->ipath_cregs->cr_pktrcvcnt);
|
|
|
- cntrs->local_link_integrity_errors = dd->ipath_lli_errors;
|
|
|
- cntrs->excessive_buffer_overrun_errors = 0; /* XXX */
|
|
|
-
|
|
|
- ret = 0;
|
|
|
-
|
|
|
-bail:
|
|
|
- return ret;
|
|
|
-}
|
|
|
-
|
|
|
-int ipath_layer_want_buffer(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- set_bit(IPATH_S_PIOINTBUFAVAIL, &dd->ipath_sendctrl);
|
|
|
- ipath_write_kreg(dd, dd->ipath_kregs->kr_sendctrl,
|
|
|
- dd->ipath_sendctrl);
|
|
|
-
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
int ipath_layer_send_hdr(struct ipath_devdata *dd, struct ether_header *hdr)
|
|
|
{
|
|
|
int ret = 0;
|
|
@@ -985,361 +367,3 @@ int ipath_layer_set_piointbufavail_int(struct ipath_devdata *dd)
|
|
|
}
|
|
|
|
|
|
EXPORT_SYMBOL_GPL(ipath_layer_set_piointbufavail_int);
|
|
|
-
|
|
|
-int ipath_layer_enable_timer(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- /*
|
|
|
- * HT-400 has a design flaw where the chip and kernel idea
|
|
|
- * of the tail register don't always agree, and therefore we won't
|
|
|
- * get an interrupt on the next packet received.
|
|
|
- * If the board supports per packet receive interrupts, use it.
|
|
|
- * Otherwise, the timer function periodically checks for packets
|
|
|
- * to cover this case.
|
|
|
- * Either way, the timer is needed for verbs layer related
|
|
|
- * processing.
|
|
|
- */
|
|
|
- if (dd->ipath_flags & IPATH_GPIO_INTR) {
|
|
|
- ipath_write_kreg(dd, dd->ipath_kregs->kr_debugportselect,
|
|
|
- 0x2074076542310ULL);
|
|
|
- /* Enable GPIO bit 2 interrupt */
|
|
|
- ipath_write_kreg(dd, dd->ipath_kregs->kr_gpio_mask,
|
|
|
- (u64) (1 << 2));
|
|
|
- }
|
|
|
-
|
|
|
- init_timer(&dd->verbs_timer);
|
|
|
- dd->verbs_timer.function = __ipath_verbs_timer;
|
|
|
- dd->verbs_timer.data = (unsigned long)dd;
|
|
|
- dd->verbs_timer.expires = jiffies + 1;
|
|
|
- add_timer(&dd->verbs_timer);
|
|
|
-
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-int ipath_layer_disable_timer(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- /* Disable GPIO bit 2 interrupt */
|
|
|
- if (dd->ipath_flags & IPATH_GPIO_INTR)
|
|
|
- ipath_write_kreg(dd, dd->ipath_kregs->kr_gpio_mask, 0);
|
|
|
-
|
|
|
- del_timer_sync(&dd->verbs_timer);
|
|
|
-
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * ipath_layer_set_verbs_flags - set the verbs layer flags
|
|
|
- * @dd: the infinipath device
|
|
|
- * @flags: the flags to set
|
|
|
- */
|
|
|
-int ipath_layer_set_verbs_flags(struct ipath_devdata *dd, unsigned flags)
|
|
|
-{
|
|
|
- struct ipath_devdata *ss;
|
|
|
- unsigned long lflags;
|
|
|
-
|
|
|
- spin_lock_irqsave(&ipath_devs_lock, lflags);
|
|
|
-
|
|
|
- list_for_each_entry(ss, &ipath_dev_list, ipath_list) {
|
|
|
- if (!(ss->ipath_flags & IPATH_INITTED))
|
|
|
- continue;
|
|
|
- if ((flags & IPATH_VERBS_KERNEL_SMA) &&
|
|
|
- !(*ss->ipath_statusp & IPATH_STATUS_SMA))
|
|
|
- *ss->ipath_statusp |= IPATH_STATUS_OIB_SMA;
|
|
|
- else
|
|
|
- *ss->ipath_statusp &= ~IPATH_STATUS_OIB_SMA;
|
|
|
- }
|
|
|
-
|
|
|
- spin_unlock_irqrestore(&ipath_devs_lock, lflags);
|
|
|
-
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * ipath_layer_get_npkeys - return the size of the PKEY table for port 0
|
|
|
- * @dd: the infinipath device
|
|
|
- */
|
|
|
-unsigned ipath_layer_get_npkeys(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return ARRAY_SIZE(dd->ipath_pd[0]->port_pkeys);
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * ipath_layer_get_pkey - return the indexed PKEY from the port 0 PKEY table
|
|
|
- * @dd: the infinipath device
|
|
|
- * @index: the PKEY index
|
|
|
- */
|
|
|
-unsigned ipath_layer_get_pkey(struct ipath_devdata *dd, unsigned index)
|
|
|
-{
|
|
|
- unsigned ret;
|
|
|
-
|
|
|
- if (index >= ARRAY_SIZE(dd->ipath_pd[0]->port_pkeys))
|
|
|
- ret = 0;
|
|
|
- else
|
|
|
- ret = dd->ipath_pd[0]->port_pkeys[index];
|
|
|
-
|
|
|
- return ret;
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * ipath_layer_get_pkeys - return the PKEY table for port 0
|
|
|
- * @dd: the infinipath device
|
|
|
- * @pkeys: the pkey table is placed here
|
|
|
- */
|
|
|
-int ipath_layer_get_pkeys(struct ipath_devdata *dd, u16 * pkeys)
|
|
|
-{
|
|
|
- struct ipath_portdata *pd = dd->ipath_pd[0];
|
|
|
-
|
|
|
- memcpy(pkeys, pd->port_pkeys, sizeof(pd->port_pkeys));
|
|
|
-
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * rm_pkey - decrecment the reference count for the given PKEY
|
|
|
- * @dd: the infinipath device
|
|
|
- * @key: the PKEY index
|
|
|
- *
|
|
|
- * Return true if this was the last reference and the hardware table entry
|
|
|
- * needs to be changed.
|
|
|
- */
|
|
|
-static int rm_pkey(struct ipath_devdata *dd, u16 key)
|
|
|
-{
|
|
|
- int i;
|
|
|
- int ret;
|
|
|
-
|
|
|
- for (i = 0; i < ARRAY_SIZE(dd->ipath_pkeys); i++) {
|
|
|
- if (dd->ipath_pkeys[i] != key)
|
|
|
- continue;
|
|
|
- if (atomic_dec_and_test(&dd->ipath_pkeyrefs[i])) {
|
|
|
- dd->ipath_pkeys[i] = 0;
|
|
|
- ret = 1;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- break;
|
|
|
- }
|
|
|
-
|
|
|
- ret = 0;
|
|
|
-
|
|
|
-bail:
|
|
|
- return ret;
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * add_pkey - add the given PKEY to the hardware table
|
|
|
- * @dd: the infinipath device
|
|
|
- * @key: the PKEY
|
|
|
- *
|
|
|
- * Return an error code if unable to add the entry, zero if no change,
|
|
|
- * or 1 if the hardware PKEY register needs to be updated.
|
|
|
- */
|
|
|
-static int add_pkey(struct ipath_devdata *dd, u16 key)
|
|
|
-{
|
|
|
- int i;
|
|
|
- u16 lkey = key & 0x7FFF;
|
|
|
- int any = 0;
|
|
|
- int ret;
|
|
|
-
|
|
|
- if (lkey == 0x7FFF) {
|
|
|
- ret = 0;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
-
|
|
|
- /* Look for an empty slot or a matching PKEY. */
|
|
|
- for (i = 0; i < ARRAY_SIZE(dd->ipath_pkeys); i++) {
|
|
|
- if (!dd->ipath_pkeys[i]) {
|
|
|
- any++;
|
|
|
- continue;
|
|
|
- }
|
|
|
- /* If it matches exactly, try to increment the ref count */
|
|
|
- if (dd->ipath_pkeys[i] == key) {
|
|
|
- if (atomic_inc_return(&dd->ipath_pkeyrefs[i]) > 1) {
|
|
|
- ret = 0;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- /* Lost the race. Look for an empty slot below. */
|
|
|
- atomic_dec(&dd->ipath_pkeyrefs[i]);
|
|
|
- any++;
|
|
|
- }
|
|
|
- /*
|
|
|
- * It makes no sense to have both the limited and unlimited
|
|
|
- * PKEY set at the same time since the unlimited one will
|
|
|
- * disable the limited one.
|
|
|
- */
|
|
|
- if ((dd->ipath_pkeys[i] & 0x7FFF) == lkey) {
|
|
|
- ret = -EEXIST;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- }
|
|
|
- if (!any) {
|
|
|
- ret = -EBUSY;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- for (i = 0; i < ARRAY_SIZE(dd->ipath_pkeys); i++) {
|
|
|
- if (!dd->ipath_pkeys[i] &&
|
|
|
- atomic_inc_return(&dd->ipath_pkeyrefs[i]) == 1) {
|
|
|
- /* for ipathstats, etc. */
|
|
|
- ipath_stats.sps_pkeys[i] = lkey;
|
|
|
- dd->ipath_pkeys[i] = key;
|
|
|
- ret = 1;
|
|
|
- goto bail;
|
|
|
- }
|
|
|
- }
|
|
|
- ret = -EBUSY;
|
|
|
-
|
|
|
-bail:
|
|
|
- return ret;
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * ipath_layer_set_pkeys - set the PKEY table for port 0
|
|
|
- * @dd: the infinipath device
|
|
|
- * @pkeys: the PKEY table
|
|
|
- */
|
|
|
-int ipath_layer_set_pkeys(struct ipath_devdata *dd, u16 * pkeys)
|
|
|
-{
|
|
|
- struct ipath_portdata *pd;
|
|
|
- int i;
|
|
|
- int changed = 0;
|
|
|
-
|
|
|
- pd = dd->ipath_pd[0];
|
|
|
-
|
|
|
- for (i = 0; i < ARRAY_SIZE(pd->port_pkeys); i++) {
|
|
|
- u16 key = pkeys[i];
|
|
|
- u16 okey = pd->port_pkeys[i];
|
|
|
-
|
|
|
- if (key == okey)
|
|
|
- continue;
|
|
|
- /*
|
|
|
- * The value of this PKEY table entry is changing.
|
|
|
- * Remove the old entry in the hardware's array of PKEYs.
|
|
|
- */
|
|
|
- if (okey & 0x7FFF)
|
|
|
- changed |= rm_pkey(dd, okey);
|
|
|
- if (key & 0x7FFF) {
|
|
|
- int ret = add_pkey(dd, key);
|
|
|
-
|
|
|
- if (ret < 0)
|
|
|
- key = 0;
|
|
|
- else
|
|
|
- changed |= ret;
|
|
|
- }
|
|
|
- pd->port_pkeys[i] = key;
|
|
|
- }
|
|
|
- if (changed) {
|
|
|
- u64 pkey;
|
|
|
-
|
|
|
- pkey = (u64) dd->ipath_pkeys[0] |
|
|
|
- ((u64) dd->ipath_pkeys[1] << 16) |
|
|
|
- ((u64) dd->ipath_pkeys[2] << 32) |
|
|
|
- ((u64) dd->ipath_pkeys[3] << 48);
|
|
|
- ipath_cdbg(VERBOSE, "p0 new pkey reg %llx\n",
|
|
|
- (unsigned long long) pkey);
|
|
|
- ipath_write_kreg(dd, dd->ipath_kregs->kr_partitionkey,
|
|
|
- pkey);
|
|
|
- }
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * ipath_layer_get_linkdowndefaultstate - get the default linkdown state
|
|
|
- * @dd: the infinipath device
|
|
|
- *
|
|
|
- * Returns zero if the default is POLL, 1 if the default is SLEEP.
|
|
|
- */
|
|
|
-int ipath_layer_get_linkdowndefaultstate(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return !!(dd->ipath_ibcctrl & INFINIPATH_IBCC_LINKDOWNDEFAULTSTATE);
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * ipath_layer_set_linkdowndefaultstate - set the default linkdown state
|
|
|
- * @dd: the infinipath device
|
|
|
- * @sleep: the new state
|
|
|
- *
|
|
|
- * Note that this will only take effect when the link state changes.
|
|
|
- */
|
|
|
-int ipath_layer_set_linkdowndefaultstate(struct ipath_devdata *dd,
|
|
|
- int sleep)
|
|
|
-{
|
|
|
- if (sleep)
|
|
|
- dd->ipath_ibcctrl |= INFINIPATH_IBCC_LINKDOWNDEFAULTSTATE;
|
|
|
- else
|
|
|
- dd->ipath_ibcctrl &= ~INFINIPATH_IBCC_LINKDOWNDEFAULTSTATE;
|
|
|
- ipath_write_kreg(dd, dd->ipath_kregs->kr_ibcctrl,
|
|
|
- dd->ipath_ibcctrl);
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-int ipath_layer_get_phyerrthreshold(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return (dd->ipath_ibcctrl >>
|
|
|
- INFINIPATH_IBCC_PHYERRTHRESHOLD_SHIFT) &
|
|
|
- INFINIPATH_IBCC_PHYERRTHRESHOLD_MASK;
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * ipath_layer_set_phyerrthreshold - set the physical error threshold
|
|
|
- * @dd: the infinipath device
|
|
|
- * @n: the new threshold
|
|
|
- *
|
|
|
- * Note that this will only take effect when the link state changes.
|
|
|
- */
|
|
|
-int ipath_layer_set_phyerrthreshold(struct ipath_devdata *dd, unsigned n)
|
|
|
-{
|
|
|
- unsigned v;
|
|
|
-
|
|
|
- v = (dd->ipath_ibcctrl >> INFINIPATH_IBCC_PHYERRTHRESHOLD_SHIFT) &
|
|
|
- INFINIPATH_IBCC_PHYERRTHRESHOLD_MASK;
|
|
|
- if (v != n) {
|
|
|
- dd->ipath_ibcctrl &=
|
|
|
- ~(INFINIPATH_IBCC_PHYERRTHRESHOLD_MASK <<
|
|
|
- INFINIPATH_IBCC_PHYERRTHRESHOLD_SHIFT);
|
|
|
- dd->ipath_ibcctrl |=
|
|
|
- (u64) n << INFINIPATH_IBCC_PHYERRTHRESHOLD_SHIFT;
|
|
|
- ipath_write_kreg(dd, dd->ipath_kregs->kr_ibcctrl,
|
|
|
- dd->ipath_ibcctrl);
|
|
|
- }
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-int ipath_layer_get_overrunthreshold(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return (dd->ipath_ibcctrl >>
|
|
|
- INFINIPATH_IBCC_OVERRUNTHRESHOLD_SHIFT) &
|
|
|
- INFINIPATH_IBCC_OVERRUNTHRESHOLD_MASK;
|
|
|
-}
|
|
|
-
|
|
|
-/**
|
|
|
- * ipath_layer_set_overrunthreshold - set the overrun threshold
|
|
|
- * @dd: the infinipath device
|
|
|
- * @n: the new threshold
|
|
|
- *
|
|
|
- * Note that this will only take effect when the link state changes.
|
|
|
- */
|
|
|
-int ipath_layer_set_overrunthreshold(struct ipath_devdata *dd, unsigned n)
|
|
|
-{
|
|
|
- unsigned v;
|
|
|
-
|
|
|
- v = (dd->ipath_ibcctrl >> INFINIPATH_IBCC_OVERRUNTHRESHOLD_SHIFT) &
|
|
|
- INFINIPATH_IBCC_OVERRUNTHRESHOLD_MASK;
|
|
|
- if (v != n) {
|
|
|
- dd->ipath_ibcctrl &=
|
|
|
- ~(INFINIPATH_IBCC_OVERRUNTHRESHOLD_MASK <<
|
|
|
- INFINIPATH_IBCC_OVERRUNTHRESHOLD_SHIFT);
|
|
|
- dd->ipath_ibcctrl |=
|
|
|
- (u64) n << INFINIPATH_IBCC_OVERRUNTHRESHOLD_SHIFT;
|
|
|
- ipath_write_kreg(dd, dd->ipath_kregs->kr_ibcctrl,
|
|
|
- dd->ipath_ibcctrl);
|
|
|
- }
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-int ipath_layer_get_boardname(struct ipath_devdata *dd, char *name,
|
|
|
- size_t namelen)
|
|
|
-{
|
|
|
- return dd->ipath_f_get_boardname(dd, name, namelen);
|
|
|
-}
|
|
|
-
|
|
|
-u32 ipath_layer_get_rcvhdrentsize(struct ipath_devdata *dd)
|
|
|
-{
|
|
|
- return dd->ipath_rcvhdrentsize;
|
|
|
-}
|