|
@@ -28,6 +28,7 @@
|
|
|
#include "drm.h"
|
|
|
#include "nouveau_drv.h"
|
|
|
#include "nouveau_ramht.h"
|
|
|
+#include "nouveau_util.h"
|
|
|
|
|
|
#define NV04_RAMFC(c) (dev_priv->ramfc->pinst + ((c) * NV04_RAMFC__SIZE))
|
|
|
#define NV04_RAMFC__SIZE 32
|
|
@@ -284,6 +285,7 @@ nv04_fifo_init_ramxx(struct drm_device *dev)
|
|
|
static void
|
|
|
nv04_fifo_init_intr(struct drm_device *dev)
|
|
|
{
|
|
|
+ nouveau_irq_register(dev, 8, nv04_fifo_isr);
|
|
|
nv_wr32(dev, 0x002100, 0xffffffff);
|
|
|
nv_wr32(dev, 0x002140, 0xffffffff);
|
|
|
}
|
|
@@ -315,3 +317,207 @@ nv04_fifo_init(struct drm_device *dev)
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
+void
|
|
|
+nv04_fifo_fini(struct drm_device *dev)
|
|
|
+{
|
|
|
+ nv_wr32(dev, 0x2140, 0x00000000);
|
|
|
+ nouveau_irq_unregister(dev, 8);
|
|
|
+}
|
|
|
+
|
|
|
+static bool
|
|
|
+nouveau_fifo_swmthd(struct drm_device *dev, u32 chid, u32 addr, u32 data)
|
|
|
+{
|
|
|
+ struct drm_nouveau_private *dev_priv = dev->dev_private;
|
|
|
+ struct nouveau_channel *chan = NULL;
|
|
|
+ struct nouveau_gpuobj *obj;
|
|
|
+ unsigned long flags;
|
|
|
+ const int subc = (addr >> 13) & 0x7;
|
|
|
+ const int mthd = addr & 0x1ffc;
|
|
|
+ bool handled = false;
|
|
|
+ u32 engine;
|
|
|
+
|
|
|
+ spin_lock_irqsave(&dev_priv->channels.lock, flags);
|
|
|
+ if (likely(chid >= 0 && chid < dev_priv->engine.fifo.channels))
|
|
|
+ chan = dev_priv->channels.ptr[chid];
|
|
|
+ if (unlikely(!chan))
|
|
|
+ goto out;
|
|
|
+
|
|
|
+ switch (mthd) {
|
|
|
+ case 0x0000: /* bind object to subchannel */
|
|
|
+ obj = nouveau_ramht_find(chan, data);
|
|
|
+ if (unlikely(!obj || obj->engine != NVOBJ_ENGINE_SW))
|
|
|
+ break;
|
|
|
+
|
|
|
+ chan->sw_subchannel[subc] = obj->class;
|
|
|
+ engine = 0x0000000f << (subc * 4);
|
|
|
+
|
|
|
+ nv_mask(dev, NV04_PFIFO_CACHE1_ENGINE, engine, 0x00000000);
|
|
|
+ handled = true;
|
|
|
+ break;
|
|
|
+ default:
|
|
|
+ engine = nv_rd32(dev, NV04_PFIFO_CACHE1_ENGINE);
|
|
|
+ if (unlikely(((engine >> (subc * 4)) & 0xf) != 0))
|
|
|
+ break;
|
|
|
+
|
|
|
+ if (!nouveau_gpuobj_mthd_call(chan, chan->sw_subchannel[subc],
|
|
|
+ mthd, data))
|
|
|
+ handled = true;
|
|
|
+ break;
|
|
|
+ }
|
|
|
+
|
|
|
+out:
|
|
|
+ spin_unlock_irqrestore(&dev_priv->channels.lock, flags);
|
|
|
+ return handled;
|
|
|
+}
|
|
|
+
|
|
|
+void
|
|
|
+nv04_fifo_isr(struct drm_device *dev)
|
|
|
+{
|
|
|
+ struct drm_nouveau_private *dev_priv = dev->dev_private;
|
|
|
+ struct nouveau_engine *engine = &dev_priv->engine;
|
|
|
+ uint32_t status, reassign;
|
|
|
+ int cnt = 0;
|
|
|
+
|
|
|
+ reassign = nv_rd32(dev, NV03_PFIFO_CACHES) & 1;
|
|
|
+ while ((status = nv_rd32(dev, NV03_PFIFO_INTR_0)) && (cnt++ < 100)) {
|
|
|
+ uint32_t chid, get;
|
|
|
+
|
|
|
+ nv_wr32(dev, NV03_PFIFO_CACHES, 0);
|
|
|
+
|
|
|
+ chid = engine->fifo.channel_id(dev);
|
|
|
+ get = nv_rd32(dev, NV03_PFIFO_CACHE1_GET);
|
|
|
+
|
|
|
+ if (status & NV_PFIFO_INTR_CACHE_ERROR) {
|
|
|
+ uint32_t mthd, data;
|
|
|
+ int ptr;
|
|
|
+
|
|
|
+ /* NV_PFIFO_CACHE1_GET actually goes to 0xffc before
|
|
|
+ * wrapping on my G80 chips, but CACHE1 isn't big
|
|
|
+ * enough for this much data.. Tests show that it
|
|
|
+ * wraps around to the start at GET=0x800.. No clue
|
|
|
+ * as to why..
|
|
|
+ */
|
|
|
+ ptr = (get & 0x7ff) >> 2;
|
|
|
+
|
|
|
+ if (dev_priv->card_type < NV_40) {
|
|
|
+ mthd = nv_rd32(dev,
|
|
|
+ NV04_PFIFO_CACHE1_METHOD(ptr));
|
|
|
+ data = nv_rd32(dev,
|
|
|
+ NV04_PFIFO_CACHE1_DATA(ptr));
|
|
|
+ } else {
|
|
|
+ mthd = nv_rd32(dev,
|
|
|
+ NV40_PFIFO_CACHE1_METHOD(ptr));
|
|
|
+ data = nv_rd32(dev,
|
|
|
+ NV40_PFIFO_CACHE1_DATA(ptr));
|
|
|
+ }
|
|
|
+
|
|
|
+ if (!nouveau_fifo_swmthd(dev, chid, mthd, data)) {
|
|
|
+ NV_INFO(dev, "PFIFO_CACHE_ERROR - Ch %d/%d "
|
|
|
+ "Mthd 0x%04x Data 0x%08x\n",
|
|
|
+ chid, (mthd >> 13) & 7, mthd & 0x1ffc,
|
|
|
+ data);
|
|
|
+ }
|
|
|
+
|
|
|
+ nv_wr32(dev, NV04_PFIFO_CACHE1_DMA_PUSH, 0);
|
|
|
+ nv_wr32(dev, NV03_PFIFO_INTR_0,
|
|
|
+ NV_PFIFO_INTR_CACHE_ERROR);
|
|
|
+
|
|
|
+ nv_wr32(dev, NV03_PFIFO_CACHE1_PUSH0,
|
|
|
+ nv_rd32(dev, NV03_PFIFO_CACHE1_PUSH0) & ~1);
|
|
|
+ nv_wr32(dev, NV03_PFIFO_CACHE1_GET, get + 4);
|
|
|
+ nv_wr32(dev, NV03_PFIFO_CACHE1_PUSH0,
|
|
|
+ nv_rd32(dev, NV03_PFIFO_CACHE1_PUSH0) | 1);
|
|
|
+ nv_wr32(dev, NV04_PFIFO_CACHE1_HASH, 0);
|
|
|
+
|
|
|
+ nv_wr32(dev, NV04_PFIFO_CACHE1_DMA_PUSH,
|
|
|
+ nv_rd32(dev, NV04_PFIFO_CACHE1_DMA_PUSH) | 1);
|
|
|
+ nv_wr32(dev, NV04_PFIFO_CACHE1_PULL0, 1);
|
|
|
+
|
|
|
+ status &= ~NV_PFIFO_INTR_CACHE_ERROR;
|
|
|
+ }
|
|
|
+
|
|
|
+ if (status & NV_PFIFO_INTR_DMA_PUSHER) {
|
|
|
+ u32 dma_get = nv_rd32(dev, 0x003244);
|
|
|
+ u32 dma_put = nv_rd32(dev, 0x003240);
|
|
|
+ u32 push = nv_rd32(dev, 0x003220);
|
|
|
+ u32 state = nv_rd32(dev, 0x003228);
|
|
|
+
|
|
|
+ if (dev_priv->card_type == NV_50) {
|
|
|
+ u32 ho_get = nv_rd32(dev, 0x003328);
|
|
|
+ u32 ho_put = nv_rd32(dev, 0x003320);
|
|
|
+ u32 ib_get = nv_rd32(dev, 0x003334);
|
|
|
+ u32 ib_put = nv_rd32(dev, 0x003330);
|
|
|
+
|
|
|
+ if (nouveau_ratelimit())
|
|
|
+ NV_INFO(dev, "PFIFO_DMA_PUSHER - Ch %d Get 0x%02x%08x "
|
|
|
+ "Put 0x%02x%08x IbGet 0x%08x IbPut 0x%08x "
|
|
|
+ "State 0x%08x Push 0x%08x\n",
|
|
|
+ chid, ho_get, dma_get, ho_put,
|
|
|
+ dma_put, ib_get, ib_put, state,
|
|
|
+ push);
|
|
|
+
|
|
|
+ /* METHOD_COUNT, in DMA_STATE on earlier chipsets */
|
|
|
+ nv_wr32(dev, 0x003364, 0x00000000);
|
|
|
+ if (dma_get != dma_put || ho_get != ho_put) {
|
|
|
+ nv_wr32(dev, 0x003244, dma_put);
|
|
|
+ nv_wr32(dev, 0x003328, ho_put);
|
|
|
+ } else
|
|
|
+ if (ib_get != ib_put) {
|
|
|
+ nv_wr32(dev, 0x003334, ib_put);
|
|
|
+ }
|
|
|
+ } else {
|
|
|
+ NV_INFO(dev, "PFIFO_DMA_PUSHER - Ch %d Get 0x%08x "
|
|
|
+ "Put 0x%08x State 0x%08x Push 0x%08x\n",
|
|
|
+ chid, dma_get, dma_put, state, push);
|
|
|
+
|
|
|
+ if (dma_get != dma_put)
|
|
|
+ nv_wr32(dev, 0x003244, dma_put);
|
|
|
+ }
|
|
|
+
|
|
|
+ nv_wr32(dev, 0x003228, 0x00000000);
|
|
|
+ nv_wr32(dev, 0x003220, 0x00000001);
|
|
|
+ nv_wr32(dev, 0x002100, NV_PFIFO_INTR_DMA_PUSHER);
|
|
|
+ status &= ~NV_PFIFO_INTR_DMA_PUSHER;
|
|
|
+ }
|
|
|
+
|
|
|
+ if (status & NV_PFIFO_INTR_SEMAPHORE) {
|
|
|
+ uint32_t sem;
|
|
|
+
|
|
|
+ status &= ~NV_PFIFO_INTR_SEMAPHORE;
|
|
|
+ nv_wr32(dev, NV03_PFIFO_INTR_0,
|
|
|
+ NV_PFIFO_INTR_SEMAPHORE);
|
|
|
+
|
|
|
+ sem = nv_rd32(dev, NV10_PFIFO_CACHE1_SEMAPHORE);
|
|
|
+ nv_wr32(dev, NV10_PFIFO_CACHE1_SEMAPHORE, sem | 0x1);
|
|
|
+
|
|
|
+ nv_wr32(dev, NV03_PFIFO_CACHE1_GET, get + 4);
|
|
|
+ nv_wr32(dev, NV04_PFIFO_CACHE1_PULL0, 1);
|
|
|
+ }
|
|
|
+
|
|
|
+ if (dev_priv->card_type == NV_50) {
|
|
|
+ if (status & 0x00000010) {
|
|
|
+ nv50_fb_vm_trap(dev, 1, "PFIFO_BAR_FAULT");
|
|
|
+ status &= ~0x00000010;
|
|
|
+ nv_wr32(dev, 0x002100, 0x00000010);
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ if (status) {
|
|
|
+ if (nouveau_ratelimit())
|
|
|
+ NV_INFO(dev, "PFIFO_INTR 0x%08x - Ch %d\n",
|
|
|
+ status, chid);
|
|
|
+ nv_wr32(dev, NV03_PFIFO_INTR_0, status);
|
|
|
+ status = 0;
|
|
|
+ }
|
|
|
+
|
|
|
+ nv_wr32(dev, NV03_PFIFO_CACHES, reassign);
|
|
|
+ }
|
|
|
+
|
|
|
+ if (status) {
|
|
|
+ NV_INFO(dev, "PFIFO still angry after %d spins, halt\n", cnt);
|
|
|
+ nv_wr32(dev, 0x2140, 0);
|
|
|
+ nv_wr32(dev, 0x140, 0);
|
|
|
+ }
|
|
|
+
|
|
|
+ nv_wr32(dev, NV03_PMC_INTR_0, NV_PMC_INTR_0_PFIFO_PENDING);
|
|
|
+}
|