Quellcodebibliothek Statistik Leitseite products/Sources/formale Sprachen/C/Linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/   (Open Source Betriebssystem Version 6.17.9©)  Datei vom 24.10.2025 mit Größe 32 kB image not shown  

Quelle  nv10.c   Sprache: C

 
/*
 * Copyright 2007 Matthieu CASTET <castet.matthieu@free.fr>
 * All Rights Reserved.
 *
 * Permission is hereby granted, free of charge, to any person obtaining a
 * copy of this software and associated documentation files (the "Software"),
 * to deal in the Software without restriction, including without limitation
 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
 * and/or sell copies of the Software, and to permit persons to whom the
 * Software is furnished to do so, subject to the following conditions:
 *
 * The above copyright notice and this permission notice (including the next
 * paragr) shall be included in all copies or substantial portions of the
 * Software.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
 * PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
 * DEALINGS IN THE SOFTWARE.
 */

#include "nv10.h"
#include "regs.h"

#include <core/client.h>
#include <core/gpuobj.h>
#include <engine/fifo.h>
#include <engine/fifo/chan.h>
#include <subdev/fb.h>

struct pipe_state {
 u32 pipe_0x0000[0x040/4];
 u32 pipe_0x0040[0x010/4];
 u32 pipe_0x0200[0x0c0/4];
 u32 pipe_0x4400[0x080/4];
 u32 pipe_0x6400[0x3b0/4];
 u32 pipe_0x6800[0x2f0/4];
 u32 pipe_0x6c00[0x030/4];
 u32 pipe_0x7000[0x130/4];
 u32 pipe_0x7400[0x0c0/4];
 u32 pipe_0x7800[0x0c0/4];
};

static int nv10_gr_ctx_regs[] = {
 NV10_PGRAPH_CTX_SWITCH(0),
 NV10_PGRAPH_CTX_SWITCH(1),
 NV10_PGRAPH_CTX_SWITCH(2),
 NV10_PGRAPH_CTX_SWITCH(3),
 NV10_PGRAPH_CTX_SWITCH(4),
 NV10_PGRAPH_CTX_CACHE(0, 0),
 NV10_PGRAPH_CTX_CACHE(0, 1),
 NV10_PGRAPH_CTX_CACHE(0, 2),
 NV10_PGRAPH_CTX_CACHE(0, 3),
 NV10_PGRAPH_CTX_CACHE(0, 4),
 NV10_PGRAPH_CTX_CACHE(1, 0),
 NV10_PGRAPH_CTX_CACHE(1, 1),
 NV10_PGRAPH_CTX_CACHE(1, 2),
 NV10_PGRAPH_CTX_CACHE(1, 3),
 NV10_PGRAPH_CTX_CACHE(1, 4),
 NV10_PGRAPH_CTX_CACHE(2, 0),
 NV10_PGRAPH_CTX_CACHE(2, 1),
 NV10_PGRAPH_CTX_CACHE(2, 2),
 NV10_PGRAPH_CTX_CACHE(2, 3),
 NV10_PGRAPH_CTX_CACHE(2, 4),
 NV10_PGRAPH_CTX_CACHE(3, 0),
 NV10_PGRAPH_CTX_CACHE(3, 1),
 NV10_PGRAPH_CTX_CACHE(3, 2),
 NV10_PGRAPH_CTX_CACHE(3, 3),
 NV10_PGRAPH_CTX_CACHE(3, 4),
 NV10_PGRAPH_CTX_CACHE(4, 0),
 NV10_PGRAPH_CTX_CACHE(4, 1),
 NV10_PGRAPH_CTX_CACHE(4, 2),
 NV10_PGRAPH_CTX_CACHE(4, 3),
 NV10_PGRAPH_CTX_CACHE(4, 4),
 NV10_PGRAPH_CTX_CACHE(5, 0),
 NV10_PGRAPH_CTX_CACHE(5, 1),
 NV10_PGRAPH_CTX_CACHE(5, 2),
 NV10_PGRAPH_CTX_CACHE(5, 3),
 NV10_PGRAPH_CTX_CACHE(5, 4),
 NV10_PGRAPH_CTX_CACHE(6, 0),
 NV10_PGRAPH_CTX_CACHE(6, 1),
 NV10_PGRAPH_CTX_CACHE(6, 2),
 NV10_PGRAPH_CTX_CACHE(6, 3),
 NV10_PGRAPH_CTX_CACHE(6, 4),
 NV10_PGRAPH_CTX_CACHE(7, 0),
 NV10_PGRAPH_CTX_CACHE(7, 1),
 NV10_PGRAPH_CTX_CACHE(7, 2),
 NV10_PGRAPH_CTX_CACHE(7, 3),
 NV10_PGRAPH_CTX_CACHE(7, 4),
 NV10_PGRAPH_CTX_USER,
 NV04_PGRAPH_DMA_START_0,
 NV04_PGRAPH_DMA_START_1,
 NV04_PGRAPH_DMA_LENGTH,
 NV04_PGRAPH_DMA_MISC,
 NV10_PGRAPH_DMA_PITCH,
 NV04_PGRAPH_BOFFSET0,
 NV04_PGRAPH_BBASE0,
 NV04_PGRAPH_BLIMIT0,
 NV04_PGRAPH_BOFFSET1,
 NV04_PGRAPH_BBASE1,
 NV04_PGRAPH_BLIMIT1,
 NV04_PGRAPH_BOFFSET2,
 NV04_PGRAPH_BBASE2,
 NV04_PGRAPH_BLIMIT2,
 NV04_PGRAPH_BOFFSET3,
 NV04_PGRAPH_BBASE3,
 NV04_PGRAPH_BLIMIT3,
 NV04_PGRAPH_BOFFSET4,
 NV04_PGRAPH_BBASE4,
 NV04_PGRAPH_BLIMIT4,
 NV04_PGRAPH_BOFFSET5,
 NV04_PGRAPH_BBASE5,
 NV04_PGRAPH_BLIMIT5,
 NV04_PGRAPH_BPITCH0,
 NV04_PGRAPH_BPITCH1,
 NV04_PGRAPH_BPITCH2,
 NV04_PGRAPH_BPITCH3,
 NV04_PGRAPH_BPITCH4,
 NV10_PGRAPH_SURFACE,
 NV10_PGRAPH_STATE,
 NV04_PGRAPH_BSWIZZLE2,
 NV04_PGRAPH_BSWIZZLE5,
 NV04_PGRAPH_BPIXEL,
 NV10_PGRAPH_NOTIFY,
 NV04_PGRAPH_PATT_COLOR0,
 NV04_PGRAPH_PATT_COLOR1,
 NV04_PGRAPH_PATT_COLORRAM, /* 64 values from 0x400900 to 0x4009fc */
 0x00400904,
 0x00400908,
 0x0040090c,
 0x00400910,
 0x00400914,
 0x00400918,
 0x0040091c,
 0x00400920,
 0x00400924,
 0x00400928,
 0x0040092c,
 0x00400930,
 0x00400934,
 0x00400938,
 0x0040093c,
 0x00400940,
 0x00400944,
 0x00400948,
 0x0040094c,
 0x00400950,
 0x00400954,
 0x00400958,
 0x0040095c,
 0x00400960,
 0x00400964,
 0x00400968,
 0x0040096c,
 0x00400970,
 0x00400974,
 0x00400978,
 0x0040097c,
 0x00400980,
 0x00400984,
 0x00400988,
 0x0040098c,
 0x00400990,
 0x00400994,
 0x00400998,
 0x0040099c,
 0x004009a0,
 0x004009a4,
 0x004009a8,
 0x004009ac,
 0x004009b0,
 0x004009b4,
 0x004009b8,
 0x004009bc,
 0x004009c0,
 0x004009c4,
 0x004009c8,
 0x004009cc,
 0x004009d0,
 0x004009d4,
 0x004009d8,
 0x004009dc,
 0x004009e0,
 0x004009e4,
 0x004009e8,
 0x004009ec,
 0x004009f0,
 0x004009f4,
 0x004009f8,
 0x004009fc,
 NV04_PGRAPH_PATTERN, /* 2 values from 0x400808 to 0x40080c */
 0x0040080c,
 NV04_PGRAPH_PATTERN_SHAPE,
 NV03_PGRAPH_MONO_COLOR0,
 NV04_PGRAPH_ROP3,
 NV04_PGRAPH_CHROMA,
 NV04_PGRAPH_BETA_AND,
 NV04_PGRAPH_BETA_PREMULT,
 0x00400e70,
 0x00400e74,
 0x00400e78,
 0x00400e7c,
 0x00400e80,
 0x00400e84,
 0x00400e88,
 0x00400e8c,
 0x00400ea0,
 0x00400ea4,
 0x00400ea8,
 0x00400e90,
 0x00400e94,
 0x00400e98,
 0x00400e9c,
 NV10_PGRAPH_WINDOWCLIP_HORIZONTAL, /* 8 values from 0x400f00-0x400f1c */
 NV10_PGRAPH_WINDOWCLIP_VERTICAL,   /* 8 values from 0x400f20-0x400f3c */
 0x00400f04,
 0x00400f24,
 0x00400f08,
 0x00400f28,
 0x00400f0c,
 0x00400f2c,
 0x00400f10,
 0x00400f30,
 0x00400f14,
 0x00400f34,
 0x00400f18,
 0x00400f38,
 0x00400f1c,
 0x00400f3c,
 NV10_PGRAPH_XFMODE0,
 NV10_PGRAPH_XFMODE1,
 NV10_PGRAPH_GLOBALSTATE0,
 NV10_PGRAPH_GLOBALSTATE1,
 NV04_PGRAPH_STORED_FMT,
 NV04_PGRAPH_SOURCE_COLOR,
 NV03_PGRAPH_ABS_X_RAM, /* 32 values from 0x400400 to 0x40047c */
 NV03_PGRAPH_ABS_Y_RAM, /* 32 values from 0x400480 to 0x4004fc */
 0x00400404,
 0x00400484,
 0x00400408,
 0x00400488,
 0x0040040c,
 0x0040048c,
 0x00400410,
 0x00400490,
 0x00400414,
 0x00400494,
 0x00400418,
 0x00400498,
 0x0040041c,
 0x0040049c,
 0x00400420,
 0x004004a0,
 0x00400424,
 0x004004a4,
 0x00400428,
 0x004004a8,
 0x0040042c,
 0x004004ac,
 0x00400430,
 0x004004b0,
 0x00400434,
 0x004004b4,
 0x00400438,
 0x004004b8,
 0x0040043c,
 0x004004bc,
 0x00400440,
 0x004004c0,
 0x00400444,
 0x004004c4,
 0x00400448,
 0x004004c8,
 0x0040044c,
 0x004004cc,
 0x00400450,
 0x004004d0,
 0x00400454,
 0x004004d4,
 0x00400458,
 0x004004d8,
 0x0040045c,
 0x004004dc,
 0x00400460,
 0x004004e0,
 0x00400464,
 0x004004e4,
 0x00400468,
 0x004004e8,
 0x0040046c,
 0x004004ec,
 0x00400470,
 0x004004f0,
 0x00400474,
 0x004004f4,
 0x00400478,
 0x004004f8,
 0x0040047c,
 0x004004fc,
 NV03_PGRAPH_ABS_UCLIP_XMIN,
 NV03_PGRAPH_ABS_UCLIP_XMAX,
 NV03_PGRAPH_ABS_UCLIP_YMIN,
 NV03_PGRAPH_ABS_UCLIP_YMAX,
 0x00400550,
 0x00400558,
 0x00400554,
 0x0040055c,
 NV03_PGRAPH_ABS_UCLIPA_XMIN,
 NV03_PGRAPH_ABS_UCLIPA_XMAX,
 NV03_PGRAPH_ABS_UCLIPA_YMIN,
 NV03_PGRAPH_ABS_UCLIPA_YMAX,
 NV03_PGRAPH_ABS_ICLIP_XMAX,
 NV03_PGRAPH_ABS_ICLIP_YMAX,
 NV03_PGRAPH_XY_LOGIC_MISC0,
 NV03_PGRAPH_XY_LOGIC_MISC1,
 NV03_PGRAPH_XY_LOGIC_MISC2,
 NV03_PGRAPH_XY_LOGIC_MISC3,
 NV03_PGRAPH_CLIPX_0,
 NV03_PGRAPH_CLIPX_1,
 NV03_PGRAPH_CLIPY_0,
 NV03_PGRAPH_CLIPY_1,
 NV10_PGRAPH_COMBINER0_IN_ALPHA,
 NV10_PGRAPH_COMBINER1_IN_ALPHA,
 NV10_PGRAPH_COMBINER0_IN_RGB,
 NV10_PGRAPH_COMBINER1_IN_RGB,
 NV10_PGRAPH_COMBINER_COLOR0,
 NV10_PGRAPH_COMBINER_COLOR1,
 NV10_PGRAPH_COMBINER0_OUT_ALPHA,
 NV10_PGRAPH_COMBINER1_OUT_ALPHA,
 NV10_PGRAPH_COMBINER0_OUT_RGB,
 NV10_PGRAPH_COMBINER1_OUT_RGB,
 NV10_PGRAPH_COMBINER_FINAL0,
 NV10_PGRAPH_COMBINER_FINAL1,
 0x00400e00,
 0x00400e04,
 0x00400e08,
 0x00400e0c,
 0x00400e10,
 0x00400e14,
 0x00400e18,
 0x00400e1c,
 0x00400e20,
 0x00400e24,
 0x00400e28,
 0x00400e2c,
 0x00400e30,
 0x00400e34,
 0x00400e38,
 0x00400e3c,
 NV04_PGRAPH_PASSTHRU_0,
 NV04_PGRAPH_PASSTHRU_1,
 NV04_PGRAPH_PASSTHRU_2,
 NV10_PGRAPH_DIMX_TEXTURE,
 NV10_PGRAPH_WDIMX_TEXTURE,
 NV10_PGRAPH_DVD_COLORFMT,
 NV10_PGRAPH_SCALED_FORMAT,
 NV04_PGRAPH_MISC24_0,
 NV04_PGRAPH_MISC24_1,
 NV04_PGRAPH_MISC24_2,
 NV03_PGRAPH_X_MISC,
 NV03_PGRAPH_Y_MISC,
 NV04_PGRAPH_VALID1,
 NV04_PGRAPH_VALID2,
};

static int nv17_gr_ctx_regs[] = {
 NV10_PGRAPH_DEBUG_4,
 0x004006b0,
 0x00400eac,
 0x00400eb0,
 0x00400eb4,
 0x00400eb8,
 0x00400ebc,
 0x00400ec0,
 0x00400ec4,
 0x00400ec8,
 0x00400ecc,
 0x00400ed0,
 0x00400ed4,
 0x00400ed8,
 0x00400edc,
 0x00400ee0,
 0x00400a00,
 0x00400a04,
};

#define nv10_gr(p) container_of((p), struct nv10_gr, base)

struct nv10_gr {
 struct nvkm_gr base;
 struct nv10_gr_chan *chan[32];
 spinlock_t lock;
};

#define nv10_gr_chan(p) container_of((p), struct nv10_gr_chan, object)

struct nv10_gr_chan {
 struct nvkm_object object;
 struct nv10_gr *gr;
 int chid;
 int nv10[ARRAY_SIZE(nv10_gr_ctx_regs)];
 int nv17[ARRAY_SIZE(nv17_gr_ctx_regs)];
 struct pipe_state pipe_state;
 u32 lma_window[4];
};


/*******************************************************************************
 * Graphics object classes
 ******************************************************************************/


#define PIPE_SAVE(gr, state, addr)     \
 do {        \
  int __i;      \
  nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, addr);  \
  for (__i = 0; __i < ARRAY_SIZE(state); __i++)  \
   state[__i] = nvkm_rd32(device, NV10_PGRAPH_PIPE_DATA); \
 } while (0)

#define PIPE_RESTORE(gr, state, addr)     \
 do {        \
  int __i;      \
  nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, addr);  \
  for (__i = 0; __i < ARRAY_SIZE(state); __i++)  \
   nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, state[__i]); \
 } while (0)

static void
nv17_gr_mthd_lma_window(struct nv10_gr_chan *chan, u32 mthd, u32 data)
{
 struct nvkm_device *device = chan->object.engine->subdev.device;
 struct nvkm_gr *gr = &chan->gr->base;
 struct pipe_state *pipe = &chan->pipe_state;
 u32 pipe_0x0040[1], pipe_0x64c0[8], pipe_0x6a80[3], pipe_0x6ab0[3];
 u32 xfmode0, xfmode1;
 int i;

 chan->lma_window[(mthd - 0x1638) / 4] = data;

 if (mthd != 0x1644)
  return;

 nv04_gr_idle(gr);

 PIPE_SAVE(device, pipe_0x0040, 0x0040);
 PIPE_SAVE(device, pipe->pipe_0x0200, 0x0200);

 PIPE_RESTORE(device, chan->lma_window, 0x6790);

 nv04_gr_idle(gr);

 xfmode0 = nvkm_rd32(device, NV10_PGRAPH_XFMODE0);
 xfmode1 = nvkm_rd32(device, NV10_PGRAPH_XFMODE1);

 PIPE_SAVE(device, pipe->pipe_0x4400, 0x4400);
 PIPE_SAVE(device, pipe_0x64c0, 0x64c0);
 PIPE_SAVE(device, pipe_0x6ab0, 0x6ab0);
 PIPE_SAVE(device, pipe_0x6a80, 0x6a80);

 nv04_gr_idle(gr);

 nvkm_wr32(device, NV10_PGRAPH_XFMODE0, 0x10000000);
 nvkm_wr32(device, NV10_PGRAPH_XFMODE1, 0x00000000);
 nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x000064c0);
 for (i = 0; i < 4; i++)
  nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000);
 for (i = 0; i < 4; i++)
  nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000);

 nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006ab0);
 for (i = 0; i < 3; i++)
  nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000);

 nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006a80);
 for (i = 0; i < 3; i++)
  nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000);

 nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00000040);
 nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000008);

 PIPE_RESTORE(device, pipe->pipe_0x0200, 0x0200);

 nv04_gr_idle(gr);

 PIPE_RESTORE(device, pipe_0x0040, 0x0040);

 nvkm_wr32(device, NV10_PGRAPH_XFMODE0, xfmode0);
 nvkm_wr32(device, NV10_PGRAPH_XFMODE1, xfmode1);

 PIPE_RESTORE(device, pipe_0x64c0, 0x64c0);
 PIPE_RESTORE(device, pipe_0x6ab0, 0x6ab0);
 PIPE_RESTORE(device, pipe_0x6a80, 0x6a80);
 PIPE_RESTORE(device, pipe->pipe_0x4400, 0x4400);

 nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x000000c0);
 nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000);

 nv04_gr_idle(gr);
}

static void
nv17_gr_mthd_lma_enable(struct nv10_gr_chan *chan, u32 mthd, u32 data)
{
 struct nvkm_device *device = chan->object.engine->subdev.device;
 struct nvkm_gr *gr = &chan->gr->base;

 nv04_gr_idle(gr);

 nvkm_mask(device, NV10_PGRAPH_DEBUG_4, 0x00000100, 0x00000100);
 nvkm_mask(device, 0x4006b0, 0x08000000, 0x08000000);
}

static bool
nv17_gr_mthd_celcius(struct nv10_gr_chan *chan, u32 mthd, u32 data)
{
 void (*func)(struct nv10_gr_chan *, u32, u32);
 switch (mthd) {
 case 0x1638 ... 0x1644:
       func = nv17_gr_mthd_lma_window; break;
 case 0x1658: func = nv17_gr_mthd_lma_enable; break;
 default:
  return false;
 }
 func(chan, mthd, data);
 return true;
}

static bool
nv10_gr_mthd(struct nv10_gr_chan *chan, u8 class, u32 mthd, u32 data)
{
 bool (*func)(struct nv10_gr_chan *, u32, u32);
 switch (class) {
 case 0x99: func = nv17_gr_mthd_celcius; break;
 default:
  return false;
 }
 return func(chan, mthd, data);
}

/*******************************************************************************
 * PGRAPH context
 ******************************************************************************/


static struct nv10_gr_chan *
nv10_gr_channel(struct nv10_gr *gr)
{
 struct nvkm_device *device = gr->base.engine.subdev.device;
 struct nv10_gr_chan *chan = NULL;
 if (nvkm_rd32(device, 0x400144) & 0x00010000) {
  int chid = nvkm_rd32(device, 0x400148) >> 24;
  if (chid < ARRAY_SIZE(gr->chan))
   chan = gr->chan[chid];
 }
 return chan;
}

static void
nv10_gr_save_pipe(struct nv10_gr_chan *chan)
{
 struct nv10_gr *gr = chan->gr;
 struct pipe_state *pipe = &chan->pipe_state;
 struct nvkm_device *device = gr->base.engine.subdev.device;

 PIPE_SAVE(gr, pipe->pipe_0x4400, 0x4400);
 PIPE_SAVE(gr, pipe->pipe_0x0200, 0x0200);
 PIPE_SAVE(gr, pipe->pipe_0x6400, 0x6400);
 PIPE_SAVE(gr, pipe->pipe_0x6800, 0x6800);
 PIPE_SAVE(gr, pipe->pipe_0x6c00, 0x6c00);
 PIPE_SAVE(gr, pipe->pipe_0x7000, 0x7000);
 PIPE_SAVE(gr, pipe->pipe_0x7400, 0x7400);
 PIPE_SAVE(gr, pipe->pipe_0x7800, 0x7800);
 PIPE_SAVE(gr, pipe->pipe_0x0040, 0x0040);
 PIPE_SAVE(gr, pipe->pipe_0x0000, 0x0000);
}

static void
nv10_gr_load_pipe(struct nv10_gr_chan *chan)
{
 struct nv10_gr *gr = chan->gr;
 struct pipe_state *pipe = &chan->pipe_state;
 struct nvkm_device *device = gr->base.engine.subdev.device;
 u32 xfmode0, xfmode1;
 int i;

 nv04_gr_idle(&gr->base);
 /* XXX check haiku comments */
 xfmode0 = nvkm_rd32(device, NV10_PGRAPH_XFMODE0);
 xfmode1 = nvkm_rd32(device, NV10_PGRAPH_XFMODE1);
 nvkm_wr32(device, NV10_PGRAPH_XFMODE0, 0x10000000);
 nvkm_wr32(device, NV10_PGRAPH_XFMODE1, 0x00000000);
 nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x000064c0);
 for (i = 0; i < 4; i++)
  nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000);
 for (i = 0; i < 4; i++)
  nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000);

 nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006ab0);
 for (i = 0; i < 3; i++)
  nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000);

 nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006a80);
 for (i = 0; i < 3; i++)
  nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000);

 nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00000040);
 nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000008);


 PIPE_RESTORE(gr, pipe->pipe_0x0200, 0x0200);
 nv04_gr_idle(&gr->base);

 /* restore XFMODE */
 nvkm_wr32(device, NV10_PGRAPH_XFMODE0, xfmode0);
 nvkm_wr32(device, NV10_PGRAPH_XFMODE1, xfmode1);
 PIPE_RESTORE(gr, pipe->pipe_0x6400, 0x6400);
 PIPE_RESTORE(gr, pipe->pipe_0x6800, 0x6800);
 PIPE_RESTORE(gr, pipe->pipe_0x6c00, 0x6c00);
 PIPE_RESTORE(gr, pipe->pipe_0x7000, 0x7000);
 PIPE_RESTORE(gr, pipe->pipe_0x7400, 0x7400);
 PIPE_RESTORE(gr, pipe->pipe_0x7800, 0x7800);
 PIPE_RESTORE(gr, pipe->pipe_0x4400, 0x4400);
 PIPE_RESTORE(gr, pipe->pipe_0x0000, 0x0000);
 PIPE_RESTORE(gr, pipe->pipe_0x0040, 0x0040);
 nv04_gr_idle(&gr->base);
}

static void
nv10_gr_create_pipe(struct nv10_gr_chan *chan)
{
 struct nv10_gr *gr = chan->gr;
 struct nvkm_subdev *subdev = &gr->base.engine.subdev;
 struct pipe_state *pipe_state = &chan->pipe_state;
 u32 *pipe_state_addr;
 int i;
#define PIPE_INIT(addr) \
 do { \
  pipe_state_addr = pipe_state->pipe_##addr; \
 } while (0)
#define PIPE_INIT_END(addr) \
 do { \
  u32 *__end_addr = pipe_state->pipe_##addr + \
    ARRAY_SIZE(pipe_state->pipe_##addr); \
  if (pipe_state_addr != __end_addr) \
   nvkm_error(subdev, "incomplete pipe init for 0x%x : %p/%p\n", \
    addr, pipe_state_addr, __end_addr); \
 } while (0)
#define NV_WRITE_PIPE_INIT(value) *(pipe_state_addr++) = value

 PIPE_INIT(0x0200);
 for (i = 0; i < 48; i++)
  NV_WRITE_PIPE_INIT(0x00000000);
 PIPE_INIT_END(0x0200);

 PIPE_INIT(0x6400);
 for (i = 0; i < 211; i++)
  NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x3f800000);
 NV_WRITE_PIPE_INIT(0x40000000);
 NV_WRITE_PIPE_INIT(0x40000000);
 NV_WRITE_PIPE_INIT(0x40000000);
 NV_WRITE_PIPE_INIT(0x40000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x3f800000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x3f000000);
 NV_WRITE_PIPE_INIT(0x3f000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x3f800000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x3f800000);
 NV_WRITE_PIPE_INIT(0x3f800000);
 NV_WRITE_PIPE_INIT(0x3f800000);
 NV_WRITE_PIPE_INIT(0x3f800000);
 PIPE_INIT_END(0x6400);

 PIPE_INIT(0x6800);
 for (i = 0; i < 162; i++)
  NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x3f800000);
 for (i = 0; i < 25; i++)
  NV_WRITE_PIPE_INIT(0x00000000);
 PIPE_INIT_END(0x6800);

 PIPE_INIT(0x6c00);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0xbf800000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 PIPE_INIT_END(0x6c00);

 PIPE_INIT(0x7000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x7149f2ca);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x7149f2ca);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x7149f2ca);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x7149f2ca);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x7149f2ca);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x7149f2ca);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x7149f2ca);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x00000000);
 NV_WRITE_PIPE_INIT(0x7149f2ca);
 for (i = 0; i < 35; i++)
  NV_WRITE_PIPE_INIT(0x00000000);
 PIPE_INIT_END(0x7000);

 PIPE_INIT(0x7400);
 for (i = 0; i < 48; i++)
  NV_WRITE_PIPE_INIT(0x00000000);
 PIPE_INIT_END(0x7400);

 PIPE_INIT(0x7800);
 for (i = 0; i < 48; i++)
  NV_WRITE_PIPE_INIT(0x00000000);
 PIPE_INIT_END(0x7800);

 PIPE_INIT(0x4400);
 for (i = 0; i < 32; i++)
  NV_WRITE_PIPE_INIT(0x00000000);
 PIPE_INIT_END(0x4400);

 PIPE_INIT(0x0000);
 for (i = 0; i < 16; i++)
  NV_WRITE_PIPE_INIT(0x00000000);
 PIPE_INIT_END(0x0000);

 PIPE_INIT(0x0040);
 for (i = 0; i < 4; i++)
  NV_WRITE_PIPE_INIT(0x00000000);
 PIPE_INIT_END(0x0040);

#undef PIPE_INIT
#undef PIPE_INIT_END
#undef NV_WRITE_PIPE_INIT
}

static int
nv10_gr_ctx_regs_find_offset(struct nv10_gr *gr, int reg)
{
 struct nvkm_subdev *subdev = &gr->base.engine.subdev;
 int i;
 for (i = 0; i < ARRAY_SIZE(nv10_gr_ctx_regs); i++) {
  if (nv10_gr_ctx_regs[i] == reg)
   return i;
 }
 nvkm_error(subdev, "unknown offset nv10_ctx_regs %d\n", reg);
 return -1;
}

static int
nv17_gr_ctx_regs_find_offset(struct nv10_gr *gr, int reg)
{
 struct nvkm_subdev *subdev = &gr->base.engine.subdev;
 int i;
 for (i = 0; i < ARRAY_SIZE(nv17_gr_ctx_regs); i++) {
  if (nv17_gr_ctx_regs[i] == reg)
   return i;
 }
 nvkm_error(subdev, "unknown offset nv17_ctx_regs %d\n", reg);
 return -1;
}

static void
nv10_gr_load_dma_vtxbuf(struct nv10_gr_chan *chan, int chid, u32 inst)
{
 struct nv10_gr *gr = chan->gr;
 struct nvkm_device *device = gr->base.engine.subdev.device;
 u32 st2, st2_dl, st2_dh, fifo_ptr, fifo[0x60/4];
 u32 ctx_user, ctx_switch[5];
 int i, subchan = -1;

 /* NV10TCL_DMA_VTXBUF (method 0x18c) modifies hidden state
 * that cannot be restored via MMIO. Do it through the FIFO
 * instead.
 */


 /* Look for a celsius object */
 for (i = 0; i < 8; i++) {
  int class = nvkm_rd32(device, NV10_PGRAPH_CTX_CACHE(i, 0)) & 0xfff;

  if (class == 0x56 || class == 0x96 || class == 0x99) {
   subchan = i;
   break;
  }
 }

 if (subchan < 0 || !inst)
  return;

 /* Save the current ctx object */
 ctx_user = nvkm_rd32(device, NV10_PGRAPH_CTX_USER);
 for (i = 0; i < 5; i++)
  ctx_switch[i] = nvkm_rd32(device, NV10_PGRAPH_CTX_SWITCH(i));

 /* Save the FIFO state */
 st2 = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_ST2);
 st2_dl = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_ST2_DL);
 st2_dh = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_ST2_DH);
 fifo_ptr = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_FIFO_PTR);

 for (i = 0; i < ARRAY_SIZE(fifo); i++)
  fifo[i] = nvkm_rd32(device, 0x4007a0 + 4 * i);

 /* Switch to the celsius subchannel */
 for (i = 0; i < 5; i++)
  nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(i),
   nvkm_rd32(device, NV10_PGRAPH_CTX_CACHE(subchan, i)));
 nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xe000, subchan << 13);

 /* Inject NV10TCL_DMA_VTXBUF */
 nvkm_wr32(device, NV10_PGRAPH_FFINTFC_FIFO_PTR, 0);
 nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2,
  0x2c000000 | chid << 20 | subchan << 16 | 0x18c);
 nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2_DL, inst);
 nvkm_mask(device, NV10_PGRAPH_CTX_CONTROL, 0, 0x10000);
 nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000001);
 nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000000);

 /* Restore the FIFO state */
 for (i = 0; i < ARRAY_SIZE(fifo); i++)
  nvkm_wr32(device, 0x4007a0 + 4 * i, fifo[i]);

 nvkm_wr32(device, NV10_PGRAPH_FFINTFC_FIFO_PTR, fifo_ptr);
 nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2, st2);
 nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2_DL, st2_dl);
 nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2_DH, st2_dh);

 /* Restore the current ctx object */
 for (i = 0; i < 5; i++)
  nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(i), ctx_switch[i]);
 nvkm_wr32(device, NV10_PGRAPH_CTX_USER, ctx_user);
}

static int
nv10_gr_load_context(struct nv10_gr_chan *chan, int chid)
{
 struct nv10_gr *gr = chan->gr;
 struct nvkm_device *device = gr->base.engine.subdev.device;
 u32 inst;
 int i;

 for (i = 0; i < ARRAY_SIZE(nv10_gr_ctx_regs); i++)
  nvkm_wr32(device, nv10_gr_ctx_regs[i], chan->nv10[i]);

 if (device->card_type >= NV_11 && device->chipset >= 0x17) {
  for (i = 0; i < ARRAY_SIZE(nv17_gr_ctx_regs); i++)
   nvkm_wr32(device, nv17_gr_ctx_regs[i], chan->nv17[i]);
 }

 nv10_gr_load_pipe(chan);

 inst = nvkm_rd32(device, NV10_PGRAPH_GLOBALSTATE1) & 0xffff;
 nv10_gr_load_dma_vtxbuf(chan, chid, inst);

 nvkm_wr32(device, NV10_PGRAPH_CTX_CONTROL, 0x10010100);
 nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xff000000, chid << 24);
 nvkm_mask(device, NV10_PGRAPH_FFINTFC_ST2, 0x30000000, 0x00000000);
 return 0;
}

static int
nv10_gr_unload_context(struct nv10_gr_chan *chan)
{
 struct nv10_gr *gr = chan->gr;
 struct nvkm_device *device = gr->base.engine.subdev.device;
 int i;

 for (i = 0; i < ARRAY_SIZE(nv10_gr_ctx_regs); i++)
  chan->nv10[i] = nvkm_rd32(device, nv10_gr_ctx_regs[i]);

 if (device->card_type >= NV_11 && device->chipset >= 0x17) {
  for (i = 0; i < ARRAY_SIZE(nv17_gr_ctx_regs); i++)
   chan->nv17[i] = nvkm_rd32(device, nv17_gr_ctx_regs[i]);
 }

 nv10_gr_save_pipe(chan);

 nvkm_wr32(device, NV10_PGRAPH_CTX_CONTROL, 0x10000000);
 nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xff000000, 0x1f000000);
 return 0;
}

static void
nv10_gr_context_switch(struct nv10_gr *gr)
{
 struct nvkm_device *device = gr->base.engine.subdev.device;
 struct nv10_gr_chan *prev = NULL;
 struct nv10_gr_chan *next = NULL;
 int chid;

 nv04_gr_idle(&gr->base);

 /* If previous context is valid, we need to save it */
 prev = nv10_gr_channel(gr);
 if (prev)
  nv10_gr_unload_context(prev);

 /* load context for next channel */
 chid = (nvkm_rd32(device, NV04_PGRAPH_TRAPPED_ADDR) >> 20) & 0x1f;
 next = gr->chan[chid];
 if (next)
  nv10_gr_load_context(next, chid);
}

static int
nv10_gr_chan_fini(struct nvkm_object *object, bool suspend)
{
 struct nv10_gr_chan *chan = nv10_gr_chan(object);
 struct nv10_gr *gr = chan->gr;
 struct nvkm_device *device = gr->base.engine.subdev.device;
 unsigned long flags;

 spin_lock_irqsave(&gr->lock, flags);
 nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000000);
 if (nv10_gr_channel(gr) == chan)
  nv10_gr_unload_context(chan);
 nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000001);
 spin_unlock_irqrestore(&gr->lock, flags);
 return 0;
}

static void *
nv10_gr_chan_dtor(struct nvkm_object *object)
{
 struct nv10_gr_chan *chan = nv10_gr_chan(object);
 struct nv10_gr *gr = chan->gr;
 unsigned long flags;

 spin_lock_irqsave(&gr->lock, flags);
 gr->chan[chan->chid] = NULL;
 spin_unlock_irqrestore(&gr->lock, flags);
 return chan;
}

static const struct nvkm_object_func
nv10_gr_chan = {
 .dtor = nv10_gr_chan_dtor,
 .fini = nv10_gr_chan_fini,
};

#define NV_WRITE_CTX(reg, val) do { \
 int offset = nv10_gr_ctx_regs_find_offset(gr, reg); \
 if (offset > 0) \
  chan->nv10[offset] = val; \
 } while (0)

#define NV17_WRITE_CTX(reg, val) do { \
 int offset = nv17_gr_ctx_regs_find_offset(gr, reg); \
 if (offset > 0) \
  chan->nv17[offset] = val; \
 } while (0)

int
nv10_gr_chan_new(struct nvkm_gr *base, struct nvkm_chan *fifoch,
   const struct nvkm_oclass *oclass, struct nvkm_object **pobject)
{
 struct nv10_gr *gr = nv10_gr(base);
 struct nv10_gr_chan *chan;
 struct nvkm_device *device = gr->base.engine.subdev.device;
 unsigned long flags;

 if (!(chan = kzalloc(sizeof(*chan), GFP_KERNEL)))
  return -ENOMEM;
 nvkm_object_ctor(&nv10_gr_chan, oclass, &chan->object);
 chan->gr = gr;
 chan->chid = fifoch->id;
 *pobject = &chan->object;

 NV_WRITE_CTX(0x00400e88, 0x08000000);
 NV_WRITE_CTX(0x00400e9c, 0x4b7fffff);
 NV_WRITE_CTX(NV03_PGRAPH_XY_LOGIC_MISC0, 0x0001ffff);
 NV_WRITE_CTX(0x00400e10, 0x00001000);
 NV_WRITE_CTX(0x00400e14, 0x00001000);
 NV_WRITE_CTX(0x00400e30, 0x00080008);
 NV_WRITE_CTX(0x00400e34, 0x00080008);
 if (device->card_type >= NV_11 && device->chipset >= 0x17) {
  /* is it really needed ??? */
  NV17_WRITE_CTX(NV10_PGRAPH_DEBUG_4,
          nvkm_rd32(device, NV10_PGRAPH_DEBUG_4));
  NV17_WRITE_CTX(0x004006b0, nvkm_rd32(device, 0x004006b0));
  NV17_WRITE_CTX(0x00400eac, 0x0fff0000);
  NV17_WRITE_CTX(0x00400eb0, 0x0fff0000);
  NV17_WRITE_CTX(0x00400ec0, 0x00000080);
  NV17_WRITE_CTX(0x00400ed0, 0x00000080);
 }
 NV_WRITE_CTX(NV10_PGRAPH_CTX_USER, chan->chid << 24);

 nv10_gr_create_pipe(chan);

 spin_lock_irqsave(&gr->lock, flags);
 gr->chan[chan->chid] = chan;
 spin_unlock_irqrestore(&gr->lock, flags);
 return 0;
}

/*******************************************************************************
 * PGRAPH engine/subdev functions
 ******************************************************************************/


void
nv10_gr_tile(struct nvkm_gr *base, int i, struct nvkm_fb_tile *tile)
{
 struct nv10_gr *gr = nv10_gr(base);
 struct nvkm_device *device = gr->base.engine.subdev.device;
 struct nvkm_fifo *fifo = device->fifo;
 unsigned long flags;

 nvkm_fifo_pause(fifo, &flags);
 nv04_gr_idle(&gr->base);

 nvkm_wr32(device, NV10_PGRAPH_TLIMIT(i), tile->limit);
 nvkm_wr32(device, NV10_PGRAPH_TSIZE(i), tile->pitch);
 nvkm_wr32(device, NV10_PGRAPH_TILE(i), tile->addr);

 nvkm_fifo_start(fifo, &flags);
}

const struct nvkm_bitfield nv10_gr_intr_name[] = {
 { NV_PGRAPH_INTR_NOTIFY, "NOTIFY" },
 { NV_PGRAPH_INTR_ERROR,  "ERROR"  },
 {}
};

const struct nvkm_bitfield nv10_gr_nstatus[] = {
 { NV10_PGRAPH_NSTATUS_STATE_IN_USE,       "STATE_IN_USE" },
 { NV10_PGRAPH_NSTATUS_INVALID_STATE,      "INVALID_STATE" },
 { NV10_PGRAPH_NSTATUS_BAD_ARGUMENT,       "BAD_ARGUMENT" },
 { NV10_PGRAPH_NSTATUS_PROTECTION_FAULT,   "PROTECTION_FAULT" },
 {}
};

void
nv10_gr_intr(struct nvkm_gr *base)
{
 struct nv10_gr *gr = nv10_gr(base);
 struct nvkm_subdev *subdev = &gr->base.engine.subdev;
 struct nvkm_device *device = subdev->device;
 u32 stat = nvkm_rd32(device, NV03_PGRAPH_INTR);
 u32 nsource = nvkm_rd32(device, NV03_PGRAPH_NSOURCE);
 u32 nstatus = nvkm_rd32(device, NV03_PGRAPH_NSTATUS);
 u32 addr = nvkm_rd32(device, NV04_PGRAPH_TRAPPED_ADDR);
 u32 chid = (addr & 0x01f00000) >> 20;
 u32 subc = (addr & 0x00070000) >> 16;
 u32 mthd = (addr & 0x00001ffc);
 u32 data = nvkm_rd32(device, NV04_PGRAPH_TRAPPED_DATA);
 u32 class = nvkm_rd32(device, 0x400160 + subc * 4) & 0xfff;
 u32 show = stat;
 char msg[128], src[128], sta[128];
 struct nv10_gr_chan *chan;
 unsigned long flags;

 spin_lock_irqsave(&gr->lock, flags);
 chan = gr->chan[chid];

 if (stat & NV_PGRAPH_INTR_ERROR) {
  if (chan && (nsource & NV03_PGRAPH_NSOURCE_ILLEGAL_MTHD)) {
   if (!nv10_gr_mthd(chan, class, mthd, data))
    show &= ~NV_PGRAPH_INTR_ERROR;
  }
 }

 if (stat & NV_PGRAPH_INTR_CONTEXT_SWITCH) {
  nvkm_wr32(device, NV03_PGRAPH_INTR, NV_PGRAPH_INTR_CONTEXT_SWITCH);
  stat &= ~NV_PGRAPH_INTR_CONTEXT_SWITCH;
  show &= ~NV_PGRAPH_INTR_CONTEXT_SWITCH;
  nv10_gr_context_switch(gr);
 }

 nvkm_wr32(device, NV03_PGRAPH_INTR, stat);
 nvkm_wr32(device, NV04_PGRAPH_FIFO, 0x00000001);

 if (show) {
  nvkm_snprintbf(msg, sizeof(msg), nv10_gr_intr_name, show);
  nvkm_snprintbf(src, sizeof(src), nv04_gr_nsource, nsource);
  nvkm_snprintbf(sta, sizeof(sta), nv10_gr_nstatus, nstatus);
  nvkm_error(subdev, "intr %08x [%s] nsource %08x [%s] "
       "nstatus %08x [%s] ch %d [%s] subc %d "
       "class %04x mthd %04x data %08x\n",
      show, msg, nsource, src, nstatus, sta, chid,
      chan ? chan->object.client->name : "unknown",
      subc, class, mthd, data);
 }

 spin_unlock_irqrestore(&gr->lock, flags);
}

int
nv10_gr_init(struct nvkm_gr *base)
{
 struct nv10_gr *gr = nv10_gr(base);
 struct nvkm_device *device = gr->base.engine.subdev.device;

 nvkm_wr32(device, NV03_PGRAPH_INTR   , 0xFFFFFFFF);
 nvkm_wr32(device, NV03_PGRAPH_INTR_EN, 0xFFFFFFFF);

 nvkm_wr32(device, NV04_PGRAPH_DEBUG_0, 0xFFFFFFFF);
 nvkm_wr32(device, NV04_PGRAPH_DEBUG_0, 0x00000000);
 nvkm_wr32(device, NV04_PGRAPH_DEBUG_1, 0x00118700);
 /* nvkm_wr32(device, NV04_PGRAPH_DEBUG_2, 0x24E00810); */ /* 0x25f92ad9 */
 nvkm_wr32(device, NV04_PGRAPH_DEBUG_2, 0x25f92ad9);
 nvkm_wr32(device, NV04_PGRAPH_DEBUG_3, 0x55DE0830 | (1 << 29) | (1 << 31));

 if (device->card_type >= NV_11 && device->chipset >= 0x17) {
  nvkm_wr32(device, NV10_PGRAPH_DEBUG_4, 0x1f000000);
  nvkm_wr32(device, 0x400a10, 0x03ff3fb6);
  nvkm_wr32(device, 0x400838, 0x002f8684);
  nvkm_wr32(device, 0x40083c, 0x00115f3f);
  nvkm_wr32(device, 0x4006b0, 0x40000020);
 } else {
  nvkm_wr32(device, NV10_PGRAPH_DEBUG_4, 0x00000000);
 }

 nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(0), 0x00000000);
 nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(1), 0x00000000);
 nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(2), 0x00000000);
 nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(3), 0x00000000);
 nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(4), 0x00000000);
 nvkm_wr32(device, NV10_PGRAPH_STATE, 0xFFFFFFFF);

 nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xff000000, 0x1f000000);
 nvkm_wr32(device, NV10_PGRAPH_CTX_CONTROL, 0x10000100);
 nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2, 0x08000000);
 return 0;
}

int
nv10_gr_new_(const struct nvkm_gr_func *func, struct nvkm_device *device,
      enum nvkm_subdev_type type, int inst, struct nvkm_gr **pgr)
{
 struct nv10_gr *gr;

 if (!(gr = kzalloc(sizeof(*gr), GFP_KERNEL)))
  return -ENOMEM;
 spin_lock_init(&gr->lock);
 *pgr = &gr->base;

 return nvkm_gr_ctor(func, device, type, inst, true, &gr->base);
}

static const struct nvkm_gr_func
nv10_gr = {
 .init = nv10_gr_init,
 .intr = nv10_gr_intr,
 .tile = nv10_gr_tile,
 .chan_new = nv10_gr_chan_new,
 .sclass = {
  { -1, -1, 0x0012, &nv04_gr_object }, /* beta1 */
  { -1, -1, 0x0019, &nv04_gr_object }, /* clip */
  { -1, -1, 0x0030, &nv04_gr_object }, /* null */
  { -1, -1, 0x0039, &nv04_gr_object }, /* m2mf */
  { -1, -1, 0x0043, &nv04_gr_object }, /* rop */
  { -1, -1, 0x0044, &nv04_gr_object }, /* pattern */
  { -1, -1, 0x004a, &nv04_gr_object }, /* gdi */
  { -1, -1, 0x0052, &nv04_gr_object }, /* swzsurf */
  { -1, -1, 0x005f, &nv04_gr_object }, /* blit */
  { -1, -1, 0x0062, &nv04_gr_object }, /* surf2d */
  { -1, -1, 0x0072, &nv04_gr_object }, /* beta4 */
  { -1, -1, 0x0089, &nv04_gr_object }, /* sifm */
  { -1, -1, 0x008a, &nv04_gr_object }, /* ifc */
  { -1, -1, 0x009f, &nv04_gr_object }, /* blit */
  { -1, -1, 0x0093, &nv04_gr_object }, /* surf3d */
  { -1, -1, 0x0094, &nv04_gr_object }, /* ttri */
  { -1, -1, 0x0095, &nv04_gr_object }, /* mtri */
  { -1, -1, 0x0056, &nv04_gr_object }, /* celcius */
  {}
 }
};

int
nv10_gr_new(struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_gr **pgr)
{
 return nv10_gr_new_(&nv10_gr, device, type, inst, pgr);
}

Messung V0.5
C=94 H=88 G=90

¤ Dauer der Verarbeitung: 0.16 Sekunden  (vorverarbeitet)  ¤

*© Formatika GbR, Deutschland






Wurzel

Suchen

Beweissystem der NASA

Beweissystem Isabelle

NIST Cobol Testsuite

Cephes Mathematical Library

Wiener Entwicklungsmethode

Haftungshinweis

Die Informationen auf dieser Webseite wurden nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit, noch Qualität der bereit gestellten Informationen zugesichert.

Bemerkung:

Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.