/* * Copyright 2007 Matthieu CASTET * All Rights Reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragr) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER * DEALINGS IN THE SOFTWARE. */ #include #include "regs.h" #include #include #include struct pipe_state { u32 pipe_0x0000[0x040/4]; u32 pipe_0x0040[0x010/4]; u32 pipe_0x0200[0x0c0/4]; u32 pipe_0x4400[0x080/4]; u32 pipe_0x6400[0x3b0/4]; u32 pipe_0x6800[0x2f0/4]; u32 pipe_0x6c00[0x030/4]; u32 pipe_0x7000[0x130/4]; u32 pipe_0x7400[0x0c0/4]; u32 pipe_0x7800[0x0c0/4]; }; static int nv10_gr_ctx_regs[] = { NV10_PGRAPH_CTX_SWITCH(0), NV10_PGRAPH_CTX_SWITCH(1), NV10_PGRAPH_CTX_SWITCH(2), NV10_PGRAPH_CTX_SWITCH(3), NV10_PGRAPH_CTX_SWITCH(4), NV10_PGRAPH_CTX_CACHE(0, 0), NV10_PGRAPH_CTX_CACHE(0, 1), NV10_PGRAPH_CTX_CACHE(0, 2), NV10_PGRAPH_CTX_CACHE(0, 3), NV10_PGRAPH_CTX_CACHE(0, 4), NV10_PGRAPH_CTX_CACHE(1, 0), NV10_PGRAPH_CTX_CACHE(1, 1), NV10_PGRAPH_CTX_CACHE(1, 2), NV10_PGRAPH_CTX_CACHE(1, 3), NV10_PGRAPH_CTX_CACHE(1, 4), NV10_PGRAPH_CTX_CACHE(2, 0), NV10_PGRAPH_CTX_CACHE(2, 1), NV10_PGRAPH_CTX_CACHE(2, 2), NV10_PGRAPH_CTX_CACHE(2, 3), NV10_PGRAPH_CTX_CACHE(2, 4), NV10_PGRAPH_CTX_CACHE(3, 0), NV10_PGRAPH_CTX_CACHE(3, 1), NV10_PGRAPH_CTX_CACHE(3, 2), NV10_PGRAPH_CTX_CACHE(3, 3), NV10_PGRAPH_CTX_CACHE(3, 4), NV10_PGRAPH_CTX_CACHE(4, 0), NV10_PGRAPH_CTX_CACHE(4, 1), NV10_PGRAPH_CTX_CACHE(4, 2), NV10_PGRAPH_CTX_CACHE(4, 3), NV10_PGRAPH_CTX_CACHE(4, 4), NV10_PGRAPH_CTX_CACHE(5, 0), NV10_PGRAPH_CTX_CACHE(5, 1), NV10_PGRAPH_CTX_CACHE(5, 2), NV10_PGRAPH_CTX_CACHE(5, 3), NV10_PGRAPH_CTX_CACHE(5, 4), NV10_PGRAPH_CTX_CACHE(6, 0), NV10_PGRAPH_CTX_CACHE(6, 1), NV10_PGRAPH_CTX_CACHE(6, 2), NV10_PGRAPH_CTX_CACHE(6, 3), NV10_PGRAPH_CTX_CACHE(6, 4), NV10_PGRAPH_CTX_CACHE(7, 0), NV10_PGRAPH_CTX_CACHE(7, 1), NV10_PGRAPH_CTX_CACHE(7, 2), NV10_PGRAPH_CTX_CACHE(7, 3), NV10_PGRAPH_CTX_CACHE(7, 4), NV10_PGRAPH_CTX_USER, NV04_PGRAPH_DMA_START_0, NV04_PGRAPH_DMA_START_1, NV04_PGRAPH_DMA_LENGTH, NV04_PGRAPH_DMA_MISC, NV10_PGRAPH_DMA_PITCH, NV04_PGRAPH_BOFFSET0, NV04_PGRAPH_BBASE0, NV04_PGRAPH_BLIMIT0, NV04_PGRAPH_BOFFSET1, NV04_PGRAPH_BBASE1, NV04_PGRAPH_BLIMIT1, NV04_PGRAPH_BOFFSET2, NV04_PGRAPH_BBASE2, NV04_PGRAPH_BLIMIT2, NV04_PGRAPH_BOFFSET3, NV04_PGRAPH_BBASE3, NV04_PGRAPH_BLIMIT3, NV04_PGRAPH_BOFFSET4, NV04_PGRAPH_BBASE4, NV04_PGRAPH_BLIMIT4, NV04_PGRAPH_BOFFSET5, NV04_PGRAPH_BBASE5, NV04_PGRAPH_BLIMIT5, NV04_PGRAPH_BPITCH0, NV04_PGRAPH_BPITCH1, NV04_PGRAPH_BPITCH2, NV04_PGRAPH_BPITCH3, NV04_PGRAPH_BPITCH4, NV10_PGRAPH_SURFACE, NV10_PGRAPH_STATE, NV04_PGRAPH_BSWIZZLE2, NV04_PGRAPH_BSWIZZLE5, NV04_PGRAPH_BPIXEL, NV10_PGRAPH_NOTIFY, NV04_PGRAPH_PATT_COLOR0, NV04_PGRAPH_PATT_COLOR1, NV04_PGRAPH_PATT_COLORRAM, /* 64 values from 0x400900 to 0x4009fc */ 0x00400904, 0x00400908, 0x0040090c, 0x00400910, 0x00400914, 0x00400918, 0x0040091c, 0x00400920, 0x00400924, 0x00400928, 0x0040092c, 0x00400930, 0x00400934, 0x00400938, 0x0040093c, 0x00400940, 0x00400944, 0x00400948, 0x0040094c, 0x00400950, 0x00400954, 0x00400958, 0x0040095c, 0x00400960, 0x00400964, 0x00400968, 0x0040096c, 0x00400970, 0x00400974, 0x00400978, 0x0040097c, 0x00400980, 0x00400984, 0x00400988, 0x0040098c, 0x00400990, 0x00400994, 0x00400998, 0x0040099c, 0x004009a0, 0x004009a4, 0x004009a8, 0x004009ac, 0x004009b0, 0x004009b4, 0x004009b8, 0x004009bc, 0x004009c0, 0x004009c4, 0x004009c8, 0x004009cc, 0x004009d0, 0x004009d4, 0x004009d8, 0x004009dc, 0x004009e0, 0x004009e4, 0x004009e8, 0x004009ec, 0x004009f0, 0x004009f4, 0x004009f8, 0x004009fc, NV04_PGRAPH_PATTERN, /* 2 values from 0x400808 to 0x40080c */ 0x0040080c, NV04_PGRAPH_PATTERN_SHAPE, NV03_PGRAPH_MONO_COLOR0, NV04_PGRAPH_ROP3, NV04_PGRAPH_CHROMA, NV04_PGRAPH_BETA_AND, NV04_PGRAPH_BETA_PREMULT, 0x00400e70, 0x00400e74, 0x00400e78, 0x00400e7c, 0x00400e80, 0x00400e84, 0x00400e88, 0x00400e8c, 0x00400ea0, 0x00400ea4, 0x00400ea8, 0x00400e90, 0x00400e94, 0x00400e98, 0x00400e9c, NV10_PGRAPH_WINDOWCLIP_HORIZONTAL, /* 8 values from 0x400f00-0x400f1c */ NV10_PGRAPH_WINDOWCLIP_VERTICAL, /* 8 values from 0x400f20-0x400f3c */ 0x00400f04, 0x00400f24, 0x00400f08, 0x00400f28, 0x00400f0c, 0x00400f2c, 0x00400f10, 0x00400f30, 0x00400f14, 0x00400f34, 0x00400f18, 0x00400f38, 0x00400f1c, 0x00400f3c, NV10_PGRAPH_XFMODE0, NV10_PGRAPH_XFMODE1, NV10_PGRAPH_GLOBALSTATE0, NV10_PGRAPH_GLOBALSTATE1, NV04_PGRAPH_STORED_FMT, NV04_PGRAPH_SOURCE_COLOR, NV03_PGRAPH_ABS_X_RAM, /* 32 values from 0x400400 to 0x40047c */ NV03_PGRAPH_ABS_Y_RAM, /* 32 values from 0x400480 to 0x4004fc */ 0x00400404, 0x00400484, 0x00400408, 0x00400488, 0x0040040c, 0x0040048c, 0x00400410, 0x00400490, 0x00400414, 0x00400494, 0x00400418, 0x00400498, 0x0040041c, 0x0040049c, 0x00400420, 0x004004a0, 0x00400424, 0x004004a4, 0x00400428, 0x004004a8, 0x0040042c, 0x004004ac, 0x00400430, 0x004004b0, 0x00400434, 0x004004b4, 0x00400438, 0x004004b8, 0x0040043c, 0x004004bc, 0x00400440, 0x004004c0, 0x00400444, 0x004004c4, 0x00400448, 0x004004c8, 0x0040044c, 0x004004cc, 0x00400450, 0x004004d0, 0x00400454, 0x004004d4, 0x00400458, 0x004004d8, 0x0040045c, 0x004004dc, 0x00400460, 0x004004e0, 0x00400464, 0x004004e4, 0x00400468, 0x004004e8, 0x0040046c, 0x004004ec, 0x00400470, 0x004004f0, 0x00400474, 0x004004f4, 0x00400478, 0x004004f8, 0x0040047c, 0x004004fc, NV03_PGRAPH_ABS_UCLIP_XMIN, NV03_PGRAPH_ABS_UCLIP_XMAX, NV03_PGRAPH_ABS_UCLIP_YMIN, NV03_PGRAPH_ABS_UCLIP_YMAX, 0x00400550, 0x00400558, 0x00400554, 0x0040055c, NV03_PGRAPH_ABS_UCLIPA_XMIN, NV03_PGRAPH_ABS_UCLIPA_XMAX, NV03_PGRAPH_ABS_UCLIPA_YMIN, NV03_PGRAPH_ABS_UCLIPA_YMAX, NV03_PGRAPH_ABS_ICLIP_XMAX, NV03_PGRAPH_ABS_ICLIP_YMAX, NV03_PGRAPH_XY_LOGIC_MISC0, NV03_PGRAPH_XY_LOGIC_MISC1, NV03_PGRAPH_XY_LOGIC_MISC2, NV03_PGRAPH_XY_LOGIC_MISC3, NV03_PGRAPH_CLIPX_0, NV03_PGRAPH_CLIPX_1, NV03_PGRAPH_CLIPY_0, NV03_PGRAPH_CLIPY_1, NV10_PGRAPH_COMBINER0_IN_ALPHA, NV10_PGRAPH_COMBINER1_IN_ALPHA, NV10_PGRAPH_COMBINER0_IN_RGB, NV10_PGRAPH_COMBINER1_IN_RGB, NV10_PGRAPH_COMBINER_COLOR0, NV10_PGRAPH_COMBINER_COLOR1, NV10_PGRAPH_COMBINER0_OUT_ALPHA, NV10_PGRAPH_COMBINER1_OUT_ALPHA, NV10_PGRAPH_COMBINER0_OUT_RGB, NV10_PGRAPH_COMBINER1_OUT_RGB, NV10_PGRAPH_COMBINER_FINAL0, NV10_PGRAPH_COMBINER_FINAL1, 0x00400e00, 0x00400e04, 0x00400e08, 0x00400e0c, 0x00400e10, 0x00400e14, 0x00400e18, 0x00400e1c, 0x00400e20, 0x00400e24, 0x00400e28, 0x00400e2c, 0x00400e30, 0x00400e34, 0x00400e38, 0x00400e3c, NV04_PGRAPH_PASSTHRU_0, NV04_PGRAPH_PASSTHRU_1, NV04_PGRAPH_PASSTHRU_2, NV10_PGRAPH_DIMX_TEXTURE, NV10_PGRAPH_WDIMX_TEXTURE, NV10_PGRAPH_DVD_COLORFMT, NV10_PGRAPH_SCALED_FORMAT, NV04_PGRAPH_MISC24_0, NV04_PGRAPH_MISC24_1, NV04_PGRAPH_MISC24_2, NV03_PGRAPH_X_MISC, NV03_PGRAPH_Y_MISC, NV04_PGRAPH_VALID1, NV04_PGRAPH_VALID2, }; static int nv17_gr_ctx_regs[] = { NV10_PGRAPH_DEBUG_4, 0x004006b0, 0x00400eac, 0x00400eb0, 0x00400eb4, 0x00400eb8, 0x00400ebc, 0x00400ec0, 0x00400ec4, 0x00400ec8, 0x00400ecc, 0x00400ed0, 0x00400ed4, 0x00400ed8, 0x00400edc, 0x00400ee0, 0x00400a00, 0x00400a04, }; struct nv10_gr { struct nvkm_gr base; struct nv10_gr_chan *chan[32]; spinlock_t lock; }; struct nv10_gr_chan { struct nvkm_object base; int chid; int nv10[ARRAY_SIZE(nv10_gr_ctx_regs)]; int nv17[ARRAY_SIZE(nv17_gr_ctx_regs)]; struct pipe_state pipe_state; u32 lma_window[4]; }; static inline struct nv10_gr * nv10_gr(struct nv10_gr_chan *chan) { return (void *)nv_object(chan)->engine; } /******************************************************************************* * Graphics object classes ******************************************************************************/ #define PIPE_SAVE(gr, state, addr) \ do { \ int __i; \ nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, addr); \ for (__i = 0; __i < ARRAY_SIZE(state); __i++) \ state[__i] = nvkm_rd32(device, NV10_PGRAPH_PIPE_DATA); \ } while (0) #define PIPE_RESTORE(gr, state, addr) \ do { \ int __i; \ nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, addr); \ for (__i = 0; __i < ARRAY_SIZE(state); __i++) \ nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, state[__i]); \ } while (0) static struct nvkm_oclass nv10_gr_sclass[] = { { 0x0012, &nv04_gr_ofuncs }, /* beta1 */ { 0x0019, &nv04_gr_ofuncs }, /* clip */ { 0x0030, &nv04_gr_ofuncs }, /* null */ { 0x0039, &nv04_gr_ofuncs }, /* m2mf */ { 0x0043, &nv04_gr_ofuncs }, /* rop */ { 0x0044, &nv04_gr_ofuncs }, /* pattern */ { 0x004a, &nv04_gr_ofuncs }, /* gdi */ { 0x0052, &nv04_gr_ofuncs }, /* swzsurf */ { 0x005f, &nv04_gr_ofuncs }, /* blit */ { 0x0062, &nv04_gr_ofuncs }, /* surf2d */ { 0x0072, &nv04_gr_ofuncs }, /* beta4 */ { 0x0089, &nv04_gr_ofuncs }, /* sifm */ { 0x008a, &nv04_gr_ofuncs }, /* ifc */ { 0x009f, &nv04_gr_ofuncs }, /* blit */ { 0x0093, &nv04_gr_ofuncs }, /* surf3d */ { 0x0094, &nv04_gr_ofuncs }, /* ttri */ { 0x0095, &nv04_gr_ofuncs }, /* mtri */ { 0x0056, &nv04_gr_ofuncs }, /* celcius */ {}, }; static struct nvkm_oclass nv15_gr_sclass[] = { { 0x0012, &nv04_gr_ofuncs }, /* beta1 */ { 0x0019, &nv04_gr_ofuncs }, /* clip */ { 0x0030, &nv04_gr_ofuncs }, /* null */ { 0x0039, &nv04_gr_ofuncs }, /* m2mf */ { 0x0043, &nv04_gr_ofuncs }, /* rop */ { 0x0044, &nv04_gr_ofuncs }, /* pattern */ { 0x004a, &nv04_gr_ofuncs }, /* gdi */ { 0x0052, &nv04_gr_ofuncs }, /* swzsurf */ { 0x005f, &nv04_gr_ofuncs }, /* blit */ { 0x0062, &nv04_gr_ofuncs }, /* surf2d */ { 0x0072, &nv04_gr_ofuncs }, /* beta4 */ { 0x0089, &nv04_gr_ofuncs }, /* sifm */ { 0x008a, &nv04_gr_ofuncs }, /* ifc */ { 0x009f, &nv04_gr_ofuncs }, /* blit */ { 0x0093, &nv04_gr_ofuncs }, /* surf3d */ { 0x0094, &nv04_gr_ofuncs }, /* ttri */ { 0x0095, &nv04_gr_ofuncs }, /* mtri */ { 0x0096, &nv04_gr_ofuncs }, /* celcius */ {}, }; static void nv17_gr_mthd_lma_window(struct nv10_gr_chan *chan, u32 mthd, u32 data) { struct nvkm_device *device = chan->base.engine->subdev.device; struct nvkm_gr *gr = nvkm_gr(chan); struct pipe_state *pipe = &chan->pipe_state; u32 pipe_0x0040[1], pipe_0x64c0[8], pipe_0x6a80[3], pipe_0x6ab0[3]; u32 xfmode0, xfmode1; int i; chan->lma_window[(mthd - 0x1638) / 4] = data; if (mthd != 0x1644) return; nv04_gr_idle(gr); PIPE_SAVE(device, pipe_0x0040, 0x0040); PIPE_SAVE(device, pipe->pipe_0x0200, 0x0200); PIPE_RESTORE(device, chan->lma_window, 0x6790); nv04_gr_idle(gr); xfmode0 = nvkm_rd32(device, NV10_PGRAPH_XFMODE0); xfmode1 = nvkm_rd32(device, NV10_PGRAPH_XFMODE1); PIPE_SAVE(device, pipe->pipe_0x4400, 0x4400); PIPE_SAVE(device, pipe_0x64c0, 0x64c0); PIPE_SAVE(device, pipe_0x6ab0, 0x6ab0); PIPE_SAVE(device, pipe_0x6a80, 0x6a80); nv04_gr_idle(gr); nvkm_wr32(device, NV10_PGRAPH_XFMODE0, 0x10000000); nvkm_wr32(device, NV10_PGRAPH_XFMODE1, 0x00000000); nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x000064c0); for (i = 0; i < 4; i++) nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000); for (i = 0; i < 4; i++) nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000); nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006ab0); for (i = 0; i < 3; i++) nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000); nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006a80); for (i = 0; i < 3; i++) nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000); nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00000040); nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000008); PIPE_RESTORE(device, pipe->pipe_0x0200, 0x0200); nv04_gr_idle(gr); PIPE_RESTORE(device, pipe_0x0040, 0x0040); nvkm_wr32(device, NV10_PGRAPH_XFMODE0, xfmode0); nvkm_wr32(device, NV10_PGRAPH_XFMODE1, xfmode1); PIPE_RESTORE(device, pipe_0x64c0, 0x64c0); PIPE_RESTORE(device, pipe_0x6ab0, 0x6ab0); PIPE_RESTORE(device, pipe_0x6a80, 0x6a80); PIPE_RESTORE(device, pipe->pipe_0x4400, 0x4400); nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x000000c0); nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000); nv04_gr_idle(gr); } static void nv17_gr_mthd_lma_enable(struct nv10_gr_chan *chan, u32 mthd, u32 data) { struct nvkm_device *device = chan->base.engine->subdev.device; struct nvkm_gr *gr = nvkm_gr(chan); nv04_gr_idle(gr); nvkm_mask(device, NV10_PGRAPH_DEBUG_4, 0x00000100, 0x00000100); nvkm_mask(device, 0x4006b0, 0x08000000, 0x08000000); } static bool nv17_gr_mthd_celcius(struct nv10_gr_chan *chan, u32 mthd, u32 data) { void (*func)(struct nv10_gr_chan *, u32, u32); switch (mthd) { case 0x1638 ... 0x1644: func = nv17_gr_mthd_lma_window; break; case 0x1658: func = nv17_gr_mthd_lma_enable; break; default: return false; } func(chan, mthd, data); return true; } static bool nv10_gr_mthd(struct nv10_gr_chan *chan, u8 class, u32 mthd, u32 data) { bool (*func)(struct nv10_gr_chan *, u32, u32); switch (class) { case 0x99: func = nv17_gr_mthd_celcius; break; default: return false; } return func(chan, mthd, data); } static struct nvkm_oclass nv17_gr_sclass[] = { { 0x0012, &nv04_gr_ofuncs }, /* beta1 */ { 0x0019, &nv04_gr_ofuncs }, /* clip */ { 0x0030, &nv04_gr_ofuncs }, /* null */ { 0x0039, &nv04_gr_ofuncs }, /* m2mf */ { 0x0043, &nv04_gr_ofuncs }, /* rop */ { 0x0044, &nv04_gr_ofuncs }, /* pattern */ { 0x004a, &nv04_gr_ofuncs }, /* gdi */ { 0x0052, &nv04_gr_ofuncs }, /* swzsurf */ { 0x005f, &nv04_gr_ofuncs }, /* blit */ { 0x0062, &nv04_gr_ofuncs }, /* surf2d */ { 0x0072, &nv04_gr_ofuncs }, /* beta4 */ { 0x0089, &nv04_gr_ofuncs }, /* sifm */ { 0x008a, &nv04_gr_ofuncs }, /* ifc */ { 0x009f, &nv04_gr_ofuncs }, /* blit */ { 0x0093, &nv04_gr_ofuncs }, /* surf3d */ { 0x0094, &nv04_gr_ofuncs }, /* ttri */ { 0x0095, &nv04_gr_ofuncs }, /* mtri */ { 0x0099, &nv04_gr_ofuncs }, {}, }; /******************************************************************************* * PGRAPH context ******************************************************************************/ static struct nv10_gr_chan * nv10_gr_channel(struct nv10_gr *gr) { struct nvkm_device *device = gr->base.engine.subdev.device; struct nv10_gr_chan *chan = NULL; if (nvkm_rd32(device, 0x400144) & 0x00010000) { int chid = nvkm_rd32(device, 0x400148) >> 24; if (chid < ARRAY_SIZE(gr->chan)) chan = gr->chan[chid]; } return chan; } static void nv10_gr_save_pipe(struct nv10_gr_chan *chan) { struct nv10_gr *gr = nv10_gr(chan); struct pipe_state *pipe = &chan->pipe_state; struct nvkm_device *device = gr->base.engine.subdev.device; PIPE_SAVE(gr, pipe->pipe_0x4400, 0x4400); PIPE_SAVE(gr, pipe->pipe_0x0200, 0x0200); PIPE_SAVE(gr, pipe->pipe_0x6400, 0x6400); PIPE_SAVE(gr, pipe->pipe_0x6800, 0x6800); PIPE_SAVE(gr, pipe->pipe_0x6c00, 0x6c00); PIPE_SAVE(gr, pipe->pipe_0x7000, 0x7000); PIPE_SAVE(gr, pipe->pipe_0x7400, 0x7400); PIPE_SAVE(gr, pipe->pipe_0x7800, 0x7800); PIPE_SAVE(gr, pipe->pipe_0x0040, 0x0040); PIPE_SAVE(gr, pipe->pipe_0x0000, 0x0000); } static void nv10_gr_load_pipe(struct nv10_gr_chan *chan) { struct nv10_gr *gr = nv10_gr(chan); struct pipe_state *pipe = &chan->pipe_state; struct nvkm_device *device = gr->base.engine.subdev.device; u32 xfmode0, xfmode1; int i; nv04_gr_idle(gr); /* XXX check haiku comments */ xfmode0 = nvkm_rd32(device, NV10_PGRAPH_XFMODE0); xfmode1 = nvkm_rd32(device, NV10_PGRAPH_XFMODE1); nvkm_wr32(device, NV10_PGRAPH_XFMODE0, 0x10000000); nvkm_wr32(device, NV10_PGRAPH_XFMODE1, 0x00000000); nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x000064c0); for (i = 0; i < 4; i++) nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000); for (i = 0; i < 4; i++) nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000); nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006ab0); for (i = 0; i < 3; i++) nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000); nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006a80); for (i = 0; i < 3; i++) nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000); nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00000040); nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000008); PIPE_RESTORE(gr, pipe->pipe_0x0200, 0x0200); nv04_gr_idle(gr); /* restore XFMODE */ nvkm_wr32(device, NV10_PGRAPH_XFMODE0, xfmode0); nvkm_wr32(device, NV10_PGRAPH_XFMODE1, xfmode1); PIPE_RESTORE(gr, pipe->pipe_0x6400, 0x6400); PIPE_RESTORE(gr, pipe->pipe_0x6800, 0x6800); PIPE_RESTORE(gr, pipe->pipe_0x6c00, 0x6c00); PIPE_RESTORE(gr, pipe->pipe_0x7000, 0x7000); PIPE_RESTORE(gr, pipe->pipe_0x7400, 0x7400); PIPE_RESTORE(gr, pipe->pipe_0x7800, 0x7800); PIPE_RESTORE(gr, pipe->pipe_0x4400, 0x4400); PIPE_RESTORE(gr, pipe->pipe_0x0000, 0x0000); PIPE_RESTORE(gr, pipe->pipe_0x0040, 0x0040); nv04_gr_idle(gr); } static void nv10_gr_create_pipe(struct nv10_gr_chan *chan) { struct nv10_gr *gr = nv10_gr(chan); struct nvkm_subdev *subdev = &gr->base.engine.subdev; struct pipe_state *pipe_state = &chan->pipe_state; u32 *pipe_state_addr; int i; #define PIPE_INIT(addr) \ do { \ pipe_state_addr = pipe_state->pipe_##addr; \ } while (0) #define PIPE_INIT_END(addr) \ do { \ u32 *__end_addr = pipe_state->pipe_##addr + \ ARRAY_SIZE(pipe_state->pipe_##addr); \ if (pipe_state_addr != __end_addr) \ nvkm_error(subdev, "incomplete pipe init for 0x%x : %p/%p\n", \ addr, pipe_state_addr, __end_addr); \ } while (0) #define NV_WRITE_PIPE_INIT(value) *(pipe_state_addr++) = value PIPE_INIT(0x0200); for (i = 0; i < 48; i++) NV_WRITE_PIPE_INIT(0x00000000); PIPE_INIT_END(0x0200); PIPE_INIT(0x6400); for (i = 0; i < 211; i++) NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x3f800000); NV_WRITE_PIPE_INIT(0x40000000); NV_WRITE_PIPE_INIT(0x40000000); NV_WRITE_PIPE_INIT(0x40000000); NV_WRITE_PIPE_INIT(0x40000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x3f800000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x3f000000); NV_WRITE_PIPE_INIT(0x3f000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x3f800000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x3f800000); NV_WRITE_PIPE_INIT(0x3f800000); NV_WRITE_PIPE_INIT(0x3f800000); NV_WRITE_PIPE_INIT(0x3f800000); PIPE_INIT_END(0x6400); PIPE_INIT(0x6800); for (i = 0; i < 162; i++) NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x3f800000); for (i = 0; i < 25; i++) NV_WRITE_PIPE_INIT(0x00000000); PIPE_INIT_END(0x6800); PIPE_INIT(0x6c00); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0xbf800000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); PIPE_INIT_END(0x6c00); PIPE_INIT(0x7000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x7149f2ca); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x7149f2ca); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x7149f2ca); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x7149f2ca); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x7149f2ca); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x7149f2ca); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x7149f2ca); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x00000000); NV_WRITE_PIPE_INIT(0x7149f2ca); for (i = 0; i < 35; i++) NV_WRITE_PIPE_INIT(0x00000000); PIPE_INIT_END(0x7000); PIPE_INIT(0x7400); for (i = 0; i < 48; i++) NV_WRITE_PIPE_INIT(0x00000000); PIPE_INIT_END(0x7400); PIPE_INIT(0x7800); for (i = 0; i < 48; i++) NV_WRITE_PIPE_INIT(0x00000000); PIPE_INIT_END(0x7800); PIPE_INIT(0x4400); for (i = 0; i < 32; i++) NV_WRITE_PIPE_INIT(0x00000000); PIPE_INIT_END(0x4400); PIPE_INIT(0x0000); for (i = 0; i < 16; i++) NV_WRITE_PIPE_INIT(0x00000000); PIPE_INIT_END(0x0000); PIPE_INIT(0x0040); for (i = 0; i < 4; i++) NV_WRITE_PIPE_INIT(0x00000000); PIPE_INIT_END(0x0040); #undef PIPE_INIT #undef PIPE_INIT_END #undef NV_WRITE_PIPE_INIT } static int nv10_gr_ctx_regs_find_offset(struct nv10_gr *gr, int reg) { struct nvkm_subdev *subdev = &gr->base.engine.subdev; int i; for (i = 0; i < ARRAY_SIZE(nv10_gr_ctx_regs); i++) { if (nv10_gr_ctx_regs[i] == reg) return i; } nvkm_error(subdev, "unknow offset nv10_ctx_regs %d\n", reg); return -1; } static int nv17_gr_ctx_regs_find_offset(struct nv10_gr *gr, int reg) { struct nvkm_subdev *subdev = &gr->base.engine.subdev; int i; for (i = 0; i < ARRAY_SIZE(nv17_gr_ctx_regs); i++) { if (nv17_gr_ctx_regs[i] == reg) return i; } nvkm_error(subdev, "unknow offset nv17_ctx_regs %d\n", reg); return -1; } static void nv10_gr_load_dma_vtxbuf(struct nv10_gr_chan *chan, int chid, u32 inst) { struct nv10_gr *gr = nv10_gr(chan); struct nvkm_device *device = gr->base.engine.subdev.device; u32 st2, st2_dl, st2_dh, fifo_ptr, fifo[0x60/4]; u32 ctx_user, ctx_switch[5]; int i, subchan = -1; /* NV10TCL_DMA_VTXBUF (method 0x18c) modifies hidden state * that cannot be restored via MMIO. Do it through the FIFO * instead. */ /* Look for a celsius object */ for (i = 0; i < 8; i++) { int class = nvkm_rd32(device, NV10_PGRAPH_CTX_CACHE(i, 0)) & 0xfff; if (class == 0x56 || class == 0x96 || class == 0x99) { subchan = i; break; } } if (subchan < 0 || !inst) return; /* Save the current ctx object */ ctx_user = nvkm_rd32(device, NV10_PGRAPH_CTX_USER); for (i = 0; i < 5; i++) ctx_switch[i] = nvkm_rd32(device, NV10_PGRAPH_CTX_SWITCH(i)); /* Save the FIFO state */ st2 = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_ST2); st2_dl = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_ST2_DL); st2_dh = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_ST2_DH); fifo_ptr = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_FIFO_PTR); for (i = 0; i < ARRAY_SIZE(fifo); i++) fifo[i] = nvkm_rd32(device, 0x4007a0 + 4 * i); /* Switch to the celsius subchannel */ for (i = 0; i < 5; i++) nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(i), nvkm_rd32(device, NV10_PGRAPH_CTX_CACHE(subchan, i))); nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xe000, subchan << 13); /* Inject NV10TCL_DMA_VTXBUF */ nvkm_wr32(device, NV10_PGRAPH_FFINTFC_FIFO_PTR, 0); nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2, 0x2c000000 | chid << 20 | subchan << 16 | 0x18c); nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2_DL, inst); nvkm_mask(device, NV10_PGRAPH_CTX_CONTROL, 0, 0x10000); nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000001); nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000000); /* Restore the FIFO state */ for (i = 0; i < ARRAY_SIZE(fifo); i++) nvkm_wr32(device, 0x4007a0 + 4 * i, fifo[i]); nvkm_wr32(device, NV10_PGRAPH_FFINTFC_FIFO_PTR, fifo_ptr); nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2, st2); nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2_DL, st2_dl); nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2_DH, st2_dh); /* Restore the current ctx object */ for (i = 0; i < 5; i++) nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(i), ctx_switch[i]); nvkm_wr32(device, NV10_PGRAPH_CTX_USER, ctx_user); } static int nv10_gr_load_context(struct nv10_gr_chan *chan, int chid) { struct nv10_gr *gr = nv10_gr(chan); struct nvkm_device *device = gr->base.engine.subdev.device; u32 inst; int i; for (i = 0; i < ARRAY_SIZE(nv10_gr_ctx_regs); i++) nvkm_wr32(device, nv10_gr_ctx_regs[i], chan->nv10[i]); if (nv_device(gr)->card_type >= NV_11 && nv_device(gr)->chipset >= 0x17) { for (i = 0; i < ARRAY_SIZE(nv17_gr_ctx_regs); i++) nvkm_wr32(device, nv17_gr_ctx_regs[i], chan->nv17[i]); } nv10_gr_load_pipe(chan); inst = nvkm_rd32(device, NV10_PGRAPH_GLOBALSTATE1) & 0xffff; nv10_gr_load_dma_vtxbuf(chan, chid, inst); nvkm_wr32(device, NV10_PGRAPH_CTX_CONTROL, 0x10010100); nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xff000000, chid << 24); nvkm_mask(device, NV10_PGRAPH_FFINTFC_ST2, 0x30000000, 0x00000000); return 0; } static int nv10_gr_unload_context(struct nv10_gr_chan *chan) { struct nv10_gr *gr = nv10_gr(chan); struct nvkm_device *device = gr->base.engine.subdev.device; int i; for (i = 0; i < ARRAY_SIZE(nv10_gr_ctx_regs); i++) chan->nv10[i] = nvkm_rd32(device, nv10_gr_ctx_regs[i]); if (nv_device(gr)->card_type >= NV_11 && nv_device(gr)->chipset >= 0x17) { for (i = 0; i < ARRAY_SIZE(nv17_gr_ctx_regs); i++) chan->nv17[i] = nvkm_rd32(device, nv17_gr_ctx_regs[i]); } nv10_gr_save_pipe(chan); nvkm_wr32(device, NV10_PGRAPH_CTX_CONTROL, 0x10000000); nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xff000000, 0x1f000000); return 0; } static void nv10_gr_context_switch(struct nv10_gr *gr) { struct nvkm_device *device = gr->base.engine.subdev.device; struct nv10_gr_chan *prev = NULL; struct nv10_gr_chan *next = NULL; int chid; nv04_gr_idle(gr); /* If previous context is valid, we need to save it */ prev = nv10_gr_channel(gr); if (prev) nv10_gr_unload_context(prev); /* load context for next channel */ chid = (nvkm_rd32(device, NV04_PGRAPH_TRAPPED_ADDR) >> 20) & 0x1f; next = gr->chan[chid]; if (next) nv10_gr_load_context(next, chid); } #define NV_WRITE_CTX(reg, val) do { \ int offset = nv10_gr_ctx_regs_find_offset(gr, reg); \ if (offset > 0) \ chan->nv10[offset] = val; \ } while (0) #define NV17_WRITE_CTX(reg, val) do { \ int offset = nv17_gr_ctx_regs_find_offset(gr, reg); \ if (offset > 0) \ chan->nv17[offset] = val; \ } while (0) static int nv10_gr_context_ctor(struct nvkm_object *parent, struct nvkm_object *engine, struct nvkm_oclass *oclass, void *data, u32 size, struct nvkm_object **pobject) { struct nvkm_fifo_chan *fifo = (void *)parent; struct nv10_gr *gr = (void *)engine; struct nv10_gr_chan *chan; struct nvkm_device *device = gr->base.engine.subdev.device; unsigned long flags; int ret; ret = nvkm_object_create(parent, engine, oclass, 0, &chan); *pobject = nv_object(chan); if (ret) return ret; spin_lock_irqsave(&gr->lock, flags); if (gr->chan[fifo->chid]) { *pobject = nv_object(gr->chan[fifo->chid]); atomic_inc(&(*pobject)->refcount); spin_unlock_irqrestore(&gr->lock, flags); nvkm_object_destroy(&chan->base); return 1; } NV_WRITE_CTX(0x00400e88, 0x08000000); NV_WRITE_CTX(0x00400e9c, 0x4b7fffff); NV_WRITE_CTX(NV03_PGRAPH_XY_LOGIC_MISC0, 0x0001ffff); NV_WRITE_CTX(0x00400e10, 0x00001000); NV_WRITE_CTX(0x00400e14, 0x00001000); NV_WRITE_CTX(0x00400e30, 0x00080008); NV_WRITE_CTX(0x00400e34, 0x00080008); if (nv_device(gr)->card_type >= NV_11 && nv_device(gr)->chipset >= 0x17) { /* is it really needed ??? */ NV17_WRITE_CTX(NV10_PGRAPH_DEBUG_4, nvkm_rd32(device, NV10_PGRAPH_DEBUG_4)); NV17_WRITE_CTX(0x004006b0, nvkm_rd32(device, 0x004006b0)); NV17_WRITE_CTX(0x00400eac, 0x0fff0000); NV17_WRITE_CTX(0x00400eb0, 0x0fff0000); NV17_WRITE_CTX(0x00400ec0, 0x00000080); NV17_WRITE_CTX(0x00400ed0, 0x00000080); } NV_WRITE_CTX(NV10_PGRAPH_CTX_USER, chan->chid << 24); nv10_gr_create_pipe(chan); gr->chan[fifo->chid] = chan; chan->chid = fifo->chid; spin_unlock_irqrestore(&gr->lock, flags); return 0; } static void nv10_gr_context_dtor(struct nvkm_object *object) { struct nv10_gr *gr = (void *)object->engine; struct nv10_gr_chan *chan = (void *)object; unsigned long flags; spin_lock_irqsave(&gr->lock, flags); gr->chan[chan->chid] = NULL; spin_unlock_irqrestore(&gr->lock, flags); nvkm_object_destroy(&chan->base); } static int nv10_gr_context_fini(struct nvkm_object *object, bool suspend) { struct nv10_gr *gr = (void *)object->engine; struct nv10_gr_chan *chan = (void *)object; struct nvkm_device *device = gr->base.engine.subdev.device; unsigned long flags; spin_lock_irqsave(&gr->lock, flags); nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000000); if (nv10_gr_channel(gr) == chan) nv10_gr_unload_context(chan); nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000001); spin_unlock_irqrestore(&gr->lock, flags); return _nvkm_object_fini(&chan->base, suspend); } static struct nvkm_oclass nv10_gr_cclass = { .handle = NV_ENGCTX(GR, 0x10), .ofuncs = &(struct nvkm_ofuncs) { .ctor = nv10_gr_context_ctor, .dtor = nv10_gr_context_dtor, .init = _nvkm_object_init, .fini = nv10_gr_context_fini, }, }; /******************************************************************************* * PGRAPH engine/subdev functions ******************************************************************************/ static void nv10_gr_tile_prog(struct nvkm_engine *engine, int i) { struct nv10_gr *gr = (void *)engine; struct nvkm_device *device = gr->base.engine.subdev.device; struct nvkm_fifo *fifo = device->fifo; struct nvkm_fb_tile *tile = &device->fb->tile.region[i]; unsigned long flags; fifo->pause(fifo, &flags); nv04_gr_idle(gr); nvkm_wr32(device, NV10_PGRAPH_TLIMIT(i), tile->limit); nvkm_wr32(device, NV10_PGRAPH_TSIZE(i), tile->pitch); nvkm_wr32(device, NV10_PGRAPH_TILE(i), tile->addr); fifo->start(fifo, &flags); } const struct nvkm_bitfield nv10_gr_intr_name[] = { { NV_PGRAPH_INTR_NOTIFY, "NOTIFY" }, { NV_PGRAPH_INTR_ERROR, "ERROR" }, {} }; const struct nvkm_bitfield nv10_gr_nstatus[] = { { NV10_PGRAPH_NSTATUS_STATE_IN_USE, "STATE_IN_USE" }, { NV10_PGRAPH_NSTATUS_INVALID_STATE, "INVALID_STATE" }, { NV10_PGRAPH_NSTATUS_BAD_ARGUMENT, "BAD_ARGUMENT" }, { NV10_PGRAPH_NSTATUS_PROTECTION_FAULT, "PROTECTION_FAULT" }, {} }; static void nv10_gr_intr(struct nvkm_subdev *subdev) { struct nv10_gr *gr = (void *)subdev; struct nv10_gr_chan *chan = NULL; struct nvkm_device *device = gr->base.engine.subdev.device; u32 stat = nvkm_rd32(device, NV03_PGRAPH_INTR); u32 nsource = nvkm_rd32(device, NV03_PGRAPH_NSOURCE); u32 nstatus = nvkm_rd32(device, NV03_PGRAPH_NSTATUS); u32 addr = nvkm_rd32(device, NV04_PGRAPH_TRAPPED_ADDR); u32 chid = (addr & 0x01f00000) >> 20; u32 subc = (addr & 0x00070000) >> 16; u32 mthd = (addr & 0x00001ffc); u32 data = nvkm_rd32(device, NV04_PGRAPH_TRAPPED_DATA); u32 class = nvkm_rd32(device, 0x400160 + subc * 4) & 0xfff; u32 show = stat; char msg[128], src[128], sta[128]; unsigned long flags; spin_lock_irqsave(&gr->lock, flags); chan = gr->chan[chid]; if (stat & NV_PGRAPH_INTR_ERROR) { if (chan && (nsource & NV03_PGRAPH_NSOURCE_ILLEGAL_MTHD)) { if (!nv10_gr_mthd(chan, class, mthd, data)) show &= ~NV_PGRAPH_INTR_ERROR; } } if (stat & NV_PGRAPH_INTR_CONTEXT_SWITCH) { nvkm_wr32(device, NV03_PGRAPH_INTR, NV_PGRAPH_INTR_CONTEXT_SWITCH); stat &= ~NV_PGRAPH_INTR_CONTEXT_SWITCH; show &= ~NV_PGRAPH_INTR_CONTEXT_SWITCH; nv10_gr_context_switch(gr); } nvkm_wr32(device, NV03_PGRAPH_INTR, stat); nvkm_wr32(device, NV04_PGRAPH_FIFO, 0x00000001); if (show) { nvkm_snprintbf(msg, sizeof(msg), nv10_gr_intr_name, show); nvkm_snprintbf(src, sizeof(src), nv04_gr_nsource, nsource); nvkm_snprintbf(sta, sizeof(sta), nv10_gr_nstatus, nstatus); nvkm_error(subdev, "intr %08x [%s] nsource %08x [%s] " "nstatus %08x [%s] ch %d [%s] subc %d " "class %04x mthd %04x data %08x\n", show, msg, nsource, src, nstatus, sta, chid, nvkm_client_name(chan), subc, class, mthd, data); } spin_unlock_irqrestore(&gr->lock, flags); } static int nv10_gr_ctor(struct nvkm_object *parent, struct nvkm_object *engine, struct nvkm_oclass *oclass, void *data, u32 size, struct nvkm_object **pobject) { struct nv10_gr *gr; int ret; ret = nvkm_gr_create(parent, engine, oclass, true, &gr); *pobject = nv_object(gr); if (ret) return ret; nv_subdev(gr)->unit = 0x00001000; nv_subdev(gr)->intr = nv10_gr_intr; nv_engine(gr)->cclass = &nv10_gr_cclass; if (nv_device(gr)->chipset <= 0x10) nv_engine(gr)->sclass = nv10_gr_sclass; else if (nv_device(gr)->chipset < 0x17 || nv_device(gr)->card_type < NV_11) nv_engine(gr)->sclass = nv15_gr_sclass; else nv_engine(gr)->sclass = nv17_gr_sclass; nv_engine(gr)->tile_prog = nv10_gr_tile_prog; spin_lock_init(&gr->lock); return 0; } static void nv10_gr_dtor(struct nvkm_object *object) { struct nv10_gr *gr = (void *)object; nvkm_gr_destroy(&gr->base); } static int nv10_gr_init(struct nvkm_object *object) { struct nvkm_engine *engine = nv_engine(object); struct nv10_gr *gr = (void *)engine; struct nvkm_device *device = gr->base.engine.subdev.device; struct nvkm_fb *fb = device->fb; int ret, i; ret = nvkm_gr_init(&gr->base); if (ret) return ret; nvkm_wr32(device, NV03_PGRAPH_INTR , 0xFFFFFFFF); nvkm_wr32(device, NV03_PGRAPH_INTR_EN, 0xFFFFFFFF); nvkm_wr32(device, NV04_PGRAPH_DEBUG_0, 0xFFFFFFFF); nvkm_wr32(device, NV04_PGRAPH_DEBUG_0, 0x00000000); nvkm_wr32(device, NV04_PGRAPH_DEBUG_1, 0x00118700); /* nvkm_wr32(device, NV04_PGRAPH_DEBUG_2, 0x24E00810); */ /* 0x25f92ad9 */ nvkm_wr32(device, NV04_PGRAPH_DEBUG_2, 0x25f92ad9); nvkm_wr32(device, NV04_PGRAPH_DEBUG_3, 0x55DE0830 | (1 << 29) | (1 << 31)); if (nv_device(gr)->card_type >= NV_11 && nv_device(gr)->chipset >= 0x17) { nvkm_wr32(device, NV10_PGRAPH_DEBUG_4, 0x1f000000); nvkm_wr32(device, 0x400a10, 0x03ff3fb6); nvkm_wr32(device, 0x400838, 0x002f8684); nvkm_wr32(device, 0x40083c, 0x00115f3f); nvkm_wr32(device, 0x4006b0, 0x40000020); } else { nvkm_wr32(device, NV10_PGRAPH_DEBUG_4, 0x00000000); } /* Turn all the tiling regions off. */ for (i = 0; i < fb->tile.regions; i++) engine->tile_prog(engine, i); nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(0), 0x00000000); nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(1), 0x00000000); nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(2), 0x00000000); nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(3), 0x00000000); nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(4), 0x00000000); nvkm_wr32(device, NV10_PGRAPH_STATE, 0xFFFFFFFF); nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xff000000, 0x1f000000); nvkm_wr32(device, NV10_PGRAPH_CTX_CONTROL, 0x10000100); nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2, 0x08000000); return 0; } static int nv10_gr_fini(struct nvkm_object *object, bool suspend) { struct nv10_gr *gr = (void *)object; return nvkm_gr_fini(&gr->base, suspend); } struct nvkm_oclass nv10_gr_oclass = { .handle = NV_ENGINE(GR, 0x10), .ofuncs = &(struct nvkm_ofuncs) { .ctor = nv10_gr_ctor, .dtor = nv10_gr_dtor, .init = nv10_gr_init, .fini = nv10_gr_fini, }, };