2 * Copyright 2005-2006 Stephane Marchesin
5 * Permission is hereby granted, free of charge, to any person obtaining a
6 * copy of this software and associated documentation files (the "Software"),
7 * to deal in the Software without restriction, including without limitation
8 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
9 * and/or sell copies of the Software, and to permit persons to whom the
10 * Software is furnished to do so, subject to the following conditions:
12 * The above copyright notice and this permission notice (including the next
13 * paragraph) shall be included in all copies or substantial portions of the
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
20 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
21 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
22 * DEALINGS IN THE SOFTWARE.
27 #include "nouveau_drv.h"
28 #include "nouveau_drm.h"
31 /* returns the size of fifo context */
32 int nouveau_fifo_ctx_size(struct drm_device *dev)
34 struct drm_nouveau_private *dev_priv=dev->dev_private;
36 if (dev_priv->card_type >= NV_40)
38 else if (dev_priv->card_type >= NV_17)
44 /***********************************
45 * functions doing the actual work
46 ***********************************/
48 /* voir nv_xaa.c : NVResetGraphics
49 * mémoire mappée par nv_driver.c : NVMapMem
50 * voir nv_driver.c : NVPreInit
53 static int nouveau_fifo_instmem_configure(struct drm_device *dev)
55 struct drm_nouveau_private *dev_priv = dev->dev_private;
57 NV_WRITE(NV03_PFIFO_RAMHT,
58 (0x03 << 24) /* search 128 */ |
59 ((dev_priv->ramht_bits - 9) << 16) |
60 (dev_priv->ramht_offset >> 8)
63 NV_WRITE(NV03_PFIFO_RAMRO, dev_priv->ramro_offset>>8);
65 switch(dev_priv->card_type)
68 switch (dev_priv->chipset) {
77 NV_WRITE(NV40_PFIFO_RAMFC, 0x30002);
80 NV_WRITE(NV40_PFIFO_RAMFC, ((nouveau_mem_fb_amount(dev)-512*1024+dev_priv->ramfc_offset)>>16) |
86 NV_WRITE(NV03_PFIFO_RAMFC, (dev_priv->ramfc_offset>>8) |
87 (1 << 16) /* 64 Bytes entry*/);
88 /* XXX nvidia blob set bit 18, 21,23 for nv20 & nv30 */
93 NV_WRITE(NV03_PFIFO_RAMFC, dev_priv->ramfc_offset>>8);
100 int nouveau_fifo_init(struct drm_device *dev)
102 struct drm_nouveau_private *dev_priv = dev->dev_private;
105 NV_WRITE(NV03_PMC_ENABLE, NV_READ(NV03_PMC_ENABLE) &
106 ~NV_PMC_ENABLE_PFIFO);
107 NV_WRITE(NV03_PMC_ENABLE, NV_READ(NV03_PMC_ENABLE) |
108 NV_PMC_ENABLE_PFIFO);
110 /* Enable PFIFO error reporting */
111 NV_WRITE(NV03_PFIFO_INTR_0, 0xFFFFFFFF);
112 NV_WRITE(NV03_PFIFO_INTR_EN_0, 0xFFFFFFFF);
114 NV_WRITE(NV03_PFIFO_CACHES, 0x00000000);
116 ret = nouveau_fifo_instmem_configure(dev);
118 DRM_ERROR("Failed to configure instance memory\n");
122 /* FIXME remove all the stuff that's done in nouveau_fifo_alloc */
124 DRM_DEBUG("Setting defaults for remaining PFIFO regs\n");
126 /* All channels into PIO mode */
127 NV_WRITE(NV04_PFIFO_MODE, 0x00000000);
129 NV_WRITE(NV03_PFIFO_CACHE1_PUSH0, 0x00000000);
130 NV_WRITE(NV04_PFIFO_CACHE1_PULL0, 0x00000000);
131 /* Channel 0 active, PIO mode */
132 NV_WRITE(NV03_PFIFO_CACHE1_PUSH1, 0x00000000);
133 /* PUT and GET to 0 */
134 NV_WRITE(NV04_PFIFO_CACHE1_DMA_PUT, 0x00000000);
135 NV_WRITE(NV04_PFIFO_CACHE1_DMA_GET, 0x00000000);
136 /* No cmdbuf object */
137 NV_WRITE(NV04_PFIFO_CACHE1_DMA_INSTANCE, 0x00000000);
138 NV_WRITE(NV03_PFIFO_CACHE0_PUSH0, 0x00000000);
139 NV_WRITE(NV03_PFIFO_CACHE0_PULL0, 0x00000000);
140 NV_WRITE(NV04_PFIFO_SIZE, 0x0000FFFF);
141 NV_WRITE(NV04_PFIFO_CACHE1_HASH, 0x0000FFFF);
142 NV_WRITE(NV04_PFIFO_CACHE0_PULL1, 0x00000001);
143 NV_WRITE(NV04_PFIFO_CACHE1_DMA_CTL, 0x00000000);
144 NV_WRITE(NV04_PFIFO_CACHE1_DMA_STATE, 0x00000000);
145 NV_WRITE(NV04_PFIFO_CACHE1_ENGINE, 0x00000000);
147 NV_WRITE(NV04_PFIFO_CACHE1_DMA_FETCH, NV_PFIFO_CACHE1_DMA_FETCH_TRIG_112_BYTES |
148 NV_PFIFO_CACHE1_DMA_FETCH_SIZE_128_BYTES |
149 NV_PFIFO_CACHE1_DMA_FETCH_MAX_REQS_4 |
151 NV_PFIFO_CACHE1_BIG_ENDIAN |
155 NV_WRITE(NV04_PFIFO_CACHE1_DMA_PUSH, 0x00000001);
156 NV_WRITE(NV03_PFIFO_CACHE1_PUSH0, 0x00000001);
157 NV_WRITE(NV04_PFIFO_CACHE1_PULL0, 0x00000001);
158 NV_WRITE(NV04_PFIFO_CACHE1_PULL1, 0x00000001);
161 if (dev_priv->card_type >= NV_10) {
162 NV_WRITE(NV10_PGRAPH_CTX_USER, 0x0);
163 NV_WRITE(NV04_PFIFO_DELAY_0, 0xff /* retrycount*/ );
164 if (dev_priv->card_type >= NV_40)
165 NV_WRITE(NV10_PGRAPH_CTX_CONTROL, 0x00002001);
167 NV_WRITE(NV10_PGRAPH_CTX_CONTROL, 0x10110000);
169 NV_WRITE(NV04_PGRAPH_CTX_USER, 0x0);
170 NV_WRITE(NV04_PFIFO_DELAY_0, 0xff /* retrycount*/ );
171 NV_WRITE(NV04_PGRAPH_CTX_CONTROL, 0x10110000);
174 NV_WRITE(NV04_PFIFO_DMA_TIMESLICE, 0x001fffff);
175 NV_WRITE(NV03_PFIFO_CACHES, 0x00000001);
180 nouveau_fifo_pushbuf_ctxdma_init(struct nouveau_channel *chan)
182 struct drm_device *dev = chan->dev;
183 struct drm_nouveau_private *dev_priv = dev->dev_private;
184 struct mem_block *pb = chan->pushbuf_mem;
185 struct nouveau_gpuobj *pushbuf = NULL;
188 if (pb->flags & NOUVEAU_MEM_AGP) {
189 ret = nouveau_gpuobj_gart_dma_new(chan, pb->start, pb->size,
192 &chan->pushbuf_base);
194 if (pb->flags & NOUVEAU_MEM_PCI) {
195 ret = nouveau_gpuobj_dma_new(chan, NV_CLASS_DMA_IN_MEMORY,
198 NV_DMA_TARGET_PCI_NONLINEAR,
200 chan->pushbuf_base = 0;
201 } else if (dev_priv->card_type != NV_04) {
202 ret = nouveau_gpuobj_dma_new(chan, NV_CLASS_DMA_IN_MEMORY,
205 NV_DMA_TARGET_VIDMEM, &pushbuf);
206 chan->pushbuf_base = 0;
208 /* NV04 cmdbuf hack, from original ddx.. not sure of it's
209 * exact reason for existing :) PCI access to cmdbuf in
212 ret = nouveau_gpuobj_dma_new(chan, NV_CLASS_DMA_IN_MEMORY,
214 drm_get_resource_start(dev, 1),
215 pb->size, NV_DMA_ACCESS_RO,
216 NV_DMA_TARGET_PCI, &pushbuf);
217 chan->pushbuf_base = 0;
220 if ((ret = nouveau_gpuobj_ref_add(dev, chan, 0, pushbuf,
222 DRM_ERROR("Error referencing push buffer ctxdma: %d\n", ret);
223 if (pushbuf != dev_priv->gart_info.sg_ctxdma)
224 nouveau_gpuobj_del(dev, &pushbuf);
231 static struct mem_block *
232 nouveau_fifo_user_pushbuf_alloc(struct drm_device *dev)
234 struct drm_nouveau_private *dev_priv = dev->dev_private;
235 struct nouveau_config *config = &dev_priv->config;
236 struct mem_block *pb;
237 int pb_min_size = max(NV03_FIFO_SIZE,PAGE_SIZE);
239 /* Defaults for unconfigured values */
240 if (!config->cmdbuf.location)
241 config->cmdbuf.location = NOUVEAU_MEM_FB;
242 if (!config->cmdbuf.size || config->cmdbuf.size < pb_min_size)
243 config->cmdbuf.size = pb_min_size;
245 pb = nouveau_mem_alloc(dev, 0, config->cmdbuf.size,
246 config->cmdbuf.location | NOUVEAU_MEM_MAPPED,
247 (struct drm_file *)-2);
249 DRM_ERROR("Couldn't allocate DMA push buffer.\n");
254 /* allocates and initializes a fifo for user space consumption */
256 nouveau_fifo_alloc(struct drm_device *dev, struct nouveau_channel **chan_ret,
257 struct drm_file *file_priv, struct mem_block *pushbuf,
258 uint32_t vram_handle, uint32_t tt_handle)
261 struct drm_nouveau_private *dev_priv = dev->dev_private;
262 struct nouveau_engine *engine = &dev_priv->Engine;
263 struct nouveau_channel *chan;
267 * Alright, here is the full story
268 * Nvidia cards have multiple hw fifo contexts (praise them for that,
269 * no complicated crash-prone context switches)
270 * We allocate a new context for each app and let it write to it directly
271 * (woo, full userspace command submission !)
272 * When there are no more contexts, you lost
274 for (channel = 0; channel < engine->fifo.channels; channel++) {
275 if (dev_priv->fifos[channel] == NULL)
279 /* no more fifos. you lost. */
280 if (channel == engine->fifo.channels)
283 dev_priv->fifos[channel] = drm_calloc(1, sizeof(struct nouveau_channel),
285 if (!dev_priv->fifos[channel])
287 dev_priv->fifo_alloc_count++;
288 chan = dev_priv->fifos[channel];
291 chan->file_priv = file_priv;
292 chan->pushbuf_mem = pushbuf;
294 DRM_INFO("Allocating FIFO number %d\n", channel);
296 /* Locate channel's user control regs */
297 if (dev_priv->card_type < NV_40) {
298 chan->user = NV03_USER(channel);
299 chan->user_size = NV03_USER_SIZE;
300 chan->put = NV03_USER_DMA_PUT(channel);
301 chan->get = NV03_USER_DMA_GET(channel);
302 chan->ref_cnt = NV03_USER_REF_CNT(channel);
304 if (dev_priv->card_type < NV_50) {
305 chan->user = NV40_USER(channel);
306 chan->user_size = NV40_USER_SIZE;
307 chan->put = NV40_USER_DMA_PUT(channel);
308 chan->get = NV40_USER_DMA_GET(channel);
309 chan->ref_cnt = NV40_USER_REF_CNT(channel);
311 chan->user = NV50_USER(channel);
312 chan->user_size = NV50_USER_SIZE;
313 chan->put = NV50_USER_DMA_PUT(channel);
314 chan->get = NV50_USER_DMA_GET(channel);
315 chan->ref_cnt = NV50_USER_REF_CNT(channel);
318 /* Allocate space for per-channel fixed notifier memory */
319 ret = nouveau_notifier_init_channel(chan);
321 nouveau_fifo_free(chan);
325 /* Setup channel's default objects */
326 ret = nouveau_gpuobj_channel_init(chan, vram_handle, tt_handle);
328 nouveau_fifo_free(chan);
332 /* Create a dma object for the push buffer */
333 ret = nouveau_fifo_pushbuf_ctxdma_init(chan);
335 nouveau_fifo_free(chan);
339 nouveau_wait_for_idle(dev);
341 /* disable the fifo caches */
342 NV_WRITE(NV03_PFIFO_CACHES, 0x00000000);
343 NV_WRITE(NV04_PFIFO_CACHE1_DMA_PUSH, NV_READ(NV04_PFIFO_CACHE1_DMA_PUSH)&(~0x1));
344 NV_WRITE(NV03_PFIFO_CACHE1_PUSH0, 0x00000000);
345 NV_WRITE(NV04_PFIFO_CACHE1_PULL0, 0x00000000);
347 /* Create a graphics context for new channel */
348 ret = engine->graph.create_context(chan);
350 nouveau_fifo_free(chan);
354 /* Construct inital RAMFC for new channel */
355 ret = engine->fifo.create_context(chan);
357 nouveau_fifo_free(chan);
361 /* setup channel's default get/put values
362 * XXX: quite possibly extremely pointless..
364 NV_WRITE(chan->get, chan->pushbuf_base);
365 NV_WRITE(chan->put, chan->pushbuf_base);
367 /* If this is the first channel, setup PFIFO ourselves. For any
368 * other case, the GPU will handle this when it switches contexts.
370 if (dev_priv->fifo_alloc_count == 1) {
371 ret = engine->fifo.load_context(chan);
373 nouveau_fifo_free(chan);
377 ret = engine->graph.load_context(chan);
379 nouveau_fifo_free(chan);
384 NV_WRITE(NV04_PFIFO_CACHE1_DMA_PUSH,
385 NV_READ(NV04_PFIFO_CACHE1_DMA_PUSH) | 1);
386 NV_WRITE(NV03_PFIFO_CACHE1_PUSH0, 0x00000001);
387 NV_WRITE(NV04_PFIFO_CACHE1_PULL0, 0x00000001);
388 NV_WRITE(NV04_PFIFO_CACHE1_PULL1, 0x00000001);
390 /* reenable the fifo caches */
391 NV_WRITE(NV03_PFIFO_CACHES, 1);
393 DRM_INFO("%s: initialised FIFO %d\n", __func__, channel);
399 void nouveau_fifo_free(struct nouveau_channel *chan)
401 struct drm_device *dev = chan->dev;
402 struct drm_nouveau_private *dev_priv = dev->dev_private;
403 struct nouveau_engine *engine = &dev_priv->Engine;
406 DRM_INFO("%s: freeing fifo %d\n", __func__, chan->id);
408 /* Disable channel switching, if this channel isn't currenly
409 * active re-enable it if there's still pending commands.
410 * We really should do a manual context switch here, but I'm
411 * not sure I trust our ability to do this reliably yet..
413 NV_WRITE(NV03_PFIFO_CACHES, 0);
414 if (engine->fifo.channel_id(dev) != chan->id &&
415 NV_READ(chan->get) != NV_READ(chan->put)) {
416 NV_WRITE(NV03_PFIFO_CACHES, 1);
419 /* Give the channel a chance to idle, wait 2s (hopefully) */
420 t_start = engine->timer.read(dev);
421 while (NV_READ(chan->get) != NV_READ(chan->put) ||
422 NV_READ(NV03_PFIFO_CACHE1_GET) !=
423 NV_READ(NV03_PFIFO_CACHE1_PUT)) {
424 if (engine->timer.read(dev) - t_start > 2000000000ULL) {
425 DRM_ERROR("Failed to idle channel %d before destroy."
426 "Prepare for strangeness..\n", chan->id);
431 /*XXX: Maybe should wait for PGRAPH to finish with the stuff it fetched
435 /* disable the fifo caches */
436 NV_WRITE(NV03_PFIFO_CACHES, 0x00000000);
437 NV_WRITE(NV04_PFIFO_CACHE1_DMA_PUSH, NV_READ(NV04_PFIFO_CACHE1_DMA_PUSH)&(~0x1));
438 NV_WRITE(NV03_PFIFO_CACHE1_PUSH0, 0x00000000);
439 NV_WRITE(NV04_PFIFO_CACHE1_PULL0, 0x00000000);
441 /* stop the fifo, otherwise it could be running and
442 * it will crash when removing gpu objects
443 *XXX: from real-world evidence, absolutely useless..
445 NV_WRITE(chan->get, chan->pushbuf_base);
446 NV_WRITE(chan->put, chan->pushbuf_base);
448 // FIXME XXX needs more code
450 engine->fifo.destroy_context(chan);
452 /* Cleanup PGRAPH state */
453 engine->graph.destroy_context(chan);
455 /* reenable the fifo caches */
456 NV_WRITE(NV04_PFIFO_CACHE1_DMA_PUSH,
457 NV_READ(NV04_PFIFO_CACHE1_DMA_PUSH) | 1);
458 NV_WRITE(NV03_PFIFO_CACHE1_PUSH0, 0x00000001);
459 NV_WRITE(NV04_PFIFO_CACHE1_PULL0, 0x00000001);
460 NV_WRITE(NV03_PFIFO_CACHES, 0x00000001);
462 /* Deallocate push buffer */
463 nouveau_gpuobj_ref_del(dev, &chan->pushbuf);
464 if (chan->pushbuf_mem) {
465 nouveau_mem_free(dev, chan->pushbuf_mem);
466 chan->pushbuf_mem = NULL;
469 /* Destroy objects belonging to the channel */
470 nouveau_gpuobj_channel_takedown(chan);
472 nouveau_notifier_takedown_channel(chan);
474 dev_priv->fifos[chan->id] = NULL;
475 dev_priv->fifo_alloc_count--;
476 drm_free(chan, sizeof(*chan), DRM_MEM_DRIVER);
479 /* cleanups all the fifos from file_priv */
480 void nouveau_fifo_cleanup(struct drm_device *dev, struct drm_file *file_priv)
482 struct drm_nouveau_private *dev_priv = dev->dev_private;
483 struct nouveau_engine *engine = &dev_priv->Engine;
486 DRM_DEBUG("clearing FIFO enables from file_priv\n");
487 for(i = 0; i < engine->fifo.channels; i++) {
488 struct nouveau_channel *chan = dev_priv->fifos[i];
490 if (chan && chan->file_priv == file_priv)
491 nouveau_fifo_free(chan);
496 nouveau_fifo_owner(struct drm_device *dev, struct drm_file *file_priv,
499 struct drm_nouveau_private *dev_priv = dev->dev_private;
500 struct nouveau_engine *engine = &dev_priv->Engine;
502 if (channel >= engine->fifo.channels)
504 if (dev_priv->fifos[channel] == NULL)
506 return (dev_priv->fifos[channel]->file_priv == file_priv);
509 /***********************************
510 * ioctls wrapping the functions
511 ***********************************/
513 static int nouveau_ioctl_fifo_alloc(struct drm_device *dev, void *data,
514 struct drm_file *file_priv)
516 struct drm_nouveau_private *dev_priv = dev->dev_private;
517 struct drm_nouveau_channel_alloc *init = data;
518 struct drm_map_list *entry;
519 struct nouveau_channel *chan;
520 struct mem_block *pushbuf;
523 NOUVEAU_CHECK_INITIALISED_WITH_RETURN;
525 if (init->fb_ctxdma_handle == ~0 || init->tt_ctxdma_handle == ~0)
528 pushbuf = nouveau_fifo_user_pushbuf_alloc(dev);
532 res = nouveau_fifo_alloc(dev, &chan, file_priv, pushbuf,
533 init->fb_ctxdma_handle,
534 init->tt_ctxdma_handle);
537 init->channel = chan->id;
538 init->put_base = chan->pushbuf_base;
540 /* make the fifo available to user space */
541 /* first, the fifo control regs */
542 init->ctrl = dev_priv->mmio->offset + chan->user;
543 init->ctrl_size = chan->user_size;
544 res = drm_addmap(dev, init->ctrl, init->ctrl_size, _DRM_REGISTERS,
549 entry = drm_find_matching_map(dev, chan->regs);
552 init->ctrl = entry->user_token;
554 /* pass back FIFO map info to the caller */
555 init->cmdbuf = chan->pushbuf_mem->map_handle;
556 init->cmdbuf_size = chan->pushbuf_mem->size;
558 /* and the notifier block */
559 init->notifier = chan->notifier_block->map_handle;
560 init->notifier_size = chan->notifier_block->size;
565 static int nouveau_ioctl_fifo_free(struct drm_device *dev, void *data,
566 struct drm_file *file_priv)
568 struct drm_nouveau_channel_free *cfree = data;
569 struct nouveau_channel *chan;
571 NOUVEAU_CHECK_INITIALISED_WITH_RETURN;
572 NOUVEAU_GET_USER_CHANNEL_WITH_RETURN(cfree->channel, file_priv, chan);
574 nouveau_fifo_free(chan);
578 /***********************************
579 * finally, the ioctl table
580 ***********************************/
582 struct drm_ioctl_desc nouveau_ioctls[] = {
583 DRM_IOCTL_DEF(DRM_NOUVEAU_CARD_INIT, nouveau_ioctl_card_init, DRM_AUTH),
584 DRM_IOCTL_DEF(DRM_NOUVEAU_GETPARAM, nouveau_ioctl_getparam, DRM_AUTH),
585 DRM_IOCTL_DEF(DRM_NOUVEAU_SETPARAM, nouveau_ioctl_setparam, DRM_AUTH|DRM_MASTER|DRM_ROOT_ONLY),
586 DRM_IOCTL_DEF(DRM_NOUVEAU_CHANNEL_ALLOC, nouveau_ioctl_fifo_alloc, DRM_AUTH),
587 DRM_IOCTL_DEF(DRM_NOUVEAU_CHANNEL_FREE, nouveau_ioctl_fifo_free, DRM_AUTH),
588 DRM_IOCTL_DEF(DRM_NOUVEAU_GROBJ_ALLOC, nouveau_ioctl_grobj_alloc, DRM_AUTH),
589 DRM_IOCTL_DEF(DRM_NOUVEAU_NOTIFIEROBJ_ALLOC, nouveau_ioctl_notifier_alloc, DRM_AUTH),
590 DRM_IOCTL_DEF(DRM_NOUVEAU_GPUOBJ_FREE, nouveau_ioctl_gpuobj_free, DRM_AUTH),
591 DRM_IOCTL_DEF(DRM_NOUVEAU_MEM_ALLOC, nouveau_ioctl_mem_alloc, DRM_AUTH),
592 DRM_IOCTL_DEF(DRM_NOUVEAU_MEM_FREE, nouveau_ioctl_mem_free, DRM_AUTH),
595 int nouveau_max_ioctl = DRM_ARRAY_SIZE(nouveau_ioctls);