2 * Copyright 2010 Jerome Glisse <glisse@freedesktop.org>
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * on the rights to use, copy, modify, merge, publish, distribute, sub
8 * license, and/or sell copies of the Software, and to permit persons to whom
9 * the Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
19 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
20 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
21 * USE OR OTHER DEALINGS IN THE SOFTWARE.
23 #include "r600_pipe.h"
24 #include "r600_public.h"
26 #include "evergreen_compute.h"
30 #include "pipe/p_shader_tokens.h"
31 #include "util/u_blitter.h"
32 #include "util/u_debug.h"
33 #include "util/u_format_s3tc.h"
34 #include "util/u_memory.h"
35 #include "util/u_simple_shaders.h"
36 #include "util/u_upload_mgr.h"
37 #include "util/u_math.h"
38 #include "vl/vl_decoder.h"
39 #include "vl/vl_video_buffer.h"
40 #include "radeon/radeon_uvd.h"
41 #include "os/os_time.h"
43 static const struct debug_named_value debug_options[] = {
45 { "texdepth", DBG_TEX_DEPTH, "Print texture depth info" },
46 { "compute", DBG_COMPUTE, "Print compute info" },
47 { "vm", DBG_VM, "Print virtual addresses when creating resources" },
48 { "trace_cs", DBG_TRACE_CS, "Trace cs and write rlockup_<csid>.c file with faulty cs" },
51 { "fs", DBG_FS, "Print fetch shaders" },
52 { "vs", DBG_VS, "Print vertex shaders" },
53 { "gs", DBG_GS, "Print geometry shaders" },
54 { "ps", DBG_PS, "Print pixel shaders" },
55 { "cs", DBG_CS, "Print compute shaders" },
58 { "nohyperz", DBG_NO_HYPERZ, "Disable Hyper-Z" },
59 #if defined(R600_USE_LLVM)
60 { "nollvm", DBG_NO_LLVM, "Disable the LLVM shader compiler" },
62 { "nocpdma", DBG_NO_CP_DMA, "Disable CP DMA" },
63 { "nodma", DBG_NO_ASYNC_DMA, "Disable asynchronous DMA" },
64 /* GL uses the word INVALIDATE, gallium uses the word DISCARD */
65 { "noinvalrange", DBG_NO_DISCARD_RANGE, "Disable handling of INVALIDATE_RANGE map flags" },
67 DEBUG_NAMED_VALUE_END /* must be last */
73 static struct r600_fence *r600_create_fence(struct r600_context *rctx)
75 struct r600_screen *rscreen = rctx->screen;
76 struct r600_fence *fence = NULL;
78 pipe_mutex_lock(rscreen->fences.mutex);
80 if (!rscreen->fences.bo) {
81 /* Create the shared buffer object */
82 rscreen->fences.bo = (struct r600_resource*)
83 pipe_buffer_create(&rscreen->screen, PIPE_BIND_CUSTOM,
84 PIPE_USAGE_STAGING, 4096);
85 if (!rscreen->fences.bo) {
86 R600_ERR("r600: failed to create bo for fence objects\n");
89 rscreen->fences.data = r600_buffer_mmap_sync_with_rings(rctx, rscreen->fences.bo, PIPE_TRANSFER_READ_WRITE);
92 if (!LIST_IS_EMPTY(&rscreen->fences.pool)) {
93 struct r600_fence *entry;
95 /* Try to find a freed fence that has been signalled */
96 LIST_FOR_EACH_ENTRY(entry, &rscreen->fences.pool, head) {
97 if (rscreen->fences.data[entry->index] != 0) {
98 LIST_DELINIT(&entry->head);
106 /* Allocate a new fence */
107 struct r600_fence_block *block;
110 if ((rscreen->fences.next_index + 1) >= 1024) {
111 R600_ERR("r600: too many concurrent fences\n");
115 index = rscreen->fences.next_index++;
117 if (!(index % FENCE_BLOCK_SIZE)) {
118 /* Allocate a new block */
119 block = CALLOC_STRUCT(r600_fence_block);
123 LIST_ADD(&block->head, &rscreen->fences.blocks);
125 block = LIST_ENTRY(struct r600_fence_block, rscreen->fences.blocks.next, head);
128 fence = &block->fences[index % FENCE_BLOCK_SIZE];
129 fence->index = index;
132 pipe_reference_init(&fence->reference, 1);
134 rscreen->fences.data[fence->index] = 0;
135 r600_context_emit_fence(rctx, rscreen->fences.bo, fence->index, 1);
137 /* Create a dummy BO so that fence_finish without a timeout can sleep waiting for completion */
138 fence->sleep_bo = (struct r600_resource*)
139 pipe_buffer_create(&rctx->screen->screen, PIPE_BIND_CUSTOM,
140 PIPE_USAGE_STAGING, 1);
141 /* Add the fence as a dummy relocation. */
142 r600_context_bo_reloc(rctx, &rctx->rings.gfx, fence->sleep_bo, RADEON_USAGE_READWRITE);
145 pipe_mutex_unlock(rscreen->fences.mutex);
149 static void r600_flush(struct pipe_context *ctx, unsigned flags)
151 struct r600_context *rctx = (struct r600_context *)ctx;
152 struct pipe_query *render_cond = NULL;
153 unsigned render_cond_mode = 0;
155 rctx->rings.gfx.flushing = true;
156 /* Disable render condition. */
157 if (rctx->current_render_cond) {
158 render_cond = rctx->current_render_cond;
159 render_cond_mode = rctx->current_render_cond_mode;
160 ctx->render_condition(ctx, NULL, 0);
163 r600_context_flush(rctx, flags);
164 rctx->rings.gfx.flushing = false;
165 r600_begin_new_cs(rctx);
167 /* Re-enable render condition. */
169 ctx->render_condition(ctx, render_cond, render_cond_mode);
173 static void r600_flush_from_st(struct pipe_context *ctx,
174 struct pipe_fence_handle **fence,
175 enum pipe_flush_flags flags)
177 struct r600_context *rctx = (struct r600_context *)ctx;
178 struct r600_fence **rfence = (struct r600_fence**)fence;
181 fflags = flags & PIPE_FLUSH_END_OF_FRAME ? RADEON_FLUSH_END_OF_FRAME : 0;
183 *rfence = r600_create_fence(rctx);
185 /* flush gfx & dma ring, order does not matter as only one can be live */
186 if (rctx->rings.dma.cs) {
187 rctx->rings.dma.flush(rctx, fflags);
189 rctx->rings.gfx.flush(rctx, fflags);
192 static void r600_flush_gfx_ring(void *ctx, unsigned flags)
194 r600_flush((struct pipe_context*)ctx, flags);
197 static void r600_flush_dma_ring(void *ctx, unsigned flags)
199 struct r600_context *rctx = (struct r600_context *)ctx;
200 struct radeon_winsys_cs *cs = rctx->rings.dma.cs;
201 unsigned padding_dw, i;
207 /* Pad the DMA CS to a multiple of 8 dwords. */
208 padding_dw = 8 - cs->cdw % 8;
209 if (padding_dw < 8) {
210 for (i = 0; i < padding_dw; i++) {
211 cs->buf[cs->cdw++] = DMA_PACKET(DMA_PACKET_NOP, 0, 0, 0);
215 rctx->rings.dma.flushing = true;
216 rctx->ws->cs_flush(cs, flags, 0);
217 rctx->rings.dma.flushing = false;
220 boolean r600_rings_is_buffer_referenced(struct r600_context *ctx,
221 struct radeon_winsys_cs_handle *buf,
222 enum radeon_bo_usage usage)
224 if (ctx->ws->cs_is_buffer_referenced(ctx->rings.gfx.cs, buf, usage)) {
227 if (ctx->rings.dma.cs) {
228 if (ctx->ws->cs_is_buffer_referenced(ctx->rings.dma.cs, buf, usage)) {
235 void *r600_buffer_mmap_sync_with_rings(struct r600_context *ctx,
236 struct r600_resource *resource,
239 enum radeon_bo_usage rusage = RADEON_USAGE_READWRITE;
241 bool sync_flush = TRUE;
243 if (usage & PIPE_TRANSFER_UNSYNCHRONIZED) {
244 return ctx->ws->buffer_map(resource->cs_buf, NULL, usage);
247 if (!(usage & PIPE_TRANSFER_WRITE)) {
248 /* have to wait for pending read */
249 rusage = RADEON_USAGE_WRITE;
251 if (usage & PIPE_TRANSFER_DONTBLOCK) {
252 flags |= RADEON_FLUSH_ASYNC;
255 if (ctx->ws->cs_is_buffer_referenced(ctx->rings.gfx.cs, resource->cs_buf, rusage) && ctx->rings.gfx.cs->cdw) {
256 ctx->rings.gfx.flush(ctx, flags);
257 if (usage & PIPE_TRANSFER_DONTBLOCK) {
261 if (ctx->rings.dma.cs) {
262 if (ctx->ws->cs_is_buffer_referenced(ctx->rings.dma.cs, resource->cs_buf, rusage) && ctx->rings.dma.cs->cdw) {
263 ctx->rings.dma.flush(ctx, flags);
264 if (usage & PIPE_TRANSFER_DONTBLOCK) {
270 if (usage & PIPE_TRANSFER_DONTBLOCK) {
271 if (ctx->ws->buffer_is_busy(resource->buf, rusage)) {
276 /* Try to avoid busy-waiting in radeon_bo_wait. */
277 ctx->ws->cs_sync_flush(ctx->rings.gfx.cs);
278 if (ctx->rings.dma.cs) {
279 ctx->ws->cs_sync_flush(ctx->rings.dma.cs);
282 ctx->ws->buffer_wait(resource->buf, rusage);
284 /* at this point everything is synchronized */
285 return ctx->ws->buffer_map(resource->cs_buf, NULL, usage | PIPE_TRANSFER_UNSYNCHRONIZED);
288 static void r600_flush_from_winsys(void *ctx, unsigned flags)
290 struct r600_context *rctx = (struct r600_context *)ctx;
292 rctx->rings.gfx.flush(rctx, flags);
295 static void r600_flush_dma_from_winsys(void *ctx, unsigned flags)
297 struct r600_context *rctx = (struct r600_context *)ctx;
299 rctx->rings.dma.flush(rctx, flags);
302 static void r600_destroy_context(struct pipe_context *context)
304 struct r600_context *rctx = (struct r600_context *)context;
306 r600_isa_destroy(rctx->isa);
308 pipe_resource_reference((struct pipe_resource**)&rctx->dummy_cmask, NULL);
309 pipe_resource_reference((struct pipe_resource**)&rctx->dummy_fmask, NULL);
311 if (rctx->dummy_pixel_shader) {
312 rctx->context.delete_fs_state(&rctx->context, rctx->dummy_pixel_shader);
314 if (rctx->custom_dsa_flush) {
315 rctx->context.delete_depth_stencil_alpha_state(&rctx->context, rctx->custom_dsa_flush);
317 if (rctx->custom_blend_resolve) {
318 rctx->context.delete_blend_state(&rctx->context, rctx->custom_blend_resolve);
320 if (rctx->custom_blend_decompress) {
321 rctx->context.delete_blend_state(&rctx->context, rctx->custom_blend_decompress);
323 if (rctx->custom_blend_fmask_decompress) {
324 rctx->context.delete_blend_state(&rctx->context, rctx->custom_blend_fmask_decompress);
326 util_unreference_framebuffer_state(&rctx->framebuffer.state);
329 util_blitter_destroy(rctx->blitter);
331 if (rctx->uploader) {
332 u_upload_destroy(rctx->uploader);
334 if (rctx->allocator_so_filled_size) {
335 u_suballocator_destroy(rctx->allocator_so_filled_size);
337 if (rctx->allocator_fetch_shader) {
338 u_suballocator_destroy(rctx->allocator_fetch_shader);
340 util_slab_destroy(&rctx->pool_transfers);
342 r600_release_command_buffer(&rctx->start_cs_cmd);
344 if (rctx->rings.gfx.cs) {
345 rctx->ws->cs_destroy(rctx->rings.gfx.cs);
347 if (rctx->rings.dma.cs) {
348 rctx->ws->cs_destroy(rctx->rings.dma.cs);
354 static struct pipe_context *r600_create_context(struct pipe_screen *screen, void *priv)
356 struct r600_context *rctx = CALLOC_STRUCT(r600_context);
357 struct r600_screen* rscreen = (struct r600_screen *)screen;
362 util_slab_create(&rctx->pool_transfers,
363 sizeof(struct r600_transfer), 64,
364 UTIL_SLAB_SINGLETHREADED);
366 rctx->context.screen = screen;
367 rctx->context.priv = priv;
368 rctx->context.destroy = r600_destroy_context;
369 rctx->context.flush = r600_flush_from_st;
371 /* Easy accessing of screen/winsys. */
372 rctx->screen = rscreen;
373 rctx->ws = rscreen->ws;
374 rctx->family = rscreen->family;
375 rctx->chip_class = rscreen->chip_class;
376 rctx->keep_tiling_flags = rscreen->info.drm_minor >= 12;
378 LIST_INITHEAD(&rctx->active_nontimer_queries);
380 r600_init_blit_functions(rctx);
381 r600_init_query_functions(rctx);
382 r600_init_context_resource_functions(rctx);
383 r600_init_surface_functions(rctx);
385 if (rscreen->info.has_uvd) {
386 rctx->context.create_video_decoder = r600_uvd_create_decoder;
387 rctx->context.create_video_buffer = r600_video_buffer_create;
389 rctx->context.create_video_decoder = vl_create_decoder;
390 rctx->context.create_video_buffer = vl_video_buffer_create;
393 r600_init_common_state_functions(rctx);
395 switch (rctx->chip_class) {
398 r600_init_state_functions(rctx);
399 r600_init_atom_start_cs(rctx);
401 rctx->custom_dsa_flush = r600_create_db_flush_dsa(rctx);
402 rctx->custom_blend_resolve = rctx->chip_class == R700 ? r700_create_resolve_blend(rctx)
403 : r600_create_resolve_blend(rctx);
404 rctx->custom_blend_decompress = r600_create_decompress_blend(rctx);
405 rctx->has_vertex_cache = !(rctx->family == CHIP_RV610 ||
406 rctx->family == CHIP_RV620 ||
407 rctx->family == CHIP_RS780 ||
408 rctx->family == CHIP_RS880 ||
409 rctx->family == CHIP_RV710);
413 evergreen_init_state_functions(rctx);
414 evergreen_init_atom_start_cs(rctx);
415 evergreen_init_atom_start_compute_cs(rctx);
417 rctx->custom_dsa_flush = evergreen_create_db_flush_dsa(rctx);
418 rctx->custom_blend_resolve = evergreen_create_resolve_blend(rctx);
419 rctx->custom_blend_decompress = evergreen_create_decompress_blend(rctx);
420 rctx->custom_blend_fmask_decompress = evergreen_create_fmask_decompress_blend(rctx);
421 rctx->has_vertex_cache = !(rctx->family == CHIP_CEDAR ||
422 rctx->family == CHIP_PALM ||
423 rctx->family == CHIP_SUMO ||
424 rctx->family == CHIP_SUMO2 ||
425 rctx->family == CHIP_CAICOS ||
426 rctx->family == CHIP_CAYMAN ||
427 rctx->family == CHIP_ARUBA);
430 R600_ERR("Unsupported chip class %d.\n", rctx->chip_class);
434 if (rscreen->trace_bo) {
435 rctx->rings.gfx.cs = rctx->ws->cs_create(rctx->ws, RING_GFX, rscreen->trace_bo->cs_buf);
437 rctx->rings.gfx.cs = rctx->ws->cs_create(rctx->ws, RING_GFX, NULL);
439 rctx->rings.gfx.flush = r600_flush_gfx_ring;
440 rctx->ws->cs_set_flush_callback(rctx->rings.gfx.cs, r600_flush_from_winsys, rctx);
441 rctx->rings.gfx.flushing = false;
443 rctx->rings.dma.cs = NULL;
444 if (rscreen->info.r600_has_dma && !(rscreen->debug_flags & DBG_NO_ASYNC_DMA)) {
445 rctx->rings.dma.cs = rctx->ws->cs_create(rctx->ws, RING_DMA, NULL);
446 rctx->rings.dma.flush = r600_flush_dma_ring;
447 rctx->ws->cs_set_flush_callback(rctx->rings.dma.cs, r600_flush_dma_from_winsys, rctx);
448 rctx->rings.dma.flushing = false;
451 rctx->uploader = u_upload_create(&rctx->context, 1024 * 1024, 256,
452 PIPE_BIND_INDEX_BUFFER |
453 PIPE_BIND_CONSTANT_BUFFER);
457 rctx->allocator_fetch_shader = u_suballocator_create(&rctx->context, 64 * 1024, 256,
458 0, PIPE_USAGE_STATIC, FALSE);
459 if (!rctx->allocator_fetch_shader)
462 rctx->allocator_so_filled_size = u_suballocator_create(&rctx->context, 4096, 4,
463 0, PIPE_USAGE_STATIC, TRUE);
464 if (!rctx->allocator_so_filled_size)
467 rctx->isa = calloc(1, sizeof(struct r600_isa));
468 if (!rctx->isa || r600_isa_init(rctx, rctx->isa))
471 rctx->blitter = util_blitter_create(&rctx->context);
472 if (rctx->blitter == NULL)
474 util_blitter_set_texture_multisample(rctx->blitter, rscreen->has_msaa);
475 rctx->blitter->draw_rectangle = r600_draw_rectangle;
477 r600_begin_new_cs(rctx);
478 r600_get_backend_mask(rctx); /* this emits commands and must be last */
480 rctx->dummy_pixel_shader =
481 util_make_fragment_cloneinput_shader(&rctx->context, 0,
482 TGSI_SEMANTIC_GENERIC,
483 TGSI_INTERPOLATE_CONSTANT);
484 rctx->context.bind_fs_state(&rctx->context, rctx->dummy_pixel_shader);
486 return &rctx->context;
489 r600_destroy_context(&rctx->context);
496 static const char* r600_get_vendor(struct pipe_screen* pscreen)
501 static const char *r600_get_family_name(enum radeon_family family)
504 case CHIP_R600: return "AMD R600";
505 case CHIP_RV610: return "AMD RV610";
506 case CHIP_RV630: return "AMD RV630";
507 case CHIP_RV670: return "AMD RV670";
508 case CHIP_RV620: return "AMD RV620";
509 case CHIP_RV635: return "AMD RV635";
510 case CHIP_RS780: return "AMD RS780";
511 case CHIP_RS880: return "AMD RS880";
512 case CHIP_RV770: return "AMD RV770";
513 case CHIP_RV730: return "AMD RV730";
514 case CHIP_RV710: return "AMD RV710";
515 case CHIP_RV740: return "AMD RV740";
516 case CHIP_CEDAR: return "AMD CEDAR";
517 case CHIP_REDWOOD: return "AMD REDWOOD";
518 case CHIP_JUNIPER: return "AMD JUNIPER";
519 case CHIP_CYPRESS: return "AMD CYPRESS";
520 case CHIP_HEMLOCK: return "AMD HEMLOCK";
521 case CHIP_PALM: return "AMD PALM";
522 case CHIP_SUMO: return "AMD SUMO";
523 case CHIP_SUMO2: return "AMD SUMO2";
524 case CHIP_BARTS: return "AMD BARTS";
525 case CHIP_TURKS: return "AMD TURKS";
526 case CHIP_CAICOS: return "AMD CAICOS";
527 case CHIP_CAYMAN: return "AMD CAYMAN";
528 case CHIP_ARUBA: return "AMD ARUBA";
529 default: return "AMD unknown";
533 static const char* r600_get_name(struct pipe_screen* pscreen)
535 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
537 return r600_get_family_name(rscreen->family);
540 static int r600_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
542 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
543 enum radeon_family family = rscreen->family;
546 /* Supported features (boolean caps). */
547 case PIPE_CAP_NPOT_TEXTURES:
548 case PIPE_CAP_TWO_SIDED_STENCIL:
549 case PIPE_CAP_ANISOTROPIC_FILTER:
550 case PIPE_CAP_POINT_SPRITE:
551 case PIPE_CAP_OCCLUSION_QUERY:
552 case PIPE_CAP_TEXTURE_SHADOW_MAP:
553 case PIPE_CAP_TEXTURE_MIRROR_CLAMP:
554 case PIPE_CAP_BLEND_EQUATION_SEPARATE:
555 case PIPE_CAP_TEXTURE_SWIZZLE:
556 case PIPE_CAP_DEPTH_CLIP_DISABLE:
557 case PIPE_CAP_SHADER_STENCIL_EXPORT:
558 case PIPE_CAP_VERTEX_ELEMENT_INSTANCE_DIVISOR:
559 case PIPE_CAP_MIXED_COLORBUFFER_FORMATS:
560 case PIPE_CAP_TGSI_FS_COORD_ORIGIN_UPPER_LEFT:
561 case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_HALF_INTEGER:
563 case PIPE_CAP_SEAMLESS_CUBE_MAP:
564 case PIPE_CAP_PRIMITIVE_RESTART:
565 case PIPE_CAP_CONDITIONAL_RENDER:
566 case PIPE_CAP_TEXTURE_BARRIER:
567 case PIPE_CAP_VERTEX_COLOR_UNCLAMPED:
568 case PIPE_CAP_QUADS_FOLLOW_PROVOKING_VERTEX_CONVENTION:
569 case PIPE_CAP_TGSI_INSTANCEID:
570 case PIPE_CAP_VERTEX_BUFFER_OFFSET_4BYTE_ALIGNED_ONLY:
571 case PIPE_CAP_VERTEX_BUFFER_STRIDE_4BYTE_ALIGNED_ONLY:
572 case PIPE_CAP_VERTEX_ELEMENT_SRC_OFFSET_4BYTE_ALIGNED_ONLY:
573 case PIPE_CAP_USER_INDEX_BUFFERS:
574 case PIPE_CAP_USER_CONSTANT_BUFFERS:
575 case PIPE_CAP_COMPUTE:
576 case PIPE_CAP_START_INSTANCE:
577 case PIPE_CAP_MAX_DUAL_SOURCE_RENDER_TARGETS:
578 case PIPE_CAP_TEXTURE_BUFFER_OBJECTS:
579 case PIPE_CAP_PREFER_BLIT_BASED_TEXTURE_TRANSFER:
580 case PIPE_CAP_QUERY_PIPELINE_STATISTICS:
582 case PIPE_CAP_TGSI_TEXCOORD:
585 case PIPE_CAP_MIN_MAP_BUFFER_ALIGNMENT:
586 return R600_MAP_BUFFER_ALIGNMENT;
588 case PIPE_CAP_CONSTANT_BUFFER_OFFSET_ALIGNMENT:
591 case PIPE_CAP_TEXTURE_BUFFER_OFFSET_ALIGNMENT:
594 case PIPE_CAP_GLSL_FEATURE_LEVEL:
597 case PIPE_CAP_TEXTURE_MULTISAMPLE:
598 return rscreen->msaa_texture_support != MSAA_TEXTURE_SAMPLE_ZERO;
600 /* Supported except the original R600. */
601 case PIPE_CAP_INDEP_BLEND_ENABLE:
602 case PIPE_CAP_INDEP_BLEND_FUNC:
603 /* R600 doesn't support per-MRT blends */
604 return family == CHIP_R600 ? 0 : 1;
606 /* Supported on Evergreen. */
607 case PIPE_CAP_SEAMLESS_CUBE_MAP_PER_TEXTURE:
608 case PIPE_CAP_CUBE_MAP_ARRAY:
609 return family >= CHIP_CEDAR ? 1 : 0;
611 /* Unsupported features. */
612 case PIPE_CAP_TGSI_FS_COORD_ORIGIN_LOWER_LEFT:
613 case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_INTEGER:
614 case PIPE_CAP_SCALED_RESOLVE:
615 case PIPE_CAP_TGSI_CAN_COMPACT_VARYINGS:
616 case PIPE_CAP_TGSI_CAN_COMPACT_CONSTANTS:
617 case PIPE_CAP_FRAGMENT_COLOR_CLAMPED:
618 case PIPE_CAP_VERTEX_COLOR_CLAMPED:
619 case PIPE_CAP_USER_VERTEX_BUFFERS:
623 case PIPE_CAP_MAX_STREAM_OUTPUT_BUFFERS:
624 return rscreen->has_streamout ? 4 : 0;
625 case PIPE_CAP_STREAM_OUTPUT_PAUSE_RESUME:
626 return rscreen->has_streamout ? 1 : 0;
627 case PIPE_CAP_MAX_STREAM_OUTPUT_SEPARATE_COMPONENTS:
628 case PIPE_CAP_MAX_STREAM_OUTPUT_INTERLEAVED_COMPONENTS:
632 case PIPE_CAP_MAX_TEXTURE_2D_LEVELS:
633 case PIPE_CAP_MAX_TEXTURE_3D_LEVELS:
634 case PIPE_CAP_MAX_TEXTURE_CUBE_LEVELS:
635 if (family >= CHIP_CEDAR)
639 case PIPE_CAP_MAX_TEXTURE_ARRAY_LAYERS:
640 return rscreen->info.drm_minor >= 9 ?
641 (family >= CHIP_CEDAR ? 16384 : 8192) : 0;
642 case PIPE_CAP_MAX_COMBINED_SAMPLERS:
645 /* Render targets. */
646 case PIPE_CAP_MAX_RENDER_TARGETS:
647 /* XXX some r6xx are buggy and can only do 4 */
650 /* Timer queries, present when the clock frequency is non zero. */
651 case PIPE_CAP_QUERY_TIME_ELAPSED:
652 return rscreen->info.r600_clock_crystal_freq != 0;
653 case PIPE_CAP_QUERY_TIMESTAMP:
654 return rscreen->info.drm_minor >= 20 &&
655 rscreen->info.r600_clock_crystal_freq != 0;
657 case PIPE_CAP_MIN_TEXEL_OFFSET:
660 case PIPE_CAP_MAX_TEXEL_OFFSET:
663 case PIPE_CAP_TEXTURE_BORDER_COLOR_QUIRK:
664 return PIPE_QUIRK_TEXTURE_BORDER_COLOR_SWIZZLE_R600;
669 static float r600_get_paramf(struct pipe_screen* pscreen,
670 enum pipe_capf param)
672 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
673 enum radeon_family family = rscreen->family;
676 case PIPE_CAPF_MAX_LINE_WIDTH:
677 case PIPE_CAPF_MAX_LINE_WIDTH_AA:
678 case PIPE_CAPF_MAX_POINT_WIDTH:
679 case PIPE_CAPF_MAX_POINT_WIDTH_AA:
680 if (family >= CHIP_CEDAR)
684 case PIPE_CAPF_MAX_TEXTURE_ANISOTROPY:
686 case PIPE_CAPF_MAX_TEXTURE_LOD_BIAS:
688 case PIPE_CAPF_GUARD_BAND_LEFT:
689 case PIPE_CAPF_GUARD_BAND_TOP:
690 case PIPE_CAPF_GUARD_BAND_RIGHT:
691 case PIPE_CAPF_GUARD_BAND_BOTTOM:
697 static int r600_get_shader_param(struct pipe_screen* pscreen, unsigned shader, enum pipe_shader_cap param)
701 case PIPE_SHADER_FRAGMENT:
702 case PIPE_SHADER_VERTEX:
703 case PIPE_SHADER_COMPUTE:
705 case PIPE_SHADER_GEOMETRY:
706 /* XXX: support and enable geometry programs */
709 /* XXX: support tessellation on Evergreen */
714 case PIPE_SHADER_CAP_MAX_INSTRUCTIONS:
715 case PIPE_SHADER_CAP_MAX_ALU_INSTRUCTIONS:
716 case PIPE_SHADER_CAP_MAX_TEX_INSTRUCTIONS:
717 case PIPE_SHADER_CAP_MAX_TEX_INDIRECTIONS:
719 case PIPE_SHADER_CAP_MAX_CONTROL_FLOW_DEPTH:
721 case PIPE_SHADER_CAP_MAX_INPUTS:
723 case PIPE_SHADER_CAP_MAX_TEMPS:
724 return 256; /* Max native temporaries. */
725 case PIPE_SHADER_CAP_MAX_ADDRS:
726 /* XXX Isn't this equal to TEMPS? */
727 return 1; /* Max native address registers */
728 case PIPE_SHADER_CAP_MAX_CONSTS:
729 return R600_MAX_CONST_BUFFER_SIZE;
730 case PIPE_SHADER_CAP_MAX_CONST_BUFFERS:
731 return R600_MAX_USER_CONST_BUFFERS;
732 case PIPE_SHADER_CAP_MAX_PREDS:
733 return 0; /* nothing uses this */
734 case PIPE_SHADER_CAP_TGSI_CONT_SUPPORTED:
736 case PIPE_SHADER_CAP_TGSI_SQRT_SUPPORTED:
738 case PIPE_SHADER_CAP_INDIRECT_INPUT_ADDR:
739 case PIPE_SHADER_CAP_INDIRECT_OUTPUT_ADDR:
740 case PIPE_SHADER_CAP_INDIRECT_TEMP_ADDR:
741 case PIPE_SHADER_CAP_INDIRECT_CONST_ADDR:
743 case PIPE_SHADER_CAP_SUBROUTINES:
745 case PIPE_SHADER_CAP_INTEGERS:
747 case PIPE_SHADER_CAP_MAX_TEXTURE_SAMPLERS:
749 case PIPE_SHADER_CAP_PREFERRED_IR:
750 if (shader == PIPE_SHADER_COMPUTE) {
751 return PIPE_SHADER_IR_LLVM;
753 return PIPE_SHADER_IR_TGSI;
759 static int r600_get_video_param(struct pipe_screen *screen,
760 enum pipe_video_profile profile,
761 enum pipe_video_cap param)
764 case PIPE_VIDEO_CAP_SUPPORTED:
765 return vl_profile_supported(screen, profile);
766 case PIPE_VIDEO_CAP_NPOT_TEXTURES:
768 case PIPE_VIDEO_CAP_MAX_WIDTH:
769 case PIPE_VIDEO_CAP_MAX_HEIGHT:
770 return vl_video_buffer_max_size(screen);
771 case PIPE_VIDEO_CAP_PREFERED_FORMAT:
772 return PIPE_FORMAT_NV12;
773 case PIPE_VIDEO_CAP_PREFERS_INTERLACED:
775 case PIPE_VIDEO_CAP_SUPPORTS_INTERLACED:
777 case PIPE_VIDEO_CAP_SUPPORTS_PROGRESSIVE:
784 const char * r600_llvm_gpu_string(enum radeon_family family)
786 const char * gpu_family;
800 gpu_family = "rv710";
803 gpu_family = "rv730";
807 gpu_family = "rv770";
811 gpu_family = "cedar";
816 gpu_family = "redwood";
819 gpu_family = "juniper";
823 gpu_family = "cypress";
826 gpu_family = "barts";
829 gpu_family = "turks";
832 gpu_family = "caicos";
836 gpu_family = "cayman";
840 fprintf(stderr, "Chip not supported by r600 llvm "
841 "backend, please file a bug at " PACKAGE_BUGREPORT "\n");
848 static int r600_get_compute_param(struct pipe_screen *screen,
849 enum pipe_compute_cap param,
852 struct r600_screen *rscreen = (struct r600_screen *)screen;
853 //TODO: select these params by asic
855 case PIPE_COMPUTE_CAP_IR_TARGET: {
856 const char *gpu = r600_llvm_gpu_string(rscreen->family);
858 sprintf(ret, "%s-r600--", gpu);
860 return (8 + strlen(gpu)) * sizeof(char);
862 case PIPE_COMPUTE_CAP_GRID_DIMENSION:
864 uint64_t * grid_dimension = ret;
865 grid_dimension[0] = 3;
867 return 1 * sizeof(uint64_t);
869 case PIPE_COMPUTE_CAP_MAX_GRID_SIZE:
871 uint64_t * grid_size = ret;
872 grid_size[0] = 65535;
873 grid_size[1] = 65535;
876 return 3 * sizeof(uint64_t) ;
878 case PIPE_COMPUTE_CAP_MAX_BLOCK_SIZE:
880 uint64_t * block_size = ret;
885 return 3 * sizeof(uint64_t);
887 case PIPE_COMPUTE_CAP_MAX_THREADS_PER_BLOCK:
889 uint64_t * max_threads_per_block = ret;
890 *max_threads_per_block = 256;
892 return sizeof(uint64_t);
894 case PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE:
896 uint64_t * max_global_size = ret;
897 /* XXX: This is what the proprietary driver reports, we
898 * may want to use a different value. */
899 *max_global_size = 201326592;
901 return sizeof(uint64_t);
903 case PIPE_COMPUTE_CAP_MAX_INPUT_SIZE:
905 uint64_t * max_input_size = ret;
906 *max_input_size = 1024;
908 return sizeof(uint64_t);
910 case PIPE_COMPUTE_CAP_MAX_LOCAL_SIZE:
912 uint64_t * max_local_size = ret;
913 /* XXX: This is what the proprietary driver reports, we
914 * may want to use a different value. */
915 *max_local_size = 32768;
917 return sizeof(uint64_t);
919 case PIPE_COMPUTE_CAP_MAX_MEM_ALLOC_SIZE:
921 uint64_t max_global_size;
922 uint64_t * max_mem_alloc_size = ret;
923 r600_get_compute_param(screen,
924 PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE,
926 /* OpenCL requres this value be at least
927 * max(MAX_GLOBAL_SIZE / 4, 128 * 1024 *1024)
928 * I'm really not sure what value to report here, but
929 * MAX_GLOBAL_SIZE / 4 seems resonable.
931 *max_mem_alloc_size = max_global_size / 4;
933 return sizeof(uint64_t);
936 fprintf(stderr, "unknown PIPE_COMPUTE_CAP %d\n", param);
941 static void r600_destroy_screen(struct pipe_screen* pscreen)
943 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
948 pipe_mutex_destroy(rscreen->aux_context_lock);
949 rscreen->aux_context->destroy(rscreen->aux_context);
951 if (rscreen->global_pool) {
952 compute_memory_pool_delete(rscreen->global_pool);
955 if (rscreen->fences.bo) {
956 struct r600_fence_block *entry, *tmp;
958 LIST_FOR_EACH_ENTRY_SAFE(entry, tmp, &rscreen->fences.blocks, head) {
959 LIST_DEL(&entry->head);
963 rscreen->ws->buffer_unmap(rscreen->fences.bo->cs_buf);
964 pipe_resource_reference((struct pipe_resource**)&rscreen->fences.bo, NULL);
966 if (rscreen->trace_bo) {
967 rscreen->ws->buffer_unmap(rscreen->trace_bo->cs_buf);
968 pipe_resource_reference((struct pipe_resource**)&rscreen->trace_bo, NULL);
970 pipe_mutex_destroy(rscreen->fences.mutex);
972 rscreen->ws->destroy(rscreen->ws);
976 static void r600_fence_reference(struct pipe_screen *pscreen,
977 struct pipe_fence_handle **ptr,
978 struct pipe_fence_handle *fence)
980 struct r600_fence **oldf = (struct r600_fence**)ptr;
981 struct r600_fence *newf = (struct r600_fence*)fence;
983 if (pipe_reference(&(*oldf)->reference, &newf->reference)) {
984 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
985 pipe_mutex_lock(rscreen->fences.mutex);
986 pipe_resource_reference((struct pipe_resource**)&(*oldf)->sleep_bo, NULL);
987 LIST_ADDTAIL(&(*oldf)->head, &rscreen->fences.pool);
988 pipe_mutex_unlock(rscreen->fences.mutex);
994 static boolean r600_fence_signalled(struct pipe_screen *pscreen,
995 struct pipe_fence_handle *fence)
997 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
998 struct r600_fence *rfence = (struct r600_fence*)fence;
1000 return rscreen->fences.data[rfence->index] != 0;
1003 static boolean r600_fence_finish(struct pipe_screen *pscreen,
1004 struct pipe_fence_handle *fence,
1007 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
1008 struct r600_fence *rfence = (struct r600_fence*)fence;
1009 int64_t start_time = 0;
1012 if (timeout != PIPE_TIMEOUT_INFINITE) {
1013 start_time = os_time_get();
1015 /* Convert to microseconds. */
1019 while (rscreen->fences.data[rfence->index] == 0) {
1020 /* Special-case infinite timeout - wait for the dummy BO to become idle */
1021 if (timeout == PIPE_TIMEOUT_INFINITE) {
1022 rscreen->ws->buffer_wait(rfence->sleep_bo->buf, RADEON_USAGE_READWRITE);
1026 /* The dummy BO will be busy until the CS including the fence has completed, or
1027 * the GPU is reset. Don't bother continuing to spin when the BO is idle. */
1028 if (!rscreen->ws->buffer_is_busy(rfence->sleep_bo->buf, RADEON_USAGE_READWRITE))
1038 if (timeout != PIPE_TIMEOUT_INFINITE &&
1039 os_time_get() - start_time >= timeout) {
1044 return rscreen->fences.data[rfence->index] != 0;
1047 static int r600_interpret_tiling(struct r600_screen *rscreen, uint32_t tiling_config)
1049 switch ((tiling_config & 0xe) >> 1) {
1051 rscreen->tiling_info.num_channels = 1;
1054 rscreen->tiling_info.num_channels = 2;
1057 rscreen->tiling_info.num_channels = 4;
1060 rscreen->tiling_info.num_channels = 8;
1066 switch ((tiling_config & 0x30) >> 4) {
1068 rscreen->tiling_info.num_banks = 4;
1071 rscreen->tiling_info.num_banks = 8;
1077 switch ((tiling_config & 0xc0) >> 6) {
1079 rscreen->tiling_info.group_bytes = 256;
1082 rscreen->tiling_info.group_bytes = 512;
1090 static int evergreen_interpret_tiling(struct r600_screen *rscreen, uint32_t tiling_config)
1092 switch (tiling_config & 0xf) {
1094 rscreen->tiling_info.num_channels = 1;
1097 rscreen->tiling_info.num_channels = 2;
1100 rscreen->tiling_info.num_channels = 4;
1103 rscreen->tiling_info.num_channels = 8;
1109 switch ((tiling_config & 0xf0) >> 4) {
1111 rscreen->tiling_info.num_banks = 4;
1114 rscreen->tiling_info.num_banks = 8;
1117 rscreen->tiling_info.num_banks = 16;
1123 switch ((tiling_config & 0xf00) >> 8) {
1125 rscreen->tiling_info.group_bytes = 256;
1128 rscreen->tiling_info.group_bytes = 512;
1136 static int r600_init_tiling(struct r600_screen *rscreen)
1138 uint32_t tiling_config = rscreen->info.r600_tiling_config;
1140 /* set default group bytes, overridden by tiling info ioctl */
1141 if (rscreen->chip_class <= R700) {
1142 rscreen->tiling_info.group_bytes = 256;
1144 rscreen->tiling_info.group_bytes = 512;
1150 if (rscreen->chip_class <= R700) {
1151 return r600_interpret_tiling(rscreen, tiling_config);
1153 return evergreen_interpret_tiling(rscreen, tiling_config);
1157 static uint64_t r600_get_timestamp(struct pipe_screen *screen)
1159 struct r600_screen *rscreen = (struct r600_screen*)screen;
1161 return 1000000 * rscreen->ws->query_value(rscreen->ws, RADEON_TIMESTAMP) /
1162 rscreen->info.r600_clock_crystal_freq;
1165 static int r600_get_driver_query_info(struct pipe_screen *screen,
1167 struct pipe_driver_query_info *info)
1169 struct r600_screen *rscreen = (struct r600_screen*)screen;
1170 struct pipe_driver_query_info list[] = {
1171 {"draw-calls", R600_QUERY_DRAW_CALLS, 0},
1172 {"requested-VRAM", R600_QUERY_REQUESTED_VRAM, rscreen->info.vram_size, TRUE},
1173 {"requested-GTT", R600_QUERY_REQUESTED_GTT, rscreen->info.gart_size, TRUE},
1174 {"buffer-wait-time", R600_QUERY_BUFFER_WAIT_TIME, 0, FALSE}
1178 return Elements(list);
1180 if (index >= Elements(list))
1183 *info = list[index];
1187 struct pipe_screen *r600_screen_create(struct radeon_winsys *ws)
1189 struct r600_screen *rscreen = CALLOC_STRUCT(r600_screen);
1191 if (rscreen == NULL) {
1196 ws->query_info(ws, &rscreen->info);
1198 rscreen->debug_flags = debug_get_flags_option("R600_DEBUG", debug_options, 0);
1199 if (debug_get_bool_option("R600_DEBUG_COMPUTE", FALSE))
1200 rscreen->debug_flags |= DBG_COMPUTE;
1201 if (debug_get_bool_option("R600_DUMP_SHADERS", FALSE))
1202 rscreen->debug_flags |= DBG_FS | DBG_VS | DBG_GS | DBG_PS | DBG_CS;
1203 if (!debug_get_bool_option("R600_HYPERZ", TRUE))
1204 rscreen->debug_flags |= DBG_NO_HYPERZ;
1205 if (!debug_get_bool_option("R600_LLVM", TRUE))
1206 rscreen->debug_flags |= DBG_NO_LLVM;
1207 if (debug_get_bool_option("R600_PRINT_TEXDEPTH", FALSE))
1208 rscreen->debug_flags |= DBG_TEX_DEPTH;
1209 rscreen->family = rscreen->info.family;
1210 rscreen->chip_class = rscreen->info.chip_class;
1212 if (rscreen->family == CHIP_UNKNOWN) {
1213 fprintf(stderr, "r600: Unknown chipset 0x%04X\n", rscreen->info.pci_id);
1218 /* Figure out streamout kernel support. */
1219 switch (rscreen->chip_class) {
1221 if (rscreen->family < CHIP_RS780) {
1222 rscreen->has_streamout = rscreen->info.drm_minor >= 14;
1224 rscreen->has_streamout = rscreen->info.drm_minor >= 23;
1228 rscreen->has_streamout = rscreen->info.drm_minor >= 17;
1232 rscreen->has_streamout = rscreen->info.drm_minor >= 14;
1235 rscreen->has_streamout = FALSE;
1240 switch (rscreen->chip_class) {
1243 rscreen->has_msaa = rscreen->info.drm_minor >= 22;
1244 rscreen->msaa_texture_support = MSAA_TEXTURE_DECOMPRESSED;
1247 rscreen->has_msaa = rscreen->info.drm_minor >= 19;
1248 rscreen->msaa_texture_support =
1249 rscreen->info.drm_minor >= 24 ? MSAA_TEXTURE_COMPRESSED :
1250 MSAA_TEXTURE_DECOMPRESSED;
1253 rscreen->has_msaa = rscreen->info.drm_minor >= 19;
1254 /* We should be able to read compressed MSAA textures, but it doesn't work. */
1255 rscreen->msaa_texture_support = MSAA_TEXTURE_SAMPLE_ZERO;
1258 rscreen->has_msaa = FALSE;
1259 rscreen->msaa_texture_support = 0;
1263 rscreen->has_cp_dma = rscreen->info.drm_minor >= 27 &&
1264 !(rscreen->debug_flags & DBG_NO_CP_DMA);
1266 if (r600_init_tiling(rscreen)) {
1271 rscreen->screen.destroy = r600_destroy_screen;
1272 rscreen->screen.get_name = r600_get_name;
1273 rscreen->screen.get_vendor = r600_get_vendor;
1274 rscreen->screen.get_param = r600_get_param;
1275 rscreen->screen.get_shader_param = r600_get_shader_param;
1276 rscreen->screen.get_paramf = r600_get_paramf;
1277 rscreen->screen.get_compute_param = r600_get_compute_param;
1278 rscreen->screen.get_timestamp = r600_get_timestamp;
1280 if (rscreen->chip_class >= EVERGREEN) {
1281 rscreen->screen.is_format_supported = evergreen_is_format_supported;
1282 rscreen->dma_blit = &evergreen_dma_blit;
1284 rscreen->screen.is_format_supported = r600_is_format_supported;
1285 rscreen->dma_blit = &r600_dma_blit;
1287 rscreen->screen.context_create = r600_create_context;
1288 rscreen->screen.fence_reference = r600_fence_reference;
1289 rscreen->screen.fence_signalled = r600_fence_signalled;
1290 rscreen->screen.fence_finish = r600_fence_finish;
1291 rscreen->screen.get_driver_query_info = r600_get_driver_query_info;
1293 if (rscreen->info.has_uvd) {
1294 rscreen->screen.get_video_param = ruvd_get_video_param;
1295 rscreen->screen.is_video_format_supported = ruvd_is_format_supported;
1297 rscreen->screen.get_video_param = r600_get_video_param;
1298 rscreen->screen.is_video_format_supported = vl_video_buffer_is_format_supported;
1301 r600_init_screen_resource_functions(&rscreen->screen);
1303 util_format_s3tc_init();
1305 rscreen->fences.bo = NULL;
1306 rscreen->fences.data = NULL;
1307 rscreen->fences.next_index = 0;
1308 LIST_INITHEAD(&rscreen->fences.pool);
1309 LIST_INITHEAD(&rscreen->fences.blocks);
1310 pipe_mutex_init(rscreen->fences.mutex);
1312 rscreen->global_pool = compute_memory_pool_new(rscreen);
1314 rscreen->cs_count = 0;
1315 if (rscreen->info.drm_minor >= 28 && (rscreen->debug_flags & DBG_TRACE_CS)) {
1316 rscreen->trace_bo = (struct r600_resource*)pipe_buffer_create(&rscreen->screen,
1320 if (rscreen->trace_bo) {
1321 rscreen->trace_ptr = rscreen->ws->buffer_map(rscreen->trace_bo->cs_buf, NULL,
1322 PIPE_TRANSFER_UNSYNCHRONIZED);
1326 /* Create the auxiliary context. */
1327 pipe_mutex_init(rscreen->aux_context_lock);
1328 rscreen->aux_context = rscreen->screen.context_create(&rscreen->screen, NULL);
1330 #if 0 /* This is for testing whether aux_context and buffer clearing work correctly. */
1331 struct pipe_resource templ = {};
1334 templ.height0 = 2048;
1336 templ.array_size = 1;
1337 templ.target = PIPE_TEXTURE_2D;
1338 templ.format = PIPE_FORMAT_R8G8B8A8_UNORM;
1339 templ.usage = PIPE_USAGE_STATIC;
1341 struct r600_resource *res = r600_resource(rscreen->screen.resource_create(&rscreen->screen, &templ));
1342 unsigned char *map = ws->buffer_map(res->cs_buf, NULL, PIPE_TRANSFER_WRITE);
1344 memset(map, 0, 256);
1346 r600_screen_clear_buffer(rscreen, &res->b.b, 4, 4, 0xCC);
1347 r600_screen_clear_buffer(rscreen, &res->b.b, 8, 4, 0xDD);
1348 r600_screen_clear_buffer(rscreen, &res->b.b, 12, 4, 0xEE);
1349 r600_screen_clear_buffer(rscreen, &res->b.b, 20, 4, 0xFF);
1350 r600_screen_clear_buffer(rscreen, &res->b.b, 32, 20, 0x87);
1352 ws->buffer_wait(res->buf, RADEON_USAGE_WRITE);
1355 for (i = 0; i < 256; i++) {
1356 printf("%02X", map[i]);
1362 return &rscreen->screen;