2 * Copyright © 2015 Intel Corporation
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
25 #include "nir/nir_builder.h"
28 VkOffset3D src_offset;
29 VkExtent3D src_extent;
30 VkOffset3D dest_offset;
31 VkExtent3D dest_extent;
35 build_nir_vertex_shader(void)
37 const struct glsl_type *vec4 = glsl_vec4_type();
40 nir_builder_init_simple_shader(&b, NULL, MESA_SHADER_VERTEX, NULL);
41 b.shader->info.name = ralloc_strdup(b.shader, "meta_blit_vs");
43 nir_variable *pos_in = nir_variable_create(b.shader, nir_var_shader_in,
45 pos_in->data.location = VERT_ATTRIB_GENERIC0;
46 nir_variable *pos_out = nir_variable_create(b.shader, nir_var_shader_out,
48 pos_out->data.location = VARYING_SLOT_POS;
49 nir_copy_var(&b, pos_out, pos_in);
51 nir_variable *tex_pos_in = nir_variable_create(b.shader, nir_var_shader_in,
53 tex_pos_in->data.location = VERT_ATTRIB_GENERIC1;
54 nir_variable *tex_pos_out = nir_variable_create(b.shader, nir_var_shader_out,
56 tex_pos_out->data.location = VARYING_SLOT_VAR0;
57 tex_pos_out->data.interpolation = INTERP_QUALIFIER_SMOOTH;
58 nir_copy_var(&b, tex_pos_out, tex_pos_in);
64 build_nir_copy_fragment_shader(enum glsl_sampler_dim tex_dim)
66 const struct glsl_type *vec4 = glsl_vec4_type();
69 nir_builder_init_simple_shader(&b, NULL, MESA_SHADER_FRAGMENT, NULL);
70 b.shader->info.name = ralloc_strdup(b.shader, "meta_blit_fs");
72 nir_variable *tex_pos_in = nir_variable_create(b.shader, nir_var_shader_in,
74 tex_pos_in->data.location = VARYING_SLOT_VAR0;
76 /* Swizzle the array index which comes in as Z coordinate into the right
79 unsigned swz[] = { 0, (tex_dim == GLSL_SAMPLER_DIM_1D ? 2 : 1), 2 };
80 nir_ssa_def *const tex_pos =
81 nir_swizzle(&b, nir_load_var(&b, tex_pos_in), swz,
82 (tex_dim == GLSL_SAMPLER_DIM_1D ? 2 : 3), false);
84 const struct glsl_type *sampler_type =
85 glsl_sampler_type(tex_dim, false, tex_dim != GLSL_SAMPLER_DIM_3D,
86 glsl_get_base_type(vec4));
87 nir_variable *sampler = nir_variable_create(b.shader, nir_var_uniform,
88 sampler_type, "s_tex");
89 sampler->data.descriptor_set = 0;
90 sampler->data.binding = 0;
92 nir_tex_instr *tex = nir_tex_instr_create(b.shader, 1);
93 tex->sampler_dim = tex_dim;
94 tex->op = nir_texop_tex;
95 tex->src[0].src_type = nir_tex_src_coord;
96 tex->src[0].src = nir_src_for_ssa(tex_pos);
97 tex->dest_type = nir_type_float; /* TODO */
98 tex->is_array = glsl_sampler_type_is_array(sampler_type);
99 tex->coord_components = tex_pos->num_components;
100 tex->texture = nir_deref_var_create(tex, sampler);
101 tex->sampler = nir_deref_var_create(tex, sampler);
103 nir_ssa_dest_init(&tex->instr, &tex->dest, 4, "tex");
104 nir_builder_instr_insert(&b, &tex->instr);
106 nir_variable *color_out = nir_variable_create(b.shader, nir_var_shader_out,
108 color_out->data.location = FRAG_RESULT_DATA0;
109 nir_store_var(&b, color_out, &tex->dest.ssa, 4);
115 meta_prepare_blit(struct anv_cmd_buffer *cmd_buffer,
116 struct anv_meta_saved_state *saved_state)
118 anv_meta_save(saved_state, cmd_buffer,
119 (1 << VK_DYNAMIC_STATE_VIEWPORT));
123 anv_meta_begin_blit2d(struct anv_cmd_buffer *cmd_buffer,
124 struct anv_meta_saved_state *save)
126 meta_prepare_blit(cmd_buffer, save);
130 /* Returns the user-provided VkBufferImageCopy::imageOffset in units of
131 * elements rather than texels. One element equals one texel or one block
132 * if Image is uncompressed or compressed, respectively.
134 static struct VkOffset3D
135 meta_region_offset_el(const struct anv_image * image,
136 const struct VkOffset3D * offset)
138 const struct isl_format_layout * isl_layout = image->format->isl_layout;
139 return (VkOffset3D) {
140 .x = offset->x / isl_layout->bw,
141 .y = offset->y / isl_layout->bh,
142 .z = offset->z / isl_layout->bd,
146 /* Returns the user-provided VkBufferImageCopy::imageExtent in units of
147 * elements rather than texels. One element equals one texel or one block
148 * if Image is uncompressed or compressed, respectively.
150 static struct VkExtent3D
151 meta_region_extent_el(const VkFormat format,
152 const struct VkExtent3D * extent)
154 const struct isl_format_layout * isl_layout =
155 anv_format_for_vk_format(format)->isl_layout;
156 return (VkExtent3D) {
157 .width = DIV_ROUND_UP(extent->width , isl_layout->bw),
158 .height = DIV_ROUND_UP(extent->height, isl_layout->bh),
159 .depth = DIV_ROUND_UP(extent->depth , isl_layout->bd),
164 meta_emit_blit(struct anv_cmd_buffer *cmd_buffer,
165 struct anv_image *src_image,
166 struct anv_image_view *src_iview,
167 VkOffset3D src_offset,
168 VkExtent3D src_extent,
169 struct anv_image *dest_image,
170 struct anv_image_view *dest_iview,
171 VkOffset3D dest_offset,
172 VkExtent3D dest_extent,
173 VkFilter blit_filter)
175 struct anv_device *device = cmd_buffer->device;
177 struct blit_vb_data {
182 assert(src_image->samples == dest_image->samples);
184 unsigned vb_size = sizeof(struct anv_vue_header) + 3 * sizeof(*vb_data);
186 struct anv_state vb_state =
187 anv_cmd_buffer_alloc_dynamic_state(cmd_buffer, vb_size, 16);
188 memset(vb_state.map, 0, sizeof(struct anv_vue_header));
189 vb_data = vb_state.map + sizeof(struct anv_vue_header);
191 vb_data[0] = (struct blit_vb_data) {
193 dest_offset.x + dest_extent.width,
194 dest_offset.y + dest_extent.height,
197 (float)(src_offset.x + src_extent.width) / (float)src_iview->extent.width,
198 (float)(src_offset.y + src_extent.height) / (float)src_iview->extent.height,
199 (float)src_offset.z / (float)src_iview->extent.depth,
203 vb_data[1] = (struct blit_vb_data) {
206 dest_offset.y + dest_extent.height,
209 (float)src_offset.x / (float)src_iview->extent.width,
210 (float)(src_offset.y + src_extent.height) / (float)src_iview->extent.height,
211 (float)src_offset.z / (float)src_iview->extent.depth,
215 vb_data[2] = (struct blit_vb_data) {
221 (float)src_offset.x / (float)src_iview->extent.width,
222 (float)src_offset.y / (float)src_iview->extent.height,
223 (float)src_offset.z / (float)src_iview->extent.depth,
227 anv_state_clflush(vb_state);
229 struct anv_buffer vertex_buffer = {
232 .bo = &device->dynamic_state_block_pool.bo,
233 .offset = vb_state.offset,
236 anv_CmdBindVertexBuffers(anv_cmd_buffer_to_handle(cmd_buffer), 0, 2,
238 anv_buffer_to_handle(&vertex_buffer),
239 anv_buffer_to_handle(&vertex_buffer)
243 sizeof(struct anv_vue_header),
247 ANV_CALL(CreateSampler)(anv_device_to_handle(device),
248 &(VkSamplerCreateInfo) {
249 .sType = VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO,
250 .magFilter = blit_filter,
251 .minFilter = blit_filter,
252 }, &cmd_buffer->pool->alloc, &sampler);
254 VkDescriptorPool desc_pool;
255 anv_CreateDescriptorPool(anv_device_to_handle(device),
256 &(const VkDescriptorPoolCreateInfo) {
257 .sType = VK_STRUCTURE_TYPE_DESCRIPTOR_POOL_CREATE_INFO,
262 .pPoolSizes = (VkDescriptorPoolSize[]) {
264 .type = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER,
268 }, &cmd_buffer->pool->alloc, &desc_pool);
271 anv_AllocateDescriptorSets(anv_device_to_handle(device),
272 &(VkDescriptorSetAllocateInfo) {
273 .sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_ALLOCATE_INFO,
274 .descriptorPool = desc_pool,
275 .descriptorSetCount = 1,
276 .pSetLayouts = &device->meta_state.blit.ds_layout
279 anv_UpdateDescriptorSets(anv_device_to_handle(device),
281 (VkWriteDescriptorSet[]) {
283 .sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET,
286 .dstArrayElement = 0,
287 .descriptorCount = 1,
288 .descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER,
289 .pImageInfo = (VkDescriptorImageInfo[]) {
292 .imageView = anv_image_view_to_handle(src_iview),
293 .imageLayout = VK_IMAGE_LAYOUT_GENERAL,
300 anv_CreateFramebuffer(anv_device_to_handle(device),
301 &(VkFramebufferCreateInfo) {
302 .sType = VK_STRUCTURE_TYPE_FRAMEBUFFER_CREATE_INFO,
303 .attachmentCount = 1,
304 .pAttachments = (VkImageView[]) {
305 anv_image_view_to_handle(dest_iview),
307 .width = dest_iview->extent.width,
308 .height = dest_iview->extent.height,
310 }, &cmd_buffer->pool->alloc, &fb);
312 ANV_CALL(CmdBeginRenderPass)(anv_cmd_buffer_to_handle(cmd_buffer),
313 &(VkRenderPassBeginInfo) {
314 .sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO,
315 .renderPass = device->meta_state.blit.render_pass,
318 .offset = { dest_offset.x, dest_offset.y },
319 .extent = { dest_extent.width, dest_extent.height },
321 .clearValueCount = 0,
322 .pClearValues = NULL,
323 }, VK_SUBPASS_CONTENTS_INLINE);
327 switch (src_image->type) {
328 case VK_IMAGE_TYPE_1D:
329 pipeline = device->meta_state.blit.pipeline_1d_src;
331 case VK_IMAGE_TYPE_2D:
332 pipeline = device->meta_state.blit.pipeline_2d_src;
334 case VK_IMAGE_TYPE_3D:
335 pipeline = device->meta_state.blit.pipeline_3d_src;
338 unreachable(!"bad VkImageType");
341 if (cmd_buffer->state.pipeline != anv_pipeline_from_handle(pipeline)) {
342 anv_CmdBindPipeline(anv_cmd_buffer_to_handle(cmd_buffer),
343 VK_PIPELINE_BIND_POINT_GRAPHICS, pipeline);
346 anv_CmdSetViewport(anv_cmd_buffer_to_handle(cmd_buffer), 0, 1,
350 .width = dest_iview->extent.width,
351 .height = dest_iview->extent.height,
356 anv_CmdBindDescriptorSets(anv_cmd_buffer_to_handle(cmd_buffer),
357 VK_PIPELINE_BIND_POINT_GRAPHICS,
358 device->meta_state.blit.pipeline_layout, 0, 1,
361 ANV_CALL(CmdDraw)(anv_cmd_buffer_to_handle(cmd_buffer), 3, 1, 0, 0);
363 ANV_CALL(CmdEndRenderPass)(anv_cmd_buffer_to_handle(cmd_buffer));
365 /* At the point where we emit the draw call, all data from the
366 * descriptor sets, etc. has been used. We are free to delete it.
368 anv_DestroyDescriptorPool(anv_device_to_handle(device),
369 desc_pool, &cmd_buffer->pool->alloc);
370 anv_DestroySampler(anv_device_to_handle(device), sampler,
371 &cmd_buffer->pool->alloc);
372 anv_DestroyFramebuffer(anv_device_to_handle(device), fb,
373 &cmd_buffer->pool->alloc);
377 meta_finish_blit(struct anv_cmd_buffer *cmd_buffer,
378 const struct anv_meta_saved_state *saved_state)
380 anv_meta_restore(saved_state, cmd_buffer);
384 anv_meta_end_blit2d(struct anv_cmd_buffer *cmd_buffer,
385 struct anv_meta_saved_state *save)
387 meta_finish_blit(cmd_buffer, save);
391 vk_format_for_size(int bs)
393 /* The choice of UNORM and UINT formats is very intentional here. Most of
394 * the time, we want to use a UINT format to avoid any rounding error in
395 * the blit. For stencil blits, R8_UINT is required by the hardware.
396 * (It's the only format allowed in conjunction with W-tiling.) Also we
397 * intentionally use the 4-channel formats whenever we can. This is so
398 * that, when we do a RGB <-> RGBX copy, the two formats will line up even
399 * though one of them is 3/4 the size of the other. The choice of UNORM
400 * vs. UINT is also very intentional because Haswell doesn't handle 8 or
401 * 16-bit RGB UINT formats at all so we have to use UNORM there.
402 * Fortunately, the only time we should ever use two different formats in
403 * the table below is for RGB -> RGBA blits and so we will never have any
404 * UNORM/UINT mismatch.
407 case 1: return VK_FORMAT_R8_UINT;
408 case 2: return VK_FORMAT_R8G8_UINT;
409 case 3: return VK_FORMAT_R8G8B8_UNORM;
410 case 4: return VK_FORMAT_R8G8B8A8_UNORM;
411 case 6: return VK_FORMAT_R16G16B16_UNORM;
412 case 8: return VK_FORMAT_R16G16B16A16_UNORM;
413 case 12: return VK_FORMAT_R32G32B32_UINT;
414 case 16: return VK_FORMAT_R32G32B32A32_UINT;
416 unreachable("Invalid format block size");
420 static struct anv_meta_blit2d_surf
421 blit_surf_for_image(const struct anv_image* image,
422 const struct isl_surf *img_isl_surf)
424 return (struct anv_meta_blit2d_surf) {
426 .tiling = img_isl_surf->tiling,
427 .base_offset = image->offset,
428 .bs = isl_format_get_layout(img_isl_surf->format)->bs,
429 .pitch = isl_surf_get_row_pitch(img_isl_surf),
434 anv_meta_blit2d(struct anv_cmd_buffer *cmd_buffer,
435 struct anv_meta_blit2d_surf *src,
436 struct anv_meta_blit2d_surf *dst,
438 struct anv_meta_blit2d_rect *rects)
440 VkDevice vk_device = anv_device_to_handle(cmd_buffer->device);
441 VkFormat src_format = vk_format_for_size(src->bs);
442 VkFormat dst_format = vk_format_for_size(dst->bs);
444 for (unsigned r = 0; r < num_rects; ++r) {
446 /* Create VkImages */
447 VkImageCreateInfo image_info = {
448 .sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO,
449 .imageType = VK_IMAGE_TYPE_2D,
450 .format = 0, /* TEMPLATE */
452 .width = 0, /* TEMPLATE */
453 /* Pad to highest tile height to compensate for a vertical intratile offset */
454 .height = MIN(rects[r].height + 64, 1 << 14),
460 .tiling = 0, /* TEMPLATE */
461 .usage = 0, /* TEMPLATE */
463 struct anv_image_create_info anv_image_info = {
464 .vk_info = &image_info,
465 .isl_tiling_flags = 0, /* TEMPLATE */
468 anv_image_info.isl_tiling_flags = 1 << src->tiling;
469 image_info.tiling = anv_image_info.isl_tiling_flags == ISL_TILING_LINEAR_BIT ?
470 VK_IMAGE_TILING_LINEAR : VK_IMAGE_TILING_OPTIMAL;
471 image_info.usage = VK_IMAGE_USAGE_SAMPLED_BIT;
472 image_info.format = src_format,
473 image_info.extent.width = src->pitch / src->bs;
475 anv_image_create(vk_device, &anv_image_info,
476 &cmd_buffer->pool->alloc, &src_image);
478 anv_image_info.isl_tiling_flags = 1 << dst->tiling;
479 image_info.tiling = anv_image_info.isl_tiling_flags == ISL_TILING_LINEAR_BIT ?
480 VK_IMAGE_TILING_LINEAR : VK_IMAGE_TILING_OPTIMAL;
481 image_info.usage = VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
482 image_info.format = dst_format,
483 image_info.extent.width = dst->pitch / dst->bs;
485 anv_image_create(vk_device, &anv_image_info,
486 &cmd_buffer->pool->alloc, &dst_image);
488 /* We could use a vk call to bind memory, but that would require
489 * creating a dummy memory object etc. so there's really no point.
491 anv_image_from_handle(src_image)->bo = src->bo;
492 anv_image_from_handle(src_image)->offset = src->base_offset;
493 anv_image_from_handle(dst_image)->bo = dst->bo;
494 anv_image_from_handle(dst_image)->offset = dst->base_offset;
496 /* Create VkImageViews */
497 VkImageViewCreateInfo iview_info = {
498 .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
499 .image = 0, /* TEMPLATE */
500 .viewType = VK_IMAGE_VIEW_TYPE_2D,
501 .format = 0, /* TEMPLATE */
502 .subresourceRange = {
503 .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
512 iview_info.image = src_image;
513 iview_info.format = src_format;
514 VkOffset3D src_offset_el = {0};
515 isl_surf_get_image_intratile_offset_el_xy(&cmd_buffer->device->isl_dev,
516 &anv_image_from_handle(src_image)->
521 (uint32_t*)&src_offset_el.x,
522 (uint32_t*)&src_offset_el.y);
524 struct anv_image_view src_iview;
525 anv_image_view_init(&src_iview, cmd_buffer->device,
526 &iview_info, cmd_buffer, img_o, VK_IMAGE_USAGE_SAMPLED_BIT);
528 iview_info.image = dst_image;
529 iview_info.format = dst_format;
530 VkOffset3D dst_offset_el = {0};
531 isl_surf_get_image_intratile_offset_el_xy(&cmd_buffer->device->isl_dev,
532 &anv_image_from_handle(dst_image)->
537 (uint32_t*)&dst_offset_el.x,
538 (uint32_t*)&dst_offset_el.y);
539 struct anv_image_view dst_iview;
540 anv_image_view_init(&dst_iview, cmd_buffer->device,
541 &iview_info, cmd_buffer, img_o, VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT);
544 meta_emit_blit(cmd_buffer,
545 anv_image_from_handle(src_image),
548 (VkExtent3D){rects[r].width, rects[r].height, 1},
549 anv_image_from_handle(dst_image),
552 (VkExtent3D){rects[r].width, rects[r].height, 1},
555 anv_DestroyImage(vk_device, src_image, &cmd_buffer->pool->alloc);
556 anv_DestroyImage(vk_device, dst_image, &cmd_buffer->pool->alloc);
561 do_buffer_copy(struct anv_cmd_buffer *cmd_buffer,
562 struct anv_bo *src, uint64_t src_offset,
563 struct anv_bo *dest, uint64_t dest_offset,
564 int width, int height, int bs)
566 VkDevice vk_device = anv_device_to_handle(cmd_buffer->device);
567 VkFormat copy_format = vk_format_for_size(bs);
569 VkImageCreateInfo image_info = {
570 .sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO,
571 .imageType = VK_IMAGE_TYPE_2D,
572 .format = copy_format,
581 .tiling = VK_IMAGE_TILING_LINEAR,
587 image_info.usage = VK_IMAGE_USAGE_SAMPLED_BIT;
588 anv_CreateImage(vk_device, &image_info,
589 &cmd_buffer->pool->alloc, &src_image);
592 image_info.usage = VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
593 anv_CreateImage(vk_device, &image_info,
594 &cmd_buffer->pool->alloc, &dest_image);
596 /* We could use a vk call to bind memory, but that would require
597 * creating a dummy memory object etc. so there's really no point.
599 anv_image_from_handle(src_image)->bo = src;
600 anv_image_from_handle(src_image)->offset = src_offset;
601 anv_image_from_handle(dest_image)->bo = dest;
602 anv_image_from_handle(dest_image)->offset = dest_offset;
604 VkImageViewCreateInfo iview_info = {
605 .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
606 .image = 0, /* TEMPLATE */
607 .viewType = VK_IMAGE_VIEW_TYPE_2D,
608 .format = copy_format,
609 .subresourceRange = {
610 .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
618 struct anv_image_view src_iview;
619 iview_info.image = src_image;
620 anv_image_view_init(&src_iview, cmd_buffer->device,
621 &iview_info, cmd_buffer, 0, VK_IMAGE_USAGE_SAMPLED_BIT);
623 struct anv_image_view dest_iview;
624 iview_info.image = dest_image;
625 anv_image_view_init(&dest_iview, cmd_buffer->device,
626 &iview_info, cmd_buffer, 0, VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT);
628 meta_emit_blit(cmd_buffer,
629 anv_image_from_handle(src_image),
631 (VkOffset3D) { 0, 0, 0 },
632 (VkExtent3D) { width, height, 1 },
633 anv_image_from_handle(dest_image),
635 (VkOffset3D) { 0, 0, 0 },
636 (VkExtent3D) { width, height, 1 },
639 anv_DestroyImage(vk_device, src_image, &cmd_buffer->pool->alloc);
640 anv_DestroyImage(vk_device, dest_image, &cmd_buffer->pool->alloc);
643 void anv_CmdCopyBuffer(
644 VkCommandBuffer commandBuffer,
647 uint32_t regionCount,
648 const VkBufferCopy* pRegions)
650 ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
651 ANV_FROM_HANDLE(anv_buffer, src_buffer, srcBuffer);
652 ANV_FROM_HANDLE(anv_buffer, dest_buffer, destBuffer);
654 struct anv_meta_saved_state saved_state;
656 meta_prepare_blit(cmd_buffer, &saved_state);
658 for (unsigned r = 0; r < regionCount; r++) {
659 uint64_t src_offset = src_buffer->offset + pRegions[r].srcOffset;
660 uint64_t dest_offset = dest_buffer->offset + pRegions[r].dstOffset;
661 uint64_t copy_size = pRegions[r].size;
663 /* First, we compute the biggest format that can be used with the
664 * given offsets and size.
668 int fs = ffs(src_offset) - 1;
670 bs = MIN2(bs, 1 << fs);
671 assert(src_offset % bs == 0);
673 fs = ffs(dest_offset) - 1;
675 bs = MIN2(bs, 1 << fs);
676 assert(dest_offset % bs == 0);
678 fs = ffs(pRegions[r].size) - 1;
680 bs = MIN2(bs, 1 << fs);
681 assert(pRegions[r].size % bs == 0);
683 /* This is maximum possible width/height our HW can handle */
684 uint64_t max_surface_dim = 1 << 14;
686 /* First, we make a bunch of max-sized copies */
687 uint64_t max_copy_size = max_surface_dim * max_surface_dim * bs;
688 while (copy_size >= max_copy_size) {
689 do_buffer_copy(cmd_buffer, src_buffer->bo, src_offset,
690 dest_buffer->bo, dest_offset,
691 max_surface_dim, max_surface_dim, bs);
692 copy_size -= max_copy_size;
693 src_offset += max_copy_size;
694 dest_offset += max_copy_size;
697 uint64_t height = copy_size / (max_surface_dim * bs);
698 assert(height < max_surface_dim);
700 uint64_t rect_copy_size = height * max_surface_dim * bs;
701 do_buffer_copy(cmd_buffer, src_buffer->bo, src_offset,
702 dest_buffer->bo, dest_offset,
703 max_surface_dim, height, bs);
704 copy_size -= rect_copy_size;
705 src_offset += rect_copy_size;
706 dest_offset += rect_copy_size;
709 if (copy_size != 0) {
710 do_buffer_copy(cmd_buffer, src_buffer->bo, src_offset,
711 dest_buffer->bo, dest_offset,
712 copy_size / bs, 1, bs);
716 meta_finish_blit(cmd_buffer, &saved_state);
719 void anv_CmdUpdateBuffer(
720 VkCommandBuffer commandBuffer,
722 VkDeviceSize dstOffset,
723 VkDeviceSize dataSize,
724 const uint32_t* pData)
726 ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
727 ANV_FROM_HANDLE(anv_buffer, dst_buffer, dstBuffer);
728 struct anv_meta_saved_state saved_state;
730 meta_prepare_blit(cmd_buffer, &saved_state);
732 /* We can't quite grab a full block because the state stream needs a
733 * little data at the top to build its linked list.
735 const uint32_t max_update_size =
736 cmd_buffer->device->dynamic_state_block_pool.block_size - 64;
738 assert(max_update_size < (1 << 14) * 4);
741 const uint32_t copy_size = MIN2(dataSize, max_update_size);
743 struct anv_state tmp_data =
744 anv_cmd_buffer_alloc_dynamic_state(cmd_buffer, copy_size, 64);
746 memcpy(tmp_data.map, pData, copy_size);
749 if ((copy_size & 15) == 0 && (dstOffset & 15) == 0) {
751 } else if ((copy_size & 7) == 0 && (dstOffset & 7) == 0) {
754 assert((copy_size & 3) == 0 && (dstOffset & 3) == 0);
758 do_buffer_copy(cmd_buffer,
759 &cmd_buffer->device->dynamic_state_block_pool.bo,
761 dst_buffer->bo, dst_buffer->offset + dstOffset,
762 copy_size / bs, 1, bs);
764 dataSize -= copy_size;
765 dstOffset += copy_size;
766 pData = (void *)pData + copy_size;
769 meta_finish_blit(cmd_buffer, &saved_state);
773 choose_iview_format(struct anv_image *image, VkImageAspectFlagBits aspect)
775 assert(__builtin_popcount(aspect) == 1);
777 struct isl_surf *surf =
778 &anv_image_get_surface_for_aspect_mask(image, aspect)->isl;
780 /* vkCmdCopyImage behaves like memcpy. Therefore we choose identical UINT
781 * formats for the source and destination image views.
783 * From the Vulkan spec (2015-12-30):
785 * vkCmdCopyImage performs image copies in a similar manner to a host
786 * memcpy. It does not perform general-purpose conversions such as
787 * scaling, resizing, blending, color-space conversion, or format
788 * conversions. Rather, it simply copies raw image data. vkCmdCopyImage
789 * can copy between images with different formats, provided the formats
790 * are compatible as defined below.
792 * [The spec later defines compatibility as having the same number of
795 return vk_format_for_size(isl_format_layouts[surf->format].bs);
799 choose_buffer_format(VkFormat format, VkImageAspectFlagBits aspect)
801 assert(__builtin_popcount(aspect) == 1);
803 /* vkCmdCopy* commands behave like memcpy. Therefore we choose
804 * compatable UINT formats for the source and destination image views.
806 * For the buffer, we go back to the original image format and get a
807 * the format as if it were linear. This way, for RGB formats, we get
808 * an RGB format here even if the tiled image is RGBA. XXX: This doesn't
809 * work if the buffer is the destination.
811 enum isl_format linear_format = anv_get_isl_format(format, aspect,
812 VK_IMAGE_TILING_LINEAR,
815 return vk_format_for_size(isl_format_layouts[linear_format].bs);
818 void anv_CmdCopyImage(
819 VkCommandBuffer commandBuffer,
821 VkImageLayout srcImageLayout,
823 VkImageLayout destImageLayout,
824 uint32_t regionCount,
825 const VkImageCopy* pRegions)
827 ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
828 ANV_FROM_HANDLE(anv_image, src_image, srcImage);
829 ANV_FROM_HANDLE(anv_image, dest_image, destImage);
830 struct anv_meta_saved_state saved_state;
832 /* From the Vulkan 1.0 spec:
834 * vkCmdCopyImage can be used to copy image data between multisample
835 * images, but both images must have the same number of samples.
837 assert(src_image->samples == dest_image->samples);
839 meta_prepare_blit(cmd_buffer, &saved_state);
841 for (unsigned r = 0; r < regionCount; r++) {
842 assert(pRegions[r].srcSubresource.aspectMask ==
843 pRegions[r].dstSubresource.aspectMask);
845 VkImageAspectFlags aspect = pRegions[r].srcSubresource.aspectMask;
847 VkFormat src_format = choose_iview_format(src_image, aspect);
848 VkFormat dst_format = choose_iview_format(dest_image, aspect);
850 struct anv_image_view src_iview;
851 anv_image_view_init(&src_iview, cmd_buffer->device,
852 &(VkImageViewCreateInfo) {
853 .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
855 .viewType = anv_meta_get_view_type(src_image),
856 .format = src_format,
857 .subresourceRange = {
858 .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
859 .baseMipLevel = pRegions[r].srcSubresource.mipLevel,
861 .baseArrayLayer = pRegions[r].srcSubresource.baseArrayLayer,
862 .layerCount = pRegions[r].dstSubresource.layerCount,
865 cmd_buffer, 0, VK_IMAGE_USAGE_SAMPLED_BIT);
867 const uint32_t dest_base_array_slice =
868 anv_meta_get_iview_layer(dest_image, &pRegions[r].dstSubresource,
869 &pRegions[r].dstOffset);
872 unsigned num_slices_3d = pRegions[r].extent.depth;
873 unsigned num_slices_array = pRegions[r].dstSubresource.layerCount;
874 unsigned slice_3d = 0;
875 unsigned slice_array = 0;
876 while (slice_3d < num_slices_3d && slice_array < num_slices_array) {
877 VkOffset3D src_offset = pRegions[r].srcOffset;
878 src_offset.z += slice_3d + slice_array;
883 if (isl_format_is_compressed(dest_image->format->isl_format))
884 isl_surf_get_image_intratile_offset_el(&cmd_buffer->device->isl_dev,
885 &dest_image->color_surface.isl,
886 pRegions[r].dstSubresource.mipLevel,
887 pRegions[r].dstSubresource.baseArrayLayer + slice_array,
888 pRegions[r].dstOffset.z + slice_3d,
889 &img_o, &img_x, &img_y);
891 VkOffset3D dest_offset_el = meta_region_offset_el(dest_image, &pRegions[r].dstOffset);
892 dest_offset_el.x += img_x;
893 dest_offset_el.y += img_y;
894 dest_offset_el.z = 0;
896 struct anv_image_view dest_iview;
897 anv_image_view_init(&dest_iview, cmd_buffer->device,
898 &(VkImageViewCreateInfo) {
899 .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
901 .viewType = anv_meta_get_view_type(dest_image),
902 .format = dst_format,
903 .subresourceRange = {
904 .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
905 .baseMipLevel = pRegions[r].dstSubresource.mipLevel,
907 .baseArrayLayer = dest_base_array_slice +
908 slice_array + slice_3d,
912 cmd_buffer, img_o, VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT);
914 const VkExtent3D img_extent_el = meta_region_extent_el(dest_image->vk_format,
915 &pRegions[r].extent);
917 meta_emit_blit(cmd_buffer,
918 src_image, &src_iview,
921 dest_image, &dest_iview,
926 if (dest_image->type == VK_IMAGE_TYPE_3D)
933 meta_finish_blit(cmd_buffer, &saved_state);
936 void anv_CmdBlitImage(
937 VkCommandBuffer commandBuffer,
939 VkImageLayout srcImageLayout,
941 VkImageLayout destImageLayout,
942 uint32_t regionCount,
943 const VkImageBlit* pRegions,
947 ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
948 ANV_FROM_HANDLE(anv_image, src_image, srcImage);
949 ANV_FROM_HANDLE(anv_image, dest_image, destImage);
950 struct anv_meta_saved_state saved_state;
952 /* From the Vulkan 1.0 spec:
954 * vkCmdBlitImage must not be used for multisampled source or
955 * destination images. Use vkCmdResolveImage for this purpose.
957 assert(src_image->samples == 1);
958 assert(dest_image->samples == 1);
960 anv_finishme("respect VkFilter");
962 meta_prepare_blit(cmd_buffer, &saved_state);
964 for (unsigned r = 0; r < regionCount; r++) {
965 struct anv_image_view src_iview;
966 anv_image_view_init(&src_iview, cmd_buffer->device,
967 &(VkImageViewCreateInfo) {
968 .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
970 .viewType = anv_meta_get_view_type(src_image),
971 .format = src_image->vk_format,
972 .subresourceRange = {
973 .aspectMask = pRegions[r].srcSubresource.aspectMask,
974 .baseMipLevel = pRegions[r].srcSubresource.mipLevel,
976 .baseArrayLayer = pRegions[r].srcSubresource.baseArrayLayer,
980 cmd_buffer, 0, VK_IMAGE_USAGE_SAMPLED_BIT);
982 const VkOffset3D dest_offset = {
983 .x = pRegions[r].dstOffsets[0].x,
984 .y = pRegions[r].dstOffsets[0].y,
988 if (pRegions[r].dstOffsets[1].x < pRegions[r].dstOffsets[0].x ||
989 pRegions[r].dstOffsets[1].y < pRegions[r].dstOffsets[0].y ||
990 pRegions[r].srcOffsets[1].x < pRegions[r].srcOffsets[0].x ||
991 pRegions[r].srcOffsets[1].y < pRegions[r].srcOffsets[0].y)
992 anv_finishme("FINISHME: Allow flipping in blits");
994 const VkExtent3D dest_extent = {
995 .width = pRegions[r].dstOffsets[1].x - pRegions[r].dstOffsets[0].x,
996 .height = pRegions[r].dstOffsets[1].y - pRegions[r].dstOffsets[0].y,
999 const VkExtent3D src_extent = {
1000 .width = pRegions[r].srcOffsets[1].x - pRegions[r].srcOffsets[0].x,
1001 .height = pRegions[r].srcOffsets[1].y - pRegions[r].srcOffsets[0].y,
1004 const uint32_t dest_array_slice =
1005 anv_meta_get_iview_layer(dest_image, &pRegions[r].dstSubresource,
1006 &pRegions[r].dstOffsets[0]);
1008 if (pRegions[r].srcSubresource.layerCount > 1)
1009 anv_finishme("FINISHME: copy multiple array layers");
1011 if (pRegions[r].srcOffsets[0].z + 1 != pRegions[r].srcOffsets[1].z ||
1012 pRegions[r].dstOffsets[0].z + 1 != pRegions[r].dstOffsets[1].z)
1013 anv_finishme("FINISHME: copy multiple depth layers");
1015 struct anv_image_view dest_iview;
1016 anv_image_view_init(&dest_iview, cmd_buffer->device,
1017 &(VkImageViewCreateInfo) {
1018 .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
1020 .viewType = anv_meta_get_view_type(dest_image),
1021 .format = dest_image->vk_format,
1022 .subresourceRange = {
1023 .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
1024 .baseMipLevel = pRegions[r].dstSubresource.mipLevel,
1026 .baseArrayLayer = dest_array_slice,
1030 cmd_buffer, 0, VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT);
1032 meta_emit_blit(cmd_buffer,
1033 src_image, &src_iview,
1034 pRegions[r].srcOffsets[0], src_extent,
1035 dest_image, &dest_iview,
1036 dest_offset, dest_extent,
1040 meta_finish_blit(cmd_buffer, &saved_state);
1043 static struct anv_image *
1044 make_image_for_buffer(VkDevice vk_device, VkBuffer vk_buffer, VkFormat format,
1045 VkImageUsageFlags usage,
1046 VkImageType image_type,
1047 const VkAllocationCallbacks *alloc,
1048 const VkBufferImageCopy *copy)
1050 ANV_FROM_HANDLE(anv_buffer, buffer, vk_buffer);
1052 VkExtent3D extent = copy->imageExtent;
1053 if (copy->bufferRowLength)
1054 extent.width = copy->bufferRowLength;
1055 if (copy->bufferImageHeight)
1056 extent.height = copy->bufferImageHeight;
1058 extent = meta_region_extent_el(format, &extent);
1060 VkImageAspectFlags aspect = copy->imageSubresource.aspectMask;
1061 VkFormat buffer_format = choose_buffer_format(format, aspect);
1064 VkResult result = anv_CreateImage(vk_device,
1065 &(VkImageCreateInfo) {
1066 .sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO,
1067 .imageType = VK_IMAGE_TYPE_2D,
1068 .format = buffer_format,
1073 .tiling = VK_IMAGE_TILING_LINEAR,
1076 }, alloc, &vk_image);
1077 assert(result == VK_SUCCESS);
1079 ANV_FROM_HANDLE(anv_image, image, vk_image);
1081 /* We could use a vk call to bind memory, but that would require
1082 * creating a dummy memory object etc. so there's really no point.
1084 image->bo = buffer->bo;
1085 image->offset = buffer->offset + copy->bufferOffset;
1090 void anv_CmdCopyBufferToImage(
1091 VkCommandBuffer commandBuffer,
1094 VkImageLayout destImageLayout,
1095 uint32_t regionCount,
1096 const VkBufferImageCopy* pRegions)
1098 ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
1099 ANV_FROM_HANDLE(anv_image, dest_image, destImage);
1100 VkDevice vk_device = anv_device_to_handle(cmd_buffer->device);
1101 struct anv_meta_saved_state saved_state;
1103 /* The Vulkan 1.0 spec says "dstImage must have a sample count equal to
1104 * VK_SAMPLE_COUNT_1_BIT."
1106 assert(dest_image->samples == 1);
1108 meta_prepare_blit(cmd_buffer, &saved_state);
1110 for (unsigned r = 0; r < regionCount; r++) {
1111 VkImageAspectFlags aspect = pRegions[r].imageSubresource.aspectMask;
1113 VkFormat image_format = choose_iview_format(dest_image, aspect);
1115 struct anv_image *src_image =
1116 make_image_for_buffer(vk_device, srcBuffer, dest_image->vk_format,
1117 VK_IMAGE_USAGE_SAMPLED_BIT,
1118 dest_image->type, &cmd_buffer->pool->alloc,
1121 const uint32_t dest_base_array_slice =
1122 anv_meta_get_iview_layer(dest_image, &pRegions[r].imageSubresource,
1123 &pRegions[r].imageOffset);
1125 unsigned num_slices_3d = pRegions[r].imageExtent.depth;
1126 unsigned num_slices_array = pRegions[r].imageSubresource.layerCount;
1127 unsigned slice_3d = 0;
1128 unsigned slice_array = 0;
1129 while (slice_3d < num_slices_3d && slice_array < num_slices_array) {
1130 struct anv_image_view src_iview;
1131 anv_image_view_init(&src_iview, cmd_buffer->device,
1132 &(VkImageViewCreateInfo) {
1133 .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
1134 .image = anv_image_to_handle(src_image),
1135 .viewType = VK_IMAGE_VIEW_TYPE_2D,
1136 .format = src_image->vk_format,
1137 .subresourceRange = {
1138 .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
1141 .baseArrayLayer = 0,
1145 cmd_buffer, 0, VK_IMAGE_USAGE_SAMPLED_BIT);
1150 if (isl_format_is_compressed(dest_image->format->isl_format))
1151 isl_surf_get_image_intratile_offset_el(&cmd_buffer->device->isl_dev,
1152 &dest_image->color_surface.isl,
1153 pRegions[r].imageSubresource.mipLevel,
1154 pRegions[r].imageSubresource.baseArrayLayer + slice_array,
1155 pRegions[r].imageOffset.z + slice_3d,
1156 &img_o, &img_x, &img_y);
1158 VkOffset3D dest_offset_el = meta_region_offset_el(dest_image, & pRegions[r].imageOffset);
1159 dest_offset_el.x += img_x;
1160 dest_offset_el.y += img_y;
1161 dest_offset_el.z = 0;
1163 struct anv_image_view dest_iview;
1164 anv_image_view_init(&dest_iview, cmd_buffer->device,
1165 &(VkImageViewCreateInfo) {
1166 .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
1167 .image = anv_image_to_handle(dest_image),
1168 .viewType = anv_meta_get_view_type(dest_image),
1169 .format = image_format,
1170 .subresourceRange = {
1171 .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
1172 .baseMipLevel = pRegions[r].imageSubresource.mipLevel,
1174 .baseArrayLayer = dest_base_array_slice +
1175 slice_array + slice_3d,
1179 cmd_buffer, img_o, VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT);
1181 const VkExtent3D img_extent_el = meta_region_extent_el(dest_image->vk_format,
1182 &pRegions[r].imageExtent);
1184 meta_emit_blit(cmd_buffer,
1187 (VkOffset3D){0, 0, 0},
1195 /* Once we've done the blit, all of the actual information about
1196 * the image is embedded in the command buffer so we can just
1197 * increment the offset directly in the image effectively
1198 * re-binding it to different backing memory.
1200 src_image->offset += src_image->extent.width *
1201 src_image->extent.height *
1202 src_image->format->isl_layout->bs;
1204 if (dest_image->type == VK_IMAGE_TYPE_3D)
1210 anv_DestroyImage(vk_device, anv_image_to_handle(src_image),
1211 &cmd_buffer->pool->alloc);
1214 meta_finish_blit(cmd_buffer, &saved_state);
1217 void anv_CmdCopyImageToBuffer(
1218 VkCommandBuffer commandBuffer,
1220 VkImageLayout srcImageLayout,
1221 VkBuffer destBuffer,
1222 uint32_t regionCount,
1223 const VkBufferImageCopy* pRegions)
1225 ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
1226 ANV_FROM_HANDLE(anv_image, src_image, srcImage);
1227 VkDevice vk_device = anv_device_to_handle(cmd_buffer->device);
1228 struct anv_meta_saved_state saved_state;
1231 /* The Vulkan 1.0 spec says "srcImage must have a sample count equal to
1232 * VK_SAMPLE_COUNT_1_BIT."
1234 assert(src_image->samples == 1);
1236 meta_prepare_blit(cmd_buffer, &saved_state);
1238 for (unsigned r = 0; r < regionCount; r++) {
1239 VkImageAspectFlags aspect = pRegions[r].imageSubresource.aspectMask;
1241 VkFormat image_format = choose_iview_format(src_image, aspect);
1243 struct anv_image_view src_iview;
1244 anv_image_view_init(&src_iview, cmd_buffer->device,
1245 &(VkImageViewCreateInfo) {
1246 .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
1248 .viewType = anv_meta_get_view_type(src_image),
1249 .format = image_format,
1250 .subresourceRange = {
1251 .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
1252 .baseMipLevel = pRegions[r].imageSubresource.mipLevel,
1254 .baseArrayLayer = pRegions[r].imageSubresource.baseArrayLayer,
1255 .layerCount = pRegions[r].imageSubresource.layerCount,
1258 cmd_buffer, 0, VK_IMAGE_USAGE_SAMPLED_BIT);
1260 struct anv_image *dest_image =
1261 make_image_for_buffer(vk_device, destBuffer, src_image->vk_format,
1262 VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT,
1263 src_image->type, &cmd_buffer->pool->alloc,
1266 unsigned num_slices;
1267 if (src_image->type == VK_IMAGE_TYPE_3D) {
1268 assert(pRegions[r].imageSubresource.layerCount == 1);
1269 num_slices = pRegions[r].imageExtent.depth;
1271 assert(pRegions[r].imageExtent.depth == 1);
1272 num_slices = pRegions[r].imageSubresource.layerCount;
1275 for (unsigned slice = 0; slice < num_slices; slice++) {
1276 VkOffset3D src_offset = pRegions[r].imageOffset;
1277 src_offset.z += slice;
1279 struct anv_image_view dest_iview;
1280 anv_image_view_init(&dest_iview, cmd_buffer->device,
1281 &(VkImageViewCreateInfo) {
1282 .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
1283 .image = anv_image_to_handle(dest_image),
1284 .viewType = VK_IMAGE_VIEW_TYPE_2D,
1285 .format = dest_image->vk_format,
1286 .subresourceRange = {
1287 .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
1290 .baseArrayLayer = 0,
1294 cmd_buffer, 0, VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT);
1296 meta_emit_blit(cmd_buffer,
1297 anv_image_from_handle(srcImage),
1300 pRegions[r].imageExtent,
1303 (VkOffset3D) { 0, 0, 0 },
1304 pRegions[r].imageExtent,
1307 /* Once we've done the blit, all of the actual information about
1308 * the image is embedded in the command buffer so we can just
1309 * increment the offset directly in the image effectively
1310 * re-binding it to different backing memory.
1312 dest_image->offset += dest_image->extent.width *
1313 dest_image->extent.height *
1314 src_image->format->isl_layout->bs;
1317 anv_DestroyImage(vk_device, anv_image_to_handle(dest_image),
1318 &cmd_buffer->pool->alloc);
1321 meta_finish_blit(cmd_buffer, &saved_state);
1325 anv_device_finish_meta_blit_state(struct anv_device *device)
1327 anv_DestroyRenderPass(anv_device_to_handle(device),
1328 device->meta_state.blit.render_pass,
1329 &device->meta_state.alloc);
1330 anv_DestroyPipeline(anv_device_to_handle(device),
1331 device->meta_state.blit.pipeline_1d_src,
1332 &device->meta_state.alloc);
1333 anv_DestroyPipeline(anv_device_to_handle(device),
1334 device->meta_state.blit.pipeline_2d_src,
1335 &device->meta_state.alloc);
1336 anv_DestroyPipeline(anv_device_to_handle(device),
1337 device->meta_state.blit.pipeline_3d_src,
1338 &device->meta_state.alloc);
1339 anv_DestroyPipelineLayout(anv_device_to_handle(device),
1340 device->meta_state.blit.pipeline_layout,
1341 &device->meta_state.alloc);
1342 anv_DestroyDescriptorSetLayout(anv_device_to_handle(device),
1343 device->meta_state.blit.ds_layout,
1344 &device->meta_state.alloc);
1348 anv_device_init_meta_blit_state(struct anv_device *device)
1352 result = anv_CreateRenderPass(anv_device_to_handle(device),
1353 &(VkRenderPassCreateInfo) {
1354 .sType = VK_STRUCTURE_TYPE_RENDER_PASS_CREATE_INFO,
1355 .attachmentCount = 1,
1356 .pAttachments = &(VkAttachmentDescription) {
1357 .format = VK_FORMAT_UNDEFINED, /* Our shaders don't care */
1358 .loadOp = VK_ATTACHMENT_LOAD_OP_LOAD,
1359 .storeOp = VK_ATTACHMENT_STORE_OP_STORE,
1360 .initialLayout = VK_IMAGE_LAYOUT_GENERAL,
1361 .finalLayout = VK_IMAGE_LAYOUT_GENERAL,
1364 .pSubpasses = &(VkSubpassDescription) {
1365 .pipelineBindPoint = VK_PIPELINE_BIND_POINT_GRAPHICS,
1366 .inputAttachmentCount = 0,
1367 .colorAttachmentCount = 1,
1368 .pColorAttachments = &(VkAttachmentReference) {
1370 .layout = VK_IMAGE_LAYOUT_GENERAL,
1372 .pResolveAttachments = NULL,
1373 .pDepthStencilAttachment = &(VkAttachmentReference) {
1374 .attachment = VK_ATTACHMENT_UNUSED,
1375 .layout = VK_IMAGE_LAYOUT_GENERAL,
1377 .preserveAttachmentCount = 1,
1378 .pPreserveAttachments = (uint32_t[]) { 0 },
1380 .dependencyCount = 0,
1381 }, &device->meta_state.alloc, &device->meta_state.blit.render_pass);
1382 if (result != VK_SUCCESS)
1385 /* We don't use a vertex shader for blitting, but instead build and pass
1386 * the VUEs directly to the rasterization backend. However, we do need
1387 * to provide GLSL source for the vertex shader so that the compiler
1388 * does not dead-code our inputs.
1390 struct anv_shader_module vs = {
1391 .nir = build_nir_vertex_shader(),
1394 struct anv_shader_module fs_1d = {
1395 .nir = build_nir_copy_fragment_shader(GLSL_SAMPLER_DIM_1D),
1398 struct anv_shader_module fs_2d = {
1399 .nir = build_nir_copy_fragment_shader(GLSL_SAMPLER_DIM_2D),
1402 struct anv_shader_module fs_3d = {
1403 .nir = build_nir_copy_fragment_shader(GLSL_SAMPLER_DIM_3D),
1406 VkPipelineVertexInputStateCreateInfo vi_create_info = {
1407 .sType = VK_STRUCTURE_TYPE_PIPELINE_VERTEX_INPUT_STATE_CREATE_INFO,
1408 .vertexBindingDescriptionCount = 2,
1409 .pVertexBindingDescriptions = (VkVertexInputBindingDescription[]) {
1413 .inputRate = VK_VERTEX_INPUT_RATE_VERTEX
1417 .stride = 5 * sizeof(float),
1418 .inputRate = VK_VERTEX_INPUT_RATE_VERTEX
1421 .vertexAttributeDescriptionCount = 3,
1422 .pVertexAttributeDescriptions = (VkVertexInputAttributeDescription[]) {
1427 .format = VK_FORMAT_R32G32B32A32_UINT,
1434 .format = VK_FORMAT_R32G32_SFLOAT,
1438 /* Texture Coordinate */
1441 .format = VK_FORMAT_R32G32B32_SFLOAT,
1447 VkDescriptorSetLayoutCreateInfo ds_layout_info = {
1448 .sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO,
1450 .pBindings = (VkDescriptorSetLayoutBinding[]) {
1453 .descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER,
1454 .descriptorCount = 1,
1455 .stageFlags = VK_SHADER_STAGE_FRAGMENT_BIT,
1456 .pImmutableSamplers = NULL
1460 result = anv_CreateDescriptorSetLayout(anv_device_to_handle(device),
1462 &device->meta_state.alloc,
1463 &device->meta_state.blit.ds_layout);
1464 if (result != VK_SUCCESS)
1465 goto fail_render_pass;
1467 result = anv_CreatePipelineLayout(anv_device_to_handle(device),
1468 &(VkPipelineLayoutCreateInfo) {
1469 .sType = VK_STRUCTURE_TYPE_PIPELINE_LAYOUT_CREATE_INFO,
1470 .setLayoutCount = 1,
1471 .pSetLayouts = &device->meta_state.blit.ds_layout,
1473 &device->meta_state.alloc, &device->meta_state.blit.pipeline_layout);
1474 if (result != VK_SUCCESS)
1475 goto fail_descriptor_set_layout;
1477 VkPipelineShaderStageCreateInfo pipeline_shader_stages[] = {
1479 .sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO,
1480 .stage = VK_SHADER_STAGE_VERTEX_BIT,
1481 .module = anv_shader_module_to_handle(&vs),
1483 .pSpecializationInfo = NULL
1485 .sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO,
1486 .stage = VK_SHADER_STAGE_FRAGMENT_BIT,
1487 .module = VK_NULL_HANDLE, /* TEMPLATE VALUE! FILL ME IN! */
1489 .pSpecializationInfo = NULL
1493 const VkGraphicsPipelineCreateInfo vk_pipeline_info = {
1494 .sType = VK_STRUCTURE_TYPE_GRAPHICS_PIPELINE_CREATE_INFO,
1495 .stageCount = ARRAY_SIZE(pipeline_shader_stages),
1496 .pStages = pipeline_shader_stages,
1497 .pVertexInputState = &vi_create_info,
1498 .pInputAssemblyState = &(VkPipelineInputAssemblyStateCreateInfo) {
1499 .sType = VK_STRUCTURE_TYPE_PIPELINE_INPUT_ASSEMBLY_STATE_CREATE_INFO,
1500 .topology = VK_PRIMITIVE_TOPOLOGY_TRIANGLE_STRIP,
1501 .primitiveRestartEnable = false,
1503 .pViewportState = &(VkPipelineViewportStateCreateInfo) {
1504 .sType = VK_STRUCTURE_TYPE_PIPELINE_VIEWPORT_STATE_CREATE_INFO,
1508 .pRasterizationState = &(VkPipelineRasterizationStateCreateInfo) {
1509 .sType = VK_STRUCTURE_TYPE_PIPELINE_RASTERIZATION_STATE_CREATE_INFO,
1510 .rasterizerDiscardEnable = false,
1511 .polygonMode = VK_POLYGON_MODE_FILL,
1512 .cullMode = VK_CULL_MODE_NONE,
1513 .frontFace = VK_FRONT_FACE_COUNTER_CLOCKWISE
1515 .pMultisampleState = &(VkPipelineMultisampleStateCreateInfo) {
1516 .sType = VK_STRUCTURE_TYPE_PIPELINE_MULTISAMPLE_STATE_CREATE_INFO,
1517 .rasterizationSamples = 1,
1518 .sampleShadingEnable = false,
1519 .pSampleMask = (VkSampleMask[]) { UINT32_MAX },
1521 .pColorBlendState = &(VkPipelineColorBlendStateCreateInfo) {
1522 .sType = VK_STRUCTURE_TYPE_PIPELINE_COLOR_BLEND_STATE_CREATE_INFO,
1523 .attachmentCount = 1,
1524 .pAttachments = (VkPipelineColorBlendAttachmentState []) {
1526 VK_COLOR_COMPONENT_A_BIT |
1527 VK_COLOR_COMPONENT_R_BIT |
1528 VK_COLOR_COMPONENT_G_BIT |
1529 VK_COLOR_COMPONENT_B_BIT },
1532 .pDynamicState = &(VkPipelineDynamicStateCreateInfo) {
1533 .sType = VK_STRUCTURE_TYPE_PIPELINE_DYNAMIC_STATE_CREATE_INFO,
1534 .dynamicStateCount = 9,
1535 .pDynamicStates = (VkDynamicState[]) {
1536 VK_DYNAMIC_STATE_VIEWPORT,
1537 VK_DYNAMIC_STATE_SCISSOR,
1538 VK_DYNAMIC_STATE_LINE_WIDTH,
1539 VK_DYNAMIC_STATE_DEPTH_BIAS,
1540 VK_DYNAMIC_STATE_BLEND_CONSTANTS,
1541 VK_DYNAMIC_STATE_DEPTH_BOUNDS,
1542 VK_DYNAMIC_STATE_STENCIL_COMPARE_MASK,
1543 VK_DYNAMIC_STATE_STENCIL_WRITE_MASK,
1544 VK_DYNAMIC_STATE_STENCIL_REFERENCE,
1548 .layout = device->meta_state.blit.pipeline_layout,
1549 .renderPass = device->meta_state.blit.render_pass,
1553 const struct anv_graphics_pipeline_create_info anv_pipeline_info = {
1554 .color_attachment_count = -1,
1555 .use_repclear = false,
1556 .disable_viewport = true,
1557 .disable_scissor = true,
1559 .use_rectlist = true
1562 pipeline_shader_stages[1].module = anv_shader_module_to_handle(&fs_1d);
1563 result = anv_graphics_pipeline_create(anv_device_to_handle(device),
1565 &vk_pipeline_info, &anv_pipeline_info,
1566 &device->meta_state.alloc, &device->meta_state.blit.pipeline_1d_src);
1567 if (result != VK_SUCCESS)
1568 goto fail_pipeline_layout;
1570 pipeline_shader_stages[1].module = anv_shader_module_to_handle(&fs_2d);
1571 result = anv_graphics_pipeline_create(anv_device_to_handle(device),
1573 &vk_pipeline_info, &anv_pipeline_info,
1574 &device->meta_state.alloc, &device->meta_state.blit.pipeline_2d_src);
1575 if (result != VK_SUCCESS)
1576 goto fail_pipeline_1d;
1578 pipeline_shader_stages[1].module = anv_shader_module_to_handle(&fs_3d);
1579 result = anv_graphics_pipeline_create(anv_device_to_handle(device),
1581 &vk_pipeline_info, &anv_pipeline_info,
1582 &device->meta_state.alloc, &device->meta_state.blit.pipeline_3d_src);
1583 if (result != VK_SUCCESS)
1584 goto fail_pipeline_2d;
1586 ralloc_free(vs.nir);
1587 ralloc_free(fs_1d.nir);
1588 ralloc_free(fs_2d.nir);
1589 ralloc_free(fs_3d.nir);
1594 anv_DestroyPipeline(anv_device_to_handle(device),
1595 device->meta_state.blit.pipeline_2d_src,
1596 &device->meta_state.alloc);
1599 anv_DestroyPipeline(anv_device_to_handle(device),
1600 device->meta_state.blit.pipeline_1d_src,
1601 &device->meta_state.alloc);
1603 fail_pipeline_layout:
1604 anv_DestroyPipelineLayout(anv_device_to_handle(device),
1605 device->meta_state.blit.pipeline_layout,
1606 &device->meta_state.alloc);
1607 fail_descriptor_set_layout:
1608 anv_DestroyDescriptorSetLayout(anv_device_to_handle(device),
1609 device->meta_state.blit.ds_layout,
1610 &device->meta_state.alloc);
1612 anv_DestroyRenderPass(anv_device_to_handle(device),
1613 device->meta_state.blit.render_pass,
1614 &device->meta_state.alloc);
1616 ralloc_free(vs.nir);
1617 ralloc_free(fs_1d.nir);
1618 ralloc_free(fs_2d.nir);
1619 ralloc_free(fs_3d.nir);