Flutter Engine
 
Loading...
Searching...
No Matches
compute_pass_vk.cc
Go to the documentation of this file.
1// Copyright 2013 The Flutter Authors. All rights reserved.
2// Use of this source code is governed by a BSD-style license that can be
3// found in the LICENSE file.
4
6
12#include "vulkan/vulkan_structs.hpp"
13
14namespace impeller {
15
16ComputePassVK::ComputePassVK(std::shared_ptr<const Context> context,
17 std::shared_ptr<CommandBufferVK> command_buffer)
18 : ComputePass(std::move(context)),
19 command_buffer_(std::move(command_buffer)) {
20 // TOOD(dnfield): This should be moved to caps. But for now keeping this
21 // in parallel with Metal.
22 max_wg_size_ = ContextVK::Cast(*context_)
23 .GetPhysicalDevice()
24 .getProperties()
25 .limits.maxComputeWorkGroupSize;
26 is_valid_ = true;
27}
28
29ComputePassVK::~ComputePassVK() = default;
30
31bool ComputePassVK::IsValid() const {
32 return is_valid_;
33}
34
35void ComputePassVK::OnSetLabel(const std::string& label) {
36 if (label.empty()) {
37 return;
38 }
39 label_ = label;
40}
41
42// |RenderPass|
43void ComputePassVK::SetCommandLabel(std::string_view label) {
44#ifdef IMPELLER_DEBUG
45 command_buffer_->PushDebugGroup(label);
46 has_label_ = true;
47#endif // IMPELLER_DEBUG
48}
49
50// |ComputePass|
51void ComputePassVK::SetPipeline(
52 const std::shared_ptr<Pipeline<ComputePipelineDescriptor>>& pipeline) {
53 const auto& pipeline_vk = ComputePipelineVK::Cast(*pipeline);
54 const vk::CommandBuffer& command_buffer_vk =
55 command_buffer_->GetCommandBuffer();
56 command_buffer_vk.bindPipeline(vk::PipelineBindPoint::eCompute,
57 pipeline_vk.GetPipeline());
58 pipeline_layout_ = pipeline_vk.GetPipelineLayout();
59
60 auto descriptor_result = command_buffer_->AllocateDescriptorSets(
61 pipeline_vk.GetDescriptorSetLayout(), pipeline_vk.GetPipelineKey(),
62 ContextVK::Cast(*context_));
63 if (!descriptor_result.ok()) {
64 return;
65 }
66 descriptor_set_ = descriptor_result.value();
67 pipeline_valid_ = true;
68}
69
70// |ComputePass|
71fml::Status ComputePassVK::Compute(const ISize& grid_size) {
72 if (grid_size.IsEmpty() || !pipeline_valid_) {
73 bound_image_offset_ = 0u;
74 bound_buffer_offset_ = 0u;
75 descriptor_write_offset_ = 0u;
76 has_label_ = false;
77 pipeline_valid_ = false;
79 "Invalid pipeline or empty grid.");
80 }
81
82 const ContextVK& context_vk = ContextVK::Cast(*context_);
83 for (auto i = 0u; i < descriptor_write_offset_; i++) {
84 write_workspace_[i].dstSet = descriptor_set_;
85 }
86
87 context_vk.GetDevice().updateDescriptorSets(descriptor_write_offset_,
88 write_workspace_.data(), 0u, {});
89 const vk::CommandBuffer& command_buffer_vk =
90 command_buffer_->GetCommandBuffer();
91
92 command_buffer_vk.bindDescriptorSets(
93 vk::PipelineBindPoint::eCompute, // bind point
94 pipeline_layout_, // layout
95 0, // first set
96 1, // set count
97 &descriptor_set_, // sets
98 0, // offset count
99 nullptr // offsets
100 );
101
102 int64_t width = grid_size.width;
103 int64_t height = grid_size.height;
104
105 // Special case for linear processing.
106 if (height == 1) {
107 command_buffer_vk.dispatch(width, 1, 1);
108 } else {
109 while (width > max_wg_size_[0]) {
110 width = std::max(static_cast<int64_t>(1), width / 2);
111 }
112 while (height > max_wg_size_[1]) {
113 height = std::max(static_cast<int64_t>(1), height / 2);
114 }
115 command_buffer_vk.dispatch(width, height, 1);
116 }
117
118#ifdef IMPELLER_DEBUG
119 if (has_label_) {
120 command_buffer_->PopDebugGroup();
121 }
122 has_label_ = false;
123#endif // IMPELLER_DEBUG
124
125 bound_image_offset_ = 0u;
126 bound_buffer_offset_ = 0u;
127 descriptor_write_offset_ = 0u;
128 has_label_ = false;
129 pipeline_valid_ = false;
130
131 return fml::Status();
132}
133
134// |ResourceBinder|
135bool ComputePassVK::BindResource(ShaderStage stage,
136 DescriptorType type,
137 const ShaderUniformSlot& slot,
138 const ShaderMetadata* metadata,
139 BufferView view) {
140 return BindResource(slot.binding, type, view);
141}
142
143// |ResourceBinder|
144bool ComputePassVK::BindResource(ShaderStage stage,
145 DescriptorType type,
146 const SampledImageSlot& slot,
147 const ShaderMetadata* metadata,
148 std::shared_ptr<const Texture> texture,
149 raw_ptr<const Sampler> sampler) {
150 if (bound_image_offset_ >= kMaxBindings) {
151 return false;
152 }
153 if (!texture->IsValid() || !sampler) {
154 return false;
155 }
156 const TextureVK& texture_vk = TextureVK::Cast(*texture);
157 const SamplerVK& sampler_vk = SamplerVK::Cast(*sampler);
158
159 if (!command_buffer_->Track(texture)) {
160 return false;
161 }
162
163 vk::DescriptorImageInfo image_info;
164 image_info.imageLayout = vk::ImageLayout::eShaderReadOnlyOptimal;
165 image_info.sampler = sampler_vk.GetSampler();
166 image_info.imageView = texture_vk.GetImageView();
167 image_workspace_[bound_image_offset_++] = image_info;
168
169 vk::WriteDescriptorSet write_set;
170 write_set.dstBinding = slot.binding;
171 write_set.descriptorCount = 1u;
172 write_set.descriptorType = ToVKDescriptorType(type);
173 write_set.pImageInfo = &image_workspace_[bound_image_offset_ - 1];
174
175 write_workspace_[descriptor_write_offset_++] = write_set;
176 return true;
177}
178
179bool ComputePassVK::BindResource(size_t binding,
180 DescriptorType type,
181 BufferView view) {
182 if (bound_buffer_offset_ >= kMaxBindings) {
183 return false;
184 }
185
186 auto buffer = DeviceBufferVK::Cast(*view.GetBuffer()).GetBuffer();
187 if (!buffer) {
188 return false;
189 }
190
191 std::shared_ptr<const DeviceBuffer> device_buffer = view.TakeBuffer();
192 if (device_buffer && !command_buffer_->Track(device_buffer)) {
193 return false;
194 }
195
196 uint32_t offset = view.GetRange().offset;
197
198 vk::DescriptorBufferInfo buffer_info;
199 buffer_info.buffer = buffer;
200 buffer_info.offset = offset;
201 buffer_info.range = view.GetRange().length;
202 buffer_workspace_[bound_buffer_offset_++] = buffer_info;
203
204 vk::WriteDescriptorSet write_set;
205 write_set.dstBinding = binding;
206 write_set.descriptorCount = 1u;
207 write_set.descriptorType = ToVKDescriptorType(type);
208 write_set.pBufferInfo = &buffer_workspace_[bound_buffer_offset_ - 1];
209
210 write_workspace_[descriptor_write_offset_++] = write_set;
211 return true;
212}
213
214// Note:
215// https://github.com/KhronosGroup/Vulkan-Docs/wiki/Synchronization-Examples
216// Seems to suggest that anything more finely grained than a global memory
217// barrier is likely to be weakened into a global barrier. Confirming this on
218// mobile devices will require some experimentation.
219
220// |ComputePass|
221void ComputePassVK::AddBufferMemoryBarrier() {
222 vk::MemoryBarrier barrier;
223 barrier.srcAccessMask = vk::AccessFlagBits::eShaderWrite;
224 barrier.dstAccessMask = vk::AccessFlagBits::eShaderRead;
225
226 command_buffer_->GetCommandBuffer().pipelineBarrier(
227 vk::PipelineStageFlagBits::eComputeShader,
228 vk::PipelineStageFlagBits::eComputeShader, {}, 1, &barrier, 0, {}, 0, {});
229}
230
231// |ComputePass|
232void ComputePassVK::AddTextureMemoryBarrier() {
233 vk::MemoryBarrier barrier;
234 barrier.srcAccessMask = vk::AccessFlagBits::eShaderWrite;
235 barrier.dstAccessMask = vk::AccessFlagBits::eShaderRead;
236
237 command_buffer_->GetCommandBuffer().pipelineBarrier(
238 vk::PipelineStageFlagBits::eComputeShader,
239 vk::PipelineStageFlagBits::eComputeShader, {}, 1, &barrier, 0, {}, 0, {});
240}
241
242// |ComputePass|
243bool ComputePassVK::EncodeCommands() const {
244 // Since we only use global memory barrier, we don't have to worry about
245 // compute to compute dependencies across cmd buffers. Instead, we pessimize
246 // here and assume that we wrote to a storage image or buffer and that a
247 // render pass will read from it. if there are ever scenarios where we end up
248 // with compute to compute dependencies this should be revisited.
249
250 // This does not currently handle image barriers as we do not use them
251 // for anything.
252 vk::MemoryBarrier barrier;
253 barrier.srcAccessMask = vk::AccessFlagBits::eShaderWrite;
254 barrier.dstAccessMask =
255 vk::AccessFlagBits::eIndexRead | vk::AccessFlagBits::eVertexAttributeRead;
256
257 command_buffer_->GetCommandBuffer().pipelineBarrier(
258 vk::PipelineStageFlagBits::eComputeShader,
259 vk::PipelineStageFlagBits::eVertexInput, {}, 1, &barrier, 0, {}, 0, {});
260
261 return true;
262}
263
264} // namespace impeller
GLenum type
FlView * view
FlTexture * texture
DEF_SWITCHES_START aot vmservice shared library Name of the *so containing AOT compiled Dart assets for launching the service isolate vm snapshot The VM snapshot data that will be memory mapped as read only SnapshotAssetPath must be present isolate snapshot The isolate snapshot data that will be memory mapped as read only SnapshotAssetPath must be present cache dir Path to the cache directory This is different from the persistent_cache_path in embedder which is used for Skia shader cache icu native lib Path to the library file that exports the ICU data vm service The hostname IP address on which the Dart VM Service should be served If not defaults to or::depending on whether ipv6 is specified disable vm Disable the Dart VM Service The Dart VM Service is never available in release mode Bind to the IPv6 localhost address for the Dart VM Service Ignored if vm service host is set profile Make the profiler discard new samples once the profiler sample buffer is full When this flag is not the profiler sample buffer is used as a ring buffer
Definition switch_defs.h:98
constexpr vk::DescriptorType ToVKDescriptorType(DescriptorType type)
Definition formats_vk.h:292
Definition ref_ptr.h:261
int32_t height
int32_t width