avoid creating the bind group for indirect validation if buffer size is 0

This commit is contained in:
teoxoy 2024-10-17 12:55:08 +02:00 committed by Teodor Tanasoaia
parent 74ef445bca
commit 94e040bc8a
3 changed files with 185 additions and 95 deletions

View File

@ -16,7 +16,7 @@ static NUM_WORKGROUPS_BUILTIN: GpuTestConfiguration = GpuTestConfiguration::new(
) )
.run_async(|ctx| async move { .run_async(|ctx| async move {
let num_workgroups = [1, 2, 3]; let num_workgroups = [1, 2, 3];
let res = run_test(&ctx, &num_workgroups, false).await; let res = run_test(&ctx, &num_workgroups).await;
assert_eq!(res, num_workgroups); assert_eq!(res, num_workgroups);
}); });
@ -38,16 +38,16 @@ static DISCARD_DISPATCH: GpuTestConfiguration = GpuTestConfiguration::new()
.run_async(|ctx| async move { .run_async(|ctx| async move {
let max = ctx.device.limits().max_compute_workgroups_per_dimension; let max = ctx.device.limits().max_compute_workgroups_per_dimension;
let res = run_test(&ctx, &[max, max, max], false).await; let res = run_test(&ctx, &[max, max, max]).await;
assert_eq!(res, [max; 3]); assert_eq!(res, [max; 3]);
let res = run_test(&ctx, &[max + 1, 1, 1], false).await; let res = run_test(&ctx, &[max + 1, 1, 1]).await;
assert_eq!(res, [0; 3]); assert_eq!(res, [0; 3]);
let res = run_test(&ctx, &[1, max + 1, 1], false).await; let res = run_test(&ctx, &[1, max + 1, 1]).await;
assert_eq!(res, [0; 3]); assert_eq!(res, [0; 3]);
let res = run_test(&ctx, &[1, 1, max + 1], false).await; let res = run_test(&ctx, &[1, 1, max + 1]).await;
assert_eq!(res, [0; 3]); assert_eq!(res, [0; 3]);
}); });
@ -68,7 +68,24 @@ static RESET_BIND_GROUPS: GpuTestConfiguration = GpuTestConfiguration::new()
.run_async(|ctx| async move { .run_async(|ctx| async move {
ctx.device.push_error_scope(wgpu::ErrorFilter::Validation); ctx.device.push_error_scope(wgpu::ErrorFilter::Validation);
let _ = run_test(&ctx, &[0, 0, 0], true).await; let test_resources = TestResources::new(&ctx);
let indirect_buffer = ctx.device.create_buffer(&wgpu::BufferDescriptor {
label: None,
size: 12,
usage: wgpu::BufferUsages::COPY_DST | wgpu::BufferUsages::INDIRECT,
mapped_at_creation: false,
});
let mut encoder = ctx.device.create_command_encoder(&Default::default());
{
let mut compute_pass = encoder.begin_compute_pass(&Default::default());
compute_pass.set_pipeline(&test_resources.pipeline);
compute_pass.set_push_constants(0, &[0, 0, 0, 0]);
// compute_pass.set_bind_group(0, Some(&test_resources.bind_group), &[]);
compute_pass.dispatch_workgroups_indirect(&indirect_buffer, 0);
}
ctx.queue.submit(Some(encoder.finish()));
let error = pollster::block_on(ctx.device.pop_error_scope()); let error = pollster::block_on(ctx.device.pop_error_scope());
assert!(error.map_or(false, |error| { assert!(error.map_or(false, |error| {
@ -76,11 +93,59 @@ static RESET_BIND_GROUPS: GpuTestConfiguration = GpuTestConfiguration::new()
})); }));
}); });
async fn run_test( /// Make sure that zero sized buffer validation is raised.
ctx: &TestingContext, #[gpu_test]
num_workgroups: &[u32; 3], static ZERO_SIZED_BUFFER: GpuTestConfiguration = GpuTestConfiguration::new()
forget_to_set_bind_group: bool, .parameters(
) -> [u32; 3] { TestParameters::default()
.features(wgpu::Features::PUSH_CONSTANTS)
.downlevel_flags(
wgpu::DownlevelFlags::COMPUTE_SHADERS | wgpu::DownlevelFlags::INDIRECT_EXECUTION,
)
.limits(wgpu::Limits {
max_push_constant_size: 4,
..wgpu::Limits::downlevel_defaults()
}),
)
.run_async(|ctx| async move {
ctx.device.push_error_scope(wgpu::ErrorFilter::Validation);
let test_resources = TestResources::new(&ctx);
let indirect_buffer = ctx.device.create_buffer(&wgpu::BufferDescriptor {
label: None,
size: 0,
usage: wgpu::BufferUsages::COPY_DST | wgpu::BufferUsages::INDIRECT,
mapped_at_creation: false,
});
let mut encoder = ctx.device.create_command_encoder(&Default::default());
{
let mut compute_pass = encoder.begin_compute_pass(&Default::default());
compute_pass.set_pipeline(&test_resources.pipeline);
compute_pass.set_push_constants(0, &[0, 0, 0, 0]);
compute_pass.set_bind_group(0, Some(&test_resources.bind_group), &[]);
compute_pass.dispatch_workgroups_indirect(&indirect_buffer, 0);
}
ctx.queue.submit(Some(encoder.finish()));
let error = pollster::block_on(ctx.device.pop_error_scope());
assert!(error.map_or(false, |error| {
format!("{error}").contains(
"Indirect buffer uses bytes 0..12 which overruns indirect buffer of size 0",
)
}));
});
struct TestResources {
pipeline: wgpu::ComputePipeline,
out_buffer: wgpu::Buffer,
readback_buffer: wgpu::Buffer,
bind_group: wgpu::BindGroup,
}
impl TestResources {
fn new(ctx: &TestingContext) -> Self {
const SHADER_SRC: &str = " const SHADER_SRC: &str = "
struct TestOffsetPc { struct TestOffsetPc {
inner: u32, inner: u32,
@ -170,6 +235,18 @@ async fn run_test(
}], }],
}); });
Self {
pipeline,
out_buffer,
readback_buffer,
bind_group,
}
}
}
async fn run_test(ctx: &TestingContext, num_workgroups: &[u32; 3]) -> [u32; 3] {
let test_resources = TestResources::new(ctx);
let mut res = None; let mut res = None;
for (indirect_offset, indirect_buffer_size) in [ for (indirect_offset, indirect_buffer_size) in [
@ -202,19 +279,24 @@ async fn run_test(
let mut encoder = ctx.device.create_command_encoder(&Default::default()); let mut encoder = ctx.device.create_command_encoder(&Default::default());
{ {
let mut compute_pass = encoder.begin_compute_pass(&Default::default()); let mut compute_pass = encoder.begin_compute_pass(&Default::default());
compute_pass.set_pipeline(&pipeline); compute_pass.set_pipeline(&test_resources.pipeline);
compute_pass.set_push_constants(0, &[0, 0, 0, 0]); compute_pass.set_push_constants(0, &[0, 0, 0, 0]);
if !forget_to_set_bind_group { compute_pass.set_bind_group(0, Some(&test_resources.bind_group), &[]);
compute_pass.set_bind_group(0, Some(&bind_group), &[]);
}
compute_pass.dispatch_workgroups_indirect(&indirect_buffer, indirect_offset); compute_pass.dispatch_workgroups_indirect(&indirect_buffer, indirect_offset);
} }
encoder.copy_buffer_to_buffer(&out_buffer, 0, &readback_buffer, 0, 12); encoder.copy_buffer_to_buffer(
&test_resources.out_buffer,
0,
&test_resources.readback_buffer,
0,
12,
);
ctx.queue.submit(Some(encoder.finish())); ctx.queue.submit(Some(encoder.finish()));
readback_buffer test_resources
.readback_buffer
.slice(..) .slice(..)
.map_async(wgpu::MapMode::Read, |_| {}); .map_async(wgpu::MapMode::Read, |_| {});
@ -222,11 +304,11 @@ async fn run_test(
.await .await
.panic_on_timeout(); .panic_on_timeout();
let view = readback_buffer.slice(..).get_mapped_range(); let view = test_resources.readback_buffer.slice(..).get_mapped_range();
let current_res = *bytemuck::from_bytes(&view); let current_res = *bytemuck::from_bytes(&view);
drop(view); drop(view);
readback_buffer.unmap(); test_resources.readback_buffer.unmap();
if let Some(past_res) = res { if let Some(past_res) = res {
assert_eq!(past_res, current_res); assert_eq!(past_res, current_res);

View File

@ -780,7 +780,10 @@ impl Device {
let bind_group = indirect_validation let bind_group = indirect_validation
.create_src_bind_group(self.raw(), &self.limits, buffer_size, raw_buffer) .create_src_bind_group(self.raw(), &self.limits, buffer_size, raw_buffer)
.map_err(resource::CreateBufferError::IndirectValidationBindGroup)?; .map_err(resource::CreateBufferError::IndirectValidationBindGroup)?;
Ok(Snatchable::new(bind_group)) match bind_group {
Some(bind_group) => Ok(Snatchable::new(bind_group)),
None => Ok(Snatchable::empty()),
}
} else { } else {
Ok(Snatchable::empty()) Ok(Snatchable::empty())
} }

View File

@ -266,14 +266,18 @@ impl IndirectValidation {
}) })
} }
/// `Ok(None)` will only be returned if `buffer_size` is `0`.
pub fn create_src_bind_group( pub fn create_src_bind_group(
&self, &self,
device: &dyn hal::DynDevice, device: &dyn hal::DynDevice,
limits: &wgt::Limits, limits: &wgt::Limits,
buffer_size: u64, buffer_size: u64,
buffer: &dyn hal::DynBuffer, buffer: &dyn hal::DynBuffer,
) -> Result<Box<dyn hal::DynBindGroup>, DeviceError> { ) -> Result<Option<Box<dyn hal::DynBindGroup>>, DeviceError> {
let binding_size = calculate_src_buffer_binding_size(buffer_size, limits); let binding_size = calculate_src_buffer_binding_size(buffer_size, limits);
let Some(binding_size) = NonZeroU64::new(binding_size) else {
return Ok(None);
};
let hal_desc = hal::BindGroupDescriptor { let hal_desc = hal::BindGroupDescriptor {
label: None, label: None,
layout: self.src_bind_group_layout.as_ref(), layout: self.src_bind_group_layout.as_ref(),
@ -285,7 +289,7 @@ impl IndirectValidation {
buffers: &[hal::BufferBinding { buffers: &[hal::BufferBinding {
buffer, buffer,
offset: 0, offset: 0,
size: Some(NonZeroU64::new(binding_size).unwrap()), size: Some(binding_size),
}], }],
samplers: &[], samplers: &[],
textures: &[], textures: &[],
@ -294,6 +298,7 @@ impl IndirectValidation {
unsafe { unsafe {
device device
.create_bind_group(&hal_desc) .create_bind_group(&hal_desc)
.map(Some)
.map_err(DeviceError::from_hal) .map_err(DeviceError::from_hal)
} }
} }