| struct S0_atomic { |
| int x; |
| uint a; |
| int y; |
| int z; |
| }; |
| |
| struct S1_atomic { |
| int x; |
| S0_atomic a; |
| int y; |
| int z; |
| }; |
| |
| struct S2_atomic { |
| int x; |
| int y; |
| int z; |
| S1_atomic a; |
| }; |
| |
| struct compute_main_inputs { |
| uint local_invocation_index_1_param : SV_GroupIndex; |
| }; |
| |
| |
| static uint local_invocation_index_1 = 0u; |
| groupshared S2_atomic wg; |
| void compute_main_inner(uint local_invocation_index_2) { |
| wg.x = int(0); |
| wg.y = int(0); |
| wg.z = int(0); |
| wg.a.x = int(0); |
| wg.a.a.x = int(0); |
| uint v = 0u; |
| InterlockedExchange(wg.a.a.a, 0u, v); |
| wg.a.a.y = int(0); |
| wg.a.a.z = int(0); |
| wg.a.y = int(0); |
| wg.a.z = int(0); |
| GroupMemoryBarrierWithGroupSync(); |
| uint v_1 = 0u; |
| InterlockedExchange(wg.a.a.a, 1u, v_1); |
| } |
| |
| void compute_main_1() { |
| uint x_44 = local_invocation_index_1; |
| compute_main_inner(x_44); |
| } |
| |
| void compute_main_inner_1(uint local_invocation_index_1_param) { |
| if ((local_invocation_index_1_param == 0u)) { |
| wg.x = int(0); |
| wg.y = int(0); |
| wg.z = int(0); |
| wg.a.x = int(0); |
| wg.a.a.x = int(0); |
| uint v_2 = 0u; |
| InterlockedExchange(wg.a.a.a, 0u, v_2); |
| wg.a.a.y = int(0); |
| wg.a.a.z = int(0); |
| wg.a.y = int(0); |
| wg.a.z = int(0); |
| } |
| GroupMemoryBarrierWithGroupSync(); |
| local_invocation_index_1 = local_invocation_index_1_param; |
| compute_main_1(); |
| } |
| |
| [numthreads(1, 1, 1)] |
| void compute_main(compute_main_inputs inputs) { |
| compute_main_inner_1(inputs.local_invocation_index_1_param); |
| } |
| |