| struct S_atomic { |
| int x; |
| uint a[10]; |
| uint y; |
| }; |
| |
| struct compute_main_inputs { |
| uint local_invocation_index_1_param : SV_GroupIndex; |
| }; |
| |
| |
| static uint local_invocation_index_1 = 0u; |
| groupshared S_atomic wg; |
| void compute_main_inner(uint local_invocation_index_2) { |
| uint idx = 0u; |
| wg.x = 0; |
| wg.y = 0u; |
| idx = local_invocation_index_2; |
| { |
| while(true) { |
| if (!((idx < 10u))) { |
| break; |
| } |
| uint x_35 = idx; |
| uint v = 0u; |
| InterlockedExchange(wg.a[x_35], 0u, v); |
| { |
| idx = (idx + 1u); |
| } |
| continue; |
| } |
| } |
| GroupMemoryBarrierWithGroupSync(); |
| uint v_1 = 0u; |
| InterlockedExchange(wg.a[4], 1u, v_1); |
| } |
| |
| void compute_main_1() { |
| uint x_53 = local_invocation_index_1; |
| compute_main_inner(x_53); |
| } |
| |
| void compute_main_inner_1(uint local_invocation_index_1_param) { |
| if ((local_invocation_index_1_param == 0u)) { |
| wg.x = 0; |
| wg.y = 0u; |
| } |
| { |
| uint v_2 = 0u; |
| v_2 = local_invocation_index_1_param; |
| while(true) { |
| uint v_3 = v_2; |
| if ((v_3 >= 10u)) { |
| break; |
| } |
| uint v_4 = 0u; |
| InterlockedExchange(wg.a[v_3], 0u, v_4); |
| { |
| v_2 = (v_3 + 1u); |
| } |
| continue; |
| } |
| } |
| GroupMemoryBarrierWithGroupSync(); |
| local_invocation_index_1 = local_invocation_index_1_param; |
| compute_main_1(); |
| } |
| |
| [numthreads(1, 1, 1)] |
| void compute_main(compute_main_inputs inputs) { |
| compute_main_inner_1(inputs.local_invocation_index_1_param); |
| } |
| |