2023-02-28 17:59:28 -08:00
|
|
|
/*
|
|
|
|
* Copyright 2023 Conor McCarthy for CodeWeavers
|
|
|
|
*
|
|
|
|
* This library is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
|
|
* License as published by the Free Software Foundation; either
|
|
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
|
|
*
|
|
|
|
* This library is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* Lesser General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
|
|
* License along with this library; if not, write to the Free Software
|
|
|
|
* Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "vkd3d_shader_private.h"
|
2024-04-11 03:38:24 -07:00
|
|
|
#include "vkd3d_types.h"
|
2023-02-28 17:59:28 -08:00
|
|
|
|
2024-01-16 07:40:30 -08:00
|
|
|
bool vsir_program_init(struct vsir_program *program, const struct vkd3d_shader_version *version, unsigned int reserve)
|
2024-01-16 05:47:34 -08:00
|
|
|
{
|
2024-05-16 02:42:12 -07:00
|
|
|
memset(program, 0, sizeof(*program));
|
2024-01-16 07:40:30 -08:00
|
|
|
program->shader_version = *version;
|
2024-01-16 05:47:34 -08:00
|
|
|
return shader_instruction_array_init(&program->instructions, reserve);
|
|
|
|
}
|
|
|
|
|
|
|
|
void vsir_program_cleanup(struct vsir_program *program)
|
|
|
|
{
|
2024-01-18 11:41:27 -08:00
|
|
|
size_t i;
|
|
|
|
|
|
|
|
for (i = 0; i < program->block_name_count; ++i)
|
|
|
|
vkd3d_free((void *)program->block_names[i]);
|
|
|
|
vkd3d_free(program->block_names);
|
2024-01-16 05:47:34 -08:00
|
|
|
shader_instruction_array_destroy(&program->instructions);
|
2024-03-01 04:36:11 -08:00
|
|
|
shader_signature_cleanup(&program->input_signature);
|
|
|
|
shader_signature_cleanup(&program->output_signature);
|
|
|
|
shader_signature_cleanup(&program->patch_constant_signature);
|
2024-01-16 05:47:34 -08:00
|
|
|
}
|
|
|
|
|
2023-02-28 17:59:28 -08:00
|
|
|
static inline bool shader_register_is_phase_instance_id(const struct vkd3d_shader_register *reg)
|
|
|
|
{
|
|
|
|
return reg->type == VKD3DSPR_FORKINSTID || reg->type == VKD3DSPR_JOININSTID;
|
|
|
|
}
|
|
|
|
|
2024-01-22 11:32:55 -08:00
|
|
|
static bool vsir_instruction_is_dcl(const struct vkd3d_shader_instruction *instruction)
|
2023-02-28 17:59:28 -08:00
|
|
|
{
|
2024-04-24 11:34:42 -07:00
|
|
|
enum vkd3d_shader_opcode opcode = instruction->opcode;
|
|
|
|
return (VKD3DSIH_DCL <= opcode && opcode <= VKD3DSIH_DCL_VERTICES_OUT)
|
|
|
|
|| opcode == VKD3DSIH_HS_DECLS;
|
2023-02-28 17:59:28 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void vkd3d_shader_instruction_make_nop(struct vkd3d_shader_instruction *ins)
|
|
|
|
{
|
2023-09-05 02:31:31 -07:00
|
|
|
struct vkd3d_shader_location location = ins->location;
|
|
|
|
|
|
|
|
vsir_instruction_init(ins, &location, VKD3DSIH_NOP);
|
2023-02-28 17:59:28 -08:00
|
|
|
}
|
|
|
|
|
2024-01-24 05:26:35 -08:00
|
|
|
static bool vsir_instruction_init_with_params(struct vsir_program *program,
|
2024-01-10 12:23:04 -08:00
|
|
|
struct vkd3d_shader_instruction *ins, const struct vkd3d_shader_location *location,
|
2024-04-24 11:34:42 -07:00
|
|
|
enum vkd3d_shader_opcode opcode, unsigned int dst_count, unsigned int src_count)
|
2024-01-10 12:23:04 -08:00
|
|
|
{
|
2024-04-24 11:34:42 -07:00
|
|
|
vsir_instruction_init(ins, location, opcode);
|
2024-01-10 12:23:04 -08:00
|
|
|
ins->dst_count = dst_count;
|
|
|
|
ins->src_count = src_count;
|
|
|
|
|
2024-01-24 05:26:35 -08:00
|
|
|
if (!(ins->dst = vsir_program_get_dst_params(program, ins->dst_count)))
|
2024-01-10 12:23:04 -08:00
|
|
|
{
|
|
|
|
ERR("Failed to allocate %u destination parameters.\n", dst_count);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2024-01-24 05:26:35 -08:00
|
|
|
if (!(ins->src = vsir_program_get_src_params(program, ins->src_count)))
|
2024-01-10 12:23:04 -08:00
|
|
|
{
|
|
|
|
ERR("Failed to allocate %u source parameters.\n", src_count);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
memset(ins->dst, 0, sizeof(*ins->dst) * ins->dst_count);
|
|
|
|
memset(ins->src, 0, sizeof(*ins->src) * ins->src_count);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2024-04-01 21:24:03 -07:00
|
|
|
static bool get_opcode_from_rel_op(enum vkd3d_shader_rel_op rel_op, enum vkd3d_data_type data_type,
|
|
|
|
enum vkd3d_shader_opcode *opcode, bool *requires_swap)
|
|
|
|
{
|
|
|
|
switch (rel_op)
|
|
|
|
{
|
|
|
|
case VKD3D_SHADER_REL_OP_LT:
|
|
|
|
case VKD3D_SHADER_REL_OP_GT:
|
|
|
|
*requires_swap = (rel_op == VKD3D_SHADER_REL_OP_GT);
|
|
|
|
if (data_type == VKD3D_DATA_FLOAT)
|
|
|
|
{
|
|
|
|
*opcode = VKD3DSIH_LTO;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3D_SHADER_REL_OP_GE:
|
|
|
|
case VKD3D_SHADER_REL_OP_LE:
|
|
|
|
*requires_swap = (rel_op == VKD3D_SHADER_REL_OP_LE);
|
|
|
|
if (data_type == VKD3D_DATA_FLOAT)
|
|
|
|
{
|
|
|
|
*opcode = VKD3DSIH_GEO;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3D_SHADER_REL_OP_EQ:
|
|
|
|
*requires_swap = false;
|
|
|
|
if (data_type == VKD3D_DATA_FLOAT)
|
|
|
|
{
|
|
|
|
*opcode = VKD3DSIH_EQO;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3D_SHADER_REL_OP_NE:
|
|
|
|
*requires_swap = false;
|
|
|
|
if (data_type == VKD3D_DATA_FLOAT)
|
|
|
|
{
|
|
|
|
*opcode = VKD3DSIH_NEO;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
static enum vkd3d_result vsir_program_lower_ifc(struct vsir_program *program,
|
|
|
|
struct vkd3d_shader_instruction *ifc, unsigned int *tmp_idx,
|
|
|
|
struct vkd3d_shader_message_context *message_context)
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction_array *instructions = &program->instructions;
|
|
|
|
size_t pos = ifc - instructions->elements;
|
|
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
enum vkd3d_shader_opcode opcode;
|
|
|
|
bool swap;
|
|
|
|
|
|
|
|
if (!shader_instruction_array_insert_at(instructions, pos + 1, 2))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
if (*tmp_idx == ~0u)
|
|
|
|
*tmp_idx = program->temp_count++;
|
|
|
|
|
|
|
|
/* Replace ifc comparison with actual comparison, saving the result in the tmp register. */
|
|
|
|
if (!(get_opcode_from_rel_op(ifc->flags, ifc->src[0].reg.data_type, &opcode, &swap)))
|
|
|
|
{
|
|
|
|
vkd3d_shader_error(message_context, &ifc->location, VKD3D_SHADER_ERROR_VSIR_NOT_IMPLEMENTED,
|
|
|
|
"Aborting due to not yet implemented feature: opcode for rel_op %u and data type %u.",
|
|
|
|
ifc->flags, ifc->src[0].reg.data_type);
|
|
|
|
return VKD3D_ERROR_NOT_IMPLEMENTED;
|
|
|
|
}
|
|
|
|
|
|
|
|
ins = &instructions->elements[pos + 1];
|
|
|
|
if (!vsir_instruction_init_with_params(program, ins, &ifc->location, opcode, 1, 2))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
vsir_register_init(&ins->dst[0].reg, VKD3DSPR_TEMP, VKD3D_DATA_UINT, 1);
|
|
|
|
ins->dst[0].reg.dimension = VSIR_DIMENSION_VEC4;
|
|
|
|
ins->dst[0].reg.idx[0].offset = *tmp_idx;
|
|
|
|
ins->dst[0].write_mask = VKD3DSP_WRITEMASK_0;
|
|
|
|
|
|
|
|
ins->src[0] = ifc->src[swap];
|
|
|
|
ins->src[1] = ifc->src[!swap];
|
|
|
|
|
|
|
|
/* Create new if instruction using the previous result. */
|
|
|
|
ins = &instructions->elements[pos + 2];
|
|
|
|
if (!vsir_instruction_init_with_params(program, ins, &ifc->location, VKD3DSIH_IF, 0, 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
ins->flags = VKD3D_SHADER_CONDITIONAL_OP_NZ;
|
|
|
|
|
|
|
|
vsir_register_init(&ins->src[0].reg, VKD3DSPR_TEMP, VKD3D_DATA_UINT, 1);
|
|
|
|
ins->src[0].reg.dimension = VSIR_DIMENSION_VEC4;
|
|
|
|
ins->src[0].reg.idx[0].offset = *tmp_idx;
|
|
|
|
ins->src[0].swizzle = VKD3D_SHADER_SWIZZLE(X, X, X, X);
|
|
|
|
|
|
|
|
/* Make the original instruction no-op */
|
|
|
|
vkd3d_shader_instruction_make_nop(ifc);
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
static enum vkd3d_result vsir_program_lower_texkill(struct vsir_program *program,
|
|
|
|
struct vkd3d_shader_instruction *texkill, unsigned int *tmp_idx)
|
2024-01-10 12:23:04 -08:00
|
|
|
{
|
2024-04-11 03:19:34 -07:00
|
|
|
const unsigned int components_read = 3 + (program->shader_version.major >= 2);
|
2024-01-10 12:23:04 -08:00
|
|
|
struct vkd3d_shader_instruction_array *instructions = &program->instructions;
|
2024-04-11 03:19:34 -07:00
|
|
|
size_t pos = texkill - instructions->elements;
|
|
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
unsigned int j;
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
if (!shader_instruction_array_insert_at(instructions, pos + 1, components_read + 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
if (*tmp_idx == ~0u)
|
|
|
|
*tmp_idx = program->temp_count++;
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
/* tmp = ins->dst[0] < 0 */
|
|
|
|
|
|
|
|
ins = &instructions->elements[pos + 1];
|
|
|
|
if (!vsir_instruction_init_with_params(program, ins, &texkill->location, VKD3DSIH_LTO, 1, 2))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
vsir_register_init(&ins->dst[0].reg, VKD3DSPR_TEMP, VKD3D_DATA_UINT, 1);
|
|
|
|
ins->dst[0].reg.dimension = VSIR_DIMENSION_VEC4;
|
|
|
|
ins->dst[0].reg.idx[0].offset = *tmp_idx;
|
|
|
|
ins->dst[0].write_mask = VKD3DSP_WRITEMASK_ALL;
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
ins->src[0].reg = texkill->dst[0].reg;
|
|
|
|
ins->src[0].swizzle = VKD3D_SHADER_NO_SWIZZLE;
|
|
|
|
vsir_register_init(&ins->src[1].reg, VKD3DSPR_IMMCONST, VKD3D_DATA_FLOAT, 0);
|
|
|
|
ins->src[1].reg.dimension = VSIR_DIMENSION_VEC4;
|
|
|
|
ins->src[1].reg.u.immconst_f32[0] = 0.0f;
|
|
|
|
ins->src[1].reg.u.immconst_f32[1] = 0.0f;
|
|
|
|
ins->src[1].reg.u.immconst_f32[2] = 0.0f;
|
|
|
|
ins->src[1].reg.u.immconst_f32[3] = 0.0f;
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
/* tmp.x = tmp.x || tmp.y */
|
|
|
|
/* tmp.x = tmp.x || tmp.z */
|
|
|
|
/* tmp.x = tmp.x || tmp.w, if sm >= 2.0 */
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
for (j = 1; j < components_read; ++j)
|
|
|
|
{
|
|
|
|
ins = &instructions->elements[pos + 1 + j];
|
|
|
|
if (!(vsir_instruction_init_with_params(program, ins, &texkill->location, VKD3DSIH_OR, 1, 2)))
|
2024-01-10 12:23:04 -08:00
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
vsir_register_init(&ins->dst[0].reg, VKD3DSPR_TEMP, VKD3D_DATA_UINT, 1);
|
|
|
|
ins->dst[0].reg.dimension = VSIR_DIMENSION_VEC4;
|
2024-04-11 03:19:34 -07:00
|
|
|
ins->dst[0].reg.idx[0].offset = *tmp_idx;
|
|
|
|
ins->dst[0].write_mask = VKD3DSP_WRITEMASK_0;
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
vsir_register_init(&ins->src[0].reg, VKD3DSPR_TEMP, VKD3D_DATA_UINT, 1);
|
|
|
|
ins->src[0].reg.dimension = VSIR_DIMENSION_VEC4;
|
|
|
|
ins->src[0].reg.idx[0].offset = *tmp_idx;
|
|
|
|
ins->src[0].swizzle = VKD3D_SHADER_SWIZZLE(X, X, X, X);
|
|
|
|
vsir_register_init(&ins->src[1].reg, VKD3DSPR_TEMP, VKD3D_DATA_UINT, 1);
|
2024-01-10 12:23:04 -08:00
|
|
|
ins->src[1].reg.dimension = VSIR_DIMENSION_VEC4;
|
2024-04-11 03:19:34 -07:00
|
|
|
ins->src[1].reg.idx[0].offset = *tmp_idx;
|
|
|
|
ins->src[1].swizzle = vkd3d_shader_create_swizzle(j, j, j, j);
|
|
|
|
}
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
/* discard_nz tmp.x */
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
ins = &instructions->elements[pos + 1 + components_read];
|
|
|
|
if (!(vsir_instruction_init_with_params(program, ins, &texkill->location, VKD3DSIH_DISCARD, 0, 1)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
ins->flags = VKD3D_SHADER_CONDITIONAL_OP_NZ;
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
vsir_register_init(&ins->src[0].reg, VKD3DSPR_TEMP, VKD3D_DATA_UINT, 1);
|
|
|
|
ins->src[0].reg.dimension = VSIR_DIMENSION_VEC4;
|
|
|
|
ins->src[0].reg.idx[0].offset = *tmp_idx;
|
|
|
|
ins->src[0].swizzle = VKD3D_SHADER_SWIZZLE(X, X, X, X);
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
/* Make the original instruction no-op */
|
|
|
|
vkd3d_shader_instruction_make_nop(texkill);
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:38:24 -07:00
|
|
|
/* The Shader Model 5 Assembly documentation states: "If components of a mad
|
|
|
|
* instruction are tagged as precise, the hardware must execute a mad instruction
|
|
|
|
* or the exact equivalent, and it cannot split it into a multiply followed by an add."
|
|
|
|
* But DXIL.rst states the opposite: "Floating point multiply & add. This operation is
|
|
|
|
* not fused for "precise" operations."
|
|
|
|
* Windows drivers seem to conform with the latter, for SM 4-5 and SM 6. */
|
|
|
|
static enum vkd3d_result vsir_program_lower_precise_mad(struct vsir_program *program,
|
|
|
|
struct vkd3d_shader_instruction *mad, unsigned int *tmp_idx)
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction_array *instructions = &program->instructions;
|
|
|
|
struct vkd3d_shader_instruction *mul_ins, *add_ins;
|
|
|
|
size_t pos = mad - instructions->elements;
|
|
|
|
struct vkd3d_shader_dst_param *mul_dst;
|
|
|
|
|
|
|
|
if (!(mad->flags & VKD3DSI_PRECISE_XYZW))
|
|
|
|
return VKD3D_OK;
|
|
|
|
|
|
|
|
if (!shader_instruction_array_insert_at(instructions, pos + 1, 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
if (*tmp_idx == ~0u)
|
|
|
|
*tmp_idx = program->temp_count++;
|
|
|
|
|
|
|
|
mul_ins = &instructions->elements[pos];
|
|
|
|
add_ins = &instructions->elements[pos + 1];
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
mul_ins->opcode = VKD3DSIH_MUL;
|
2024-04-11 03:38:24 -07:00
|
|
|
mul_ins->src_count = 2;
|
|
|
|
|
|
|
|
if (!(vsir_instruction_init_with_params(program, add_ins, &mul_ins->location, VKD3DSIH_ADD, 1, 2)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
add_ins->flags = mul_ins->flags & VKD3DSI_PRECISE_XYZW;
|
|
|
|
|
|
|
|
mul_dst = mul_ins->dst;
|
|
|
|
*add_ins->dst = *mul_dst;
|
|
|
|
|
|
|
|
mul_dst->modifiers = 0;
|
|
|
|
vsir_register_init(&mul_dst->reg, VKD3DSPR_TEMP, mul_ins->src[0].reg.data_type, 1);
|
|
|
|
mul_dst->reg.dimension = add_ins->dst->reg.dimension;
|
|
|
|
mul_dst->reg.idx[0].offset = *tmp_idx;
|
|
|
|
|
|
|
|
add_ins->src[0].reg = mul_dst->reg;
|
|
|
|
add_ins->src[0].swizzle = vsir_swizzle_from_writemask(mul_dst->write_mask);
|
|
|
|
add_ins->src[0].modifiers = 0;
|
|
|
|
add_ins->src[1] = mul_ins->src[2];
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-04-01 21:24:03 -07:00
|
|
|
static enum vkd3d_result vsir_program_lower_instructions(struct vsir_program *program,
|
|
|
|
struct vkd3d_shader_message_context *message_context)
|
2024-04-11 03:19:34 -07:00
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction_array *instructions = &program->instructions;
|
|
|
|
unsigned int tmp_idx = ~0u, i;
|
|
|
|
enum vkd3d_result ret;
|
2024-01-10 12:23:04 -08:00
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
for (i = 0; i < instructions->count; ++i)
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction *ins = &instructions->elements[i];
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (ins->opcode)
|
2024-04-11 03:19:34 -07:00
|
|
|
{
|
2024-04-01 21:24:03 -07:00
|
|
|
case VKD3DSIH_IFC:
|
|
|
|
if ((ret = vsir_program_lower_ifc(program, ins, &tmp_idx, message_context)) < 0)
|
|
|
|
return ret;
|
|
|
|
break;
|
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
case VKD3DSIH_TEXKILL:
|
|
|
|
if ((ret = vsir_program_lower_texkill(program, ins, &tmp_idx)) < 0)
|
|
|
|
return ret;
|
|
|
|
break;
|
|
|
|
|
2024-04-11 03:38:24 -07:00
|
|
|
case VKD3DSIH_MAD:
|
|
|
|
if ((ret = vsir_program_lower_precise_mad(program, ins, &tmp_idx)) < 0)
|
|
|
|
return ret;
|
|
|
|
break;
|
|
|
|
|
2024-02-24 07:45:39 -08:00
|
|
|
case VKD3DSIH_DCL_CONSTANT_BUFFER:
|
2024-04-11 03:24:23 -07:00
|
|
|
case VKD3DSIH_DCL_TEMPS:
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
|
|
|
break;
|
|
|
|
|
2024-04-11 03:19:34 -07:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2024-01-10 12:23:04 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2023-02-28 17:59:28 -08:00
|
|
|
static void shader_register_eliminate_phase_addressing(struct vkd3d_shader_register *reg,
|
|
|
|
unsigned int instance_id)
|
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
2023-05-03 00:10:37 -07:00
|
|
|
for (i = 0; i < reg->idx_count; ++i)
|
2023-02-28 17:59:28 -08:00
|
|
|
{
|
|
|
|
if (reg->idx[i].rel_addr && shader_register_is_phase_instance_id(®->idx[i].rel_addr->reg))
|
|
|
|
{
|
|
|
|
reg->idx[i].rel_addr = NULL;
|
|
|
|
reg->idx[i].offset += instance_id;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void shader_instruction_eliminate_phase_instance_id(struct vkd3d_shader_instruction *ins,
|
|
|
|
unsigned int instance_id)
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_register *reg;
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
for (i = 0; i < ins->src_count; ++i)
|
|
|
|
{
|
|
|
|
reg = (struct vkd3d_shader_register *)&ins->src[i].reg;
|
|
|
|
if (shader_register_is_phase_instance_id(reg))
|
|
|
|
{
|
2023-11-08 02:11:39 -08:00
|
|
|
vsir_register_init(reg, VKD3DSPR_IMMCONST, reg->data_type, 0);
|
2023-12-05 14:24:36 -08:00
|
|
|
reg->u.immconst_u32[0] = instance_id;
|
2023-02-28 17:59:28 -08:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
shader_register_eliminate_phase_addressing(reg, instance_id);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < ins->dst_count; ++i)
|
|
|
|
shader_register_eliminate_phase_addressing((struct vkd3d_shader_register *)&ins->dst[i].reg, instance_id);
|
|
|
|
}
|
|
|
|
|
2023-07-20 19:52:25 -07:00
|
|
|
static const struct vkd3d_shader_varying_map *find_varying_map(
|
2023-09-19 14:04:53 -07:00
|
|
|
const struct vkd3d_shader_varying_map_info *varying_map, unsigned int signature_idx)
|
2023-07-20 19:52:25 -07:00
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
2023-09-19 14:04:53 -07:00
|
|
|
for (i = 0; i < varying_map->varying_count; ++i)
|
2023-07-20 19:52:25 -07:00
|
|
|
{
|
2023-09-19 14:04:53 -07:00
|
|
|
if (varying_map->varying_map[i].output_signature_index == signature_idx)
|
|
|
|
return &varying_map->varying_map[i];
|
2023-07-20 19:52:25 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2024-03-12 13:02:13 -07:00
|
|
|
static enum vkd3d_result vsir_program_remap_output_signature(struct vsir_program *program,
|
|
|
|
const struct vkd3d_shader_compile_info *compile_info, struct vkd3d_shader_message_context *message_context)
|
2023-07-20 19:52:25 -07:00
|
|
|
{
|
2024-03-12 13:02:13 -07:00
|
|
|
const struct vkd3d_shader_location location = {.source_name = compile_info->source_name};
|
|
|
|
struct shader_signature *signature = &program->output_signature;
|
2023-09-19 14:04:53 -07:00
|
|
|
const struct vkd3d_shader_varying_map_info *varying_map;
|
2023-07-20 19:52:25 -07:00
|
|
|
unsigned int i;
|
|
|
|
|
2023-09-19 14:04:53 -07:00
|
|
|
if (!(varying_map = vkd3d_find_struct(compile_info->next, VARYING_MAP_INFO)))
|
2023-07-20 19:52:25 -07:00
|
|
|
return VKD3D_OK;
|
|
|
|
|
|
|
|
for (i = 0; i < signature->element_count; ++i)
|
|
|
|
{
|
2023-09-19 14:04:53 -07:00
|
|
|
const struct vkd3d_shader_varying_map *map = find_varying_map(varying_map, i);
|
2023-07-20 19:52:25 -07:00
|
|
|
struct signature_element *e = &signature->elements[i];
|
|
|
|
|
|
|
|
if (map)
|
|
|
|
{
|
|
|
|
unsigned int input_mask = map->input_mask;
|
|
|
|
|
|
|
|
e->target_location = map->input_register_index;
|
|
|
|
|
|
|
|
/* It is illegal in Vulkan if the next shader uses the same varying
|
|
|
|
* location with a different mask. */
|
|
|
|
if (input_mask && input_mask != e->mask)
|
|
|
|
{
|
2024-03-12 13:02:13 -07:00
|
|
|
vkd3d_shader_error(message_context, &location, VKD3D_SHADER_ERROR_VSIR_NOT_IMPLEMENTED,
|
2023-07-20 19:52:25 -07:00
|
|
|
"Aborting due to not yet implemented feature: "
|
|
|
|
"Output mask %#x does not match input mask %#x.",
|
|
|
|
e->mask, input_mask);
|
|
|
|
return VKD3D_ERROR_NOT_IMPLEMENTED;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2023-07-31 10:46:55 -07:00
|
|
|
e->target_location = SIGNATURE_TARGET_LOCATION_UNUSED;
|
2023-07-20 19:52:25 -07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-09-19 14:04:53 -07:00
|
|
|
for (i = 0; i < varying_map->varying_count; ++i)
|
2023-07-20 19:52:25 -07:00
|
|
|
{
|
2023-09-19 14:04:53 -07:00
|
|
|
if (varying_map->varying_map[i].output_signature_index >= signature->element_count)
|
2023-07-20 19:52:25 -07:00
|
|
|
{
|
2024-03-12 13:02:13 -07:00
|
|
|
vkd3d_shader_error(message_context, &location, VKD3D_SHADER_ERROR_VSIR_NOT_IMPLEMENTED,
|
2023-07-20 19:52:25 -07:00
|
|
|
"Aborting due to not yet implemented feature: "
|
|
|
|
"The next stage consumes varyings not written by this stage.");
|
|
|
|
return VKD3D_ERROR_NOT_IMPLEMENTED;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
struct hull_flattener
|
2023-04-04 00:55:05 -07:00
|
|
|
{
|
2023-05-05 00:23:56 -07:00
|
|
|
struct vkd3d_shader_instruction_array instructions;
|
|
|
|
|
|
|
|
unsigned int instance_count;
|
|
|
|
unsigned int phase_body_idx;
|
|
|
|
enum vkd3d_shader_opcode phase;
|
2023-09-05 02:31:31 -07:00
|
|
|
struct vkd3d_shader_location last_ret_location;
|
2023-05-05 00:23:56 -07:00
|
|
|
};
|
2023-04-04 00:55:05 -07:00
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
static bool flattener_is_in_fork_or_join_phase(const struct hull_flattener *flattener)
|
2023-02-28 17:59:28 -08:00
|
|
|
{
|
2023-05-05 00:23:56 -07:00
|
|
|
return flattener->phase == VKD3DSIH_HS_FORK_PHASE || flattener->phase == VKD3DSIH_HS_JOIN_PHASE;
|
2023-02-28 17:59:28 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
struct shader_phase_location
|
|
|
|
{
|
|
|
|
unsigned int index;
|
|
|
|
unsigned int instance_count;
|
|
|
|
unsigned int instruction_count;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct shader_phase_location_array
|
|
|
|
{
|
|
|
|
/* Unlikely worst case: one phase for each component of each output register. */
|
|
|
|
struct shader_phase_location locations[MAX_REG_OUTPUT * VKD3D_VEC4_SIZE];
|
|
|
|
unsigned int count;
|
|
|
|
};
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
static void flattener_eliminate_phase_related_dcls(struct hull_flattener *normaliser,
|
2023-02-28 17:59:28 -08:00
|
|
|
unsigned int index, struct shader_phase_location_array *locations)
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction *ins = &normaliser->instructions.elements[index];
|
|
|
|
struct shader_phase_location *loc;
|
|
|
|
bool b;
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
if (ins->opcode == VKD3DSIH_HS_FORK_PHASE || ins->opcode == VKD3DSIH_HS_JOIN_PHASE)
|
2023-02-28 17:59:28 -08:00
|
|
|
{
|
2023-05-05 00:23:56 -07:00
|
|
|
b = flattener_is_in_fork_or_join_phase(normaliser);
|
2023-02-28 17:59:28 -08:00
|
|
|
/* Reset the phase info. */
|
|
|
|
normaliser->phase_body_idx = ~0u;
|
2024-04-24 11:34:42 -07:00
|
|
|
normaliser->phase = ins->opcode;
|
2023-02-28 17:59:28 -08:00
|
|
|
normaliser->instance_count = 1;
|
|
|
|
/* Leave the first occurrence and delete the rest. */
|
|
|
|
if (b)
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
|
|
|
return;
|
|
|
|
}
|
2024-04-24 11:34:42 -07:00
|
|
|
else if (ins->opcode == VKD3DSIH_DCL_HS_FORK_PHASE_INSTANCE_COUNT
|
|
|
|
|| ins->opcode == VKD3DSIH_DCL_HS_JOIN_PHASE_INSTANCE_COUNT)
|
2023-02-28 17:59:28 -08:00
|
|
|
{
|
|
|
|
normaliser->instance_count = ins->declaration.count + !ins->declaration.count;
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
|
|
|
return;
|
|
|
|
}
|
2024-04-24 11:34:42 -07:00
|
|
|
else if (ins->opcode == VKD3DSIH_DCL_INPUT && shader_register_is_phase_instance_id(
|
2023-02-28 17:59:28 -08:00
|
|
|
&ins->declaration.dst.reg))
|
|
|
|
{
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2024-01-22 11:32:55 -08:00
|
|
|
if (normaliser->phase == VKD3DSIH_INVALID || vsir_instruction_is_dcl(ins))
|
2023-02-28 17:59:28 -08:00
|
|
|
return;
|
|
|
|
|
|
|
|
if (normaliser->phase_body_idx == ~0u)
|
|
|
|
normaliser->phase_body_idx = index;
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
if (ins->opcode == VKD3DSIH_RET)
|
2023-02-28 17:59:28 -08:00
|
|
|
{
|
2023-09-05 02:31:31 -07:00
|
|
|
normaliser->last_ret_location = ins->location;
|
2023-02-28 17:59:28 -08:00
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
|
|
|
if (locations->count >= ARRAY_SIZE(locations->locations))
|
|
|
|
{
|
|
|
|
FIXME("Insufficient space for phase location.\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
loc = &locations->locations[locations->count++];
|
|
|
|
loc->index = normaliser->phase_body_idx;
|
|
|
|
loc->instance_count = normaliser->instance_count;
|
|
|
|
loc->instruction_count = index - normaliser->phase_body_idx;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
static enum vkd3d_result flattener_flatten_phases(struct hull_flattener *normaliser,
|
2023-02-28 17:59:28 -08:00
|
|
|
struct shader_phase_location_array *locations)
|
|
|
|
{
|
|
|
|
struct shader_phase_location *loc;
|
|
|
|
unsigned int i, j, k, end, count;
|
|
|
|
|
|
|
|
for (i = 0, count = 0; i < locations->count; ++i)
|
|
|
|
count += (locations->locations[i].instance_count - 1) * locations->locations[i].instruction_count;
|
|
|
|
|
|
|
|
if (!shader_instruction_array_reserve(&normaliser->instructions, normaliser->instructions.count + count))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
end = normaliser->instructions.count;
|
|
|
|
normaliser->instructions.count += count;
|
|
|
|
|
|
|
|
for (i = locations->count; i > 0; --i)
|
|
|
|
{
|
|
|
|
loc = &locations->locations[i - 1];
|
|
|
|
j = loc->index + loc->instruction_count;
|
|
|
|
memmove(&normaliser->instructions.elements[j + count], &normaliser->instructions.elements[j],
|
|
|
|
(end - j) * sizeof(*normaliser->instructions.elements));
|
|
|
|
end = j;
|
|
|
|
count -= (loc->instance_count - 1) * loc->instruction_count;
|
|
|
|
loc->index += count;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0, count = 0; i < locations->count; ++i)
|
|
|
|
{
|
|
|
|
loc = &locations->locations[i];
|
|
|
|
/* Make a copy of the non-dcl instructions for each instance. */
|
|
|
|
for (j = 1; j < loc->instance_count; ++j)
|
|
|
|
{
|
|
|
|
for (k = 0; k < loc->instruction_count; ++k)
|
|
|
|
{
|
|
|
|
if (!shader_instruction_array_clone_instruction(&normaliser->instructions,
|
|
|
|
loc->index + loc->instruction_count * j + k, loc->index + k))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/* Replace each reference to the instance id with a constant instance id. */
|
|
|
|
for (j = 0; j < loc->instance_count; ++j)
|
|
|
|
{
|
|
|
|
for (k = 0; k < loc->instruction_count; ++k)
|
|
|
|
shader_instruction_eliminate_phase_instance_id(
|
|
|
|
&normaliser->instructions.elements[loc->index + loc->instruction_count * j + k], j);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2023-08-29 10:20:44 -07:00
|
|
|
void vsir_register_init(struct vkd3d_shader_register *reg, enum vkd3d_shader_register_type reg_type,
|
2023-05-03 00:10:37 -07:00
|
|
|
enum vkd3d_data_type data_type, unsigned int idx_count)
|
2023-04-04 00:55:05 -07:00
|
|
|
{
|
|
|
|
reg->type = reg_type;
|
|
|
|
reg->precision = VKD3D_SHADER_REGISTER_PRECISION_DEFAULT;
|
|
|
|
reg->non_uniform = false;
|
|
|
|
reg->data_type = data_type;
|
|
|
|
reg->idx[0].offset = ~0u;
|
|
|
|
reg->idx[0].rel_addr = NULL;
|
2023-11-12 22:52:45 -08:00
|
|
|
reg->idx[0].is_in_bounds = false;
|
2023-04-04 00:55:05 -07:00
|
|
|
reg->idx[1].offset = ~0u;
|
|
|
|
reg->idx[1].rel_addr = NULL;
|
2023-11-12 22:52:45 -08:00
|
|
|
reg->idx[1].is_in_bounds = false;
|
2023-04-04 00:55:05 -07:00
|
|
|
reg->idx[2].offset = ~0u;
|
|
|
|
reg->idx[2].rel_addr = NULL;
|
2023-11-12 22:52:45 -08:00
|
|
|
reg->idx[2].is_in_bounds = false;
|
2023-05-03 00:10:37 -07:00
|
|
|
reg->idx_count = idx_count;
|
2023-08-14 10:21:41 -07:00
|
|
|
reg->dimension = VSIR_DIMENSION_SCALAR;
|
2023-11-21 18:42:23 -08:00
|
|
|
reg->alignment = 0;
|
2023-04-04 00:55:05 -07:00
|
|
|
}
|
|
|
|
|
2023-11-23 19:14:44 -08:00
|
|
|
void vsir_src_param_init(struct vkd3d_shader_src_param *param, enum vkd3d_shader_register_type reg_type,
|
2024-01-04 06:55:51 -08:00
|
|
|
enum vkd3d_data_type data_type, unsigned int idx_count)
|
|
|
|
{
|
|
|
|
vsir_register_init(¶m->reg, reg_type, data_type, idx_count);
|
|
|
|
param->swizzle = 0;
|
|
|
|
param->modifiers = VKD3DSPSM_NONE;
|
|
|
|
}
|
|
|
|
|
2024-01-03 07:18:50 -08:00
|
|
|
void vsir_dst_param_init(struct vkd3d_shader_dst_param *param, enum vkd3d_shader_register_type reg_type,
|
|
|
|
enum vkd3d_data_type data_type, unsigned int idx_count)
|
|
|
|
{
|
|
|
|
vsir_register_init(¶m->reg, reg_type, data_type, idx_count);
|
|
|
|
param->write_mask = VKD3DSP_WRITEMASK_0;
|
|
|
|
param->modifiers = VKD3DSPDM_NONE;
|
|
|
|
param->shift = 0;
|
|
|
|
}
|
|
|
|
|
2024-01-11 19:20:32 -08:00
|
|
|
void vsir_src_param_init_label(struct vkd3d_shader_src_param *param, unsigned int label_id)
|
2024-01-04 06:55:51 -08:00
|
|
|
{
|
2024-03-14 14:40:12 -07:00
|
|
|
vsir_src_param_init(param, VKD3DSPR_LABEL, VKD3D_DATA_UNUSED, 1);
|
2024-01-04 06:55:51 -08:00
|
|
|
param->reg.dimension = VSIR_DIMENSION_NONE;
|
|
|
|
param->reg.idx[0].offset = label_id;
|
|
|
|
}
|
|
|
|
|
2024-01-03 07:18:50 -08:00
|
|
|
static void src_param_init_ssa_bool(struct vkd3d_shader_src_param *src, unsigned int idx)
|
|
|
|
{
|
|
|
|
vsir_src_param_init(src, VKD3DSPR_SSA, VKD3D_DATA_BOOL, 1);
|
|
|
|
src->reg.idx[0].offset = idx;
|
|
|
|
}
|
|
|
|
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
static void src_param_init_temp_bool(struct vkd3d_shader_src_param *src, unsigned int idx)
|
|
|
|
{
|
|
|
|
vsir_src_param_init(src, VKD3DSPR_TEMP, VKD3D_DATA_BOOL, 1);
|
|
|
|
src->reg.idx[0].offset = idx;
|
|
|
|
}
|
|
|
|
|
2024-01-03 07:18:50 -08:00
|
|
|
static void dst_param_init_ssa_bool(struct vkd3d_shader_dst_param *dst, unsigned int idx)
|
|
|
|
{
|
|
|
|
vsir_dst_param_init(dst, VKD3DSPR_SSA, VKD3D_DATA_BOOL, 1);
|
|
|
|
dst->reg.idx[0].offset = idx;
|
|
|
|
}
|
|
|
|
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
static void dst_param_init_temp_bool(struct vkd3d_shader_dst_param *dst, unsigned int idx)
|
|
|
|
{
|
|
|
|
vsir_dst_param_init(dst, VKD3DSPR_TEMP, VKD3D_DATA_BOOL, 1);
|
|
|
|
dst->reg.idx[0].offset = idx;
|
|
|
|
}
|
|
|
|
|
2024-01-16 14:56:43 -08:00
|
|
|
static void dst_param_init_temp_uint(struct vkd3d_shader_dst_param *dst, unsigned int idx)
|
|
|
|
{
|
|
|
|
vsir_dst_param_init(dst, VKD3DSPR_TEMP, VKD3D_DATA_UINT, 1);
|
|
|
|
dst->reg.idx[0].offset = idx;
|
|
|
|
dst->write_mask = VKD3DSP_WRITEMASK_0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void src_param_init_temp_uint(struct vkd3d_shader_src_param *src, unsigned int idx)
|
|
|
|
{
|
|
|
|
vsir_src_param_init(src, VKD3DSPR_TEMP, VKD3D_DATA_UINT, 1);
|
|
|
|
src->reg.idx[0].offset = idx;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void src_param_init_const_uint(struct vkd3d_shader_src_param *src, uint32_t value)
|
|
|
|
{
|
|
|
|
vsir_src_param_init(src, VKD3DSPR_IMMCONST, VKD3D_DATA_UINT, 0);
|
|
|
|
src->reg.u.immconst_u32[0] = value;
|
|
|
|
}
|
|
|
|
|
2023-09-05 02:31:31 -07:00
|
|
|
void vsir_instruction_init(struct vkd3d_shader_instruction *ins, const struct vkd3d_shader_location *location,
|
2024-04-24 11:34:42 -07:00
|
|
|
enum vkd3d_shader_opcode opcode)
|
2023-02-28 17:59:28 -08:00
|
|
|
{
|
|
|
|
memset(ins, 0, sizeof(*ins));
|
2023-09-05 02:31:31 -07:00
|
|
|
ins->location = *location;
|
2024-04-24 11:34:42 -07:00
|
|
|
ins->opcode = opcode;
|
2023-02-28 17:59:28 -08:00
|
|
|
}
|
|
|
|
|
2024-01-24 05:26:35 -08:00
|
|
|
static bool vsir_instruction_init_label(struct vkd3d_shader_instruction *ins,
|
|
|
|
const struct vkd3d_shader_location *location, unsigned int label_id, struct vsir_program *program)
|
2024-01-04 06:55:51 -08:00
|
|
|
{
|
|
|
|
struct vkd3d_shader_src_param *src_param;
|
|
|
|
|
2024-01-24 05:26:35 -08:00
|
|
|
if (!(src_param = vsir_program_get_src_params(program, 1)))
|
2024-01-04 06:55:51 -08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
vsir_src_param_init_label(src_param, label_id);
|
|
|
|
|
|
|
|
vsir_instruction_init(ins, location, VKD3DSIH_LABEL);
|
|
|
|
ins->src = src_param;
|
|
|
|
ins->src_count = 1;
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2023-07-02 18:23:07 -07:00
|
|
|
static enum vkd3d_result instruction_array_flatten_hull_shader_phases(struct vkd3d_shader_instruction_array *src_instructions)
|
2023-02-28 17:59:28 -08:00
|
|
|
{
|
2023-05-05 00:23:56 -07:00
|
|
|
struct hull_flattener flattener = {*src_instructions};
|
|
|
|
struct vkd3d_shader_instruction_array *instructions;
|
2023-02-28 17:59:28 -08:00
|
|
|
struct shader_phase_location_array locations;
|
|
|
|
enum vkd3d_result result = VKD3D_OK;
|
|
|
|
unsigned int i;
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
instructions = &flattener.instructions;
|
|
|
|
|
|
|
|
flattener.phase = VKD3DSIH_INVALID;
|
2023-02-28 17:59:28 -08:00
|
|
|
for (i = 0, locations.count = 0; i < instructions->count; ++i)
|
2023-05-05 00:23:56 -07:00
|
|
|
flattener_eliminate_phase_related_dcls(&flattener, i, &locations);
|
2023-02-28 17:59:28 -08:00
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
if ((result = flattener_flatten_phases(&flattener, &locations)) < 0)
|
2023-02-28 17:59:28 -08:00
|
|
|
return result;
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
if (flattener.phase != VKD3DSIH_INVALID)
|
2023-02-28 17:59:28 -08:00
|
|
|
{
|
2023-05-05 00:23:56 -07:00
|
|
|
if (!shader_instruction_array_reserve(&flattener.instructions, flattener.instructions.count + 1))
|
2023-02-28 17:59:28 -08:00
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
2023-09-05 02:31:31 -07:00
|
|
|
vsir_instruction_init(&instructions->elements[instructions->count++], &flattener.last_ret_location, VKD3DSIH_RET);
|
2023-02-28 17:59:28 -08:00
|
|
|
}
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
*src_instructions = flattener.instructions;
|
2023-02-28 17:59:28 -08:00
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
struct control_point_normaliser
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction_array instructions;
|
|
|
|
enum vkd3d_shader_opcode phase;
|
|
|
|
struct vkd3d_shader_src_param *outpointid_param;
|
|
|
|
};
|
|
|
|
|
|
|
|
static bool control_point_normaliser_is_in_control_point_phase(const struct control_point_normaliser *normaliser)
|
|
|
|
{
|
|
|
|
return normaliser->phase == VKD3DSIH_HS_CONTROL_POINT_PHASE;
|
|
|
|
}
|
|
|
|
|
2023-06-14 21:31:16 -07:00
|
|
|
struct vkd3d_shader_src_param *instruction_array_create_outpointid_param(
|
2023-05-05 00:23:56 -07:00
|
|
|
struct vkd3d_shader_instruction_array *instructions)
|
2023-04-04 00:55:05 -07:00
|
|
|
{
|
|
|
|
struct vkd3d_shader_src_param *rel_addr;
|
|
|
|
|
2023-06-14 21:31:16 -07:00
|
|
|
if (instructions->outpointid_param)
|
|
|
|
return instructions->outpointid_param;
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
if (!(rel_addr = shader_src_param_allocator_get(&instructions->src_params, 1)))
|
2023-04-04 00:55:05 -07:00
|
|
|
return NULL;
|
|
|
|
|
2023-08-29 10:20:44 -07:00
|
|
|
vsir_register_init(&rel_addr->reg, VKD3DSPR_OUTPOINTID, VKD3D_DATA_UINT, 0);
|
2023-04-04 00:55:05 -07:00
|
|
|
rel_addr->swizzle = 0;
|
|
|
|
rel_addr->modifiers = 0;
|
|
|
|
|
2023-06-14 21:31:16 -07:00
|
|
|
instructions->outpointid_param = rel_addr;
|
2023-04-04 00:55:05 -07:00
|
|
|
return rel_addr;
|
|
|
|
}
|
|
|
|
|
2023-05-03 00:10:37 -07:00
|
|
|
static void shader_dst_param_normalise_outpointid(struct vkd3d_shader_dst_param *dst_param,
|
2023-05-05 00:23:56 -07:00
|
|
|
struct control_point_normaliser *normaliser)
|
2023-04-04 00:55:05 -07:00
|
|
|
{
|
|
|
|
struct vkd3d_shader_register *reg = &dst_param->reg;
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
if (control_point_normaliser_is_in_control_point_phase(normaliser) && reg->type == VKD3DSPR_OUTPUT)
|
2023-04-04 00:55:05 -07:00
|
|
|
{
|
2023-05-03 00:10:37 -07:00
|
|
|
/* The TPF reader validates idx_count. */
|
|
|
|
assert(reg->idx_count == 1);
|
2023-04-04 00:55:05 -07:00
|
|
|
reg->idx[1] = reg->idx[0];
|
|
|
|
/* The control point id param is implicit here. Avoid later complications by inserting it. */
|
|
|
|
reg->idx[0].offset = 0;
|
|
|
|
reg->idx[0].rel_addr = normaliser->outpointid_param;
|
2023-05-03 00:10:37 -07:00
|
|
|
++reg->idx_count;
|
2023-04-04 00:55:05 -07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-05-03 00:10:37 -07:00
|
|
|
static void shader_dst_param_io_init(struct vkd3d_shader_dst_param *param, const struct signature_element *e,
|
|
|
|
enum vkd3d_shader_register_type reg_type, unsigned int idx_count)
|
2023-04-04 00:28:48 -07:00
|
|
|
{
|
|
|
|
param->write_mask = e->mask;
|
|
|
|
param->modifiers = 0;
|
|
|
|
param->shift = 0;
|
2023-08-29 10:20:44 -07:00
|
|
|
vsir_register_init(¶m->reg, reg_type, vkd3d_data_type_from_component_type(e->component_type), idx_count);
|
2023-04-04 00:28:48 -07:00
|
|
|
}
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
static enum vkd3d_result control_point_normaliser_emit_hs_input(struct control_point_normaliser *normaliser,
|
2023-09-05 02:31:31 -07:00
|
|
|
const struct shader_signature *s, unsigned int input_control_point_count, unsigned int dst,
|
|
|
|
const struct vkd3d_shader_location *location)
|
2023-04-04 00:28:48 -07:00
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
struct vkd3d_shader_dst_param *param;
|
2023-04-03 23:22:57 -07:00
|
|
|
const struct signature_element *e;
|
2023-04-04 00:28:48 -07:00
|
|
|
unsigned int i, count;
|
|
|
|
|
|
|
|
for (i = 0, count = 1; i < s->element_count; ++i)
|
|
|
|
count += !!s->elements[i].used_mask;
|
|
|
|
|
|
|
|
if (!shader_instruction_array_reserve(&normaliser->instructions, normaliser->instructions.count + count))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
memmove(&normaliser->instructions.elements[dst + count], &normaliser->instructions.elements[dst],
|
|
|
|
(normaliser->instructions.count - dst) * sizeof(*normaliser->instructions.elements));
|
|
|
|
normaliser->instructions.count += count;
|
|
|
|
|
|
|
|
ins = &normaliser->instructions.elements[dst];
|
2023-09-05 02:31:31 -07:00
|
|
|
vsir_instruction_init(ins, location, VKD3DSIH_HS_CONTROL_POINT_PHASE);
|
2023-04-04 00:28:48 -07:00
|
|
|
ins->flags = 1;
|
|
|
|
++ins;
|
|
|
|
|
|
|
|
for (i = 0; i < s->element_count; ++i)
|
|
|
|
{
|
|
|
|
e = &s->elements[i];
|
|
|
|
if (!e->used_mask)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (e->sysval_semantic != VKD3D_SHADER_SV_NONE)
|
|
|
|
{
|
2023-09-05 02:31:31 -07:00
|
|
|
vsir_instruction_init(ins, location, VKD3DSIH_DCL_INPUT_SIV);
|
2023-04-04 00:28:48 -07:00
|
|
|
param = &ins->declaration.register_semantic.reg;
|
|
|
|
ins->declaration.register_semantic.sysval_semantic = vkd3d_siv_from_sysval(e->sysval_semantic);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2023-09-05 02:31:31 -07:00
|
|
|
vsir_instruction_init(ins, location, VKD3DSIH_DCL_INPUT);
|
2023-04-04 00:28:48 -07:00
|
|
|
param = &ins->declaration.dst;
|
|
|
|
}
|
|
|
|
|
2023-05-03 00:10:37 -07:00
|
|
|
shader_dst_param_io_init(param, e, VKD3DSPR_INPUT, 2);
|
2023-04-04 00:28:48 -07:00
|
|
|
param->reg.idx[0].offset = input_control_point_count;
|
2023-10-24 14:55:16 -07:00
|
|
|
param->reg.idx[1].offset = e->register_index;
|
|
|
|
param->write_mask = e->mask;
|
2023-04-04 00:28:48 -07:00
|
|
|
|
|
|
|
++ins;
|
|
|
|
}
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2023-07-02 18:23:07 -07:00
|
|
|
static enum vkd3d_result instruction_array_normalise_hull_shader_control_point_io(
|
2023-05-05 00:23:56 -07:00
|
|
|
struct vkd3d_shader_instruction_array *src_instructions, const struct shader_signature *input_signature)
|
2023-04-04 00:55:05 -07:00
|
|
|
{
|
2023-05-05 00:23:56 -07:00
|
|
|
struct vkd3d_shader_instruction_array *instructions;
|
|
|
|
struct control_point_normaliser normaliser;
|
2023-04-04 00:28:48 -07:00
|
|
|
unsigned int input_control_point_count;
|
2023-12-07 06:34:10 -08:00
|
|
|
struct vkd3d_shader_location location;
|
2023-04-04 00:55:05 -07:00
|
|
|
struct vkd3d_shader_instruction *ins;
|
2023-05-05 00:23:56 -07:00
|
|
|
enum vkd3d_result ret;
|
2023-04-04 00:55:05 -07:00
|
|
|
unsigned int i, j;
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
if (!(normaliser.outpointid_param = instruction_array_create_outpointid_param(src_instructions)))
|
2023-04-04 00:55:05 -07:00
|
|
|
{
|
|
|
|
ERR("Failed to allocate src param.\n");
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
}
|
2023-05-05 00:23:56 -07:00
|
|
|
normaliser.instructions = *src_instructions;
|
|
|
|
instructions = &normaliser.instructions;
|
|
|
|
normaliser.phase = VKD3DSIH_INVALID;
|
2023-04-04 00:55:05 -07:00
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
for (i = 0; i < normaliser.instructions.count; ++i)
|
2023-04-04 00:55:05 -07:00
|
|
|
{
|
|
|
|
ins = &instructions->elements[i];
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (ins->opcode)
|
2023-04-04 00:55:05 -07:00
|
|
|
{
|
|
|
|
case VKD3DSIH_HS_CONTROL_POINT_PHASE:
|
|
|
|
case VKD3DSIH_HS_FORK_PHASE:
|
|
|
|
case VKD3DSIH_HS_JOIN_PHASE:
|
2024-04-24 11:34:42 -07:00
|
|
|
normaliser.phase = ins->opcode;
|
2023-04-04 00:55:05 -07:00
|
|
|
break;
|
|
|
|
default:
|
2024-01-22 11:32:55 -08:00
|
|
|
if (vsir_instruction_is_dcl(ins))
|
2023-04-04 00:55:05 -07:00
|
|
|
break;
|
|
|
|
for (j = 0; j < ins->dst_count; ++j)
|
2024-01-11 02:19:39 -08:00
|
|
|
shader_dst_param_normalise_outpointid(&ins->dst[j], &normaliser);
|
2023-04-04 00:55:05 -07:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
normaliser.phase = VKD3DSIH_INVALID;
|
2023-04-04 00:28:48 -07:00
|
|
|
input_control_point_count = 1;
|
|
|
|
|
|
|
|
for (i = 0; i < instructions->count; ++i)
|
|
|
|
{
|
|
|
|
ins = &instructions->elements[i];
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (ins->opcode)
|
2023-04-04 00:28:48 -07:00
|
|
|
{
|
|
|
|
case VKD3DSIH_DCL_INPUT_CONTROL_POINT_COUNT:
|
|
|
|
input_control_point_count = ins->declaration.count;
|
|
|
|
break;
|
|
|
|
case VKD3DSIH_HS_CONTROL_POINT_PHASE:
|
2023-05-05 00:23:56 -07:00
|
|
|
*src_instructions = normaliser.instructions;
|
2023-04-04 00:28:48 -07:00
|
|
|
return VKD3D_OK;
|
|
|
|
case VKD3DSIH_HS_FORK_PHASE:
|
|
|
|
case VKD3DSIH_HS_JOIN_PHASE:
|
2023-12-07 06:34:10 -08:00
|
|
|
/* ins may be relocated if the instruction array expands. */
|
|
|
|
location = ins->location;
|
2023-05-05 00:23:56 -07:00
|
|
|
ret = control_point_normaliser_emit_hs_input(&normaliser, input_signature,
|
2023-12-07 06:34:10 -08:00
|
|
|
input_control_point_count, i, &location);
|
2023-05-05 00:23:56 -07:00
|
|
|
*src_instructions = normaliser.instructions;
|
|
|
|
return ret;
|
2023-04-04 00:28:48 -07:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-05-05 00:23:56 -07:00
|
|
|
*src_instructions = normaliser.instructions;
|
2023-04-04 00:55:05 -07:00
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
2023-04-25 22:27:08 -07:00
|
|
|
|
|
|
|
struct io_normaliser
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction_array instructions;
|
|
|
|
enum vkd3d_shader_type shader_type;
|
2023-10-27 10:44:08 -07:00
|
|
|
uint8_t major;
|
2023-04-25 22:27:08 -07:00
|
|
|
struct shader_signature *input_signature;
|
|
|
|
struct shader_signature *output_signature;
|
|
|
|
struct shader_signature *patch_constant_signature;
|
|
|
|
|
|
|
|
unsigned int instance_count;
|
|
|
|
unsigned int phase_body_idx;
|
|
|
|
enum vkd3d_shader_opcode phase;
|
|
|
|
unsigned int output_control_point_count;
|
|
|
|
|
|
|
|
struct vkd3d_shader_src_param *outpointid_param;
|
|
|
|
|
|
|
|
struct vkd3d_shader_dst_param *input_dcl_params[MAX_REG_OUTPUT];
|
|
|
|
struct vkd3d_shader_dst_param *output_dcl_params[MAX_REG_OUTPUT];
|
|
|
|
struct vkd3d_shader_dst_param *pc_dcl_params[MAX_REG_OUTPUT];
|
|
|
|
uint8_t input_range_map[MAX_REG_OUTPUT][VKD3D_VEC4_SIZE];
|
|
|
|
uint8_t output_range_map[MAX_REG_OUTPUT][VKD3D_VEC4_SIZE];
|
|
|
|
uint8_t pc_range_map[MAX_REG_OUTPUT][VKD3D_VEC4_SIZE];
|
2023-07-16 21:34:04 -07:00
|
|
|
|
|
|
|
bool use_vocp;
|
2023-04-25 22:27:08 -07:00
|
|
|
};
|
|
|
|
|
|
|
|
static bool io_normaliser_is_in_fork_or_join_phase(const struct io_normaliser *normaliser)
|
|
|
|
{
|
|
|
|
return normaliser->phase == VKD3DSIH_HS_FORK_PHASE || normaliser->phase == VKD3DSIH_HS_JOIN_PHASE;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool io_normaliser_is_in_control_point_phase(const struct io_normaliser *normaliser)
|
|
|
|
{
|
|
|
|
return normaliser->phase == VKD3DSIH_HS_CONTROL_POINT_PHASE;
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned int shader_signature_find_element_for_reg(const struct shader_signature *signature,
|
|
|
|
unsigned int reg_idx, unsigned int write_mask)
|
|
|
|
{
|
2024-01-03 22:05:15 -08:00
|
|
|
unsigned int i, base_write_mask;
|
2023-04-25 22:27:08 -07:00
|
|
|
|
|
|
|
for (i = 0; i < signature->element_count; ++i)
|
|
|
|
{
|
|
|
|
struct signature_element *e = &signature->elements[i];
|
|
|
|
if (e->register_index <= reg_idx && e->register_index + e->register_count > reg_idx
|
|
|
|
&& (e->mask & write_mask) == write_mask)
|
|
|
|
{
|
|
|
|
return i;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-01-03 22:05:15 -08:00
|
|
|
/* Validated in the TPF reader, but failure in signature_element_range_expand_mask()
|
|
|
|
* can land us here on an unmatched vector mask. */
|
|
|
|
FIXME("Failed to find signature element for register index %u, mask %#x; using scalar mask.\n",
|
|
|
|
reg_idx, write_mask);
|
|
|
|
base_write_mask = 1u << vsir_write_mask_get_component_idx(write_mask);
|
|
|
|
if (base_write_mask != write_mask)
|
|
|
|
return shader_signature_find_element_for_reg(signature, reg_idx, base_write_mask);
|
|
|
|
|
2023-04-25 22:27:08 -07:00
|
|
|
vkd3d_unreachable();
|
|
|
|
}
|
|
|
|
|
2023-10-19 19:34:41 -07:00
|
|
|
struct signature_element *vsir_signature_find_element_for_reg(const struct shader_signature *signature,
|
|
|
|
unsigned int reg_idx, unsigned int write_mask)
|
|
|
|
{
|
|
|
|
return &signature->elements[shader_signature_find_element_for_reg(signature, reg_idx, write_mask)];
|
|
|
|
}
|
|
|
|
|
2023-04-25 22:27:08 -07:00
|
|
|
static unsigned int range_map_get_register_count(uint8_t range_map[][VKD3D_VEC4_SIZE],
|
2023-12-05 13:19:07 -08:00
|
|
|
unsigned int register_idx, uint32_t write_mask)
|
2023-04-25 22:27:08 -07:00
|
|
|
{
|
2023-12-05 13:19:07 -08:00
|
|
|
return range_map[register_idx][vsir_write_mask_get_component_idx(write_mask)];
|
2023-04-25 22:27:08 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void range_map_set_register_range(uint8_t range_map[][VKD3D_VEC4_SIZE], unsigned int register_idx,
|
2023-12-05 13:19:07 -08:00
|
|
|
unsigned int register_count, uint32_t write_mask, bool is_dcl_indexrange)
|
2023-04-25 22:27:08 -07:00
|
|
|
{
|
|
|
|
unsigned int i, j, r, c, component_idx, component_count;
|
|
|
|
|
|
|
|
assert(write_mask <= VKD3DSP_WRITEMASK_ALL);
|
2023-12-05 13:19:07 -08:00
|
|
|
component_idx = vsir_write_mask_get_component_idx(write_mask);
|
2023-12-05 13:35:37 -08:00
|
|
|
component_count = vsir_write_mask_component_count(write_mask);
|
2023-04-25 22:27:08 -07:00
|
|
|
|
|
|
|
assert(register_idx < MAX_REG_OUTPUT && MAX_REG_OUTPUT - register_idx >= register_count);
|
|
|
|
|
|
|
|
if (range_map[register_idx][component_idx] > register_count && is_dcl_indexrange)
|
|
|
|
{
|
|
|
|
/* Validated in the TPF reader. */
|
|
|
|
assert(range_map[register_idx][component_idx] != UINT8_MAX);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
if (range_map[register_idx][component_idx] == register_count)
|
|
|
|
{
|
|
|
|
/* Already done. This happens when fxc splits a register declaration by
|
|
|
|
* component(s). The dcl_indexrange instructions are split too. */
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
range_map[register_idx][component_idx] = register_count;
|
|
|
|
|
|
|
|
for (i = 0; i < register_count; ++i)
|
|
|
|
{
|
|
|
|
r = register_idx + i;
|
|
|
|
for (j = !i; j < component_count; ++j)
|
|
|
|
{
|
|
|
|
c = component_idx + j;
|
|
|
|
/* A synthetic patch constant range which overlaps an existing range can start upstream of it
|
|
|
|
* for fork/join phase instancing, but ranges declared by dcl_indexrange should not overlap.
|
|
|
|
* The latter is validated in the TPF reader. */
|
|
|
|
assert(!range_map[r][c] || !is_dcl_indexrange);
|
|
|
|
range_map[r][c] = UINT8_MAX;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void io_normaliser_add_index_range(struct io_normaliser *normaliser,
|
|
|
|
const struct vkd3d_shader_instruction *ins)
|
|
|
|
{
|
|
|
|
const struct vkd3d_shader_index_range *range = &ins->declaration.index_range;
|
|
|
|
const struct vkd3d_shader_register *reg = &range->dst.reg;
|
|
|
|
unsigned int reg_idx, write_mask, element_idx;
|
|
|
|
const struct shader_signature *signature;
|
|
|
|
uint8_t (*range_map)[VKD3D_VEC4_SIZE];
|
|
|
|
|
|
|
|
switch (reg->type)
|
|
|
|
{
|
|
|
|
case VKD3DSPR_INPUT:
|
|
|
|
case VKD3DSPR_INCONTROLPOINT:
|
|
|
|
range_map = normaliser->input_range_map;
|
|
|
|
signature = normaliser->input_signature;
|
|
|
|
break;
|
|
|
|
case VKD3DSPR_OUTCONTROLPOINT:
|
|
|
|
range_map = normaliser->output_range_map;
|
|
|
|
signature = normaliser->output_signature;
|
|
|
|
break;
|
|
|
|
case VKD3DSPR_OUTPUT:
|
|
|
|
if (!io_normaliser_is_in_fork_or_join_phase(normaliser))
|
|
|
|
{
|
|
|
|
range_map = normaliser->output_range_map;
|
|
|
|
signature = normaliser->output_signature;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
/* fall through */
|
|
|
|
case VKD3DSPR_PATCHCONST:
|
|
|
|
range_map = normaliser->pc_range_map;
|
|
|
|
signature = normaliser->patch_constant_signature;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
/* Validated in the TPF reader. */
|
|
|
|
vkd3d_unreachable();
|
|
|
|
}
|
|
|
|
|
|
|
|
reg_idx = reg->idx[reg->idx_count - 1].offset;
|
|
|
|
write_mask = range->dst.write_mask;
|
|
|
|
element_idx = shader_signature_find_element_for_reg(signature, reg_idx, write_mask);
|
|
|
|
range_map_set_register_range(range_map, reg_idx, range->register_count,
|
|
|
|
signature->elements[element_idx].mask, true);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int signature_element_mask_compare(const void *a, const void *b)
|
|
|
|
{
|
|
|
|
const struct signature_element *e = a, *f = b;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
return (ret = vkd3d_u32_compare(e->mask, f->mask)) ? ret : vkd3d_u32_compare(e->register_index, f->register_index);
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool sysval_semantics_should_merge(const struct signature_element *e, const struct signature_element *f)
|
|
|
|
{
|
|
|
|
if (e->sysval_semantic < VKD3D_SHADER_SV_TESS_FACTOR_QUADEDGE
|
|
|
|
|| e->sysval_semantic > VKD3D_SHADER_SV_TESS_FACTOR_LINEDEN)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return e->sysval_semantic == f->sysval_semantic
|
|
|
|
/* Line detail and density must be merged together to match the SPIR-V array.
|
|
|
|
* This deletes one of the two sysvals, but these are not used. */
|
|
|
|
|| (e->sysval_semantic == VKD3D_SHADER_SV_TESS_FACTOR_LINEDET
|
|
|
|
&& f->sysval_semantic == VKD3D_SHADER_SV_TESS_FACTOR_LINEDEN)
|
|
|
|
|| (e->sysval_semantic == VKD3D_SHADER_SV_TESS_FACTOR_LINEDEN
|
|
|
|
&& f->sysval_semantic == VKD3D_SHADER_SV_TESS_FACTOR_LINEDET);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Merge tess factor sysvals because they are an array in SPIR-V. */
|
|
|
|
static void shader_signature_map_patch_constant_index_ranges(struct shader_signature *s,
|
|
|
|
uint8_t range_map[][VKD3D_VEC4_SIZE])
|
|
|
|
{
|
|
|
|
struct signature_element *e, *f;
|
|
|
|
unsigned int i, j, register_count;
|
|
|
|
|
|
|
|
qsort(s->elements, s->element_count, sizeof(s->elements[0]), signature_element_mask_compare);
|
|
|
|
|
|
|
|
for (i = 0; i < s->element_count; i += register_count)
|
|
|
|
{
|
|
|
|
e = &s->elements[i];
|
|
|
|
register_count = 1;
|
|
|
|
|
|
|
|
if (!e->sysval_semantic)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
for (j = i + 1; j < s->element_count; ++j, ++register_count)
|
|
|
|
{
|
|
|
|
f = &s->elements[j];
|
|
|
|
if (f->register_index != e->register_index + register_count || !sysval_semantics_should_merge(e, f))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (register_count < 2)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
range_map_set_register_range(range_map, e->register_index, register_count, e->mask, false);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static int signature_element_register_compare(const void *a, const void *b)
|
|
|
|
{
|
|
|
|
const struct signature_element *e = a, *f = b;
|
|
|
|
|
|
|
|
return vkd3d_u32_compare(e->register_index, f->register_index);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int signature_element_index_compare(const void *a, const void *b)
|
|
|
|
{
|
|
|
|
const struct signature_element *e = a, *f = b;
|
|
|
|
|
|
|
|
return vkd3d_u32_compare(e->sort_index, f->sort_index);
|
|
|
|
}
|
|
|
|
|
2024-01-03 22:05:15 -08:00
|
|
|
static unsigned int signature_element_range_expand_mask(struct signature_element *e, unsigned int register_count,
|
|
|
|
uint8_t range_map[][VKD3D_VEC4_SIZE])
|
|
|
|
{
|
|
|
|
unsigned int i, j, component_idx, component_count, merged_write_mask = e->mask;
|
|
|
|
|
|
|
|
/* dcl_indexrange instructions can declare a subset of the full mask, and the masks of
|
|
|
|
* the elements within the range may differ. TPF's handling of arrayed inputs with
|
|
|
|
* dcl_indexrange is really just a hack. Here we create a mask which covers all element
|
|
|
|
* masks, and check for collisions with other ranges. */
|
|
|
|
|
|
|
|
for (i = 1; i < register_count; ++i)
|
|
|
|
merged_write_mask |= e[i].mask;
|
|
|
|
|
|
|
|
if (merged_write_mask == e->mask)
|
|
|
|
return merged_write_mask;
|
|
|
|
|
|
|
|
/* Reaching this point is very rare to begin with, and collisions are even rarer or
|
|
|
|
* impossible. If the latter shows up, the fallback in shader_signature_find_element_for_reg()
|
|
|
|
* may be sufficient. */
|
|
|
|
|
|
|
|
component_idx = vsir_write_mask_get_component_idx(e->mask);
|
|
|
|
component_count = vsir_write_mask_component_count(e->mask);
|
|
|
|
|
|
|
|
for (i = e->register_index; i < e->register_index + register_count; ++i)
|
|
|
|
{
|
|
|
|
for (j = 0; j < component_idx; ++j)
|
|
|
|
if (range_map[i][j])
|
|
|
|
break;
|
|
|
|
for (j = component_idx + component_count; j < VKD3D_VEC4_SIZE; ++j)
|
|
|
|
if (range_map[i][j])
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (i == register_count)
|
|
|
|
{
|
|
|
|
WARN("Expanding mask %#x to %#x for %s, base reg %u, count %u.\n", e->mask, merged_write_mask,
|
|
|
|
e->semantic_name, e->register_index, register_count);
|
|
|
|
return merged_write_mask;
|
|
|
|
}
|
|
|
|
|
|
|
|
WARN("Cannot expand mask %#x to %#x for %s, base reg %u, count %u.\n", e->mask, merged_write_mask,
|
|
|
|
e->semantic_name, e->register_index, register_count);
|
|
|
|
return e->mask;
|
|
|
|
}
|
|
|
|
|
2023-04-25 22:27:08 -07:00
|
|
|
static bool shader_signature_merge(struct shader_signature *s, uint8_t range_map[][VKD3D_VEC4_SIZE],
|
|
|
|
bool is_patch_constant)
|
|
|
|
{
|
|
|
|
unsigned int i, j, element_count, new_count, register_count;
|
|
|
|
struct signature_element *elements;
|
|
|
|
struct signature_element *e, *f;
|
2024-01-07 19:52:46 -08:00
|
|
|
bool used;
|
2023-04-25 22:27:08 -07:00
|
|
|
|
|
|
|
element_count = s->element_count;
|
|
|
|
if (!(elements = vkd3d_malloc(element_count * sizeof(*elements))))
|
|
|
|
return false;
|
|
|
|
memcpy(elements, s->elements, element_count * sizeof(*elements));
|
|
|
|
|
|
|
|
qsort(elements, element_count, sizeof(elements[0]), signature_element_register_compare);
|
|
|
|
|
|
|
|
for (i = 0, new_count = 0; i < element_count; i = j, elements[new_count++] = *e)
|
|
|
|
{
|
|
|
|
e = &elements[i];
|
|
|
|
j = i + 1;
|
|
|
|
|
|
|
|
if (e->register_index == ~0u)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
/* Do not merge if the register index will be relative-addressed. */
|
|
|
|
if (range_map_get_register_count(range_map, e->register_index, e->mask) > 1)
|
|
|
|
continue;
|
|
|
|
|
2024-01-07 19:52:46 -08:00
|
|
|
used = e->used_mask;
|
|
|
|
|
2023-04-25 22:27:08 -07:00
|
|
|
for (; j < element_count; ++j)
|
|
|
|
{
|
|
|
|
f = &elements[j];
|
|
|
|
|
|
|
|
/* Merge different components of the same register unless sysvals are different,
|
2024-01-03 17:23:16 -08:00
|
|
|
* or it will be relative-addressed. */
|
2023-04-25 22:27:08 -07:00
|
|
|
if (f->register_index != e->register_index || f->sysval_semantic != e->sysval_semantic
|
|
|
|
|| range_map_get_register_count(range_map, f->register_index, f->mask) > 1)
|
|
|
|
break;
|
|
|
|
|
|
|
|
TRACE("Merging %s, reg %u, mask %#x, sysval %#x with %s, mask %#x, sysval %#x.\n", e->semantic_name,
|
|
|
|
e->register_index, e->mask, e->sysval_semantic, f->semantic_name, f->mask, f->sysval_semantic);
|
|
|
|
assert(!(e->mask & f->mask));
|
|
|
|
|
|
|
|
e->mask |= f->mask;
|
|
|
|
e->used_mask |= f->used_mask;
|
|
|
|
e->semantic_index = min(e->semantic_index, f->semantic_index);
|
2024-01-07 19:52:46 -08:00
|
|
|
|
|
|
|
/* The first element may have no interpolation mode if it is unused. Elements which
|
|
|
|
* actually have different interpolation modes are assigned different registers. */
|
|
|
|
if (f->used_mask && !used)
|
|
|
|
{
|
|
|
|
if (e->interpolation_mode && e->interpolation_mode != f->interpolation_mode)
|
|
|
|
FIXME("Mismatching interpolation modes %u and %u.\n", e->interpolation_mode, f->interpolation_mode);
|
|
|
|
else
|
|
|
|
e->interpolation_mode = f->interpolation_mode;
|
|
|
|
}
|
2023-04-25 22:27:08 -07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
element_count = new_count;
|
2023-06-16 04:44:50 -07:00
|
|
|
vkd3d_free(s->elements);
|
2023-04-25 22:27:08 -07:00
|
|
|
s->elements = elements;
|
|
|
|
s->element_count = element_count;
|
|
|
|
|
|
|
|
if (is_patch_constant)
|
|
|
|
shader_signature_map_patch_constant_index_ranges(s, range_map);
|
|
|
|
|
|
|
|
for (i = 0, new_count = 0; i < element_count; i += register_count, elements[new_count++] = *e)
|
|
|
|
{
|
|
|
|
e = &elements[i];
|
|
|
|
register_count = 1;
|
|
|
|
|
|
|
|
if (e->register_index >= MAX_REG_OUTPUT)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
register_count = range_map_get_register_count(range_map, e->register_index, e->mask);
|
|
|
|
assert(register_count != UINT8_MAX);
|
|
|
|
register_count += !register_count;
|
|
|
|
|
|
|
|
if (register_count > 1)
|
|
|
|
{
|
|
|
|
TRACE("Merging %s, base reg %u, count %u.\n", e->semantic_name, e->register_index, register_count);
|
|
|
|
e->register_count = register_count;
|
2024-01-03 22:05:15 -08:00
|
|
|
e->mask = signature_element_range_expand_mask(e, register_count, range_map);
|
2023-04-25 22:27:08 -07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
element_count = new_count;
|
|
|
|
|
|
|
|
/* Restoring the original order is required for sensible trace output. */
|
|
|
|
qsort(elements, element_count, sizeof(elements[0]), signature_element_index_compare);
|
|
|
|
|
|
|
|
s->element_count = element_count;
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned int shader_register_normalise_arrayed_addressing(struct vkd3d_shader_register *reg,
|
|
|
|
unsigned int id_idx, unsigned int register_index)
|
|
|
|
{
|
|
|
|
assert(id_idx < ARRAY_SIZE(reg->idx) - 1);
|
|
|
|
|
|
|
|
/* For a relative-addressed register index, move the id up a slot to separate it from the address,
|
|
|
|
* because rel_addr can be replaced with a constant offset in some cases. */
|
|
|
|
if (reg->idx[id_idx].rel_addr)
|
|
|
|
{
|
|
|
|
reg->idx[id_idx + 1].rel_addr = NULL;
|
|
|
|
reg->idx[id_idx + 1].offset = reg->idx[id_idx].offset;
|
|
|
|
reg->idx[id_idx].offset -= register_index;
|
2024-01-09 19:51:59 -08:00
|
|
|
if (id_idx)
|
|
|
|
{
|
|
|
|
/* idx[id_idx] now contains the array index, which must be moved below the control point id. */
|
|
|
|
struct vkd3d_shader_register_index tmp = reg->idx[id_idx];
|
|
|
|
reg->idx[id_idx] = reg->idx[id_idx - 1];
|
|
|
|
reg->idx[id_idx - 1] = tmp;
|
|
|
|
}
|
2023-04-25 22:27:08 -07:00
|
|
|
++id_idx;
|
|
|
|
}
|
|
|
|
/* Otherwise we have no address for the arrayed register, so insert one. This happens e.g. where
|
|
|
|
* tessellation level registers are merged into an array because they're an array in SPIR-V. */
|
|
|
|
else
|
|
|
|
{
|
|
|
|
++id_idx;
|
|
|
|
memmove(®->idx[1], ®->idx[0], id_idx * sizeof(reg->idx[0]));
|
|
|
|
reg->idx[0].rel_addr = NULL;
|
|
|
|
reg->idx[0].offset = reg->idx[id_idx].offset - register_index;
|
|
|
|
}
|
|
|
|
|
|
|
|
return id_idx;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool shader_dst_param_io_normalise(struct vkd3d_shader_dst_param *dst_param, bool is_io_dcl,
|
|
|
|
struct io_normaliser *normaliser)
|
|
|
|
{
|
|
|
|
unsigned int id_idx, reg_idx, write_mask, element_idx;
|
|
|
|
struct vkd3d_shader_register *reg = &dst_param->reg;
|
|
|
|
struct vkd3d_shader_dst_param **dcl_params;
|
|
|
|
const struct shader_signature *signature;
|
|
|
|
const struct signature_element *e;
|
|
|
|
|
2023-10-27 10:44:08 -07:00
|
|
|
switch (reg->type)
|
2023-04-25 22:27:08 -07:00
|
|
|
{
|
2023-10-27 10:44:08 -07:00
|
|
|
case VKD3DSPR_OUTPUT:
|
|
|
|
reg_idx = reg->idx[reg->idx_count - 1].offset;
|
|
|
|
if (io_normaliser_is_in_fork_or_join_phase(normaliser))
|
|
|
|
{
|
|
|
|
signature = normaliser->patch_constant_signature;
|
|
|
|
/* Convert patch constant outputs to the patch constant register type to avoid the need
|
|
|
|
* to convert compiler symbols when accessed as inputs in a later stage. */
|
|
|
|
reg->type = VKD3DSPR_PATCHCONST;
|
|
|
|
dcl_params = normaliser->pc_dcl_params;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
signature = normaliser->output_signature;
|
|
|
|
dcl_params = normaliser->output_dcl_params;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSPR_PATCHCONST:
|
|
|
|
reg_idx = reg->idx[reg->idx_count - 1].offset;
|
|
|
|
signature = normaliser->patch_constant_signature;
|
|
|
|
dcl_params = normaliser->pc_dcl_params;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSPR_COLOROUT:
|
|
|
|
reg_idx = reg->idx[0].offset;
|
|
|
|
signature = normaliser->output_signature;
|
|
|
|
reg->type = VKD3DSPR_OUTPUT;
|
|
|
|
dcl_params = normaliser->output_dcl_params;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSPR_INCONTROLPOINT:
|
|
|
|
case VKD3DSPR_INPUT:
|
|
|
|
reg_idx = reg->idx[reg->idx_count - 1].offset;
|
|
|
|
signature = normaliser->input_signature;
|
|
|
|
reg->type = VKD3DSPR_INPUT;
|
|
|
|
dcl_params = normaliser->input_dcl_params;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSPR_ATTROUT:
|
|
|
|
reg_idx = SM1_COLOR_REGISTER_OFFSET + reg->idx[0].offset;
|
|
|
|
signature = normaliser->output_signature;
|
|
|
|
reg->type = VKD3DSPR_OUTPUT;
|
|
|
|
dcl_params = normaliser->output_dcl_params;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSPR_RASTOUT:
|
2023-10-27 10:44:38 -07:00
|
|
|
reg_idx = SM1_RASTOUT_REGISTER_OFFSET + reg->idx[0].offset;
|
2023-10-27 10:44:08 -07:00
|
|
|
signature = normaliser->output_signature;
|
|
|
|
reg->type = VKD3DSPR_OUTPUT;
|
|
|
|
dcl_params = normaliser->output_dcl_params;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
return true;
|
2023-04-25 22:27:08 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
id_idx = reg->idx_count - 1;
|
|
|
|
write_mask = dst_param->write_mask;
|
|
|
|
element_idx = shader_signature_find_element_for_reg(signature, reg_idx, write_mask);
|
|
|
|
e = &signature->elements[element_idx];
|
|
|
|
|
2023-12-05 13:19:07 -08:00
|
|
|
dst_param->write_mask >>= vsir_write_mask_get_component_idx(e->mask);
|
2023-04-25 22:27:08 -07:00
|
|
|
if (is_io_dcl)
|
|
|
|
{
|
|
|
|
/* Validated in the TPF reader. */
|
|
|
|
assert(element_idx < ARRAY_SIZE(normaliser->input_dcl_params));
|
|
|
|
|
|
|
|
if (dcl_params[element_idx])
|
|
|
|
{
|
|
|
|
/* Merge split declarations into a single one. */
|
|
|
|
dcl_params[element_idx]->write_mask |= dst_param->write_mask;
|
|
|
|
/* Turn this into a nop. */
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
dcl_params[element_idx] = dst_param;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (io_normaliser_is_in_control_point_phase(normaliser) && reg->type == VKD3DSPR_OUTPUT)
|
|
|
|
{
|
|
|
|
if (is_io_dcl)
|
|
|
|
{
|
|
|
|
/* Emit an array size for the control points for consistency with inputs. */
|
|
|
|
reg->idx[0].offset = normaliser->output_control_point_count;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
/* The control point id param. */
|
|
|
|
assert(reg->idx[0].rel_addr);
|
|
|
|
}
|
|
|
|
id_idx = 1;
|
|
|
|
}
|
|
|
|
|
2023-10-19 19:29:30 -07:00
|
|
|
if ((e->register_count > 1 || vsir_sysval_semantic_is_tess_factor(e->sysval_semantic)))
|
2023-04-25 22:27:08 -07:00
|
|
|
{
|
|
|
|
if (is_io_dcl)
|
|
|
|
{
|
|
|
|
/* For control point I/O, idx 0 contains the control point count.
|
|
|
|
* Ensure it is moved up to the next slot. */
|
|
|
|
reg->idx[id_idx].offset = reg->idx[0].offset;
|
|
|
|
reg->idx[0].offset = e->register_count;
|
|
|
|
++id_idx;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
id_idx = shader_register_normalise_arrayed_addressing(reg, id_idx, e->register_index);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Replace the register index with the signature element index */
|
|
|
|
reg->idx[id_idx].offset = element_idx;
|
|
|
|
reg->idx_count = id_idx + 1;
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void shader_src_param_io_normalise(struct vkd3d_shader_src_param *src_param,
|
|
|
|
struct io_normaliser *normaliser)
|
|
|
|
{
|
|
|
|
unsigned int i, id_idx, reg_idx, write_mask, element_idx, component_idx;
|
|
|
|
struct vkd3d_shader_register *reg = &src_param->reg;
|
|
|
|
const struct shader_signature *signature;
|
|
|
|
const struct signature_element *e;
|
|
|
|
|
|
|
|
/* Input/output registers from one phase can be used as inputs in
|
|
|
|
* subsequent phases. Specifically:
|
|
|
|
*
|
|
|
|
* - Control phase inputs are available as "vicp" in fork and join
|
|
|
|
* phases.
|
|
|
|
* - Control phase outputs are available as "vocp" in fork and join
|
|
|
|
* phases.
|
|
|
|
* - Fork phase patch constants are available as "vpc" in join
|
|
|
|
* phases.
|
|
|
|
*
|
|
|
|
* We handle "vicp" here by converting INCONTROLPOINT src registers to
|
|
|
|
* type INPUT so they match the control phase declarations. We handle
|
|
|
|
* "vocp" by converting OUTCONTROLPOINT registers to type OUTPUT.
|
|
|
|
* Merging fork and join phases handles "vpc". */
|
|
|
|
|
|
|
|
switch (reg->type)
|
|
|
|
{
|
|
|
|
case VKD3DSPR_PATCHCONST:
|
2023-10-27 10:44:08 -07:00
|
|
|
reg_idx = reg->idx[reg->idx_count - 1].offset;
|
2023-04-25 22:27:08 -07:00
|
|
|
signature = normaliser->patch_constant_signature;
|
|
|
|
break;
|
2023-10-27 10:44:08 -07:00
|
|
|
|
2023-04-25 22:27:08 -07:00
|
|
|
case VKD3DSPR_INCONTROLPOINT:
|
2023-07-16 21:56:22 -07:00
|
|
|
reg->type = VKD3DSPR_INPUT;
|
2023-04-25 22:27:08 -07:00
|
|
|
/* fall through */
|
|
|
|
case VKD3DSPR_INPUT:
|
2023-10-27 10:44:08 -07:00
|
|
|
if (normaliser->major < 3 && normaliser->shader_type == VKD3D_SHADER_TYPE_PIXEL)
|
|
|
|
reg_idx = SM1_COLOR_REGISTER_OFFSET + reg->idx[0].offset;
|
|
|
|
else
|
|
|
|
reg_idx = reg->idx[reg->idx_count - 1].offset;
|
2023-04-25 22:27:08 -07:00
|
|
|
signature = normaliser->input_signature;
|
|
|
|
break;
|
2023-10-27 10:44:08 -07:00
|
|
|
|
2023-04-25 22:27:08 -07:00
|
|
|
case VKD3DSPR_OUTCONTROLPOINT:
|
2023-07-16 21:56:22 -07:00
|
|
|
reg->type = VKD3DSPR_OUTPUT;
|
2023-04-25 22:27:08 -07:00
|
|
|
/* fall through */
|
|
|
|
case VKD3DSPR_OUTPUT:
|
2023-10-27 10:44:08 -07:00
|
|
|
reg_idx = reg->idx[reg->idx_count - 1].offset;
|
2023-04-25 22:27:08 -07:00
|
|
|
signature = normaliser->output_signature;
|
|
|
|
break;
|
2023-10-27 10:44:08 -07:00
|
|
|
|
|
|
|
case VKD3DSPR_TEXTURE:
|
|
|
|
if (normaliser->shader_type != VKD3D_SHADER_TYPE_PIXEL)
|
|
|
|
return;
|
|
|
|
reg->type = VKD3DSPR_INPUT;
|
|
|
|
reg_idx = reg->idx[0].offset;
|
|
|
|
signature = normaliser->input_signature;
|
|
|
|
break;
|
|
|
|
|
2023-04-25 22:27:08 -07:00
|
|
|
default:
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
id_idx = reg->idx_count - 1;
|
2023-12-05 14:05:40 -08:00
|
|
|
write_mask = VKD3DSP_WRITEMASK_0 << vsir_swizzle_get_component(src_param->swizzle, 0);
|
2023-04-25 22:27:08 -07:00
|
|
|
element_idx = shader_signature_find_element_for_reg(signature, reg_idx, write_mask);
|
|
|
|
|
|
|
|
e = &signature->elements[element_idx];
|
2023-10-19 19:29:30 -07:00
|
|
|
if ((e->register_count > 1 || vsir_sysval_semantic_is_tess_factor(e->sysval_semantic)))
|
2023-04-25 22:27:08 -07:00
|
|
|
id_idx = shader_register_normalise_arrayed_addressing(reg, id_idx, e->register_index);
|
|
|
|
reg->idx[id_idx].offset = element_idx;
|
|
|
|
reg->idx_count = id_idx + 1;
|
|
|
|
|
2023-12-05 13:19:07 -08:00
|
|
|
if ((component_idx = vsir_write_mask_get_component_idx(e->mask)))
|
2023-04-25 22:27:08 -07:00
|
|
|
{
|
|
|
|
for (i = 0; i < VKD3D_VEC4_SIZE; ++i)
|
2023-12-05 14:05:40 -08:00
|
|
|
if (vsir_swizzle_get_component(src_param->swizzle, i))
|
2023-04-25 22:27:08 -07:00
|
|
|
src_param->swizzle -= component_idx << VKD3D_SHADER_SWIZZLE_SHIFT(i);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void shader_instruction_normalise_io_params(struct vkd3d_shader_instruction *ins,
|
|
|
|
struct io_normaliser *normaliser)
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_register *reg;
|
|
|
|
unsigned int i;
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (ins->opcode)
|
2023-04-25 22:27:08 -07:00
|
|
|
{
|
|
|
|
case VKD3DSIH_DCL_INPUT:
|
|
|
|
if (normaliser->shader_type == VKD3D_SHADER_TYPE_HULL)
|
|
|
|
{
|
|
|
|
reg = &ins->declaration.dst.reg;
|
2023-07-16 21:34:04 -07:00
|
|
|
|
|
|
|
if (reg->type == VKD3DSPR_OUTCONTROLPOINT)
|
|
|
|
normaliser->use_vocp = true;
|
|
|
|
|
2023-04-25 22:27:08 -07:00
|
|
|
/* We don't need to keep OUTCONTROLPOINT or PATCHCONST input declarations since their
|
|
|
|
* equivalents were declared earlier, but INCONTROLPOINT may be the first occurrence. */
|
|
|
|
if (reg->type == VKD3DSPR_OUTCONTROLPOINT || reg->type == VKD3DSPR_PATCHCONST)
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
|
|
|
else if (reg->type == VKD3DSPR_INCONTROLPOINT)
|
|
|
|
reg->type = VKD3DSPR_INPUT;
|
|
|
|
}
|
|
|
|
/* fall through */
|
|
|
|
case VKD3DSIH_DCL_INPUT_PS:
|
|
|
|
case VKD3DSIH_DCL_OUTPUT:
|
2023-09-08 07:07:45 -07:00
|
|
|
if (!shader_dst_param_io_normalise(&ins->declaration.dst, true, normaliser))
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
2023-04-25 22:27:08 -07:00
|
|
|
break;
|
|
|
|
case VKD3DSIH_DCL_INPUT_SGV:
|
|
|
|
case VKD3DSIH_DCL_INPUT_SIV:
|
|
|
|
case VKD3DSIH_DCL_INPUT_PS_SGV:
|
|
|
|
case VKD3DSIH_DCL_INPUT_PS_SIV:
|
|
|
|
case VKD3DSIH_DCL_OUTPUT_SIV:
|
2023-09-08 07:07:45 -07:00
|
|
|
if (!shader_dst_param_io_normalise(&ins->declaration.register_semantic.reg, true, normaliser))
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
2023-04-25 22:27:08 -07:00
|
|
|
break;
|
|
|
|
case VKD3DSIH_HS_CONTROL_POINT_PHASE:
|
|
|
|
case VKD3DSIH_HS_FORK_PHASE:
|
|
|
|
case VKD3DSIH_HS_JOIN_PHASE:
|
2024-04-24 11:34:42 -07:00
|
|
|
normaliser->phase = ins->opcode;
|
2023-04-25 22:27:08 -07:00
|
|
|
memset(normaliser->input_dcl_params, 0, sizeof(normaliser->input_dcl_params));
|
|
|
|
memset(normaliser->output_dcl_params, 0, sizeof(normaliser->output_dcl_params));
|
|
|
|
memset(normaliser->pc_dcl_params, 0, sizeof(normaliser->pc_dcl_params));
|
|
|
|
break;
|
|
|
|
default:
|
2024-01-22 11:32:55 -08:00
|
|
|
if (vsir_instruction_is_dcl(ins))
|
2023-04-25 22:27:08 -07:00
|
|
|
break;
|
|
|
|
for (i = 0; i < ins->dst_count; ++i)
|
2024-01-11 02:19:39 -08:00
|
|
|
shader_dst_param_io_normalise(&ins->dst[i], false, normaliser);
|
2023-04-25 22:27:08 -07:00
|
|
|
for (i = 0; i < ins->src_count; ++i)
|
2024-01-11 02:19:39 -08:00
|
|
|
shader_src_param_io_normalise(&ins->src[i], normaliser);
|
2023-04-25 22:27:08 -07:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-03-12 12:31:56 -07:00
|
|
|
static enum vkd3d_result vsir_program_normalise_io_registers(struct vsir_program *program)
|
2023-04-25 22:27:08 -07:00
|
|
|
{
|
2024-03-12 12:31:56 -07:00
|
|
|
struct io_normaliser normaliser = {program->instructions};
|
2023-04-25 22:27:08 -07:00
|
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
bool has_control_point_phase;
|
|
|
|
unsigned int i, j;
|
|
|
|
|
|
|
|
normaliser.phase = VKD3DSIH_INVALID;
|
2024-01-17 11:39:55 -08:00
|
|
|
normaliser.shader_type = program->shader_version.type;
|
|
|
|
normaliser.major = program->shader_version.major;
|
2024-03-01 04:36:11 -08:00
|
|
|
normaliser.input_signature = &program->input_signature;
|
|
|
|
normaliser.output_signature = &program->output_signature;
|
|
|
|
normaliser.patch_constant_signature = &program->patch_constant_signature;
|
2023-04-25 22:27:08 -07:00
|
|
|
|
2024-01-17 11:39:55 -08:00
|
|
|
for (i = 0, has_control_point_phase = false; i < program->instructions.count; ++i)
|
2023-04-25 22:27:08 -07:00
|
|
|
{
|
2024-01-17 11:39:55 -08:00
|
|
|
ins = &program->instructions.elements[i];
|
2023-04-25 22:27:08 -07:00
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (ins->opcode)
|
2023-04-25 22:27:08 -07:00
|
|
|
{
|
|
|
|
case VKD3DSIH_DCL_OUTPUT_CONTROL_POINT_COUNT:
|
|
|
|
normaliser.output_control_point_count = ins->declaration.count;
|
|
|
|
break;
|
|
|
|
case VKD3DSIH_DCL_INDEX_RANGE:
|
|
|
|
io_normaliser_add_index_range(&normaliser, ins);
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
|
|
|
break;
|
|
|
|
case VKD3DSIH_HS_CONTROL_POINT_PHASE:
|
|
|
|
has_control_point_phase = true;
|
|
|
|
/* fall through */
|
|
|
|
case VKD3DSIH_HS_FORK_PHASE:
|
|
|
|
case VKD3DSIH_HS_JOIN_PHASE:
|
2024-04-24 11:34:42 -07:00
|
|
|
normaliser.phase = ins->opcode;
|
2023-04-25 22:27:08 -07:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (normaliser.shader_type == VKD3D_SHADER_TYPE_HULL && !has_control_point_phase)
|
|
|
|
{
|
|
|
|
/* Inputs and outputs must match for the default phase, so merge ranges must match too. */
|
|
|
|
for (i = 0; i < MAX_REG_OUTPUT; ++i)
|
|
|
|
{
|
|
|
|
for (j = 0; j < VKD3D_VEC4_SIZE; ++j)
|
|
|
|
{
|
|
|
|
if (!normaliser.input_range_map[i][j] && normaliser.output_range_map[i][j])
|
|
|
|
normaliser.input_range_map[i][j] = normaliser.output_range_map[i][j];
|
|
|
|
else if (normaliser.input_range_map[i][j] && !normaliser.output_range_map[i][j])
|
|
|
|
normaliser.output_range_map[i][j] = normaliser.input_range_map[i][j];
|
|
|
|
else assert(normaliser.input_range_map[i][j] == normaliser.output_range_map[i][j]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-03-01 04:36:11 -08:00
|
|
|
if (!shader_signature_merge(&program->input_signature, normaliser.input_range_map, false)
|
|
|
|
|| !shader_signature_merge(&program->output_signature, normaliser.output_range_map, false)
|
|
|
|
|| !shader_signature_merge(&program->patch_constant_signature, normaliser.pc_range_map, true))
|
2023-04-25 22:27:08 -07:00
|
|
|
{
|
2024-01-17 11:39:55 -08:00
|
|
|
program->instructions = normaliser.instructions;
|
2023-04-25 22:27:08 -07:00
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
}
|
|
|
|
|
|
|
|
normaliser.phase = VKD3DSIH_INVALID;
|
|
|
|
for (i = 0; i < normaliser.instructions.count; ++i)
|
|
|
|
shader_instruction_normalise_io_params(&normaliser.instructions.elements[i], &normaliser);
|
|
|
|
|
2024-01-17 11:39:55 -08:00
|
|
|
program->instructions = normaliser.instructions;
|
|
|
|
program->use_vocp = normaliser.use_vocp;
|
2023-04-25 22:27:08 -07:00
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
2023-07-02 18:23:07 -07:00
|
|
|
|
2023-07-09 15:46:04 -07:00
|
|
|
struct flat_constant_def
|
|
|
|
{
|
|
|
|
enum vkd3d_shader_d3dbc_constant_register set;
|
|
|
|
uint32_t index;
|
|
|
|
uint32_t value[4];
|
|
|
|
};
|
|
|
|
|
|
|
|
struct flat_constants_normaliser
|
|
|
|
{
|
|
|
|
struct flat_constant_def *defs;
|
|
|
|
size_t def_count, defs_capacity;
|
|
|
|
};
|
|
|
|
|
|
|
|
static bool get_flat_constant_register_type(const struct vkd3d_shader_register *reg,
|
|
|
|
enum vkd3d_shader_d3dbc_constant_register *set, uint32_t *index)
|
|
|
|
{
|
|
|
|
static const struct
|
|
|
|
{
|
|
|
|
enum vkd3d_shader_register_type type;
|
|
|
|
enum vkd3d_shader_d3dbc_constant_register set;
|
|
|
|
uint32_t offset;
|
|
|
|
}
|
|
|
|
regs[] =
|
|
|
|
{
|
|
|
|
{VKD3DSPR_CONST, VKD3D_SHADER_D3DBC_FLOAT_CONSTANT_REGISTER, 0},
|
|
|
|
{VKD3DSPR_CONST2, VKD3D_SHADER_D3DBC_FLOAT_CONSTANT_REGISTER, 2048},
|
|
|
|
{VKD3DSPR_CONST3, VKD3D_SHADER_D3DBC_FLOAT_CONSTANT_REGISTER, 4096},
|
|
|
|
{VKD3DSPR_CONST4, VKD3D_SHADER_D3DBC_FLOAT_CONSTANT_REGISTER, 6144},
|
|
|
|
{VKD3DSPR_CONSTINT, VKD3D_SHADER_D3DBC_INT_CONSTANT_REGISTER, 0},
|
|
|
|
{VKD3DSPR_CONSTBOOL, VKD3D_SHADER_D3DBC_BOOL_CONSTANT_REGISTER, 0},
|
|
|
|
};
|
|
|
|
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(regs); ++i)
|
|
|
|
{
|
|
|
|
if (reg->type == regs[i].type)
|
|
|
|
{
|
|
|
|
if (reg->idx[0].rel_addr)
|
|
|
|
{
|
|
|
|
FIXME("Unhandled relative address.\n");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
*set = regs[i].set;
|
|
|
|
*index = regs[i].offset + reg->idx[0].offset;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void shader_register_normalise_flat_constants(struct vkd3d_shader_src_param *param,
|
|
|
|
const struct flat_constants_normaliser *normaliser)
|
|
|
|
{
|
|
|
|
enum vkd3d_shader_d3dbc_constant_register set;
|
|
|
|
uint32_t index;
|
|
|
|
size_t i, j;
|
|
|
|
|
|
|
|
if (!get_flat_constant_register_type(¶m->reg, &set, &index))
|
|
|
|
return;
|
|
|
|
|
|
|
|
for (i = 0; i < normaliser->def_count; ++i)
|
|
|
|
{
|
|
|
|
if (normaliser->defs[i].set == set && normaliser->defs[i].index == index)
|
|
|
|
{
|
|
|
|
param->reg.type = VKD3DSPR_IMMCONST;
|
|
|
|
param->reg.idx_count = 0;
|
2023-08-14 10:21:41 -07:00
|
|
|
param->reg.dimension = VSIR_DIMENSION_VEC4;
|
2023-07-09 15:46:04 -07:00
|
|
|
for (j = 0; j < 4; ++j)
|
2023-12-05 14:24:36 -08:00
|
|
|
param->reg.u.immconst_u32[j] = normaliser->defs[i].value[j];
|
2023-07-09 15:46:04 -07:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
param->reg.type = VKD3DSPR_CONSTBUFFER;
|
|
|
|
param->reg.idx[0].offset = set; /* register ID */
|
|
|
|
param->reg.idx[1].offset = set; /* register index */
|
|
|
|
param->reg.idx[2].offset = index; /* buffer index */
|
|
|
|
param->reg.idx_count = 3;
|
|
|
|
}
|
|
|
|
|
2024-01-16 05:47:34 -08:00
|
|
|
static enum vkd3d_result instruction_array_normalise_flat_constants(struct vsir_program *program)
|
2023-07-09 15:46:04 -07:00
|
|
|
{
|
2024-01-16 05:47:34 -08:00
|
|
|
struct flat_constants_normaliser normaliser = {0};
|
2023-07-09 15:46:04 -07:00
|
|
|
unsigned int i, j;
|
|
|
|
|
2024-01-16 05:47:34 -08:00
|
|
|
for (i = 0; i < program->instructions.count; ++i)
|
2023-07-09 15:46:04 -07:00
|
|
|
{
|
2024-01-16 05:47:34 -08:00
|
|
|
struct vkd3d_shader_instruction *ins = &program->instructions.elements[i];
|
2023-07-09 15:46:04 -07:00
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
if (ins->opcode == VKD3DSIH_DEF || ins->opcode == VKD3DSIH_DEFI || ins->opcode == VKD3DSIH_DEFB)
|
2023-07-09 15:46:04 -07:00
|
|
|
{
|
|
|
|
struct flat_constant_def *def;
|
|
|
|
|
|
|
|
if (!vkd3d_array_reserve((void **)&normaliser.defs, &normaliser.defs_capacity,
|
|
|
|
normaliser.def_count + 1, sizeof(*normaliser.defs)))
|
|
|
|
{
|
|
|
|
vkd3d_free(normaliser.defs);
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
}
|
|
|
|
|
|
|
|
def = &normaliser.defs[normaliser.def_count++];
|
|
|
|
|
|
|
|
get_flat_constant_register_type((struct vkd3d_shader_register *)&ins->dst[0].reg, &def->set, &def->index);
|
|
|
|
for (j = 0; j < 4; ++j)
|
2023-12-05 14:24:36 -08:00
|
|
|
def->value[j] = ins->src[0].reg.u.immconst_u32[j];
|
2023-07-09 15:46:04 -07:00
|
|
|
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
for (j = 0; j < ins->src_count; ++j)
|
2024-01-11 02:19:39 -08:00
|
|
|
shader_register_normalise_flat_constants(&ins->src[j], &normaliser);
|
2023-07-09 15:46:04 -07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
vkd3d_free(normaliser.defs);
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-01-16 05:47:34 -08:00
|
|
|
static void remove_dead_code(struct vsir_program *program)
|
2023-09-20 13:10:39 -07:00
|
|
|
{
|
|
|
|
size_t i, depth = 0;
|
|
|
|
bool dead = false;
|
|
|
|
|
2024-01-16 05:47:34 -08:00
|
|
|
for (i = 0; i < program->instructions.count; ++i)
|
2023-09-20 13:10:39 -07:00
|
|
|
{
|
2024-01-16 05:47:34 -08:00
|
|
|
struct vkd3d_shader_instruction *ins = &program->instructions.elements[i];
|
2023-09-20 13:10:39 -07:00
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (ins->opcode)
|
2023-09-20 13:10:39 -07:00
|
|
|
{
|
|
|
|
case VKD3DSIH_IF:
|
|
|
|
case VKD3DSIH_LOOP:
|
|
|
|
case VKD3DSIH_SWITCH:
|
|
|
|
if (dead)
|
|
|
|
{
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
|
|
|
++depth;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_ENDIF:
|
|
|
|
case VKD3DSIH_ENDLOOP:
|
|
|
|
case VKD3DSIH_ENDSWITCH:
|
|
|
|
case VKD3DSIH_ELSE:
|
|
|
|
if (dead)
|
|
|
|
{
|
|
|
|
if (depth > 0)
|
|
|
|
{
|
2024-04-24 11:34:42 -07:00
|
|
|
if (ins->opcode != VKD3DSIH_ELSE)
|
2023-09-20 13:10:39 -07:00
|
|
|
--depth;
|
2023-11-08 20:07:03 -08:00
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
2023-09-20 13:10:39 -07:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
dead = false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
/* `depth' is counted with respect to where the dead code
|
|
|
|
* segment began. So it starts at zero and it signals the
|
|
|
|
* termination of the dead code segment when it would
|
|
|
|
* become negative. */
|
|
|
|
case VKD3DSIH_BREAK:
|
|
|
|
case VKD3DSIH_RET:
|
|
|
|
case VKD3DSIH_CONTINUE:
|
|
|
|
if (dead)
|
|
|
|
{
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
dead = true;
|
|
|
|
depth = 0;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
/* If `case' or `default' appears at zero depth, it means
|
|
|
|
* that they are a possible target for the corresponding
|
|
|
|
* switch, so the code is live again. */
|
|
|
|
case VKD3DSIH_CASE:
|
|
|
|
case VKD3DSIH_DEFAULT:
|
|
|
|
if (dead)
|
|
|
|
{
|
|
|
|
if (depth == 0)
|
|
|
|
dead = false;
|
|
|
|
else
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
/* Phase instructions can only appear in hull shaders and
|
|
|
|
* outside of any block. When a phase returns, control is
|
|
|
|
* moved to the following phase, so they make code live
|
|
|
|
* again. */
|
|
|
|
case VKD3DSIH_HS_CONTROL_POINT_PHASE:
|
|
|
|
case VKD3DSIH_HS_FORK_PHASE:
|
|
|
|
case VKD3DSIH_HS_JOIN_PHASE:
|
|
|
|
dead = false;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
if (dead)
|
|
|
|
vkd3d_shader_instruction_make_nop(ins);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-03-12 13:43:41 -07:00
|
|
|
static enum vkd3d_result vsir_program_normalise_combined_samplers(struct vsir_program *program,
|
|
|
|
struct vkd3d_shader_message_context *message_context)
|
2023-08-11 14:46:03 -07:00
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
2024-03-12 13:43:41 -07:00
|
|
|
for (i = 0; i < program->instructions.count; ++i)
|
2023-08-11 14:46:03 -07:00
|
|
|
{
|
2024-03-12 13:43:41 -07:00
|
|
|
struct vkd3d_shader_instruction *ins = &program->instructions.elements[i];
|
2023-08-11 14:46:03 -07:00
|
|
|
struct vkd3d_shader_src_param *srcs;
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (ins->opcode)
|
2023-08-11 14:46:03 -07:00
|
|
|
{
|
|
|
|
case VKD3DSIH_TEX:
|
2024-03-12 13:43:41 -07:00
|
|
|
if (!(srcs = shader_src_param_allocator_get(&program->instructions.src_params, 3)))
|
2023-08-11 14:46:03 -07:00
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
memset(srcs, 0, sizeof(*srcs) * 3);
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
ins->opcode = VKD3DSIH_SAMPLE;
|
2023-08-11 14:46:03 -07:00
|
|
|
|
|
|
|
srcs[0] = ins->src[0];
|
|
|
|
|
|
|
|
srcs[1].reg.type = VKD3DSPR_RESOURCE;
|
|
|
|
srcs[1].reg.idx[0] = ins->src[1].reg.idx[0];
|
|
|
|
srcs[1].reg.idx[1] = ins->src[1].reg.idx[0];
|
|
|
|
srcs[1].reg.idx_count = 2;
|
|
|
|
srcs[1].reg.data_type = VKD3D_DATA_RESOURCE;
|
2024-05-24 03:58:39 -07:00
|
|
|
srcs[1].reg.dimension = VSIR_DIMENSION_VEC4;
|
2023-08-11 14:46:03 -07:00
|
|
|
srcs[1].swizzle = VKD3D_SHADER_NO_SWIZZLE;
|
|
|
|
|
|
|
|
srcs[2].reg.type = VKD3DSPR_SAMPLER;
|
|
|
|
srcs[2].reg.idx[0] = ins->src[1].reg.idx[0];
|
|
|
|
srcs[2].reg.idx[1] = ins->src[1].reg.idx[0];
|
|
|
|
srcs[2].reg.idx_count = 2;
|
|
|
|
srcs[2].reg.data_type = VKD3D_DATA_SAMPLER;
|
|
|
|
|
|
|
|
ins->src = srcs;
|
|
|
|
ins->src_count = 3;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_TEXBEM:
|
|
|
|
case VKD3DSIH_TEXBEML:
|
|
|
|
case VKD3DSIH_TEXCOORD:
|
|
|
|
case VKD3DSIH_TEXDEPTH:
|
|
|
|
case VKD3DSIH_TEXDP3:
|
|
|
|
case VKD3DSIH_TEXDP3TEX:
|
|
|
|
case VKD3DSIH_TEXLDD:
|
|
|
|
case VKD3DSIH_TEXLDL:
|
|
|
|
case VKD3DSIH_TEXM3x2PAD:
|
|
|
|
case VKD3DSIH_TEXM3x2TEX:
|
|
|
|
case VKD3DSIH_TEXM3x3DIFF:
|
|
|
|
case VKD3DSIH_TEXM3x3PAD:
|
|
|
|
case VKD3DSIH_TEXM3x3SPEC:
|
|
|
|
case VKD3DSIH_TEXM3x3TEX:
|
|
|
|
case VKD3DSIH_TEXM3x3VSPEC:
|
|
|
|
case VKD3DSIH_TEXREG2AR:
|
|
|
|
case VKD3DSIH_TEXREG2GB:
|
|
|
|
case VKD3DSIH_TEXREG2RGB:
|
2024-03-12 13:43:41 -07:00
|
|
|
vkd3d_shader_error(message_context, &ins->location, VKD3D_SHADER_ERROR_VSIR_NOT_IMPLEMENTED,
|
2023-08-11 14:46:03 -07:00
|
|
|
"Aborting due to not yet implemented feature: "
|
2024-04-24 11:34:42 -07:00
|
|
|
"Combined sampler instruction %#x.", ins->opcode);
|
2023-08-11 14:46:03 -07:00
|
|
|
return VKD3D_ERROR_NOT_IMPLEMENTED;
|
|
|
|
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-01-04 06:55:51 -08:00
|
|
|
struct cf_flattener_if_info
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_src_param *false_param;
|
2023-11-07 19:14:01 -08:00
|
|
|
unsigned int id;
|
2024-01-04 06:55:51 -08:00
|
|
|
uint32_t merge_block_id;
|
|
|
|
unsigned int else_block_id;
|
|
|
|
};
|
|
|
|
|
2023-12-11 21:34:51 -08:00
|
|
|
struct cf_flattener_loop_info
|
|
|
|
{
|
|
|
|
unsigned int header_block_id;
|
|
|
|
unsigned int continue_block_id;
|
|
|
|
uint32_t merge_block_id;
|
|
|
|
};
|
|
|
|
|
2023-12-11 21:58:08 -08:00
|
|
|
struct cf_flattener_switch_case
|
|
|
|
{
|
|
|
|
unsigned int value;
|
|
|
|
unsigned int block_id;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct cf_flattener_switch_info
|
|
|
|
{
|
|
|
|
size_t ins_location;
|
|
|
|
const struct vkd3d_shader_src_param *condition;
|
2023-11-07 19:14:01 -08:00
|
|
|
unsigned int id;
|
2023-12-11 21:58:08 -08:00
|
|
|
unsigned int merge_block_id;
|
|
|
|
unsigned int default_block_id;
|
|
|
|
struct cf_flattener_switch_case *cases;
|
|
|
|
size_t cases_size;
|
|
|
|
unsigned int cases_count;
|
|
|
|
};
|
|
|
|
|
2024-01-04 06:55:51 -08:00
|
|
|
struct cf_flattener_info
|
|
|
|
{
|
|
|
|
union
|
|
|
|
{
|
|
|
|
struct cf_flattener_if_info if_;
|
2023-12-11 21:34:51 -08:00
|
|
|
struct cf_flattener_loop_info loop;
|
2023-12-11 21:58:08 -08:00
|
|
|
struct cf_flattener_switch_info switch_;
|
2024-01-04 06:55:51 -08:00
|
|
|
} u;
|
|
|
|
|
|
|
|
enum
|
|
|
|
{
|
|
|
|
VKD3D_BLOCK_IF,
|
2023-12-11 21:34:51 -08:00
|
|
|
VKD3D_BLOCK_LOOP,
|
|
|
|
VKD3D_BLOCK_SWITCH,
|
2024-01-04 06:55:51 -08:00
|
|
|
} current_block;
|
|
|
|
bool inside_block;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct cf_flattener
|
|
|
|
{
|
2024-03-12 14:37:04 -07:00
|
|
|
struct vsir_program *program;
|
2024-01-04 06:55:51 -08:00
|
|
|
|
|
|
|
struct vkd3d_shader_location location;
|
2024-03-12 14:37:04 -07:00
|
|
|
enum vkd3d_result status;
|
2024-01-04 06:55:51 -08:00
|
|
|
|
|
|
|
struct vkd3d_shader_instruction *instructions;
|
|
|
|
size_t instruction_capacity;
|
|
|
|
size_t instruction_count;
|
|
|
|
|
|
|
|
unsigned int block_id;
|
2023-11-07 19:14:01 -08:00
|
|
|
const char **block_names;
|
|
|
|
size_t block_name_capacity;
|
|
|
|
size_t block_name_count;
|
|
|
|
|
|
|
|
unsigned int branch_id;
|
|
|
|
unsigned int loop_id;
|
|
|
|
unsigned int switch_id;
|
2024-01-04 06:55:51 -08:00
|
|
|
|
|
|
|
unsigned int control_flow_depth;
|
|
|
|
struct cf_flattener_info *control_flow_info;
|
|
|
|
size_t control_flow_info_size;
|
|
|
|
};
|
|
|
|
|
2024-03-12 14:37:04 -07:00
|
|
|
static void cf_flattener_set_error(struct cf_flattener *flattener, enum vkd3d_result error)
|
|
|
|
{
|
|
|
|
if (flattener->status != VKD3D_OK)
|
|
|
|
return;
|
|
|
|
flattener->status = error;
|
|
|
|
}
|
|
|
|
|
2024-01-04 06:55:51 -08:00
|
|
|
static struct vkd3d_shader_instruction *cf_flattener_require_space(struct cf_flattener *flattener, size_t count)
|
|
|
|
{
|
|
|
|
if (!vkd3d_array_reserve((void **)&flattener->instructions, &flattener->instruction_capacity,
|
|
|
|
flattener->instruction_count + count, sizeof(*flattener->instructions)))
|
|
|
|
{
|
|
|
|
ERR("Failed to allocate instructions.\n");
|
2024-03-12 14:37:04 -07:00
|
|
|
cf_flattener_set_error(flattener, VKD3D_ERROR_OUT_OF_MEMORY);
|
2024-01-04 06:55:51 -08:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
return &flattener->instructions[flattener->instruction_count];
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool cf_flattener_copy_instruction(struct cf_flattener *flattener,
|
|
|
|
const struct vkd3d_shader_instruction *instruction)
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction *dst_ins;
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
if (instruction->opcode == VKD3DSIH_NOP)
|
2024-01-04 06:55:51 -08:00
|
|
|
return true;
|
|
|
|
|
|
|
|
if (!(dst_ins = cf_flattener_require_space(flattener, 1)))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
*dst_ins = *instruction;
|
|
|
|
++flattener->instruction_count;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned int cf_flattener_alloc_block_id(struct cf_flattener *flattener)
|
|
|
|
{
|
|
|
|
return ++flattener->block_id;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct vkd3d_shader_src_param *instruction_src_params_alloc(struct vkd3d_shader_instruction *ins,
|
|
|
|
unsigned int count, struct cf_flattener *flattener)
|
|
|
|
{
|
2024-01-24 05:26:35 -08:00
|
|
|
struct vkd3d_shader_src_param *params;
|
|
|
|
|
2024-03-12 14:37:04 -07:00
|
|
|
if (!(params = vsir_program_get_src_params(flattener->program, count)))
|
2024-01-04 06:55:51 -08:00
|
|
|
{
|
2024-03-12 14:37:04 -07:00
|
|
|
cf_flattener_set_error(flattener, VKD3D_ERROR_OUT_OF_MEMORY);
|
2024-01-04 06:55:51 -08:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
ins->src = params;
|
|
|
|
ins->src_count = count;
|
|
|
|
return params;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void cf_flattener_emit_label(struct cf_flattener *flattener, unsigned int label_id)
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
|
|
|
|
if (!(ins = cf_flattener_require_space(flattener, 1)))
|
|
|
|
return;
|
2024-03-12 14:37:04 -07:00
|
|
|
if (vsir_instruction_init_label(ins, &flattener->location, label_id, flattener->program))
|
2024-01-04 06:55:51 -08:00
|
|
|
++flattener->instruction_count;
|
|
|
|
else
|
2024-03-12 14:37:04 -07:00
|
|
|
cf_flattener_set_error(flattener, VKD3D_ERROR_OUT_OF_MEMORY);
|
2024-01-04 06:55:51 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* For conditional branches, this returns the false target branch parameter. */
|
|
|
|
static struct vkd3d_shader_src_param *cf_flattener_emit_branch(struct cf_flattener *flattener,
|
2023-12-11 21:34:51 -08:00
|
|
|
unsigned int merge_block_id, unsigned int continue_block_id,
|
|
|
|
const struct vkd3d_shader_src_param *condition, unsigned int true_id, unsigned int false_id,
|
2024-01-04 06:55:51 -08:00
|
|
|
unsigned int flags)
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_src_param *src_params, *false_branch_param;
|
|
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
|
|
|
|
if (!(ins = cf_flattener_require_space(flattener, 1)))
|
|
|
|
return NULL;
|
|
|
|
vsir_instruction_init(ins, &flattener->location, VKD3DSIH_BRANCH);
|
|
|
|
|
|
|
|
if (condition)
|
|
|
|
{
|
2023-12-11 21:34:51 -08:00
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 4 + !!continue_block_id, flattener)))
|
2024-01-04 06:55:51 -08:00
|
|
|
return NULL;
|
|
|
|
src_params[0] = *condition;
|
|
|
|
if (flags == VKD3D_SHADER_CONDITIONAL_OP_Z)
|
|
|
|
{
|
|
|
|
vsir_src_param_init_label(&src_params[1], false_id);
|
|
|
|
vsir_src_param_init_label(&src_params[2], true_id);
|
|
|
|
false_branch_param = &src_params[1];
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
vsir_src_param_init_label(&src_params[1], true_id);
|
|
|
|
vsir_src_param_init_label(&src_params[2], false_id);
|
|
|
|
false_branch_param = &src_params[2];
|
|
|
|
}
|
|
|
|
vsir_src_param_init_label(&src_params[3], merge_block_id);
|
2023-12-11 21:34:51 -08:00
|
|
|
if (continue_block_id)
|
|
|
|
vsir_src_param_init_label(&src_params[4], continue_block_id);
|
2024-01-04 06:55:51 -08:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2023-12-11 21:34:51 -08:00
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, merge_block_id ? 3 : 1, flattener)))
|
2024-01-04 06:55:51 -08:00
|
|
|
return NULL;
|
|
|
|
vsir_src_param_init_label(&src_params[0], true_id);
|
2023-12-11 21:34:51 -08:00
|
|
|
if (merge_block_id)
|
|
|
|
{
|
|
|
|
/* An unconditional branch may only have merge information for a loop, which
|
|
|
|
* must have both a merge block and continue block. */
|
|
|
|
vsir_src_param_init_label(&src_params[1], merge_block_id);
|
|
|
|
vsir_src_param_init_label(&src_params[2], continue_block_id);
|
|
|
|
}
|
2024-01-04 06:55:51 -08:00
|
|
|
false_branch_param = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
++flattener->instruction_count;
|
|
|
|
|
|
|
|
return false_branch_param;
|
|
|
|
}
|
|
|
|
|
2023-12-11 21:34:51 -08:00
|
|
|
static void cf_flattener_emit_conditional_branch_and_merge(struct cf_flattener *flattener,
|
|
|
|
const struct vkd3d_shader_src_param *condition, unsigned int true_id, unsigned int flags)
|
|
|
|
{
|
|
|
|
unsigned int merge_block_id;
|
|
|
|
|
|
|
|
merge_block_id = cf_flattener_alloc_block_id(flattener);
|
|
|
|
cf_flattener_emit_branch(flattener, merge_block_id, 0, condition, true_id, merge_block_id, flags);
|
|
|
|
cf_flattener_emit_label(flattener, merge_block_id);
|
|
|
|
}
|
|
|
|
|
2024-01-04 06:55:51 -08:00
|
|
|
static void cf_flattener_emit_unconditional_branch(struct cf_flattener *flattener, unsigned int target_block_id)
|
|
|
|
{
|
2023-12-11 21:34:51 -08:00
|
|
|
cf_flattener_emit_branch(flattener, 0, 0, NULL, target_block_id, 0, 0);
|
2024-01-04 06:55:51 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
static struct cf_flattener_info *cf_flattener_push_control_flow_level(struct cf_flattener *flattener)
|
|
|
|
{
|
|
|
|
if (!vkd3d_array_reserve((void **)&flattener->control_flow_info, &flattener->control_flow_info_size,
|
|
|
|
flattener->control_flow_depth + 1, sizeof(*flattener->control_flow_info)))
|
|
|
|
{
|
|
|
|
ERR("Failed to allocate control flow info structure.\n");
|
2024-03-12 14:37:04 -07:00
|
|
|
cf_flattener_set_error(flattener, VKD3D_ERROR_OUT_OF_MEMORY);
|
2024-01-04 06:55:51 -08:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
return &flattener->control_flow_info[flattener->control_flow_depth++];
|
|
|
|
}
|
|
|
|
|
|
|
|
static void cf_flattener_pop_control_flow_level(struct cf_flattener *flattener)
|
|
|
|
{
|
|
|
|
struct cf_flattener_info *cf_info;
|
|
|
|
|
|
|
|
cf_info = &flattener->control_flow_info[--flattener->control_flow_depth];
|
|
|
|
memset(cf_info, 0, sizeof(*cf_info));
|
|
|
|
}
|
|
|
|
|
2023-12-11 21:34:51 -08:00
|
|
|
static struct cf_flattener_info *cf_flattener_find_innermost_loop(struct cf_flattener *flattener)
|
|
|
|
{
|
|
|
|
int depth;
|
|
|
|
|
|
|
|
for (depth = flattener->control_flow_depth - 1; depth >= 0; --depth)
|
|
|
|
{
|
|
|
|
if (flattener->control_flow_info[depth].current_block == VKD3D_BLOCK_LOOP)
|
|
|
|
return &flattener->control_flow_info[depth];
|
|
|
|
}
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct cf_flattener_info *cf_flattener_find_innermost_breakable_cf_construct(struct cf_flattener *flattener)
|
|
|
|
{
|
|
|
|
int depth;
|
|
|
|
|
|
|
|
for (depth = flattener->control_flow_depth - 1; depth >= 0; --depth)
|
|
|
|
{
|
|
|
|
if (flattener->control_flow_info[depth].current_block == VKD3D_BLOCK_LOOP
|
|
|
|
|| flattener->control_flow_info[depth].current_block == VKD3D_BLOCK_SWITCH)
|
|
|
|
return &flattener->control_flow_info[depth];
|
|
|
|
}
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2023-11-07 19:14:01 -08:00
|
|
|
static void VKD3D_PRINTF_FUNC(3, 4) cf_flattener_create_block_name(struct cf_flattener *flattener,
|
|
|
|
unsigned int block_id, const char *fmt, ...)
|
|
|
|
{
|
|
|
|
struct vkd3d_string_buffer buffer;
|
|
|
|
size_t block_name_count;
|
|
|
|
va_list args;
|
|
|
|
|
|
|
|
--block_id;
|
|
|
|
|
|
|
|
block_name_count = max(flattener->block_name_count, block_id + 1);
|
|
|
|
if (!vkd3d_array_reserve((void **)&flattener->block_names, &flattener->block_name_capacity,
|
|
|
|
block_name_count, sizeof(*flattener->block_names)))
|
|
|
|
return;
|
|
|
|
memset(&flattener->block_names[flattener->block_name_count], 0,
|
|
|
|
(block_name_count - flattener->block_name_count) * sizeof(*flattener->block_names));
|
|
|
|
flattener->block_name_count = block_name_count;
|
|
|
|
|
|
|
|
vkd3d_string_buffer_init(&buffer);
|
|
|
|
va_start(args, fmt);
|
|
|
|
vkd3d_string_buffer_vprintf(&buffer, fmt, args);
|
|
|
|
va_end(args);
|
|
|
|
|
|
|
|
flattener->block_names[block_id] = buffer.buffer;
|
|
|
|
}
|
|
|
|
|
2024-03-12 14:37:04 -07:00
|
|
|
static enum vkd3d_result cf_flattener_iterate_instruction_array(struct cf_flattener *flattener,
|
|
|
|
struct vkd3d_shader_message_context *message_context)
|
2024-01-04 06:55:51 -08:00
|
|
|
{
|
2024-01-11 19:20:32 -08:00
|
|
|
bool main_block_open, is_hull_shader, after_declarations_section;
|
2024-01-04 06:55:51 -08:00
|
|
|
struct vkd3d_shader_instruction_array *instructions;
|
2024-03-12 14:37:04 -07:00
|
|
|
struct vsir_program *program = flattener->program;
|
2024-01-04 06:55:51 -08:00
|
|
|
struct vkd3d_shader_instruction *dst_ins;
|
|
|
|
size_t i;
|
|
|
|
|
2024-01-16 07:40:30 -08:00
|
|
|
instructions = &program->instructions;
|
|
|
|
is_hull_shader = program->shader_version.type == VKD3D_SHADER_TYPE_HULL;
|
2024-01-11 19:20:32 -08:00
|
|
|
main_block_open = !is_hull_shader;
|
|
|
|
after_declarations_section = is_hull_shader;
|
2024-01-04 06:55:51 -08:00
|
|
|
|
2024-01-11 19:20:32 -08:00
|
|
|
if (!cf_flattener_require_space(flattener, instructions->count + 1))
|
2024-01-04 06:55:51 -08:00
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
for (i = 0; i < instructions->count; ++i)
|
|
|
|
{
|
2023-12-11 21:34:51 -08:00
|
|
|
unsigned int loop_header_block_id, loop_body_block_id, continue_block_id, merge_block_id, true_block_id;
|
2024-01-04 06:55:51 -08:00
|
|
|
const struct vkd3d_shader_instruction *instruction = &instructions->elements[i];
|
|
|
|
const struct vkd3d_shader_src_param *src = instruction->src;
|
|
|
|
struct cf_flattener_info *cf_info;
|
|
|
|
|
|
|
|
flattener->location = instruction->location;
|
|
|
|
|
2024-01-11 19:20:32 -08:00
|
|
|
/* Declarations should occur before the first code block, which in hull shaders is marked by the first
|
2024-02-20 06:44:23 -08:00
|
|
|
* phase instruction, and in all other shader types begins with the first label instruction.
|
|
|
|
* Declaring an indexable temp with function scope is not considered a declaration,
|
|
|
|
* because it needs to live inside a function. */
|
2024-04-24 11:34:42 -07:00
|
|
|
if (!after_declarations_section && instruction->opcode != VKD3DSIH_NOP)
|
2024-01-11 19:20:32 -08:00
|
|
|
{
|
2024-04-24 11:34:42 -07:00
|
|
|
bool is_function_indexable = instruction->opcode == VKD3DSIH_DCL_INDEXABLE_TEMP
|
2024-02-20 06:44:23 -08:00
|
|
|
&& instruction->declaration.indexable_temp.has_function_scope;
|
|
|
|
|
|
|
|
if (!vsir_instruction_is_dcl(instruction) || is_function_indexable)
|
|
|
|
{
|
|
|
|
after_declarations_section = true;
|
|
|
|
cf_flattener_emit_label(flattener, cf_flattener_alloc_block_id(flattener));
|
|
|
|
}
|
2024-01-11 19:20:32 -08:00
|
|
|
}
|
|
|
|
|
2024-01-04 06:55:51 -08:00
|
|
|
cf_info = flattener->control_flow_depth
|
|
|
|
? &flattener->control_flow_info[flattener->control_flow_depth - 1] : NULL;
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (instruction->opcode)
|
2024-01-04 06:55:51 -08:00
|
|
|
{
|
2024-01-11 19:20:32 -08:00
|
|
|
case VKD3DSIH_HS_CONTROL_POINT_PHASE:
|
|
|
|
case VKD3DSIH_HS_FORK_PHASE:
|
|
|
|
case VKD3DSIH_HS_JOIN_PHASE:
|
|
|
|
if (!cf_flattener_copy_instruction(flattener, instruction))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
2024-04-24 11:34:42 -07:00
|
|
|
if (instruction->opcode != VKD3DSIH_HS_CONTROL_POINT_PHASE || !instruction->flags)
|
2024-01-11 19:20:32 -08:00
|
|
|
after_declarations_section = false;
|
|
|
|
break;
|
|
|
|
|
2024-01-04 06:55:51 -08:00
|
|
|
case VKD3DSIH_LABEL:
|
2024-03-12 14:37:04 -07:00
|
|
|
vkd3d_shader_error(message_context, &instruction->location,
|
|
|
|
VKD3D_SHADER_ERROR_VSIR_NOT_IMPLEMENTED,
|
2024-01-04 06:55:51 -08:00
|
|
|
"Aborting due to not yet implemented feature: Label instruction.");
|
|
|
|
return VKD3D_ERROR_NOT_IMPLEMENTED;
|
|
|
|
|
|
|
|
case VKD3DSIH_IF:
|
|
|
|
if (!(cf_info = cf_flattener_push_control_flow_level(flattener)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
true_block_id = cf_flattener_alloc_block_id(flattener);
|
|
|
|
merge_block_id = cf_flattener_alloc_block_id(flattener);
|
2023-12-11 21:34:51 -08:00
|
|
|
cf_info->u.if_.false_param = cf_flattener_emit_branch(flattener, merge_block_id, 0,
|
2024-01-04 06:55:51 -08:00
|
|
|
src, true_block_id, merge_block_id, instruction->flags);
|
|
|
|
if (!cf_info->u.if_.false_param)
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
cf_flattener_emit_label(flattener, true_block_id);
|
|
|
|
|
2023-11-07 19:14:01 -08:00
|
|
|
cf_info->u.if_.id = flattener->branch_id;
|
2024-01-04 06:55:51 -08:00
|
|
|
cf_info->u.if_.merge_block_id = merge_block_id;
|
|
|
|
cf_info->u.if_.else_block_id = 0;
|
|
|
|
cf_info->inside_block = true;
|
|
|
|
cf_info->current_block = VKD3D_BLOCK_IF;
|
2023-11-07 19:14:01 -08:00
|
|
|
|
|
|
|
cf_flattener_create_block_name(flattener, merge_block_id, "branch%u_merge", flattener->branch_id);
|
|
|
|
cf_flattener_create_block_name(flattener, true_block_id, "branch%u_true", flattener->branch_id);
|
|
|
|
++flattener->branch_id;
|
2024-01-04 06:55:51 -08:00
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_ELSE:
|
|
|
|
if (cf_info->inside_block)
|
|
|
|
cf_flattener_emit_unconditional_branch(flattener, cf_info->u.if_.merge_block_id);
|
|
|
|
|
|
|
|
cf_info->u.if_.else_block_id = cf_flattener_alloc_block_id(flattener);
|
|
|
|
cf_info->u.if_.false_param->reg.idx[0].offset = cf_info->u.if_.else_block_id;
|
|
|
|
|
2023-11-07 19:14:01 -08:00
|
|
|
cf_flattener_create_block_name(flattener,
|
|
|
|
cf_info->u.if_.else_block_id, "branch%u_false", cf_info->u.if_.id);
|
2024-01-04 06:55:51 -08:00
|
|
|
cf_flattener_emit_label(flattener, cf_info->u.if_.else_block_id);
|
|
|
|
|
|
|
|
cf_info->inside_block = true;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_ENDIF:
|
|
|
|
if (cf_info->inside_block)
|
|
|
|
cf_flattener_emit_unconditional_branch(flattener, cf_info->u.if_.merge_block_id);
|
|
|
|
|
|
|
|
cf_flattener_emit_label(flattener, cf_info->u.if_.merge_block_id);
|
|
|
|
|
|
|
|
cf_flattener_pop_control_flow_level(flattener);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_LOOP:
|
2023-12-11 21:34:51 -08:00
|
|
|
if (!(cf_info = cf_flattener_push_control_flow_level(flattener)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
loop_header_block_id = cf_flattener_alloc_block_id(flattener);
|
|
|
|
loop_body_block_id = cf_flattener_alloc_block_id(flattener);
|
|
|
|
continue_block_id = cf_flattener_alloc_block_id(flattener);
|
|
|
|
merge_block_id = cf_flattener_alloc_block_id(flattener);
|
|
|
|
|
|
|
|
cf_flattener_emit_unconditional_branch(flattener, loop_header_block_id);
|
|
|
|
cf_flattener_emit_label(flattener, loop_header_block_id);
|
|
|
|
cf_flattener_emit_branch(flattener, merge_block_id, continue_block_id,
|
|
|
|
NULL, loop_body_block_id, 0, 0);
|
|
|
|
|
|
|
|
cf_flattener_emit_label(flattener, loop_body_block_id);
|
|
|
|
|
|
|
|
cf_info->u.loop.header_block_id = loop_header_block_id;
|
|
|
|
cf_info->u.loop.continue_block_id = continue_block_id;
|
|
|
|
cf_info->u.loop.merge_block_id = merge_block_id;
|
|
|
|
cf_info->current_block = VKD3D_BLOCK_LOOP;
|
|
|
|
cf_info->inside_block = true;
|
2023-11-07 19:14:01 -08:00
|
|
|
|
|
|
|
cf_flattener_create_block_name(flattener, loop_header_block_id, "loop%u_header", flattener->loop_id);
|
|
|
|
cf_flattener_create_block_name(flattener, loop_body_block_id, "loop%u_body", flattener->loop_id);
|
|
|
|
cf_flattener_create_block_name(flattener, continue_block_id, "loop%u_continue", flattener->loop_id);
|
|
|
|
cf_flattener_create_block_name(flattener, merge_block_id, "loop%u_merge", flattener->loop_id);
|
|
|
|
++flattener->loop_id;
|
2023-12-11 21:34:51 -08:00
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_ENDLOOP:
|
|
|
|
if (cf_info->inside_block)
|
|
|
|
cf_flattener_emit_unconditional_branch(flattener, cf_info->u.loop.continue_block_id);
|
|
|
|
|
|
|
|
cf_flattener_emit_label(flattener, cf_info->u.loop.continue_block_id);
|
|
|
|
cf_flattener_emit_unconditional_branch(flattener, cf_info->u.loop.header_block_id);
|
|
|
|
cf_flattener_emit_label(flattener, cf_info->u.loop.merge_block_id);
|
|
|
|
|
|
|
|
cf_flattener_pop_control_flow_level(flattener);
|
|
|
|
break;
|
|
|
|
|
2024-01-04 06:55:51 -08:00
|
|
|
case VKD3DSIH_SWITCH:
|
2023-12-11 21:34:51 -08:00
|
|
|
if (!(cf_info = cf_flattener_push_control_flow_level(flattener)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
2023-12-11 21:58:08 -08:00
|
|
|
merge_block_id = cf_flattener_alloc_block_id(flattener);
|
|
|
|
|
|
|
|
cf_info->u.switch_.ins_location = flattener->instruction_count;
|
|
|
|
cf_info->u.switch_.condition = src;
|
2023-12-11 21:34:51 -08:00
|
|
|
|
2023-12-11 21:58:08 -08:00
|
|
|
if (!(dst_ins = cf_flattener_require_space(flattener, 1)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
vsir_instruction_init(dst_ins, &instruction->location, VKD3DSIH_SWITCH_MONOLITHIC);
|
|
|
|
++flattener->instruction_count;
|
|
|
|
|
2023-11-07 19:14:01 -08:00
|
|
|
cf_info->u.switch_.id = flattener->switch_id;
|
2023-12-11 21:58:08 -08:00
|
|
|
cf_info->u.switch_.merge_block_id = merge_block_id;
|
|
|
|
cf_info->u.switch_.cases = NULL;
|
|
|
|
cf_info->u.switch_.cases_size = 0;
|
|
|
|
cf_info->u.switch_.cases_count = 0;
|
|
|
|
cf_info->u.switch_.default_block_id = 0;
|
|
|
|
cf_info->inside_block = false;
|
2023-12-11 21:34:51 -08:00
|
|
|
cf_info->current_block = VKD3D_BLOCK_SWITCH;
|
|
|
|
|
2023-11-07 19:14:01 -08:00
|
|
|
cf_flattener_create_block_name(flattener, merge_block_id, "switch%u_merge", flattener->switch_id);
|
|
|
|
++flattener->switch_id;
|
|
|
|
|
2023-12-11 21:58:08 -08:00
|
|
|
if (!vkd3d_array_reserve((void **)&cf_info->u.switch_.cases, &cf_info->u.switch_.cases_size,
|
|
|
|
10, sizeof(*cf_info->u.switch_.cases)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
2024-01-04 06:55:51 -08:00
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_ENDSWITCH:
|
2023-12-11 21:58:08 -08:00
|
|
|
{
|
|
|
|
struct vkd3d_shader_src_param *src_params;
|
|
|
|
unsigned int j;
|
|
|
|
|
|
|
|
if (!cf_info->u.switch_.default_block_id)
|
|
|
|
cf_info->u.switch_.default_block_id = cf_info->u.switch_.merge_block_id;
|
|
|
|
|
|
|
|
cf_flattener_emit_label(flattener, cf_info->u.switch_.merge_block_id);
|
|
|
|
|
|
|
|
/* The SWITCH instruction is completed when the endswitch
|
|
|
|
* instruction is processed because we do not know the number
|
|
|
|
* of case statements or the default block id in advance.*/
|
|
|
|
dst_ins = &flattener->instructions[cf_info->u.switch_.ins_location];
|
|
|
|
if (!(src_params = instruction_src_params_alloc(dst_ins, cf_info->u.switch_.cases_count * 2 + 3, flattener)))
|
|
|
|
{
|
|
|
|
vkd3d_free(cf_info->u.switch_.cases);
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
}
|
|
|
|
src_params[0] = *cf_info->u.switch_.condition;
|
|
|
|
vsir_src_param_init_label(&src_params[1], cf_info->u.switch_.default_block_id);
|
|
|
|
vsir_src_param_init_label(&src_params[2], cf_info->u.switch_.merge_block_id);
|
|
|
|
for (j = 0; j < cf_info->u.switch_.cases_count; ++j)
|
|
|
|
{
|
|
|
|
unsigned int index = j * 2 + 3;
|
|
|
|
vsir_src_param_init(&src_params[index], VKD3DSPR_IMMCONST, VKD3D_DATA_UINT, 0);
|
|
|
|
src_params[index].reg.u.immconst_u32[0] = cf_info->u.switch_.cases[j].value;
|
|
|
|
vsir_src_param_init_label(&src_params[index + 1], cf_info->u.switch_.cases[j].block_id);
|
|
|
|
}
|
|
|
|
vkd3d_free(cf_info->u.switch_.cases);
|
|
|
|
|
2023-12-11 21:34:51 -08:00
|
|
|
cf_flattener_pop_control_flow_level(flattener);
|
2023-12-11 21:58:08 -08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case VKD3DSIH_CASE:
|
|
|
|
{
|
|
|
|
unsigned int label_id, value;
|
|
|
|
|
|
|
|
if (src->swizzle != VKD3D_SHADER_SWIZZLE(X, X, X, X))
|
|
|
|
{
|
|
|
|
WARN("Unexpected src swizzle %#x.\n", src->swizzle);
|
2024-03-12 14:37:04 -07:00
|
|
|
vkd3d_shader_error(message_context, &instruction->location,
|
|
|
|
VKD3D_SHADER_ERROR_VSIR_INVALID_SWIZZLE,
|
2023-12-11 21:58:08 -08:00
|
|
|
"The swizzle for a switch case value is not scalar X.");
|
2024-03-12 14:37:04 -07:00
|
|
|
cf_flattener_set_error(flattener, VKD3D_ERROR_INVALID_SHADER);
|
2023-12-11 21:58:08 -08:00
|
|
|
}
|
|
|
|
value = *src->reg.u.immconst_u32;
|
2023-12-11 21:34:51 -08:00
|
|
|
|
2023-12-11 21:58:08 -08:00
|
|
|
if (!vkd3d_array_reserve((void **)&cf_info->u.switch_.cases, &cf_info->u.switch_.cases_size,
|
|
|
|
cf_info->u.switch_.cases_count + 1, sizeof(*cf_info->u.switch_.cases)))
|
2024-01-04 06:55:51 -08:00
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
2023-12-11 21:58:08 -08:00
|
|
|
|
|
|
|
label_id = cf_flattener_alloc_block_id(flattener);
|
|
|
|
if (cf_info->inside_block) /* fall-through */
|
|
|
|
cf_flattener_emit_unconditional_branch(flattener, label_id);
|
|
|
|
|
|
|
|
cf_info->u.switch_.cases[cf_info->u.switch_.cases_count].value = value;
|
|
|
|
cf_info->u.switch_.cases[cf_info->u.switch_.cases_count].block_id = label_id;
|
|
|
|
++cf_info->u.switch_.cases_count;
|
|
|
|
|
|
|
|
cf_flattener_emit_label(flattener, label_id);
|
2023-11-07 19:14:01 -08:00
|
|
|
cf_flattener_create_block_name(flattener, label_id, "switch%u_case%u", cf_info->u.switch_.id, value);
|
2023-12-11 21:58:08 -08:00
|
|
|
cf_info->inside_block = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case VKD3DSIH_DEFAULT:
|
|
|
|
cf_info->u.switch_.default_block_id = cf_flattener_alloc_block_id(flattener);
|
|
|
|
if (cf_info->inside_block) /* fall-through */
|
|
|
|
cf_flattener_emit_unconditional_branch(flattener, cf_info->u.switch_.default_block_id);
|
|
|
|
|
|
|
|
cf_flattener_emit_label(flattener, cf_info->u.switch_.default_block_id);
|
2023-11-07 19:14:01 -08:00
|
|
|
|
|
|
|
cf_flattener_create_block_name(flattener, cf_info->u.switch_.default_block_id,
|
|
|
|
"switch%u_default", cf_info->u.switch_.id);
|
2023-12-11 21:58:08 -08:00
|
|
|
cf_info->inside_block = true;
|
2024-01-04 06:55:51 -08:00
|
|
|
break;
|
|
|
|
|
2023-12-11 21:34:51 -08:00
|
|
|
case VKD3DSIH_BREAK:
|
|
|
|
{
|
|
|
|
struct cf_flattener_info *breakable_cf_info;
|
|
|
|
|
|
|
|
if (!(breakable_cf_info = cf_flattener_find_innermost_breakable_cf_construct(flattener)))
|
|
|
|
{
|
|
|
|
FIXME("Unhandled break instruction.\n");
|
|
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (breakable_cf_info->current_block == VKD3D_BLOCK_LOOP)
|
|
|
|
{
|
|
|
|
cf_flattener_emit_unconditional_branch(flattener, breakable_cf_info->u.loop.merge_block_id);
|
|
|
|
}
|
|
|
|
else if (breakable_cf_info->current_block == VKD3D_BLOCK_SWITCH)
|
|
|
|
{
|
2023-12-11 21:58:08 -08:00
|
|
|
cf_flattener_emit_unconditional_branch(flattener, breakable_cf_info->u.switch_.merge_block_id);
|
2023-12-11 21:34:51 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
cf_info->inside_block = false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case VKD3DSIH_BREAKP:
|
|
|
|
{
|
|
|
|
struct cf_flattener_info *loop_cf_info;
|
|
|
|
|
|
|
|
if (!(loop_cf_info = cf_flattener_find_innermost_loop(flattener)))
|
|
|
|
{
|
|
|
|
ERR("Invalid 'breakc' instruction outside loop.\n");
|
|
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
}
|
|
|
|
|
|
|
|
cf_flattener_emit_conditional_branch_and_merge(flattener,
|
|
|
|
src, loop_cf_info->u.loop.merge_block_id, instruction->flags);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case VKD3DSIH_CONTINUE:
|
|
|
|
{
|
|
|
|
struct cf_flattener_info *loop_cf_info;
|
|
|
|
|
|
|
|
if (!(loop_cf_info = cf_flattener_find_innermost_loop(flattener)))
|
|
|
|
{
|
|
|
|
ERR("Invalid 'continue' instruction outside loop.\n");
|
|
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
}
|
|
|
|
|
|
|
|
cf_flattener_emit_unconditional_branch(flattener, loop_cf_info->u.loop.continue_block_id);
|
|
|
|
|
|
|
|
cf_info->inside_block = false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case VKD3DSIH_CONTINUEP:
|
|
|
|
{
|
|
|
|
struct cf_flattener_info *loop_cf_info;
|
|
|
|
|
|
|
|
if (!(loop_cf_info = cf_flattener_find_innermost_loop(flattener)))
|
|
|
|
{
|
|
|
|
ERR("Invalid 'continuec' instruction outside loop.\n");
|
|
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
}
|
|
|
|
|
|
|
|
cf_flattener_emit_conditional_branch_and_merge(flattener,
|
|
|
|
src, loop_cf_info->u.loop.continue_block_id, instruction->flags);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2024-01-04 06:55:51 -08:00
|
|
|
case VKD3DSIH_RET:
|
|
|
|
if (!cf_flattener_copy_instruction(flattener, instruction))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
if (cf_info)
|
|
|
|
cf_info->inside_block = false;
|
2023-12-11 21:34:51 -08:00
|
|
|
else
|
2024-01-04 06:55:51 -08:00
|
|
|
main_block_open = false;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
if (!cf_flattener_copy_instruction(flattener, instruction))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (main_block_open)
|
|
|
|
{
|
|
|
|
if (!(dst_ins = cf_flattener_require_space(flattener, 1)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
vsir_instruction_init(dst_ins, &flattener->location, VKD3DSIH_RET);
|
|
|
|
++flattener->instruction_count;
|
|
|
|
}
|
|
|
|
|
2024-03-12 14:37:04 -07:00
|
|
|
return flattener->status;
|
2024-01-04 06:55:51 -08:00
|
|
|
}
|
|
|
|
|
2024-03-12 14:37:04 -07:00
|
|
|
static enum vkd3d_result vsir_program_flatten_control_flow_constructs(struct vsir_program *program,
|
|
|
|
struct vkd3d_shader_message_context *message_context)
|
2024-01-04 06:55:51 -08:00
|
|
|
{
|
2024-03-12 14:37:04 -07:00
|
|
|
struct cf_flattener flattener = {.program = program};
|
2024-01-04 06:55:51 -08:00
|
|
|
enum vkd3d_result result;
|
|
|
|
|
2024-03-12 14:37:04 -07:00
|
|
|
if ((result = cf_flattener_iterate_instruction_array(&flattener, message_context)) >= 0)
|
2024-01-04 06:55:51 -08:00
|
|
|
{
|
2024-03-12 14:37:04 -07:00
|
|
|
vkd3d_free(program->instructions.elements);
|
2024-01-18 11:34:52 -08:00
|
|
|
program->instructions.elements = flattener.instructions;
|
|
|
|
program->instructions.capacity = flattener.instruction_capacity;
|
|
|
|
program->instructions.count = flattener.instruction_count;
|
|
|
|
program->block_count = flattener.block_id;
|
2024-01-04 06:55:51 -08:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
vkd3d_free(flattener.instructions);
|
|
|
|
}
|
|
|
|
|
|
|
|
vkd3d_free(flattener.control_flow_info);
|
2024-01-24 12:25:38 -08:00
|
|
|
/* Simpler to always free these in vsir_program_cleanup(). */
|
2024-01-18 11:41:27 -08:00
|
|
|
program->block_names = flattener.block_names;
|
|
|
|
program->block_name_count = flattener.block_name_count;
|
2024-01-04 06:55:51 -08:00
|
|
|
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2024-01-03 07:18:50 -08:00
|
|
|
static unsigned int label_from_src_param(const struct vkd3d_shader_src_param *param)
|
|
|
|
{
|
|
|
|
assert(param->reg.type == VKD3DSPR_LABEL);
|
|
|
|
return param->reg.idx[0].offset;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool reserve_instructions(struct vkd3d_shader_instruction **instructions, size_t *capacity, size_t count)
|
|
|
|
{
|
|
|
|
if (!vkd3d_array_reserve((void **)instructions, capacity, count, sizeof(**instructions)))
|
|
|
|
{
|
|
|
|
ERR("Failed to allocate instructions.\n");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2024-01-15 08:02:07 -08:00
|
|
|
/* A record represents replacing a jump from block `switch_label' to
|
|
|
|
* block `target_label' with a jump from block `if_label' to block
|
|
|
|
* `target_label'. */
|
|
|
|
struct lower_switch_to_if_ladder_block_mapping
|
|
|
|
{
|
|
|
|
unsigned int switch_label;
|
|
|
|
unsigned int if_label;
|
|
|
|
unsigned int target_label;
|
|
|
|
};
|
|
|
|
|
|
|
|
static bool lower_switch_to_if_ladder_add_block_mapping(struct lower_switch_to_if_ladder_block_mapping **block_map,
|
|
|
|
size_t *map_capacity, size_t *map_count, unsigned int switch_label, unsigned int if_label, unsigned int target_label)
|
|
|
|
{
|
|
|
|
if (!vkd3d_array_reserve((void **)block_map, map_capacity, *map_count + 1, sizeof(**block_map)))
|
|
|
|
{
|
|
|
|
ERR("Failed to allocate block mapping.\n");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
(*block_map)[*map_count].switch_label = switch_label;
|
|
|
|
(*block_map)[*map_count].if_label = if_label;
|
|
|
|
(*block_map)[*map_count].target_label = target_label;
|
|
|
|
|
|
|
|
*map_count += 1;
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2024-01-24 05:26:35 -08:00
|
|
|
static enum vkd3d_result lower_switch_to_if_ladder(struct vsir_program *program)
|
2024-01-03 07:18:50 -08:00
|
|
|
{
|
2024-01-24 05:26:35 -08:00
|
|
|
unsigned int block_count = program->block_count, ssa_count = program->ssa_count, current_label = 0, if_label;
|
2024-01-15 08:02:07 -08:00
|
|
|
size_t ins_capacity = 0, ins_count = 0, i, map_capacity = 0, map_count = 0;
|
2024-01-03 07:18:50 -08:00
|
|
|
struct vkd3d_shader_instruction *instructions = NULL;
|
2024-01-15 08:02:07 -08:00
|
|
|
struct lower_switch_to_if_ladder_block_mapping *block_map = NULL;
|
2024-01-03 07:18:50 -08:00
|
|
|
|
2024-01-24 05:26:35 -08:00
|
|
|
if (!reserve_instructions(&instructions, &ins_capacity, program->instructions.count))
|
2024-01-03 07:18:50 -08:00
|
|
|
goto fail;
|
|
|
|
|
2024-01-15 08:02:07 -08:00
|
|
|
/* First subpass: convert SWITCH_MONOLITHIC instructions to
|
|
|
|
* selection ladders, keeping a map between blocks before and
|
|
|
|
* after the subpass. */
|
2024-01-24 05:26:35 -08:00
|
|
|
for (i = 0; i < program->instructions.count; ++i)
|
2024-01-03 07:18:50 -08:00
|
|
|
{
|
2024-01-24 05:26:35 -08:00
|
|
|
struct vkd3d_shader_instruction *ins = &program->instructions.elements[i];
|
2024-01-03 07:18:50 -08:00
|
|
|
unsigned int case_count, j, default_label;
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (ins->opcode)
|
2024-01-03 07:18:50 -08:00
|
|
|
{
|
2024-01-15 08:02:07 -08:00
|
|
|
case VKD3DSIH_LABEL:
|
|
|
|
current_label = label_from_src_param(&ins->src[0]);
|
|
|
|
if (!reserve_instructions(&instructions, &ins_capacity, ins_count + 1))
|
|
|
|
goto fail;
|
|
|
|
instructions[ins_count++] = *ins;
|
|
|
|
continue;
|
|
|
|
|
2024-01-03 07:18:50 -08:00
|
|
|
case VKD3DSIH_SWITCH_MONOLITHIC:
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
if (!reserve_instructions(&instructions, &ins_capacity, ins_count + 1))
|
|
|
|
goto fail;
|
|
|
|
instructions[ins_count++] = *ins;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
case_count = (ins->src_count - 3) / 2;
|
|
|
|
default_label = label_from_src_param(&ins->src[1]);
|
|
|
|
|
|
|
|
/* In principle we can have a switch with no cases, and we
|
|
|
|
* just have to jump to the default label. */
|
|
|
|
if (case_count == 0)
|
|
|
|
{
|
|
|
|
if (!reserve_instructions(&instructions, &ins_capacity, ins_count + 1))
|
|
|
|
goto fail;
|
|
|
|
|
2024-01-24 05:26:35 -08:00
|
|
|
if (!vsir_instruction_init_with_params(program, &instructions[ins_count],
|
|
|
|
&ins->location, VKD3DSIH_BRANCH, 0, 1))
|
2024-01-03 07:18:50 -08:00
|
|
|
goto fail;
|
|
|
|
vsir_src_param_init_label(&instructions[ins_count].src[0], default_label);
|
|
|
|
++ins_count;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!reserve_instructions(&instructions, &ins_capacity, ins_count + 3 * case_count - 1))
|
|
|
|
goto fail;
|
|
|
|
|
2024-01-15 08:02:07 -08:00
|
|
|
if_label = current_label;
|
|
|
|
|
2024-01-03 07:18:50 -08:00
|
|
|
for (j = 0; j < case_count; ++j)
|
|
|
|
{
|
|
|
|
unsigned int fallthrough_label, case_label = label_from_src_param(&ins->src[3 + 2 * j + 1]);
|
|
|
|
|
2024-01-24 05:26:35 -08:00
|
|
|
if (!vsir_instruction_init_with_params(program,
|
|
|
|
&instructions[ins_count], &ins->location, VKD3DSIH_IEQ, 1, 2))
|
2024-01-03 07:18:50 -08:00
|
|
|
goto fail;
|
|
|
|
dst_param_init_ssa_bool(&instructions[ins_count].dst[0], ssa_count);
|
|
|
|
instructions[ins_count].src[0] = ins->src[0];
|
|
|
|
instructions[ins_count].src[1] = ins->src[3 + 2 * j];
|
|
|
|
++ins_count;
|
|
|
|
|
|
|
|
/* For all cases except the last one we fall through to
|
|
|
|
* the following case; the last one has to jump to the
|
|
|
|
* default label. */
|
|
|
|
if (j == case_count - 1)
|
|
|
|
fallthrough_label = default_label;
|
|
|
|
else
|
|
|
|
fallthrough_label = block_count + 1;
|
|
|
|
|
2024-01-24 05:26:35 -08:00
|
|
|
if (!vsir_instruction_init_with_params(program, &instructions[ins_count],
|
|
|
|
&ins->location, VKD3DSIH_BRANCH, 0, 3))
|
2024-01-03 07:18:50 -08:00
|
|
|
goto fail;
|
|
|
|
src_param_init_ssa_bool(&instructions[ins_count].src[0], ssa_count);
|
|
|
|
vsir_src_param_init_label(&instructions[ins_count].src[1], case_label);
|
|
|
|
vsir_src_param_init_label(&instructions[ins_count].src[2], fallthrough_label);
|
|
|
|
++ins_count;
|
|
|
|
|
|
|
|
++ssa_count;
|
|
|
|
|
2024-01-15 08:02:07 -08:00
|
|
|
if (!lower_switch_to_if_ladder_add_block_mapping(&block_map, &map_capacity, &map_count,
|
|
|
|
current_label, if_label, case_label))
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
if (j == case_count - 1)
|
|
|
|
{
|
|
|
|
if (!lower_switch_to_if_ladder_add_block_mapping(&block_map, &map_capacity, &map_count,
|
|
|
|
current_label, if_label, default_label))
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
else
|
2024-01-03 07:18:50 -08:00
|
|
|
{
|
2024-01-24 05:26:35 -08:00
|
|
|
if (!vsir_instruction_init_with_params(program,
|
|
|
|
&instructions[ins_count], &ins->location, VKD3DSIH_LABEL, 0, 1))
|
2024-01-03 07:18:50 -08:00
|
|
|
goto fail;
|
|
|
|
vsir_src_param_init_label(&instructions[ins_count].src[0], ++block_count);
|
|
|
|
++ins_count;
|
2024-01-15 08:02:07 -08:00
|
|
|
|
|
|
|
if_label = block_count;
|
2024-01-03 07:18:50 -08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-01-24 05:26:35 -08:00
|
|
|
vkd3d_free(program->instructions.elements);
|
2024-01-15 08:02:07 -08:00
|
|
|
vkd3d_free(block_map);
|
2024-01-24 05:26:35 -08:00
|
|
|
program->instructions.elements = instructions;
|
|
|
|
program->instructions.capacity = ins_capacity;
|
|
|
|
program->instructions.count = ins_count;
|
|
|
|
program->block_count = block_count;
|
|
|
|
program->ssa_count = ssa_count;
|
2024-01-03 07:18:50 -08:00
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
vkd3d_free(instructions);
|
2024-01-15 08:02:07 -08:00
|
|
|
vkd3d_free(block_map);
|
2024-01-03 07:18:50 -08:00
|
|
|
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
}
|
|
|
|
|
2024-04-01 19:55:05 -07:00
|
|
|
struct ssas_to_temps_alloc
|
|
|
|
{
|
|
|
|
unsigned int *table;
|
|
|
|
unsigned int next_temp_idx;
|
|
|
|
};
|
|
|
|
|
|
|
|
static bool ssas_to_temps_alloc_init(struct ssas_to_temps_alloc *alloc, unsigned int ssa_count, unsigned int temp_count)
|
|
|
|
{
|
|
|
|
size_t i = ssa_count * sizeof(*alloc->table);
|
|
|
|
|
|
|
|
if (!(alloc->table = vkd3d_malloc(i)))
|
|
|
|
{
|
|
|
|
ERR("Failed to allocate SSA table.\n");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
memset(alloc->table, 0xff, i);
|
|
|
|
|
|
|
|
alloc->next_temp_idx = temp_count;
|
|
|
|
return true;
|
|
|
|
}
|
2024-01-09 06:56:45 -08:00
|
|
|
|
|
|
|
/* This is idempotent: it can be safely applied more than once on the
|
|
|
|
* same register. */
|
2024-04-01 19:55:05 -07:00
|
|
|
static void materialize_ssas_to_temps_process_reg(struct vsir_program *program, struct ssas_to_temps_alloc *alloc,
|
|
|
|
struct vkd3d_shader_register *reg)
|
2024-01-09 06:56:45 -08:00
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
2024-04-01 19:55:05 -07:00
|
|
|
if (reg->type == VKD3DSPR_SSA && alloc->table[reg->idx[0].offset] != UINT_MAX)
|
2024-01-09 06:56:45 -08:00
|
|
|
{
|
|
|
|
reg->type = VKD3DSPR_TEMP;
|
2024-04-01 19:55:05 -07:00
|
|
|
reg->idx[0].offset = alloc->table[reg->idx[0].offset];
|
2024-01-09 06:56:45 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < reg->idx_count; ++i)
|
|
|
|
if (reg->idx[i].rel_addr)
|
2024-04-01 19:55:05 -07:00
|
|
|
materialize_ssas_to_temps_process_reg(program, alloc, ®->idx[i].rel_addr->reg);
|
2024-01-09 06:56:45 -08:00
|
|
|
}
|
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
struct ssas_to_temps_block_info
|
2024-01-09 06:56:45 -08:00
|
|
|
{
|
2024-04-02 17:35:45 -07:00
|
|
|
struct phi_incoming_to_temp
|
2024-01-09 06:56:45 -08:00
|
|
|
{
|
2024-04-02 17:35:45 -07:00
|
|
|
struct vkd3d_shader_src_param *src;
|
|
|
|
struct vkd3d_shader_dst_param *dst;
|
|
|
|
} *incomings;
|
|
|
|
size_t incoming_capacity;
|
|
|
|
size_t incoming_count;
|
|
|
|
};
|
2024-01-09 06:56:45 -08:00
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
static void ssas_to_temps_block_info_cleanup(struct ssas_to_temps_block_info *block_info,
|
|
|
|
size_t count)
|
2024-01-09 06:56:45 -08:00
|
|
|
{
|
2024-04-02 17:35:45 -07:00
|
|
|
size_t i;
|
2024-01-09 06:56:45 -08:00
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
for (i = 0; i < count; ++i)
|
|
|
|
vkd3d_free(block_info[i].incomings);
|
2024-01-09 06:56:45 -08:00
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
vkd3d_free(block_info);
|
2024-01-09 06:56:45 -08:00
|
|
|
}
|
|
|
|
|
2024-04-01 19:55:05 -07:00
|
|
|
static enum vkd3d_result vsir_program_materialise_phi_ssas_to_temps(struct vsir_program *program)
|
2024-01-09 06:56:45 -08:00
|
|
|
{
|
2024-04-02 17:35:45 -07:00
|
|
|
size_t ins_capacity = 0, ins_count = 0, phi_count, incoming_count, i;
|
|
|
|
struct ssas_to_temps_block_info *info, *block_info = NULL;
|
2024-01-09 06:56:45 -08:00
|
|
|
struct vkd3d_shader_instruction *instructions = NULL;
|
2024-04-01 19:55:05 -07:00
|
|
|
struct ssas_to_temps_alloc alloc = {0};
|
2024-01-09 06:56:45 -08:00
|
|
|
unsigned int current_label = 0;
|
2024-01-09 06:56:45 -08:00
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
if (!(block_info = vkd3d_calloc(program->block_count, sizeof(*block_info))))
|
2024-01-09 06:56:45 -08:00
|
|
|
{
|
2024-04-02 17:35:45 -07:00
|
|
|
ERR("Failed to allocate block info array.\n");
|
2024-01-09 06:56:45 -08:00
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
2024-04-01 19:55:05 -07:00
|
|
|
if (!ssas_to_temps_alloc_init(&alloc, program->ssa_count, program->temp_count))
|
|
|
|
goto fail;
|
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
for (i = 0, phi_count = 0, incoming_count = 0; i < program->instructions.count; ++i)
|
2024-01-09 06:56:45 -08:00
|
|
|
{
|
2024-03-12 11:40:45 -07:00
|
|
|
struct vkd3d_shader_instruction *ins = &program->instructions.elements[i];
|
2024-04-01 19:55:05 -07:00
|
|
|
unsigned int j, temp_idx;
|
2024-01-09 06:56:45 -08:00
|
|
|
|
2024-04-01 19:55:05 -07:00
|
|
|
/* Only phi src/dst SSA values need be converted here. Structurisation may
|
|
|
|
* introduce new cases of undominated SSA use, which will be handled later. */
|
2024-04-24 11:34:42 -07:00
|
|
|
if (ins->opcode != VKD3DSIH_PHI)
|
2024-04-02 17:35:45 -07:00
|
|
|
continue;
|
|
|
|
++phi_count;
|
|
|
|
|
2024-04-01 19:55:05 -07:00
|
|
|
temp_idx = alloc.next_temp_idx++;
|
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
for (j = 0; j < ins->src_count; j += 2)
|
2024-01-09 06:56:45 -08:00
|
|
|
{
|
2024-04-02 17:35:45 -07:00
|
|
|
struct phi_incoming_to_temp *incoming;
|
|
|
|
unsigned int label;
|
2024-01-09 06:56:45 -08:00
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
label = label_from_src_param(&ins->src[j + 1]);
|
|
|
|
assert(label);
|
2024-01-09 06:56:45 -08:00
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
info = &block_info[label - 1];
|
|
|
|
|
|
|
|
if (!(vkd3d_array_reserve((void **)&info->incomings, &info->incoming_capacity, info->incoming_count + 1,
|
|
|
|
sizeof(*info->incomings))))
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
incoming = &info->incomings[info->incoming_count++];
|
|
|
|
incoming->src = &ins->src[j];
|
|
|
|
incoming->dst = ins->dst;
|
|
|
|
|
2024-04-01 19:55:05 -07:00
|
|
|
alloc.table[ins->dst->reg.idx[0].offset] = temp_idx;
|
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
++incoming_count;
|
2024-01-09 06:56:45 -08:00
|
|
|
}
|
2024-04-02 17:35:45 -07:00
|
|
|
|
2024-04-01 19:55:05 -07:00
|
|
|
materialize_ssas_to_temps_process_reg(program, &alloc, &ins->dst->reg);
|
2024-01-09 06:56:45 -08:00
|
|
|
}
|
|
|
|
|
2024-04-01 19:55:05 -07:00
|
|
|
if (!phi_count)
|
|
|
|
goto done;
|
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
if (!reserve_instructions(&instructions, &ins_capacity, program->instructions.count + incoming_count - phi_count))
|
|
|
|
goto fail;
|
|
|
|
|
2024-03-12 11:40:45 -07:00
|
|
|
for (i = 0; i < program->instructions.count; ++i)
|
2024-01-09 06:56:45 -08:00
|
|
|
{
|
2024-04-02 17:35:45 -07:00
|
|
|
struct vkd3d_shader_instruction *mov_ins, *ins = &program->instructions.elements[i];
|
2024-01-09 06:56:45 -08:00
|
|
|
size_t j;
|
2024-01-09 06:56:45 -08:00
|
|
|
|
|
|
|
for (j = 0; j < ins->dst_count; ++j)
|
2024-04-01 19:55:05 -07:00
|
|
|
materialize_ssas_to_temps_process_reg(program, &alloc, &ins->dst[j].reg);
|
2024-01-09 06:56:45 -08:00
|
|
|
|
|
|
|
for (j = 0; j < ins->src_count; ++j)
|
2024-04-01 19:55:05 -07:00
|
|
|
materialize_ssas_to_temps_process_reg(program, &alloc, &ins->src[j].reg);
|
2024-01-09 06:56:45 -08:00
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (ins->opcode)
|
2024-01-09 06:56:45 -08:00
|
|
|
{
|
|
|
|
case VKD3DSIH_LABEL:
|
|
|
|
current_label = label_from_src_param(&ins->src[0]);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_BRANCH:
|
2024-04-02 19:20:09 -07:00
|
|
|
case VKD3DSIH_SWITCH_MONOLITHIC:
|
2024-04-02 17:35:45 -07:00
|
|
|
info = &block_info[current_label - 1];
|
2024-01-09 06:56:45 -08:00
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
for (j = 0; j < info->incoming_count; ++j)
|
2024-01-09 06:56:45 -08:00
|
|
|
{
|
2024-04-02 17:35:45 -07:00
|
|
|
struct phi_incoming_to_temp *incoming = &info->incomings[j];
|
2024-01-09 06:56:45 -08:00
|
|
|
|
2024-04-02 17:35:45 -07:00
|
|
|
mov_ins = &instructions[ins_count++];
|
|
|
|
if (!vsir_instruction_init_with_params(program, mov_ins, &ins->location, VKD3DSIH_MOV, 1, 0))
|
2024-01-09 06:56:45 -08:00
|
|
|
goto fail;
|
2024-04-02 17:35:45 -07:00
|
|
|
*mov_ins->dst = *incoming->dst;
|
|
|
|
mov_ins->src = incoming->src;
|
|
|
|
mov_ins->src_count = 1;
|
2024-01-09 06:56:45 -08:00
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_PHI:
|
|
|
|
continue;
|
|
|
|
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2024-01-09 06:56:45 -08:00
|
|
|
instructions[ins_count++] = *ins;
|
|
|
|
}
|
|
|
|
|
2024-03-12 11:40:45 -07:00
|
|
|
vkd3d_free(program->instructions.elements);
|
|
|
|
program->instructions.elements = instructions;
|
|
|
|
program->instructions.capacity = ins_capacity;
|
|
|
|
program->instructions.count = ins_count;
|
2024-04-01 19:55:05 -07:00
|
|
|
program->temp_count = alloc.next_temp_idx;
|
|
|
|
done:
|
2024-04-02 17:35:45 -07:00
|
|
|
ssas_to_temps_block_info_cleanup(block_info, program->block_count);
|
2024-04-01 19:55:05 -07:00
|
|
|
vkd3d_free(alloc.table);
|
2024-04-02 17:35:45 -07:00
|
|
|
|
2024-01-09 06:56:45 -08:00
|
|
|
return VKD3D_OK;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
vkd3d_free(instructions);
|
2024-04-02 17:35:45 -07:00
|
|
|
ssas_to_temps_block_info_cleanup(block_info, program->block_count);
|
2024-04-01 19:55:05 -07:00
|
|
|
vkd3d_free(alloc.table);
|
2024-01-09 06:56:45 -08:00
|
|
|
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
}
|
|
|
|
|
2024-01-29 07:41:31 -08:00
|
|
|
struct vsir_block_list
|
|
|
|
{
|
|
|
|
struct vsir_block **blocks;
|
|
|
|
size_t count, capacity;
|
|
|
|
};
|
|
|
|
|
|
|
|
static void vsir_block_list_init(struct vsir_block_list *list)
|
|
|
|
{
|
|
|
|
memset(list, 0, sizeof(*list));
|
|
|
|
}
|
|
|
|
|
|
|
|
static void vsir_block_list_cleanup(struct vsir_block_list *list)
|
|
|
|
{
|
|
|
|
vkd3d_free(list->blocks);
|
|
|
|
}
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
static enum vkd3d_result vsir_block_list_add_checked(struct vsir_block_list *list, struct vsir_block *block)
|
2024-01-29 07:41:31 -08:00
|
|
|
{
|
|
|
|
if (!vkd3d_array_reserve((void **)&list->blocks, &list->capacity, list->count + 1, sizeof(*list->blocks)))
|
|
|
|
{
|
|
|
|
ERR("Cannot extend block list.\n");
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
}
|
|
|
|
|
|
|
|
list->blocks[list->count++] = block;
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
static enum vkd3d_result vsir_block_list_add(struct vsir_block_list *list, struct vsir_block *block)
|
|
|
|
{
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
for (i = 0; i < list->count; ++i)
|
|
|
|
if (block == list->blocks[i])
|
|
|
|
return VKD3D_FALSE;
|
|
|
|
|
|
|
|
return vsir_block_list_add_checked(list, block);
|
|
|
|
}
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
/* It is guaranteed that the relative order is kept. */
|
|
|
|
static void vsir_block_list_remove_index(struct vsir_block_list *list, size_t idx)
|
|
|
|
{
|
|
|
|
--list->count;
|
|
|
|
memmove(&list->blocks[idx], &list->blocks[idx + 1], (list->count - idx) * sizeof(*list->blocks));
|
|
|
|
}
|
|
|
|
|
2024-01-29 07:41:31 -08:00
|
|
|
struct vsir_block
|
|
|
|
{
|
2024-02-05 02:15:24 -08:00
|
|
|
unsigned int label, order_pos;
|
2024-01-29 07:41:31 -08:00
|
|
|
/* `begin' points to the instruction immediately following the
|
|
|
|
* LABEL that introduces the block. `end' points to the terminator
|
|
|
|
* instruction (either BRANCH or RET). They can coincide, meaning
|
|
|
|
* that the block is empty. */
|
|
|
|
struct vkd3d_shader_instruction *begin, *end;
|
|
|
|
struct vsir_block_list predecessors, successors;
|
2024-01-30 03:51:04 -08:00
|
|
|
uint32_t *dominates;
|
2024-01-29 07:41:31 -08:00
|
|
|
};
|
|
|
|
|
2024-01-30 03:51:04 -08:00
|
|
|
static enum vkd3d_result vsir_block_init(struct vsir_block *block, unsigned int label, size_t block_count)
|
2024-01-29 07:41:31 -08:00
|
|
|
{
|
2024-01-30 03:51:04 -08:00
|
|
|
size_t byte_count;
|
|
|
|
|
|
|
|
if (block_count > SIZE_MAX - (sizeof(*block->dominates) * CHAR_BIT - 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
2024-04-17 04:12:13 -07:00
|
|
|
byte_count = VKD3D_BITMAP_SIZE(block_count) * sizeof(*block->dominates);
|
2024-01-30 03:51:04 -08:00
|
|
|
|
2024-01-29 07:41:31 -08:00
|
|
|
assert(label);
|
|
|
|
memset(block, 0, sizeof(*block));
|
|
|
|
block->label = label;
|
|
|
|
vsir_block_list_init(&block->predecessors);
|
|
|
|
vsir_block_list_init(&block->successors);
|
2024-01-30 03:51:04 -08:00
|
|
|
|
|
|
|
if (!(block->dominates = vkd3d_malloc(byte_count)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
memset(block->dominates, 0xff, byte_count);
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
2024-01-29 07:41:31 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void vsir_block_cleanup(struct vsir_block *block)
|
|
|
|
{
|
2024-01-30 03:51:04 -08:00
|
|
|
if (block->label == 0)
|
|
|
|
return;
|
2024-01-29 07:41:31 -08:00
|
|
|
vsir_block_list_cleanup(&block->predecessors);
|
|
|
|
vsir_block_list_cleanup(&block->successors);
|
2024-01-30 03:51:04 -08:00
|
|
|
vkd3d_free(block->dominates);
|
2024-01-29 07:41:31 -08:00
|
|
|
}
|
|
|
|
|
2024-01-30 06:55:12 -08:00
|
|
|
static int block_compare(const void *ptr1, const void *ptr2)
|
|
|
|
{
|
|
|
|
const struct vsir_block *block1 = *(const struct vsir_block **)ptr1;
|
|
|
|
const struct vsir_block *block2 = *(const struct vsir_block **)ptr2;
|
|
|
|
|
|
|
|
return vkd3d_u32_compare(block1->label, block2->label);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void vsir_block_list_sort(struct vsir_block_list *list)
|
|
|
|
{
|
|
|
|
qsort(list->blocks, list->count, sizeof(*list->blocks), block_compare);
|
|
|
|
}
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
static bool vsir_block_list_search(struct vsir_block_list *list, struct vsir_block *block)
|
|
|
|
{
|
|
|
|
return !!bsearch(&block, list->blocks, list->count, sizeof(*list->blocks), block_compare);
|
|
|
|
}
|
|
|
|
|
2024-02-06 06:08:49 -08:00
|
|
|
struct vsir_cfg_structure_list
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure *structures;
|
|
|
|
size_t count, capacity;
|
|
|
|
unsigned int end;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct vsir_cfg_structure
|
|
|
|
{
|
|
|
|
enum vsir_cfg_structure_type
|
|
|
|
{
|
|
|
|
/* Execute a block of the original VSIR program. */
|
|
|
|
STRUCTURE_TYPE_BLOCK,
|
|
|
|
/* Execute a loop, which is identified by an index. */
|
|
|
|
STRUCTURE_TYPE_LOOP,
|
2024-02-20 14:29:48 -08:00
|
|
|
/* Execute a selection construct. */
|
|
|
|
STRUCTURE_TYPE_SELECTION,
|
2024-02-06 06:08:49 -08:00
|
|
|
/* Execute a `return' or a (possibly) multilevel `break' or
|
|
|
|
* `continue', targeting a loop by its index. If `condition'
|
|
|
|
* is non-NULL, then the jump is conditional (this is
|
|
|
|
* currently not allowed for `return'). */
|
|
|
|
STRUCTURE_TYPE_JUMP,
|
2024-02-06 06:08:49 -08:00
|
|
|
} type;
|
|
|
|
union
|
|
|
|
{
|
|
|
|
struct vsir_block *block;
|
2024-04-10 08:20:05 -07:00
|
|
|
struct vsir_cfg_structure_loop
|
2024-02-06 06:08:49 -08:00
|
|
|
{
|
|
|
|
struct vsir_cfg_structure_list body;
|
|
|
|
unsigned idx;
|
2024-03-18 04:38:51 -07:00
|
|
|
bool needs_trampoline;
|
|
|
|
struct vsir_cfg_structure *outer_loop;
|
2024-02-06 06:08:49 -08:00
|
|
|
} loop;
|
2024-04-10 08:30:16 -07:00
|
|
|
struct vsir_cfg_structure_selection
|
2024-02-20 14:29:48 -08:00
|
|
|
{
|
|
|
|
struct vkd3d_shader_src_param *condition;
|
|
|
|
struct vsir_cfg_structure_list if_body;
|
|
|
|
struct vsir_cfg_structure_list else_body;
|
|
|
|
bool invert_condition;
|
|
|
|
} selection;
|
2024-04-10 08:37:51 -07:00
|
|
|
struct vsir_cfg_structure_jump
|
2024-02-06 06:08:49 -08:00
|
|
|
{
|
|
|
|
enum vsir_cfg_jump_type
|
|
|
|
{
|
|
|
|
/* NONE is available as an intermediate value, but it
|
|
|
|
* is not allowed in valid structured programs. */
|
|
|
|
JUMP_NONE,
|
|
|
|
JUMP_BREAK,
|
|
|
|
JUMP_CONTINUE,
|
|
|
|
JUMP_RET,
|
|
|
|
} type;
|
|
|
|
unsigned int target;
|
|
|
|
struct vkd3d_shader_src_param *condition;
|
|
|
|
bool invert_condition;
|
2024-03-18 04:50:02 -07:00
|
|
|
bool needs_launcher;
|
2024-02-06 06:08:49 -08:00
|
|
|
} jump;
|
2024-02-06 06:08:49 -08:00
|
|
|
} u;
|
|
|
|
};
|
|
|
|
|
|
|
|
static void vsir_cfg_structure_init(struct vsir_cfg_structure *structure, enum vsir_cfg_structure_type type);
|
|
|
|
static void vsir_cfg_structure_cleanup(struct vsir_cfg_structure *structure);
|
|
|
|
|
|
|
|
static void vsir_cfg_structure_list_cleanup(struct vsir_cfg_structure_list *list)
|
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
for (i = 0; i < list->count; ++i)
|
|
|
|
vsir_cfg_structure_cleanup(&list->structures[i]);
|
|
|
|
vkd3d_free(list->structures);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct vsir_cfg_structure *vsir_cfg_structure_list_append(struct vsir_cfg_structure_list *list,
|
|
|
|
enum vsir_cfg_structure_type type)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure *ret;
|
|
|
|
|
|
|
|
if (!vkd3d_array_reserve((void **)&list->structures, &list->capacity, list->count + 1,
|
|
|
|
sizeof(*list->structures)))
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
ret = &list->structures[list->count++];
|
|
|
|
|
|
|
|
vsir_cfg_structure_init(ret, type);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2024-02-20 14:29:48 -08:00
|
|
|
static enum vkd3d_result vsir_cfg_structure_list_append_from_region(struct vsir_cfg_structure_list *list,
|
|
|
|
struct vsir_cfg_structure *begin, size_t size)
|
|
|
|
{
|
|
|
|
if (!vkd3d_array_reserve((void **)&list->structures, &list->capacity, list->count + size,
|
|
|
|
sizeof(*list->structures)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
memcpy(&list->structures[list->count], begin, size * sizeof(*begin));
|
|
|
|
|
|
|
|
list->count += size;
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-02-06 06:08:49 -08:00
|
|
|
static void vsir_cfg_structure_init(struct vsir_cfg_structure *structure, enum vsir_cfg_structure_type type)
|
|
|
|
{
|
|
|
|
memset(structure, 0, sizeof(*structure));
|
|
|
|
structure->type = type;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void vsir_cfg_structure_cleanup(struct vsir_cfg_structure *structure)
|
|
|
|
{
|
2024-02-20 14:29:48 -08:00
|
|
|
switch (structure->type)
|
|
|
|
{
|
|
|
|
case STRUCTURE_TYPE_LOOP:
|
|
|
|
vsir_cfg_structure_list_cleanup(&structure->u.loop.body);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case STRUCTURE_TYPE_SELECTION:
|
|
|
|
vsir_cfg_structure_list_cleanup(&structure->u.selection.if_body);
|
|
|
|
vsir_cfg_structure_list_cleanup(&structure->u.selection.else_body);
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2024-02-06 06:08:49 -08:00
|
|
|
}
|
|
|
|
|
2024-04-10 06:58:34 -07:00
|
|
|
struct vsir_cfg_emit_target
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction *instructions;
|
|
|
|
size_t ins_capacity, ins_count;
|
|
|
|
unsigned int jump_target_temp_idx;
|
|
|
|
unsigned int temp_count;
|
|
|
|
};
|
|
|
|
|
2024-01-29 07:41:31 -08:00
|
|
|
struct vsir_cfg
|
|
|
|
{
|
2024-02-23 02:43:47 -08:00
|
|
|
struct vkd3d_shader_message_context *message_context;
|
2024-01-29 07:41:31 -08:00
|
|
|
struct vsir_program *program;
|
2024-04-12 07:19:49 -07:00
|
|
|
size_t function_begin;
|
|
|
|
size_t function_end;
|
2024-01-29 07:41:31 -08:00
|
|
|
struct vsir_block *blocks;
|
|
|
|
struct vsir_block *entry;
|
|
|
|
size_t block_count;
|
2024-02-21 02:48:53 -08:00
|
|
|
struct vkd3d_string_buffer debug_buffer;
|
2024-01-30 06:55:12 -08:00
|
|
|
|
|
|
|
struct vsir_block_list *loops;
|
|
|
|
size_t loops_count, loops_capacity;
|
2024-01-30 06:55:12 -08:00
|
|
|
size_t *loops_by_header;
|
2024-02-02 14:35:21 -08:00
|
|
|
|
|
|
|
struct vsir_block_list order;
|
2024-02-04 11:21:20 -08:00
|
|
|
struct cfg_loop_interval
|
|
|
|
{
|
|
|
|
/* `begin' is the position of the first block of the loop in
|
|
|
|
* the topological sort; `end' is the position of the first
|
|
|
|
* block after the loop. In other words, `begin' is where a
|
|
|
|
* `continue' instruction would jump and `end' is where a
|
|
|
|
* `break' instruction would jump. */
|
|
|
|
unsigned int begin, end;
|
2024-02-05 02:15:24 -08:00
|
|
|
/* Each loop interval can be natural or synthetic. Natural
|
|
|
|
* intervals are added to represent loops given by CFG back
|
|
|
|
* edges. Synthetic intervals do not correspond to loops in
|
|
|
|
* the input CFG, but are added to leverage their `break'
|
|
|
|
* instruction in order to execute forward edges.
|
|
|
|
*
|
|
|
|
* For a synthetic loop interval it's not really important
|
|
|
|
* which one is the `begin' block, since we don't need to
|
|
|
|
* execute `continue' for them. So we have some leeway for
|
|
|
|
* moving it provided that these conditions are met: 1. the
|
|
|
|
* interval must contain all `break' instructions that target
|
|
|
|
* it, which in practice means that `begin' can be moved
|
|
|
|
* backward and not forward; 2. intervals must remain properly
|
|
|
|
* nested (for each pair of intervals, either one contains the
|
|
|
|
* other or they are disjoint).
|
|
|
|
*
|
|
|
|
* Subject to these conditions, we try to reuse the same loop
|
|
|
|
* as much as possible (if many forward edges target the same
|
|
|
|
* block), but we still try to keep `begin' as forward as
|
|
|
|
* possible, to keep the loop scope as small as possible. */
|
|
|
|
bool synthetic;
|
2024-03-13 05:39:24 -07:00
|
|
|
/* The number of jump instructions (both conditional and
|
|
|
|
* unconditional) that target this loop. */
|
|
|
|
unsigned int target_count;
|
2024-02-04 11:21:20 -08:00
|
|
|
} *loop_intervals;
|
|
|
|
size_t loop_interval_count, loop_interval_capacity;
|
2024-02-06 06:08:49 -08:00
|
|
|
|
|
|
|
struct vsir_cfg_structure_list structured_program;
|
2024-02-20 03:47:19 -08:00
|
|
|
|
2024-04-10 06:58:34 -07:00
|
|
|
struct vsir_cfg_emit_target *target;
|
2024-01-29 07:41:31 -08:00
|
|
|
};
|
|
|
|
|
|
|
|
static void vsir_cfg_cleanup(struct vsir_cfg *cfg)
|
|
|
|
{
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
for (i = 0; i < cfg->block_count; ++i)
|
|
|
|
vsir_block_cleanup(&cfg->blocks[i]);
|
|
|
|
|
2024-01-30 06:55:12 -08:00
|
|
|
for (i = 0; i < cfg->loops_count; ++i)
|
|
|
|
vsir_block_list_cleanup(&cfg->loops[i]);
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
vsir_block_list_cleanup(&cfg->order);
|
|
|
|
|
2024-02-06 06:08:49 -08:00
|
|
|
vsir_cfg_structure_list_cleanup(&cfg->structured_program);
|
|
|
|
|
2024-01-29 07:41:31 -08:00
|
|
|
vkd3d_free(cfg->blocks);
|
2024-01-30 06:55:12 -08:00
|
|
|
vkd3d_free(cfg->loops);
|
2024-01-30 06:55:12 -08:00
|
|
|
vkd3d_free(cfg->loops_by_header);
|
2024-02-04 11:21:20 -08:00
|
|
|
vkd3d_free(cfg->loop_intervals);
|
2024-02-21 02:48:53 -08:00
|
|
|
|
|
|
|
if (TRACE_ON())
|
|
|
|
vkd3d_string_buffer_cleanup(&cfg->debug_buffer);
|
2024-01-29 07:41:31 -08:00
|
|
|
}
|
|
|
|
|
2024-02-04 11:21:20 -08:00
|
|
|
static enum vkd3d_result vsir_cfg_add_loop_interval(struct vsir_cfg *cfg, unsigned int begin,
|
2024-02-05 02:15:24 -08:00
|
|
|
unsigned int end, bool synthetic)
|
2024-02-04 11:21:20 -08:00
|
|
|
{
|
|
|
|
struct cfg_loop_interval *interval;
|
|
|
|
|
|
|
|
if (!vkd3d_array_reserve((void **)&cfg->loop_intervals, &cfg->loop_interval_capacity,
|
|
|
|
cfg->loop_interval_count + 1, sizeof(*cfg->loop_intervals)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
interval = &cfg->loop_intervals[cfg->loop_interval_count++];
|
|
|
|
|
|
|
|
interval->begin = begin;
|
|
|
|
interval->end = end;
|
2024-02-05 02:15:24 -08:00
|
|
|
interval->synthetic = synthetic;
|
2024-03-13 05:39:24 -07:00
|
|
|
interval->target_count = 0;
|
2024-02-04 11:21:20 -08:00
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-02-22 14:50:21 -08:00
|
|
|
static bool vsir_block_dominates(struct vsir_block *b1, struct vsir_block *b2)
|
|
|
|
{
|
|
|
|
return bitmap_is_set(b1->dominates, b2->label - 1);
|
|
|
|
}
|
|
|
|
|
2024-01-29 07:41:31 -08:00
|
|
|
static enum vkd3d_result vsir_cfg_add_edge(struct vsir_cfg *cfg, struct vsir_block *block,
|
|
|
|
struct vkd3d_shader_src_param *successor_param)
|
|
|
|
{
|
|
|
|
unsigned int target = label_from_src_param(successor_param);
|
|
|
|
struct vsir_block *successor = &cfg->blocks[target - 1];
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
|
|
|
|
assert(successor->label != 0);
|
|
|
|
|
|
|
|
if ((ret = vsir_block_list_add(&block->successors, successor)) < 0)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
if ((ret = vsir_block_list_add(&successor->predecessors, block)) < 0)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-01-30 05:59:52 -08:00
|
|
|
static void vsir_cfg_dump_dot(struct vsir_cfg *cfg)
|
|
|
|
{
|
|
|
|
size_t i, j;
|
|
|
|
|
|
|
|
TRACE("digraph cfg {\n");
|
|
|
|
|
|
|
|
for (i = 0; i < cfg->block_count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_block *block = &cfg->blocks[i];
|
|
|
|
const char *shape;
|
|
|
|
|
|
|
|
if (block->label == 0)
|
|
|
|
continue;
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (block->end->opcode)
|
2024-01-30 05:59:52 -08:00
|
|
|
{
|
|
|
|
case VKD3DSIH_RET:
|
|
|
|
shape = "trapezium";
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_BRANCH:
|
|
|
|
shape = vsir_register_is_label(&block->end->src[0].reg) ? "ellipse" : "box";
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
vkd3d_unreachable();
|
|
|
|
}
|
|
|
|
|
|
|
|
TRACE(" n%u [label=\"%u\", shape=\"%s\"];\n", block->label, block->label, shape);
|
|
|
|
|
|
|
|
for (j = 0; j < block->successors.count; ++j)
|
|
|
|
TRACE(" n%u -> n%u;\n", block->label, block->successors.blocks[j]->label);
|
|
|
|
}
|
|
|
|
|
|
|
|
TRACE("}\n");
|
|
|
|
}
|
|
|
|
|
2024-02-06 06:08:49 -08:00
|
|
|
static void vsir_cfg_structure_list_dump(struct vsir_cfg *cfg, struct vsir_cfg_structure_list *list);
|
|
|
|
|
|
|
|
static void vsir_cfg_structure_dump(struct vsir_cfg *cfg, struct vsir_cfg_structure *structure)
|
|
|
|
{
|
|
|
|
switch (structure->type)
|
|
|
|
{
|
|
|
|
case STRUCTURE_TYPE_BLOCK:
|
|
|
|
TRACE("%sblock %u\n", cfg->debug_buffer.buffer, structure->u.block->label);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case STRUCTURE_TYPE_LOOP:
|
|
|
|
TRACE("%s%u : loop {\n", cfg->debug_buffer.buffer, structure->u.loop.idx);
|
|
|
|
|
|
|
|
vsir_cfg_structure_list_dump(cfg, &structure->u.loop.body);
|
|
|
|
|
2024-03-18 04:38:51 -07:00
|
|
|
TRACE("%s} # %u%s\n", cfg->debug_buffer.buffer, structure->u.loop.idx,
|
|
|
|
structure->u.loop.needs_trampoline ? ", tramp" : "");
|
2024-02-06 06:08:49 -08:00
|
|
|
break;
|
|
|
|
|
2024-02-20 14:29:48 -08:00
|
|
|
case STRUCTURE_TYPE_SELECTION:
|
|
|
|
TRACE("%sif {\n", cfg->debug_buffer.buffer);
|
|
|
|
|
|
|
|
vsir_cfg_structure_list_dump(cfg, &structure->u.selection.if_body);
|
|
|
|
|
|
|
|
if (structure->u.selection.else_body.count == 0)
|
|
|
|
{
|
|
|
|
TRACE("%s}\n", cfg->debug_buffer.buffer);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
TRACE("%s} else {\n", cfg->debug_buffer.buffer);
|
|
|
|
|
|
|
|
vsir_cfg_structure_list_dump(cfg, &structure->u.selection.else_body);
|
|
|
|
|
|
|
|
TRACE("%s}\n", cfg->debug_buffer.buffer);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
2024-02-06 06:08:49 -08:00
|
|
|
case STRUCTURE_TYPE_JUMP:
|
|
|
|
{
|
|
|
|
const char *type_str;
|
|
|
|
|
|
|
|
switch (structure->u.jump.type)
|
|
|
|
{
|
|
|
|
case JUMP_RET:
|
|
|
|
TRACE("%sret\n", cfg->debug_buffer.buffer);
|
|
|
|
return;
|
|
|
|
|
|
|
|
case JUMP_BREAK:
|
|
|
|
type_str = "break";
|
|
|
|
break;
|
|
|
|
|
|
|
|
case JUMP_CONTINUE:
|
|
|
|
type_str = "continue";
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
vkd3d_unreachable();
|
|
|
|
}
|
|
|
|
|
2024-03-18 04:50:02 -07:00
|
|
|
TRACE("%s%s%s %u%s\n", cfg->debug_buffer.buffer, type_str,
|
|
|
|
structure->u.jump.condition ? "c" : "", structure->u.jump.target,
|
|
|
|
structure->u.jump.needs_launcher ? " # launch" : "");
|
2024-02-06 06:08:49 -08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
default:
|
|
|
|
vkd3d_unreachable();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void vsir_cfg_structure_list_dump(struct vsir_cfg *cfg, struct vsir_cfg_structure_list *list)
|
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
vkd3d_string_buffer_printf(&cfg->debug_buffer, " ");
|
|
|
|
|
|
|
|
for (i = 0; i < list->count; ++i)
|
|
|
|
vsir_cfg_structure_dump(cfg, &list->structures[i]);
|
|
|
|
|
|
|
|
vkd3d_string_buffer_truncate(&cfg->debug_buffer, cfg->debug_buffer.content_size - 2);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void vsir_cfg_dump_structured_program(struct vsir_cfg *cfg)
|
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
for (i = 0; i < cfg->structured_program.count; ++i)
|
|
|
|
vsir_cfg_structure_dump(cfg, &cfg->structured_program.structures[i]);
|
|
|
|
}
|
|
|
|
|
2024-02-23 02:43:47 -08:00
|
|
|
static enum vkd3d_result vsir_cfg_init(struct vsir_cfg *cfg, struct vsir_program *program,
|
2024-04-11 07:33:19 -07:00
|
|
|
struct vkd3d_shader_message_context *message_context, struct vsir_cfg_emit_target *target,
|
|
|
|
size_t *pos)
|
2024-01-29 07:41:31 -08:00
|
|
|
{
|
|
|
|
struct vsir_block *current_block = NULL;
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
memset(cfg, 0, sizeof(*cfg));
|
2024-02-23 02:43:47 -08:00
|
|
|
cfg->message_context = message_context;
|
2024-01-29 07:41:31 -08:00
|
|
|
cfg->program = program;
|
|
|
|
cfg->block_count = program->block_count;
|
2024-04-10 06:58:34 -07:00
|
|
|
cfg->target = target;
|
2024-04-12 07:19:49 -07:00
|
|
|
cfg->function_begin = *pos;
|
2024-01-29 07:41:31 -08:00
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
vsir_block_list_init(&cfg->order);
|
|
|
|
|
2024-01-29 07:41:31 -08:00
|
|
|
if (!(cfg->blocks = vkd3d_calloc(cfg->block_count, sizeof(*cfg->blocks))))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
2024-02-21 02:48:53 -08:00
|
|
|
if (TRACE_ON())
|
|
|
|
vkd3d_string_buffer_init(&cfg->debug_buffer);
|
|
|
|
|
2024-04-11 07:33:19 -07:00
|
|
|
for (i = *pos; i < program->instructions.count; ++i)
|
2024-01-29 07:41:31 -08:00
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction *instruction = &program->instructions.elements[i];
|
2024-04-11 07:33:19 -07:00
|
|
|
bool finish = false;
|
2024-01-29 07:41:31 -08:00
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (instruction->opcode)
|
2024-01-29 07:41:31 -08:00
|
|
|
{
|
|
|
|
case VKD3DSIH_PHI:
|
|
|
|
case VKD3DSIH_SWITCH_MONOLITHIC:
|
|
|
|
vkd3d_unreachable();
|
|
|
|
|
|
|
|
case VKD3DSIH_LABEL:
|
|
|
|
{
|
|
|
|
unsigned int label = label_from_src_param(&instruction->src[0]);
|
|
|
|
|
|
|
|
assert(!current_block);
|
|
|
|
assert(label > 0);
|
|
|
|
assert(label <= cfg->block_count);
|
|
|
|
current_block = &cfg->blocks[label - 1];
|
2024-01-30 03:51:04 -08:00
|
|
|
assert(current_block->label == 0);
|
|
|
|
if ((ret = vsir_block_init(current_block, label, program->block_count)) < 0)
|
|
|
|
goto fail;
|
2024-01-29 07:41:31 -08:00
|
|
|
current_block->begin = &program->instructions.elements[i + 1];
|
|
|
|
if (!cfg->entry)
|
|
|
|
cfg->entry = current_block;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case VKD3DSIH_BRANCH:
|
|
|
|
case VKD3DSIH_RET:
|
|
|
|
assert(current_block);
|
|
|
|
current_block->end = instruction;
|
|
|
|
current_block = NULL;
|
|
|
|
break;
|
|
|
|
|
2024-04-11 07:33:19 -07:00
|
|
|
case VKD3DSIH_HS_CONTROL_POINT_PHASE:
|
|
|
|
case VKD3DSIH_HS_FORK_PHASE:
|
|
|
|
case VKD3DSIH_HS_JOIN_PHASE:
|
|
|
|
assert(!current_block);
|
|
|
|
finish = true;
|
|
|
|
break;
|
|
|
|
|
2024-01-29 07:41:31 -08:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2024-04-11 07:33:19 -07:00
|
|
|
|
|
|
|
if (finish)
|
|
|
|
break;
|
2024-01-29 07:41:31 -08:00
|
|
|
}
|
|
|
|
|
2024-04-11 07:33:19 -07:00
|
|
|
*pos = i;
|
2024-04-12 07:19:49 -07:00
|
|
|
cfg->function_end = *pos;
|
2024-04-11 07:33:19 -07:00
|
|
|
|
2024-01-29 07:41:31 -08:00
|
|
|
for (i = 0; i < cfg->block_count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_block *block = &cfg->blocks[i];
|
|
|
|
|
|
|
|
if (block->label == 0)
|
|
|
|
continue;
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (block->end->opcode)
|
2024-01-29 07:41:31 -08:00
|
|
|
{
|
|
|
|
case VKD3DSIH_RET:
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_BRANCH:
|
|
|
|
if (vsir_register_is_label(&block->end->src[0].reg))
|
|
|
|
{
|
|
|
|
if ((ret = vsir_cfg_add_edge(cfg, block, &block->end->src[0])) < 0)
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
if ((ret = vsir_cfg_add_edge(cfg, block, &block->end->src[1])) < 0)
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
if ((ret = vsir_cfg_add_edge(cfg, block, &block->end->src[2])) < 0)
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
vkd3d_unreachable();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-01-30 05:59:52 -08:00
|
|
|
if (TRACE_ON())
|
|
|
|
vsir_cfg_dump_dot(cfg);
|
|
|
|
|
2024-01-29 07:41:31 -08:00
|
|
|
return VKD3D_OK;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
vsir_cfg_cleanup(cfg);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2024-01-30 03:51:04 -08:00
|
|
|
/* Block A dominates block B if every path from the entry point to B
|
|
|
|
* must pass through A. Naively compute the set of blocks that are
|
|
|
|
* dominated by `reference' by running a graph visit starting from the
|
|
|
|
* entry point (which must be the initial value of `current') and
|
|
|
|
* avoiding `reference'. Running this for all the blocks takes
|
|
|
|
* quadratic time: if in the future something better is sought after,
|
|
|
|
* the standard tool seems to be the Lengauer-Tarjan algorithm. */
|
|
|
|
static void vsir_cfg_compute_dominators_recurse(struct vsir_block *current, struct vsir_block *reference)
|
|
|
|
{
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
assert(current->label != 0);
|
|
|
|
|
|
|
|
if (current == reference)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (!bitmap_is_set(reference->dominates, current->label - 1))
|
|
|
|
return;
|
|
|
|
|
|
|
|
bitmap_clear(reference->dominates, current->label - 1);
|
|
|
|
|
|
|
|
for (i = 0; i < current->successors.count; ++i)
|
|
|
|
vsir_cfg_compute_dominators_recurse(current->successors.blocks[i], reference);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void vsir_cfg_compute_dominators(struct vsir_cfg *cfg)
|
|
|
|
{
|
2024-01-30 06:07:50 -08:00
|
|
|
size_t i, j;
|
|
|
|
|
2024-01-30 03:51:04 -08:00
|
|
|
for (i = 0; i < cfg->block_count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_block *block = &cfg->blocks[i];
|
|
|
|
|
|
|
|
if (block->label == 0)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
vsir_cfg_compute_dominators_recurse(cfg->entry, block);
|
2024-01-30 06:07:50 -08:00
|
|
|
|
|
|
|
if (TRACE_ON())
|
|
|
|
{
|
2024-02-21 02:48:53 -08:00
|
|
|
vkd3d_string_buffer_printf(&cfg->debug_buffer, "Block %u dominates:", block->label);
|
2024-01-30 06:07:50 -08:00
|
|
|
for (j = 0; j < cfg->block_count; j++)
|
|
|
|
{
|
|
|
|
struct vsir_block *block2 = &cfg->blocks[j];
|
|
|
|
|
|
|
|
if (block2->label == 0)
|
|
|
|
continue;
|
|
|
|
|
2024-02-22 14:50:21 -08:00
|
|
|
if (vsir_block_dominates(block, block2))
|
2024-02-21 02:48:53 -08:00
|
|
|
vkd3d_string_buffer_printf(&cfg->debug_buffer, " %u", block2->label);
|
2024-01-30 06:07:50 -08:00
|
|
|
}
|
2024-02-21 02:48:53 -08:00
|
|
|
TRACE("%s\n", cfg->debug_buffer.buffer);
|
|
|
|
vkd3d_string_buffer_clear(&cfg->debug_buffer);
|
2024-01-30 06:07:50 -08:00
|
|
|
}
|
2024-01-30 03:51:04 -08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-01-30 06:55:12 -08:00
|
|
|
/* A back edge is an edge X -> Y for which block Y dominates block
|
|
|
|
* X. All the other edges are forward edges, and it is required that
|
|
|
|
* the input CFG is reducible, i.e., it is acyclic once you strip away
|
|
|
|
* the back edges.
|
|
|
|
*
|
|
|
|
* Each back edge X -> Y defines a loop: block X is the header block,
|
|
|
|
* block Y is the back edge block, and the loop consists of all the
|
|
|
|
* blocks which are dominated by the header block and have a path to
|
|
|
|
* the back edge block that doesn't pass through the header block
|
|
|
|
* (including the header block itself). It can be proved that all the
|
|
|
|
* blocks in such a path (connecting a loop block to the back edge
|
|
|
|
* block without passing through the header block) belong to the same
|
|
|
|
* loop.
|
|
|
|
*
|
2024-02-04 11:21:20 -08:00
|
|
|
* If the input CFG is reducible its loops are properly nested (i.e.,
|
|
|
|
* each two loops are either disjoint or one is contained in the
|
|
|
|
* other), provided that each block has at most one incoming back
|
|
|
|
* edge. If this condition does not hold, a synthetic block can be
|
|
|
|
* introduced as the only back edge block for the given header block,
|
|
|
|
* with all the previous back edge now being forward edges to the
|
|
|
|
* synthetic block. This is not currently implemented (but it is
|
|
|
|
* rarely found in practice anyway). */
|
2024-01-30 06:55:12 -08:00
|
|
|
static enum vkd3d_result vsir_cfg_scan_loop(struct vsir_block_list *loop, struct vsir_block *block,
|
|
|
|
struct vsir_block *header)
|
|
|
|
{
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
if ((ret = vsir_block_list_add(loop, block)) < 0)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
if (ret == VKD3D_FALSE || block == header)
|
|
|
|
return VKD3D_OK;
|
|
|
|
|
|
|
|
for (i = 0; i < block->predecessors.count; ++i)
|
|
|
|
{
|
|
|
|
if ((ret = vsir_cfg_scan_loop(loop, block->predecessors.blocks[i], header)) < 0)
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
static enum vkd3d_result vsir_cfg_compute_loops(struct vsir_cfg *cfg)
|
|
|
|
{
|
2024-01-30 06:55:12 -08:00
|
|
|
size_t i, j, k;
|
2024-01-30 06:55:12 -08:00
|
|
|
|
2024-01-30 06:55:12 -08:00
|
|
|
if (!(cfg->loops_by_header = vkd3d_calloc(cfg->block_count, sizeof(*cfg->loops_by_header))))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
memset(cfg->loops_by_header, 0xff, cfg->block_count * sizeof(*cfg->loops_by_header));
|
|
|
|
|
2024-01-30 06:55:12 -08:00
|
|
|
for (i = 0; i < cfg->block_count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_block *block = &cfg->blocks[i];
|
|
|
|
|
|
|
|
if (block->label == 0)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
for (j = 0; j < block->successors.count; ++j)
|
|
|
|
{
|
|
|
|
struct vsir_block *header = block->successors.blocks[j];
|
|
|
|
struct vsir_block_list *loop;
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
|
|
|
|
/* Is this a back edge? */
|
|
|
|
if (!vsir_block_dominates(header, block))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (!vkd3d_array_reserve((void **)&cfg->loops, &cfg->loops_capacity, cfg->loops_count + 1, sizeof(*cfg->loops)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
2024-01-30 06:55:12 -08:00
|
|
|
loop = &cfg->loops[cfg->loops_count];
|
2024-01-30 06:55:12 -08:00
|
|
|
vsir_block_list_init(loop);
|
|
|
|
|
|
|
|
if ((ret = vsir_cfg_scan_loop(loop, block, header)) < 0)
|
|
|
|
return ret;
|
2024-01-30 06:55:12 -08:00
|
|
|
|
2024-01-30 06:55:12 -08:00
|
|
|
vsir_block_list_sort(loop);
|
|
|
|
|
2024-01-30 06:55:12 -08:00
|
|
|
if (TRACE_ON())
|
|
|
|
{
|
|
|
|
vkd3d_string_buffer_printf(&cfg->debug_buffer, "Back edge %u -> %u with loop:", block->label, header->label);
|
|
|
|
|
|
|
|
for (k = 0; k < loop->count; ++k)
|
|
|
|
vkd3d_string_buffer_printf(&cfg->debug_buffer, " %u", loop->blocks[k]->label);
|
|
|
|
|
|
|
|
TRACE("%s\n", cfg->debug_buffer.buffer);
|
|
|
|
vkd3d_string_buffer_clear(&cfg->debug_buffer);
|
|
|
|
}
|
|
|
|
|
2024-01-30 06:55:12 -08:00
|
|
|
if (cfg->loops_by_header[header->label - 1] != SIZE_MAX)
|
|
|
|
{
|
|
|
|
FIXME("Block %u is header to more than one loop, this is not implemented.\n", header->label);
|
|
|
|
vkd3d_shader_error(cfg->message_context, &header->begin->location, VKD3D_SHADER_ERROR_VSIR_NOT_IMPLEMENTED,
|
|
|
|
"Block %u is header to more than one loop, this is not implemented.", header->label);
|
|
|
|
return VKD3D_ERROR_NOT_IMPLEMENTED;
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg->loops_by_header[header->label - 1] = cfg->loops_count;
|
|
|
|
|
|
|
|
++cfg->loops_count;
|
2024-01-30 06:55:12 -08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
struct vsir_cfg_node_sorter
|
|
|
|
{
|
|
|
|
struct vsir_cfg *cfg;
|
2024-02-02 14:35:21 -08:00
|
|
|
struct vsir_cfg_node_sorter_stack_item
|
|
|
|
{
|
|
|
|
struct vsir_block_list *loop;
|
|
|
|
unsigned int seen_count;
|
2024-02-04 11:21:20 -08:00
|
|
|
unsigned int begin;
|
2024-02-02 14:35:21 -08:00
|
|
|
} *stack;
|
|
|
|
size_t stack_count, stack_capacity;
|
2024-02-02 14:35:21 -08:00
|
|
|
struct vsir_block_list available_blocks;
|
|
|
|
};
|
|
|
|
|
|
|
|
/* Topologically sort the blocks according to the forward edges. By
|
|
|
|
* definition if the input CFG is reducible then its forward edges
|
|
|
|
* form a DAG, so a topological sorting exists. In order to compute it
|
|
|
|
* we keep an array with the incoming degree for each block and an
|
|
|
|
* available list of all the blocks whose incoming degree has reached
|
|
|
|
* zero. At each step we pick a block from the available list and
|
|
|
|
* strip it away from the graph, updating the incoming degrees and
|
2024-02-02 14:35:21 -08:00
|
|
|
* available list.
|
|
|
|
*
|
|
|
|
* In principle at each step we can pick whatever node we want from
|
|
|
|
* the available list, and will get a topological sort
|
|
|
|
* anyway. However, we use these two criteria to give to the computed
|
|
|
|
* order additional properties:
|
|
|
|
*
|
|
|
|
* 1. we keep track of which loops we're into, and pick blocks
|
|
|
|
* belonging to the current innermost loop, so that loops are kept
|
|
|
|
* contiguous in the order; this can always be done when the input
|
|
|
|
* CFG is reducible;
|
|
|
|
*
|
|
|
|
* 2. subject to the requirement above, we always pick the most
|
|
|
|
* recently added block to the available list, because this tends
|
|
|
|
* to keep related blocks and require fewer control flow
|
|
|
|
* primitives.
|
|
|
|
*/
|
2024-02-02 14:35:21 -08:00
|
|
|
static enum vkd3d_result vsir_cfg_sort_nodes(struct vsir_cfg *cfg)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_node_sorter sorter = { .cfg = cfg };
|
|
|
|
unsigned int *in_degrees = NULL;
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
if (!(in_degrees = vkd3d_calloc(cfg->block_count, sizeof(*in_degrees))))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
for (i = 0; i < cfg->block_count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_block *block = &cfg->blocks[i];
|
|
|
|
|
|
|
|
if (block->label == 0)
|
|
|
|
{
|
|
|
|
in_degrees[i] = UINT_MAX;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
in_degrees[i] = block->predecessors.count;
|
|
|
|
|
|
|
|
/* Do not count back edges. */
|
|
|
|
if (cfg->loops_by_header[i] != SIZE_MAX)
|
|
|
|
{
|
|
|
|
assert(in_degrees[i] > 0);
|
|
|
|
in_degrees[i] -= 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (in_degrees[i] == 0 && block != cfg->entry)
|
|
|
|
{
|
|
|
|
WARN("Unexpected entry point %u.\n", block->label);
|
|
|
|
vkd3d_shader_error(cfg->message_context, &block->begin->location, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW,
|
|
|
|
"Block %u is unreachable from the entry point.", block->label);
|
|
|
|
ret = VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (in_degrees[cfg->entry->label - 1] != 0)
|
|
|
|
{
|
|
|
|
WARN("Entry point has %u incoming forward edges.\n", in_degrees[cfg->entry->label - 1]);
|
|
|
|
vkd3d_shader_error(cfg->message_context, &cfg->entry->begin->location, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW,
|
|
|
|
"The entry point block has %u incoming forward edges.", in_degrees[cfg->entry->label - 1]);
|
|
|
|
ret = VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
|
|
|
vsir_block_list_init(&sorter.available_blocks);
|
|
|
|
|
2024-03-20 09:13:45 -07:00
|
|
|
if ((ret = vsir_block_list_add_checked(&sorter.available_blocks, cfg->entry)) < 0)
|
2024-02-02 14:35:21 -08:00
|
|
|
goto fail;
|
|
|
|
|
|
|
|
while (sorter.available_blocks.count != 0)
|
|
|
|
{
|
2024-02-02 14:35:21 -08:00
|
|
|
struct vsir_cfg_node_sorter_stack_item *inner_stack_item = NULL;
|
|
|
|
struct vsir_block *block;
|
|
|
|
size_t new_seen_count;
|
|
|
|
|
|
|
|
if (sorter.stack_count != 0)
|
|
|
|
inner_stack_item = &sorter.stack[sorter.stack_count - 1];
|
|
|
|
|
|
|
|
for (i = sorter.available_blocks.count - 1; ; --i)
|
|
|
|
{
|
|
|
|
if (i == SIZE_MAX)
|
|
|
|
{
|
|
|
|
ERR("Couldn't find any viable next block, is the input CFG reducible?\n");
|
|
|
|
ret = VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
|
|
|
block = sorter.available_blocks.blocks[i];
|
|
|
|
|
|
|
|
if (!inner_stack_item || vsir_block_list_search(inner_stack_item->loop, block))
|
|
|
|
break;
|
|
|
|
}
|
2024-02-02 14:35:21 -08:00
|
|
|
|
2024-03-20 08:58:25 -07:00
|
|
|
/* If the node is a loop header, open the loop. */
|
|
|
|
if (sorter.cfg->loops_by_header[block->label - 1] != SIZE_MAX)
|
|
|
|
{
|
|
|
|
struct vsir_block_list *loop = &sorter.cfg->loops[sorter.cfg->loops_by_header[block->label - 1]];
|
|
|
|
|
|
|
|
if (loop)
|
|
|
|
{
|
|
|
|
if (!vkd3d_array_reserve((void **)&sorter.stack, &sorter.stack_capacity,
|
|
|
|
sorter.stack_count + 1, sizeof(*sorter.stack)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
inner_stack_item = &sorter.stack[sorter.stack_count++];
|
|
|
|
inner_stack_item->loop = loop;
|
|
|
|
inner_stack_item->seen_count = 0;
|
|
|
|
inner_stack_item->begin = sorter.cfg->order.count;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
vsir_block_list_remove_index(&sorter.available_blocks, i);
|
2024-02-05 02:15:24 -08:00
|
|
|
block->order_pos = cfg->order.count;
|
2024-02-02 14:35:21 -08:00
|
|
|
if ((ret = vsir_block_list_add_checked(&cfg->order, block)) < 0)
|
|
|
|
goto fail;
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
/* Close loops: since each loop is a strict subset of any
|
|
|
|
* outer loop, we just need to track how many blocks we've
|
|
|
|
* seen; when I close a loop I mark the same number of seen
|
|
|
|
* blocks for the next outer loop. */
|
|
|
|
new_seen_count = 1;
|
|
|
|
while (sorter.stack_count != 0)
|
|
|
|
{
|
|
|
|
inner_stack_item = &sorter.stack[sorter.stack_count - 1];
|
|
|
|
|
|
|
|
inner_stack_item->seen_count += new_seen_count;
|
|
|
|
|
|
|
|
assert(inner_stack_item->seen_count <= inner_stack_item->loop->count);
|
|
|
|
if (inner_stack_item->seen_count != inner_stack_item->loop->count)
|
|
|
|
break;
|
|
|
|
|
2024-02-04 11:21:20 -08:00
|
|
|
if ((ret = vsir_cfg_add_loop_interval(cfg, inner_stack_item->begin,
|
2024-02-05 02:15:24 -08:00
|
|
|
cfg->order.count, false)) < 0)
|
2024-02-04 11:21:20 -08:00
|
|
|
goto fail;
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
new_seen_count = inner_stack_item->loop->count;
|
|
|
|
--sorter.stack_count;
|
|
|
|
}
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
/* Remove (forward) edges and make new nodes available. */
|
|
|
|
for (i = 0; i < block->successors.count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_block *successor = block->successors.blocks[i];
|
|
|
|
|
|
|
|
if (vsir_block_dominates(successor, block))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
assert(in_degrees[successor->label - 1] > 0);
|
|
|
|
--in_degrees[successor->label - 1];
|
|
|
|
|
|
|
|
if (in_degrees[successor->label - 1] == 0)
|
|
|
|
{
|
2024-03-20 09:13:45 -07:00
|
|
|
if ((ret = vsir_block_list_add_checked(&sorter.available_blocks, successor)) < 0)
|
2024-02-02 14:35:21 -08:00
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (cfg->order.count != cfg->block_count)
|
|
|
|
{
|
|
|
|
/* There is a cycle of forward edges. */
|
|
|
|
WARN("The control flow graph is not reducible.\n");
|
|
|
|
vkd3d_shader_error(cfg->message_context, &cfg->entry->begin->location, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW,
|
|
|
|
"The control flow graph is not reducible.");
|
|
|
|
ret = VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
assert(sorter.stack_count == 0);
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
vkd3d_free(in_degrees);
|
2024-02-02 14:35:21 -08:00
|
|
|
vkd3d_free(sorter.stack);
|
2024-02-02 14:35:21 -08:00
|
|
|
vsir_block_list_cleanup(&sorter.available_blocks);
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
if (TRACE_ON())
|
|
|
|
{
|
|
|
|
vkd3d_string_buffer_printf(&cfg->debug_buffer, "Block order:");
|
|
|
|
|
|
|
|
for (i = 0; i < cfg->order.count; ++i)
|
|
|
|
vkd3d_string_buffer_printf(&cfg->debug_buffer, " %u", cfg->order.blocks[i]->label);
|
|
|
|
|
|
|
|
TRACE("%s\n", cfg->debug_buffer.buffer);
|
|
|
|
vkd3d_string_buffer_clear(&cfg->debug_buffer);
|
|
|
|
}
|
|
|
|
|
2024-02-02 14:35:21 -08:00
|
|
|
return VKD3D_OK;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
vkd3d_free(in_degrees);
|
2024-02-02 14:35:21 -08:00
|
|
|
vkd3d_free(sorter.stack);
|
2024-02-02 14:35:21 -08:00
|
|
|
vsir_block_list_cleanup(&sorter.available_blocks);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2024-02-06 05:02:59 -08:00
|
|
|
/* Sort loop intervals first by ascending begin time and then by
|
|
|
|
* descending end time, so that inner intervals appear after outer
|
|
|
|
* ones and disjoint intervals appear in their proper order. */
|
|
|
|
static int compare_loop_intervals(const void *ptr1, const void *ptr2)
|
|
|
|
{
|
|
|
|
const struct cfg_loop_interval *interval1 = ptr1;
|
|
|
|
const struct cfg_loop_interval *interval2 = ptr2;
|
|
|
|
|
|
|
|
if (interval1->begin != interval2->begin)
|
|
|
|
return vkd3d_u32_compare(interval1->begin, interval2->begin);
|
|
|
|
|
|
|
|
return -vkd3d_u32_compare(interval1->end, interval2->end);
|
|
|
|
}
|
|
|
|
|
2024-02-05 02:15:24 -08:00
|
|
|
static enum vkd3d_result vsir_cfg_generate_synthetic_loop_intervals(struct vsir_cfg *cfg)
|
|
|
|
{
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
size_t i, j, k;
|
|
|
|
|
|
|
|
for (i = 0; i < cfg->block_count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_block *block = &cfg->blocks[i];
|
|
|
|
|
|
|
|
if (block->label == 0)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
for (j = 0; j < block->successors.count; ++j)
|
|
|
|
{
|
|
|
|
struct vsir_block *successor = block->successors.blocks[j];
|
|
|
|
struct cfg_loop_interval *extend = NULL;
|
|
|
|
unsigned int begin;
|
|
|
|
enum
|
|
|
|
{
|
|
|
|
ACTION_DO_NOTHING,
|
|
|
|
ACTION_CREATE_NEW,
|
|
|
|
ACTION_EXTEND,
|
|
|
|
} action = ACTION_CREATE_NEW;
|
|
|
|
|
|
|
|
/* We've already contructed loop intervals for the back
|
|
|
|
* edges, there's nothing more to do. */
|
|
|
|
if (vsir_block_dominates(successor, block))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
assert(block->order_pos < successor->order_pos);
|
|
|
|
|
|
|
|
/* Jumping from a block to the following one is always
|
|
|
|
* possible, so nothing to do. */
|
|
|
|
if (block->order_pos + 1 == successor->order_pos)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
/* Let's look for a loop interval that already breaks at
|
|
|
|
* `successor' and either contains or can be extended to
|
|
|
|
* contain `block'. */
|
|
|
|
for (k = 0; k < cfg->loop_interval_count; ++k)
|
|
|
|
{
|
|
|
|
struct cfg_loop_interval *interval = &cfg->loop_intervals[k];
|
|
|
|
|
|
|
|
if (interval->end != successor->order_pos)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (interval->begin <= block->order_pos)
|
|
|
|
{
|
|
|
|
action = ACTION_DO_NOTHING;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (interval->synthetic)
|
|
|
|
{
|
|
|
|
action = ACTION_EXTEND;
|
|
|
|
extend = interval;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (action == ACTION_DO_NOTHING)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
/* Ok, we have to decide where the new or replacing
|
|
|
|
* interval has to begin. These are the rules: 1. it must
|
|
|
|
* begin before `block'; 2. intervals must be properly
|
|
|
|
* nested; 3. the new interval should begin as late as
|
|
|
|
* possible, to limit control flow depth and extension. */
|
|
|
|
begin = block->order_pos;
|
|
|
|
|
|
|
|
/* Our candidate interval is always [begin,
|
|
|
|
* successor->order_pos), and we move `begin' backward
|
|
|
|
* until the candidate interval contains all the intervals
|
|
|
|
* whose endpoint lies in the candidate interval
|
|
|
|
* itself. */
|
|
|
|
for (k = 0; k < cfg->loop_interval_count; ++k)
|
|
|
|
{
|
|
|
|
struct cfg_loop_interval *interval = &cfg->loop_intervals[k];
|
|
|
|
|
|
|
|
if (begin < interval->end && interval->end < successor->order_pos)
|
|
|
|
begin = min(begin, interval->begin);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* New we have to care about the intervals whose begin
|
|
|
|
* point lies in the candidate interval. We cannot move
|
|
|
|
* the candidate interval endpoint, because it is
|
|
|
|
* important that the loop break target matches
|
|
|
|
* `successor'. So we have to move that interval's begin
|
|
|
|
* point to the begin point of the candidate interval,
|
|
|
|
* i.e. `begin'. But what if the interval we should extend
|
|
|
|
* backward is not synthetic? This cannot happen,
|
|
|
|
* fortunately, because it would mean that there is a jump
|
|
|
|
* entering a loop via a block which is not the loop
|
|
|
|
* header, so the CFG would not be reducible. */
|
|
|
|
for (k = 0; k < cfg->loop_interval_count; ++k)
|
|
|
|
{
|
|
|
|
struct cfg_loop_interval *interval = &cfg->loop_intervals[k];
|
|
|
|
|
|
|
|
if (interval->begin < successor->order_pos && successor->order_pos < interval->end)
|
|
|
|
{
|
|
|
|
if (interval->synthetic)
|
|
|
|
interval->begin = min(begin, interval->begin);
|
|
|
|
assert(begin >= interval->begin);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (action == ACTION_EXTEND)
|
|
|
|
extend->begin = begin;
|
|
|
|
else if ((ret = vsir_cfg_add_loop_interval(cfg, begin, successor->order_pos, true)) < 0)
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-02-06 05:02:59 -08:00
|
|
|
qsort(cfg->loop_intervals, cfg->loop_interval_count, sizeof(*cfg->loop_intervals), compare_loop_intervals);
|
|
|
|
|
2024-02-05 02:15:24 -08:00
|
|
|
if (TRACE_ON())
|
|
|
|
for (i = 0; i < cfg->loop_interval_count; ++i)
|
|
|
|
TRACE("%s loop interval %u - %u\n", cfg->loop_intervals[i].synthetic ? "Synthetic" : "Natural",
|
|
|
|
cfg->loop_intervals[i].begin, cfg->loop_intervals[i].end);
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-02-06 06:08:49 -08:00
|
|
|
struct vsir_cfg_edge_action
|
|
|
|
{
|
|
|
|
enum vsir_cfg_jump_type jump_type;
|
|
|
|
unsigned int target;
|
|
|
|
struct vsir_block *successor;
|
|
|
|
};
|
|
|
|
|
|
|
|
static void vsir_cfg_compute_edge_action(struct vsir_cfg *cfg, struct vsir_block *block,
|
|
|
|
struct vsir_block *successor, struct vsir_cfg_edge_action *action)
|
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
action->target = UINT_MAX;
|
|
|
|
action->successor = successor;
|
|
|
|
|
|
|
|
if (successor->order_pos <= block->order_pos)
|
|
|
|
{
|
|
|
|
/* The successor is before the current block, so we have to
|
|
|
|
* use `continue'. The target loop is the innermost that
|
|
|
|
* contains the current block and has the successor as
|
|
|
|
* `continue' target. */
|
|
|
|
for (i = 0; i < cfg->loop_interval_count; ++i)
|
|
|
|
{
|
|
|
|
struct cfg_loop_interval *interval = &cfg->loop_intervals[i];
|
|
|
|
|
|
|
|
if (interval->begin == successor->order_pos && block->order_pos < interval->end)
|
|
|
|
action->target = i;
|
|
|
|
|
|
|
|
if (interval->begin > successor->order_pos)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(action->target != UINT_MAX);
|
|
|
|
action->jump_type = JUMP_CONTINUE;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
/* The successor is after the current block, so we have to use
|
|
|
|
* `break', or possibly just jump to the following block. The
|
|
|
|
* target loop is the outermost that contains the current
|
|
|
|
* block and has the successor as `break' target. */
|
|
|
|
for (i = 0; i < cfg->loop_interval_count; ++i)
|
|
|
|
{
|
|
|
|
struct cfg_loop_interval *interval = &cfg->loop_intervals[i];
|
|
|
|
|
|
|
|
if (interval->begin <= block->order_pos && interval->end == successor->order_pos)
|
|
|
|
{
|
|
|
|
action->target = i;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (action->target == UINT_MAX)
|
|
|
|
{
|
|
|
|
assert(successor->order_pos == block->order_pos + 1);
|
|
|
|
action->jump_type = JUMP_NONE;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
action->jump_type = JUMP_BREAK;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-02-06 06:08:49 -08:00
|
|
|
static enum vkd3d_result vsir_cfg_build_structured_program(struct vsir_cfg *cfg)
|
|
|
|
{
|
|
|
|
unsigned int i, stack_depth = 1, open_interval_idx = 0;
|
|
|
|
struct vsir_cfg_structure_list **stack = NULL;
|
|
|
|
|
|
|
|
/* It's enough to allocate up to the maximum interval stacking
|
|
|
|
* depth (plus one for the full program), but this is simpler. */
|
|
|
|
if (!(stack = vkd3d_calloc(cfg->loop_interval_count + 1, sizeof(*stack))))
|
|
|
|
goto fail;
|
|
|
|
cfg->structured_program.end = cfg->order.count;
|
|
|
|
stack[0] = &cfg->structured_program;
|
|
|
|
|
|
|
|
for (i = 0; i < cfg->order.count; ++i)
|
|
|
|
{
|
2024-02-06 06:08:49 -08:00
|
|
|
struct vsir_block *block = cfg->order.blocks[i];
|
2024-02-06 06:08:49 -08:00
|
|
|
struct vsir_cfg_structure *structure;
|
|
|
|
|
|
|
|
assert(stack_depth > 0);
|
|
|
|
|
|
|
|
/* Open loop intervals. */
|
|
|
|
while (open_interval_idx < cfg->loop_interval_count)
|
|
|
|
{
|
|
|
|
struct cfg_loop_interval *interval = &cfg->loop_intervals[open_interval_idx];
|
|
|
|
|
|
|
|
if (interval->begin != i)
|
|
|
|
break;
|
|
|
|
|
|
|
|
if (!(structure = vsir_cfg_structure_list_append(stack[stack_depth - 1], STRUCTURE_TYPE_LOOP)))
|
|
|
|
goto fail;
|
|
|
|
structure->u.loop.idx = open_interval_idx++;
|
|
|
|
|
|
|
|
structure->u.loop.body.end = interval->end;
|
|
|
|
stack[stack_depth++] = &structure->u.loop.body;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Execute the block. */
|
|
|
|
if (!(structure = vsir_cfg_structure_list_append(stack[stack_depth - 1], STRUCTURE_TYPE_BLOCK)))
|
|
|
|
goto fail;
|
2024-02-06 06:08:49 -08:00
|
|
|
structure->u.block = block;
|
2024-02-06 06:08:49 -08:00
|
|
|
|
2024-02-06 06:08:49 -08:00
|
|
|
/* Generate between zero and two jump instructions. */
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (block->end->opcode)
|
2024-02-06 06:08:49 -08:00
|
|
|
{
|
|
|
|
case VKD3DSIH_BRANCH:
|
|
|
|
{
|
|
|
|
struct vsir_cfg_edge_action action_true, action_false;
|
|
|
|
bool invert_condition = false;
|
|
|
|
|
|
|
|
if (vsir_register_is_label(&block->end->src[0].reg))
|
|
|
|
{
|
|
|
|
unsigned int target = label_from_src_param(&block->end->src[0]);
|
|
|
|
struct vsir_block *successor = &cfg->blocks[target - 1];
|
|
|
|
|
|
|
|
vsir_cfg_compute_edge_action(cfg, block, successor, &action_true);
|
|
|
|
action_false = action_true;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
unsigned int target = label_from_src_param(&block->end->src[1]);
|
|
|
|
struct vsir_block *successor = &cfg->blocks[target - 1];
|
|
|
|
|
|
|
|
vsir_cfg_compute_edge_action(cfg, block, successor, &action_true);
|
|
|
|
|
|
|
|
target = label_from_src_param(&block->end->src[2]);
|
|
|
|
successor = &cfg->blocks[target - 1];
|
|
|
|
|
|
|
|
vsir_cfg_compute_edge_action(cfg, block, successor, &action_false);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* This will happen if the branch is unconditional,
|
|
|
|
* but also if it's conditional with the same target
|
|
|
|
* in both branches, which can happen in some corner
|
|
|
|
* cases, e.g. when converting switch instructions to
|
|
|
|
* selection ladders. */
|
|
|
|
if (action_true.successor == action_false.successor)
|
|
|
|
{
|
|
|
|
assert(action_true.jump_type == action_false.jump_type);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
/* At most one branch can just fall through to the
|
|
|
|
* next block, in which case we make sure it's the
|
|
|
|
* false branch. */
|
|
|
|
if (action_true.jump_type == JUMP_NONE)
|
2024-02-29 07:00:51 -08:00
|
|
|
{
|
|
|
|
invert_condition = true;
|
|
|
|
}
|
|
|
|
else if (stack_depth >= 2)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure_list *inner_loop_frame = stack[stack_depth - 2];
|
|
|
|
struct vsir_cfg_structure *inner_loop = &inner_loop_frame->structures[inner_loop_frame->count - 1];
|
|
|
|
|
|
|
|
assert(inner_loop->type == STRUCTURE_TYPE_LOOP);
|
|
|
|
|
|
|
|
/* Otherwise, if one of the branches is
|
|
|
|
* continueing the inner loop we're inside,
|
|
|
|
* make sure it's the false branch (because it
|
|
|
|
* will be optimized out later). */
|
|
|
|
if (action_true.jump_type == JUMP_CONTINUE && action_true.target == inner_loop->u.loop.idx)
|
|
|
|
invert_condition = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (invert_condition)
|
2024-02-06 06:08:49 -08:00
|
|
|
{
|
|
|
|
struct vsir_cfg_edge_action tmp = action_true;
|
|
|
|
action_true = action_false;
|
|
|
|
action_false = tmp;
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(action_true.jump_type != JUMP_NONE);
|
|
|
|
|
|
|
|
if (!(structure = vsir_cfg_structure_list_append(stack[stack_depth - 1], STRUCTURE_TYPE_JUMP)))
|
|
|
|
goto fail;
|
|
|
|
structure->u.jump.type = action_true.jump_type;
|
|
|
|
structure->u.jump.target = action_true.target;
|
|
|
|
structure->u.jump.condition = &block->end->src[0];
|
|
|
|
structure->u.jump.invert_condition = invert_condition;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (action_false.jump_type != JUMP_NONE)
|
|
|
|
{
|
|
|
|
if (!(structure = vsir_cfg_structure_list_append(stack[stack_depth - 1], STRUCTURE_TYPE_JUMP)))
|
|
|
|
goto fail;
|
|
|
|
structure->u.jump.type = action_false.jump_type;
|
|
|
|
structure->u.jump.target = action_false.target;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case VKD3DSIH_RET:
|
|
|
|
if (!(structure = vsir_cfg_structure_list_append(stack[stack_depth - 1], STRUCTURE_TYPE_JUMP)))
|
|
|
|
goto fail;
|
|
|
|
structure->u.jump.type = JUMP_RET;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
vkd3d_unreachable();
|
|
|
|
}
|
2024-02-06 06:08:49 -08:00
|
|
|
|
|
|
|
/* Close loop intervals. */
|
|
|
|
while (stack_depth > 0)
|
|
|
|
{
|
|
|
|
if (stack[stack_depth - 1]->end != i + 1)
|
|
|
|
break;
|
|
|
|
|
|
|
|
--stack_depth;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(stack_depth == 0);
|
|
|
|
assert(open_interval_idx == cfg->loop_interval_count);
|
|
|
|
|
2024-02-06 06:08:49 -08:00
|
|
|
if (TRACE_ON())
|
|
|
|
vsir_cfg_dump_structured_program(cfg);
|
|
|
|
|
2024-02-06 06:08:49 -08:00
|
|
|
vkd3d_free(stack);
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
vkd3d_free(stack);
|
|
|
|
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
}
|
|
|
|
|
2024-03-13 05:39:24 -07:00
|
|
|
static void vsir_cfg_remove_trailing_continue(struct vsir_cfg *cfg,
|
|
|
|
struct vsir_cfg_structure_list *list, unsigned int target)
|
2024-02-20 14:29:48 -08:00
|
|
|
{
|
|
|
|
struct vsir_cfg_structure *last = &list->structures[list->count - 1];
|
|
|
|
|
|
|
|
if (last->type == STRUCTURE_TYPE_JUMP && last->u.jump.type == JUMP_CONTINUE
|
|
|
|
&& !last->u.jump.condition && last->u.jump.target == target)
|
2024-03-13 05:39:24 -07:00
|
|
|
{
|
2024-02-20 14:29:48 -08:00
|
|
|
--list->count;
|
2024-03-13 05:39:24 -07:00
|
|
|
assert(cfg->loop_intervals[target].target_count > 0);
|
|
|
|
--cfg->loop_intervals[target].target_count;
|
|
|
|
}
|
2024-02-20 14:29:48 -08:00
|
|
|
}
|
|
|
|
|
2024-03-13 05:09:51 -07:00
|
|
|
static struct vsir_cfg_structure *vsir_cfg_get_trailing_break(struct vsir_cfg_structure_list *list)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure *structure;
|
|
|
|
size_t count = list->count;
|
|
|
|
|
|
|
|
if (count == 0)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
structure = &list->structures[count - 1];
|
|
|
|
|
|
|
|
if (structure->type != STRUCTURE_TYPE_JUMP || structure->u.jump.type != JUMP_BREAK
|
|
|
|
|| structure->u.jump.condition)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
return structure;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* When the last instruction in both branches of a selection construct
|
|
|
|
* is an unconditional break, any of them can be moved after the
|
|
|
|
* selection construct. If they break the same loop both of them can
|
|
|
|
* be moved out, otherwise we can choose which one: we choose the one
|
|
|
|
* that breaks the innermost loop, because we hope to eventually
|
|
|
|
* remove the loop itself.
|
|
|
|
*
|
|
|
|
* In principle a similar movement could be done when the last
|
|
|
|
* instructions are continue and continue, or continue and break. But
|
|
|
|
* in practice I don't think those situations can happen given the
|
|
|
|
* previous passes we do on the program, so we don't care. */
|
2024-03-13 05:39:24 -07:00
|
|
|
static enum vkd3d_result vsir_cfg_move_breaks_out_of_selections(struct vsir_cfg *cfg,
|
|
|
|
struct vsir_cfg_structure_list *list)
|
2024-03-13 05:09:51 -07:00
|
|
|
{
|
|
|
|
struct vsir_cfg_structure *selection, *if_break, *else_break, *new_break;
|
|
|
|
unsigned int if_target, else_target, max_target;
|
|
|
|
size_t pos = list->count - 1;
|
|
|
|
|
|
|
|
selection = &list->structures[pos];
|
|
|
|
assert(selection->type == STRUCTURE_TYPE_SELECTION);
|
|
|
|
|
|
|
|
if_break = vsir_cfg_get_trailing_break(&selection->u.selection.if_body);
|
|
|
|
else_break = vsir_cfg_get_trailing_break(&selection->u.selection.else_body);
|
|
|
|
|
|
|
|
if (!if_break || !else_break)
|
|
|
|
return VKD3D_OK;
|
|
|
|
|
|
|
|
if_target = if_break->u.jump.target;
|
|
|
|
else_target = else_break->u.jump.target;
|
|
|
|
max_target = max(if_target, else_target);
|
|
|
|
|
|
|
|
if (!(new_break = vsir_cfg_structure_list_append(list, STRUCTURE_TYPE_JUMP)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
new_break->u.jump.type = JUMP_BREAK;
|
|
|
|
new_break->u.jump.target = max_target;
|
2024-03-13 05:39:24 -07:00
|
|
|
++cfg->loop_intervals[max_target].target_count;
|
2024-03-13 05:09:51 -07:00
|
|
|
|
|
|
|
/* Pointer `selection' could have been invalidated by the append
|
|
|
|
* operation. */
|
|
|
|
selection = &list->structures[pos];
|
|
|
|
assert(selection->type == STRUCTURE_TYPE_SELECTION);
|
|
|
|
|
|
|
|
if (if_target == max_target)
|
2024-03-13 05:39:24 -07:00
|
|
|
{
|
2024-03-13 05:09:51 -07:00
|
|
|
--selection->u.selection.if_body.count;
|
2024-03-13 05:39:24 -07:00
|
|
|
assert(cfg->loop_intervals[if_target].target_count > 0);
|
|
|
|
--cfg->loop_intervals[if_target].target_count;
|
|
|
|
}
|
|
|
|
|
2024-03-13 05:09:51 -07:00
|
|
|
if (else_target == max_target)
|
2024-03-13 05:39:24 -07:00
|
|
|
{
|
2024-03-13 05:09:51 -07:00
|
|
|
--selection->u.selection.else_body.count;
|
2024-03-13 05:39:24 -07:00
|
|
|
assert(cfg->loop_intervals[else_target].target_count > 0);
|
|
|
|
--cfg->loop_intervals[else_target].target_count;
|
|
|
|
}
|
2024-03-13 05:09:51 -07:00
|
|
|
|
2024-03-15 05:06:27 -07:00
|
|
|
/* If a branch becomes empty, make it the else branch, so we save a block. */
|
|
|
|
if (selection->u.selection.if_body.count == 0)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure_list tmp;
|
|
|
|
|
|
|
|
selection->u.selection.invert_condition = !selection->u.selection.invert_condition;
|
|
|
|
tmp = selection->u.selection.if_body;
|
|
|
|
selection->u.selection.if_body = selection->u.selection.else_body;
|
|
|
|
selection->u.selection.else_body = tmp;
|
|
|
|
}
|
|
|
|
|
2024-03-13 05:09:51 -07:00
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-03-14 08:26:52 -07:00
|
|
|
static enum vkd3d_result vsir_cfg_move_breaks_out_of_selections_recursively(struct vsir_cfg *cfg,
|
|
|
|
struct vsir_cfg_structure_list *list)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure *trailing;
|
|
|
|
|
|
|
|
if (list->count == 0)
|
|
|
|
return VKD3D_OK;
|
|
|
|
|
|
|
|
trailing = &list->structures[list->count - 1];
|
|
|
|
|
|
|
|
if (trailing->type != STRUCTURE_TYPE_SELECTION)
|
|
|
|
return VKD3D_OK;
|
|
|
|
|
|
|
|
vsir_cfg_move_breaks_out_of_selections_recursively(cfg, &trailing->u.selection.if_body);
|
|
|
|
vsir_cfg_move_breaks_out_of_selections_recursively(cfg, &trailing->u.selection.else_body);
|
|
|
|
|
|
|
|
return vsir_cfg_move_breaks_out_of_selections(cfg, list);
|
|
|
|
}
|
|
|
|
|
2024-03-13 05:39:24 -07:00
|
|
|
static enum vkd3d_result vsir_cfg_synthesize_selections(struct vsir_cfg *cfg,
|
|
|
|
struct vsir_cfg_structure_list *list)
|
2024-02-20 14:29:48 -08:00
|
|
|
{
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
for (i = 0; i < list->count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure *structure = &list->structures[i], new_selection, *new_jump;
|
|
|
|
|
|
|
|
if (structure->type != STRUCTURE_TYPE_JUMP || !structure->u.jump.condition)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
vsir_cfg_structure_init(&new_selection, STRUCTURE_TYPE_SELECTION);
|
|
|
|
new_selection.u.selection.condition = structure->u.jump.condition;
|
|
|
|
new_selection.u.selection.invert_condition = structure->u.jump.invert_condition;
|
|
|
|
|
|
|
|
if (!(new_jump = vsir_cfg_structure_list_append(&new_selection.u.selection.if_body,
|
|
|
|
STRUCTURE_TYPE_JUMP)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
new_jump->u.jump.type = structure->u.jump.type;
|
|
|
|
new_jump->u.jump.target = structure->u.jump.target;
|
|
|
|
|
|
|
|
/* Move the rest of the structure list in the else branch
|
|
|
|
* rather than leaving it after the selection construct. The
|
|
|
|
* reason is that this is more conducive to further
|
|
|
|
* optimization, because all the conditional `break's appear
|
|
|
|
* as the last instruction of a branch of a cascade of
|
|
|
|
* selection constructs at the end of the structure list we're
|
|
|
|
* processing, instead of being buried in the middle of the
|
|
|
|
* structure list itself. */
|
|
|
|
if ((ret = vsir_cfg_structure_list_append_from_region(&new_selection.u.selection.else_body,
|
|
|
|
&list->structures[i + 1], list->count - i - 1)) < 0)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
*structure = new_selection;
|
|
|
|
list->count = i + 1;
|
|
|
|
|
2024-03-13 05:39:24 -07:00
|
|
|
if ((ret = vsir_cfg_synthesize_selections(cfg, &structure->u.selection.else_body)) < 0)
|
2024-02-20 14:29:48 -08:00
|
|
|
return ret;
|
|
|
|
|
2024-03-13 05:39:24 -07:00
|
|
|
if ((ret = vsir_cfg_move_breaks_out_of_selections(cfg, list)) < 0)
|
2024-03-13 05:09:51 -07:00
|
|
|
return ret;
|
|
|
|
|
2024-02-20 14:29:48 -08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-03-13 14:43:50 -07:00
|
|
|
static enum vkd3d_result vsir_cfg_append_loop(struct vsir_cfg *cfg,
|
|
|
|
struct vsir_cfg_structure_list *new_list, struct vsir_cfg_structure *loop)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure_list *loop_body = &loop->u.loop.body;
|
|
|
|
unsigned int target, loop_idx = loop->u.loop.idx;
|
|
|
|
struct vsir_cfg_structure *trailing_break;
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
|
|
|
|
trailing_break = vsir_cfg_get_trailing_break(loop_body);
|
|
|
|
|
|
|
|
/* If the loop's last instruction is not a break, we cannot remove
|
|
|
|
* the loop itself. */
|
|
|
|
if (!trailing_break)
|
|
|
|
{
|
|
|
|
if ((ret = vsir_cfg_structure_list_append_from_region(new_list, loop, 1)) < 0)
|
|
|
|
return ret;
|
|
|
|
memset(loop, 0, sizeof(*loop));
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
target = trailing_break->u.jump.target;
|
|
|
|
assert(cfg->loop_intervals[target].target_count > 0);
|
|
|
|
|
|
|
|
/* If the loop is not targeted by any jump, we can remove it. The
|
|
|
|
* trailing `break' then targets another loop, so we have to keep
|
|
|
|
* it. */
|
|
|
|
if (cfg->loop_intervals[loop_idx].target_count == 0)
|
|
|
|
{
|
|
|
|
if ((ret = vsir_cfg_structure_list_append_from_region(new_list,
|
|
|
|
&loop_body->structures[0], loop_body->count)) < 0)
|
|
|
|
return ret;
|
|
|
|
loop_body->count = 0;
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* If the loop is targeted only by its own trailing `break'
|
|
|
|
* instruction, then we can remove it together with the `break'
|
|
|
|
* itself. */
|
|
|
|
if (target == loop_idx && cfg->loop_intervals[loop_idx].target_count == 1)
|
|
|
|
{
|
|
|
|
--cfg->loop_intervals[loop_idx].target_count;
|
|
|
|
if ((ret = vsir_cfg_structure_list_append_from_region(new_list,
|
|
|
|
&loop_body->structures[0], loop_body->count - 1)) < 0)
|
|
|
|
return ret;
|
|
|
|
loop_body->count = 0;
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((ret = vsir_cfg_structure_list_append_from_region(new_list, loop, 1)) < 0)
|
|
|
|
return ret;
|
|
|
|
memset(loop, 0, sizeof(*loop));
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-02-20 14:29:48 -08:00
|
|
|
static enum vkd3d_result vsir_cfg_optimize_recurse(struct vsir_cfg *cfg, struct vsir_cfg_structure_list *list)
|
|
|
|
{
|
2024-03-14 07:04:29 -07:00
|
|
|
struct vsir_cfg_structure_list old_list = *list, *new_list = list;
|
2024-02-20 14:29:48 -08:00
|
|
|
enum vkd3d_result ret;
|
|
|
|
size_t i;
|
|
|
|
|
2024-03-14 07:04:29 -07:00
|
|
|
memset(new_list, 0, sizeof(*new_list));
|
|
|
|
|
|
|
|
for (i = 0; i < old_list.count; ++i)
|
2024-02-20 14:29:48 -08:00
|
|
|
{
|
2024-03-14 07:33:19 -07:00
|
|
|
struct vsir_cfg_structure *loop = &old_list.structures[i], *selection;
|
2024-02-20 14:29:48 -08:00
|
|
|
struct vsir_cfg_structure_list *loop_body;
|
|
|
|
|
|
|
|
if (loop->type != STRUCTURE_TYPE_LOOP)
|
2024-03-13 14:43:50 -07:00
|
|
|
{
|
2024-03-14 07:04:29 -07:00
|
|
|
if ((ret = vsir_cfg_structure_list_append_from_region(new_list, loop, 1)) < 0)
|
|
|
|
goto out;
|
2024-03-13 14:43:50 -07:00
|
|
|
memset(loop, 0, sizeof(*loop));
|
2024-02-20 14:29:48 -08:00
|
|
|
continue;
|
2024-03-13 14:43:50 -07:00
|
|
|
}
|
2024-02-20 14:29:48 -08:00
|
|
|
|
|
|
|
loop_body = &loop->u.loop.body;
|
|
|
|
|
|
|
|
if (loop_body->count == 0)
|
2024-03-13 14:43:50 -07:00
|
|
|
{
|
2024-03-14 07:04:29 -07:00
|
|
|
if ((ret = vsir_cfg_structure_list_append_from_region(new_list, loop, 1)) < 0)
|
|
|
|
goto out;
|
2024-03-13 14:43:50 -07:00
|
|
|
memset(loop, 0, sizeof(*loop));
|
2024-02-20 14:29:48 -08:00
|
|
|
continue;
|
2024-03-13 14:43:50 -07:00
|
|
|
}
|
2024-02-20 14:29:48 -08:00
|
|
|
|
2024-03-13 05:39:24 -07:00
|
|
|
vsir_cfg_remove_trailing_continue(cfg, loop_body, loop->u.loop.idx);
|
2024-02-20 14:29:48 -08:00
|
|
|
|
|
|
|
if ((ret = vsir_cfg_optimize_recurse(cfg, loop_body)) < 0)
|
2024-03-14 07:04:29 -07:00
|
|
|
goto out;
|
2024-02-20 14:29:48 -08:00
|
|
|
|
2024-03-13 05:39:24 -07:00
|
|
|
if ((ret = vsir_cfg_synthesize_selections(cfg, loop_body)) < 0)
|
2024-03-14 07:04:29 -07:00
|
|
|
goto out;
|
2024-03-13 14:43:50 -07:00
|
|
|
|
2024-03-14 07:04:29 -07:00
|
|
|
if ((ret = vsir_cfg_append_loop(cfg, new_list, loop)) < 0)
|
|
|
|
goto out;
|
2024-03-14 07:33:19 -07:00
|
|
|
|
|
|
|
/* If the last pushed instruction is a selection and one of the branches terminates with a
|
|
|
|
* `break', start pushing to the other branch, in the hope of eventually push a `break'
|
|
|
|
* there too and be able to remove a loop. */
|
|
|
|
if (new_list->count == 0)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
selection = &new_list->structures[new_list->count - 1];
|
|
|
|
|
|
|
|
if (selection->type == STRUCTURE_TYPE_SELECTION)
|
|
|
|
{
|
|
|
|
if (vsir_cfg_get_trailing_break(&selection->u.selection.if_body))
|
|
|
|
new_list = &selection->u.selection.else_body;
|
|
|
|
else if (vsir_cfg_get_trailing_break(&selection->u.selection.else_body))
|
|
|
|
new_list = &selection->u.selection.if_body;
|
|
|
|
}
|
2024-02-20 14:29:48 -08:00
|
|
|
}
|
|
|
|
|
2024-03-14 08:26:52 -07:00
|
|
|
ret = vsir_cfg_move_breaks_out_of_selections_recursively(cfg, list);
|
2024-03-13 14:43:50 -07:00
|
|
|
|
2024-03-14 07:04:29 -07:00
|
|
|
out:
|
|
|
|
vsir_cfg_structure_list_cleanup(&old_list);
|
2024-03-13 14:43:50 -07:00
|
|
|
|
|
|
|
return ret;
|
2024-02-20 14:29:48 -08:00
|
|
|
}
|
|
|
|
|
2024-03-13 05:39:24 -07:00
|
|
|
static void vsir_cfg_count_targets(struct vsir_cfg *cfg, struct vsir_cfg_structure_list *list)
|
|
|
|
{
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
for (i = 0; i < list->count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure *structure = &list->structures[i];
|
|
|
|
|
|
|
|
switch (structure->type)
|
|
|
|
{
|
|
|
|
case STRUCTURE_TYPE_BLOCK:
|
|
|
|
break;
|
|
|
|
|
|
|
|
case STRUCTURE_TYPE_LOOP:
|
|
|
|
vsir_cfg_count_targets(cfg, &structure->u.loop.body);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case STRUCTURE_TYPE_SELECTION:
|
|
|
|
vsir_cfg_count_targets(cfg, &structure->u.selection.if_body);
|
|
|
|
vsir_cfg_count_targets(cfg, &structure->u.selection.else_body);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case STRUCTURE_TYPE_JUMP:
|
|
|
|
if (structure->u.jump.type == JUMP_BREAK || structure->u.jump.type == JUMP_CONTINUE)
|
|
|
|
++cfg->loop_intervals[structure->u.jump.target].target_count;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-03-18 04:38:51 -07:00
|
|
|
/* Trampolines are code gadgets used to emulate multilevel jumps (which are not natively supported
|
|
|
|
* by SPIR-V). A trampoline is inserted just after a loop and checks whether control has reached the
|
|
|
|
* intended site (i.e., we just jumped out of the target block) or if other levels of jumping are
|
|
|
|
* needed. For each jump a trampoline is required for all the loops between the jump itself and the
|
|
|
|
* target loop, excluding the target loop itself. */
|
|
|
|
static void vsir_cfg_mark_trampolines(struct vsir_cfg *cfg, struct vsir_cfg_structure_list *list,
|
|
|
|
struct vsir_cfg_structure *loop)
|
|
|
|
{
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
for (i = 0; i < list->count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure *structure = &list->structures[i];
|
|
|
|
|
|
|
|
switch (structure->type)
|
|
|
|
{
|
|
|
|
case STRUCTURE_TYPE_BLOCK:
|
|
|
|
break;
|
|
|
|
|
|
|
|
case STRUCTURE_TYPE_LOOP:
|
|
|
|
structure->u.loop.outer_loop = loop;
|
|
|
|
vsir_cfg_mark_trampolines(cfg, &structure->u.loop.body, structure);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case STRUCTURE_TYPE_SELECTION:
|
|
|
|
vsir_cfg_mark_trampolines(cfg, &structure->u.selection.if_body, loop);
|
|
|
|
vsir_cfg_mark_trampolines(cfg, &structure->u.selection.else_body, loop);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case STRUCTURE_TYPE_JUMP:
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure *l;
|
|
|
|
if (structure->u.jump.type != JUMP_BREAK && structure->u.jump.type != JUMP_CONTINUE)
|
|
|
|
break;
|
|
|
|
for (l = loop; l && l->u.loop.idx != structure->u.jump.target; l = l->u.loop.outer_loop)
|
|
|
|
{
|
|
|
|
assert(l->type == STRUCTURE_TYPE_LOOP);
|
|
|
|
l->u.loop.needs_trampoline = true;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-03-18 04:50:02 -07:00
|
|
|
/* Launchers are the counterpart of trampolines. A launcher is inserted just before a jump, and
|
|
|
|
* writes in a well-known variable what is the target of the jump. Trampolines will then read that
|
|
|
|
* variable to decide how to redirect the jump to its intended target. A launcher is needed each
|
|
|
|
* time the innermost loop containing the jump itself has a trampoline (independently of whether the
|
|
|
|
* jump is targeting that loop or not). */
|
|
|
|
static void vsir_cfg_mark_launchers(struct vsir_cfg *cfg, struct vsir_cfg_structure_list *list,
|
|
|
|
struct vsir_cfg_structure *loop)
|
|
|
|
{
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
for (i = 0; i < list->count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure *structure = &list->structures[i];
|
|
|
|
|
|
|
|
switch (structure->type)
|
|
|
|
{
|
|
|
|
case STRUCTURE_TYPE_BLOCK:
|
|
|
|
break;
|
|
|
|
|
|
|
|
case STRUCTURE_TYPE_LOOP:
|
|
|
|
vsir_cfg_mark_launchers(cfg, &structure->u.loop.body, structure);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case STRUCTURE_TYPE_SELECTION:
|
|
|
|
vsir_cfg_mark_launchers(cfg, &structure->u.selection.if_body, loop);
|
|
|
|
vsir_cfg_mark_launchers(cfg, &structure->u.selection.else_body, loop);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case STRUCTURE_TYPE_JUMP:
|
|
|
|
if (structure->u.jump.type != JUMP_BREAK && structure->u.jump.type != JUMP_CONTINUE)
|
|
|
|
break;
|
|
|
|
assert(loop && loop->type == STRUCTURE_TYPE_LOOP);
|
|
|
|
if (loop->u.loop.needs_trampoline)
|
|
|
|
structure->u.jump.needs_launcher = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-02-20 14:29:48 -08:00
|
|
|
static enum vkd3d_result vsir_cfg_optimize(struct vsir_cfg *cfg)
|
|
|
|
{
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
|
2024-03-13 05:39:24 -07:00
|
|
|
vsir_cfg_count_targets(cfg, &cfg->structured_program);
|
|
|
|
|
2024-02-20 14:29:48 -08:00
|
|
|
ret = vsir_cfg_optimize_recurse(cfg, &cfg->structured_program);
|
|
|
|
|
2024-03-18 04:50:02 -07:00
|
|
|
/* Trampolines and launchers cannot be marked with the same pass,
|
|
|
|
* because a jump might have to be marked as launcher even when it
|
|
|
|
* targets its innermost loop, if other jumps in the same loop
|
|
|
|
* need a trampoline anyway. So launchers can be discovered only
|
|
|
|
* once all the trampolines are known. */
|
2024-03-18 04:38:51 -07:00
|
|
|
vsir_cfg_mark_trampolines(cfg, &cfg->structured_program, NULL);
|
2024-03-18 04:50:02 -07:00
|
|
|
vsir_cfg_mark_launchers(cfg, &cfg->structured_program, NULL);
|
2024-03-18 04:38:51 -07:00
|
|
|
|
2024-02-20 14:29:48 -08:00
|
|
|
if (TRACE_ON())
|
|
|
|
vsir_cfg_dump_structured_program(cfg);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
static enum vkd3d_result vsir_cfg_structure_list_emit(struct vsir_cfg *cfg,
|
|
|
|
struct vsir_cfg_structure_list *list, unsigned int loop_idx);
|
|
|
|
|
2024-04-10 08:13:02 -07:00
|
|
|
static enum vkd3d_result vsir_cfg_structure_list_emit_block(struct vsir_cfg *cfg,
|
|
|
|
struct vsir_block *block)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_emit_target *target = cfg->target;
|
|
|
|
|
|
|
|
if (!reserve_instructions(&target->instructions, &target->ins_capacity,
|
|
|
|
target->ins_count + (block->end - block->begin)))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
memcpy(&target->instructions[target->ins_count], block->begin,
|
|
|
|
(char *)block->end - (char *)block->begin);
|
|
|
|
|
|
|
|
target->ins_count += block->end - block->begin;
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
static enum vkd3d_result vsir_cfg_structure_list_emit_loop(struct vsir_cfg *cfg,
|
|
|
|
struct vsir_cfg_structure_loop *loop, unsigned int loop_idx)
|
2024-02-20 03:47:19 -08:00
|
|
|
{
|
2024-04-10 06:58:34 -07:00
|
|
|
struct vsir_cfg_emit_target *target = cfg->target;
|
2024-02-20 03:47:19 -08:00
|
|
|
const struct vkd3d_shader_location no_loc = {0};
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
if (!reserve_instructions(&target->instructions, &target->ins_capacity, target->ins_count + 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
2024-02-20 03:47:19 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
vsir_instruction_init(&target->instructions[target->ins_count++], &no_loc, VKD3DSIH_LOOP);
|
2024-02-20 03:47:19 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
if ((ret = vsir_cfg_structure_list_emit(cfg, &loop->body, loop->idx)) < 0)
|
|
|
|
return ret;
|
2024-02-20 03:47:19 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
if (!reserve_instructions(&target->instructions, &target->ins_capacity, target->ins_count + 5))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
2024-02-20 03:47:19 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
vsir_instruction_init(&target->instructions[target->ins_count++], &no_loc, VKD3DSIH_ENDLOOP);
|
2024-02-20 03:47:19 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
/* Add a trampoline to implement multilevel jumping depending on the stored
|
|
|
|
* jump_target value. */
|
2024-03-18 04:38:51 -07:00
|
|
|
if (loop->needs_trampoline)
|
2024-04-10 08:20:05 -07:00
|
|
|
{
|
|
|
|
/* If the multilevel jump is a `continue' and the target is the loop we're inside
|
|
|
|
* right now, then we can finally do the `continue'. */
|
|
|
|
const unsigned int outer_continue_target = loop_idx << 1 | 1;
|
|
|
|
/* If the multilevel jump is a `continue' to any other target, or if it is a `break'
|
|
|
|
* and the target is not the loop we just finished emitting, then it means that
|
|
|
|
* we have to reach an outer loop, so we keep breaking. */
|
|
|
|
const unsigned int inner_break_target = loop->idx << 1;
|
|
|
|
|
|
|
|
if (!vsir_instruction_init_with_params(cfg->program, &target->instructions[target->ins_count],
|
|
|
|
&no_loc, VKD3DSIH_IEQ, 1, 2))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
2024-02-20 03:47:19 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
dst_param_init_temp_bool(&target->instructions[target->ins_count].dst[0], target->temp_count);
|
|
|
|
src_param_init_temp_uint(&target->instructions[target->ins_count].src[0], target->jump_target_temp_idx);
|
|
|
|
src_param_init_const_uint(&target->instructions[target->ins_count].src[1], outer_continue_target);
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
++target->ins_count;
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
if (!vsir_instruction_init_with_params(cfg->program, &target->instructions[target->ins_count],
|
|
|
|
&no_loc, VKD3DSIH_CONTINUEP, 0, 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
src_param_init_temp_bool(&target->instructions[target->ins_count].src[0], target->temp_count);
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
++target->ins_count;
|
|
|
|
++target->temp_count;
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
if (!vsir_instruction_init_with_params(cfg->program, &target->instructions[target->ins_count],
|
|
|
|
&no_loc, VKD3DSIH_IEQ, 1, 2))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
dst_param_init_temp_bool(&target->instructions[target->ins_count].dst[0], target->temp_count);
|
|
|
|
src_param_init_temp_uint(&target->instructions[target->ins_count].src[0], target->jump_target_temp_idx);
|
|
|
|
src_param_init_const_uint(&target->instructions[target->ins_count].src[1], inner_break_target);
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
++target->ins_count;
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
if (!vsir_instruction_init_with_params(cfg->program, &target->instructions[target->ins_count],
|
|
|
|
&no_loc, VKD3DSIH_BREAKP, 0, 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
target->instructions[target->ins_count].flags |= VKD3D_SHADER_CONDITIONAL_OP_Z;
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
src_param_init_temp_bool(&target->instructions[target->ins_count].src[0], target->temp_count);
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
++target->ins_count;
|
|
|
|
++target->temp_count;
|
|
|
|
}
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:30:16 -07:00
|
|
|
static enum vkd3d_result vsir_cfg_structure_list_emit_selection(struct vsir_cfg *cfg,
|
|
|
|
struct vsir_cfg_structure_selection *selection, unsigned int loop_idx)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_emit_target *target = cfg->target;
|
|
|
|
const struct vkd3d_shader_location no_loc = {0};
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
|
|
|
|
if (!reserve_instructions(&target->instructions, &target->ins_capacity, target->ins_count + 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
if (!vsir_instruction_init_with_params(cfg->program, &target->instructions[target->ins_count],
|
|
|
|
&no_loc, VKD3DSIH_IF, 0, 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
target->instructions[target->ins_count].src[0] = *selection->condition;
|
|
|
|
|
|
|
|
if (selection->invert_condition)
|
|
|
|
target->instructions[target->ins_count].flags |= VKD3D_SHADER_CONDITIONAL_OP_Z;
|
|
|
|
|
|
|
|
++target->ins_count;
|
|
|
|
|
|
|
|
if ((ret = vsir_cfg_structure_list_emit(cfg, &selection->if_body, loop_idx)) < 0)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
if (selection->else_body.count != 0)
|
|
|
|
{
|
|
|
|
if (!reserve_instructions(&target->instructions, &target->ins_capacity, target->ins_count + 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
vsir_instruction_init(&target->instructions[target->ins_count++], &no_loc, VKD3DSIH_ELSE);
|
|
|
|
|
|
|
|
if ((ret = vsir_cfg_structure_list_emit(cfg, &selection->else_body, loop_idx)) < 0)
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!reserve_instructions(&target->instructions, &target->ins_capacity, target->ins_count + 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
vsir_instruction_init(&target->instructions[target->ins_count++], &no_loc, VKD3DSIH_ENDIF);
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-04-10 08:37:51 -07:00
|
|
|
static enum vkd3d_result vsir_cfg_structure_list_emit_jump(struct vsir_cfg *cfg,
|
|
|
|
struct vsir_cfg_structure_jump *jump, unsigned int loop_idx)
|
2024-04-10 08:20:05 -07:00
|
|
|
{
|
|
|
|
struct vsir_cfg_emit_target *target = cfg->target;
|
|
|
|
const struct vkd3d_shader_location no_loc = {0};
|
2024-04-10 08:37:51 -07:00
|
|
|
/* Encode the jump target as the loop index plus a bit to remember whether
|
|
|
|
* we're breaking or continueing. */
|
|
|
|
unsigned int jump_target = jump->target << 1;
|
|
|
|
enum vkd3d_shader_opcode opcode;
|
|
|
|
|
|
|
|
switch (jump->type)
|
|
|
|
{
|
|
|
|
case JUMP_CONTINUE:
|
|
|
|
/* If we're continueing the loop we're directly inside, then we can emit a
|
|
|
|
* `continue'. Otherwise we first have to break all the loops between here
|
|
|
|
* and the loop to continue, recording our intention to continue
|
|
|
|
* in the lowest bit of jump_target. */
|
|
|
|
if (jump->target == loop_idx)
|
|
|
|
{
|
|
|
|
opcode = jump->condition ? VKD3DSIH_CONTINUEP : VKD3DSIH_CONTINUE;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
jump_target |= 1;
|
|
|
|
/* fall through */
|
|
|
|
|
|
|
|
case JUMP_BREAK:
|
|
|
|
opcode = jump->condition ? VKD3DSIH_BREAKP : VKD3DSIH_BREAK;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case JUMP_RET:
|
|
|
|
assert(!jump->condition);
|
|
|
|
opcode = VKD3DSIH_RET;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
vkd3d_unreachable();
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!reserve_instructions(&target->instructions, &target->ins_capacity, target->ins_count + 2))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
2024-03-18 04:50:02 -07:00
|
|
|
if (jump->needs_launcher)
|
2024-04-10 08:37:51 -07:00
|
|
|
{
|
|
|
|
if (!vsir_instruction_init_with_params(cfg->program, &target->instructions[target->ins_count],
|
|
|
|
&no_loc, VKD3DSIH_MOV, 1, 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
dst_param_init_temp_uint(&target->instructions[target->ins_count].dst[0], target->jump_target_temp_idx);
|
|
|
|
src_param_init_const_uint(&target->instructions[target->ins_count].src[0], jump_target);
|
|
|
|
|
|
|
|
++target->ins_count;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!vsir_instruction_init_with_params(cfg->program, &target->instructions[target->ins_count],
|
|
|
|
&no_loc, opcode, 0, !!jump->condition))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
|
|
|
if (jump->invert_condition)
|
|
|
|
target->instructions[target->ins_count].flags |= VKD3D_SHADER_CONDITIONAL_OP_Z;
|
|
|
|
|
|
|
|
if (jump->condition)
|
|
|
|
target->instructions[target->ins_count].src[0] = *jump->condition;
|
|
|
|
|
|
|
|
++target->ins_count;
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
static enum vkd3d_result vsir_cfg_structure_list_emit(struct vsir_cfg *cfg,
|
|
|
|
struct vsir_cfg_structure_list *list, unsigned int loop_idx)
|
|
|
|
{
|
2024-04-10 08:20:05 -07:00
|
|
|
enum vkd3d_result ret;
|
|
|
|
size_t i;
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
for (i = 0; i < list->count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_structure *structure = &list->structures[i];
|
vkd3d-shader/ir: Emit multilevel jumps in the structured program.
The new structurizer therefore reaches feature parity with the
older simple one, except for a couple of points:
* the old structurizer accepts any CFG, without requiring reducibility;
however, the DXIL specification requires the CFG to be reducible
anyway, so we're not really losing anything;
* the new structurizer additionally requires that no block has two
incoming back arrows; AFAIK this is condition that can happen,
but in practice it seems to be rare; also, it's not hard to add
support for it, as soon as it is decided it is useful.
On the other hand, the new structurizer makes use of the merging
information that are reconstructed from the CFG, which is important
for downstream optimization and fundamental for correctly emitting
tangled instructions.
Taking these considerations into account, the old structurizer is
considered superseded and is therefore removed.
2024-02-20 05:07:16 -08:00
|
|
|
|
2024-04-10 08:20:05 -07:00
|
|
|
switch (structure->type)
|
|
|
|
{
|
|
|
|
case STRUCTURE_TYPE_BLOCK:
|
|
|
|
if ((ret = vsir_cfg_structure_list_emit_block(cfg, structure->u.block)) < 0)
|
|
|
|
return ret;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case STRUCTURE_TYPE_LOOP:
|
|
|
|
if ((ret = vsir_cfg_structure_list_emit_loop(cfg, &structure->u.loop, loop_idx)) < 0)
|
|
|
|
return ret;
|
2024-02-20 03:47:19 -08:00
|
|
|
break;
|
|
|
|
|
2024-02-20 14:29:48 -08:00
|
|
|
case STRUCTURE_TYPE_SELECTION:
|
2024-04-10 08:30:16 -07:00
|
|
|
if ((ret = vsir_cfg_structure_list_emit_selection(cfg, &structure->u.selection,
|
|
|
|
loop_idx)) < 0)
|
2024-02-20 14:29:48 -08:00
|
|
|
return ret;
|
|
|
|
break;
|
|
|
|
|
2024-02-20 03:47:19 -08:00
|
|
|
case STRUCTURE_TYPE_JUMP:
|
2024-04-10 08:37:51 -07:00
|
|
|
if ((ret = vsir_cfg_structure_list_emit_jump(cfg, &structure->u.jump,
|
|
|
|
loop_idx)) < 0)
|
|
|
|
return ret;
|
2024-02-20 03:47:19 -08:00
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
vkd3d_unreachable();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
static enum vkd3d_result vsir_cfg_emit_structured_program(struct vsir_cfg *cfg)
|
|
|
|
{
|
2024-04-10 06:58:34 -07:00
|
|
|
return vsir_cfg_structure_list_emit(cfg, &cfg->structured_program, UINT_MAX);
|
2024-02-20 03:47:19 -08:00
|
|
|
}
|
|
|
|
|
2024-04-10 06:58:34 -07:00
|
|
|
static enum vkd3d_result vsir_program_structurize_function(struct vsir_program *program,
|
2024-04-11 07:33:19 -07:00
|
|
|
struct vkd3d_shader_message_context *message_context, struct vsir_cfg_emit_target *target,
|
|
|
|
size_t *pos)
|
2024-04-10 06:21:52 -07:00
|
|
|
{
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
struct vsir_cfg cfg;
|
|
|
|
|
2024-04-11 07:33:19 -07:00
|
|
|
if ((ret = vsir_cfg_init(&cfg, program, message_context, target, pos)) < 0)
|
2024-04-10 06:21:52 -07:00
|
|
|
return ret;
|
|
|
|
|
|
|
|
vsir_cfg_compute_dominators(&cfg);
|
|
|
|
|
|
|
|
if ((ret = vsir_cfg_compute_loops(&cfg)) < 0)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
if ((ret = vsir_cfg_sort_nodes(&cfg)) < 0)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
if ((ret = vsir_cfg_generate_synthetic_loop_intervals(&cfg)) < 0)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
if ((ret = vsir_cfg_build_structured_program(&cfg)) < 0)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
if ((ret = vsir_cfg_optimize(&cfg)) < 0)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
ret = vsir_cfg_emit_structured_program(&cfg);
|
|
|
|
|
|
|
|
out:
|
|
|
|
vsir_cfg_cleanup(&cfg);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2024-04-10 06:58:34 -07:00
|
|
|
static enum vkd3d_result vsir_program_structurize(struct vsir_program *program,
|
|
|
|
struct vkd3d_shader_message_context *message_context)
|
|
|
|
{
|
|
|
|
struct vsir_cfg_emit_target target = {0};
|
|
|
|
enum vkd3d_result ret;
|
2024-04-11 07:33:19 -07:00
|
|
|
size_t i;
|
2024-04-10 06:58:34 -07:00
|
|
|
|
|
|
|
target.jump_target_temp_idx = program->temp_count;
|
|
|
|
target.temp_count = program->temp_count + 1;
|
|
|
|
|
|
|
|
if (!reserve_instructions(&target.instructions, &target.ins_capacity, program->instructions.count))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
|
2024-04-11 07:33:19 -07:00
|
|
|
for (i = 0; i < program->instructions.count;)
|
2024-04-10 06:58:34 -07:00
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction *ins = &program->instructions.elements[i];
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (ins->opcode)
|
2024-04-11 07:33:19 -07:00
|
|
|
{
|
|
|
|
case VKD3DSIH_LABEL:
|
|
|
|
assert(program->shader_version.type != VKD3D_SHADER_TYPE_HULL);
|
|
|
|
TRACE("Structurizing a non-hull shader.\n");
|
|
|
|
if ((ret = vsir_program_structurize_function(program, message_context,
|
|
|
|
&target, &i)) < 0)
|
|
|
|
goto fail;
|
|
|
|
assert(i == program->instructions.count);
|
|
|
|
break;
|
2024-04-10 06:58:34 -07:00
|
|
|
|
2024-04-11 07:33:19 -07:00
|
|
|
case VKD3DSIH_HS_CONTROL_POINT_PHASE:
|
|
|
|
case VKD3DSIH_HS_FORK_PHASE:
|
|
|
|
case VKD3DSIH_HS_JOIN_PHASE:
|
|
|
|
assert(program->shader_version.type == VKD3D_SHADER_TYPE_HULL);
|
2024-04-24 11:34:42 -07:00
|
|
|
TRACE("Structurizing phase %u of a hull shader.\n", ins->opcode);
|
2024-04-11 07:33:19 -07:00
|
|
|
target.instructions[target.ins_count++] = *ins;
|
|
|
|
++i;
|
|
|
|
if ((ret = vsir_program_structurize_function(program, message_context,
|
|
|
|
&target, &i)) < 0)
|
|
|
|
goto fail;
|
|
|
|
break;
|
2024-04-10 06:58:34 -07:00
|
|
|
|
2024-04-11 07:33:19 -07:00
|
|
|
default:
|
|
|
|
if (!reserve_instructions(&target.instructions, &target.ins_capacity, target.ins_count + 1))
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
target.instructions[target.ins_count++] = *ins;
|
|
|
|
++i;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
2024-04-10 06:58:34 -07:00
|
|
|
|
|
|
|
vkd3d_free(program->instructions.elements);
|
|
|
|
program->instructions.elements = target.instructions;
|
|
|
|
program->instructions.capacity = target.ins_capacity;
|
|
|
|
program->instructions.count = target.ins_count;
|
|
|
|
program->temp_count = target.temp_count;
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
vkd3d_free(target.instructions);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2024-04-01 19:55:05 -07:00
|
|
|
static void register_map_undominated_use(struct vkd3d_shader_register *reg, struct ssas_to_temps_alloc *alloc,
|
|
|
|
struct vsir_block *block, struct vsir_block **origin_blocks)
|
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
2024-04-24 01:07:16 -07:00
|
|
|
if (register_is_ssa(reg))
|
|
|
|
{
|
|
|
|
i = reg->idx[0].offset;
|
|
|
|
if (alloc->table[i] == UINT_MAX && !vsir_block_dominates(origin_blocks[i], block))
|
|
|
|
alloc->table[i] = alloc->next_temp_idx++;
|
|
|
|
}
|
2024-04-01 19:55:05 -07:00
|
|
|
|
|
|
|
for (i = 0; i < reg->idx_count; ++i)
|
|
|
|
if (reg->idx[i].rel_addr)
|
|
|
|
register_map_undominated_use(®->idx[i].rel_addr->reg, alloc, block, origin_blocks);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Drivers are not necessarily optimised to handle very large numbers of temps. For example,
|
|
|
|
* using them only where necessary fixes stuttering issues in Horizon Zero Dawn on RADV.
|
|
|
|
* This can also result in the backend emitting less code because temps typically need an
|
|
|
|
* access chain and a load/store. Conversion of phi SSA values to temps should eliminate all
|
|
|
|
* undominated SSA use, but structurisation may create new occurrences. */
|
|
|
|
static enum vkd3d_result vsir_cfg_materialize_undominated_ssas_to_temps(struct vsir_cfg *cfg)
|
|
|
|
{
|
|
|
|
struct vsir_program *program = cfg->program;
|
|
|
|
struct ssas_to_temps_alloc alloc = {0};
|
|
|
|
struct vsir_block **origin_blocks;
|
|
|
|
unsigned int j;
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
if (!(origin_blocks = vkd3d_calloc(program->ssa_count, sizeof(*origin_blocks))))
|
|
|
|
{
|
|
|
|
ERR("Failed to allocate origin block array.\n");
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
}
|
|
|
|
if (!ssas_to_temps_alloc_init(&alloc, program->ssa_count, program->temp_count))
|
|
|
|
{
|
|
|
|
vkd3d_free(origin_blocks);
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < cfg->block_count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_block *block = &cfg->blocks[i];
|
|
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
|
2024-04-19 04:13:15 -07:00
|
|
|
if (block->label == 0)
|
|
|
|
continue;
|
|
|
|
|
2024-04-01 19:55:05 -07:00
|
|
|
for (ins = block->begin; ins <= block->end; ++ins)
|
|
|
|
{
|
|
|
|
for (j = 0; j < ins->dst_count; ++j)
|
|
|
|
{
|
|
|
|
if (register_is_ssa(&ins->dst[j].reg))
|
|
|
|
origin_blocks[ins->dst[j].reg.idx[0].offset] = block;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < cfg->block_count; ++i)
|
|
|
|
{
|
|
|
|
struct vsir_block *block = &cfg->blocks[i];
|
|
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
|
2024-04-19 04:13:15 -07:00
|
|
|
if (block->label == 0)
|
|
|
|
continue;
|
|
|
|
|
2024-04-01 19:55:05 -07:00
|
|
|
for (ins = block->begin; ins <= block->end; ++ins)
|
|
|
|
{
|
|
|
|
for (j = 0; j < ins->src_count; ++j)
|
|
|
|
register_map_undominated_use(&ins->src[j].reg, &alloc, block, origin_blocks);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (alloc.next_temp_idx == program->temp_count)
|
|
|
|
goto done;
|
|
|
|
|
|
|
|
TRACE("Emitting temps for %u values with undominated usage.\n", alloc.next_temp_idx - program->temp_count);
|
|
|
|
|
2024-04-12 07:19:49 -07:00
|
|
|
for (i = cfg->function_begin; i < cfg->function_end; ++i)
|
2024-04-01 19:55:05 -07:00
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction *ins = &program->instructions.elements[i];
|
|
|
|
|
|
|
|
for (j = 0; j < ins->dst_count; ++j)
|
|
|
|
materialize_ssas_to_temps_process_reg(program, &alloc, &ins->dst[j].reg);
|
|
|
|
|
|
|
|
for (j = 0; j < ins->src_count; ++j)
|
|
|
|
materialize_ssas_to_temps_process_reg(program, &alloc, &ins->src[j].reg);
|
|
|
|
}
|
|
|
|
|
|
|
|
program->temp_count = alloc.next_temp_idx;
|
|
|
|
done:
|
|
|
|
vkd3d_free(origin_blocks);
|
|
|
|
vkd3d_free(alloc.table);
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2024-04-12 07:19:49 -07:00
|
|
|
static enum vkd3d_result vsir_program_materialize_undominated_ssas_to_temps_in_function(
|
|
|
|
struct vsir_program *program, struct vkd3d_shader_message_context *message_context,
|
|
|
|
size_t *pos)
|
2024-04-10 06:26:50 -07:00
|
|
|
{
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
struct vsir_cfg cfg;
|
2024-04-11 07:33:19 -07:00
|
|
|
|
2024-04-12 07:19:49 -07:00
|
|
|
if ((ret = vsir_cfg_init(&cfg, program, message_context, NULL, pos)) < 0)
|
2024-04-10 06:26:50 -07:00
|
|
|
return ret;
|
|
|
|
|
|
|
|
vsir_cfg_compute_dominators(&cfg);
|
|
|
|
|
|
|
|
ret = vsir_cfg_materialize_undominated_ssas_to_temps(&cfg);
|
|
|
|
|
|
|
|
vsir_cfg_cleanup(&cfg);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2024-04-12 07:19:49 -07:00
|
|
|
static enum vkd3d_result vsir_program_materialize_undominated_ssas_to_temps(struct vsir_program *program,
|
|
|
|
struct vkd3d_shader_message_context *message_context)
|
|
|
|
{
|
|
|
|
enum vkd3d_result ret;
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
for (i = 0; i < program->instructions.count;)
|
|
|
|
{
|
|
|
|
struct vkd3d_shader_instruction *ins = &program->instructions.elements[i];
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (ins->opcode)
|
2024-04-12 07:19:49 -07:00
|
|
|
{
|
|
|
|
case VKD3DSIH_LABEL:
|
|
|
|
assert(program->shader_version.type != VKD3D_SHADER_TYPE_HULL);
|
|
|
|
TRACE("Materializing undominated SSAs in a non-hull shader.\n");
|
|
|
|
if ((ret = vsir_program_materialize_undominated_ssas_to_temps_in_function(
|
|
|
|
program, message_context, &i)) < 0)
|
|
|
|
return ret;
|
|
|
|
assert(i == program->instructions.count);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_HS_CONTROL_POINT_PHASE:
|
|
|
|
case VKD3DSIH_HS_FORK_PHASE:
|
|
|
|
case VKD3DSIH_HS_JOIN_PHASE:
|
|
|
|
assert(program->shader_version.type == VKD3D_SHADER_TYPE_HULL);
|
2024-04-24 11:34:42 -07:00
|
|
|
TRACE("Materializing undominated SSAs in phase %u of a hull shader.\n", ins->opcode);
|
2024-04-12 07:19:49 -07:00
|
|
|
++i;
|
|
|
|
if ((ret = vsir_program_materialize_undominated_ssas_to_temps_in_function(
|
|
|
|
program, message_context, &i)) < 0)
|
|
|
|
return ret;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
++i;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return VKD3D_OK;
|
|
|
|
}
|
|
|
|
|
2023-08-29 12:20:35 -07:00
|
|
|
struct validation_context
|
|
|
|
{
|
2024-03-13 06:37:45 -07:00
|
|
|
struct vkd3d_shader_message_context *message_context;
|
2024-01-16 07:40:30 -08:00
|
|
|
const struct vsir_program *program;
|
2023-08-29 12:35:01 -07:00
|
|
|
size_t instruction_idx;
|
2024-03-13 06:37:45 -07:00
|
|
|
struct vkd3d_shader_location null_location;
|
2024-01-10 02:58:44 -08:00
|
|
|
bool invalid_instruction_idx;
|
2024-03-13 06:37:45 -07:00
|
|
|
enum vkd3d_result status;
|
2023-10-24 17:55:37 -07:00
|
|
|
bool dcl_temps_found;
|
2023-11-06 14:11:03 -08:00
|
|
|
enum vkd3d_shader_opcode phase;
|
2024-01-12 00:28:14 -08:00
|
|
|
enum cf_type
|
|
|
|
{
|
|
|
|
CF_TYPE_UNKNOWN = 0,
|
|
|
|
CF_TYPE_STRUCTURED,
|
|
|
|
CF_TYPE_BLOCKS,
|
|
|
|
} cf_type;
|
2024-01-12 02:20:26 -08:00
|
|
|
bool inside_block;
|
2023-11-06 04:47:34 -08:00
|
|
|
|
2024-01-10 02:27:03 -08:00
|
|
|
struct validation_context_temp_data
|
|
|
|
{
|
|
|
|
enum vsir_dimension dimension;
|
|
|
|
size_t first_seen;
|
|
|
|
} *temps;
|
|
|
|
|
2024-01-10 02:31:40 -08:00
|
|
|
struct validation_context_ssa_data
|
|
|
|
{
|
|
|
|
enum vsir_dimension dimension;
|
2024-01-25 05:32:01 -08:00
|
|
|
enum vkd3d_data_type data_type;
|
2024-01-10 02:31:40 -08:00
|
|
|
size_t first_seen;
|
2024-01-10 02:58:44 -08:00
|
|
|
uint32_t write_mask;
|
|
|
|
uint32_t read_mask;
|
|
|
|
size_t first_assigned;
|
2024-01-10 02:31:40 -08:00
|
|
|
} *ssas;
|
|
|
|
|
2023-11-06 04:47:34 -08:00
|
|
|
enum vkd3d_shader_opcode *blocks;
|
|
|
|
size_t depth;
|
|
|
|
size_t blocks_capacity;
|
2023-08-29 12:20:35 -07:00
|
|
|
};
|
|
|
|
|
2023-08-29 12:35:01 -07:00
|
|
|
static void VKD3D_PRINTF_FUNC(3, 4) validator_error(struct validation_context *ctx,
|
|
|
|
enum vkd3d_shader_error error, const char *format, ...)
|
|
|
|
{
|
|
|
|
struct vkd3d_string_buffer buf;
|
|
|
|
va_list args;
|
|
|
|
|
|
|
|
vkd3d_string_buffer_init(&buf);
|
|
|
|
|
|
|
|
va_start(args, format);
|
|
|
|
vkd3d_string_buffer_vprintf(&buf, format, args);
|
|
|
|
va_end(args);
|
|
|
|
|
2024-01-10 02:58:44 -08:00
|
|
|
if (ctx->invalid_instruction_idx)
|
|
|
|
{
|
2024-03-13 06:37:45 -07:00
|
|
|
vkd3d_shader_error(ctx->message_context, &ctx->null_location, error, "%s", buf.buffer);
|
2024-01-10 02:58:44 -08:00
|
|
|
ERR("VSIR validation error: %s\n", buf.buffer);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2024-03-13 06:37:45 -07:00
|
|
|
const struct vkd3d_shader_instruction *ins = &ctx->program->instructions.elements[ctx->instruction_idx];
|
|
|
|
vkd3d_shader_error(ctx->message_context, &ins->location, error,
|
|
|
|
"instruction %zu: %s", ctx->instruction_idx + 1, buf.buffer);
|
2024-01-10 02:58:44 -08:00
|
|
|
ERR("VSIR validation error: instruction %zu: %s\n", ctx->instruction_idx + 1, buf.buffer);
|
|
|
|
}
|
2023-08-29 12:35:01 -07:00
|
|
|
|
|
|
|
vkd3d_string_buffer_cleanup(&buf);
|
2024-03-13 06:37:45 -07:00
|
|
|
|
|
|
|
if (!ctx->status)
|
|
|
|
ctx->status = VKD3D_ERROR_INVALID_SHADER;
|
2023-08-29 12:35:01 -07:00
|
|
|
}
|
|
|
|
|
2023-11-08 14:10:06 -08:00
|
|
|
static void vsir_validate_src_param(struct validation_context *ctx,
|
|
|
|
const struct vkd3d_shader_src_param *src);
|
|
|
|
|
2023-08-29 12:58:31 -07:00
|
|
|
static void vsir_validate_register(struct validation_context *ctx,
|
|
|
|
const struct vkd3d_shader_register *reg)
|
|
|
|
{
|
2024-01-18 04:05:24 -08:00
|
|
|
unsigned int i;
|
2023-11-07 14:21:11 -08:00
|
|
|
|
2023-08-29 12:58:31 -07:00
|
|
|
if (reg->type >= VKD3DSPR_COUNT)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE, "Invalid register type %#x.",
|
|
|
|
reg->type);
|
2023-10-24 17:37:40 -07:00
|
|
|
|
|
|
|
if (reg->precision >= VKD3D_SHADER_REGISTER_PRECISION_COUNT)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_PRECISION, "Invalid register precision %#x.",
|
|
|
|
reg->precision);
|
2023-10-24 17:41:09 -07:00
|
|
|
|
|
|
|
if (reg->data_type >= VKD3D_DATA_COUNT)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DATA_TYPE, "Invalid register data type %#x.",
|
|
|
|
reg->data_type);
|
2023-10-24 17:43:31 -07:00
|
|
|
|
|
|
|
if (reg->dimension >= VSIR_DIMENSION_COUNT)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DIMENSION, "Invalid register dimension %#x.",
|
|
|
|
reg->dimension);
|
2023-10-24 17:46:07 -07:00
|
|
|
|
|
|
|
if (reg->idx_count > ARRAY_SIZE(reg->idx))
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX_COUNT, "Invalid register index count %u.",
|
|
|
|
reg->idx_count);
|
2023-10-24 17:55:37 -07:00
|
|
|
|
2023-11-08 14:10:06 -08:00
|
|
|
for (i = 0; i < min(reg->idx_count, ARRAY_SIZE(reg->idx)); ++i)
|
|
|
|
{
|
|
|
|
const struct vkd3d_shader_src_param *param = reg->idx[i].rel_addr;
|
|
|
|
if (reg->idx[i].rel_addr)
|
|
|
|
vsir_validate_src_param(ctx, param);
|
|
|
|
}
|
|
|
|
|
2023-10-24 17:55:37 -07:00
|
|
|
switch (reg->type)
|
|
|
|
{
|
|
|
|
case VKD3DSPR_TEMP:
|
2024-01-10 02:27:03 -08:00
|
|
|
{
|
|
|
|
struct validation_context_temp_data *data;
|
|
|
|
|
2023-10-24 17:55:37 -07:00
|
|
|
if (reg->idx_count != 1)
|
2024-01-10 02:02:56 -08:00
|
|
|
{
|
2023-10-24 17:55:37 -07:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX_COUNT, "Invalid index count %u for a TEMP register.",
|
|
|
|
reg->idx_count);
|
2024-01-10 02:02:56 -08:00
|
|
|
break;
|
|
|
|
}
|
2023-10-24 17:55:37 -07:00
|
|
|
|
2024-01-10 02:02:56 -08:00
|
|
|
if (reg->idx[0].rel_addr)
|
2023-10-24 17:55:37 -07:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX, "Non-NULL relative address for a TEMP register.");
|
|
|
|
|
2024-03-13 06:37:45 -07:00
|
|
|
if (reg->idx[0].offset >= ctx->program->temp_count)
|
2024-01-10 02:27:03 -08:00
|
|
|
{
|
2023-11-07 14:21:11 -08:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX, "TEMP register index %u exceeds the maximum count %u.",
|
2024-03-13 06:37:45 -07:00
|
|
|
reg->idx[0].offset, ctx->program->temp_count);
|
2024-01-10 02:27:03 -08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
data = &ctx->temps[reg->idx[0].offset];
|
|
|
|
|
|
|
|
if (reg->dimension == VSIR_DIMENSION_NONE)
|
|
|
|
{
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DIMENSION, "Invalid dimension NONE for a TEMP register.");
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* TEMP registers can be scalar or vec4, provided that
|
|
|
|
* each individual register always appears with the same
|
|
|
|
* dimension. */
|
|
|
|
if (data->dimension == VSIR_DIMENSION_NONE)
|
|
|
|
{
|
|
|
|
data->dimension = reg->dimension;
|
|
|
|
data->first_seen = ctx->instruction_idx;
|
|
|
|
}
|
|
|
|
else if (data->dimension != reg->dimension)
|
|
|
|
{
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DIMENSION, "Invalid dimension %#x for a TEMP register: "
|
|
|
|
"it has already been seen with dimension %#x at instruction %zu.",
|
|
|
|
reg->dimension, data->dimension, data->first_seen);
|
|
|
|
}
|
2023-10-24 17:55:37 -07:00
|
|
|
break;
|
2024-01-10 02:27:03 -08:00
|
|
|
}
|
2023-10-24 17:55:37 -07:00
|
|
|
|
2023-12-21 06:54:13 -08:00
|
|
|
case VKD3DSPR_SSA:
|
2024-01-10 02:31:40 -08:00
|
|
|
{
|
|
|
|
struct validation_context_ssa_data *data;
|
|
|
|
|
2023-12-21 06:54:13 -08:00
|
|
|
if (reg->idx_count != 1)
|
2024-01-10 02:02:56 -08:00
|
|
|
{
|
2023-12-21 06:54:13 -08:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX_COUNT, "Invalid index count %u for a SSA register.",
|
|
|
|
reg->idx_count);
|
2024-01-10 02:02:56 -08:00
|
|
|
break;
|
|
|
|
}
|
2023-12-21 06:54:13 -08:00
|
|
|
|
2024-01-10 02:02:56 -08:00
|
|
|
if (reg->idx[0].rel_addr)
|
2023-12-21 06:54:13 -08:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX, "Non-NULL relative address for a SSA register.");
|
|
|
|
|
2024-01-17 12:14:41 -08:00
|
|
|
if (reg->idx[0].offset >= ctx->program->ssa_count)
|
2024-01-10 02:31:40 -08:00
|
|
|
{
|
2024-01-17 12:14:41 -08:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX,
|
|
|
|
"SSA register index %u exceeds the maximum count %u.",
|
|
|
|
reg->idx[0].offset, ctx->program->ssa_count);
|
2024-01-10 02:31:40 -08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
data = &ctx->ssas[reg->idx[0].offset];
|
|
|
|
|
|
|
|
if (reg->dimension == VSIR_DIMENSION_NONE)
|
|
|
|
{
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DIMENSION, "Invalid dimension NONE for a SSA register.");
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* SSA registers can be scalar or vec4, provided that each
|
|
|
|
* individual register always appears with the same
|
|
|
|
* dimension. */
|
|
|
|
if (data->dimension == VSIR_DIMENSION_NONE)
|
|
|
|
{
|
|
|
|
data->dimension = reg->dimension;
|
2024-01-25 05:32:01 -08:00
|
|
|
data->data_type = reg->data_type;
|
2024-01-10 02:31:40 -08:00
|
|
|
data->first_seen = ctx->instruction_idx;
|
|
|
|
}
|
2024-01-25 05:32:01 -08:00
|
|
|
else
|
2024-01-10 02:31:40 -08:00
|
|
|
{
|
2024-01-25 05:32:01 -08:00
|
|
|
if (data->dimension != reg->dimension)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DIMENSION, "Invalid dimension %#x for a SSA register: "
|
|
|
|
"it has already been seen with dimension %#x at instruction %zu.",
|
|
|
|
reg->dimension, data->dimension, data->first_seen);
|
|
|
|
|
|
|
|
if (data_type_is_64_bit(data->data_type) != data_type_is_64_bit(reg->data_type))
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DATA_TYPE, "Invalid data type %#x for a SSA register: "
|
|
|
|
"it has already been seen with data type %#x at instruction %zu.",
|
|
|
|
reg->data_type, data->data_type, data->first_seen);
|
2024-01-10 02:31:40 -08:00
|
|
|
}
|
2023-12-21 06:54:13 -08:00
|
|
|
break;
|
2024-01-10 02:31:40 -08:00
|
|
|
}
|
2023-12-21 06:54:13 -08:00
|
|
|
|
2023-12-22 05:36:30 -08:00
|
|
|
case VKD3DSPR_LABEL:
|
|
|
|
if (reg->precision != VKD3D_SHADER_REGISTER_PRECISION_DEFAULT)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_PRECISION, "Invalid precision %#x for a LABEL register.",
|
|
|
|
reg->precision);
|
|
|
|
|
2024-03-14 14:40:12 -07:00
|
|
|
if (reg->data_type != VKD3D_DATA_UNUSED)
|
2023-12-22 05:36:30 -08:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DATA_TYPE, "Invalid data type %#x for a LABEL register.",
|
|
|
|
reg->data_type);
|
|
|
|
|
|
|
|
if (reg->dimension != VSIR_DIMENSION_NONE)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DIMENSION, "Invalid dimension %#x for a LABEL register.",
|
|
|
|
reg->dimension);
|
|
|
|
|
|
|
|
if (reg->idx_count != 1)
|
|
|
|
{
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX_COUNT, "Invalid index count %u for a LABEL register.",
|
|
|
|
reg->idx_count);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (reg->idx[0].rel_addr)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX, "Non-NULL relative address for a LABEL register.");
|
|
|
|
|
|
|
|
/* Index == 0 is invalid, but it is temporarily allowed
|
|
|
|
* for intermediate stages. Once we support validation
|
|
|
|
* dialects we can selectively check for that. */
|
2024-01-18 11:34:52 -08:00
|
|
|
if (reg->idx[0].offset > ctx->program->block_count)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX,
|
|
|
|
"LABEL register index %u exceeds the maximum count %u.",
|
|
|
|
reg->idx[0].offset, ctx->program->block_count);
|
2023-12-22 05:36:30 -08:00
|
|
|
break;
|
|
|
|
|
2023-11-08 14:12:30 -08:00
|
|
|
case VKD3DSPR_NULL:
|
|
|
|
if (reg->idx_count != 0)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX_COUNT, "Invalid index count %u for a NULL register.",
|
|
|
|
reg->idx_count);
|
|
|
|
break;
|
|
|
|
|
2023-11-08 14:17:05 -08:00
|
|
|
case VKD3DSPR_IMMCONST:
|
|
|
|
if (reg->idx_count != 0)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX_COUNT, "Invalid index count %u for a IMMCONST register.",
|
|
|
|
reg->idx_count);
|
|
|
|
break;
|
|
|
|
|
2023-11-08 14:17:39 -08:00
|
|
|
case VKD3DSPR_IMMCONST64:
|
|
|
|
if (reg->idx_count != 0)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_INDEX_COUNT, "Invalid index count %u for a IMMCONST64 register.",
|
|
|
|
reg->idx_count);
|
|
|
|
break;
|
|
|
|
|
2023-10-24 17:55:37 -07:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2023-08-29 12:58:31 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void vsir_validate_dst_param(struct validation_context *ctx,
|
|
|
|
const struct vkd3d_shader_dst_param *dst)
|
|
|
|
{
|
|
|
|
vsir_validate_register(ctx, &dst->reg);
|
2023-08-29 13:17:28 -07:00
|
|
|
|
|
|
|
if (dst->write_mask & ~VKD3DSP_WRITEMASK_ALL)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_WRITE_MASK, "Destination has invalid write mask %#x.",
|
|
|
|
dst->write_mask);
|
|
|
|
|
2023-12-21 07:04:10 -08:00
|
|
|
switch (dst->reg.dimension)
|
|
|
|
{
|
|
|
|
case VSIR_DIMENSION_SCALAR:
|
|
|
|
if (dst->write_mask != VKD3DSP_WRITEMASK_0)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_WRITE_MASK, "Scalar destination has invalid write mask %#x.",
|
|
|
|
dst->write_mask);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VSIR_DIMENSION_VEC4:
|
|
|
|
if (dst->write_mask == 0)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_WRITE_MASK, "Vec4 destination has empty write mask.");
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
if (dst->write_mask != 0)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_WRITE_MASK, "Destination of dimension %u has invalid write mask %#x.",
|
|
|
|
dst->reg.dimension, dst->write_mask);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2023-08-29 13:17:28 -07:00
|
|
|
if (dst->modifiers & ~VKD3DSPDM_MASK)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_MODIFIERS, "Destination has invalid modifiers %#x.",
|
|
|
|
dst->modifiers);
|
|
|
|
|
|
|
|
switch (dst->shift)
|
|
|
|
{
|
|
|
|
case 0:
|
|
|
|
case 1:
|
|
|
|
case 2:
|
|
|
|
case 3:
|
|
|
|
case 13:
|
|
|
|
case 14:
|
|
|
|
case 15:
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_SHIFT, "Destination has invalid shift %#x.",
|
|
|
|
dst->shift);
|
|
|
|
}
|
2024-01-10 02:58:44 -08:00
|
|
|
|
2024-01-18 15:02:13 -08:00
|
|
|
switch (dst->reg.type)
|
2024-01-10 02:58:44 -08:00
|
|
|
{
|
2024-01-18 15:02:13 -08:00
|
|
|
case VKD3DSPR_SSA:
|
2024-03-13 06:37:45 -07:00
|
|
|
if (dst->reg.idx[0].offset < ctx->program->ssa_count)
|
2024-01-18 15:02:13 -08:00
|
|
|
{
|
|
|
|
struct validation_context_ssa_data *data = &ctx->ssas[dst->reg.idx[0].offset];
|
2024-01-10 02:58:44 -08:00
|
|
|
|
2024-01-18 15:02:13 -08:00
|
|
|
if (data->write_mask == 0)
|
|
|
|
{
|
|
|
|
data->write_mask = dst->write_mask;
|
|
|
|
data->first_assigned = ctx->instruction_idx;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_SSA_USAGE,
|
|
|
|
"SSA register is already assigned at instruction %zu.",
|
|
|
|
data->first_assigned);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
2024-01-12 06:39:52 -08:00
|
|
|
case VKD3DSPR_IMMCONST:
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE,
|
|
|
|
"Invalid IMMCONST register used as destination parameter.");
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSPR_IMMCONST64:
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE,
|
|
|
|
"Invalid IMMCONST64 register used as destination parameter.");
|
|
|
|
break;
|
|
|
|
|
2024-01-18 15:02:13 -08:00
|
|
|
default:
|
|
|
|
break;
|
2024-01-10 02:58:44 -08:00
|
|
|
}
|
2023-08-29 12:58:31 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void vsir_validate_src_param(struct validation_context *ctx,
|
|
|
|
const struct vkd3d_shader_src_param *src)
|
|
|
|
{
|
|
|
|
vsir_validate_register(ctx, &src->reg);
|
2023-08-29 13:23:02 -07:00
|
|
|
|
|
|
|
if (src->swizzle & ~0x03030303u)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_SWIZZLE, "Source has invalid swizzle %#x.",
|
|
|
|
src->swizzle);
|
|
|
|
|
2023-12-21 06:58:46 -08:00
|
|
|
if (src->reg.dimension != VSIR_DIMENSION_VEC4 && src->swizzle != 0)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_SWIZZLE, "Source of dimension %u has invalid swizzle %#x.",
|
|
|
|
src->reg.dimension, src->swizzle);
|
|
|
|
|
2023-08-29 13:23:02 -07:00
|
|
|
if (src->modifiers >= VKD3DSPSM_COUNT)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_MODIFIERS, "Source has invalid modifiers %#x.",
|
|
|
|
src->modifiers);
|
2024-01-10 02:58:44 -08:00
|
|
|
|
2024-01-18 15:02:13 -08:00
|
|
|
switch (src->reg.type)
|
2024-01-10 02:58:44 -08:00
|
|
|
{
|
2024-01-18 15:02:13 -08:00
|
|
|
case VKD3DSPR_SSA:
|
2024-03-13 06:37:45 -07:00
|
|
|
if (src->reg.idx[0].offset < ctx->program->ssa_count)
|
2024-01-18 15:02:13 -08:00
|
|
|
{
|
|
|
|
struct validation_context_ssa_data *data = &ctx->ssas[src->reg.idx[0].offset];
|
|
|
|
unsigned int i;
|
2024-01-10 02:58:44 -08:00
|
|
|
|
2024-01-18 15:02:13 -08:00
|
|
|
for (i = 0; i < VKD3D_VEC4_SIZE; ++i)
|
|
|
|
data->read_mask |= (1u << vsir_swizzle_get_component(src->swizzle, i));
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
break;
|
2024-01-10 02:58:44 -08:00
|
|
|
}
|
2023-08-29 12:58:31 -07:00
|
|
|
}
|
|
|
|
|
2023-10-24 18:08:00 -07:00
|
|
|
static void vsir_validate_dst_count(struct validation_context *ctx,
|
|
|
|
const struct vkd3d_shader_instruction *instruction, unsigned int count)
|
|
|
|
{
|
|
|
|
if (instruction->dst_count != count)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DEST_COUNT,
|
|
|
|
"Invalid destination count %u for an instruction of type %#x, expected %u.",
|
2024-04-24 11:34:42 -07:00
|
|
|
instruction->dst_count, instruction->opcode, count);
|
2023-10-24 18:08:00 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void vsir_validate_src_count(struct validation_context *ctx,
|
|
|
|
const struct vkd3d_shader_instruction *instruction, unsigned int count)
|
|
|
|
{
|
|
|
|
if (instruction->src_count != count)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_SOURCE_COUNT,
|
|
|
|
"Invalid source count %u for an instruction of type %#x, expected %u.",
|
2024-04-24 11:34:42 -07:00
|
|
|
instruction->src_count, instruction->opcode, count);
|
2023-10-24 18:08:00 -07:00
|
|
|
}
|
|
|
|
|
2023-12-22 08:11:41 -08:00
|
|
|
static bool vsir_validate_src_min_count(struct validation_context *ctx,
|
|
|
|
const struct vkd3d_shader_instruction *instruction, unsigned int count)
|
|
|
|
{
|
|
|
|
if (instruction->src_count < count)
|
|
|
|
{
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_SOURCE_COUNT,
|
|
|
|
"Invalid source count %u for an instruction of type %#x, expected at least %u.",
|
2024-04-24 11:34:42 -07:00
|
|
|
instruction->src_count, instruction->opcode, count);
|
2023-12-22 08:11:41 -08:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool vsir_validate_src_max_count(struct validation_context *ctx,
|
|
|
|
const struct vkd3d_shader_instruction *instruction, unsigned int count)
|
|
|
|
{
|
|
|
|
if (instruction->src_count > count)
|
|
|
|
{
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_SOURCE_COUNT,
|
|
|
|
"Invalid source count %u for an instruction of type %#x, expected at most %u.",
|
2024-04-24 11:34:42 -07:00
|
|
|
instruction->src_count, instruction->opcode, count);
|
2023-12-22 08:11:41 -08:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2024-01-12 00:28:14 -08:00
|
|
|
static const char *name_from_cf_type(enum cf_type type)
|
|
|
|
{
|
|
|
|
switch (type)
|
|
|
|
{
|
|
|
|
case CF_TYPE_STRUCTURED:
|
|
|
|
return "structured";
|
|
|
|
case CF_TYPE_BLOCKS:
|
|
|
|
return "block-based";
|
|
|
|
default:
|
|
|
|
vkd3d_unreachable();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void vsir_validate_cf_type(struct validation_context *ctx,
|
|
|
|
const struct vkd3d_shader_instruction *instruction, enum cf_type expected_type)
|
|
|
|
{
|
|
|
|
assert(ctx->cf_type != CF_TYPE_UNKNOWN);
|
|
|
|
assert(expected_type != CF_TYPE_UNKNOWN);
|
|
|
|
if (ctx->cf_type != expected_type)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW, "Invalid instruction %#x in %s shader.",
|
2024-04-24 11:34:42 -07:00
|
|
|
instruction->opcode, name_from_cf_type(ctx->cf_type));
|
2024-01-12 00:28:14 -08:00
|
|
|
}
|
|
|
|
|
2023-08-29 12:35:01 -07:00
|
|
|
static void vsir_validate_instruction(struct validation_context *ctx)
|
|
|
|
{
|
2024-01-16 07:40:30 -08:00
|
|
|
const struct vkd3d_shader_version *version = &ctx->program->shader_version;
|
2024-01-16 05:47:34 -08:00
|
|
|
const struct vkd3d_shader_instruction *instruction;
|
2023-08-29 12:58:31 -07:00
|
|
|
size_t i;
|
2023-08-29 12:35:01 -07:00
|
|
|
|
2024-01-16 07:40:30 -08:00
|
|
|
instruction = &ctx->program->instructions.elements[ctx->instruction_idx];
|
2023-08-29 12:35:01 -07:00
|
|
|
|
2023-08-29 12:58:31 -07:00
|
|
|
for (i = 0; i < instruction->dst_count; ++i)
|
|
|
|
vsir_validate_dst_param(ctx, &instruction->dst[i]);
|
|
|
|
|
|
|
|
for (i = 0; i < instruction->src_count; ++i)
|
|
|
|
vsir_validate_src_param(ctx, &instruction->src[i]);
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
if (instruction->opcode >= VKD3DSIH_INVALID)
|
2023-08-29 12:35:01 -07:00
|
|
|
{
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_HANDLER, "Invalid instruction handler %#x.",
|
2024-04-24 11:34:42 -07:00
|
|
|
instruction->opcode);
|
2023-08-29 12:35:01 -07:00
|
|
|
}
|
2023-10-24 18:08:00 -07:00
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (instruction->opcode)
|
2023-11-06 14:11:03 -08:00
|
|
|
{
|
|
|
|
case VKD3DSIH_HS_DECLS:
|
|
|
|
case VKD3DSIH_HS_CONTROL_POINT_PHASE:
|
|
|
|
case VKD3DSIH_HS_FORK_PHASE:
|
|
|
|
case VKD3DSIH_HS_JOIN_PHASE:
|
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 0);
|
2024-01-16 07:40:30 -08:00
|
|
|
if (version->type != VKD3D_SHADER_TYPE_HULL)
|
2024-04-24 11:34:42 -07:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_HANDLER,
|
|
|
|
"Phase instruction %#x is only valid in a hull shader.",
|
|
|
|
instruction->opcode);
|
2023-11-06 14:11:03 -08:00
|
|
|
if (ctx->depth != 0)
|
2024-04-24 11:34:42 -07:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW,
|
|
|
|
"Phase instruction %#x must appear to top level.",
|
|
|
|
instruction->opcode);
|
|
|
|
ctx->phase = instruction->opcode;
|
2023-11-07 14:21:11 -08:00
|
|
|
ctx->dcl_temps_found = false;
|
2023-11-06 14:11:03 -08:00
|
|
|
return;
|
|
|
|
|
2024-04-04 20:54:25 -07:00
|
|
|
case VKD3DSIH_DCL_HS_MAX_TESSFACTOR:
|
|
|
|
/* Exclude non-finite values. */
|
|
|
|
if (!(instruction->declaration.max_tessellation_factor >= 1.0f
|
|
|
|
&& instruction->declaration.max_tessellation_factor <= 64.0f))
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_TESSELLATION, "Max tessellation factor %f is invalid.",
|
|
|
|
instruction->declaration.max_tessellation_factor);
|
|
|
|
return;
|
|
|
|
|
2024-04-18 19:45:09 -07:00
|
|
|
case VKD3DSIH_DCL_INPUT_PRIMITIVE:
|
|
|
|
if (instruction->declaration.primitive_type.type == VKD3D_PT_UNDEFINED
|
|
|
|
|| instruction->declaration.primitive_type.type >= VKD3D_PT_COUNT)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_GS, "GS input primitive %u is invalid.",
|
|
|
|
instruction->declaration.primitive_type.type);
|
|
|
|
return;
|
|
|
|
|
|
|
|
case VKD3DSIH_DCL_VERTICES_OUT:
|
|
|
|
if (instruction->declaration.count > 1024)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_GS, "GS output vertex count %u is invalid.",
|
|
|
|
instruction->declaration.count);
|
|
|
|
return;
|
|
|
|
|
|
|
|
case VKD3DSIH_DCL_OUTPUT_TOPOLOGY:
|
|
|
|
if (instruction->declaration.primitive_type.type == VKD3D_PT_UNDEFINED
|
|
|
|
|| instruction->declaration.primitive_type.type >= VKD3D_PT_COUNT)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_GS, "GS output primitive %u is invalid.",
|
|
|
|
instruction->declaration.primitive_type.type);
|
|
|
|
return;
|
|
|
|
|
|
|
|
case VKD3DSIH_DCL_GS_INSTANCES:
|
|
|
|
if (!instruction->declaration.count || instruction->declaration.count > 32)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_GS, "GS instance count %u is invalid.",
|
|
|
|
instruction->declaration.count);
|
|
|
|
return;
|
|
|
|
|
2024-04-04 20:54:25 -07:00
|
|
|
case VKD3DSIH_DCL_OUTPUT_CONTROL_POINT_COUNT:
|
|
|
|
if (!instruction->declaration.count || instruction->declaration.count > 32)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_TESSELLATION, "Output control point count %u is invalid.",
|
|
|
|
instruction->declaration.count);
|
|
|
|
return;
|
|
|
|
|
|
|
|
case VKD3DSIH_DCL_TESSELLATOR_DOMAIN:
|
|
|
|
if (instruction->declaration.tessellator_domain == VKD3D_TESSELLATOR_DOMAIN_INVALID
|
|
|
|
|| instruction->declaration.tessellator_domain >= VKD3D_TESSELLATOR_DOMAIN_COUNT)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_TESSELLATION,
|
|
|
|
"Tessellator domain %#x is invalid.", instruction->declaration.tessellator_domain);
|
|
|
|
return;
|
|
|
|
|
|
|
|
case VKD3DSIH_DCL_TESSELLATOR_OUTPUT_PRIMITIVE:
|
|
|
|
if (!instruction->declaration.tessellator_output_primitive
|
|
|
|
|| instruction->declaration.tessellator_output_primitive > VKD3D_SHADER_TESSELLATOR_OUTPUT_TRIANGLE_CCW)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_TESSELLATION,
|
|
|
|
"Tessellator output primitive %#x is invalid.", instruction->declaration.tessellator_output_primitive);
|
|
|
|
return;
|
|
|
|
|
|
|
|
case VKD3DSIH_DCL_TESSELLATOR_PARTITIONING:
|
|
|
|
if (!instruction->declaration.tessellator_partitioning
|
|
|
|
|| instruction->declaration.tessellator_partitioning > VKD3D_SHADER_TESSELLATOR_PARTITIONING_FRACTIONAL_EVEN)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_TESSELLATION,
|
|
|
|
"Tessellator partitioning %#x is invalid.", instruction->declaration.tessellator_partitioning);
|
|
|
|
return;
|
|
|
|
|
2023-11-06 14:11:03 -08:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2024-04-23 19:28:45 -07:00
|
|
|
/* Only DCL instructions may occur outside hull shader phases. */
|
|
|
|
if (!vsir_instruction_is_dcl(instruction) && version->type == VKD3D_SHADER_TYPE_HULL
|
|
|
|
&& ctx->phase == VKD3DSIH_INVALID)
|
2024-01-16 07:40:30 -08:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_HANDLER,
|
|
|
|
"Instruction %#x appear before any phase instruction in a hull shader.",
|
2024-04-24 11:34:42 -07:00
|
|
|
instruction->opcode);
|
2023-11-06 14:11:03 -08:00
|
|
|
|
2024-01-12 00:28:14 -08:00
|
|
|
/* We support two different control flow types in shaders:
|
|
|
|
* block-based, like DXIL and SPIR-V, and structured, like D3DBC
|
|
|
|
* and TPF. The shader is detected as block-based when its first
|
|
|
|
* instruction, except for DCL_* and phases, is a LABEL. Currently
|
|
|
|
* we mandate that each shader is either purely block-based or
|
|
|
|
* purely structured. In principle we could allow structured
|
|
|
|
* constructs in a block, provided they are confined in a single
|
|
|
|
* block, but need for that hasn't arisen yet, so we don't. */
|
2024-01-22 11:35:28 -08:00
|
|
|
if (ctx->cf_type == CF_TYPE_UNKNOWN && !vsir_instruction_is_dcl(instruction))
|
2024-01-12 00:28:14 -08:00
|
|
|
{
|
2024-04-24 11:34:42 -07:00
|
|
|
if (instruction->opcode == VKD3DSIH_LABEL)
|
2024-01-12 00:28:14 -08:00
|
|
|
ctx->cf_type = CF_TYPE_BLOCKS;
|
|
|
|
else
|
|
|
|
ctx->cf_type = CF_TYPE_STRUCTURED;
|
|
|
|
}
|
|
|
|
|
2024-01-12 02:20:26 -08:00
|
|
|
if (ctx->cf_type == CF_TYPE_BLOCKS && !vsir_instruction_is_dcl(instruction))
|
|
|
|
{
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (instruction->opcode)
|
2024-01-12 02:20:26 -08:00
|
|
|
{
|
|
|
|
case VKD3DSIH_LABEL:
|
|
|
|
if (ctx->inside_block)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW, "Invalid LABEL instruction inside a block.");
|
|
|
|
ctx->inside_block = true;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_RET:
|
|
|
|
case VKD3DSIH_BRANCH:
|
|
|
|
case VKD3DSIH_SWITCH_MONOLITHIC:
|
|
|
|
if (!ctx->inside_block)
|
2024-04-24 11:34:42 -07:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW,
|
|
|
|
"Invalid instruction %#x outside any block.",
|
|
|
|
instruction->opcode);
|
2024-01-12 02:20:26 -08:00
|
|
|
ctx->inside_block = false;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
if (!ctx->inside_block)
|
2024-04-24 11:34:42 -07:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW,
|
|
|
|
"Invalid instruction %#x outside any block.",
|
|
|
|
instruction->opcode);
|
2024-01-12 02:20:26 -08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-04-24 11:34:42 -07:00
|
|
|
switch (instruction->opcode)
|
2023-10-24 18:08:00 -07:00
|
|
|
{
|
|
|
|
case VKD3DSIH_DCL_TEMPS:
|
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 0);
|
2023-11-07 14:21:11 -08:00
|
|
|
if (ctx->dcl_temps_found)
|
2023-10-24 17:55:37 -07:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_DUPLICATE_DCL_TEMPS, "Duplicate DCL_TEMPS instruction.");
|
2024-01-17 12:28:23 -08:00
|
|
|
if (instruction->declaration.count > ctx->program->temp_count)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DCL_TEMPS,
|
|
|
|
"Invalid DCL_TEMPS count %u, expected at most %u.",
|
|
|
|
instruction->declaration.count, ctx->program->temp_count);
|
2023-11-07 14:21:11 -08:00
|
|
|
ctx->dcl_temps_found = true;
|
2023-10-24 18:08:00 -07:00
|
|
|
break;
|
|
|
|
|
2023-11-06 04:47:34 -08:00
|
|
|
case VKD3DSIH_IF:
|
2024-01-12 00:28:14 -08:00
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_STRUCTURED);
|
2023-11-06 04:47:34 -08:00
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 1);
|
|
|
|
if (!vkd3d_array_reserve((void **)&ctx->blocks, &ctx->blocks_capacity, ctx->depth + 1, sizeof(*ctx->blocks)))
|
|
|
|
return;
|
2024-04-24 11:34:42 -07:00
|
|
|
ctx->blocks[ctx->depth++] = instruction->opcode;
|
2023-11-06 04:47:34 -08:00
|
|
|
break;
|
2023-11-16 13:32:13 -08:00
|
|
|
|
|
|
|
case VKD3DSIH_IFC:
|
2024-01-12 00:28:14 -08:00
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_STRUCTURED);
|
2023-11-16 13:32:13 -08:00
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 2);
|
|
|
|
if (!vkd3d_array_reserve((void **)&ctx->blocks, &ctx->blocks_capacity, ctx->depth + 1, sizeof(*ctx->blocks)))
|
|
|
|
return;
|
|
|
|
ctx->blocks[ctx->depth++] = VKD3DSIH_IF;
|
|
|
|
break;
|
2023-11-06 04:47:34 -08:00
|
|
|
|
|
|
|
case VKD3DSIH_ELSE:
|
2024-01-12 00:28:14 -08:00
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_STRUCTURED);
|
2023-11-06 04:47:34 -08:00
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 0);
|
|
|
|
if (ctx->depth == 0 || ctx->blocks[ctx->depth - 1] != VKD3DSIH_IF)
|
2024-01-12 00:28:14 -08:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW, "ELSE instruction doesn't terminate IF block.");
|
2023-11-06 04:47:34 -08:00
|
|
|
else
|
2024-04-24 11:34:42 -07:00
|
|
|
ctx->blocks[ctx->depth - 1] = instruction->opcode;
|
2023-11-06 04:47:34 -08:00
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_ENDIF:
|
2024-01-12 00:28:14 -08:00
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_STRUCTURED);
|
2023-11-06 04:47:34 -08:00
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 0);
|
|
|
|
if (ctx->depth == 0 || (ctx->blocks[ctx->depth - 1] != VKD3DSIH_IF && ctx->blocks[ctx->depth - 1] != VKD3DSIH_ELSE))
|
2024-01-12 00:28:14 -08:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW, "ENDIF instruction doesn't terminate IF/ELSE block.");
|
2023-11-06 04:47:34 -08:00
|
|
|
else
|
|
|
|
--ctx->depth;
|
|
|
|
break;
|
|
|
|
|
2023-11-06 04:54:07 -08:00
|
|
|
case VKD3DSIH_LOOP:
|
2024-01-12 00:28:14 -08:00
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_STRUCTURED);
|
2023-11-06 04:54:07 -08:00
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
2024-01-16 07:40:30 -08:00
|
|
|
vsir_validate_src_count(ctx, instruction, version->major <= 3 ? 2 : 0);
|
2023-11-06 04:54:07 -08:00
|
|
|
if (!vkd3d_array_reserve((void **)&ctx->blocks, &ctx->blocks_capacity, ctx->depth + 1, sizeof(*ctx->blocks)))
|
|
|
|
return;
|
2024-04-24 11:34:42 -07:00
|
|
|
ctx->blocks[ctx->depth++] = instruction->opcode;
|
2023-11-06 04:54:07 -08:00
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_ENDLOOP:
|
2024-01-12 00:28:14 -08:00
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_STRUCTURED);
|
2023-11-06 04:54:07 -08:00
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 0);
|
|
|
|
if (ctx->depth == 0 || ctx->blocks[ctx->depth - 1] != VKD3DSIH_LOOP)
|
2024-01-12 00:28:14 -08:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW, "ENDLOOP instruction doesn't terminate LOOP block.");
|
2023-11-06 04:54:07 -08:00
|
|
|
else
|
|
|
|
--ctx->depth;
|
|
|
|
break;
|
|
|
|
|
2023-11-06 04:56:38 -08:00
|
|
|
case VKD3DSIH_REP:
|
2024-01-12 00:28:14 -08:00
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_STRUCTURED);
|
2023-11-06 04:56:38 -08:00
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 1);
|
|
|
|
if (!vkd3d_array_reserve((void **)&ctx->blocks, &ctx->blocks_capacity, ctx->depth + 1, sizeof(*ctx->blocks)))
|
|
|
|
return;
|
2024-04-24 11:34:42 -07:00
|
|
|
ctx->blocks[ctx->depth++] = instruction->opcode;
|
2023-11-06 04:56:38 -08:00
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_ENDREP:
|
2024-01-12 00:28:14 -08:00
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_STRUCTURED);
|
2023-11-06 04:56:38 -08:00
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 0);
|
|
|
|
if (ctx->depth == 0 || ctx->blocks[ctx->depth - 1] != VKD3DSIH_REP)
|
2024-01-12 00:28:14 -08:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW, "ENDREP instruction doesn't terminate REP block.");
|
2023-11-06 04:56:38 -08:00
|
|
|
else
|
|
|
|
--ctx->depth;
|
|
|
|
break;
|
|
|
|
|
2023-11-06 04:57:25 -08:00
|
|
|
case VKD3DSIH_SWITCH:
|
2024-01-12 00:28:14 -08:00
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_STRUCTURED);
|
2023-11-06 04:57:25 -08:00
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 1);
|
|
|
|
if (!vkd3d_array_reserve((void **)&ctx->blocks, &ctx->blocks_capacity, ctx->depth + 1, sizeof(*ctx->blocks)))
|
|
|
|
return;
|
2024-04-24 11:34:42 -07:00
|
|
|
ctx->blocks[ctx->depth++] = instruction->opcode;
|
2023-11-06 04:57:25 -08:00
|
|
|
break;
|
|
|
|
|
|
|
|
case VKD3DSIH_ENDSWITCH:
|
2024-01-12 00:28:14 -08:00
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_STRUCTURED);
|
2023-11-06 04:57:25 -08:00
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 0);
|
|
|
|
if (ctx->depth == 0 || ctx->blocks[ctx->depth - 1] != VKD3DSIH_SWITCH)
|
2024-01-12 00:28:14 -08:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW, "ENDSWITCH instruction doesn't terminate SWITCH block.");
|
2023-11-06 04:57:25 -08:00
|
|
|
else
|
|
|
|
--ctx->depth;
|
|
|
|
break;
|
|
|
|
|
2024-01-12 02:10:22 -08:00
|
|
|
case VKD3DSIH_RET:
|
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 0);
|
|
|
|
break;
|
|
|
|
|
2023-12-22 06:58:55 -08:00
|
|
|
case VKD3DSIH_LABEL:
|
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_BLOCKS);
|
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
vsir_validate_src_count(ctx, instruction, 1);
|
|
|
|
if (instruction->src_count >= 1 && !vsir_register_is_label(&instruction->src[0].reg))
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE,
|
|
|
|
"Invalid register of type %#x in a LABEL instruction, expected LABEL.",
|
|
|
|
instruction->src[0].reg.type);
|
|
|
|
break;
|
|
|
|
|
2023-12-22 08:11:41 -08:00
|
|
|
case VKD3DSIH_BRANCH:
|
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_BLOCKS);
|
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
if (!vsir_validate_src_min_count(ctx, instruction, 1))
|
|
|
|
break;
|
|
|
|
if (vsir_register_is_label(&instruction->src[0].reg))
|
|
|
|
{
|
|
|
|
/* Unconditional branch: parameters are jump label,
|
|
|
|
* optional merge label, optional continue label. */
|
|
|
|
vsir_validate_src_max_count(ctx, instruction, 3);
|
|
|
|
|
|
|
|
for (i = 0; i < instruction->src_count; ++i)
|
|
|
|
{
|
|
|
|
if (!vsir_register_is_label(&instruction->src[i].reg))
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE,
|
|
|
|
"Invalid register of type %#x in unconditional BRANCH instruction, expected LABEL.",
|
|
|
|
instruction->src[i].reg.type);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
/* Conditional branch: parameters are condition, true
|
|
|
|
* jump label, false jump label, optional merge label,
|
|
|
|
* optional continue label. */
|
|
|
|
vsir_validate_src_min_count(ctx, instruction, 3);
|
|
|
|
vsir_validate_src_max_count(ctx, instruction, 5);
|
|
|
|
|
|
|
|
for (i = 1; i < instruction->src_count; ++i)
|
|
|
|
{
|
|
|
|
if (!vsir_register_is_label(&instruction->src[i].reg))
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE,
|
|
|
|
"Invalid register of type %#x in conditional BRANCH instruction, expected LABEL.",
|
|
|
|
instruction->src[i].reg.type);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
2024-01-12 01:50:30 -08:00
|
|
|
case VKD3DSIH_SWITCH_MONOLITHIC:
|
|
|
|
{
|
|
|
|
unsigned int case_count;
|
|
|
|
|
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_BLOCKS);
|
|
|
|
vsir_validate_dst_count(ctx, instruction, 0);
|
|
|
|
/* Parameters are source, default label, merge label and
|
|
|
|
* then pairs of constant value and case label. */
|
|
|
|
if (!vsir_validate_src_min_count(ctx, instruction, 3))
|
|
|
|
break;
|
|
|
|
if (instruction->src_count % 2 != 1)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_SOURCE_COUNT,
|
|
|
|
"Invalid source count %u for a monolithic SWITCH instruction, it must be an odd number.",
|
|
|
|
instruction->src_count);
|
|
|
|
|
|
|
|
if (!vsir_register_is_label(&instruction->src[1].reg))
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE,
|
|
|
|
"Invalid default label register of type %#x in monolithic SWITCH instruction, expected LABEL.",
|
|
|
|
instruction->src[1].reg.type);
|
|
|
|
|
|
|
|
if (!vsir_register_is_label(&instruction->src[2].reg))
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE,
|
|
|
|
"Invalid merge label register of type %#x in monolithic SWITCH instruction, expected LABEL.",
|
|
|
|
instruction->src[2].reg.type);
|
|
|
|
|
|
|
|
case_count = (instruction->src_count - 3) / 2;
|
|
|
|
|
|
|
|
for (i = 0; i < case_count; ++i)
|
|
|
|
{
|
|
|
|
unsigned int value_idx = 3 + 2 * i;
|
|
|
|
unsigned int label_idx = 3 + 2 * i + 1;
|
|
|
|
|
|
|
|
if (!register_is_constant(&instruction->src[value_idx].reg))
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE,
|
|
|
|
"Invalid value register for case %zu of type %#x in monolithic SWITCH instruction, "
|
|
|
|
"expected IMMCONST or IMMCONST64.", i, instruction->src[value_idx].reg.type);
|
|
|
|
|
|
|
|
if (!vsir_register_is_label(&instruction->src[label_idx].reg))
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE,
|
|
|
|
"Invalid label register for case %zu of type %#x in monolithic SWITCH instruction, "
|
|
|
|
"expected LABEL.", i, instruction->src[value_idx].reg.type);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2024-01-15 03:43:26 -08:00
|
|
|
case VKD3DSIH_PHI:
|
|
|
|
{
|
|
|
|
unsigned int incoming_count;
|
|
|
|
|
|
|
|
vsir_validate_cf_type(ctx, instruction, CF_TYPE_BLOCKS);
|
|
|
|
vsir_validate_dst_count(ctx, instruction, 1);
|
|
|
|
vsir_validate_src_min_count(ctx, instruction, 2);
|
|
|
|
if (instruction->src_count % 2 != 0)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_SOURCE_COUNT,
|
|
|
|
"Invalid source count %u for a PHI instruction, it must be an even number.",
|
|
|
|
instruction->src_count);
|
|
|
|
incoming_count = instruction->src_count / 2;
|
|
|
|
|
|
|
|
if (!register_is_ssa(&instruction->dst[0].reg))
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE,
|
|
|
|
"Invalid destination of type %#x in PHI instruction, expected SSA.",
|
|
|
|
instruction->dst[0].reg.type);
|
|
|
|
|
|
|
|
if (instruction->dst[0].reg.dimension != VSIR_DIMENSION_SCALAR)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DIMENSION,
|
|
|
|
"Invalid destination dimension %#x in PHI instruction, expected scalar.",
|
|
|
|
instruction->dst[0].reg.dimension);
|
|
|
|
|
|
|
|
if (instruction->dst[0].modifiers != VKD3DSPDM_NONE)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_MODIFIERS,
|
|
|
|
"Invalid modifiers %#x for the destination of a PHI instruction, expected none.",
|
|
|
|
instruction->dst[0].modifiers);
|
|
|
|
|
|
|
|
if (instruction->dst[0].shift != 0)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_SHIFT,
|
|
|
|
"Invalid shift %#x for the destination of a PHI instruction, expected none.",
|
|
|
|
instruction->dst[0].shift);
|
|
|
|
|
|
|
|
for (i = 0; i < incoming_count; ++i)
|
|
|
|
{
|
|
|
|
unsigned int value_idx = 2 * i;
|
|
|
|
unsigned int label_idx = 2 * i + 1;
|
|
|
|
|
2024-04-23 07:23:12 -07:00
|
|
|
if (!register_is_constant_or_undef(&instruction->src[value_idx].reg)
|
|
|
|
&& !register_is_ssa(&instruction->src[value_idx].reg))
|
2024-01-15 03:43:26 -08:00
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE,
|
|
|
|
"Invalid value register for incoming %zu of type %#x in PHI instruction, "
|
|
|
|
"expected SSA, IMMCONST or IMMCONST64.", i, instruction->src[value_idx].reg.type);
|
|
|
|
|
|
|
|
if (instruction->src[value_idx].reg.dimension != VSIR_DIMENSION_SCALAR)
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_DIMENSION,
|
|
|
|
"Invalid value dimension %#x for incoming %zu in PHI instruction, expected scalar.",
|
|
|
|
instruction->src[value_idx].reg.dimension, i);
|
|
|
|
|
|
|
|
if (!vsir_register_is_label(&instruction->src[label_idx].reg))
|
|
|
|
validator_error(ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_REGISTER_TYPE,
|
|
|
|
"Invalid label register for case %zu of type %#x in PHI instruction, "
|
|
|
|
"expected LABEL.", i, instruction->src[value_idx].reg.type);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2023-10-24 18:08:00 -07:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2023-08-29 12:35:01 -07:00
|
|
|
}
|
|
|
|
|
2024-03-13 06:37:45 -07:00
|
|
|
enum vkd3d_result vsir_program_validate(struct vsir_program *program, uint64_t config_flags,
|
|
|
|
const char *source_name, struct vkd3d_shader_message_context *message_context)
|
2023-08-29 12:20:35 -07:00
|
|
|
{
|
2023-11-06 14:11:03 -08:00
|
|
|
struct validation_context ctx =
|
|
|
|
{
|
2024-03-13 06:37:45 -07:00
|
|
|
.message_context = message_context,
|
|
|
|
.program = program,
|
|
|
|
.null_location = {.source_name = source_name},
|
|
|
|
.status = VKD3D_OK,
|
2023-11-06 14:11:03 -08:00
|
|
|
.phase = VKD3DSIH_INVALID,
|
|
|
|
};
|
2024-01-10 02:58:44 -08:00
|
|
|
unsigned int i;
|
2023-08-29 12:35:01 -07:00
|
|
|
|
2024-03-13 06:37:45 -07:00
|
|
|
if (!(config_flags & VKD3D_SHADER_CONFIG_FLAG_FORCE_VALIDATION))
|
2024-01-10 02:27:03 -08:00
|
|
|
return VKD3D_OK;
|
|
|
|
|
2024-01-17 12:28:23 -08:00
|
|
|
if (!(ctx.temps = vkd3d_calloc(ctx.program->temp_count, sizeof(*ctx.temps))))
|
2024-01-10 02:27:03 -08:00
|
|
|
goto fail;
|
2023-08-29 12:35:01 -07:00
|
|
|
|
2024-01-17 12:14:41 -08:00
|
|
|
if (!(ctx.ssas = vkd3d_calloc(ctx.program->ssa_count, sizeof(*ctx.ssas))))
|
2024-01-10 02:31:40 -08:00
|
|
|
goto fail;
|
|
|
|
|
2024-03-13 06:37:45 -07:00
|
|
|
for (ctx.instruction_idx = 0; ctx.instruction_idx < program->instructions.count; ++ctx.instruction_idx)
|
2023-08-29 12:35:01 -07:00
|
|
|
vsir_validate_instruction(&ctx);
|
2023-11-06 04:47:34 -08:00
|
|
|
|
2024-01-10 02:58:44 -08:00
|
|
|
ctx.invalid_instruction_idx = true;
|
|
|
|
|
2023-11-06 04:47:34 -08:00
|
|
|
if (ctx.depth != 0)
|
2024-01-12 00:28:14 -08:00
|
|
|
validator_error(&ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW, "%zu nested blocks were not closed.", ctx.depth);
|
2023-11-06 04:47:34 -08:00
|
|
|
|
2024-01-12 02:20:26 -08:00
|
|
|
if (ctx.inside_block)
|
|
|
|
validator_error(&ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_CONTROL_FLOW, "Last block was not closed.");
|
|
|
|
|
2024-01-17 12:14:41 -08:00
|
|
|
for (i = 0; i < ctx.program->ssa_count; ++i)
|
2024-01-10 02:58:44 -08:00
|
|
|
{
|
|
|
|
struct validation_context_ssa_data *data = &ctx.ssas[i];
|
|
|
|
|
|
|
|
if ((data->write_mask | data->read_mask) != data->write_mask)
|
|
|
|
validator_error(&ctx, VKD3D_SHADER_ERROR_VSIR_INVALID_SSA_USAGE,
|
|
|
|
"SSA register %u has invalid read mask %#x, which is not a subset of the write mask %#x "
|
|
|
|
"at the point of definition.", i, data->read_mask, data->write_mask);
|
|
|
|
}
|
|
|
|
|
2024-01-10 02:14:39 -08:00
|
|
|
vkd3d_free(ctx.blocks);
|
2024-01-10 02:27:03 -08:00
|
|
|
vkd3d_free(ctx.temps);
|
2024-01-10 02:31:40 -08:00
|
|
|
vkd3d_free(ctx.ssas);
|
2024-01-10 02:27:03 -08:00
|
|
|
|
2024-03-13 06:37:45 -07:00
|
|
|
return ctx.status;
|
2024-01-10 02:27:03 -08:00
|
|
|
|
|
|
|
fail:
|
|
|
|
vkd3d_free(ctx.blocks);
|
|
|
|
vkd3d_free(ctx.temps);
|
2024-01-10 02:31:40 -08:00
|
|
|
vkd3d_free(ctx.ssas);
|
2024-01-10 02:27:03 -08:00
|
|
|
|
|
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
2023-08-29 12:20:35 -07:00
|
|
|
}
|
2024-04-02 01:01:13 -07:00
|
|
|
|
|
|
|
enum vkd3d_result vsir_program_normalise(struct vsir_program *program, uint64_t config_flags,
|
|
|
|
const struct vkd3d_shader_compile_info *compile_info, struct vkd3d_shader_message_context *message_context)
|
|
|
|
{
|
|
|
|
enum vkd3d_result result = VKD3D_OK;
|
|
|
|
|
2024-04-01 21:24:03 -07:00
|
|
|
if ((result = vsir_program_lower_instructions(program, message_context)) < 0)
|
2024-04-02 01:01:13 -07:00
|
|
|
return result;
|
|
|
|
|
|
|
|
if (program->shader_version.major >= 6)
|
|
|
|
{
|
2024-04-02 19:20:09 -07:00
|
|
|
if ((result = vsir_program_materialise_phi_ssas_to_temps(program)) < 0)
|
2024-04-02 01:01:13 -07:00
|
|
|
return result;
|
|
|
|
|
2024-04-02 19:20:09 -07:00
|
|
|
if ((result = lower_switch_to_if_ladder(program)) < 0)
|
2024-04-02 01:01:13 -07:00
|
|
|
return result;
|
|
|
|
|
2024-04-10 06:21:52 -07:00
|
|
|
if ((result = vsir_program_structurize(program, message_context)) < 0)
|
2024-04-02 01:01:13 -07:00
|
|
|
return result;
|
2024-04-01 19:55:05 -07:00
|
|
|
|
|
|
|
if ((result = vsir_program_flatten_control_flow_constructs(program, message_context)) < 0)
|
|
|
|
return result;
|
|
|
|
|
2024-04-10 06:26:50 -07:00
|
|
|
if ((result = vsir_program_materialize_undominated_ssas_to_temps(program, message_context)) < 0)
|
2024-04-01 19:55:05 -07:00
|
|
|
return result;
|
2024-04-02 01:01:13 -07:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
if (program->shader_version.type != VKD3D_SHADER_TYPE_PIXEL)
|
|
|
|
{
|
|
|
|
if ((result = vsir_program_remap_output_signature(program, compile_info, message_context)) < 0)
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (program->shader_version.type == VKD3D_SHADER_TYPE_HULL)
|
|
|
|
{
|
|
|
|
if ((result = instruction_array_flatten_hull_shader_phases(&program->instructions)) < 0)
|
|
|
|
return result;
|
|
|
|
|
|
|
|
if ((result = instruction_array_normalise_hull_shader_control_point_io(&program->instructions,
|
|
|
|
&program->input_signature)) < 0)
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((result = vsir_program_normalise_io_registers(program)) < 0)
|
|
|
|
return result;
|
|
|
|
|
|
|
|
if ((result = instruction_array_normalise_flat_constants(program)) < 0)
|
|
|
|
return result;
|
|
|
|
|
|
|
|
remove_dead_code(program);
|
|
|
|
|
|
|
|
if ((result = vsir_program_normalise_combined_samplers(program, message_context)) < 0)
|
|
|
|
return result;
|
|
|
|
|
2024-02-24 06:13:06 -08:00
|
|
|
if (compile_info->target_type != VKD3D_SHADER_TARGET_GLSL
|
|
|
|
&& (result = vsir_program_flatten_control_flow_constructs(program, message_context)) < 0)
|
2024-04-01 19:55:05 -07:00
|
|
|
return result;
|
|
|
|
}
|
2024-04-02 01:01:13 -07:00
|
|
|
|
|
|
|
if (TRACE_ON())
|
|
|
|
vkd3d_shader_trace(program);
|
|
|
|
|
|
|
|
if ((result = vsir_program_validate(program, config_flags,
|
|
|
|
compile_info->source_name, message_context)) < 0)
|
|
|
|
return result;
|
|
|
|
|
|
|
|
return result;
|
|
|
|
}
|