mirror of
https://gitlab.winehq.org/wine/vkd3d.git
synced 2025-01-28 13:05:02 -08:00
vkd3d-shader/hlsl: Store temp declarations in the vsir program.
Move the temp allocation back to hlsl_codegen.c. Note that the DCL_TEMPS instructions wouldn't be necessary if we had the capacity to store the temp_count for both the main program and the patch constant program (or more generally speaking, a temp_count for all phases). The plan is to eventually also move the HS_CONTROL_POINT and HS_FORK_PHASE markers to the vsir_program, making it able to contain both functions.
This commit is contained in:
parent
158bf794e6
commit
23be6ed0dd
Notes:
Henri Verbeet
2024-10-24 21:02:04 +02:00
Approved-by: Elizabeth Figura (@zfigura) Approved-by: Henri Verbeet (@hverbeet) Merge-Request: https://gitlab.winehq.org/wine/vkd3d/-/merge_requests/1210
@ -7542,10 +7542,101 @@ static void sm1_generate_vsir(struct hlsl_ctx *ctx, struct hlsl_ir_function_decl
|
|||||||
sm1_generate_vsir_block(ctx, &entry_func->body, program);
|
sm1_generate_vsir_block(ctx, &entry_func->body, program);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void add_last_vsir_instr_to_block(struct hlsl_ctx *ctx, struct vsir_program *program, struct hlsl_block *block)
|
||||||
|
{
|
||||||
|
struct vkd3d_shader_location *loc;
|
||||||
|
struct hlsl_ir_node *vsir_instr;
|
||||||
|
|
||||||
|
loc = &program->instructions.elements[program->instructions.count - 1].location;
|
||||||
|
|
||||||
|
if (!(vsir_instr = hlsl_new_vsir_instruction_ref(ctx, program->instructions.count - 1, NULL, NULL, loc)))
|
||||||
|
{
|
||||||
|
ctx->result = VKD3D_ERROR_OUT_OF_MEMORY;
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
hlsl_block_add_instr(block, vsir_instr);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void sm4_generate_vsir_instr_dcl_temps(struct hlsl_ctx *ctx, struct vsir_program *program,
|
||||||
|
uint32_t temp_count, struct hlsl_block *block, const struct vkd3d_shader_location *loc)
|
||||||
|
{
|
||||||
|
struct vkd3d_shader_instruction *ins;
|
||||||
|
|
||||||
|
if (!(ins = generate_vsir_add_program_instruction(ctx, program, loc, VKD3DSIH_DCL_TEMPS, 0, 0)))
|
||||||
|
return;
|
||||||
|
|
||||||
|
ins->declaration.count = temp_count;
|
||||||
|
|
||||||
|
add_last_vsir_instr_to_block(ctx, program, block);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void sm4_generate_vsir_instr_dcl_indexable_temp(struct hlsl_ctx *ctx,
|
||||||
|
struct vsir_program *program, struct hlsl_block *block, uint32_t idx,
|
||||||
|
uint32_t size, uint32_t comp_count, const struct vkd3d_shader_location *loc)
|
||||||
|
{
|
||||||
|
struct vkd3d_shader_instruction *ins;
|
||||||
|
|
||||||
|
if (!(ins = generate_vsir_add_program_instruction(ctx, program, loc, VKD3DSIH_DCL_INDEXABLE_TEMP, 0, 0)))
|
||||||
|
return;
|
||||||
|
|
||||||
|
ins->declaration.indexable_temp.register_idx = idx;
|
||||||
|
ins->declaration.indexable_temp.register_size = size;
|
||||||
|
ins->declaration.indexable_temp.alignment = 0;
|
||||||
|
ins->declaration.indexable_temp.data_type = VKD3D_DATA_FLOAT;
|
||||||
|
ins->declaration.indexable_temp.component_count = comp_count;
|
||||||
|
ins->declaration.indexable_temp.has_function_scope = false;
|
||||||
|
|
||||||
|
add_last_vsir_instr_to_block(ctx, program, block);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void sm4_generate_vsir_add_function(struct hlsl_ctx *ctx,
|
||||||
|
struct hlsl_ir_function_decl *func, uint64_t config_flags, struct vsir_program *program)
|
||||||
|
{
|
||||||
|
struct hlsl_block block = {0};
|
||||||
|
struct hlsl_scope *scope;
|
||||||
|
struct hlsl_ir_var *var;
|
||||||
|
uint32_t temp_count;
|
||||||
|
|
||||||
|
compute_liveness(ctx, func);
|
||||||
|
mark_indexable_vars(ctx, func);
|
||||||
|
temp_count = allocate_temp_registers(ctx, func);
|
||||||
|
if (ctx->result)
|
||||||
|
return;
|
||||||
|
program->temp_count = max(program->temp_count, temp_count);
|
||||||
|
|
||||||
|
hlsl_block_init(&block);
|
||||||
|
|
||||||
|
if (temp_count)
|
||||||
|
sm4_generate_vsir_instr_dcl_temps(ctx, program, temp_count, &block, &func->loc);
|
||||||
|
|
||||||
|
LIST_FOR_EACH_ENTRY(scope, &ctx->scopes, struct hlsl_scope, entry)
|
||||||
|
{
|
||||||
|
LIST_FOR_EACH_ENTRY(var, &scope->vars, struct hlsl_ir_var, scope_entry)
|
||||||
|
{
|
||||||
|
if (var->is_uniform || var->is_input_semantic || var->is_output_semantic)
|
||||||
|
continue;
|
||||||
|
if (!var->regs[HLSL_REGSET_NUMERIC].allocated)
|
||||||
|
continue;
|
||||||
|
|
||||||
|
if (var->indexable)
|
||||||
|
{
|
||||||
|
unsigned int id = var->regs[HLSL_REGSET_NUMERIC].id;
|
||||||
|
unsigned int size = align(var->data_type->reg_size[HLSL_REGSET_NUMERIC], 4) / 4;
|
||||||
|
|
||||||
|
sm4_generate_vsir_instr_dcl_indexable_temp(ctx, program, &block, id, size, 4, &var->loc);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
list_move_head(&func->body.instrs, &block.instrs);
|
||||||
|
|
||||||
|
hlsl_block_cleanup(&block);
|
||||||
|
}
|
||||||
|
|
||||||
/* OBJECTIVE: Translate all the information from ctx and entry_func to the
|
/* OBJECTIVE: Translate all the information from ctx and entry_func to the
|
||||||
* vsir_program, so it can be used as input to tpf_compile() without relying
|
* vsir_program, so it can be used as input to tpf_compile() without relying
|
||||||
* on ctx and entry_func. */
|
* on ctx and entry_func. */
|
||||||
static void sm4_generate_vsir(struct hlsl_ctx *ctx, struct hlsl_ir_function_decl *entry_func,
|
static void sm4_generate_vsir(struct hlsl_ctx *ctx, struct hlsl_ir_function_decl *func,
|
||||||
uint64_t config_flags, struct vsir_program *program)
|
uint64_t config_flags, struct vsir_program *program)
|
||||||
{
|
{
|
||||||
struct vkd3d_shader_version version = {0};
|
struct vkd3d_shader_version version = {0};
|
||||||
@ -7560,7 +7651,7 @@ static void sm4_generate_vsir(struct hlsl_ctx *ctx, struct hlsl_ir_function_decl
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
generate_vsir_signature(ctx, program, entry_func);
|
generate_vsir_signature(ctx, program, func);
|
||||||
if (version.type == VKD3D_SHADER_TYPE_HULL)
|
if (version.type == VKD3D_SHADER_TYPE_HULL)
|
||||||
generate_vsir_signature(ctx, program, ctx->patch_constant_func);
|
generate_vsir_signature(ctx, program, ctx->patch_constant_func);
|
||||||
|
|
||||||
@ -7570,6 +7661,10 @@ static void sm4_generate_vsir(struct hlsl_ctx *ctx, struct hlsl_ir_function_decl
|
|||||||
program->thread_group_size.y = ctx->thread_count[1];
|
program->thread_group_size.y = ctx->thread_count[1];
|
||||||
program->thread_group_size.z = ctx->thread_count[2];
|
program->thread_group_size.z = ctx->thread_count[2];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
sm4_generate_vsir_add_function(ctx, func, config_flags, program);
|
||||||
|
if (version.type == VKD3D_SHADER_TYPE_HULL)
|
||||||
|
sm4_generate_vsir_add_function(ctx, ctx->patch_constant_func, config_flags, program);
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct hlsl_ir_jump *loop_unrolling_find_jump(struct hlsl_block *block, struct hlsl_ir_node *stop_point,
|
static struct hlsl_ir_jump *loop_unrolling_find_jump(struct hlsl_block *block, struct hlsl_ir_node *stop_point,
|
||||||
|
@ -4935,27 +4935,26 @@ static void tpf_write_dcl_semantic(const struct tpf_compiler *tpf,
|
|||||||
write_sm4_instruction(tpf, &instr);
|
write_sm4_instruction(tpf, &instr);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void write_sm4_dcl_temps(const struct tpf_compiler *tpf, uint32_t temp_count)
|
static void tpf_dcl_temps(const struct tpf_compiler *tpf, unsigned int count)
|
||||||
{
|
{
|
||||||
struct sm4_instruction instr =
|
struct sm4_instruction instr =
|
||||||
{
|
{
|
||||||
.opcode = VKD3D_SM4_OP_DCL_TEMPS,
|
.opcode = VKD3D_SM4_OP_DCL_TEMPS,
|
||||||
|
|
||||||
.idx = {temp_count},
|
.idx = {count},
|
||||||
.idx_count = 1,
|
.idx_count = 1,
|
||||||
};
|
};
|
||||||
|
|
||||||
write_sm4_instruction(tpf, &instr);
|
write_sm4_instruction(tpf, &instr);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void write_sm4_dcl_indexable_temp(const struct tpf_compiler *tpf, uint32_t idx,
|
static void tpf_dcl_indexable_temp(const struct tpf_compiler *tpf, const struct vkd3d_shader_indexable_temp *temp)
|
||||||
uint32_t size, uint32_t comp_count)
|
|
||||||
{
|
{
|
||||||
struct sm4_instruction instr =
|
struct sm4_instruction instr =
|
||||||
{
|
{
|
||||||
.opcode = VKD3D_SM4_OP_DCL_INDEXABLE_TEMP,
|
.opcode = VKD3D_SM4_OP_DCL_INDEXABLE_TEMP,
|
||||||
|
|
||||||
.idx = {idx, size, comp_count},
|
.idx = {temp->register_idx, temp->register_size, temp->component_count},
|
||||||
.idx_count = 3,
|
.idx_count = 3,
|
||||||
};
|
};
|
||||||
|
|
||||||
@ -6447,9 +6446,28 @@ static void write_sm4_swizzle(const struct tpf_compiler *tpf, const struct hlsl_
|
|||||||
write_sm4_instruction(tpf, &instr);
|
write_sm4_instruction(tpf, &instr);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void tpf_handle_instruction(const struct tpf_compiler *tpf, const struct vkd3d_shader_instruction *ins)
|
||||||
|
{
|
||||||
|
switch (ins->opcode)
|
||||||
|
{
|
||||||
|
case VKD3DSIH_DCL_TEMPS:
|
||||||
|
tpf_dcl_temps(tpf, ins->declaration.count);
|
||||||
|
break;
|
||||||
|
|
||||||
|
case VKD3DSIH_DCL_INDEXABLE_TEMP:
|
||||||
|
tpf_dcl_indexable_temp(tpf, &ins->declaration.indexable_temp);
|
||||||
|
break;
|
||||||
|
|
||||||
|
default:
|
||||||
|
vkd3d_unreachable();
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
static void write_sm4_block(const struct tpf_compiler *tpf, const struct hlsl_block *block)
|
static void write_sm4_block(const struct tpf_compiler *tpf, const struct hlsl_block *block)
|
||||||
{
|
{
|
||||||
const struct hlsl_ir_node *instr;
|
const struct hlsl_ir_node *instr;
|
||||||
|
unsigned int vsir_instr_idx;
|
||||||
|
|
||||||
LIST_FOR_EACH_ENTRY(instr, &block->instrs, struct hlsl_ir_node, entry)
|
LIST_FOR_EACH_ENTRY(instr, &block->instrs, struct hlsl_ir_node, entry)
|
||||||
{
|
{
|
||||||
@ -6515,6 +6533,11 @@ static void write_sm4_block(const struct tpf_compiler *tpf, const struct hlsl_bl
|
|||||||
write_sm4_swizzle(tpf, hlsl_ir_swizzle(instr));
|
write_sm4_swizzle(tpf, hlsl_ir_swizzle(instr));
|
||||||
break;
|
break;
|
||||||
|
|
||||||
|
case HLSL_IR_VSIR_INSTRUCTION_REF:
|
||||||
|
vsir_instr_idx = hlsl_ir_vsir_instruction_ref(instr)->vsir_instr_idx;
|
||||||
|
tpf_handle_instruction(tpf, &tpf->program->instructions.elements[vsir_instr_idx]);
|
||||||
|
break;
|
||||||
|
|
||||||
default:
|
default:
|
||||||
hlsl_fixme(tpf->ctx, &instr->loc, "Instruction type %s.", hlsl_node_type_to_string(instr->type));
|
hlsl_fixme(tpf->ctx, &instr->loc, "Instruction type %s.", hlsl_node_type_to_string(instr->type));
|
||||||
}
|
}
|
||||||
@ -6524,15 +6547,7 @@ static void write_sm4_block(const struct tpf_compiler *tpf, const struct hlsl_bl
|
|||||||
static void tpf_write_shader_function(struct tpf_compiler *tpf, struct hlsl_ir_function_decl *func)
|
static void tpf_write_shader_function(struct tpf_compiler *tpf, struct hlsl_ir_function_decl *func)
|
||||||
{
|
{
|
||||||
struct hlsl_ctx *ctx = tpf->ctx;
|
struct hlsl_ctx *ctx = tpf->ctx;
|
||||||
const struct hlsl_scope *scope;
|
|
||||||
const struct hlsl_ir_var *var;
|
const struct hlsl_ir_var *var;
|
||||||
uint32_t temp_count;
|
|
||||||
|
|
||||||
compute_liveness(ctx, func);
|
|
||||||
mark_indexable_vars(ctx, func);
|
|
||||||
temp_count = allocate_temp_registers(ctx, func);
|
|
||||||
if (ctx->result)
|
|
||||||
return;
|
|
||||||
|
|
||||||
LIST_FOR_EACH_ENTRY(var, &func->extern_vars, struct hlsl_ir_var, extern_entry)
|
LIST_FOR_EACH_ENTRY(var, &func->extern_vars, struct hlsl_ir_var, extern_entry)
|
||||||
{
|
{
|
||||||
@ -6544,28 +6559,6 @@ static void tpf_write_shader_function(struct tpf_compiler *tpf, struct hlsl_ir_f
|
|||||||
if (tpf->program->shader_version.type == VKD3D_SHADER_TYPE_COMPUTE)
|
if (tpf->program->shader_version.type == VKD3D_SHADER_TYPE_COMPUTE)
|
||||||
tpf_dcl_thread_group(tpf, &tpf->program->thread_group_size);
|
tpf_dcl_thread_group(tpf, &tpf->program->thread_group_size);
|
||||||
|
|
||||||
if (temp_count)
|
|
||||||
write_sm4_dcl_temps(tpf, temp_count);
|
|
||||||
|
|
||||||
LIST_FOR_EACH_ENTRY(scope, &ctx->scopes, struct hlsl_scope, entry)
|
|
||||||
{
|
|
||||||
LIST_FOR_EACH_ENTRY(var, &scope->vars, struct hlsl_ir_var, scope_entry)
|
|
||||||
{
|
|
||||||
if (var->is_uniform || var->is_input_semantic || var->is_output_semantic)
|
|
||||||
continue;
|
|
||||||
if (!var->regs[HLSL_REGSET_NUMERIC].allocated)
|
|
||||||
continue;
|
|
||||||
|
|
||||||
if (var->indexable)
|
|
||||||
{
|
|
||||||
unsigned int id = var->regs[HLSL_REGSET_NUMERIC].id;
|
|
||||||
unsigned int size = align(var->data_type->reg_size[HLSL_REGSET_NUMERIC], 4) / 4;
|
|
||||||
|
|
||||||
write_sm4_dcl_indexable_temp(tpf, id, size, 4);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
write_sm4_block(tpf, &func->body);
|
write_sm4_block(tpf, &func->body);
|
||||||
|
|
||||||
write_sm4_ret(tpf);
|
write_sm4_ret(tpf);
|
||||||
|
Loading…
x
Reference in New Issue
Block a user