mirror of
https://gitlab.winehq.org/wine/vkd3d.git
synced 2024-11-21 16:46:41 -08:00
da7c9694f1
DXIL handles are used to represent descriptors. Currently they are translated to registers of the appropriate type (depending on the descriptor type) and three indices. The first two indices are used to represent the descriptor itself (through its signature and array index), and are filled when the handle itself is created. The last index is used with constant buffers to address the data inside the buffer itself, and it goes unused with other descriptor types. As currently implemented, however, registers for descriptors other than constant buffers are still created with three indices, even if the last one is useless and set to -1. In the interest of creating more sensible VSIR code, this is now removed: DXIL handles are created with just two indices; a third one is added when accessing constant buffers, and nothing is changed for other descriptor types.
10606 lines
379 KiB
C
10606 lines
379 KiB
C
/*
|
|
* Copyright 2023 Conor McCarthy for CodeWeavers
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA
|
|
*/
|
|
|
|
#include "vkd3d_shader_private.h"
|
|
|
|
#define VKD3D_SM6_VERSION_MAJOR(version) (((version) >> 4) & 0xf)
|
|
#define VKD3D_SM6_VERSION_MINOR(version) (((version) >> 0) & 0xf)
|
|
/* Two seems to be the maximum but leave some extra room. */
|
|
#define VKD3D_SM6_MAX_METADATA_TABLES 4
|
|
|
|
#define BITCODE_MAGIC VKD3D_MAKE_TAG('B', 'C', 0xc0, 0xde)
|
|
#define DXIL_OP_MAX_OPERANDS 17
|
|
static const uint64_t MAX_ALIGNMENT_EXPONENT = 29;
|
|
static const uint64_t GLOBALVAR_FLAG_IS_CONSTANT = 1;
|
|
static const uint64_t GLOBALVAR_FLAG_EXPLICIT_TYPE = 2;
|
|
static const unsigned int GLOBALVAR_ADDRESS_SPACE_SHIFT = 2;
|
|
static const uint64_t ALLOCA_FLAG_IN_ALLOCA = 0x20;
|
|
static const uint64_t ALLOCA_FLAG_EXPLICIT_TYPE = 0x40;
|
|
static const uint64_t ALLOCA_ALIGNMENT_MASK = 0x1f;
|
|
static const unsigned int SHADER_DESCRIPTOR_TYPE_COUNT = 4;
|
|
static const size_t MAX_IR_INSTRUCTIONS_PER_DXIL_INSTRUCTION = 11;
|
|
|
|
static const unsigned int dx_max_thread_group_size[3] = {1024, 1024, 64};
|
|
|
|
static const unsigned int MAX_GS_INSTANCE_COUNT = 32; /* kMaxGSInstanceCount */
|
|
static const unsigned int MAX_GS_OUTPUT_TOTAL_SCALARS = 1024; /* kMaxGSOutputTotalScalars */
|
|
static const unsigned int MAX_GS_OUTPUT_STREAMS = 4;
|
|
|
|
#define VKD3D_SHADER_SWIZZLE_64_MASK \
|
|
(VKD3D_SHADER_SWIZZLE_MASK << VKD3D_SHADER_SWIZZLE_SHIFT(0) \
|
|
| VKD3D_SHADER_SWIZZLE_MASK << VKD3D_SHADER_SWIZZLE_SHIFT(1))
|
|
|
|
enum bitcode_block_id
|
|
{
|
|
BLOCKINFO_BLOCK = 0,
|
|
MODULE_BLOCK = 8,
|
|
PARAMATTR_BLOCK = 9,
|
|
PARAMATTR_GROUP_BLOCK = 10,
|
|
CONSTANTS_BLOCK = 11,
|
|
FUNCTION_BLOCK = 12,
|
|
VALUE_SYMTAB_BLOCK = 14,
|
|
METADATA_BLOCK = 15,
|
|
METADATA_ATTACHMENT_BLOCK = 16,
|
|
TYPE_BLOCK = 17,
|
|
USELIST_BLOCK = 18,
|
|
};
|
|
|
|
enum bitcode_blockinfo_code
|
|
{
|
|
SETBID = 1,
|
|
BLOCKNAME = 2,
|
|
SETRECORDNAME = 3,
|
|
};
|
|
|
|
enum bitcode_block_abbreviation
|
|
{
|
|
END_BLOCK = 0,
|
|
ENTER_SUBBLOCK = 1,
|
|
DEFINE_ABBREV = 2,
|
|
UNABBREV_RECORD = 3,
|
|
};
|
|
|
|
enum bitcode_abbrev_type
|
|
{
|
|
ABBREV_FIXED = 1,
|
|
ABBREV_VBR = 2,
|
|
ABBREV_ARRAY = 3,
|
|
ABBREV_CHAR = 4,
|
|
ABBREV_BLOB = 5,
|
|
};
|
|
|
|
enum bitcode_address_space
|
|
{
|
|
ADDRESS_SPACE_DEFAULT,
|
|
ADDRESS_SPACE_DEVICEMEM,
|
|
ADDRESS_SPACE_CBUFFER,
|
|
ADDRESS_SPACE_GROUPSHARED,
|
|
};
|
|
|
|
enum bitcode_module_code
|
|
{
|
|
MODULE_CODE_VERSION = 1,
|
|
MODULE_CODE_GLOBALVAR = 7,
|
|
MODULE_CODE_FUNCTION = 8,
|
|
};
|
|
|
|
enum bitcode_constant_code
|
|
{
|
|
CST_CODE_SETTYPE = 1,
|
|
CST_CODE_NULL = 2,
|
|
CST_CODE_UNDEF = 3,
|
|
CST_CODE_INTEGER = 4,
|
|
CST_CODE_FLOAT = 6,
|
|
CST_CODE_STRING = 8,
|
|
CST_CODE_CE_CAST = 11,
|
|
CST_CODE_CE_GEP = 12,
|
|
CST_CODE_CE_INBOUNDS_GEP = 20,
|
|
CST_CODE_DATA = 22,
|
|
};
|
|
|
|
enum bitcode_function_code
|
|
{
|
|
FUNC_CODE_DECLAREBLOCKS = 1,
|
|
FUNC_CODE_INST_BINOP = 2,
|
|
FUNC_CODE_INST_CAST = 3,
|
|
FUNC_CODE_INST_RET = 10,
|
|
FUNC_CODE_INST_BR = 11,
|
|
FUNC_CODE_INST_SWITCH = 12,
|
|
FUNC_CODE_INST_PHI = 16,
|
|
FUNC_CODE_INST_ALLOCA = 19,
|
|
FUNC_CODE_INST_LOAD = 20,
|
|
FUNC_CODE_INST_EXTRACTVAL = 26,
|
|
FUNC_CODE_INST_CMP2 = 28,
|
|
FUNC_CODE_INST_VSELECT = 29,
|
|
FUNC_CODE_INST_CALL = 34,
|
|
FUNC_CODE_INST_ATOMICRMW = 38,
|
|
FUNC_CODE_INST_LOADATOMIC = 41,
|
|
FUNC_CODE_INST_GEP = 43,
|
|
FUNC_CODE_INST_STORE = 44,
|
|
FUNC_CODE_INST_STOREATOMIC = 45,
|
|
FUNC_CODE_INST_CMPXCHG = 46,
|
|
};
|
|
|
|
enum bitcode_metadata_code
|
|
{
|
|
METADATA_STRING = 1,
|
|
METADATA_VALUE = 2,
|
|
METADATA_NODE = 3,
|
|
METADATA_NAME = 4,
|
|
METADATA_DISTINCT_NODE = 5,
|
|
METADATA_KIND = 6,
|
|
METADATA_LOCATION = 7,
|
|
METADATA_NAMED_NODE = 10,
|
|
METADATA_ATTACHMENT = 11,
|
|
};
|
|
|
|
enum bitcode_type_code
|
|
{
|
|
TYPE_CODE_NUMENTRY = 1,
|
|
TYPE_CODE_VOID = 2,
|
|
TYPE_CODE_FLOAT = 3,
|
|
TYPE_CODE_DOUBLE = 4,
|
|
TYPE_CODE_LABEL = 5,
|
|
TYPE_CODE_INTEGER = 7,
|
|
TYPE_CODE_POINTER = 8,
|
|
TYPE_CODE_HALF = 10,
|
|
TYPE_CODE_ARRAY = 11,
|
|
TYPE_CODE_VECTOR = 12,
|
|
TYPE_CODE_METADATA = 16,
|
|
TYPE_CODE_STRUCT_ANON = 18,
|
|
TYPE_CODE_STRUCT_NAME = 19,
|
|
TYPE_CODE_STRUCT_NAMED = 20,
|
|
TYPE_CODE_FUNCTION = 21,
|
|
};
|
|
|
|
enum bitcode_value_symtab_code
|
|
{
|
|
VST_CODE_ENTRY = 1,
|
|
VST_CODE_BBENTRY = 2,
|
|
};
|
|
|
|
enum bitcode_linkage
|
|
{
|
|
LINKAGE_EXTERNAL = 0,
|
|
LINKAGE_APPENDING = 2,
|
|
LINKAGE_INTERNAL = 3,
|
|
};
|
|
|
|
enum dxil_resource_kind
|
|
{
|
|
RESOURCE_KIND_INVALID = 0,
|
|
RESOURCE_KIND_TEXTURE1D = 1,
|
|
RESOURCE_KIND_TEXTURE2D = 2,
|
|
RESOURCE_KIND_TEXTURE2DMS = 3,
|
|
RESOURCE_KIND_TEXTURE3D = 4,
|
|
RESOURCE_KIND_TEXTURECUBE = 5,
|
|
RESOURCE_KIND_TEXTURE1DARRAY = 6,
|
|
RESOURCE_KIND_TEXTURE2DARRAY = 7,
|
|
RESOURCE_KIND_TEXTURE2DMSARRAY = 8,
|
|
RESOURCE_KIND_TEXTURECUBEARRAY = 9,
|
|
RESOURCE_KIND_TYPEDBUFFER = 10,
|
|
RESOURCE_KIND_RAWBUFFER = 11,
|
|
RESOURCE_KIND_STRUCTUREDBUFFER = 12,
|
|
RESOURCE_KIND_CBUFFER = 13,
|
|
RESOURCE_KIND_SAMPLER = 14,
|
|
RESOURCE_KIND_TBUFFER = 15,
|
|
RESOURCE_KIND_RTACCELERATIONSTRUCTURE = 16,
|
|
RESOURCE_KIND_FEEDBACKTEXTURE2D = 17,
|
|
RESOURCE_KIND_FEEDBACKTEXTURE2DARRAY = 18,
|
|
};
|
|
|
|
enum dxil_resource_tag
|
|
{
|
|
RESOURCE_TAG_ELEMENT_TYPE = 0,
|
|
RESOURCE_TAG_ELEMENT_STRIDE = 1,
|
|
RESOURCE_TAG_SAMPLER_FEEDBACK_KIND = 2,
|
|
RESOURCE_TAG_ENABLE_ATOMIC_64 = 3,
|
|
};
|
|
|
|
enum dxil_component_type
|
|
{
|
|
COMPONENT_TYPE_INVALID = 0,
|
|
COMPONENT_TYPE_I1 = 1,
|
|
COMPONENT_TYPE_I16 = 2,
|
|
COMPONENT_TYPE_U16 = 3,
|
|
COMPONENT_TYPE_I32 = 4,
|
|
COMPONENT_TYPE_U32 = 5,
|
|
COMPONENT_TYPE_I64 = 6,
|
|
COMPONENT_TYPE_U64 = 7,
|
|
COMPONENT_TYPE_F16 = 8,
|
|
COMPONENT_TYPE_F32 = 9,
|
|
COMPONENT_TYPE_F64 = 10,
|
|
COMPONENT_TYPE_SNORMF16 = 11,
|
|
COMPONENT_TYPE_UNORMF16 = 12,
|
|
COMPONENT_TYPE_SNORMF32 = 13,
|
|
COMPONENT_TYPE_UNORMF32 = 14,
|
|
COMPONENT_TYPE_SNORMF64 = 15,
|
|
COMPONENT_TYPE_UNORMF64 = 16,
|
|
COMPONENT_TYPE_PACKEDS8X32 = 17,
|
|
COMPONENT_TYPE_PACKEDU8X32 = 18,
|
|
};
|
|
|
|
enum dxil_sampler_kind
|
|
{
|
|
SAMPLER_KIND_DEFAULT = 0,
|
|
SAMPLER_KIND_COMPARISON = 1,
|
|
SAMPLER_KIND_MONO = 2,
|
|
};
|
|
|
|
enum dxil_semantic_kind
|
|
{
|
|
SEMANTIC_KIND_ARBITRARY = 0,
|
|
SEMANTIC_KIND_VERTEXID = 1,
|
|
SEMANTIC_KIND_INSTANCEID = 2,
|
|
SEMANTIC_KIND_POSITION = 3,
|
|
SEMANTIC_KIND_RTARRAYINDEX = 4,
|
|
SEMANTIC_KIND_VIEWPORTARRAYINDEX = 5,
|
|
SEMANTIC_KIND_CLIPDISTANCE = 6,
|
|
SEMANTIC_KIND_CULLDISTANCE = 7,
|
|
SEMANTIC_KIND_OUTPUTCONTROLPOINTID = 8,
|
|
SEMANTIC_KIND_DOMAINLOCATION = 9,
|
|
SEMANTIC_KIND_PRIMITIVEID = 10,
|
|
SEMANTIC_KIND_GSINSTANCEID = 11,
|
|
SEMANTIC_KIND_SAMPLEINDEX = 12,
|
|
SEMANTIC_KIND_ISFRONTFACE = 13,
|
|
SEMANTIC_KIND_COVERAGE = 14,
|
|
SEMANTIC_KIND_INNERCOVERAGE = 15,
|
|
SEMANTIC_KIND_TARGET = 16,
|
|
SEMANTIC_KIND_DEPTH = 17,
|
|
SEMANTIC_KIND_DEPTHLESSEQUAL = 18,
|
|
SEMANTIC_KIND_DEPTHGREATEREQUAL = 19,
|
|
SEMANTIC_KIND_STENCILREF = 20,
|
|
SEMANTIC_KIND_DISPATCHTHREADID = 21,
|
|
SEMANTIC_KIND_GROUPID = 22,
|
|
SEMANTIC_KIND_GROUPINDEX = 23,
|
|
SEMANTIC_KIND_GROUPTHREADID = 24,
|
|
SEMANTIC_KIND_TESSFACTOR = 25,
|
|
SEMANTIC_KIND_INSIDETESSFACTOR = 26,
|
|
SEMANTIC_KIND_VIEWID = 27,
|
|
SEMANTIC_KIND_BARYCENTRICS = 28,
|
|
SEMANTIC_KIND_SHADINGRATE = 29,
|
|
SEMANTIC_KIND_CULLPRIMITIVE = 30,
|
|
SEMANTIC_KIND_COUNT = 31,
|
|
SEMANTIC_KIND_INVALID = SEMANTIC_KIND_COUNT,
|
|
};
|
|
|
|
enum dxil_element_additional_tag
|
|
{
|
|
ADDITIONAL_TAG_STREAM_INDEX = 0,
|
|
ADDITIONAL_TAG_GLOBAL_SYMBOL = 1, /* not used */
|
|
ADDITIONAL_TAG_RELADDR_MASK = 2,
|
|
ADDITIONAL_TAG_USED_MASK = 3,
|
|
};
|
|
|
|
enum dxil_input_primitive
|
|
{
|
|
INPUT_PRIMITIVE_UNDEFINED = 0,
|
|
INPUT_PRIMITIVE_POINT = 1,
|
|
INPUT_PRIMITIVE_LINE = 2,
|
|
INPUT_PRIMITIVE_TRIANGLE = 3,
|
|
INPUT_PRIMITIVE_LINEWITHADJACENCY = 6,
|
|
INPUT_PRIMITIVE_TRIANGLEWITHADJACENY = 7,
|
|
INPUT_PRIMITIVE_PATCH1 = 8,
|
|
INPUT_PRIMITIVE_PATCH32 = 39,
|
|
};
|
|
|
|
enum dxil_shader_properties_tag
|
|
{
|
|
SHADER_PROPERTIES_FLAGS = 0,
|
|
SHADER_PROPERTIES_GEOMETRY = 1,
|
|
SHADER_PROPERTIES_DOMAIN = 2,
|
|
SHADER_PROPERTIES_HULL = 3,
|
|
SHADER_PROPERTIES_COMPUTE = 4,
|
|
SHADER_PROPERTIES_AUTO_BINDING_SPACE = 5,
|
|
SHADER_PROPERTIES_RAY_PAYLOAD_SIZE = 6,
|
|
SHADER_PROPERTIES_RAY_ATTRIB_SIZE = 7,
|
|
SHADER_PROPERTIES_SHADER_KIND = 8,
|
|
SHADER_PROPERTIES_MESH = 9,
|
|
SHADER_PROPERTIES_AMPLIFICATION = 10,
|
|
SHADER_PROPERTIES_WAVE_SIZE = 11,
|
|
SHADER_PROPERTIES_ENTRY_ROOT_SIG = 12,
|
|
};
|
|
|
|
enum dxil_binop_code
|
|
{
|
|
BINOP_ADD = 0,
|
|
BINOP_SUB = 1,
|
|
BINOP_MUL = 2,
|
|
BINOP_UDIV = 3,
|
|
BINOP_SDIV = 4,
|
|
BINOP_UREM = 5,
|
|
BINOP_SREM = 6,
|
|
BINOP_SHL = 7,
|
|
BINOP_LSHR = 8,
|
|
BINOP_ASHR = 9,
|
|
BINOP_AND = 10,
|
|
BINOP_OR = 11,
|
|
BINOP_XOR = 12
|
|
};
|
|
|
|
enum dxil_fast_fp_flags
|
|
{
|
|
FP_ALLOW_UNSAFE_ALGEBRA = 0x1,
|
|
FP_NO_NAN = 0x2,
|
|
FP_NO_INF = 0x4,
|
|
FP_NO_SIGNED_ZEROS = 0x8,
|
|
FP_ALLOW_RECIPROCAL = 0x10,
|
|
};
|
|
|
|
enum dxil_overflowing_binop_flags
|
|
{
|
|
/* Operation is known to never overflow. */
|
|
OB_NO_UNSIGNED_WRAP = 0x1,
|
|
OB_NO_SIGNED_WRAP = 0x2,
|
|
};
|
|
|
|
enum dxil_possibly_exact_binop_flags
|
|
{
|
|
/* "A udiv or sdiv instruction, which can be marked as "exact", indicating that no bits are destroyed." */
|
|
PEB_EXACT = 0x1,
|
|
};
|
|
|
|
enum dx_intrinsic_opcode
|
|
{
|
|
DX_LOAD_INPUT = 4,
|
|
DX_STORE_OUTPUT = 5,
|
|
DX_FABS = 6,
|
|
DX_SATURATE = 7,
|
|
DX_ISNAN = 8,
|
|
DX_ISINF = 9,
|
|
DX_ISFINITE = 10,
|
|
DX_COS = 12,
|
|
DX_SIN = 13,
|
|
DX_TAN = 14,
|
|
DX_ACOS = 15,
|
|
DX_ASIN = 16,
|
|
DX_ATAN = 17,
|
|
DX_HCOS = 18,
|
|
DX_HSIN = 19,
|
|
DX_HTAN = 20,
|
|
DX_EXP = 21,
|
|
DX_FRC = 22,
|
|
DX_LOG = 23,
|
|
DX_SQRT = 24,
|
|
DX_RSQRT = 25,
|
|
DX_ROUND_NE = 26,
|
|
DX_ROUND_NI = 27,
|
|
DX_ROUND_PI = 28,
|
|
DX_ROUND_Z = 29,
|
|
DX_BFREV = 30,
|
|
DX_COUNT_BITS = 31,
|
|
DX_FIRST_BIT_LO = 32,
|
|
DX_FIRST_BIT_HI = 33,
|
|
DX_FIRST_BIT_SHI = 34,
|
|
DX_FMAX = 35,
|
|
DX_FMIN = 36,
|
|
DX_IMAX = 37,
|
|
DX_IMIN = 38,
|
|
DX_UMAX = 39,
|
|
DX_UMIN = 40,
|
|
DX_FMAD = 46,
|
|
DX_FMA = 47,
|
|
DX_IMAD = 48,
|
|
DX_UMAD = 49,
|
|
DX_IBFE = 51,
|
|
DX_UBFE = 52,
|
|
DX_DOT2 = 54,
|
|
DX_DOT3 = 55,
|
|
DX_DOT4 = 56,
|
|
DX_CREATE_HANDLE = 57,
|
|
DX_CBUFFER_LOAD_LEGACY = 59,
|
|
DX_SAMPLE = 60,
|
|
DX_SAMPLE_B = 61,
|
|
DX_SAMPLE_LOD = 62,
|
|
DX_SAMPLE_GRAD = 63,
|
|
DX_SAMPLE_C = 64,
|
|
DX_SAMPLE_C_LZ = 65,
|
|
DX_TEXTURE_LOAD = 66,
|
|
DX_TEXTURE_STORE = 67,
|
|
DX_BUFFER_LOAD = 68,
|
|
DX_BUFFER_STORE = 69,
|
|
DX_BUFFER_UPDATE_COUNTER = 70,
|
|
DX_GET_DIMENSIONS = 72,
|
|
DX_TEXTURE_GATHER = 73,
|
|
DX_TEXTURE_GATHER_CMP = 74,
|
|
DX_TEX2DMS_GET_SAMPLE_POS = 75,
|
|
DX_RT_GET_SAMPLE_POS = 76,
|
|
DX_RT_GET_SAMPLE_COUNT = 77,
|
|
DX_ATOMIC_BINOP = 78,
|
|
DX_ATOMIC_CMP_XCHG = 79,
|
|
DX_BARRIER = 80,
|
|
DX_CALCULATE_LOD = 81,
|
|
DX_DISCARD = 82,
|
|
DX_DERIV_COARSEX = 83,
|
|
DX_DERIV_COARSEY = 84,
|
|
DX_DERIV_FINEX = 85,
|
|
DX_DERIV_FINEY = 86,
|
|
DX_COVERAGE = 91,
|
|
DX_THREAD_ID = 93,
|
|
DX_GROUP_ID = 94,
|
|
DX_THREAD_ID_IN_GROUP = 95,
|
|
DX_FLATTENED_THREAD_ID_IN_GROUP = 96,
|
|
DX_EMIT_STREAM = 97,
|
|
DX_CUT_STREAM = 98,
|
|
DX_EMIT_THEN_CUT_STREAM = 99,
|
|
DX_MAKE_DOUBLE = 101,
|
|
DX_SPLIT_DOUBLE = 102,
|
|
DX_LOAD_OUTPUT_CONTROL_POINT = 103,
|
|
DX_LOAD_PATCH_CONSTANT = 104,
|
|
DX_DOMAIN_LOCATION = 105,
|
|
DX_STORE_PATCH_CONSTANT = 106,
|
|
DX_OUTPUT_CONTROL_POINT_ID = 107,
|
|
DX_PRIMITIVE_ID = 108,
|
|
DX_WAVE_IS_FIRST_LANE = 110,
|
|
DX_WAVE_GET_LANE_INDEX = 111,
|
|
DX_WAVE_GET_LANE_COUNT = 112,
|
|
DX_WAVE_ANY_TRUE = 113,
|
|
DX_WAVE_ALL_TRUE = 114,
|
|
DX_WAVE_ACTIVE_ALL_EQUAL = 115,
|
|
DX_WAVE_ACTIVE_BALLOT = 116,
|
|
DX_WAVE_READ_LANE_AT = 117,
|
|
DX_WAVE_READ_LANE_FIRST = 118,
|
|
DX_WAVE_ACTIVE_OP = 119,
|
|
DX_WAVE_ACTIVE_BIT = 120,
|
|
DX_WAVE_PREFIX_OP = 121,
|
|
DX_QUAD_READ_LANE_AT = 122,
|
|
DX_QUAD_OP = 123,
|
|
DX_LEGACY_F32TOF16 = 130,
|
|
DX_LEGACY_F16TOF32 = 131,
|
|
DX_WAVE_ALL_BIT_COUNT = 135,
|
|
DX_WAVE_PREFIX_BIT_COUNT = 136,
|
|
DX_RAW_BUFFER_LOAD = 139,
|
|
DX_RAW_BUFFER_STORE = 140,
|
|
};
|
|
|
|
enum dxil_cast_code
|
|
{
|
|
CAST_TRUNC = 0,
|
|
CAST_ZEXT = 1,
|
|
CAST_SEXT = 2,
|
|
CAST_FPTOUI = 3,
|
|
CAST_FPTOSI = 4,
|
|
CAST_UITOFP = 5,
|
|
CAST_SITOFP = 6,
|
|
CAST_FPTRUNC = 7,
|
|
CAST_FPEXT = 8,
|
|
CAST_PTRTOINT = 9,
|
|
CAST_INTTOPTR = 10,
|
|
CAST_BITCAST = 11,
|
|
CAST_ADDRSPACECAST = 12,
|
|
};
|
|
|
|
enum dxil_predicate
|
|
{
|
|
FCMP_FALSE = 0,
|
|
FCMP_OEQ = 1,
|
|
FCMP_OGT = 2,
|
|
FCMP_OGE = 3,
|
|
FCMP_OLT = 4,
|
|
FCMP_OLE = 5,
|
|
FCMP_ONE = 6,
|
|
FCMP_ORD = 7,
|
|
FCMP_UNO = 8,
|
|
FCMP_UEQ = 9,
|
|
FCMP_UGT = 10,
|
|
FCMP_UGE = 11,
|
|
FCMP_ULT = 12,
|
|
FCMP_ULE = 13,
|
|
FCMP_UNE = 14,
|
|
FCMP_TRUE = 15,
|
|
ICMP_EQ = 32,
|
|
ICMP_NE = 33,
|
|
ICMP_UGT = 34,
|
|
ICMP_UGE = 35,
|
|
ICMP_ULT = 36,
|
|
ICMP_ULE = 37,
|
|
ICMP_SGT = 38,
|
|
ICMP_SGE = 39,
|
|
ICMP_SLT = 40,
|
|
ICMP_SLE = 41,
|
|
};
|
|
|
|
enum dxil_rmw_code
|
|
{
|
|
RMW_XCHG = 0,
|
|
RMW_ADD = 1,
|
|
RMW_SUB = 2,
|
|
RMW_AND = 3,
|
|
RMW_NAND = 4,
|
|
RMW_OR = 5,
|
|
RMW_XOR = 6,
|
|
RMW_MAX = 7,
|
|
RMW_MIN = 8,
|
|
RMW_UMAX = 9,
|
|
RMW_UMIN = 10,
|
|
};
|
|
|
|
enum dxil_atomic_ordering
|
|
{
|
|
ORDERING_NOTATOMIC = 0,
|
|
ORDERING_UNORDERED = 1,
|
|
ORDERING_MONOTONIC = 2,
|
|
ORDERING_ACQUIRE = 3,
|
|
ORDERING_RELEASE = 4,
|
|
ORDERING_ACQREL = 5,
|
|
ORDERING_SEQCST = 6,
|
|
};
|
|
|
|
enum dxil_atomic_binop_code
|
|
{
|
|
ATOMIC_BINOP_ADD,
|
|
ATOMIC_BINOP_AND,
|
|
ATOMIC_BINOP_OR,
|
|
ATOMIC_BINOP_XOR,
|
|
ATOMIC_BINOP_IMIN,
|
|
ATOMIC_BINOP_IMAX,
|
|
ATOMIC_BINOP_UMIN,
|
|
ATOMIC_BINOP_UMAX,
|
|
ATOMIC_BINOP_XCHG,
|
|
ATOMIC_BINOP_INVALID,
|
|
};
|
|
|
|
enum dxil_sync_flags
|
|
{
|
|
SYNC_THREAD_GROUP = 0x1,
|
|
SYNC_GLOBAL_UAV = 0x2,
|
|
SYNC_THREAD_GROUP_UAV = 0x4,
|
|
SYNC_GROUP_SHARED_MEMORY = 0x8,
|
|
};
|
|
|
|
enum dxil_wave_bit_op_kind
|
|
{
|
|
WAVE_BIT_OP_AND = 0,
|
|
WAVE_BIT_OP_OR = 1,
|
|
WAVE_BIT_OP_XOR = 2,
|
|
};
|
|
|
|
enum dxil_wave_op_kind
|
|
{
|
|
WAVE_OP_ADD = 0,
|
|
WAVE_OP_MUL = 1,
|
|
WAVE_OP_MIN = 2,
|
|
WAVE_OP_MAX = 3,
|
|
};
|
|
|
|
enum dxil_quad_op_kind
|
|
{
|
|
QUAD_READ_ACROSS_X = 0,
|
|
QUAD_READ_ACROSS_Y = 1,
|
|
QUAD_READ_ACROSS_D = 2,
|
|
};
|
|
|
|
struct sm6_pointer_info
|
|
{
|
|
const struct sm6_type *type;
|
|
enum bitcode_address_space addr_space;
|
|
};
|
|
|
|
struct sm6_struct_info
|
|
{
|
|
const char *name;
|
|
unsigned int elem_count;
|
|
const struct sm6_type *elem_types[];
|
|
};
|
|
|
|
struct sm6_function_info
|
|
{
|
|
const struct sm6_type *ret_type;
|
|
unsigned int param_count;
|
|
const struct sm6_type *param_types[];
|
|
};
|
|
|
|
struct sm6_array_info
|
|
{
|
|
unsigned int count;
|
|
const struct sm6_type *elem_type;
|
|
};
|
|
|
|
enum sm6_type_class
|
|
{
|
|
TYPE_CLASS_VOID,
|
|
TYPE_CLASS_INTEGER,
|
|
TYPE_CLASS_FLOAT,
|
|
TYPE_CLASS_POINTER,
|
|
TYPE_CLASS_STRUCT,
|
|
TYPE_CLASS_FUNCTION,
|
|
TYPE_CLASS_VECTOR,
|
|
TYPE_CLASS_ARRAY,
|
|
TYPE_CLASS_LABEL,
|
|
TYPE_CLASS_METADATA,
|
|
};
|
|
|
|
struct sm6_type
|
|
{
|
|
enum sm6_type_class class;
|
|
union
|
|
{
|
|
unsigned int width;
|
|
struct sm6_pointer_info pointer;
|
|
struct sm6_struct_info *struc;
|
|
struct sm6_function_info *function;
|
|
struct sm6_array_info array;
|
|
} u;
|
|
};
|
|
|
|
enum sm6_value_type
|
|
{
|
|
VALUE_TYPE_FUNCTION,
|
|
VALUE_TYPE_REG,
|
|
VALUE_TYPE_ICB,
|
|
VALUE_TYPE_HANDLE,
|
|
};
|
|
|
|
struct sm6_function_data
|
|
{
|
|
const char *name;
|
|
bool is_prototype;
|
|
unsigned int attribs_id;
|
|
};
|
|
|
|
struct sm6_handle_data
|
|
{
|
|
const struct sm6_descriptor_info *d;
|
|
struct vkd3d_shader_register reg;
|
|
};
|
|
|
|
struct sm6_value
|
|
{
|
|
const struct sm6_type *type;
|
|
enum sm6_value_type value_type;
|
|
unsigned int structure_stride;
|
|
bool is_undefined;
|
|
bool is_back_ref;
|
|
union
|
|
{
|
|
struct sm6_function_data function;
|
|
struct vkd3d_shader_register reg;
|
|
const struct vkd3d_shader_immediate_constant_buffer *icb;
|
|
struct sm6_handle_data handle;
|
|
} u;
|
|
};
|
|
|
|
struct dxil_record
|
|
{
|
|
unsigned int code;
|
|
unsigned int operand_count;
|
|
const struct dxil_record *attachment;
|
|
uint64_t operands[];
|
|
};
|
|
|
|
struct sm6_symbol
|
|
{
|
|
unsigned int id;
|
|
const char *name;
|
|
};
|
|
|
|
struct incoming_value
|
|
{
|
|
const struct sm6_block *block;
|
|
struct vkd3d_shader_register reg;
|
|
};
|
|
|
|
struct sm6_phi
|
|
{
|
|
struct vkd3d_shader_register reg;
|
|
struct incoming_value *incoming;
|
|
size_t incoming_capacity;
|
|
size_t incoming_count;
|
|
};
|
|
|
|
enum sm6_block_terminator_type
|
|
{
|
|
TERMINATOR_UNCOND_BR,
|
|
TERMINATOR_COND_BR,
|
|
TERMINATOR_SWITCH,
|
|
TERMINATOR_RET,
|
|
};
|
|
|
|
struct terminator_case
|
|
{
|
|
const struct sm6_block *block;
|
|
uint64_t value;
|
|
bool is_default;
|
|
};
|
|
|
|
struct sm6_block_terminator
|
|
{
|
|
struct vkd3d_shader_register conditional_reg;
|
|
enum sm6_block_terminator_type type;
|
|
const struct sm6_block *true_block;
|
|
const struct sm6_block *false_block;
|
|
struct terminator_case *cases;
|
|
unsigned int case_count;
|
|
};
|
|
|
|
struct sm6_block
|
|
{
|
|
struct vkd3d_shader_instruction *instructions;
|
|
size_t instruction_capacity;
|
|
size_t instruction_count;
|
|
|
|
/* A nonzero id. */
|
|
unsigned int id;
|
|
|
|
struct sm6_phi *phi;
|
|
size_t phi_capacity;
|
|
size_t phi_count;
|
|
|
|
struct sm6_block_terminator terminator;
|
|
};
|
|
|
|
struct sm6_function
|
|
{
|
|
const struct sm6_value *declaration;
|
|
|
|
struct sm6_block **blocks;
|
|
size_t block_capacity;
|
|
size_t block_count;
|
|
|
|
size_t value_count;
|
|
};
|
|
|
|
struct dxil_block
|
|
{
|
|
const struct dxil_block *parent;
|
|
enum bitcode_block_id id;
|
|
unsigned int abbrev_len;
|
|
unsigned int start;
|
|
unsigned int length;
|
|
unsigned int level;
|
|
|
|
/* The abbrev, block and record structs are not relocatable. */
|
|
struct dxil_abbrev **abbrevs;
|
|
size_t abbrev_capacity;
|
|
size_t abbrev_count;
|
|
unsigned int blockinfo_bid;
|
|
bool has_bid;
|
|
|
|
struct dxil_block **child_blocks;
|
|
size_t child_block_capacity;
|
|
size_t child_block_count;
|
|
|
|
struct dxil_record **records;
|
|
size_t record_capacity;
|
|
size_t record_count;
|
|
};
|
|
|
|
enum sm6_metadata_type
|
|
{
|
|
VKD3D_METADATA_KIND,
|
|
VKD3D_METADATA_NODE,
|
|
VKD3D_METADATA_STRING,
|
|
VKD3D_METADATA_VALUE,
|
|
};
|
|
|
|
struct sm6_metadata_node
|
|
{
|
|
bool is_distinct;
|
|
unsigned int operand_count;
|
|
struct sm6_metadata_value *operands[];
|
|
};
|
|
|
|
struct sm6_metadata_kind
|
|
{
|
|
uint64_t id;
|
|
char *name;
|
|
};
|
|
|
|
struct sm6_metadata_value
|
|
{
|
|
enum sm6_metadata_type type;
|
|
const struct sm6_type *value_type;
|
|
union
|
|
{
|
|
char *string_value;
|
|
const struct sm6_value *value;
|
|
struct sm6_metadata_node *node;
|
|
struct sm6_metadata_kind kind;
|
|
} u;
|
|
};
|
|
|
|
struct sm6_metadata_table
|
|
{
|
|
struct sm6_metadata_value *values;
|
|
unsigned int count;
|
|
};
|
|
|
|
struct sm6_named_metadata
|
|
{
|
|
char *name;
|
|
struct sm6_metadata_value value;
|
|
};
|
|
|
|
struct sm6_descriptor_info
|
|
{
|
|
enum vkd3d_shader_descriptor_type type;
|
|
unsigned int id;
|
|
struct vkd3d_shader_register_range range;
|
|
enum vkd3d_shader_resource_type resource_type;
|
|
enum dxil_resource_kind kind;
|
|
enum vkd3d_data_type resource_data_type;
|
|
enum vkd3d_shader_register_type reg_type;
|
|
enum vkd3d_data_type reg_data_type;
|
|
};
|
|
|
|
struct sm6_parser
|
|
{
|
|
const uint32_t *ptr, *start, *end;
|
|
unsigned int bitpos;
|
|
|
|
struct dxil_block root_block;
|
|
struct dxil_block *current_block;
|
|
|
|
struct dxil_global_abbrev **abbrevs;
|
|
size_t abbrev_capacity;
|
|
size_t abbrev_count;
|
|
|
|
struct sm6_type *types;
|
|
size_t type_count;
|
|
struct sm6_type *bool_type;
|
|
struct sm6_type *metadata_type;
|
|
struct sm6_type *handle_type;
|
|
|
|
struct sm6_symbol *global_symbols;
|
|
size_t global_symbol_count;
|
|
|
|
const char *entry_point;
|
|
const char *patch_constant_function;
|
|
|
|
struct vkd3d_shader_dst_param *output_params;
|
|
struct vkd3d_shader_dst_param *input_params;
|
|
struct vkd3d_shader_dst_param *patch_constant_params;
|
|
uint32_t io_regs_declared[VKD3D_BITMAP_SIZE(VKD3DSPR_COUNT)];
|
|
|
|
struct sm6_function *functions;
|
|
size_t function_count;
|
|
|
|
struct sm6_metadata_table metadata_tables[VKD3D_SM6_MAX_METADATA_TABLES];
|
|
struct sm6_named_metadata *named_metadata;
|
|
unsigned int named_metadata_count;
|
|
|
|
struct sm6_descriptor_info *descriptors;
|
|
size_t descriptor_capacity;
|
|
size_t descriptor_count;
|
|
|
|
unsigned int indexable_temp_count;
|
|
unsigned int icb_count;
|
|
unsigned int tgsm_count;
|
|
|
|
struct sm6_value *values;
|
|
size_t value_count;
|
|
size_t value_capacity;
|
|
size_t cur_max_value;
|
|
unsigned int ssa_next_id;
|
|
|
|
struct vkd3d_shader_parser p;
|
|
};
|
|
|
|
struct dxil_abbrev_operand
|
|
{
|
|
uint64_t context;
|
|
bool (*read_operand)(struct sm6_parser *sm6, uint64_t context, uint64_t *operand);
|
|
};
|
|
|
|
struct dxil_abbrev
|
|
{
|
|
unsigned int count;
|
|
bool is_array;
|
|
struct dxil_abbrev_operand operands[];
|
|
};
|
|
|
|
struct dxil_global_abbrev
|
|
{
|
|
unsigned int block_id;
|
|
struct dxil_abbrev abbrev;
|
|
};
|
|
|
|
static const uint64_t CALL_CONV_FLAG_EXPLICIT_TYPE = 1ull << 15;
|
|
|
|
static size_t size_add_with_overflow_check(size_t a, size_t b)
|
|
{
|
|
size_t i = a + b;
|
|
return (i < a) ? SIZE_MAX : i;
|
|
}
|
|
|
|
static bool sm6_parser_is_end(struct sm6_parser *sm6)
|
|
{
|
|
return sm6->ptr == sm6->end;
|
|
}
|
|
|
|
static uint32_t sm6_parser_read_uint32(struct sm6_parser *sm6)
|
|
{
|
|
if (sm6_parser_is_end(sm6))
|
|
{
|
|
sm6->p.failed = true;
|
|
return 0;
|
|
}
|
|
return *sm6->ptr++;
|
|
}
|
|
|
|
static uint32_t sm6_parser_read_bits(struct sm6_parser *sm6, unsigned int length)
|
|
{
|
|
unsigned int l, prev_len = 0;
|
|
uint32_t bits;
|
|
|
|
if (!length)
|
|
return 0;
|
|
|
|
VKD3D_ASSERT(length < 32);
|
|
|
|
if (sm6_parser_is_end(sm6))
|
|
{
|
|
sm6->p.failed = true;
|
|
return 0;
|
|
}
|
|
|
|
VKD3D_ASSERT(sm6->bitpos < 32);
|
|
bits = *sm6->ptr >> sm6->bitpos;
|
|
l = 32 - sm6->bitpos;
|
|
if (l <= length)
|
|
{
|
|
++sm6->ptr;
|
|
if (sm6_parser_is_end(sm6) && l < length)
|
|
{
|
|
sm6->p.failed = true;
|
|
return bits;
|
|
}
|
|
sm6->bitpos = 0;
|
|
bits |= *sm6->ptr << l;
|
|
prev_len = l;
|
|
}
|
|
sm6->bitpos += length - prev_len;
|
|
|
|
return bits & ((1 << length) - 1);
|
|
}
|
|
|
|
static uint64_t sm6_parser_read_vbr(struct sm6_parser *sm6, unsigned int length)
|
|
{
|
|
unsigned int bits, flag, mask, shift = 0;
|
|
uint64_t result = 0;
|
|
|
|
if (!length)
|
|
return 0;
|
|
|
|
if (sm6_parser_is_end(sm6))
|
|
{
|
|
sm6->p.failed = true;
|
|
return 0;
|
|
}
|
|
|
|
flag = 1 << (length - 1);
|
|
mask = flag - 1;
|
|
do
|
|
{
|
|
bits = sm6_parser_read_bits(sm6, length);
|
|
result |= (uint64_t)(bits & mask) << shift;
|
|
shift += length - 1;
|
|
} while ((bits & flag) && !sm6->p.failed && shift < 64);
|
|
|
|
sm6->p.failed |= !!(bits & flag);
|
|
|
|
return result;
|
|
}
|
|
|
|
static void sm6_parser_align_32(struct sm6_parser *sm6)
|
|
{
|
|
if (!sm6->bitpos)
|
|
return;
|
|
|
|
if (sm6_parser_is_end(sm6))
|
|
{
|
|
sm6->p.failed = true;
|
|
return;
|
|
}
|
|
|
|
++sm6->ptr;
|
|
sm6->bitpos = 0;
|
|
}
|
|
|
|
static bool dxil_block_handle_blockinfo_record(struct dxil_block *block, struct dxil_record *record)
|
|
{
|
|
/* BLOCKINFO blocks must only occur immediately below the module root block. */
|
|
if (block->level > 1)
|
|
{
|
|
WARN("Invalid blockinfo block level %u.\n", block->level);
|
|
return false;
|
|
}
|
|
|
|
switch (record->code)
|
|
{
|
|
case SETBID:
|
|
if (!record->operand_count)
|
|
{
|
|
WARN("Missing id operand.\n");
|
|
return false;
|
|
}
|
|
if (record->operands[0] > UINT_MAX)
|
|
WARN("Truncating block id %"PRIu64".\n", record->operands[0]);
|
|
block->blockinfo_bid = record->operands[0];
|
|
block->has_bid = true;
|
|
break;
|
|
case BLOCKNAME:
|
|
case SETRECORDNAME:
|
|
break;
|
|
default:
|
|
FIXME("Unhandled BLOCKINFO record type %u.\n", record->code);
|
|
break;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static enum vkd3d_result dxil_block_add_record(struct dxil_block *block, struct dxil_record *record)
|
|
{
|
|
unsigned int reserve;
|
|
|
|
switch (block->id)
|
|
{
|
|
/* Rough initial reserve sizes for small shaders. */
|
|
case CONSTANTS_BLOCK: reserve = 32; break;
|
|
case FUNCTION_BLOCK: reserve = 128; break;
|
|
case METADATA_BLOCK: reserve = 32; break;
|
|
case TYPE_BLOCK: reserve = 32; break;
|
|
default: reserve = 8; break;
|
|
}
|
|
reserve = max(reserve, block->record_count + 1);
|
|
if (!vkd3d_array_reserve((void **)&block->records, &block->record_capacity, reserve, sizeof(*block->records)))
|
|
{
|
|
ERR("Failed to allocate %u records.\n", reserve);
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
if (block->id == BLOCKINFO_BLOCK && !dxil_block_handle_blockinfo_record(block, record))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
block->records[block->record_count++] = record;
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_read_unabbrev_record(struct sm6_parser *sm6)
|
|
{
|
|
struct dxil_block *block = sm6->current_block;
|
|
enum vkd3d_result ret = VKD3D_OK;
|
|
unsigned int code, count, i;
|
|
struct dxil_record *record;
|
|
|
|
code = sm6_parser_read_vbr(sm6, 6);
|
|
|
|
count = sm6_parser_read_vbr(sm6, 6);
|
|
if (!(record = vkd3d_malloc(sizeof(*record) + count * sizeof(record->operands[0]))))
|
|
{
|
|
ERR("Failed to allocate record with %u operands.\n", count);
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
record->code = code;
|
|
record->operand_count = count;
|
|
record->attachment = NULL;
|
|
|
|
for (i = 0; i < count; ++i)
|
|
record->operands[i] = sm6_parser_read_vbr(sm6, 6);
|
|
if (sm6->p.failed)
|
|
ret = VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
if (ret < 0 || (ret = dxil_block_add_record(block, record)) < 0)
|
|
vkd3d_free(record);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static bool sm6_parser_read_literal_operand(struct sm6_parser *sm6, uint64_t context, uint64_t *op)
|
|
{
|
|
*op = context;
|
|
return !sm6->p.failed;
|
|
}
|
|
|
|
static bool sm6_parser_read_fixed_operand(struct sm6_parser *sm6, uint64_t context, uint64_t *op)
|
|
{
|
|
*op = sm6_parser_read_bits(sm6, context);
|
|
return !sm6->p.failed;
|
|
}
|
|
|
|
static bool sm6_parser_read_vbr_operand(struct sm6_parser *sm6, uint64_t context, uint64_t *op)
|
|
{
|
|
*op = sm6_parser_read_vbr(sm6, context);
|
|
return !sm6->p.failed;
|
|
}
|
|
|
|
static bool sm6_parser_read_char6_operand(struct sm6_parser *sm6, uint64_t context, uint64_t *op)
|
|
{
|
|
*op = "abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789._"[sm6_parser_read_bits(sm6, 6)];
|
|
return !sm6->p.failed;
|
|
}
|
|
|
|
static bool sm6_parser_read_blob_operand(struct sm6_parser *sm6, uint64_t context, uint64_t *op)
|
|
{
|
|
int count = sm6_parser_read_vbr(sm6, 6);
|
|
sm6_parser_align_32(sm6);
|
|
for (; count > 0; count -= 4)
|
|
sm6_parser_read_uint32(sm6);
|
|
FIXME("Unhandled blob operand.\n");
|
|
return false;
|
|
}
|
|
|
|
static enum vkd3d_result dxil_abbrev_init(struct dxil_abbrev *abbrev, unsigned int count, struct sm6_parser *sm6)
|
|
{
|
|
enum bitcode_abbrev_type prev_type, type;
|
|
unsigned int i;
|
|
|
|
abbrev->is_array = false;
|
|
|
|
for (i = 0, prev_type = 0; i < count && !sm6->p.failed; ++i)
|
|
{
|
|
if (sm6_parser_read_bits(sm6, 1))
|
|
{
|
|
if (prev_type == ABBREV_ARRAY)
|
|
{
|
|
WARN("Unexpected literal abbreviation after array.\n");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
abbrev->operands[i].context = sm6_parser_read_vbr(sm6, 8);
|
|
abbrev->operands[i].read_operand = sm6_parser_read_literal_operand;
|
|
continue;
|
|
}
|
|
|
|
switch (type = sm6_parser_read_bits(sm6, 3))
|
|
{
|
|
case ABBREV_FIXED:
|
|
case ABBREV_VBR:
|
|
abbrev->operands[i].context = sm6_parser_read_vbr(sm6, 5);
|
|
abbrev->operands[i].read_operand = (type == ABBREV_FIXED) ? sm6_parser_read_fixed_operand
|
|
: sm6_parser_read_vbr_operand;
|
|
break;
|
|
|
|
case ABBREV_ARRAY:
|
|
if (prev_type == ABBREV_ARRAY || i != count - 2)
|
|
{
|
|
WARN("Unexpected array abbreviation.\n");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
abbrev->is_array = true;
|
|
--i;
|
|
--count;
|
|
break;
|
|
|
|
case ABBREV_CHAR:
|
|
abbrev->operands[i].read_operand = sm6_parser_read_char6_operand;
|
|
break;
|
|
|
|
case ABBREV_BLOB:
|
|
if (prev_type == ABBREV_ARRAY || i != count - 1)
|
|
{
|
|
WARN("Unexpected blob abbreviation.\n");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
abbrev->operands[i].read_operand = sm6_parser_read_blob_operand;
|
|
break;
|
|
}
|
|
|
|
prev_type = type;
|
|
}
|
|
|
|
abbrev->count = count;
|
|
|
|
return sm6->p.failed ? VKD3D_ERROR_INVALID_SHADER : VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_add_global_abbrev(struct sm6_parser *sm6)
|
|
{
|
|
struct dxil_block *block = sm6->current_block;
|
|
unsigned int count = sm6_parser_read_vbr(sm6, 5);
|
|
struct dxil_global_abbrev *global_abbrev;
|
|
enum vkd3d_result ret;
|
|
|
|
VKD3D_ASSERT(block->id == BLOCKINFO_BLOCK);
|
|
|
|
if (!vkd3d_array_reserve((void **)&sm6->abbrevs, &sm6->abbrev_capacity, sm6->abbrev_count + 1, sizeof(*sm6->abbrevs))
|
|
|| !(global_abbrev = vkd3d_malloc(sizeof(*global_abbrev) + count * sizeof(global_abbrev->abbrev.operands[0]))))
|
|
{
|
|
ERR("Failed to allocate global abbreviation.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
if ((ret = dxil_abbrev_init(&global_abbrev->abbrev, count, sm6)) < 0)
|
|
{
|
|
vkd3d_free(global_abbrev);
|
|
return ret;
|
|
}
|
|
|
|
if (!block->has_bid)
|
|
{
|
|
WARN("Missing blockinfo block id.\n");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
if (block->blockinfo_bid == MODULE_BLOCK)
|
|
{
|
|
FIXME("Unhandled global abbreviation for module block.\n");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
global_abbrev->block_id = block->blockinfo_bid;
|
|
|
|
sm6->abbrevs[sm6->abbrev_count++] = global_abbrev;
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_add_block_abbrev(struct sm6_parser *sm6)
|
|
{
|
|
struct dxil_block *block = sm6->current_block;
|
|
struct dxil_abbrev *abbrev;
|
|
enum vkd3d_result ret;
|
|
unsigned int count;
|
|
|
|
if (block->id == BLOCKINFO_BLOCK)
|
|
return sm6_parser_add_global_abbrev(sm6);
|
|
|
|
count = sm6_parser_read_vbr(sm6, 5);
|
|
if (!vkd3d_array_reserve((void **)&block->abbrevs, &block->abbrev_capacity, block->abbrev_count + 1, sizeof(*block->abbrevs))
|
|
|| !(abbrev = vkd3d_malloc(sizeof(*abbrev) + count * sizeof(abbrev->operands[0]))))
|
|
{
|
|
ERR("Failed to allocate block abbreviation.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
if ((ret = dxil_abbrev_init(abbrev, count, sm6)) < 0)
|
|
{
|
|
vkd3d_free(abbrev);
|
|
return ret;
|
|
}
|
|
|
|
block->abbrevs[block->abbrev_count++] = abbrev;
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_read_abbrev_record(struct sm6_parser *sm6, unsigned int abbrev_id)
|
|
{
|
|
enum vkd3d_result ret = VKD3D_ERROR_INVALID_SHADER;
|
|
struct dxil_block *block = sm6->current_block;
|
|
struct dxil_record *temp, *record;
|
|
unsigned int i, count, array_len;
|
|
struct dxil_abbrev *abbrev;
|
|
uint64_t code;
|
|
|
|
if (abbrev_id >= block->abbrev_count)
|
|
{
|
|
WARN("Invalid abbreviation id %u.\n", abbrev_id);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
abbrev = block->abbrevs[abbrev_id];
|
|
if (!(count = abbrev->count))
|
|
return VKD3D_OK;
|
|
if (count == 1 && abbrev->is_array)
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
/* First operand is the record code. The array is included in the count, but will be done separately. */
|
|
count -= abbrev->is_array + 1;
|
|
if (!(record = vkd3d_malloc(sizeof(*record) + count * sizeof(record->operands[0]))))
|
|
{
|
|
ERR("Failed to allocate record with %u operands.\n", count);
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
if (!abbrev->operands[0].read_operand(sm6, abbrev->operands[0].context, &code))
|
|
goto fail;
|
|
if (code > UINT_MAX)
|
|
FIXME("Truncating 64-bit record code %#"PRIx64".\n", code);
|
|
record->code = code;
|
|
|
|
for (i = 0; i < count; ++i)
|
|
if (!abbrev->operands[i + 1].read_operand(sm6, abbrev->operands[i + 1].context, &record->operands[i]))
|
|
goto fail;
|
|
record->operand_count = count;
|
|
record->attachment = NULL;
|
|
|
|
/* An array can occur only as the last operand. */
|
|
if (abbrev->is_array)
|
|
{
|
|
array_len = sm6_parser_read_vbr(sm6, 6);
|
|
if (!(temp = vkd3d_realloc(record, sizeof(*record) + (count + array_len) * sizeof(record->operands[0]))))
|
|
{
|
|
ERR("Failed to allocate record with %u operands.\n", count + array_len);
|
|
ret = VKD3D_ERROR_OUT_OF_MEMORY;
|
|
goto fail;
|
|
}
|
|
record = temp;
|
|
|
|
for (i = 0; i < array_len; ++i)
|
|
{
|
|
if (!abbrev->operands[count + 1].read_operand(sm6, abbrev->operands[count + 1].context,
|
|
&record->operands[count + i]))
|
|
{
|
|
goto fail;
|
|
}
|
|
}
|
|
record->operand_count += array_len;
|
|
}
|
|
|
|
if ((ret = dxil_block_add_record(block, record)) < 0)
|
|
goto fail;
|
|
|
|
return VKD3D_OK;
|
|
|
|
fail:
|
|
vkd3d_free(record);
|
|
return ret;
|
|
}
|
|
|
|
static enum vkd3d_result dxil_block_init(struct dxil_block *block, const struct dxil_block *parent,
|
|
struct sm6_parser *sm6);
|
|
|
|
static enum vkd3d_result dxil_block_read(struct dxil_block *parent, struct sm6_parser *sm6)
|
|
{
|
|
unsigned int reserve = (parent->id == MODULE_BLOCK) ? 12 : 2;
|
|
struct dxil_block *block;
|
|
enum vkd3d_result ret;
|
|
|
|
sm6->current_block = parent;
|
|
|
|
do
|
|
{
|
|
unsigned int abbrev_id = sm6_parser_read_bits(sm6, parent->abbrev_len);
|
|
|
|
switch (abbrev_id)
|
|
{
|
|
case END_BLOCK:
|
|
sm6_parser_align_32(sm6);
|
|
return VKD3D_OK;
|
|
|
|
case ENTER_SUBBLOCK:
|
|
if (parent->id != MODULE_BLOCK && parent->id != FUNCTION_BLOCK)
|
|
{
|
|
WARN("Invalid subblock parent id %u.\n", parent->id);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (!vkd3d_array_reserve((void **)&parent->child_blocks, &parent->child_block_capacity,
|
|
max(reserve, parent->child_block_count + 1), sizeof(*parent->child_blocks))
|
|
|| !(block = vkd3d_calloc(1, sizeof(*block))))
|
|
{
|
|
ERR("Failed to allocate block.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
if ((ret = dxil_block_init(block, parent, sm6)) < 0)
|
|
{
|
|
vkd3d_free(block);
|
|
return ret;
|
|
}
|
|
|
|
parent->child_blocks[parent->child_block_count++] = block;
|
|
sm6->current_block = parent;
|
|
break;
|
|
|
|
case DEFINE_ABBREV:
|
|
if ((ret = sm6_parser_add_block_abbrev(sm6)) < 0)
|
|
return ret;
|
|
break;
|
|
|
|
case UNABBREV_RECORD:
|
|
if ((ret = sm6_parser_read_unabbrev_record(sm6)) < 0)
|
|
{
|
|
WARN("Failed to read unabbreviated record.\n");
|
|
return ret;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
if ((ret = sm6_parser_read_abbrev_record(sm6, abbrev_id - 4)) < 0)
|
|
{
|
|
WARN("Failed to read abbreviated record.\n");
|
|
return ret;
|
|
}
|
|
break;
|
|
}
|
|
} while (!sm6->p.failed);
|
|
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
static size_t sm6_parser_compute_global_abbrev_count_for_block_id(struct sm6_parser *sm6,
|
|
unsigned int block_id)
|
|
{
|
|
size_t i, count;
|
|
|
|
for (i = 0, count = 0; i < sm6->abbrev_count; ++i)
|
|
count += sm6->abbrevs[i]->block_id == block_id;
|
|
|
|
return count;
|
|
}
|
|
|
|
static void dxil_block_destroy(struct dxil_block *block)
|
|
{
|
|
size_t i;
|
|
|
|
for (i = 0; i < block->record_count; ++i)
|
|
vkd3d_free(block->records[i]);
|
|
vkd3d_free(block->records);
|
|
|
|
for (i = 0; i < block->child_block_count; ++i)
|
|
{
|
|
dxil_block_destroy(block->child_blocks[i]);
|
|
vkd3d_free(block->child_blocks[i]);
|
|
}
|
|
vkd3d_free(block->child_blocks);
|
|
|
|
block->records = NULL;
|
|
block->record_count = 0;
|
|
block->child_blocks = NULL;
|
|
block->child_block_count = 0;
|
|
}
|
|
|
|
static enum vkd3d_result dxil_block_init(struct dxil_block *block, const struct dxil_block *parent,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
size_t i, abbrev_count = 0;
|
|
enum vkd3d_result ret;
|
|
|
|
block->parent = parent;
|
|
block->level = parent ? parent->level + 1 : 0;
|
|
block->id = sm6_parser_read_vbr(sm6, 8);
|
|
block->abbrev_len = sm6_parser_read_vbr(sm6, 4);
|
|
sm6_parser_align_32(sm6);
|
|
block->length = sm6_parser_read_uint32(sm6);
|
|
block->start = sm6->ptr - sm6->start;
|
|
|
|
if (sm6->p.failed)
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
if ((block->abbrev_count = sm6_parser_compute_global_abbrev_count_for_block_id(sm6, block->id)))
|
|
{
|
|
if (!vkd3d_array_reserve((void **)&block->abbrevs, &block->abbrev_capacity,
|
|
block->abbrev_count, sizeof(*block->abbrevs)))
|
|
{
|
|
ERR("Failed to allocate block abbreviations.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
for (i = 0; i < sm6->abbrev_count; ++i)
|
|
if (sm6->abbrevs[i]->block_id == block->id)
|
|
block->abbrevs[abbrev_count++] = &sm6->abbrevs[i]->abbrev;
|
|
|
|
VKD3D_ASSERT(abbrev_count == block->abbrev_count);
|
|
}
|
|
|
|
if ((ret = dxil_block_read(block, sm6)) < 0)
|
|
dxil_block_destroy(block);
|
|
|
|
for (i = abbrev_count; i < block->abbrev_count; ++i)
|
|
vkd3d_free(block->abbrevs[i]);
|
|
vkd3d_free(block->abbrevs);
|
|
block->abbrevs = NULL;
|
|
block->abbrev_count = 0;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static size_t dxil_block_compute_function_count(const struct dxil_block *root)
|
|
{
|
|
size_t i, count;
|
|
|
|
for (i = 0, count = 0; i < root->child_block_count; ++i)
|
|
count += root->child_blocks[i]->id == FUNCTION_BLOCK;
|
|
|
|
return count;
|
|
}
|
|
|
|
static size_t dxil_block_compute_module_decl_count(const struct dxil_block *block)
|
|
{
|
|
size_t i, count;
|
|
|
|
for (i = 0, count = 0; i < block->record_count; ++i)
|
|
count += block->records[i]->code == MODULE_CODE_FUNCTION || block->records[i]->code == MODULE_CODE_GLOBALVAR;
|
|
return count;
|
|
}
|
|
|
|
static size_t dxil_block_compute_constants_count(const struct dxil_block *block)
|
|
{
|
|
size_t i, count;
|
|
|
|
for (i = 0, count = 0; i < block->record_count; ++i)
|
|
count += block->records[i]->code != CST_CODE_SETTYPE;
|
|
return count;
|
|
}
|
|
|
|
static void dxil_global_abbrevs_cleanup(struct dxil_global_abbrev **abbrevs, size_t count)
|
|
{
|
|
size_t i;
|
|
|
|
for (i = 0; i < count; ++i)
|
|
vkd3d_free(abbrevs[i]);
|
|
vkd3d_free(abbrevs);
|
|
}
|
|
|
|
static const struct dxil_block *sm6_parser_get_level_one_block(const struct sm6_parser *sm6,
|
|
enum bitcode_block_id id, bool *is_unique)
|
|
{
|
|
const struct dxil_block *block, *found = NULL;
|
|
size_t i;
|
|
|
|
for (i = 0, *is_unique = true; i < sm6->root_block.child_block_count; ++i)
|
|
{
|
|
block = sm6->root_block.child_blocks[i];
|
|
if (block->id != id)
|
|
continue;
|
|
|
|
if (!found)
|
|
found = block;
|
|
else
|
|
*is_unique = false;
|
|
}
|
|
|
|
return found;
|
|
}
|
|
|
|
static char *dxil_record_to_string(const struct dxil_record *record, unsigned int offset, struct sm6_parser *sm6)
|
|
{
|
|
unsigned int i;
|
|
char *str;
|
|
|
|
VKD3D_ASSERT(offset <= record->operand_count);
|
|
if (!(str = vkd3d_calloc(record->operand_count - offset + 1, 1)))
|
|
{
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating a string of length %u.", record->operand_count - offset);
|
|
return NULL;
|
|
}
|
|
|
|
for (i = offset; i < record->operand_count; ++i)
|
|
str[i - offset] = record->operands[i];
|
|
|
|
return str;
|
|
}
|
|
|
|
static bool dxil_record_validate_operand_min_count(const struct dxil_record *record, unsigned int min_count,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
if (record->operand_count >= min_count)
|
|
return true;
|
|
|
|
WARN("Invalid operand count %u for code %u.\n", record->operand_count, record->code);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Invalid operand count %u for record code %u.", record->operand_count, record->code);
|
|
return false;
|
|
}
|
|
|
|
static void dxil_record_validate_operand_max_count(const struct dxil_record *record, unsigned int max_count,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
if (record->operand_count <= max_count)
|
|
return;
|
|
|
|
WARN("Ignoring %u extra operands for code %u.\n", record->operand_count - max_count, record->code);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring %u extra operands for record code %u.", record->operand_count - max_count, record->code);
|
|
}
|
|
|
|
static bool dxil_record_validate_operand_count(const struct dxil_record *record, unsigned int min_count,
|
|
unsigned int max_count, struct sm6_parser *sm6)
|
|
{
|
|
dxil_record_validate_operand_max_count(record, max_count, sm6);
|
|
return dxil_record_validate_operand_min_count(record, min_count, sm6);
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_type_table_init(struct sm6_parser *sm6)
|
|
{
|
|
const struct dxil_record *record;
|
|
size_t i, type_count, type_index;
|
|
const struct dxil_block *block;
|
|
char *struct_name = NULL;
|
|
unsigned int j, count;
|
|
struct sm6_type *type;
|
|
uint64_t type_id;
|
|
bool is_unique;
|
|
|
|
sm6->p.location.line = 0;
|
|
sm6->p.location.column = 0;
|
|
|
|
if (!(block = sm6_parser_get_level_one_block(sm6, TYPE_BLOCK, &is_unique)))
|
|
{
|
|
WARN("No type definitions found.\n");
|
|
return VKD3D_OK;
|
|
}
|
|
if (!is_unique)
|
|
WARN("Ignoring invalid extra type table(s).\n");
|
|
|
|
sm6->p.location.line = block->id;
|
|
|
|
type_count = 0;
|
|
for (i = 0; i < block->record_count; ++i)
|
|
type_count += block->records[i]->code != TYPE_CODE_NUMENTRY && block->records[i]->code != TYPE_CODE_STRUCT_NAME;
|
|
|
|
/* The type array must not be relocated. */
|
|
if (!(sm6->types = vkd3d_calloc(type_count, sizeof(*sm6->types))))
|
|
{
|
|
ERR("Failed to allocate type array.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
for (i = 0; i < block->record_count; ++i)
|
|
{
|
|
sm6->p.location.column = i;
|
|
record = block->records[i];
|
|
|
|
type = &sm6->types[sm6->type_count];
|
|
type_index = sm6->type_count;
|
|
|
|
switch (record->code)
|
|
{
|
|
case TYPE_CODE_ARRAY:
|
|
case TYPE_CODE_VECTOR:
|
|
if (!dxil_record_validate_operand_count(record, 2, 2, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
type->class = record->code == TYPE_CODE_ARRAY ? TYPE_CLASS_ARRAY : TYPE_CLASS_VECTOR;
|
|
|
|
if (!(type->u.array.count = record->operands[0]))
|
|
{
|
|
TRACE("Setting unbounded for type %zu.\n", type_index);
|
|
type->u.array.count = UINT_MAX;
|
|
}
|
|
|
|
if ((type_id = record->operands[1]) >= type_count)
|
|
{
|
|
WARN("Invalid contained type id %"PRIu64" for type %zu.\n", type_id, type_index);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
type->u.array.elem_type = &sm6->types[type_id];
|
|
break;
|
|
|
|
case TYPE_CODE_DOUBLE:
|
|
dxil_record_validate_operand_max_count(record, 0, sm6);
|
|
type->class = TYPE_CLASS_FLOAT;
|
|
type->u.width = 64;
|
|
break;
|
|
|
|
case TYPE_CODE_FLOAT:
|
|
dxil_record_validate_operand_max_count(record, 0, sm6);
|
|
type->class = TYPE_CLASS_FLOAT;
|
|
type->u.width = 32;
|
|
break;
|
|
|
|
case TYPE_CODE_FUNCTION:
|
|
if (!dxil_record_validate_operand_min_count(record, 2, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
if (record->operands[0])
|
|
FIXME("Unhandled vararg function type %zu.\n", type_index);
|
|
|
|
type->class = TYPE_CLASS_FUNCTION;
|
|
|
|
if ((type_id = record->operands[1]) >= type_count)
|
|
{
|
|
WARN("Invalid return type id %"PRIu64" for type %zu.\n", type_id, type_index);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
count = record->operand_count - 2;
|
|
if (vkd3d_object_range_overflow(sizeof(type->u.function), count, sizeof(type->u.function->param_types[0]))
|
|
|| !(type->u.function = vkd3d_malloc(offsetof(struct sm6_function_info, param_types[count]))))
|
|
{
|
|
ERR("Failed to allocate function parameter types.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
type->u.function->ret_type = &sm6->types[type_id];
|
|
type->u.function->param_count = count;
|
|
for (j = 0; j < count; ++j)
|
|
{
|
|
if ((type_id = record->operands[j + 2]) >= type_count)
|
|
{
|
|
WARN("Invalid parameter type id %"PRIu64" for type %zu.\n", type_id, type_index);
|
|
vkd3d_free(type->u.function);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
type->u.function->param_types[j] = &sm6->types[type_id];
|
|
}
|
|
break;
|
|
|
|
case TYPE_CODE_HALF:
|
|
dxil_record_validate_operand_max_count(record, 0, sm6);
|
|
type->class = TYPE_CLASS_FLOAT;
|
|
type->u.width = 16;
|
|
break;
|
|
|
|
case TYPE_CODE_INTEGER:
|
|
{
|
|
uint64_t width;
|
|
|
|
if (!dxil_record_validate_operand_count(record, 1, 1, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
type->class = TYPE_CLASS_INTEGER;
|
|
|
|
switch ((width = record->operands[0]))
|
|
{
|
|
case 1:
|
|
sm6->bool_type = type;
|
|
break;
|
|
case 8:
|
|
case 16:
|
|
case 32:
|
|
case 64:
|
|
break;
|
|
default:
|
|
WARN("Invalid integer width %"PRIu64" for type %zu.\n", width, type_index);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
type->u.width = width;
|
|
break;
|
|
}
|
|
|
|
case TYPE_CODE_LABEL:
|
|
type->class = TYPE_CLASS_LABEL;
|
|
break;
|
|
|
|
case TYPE_CODE_METADATA:
|
|
type->class = TYPE_CLASS_METADATA;
|
|
sm6->metadata_type = type;
|
|
break;
|
|
|
|
case TYPE_CODE_NUMENTRY:
|
|
continue;
|
|
|
|
case TYPE_CODE_POINTER:
|
|
if (!dxil_record_validate_operand_count(record, 1, 2, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
type->class = TYPE_CLASS_POINTER;
|
|
|
|
if ((type_id = record->operands[0]) >= type_count)
|
|
{
|
|
WARN("Invalid pointee type id %"PRIu64" for type %zu.\n", type_id, type_index);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
type->u.pointer.type = &sm6->types[type_id];
|
|
type->u.pointer.addr_space = (record->operand_count > 1) ? record->operands[1] : ADDRESS_SPACE_DEFAULT;
|
|
break;
|
|
|
|
case TYPE_CODE_STRUCT_ANON:
|
|
case TYPE_CODE_STRUCT_NAMED:
|
|
if (!dxil_record_validate_operand_min_count(record, 1, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
if (record->code == TYPE_CODE_STRUCT_NAMED && !struct_name)
|
|
{
|
|
WARN("Missing struct name before struct type %zu.\n", type_index);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
type->class = TYPE_CLASS_STRUCT;
|
|
|
|
count = record->operand_count - 1;
|
|
if (vkd3d_object_range_overflow(sizeof(type->u.struc), count, sizeof(type->u.struc->elem_types[0]))
|
|
|| !(type->u.struc = vkd3d_malloc(offsetof(struct sm6_struct_info, elem_types[count]))))
|
|
{
|
|
ERR("Failed to allocate struct element types.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
if (record->operands[0])
|
|
FIXME("Ignoring struct packed attribute.\n");
|
|
|
|
type->u.struc->elem_count = count;
|
|
for (j = 0; j < count; ++j)
|
|
{
|
|
if ((type_id = record->operands[j + 1]) >= type_count)
|
|
{
|
|
WARN("Invalid contained type id %"PRIu64" for type %zu.\n", type_id, type_index);
|
|
vkd3d_free(type->u.struc);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
type->u.struc->elem_types[j] = &sm6->types[type_id];
|
|
}
|
|
|
|
if (record->code == TYPE_CODE_STRUCT_ANON)
|
|
{
|
|
type->u.struc->name = NULL;
|
|
break;
|
|
}
|
|
|
|
if (!strcmp(struct_name, "dx.types.Handle"))
|
|
sm6->handle_type = type;
|
|
|
|
type->u.struc->name = struct_name;
|
|
struct_name = NULL;
|
|
break;
|
|
|
|
case TYPE_CODE_STRUCT_NAME:
|
|
if (!(struct_name = dxil_record_to_string(record, 0, sm6)))
|
|
{
|
|
ERR("Failed to allocate struct name.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
if (!struct_name[0])
|
|
WARN("Struct name is empty for type %zu.\n", type_index);
|
|
continue;
|
|
|
|
case TYPE_CODE_VOID:
|
|
dxil_record_validate_operand_max_count(record, 0, sm6);
|
|
type->class = TYPE_CLASS_VOID;
|
|
break;
|
|
|
|
default:
|
|
FIXME("Unhandled type %u at index %zu.\n", record->code, type_index);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
++sm6->type_count;
|
|
}
|
|
|
|
VKD3D_ASSERT(sm6->type_count == type_count);
|
|
|
|
if (struct_name)
|
|
{
|
|
WARN("Unused struct name %s.\n", struct_name);
|
|
vkd3d_free(struct_name);
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static inline bool sm6_type_is_void(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_VOID;
|
|
}
|
|
|
|
static inline bool sm6_type_is_integer(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_INTEGER;
|
|
}
|
|
|
|
static bool sm6_type_is_bool_i16_i32_i64(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_INTEGER && (type->u.width == 1 || type->u.width >= 16);
|
|
}
|
|
|
|
static bool sm6_type_is_i16_i32_i64(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_INTEGER && type->u.width >= 16;
|
|
}
|
|
|
|
static bool sm6_type_is_bool(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_INTEGER && type->u.width == 1;
|
|
}
|
|
|
|
static inline bool sm6_type_is_i8(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_INTEGER && type->u.width == 8;
|
|
}
|
|
|
|
static inline bool sm6_type_is_i32(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_INTEGER && type->u.width == 32;
|
|
}
|
|
|
|
static bool sm6_type_is_float(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_FLOAT && type->u.width == 32;
|
|
}
|
|
|
|
static bool sm6_type_is_f16_f32(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_FLOAT && (type->u.width == 16 || type->u.width == 32);
|
|
}
|
|
|
|
static bool sm6_type_is_double(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_FLOAT && type->u.width == 64;
|
|
}
|
|
|
|
static inline bool sm6_type_is_floating_point(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_FLOAT;
|
|
}
|
|
|
|
static bool sm6_type_is_scalar(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_INTEGER || type->class == TYPE_CLASS_FLOAT || type->class == TYPE_CLASS_POINTER;
|
|
}
|
|
|
|
static inline bool sm6_type_is_numeric(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_INTEGER || type->class == TYPE_CLASS_FLOAT;
|
|
}
|
|
|
|
static inline bool sm6_type_is_pointer(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_POINTER;
|
|
}
|
|
|
|
static bool sm6_type_is_aggregate(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_STRUCT || type->class == TYPE_CLASS_VECTOR || type->class == TYPE_CLASS_ARRAY;
|
|
}
|
|
|
|
static bool sm6_type_is_numeric_aggregate(const struct sm6_type *type)
|
|
{
|
|
unsigned int i;
|
|
|
|
switch (type->class)
|
|
{
|
|
case TYPE_CLASS_ARRAY:
|
|
case TYPE_CLASS_VECTOR:
|
|
return sm6_type_is_numeric(type->u.array.elem_type);
|
|
|
|
case TYPE_CLASS_STRUCT:
|
|
/* Do not handle nested structs. Support can be added if they show up. */
|
|
for (i = 0; i < type->u.struc->elem_count; ++i)
|
|
if (!sm6_type_is_numeric(type->u.struc->elem_types[i]))
|
|
return false;
|
|
return true;
|
|
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static bool sm6_type_is_array(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_ARRAY;
|
|
}
|
|
|
|
static inline bool sm6_type_is_struct(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_STRUCT;
|
|
}
|
|
|
|
static inline bool sm6_type_is_function(const struct sm6_type *type)
|
|
{
|
|
return type->class == TYPE_CLASS_FUNCTION;
|
|
}
|
|
|
|
static inline bool sm6_type_is_function_pointer(const struct sm6_type *type)
|
|
{
|
|
return sm6_type_is_pointer(type) && sm6_type_is_function(type->u.pointer.type);
|
|
}
|
|
|
|
static inline bool sm6_type_is_handle(const struct sm6_type *type)
|
|
{
|
|
return sm6_type_is_struct(type) && !strcmp(type->u.struc->name, "dx.types.Handle");
|
|
}
|
|
|
|
static inline const struct sm6_type *sm6_type_get_element_type(const struct sm6_type *type)
|
|
{
|
|
return (type->class == TYPE_CLASS_ARRAY || type->class == TYPE_CLASS_VECTOR) ? type->u.array.elem_type : type;
|
|
}
|
|
|
|
static const struct sm6_type *sm6_type_get_pointer_to_type(const struct sm6_type *type,
|
|
enum bitcode_address_space addr_space, struct sm6_parser *sm6)
|
|
{
|
|
size_t i, start = type - sm6->types;
|
|
const struct sm6_type *pointer_type;
|
|
|
|
/* DXC seems usually to place the pointer type immediately after its pointee. */
|
|
for (i = (start + 1) % sm6->type_count; i != start; i = (i + 1) % sm6->type_count)
|
|
{
|
|
pointer_type = &sm6->types[i];
|
|
if (sm6_type_is_pointer(pointer_type) && pointer_type->u.pointer.type == type
|
|
&& pointer_type->u.pointer.addr_space == addr_space)
|
|
return pointer_type;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static const struct sm6_type *sm6_type_get_cmpxchg_result_struct(struct sm6_parser *sm6)
|
|
{
|
|
const struct sm6_type *type;
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < sm6->type_count; ++i)
|
|
{
|
|
type = &sm6->types[i];
|
|
if (sm6_type_is_struct(type) && type->u.struc->elem_count == 2
|
|
&& sm6_type_is_i32(type->u.struc->elem_types[0])
|
|
&& sm6_type_is_bool(type->u.struc->elem_types[1]))
|
|
{
|
|
return type;
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/* Call for aggregate types only. */
|
|
static const struct sm6_type *sm6_type_get_element_type_at_index(const struct sm6_type *type, uint64_t elem_idx)
|
|
{
|
|
switch (type->class)
|
|
{
|
|
case TYPE_CLASS_ARRAY:
|
|
case TYPE_CLASS_VECTOR:
|
|
if (elem_idx >= type->u.array.count)
|
|
return NULL;
|
|
return type->u.array.elem_type;
|
|
|
|
case TYPE_CLASS_STRUCT:
|
|
if (elem_idx >= type->u.struc->elem_count)
|
|
return NULL;
|
|
return type->u.struc->elem_types[elem_idx];
|
|
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
/* Never returns null for elem_idx 0. */
|
|
static const struct sm6_type *sm6_type_get_scalar_type(const struct sm6_type *type, unsigned int elem_idx)
|
|
{
|
|
switch (type->class)
|
|
{
|
|
case TYPE_CLASS_ARRAY:
|
|
case TYPE_CLASS_VECTOR:
|
|
if (elem_idx >= type->u.array.count)
|
|
return NULL;
|
|
return sm6_type_get_scalar_type(type->u.array.elem_type, 0);
|
|
|
|
case TYPE_CLASS_POINTER:
|
|
return sm6_type_get_scalar_type(type->u.pointer.type, 0);
|
|
|
|
case TYPE_CLASS_STRUCT:
|
|
if (elem_idx >= type->u.struc->elem_count)
|
|
return NULL;
|
|
return sm6_type_get_scalar_type(type->u.struc->elem_types[elem_idx], 0);
|
|
|
|
default:
|
|
return type;
|
|
}
|
|
}
|
|
|
|
static unsigned int sm6_type_max_vector_size(const struct sm6_type *type)
|
|
{
|
|
return min((VKD3D_VEC4_SIZE * sizeof(uint32_t) * CHAR_BIT) / type->u.width, VKD3D_VEC4_SIZE);
|
|
}
|
|
|
|
static const struct sm6_type *sm6_parser_get_type(struct sm6_parser *sm6, uint64_t type_id)
|
|
{
|
|
if (type_id >= sm6->type_count)
|
|
{
|
|
WARN("Invalid type index %"PRIu64" at %zu.\n", type_id, sm6->value_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_TYPE_ID,
|
|
"DXIL type id %"PRIu64" is invalid.", type_id);
|
|
return NULL;
|
|
}
|
|
return &sm6->types[type_id];
|
|
}
|
|
|
|
static bool resource_kind_is_texture(enum dxil_resource_kind kind)
|
|
{
|
|
return kind >= RESOURCE_KIND_TEXTURE1D && kind <= RESOURCE_KIND_TEXTURECUBEARRAY;
|
|
}
|
|
|
|
static bool resource_kind_is_multisampled(enum dxil_resource_kind kind)
|
|
{
|
|
return kind == RESOURCE_KIND_TEXTURE2DMS || kind == RESOURCE_KIND_TEXTURE2DMSARRAY;
|
|
}
|
|
|
|
static int global_symbol_compare(const void *a, const void *b)
|
|
{
|
|
return vkd3d_u32_compare(((const struct sm6_symbol *)a)->id, ((const struct sm6_symbol *)b)->id);
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_symtab_init(struct sm6_parser *sm6)
|
|
{
|
|
const struct dxil_record *record;
|
|
const struct dxil_block *block;
|
|
struct sm6_symbol *symbol;
|
|
size_t i, count;
|
|
bool is_unique;
|
|
|
|
sm6->p.location.line = 0;
|
|
sm6->p.location.column = 0;
|
|
|
|
if (!(block = sm6_parser_get_level_one_block(sm6, VALUE_SYMTAB_BLOCK, &is_unique)))
|
|
{
|
|
/* There should always be at least one symbol: the name of the entry point function. */
|
|
WARN("No value symtab block found.\n");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
if (!is_unique)
|
|
FIXME("Ignoring extra value symtab block(s).\n");
|
|
|
|
sm6->p.location.line = block->id;
|
|
|
|
for (i = 0, count = 0; i < block->record_count; ++i)
|
|
count += block->records[i]->code == VST_CODE_ENTRY;
|
|
|
|
if (!(sm6->global_symbols = vkd3d_calloc(count, sizeof(*sm6->global_symbols))))
|
|
{
|
|
ERR("Failed to allocate global symbols.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
for (i = 0; i < block->record_count; ++i)
|
|
{
|
|
sm6->p.location.column = i;
|
|
record = block->records[i];
|
|
|
|
if (record->code != VST_CODE_ENTRY)
|
|
{
|
|
FIXME("Unhandled symtab code %u.\n", record->code);
|
|
continue;
|
|
}
|
|
if (!dxil_record_validate_operand_min_count(record, 1, sm6))
|
|
continue;
|
|
|
|
symbol = &sm6->global_symbols[sm6->global_symbol_count];
|
|
symbol->id = record->operands[0];
|
|
if (!(symbol->name = dxil_record_to_string(record, 1, sm6)))
|
|
{
|
|
ERR("Failed to allocate symbol name.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
++sm6->global_symbol_count;
|
|
}
|
|
|
|
sm6->p.location.column = block->record_count;
|
|
|
|
qsort(sm6->global_symbols, sm6->global_symbol_count, sizeof(*sm6->global_symbols), global_symbol_compare);
|
|
for (i = 1; i < sm6->global_symbol_count; ++i)
|
|
{
|
|
if (sm6->global_symbols[i].id == sm6->global_symbols[i - 1].id)
|
|
{
|
|
WARN("Invalid duplicate symbol id %u.\n", sm6->global_symbols[i].id);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static const char *sm6_parser_get_global_symbol_name(const struct sm6_parser *sm6, size_t id)
|
|
{
|
|
size_t i, start;
|
|
|
|
/* id == array index is normally true */
|
|
i = start = id % sm6->global_symbol_count;
|
|
do
|
|
{
|
|
if (sm6->global_symbols[i].id == id)
|
|
return sm6->global_symbols[i].name;
|
|
i = (i + 1) % sm6->global_symbol_count;
|
|
} while (i != start);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static unsigned int register_get_uint_value(const struct vkd3d_shader_register *reg)
|
|
{
|
|
if (!register_is_constant(reg) || (!data_type_is_integer(reg->data_type) && !data_type_is_bool(reg->data_type)))
|
|
return UINT_MAX;
|
|
|
|
if (reg->dimension == VSIR_DIMENSION_VEC4)
|
|
WARN("Returning vec4.x.\n");
|
|
|
|
if (reg->type == VKD3DSPR_IMMCONST64)
|
|
{
|
|
if (reg->u.immconst_u64[0] > UINT_MAX)
|
|
FIXME("Truncating 64-bit value.\n");
|
|
return reg->u.immconst_u64[0];
|
|
}
|
|
|
|
return reg->u.immconst_u32[0];
|
|
}
|
|
|
|
static uint64_t register_get_uint64_value(const struct vkd3d_shader_register *reg)
|
|
{
|
|
if (!register_is_constant(reg) || !data_type_is_integer(reg->data_type))
|
|
return UINT64_MAX;
|
|
|
|
if (reg->dimension == VSIR_DIMENSION_VEC4)
|
|
WARN("Returning vec4.x.\n");
|
|
|
|
return (reg->type == VKD3DSPR_IMMCONST64) ? reg->u.immconst_u64[0] : reg->u.immconst_u32[0];
|
|
}
|
|
|
|
static inline bool sm6_value_is_function_dcl(const struct sm6_value *value)
|
|
{
|
|
return value->value_type == VALUE_TYPE_FUNCTION;
|
|
}
|
|
|
|
static inline bool sm6_value_is_dx_intrinsic_dcl(const struct sm6_value *fn)
|
|
{
|
|
VKD3D_ASSERT(sm6_value_is_function_dcl(fn));
|
|
return fn->u.function.is_prototype && !strncmp(fn->u.function.name, "dx.op.", 6);
|
|
}
|
|
|
|
static inline struct sm6_value *sm6_parser_get_current_value(const struct sm6_parser *sm6)
|
|
{
|
|
VKD3D_ASSERT(sm6->value_count < sm6->value_capacity);
|
|
return &sm6->values[sm6->value_count];
|
|
}
|
|
|
|
static inline bool sm6_value_is_register(const struct sm6_value *value)
|
|
{
|
|
return value->value_type == VALUE_TYPE_REG;
|
|
}
|
|
|
|
static bool sm6_value_is_handle(const struct sm6_value *value)
|
|
{
|
|
return value->value_type == VALUE_TYPE_HANDLE;
|
|
}
|
|
|
|
static inline bool sm6_value_is_constant(const struct sm6_value *value)
|
|
{
|
|
return sm6_value_is_register(value) && register_is_constant(&value->u.reg);
|
|
}
|
|
|
|
static bool sm6_value_is_constant_zero(const struct sm6_value *value)
|
|
{
|
|
/* Constant vectors do not occur. */
|
|
return sm6_value_is_register(value) && register_is_scalar_constant_zero(&value->u.reg);
|
|
}
|
|
|
|
static inline bool sm6_value_is_undef(const struct sm6_value *value)
|
|
{
|
|
return sm6_value_is_register(value) && value->u.reg.type == VKD3DSPR_UNDEF;
|
|
}
|
|
|
|
static bool sm6_value_vector_is_constant_or_undef(const struct sm6_value **values, unsigned int count)
|
|
{
|
|
unsigned int i;
|
|
for (i = 0; i < count; ++i)
|
|
if (!sm6_value_is_constant(values[i]) && !sm6_value_is_undef(values[i]))
|
|
return false;
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_value_is_icb(const struct sm6_value *value)
|
|
{
|
|
return value->value_type == VALUE_TYPE_ICB;
|
|
}
|
|
|
|
static bool sm6_value_is_ssa(const struct sm6_value *value)
|
|
{
|
|
return sm6_value_is_register(value) && register_is_ssa(&value->u.reg);
|
|
}
|
|
|
|
static bool sm6_value_is_numeric_array(const struct sm6_value *value)
|
|
{
|
|
return sm6_value_is_register(value) && register_is_numeric_array(&value->u.reg);
|
|
}
|
|
|
|
static inline unsigned int sm6_value_get_constant_uint(const struct sm6_value *value)
|
|
{
|
|
if (!sm6_value_is_constant(value))
|
|
return UINT_MAX;
|
|
return register_get_uint_value(&value->u.reg);
|
|
}
|
|
|
|
static uint64_t sm6_value_get_constant_uint64(const struct sm6_value *value)
|
|
{
|
|
if (!sm6_value_is_constant(value))
|
|
return UINT64_MAX;
|
|
return register_get_uint64_value(&value->u.reg);
|
|
}
|
|
|
|
static unsigned int sm6_parser_alloc_ssa_id(struct sm6_parser *sm6)
|
|
{
|
|
return sm6->ssa_next_id++;
|
|
}
|
|
|
|
static void instruction_init_with_resource(struct vkd3d_shader_instruction *ins,
|
|
enum vkd3d_shader_opcode handler_idx, const struct sm6_value *resource, struct sm6_parser *sm6)
|
|
{
|
|
vsir_instruction_init(ins, &sm6->p.location, handler_idx);
|
|
ins->resource_type = resource->u.handle.d->resource_type;
|
|
ins->raw = resource->u.handle.d->kind == RESOURCE_KIND_RAWBUFFER;
|
|
ins->structured = resource->u.handle.d->kind == RESOURCE_KIND_STRUCTUREDBUFFER;
|
|
}
|
|
|
|
static struct vkd3d_shader_src_param *instruction_src_params_alloc(struct vkd3d_shader_instruction *ins,
|
|
unsigned int count, struct sm6_parser *sm6)
|
|
{
|
|
struct vkd3d_shader_src_param *params;
|
|
|
|
if (!(params = vsir_program_get_src_params(sm6->p.program, count)))
|
|
{
|
|
ERR("Failed to allocate src params.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating instruction src parameters.");
|
|
return NULL;
|
|
}
|
|
ins->src = params;
|
|
ins->src_count = count;
|
|
return params;
|
|
}
|
|
|
|
static struct vkd3d_shader_dst_param *instruction_dst_params_alloc(struct vkd3d_shader_instruction *ins,
|
|
unsigned int count, struct sm6_parser *sm6)
|
|
{
|
|
struct vkd3d_shader_dst_param *params;
|
|
|
|
if (!(params = vsir_program_get_dst_params(sm6->p.program, count)))
|
|
{
|
|
ERR("Failed to allocate dst params.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating instruction dst parameters.");
|
|
return NULL;
|
|
}
|
|
ins->dst = params;
|
|
ins->dst_count = count;
|
|
return params;
|
|
}
|
|
|
|
static void register_init_with_id(struct vkd3d_shader_register *reg,
|
|
enum vkd3d_shader_register_type reg_type, enum vkd3d_data_type data_type, unsigned int id)
|
|
{
|
|
vsir_register_init(reg, reg_type, data_type, 1);
|
|
reg->idx[0].offset = id;
|
|
}
|
|
|
|
static enum vkd3d_data_type vkd3d_data_type_from_sm6_type(const struct sm6_type *type)
|
|
{
|
|
if (type->class == TYPE_CLASS_INTEGER)
|
|
{
|
|
switch (type->u.width)
|
|
{
|
|
case 1:
|
|
return VKD3D_DATA_BOOL;
|
|
case 8:
|
|
return VKD3D_DATA_UINT8;
|
|
case 16:
|
|
return VKD3D_DATA_UINT16;
|
|
case 32:
|
|
return VKD3D_DATA_UINT;
|
|
case 64:
|
|
return VKD3D_DATA_UINT64;
|
|
default:
|
|
FIXME("Unhandled width %u.\n", type->u.width);
|
|
return VKD3D_DATA_UINT;
|
|
}
|
|
}
|
|
else if (type->class == TYPE_CLASS_FLOAT)
|
|
{
|
|
switch (type->u.width)
|
|
{
|
|
case 16:
|
|
return VKD3D_DATA_HALF;
|
|
case 32:
|
|
return VKD3D_DATA_FLOAT;
|
|
case 64:
|
|
return VKD3D_DATA_DOUBLE;
|
|
default:
|
|
FIXME("Unhandled width %u.\n", type->u.width);
|
|
return VKD3D_DATA_FLOAT;
|
|
}
|
|
}
|
|
|
|
FIXME("Unhandled type %u.\n", type->class);
|
|
return VKD3D_DATA_UINT;
|
|
}
|
|
|
|
static void register_init_ssa_vector(struct vkd3d_shader_register *reg, const struct sm6_type *type,
|
|
unsigned int component_count, struct sm6_value *value, struct sm6_parser *sm6)
|
|
{
|
|
enum vkd3d_data_type data_type;
|
|
unsigned int id;
|
|
|
|
if (value && register_is_ssa(&value->u.reg) && value->u.reg.idx[0].offset)
|
|
{
|
|
id = value->u.reg.idx[0].offset;
|
|
TRACE("Using forward-allocated id %u.\n", id);
|
|
}
|
|
else
|
|
{
|
|
id = sm6_parser_alloc_ssa_id(sm6);
|
|
}
|
|
data_type = vkd3d_data_type_from_sm6_type(sm6_type_get_scalar_type(type, 0));
|
|
register_init_with_id(reg, VKD3DSPR_SSA, data_type, id);
|
|
reg->dimension = component_count > 1 ? VSIR_DIMENSION_VEC4 : VSIR_DIMENSION_SCALAR;
|
|
}
|
|
|
|
static void register_init_ssa_scalar(struct vkd3d_shader_register *reg, const struct sm6_type *type,
|
|
struct sm6_value *value, struct sm6_parser *sm6)
|
|
{
|
|
register_init_ssa_vector(reg, sm6_type_get_scalar_type(type, 0), 1, value, sm6);
|
|
}
|
|
|
|
static void register_make_constant_uint(struct vkd3d_shader_register *reg, unsigned int value)
|
|
{
|
|
vsir_register_init(reg, VKD3DSPR_IMMCONST, VKD3D_DATA_UINT, 0);
|
|
reg->u.immconst_u32[0] = value;
|
|
}
|
|
|
|
static void dst_param_init(struct vkd3d_shader_dst_param *param)
|
|
{
|
|
param->write_mask = VKD3DSP_WRITEMASK_0;
|
|
param->modifiers = 0;
|
|
param->shift = 0;
|
|
}
|
|
|
|
static void dst_param_init_with_mask(struct vkd3d_shader_dst_param *param, unsigned int mask)
|
|
{
|
|
param->write_mask = mask;
|
|
param->modifiers = 0;
|
|
param->shift = 0;
|
|
}
|
|
|
|
static inline void dst_param_init_scalar(struct vkd3d_shader_dst_param *param, unsigned int component_idx)
|
|
{
|
|
param->write_mask = 1u << component_idx;
|
|
param->modifiers = 0;
|
|
param->shift = 0;
|
|
}
|
|
|
|
static void dst_param_init_vector(struct vkd3d_shader_dst_param *param, unsigned int component_count)
|
|
{
|
|
param->write_mask = (1u << component_count) - 1;
|
|
param->modifiers = 0;
|
|
param->shift = 0;
|
|
}
|
|
|
|
static void dst_param_init_ssa_scalar(struct vkd3d_shader_dst_param *param, const struct sm6_type *type,
|
|
struct sm6_value *value, struct sm6_parser *sm6)
|
|
{
|
|
dst_param_init(param);
|
|
register_init_ssa_scalar(¶m->reg, type, value, sm6);
|
|
}
|
|
|
|
static inline void src_param_init(struct vkd3d_shader_src_param *param)
|
|
{
|
|
param->swizzle = VKD3D_SHADER_SWIZZLE(X, X, X, X);
|
|
param->modifiers = VKD3DSPSM_NONE;
|
|
}
|
|
|
|
static void src_param_init_scalar(struct vkd3d_shader_src_param *param, unsigned int component_idx)
|
|
{
|
|
param->swizzle = vkd3d_shader_create_swizzle(component_idx, component_idx, component_idx, component_idx);
|
|
if (data_type_is_64_bit(param->reg.data_type))
|
|
param->swizzle &= VKD3D_SHADER_SWIZZLE_64_MASK;
|
|
param->modifiers = VKD3DSPSM_NONE;
|
|
}
|
|
|
|
static void src_param_init_vector(struct vkd3d_shader_src_param *param, unsigned int component_count)
|
|
{
|
|
param->swizzle = VKD3D_SHADER_NO_SWIZZLE & ((1ull << VKD3D_SHADER_SWIZZLE_SHIFT(component_count)) - 1);
|
|
param->modifiers = VKD3DSPSM_NONE;
|
|
}
|
|
|
|
static void src_param_init_from_value(struct vkd3d_shader_src_param *param, const struct sm6_value *src)
|
|
{
|
|
src_param_init(param);
|
|
param->reg = src->u.reg;
|
|
}
|
|
|
|
static void src_param_init_vector_from_reg(struct vkd3d_shader_src_param *param,
|
|
const struct vkd3d_shader_register *reg)
|
|
{
|
|
param->swizzle = (reg->dimension == VSIR_DIMENSION_VEC4) ? VKD3D_SHADER_NO_SWIZZLE : VKD3D_SHADER_SWIZZLE(X, X, X, X);
|
|
param->modifiers = VKD3DSPSM_NONE;
|
|
param->reg = *reg;
|
|
}
|
|
|
|
static void src_param_make_constant_uint(struct vkd3d_shader_src_param *param, unsigned int value)
|
|
{
|
|
src_param_init(param);
|
|
register_make_constant_uint(¶m->reg, value);
|
|
}
|
|
|
|
static void register_index_address_init(struct vkd3d_shader_register_index *idx, const struct sm6_value *address,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
if (sm6_value_is_constant(address))
|
|
{
|
|
idx->offset = sm6_value_get_constant_uint(address);
|
|
idx->rel_addr = NULL;
|
|
}
|
|
else if (sm6_value_is_undef(address))
|
|
{
|
|
idx->offset = 0;
|
|
idx->rel_addr = NULL;
|
|
}
|
|
else
|
|
{
|
|
struct vkd3d_shader_src_param *rel_addr = vsir_program_get_src_params(sm6->p.program, 1);
|
|
if (rel_addr)
|
|
src_param_init_from_value(rel_addr, address);
|
|
idx->offset = 0;
|
|
idx->rel_addr = rel_addr;
|
|
}
|
|
}
|
|
|
|
static bool instruction_dst_param_init_ssa_scalar(struct vkd3d_shader_instruction *ins, struct sm6_parser *sm6)
|
|
{
|
|
struct sm6_value *dst = sm6_parser_get_current_value(sm6);
|
|
struct vkd3d_shader_dst_param *param;
|
|
|
|
if (!(param = instruction_dst_params_alloc(ins, 1, sm6)))
|
|
return false;
|
|
|
|
dst_param_init_ssa_scalar(param, dst->type, dst, sm6);
|
|
param->write_mask = VKD3DSP_WRITEMASK_0;
|
|
dst->u.reg = param->reg;
|
|
return true;
|
|
}
|
|
|
|
static void instruction_dst_param_init_ssa_vector(struct vkd3d_shader_instruction *ins,
|
|
unsigned int component_count, struct sm6_parser *sm6)
|
|
{
|
|
struct vkd3d_shader_dst_param *param = instruction_dst_params_alloc(ins, 1, sm6);
|
|
struct sm6_value *dst = sm6_parser_get_current_value(sm6);
|
|
|
|
dst_param_init_vector(param, component_count);
|
|
register_init_ssa_vector(¶m->reg, sm6_type_get_scalar_type(dst->type, 0), component_count, dst, sm6);
|
|
dst->u.reg = param->reg;
|
|
}
|
|
|
|
static bool instruction_dst_param_init_temp_vector(struct vkd3d_shader_instruction *ins, struct sm6_parser *sm6)
|
|
{
|
|
struct sm6_value *dst = sm6_parser_get_current_value(sm6);
|
|
struct vkd3d_shader_dst_param *param;
|
|
|
|
if (!(param = instruction_dst_params_alloc(ins, 1, sm6)))
|
|
return false;
|
|
|
|
vsir_dst_param_init(param, VKD3DSPR_TEMP, vkd3d_data_type_from_sm6_type(sm6_type_get_scalar_type(dst->type, 0)), 1);
|
|
param->write_mask = VKD3DSP_WRITEMASK_ALL;
|
|
param->reg.idx[0].offset = 0;
|
|
param->reg.dimension = VSIR_DIMENSION_VEC4;
|
|
dst->u.reg = param->reg;
|
|
|
|
return true;
|
|
}
|
|
|
|
/* Recurse through the block tree while maintaining a current value count. The current
|
|
* count is the sum of the global count plus all declarations within the current function.
|
|
* Store into value_capacity the highest count seen. */
|
|
static size_t sm6_parser_compute_max_value_count(struct sm6_parser *sm6,
|
|
const struct dxil_block *block, size_t value_count)
|
|
{
|
|
size_t i, old_value_count = value_count;
|
|
|
|
if (block->id == MODULE_BLOCK)
|
|
value_count = size_add_with_overflow_check(value_count, dxil_block_compute_module_decl_count(block));
|
|
|
|
for (i = 0; i < block->child_block_count; ++i)
|
|
value_count = sm6_parser_compute_max_value_count(sm6, block->child_blocks[i], value_count);
|
|
|
|
switch (block->id)
|
|
{
|
|
case CONSTANTS_BLOCK:
|
|
/* Function local constants are contained in a child block of the function block. */
|
|
value_count = size_add_with_overflow_check(value_count, dxil_block_compute_constants_count(block));
|
|
break;
|
|
case FUNCTION_BLOCK:
|
|
/* A function must start with a block count, which emits no value. This formula is likely to
|
|
* overestimate the value count somewhat, but this should be no problem. */
|
|
value_count = size_add_with_overflow_check(value_count, max(block->record_count, 1u) - 1);
|
|
sm6->value_capacity = max(sm6->value_capacity, value_count);
|
|
sm6->functions[sm6->function_count++].value_count = value_count;
|
|
/* The value count returns to its previous value after handling a function. */
|
|
if (value_count < SIZE_MAX)
|
|
value_count = old_value_count;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return value_count;
|
|
}
|
|
|
|
static size_t sm6_parser_get_value_index(struct sm6_parser *sm6, uint64_t idx)
|
|
{
|
|
size_t i;
|
|
|
|
/* The value relative index is 32 bits. */
|
|
if (idx > UINT32_MAX)
|
|
WARN("Ignoring upper 32 bits of relative index.\n");
|
|
i = (uint32_t)sm6->value_count - (uint32_t)idx;
|
|
|
|
/* This may underflow to produce a forward reference, but it must not exceed the final value count. */
|
|
if (i >= sm6->cur_max_value)
|
|
{
|
|
WARN("Invalid value index %"PRIx64" at %zu.\n", idx, sm6->value_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid value relative index %u.", (unsigned int)idx);
|
|
return SIZE_MAX;
|
|
}
|
|
if (i == sm6->value_count)
|
|
{
|
|
WARN("Invalid value self-reference at %zu.\n", sm6->value_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND, "Invalid value self-reference.");
|
|
return SIZE_MAX;
|
|
}
|
|
|
|
return i;
|
|
}
|
|
|
|
static bool sm6_value_validate_is_register(const struct sm6_value *value, struct sm6_parser *sm6)
|
|
{
|
|
if (!sm6_value_is_register(value))
|
|
{
|
|
WARN("Operand of type %u is not a register.\n", value->value_type);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"A register operand passed to a DXIL instruction is not a register.");
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_value_validate_is_handle(const struct sm6_value *value, struct sm6_parser *sm6)
|
|
{
|
|
if (!sm6_value_is_handle(value))
|
|
{
|
|
WARN("Handle parameter of type %u is not a handle.\n", value->value_type);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCE_HANDLE,
|
|
"A handle parameter passed to a DX intrinsic function is not a handle.");
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_value_validate_is_texture_handle(const struct sm6_value *value, enum dx_intrinsic_opcode op,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
enum dxil_resource_kind kind;
|
|
|
|
if (!sm6_value_validate_is_handle(value, sm6))
|
|
return false;
|
|
|
|
kind = value->u.handle.d->kind;
|
|
if (kind < RESOURCE_KIND_TEXTURE1D || kind > RESOURCE_KIND_TEXTURECUBEARRAY)
|
|
{
|
|
WARN("Resource kind %u for op %u is not a texture.\n", kind, op);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCE_HANDLE,
|
|
"Resource kind %u for texture operation %u is not a texture.", kind, op);
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_value_validate_is_texture_2dms_handle(const struct sm6_value *value, enum dx_intrinsic_opcode op,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
enum dxil_resource_kind kind;
|
|
|
|
if (!sm6_value_validate_is_handle(value, sm6))
|
|
return false;
|
|
|
|
kind = value->u.handle.d->kind;
|
|
if (!resource_kind_is_multisampled(kind))
|
|
{
|
|
WARN("Resource kind %u for op %u is not a 2DMS texture.\n", kind, op);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCE_HANDLE,
|
|
"Resource kind %u for texture operation %u is not a 2DMS texture.", kind, op);
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_value_validate_is_sampler_handle(const struct sm6_value *value, enum dx_intrinsic_opcode op,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
enum dxil_resource_kind kind;
|
|
|
|
if (!sm6_value_validate_is_handle(value, sm6))
|
|
return false;
|
|
|
|
kind = value->u.handle.d->kind;
|
|
if (kind != RESOURCE_KIND_SAMPLER)
|
|
{
|
|
WARN("Resource kind %u for op %u is not a sampler.\n", kind, op);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCE_HANDLE,
|
|
"Resource kind %u for sample operation %u is not a sampler.", kind, op);
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_value_validate_is_pointer(const struct sm6_value *value, struct sm6_parser *sm6)
|
|
{
|
|
if (!sm6_type_is_pointer(value->type))
|
|
{
|
|
WARN("Operand result type class %u is not a pointer.\n", value->type->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"A pointer operand passed to a DXIL instruction is not a pointer.");
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_value_validate_is_backward_ref(const struct sm6_value *value, struct sm6_parser *sm6)
|
|
{
|
|
if (!value->is_back_ref)
|
|
{
|
|
FIXME("Forward-referenced pointers are not supported.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Forward-referenced pointer declarations are not supported.");
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_value_validate_is_numeric(const struct sm6_value *value, struct sm6_parser *sm6)
|
|
{
|
|
if (!sm6_type_is_numeric(value->type))
|
|
{
|
|
WARN("Operand result type class %u is not numeric.\n", value->type->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"A numeric operand passed to a DXIL instruction is not numeric.");
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_value_validate_is_bool(const struct sm6_value *value, struct sm6_parser *sm6)
|
|
{
|
|
const struct sm6_type *type = value->type;
|
|
if (!sm6_type_is_bool(type))
|
|
{
|
|
WARN("Operand of type class %u is not bool.\n", type->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"A bool operand of type class %u passed to a DXIL instruction is not a bool.", type->class);
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_value_validate_is_pointer_to_i32(const struct sm6_value *value, struct sm6_parser *sm6)
|
|
{
|
|
if (!sm6_type_is_pointer(value->type) || !sm6_type_is_i32(value->type->u.pointer.type))
|
|
{
|
|
WARN("Operand result type %u is not a pointer to i32.\n", value->type->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"An int32 pointer operand passed to a DXIL instruction is not an int32 pointer.");
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_value_validate_is_i32(const struct sm6_value *value, struct sm6_parser *sm6)
|
|
{
|
|
if (!sm6_type_is_i32(value->type))
|
|
{
|
|
WARN("Operand result type %u is not i32.\n", value->type->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"An int32 operand passed to a DXIL instruction is not an int32.");
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static const struct sm6_value *sm6_parser_get_value_safe(struct sm6_parser *sm6, unsigned int idx)
|
|
{
|
|
if (idx < sm6->value_count)
|
|
return &sm6->values[idx];
|
|
|
|
WARN("Invalid value index %u.\n", idx);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid value index %u.", idx);
|
|
return NULL;
|
|
}
|
|
|
|
static size_t sm6_parser_get_value_idx_by_ref(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
const struct sm6_type *fwd_type, unsigned int *rec_idx)
|
|
{
|
|
struct sm6_value *value;
|
|
unsigned int idx;
|
|
uint64_t val_ref;
|
|
size_t operand;
|
|
|
|
idx = *rec_idx;
|
|
if (!dxil_record_validate_operand_min_count(record, idx + 1, sm6))
|
|
return SIZE_MAX;
|
|
val_ref = record->operands[idx++];
|
|
|
|
operand = sm6_parser_get_value_index(sm6, val_ref);
|
|
if (operand == SIZE_MAX)
|
|
return SIZE_MAX;
|
|
|
|
if (operand >= sm6->value_count && !fwd_type)
|
|
{
|
|
/* Forward references are followed by a type id unless an earlier operand set the type,
|
|
* or it is contained in a function declaration. */
|
|
if (!dxil_record_validate_operand_min_count(record, idx + 1, sm6))
|
|
return SIZE_MAX;
|
|
if (!(fwd_type = sm6_parser_get_type(sm6, record->operands[idx++])))
|
|
return SIZE_MAX;
|
|
}
|
|
*rec_idx = idx;
|
|
|
|
if (fwd_type)
|
|
{
|
|
value = &sm6->values[operand];
|
|
if (value->type)
|
|
{
|
|
if (value->type != fwd_type)
|
|
{
|
|
WARN("Value already has a mismatching type.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"The type of a source value does not match the predefined type.");
|
|
}
|
|
}
|
|
else
|
|
{
|
|
value->type = fwd_type;
|
|
value->value_type = VALUE_TYPE_REG;
|
|
register_init_with_id(&value->u.reg, VKD3DSPR_SSA, vkd3d_data_type_from_sm6_type(
|
|
sm6_type_get_scalar_type(fwd_type, 0)), sm6_parser_alloc_ssa_id(sm6));
|
|
value->u.reg.dimension = sm6_type_is_scalar(fwd_type) ? VSIR_DIMENSION_SCALAR
|
|
: VSIR_DIMENSION_VEC4;
|
|
}
|
|
}
|
|
|
|
return operand;
|
|
}
|
|
|
|
static const struct sm6_value *sm6_parser_get_value_by_ref(struct sm6_parser *sm6,
|
|
const struct dxil_record *record, const struct sm6_type *type, unsigned int *rec_idx)
|
|
{
|
|
size_t operand = sm6_parser_get_value_idx_by_ref(sm6, record, type, rec_idx);
|
|
return operand == SIZE_MAX ? NULL : &sm6->values[operand];
|
|
}
|
|
|
|
static bool sm6_parser_declare_function(struct sm6_parser *sm6, const struct dxil_record *record)
|
|
{
|
|
const struct sm6_type *type, *ret_type;
|
|
const unsigned int max_count = 15;
|
|
struct sm6_value *fn;
|
|
unsigned int i, j;
|
|
|
|
if (!dxil_record_validate_operand_count(record, 8, max_count, sm6))
|
|
return false;
|
|
|
|
fn = sm6_parser_get_current_value(sm6);
|
|
fn->value_type = VALUE_TYPE_FUNCTION;
|
|
if (!(fn->u.function.name = sm6_parser_get_global_symbol_name(sm6, sm6->value_count)))
|
|
{
|
|
WARN("Missing symbol name for function %zu.\n", sm6->value_count);
|
|
fn->u.function.name = "";
|
|
}
|
|
|
|
if (!(type = sm6_parser_get_type(sm6, record->operands[0])))
|
|
return false;
|
|
if (!sm6_type_is_function(type))
|
|
{
|
|
WARN("Type is not a function.\n");
|
|
return false;
|
|
}
|
|
ret_type = type->u.function->ret_type;
|
|
|
|
if (!(fn->type = sm6_type_get_pointer_to_type(type, ADDRESS_SPACE_DEFAULT, sm6)))
|
|
{
|
|
WARN("Failed to get pointer type for type %u.\n", type->class);
|
|
return false;
|
|
}
|
|
|
|
if (record->operands[1])
|
|
WARN("Ignoring calling convention %#"PRIx64".\n", record->operands[1]);
|
|
|
|
fn->u.function.is_prototype = !!record->operands[2];
|
|
|
|
if (record->operands[3])
|
|
WARN("Ignoring linkage %#"PRIx64".\n", record->operands[3]);
|
|
|
|
if (record->operands[4] > UINT_MAX)
|
|
WARN("Invalid attributes id %#"PRIx64".\n", record->operands[4]);
|
|
/* 1-based index. */
|
|
if ((fn->u.function.attribs_id = record->operands[4]))
|
|
TRACE("Ignoring function attributes.\n");
|
|
|
|
/* These always seem to be zero. */
|
|
for (i = 5, j = 0; i < min(record->operand_count, max_count); ++i)
|
|
j += !!record->operands[i];
|
|
if (j)
|
|
WARN("Ignoring %u operands.\n", j);
|
|
|
|
if (sm6_value_is_dx_intrinsic_dcl(fn) && !sm6_type_is_void(ret_type) && !sm6_type_is_numeric(ret_type)
|
|
&& !sm6_type_is_numeric_aggregate(ret_type) && !sm6_type_is_handle(ret_type))
|
|
{
|
|
WARN("Unexpected return type for dx intrinsic function '%s'.\n", fn->u.function.name);
|
|
}
|
|
|
|
++sm6->value_count;
|
|
|
|
return true;
|
|
}
|
|
|
|
static inline uint64_t decode_rotated_signed_value(uint64_t value)
|
|
{
|
|
if (value != 1)
|
|
{
|
|
bool neg = value & 1;
|
|
value >>= 1;
|
|
return neg ? -value : value;
|
|
}
|
|
return value << 63;
|
|
}
|
|
|
|
static float bitcast_uint_to_float(unsigned int value)
|
|
{
|
|
union
|
|
{
|
|
uint32_t uint32_value;
|
|
float float_value;
|
|
} u;
|
|
|
|
u.uint32_value = value;
|
|
return u.float_value;
|
|
}
|
|
|
|
static inline double bitcast_uint64_to_double(uint64_t value)
|
|
{
|
|
union
|
|
{
|
|
uint64_t uint64_value;
|
|
double double_value;
|
|
} u;
|
|
|
|
u.uint64_value = value;
|
|
return u.double_value;
|
|
}
|
|
|
|
static float register_get_float_value(const struct vkd3d_shader_register *reg)
|
|
{
|
|
if (!register_is_constant(reg) || !data_type_is_floating_point(reg->data_type))
|
|
return 0.0;
|
|
|
|
if (reg->dimension == VSIR_DIMENSION_VEC4)
|
|
WARN("Returning vec4.x.\n");
|
|
|
|
if (reg->type == VKD3DSPR_IMMCONST64)
|
|
{
|
|
WARN("Truncating double to float.\n");
|
|
return bitcast_uint64_to_double(reg->u.immconst_u64[0]);
|
|
}
|
|
|
|
return bitcast_uint_to_float(reg->u.immconst_u32[0]);
|
|
}
|
|
|
|
static enum vkd3d_result value_allocate_constant_array(struct sm6_value *dst, const struct sm6_type *type,
|
|
const uint64_t *operands, struct sm6_parser *sm6)
|
|
{
|
|
struct vkd3d_shader_immediate_constant_buffer *icb;
|
|
const struct sm6_type *elem_type;
|
|
unsigned int i, size, count;
|
|
|
|
elem_type = type->u.array.elem_type;
|
|
/* Multidimensional arrays are emitted in flattened form. */
|
|
if (elem_type->class != TYPE_CLASS_INTEGER && elem_type->class != TYPE_CLASS_FLOAT)
|
|
{
|
|
FIXME("Unhandled element type %u for data array.\n", elem_type->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"The element data type for an immediate constant buffer is not scalar integer or floating point.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
/* Arrays of bool are not used in DXIL. dxc will emit an array of int32 instead if necessary. */
|
|
if (!(size = elem_type->u.width / CHAR_BIT))
|
|
{
|
|
WARN("Invalid data type width %u.\n", elem_type->u.width);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"An immediate constant buffer is declared with boolean elements.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
size = max(size, sizeof(icb->data[0]));
|
|
count = operands ? type->u.array.count * size / sizeof(icb->data[0]) : 0;
|
|
|
|
if (!(icb = vkd3d_malloc(offsetof(struct vkd3d_shader_immediate_constant_buffer, data[count]))))
|
|
{
|
|
ERR("Failed to allocate buffer, count %u.\n", count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating an immediate constant buffer of count %u.", count);
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
if (!shader_instruction_array_add_icb(&sm6->p.program->instructions, icb))
|
|
{
|
|
ERR("Failed to store icb object.\n");
|
|
vkd3d_free(icb);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory storing an immediate constant buffer object.");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
dst->value_type = VALUE_TYPE_ICB;
|
|
dst->u.icb = icb;
|
|
|
|
icb->register_idx = sm6->icb_count++;
|
|
icb->data_type = vkd3d_data_type_from_sm6_type(elem_type);
|
|
icb->element_count = type->u.array.count;
|
|
icb->component_count = 1;
|
|
icb->is_null = !operands;
|
|
|
|
if (!operands)
|
|
return VKD3D_OK;
|
|
|
|
count = type->u.array.count;
|
|
if (size > sizeof(icb->data[0]))
|
|
{
|
|
uint64_t *data = (uint64_t *)icb->data;
|
|
for (i = 0; i < count; ++i)
|
|
data[i] = operands[i];
|
|
}
|
|
else
|
|
{
|
|
for (i = 0; i < count; ++i)
|
|
icb->data[i] = operands[i];
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_init_constexpr_gep(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct sm6_value *dst)
|
|
{
|
|
const struct sm6_type *elem_type, *pointee_type, *gep_type, *ptr_type;
|
|
struct sm6_value *operands[3];
|
|
unsigned int i, j, offset;
|
|
uint64_t value;
|
|
|
|
i = 0;
|
|
pointee_type = (record->operand_count & 1) ? sm6_parser_get_type(sm6, record->operands[i++]) : NULL;
|
|
|
|
if (!dxil_record_validate_operand_count(record, i + 6, i + 6, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
for (j = 0; i < record->operand_count; i += 2, ++j)
|
|
{
|
|
if (!(elem_type = sm6_parser_get_type(sm6, record->operands[i])))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
if ((value = record->operands[i + 1]) >= sm6->cur_max_value)
|
|
{
|
|
WARN("Invalid value index %"PRIu64".\n", value);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid value index %"PRIu64".", value);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
else if (value == sm6->value_count)
|
|
{
|
|
WARN("Invalid value self-reference at %"PRIu64".\n", value);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid value self-reference for a constexpr GEP.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
operands[j] = &sm6->values[value];
|
|
if (value > sm6->value_count)
|
|
{
|
|
operands[j]->type = elem_type;
|
|
}
|
|
else if (operands[j]->type != elem_type)
|
|
{
|
|
WARN("Type mismatch.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"Type mismatch in constexpr GEP elements.");
|
|
}
|
|
}
|
|
|
|
if (operands[0]->u.reg.idx_count > 1)
|
|
{
|
|
WARN("Unsupported stacked GEP.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"A GEP instruction on the result of a previous GEP is unsupported.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (!sm6_value_is_constant_zero(operands[1]))
|
|
{
|
|
WARN("Expected constant zero.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"The pointer dereference index for a constexpr GEP instruction is not constant zero.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
if (!sm6_value_is_constant(operands[2]) || !sm6_type_is_integer(operands[2]->type))
|
|
{
|
|
WARN("Element index is not constant int.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"A constexpr GEP element index is not a constant integer.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
dst->structure_stride = operands[0]->structure_stride;
|
|
|
|
ptr_type = operands[0]->type;
|
|
if (!sm6_type_is_pointer(ptr_type))
|
|
{
|
|
WARN("Constexpr GEP base value is not a pointer.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"A constexpr GEP base value is not a pointer.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (!pointee_type)
|
|
{
|
|
pointee_type = ptr_type->u.pointer.type;
|
|
}
|
|
else if (pointee_type != ptr_type->u.pointer.type)
|
|
{
|
|
WARN("Explicit pointee type mismatch.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"Explicit pointee type for constexpr GEP does not match the element type.");
|
|
}
|
|
|
|
offset = sm6_value_get_constant_uint(operands[2]);
|
|
if (!(gep_type = sm6_type_get_element_type_at_index(pointee_type, offset)))
|
|
{
|
|
WARN("Failed to get element type.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Failed to get the element type of a constexpr GEP.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (!(dst->type = sm6_type_get_pointer_to_type(gep_type, ptr_type->u.pointer.addr_space, sm6)))
|
|
{
|
|
WARN("Failed to get pointer type for type %u.\n", gep_type->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_MODULE,
|
|
"Module does not define a pointer type for a constexpr GEP result.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
dst->u.reg = operands[0]->u.reg;
|
|
dst->u.reg.idx[1].offset = offset;
|
|
dst->u.reg.idx[1].is_in_bounds = record->code == CST_CODE_CE_INBOUNDS_GEP;
|
|
dst->u.reg.idx_count = 2;
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_constants_init(struct sm6_parser *sm6, const struct dxil_block *block)
|
|
{
|
|
enum vkd3d_shader_register_type reg_type = VKD3DSPR_INVALID;
|
|
const struct sm6_type *type, *elem_type, *ptr_type;
|
|
size_t i, base_value_idx, value_idx;
|
|
enum vkd3d_data_type reg_data_type;
|
|
const struct dxil_record *record;
|
|
const struct sm6_value *src;
|
|
enum vkd3d_result ret;
|
|
struct sm6_value *dst;
|
|
uint64_t value;
|
|
|
|
for (i = 0, type = NULL, base_value_idx = sm6->value_count; i < block->record_count; ++i)
|
|
{
|
|
sm6->p.location.column = i;
|
|
record = block->records[i];
|
|
value_idx = sm6->value_count;
|
|
|
|
if (record->code == CST_CODE_SETTYPE)
|
|
{
|
|
if (!dxil_record_validate_operand_count(record, 1, 1, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
if (!(type = sm6_parser_get_type(sm6, record->operands[0])))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
elem_type = sm6_type_get_element_type(type);
|
|
if (sm6_type_is_numeric(elem_type))
|
|
{
|
|
reg_data_type = vkd3d_data_type_from_sm6_type(elem_type);
|
|
reg_type = elem_type->u.width > 32 ? VKD3DSPR_IMMCONST64 : VKD3DSPR_IMMCONST;
|
|
}
|
|
else
|
|
{
|
|
reg_data_type = VKD3D_DATA_UNUSED;
|
|
reg_type = VKD3DSPR_INVALID;
|
|
}
|
|
|
|
if (i == block->record_count - 1)
|
|
WARN("Unused SETTYPE record.\n");
|
|
|
|
continue;
|
|
}
|
|
|
|
if (!type)
|
|
{
|
|
WARN("Constant record %zu has no type.\n", value_idx);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
dst = sm6_parser_get_current_value(sm6);
|
|
dst->type = type;
|
|
dst->value_type = VALUE_TYPE_REG;
|
|
dst->is_back_ref = true;
|
|
vsir_register_init(&dst->u.reg, reg_type, reg_data_type, 0);
|
|
|
|
switch (record->code)
|
|
{
|
|
case CST_CODE_NULL:
|
|
if (sm6_type_is_array(type)
|
|
&& (ret = value_allocate_constant_array(dst, type, NULL, sm6)) < 0)
|
|
{
|
|
return ret;
|
|
}
|
|
/* For non-aggregates, register constant data is already zero-filled. */
|
|
break;
|
|
|
|
case CST_CODE_INTEGER:
|
|
if (!dxil_record_validate_operand_count(record, 1, 1, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
if (!sm6_type_is_integer(type))
|
|
{
|
|
WARN("Invalid integer of non-integer type %u at constant idx %zu.\n", type->class, value_idx);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
value = decode_rotated_signed_value(record->operands[0]);
|
|
if (type->u.width <= 32)
|
|
dst->u.reg.u.immconst_u32[0] = value & ((1ull << type->u.width) - 1);
|
|
else
|
|
dst->u.reg.u.immconst_u64[0] = value;
|
|
|
|
break;
|
|
|
|
case CST_CODE_FLOAT:
|
|
if (!dxil_record_validate_operand_count(record, 1, 1, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
if (!sm6_type_is_floating_point(type))
|
|
{
|
|
WARN("Invalid float of non-fp type %u at constant idx %zu.\n", type->class, value_idx);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (type->u.width == 16)
|
|
dst->u.reg.u.immconst_u32[0] = record->operands[0];
|
|
else if (type->u.width == 32)
|
|
dst->u.reg.u.immconst_f32[0] = bitcast_uint_to_float(record->operands[0]);
|
|
else if (type->u.width == 64)
|
|
dst->u.reg.u.immconst_f64[0] = bitcast_uint64_to_double(record->operands[0]);
|
|
else
|
|
vkd3d_unreachable();
|
|
|
|
break;
|
|
|
|
case CST_CODE_DATA:
|
|
if (!sm6_type_is_array(type))
|
|
{
|
|
WARN("Invalid type %u for data constant idx %zu.\n", type->class, value_idx);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"The type of a constant array is not an array type.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
if (!dxil_record_validate_operand_count(record, type->u.array.count, type->u.array.count, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
if ((ret = value_allocate_constant_array(dst, type, record->operands, sm6)) < 0)
|
|
return ret;
|
|
|
|
break;
|
|
|
|
case CST_CODE_CE_GEP:
|
|
case CST_CODE_CE_INBOUNDS_GEP:
|
|
if ((ret = sm6_parser_init_constexpr_gep(sm6, record, dst)) < 0)
|
|
return ret;
|
|
break;
|
|
|
|
case CST_CODE_CE_CAST:
|
|
if (!dxil_record_validate_operand_count(record, 3, 3, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
if ((value = record->operands[0]) != CAST_BITCAST)
|
|
{
|
|
WARN("Unhandled constexpr cast op %"PRIu64".\n", value);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Constexpr cast op %"PRIu64" is unhandled.", value);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
ptr_type = sm6_parser_get_type(sm6, record->operands[1]);
|
|
if (!sm6_type_is_pointer(ptr_type))
|
|
{
|
|
WARN("Constexpr cast at constant idx %zu is not a pointer.\n", value_idx);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Constexpr cast source operand is not a pointer.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if ((value = record->operands[2]) >= sm6->cur_max_value)
|
|
{
|
|
WARN("Invalid value index %"PRIu64".\n", value);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid value index %"PRIu64".", value);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
else if (value == value_idx)
|
|
{
|
|
WARN("Invalid value self-reference at %"PRIu64".\n", value);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid value self-reference for a constexpr cast.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
/* Resolve later in case forward refs exist. */
|
|
dst->type = type;
|
|
dst->u.reg.type = VKD3DSPR_COUNT;
|
|
dst->u.reg.idx[0].offset = value;
|
|
break;
|
|
|
|
case CST_CODE_UNDEF:
|
|
dxil_record_validate_operand_max_count(record, 0, sm6);
|
|
dst->u.reg.type = VKD3DSPR_UNDEF;
|
|
/* Mark as explicitly undefined, not the result of a missing constant code or instruction. */
|
|
dst->is_undefined = true;
|
|
break;
|
|
|
|
default:
|
|
FIXME("Unhandled constant code %u.\n", record->code);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Constant code %u is unhandled.", record->code);
|
|
dst->u.reg.type = VKD3DSPR_UNDEF;
|
|
break;
|
|
}
|
|
|
|
if (record->attachment)
|
|
{
|
|
WARN("Ignoring metadata attachment.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_ATTACHMENT,
|
|
"Ignoring a metadata attachment for a constant.");
|
|
}
|
|
|
|
++sm6->value_count;
|
|
}
|
|
|
|
/* Resolve cast forward refs. */
|
|
for (i = base_value_idx; i < sm6->value_count; ++i)
|
|
{
|
|
dst = &sm6->values[i];
|
|
if (dst->u.reg.type != VKD3DSPR_COUNT)
|
|
continue;
|
|
|
|
type = dst->type;
|
|
|
|
src = &sm6->values[dst->u.reg.idx[0].offset];
|
|
if (!sm6_value_is_numeric_array(src))
|
|
{
|
|
WARN("Value is not an array.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Constexpr cast source value is not a global array element.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
*dst = *src;
|
|
dst->type = type;
|
|
dst->u.reg.data_type = vkd3d_data_type_from_sm6_type(type->u.pointer.type);
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static bool bitcode_parse_alignment(uint64_t encoded_alignment, unsigned int *alignment)
|
|
{
|
|
if (encoded_alignment > MAX_ALIGNMENT_EXPONENT + 1)
|
|
{
|
|
*alignment = 0;
|
|
return false;
|
|
}
|
|
*alignment = (1u << encoded_alignment) >> 1;
|
|
return true;
|
|
}
|
|
|
|
static struct vkd3d_shader_instruction *sm6_parser_require_space(struct sm6_parser *sm6, size_t extra)
|
|
{
|
|
struct vkd3d_shader_instruction_array *instructions = &sm6->p.program->instructions;
|
|
|
|
if (!shader_instruction_array_reserve(instructions, instructions->count + extra))
|
|
{
|
|
ERR("Failed to allocate instruction.\n");
|
|
return NULL;
|
|
}
|
|
return &instructions->elements[instructions->count];
|
|
}
|
|
|
|
/* Space should be reserved before calling this. It is intended to require no checking of the returned pointer. */
|
|
static struct vkd3d_shader_instruction *sm6_parser_add_instruction(struct sm6_parser *sm6,
|
|
enum vkd3d_shader_opcode handler_idx)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = sm6_parser_require_space(sm6, 1);
|
|
VKD3D_ASSERT(ins);
|
|
vsir_instruction_init(ins, &sm6->p.location, handler_idx);
|
|
++sm6->p.program->instructions.count;
|
|
return ins;
|
|
}
|
|
|
|
static void sm6_parser_declare_icb(struct sm6_parser *sm6, const struct sm6_type *elem_type, unsigned int count,
|
|
unsigned int alignment, unsigned int init, struct sm6_value *dst)
|
|
{
|
|
enum vkd3d_data_type data_type = vkd3d_data_type_from_sm6_type(elem_type);
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_DCL_IMMEDIATE_CONSTANT_BUFFER);
|
|
/* The icb value index will be resolved later so forward references can be handled. */
|
|
ins->declaration.icb = (void *)(intptr_t)init;
|
|
register_init_with_id(&dst->u.reg, VKD3DSPR_IMMCONSTBUFFER, data_type, init);
|
|
}
|
|
|
|
static void sm6_parser_declare_indexable_temp(struct sm6_parser *sm6, const struct sm6_type *elem_type,
|
|
unsigned int count, unsigned int alignment, bool has_function_scope, unsigned int init,
|
|
struct vkd3d_shader_instruction *ins, struct sm6_value *dst)
|
|
{
|
|
enum vkd3d_data_type data_type = vkd3d_data_type_from_sm6_type(elem_type);
|
|
|
|
if (ins)
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_DCL_INDEXABLE_TEMP);
|
|
else
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_DCL_INDEXABLE_TEMP);
|
|
ins->declaration.indexable_temp.register_idx = sm6->indexable_temp_count++;
|
|
ins->declaration.indexable_temp.register_size = count;
|
|
ins->declaration.indexable_temp.alignment = alignment;
|
|
ins->declaration.indexable_temp.data_type = data_type;
|
|
ins->declaration.indexable_temp.component_count = 1;
|
|
ins->declaration.indexable_temp.has_function_scope = has_function_scope;
|
|
/* The initialiser value index will be resolved later so forward references can be handled. */
|
|
ins->declaration.indexable_temp.initialiser = (void *)(uintptr_t)init;
|
|
|
|
register_init_with_id(&dst->u.reg, VKD3DSPR_IDXTEMP, data_type, ins->declaration.indexable_temp.register_idx);
|
|
}
|
|
|
|
static void sm6_parser_declare_tgsm_raw(struct sm6_parser *sm6, const struct sm6_type *elem_type,
|
|
unsigned int alignment, unsigned int init, struct sm6_value *dst)
|
|
{
|
|
enum vkd3d_data_type data_type = vkd3d_data_type_from_sm6_type(elem_type);
|
|
struct vkd3d_shader_instruction *ins;
|
|
unsigned int byte_count;
|
|
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_DCL_TGSM_RAW);
|
|
dst_param_init(&ins->declaration.tgsm_raw.reg);
|
|
register_init_with_id(&ins->declaration.tgsm_raw.reg.reg, VKD3DSPR_GROUPSHAREDMEM, data_type, sm6->tgsm_count++);
|
|
dst->u.reg = ins->declaration.tgsm_raw.reg.reg;
|
|
dst->structure_stride = 0;
|
|
ins->declaration.tgsm_raw.alignment = alignment;
|
|
byte_count = elem_type->u.width / 8u;
|
|
if (byte_count != 4)
|
|
{
|
|
FIXME("Unsupported byte count %u.\n", byte_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Raw TGSM byte count %u is not supported.", byte_count);
|
|
}
|
|
ins->declaration.tgsm_raw.byte_count = byte_count;
|
|
/* The initialiser value index will be resolved later when forward references can be handled. */
|
|
ins->flags = init;
|
|
}
|
|
|
|
static void sm6_parser_declare_tgsm_structured(struct sm6_parser *sm6, const struct sm6_type *elem_type,
|
|
unsigned int count, unsigned int alignment, unsigned int init, struct sm6_value *dst)
|
|
{
|
|
enum vkd3d_data_type data_type = vkd3d_data_type_from_sm6_type(elem_type);
|
|
struct vkd3d_shader_instruction *ins;
|
|
unsigned int structure_stride;
|
|
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_DCL_TGSM_STRUCTURED);
|
|
dst_param_init(&ins->declaration.tgsm_structured.reg);
|
|
register_init_with_id(&ins->declaration.tgsm_structured.reg.reg, VKD3DSPR_GROUPSHAREDMEM,
|
|
data_type, sm6->tgsm_count++);
|
|
dst->u.reg = ins->declaration.tgsm_structured.reg.reg;
|
|
structure_stride = elem_type->u.width / 8u;
|
|
if (structure_stride != 4)
|
|
{
|
|
FIXME("Unsupported structure stride %u.\n", structure_stride);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Structured TGSM byte stride %u is not supported.", structure_stride);
|
|
}
|
|
dst->structure_stride = structure_stride;
|
|
ins->declaration.tgsm_structured.alignment = alignment;
|
|
ins->declaration.tgsm_structured.byte_stride = structure_stride;
|
|
ins->declaration.tgsm_structured.structure_count = count;
|
|
/* The initialiser value index will be resolved later when forward references can be handled. */
|
|
ins->flags = init;
|
|
}
|
|
|
|
static bool sm6_parser_declare_global(struct sm6_parser *sm6, const struct dxil_record *record)
|
|
{
|
|
const struct sm6_type *type, *scalar_type;
|
|
unsigned int alignment, count;
|
|
uint64_t address_space, init;
|
|
struct sm6_value *dst;
|
|
bool is_constant;
|
|
|
|
if (!dxil_record_validate_operand_min_count(record, 6, sm6))
|
|
return false;
|
|
|
|
if (!(type = sm6_parser_get_type(sm6, record->operands[0])))
|
|
return false;
|
|
if (sm6_type_is_array(type))
|
|
{
|
|
if (!sm6_type_is_scalar(type->u.array.elem_type))
|
|
{
|
|
FIXME("Unsupported nested type class %u.\n", type->u.array.elem_type->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Global array variables with nested type class %u are not supported.",
|
|
type->u.array.elem_type->class);
|
|
return false;
|
|
}
|
|
count = type->u.array.count;
|
|
scalar_type = type->u.array.elem_type;
|
|
}
|
|
else if (sm6_type_is_scalar(type))
|
|
{
|
|
count = 1;
|
|
scalar_type = type;
|
|
}
|
|
else
|
|
{
|
|
FIXME("Unsupported type class %u.\n", type->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Global variables of type class %u are not supported.", type->class);
|
|
return false;
|
|
}
|
|
|
|
is_constant = record->operands[1] & GLOBALVAR_FLAG_IS_CONSTANT;
|
|
|
|
if (record->operands[1] & GLOBALVAR_FLAG_EXPLICIT_TYPE)
|
|
{
|
|
address_space = record->operands[1] >> GLOBALVAR_ADDRESS_SPACE_SHIFT;
|
|
|
|
if (!(type = sm6_type_get_pointer_to_type(type, address_space, sm6)))
|
|
{
|
|
WARN("Failed to get pointer type for type class %u, address space %"PRIu64".\n",
|
|
type->class, address_space);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_MODULE,
|
|
"Module does not define a pointer type for a global variable.");
|
|
return false;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if (!sm6_type_is_pointer(type))
|
|
{
|
|
WARN("Type is not a pointer.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"The type of a global variable is not a pointer.");
|
|
return false;
|
|
}
|
|
address_space = type->u.pointer.addr_space;
|
|
}
|
|
|
|
if ((init = record->operands[2]))
|
|
{
|
|
if (init - 1 >= sm6->value_capacity)
|
|
{
|
|
WARN("Invalid value index %"PRIu64" for initialiser.", init - 1);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Global variable initialiser value index %"PRIu64" is invalid.", init - 1);
|
|
return false;
|
|
}
|
|
}
|
|
|
|
/* LINKAGE_EXTERNAL is common but not relevant here. */
|
|
if (record->operands[3] != LINKAGE_EXTERNAL && record->operands[3] != LINKAGE_INTERNAL)
|
|
WARN("Ignoring linkage %"PRIu64".\n", record->operands[3]);
|
|
|
|
if (!bitcode_parse_alignment(record->operands[4], &alignment))
|
|
WARN("Invalid alignment %"PRIu64".\n", record->operands[4]);
|
|
|
|
if (record->operands[5])
|
|
WARN("Ignoring section code %"PRIu64".\n", record->operands[5]);
|
|
|
|
if (!sm6_parser_get_global_symbol_name(sm6, sm6->value_count))
|
|
WARN("Missing symbol name for global variable at index %zu.\n", sm6->value_count);
|
|
/* TODO: store global symbol names in struct vkd3d_shader_desc? */
|
|
|
|
if (record->operand_count > 6 && record->operands[6])
|
|
WARN("Ignoring visibility %"PRIu64".\n", record->operands[6]);
|
|
if (record->operand_count > 7 && record->operands[7])
|
|
WARN("Ignoring thread local mode %"PRIu64".\n", record->operands[7]);
|
|
/* record->operands[8] contains unnamed_addr, a flag indicating the address
|
|
* is not important, only the content is. This info is not relevant. */
|
|
if (record->operand_count > 9 && record->operands[9])
|
|
WARN("Ignoring external_init %"PRIu64".\n", record->operands[9]);
|
|
if (record->operand_count > 10 && record->operands[10])
|
|
WARN("Ignoring dll storage class %"PRIu64".\n", record->operands[10]);
|
|
if (record->operand_count > 11 && record->operands[11])
|
|
WARN("Ignoring comdat %"PRIu64".\n", record->operands[11]);
|
|
|
|
dst = sm6_parser_get_current_value(sm6);
|
|
dst->type = type;
|
|
dst->value_type = VALUE_TYPE_REG;
|
|
dst->is_back_ref = true;
|
|
|
|
if (is_constant && !init)
|
|
{
|
|
WARN("Constant array has no initialiser.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"A constant global variable has no initialiser.");
|
|
return false;
|
|
}
|
|
|
|
if (address_space == ADDRESS_SPACE_DEFAULT)
|
|
{
|
|
if (is_constant)
|
|
sm6_parser_declare_icb(sm6, scalar_type, count, alignment, init, dst);
|
|
else
|
|
sm6_parser_declare_indexable_temp(sm6, scalar_type, count, alignment, false, init, NULL, dst);
|
|
}
|
|
else if (address_space == ADDRESS_SPACE_GROUPSHARED)
|
|
{
|
|
if (!sm6_type_is_numeric(scalar_type))
|
|
{
|
|
WARN("Unsupported type class %u.\n", scalar_type->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"TGSM variables of type class %u are not supported.", scalar_type->class);
|
|
return false;
|
|
}
|
|
if (count == 1)
|
|
sm6_parser_declare_tgsm_raw(sm6, scalar_type, alignment, init, dst);
|
|
else
|
|
sm6_parser_declare_tgsm_structured(sm6, scalar_type, count, alignment, init, dst);
|
|
}
|
|
else
|
|
{
|
|
FIXME("Unhandled address space %"PRIu64".\n", address_space);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Global variables with address space %"PRIu64" are not supported.", address_space);
|
|
return false;
|
|
}
|
|
|
|
++sm6->value_count;
|
|
return true;
|
|
}
|
|
|
|
static const struct vkd3d_shader_immediate_constant_buffer *resolve_forward_initialiser(
|
|
size_t index, struct sm6_parser *sm6)
|
|
{
|
|
const struct sm6_value *value;
|
|
|
|
VKD3D_ASSERT(index);
|
|
--index;
|
|
if (!(value = sm6_parser_get_value_safe(sm6, index)) || (!sm6_value_is_icb(value) && !sm6_value_is_undef(value)))
|
|
{
|
|
WARN("Invalid initialiser index %zu.\n", index);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Global variable initialiser value index %zu is invalid.", index);
|
|
return NULL;
|
|
}
|
|
else if (sm6_value_is_icb(value))
|
|
{
|
|
return value->u.icb;
|
|
}
|
|
/* In VSIR, initialisation with undefined values of objects is implied, not explicit. */
|
|
return NULL;
|
|
}
|
|
|
|
static bool resolve_forward_zero_initialiser(size_t index, struct sm6_parser *sm6)
|
|
{
|
|
const struct sm6_value *value;
|
|
|
|
if (!index)
|
|
return false;
|
|
|
|
--index;
|
|
if (!(value = sm6_parser_get_value_safe(sm6, index))
|
|
|| (!sm6_value_is_icb(value) && !sm6_value_is_constant(value) && !sm6_value_is_undef(value)))
|
|
{
|
|
WARN("Invalid initialiser index %zu.\n", index);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"TGSM initialiser value index %zu is invalid.", index);
|
|
return false;
|
|
}
|
|
else if ((sm6_value_is_icb(value) && value->u.icb->is_null) || sm6_value_is_constant_zero(value))
|
|
{
|
|
return true;
|
|
}
|
|
else if (sm6_value_is_undef(value))
|
|
{
|
|
/* In VSIR, initialisation with undefined values of objects is implied, not explicit. */
|
|
return false;
|
|
}
|
|
|
|
FIXME("Non-zero initialisers are not supported.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Non-zero TGSM initialisers are not supported.");
|
|
return false;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_globals_init(struct sm6_parser *sm6)
|
|
{
|
|
size_t i, count, base_value_idx = sm6->value_count;
|
|
const struct dxil_block *block = &sm6->root_block;
|
|
struct vkd3d_shader_instruction *ins;
|
|
const struct dxil_record *record;
|
|
enum vkd3d_result ret;
|
|
uint64_t version;
|
|
|
|
sm6->p.location.line = block->id;
|
|
sm6->p.location.column = 0;
|
|
|
|
for (i = 0, count = 0; i < block->record_count; ++i)
|
|
count += block->records[i]->code == MODULE_CODE_GLOBALVAR;
|
|
sm6_parser_require_space(sm6, count);
|
|
|
|
for (i = 0; i < block->record_count; ++i)
|
|
{
|
|
sm6->p.location.column = i;
|
|
record = block->records[i];
|
|
switch (record->code)
|
|
{
|
|
case MODULE_CODE_FUNCTION:
|
|
if (!sm6_parser_declare_function(sm6, record))
|
|
{
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_FUNCTION_DCL,
|
|
"A DXIL function declaration is invalid.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
break;
|
|
|
|
case MODULE_CODE_GLOBALVAR:
|
|
if (!sm6_parser_declare_global(sm6, record))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
break;
|
|
|
|
case MODULE_CODE_VERSION:
|
|
if (!dxil_record_validate_operand_count(record, 1, 1, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
if ((version = record->operands[0]) != 1)
|
|
{
|
|
FIXME("Unsupported format version %#"PRIx64".\n", version);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_UNSUPPORTED_BITCODE_FORMAT,
|
|
"Bitcode format version %#"PRIx64" is unsupported.", version);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < block->child_block_count; ++i)
|
|
{
|
|
if (block->child_blocks[i]->id == CONSTANTS_BLOCK
|
|
&& (ret = sm6_parser_constants_init(sm6, block->child_blocks[i])) < 0)
|
|
return ret;
|
|
}
|
|
|
|
/* Resolve initialiser forward references. */
|
|
for (i = 0; i < sm6->p.program->instructions.count; ++i)
|
|
{
|
|
ins = &sm6->p.program->instructions.elements[i];
|
|
if (ins->opcode == VKD3DSIH_DCL_INDEXABLE_TEMP && ins->declaration.indexable_temp.initialiser)
|
|
{
|
|
ins->declaration.indexable_temp.initialiser = resolve_forward_initialiser(
|
|
(uintptr_t)ins->declaration.indexable_temp.initialiser, sm6);
|
|
}
|
|
else if (ins->opcode == VKD3DSIH_DCL_IMMEDIATE_CONSTANT_BUFFER)
|
|
{
|
|
ins->declaration.icb = resolve_forward_initialiser((uintptr_t)ins->declaration.icb, sm6);
|
|
}
|
|
else if (ins->opcode == VKD3DSIH_DCL_TGSM_RAW)
|
|
{
|
|
ins->declaration.tgsm_raw.zero_init = resolve_forward_zero_initialiser(ins->flags, sm6);
|
|
ins->flags = 0;
|
|
}
|
|
else if (ins->opcode == VKD3DSIH_DCL_TGSM_STRUCTURED)
|
|
{
|
|
ins->declaration.tgsm_structured.zero_init = resolve_forward_zero_initialiser(ins->flags, sm6);
|
|
ins->flags = 0;
|
|
}
|
|
}
|
|
for (i = base_value_idx; i < sm6->value_count; ++i)
|
|
{
|
|
const struct vkd3d_shader_immediate_constant_buffer *icb;
|
|
struct sm6_value *value = &sm6->values[i];
|
|
|
|
if (!sm6_value_is_register(value) || value->u.reg.type != VKD3DSPR_IMMCONSTBUFFER)
|
|
continue;
|
|
|
|
if ((icb = resolve_forward_initialiser(value->u.reg.idx[0].offset, sm6)))
|
|
value->u.reg.idx[0].offset = icb->register_idx;
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static void dst_param_io_init(struct vkd3d_shader_dst_param *param,
|
|
const struct signature_element *e, enum vkd3d_shader_register_type reg_type)
|
|
{
|
|
enum vkd3d_shader_component_type component_type;
|
|
|
|
param->write_mask = e->mask;
|
|
param->modifiers = 0;
|
|
param->shift = 0;
|
|
/* DXIL types do not have signedness. Load signed elements as unsigned. */
|
|
component_type = e->component_type == VKD3D_SHADER_COMPONENT_INT ? VKD3D_SHADER_COMPONENT_UINT : e->component_type;
|
|
vsir_register_init(¶m->reg, reg_type, vkd3d_data_type_from_component_type(component_type), 0);
|
|
param->reg.dimension = VSIR_DIMENSION_VEC4;
|
|
}
|
|
|
|
static void src_params_init_from_operands(struct vkd3d_shader_src_param *src_params,
|
|
const struct sm6_value **operands, unsigned int count)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < count; ++i)
|
|
src_param_init_from_value(&src_params[i], operands[i]);
|
|
}
|
|
|
|
static enum vkd3d_shader_register_type register_type_from_dxil_semantic_kind(
|
|
enum vkd3d_shader_sysval_semantic sysval_semantic)
|
|
{
|
|
switch (sysval_semantic)
|
|
{
|
|
case VKD3D_SHADER_SV_COVERAGE:
|
|
return VKD3DSPR_COVERAGE;
|
|
case VKD3D_SHADER_SV_DEPTH:
|
|
return VKD3DSPR_DEPTHOUT;
|
|
case VKD3D_SHADER_SV_DEPTH_GREATER_EQUAL:
|
|
return VKD3DSPR_DEPTHOUTGE;
|
|
case VKD3D_SHADER_SV_DEPTH_LESS_EQUAL:
|
|
return VKD3DSPR_DEPTHOUTLE;
|
|
default:
|
|
return VKD3DSPR_INVALID;
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_init_signature(struct sm6_parser *sm6, const struct shader_signature *s,
|
|
bool is_input, enum vkd3d_shader_register_type reg_type, struct vkd3d_shader_dst_param *params)
|
|
{
|
|
enum vkd3d_shader_type shader_type = sm6->p.program->shader_version.type;
|
|
bool is_patch_constant, is_control_point;
|
|
struct vkd3d_shader_dst_param *param;
|
|
const struct signature_element *e;
|
|
unsigned int i, count;
|
|
|
|
is_patch_constant = reg_type == VKD3DSPR_PATCHCONST;
|
|
|
|
is_control_point = false;
|
|
if (!is_patch_constant)
|
|
{
|
|
switch (shader_type)
|
|
{
|
|
case VKD3D_SHADER_TYPE_DOMAIN:
|
|
case VKD3D_SHADER_TYPE_GEOMETRY:
|
|
is_control_point = is_input;
|
|
break;
|
|
|
|
case VKD3D_SHADER_TYPE_HULL:
|
|
is_control_point = true;
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < s->element_count; ++i)
|
|
{
|
|
e = &s->elements[i];
|
|
|
|
param = ¶ms[i];
|
|
|
|
if (e->register_index == UINT_MAX)
|
|
{
|
|
dst_param_io_init(param, e, register_type_from_dxil_semantic_kind(e->sysval_semantic));
|
|
continue;
|
|
}
|
|
|
|
dst_param_io_init(param, e, reg_type);
|
|
count = 0;
|
|
|
|
if (is_control_point)
|
|
{
|
|
if (reg_type == VKD3DSPR_OUTPUT)
|
|
param->reg.idx[count].rel_addr = vsir_program_create_outpointid_param(sm6->p.program);
|
|
param->reg.idx[count++].offset = 0;
|
|
}
|
|
|
|
if (e->register_count > 1 || (is_patch_constant && vsir_sysval_semantic_is_tess_factor(e->sysval_semantic)))
|
|
param->reg.idx[count++].offset = 0;
|
|
|
|
VKD3D_ASSERT(count < ARRAY_SIZE(param->reg.idx));
|
|
param->reg.idx[count++].offset = i;
|
|
param->reg.idx_count = count;
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_init_output_signature(struct sm6_parser *sm6, const struct shader_signature *output_signature)
|
|
{
|
|
sm6_parser_init_signature(sm6, output_signature, false, VKD3DSPR_OUTPUT, sm6->output_params);
|
|
}
|
|
|
|
static void sm6_parser_init_input_signature(struct sm6_parser *sm6, const struct shader_signature *input_signature)
|
|
{
|
|
sm6_parser_init_signature(sm6, input_signature, true, VKD3DSPR_INPUT, sm6->input_params);
|
|
}
|
|
|
|
static void sm6_parser_init_patch_constant_signature(struct sm6_parser *sm6,
|
|
const struct shader_signature *patch_constant_signature)
|
|
{
|
|
bool is_input = sm6->p.program->shader_version.type == VKD3D_SHADER_TYPE_DOMAIN;
|
|
|
|
sm6_parser_init_signature(sm6, patch_constant_signature, is_input, VKD3DSPR_PATCHCONST,
|
|
sm6->patch_constant_params);
|
|
}
|
|
|
|
static const struct sm6_value *sm6_parser_next_function_definition(struct sm6_parser *sm6)
|
|
{
|
|
size_t i, count = sm6->function_count;
|
|
|
|
for (i = 0; i < sm6->value_count; ++i)
|
|
{
|
|
if (sm6_type_is_function_pointer(sm6->values[i].type) && !sm6->values[i].u.function.is_prototype && !count--)
|
|
break;
|
|
}
|
|
if (i == sm6->value_count)
|
|
return NULL;
|
|
|
|
++sm6->function_count;
|
|
return &sm6->values[i];
|
|
}
|
|
|
|
static struct sm6_block *sm6_block_create()
|
|
{
|
|
struct sm6_block *block = vkd3d_calloc(1, sizeof(*block));
|
|
return block;
|
|
}
|
|
|
|
static struct sm6_phi *sm6_block_phi_require_space(struct sm6_block *block, struct sm6_parser *sm6)
|
|
{
|
|
struct sm6_phi *phi;
|
|
|
|
if (!vkd3d_array_reserve((void **)&block->phi, &block->phi_capacity, block->phi_count + 1, sizeof(*block->phi)))
|
|
{
|
|
ERR("Failed to allocate phi array.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating a phi instruction.");
|
|
return NULL;
|
|
}
|
|
phi = &block->phi[block->phi_count++];
|
|
|
|
phi->incoming = NULL;
|
|
phi->incoming_capacity = 0;
|
|
phi->incoming_count = 0;
|
|
|
|
return phi;
|
|
}
|
|
|
|
struct function_emission_state
|
|
{
|
|
struct sm6_block *code_block;
|
|
struct vkd3d_shader_instruction *ins;
|
|
unsigned int temp_idx;
|
|
};
|
|
|
|
static bool sm6_parser_emit_reg_composite_construct(struct sm6_parser *sm6, const struct vkd3d_shader_register **operand_regs,
|
|
unsigned int component_count, struct function_emission_state *state, struct vkd3d_shader_register *reg);
|
|
|
|
static void sm6_parser_emit_alloca(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct vkd3d_shader_instruction *ins, struct sm6_value *dst)
|
|
{
|
|
const struct sm6_type *type[2], *elem_type;
|
|
const struct sm6_value *size;
|
|
unsigned int i, alignment;
|
|
uint64_t packed_operands;
|
|
|
|
if (!dxil_record_validate_operand_count(record, 4, 4, sm6))
|
|
return;
|
|
|
|
for (i = 0; i < 2; ++i)
|
|
{
|
|
if (!(type[i] = sm6_parser_get_type(sm6, record->operands[i])))
|
|
return;
|
|
}
|
|
|
|
packed_operands = record->operands[3];
|
|
if (packed_operands & ALLOCA_FLAG_IN_ALLOCA)
|
|
WARN("Ignoring in_alloca flag.\n");
|
|
if (!(packed_operands & ALLOCA_FLAG_EXPLICIT_TYPE))
|
|
{
|
|
FIXME("Unhandled implicit type.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Implicit result type for ALLOCA instructions is not supported.");
|
|
return;
|
|
}
|
|
packed_operands &= ~(ALLOCA_FLAG_IN_ALLOCA | ALLOCA_FLAG_EXPLICIT_TYPE);
|
|
|
|
if (!sm6_type_is_array(type[0]) || !sm6_type_is_numeric(elem_type = type[0]->u.array.elem_type))
|
|
{
|
|
WARN("Type is not a numeric array.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Result type of an ALLOCA instruction is not a numeric array.");
|
|
return;
|
|
}
|
|
/* The second type operand is the type of the allocation size operand, in case it is a
|
|
* forward reference. We only support a constant size, so no forward ref support is needed. */
|
|
if (!sm6_type_is_integer(type[1]))
|
|
{
|
|
WARN("Size operand type is not scalar integer.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"The type of the allocation size operand of an ALLOCA instruction is not scalar integer.");
|
|
return;
|
|
}
|
|
|
|
if (!(dst->type = sm6_type_get_pointer_to_type(type[0], ADDRESS_SPACE_DEFAULT, sm6)))
|
|
{
|
|
WARN("Failed to get pointer type for type class %u.\n", type[0]->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_MODULE,
|
|
"Module does not define a pointer type for an ALLOCA instruction.");
|
|
return;
|
|
}
|
|
|
|
if (!(size = sm6_parser_get_value_safe(sm6, record->operands[2])))
|
|
return;
|
|
/* A size of 1 means one instance of type[0], i.e. one array. */
|
|
if (sm6_value_get_constant_uint(size) != 1)
|
|
{
|
|
FIXME("Allocation size is not 1.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"ALLOCA instruction allocation sizes other than 1 are not supported.");
|
|
return;
|
|
}
|
|
|
|
if (!bitcode_parse_alignment(packed_operands & ALLOCA_ALIGNMENT_MASK, &alignment))
|
|
WARN("Invalid alignment %"PRIu64".\n", packed_operands);
|
|
packed_operands &= ~ALLOCA_ALIGNMENT_MASK;
|
|
|
|
if (packed_operands)
|
|
WARN("Ignoring flags %#"PRIx64".\n", packed_operands);
|
|
|
|
sm6_parser_declare_indexable_temp(sm6, elem_type, type[0]->u.array.count, alignment, true, 0, ins, dst);
|
|
}
|
|
|
|
static enum vkd3d_shader_opcode map_dx_atomicrmw_op(uint64_t code)
|
|
{
|
|
switch (code)
|
|
{
|
|
case RMW_ADD:
|
|
return VKD3DSIH_IMM_ATOMIC_IADD;
|
|
case RMW_AND:
|
|
return VKD3DSIH_IMM_ATOMIC_AND;
|
|
case RMW_MAX:
|
|
return VKD3DSIH_IMM_ATOMIC_IMAX;
|
|
case RMW_MIN:
|
|
return VKD3DSIH_IMM_ATOMIC_IMIN;
|
|
case RMW_OR:
|
|
return VKD3DSIH_IMM_ATOMIC_OR;
|
|
case RMW_UMAX:
|
|
return VKD3DSIH_IMM_ATOMIC_UMAX;
|
|
case RMW_UMIN:
|
|
return VKD3DSIH_IMM_ATOMIC_UMIN;
|
|
case RMW_XCHG:
|
|
return VKD3DSIH_IMM_ATOMIC_EXCH;
|
|
case RMW_XOR:
|
|
return VKD3DSIH_IMM_ATOMIC_XOR;
|
|
default:
|
|
/* DXIL currently doesn't use SUB and NAND. */
|
|
return VKD3DSIH_INVALID;
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_atomicrmw(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct function_emission_state *state, struct sm6_value *dst)
|
|
{
|
|
struct vkd3d_shader_register coord, const_offset, const_zero;
|
|
const struct vkd3d_shader_register *regs[2];
|
|
struct vkd3d_shader_dst_param *dst_params;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_instruction *ins;
|
|
const struct sm6_value *ptr, *src;
|
|
enum vkd3d_shader_opcode op;
|
|
unsigned int i = 0;
|
|
bool is_volatile;
|
|
uint64_t code;
|
|
|
|
if (!(ptr = sm6_parser_get_value_by_ref(sm6, record, NULL, &i))
|
|
|| !sm6_value_validate_is_pointer_to_i32(ptr, sm6)
|
|
|| !sm6_value_validate_is_backward_ref(ptr, sm6))
|
|
return;
|
|
|
|
if (ptr->u.reg.type != VKD3DSPR_GROUPSHAREDMEM)
|
|
{
|
|
WARN("Register is not groupshared.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"The destination register for an atomicrmw instruction is not groupshared memory.");
|
|
return;
|
|
}
|
|
|
|
dst->type = ptr->type->u.pointer.type;
|
|
|
|
if (!(src = sm6_parser_get_value_by_ref(sm6, record, dst->type, &i)))
|
|
return;
|
|
|
|
if (!dxil_record_validate_operand_count(record, i + 4, i + 4, sm6))
|
|
return;
|
|
|
|
if ((op = map_dx_atomicrmw_op(code = record->operands[i++])) == VKD3DSIH_INVALID)
|
|
{
|
|
FIXME("Unhandled atomicrmw op %"PRIu64".\n", code);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Operation %"PRIu64" for an atomicrmw instruction is unhandled.", code);
|
|
return;
|
|
}
|
|
|
|
is_volatile = record->operands[i++];
|
|
|
|
/* It's currently not possible to specify an atomic ordering in HLSL, and it defaults to seq_cst. */
|
|
if ((code = record->operands[i++]) != ORDERING_SEQCST)
|
|
FIXME("Unhandled atomic ordering %"PRIu64".\n", code);
|
|
|
|
if ((code = record->operands[i]) != 1)
|
|
WARN("Ignoring synchronisation scope %"PRIu64".\n", code);
|
|
|
|
if (ptr->structure_stride)
|
|
{
|
|
if (ptr->u.reg.idx[1].rel_addr)
|
|
{
|
|
regs[0] = &ptr->u.reg.idx[1].rel_addr->reg;
|
|
}
|
|
else
|
|
{
|
|
register_make_constant_uint(&const_offset, ptr->u.reg.idx[1].offset);
|
|
regs[0] = &const_offset;
|
|
}
|
|
register_make_constant_uint(&const_zero, 0);
|
|
regs[1] = &const_zero;
|
|
if (!sm6_parser_emit_reg_composite_construct(sm6, regs, 2, state, &coord))
|
|
return;
|
|
}
|
|
|
|
ins = state->ins;
|
|
vsir_instruction_init(ins, &sm6->p.location, op);
|
|
ins->flags = is_volatile ? VKD3DARF_SEQ_CST | VKD3DARF_VOLATILE : VKD3DARF_SEQ_CST;
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 2, sm6)))
|
|
return;
|
|
if (ptr->structure_stride)
|
|
src_param_init_vector_from_reg(&src_params[0], &coord);
|
|
else
|
|
src_param_make_constant_uint(&src_params[0], 0);
|
|
src_param_init_from_value(&src_params[1], src);
|
|
|
|
dst_params = instruction_dst_params_alloc(ins, 2, sm6);
|
|
register_init_ssa_scalar(&dst_params[0].reg, dst->type, dst, sm6);
|
|
dst_param_init(&dst_params[0]);
|
|
|
|
dst_params[1].reg = ptr->u.reg;
|
|
dst_params[1].reg.data_type = VKD3D_DATA_UNUSED;
|
|
dst_params[1].reg.idx[1].rel_addr = NULL;
|
|
dst_params[1].reg.idx[1].offset = ~0u;
|
|
dst_params[1].reg.idx_count = 1;
|
|
dst_param_init(&dst_params[1]);
|
|
|
|
dst->u.reg = dst_params[0].reg;
|
|
}
|
|
|
|
static enum vkd3d_shader_opcode map_binary_op(uint64_t code, const struct sm6_type *type_a,
|
|
const struct sm6_type *type_b, struct sm6_parser *sm6)
|
|
{
|
|
bool is_int = sm6_type_is_bool_i16_i32_i64(type_a);
|
|
bool is_bool = sm6_type_is_bool(type_a);
|
|
enum vkd3d_shader_opcode op;
|
|
bool is_valid;
|
|
|
|
if (!is_int && !sm6_type_is_floating_point(type_a))
|
|
{
|
|
WARN("Argument type %u is not bool, int16/32/64 or floating point.\n", type_a->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"An argument to a binary operation is not bool, int16/32/64 or floating point.");
|
|
return VKD3DSIH_INVALID;
|
|
}
|
|
if (type_a != type_b)
|
|
{
|
|
WARN("Type mismatch, type %u width %u vs type %u width %u.\n", type_a->class,
|
|
type_a->u.width, type_b->class, type_b->u.width);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"Type mismatch in binary operation arguments.");
|
|
}
|
|
|
|
switch (code)
|
|
{
|
|
case BINOP_ADD:
|
|
case BINOP_SUB:
|
|
/* NEG is applied later for subtraction. */
|
|
op = is_int ? VKD3DSIH_IADD : VKD3DSIH_ADD;
|
|
is_valid = !is_bool;
|
|
break;
|
|
case BINOP_AND:
|
|
op = VKD3DSIH_AND;
|
|
is_valid = is_int;
|
|
break;
|
|
case BINOP_ASHR:
|
|
op = VKD3DSIH_ISHR;
|
|
is_valid = is_int && !is_bool;
|
|
break;
|
|
case BINOP_LSHR:
|
|
op = VKD3DSIH_USHR;
|
|
is_valid = is_int && !is_bool;
|
|
break;
|
|
case BINOP_MUL:
|
|
op = is_int ? VKD3DSIH_UMUL : VKD3DSIH_MUL;
|
|
is_valid = !is_bool;
|
|
break;
|
|
case BINOP_OR:
|
|
op = VKD3DSIH_OR;
|
|
is_valid = is_int;
|
|
break;
|
|
case BINOP_SDIV:
|
|
op = is_int ? VKD3DSIH_IDIV : VKD3DSIH_DIV;
|
|
is_valid = !is_bool;
|
|
break;
|
|
case BINOP_SREM:
|
|
op = is_int ? VKD3DSIH_IDIV : VKD3DSIH_FREM;
|
|
is_valid = !is_bool;
|
|
break;
|
|
case BINOP_SHL:
|
|
op = VKD3DSIH_ISHL;
|
|
is_valid = is_int && !is_bool;
|
|
break;
|
|
case BINOP_UDIV:
|
|
case BINOP_UREM:
|
|
op = VKD3DSIH_UDIV;
|
|
is_valid = is_int && !is_bool;
|
|
break;
|
|
case BINOP_XOR:
|
|
op = VKD3DSIH_XOR;
|
|
is_valid = is_int;
|
|
break;
|
|
default:
|
|
FIXME("Unhandled binary op %#"PRIx64".\n", code);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Binary operation %#"PRIx64" is unhandled.", code);
|
|
return VKD3DSIH_INVALID;
|
|
}
|
|
|
|
if (!is_valid)
|
|
{
|
|
WARN("Invalid operation %u for type %u, width %u.\n", op, type_a->class, type_a->u.width);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_INVALID_OPERATION,
|
|
"Binary operation %u is invalid on type class %u, width %u.", op, type_a->class, type_a->u.width);
|
|
}
|
|
|
|
return op;
|
|
}
|
|
|
|
static void sm6_parser_emit_binop(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct vkd3d_shader_instruction *ins, struct sm6_value *dst)
|
|
{
|
|
struct vkd3d_shader_src_param *src_params;
|
|
enum vkd3d_shader_opcode handler_idx;
|
|
const struct sm6_value *a, *b;
|
|
uint64_t code, flags;
|
|
bool silence_warning;
|
|
unsigned int i = 0;
|
|
|
|
a = sm6_parser_get_value_by_ref(sm6, record, NULL, &i);
|
|
if (!a)
|
|
return;
|
|
b = sm6_parser_get_value_by_ref(sm6, record, a->type, &i);
|
|
if (!b)
|
|
return;
|
|
|
|
if (!dxil_record_validate_operand_count(record, i + 1, i + 2, sm6))
|
|
return;
|
|
|
|
code = record->operands[i++];
|
|
if ((handler_idx = map_binary_op(code, a->type, b->type, sm6)) == VKD3DSIH_INVALID)
|
|
return;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, handler_idx);
|
|
|
|
flags = (record->operand_count > i) ? record->operands[i] : 0;
|
|
silence_warning = false;
|
|
|
|
switch (handler_idx)
|
|
{
|
|
case VKD3DSIH_ADD:
|
|
case VKD3DSIH_MUL:
|
|
case VKD3DSIH_DIV:
|
|
case VKD3DSIH_FREM:
|
|
if (!(flags & FP_ALLOW_UNSAFE_ALGEBRA))
|
|
ins->flags |= VKD3DSI_PRECISE_X;
|
|
flags &= ~FP_ALLOW_UNSAFE_ALGEBRA;
|
|
/* SPIR-V FPFastMathMode is only available in the Kernel execution model. */
|
|
silence_warning = !(flags & ~(FP_NO_NAN | FP_NO_INF | FP_NO_SIGNED_ZEROS | FP_ALLOW_RECIPROCAL));
|
|
break;
|
|
case VKD3DSIH_IADD:
|
|
case VKD3DSIH_UMUL:
|
|
case VKD3DSIH_ISHL:
|
|
silence_warning = !(flags & ~(OB_NO_UNSIGNED_WRAP | OB_NO_SIGNED_WRAP));
|
|
break;
|
|
case VKD3DSIH_ISHR:
|
|
case VKD3DSIH_USHR:
|
|
case VKD3DSIH_IDIV:
|
|
case VKD3DSIH_UDIV:
|
|
silence_warning = !(flags & ~PEB_EXACT);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
/* The above flags are very common and cause warning spam. */
|
|
if (flags && silence_warning)
|
|
{
|
|
TRACE("Ignoring flags %#"PRIx64".\n", flags);
|
|
}
|
|
else if (flags)
|
|
{
|
|
WARN("Ignoring flags %#"PRIx64".\n", flags);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring flags %#"PRIx64" for a binary operation.", flags);
|
|
}
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 2, sm6)))
|
|
return;
|
|
src_param_init_from_value(&src_params[0], a);
|
|
src_param_init_from_value(&src_params[1], b);
|
|
if (code == BINOP_SUB)
|
|
src_params[1].modifiers = VKD3DSPSM_NEG;
|
|
|
|
dst->type = a->type;
|
|
|
|
if (handler_idx == VKD3DSIH_UMUL || handler_idx == VKD3DSIH_UDIV || handler_idx == VKD3DSIH_IDIV)
|
|
{
|
|
struct vkd3d_shader_dst_param *dst_params = instruction_dst_params_alloc(ins, 2, sm6);
|
|
unsigned int index = code != BINOP_UDIV && code != BINOP_SDIV;
|
|
|
|
dst_param_init(&dst_params[0]);
|
|
dst_param_init(&dst_params[1]);
|
|
register_init_ssa_scalar(&dst_params[index].reg, a->type, dst, sm6);
|
|
vsir_register_init(&dst_params[index ^ 1].reg, VKD3DSPR_NULL, VKD3D_DATA_UNUSED, 0);
|
|
dst->u.reg = dst_params[index].reg;
|
|
}
|
|
else
|
|
{
|
|
if (handler_idx == VKD3DSIH_ISHL || handler_idx == VKD3DSIH_ISHR || handler_idx == VKD3DSIH_USHR)
|
|
{
|
|
/* DXC emits AND instructions where necessary to mask shift counts. Shift binops
|
|
* do not imply masking the shift as the TPF equivalents do. */
|
|
ins->flags |= VKD3DSI_SHIFT_UNMASKED;
|
|
}
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
}
|
|
|
|
static const struct sm6_block *sm6_function_get_block(const struct sm6_function *function, uint64_t index,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
if (index >= function->block_count)
|
|
{
|
|
WARN("Invalid code block index %#"PRIx64".\n", index);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Invalid code block index %#"PRIx64" for a control flow instruction.", index);
|
|
return NULL;
|
|
}
|
|
return function->blocks[index];
|
|
}
|
|
|
|
static void sm6_parser_emit_br(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct sm6_function *function, struct sm6_block *code_block, struct vkd3d_shader_instruction *ins)
|
|
{
|
|
const struct sm6_value *value;
|
|
unsigned int i = 2;
|
|
|
|
if (record->operand_count != 1 && record->operand_count < 3)
|
|
{
|
|
WARN("Invalid operand count %u.\n", record->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Invalid operand count %u for a branch instruction.", record->operand_count);
|
|
return;
|
|
}
|
|
|
|
if (record->operand_count == 1)
|
|
{
|
|
code_block->terminator.type = TERMINATOR_UNCOND_BR;
|
|
code_block->terminator.true_block = sm6_function_get_block(function, record->operands[0], sm6);
|
|
}
|
|
else
|
|
{
|
|
if (!sm6->bool_type)
|
|
{
|
|
WARN("Bool type not found.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_MODULE,
|
|
"Module does not define a boolean type for conditions.");
|
|
return;
|
|
}
|
|
if (!(value = sm6_parser_get_value_by_ref(sm6, record, sm6->bool_type, &i))
|
|
|| !sm6_value_validate_is_bool(value, sm6))
|
|
return;
|
|
dxil_record_validate_operand_max_count(record, i, sm6);
|
|
|
|
code_block->terminator.type = TERMINATOR_COND_BR;
|
|
code_block->terminator.conditional_reg = value->u.reg;
|
|
code_block->terminator.true_block = sm6_function_get_block(function, record->operands[0], sm6);
|
|
code_block->terminator.false_block = sm6_function_get_block(function, record->operands[1], sm6);
|
|
}
|
|
|
|
ins->opcode = VKD3DSIH_NOP;
|
|
}
|
|
|
|
static bool sm6_parser_emit_reg_composite_construct(struct sm6_parser *sm6, const struct vkd3d_shader_register **operand_regs,
|
|
unsigned int component_count, struct function_emission_state *state, struct vkd3d_shader_register *reg)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_dst_param *dst_param;
|
|
bool all_constant = true;
|
|
unsigned int i;
|
|
|
|
if (component_count == 1)
|
|
{
|
|
*reg = *operand_regs[0];
|
|
return true;
|
|
}
|
|
|
|
for (i = 0; i < component_count; ++i)
|
|
all_constant &= register_is_constant(operand_regs[i]);
|
|
|
|
if (all_constant)
|
|
{
|
|
vsir_register_init(reg, VKD3DSPR_IMMCONST, operand_regs[0]->data_type, 0);
|
|
reg->dimension = VSIR_DIMENSION_VEC4;
|
|
for (i = 0; i < component_count; ++i)
|
|
reg->u.immconst_u32[i] = operand_regs[i]->u.immconst_u32[0];
|
|
for (; i < VKD3D_VEC4_SIZE; ++i)
|
|
reg->u.immconst_u32[i] = 0;
|
|
return true;
|
|
}
|
|
|
|
register_init_with_id(reg, VKD3DSPR_TEMP, operand_regs[0]->data_type, state->temp_idx++);
|
|
reg->dimension = VSIR_DIMENSION_VEC4;
|
|
|
|
for (i = 0; i < component_count; ++i, ++ins)
|
|
{
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return false;
|
|
|
|
src_param_init(&src_params[0]);
|
|
src_params[0].reg = *operand_regs[i];
|
|
|
|
if (!(dst_param = instruction_dst_params_alloc(ins, 1, sm6)))
|
|
return false;
|
|
|
|
dst_param_init_scalar(dst_param, i);
|
|
dst_param->reg = *reg;
|
|
}
|
|
|
|
state->ins = ins;
|
|
state->code_block->instruction_count += component_count;
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_parser_emit_composite_construct(struct sm6_parser *sm6, const struct sm6_value **operands,
|
|
unsigned int component_count, struct function_emission_state *state, struct vkd3d_shader_register *reg)
|
|
{
|
|
const struct vkd3d_shader_register *operand_regs[VKD3D_VEC4_SIZE];
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < component_count; ++i)
|
|
operand_regs[i] = &operands[i]->u.reg;
|
|
|
|
return sm6_parser_emit_reg_composite_construct(sm6, operand_regs, component_count, state, reg);
|
|
}
|
|
|
|
static bool sm6_parser_emit_coordinate_construct(struct sm6_parser *sm6, const struct sm6_value **operands,
|
|
unsigned int max_operands, const struct sm6_value *z_operand, struct function_emission_state *state,
|
|
struct vkd3d_shader_register *reg)
|
|
{
|
|
const struct vkd3d_shader_register *operand_regs[VKD3D_VEC4_SIZE];
|
|
unsigned int component_count;
|
|
|
|
for (component_count = 0; component_count < max_operands; ++component_count)
|
|
{
|
|
if (!z_operand && operands[component_count]->is_undefined)
|
|
break;
|
|
operand_regs[component_count] = &operands[component_count]->u.reg;
|
|
}
|
|
if (z_operand)
|
|
{
|
|
operand_regs[component_count++] = &z_operand->u.reg;
|
|
}
|
|
|
|
return sm6_parser_emit_reg_composite_construct(sm6, operand_regs, component_count, state, reg);
|
|
}
|
|
|
|
static enum vkd3d_shader_opcode sm6_dx_map_void_op(enum dx_intrinsic_opcode op)
|
|
{
|
|
switch (op)
|
|
{
|
|
case DX_WAVE_IS_FIRST_LANE:
|
|
return VKD3DSIH_WAVE_IS_FIRST_LANE;
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_void(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
vsir_instruction_init(ins, &sm6->p.location, sm6_dx_map_void_op(op));
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static enum vkd3d_shader_opcode map_dx_unary_op(enum dx_intrinsic_opcode op)
|
|
{
|
|
switch (op)
|
|
{
|
|
case DX_ISNAN:
|
|
return VKD3DSIH_ISNAN;
|
|
case DX_ISINF:
|
|
return VKD3DSIH_ISINF;
|
|
case DX_ISFINITE:
|
|
return VKD3DSIH_ISFINITE;
|
|
case DX_TAN:
|
|
return VKD3DSIH_TAN;
|
|
case DX_ACOS:
|
|
return VKD3DSIH_ACOS;
|
|
case DX_ASIN:
|
|
return VKD3DSIH_ASIN;
|
|
case DX_ATAN:
|
|
return VKD3DSIH_ATAN;
|
|
case DX_HCOS:
|
|
return VKD3DSIH_HCOS;
|
|
case DX_HSIN:
|
|
return VKD3DSIH_HSIN;
|
|
case DX_HTAN:
|
|
return VKD3DSIH_HTAN;
|
|
case DX_EXP:
|
|
return VKD3DSIH_EXP;
|
|
case DX_FRC:
|
|
return VKD3DSIH_FRC;
|
|
case DX_LOG:
|
|
return VKD3DSIH_LOG;
|
|
case DX_SQRT:
|
|
return VKD3DSIH_SQRT;
|
|
case DX_RSQRT:
|
|
return VKD3DSIH_RSQ;
|
|
case DX_ROUND_NE:
|
|
return VKD3DSIH_ROUND_NE;
|
|
case DX_ROUND_NI:
|
|
return VKD3DSIH_ROUND_NI;
|
|
case DX_ROUND_PI:
|
|
return VKD3DSIH_ROUND_PI;
|
|
case DX_ROUND_Z:
|
|
return VKD3DSIH_ROUND_Z;
|
|
case DX_BFREV:
|
|
return VKD3DSIH_BFREV;
|
|
case DX_COUNT_BITS:
|
|
return VKD3DSIH_COUNTBITS;
|
|
case DX_FIRST_BIT_LO:
|
|
return VKD3DSIH_FIRSTBIT_LO;
|
|
case DX_FIRST_BIT_HI:
|
|
return VKD3DSIH_FIRSTBIT_HI;
|
|
case DX_FIRST_BIT_SHI:
|
|
return VKD3DSIH_FIRSTBIT_SHI;
|
|
case DX_DERIV_COARSEX:
|
|
return VKD3DSIH_DSX_COARSE;
|
|
case DX_DERIV_COARSEY:
|
|
return VKD3DSIH_DSY_COARSE;
|
|
case DX_DERIV_FINEX:
|
|
return VKD3DSIH_DSX_FINE;
|
|
case DX_DERIV_FINEY:
|
|
return VKD3DSIH_DSY_FINE;
|
|
case DX_LEGACY_F32TOF16:
|
|
return VKD3DSIH_F32TOF16;
|
|
case DX_LEGACY_F16TOF32:
|
|
return VKD3DSIH_F16TOF32;
|
|
case DX_WAVE_ACTIVE_ALL_EQUAL:
|
|
return VKD3DSIH_WAVE_ACTIVE_ALL_EQUAL;
|
|
case DX_WAVE_ALL_BIT_COUNT:
|
|
return VKD3DSIH_WAVE_ALL_BIT_COUNT;
|
|
case DX_WAVE_ALL_TRUE:
|
|
return VKD3DSIH_WAVE_ALL_TRUE;
|
|
case DX_WAVE_ANY_TRUE:
|
|
return VKD3DSIH_WAVE_ANY_TRUE;
|
|
case DX_WAVE_PREFIX_BIT_COUNT:
|
|
return VKD3DSIH_WAVE_PREFIX_BIT_COUNT;
|
|
case DX_WAVE_READ_LANE_FIRST:
|
|
return VKD3DSIH_WAVE_READ_LANE_FIRST;
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_unary(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, map_dx_unary_op(op));
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_from_value(src_param, operands[0]);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static enum vkd3d_shader_opcode map_dx_binary_op(enum dx_intrinsic_opcode op, const struct sm6_type *type)
|
|
{
|
|
switch (op)
|
|
{
|
|
case DX_FMAX:
|
|
return type->u.width == 64 ? VKD3DSIH_DMAX : VKD3DSIH_MAX;
|
|
case DX_FMIN:
|
|
return type->u.width == 64 ? VKD3DSIH_DMIN : VKD3DSIH_MIN;
|
|
case DX_IMAX:
|
|
return VKD3DSIH_IMAX;
|
|
case DX_IMIN:
|
|
return VKD3DSIH_IMIN;
|
|
case DX_QUAD_READ_LANE_AT:
|
|
return VKD3DSIH_QUAD_READ_LANE_AT;
|
|
case DX_UMAX:
|
|
return VKD3DSIH_UMAX;
|
|
case DX_UMIN:
|
|
return VKD3DSIH_UMIN;
|
|
case DX_WAVE_READ_LANE_AT:
|
|
return VKD3DSIH_WAVE_READ_LANE_AT;
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_binary(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, map_dx_binary_op(op, operands[0]->type));
|
|
if (!(src_params = instruction_src_params_alloc(ins, 2, sm6)))
|
|
return;
|
|
src_param_init_from_value(&src_params[0], operands[0]);
|
|
src_param_init_from_value(&src_params[1], operands[1]);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static enum vkd3d_shader_opcode map_dx_atomic_binop(const struct sm6_value *operand, struct sm6_parser *sm6)
|
|
{
|
|
uint64_t code = sm6_value_get_constant_uint(operand);
|
|
|
|
switch (code)
|
|
{
|
|
case ATOMIC_BINOP_ADD:
|
|
return VKD3DSIH_IMM_ATOMIC_IADD;
|
|
case ATOMIC_BINOP_AND:
|
|
return VKD3DSIH_IMM_ATOMIC_AND;
|
|
case ATOMIC_BINOP_IMAX:
|
|
return VKD3DSIH_IMM_ATOMIC_IMAX;
|
|
case ATOMIC_BINOP_IMIN:
|
|
return VKD3DSIH_IMM_ATOMIC_IMIN;
|
|
case ATOMIC_BINOP_OR:
|
|
return VKD3DSIH_IMM_ATOMIC_OR;
|
|
case ATOMIC_BINOP_UMAX:
|
|
return VKD3DSIH_IMM_ATOMIC_UMAX;
|
|
case ATOMIC_BINOP_UMIN:
|
|
return VKD3DSIH_IMM_ATOMIC_UMIN;
|
|
case ATOMIC_BINOP_XCHG:
|
|
return VKD3DSIH_IMM_ATOMIC_EXCH;
|
|
case ATOMIC_BINOP_XOR:
|
|
return VKD3DSIH_IMM_ATOMIC_XOR;
|
|
/* DXIL currently doesn't use SUB and NAND. */
|
|
default:
|
|
FIXME("Unhandled atomic binop %"PRIu64".\n", code);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Operation %"PRIu64" for an atomic binop instruction is unhandled.", code);
|
|
return VKD3DSIH_INVALID;
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_atomic_binop(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct sm6_value *dst = sm6_parser_get_current_value(sm6);
|
|
enum vkd3d_shader_resource_type resource_type;
|
|
bool is_cmp_xchg = op == DX_ATOMIC_CMP_XCHG;
|
|
unsigned int i, coord_idx, coord_count = 1;
|
|
struct vkd3d_shader_dst_param *dst_params;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
enum vkd3d_shader_opcode handler_idx;
|
|
struct vkd3d_shader_instruction *ins;
|
|
const struct sm6_value *resource;
|
|
struct vkd3d_shader_register reg;
|
|
|
|
resource = operands[0];
|
|
if (!sm6_value_validate_is_handle(resource, sm6))
|
|
return;
|
|
|
|
if (is_cmp_xchg)
|
|
handler_idx = VKD3DSIH_IMM_ATOMIC_CMP_EXCH;
|
|
else if ((handler_idx = map_dx_atomic_binop(operands[1], sm6)) == VKD3DSIH_INVALID)
|
|
return;
|
|
|
|
coord_idx = 2 - is_cmp_xchg;
|
|
resource_type = resource->u.handle.d->resource_type;
|
|
if (resource_type != VKD3D_SHADER_RESOURCE_BUFFER || resource->u.handle.d->kind == RESOURCE_KIND_STRUCTUREDBUFFER)
|
|
{
|
|
coord_count = 2 + (resource_type != VKD3D_SHADER_RESOURCE_BUFFER);
|
|
if (!sm6_parser_emit_coordinate_construct(sm6, &operands[coord_idx], coord_count, NULL, state, ®))
|
|
return;
|
|
}
|
|
else
|
|
{
|
|
reg = operands[coord_idx]->u.reg;
|
|
}
|
|
|
|
for (i = coord_idx + coord_count; i < coord_idx + 3; ++i)
|
|
{
|
|
if (!operands[i]->is_undefined)
|
|
{
|
|
WARN("Ignoring unexpected operand.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring an unexpected defined operand value for atomic instruction %u.", handler_idx);
|
|
break;
|
|
}
|
|
}
|
|
|
|
ins = state->ins;
|
|
vsir_instruction_init(ins, &sm6->p.location, handler_idx);
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 2 + is_cmp_xchg, sm6)))
|
|
return;
|
|
src_param_init_vector_from_reg(&src_params[0], ®);
|
|
if (is_cmp_xchg)
|
|
src_param_init_from_value(&src_params[1], operands[4]);
|
|
src_param_init_from_value(&src_params[1 + is_cmp_xchg], operands[5]);
|
|
|
|
dst_params = instruction_dst_params_alloc(ins, 2, sm6);
|
|
dst_param_init(&dst_params[0]);
|
|
register_init_ssa_scalar(&dst_params[0].reg, dst->type, dst, sm6);
|
|
dst_param_init(&dst_params[1]);
|
|
dst_params[1].reg = resource->u.handle.reg;
|
|
|
|
dst->u.reg = dst_params[0].reg;
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_barrier(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
enum dxil_sync_flags flags;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_SYNC);
|
|
flags = sm6_value_get_constant_uint(operands[0]);
|
|
ins->flags = flags & (SYNC_THREAD_GROUP | SYNC_THREAD_GROUP_UAV);
|
|
if (flags & SYNC_GLOBAL_UAV)
|
|
ins->flags |= VKD3DSSF_GLOBAL_UAV;
|
|
if (flags & SYNC_GROUP_SHARED_MEMORY)
|
|
ins->flags |= VKD3DSSF_GROUP_SHARED_MEMORY;
|
|
if (flags &= ~(SYNC_THREAD_GROUP | SYNC_GLOBAL_UAV | SYNC_THREAD_GROUP_UAV | SYNC_GROUP_SHARED_MEMORY))
|
|
{
|
|
FIXME("Unhandled flags %#x.\n", flags);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Barrier flags %#x are unhandled.", flags);
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_buffer_update_counter(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
const struct sm6_value *resource;
|
|
unsigned int i;
|
|
int8_t inc;
|
|
|
|
resource = operands[0];
|
|
if (!sm6_value_validate_is_handle(resource, sm6))
|
|
return;
|
|
|
|
if (!sm6_value_is_constant(operands[1]))
|
|
{
|
|
FIXME("Unsupported dynamic update operand.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"A dynamic update value for a UAV counter operation is not supported.");
|
|
return;
|
|
}
|
|
i = sm6_value_get_constant_uint(operands[1]);
|
|
if (i != 1 && i != 255)
|
|
{
|
|
WARN("Unexpected update value %#x.\n", i);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Update value %#x for a UAV counter operation is not supported.", i);
|
|
}
|
|
inc = i;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, (inc < 0) ? VKD3DSIH_IMM_ATOMIC_CONSUME : VKD3DSIH_IMM_ATOMIC_ALLOC);
|
|
if (!(src_params = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_vector_from_reg(&src_params[0], &resource->u.handle.reg);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_calculate_lod(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
const struct sm6_value *resource, *sampler;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_instruction *ins;
|
|
struct vkd3d_shader_register coord;
|
|
unsigned int clamp;
|
|
|
|
resource = operands[0];
|
|
sampler = operands[1];
|
|
if (!sm6_value_validate_is_texture_handle(resource, op, sm6)
|
|
|| !sm6_value_validate_is_sampler_handle(sampler, op, sm6))
|
|
{
|
|
return;
|
|
}
|
|
|
|
if (!sm6_parser_emit_coordinate_construct(sm6, &operands[2], 3, NULL, state, &coord))
|
|
return;
|
|
|
|
clamp = sm6_value_get_constant_uint(operands[5]);
|
|
|
|
ins = state->ins;
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_LOD);
|
|
if (!(src_params = instruction_src_params_alloc(ins, 3, sm6)))
|
|
return;
|
|
src_param_init_vector_from_reg(&src_params[0], &coord);
|
|
src_params[1].reg = resource->u.handle.reg;
|
|
src_param_init_scalar(&src_params[1], !clamp);
|
|
src_param_init_vector_from_reg(&src_params[2], &sampler->u.handle.reg);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_cbuffer_load(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct sm6_value *dst = sm6_parser_get_current_value(sm6);
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
const struct sm6_value *buffer;
|
|
const struct sm6_type *type;
|
|
|
|
buffer = operands[0];
|
|
if (!sm6_value_validate_is_handle(buffer, sm6))
|
|
return;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_vector_from_reg(src_param, &buffer->u.handle.reg);
|
|
/* Differently from other descriptors, constant buffers require an
|
|
* additional index, used to index within the constant buffer itself. */
|
|
src_param->reg.idx_count = 3;
|
|
register_index_address_init(&src_param->reg.idx[2], operands[1], sm6);
|
|
|
|
type = sm6_type_get_scalar_type(dst->type, 0);
|
|
VKD3D_ASSERT(type);
|
|
src_param->reg.data_type = vkd3d_data_type_from_sm6_type(type);
|
|
if (data_type_is_64_bit(src_param->reg.data_type))
|
|
src_param->swizzle = vsir_swizzle_64_from_32(src_param->swizzle);
|
|
|
|
instruction_dst_param_init_ssa_vector(ins, sm6_type_max_vector_size(type), sm6);
|
|
}
|
|
|
|
static void sm6_parser_dcl_register_builtin(struct sm6_parser *sm6, enum vkd3d_shader_opcode handler_idx,
|
|
enum vkd3d_shader_register_type reg_type, enum vkd3d_data_type data_type, unsigned int component_count)
|
|
{
|
|
struct vkd3d_shader_dst_param *dst_param;
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
if (!bitmap_is_set(sm6->io_regs_declared, reg_type))
|
|
{
|
|
bitmap_set(sm6->io_regs_declared, reg_type);
|
|
ins = sm6_parser_add_instruction(sm6, handler_idx);
|
|
dst_param = &ins->declaration.dst;
|
|
vsir_register_init(&dst_param->reg, reg_type, data_type, 0);
|
|
dst_param_init_vector(dst_param, component_count);
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_input_register_mov(struct sm6_parser *sm6,
|
|
struct vkd3d_shader_instruction *ins, enum vkd3d_shader_register_type reg_type, enum vkd3d_data_type data_type)
|
|
{
|
|
struct vkd3d_shader_src_param *src_param;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
sm6_parser_dcl_register_builtin(sm6, VKD3DSIH_DCL_INPUT, reg_type, data_type, 1);
|
|
vsir_register_init(&src_param->reg, reg_type, data_type, 0);
|
|
src_param_init(src_param);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_coverage(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
sm6_parser_emit_dx_input_register_mov(sm6, state->ins, VKD3DSPR_COVERAGE, VKD3D_DATA_UINT);
|
|
}
|
|
|
|
static const struct sm6_descriptor_info *sm6_parser_get_descriptor(struct sm6_parser *sm6,
|
|
enum vkd3d_shader_descriptor_type type, unsigned int id, const struct sm6_value *address)
|
|
{
|
|
const struct sm6_descriptor_info *d;
|
|
unsigned int register_index;
|
|
size_t i;
|
|
|
|
for (i = 0; i < sm6->descriptor_count; ++i)
|
|
{
|
|
d = &sm6->descriptors[i];
|
|
|
|
if (d->type != type || d->id != id)
|
|
continue;
|
|
|
|
if (!sm6_value_is_constant(address))
|
|
return d;
|
|
|
|
register_index = sm6_value_get_constant_uint(address);
|
|
if (register_index >= d->range.first && register_index <= d->range.last)
|
|
return d;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_create_handle(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
enum vkd3d_shader_descriptor_type type;
|
|
const struct sm6_descriptor_info *d;
|
|
struct vkd3d_shader_register *reg;
|
|
struct sm6_value *dst;
|
|
unsigned int id;
|
|
|
|
type = sm6_value_get_constant_uint(operands[0]);
|
|
id = sm6_value_get_constant_uint(operands[1]);
|
|
if (!(d = sm6_parser_get_descriptor(sm6, type, id, operands[2])))
|
|
{
|
|
WARN("Failed to find resource type %#x, id %#x.\n", type, id);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Descriptor for resource type %#x, id %#x was not found.", type, id);
|
|
return;
|
|
}
|
|
|
|
dst = sm6_parser_get_current_value(sm6);
|
|
dst->value_type = VALUE_TYPE_HANDLE;
|
|
dst->u.handle.d = d;
|
|
|
|
reg = &dst->u.handle.reg;
|
|
vsir_register_init(reg, d->reg_type, d->reg_data_type, 2);
|
|
reg->dimension = VSIR_DIMENSION_VEC4;
|
|
reg->idx[0].offset = id;
|
|
register_index_address_init(®->idx[1], operands[2], sm6);
|
|
reg->non_uniform = !!sm6_value_get_constant_uint(operands[3]);
|
|
|
|
/* NOP is used to flag no instruction emitted. */
|
|
ins->opcode = VKD3DSIH_NOP;
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_stream(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
unsigned int i;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, (op == DX_CUT_STREAM) ? VKD3DSIH_CUT_STREAM : VKD3DSIH_EMIT_STREAM);
|
|
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
|
|
i = sm6_value_get_constant_uint(operands[0]);
|
|
if (i >= MAX_GS_OUTPUT_STREAMS)
|
|
{
|
|
WARN("Invalid stream index %u.\n", i);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Output stream index %u is invalid.", i);
|
|
}
|
|
|
|
/* VKD3D_DATA_UNUSED would be more reasonable, but TPF uses data type 0 here. */
|
|
register_init_with_id(&src_param->reg, VKD3DSPR_STREAM, 0, i);
|
|
src_param_init(src_param);
|
|
|
|
if (op == DX_EMIT_THEN_CUT_STREAM)
|
|
{
|
|
++state->ins;
|
|
++state->code_block->instruction_count;
|
|
sm6_parser_emit_dx_stream(sm6, DX_CUT_STREAM, operands, state);
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_discard(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_DISCARD);
|
|
|
|
if ((src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
src_param_init_from_value(src_param, operands[0]);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_domain_location(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
unsigned int component_idx;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
|
|
if ((component_idx = sm6_value_get_constant_uint(operands[0])) >= 3)
|
|
{
|
|
WARN("Invalid component index %u.\n", component_idx);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid domain location component index %u.", component_idx);
|
|
component_idx = 0;
|
|
}
|
|
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
sm6_parser_dcl_register_builtin(sm6, VKD3DSIH_DCL_INPUT, VKD3DSPR_TESSCOORD, VKD3D_DATA_FLOAT, 3);
|
|
vsir_register_init(&src_param->reg, VKD3DSPR_TESSCOORD, VKD3D_DATA_FLOAT, 0);
|
|
src_param->reg.dimension = VSIR_DIMENSION_VEC4;
|
|
src_param_init_scalar(src_param, component_idx);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_dot(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_instruction *ins;
|
|
struct vkd3d_shader_register regs[2];
|
|
enum vkd3d_shader_opcode handler_idx;
|
|
unsigned int component_count;
|
|
|
|
switch (op)
|
|
{
|
|
case DX_DOT2:
|
|
handler_idx = VKD3DSIH_DP2;
|
|
component_count = 2;
|
|
break;
|
|
case DX_DOT3:
|
|
handler_idx = VKD3DSIH_DP3;
|
|
component_count = 3;
|
|
break;
|
|
case DX_DOT4:
|
|
handler_idx = VKD3DSIH_DP4;
|
|
component_count = 4;
|
|
break;
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
|
|
if (!sm6_parser_emit_composite_construct(sm6, &operands[0], component_count, state, ®s[0]))
|
|
return;
|
|
if (!sm6_parser_emit_composite_construct(sm6, &operands[component_count], component_count, state, ®s[1]))
|
|
return;
|
|
|
|
ins = state->ins;
|
|
vsir_instruction_init(ins, &sm6->p.location, handler_idx);
|
|
if (!(src_params = instruction_src_params_alloc(ins, 2, sm6)))
|
|
return;
|
|
src_param_init_vector_from_reg(&src_params[0], ®s[0]);
|
|
src_param_init_vector_from_reg(&src_params[1], ®s[1]);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_fabs(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_from_value(src_param, operands[0]);
|
|
src_param->modifiers = VKD3DSPSM_ABS;
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_compute_builtin(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
unsigned int component_count = 3, component_idx = 0;
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
enum vkd3d_shader_register_type reg_type;
|
|
|
|
switch (op)
|
|
{
|
|
case DX_THREAD_ID:
|
|
reg_type = VKD3DSPR_THREADID;
|
|
break;
|
|
case DX_GROUP_ID:
|
|
reg_type = VKD3DSPR_THREADGROUPID;
|
|
break;
|
|
case DX_THREAD_ID_IN_GROUP:
|
|
reg_type = VKD3DSPR_LOCALTHREADID;
|
|
break;
|
|
case DX_FLATTENED_THREAD_ID_IN_GROUP:
|
|
reg_type = VKD3DSPR_LOCALTHREADINDEX;
|
|
component_count = 1;
|
|
break;
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
|
|
sm6_parser_dcl_register_builtin(sm6, VKD3DSIH_DCL_INPUT, reg_type, VKD3D_DATA_UINT, component_count);
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
vsir_register_init(&src_param->reg, reg_type, VKD3D_DATA_UINT, 0);
|
|
if (component_count > 1)
|
|
{
|
|
src_param->reg.dimension = VSIR_DIMENSION_VEC4;
|
|
component_idx = sm6_value_get_constant_uint(operands[0]);
|
|
}
|
|
src_param_init_scalar(src_param, component_idx);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static enum vkd3d_shader_opcode sm6_dx_map_ma_op(enum dx_intrinsic_opcode op, const struct sm6_type *type)
|
|
{
|
|
switch (op)
|
|
{
|
|
case DX_FMA:
|
|
return VKD3DSIH_DFMA;
|
|
case DX_FMAD:
|
|
return VKD3DSIH_MAD;
|
|
case DX_IMAD:
|
|
case DX_UMAD:
|
|
return VKD3DSIH_IMAD;
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_ma(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
unsigned int i;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, sm6_dx_map_ma_op(op, operands[0]->type));
|
|
if (!(src_params = instruction_src_params_alloc(ins, 3, sm6)))
|
|
return;
|
|
for (i = 0; i < 3; ++i)
|
|
src_param_init_from_value(&src_params[i], operands[i]);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_get_dimensions(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
unsigned int is_texture, component_count;
|
|
enum dxil_resource_kind resource_kind;
|
|
struct vkd3d_shader_dst_param *dst;
|
|
const struct sm6_value *resource;
|
|
|
|
resource = operands[0];
|
|
if (!sm6_value_validate_is_handle(resource, sm6))
|
|
return;
|
|
is_texture = resource->u.handle.d->resource_type != VKD3D_SHADER_RESOURCE_BUFFER;
|
|
resource_kind = resource->u.handle.d->kind;
|
|
|
|
instruction_init_with_resource(ins, is_texture ? VKD3DSIH_RESINFO : VKD3DSIH_BUFINFO, resource, sm6);
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 1 + is_texture, sm6)))
|
|
return;
|
|
src_param_init_vector_from_reg(&src_params[is_texture], &resource->u.handle.reg);
|
|
|
|
if (is_texture)
|
|
{
|
|
ins->flags = VKD3DSI_RESINFO_UINT;
|
|
src_param_init_from_value(&src_params[0], operands[1]);
|
|
component_count = VKD3D_VEC4_SIZE;
|
|
|
|
if (resource_kind_is_multisampled(resource_kind))
|
|
{
|
|
instruction_dst_param_init_temp_vector(ins++, sm6);
|
|
state->temp_idx = 1;
|
|
|
|
/* DXIL does not have an intrinsic for sample info, and resinfo is expected to return
|
|
* the sample count in .w for MS textures. The result is always a struct of 4 x uint32. */
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_SAMPLE_INFO);
|
|
ins->flags = VKD3DSI_SAMPLE_INFO_UINT;
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_vector_from_reg(&src_params[0], &resource->u.handle.reg);
|
|
src_params[0].swizzle = VKD3D_SHADER_SWIZZLE(X, X, X, X);
|
|
|
|
if (!instruction_dst_param_init_temp_vector(ins, sm6))
|
|
return;
|
|
dst = ins->dst;
|
|
dst->write_mask = VKD3DSP_WRITEMASK_3;
|
|
|
|
/* Move the result to an SSA in case another instruction overwrites r0 before
|
|
* the components are extracted for use. */
|
|
++ins;
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
if (!(src_params = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_vector_from_reg(&src_params[0], &dst->reg);
|
|
|
|
state->ins = ins;
|
|
state->code_block->instruction_count += 2;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if (!operands[1]->is_undefined)
|
|
{
|
|
WARN("Ignoring unexpected operand.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring an unexpected defined LOD value for buffer GetDimensions.");
|
|
}
|
|
component_count = 1 + (resource_kind == RESOURCE_KIND_STRUCTUREDBUFFER);
|
|
}
|
|
|
|
instruction_dst_param_init_ssa_vector(ins, component_count, sm6);
|
|
}
|
|
|
|
static enum vkd3d_shader_opcode sm6_dx_map_tertiary_op(enum dx_intrinsic_opcode op)
|
|
{
|
|
switch (op)
|
|
{
|
|
case DX_IBFE:
|
|
return VKD3DSIH_IBFE;
|
|
case DX_UBFE:
|
|
return VKD3DSIH_UBFE;
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_tertiary(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
unsigned int i;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, sm6_dx_map_tertiary_op(op));
|
|
if (!(src_params = instruction_src_params_alloc(ins, 3, sm6)))
|
|
return;
|
|
for (i = 0; i < 3; ++i)
|
|
src_param_init_from_value(&src_params[i], operands[i]);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_load_input(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
bool is_control_point = op == DX_LOAD_OUTPUT_CONTROL_POINT;
|
|
bool is_patch_constant = op == DX_LOAD_PATCH_CONSTANT;
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vsir_program *program = sm6->p.program;
|
|
unsigned int count, row_index, column_index;
|
|
const struct vkd3d_shader_dst_param *params;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
const struct shader_signature *signature;
|
|
const struct signature_element *e;
|
|
|
|
row_index = sm6_value_get_constant_uint(operands[0]);
|
|
column_index = sm6_value_get_constant_uint(operands[2]);
|
|
|
|
if (is_control_point && operands[3]->is_undefined)
|
|
{
|
|
/* dxcompiler will compile source which does this, so let it pass. */
|
|
WARN("Control point id is undefined.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_UNDEFINED_OPERAND,
|
|
"The index for a control point load is undefined.");
|
|
}
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
|
|
if (is_patch_constant)
|
|
{
|
|
signature = &program->patch_constant_signature;
|
|
params = sm6->patch_constant_params;
|
|
}
|
|
else if (is_control_point)
|
|
{
|
|
signature = &program->output_signature;
|
|
params = sm6->output_params;
|
|
}
|
|
else
|
|
{
|
|
signature = &program->input_signature;
|
|
params = sm6->input_params;
|
|
}
|
|
if (row_index >= signature->element_count)
|
|
{
|
|
WARN("Invalid row index %u.\n", row_index);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid input row index %u.", row_index);
|
|
return;
|
|
}
|
|
e = &signature->elements[row_index];
|
|
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param->reg = params[row_index].reg;
|
|
src_param_init_scalar(src_param, column_index);
|
|
count = 0;
|
|
|
|
if (e->register_count > 1)
|
|
register_index_address_init(&src_param->reg.idx[count++], operands[1], sm6);
|
|
|
|
if (!is_patch_constant && !operands[3]->is_undefined)
|
|
{
|
|
VKD3D_ASSERT(src_param->reg.idx_count > count);
|
|
register_index_address_init(&src_param->reg.idx[count], operands[3], sm6);
|
|
}
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_make_double(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_instruction *ins;
|
|
struct vkd3d_shader_register reg;
|
|
|
|
if (!sm6_parser_emit_composite_construct(sm6, &operands[0], 2, state, ®))
|
|
return;
|
|
|
|
ins = state->ins;
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
if (!(src_params = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_params[0].reg = reg;
|
|
src_param_init_vector(&src_params[0], 2);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_output_control_point_id(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
sm6_parser_emit_dx_input_register_mov(sm6, state->ins, VKD3DSPR_OUTPOINTID, VKD3D_DATA_UINT);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_primitive_id(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
sm6_parser_emit_dx_input_register_mov(sm6, state->ins, VKD3DSPR_PRIMID, VKD3D_DATA_UINT);
|
|
}
|
|
|
|
static enum vkd3d_shader_opcode dx_map_quad_op(enum dxil_quad_op_kind op)
|
|
{
|
|
switch (op)
|
|
{
|
|
case QUAD_READ_ACROSS_X:
|
|
return VKD3DSIH_QUAD_READ_ACROSS_X;
|
|
case QUAD_READ_ACROSS_Y:
|
|
return VKD3DSIH_QUAD_READ_ACROSS_Y;
|
|
case QUAD_READ_ACROSS_D:
|
|
return VKD3DSIH_QUAD_READ_ACROSS_D;
|
|
default:
|
|
return VKD3DSIH_INVALID;
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_quad_op(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
enum vkd3d_shader_opcode opcode;
|
|
enum dxil_quad_op_kind quad_op;
|
|
|
|
quad_op = sm6_value_get_constant_uint(operands[1]);
|
|
if ((opcode = dx_map_quad_op(quad_op)) == VKD3DSIH_INVALID)
|
|
{
|
|
FIXME("Unhandled quad op kind %u.\n", quad_op);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_UNHANDLED_INTRINSIC,
|
|
"Quad op kind %u is unhandled.", quad_op);
|
|
return;
|
|
}
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, opcode);
|
|
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_from_value(src_param, operands[0]);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_raw_buffer_load(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
unsigned int operand_count, write_mask, component_count = VKD3D_VEC4_SIZE;
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
const struct sm6_value *resource;
|
|
bool raw;
|
|
|
|
resource = operands[0];
|
|
if (!sm6_value_validate_is_handle(resource, sm6))
|
|
return;
|
|
raw = resource->u.handle.d->kind == RESOURCE_KIND_RAWBUFFER;
|
|
|
|
if (op == DX_RAW_BUFFER_LOAD)
|
|
{
|
|
write_mask = sm6_value_get_constant_uint(operands[3]);
|
|
if (!write_mask || write_mask > VKD3DSP_WRITEMASK_ALL)
|
|
{
|
|
WARN("Invalid write mask %#x.\n", write_mask);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Write mask %#x for a raw/structured buffer load operation is invalid.", write_mask);
|
|
return;
|
|
}
|
|
else if (write_mask & (write_mask + 1))
|
|
{
|
|
FIXME("Unhandled write mask %#x.\n", write_mask);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Write mask %#x for a raw/structured buffer load operation is unhandled.", write_mask);
|
|
}
|
|
component_count = vsir_write_mask_component_count(write_mask);
|
|
}
|
|
|
|
instruction_init_with_resource(ins, raw ? VKD3DSIH_LD_RAW : VKD3DSIH_LD_STRUCTURED, resource, sm6);
|
|
operand_count = 2 + !raw;
|
|
if (!(src_params = instruction_src_params_alloc(ins, operand_count, sm6)))
|
|
return;
|
|
src_params_init_from_operands(src_params, &operands[1], operand_count - 1);
|
|
src_param_init_vector_from_reg(&src_params[operand_count - 1], &resource->u.handle.reg);
|
|
|
|
instruction_dst_param_init_ssa_vector(ins, component_count, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_raw_buffer_store(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
unsigned int write_mask, component_count, alignment = 0, operand_count;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_dst_param *dst_param;
|
|
struct vkd3d_shader_instruction *ins;
|
|
struct vkd3d_shader_register data;
|
|
const struct sm6_value *resource;
|
|
bool raw;
|
|
|
|
resource = operands[0];
|
|
if (!sm6_value_validate_is_handle(resource, sm6))
|
|
return;
|
|
raw = resource->u.handle.d->kind == RESOURCE_KIND_RAWBUFFER;
|
|
|
|
write_mask = sm6_value_get_constant_uint(operands[7]);
|
|
if (!write_mask || write_mask > VKD3DSP_WRITEMASK_ALL)
|
|
{
|
|
WARN("Invalid write mask %#x.\n", write_mask);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Write mask %#x for a raw/structured buffer store operation is invalid.", write_mask);
|
|
return;
|
|
}
|
|
else if (write_mask & (write_mask + 1))
|
|
{
|
|
/* In this case, it is unclear which source operands will be defined unless we encounter it in a shader. */
|
|
FIXME("Unhandled write mask %#x.\n", write_mask);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Write mask %#x for a raw/structured buffer store operation is unhandled.", write_mask);
|
|
}
|
|
component_count = vsir_write_mask_component_count(write_mask);
|
|
|
|
if (op == DX_RAW_BUFFER_STORE)
|
|
{
|
|
if (!raw && resource->u.handle.d->kind != RESOURCE_KIND_STRUCTUREDBUFFER)
|
|
{
|
|
WARN("Resource is not a raw or structured buffer.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_INVALID_OPERATION,
|
|
"Resource for a raw buffer store is not a raw or structured buffer.");
|
|
}
|
|
|
|
alignment = sm6_value_get_constant_uint(operands[8]);
|
|
if (alignment & (alignment - 1))
|
|
{
|
|
FIXME("Invalid alignment %#x.\n", alignment);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Alignment %#x for a raw/structured buffer store operation is invalid.", alignment);
|
|
}
|
|
}
|
|
|
|
if (!sm6_parser_emit_composite_construct(sm6, &operands[3], component_count, state, &data))
|
|
return;
|
|
|
|
ins = state->ins;
|
|
vsir_instruction_init(ins, &sm6->p.location, raw ? VKD3DSIH_STORE_RAW : VKD3DSIH_STORE_STRUCTURED);
|
|
operand_count = 2 + !raw;
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, operand_count, sm6)))
|
|
return;
|
|
src_params_init_from_operands(src_params, &operands[1], operand_count - 1);
|
|
data.data_type = VKD3D_DATA_UINT;
|
|
src_param_init_vector_from_reg(&src_params[operand_count - 1], &data);
|
|
|
|
dst_param = instruction_dst_params_alloc(ins, 1, sm6);
|
|
dst_param_init_with_mask(dst_param, write_mask);
|
|
dst_param->reg = resource->u.handle.reg;
|
|
dst_param->reg.alignment = alignment;
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_buffer_load(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
const struct sm6_value *resource;
|
|
|
|
resource = operands[0];
|
|
if (!sm6_value_validate_is_handle(resource, sm6))
|
|
return;
|
|
|
|
if (resource->u.handle.d->kind == RESOURCE_KIND_RAWBUFFER
|
|
|| resource->u.handle.d->kind == RESOURCE_KIND_STRUCTUREDBUFFER)
|
|
{
|
|
return sm6_parser_emit_dx_raw_buffer_load(sm6, op, operands, state);
|
|
}
|
|
|
|
if (resource->u.handle.d->kind != RESOURCE_KIND_TYPEDBUFFER)
|
|
{
|
|
WARN("Resource is not a typed buffer.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_INVALID_OPERATION,
|
|
"Resource for a typed buffer load is not a typed buffer.");
|
|
}
|
|
|
|
instruction_init_with_resource(ins, (resource->u.handle.d->type == VKD3D_SHADER_DESCRIPTOR_TYPE_UAV)
|
|
? VKD3DSIH_LD_UAV_TYPED : VKD3DSIH_LD, resource, sm6);
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 2, sm6)))
|
|
return;
|
|
src_param_init_from_value(&src_params[0], operands[1]);
|
|
if (!sm6_value_is_undef(operands[2]))
|
|
{
|
|
/* Constant zero would be ok, but is not worth checking for unless it shows up. */
|
|
WARN("Ignoring structure offset.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring structure offset for a typed buffer load.");
|
|
}
|
|
src_param_init_vector_from_reg(&src_params[1], &resource->u.handle.reg);
|
|
|
|
instruction_dst_param_init_ssa_vector(ins, VKD3D_VEC4_SIZE, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_buffer_store(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_dst_param *dst_param;
|
|
unsigned int write_mask, component_count;
|
|
struct vkd3d_shader_instruction *ins;
|
|
struct vkd3d_shader_register texel;
|
|
const struct sm6_value *resource;
|
|
|
|
resource = operands[0];
|
|
if (!sm6_value_validate_is_handle(resource, sm6))
|
|
return;
|
|
|
|
if (resource->u.handle.d->kind == RESOURCE_KIND_RAWBUFFER
|
|
|| resource->u.handle.d->kind == RESOURCE_KIND_STRUCTUREDBUFFER)
|
|
{
|
|
return sm6_parser_emit_dx_raw_buffer_store(sm6, op, operands, state);
|
|
}
|
|
|
|
if (resource->u.handle.d->kind != RESOURCE_KIND_TYPEDBUFFER)
|
|
{
|
|
WARN("Resource is not a typed buffer.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_INVALID_OPERATION,
|
|
"Resource for a typed buffer store is not a typed buffer.");
|
|
}
|
|
|
|
write_mask = sm6_value_get_constant_uint(operands[7]);
|
|
if (!write_mask || write_mask > VKD3DSP_WRITEMASK_ALL)
|
|
{
|
|
WARN("Invalid write mask %#x.\n", write_mask);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Write mask %#x for a typed buffer store operation is invalid.", write_mask);
|
|
return;
|
|
}
|
|
else if (write_mask & (write_mask + 1))
|
|
{
|
|
/* In this case, it is unclear which source operands will be defined unless we encounter it in a shader. */
|
|
FIXME("Unhandled write mask %#x.\n", write_mask);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Write mask %#x for a typed buffer store operation is unhandled.", write_mask);
|
|
}
|
|
component_count = vsir_write_mask_component_count(write_mask);
|
|
|
|
if (!sm6_parser_emit_composite_construct(sm6, &operands[3], component_count, state, &texel))
|
|
return;
|
|
|
|
ins = state->ins;
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_STORE_UAV_TYPED);
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 2, sm6)))
|
|
return;
|
|
src_param_init_from_value(&src_params[0], operands[1]);
|
|
if (!sm6_value_is_undef(operands[2]))
|
|
{
|
|
/* Constant zero would have no effect, but is not worth checking for unless it shows up. */
|
|
WARN("Ignoring structure offset.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring structure offset for a typed buffer store.");
|
|
}
|
|
src_param_init_vector_from_reg(&src_params[1], &texel);
|
|
|
|
dst_param = instruction_dst_params_alloc(ins, 1, sm6);
|
|
dst_param_init_with_mask(dst_param, write_mask);
|
|
dst_param->reg = resource->u.handle.reg;
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_get_sample_count(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_SAMPLE_INFO);
|
|
ins->flags = VKD3DSI_SAMPLE_INFO_UINT;
|
|
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
vsir_register_init(&src_param->reg, VKD3DSPR_RASTERIZER, VKD3D_DATA_FLOAT, 0);
|
|
src_param->reg.dimension = VSIR_DIMENSION_VEC4;
|
|
src_param_init(src_param);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
ins->dst->reg.data_type = VKD3D_DATA_FLOAT;
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_get_sample_pos(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
const struct sm6_value *resource = NULL;
|
|
|
|
if (op == DX_TEX2DMS_GET_SAMPLE_POS)
|
|
{
|
|
resource = operands[0];
|
|
if (!sm6_value_validate_is_texture_2dms_handle(resource, op, sm6))
|
|
return;
|
|
}
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_SAMPLE_POS);
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 2, sm6)))
|
|
return;
|
|
if (op == DX_TEX2DMS_GET_SAMPLE_POS)
|
|
{
|
|
src_param_init_vector_from_reg(&src_params[0], &resource->u.handle.reg);
|
|
src_param_init_from_value(&src_params[1], operands[1]);
|
|
}
|
|
else
|
|
{
|
|
src_param_init_vector(&src_params[0], 2);
|
|
vsir_register_init(&src_params[0].reg, VKD3DSPR_RASTERIZER, VKD3D_DATA_FLOAT, 0);
|
|
src_params[0].reg.dimension = VSIR_DIMENSION_VEC4;
|
|
src_param_init_from_value(&src_params[1], operands[0]);
|
|
}
|
|
|
|
instruction_dst_param_init_ssa_vector(ins, 2, sm6);
|
|
}
|
|
|
|
static unsigned int sm6_value_get_texel_offset(const struct sm6_value *value)
|
|
{
|
|
return sm6_value_is_undef(value) ? 0 : sm6_value_get_constant_uint(value);
|
|
}
|
|
|
|
static void instruction_set_texel_offset(struct vkd3d_shader_instruction *ins,
|
|
const struct sm6_value **operands, struct sm6_parser *sm6)
|
|
{
|
|
ins->texel_offset.u = sm6_value_get_texel_offset(operands[0]);
|
|
ins->texel_offset.v = sm6_value_get_texel_offset(operands[1]);
|
|
ins->texel_offset.w = sm6_value_get_texel_offset(operands[2]);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_sample(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
unsigned int clamp_idx = 0, component_count = VKD3D_VEC4_SIZE;
|
|
struct vkd3d_shader_register coord, ddx, ddy;
|
|
const struct sm6_value *resource, *sampler;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
resource = operands[0];
|
|
sampler = operands[1];
|
|
if (!sm6_value_validate_is_texture_handle(resource, op, sm6)
|
|
|| !sm6_value_validate_is_sampler_handle(sampler, op, sm6))
|
|
{
|
|
return;
|
|
}
|
|
|
|
if (!sm6_parser_emit_coordinate_construct(sm6, &operands[2], VKD3D_VEC4_SIZE, NULL, state, &coord))
|
|
return;
|
|
|
|
if (op == DX_SAMPLE_GRAD)
|
|
{
|
|
if (!sm6_parser_emit_coordinate_construct(sm6, &operands[9], 3, NULL, state, &ddx))
|
|
return;
|
|
if (!sm6_parser_emit_coordinate_construct(sm6, &operands[12], 3, NULL, state, &ddy))
|
|
return;
|
|
}
|
|
|
|
ins = state->ins;
|
|
switch (op)
|
|
{
|
|
case DX_SAMPLE:
|
|
instruction_init_with_resource(ins, VKD3DSIH_SAMPLE, resource, sm6);
|
|
src_params = instruction_src_params_alloc(ins, 3, sm6);
|
|
clamp_idx = 9;
|
|
break;
|
|
case DX_SAMPLE_B:
|
|
clamp_idx = 10;
|
|
/* fall through */
|
|
case DX_SAMPLE_LOD:
|
|
instruction_init_with_resource(ins, (op == DX_SAMPLE_B) ? VKD3DSIH_SAMPLE_B : VKD3DSIH_SAMPLE_LOD,
|
|
resource, sm6);
|
|
src_params = instruction_src_params_alloc(ins, 4, sm6);
|
|
src_param_init_from_value(&src_params[3], operands[9]);
|
|
break;
|
|
case DX_SAMPLE_C:
|
|
clamp_idx = 10;
|
|
/* fall through */
|
|
case DX_SAMPLE_C_LZ:
|
|
instruction_init_with_resource(ins, (op == DX_SAMPLE_C_LZ) ? VKD3DSIH_SAMPLE_C_LZ : VKD3DSIH_SAMPLE_C,
|
|
resource, sm6);
|
|
src_params = instruction_src_params_alloc(ins, 4, sm6);
|
|
src_param_init_from_value(&src_params[3], operands[9]);
|
|
component_count = 1;
|
|
break;
|
|
case DX_SAMPLE_GRAD:
|
|
instruction_init_with_resource(ins, VKD3DSIH_SAMPLE_GRAD, resource, sm6);
|
|
src_params = instruction_src_params_alloc(ins, 5, sm6);
|
|
src_param_init_vector_from_reg(&src_params[3], &ddx);
|
|
src_param_init_vector_from_reg(&src_params[4], &ddy);
|
|
clamp_idx = 15;
|
|
break;
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
|
|
if (!src_params)
|
|
return;
|
|
|
|
if (clamp_idx && !sm6_value_is_undef(operands[clamp_idx]))
|
|
{
|
|
FIXME("Ignoring LOD clamp value.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring LOD clamp value for a sample operation.");
|
|
}
|
|
|
|
src_param_init_vector_from_reg(&src_params[0], &coord);
|
|
src_param_init_vector_from_reg(&src_params[1], &resource->u.handle.reg);
|
|
src_param_init_vector_from_reg(&src_params[2], &sampler->u.handle.reg);
|
|
instruction_set_texel_offset(ins, &operands[6], sm6);
|
|
|
|
instruction_dst_param_init_ssa_vector(ins, component_count, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_saturate(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_from_value(src_param, operands[0]);
|
|
|
|
if (instruction_dst_param_init_ssa_scalar(ins, sm6))
|
|
ins->dst->modifiers = VKD3DSPDM_SATURATE;
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_sincos(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct sm6_value *dst = sm6_parser_get_current_value(sm6);
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_dst_param *dst_params;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
unsigned int index;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_SINCOS);
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_from_value(src_param, operands[0]);
|
|
|
|
index = op == DX_COS;
|
|
dst_params = instruction_dst_params_alloc(ins, 2, sm6);
|
|
dst_param_init(&dst_params[0]);
|
|
dst_param_init(&dst_params[1]);
|
|
register_init_ssa_scalar(&dst_params[index].reg, dst->type, dst, sm6);
|
|
vsir_register_init(&dst_params[index ^ 1].reg, VKD3DSPR_NULL, VKD3D_DATA_UNUSED, 0);
|
|
dst->u.reg = dst_params[index].reg;
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_split_double(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_from_value(src_param, operands[0]);
|
|
|
|
instruction_dst_param_init_ssa_vector(ins, 2, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_store_output(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
bool is_patch_constant = op == DX_STORE_PATCH_CONSTANT;
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vsir_program *program = sm6->p.program;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
struct vkd3d_shader_dst_param *dst_param;
|
|
const struct shader_signature *signature;
|
|
unsigned int row_index, column_index;
|
|
const struct signature_element *e;
|
|
const struct sm6_value *value;
|
|
|
|
row_index = sm6_value_get_constant_uint(operands[0]);
|
|
column_index = sm6_value_get_constant_uint(operands[2]);
|
|
|
|
signature = is_patch_constant ? &program->patch_constant_signature : &program->output_signature;
|
|
if (row_index >= signature->element_count)
|
|
{
|
|
WARN("Invalid row index %u.\n", row_index);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid output row index %u.", row_index);
|
|
return;
|
|
}
|
|
e = &signature->elements[row_index];
|
|
|
|
if (column_index >= VKD3D_VEC4_SIZE)
|
|
{
|
|
WARN("Invalid column index %u.\n", column_index);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid output column index %u.", column_index);
|
|
return;
|
|
}
|
|
|
|
value = operands[3];
|
|
if (!sm6_value_is_register(value))
|
|
{
|
|
WARN("Source value is not a register.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Expected store operation source to be a register.");
|
|
return;
|
|
}
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
|
|
if (!(dst_param = instruction_dst_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
dst_param_init_scalar(dst_param, column_index);
|
|
dst_param->reg = is_patch_constant ? sm6->patch_constant_params[row_index].reg : sm6->output_params[row_index].reg;
|
|
if (e->register_count > 1)
|
|
register_index_address_init(&dst_param->reg.idx[0], operands[1], sm6);
|
|
|
|
if (e->register_index == UINT_MAX)
|
|
{
|
|
sm6_parser_dcl_register_builtin(sm6, VKD3DSIH_DCL_OUTPUT, dst_param->reg.type,
|
|
dst_param->reg.data_type, vsir_write_mask_component_count(e->mask));
|
|
}
|
|
|
|
if ((src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
src_param_init_from_value(src_param, value);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_texture_gather(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_register coord, offset;
|
|
const struct sm6_value *resource, *sampler;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_instruction *ins;
|
|
unsigned int swizzle;
|
|
bool extended_offset;
|
|
|
|
resource = operands[0];
|
|
sampler = operands[1];
|
|
if (!sm6_value_validate_is_texture_handle(resource, op, sm6)
|
|
|| !sm6_value_validate_is_sampler_handle(sampler, op, sm6))
|
|
{
|
|
return;
|
|
}
|
|
|
|
if (!sm6_parser_emit_coordinate_construct(sm6, &operands[2], VKD3D_VEC4_SIZE, NULL, state, &coord))
|
|
return;
|
|
|
|
if ((extended_offset = !sm6_value_vector_is_constant_or_undef(&operands[6], 2))
|
|
&& !sm6_parser_emit_coordinate_construct(sm6, &operands[6], 2, NULL, state, &offset))
|
|
{
|
|
return;
|
|
}
|
|
|
|
ins = state->ins;
|
|
if (op == DX_TEXTURE_GATHER)
|
|
{
|
|
instruction_init_with_resource(ins, extended_offset ? VKD3DSIH_GATHER4_PO : VKD3DSIH_GATHER4, resource, sm6);
|
|
if (!(src_params = instruction_src_params_alloc(ins, 3 + extended_offset, sm6)))
|
|
return;
|
|
}
|
|
else
|
|
{
|
|
instruction_init_with_resource(ins, extended_offset ? VKD3DSIH_GATHER4_PO_C : VKD3DSIH_GATHER4_C, resource, sm6);
|
|
if (!(src_params = instruction_src_params_alloc(ins, 4 + extended_offset, sm6)))
|
|
return;
|
|
src_param_init_from_value(&src_params[3 + extended_offset], operands[9]);
|
|
}
|
|
|
|
src_param_init_vector_from_reg(&src_params[0], &coord);
|
|
if (extended_offset)
|
|
src_param_init_vector_from_reg(&src_params[1], &offset);
|
|
else
|
|
instruction_set_texel_offset(ins, &operands[6], sm6);
|
|
src_param_init_vector_from_reg(&src_params[1 + extended_offset], &resource->u.handle.reg);
|
|
src_param_init_vector_from_reg(&src_params[2 + extended_offset], &sampler->u.handle.reg);
|
|
/* Swizzle stored in the sampler parameter is the scalar component index to be gathered. */
|
|
swizzle = sm6_value_get_constant_uint(operands[8]);
|
|
if (swizzle >= VKD3D_VEC4_SIZE)
|
|
{
|
|
WARN("Invalid swizzle %#x.\n", swizzle);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Swizzle %#x for a texture gather operation is invalid.", swizzle);
|
|
}
|
|
src_params[2 + extended_offset].swizzle = swizzle;
|
|
|
|
instruction_dst_param_init_ssa_vector(ins, VKD3D_VEC4_SIZE, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_texture_load(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
const struct sm6_value *resource, *mip_level_or_sample_count;
|
|
enum vkd3d_shader_resource_type resource_type;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_instruction *ins;
|
|
struct vkd3d_shader_register coord;
|
|
bool is_multisample, is_uav;
|
|
unsigned int i;
|
|
|
|
resource = operands[0];
|
|
if (!sm6_value_validate_is_texture_handle(resource, op, sm6))
|
|
return;
|
|
|
|
resource_type = resource->u.handle.d->resource_type;
|
|
is_multisample = resource_type == VKD3D_SHADER_RESOURCE_TEXTURE_2DMS
|
|
|| resource_type == VKD3D_SHADER_RESOURCE_TEXTURE_2DMSARRAY;
|
|
is_uav = resource->u.handle.d->type == VKD3D_SHADER_DESCRIPTOR_TYPE_UAV;
|
|
|
|
mip_level_or_sample_count = (resource_type != VKD3D_SHADER_RESOURCE_BUFFER) ? operands[1] : NULL;
|
|
if (!sm6_parser_emit_coordinate_construct(sm6, &operands[2], 3,
|
|
is_multisample ? NULL : mip_level_or_sample_count, state, &coord))
|
|
{
|
|
return;
|
|
}
|
|
|
|
ins = state->ins;
|
|
instruction_init_with_resource(ins, is_uav ? VKD3DSIH_LD_UAV_TYPED
|
|
: is_multisample ? VKD3DSIH_LD2DMS : VKD3DSIH_LD, resource, sm6);
|
|
instruction_set_texel_offset(ins, &operands[5], sm6);
|
|
|
|
for (i = 0; i < VKD3D_VEC4_SIZE; ++i)
|
|
ins->resource_data_type[i] = resource->u.handle.d->resource_data_type;
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 2 + is_multisample, sm6)))
|
|
return;
|
|
src_param_init_vector_from_reg(&src_params[0], &coord);
|
|
src_param_init_vector_from_reg(&src_params[1], &resource->u.handle.reg);
|
|
if (is_multisample)
|
|
src_param_init_from_value(&src_params[2], mip_level_or_sample_count);
|
|
|
|
instruction_dst_param_init_ssa_vector(ins, VKD3D_VEC4_SIZE, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_texture_store(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_register coord, texel;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_dst_param *dst_param;
|
|
unsigned int write_mask, component_count;
|
|
struct vkd3d_shader_instruction *ins;
|
|
const struct sm6_value *resource;
|
|
|
|
resource = operands[0];
|
|
if (!sm6_value_validate_is_texture_handle(resource, op, sm6))
|
|
return;
|
|
|
|
if (!sm6_parser_emit_coordinate_construct(sm6, &operands[1], 3, NULL, state, &coord))
|
|
return;
|
|
|
|
write_mask = sm6_value_get_constant_uint(operands[8]);
|
|
if (!write_mask || write_mask > VKD3DSP_WRITEMASK_ALL)
|
|
{
|
|
WARN("Invalid write mask %#x.\n", write_mask);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Write mask %#x for a texture store operation is invalid.", write_mask);
|
|
return;
|
|
}
|
|
else if (write_mask & (write_mask + 1))
|
|
{
|
|
/* In this case, it is unclear which source operands will be defined unless we encounter it in a shader. */
|
|
FIXME("Unhandled write mask %#x.\n", write_mask);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Write mask %#x for a texture store operation is unhandled.", write_mask);
|
|
}
|
|
component_count = vsir_write_mask_component_count(write_mask);
|
|
|
|
if (!sm6_parser_emit_composite_construct(sm6, &operands[4], component_count, state, &texel))
|
|
return;
|
|
|
|
ins = state->ins;
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_STORE_UAV_TYPED);
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 2, sm6)))
|
|
return;
|
|
src_param_init_vector_from_reg(&src_params[0], &coord);
|
|
src_param_init_vector_from_reg(&src_params[1], &texel);
|
|
|
|
dst_param = instruction_dst_params_alloc(ins, 1, sm6);
|
|
dst_param->reg = resource->u.handle.reg;
|
|
dst_param_init_with_mask(dst_param, write_mask);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_wave_active_ballot(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_WAVE_ACTIVE_BALLOT);
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_from_value(src_param, operands[0]);
|
|
|
|
instruction_dst_param_init_ssa_vector(ins, VKD3D_VEC4_SIZE, sm6);
|
|
}
|
|
|
|
static enum vkd3d_shader_opcode sm6_dx_map_wave_bit_op(enum dxil_wave_bit_op_kind op,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
switch (op)
|
|
{
|
|
case WAVE_BIT_OP_AND:
|
|
return VKD3DSIH_WAVE_ACTIVE_BIT_AND;
|
|
case WAVE_BIT_OP_OR:
|
|
return VKD3DSIH_WAVE_ACTIVE_BIT_OR;
|
|
case WAVE_BIT_OP_XOR:
|
|
return VKD3DSIH_WAVE_ACTIVE_BIT_XOR;
|
|
default:
|
|
FIXME("Unhandled wave bit op %u.\n", op);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_UNHANDLED_INTRINSIC,
|
|
"Wave bit operation %u is unhandled.\n", op);
|
|
return VKD3DSIH_INVALID;
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_wave_active_bit(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
enum dxil_wave_bit_op_kind wave_op;
|
|
enum vkd3d_shader_opcode opcode;
|
|
|
|
wave_op = sm6_value_get_constant_uint(operands[1]);
|
|
|
|
if ((opcode = sm6_dx_map_wave_bit_op(wave_op, sm6)) == VKD3DSIH_INVALID)
|
|
return;
|
|
vsir_instruction_init(ins, &sm6->p.location, opcode);
|
|
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_from_value(src_param, operands[0]);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static enum vkd3d_shader_opcode sm6_dx_map_wave_op(enum dxil_wave_op_kind op, bool is_signed, bool is_float,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
switch (op)
|
|
{
|
|
case WAVE_OP_ADD:
|
|
return VKD3DSIH_WAVE_OP_ADD;
|
|
case WAVE_OP_MUL:
|
|
return VKD3DSIH_WAVE_OP_MUL;
|
|
case WAVE_OP_MIN:
|
|
if (is_float)
|
|
return VKD3DSIH_WAVE_OP_MIN;
|
|
return is_signed ? VKD3DSIH_WAVE_OP_IMIN : VKD3DSIH_WAVE_OP_UMIN;
|
|
case WAVE_OP_MAX:
|
|
if (is_float)
|
|
return VKD3DSIH_WAVE_OP_MAX;
|
|
return is_signed ? VKD3DSIH_WAVE_OP_IMAX : VKD3DSIH_WAVE_OP_UMAX;
|
|
default:
|
|
FIXME("Unhandled wave op %u.\n", op);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_UNHANDLED_INTRINSIC,
|
|
"Wave operation %u is unhandled.\n", op);
|
|
return VKD3DSIH_INVALID;
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_wave_op(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
struct vkd3d_shader_instruction *ins = state->ins;
|
|
struct vkd3d_shader_src_param *src_param;
|
|
enum vkd3d_shader_opcode opcode;
|
|
enum dxil_wave_op_kind wave_op;
|
|
bool is_signed;
|
|
|
|
wave_op = sm6_value_get_constant_uint(operands[1]);
|
|
is_signed = !sm6_value_get_constant_uint(operands[2]);
|
|
opcode = sm6_dx_map_wave_op(wave_op, is_signed, sm6_type_is_floating_point(operands[0]->type), sm6);
|
|
|
|
if (opcode == VKD3DSIH_INVALID)
|
|
return;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, opcode);
|
|
ins->flags = (op == DX_WAVE_PREFIX_OP) ? VKD3DSI_WAVE_PREFIX : 0;
|
|
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_from_value(src_param, operands[0]);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_dx_wave_builtin(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const struct sm6_value **operands, struct function_emission_state *state)
|
|
{
|
|
enum vkd3d_shader_register_type type;
|
|
|
|
switch (op)
|
|
{
|
|
case DX_WAVE_GET_LANE_COUNT:
|
|
type = VKD3DSPR_WAVELANECOUNT;
|
|
break;
|
|
case DX_WAVE_GET_LANE_INDEX:
|
|
type = VKD3DSPR_WAVELANEINDEX;
|
|
break;
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
|
|
sm6_parser_emit_dx_input_register_mov(sm6, state->ins, type, VKD3D_DATA_UINT);
|
|
}
|
|
|
|
struct sm6_dx_opcode_info
|
|
{
|
|
const char *ret_type;
|
|
const char *operand_info;
|
|
void (*handler)(struct sm6_parser *, enum dx_intrinsic_opcode, const struct sm6_value **,
|
|
struct function_emission_state *);
|
|
};
|
|
|
|
/*
|
|
1 -> int1
|
|
8 -> int8
|
|
b -> constant int1
|
|
c -> constant int8/16/32
|
|
C -> constant or undefined int8/16/32
|
|
i -> int32
|
|
m -> int16/32/64
|
|
n -> any numeric
|
|
f -> float
|
|
d -> double
|
|
e -> half/float
|
|
g -> half/float/double
|
|
H -> handle
|
|
D -> Dimensions
|
|
S -> splitdouble
|
|
V -> 4 x i32
|
|
v -> void
|
|
o -> overloaded
|
|
R -> matches the return type
|
|
*/
|
|
static const struct sm6_dx_opcode_info sm6_dx_op_table[] =
|
|
{
|
|
[DX_ACOS ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_ASIN ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_ATAN ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_ATOMIC_BINOP ] = {"o", "HciiiR", sm6_parser_emit_dx_atomic_binop},
|
|
[DX_ATOMIC_CMP_XCHG ] = {"o", "HiiiRR", sm6_parser_emit_dx_atomic_binop},
|
|
[DX_BARRIER ] = {"v", "c", sm6_parser_emit_dx_barrier},
|
|
[DX_BFREV ] = {"m", "R", sm6_parser_emit_dx_unary},
|
|
[DX_BUFFER_LOAD ] = {"o", "Hii", sm6_parser_emit_dx_buffer_load},
|
|
[DX_BUFFER_STORE ] = {"v", "Hiiooooc", sm6_parser_emit_dx_buffer_store},
|
|
[DX_BUFFER_UPDATE_COUNTER ] = {"i", "H8", sm6_parser_emit_dx_buffer_update_counter},
|
|
[DX_CALCULATE_LOD ] = {"f", "HHfffb", sm6_parser_emit_dx_calculate_lod},
|
|
[DX_CBUFFER_LOAD_LEGACY ] = {"o", "Hi", sm6_parser_emit_dx_cbuffer_load},
|
|
[DX_COS ] = {"g", "R", sm6_parser_emit_dx_sincos},
|
|
[DX_COUNT_BITS ] = {"i", "m", sm6_parser_emit_dx_unary},
|
|
[DX_COVERAGE ] = {"i", "", sm6_parser_emit_dx_coverage},
|
|
[DX_CREATE_HANDLE ] = {"H", "ccib", sm6_parser_emit_dx_create_handle},
|
|
[DX_CUT_STREAM ] = {"v", "c", sm6_parser_emit_dx_stream},
|
|
[DX_DERIV_COARSEX ] = {"e", "R", sm6_parser_emit_dx_unary},
|
|
[DX_DERIV_COARSEY ] = {"e", "R", sm6_parser_emit_dx_unary},
|
|
[DX_DERIV_FINEX ] = {"e", "R", sm6_parser_emit_dx_unary},
|
|
[DX_DERIV_FINEY ] = {"e", "R", sm6_parser_emit_dx_unary},
|
|
[DX_DISCARD ] = {"v", "1", sm6_parser_emit_dx_discard},
|
|
[DX_DOMAIN_LOCATION ] = {"f", "c", sm6_parser_emit_dx_domain_location},
|
|
[DX_DOT2 ] = {"g", "RRRR", sm6_parser_emit_dx_dot},
|
|
[DX_DOT3 ] = {"g", "RRRRRR", sm6_parser_emit_dx_dot},
|
|
[DX_DOT4 ] = {"g", "RRRRRRRR", sm6_parser_emit_dx_dot},
|
|
[DX_EMIT_STREAM ] = {"v", "c", sm6_parser_emit_dx_stream},
|
|
[DX_EMIT_THEN_CUT_STREAM ] = {"v", "c", sm6_parser_emit_dx_stream},
|
|
[DX_EXP ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_FABS ] = {"g", "R", sm6_parser_emit_dx_fabs},
|
|
[DX_FIRST_BIT_HI ] = {"i", "m", sm6_parser_emit_dx_unary},
|
|
[DX_FIRST_BIT_LO ] = {"i", "m", sm6_parser_emit_dx_unary},
|
|
[DX_FIRST_BIT_SHI ] = {"i", "m", sm6_parser_emit_dx_unary},
|
|
[DX_FLATTENED_THREAD_ID_IN_GROUP ] = {"i", "", sm6_parser_emit_dx_compute_builtin},
|
|
[DX_FMA ] = {"g", "RRR", sm6_parser_emit_dx_ma},
|
|
[DX_FMAD ] = {"g", "RRR", sm6_parser_emit_dx_ma},
|
|
[DX_FMAX ] = {"g", "RR", sm6_parser_emit_dx_binary},
|
|
[DX_FMIN ] = {"g", "RR", sm6_parser_emit_dx_binary},
|
|
[DX_FRC ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_GET_DIMENSIONS ] = {"D", "Hi", sm6_parser_emit_dx_get_dimensions},
|
|
[DX_GROUP_ID ] = {"i", "c", sm6_parser_emit_dx_compute_builtin},
|
|
[DX_IBFE ] = {"m", "iiR", sm6_parser_emit_dx_tertiary},
|
|
[DX_HCOS ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_HSIN ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_HTAN ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_IMAD ] = {"m", "RRR", sm6_parser_emit_dx_ma},
|
|
[DX_IMAX ] = {"m", "RR", sm6_parser_emit_dx_binary},
|
|
[DX_IMIN ] = {"m", "RR", sm6_parser_emit_dx_binary},
|
|
[DX_ISFINITE ] = {"1", "g", sm6_parser_emit_dx_unary},
|
|
[DX_ISINF ] = {"1", "g", sm6_parser_emit_dx_unary},
|
|
[DX_ISNAN ] = {"1", "g", sm6_parser_emit_dx_unary},
|
|
[DX_LEGACY_F16TOF32 ] = {"f", "i", sm6_parser_emit_dx_unary},
|
|
[DX_LEGACY_F32TOF16 ] = {"i", "f", sm6_parser_emit_dx_unary},
|
|
[DX_LOAD_INPUT ] = {"o", "ii8i", sm6_parser_emit_dx_load_input},
|
|
[DX_LOAD_OUTPUT_CONTROL_POINT ] = {"o", "ii8i", sm6_parser_emit_dx_load_input},
|
|
[DX_LOAD_PATCH_CONSTANT ] = {"o", "ii8", sm6_parser_emit_dx_load_input},
|
|
[DX_LOG ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_MAKE_DOUBLE ] = {"d", "ii", sm6_parser_emit_dx_make_double},
|
|
[DX_OUTPUT_CONTROL_POINT_ID ] = {"i", "", sm6_parser_emit_dx_output_control_point_id},
|
|
[DX_PRIMITIVE_ID ] = {"i", "", sm6_parser_emit_dx_primitive_id},
|
|
[DX_QUAD_OP ] = {"n", "Rc", sm6_parser_emit_dx_quad_op},
|
|
[DX_QUAD_READ_LANE_AT ] = {"n", "Ri", sm6_parser_emit_dx_binary},
|
|
[DX_RAW_BUFFER_LOAD ] = {"o", "Hii8i", sm6_parser_emit_dx_raw_buffer_load},
|
|
[DX_RAW_BUFFER_STORE ] = {"v", "Hiioooocc", sm6_parser_emit_dx_raw_buffer_store},
|
|
[DX_ROUND_NE ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_ROUND_NI ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_ROUND_PI ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_ROUND_Z ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_RSQRT ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_RT_GET_SAMPLE_COUNT ] = {"i", "", sm6_parser_emit_dx_get_sample_count},
|
|
[DX_RT_GET_SAMPLE_POS ] = {"o", "i", sm6_parser_emit_dx_get_sample_pos},
|
|
[DX_SAMPLE ] = {"o", "HHffffiiif", sm6_parser_emit_dx_sample},
|
|
[DX_SAMPLE_B ] = {"o", "HHffffiiiff", sm6_parser_emit_dx_sample},
|
|
[DX_SAMPLE_C ] = {"o", "HHffffiiiff", sm6_parser_emit_dx_sample},
|
|
[DX_SAMPLE_C_LZ ] = {"o", "HHffffiiif", sm6_parser_emit_dx_sample},
|
|
[DX_SAMPLE_GRAD ] = {"o", "HHffffiiifffffff", sm6_parser_emit_dx_sample},
|
|
[DX_SAMPLE_LOD ] = {"o", "HHffffiiif", sm6_parser_emit_dx_sample},
|
|
[DX_SATURATE ] = {"g", "R", sm6_parser_emit_dx_saturate},
|
|
[DX_SIN ] = {"g", "R", sm6_parser_emit_dx_sincos},
|
|
[DX_SPLIT_DOUBLE ] = {"S", "d", sm6_parser_emit_dx_split_double},
|
|
[DX_SQRT ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_STORE_OUTPUT ] = {"v", "ii8o", sm6_parser_emit_dx_store_output},
|
|
[DX_STORE_PATCH_CONSTANT ] = {"v", "ii8o", sm6_parser_emit_dx_store_output},
|
|
[DX_TAN ] = {"g", "R", sm6_parser_emit_dx_unary},
|
|
[DX_TEX2DMS_GET_SAMPLE_POS ] = {"o", "Hi", sm6_parser_emit_dx_get_sample_pos},
|
|
[DX_TEXTURE_GATHER ] = {"o", "HHffffiic", sm6_parser_emit_dx_texture_gather},
|
|
[DX_TEXTURE_GATHER_CMP ] = {"o", "HHffffiicf", sm6_parser_emit_dx_texture_gather},
|
|
[DX_TEXTURE_LOAD ] = {"o", "HiiiiCCC", sm6_parser_emit_dx_texture_load},
|
|
[DX_TEXTURE_STORE ] = {"v", "Hiiiooooc", sm6_parser_emit_dx_texture_store},
|
|
[DX_THREAD_ID ] = {"i", "c", sm6_parser_emit_dx_compute_builtin},
|
|
[DX_THREAD_ID_IN_GROUP ] = {"i", "c", sm6_parser_emit_dx_compute_builtin},
|
|
[DX_UBFE ] = {"m", "iiR", sm6_parser_emit_dx_tertiary},
|
|
[DX_UMAD ] = {"m", "RRR", sm6_parser_emit_dx_ma},
|
|
[DX_UMAX ] = {"m", "RR", sm6_parser_emit_dx_binary},
|
|
[DX_UMIN ] = {"m", "RR", sm6_parser_emit_dx_binary},
|
|
[DX_WAVE_ACTIVE_ALL_EQUAL ] = {"1", "n", sm6_parser_emit_dx_unary},
|
|
[DX_WAVE_ACTIVE_BALLOT ] = {"V", "1", sm6_parser_emit_dx_wave_active_ballot},
|
|
[DX_WAVE_ACTIVE_BIT ] = {"m", "Rc", sm6_parser_emit_dx_wave_active_bit},
|
|
[DX_WAVE_ACTIVE_OP ] = {"n", "Rcc", sm6_parser_emit_dx_wave_op},
|
|
[DX_WAVE_ALL_BIT_COUNT ] = {"i", "1", sm6_parser_emit_dx_unary},
|
|
[DX_WAVE_ALL_TRUE ] = {"1", "1", sm6_parser_emit_dx_unary},
|
|
[DX_WAVE_ANY_TRUE ] = {"1", "1", sm6_parser_emit_dx_unary},
|
|
[DX_WAVE_GET_LANE_COUNT ] = {"i", "", sm6_parser_emit_dx_wave_builtin},
|
|
[DX_WAVE_GET_LANE_INDEX ] = {"i", "", sm6_parser_emit_dx_wave_builtin},
|
|
[DX_WAVE_IS_FIRST_LANE ] = {"1", "", sm6_parser_emit_dx_void},
|
|
[DX_WAVE_PREFIX_BIT_COUNT ] = {"i", "1", sm6_parser_emit_dx_unary},
|
|
[DX_WAVE_PREFIX_OP ] = {"n", "Rcc", sm6_parser_emit_dx_wave_op},
|
|
[DX_WAVE_READ_LANE_AT ] = {"n", "Ri", sm6_parser_emit_dx_binary},
|
|
[DX_WAVE_READ_LANE_FIRST ] = {"n", "R", sm6_parser_emit_dx_unary},
|
|
};
|
|
|
|
static bool sm6_parser_validate_operand_type(struct sm6_parser *sm6, const struct sm6_value *value, char info_type,
|
|
const struct sm6_type *ret_type, bool is_return)
|
|
{
|
|
const struct sm6_type *type = value->type;
|
|
|
|
if (info_type != 'H' && !sm6_value_is_register(value))
|
|
return false;
|
|
|
|
switch (info_type)
|
|
{
|
|
case 0:
|
|
FIXME("Invalid operand count.\n");
|
|
return false;
|
|
case '1':
|
|
return sm6_type_is_bool(type);
|
|
case '8':
|
|
return sm6_type_is_i8(type);
|
|
case 'b':
|
|
return sm6_value_is_constant(value) && sm6_type_is_bool(type);
|
|
case 'c':
|
|
return sm6_value_is_constant(value) && sm6_type_is_integer(type) && type->u.width >= 8
|
|
&& type->u.width <= 32;
|
|
case 'C':
|
|
return (sm6_value_is_constant(value) || sm6_value_is_undef(value))
|
|
&& sm6_type_is_integer(type) && type->u.width >= 8 && type->u.width <= 32;
|
|
case 'i':
|
|
return sm6_type_is_i32(type);
|
|
case 'm':
|
|
return sm6_type_is_i16_i32_i64(type);
|
|
case 'n':
|
|
return sm6_type_is_numeric(type);
|
|
case 'f':
|
|
return sm6_type_is_float(type);
|
|
case 'd':
|
|
return sm6_type_is_double(type);
|
|
case 'e':
|
|
return sm6_type_is_f16_f32(type);
|
|
case 'g':
|
|
return sm6_type_is_floating_point(type);
|
|
case 'H':
|
|
return (is_return || sm6_value_is_handle(value)) && type == sm6->handle_type;
|
|
case 'D':
|
|
return sm6_type_is_struct(type) && !strcmp(type->u.struc->name, "dx.types.Dimensions");
|
|
case 'S':
|
|
return sm6_type_is_struct(type) && !strcmp(type->u.struc->name, "dx.types.splitdouble");
|
|
case 'V':
|
|
return sm6_type_is_struct(type) && !strcmp(type->u.struc->name, "dx.types.fouri32");
|
|
case 'v':
|
|
return !type;
|
|
case 'o':
|
|
/* TODO: some type checking may be possible */
|
|
return true;
|
|
case 'R':
|
|
return type == ret_type;
|
|
default:
|
|
FIXME("Unhandled operand code '%c'.\n", info_type);
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static bool sm6_parser_validate_dx_op(struct sm6_parser *sm6, enum dx_intrinsic_opcode op, const char *name,
|
|
const struct sm6_value **operands, unsigned int operand_count, struct sm6_value *dst)
|
|
{
|
|
const struct sm6_dx_opcode_info *info;
|
|
unsigned int i;
|
|
|
|
info = &sm6_dx_op_table[op];
|
|
|
|
VKD3D_ASSERT(info->ret_type[0]);
|
|
if (!sm6_parser_validate_operand_type(sm6, dst, info->ret_type[0], NULL, true))
|
|
{
|
|
WARN("Failed to validate return type for dx intrinsic id %u, '%s'.\n", op, name);
|
|
/* Return type validation failure is not so critical. We only need to set
|
|
* a data type for the SSA result. */
|
|
}
|
|
|
|
for (i = 0; i < operand_count; ++i)
|
|
{
|
|
const struct sm6_value *value = operands[i];
|
|
if (!sm6_parser_validate_operand_type(sm6, value, info->operand_info[i], dst->type, false))
|
|
{
|
|
WARN("Failed to validate operand %u for dx intrinsic id %u, '%s'.\n", i + 1, op, name);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Operand %u for call to dx intrinsic function '%s' is invalid.", i + 1, name);
|
|
return false;
|
|
}
|
|
}
|
|
if (info->operand_info[operand_count])
|
|
{
|
|
WARN("Missing operands for dx intrinsic id %u, '%s'.\n", op, name);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Call to dx intrinsic function '%s' has missing operands.", name);
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static void sm6_parser_emit_unhandled(struct sm6_parser *sm6, struct vkd3d_shader_instruction *ins,
|
|
struct sm6_value *dst)
|
|
{
|
|
const struct sm6_type *type;
|
|
|
|
ins->opcode = VKD3DSIH_NOP;
|
|
|
|
if (!dst->type)
|
|
return;
|
|
|
|
type = sm6_type_get_scalar_type(dst->type, 0);
|
|
vsir_register_init(&dst->u.reg, VKD3DSPR_UNDEF, vkd3d_data_type_from_sm6_type(type), 0);
|
|
/* dst->is_undefined is not set here because it flags only explicitly undefined values. */
|
|
}
|
|
|
|
static void sm6_parser_decode_dx_op(struct sm6_parser *sm6, enum dx_intrinsic_opcode op,
|
|
const char *name, const struct sm6_value **operands, unsigned int operand_count,
|
|
struct function_emission_state *state, struct sm6_value *dst)
|
|
{
|
|
if (op >= ARRAY_SIZE(sm6_dx_op_table) || !sm6_dx_op_table[op].operand_info)
|
|
{
|
|
FIXME("Unhandled dx intrinsic function id %u, '%s'.\n", op, name);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_UNHANDLED_INTRINSIC,
|
|
"Call to intrinsic function %s is unhandled.", name);
|
|
sm6_parser_emit_unhandled(sm6, state->ins, dst);
|
|
return;
|
|
}
|
|
|
|
if (sm6_parser_validate_dx_op(sm6, op, name, operands, operand_count, dst))
|
|
sm6_dx_op_table[op].handler(sm6, op, operands, state);
|
|
else
|
|
sm6_parser_emit_unhandled(sm6, state->ins, dst);
|
|
}
|
|
|
|
static void sm6_parser_emit_call(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct function_emission_state *state, struct sm6_value *dst)
|
|
{
|
|
const struct sm6_value *operands[DXIL_OP_MAX_OPERANDS];
|
|
const struct sm6_value *fn_value, *op_value;
|
|
unsigned int i = 1, j, operand_count;
|
|
const struct sm6_type *type = NULL;
|
|
uint64_t call_conv;
|
|
|
|
if (!dxil_record_validate_operand_min_count(record, 2, sm6))
|
|
return;
|
|
|
|
/* TODO: load the 1-based attributes index from record->operands[0] and validate against attribute count. */
|
|
|
|
if ((call_conv = record->operands[i++]) & CALL_CONV_FLAG_EXPLICIT_TYPE)
|
|
type = sm6_parser_get_type(sm6, record->operands[i++]);
|
|
if (call_conv &= ~CALL_CONV_FLAG_EXPLICIT_TYPE)
|
|
WARN("Ignoring calling convention %#"PRIx64".\n", call_conv);
|
|
|
|
if (!(fn_value = sm6_parser_get_value_by_ref(sm6, record, NULL, &i)))
|
|
return;
|
|
if (!sm6_value_is_function_dcl(fn_value))
|
|
{
|
|
WARN("Function target value is not a function declaration.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Function call target value is not a function declaration.");
|
|
return;
|
|
}
|
|
|
|
if (type && type != fn_value->type->u.pointer.type)
|
|
WARN("Explicit call type does not match function type.\n");
|
|
type = fn_value->type->u.pointer.type;
|
|
|
|
if (!sm6_type_is_void(type->u.function->ret_type))
|
|
dst->type = type->u.function->ret_type;
|
|
|
|
operand_count = type->u.function->param_count;
|
|
if (operand_count > ARRAY_SIZE(operands))
|
|
{
|
|
WARN("Ignoring %zu operands.\n", operand_count - ARRAY_SIZE(operands));
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring %zu operands for function call.", operand_count - ARRAY_SIZE(operands));
|
|
operand_count = ARRAY_SIZE(operands);
|
|
}
|
|
|
|
for (j = 0; j < operand_count; ++j)
|
|
{
|
|
if (!(operands[j] = sm6_parser_get_value_by_ref(sm6, record, type->u.function->param_types[j], &i)))
|
|
return;
|
|
}
|
|
if ((j = record->operand_count - i))
|
|
{
|
|
WARN("Ignoring %u operands beyond the function parameter list.\n", j);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring %u function call operands beyond the parameter list.", j);
|
|
}
|
|
|
|
if (!fn_value->u.function.is_prototype)
|
|
{
|
|
FIXME("Unhandled call to local function.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Call to a local function is unsupported.");
|
|
return;
|
|
}
|
|
if (!sm6_value_is_dx_intrinsic_dcl(fn_value))
|
|
WARN("External function is not a dx intrinsic.\n");
|
|
|
|
if (!operand_count)
|
|
{
|
|
WARN("Missing dx intrinsic function id.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"The id for a dx intrinsic function is missing.");
|
|
return;
|
|
}
|
|
|
|
op_value = operands[0];
|
|
if (!sm6_value_is_constant(op_value) || !sm6_type_is_integer(op_value->type))
|
|
{
|
|
WARN("dx intrinsic function id is not a constant int.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Expected a constant integer dx intrinsic function id.");
|
|
return;
|
|
}
|
|
sm6_parser_decode_dx_op(sm6, register_get_uint_value(&op_value->u.reg),
|
|
fn_value->u.function.name, &operands[1], operand_count - 1, state, dst);
|
|
}
|
|
|
|
static enum vkd3d_shader_opcode sm6_map_cast_op(uint64_t code, const struct sm6_type *from,
|
|
const struct sm6_type *to, struct sm6_parser *sm6)
|
|
{
|
|
enum vkd3d_shader_opcode op = VKD3DSIH_INVALID;
|
|
bool from_int, to_int, from_fp, to_fp;
|
|
bool is_valid = false;
|
|
|
|
from_int = sm6_type_is_integer(from);
|
|
to_int = sm6_type_is_integer(to);
|
|
from_fp = sm6_type_is_floating_point(from);
|
|
to_fp = sm6_type_is_floating_point(to);
|
|
|
|
/* NOTE: DXIL currently doesn't use vectors here. */
|
|
if ((!from_int && !from_fp) || (!to_int && !to_fp))
|
|
{
|
|
FIXME("Unhandled cast of type class %u to type class %u.\n", from->class, to->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Cast of type class %u to type class %u is not implemented.", from->class, to->class);
|
|
return VKD3DSIH_INVALID;
|
|
}
|
|
if (to->u.width == 8 || from->u.width == 8)
|
|
{
|
|
FIXME("Unhandled 8-bit value.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Cast to/from an 8-bit type is not implemented.");
|
|
return VKD3DSIH_INVALID;
|
|
}
|
|
|
|
/* DXC emits minimum precision types as 16-bit. These must be emitted
|
|
* as 32-bit in VSIR, so all width extensions to 32 bits are no-ops. */
|
|
switch (code)
|
|
{
|
|
case CAST_TRUNC:
|
|
/* nop or min precision. TODO: native 16-bit */
|
|
if (to->u.width == from->u.width || (to->u.width == 16 && from->u.width == 32))
|
|
op = VKD3DSIH_NOP;
|
|
else
|
|
op = VKD3DSIH_UTOU;
|
|
is_valid = from_int && to_int && to->u.width <= from->u.width;
|
|
break;
|
|
case CAST_ZEXT:
|
|
case CAST_SEXT:
|
|
/* nop or min precision. TODO: native 16-bit.
|
|
* Extension instructions could be emitted for min precision, but in Windows
|
|
* the AMD RX 580 simply drops such instructions, which makes sense as no
|
|
* assumptions should be made about any behaviour which depends on bit width. */
|
|
if (to->u.width == from->u.width || (to->u.width == 32 && from->u.width == 16))
|
|
{
|
|
op = VKD3DSIH_NOP;
|
|
is_valid = from_int && to_int;
|
|
}
|
|
else if (to->u.width > from->u.width)
|
|
{
|
|
op = (code == CAST_ZEXT) ? VKD3DSIH_UTOU : VKD3DSIH_ITOI;
|
|
VKD3D_ASSERT(from->u.width == 1 || to->u.width == 64);
|
|
is_valid = from_int && to_int;
|
|
}
|
|
break;
|
|
case CAST_FPTOUI:
|
|
op = VKD3DSIH_FTOU;
|
|
is_valid = from_fp && to_int && to->u.width > 1;
|
|
break;
|
|
case CAST_FPTOSI:
|
|
op = VKD3DSIH_FTOI;
|
|
is_valid = from_fp && to_int && to->u.width > 1;
|
|
break;
|
|
case CAST_UITOFP:
|
|
op = VKD3DSIH_UTOF;
|
|
is_valid = from_int && to_fp;
|
|
break;
|
|
case CAST_SITOFP:
|
|
op = VKD3DSIH_ITOF;
|
|
is_valid = from_int && to_fp;
|
|
break;
|
|
case CAST_FPTRUNC:
|
|
/* TODO: native 16-bit */
|
|
op = (from->u.width == 64) ? VKD3DSIH_DTOF : VKD3DSIH_NOP;
|
|
is_valid = from_fp && to_fp;
|
|
break;
|
|
case CAST_FPEXT:
|
|
/* TODO: native 16-bit */
|
|
op = (to->u.width == 64) ? VKD3DSIH_FTOD : VKD3DSIH_NOP;
|
|
is_valid = from_fp && to_fp;
|
|
break;
|
|
case CAST_BITCAST:
|
|
op = VKD3DSIH_MOV;
|
|
is_valid = to->u.width == from->u.width;
|
|
break;
|
|
default:
|
|
FIXME("Unhandled cast op %"PRIu64".\n", code);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Cast operation %"PRIu64" is unhandled.\n", code);
|
|
return VKD3DSIH_INVALID;
|
|
}
|
|
|
|
if (!is_valid)
|
|
{
|
|
FIXME("Invalid types %u and/or %u for op %"PRIu64".\n", from->class, to->class, code);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Cast operation %"PRIu64" from type class %u, width %u to type class %u, width %u is invalid.\n",
|
|
code, from->class, from->u.width, to->class, to->u.width);
|
|
return VKD3DSIH_INVALID;
|
|
}
|
|
|
|
return op;
|
|
}
|
|
|
|
static void sm6_parser_emit_cast(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct vkd3d_shader_instruction *ins, struct sm6_value *dst)
|
|
{
|
|
struct vkd3d_shader_src_param *src_param;
|
|
enum vkd3d_shader_opcode handler_idx;
|
|
const struct sm6_value *value;
|
|
const struct sm6_type *type;
|
|
unsigned int i = 0;
|
|
|
|
if (!(value = sm6_parser_get_value_by_ref(sm6, record, NULL, &i)))
|
|
return;
|
|
|
|
if (!dxil_record_validate_operand_count(record, i + 2, i + 2, sm6))
|
|
return;
|
|
|
|
if (!(type = sm6_parser_get_type(sm6, record->operands[i++])))
|
|
return;
|
|
|
|
dst->type = type;
|
|
|
|
if (sm6_type_is_pointer(type))
|
|
{
|
|
*dst = *value;
|
|
dst->type = type;
|
|
ins->opcode = VKD3DSIH_NOP;
|
|
return;
|
|
}
|
|
|
|
if ((handler_idx = sm6_map_cast_op(record->operands[i], value->type, type, sm6)) == VKD3DSIH_INVALID)
|
|
return;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, handler_idx);
|
|
|
|
if (handler_idx == VKD3DSIH_NOP)
|
|
{
|
|
dst->u.reg = value->u.reg;
|
|
/* Set the result type for casts from 16-bit min precision. */
|
|
if (type->u.width != 16)
|
|
dst->u.reg.data_type = vkd3d_data_type_from_sm6_type(type);
|
|
return;
|
|
}
|
|
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param_init_from_value(src_param, value);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
|
|
/* bitcast */
|
|
if (handler_idx == VKD3DSIH_MOV)
|
|
src_param->reg.data_type = dst->u.reg.data_type;
|
|
}
|
|
|
|
struct sm6_cmp_info
|
|
{
|
|
enum vkd3d_shader_opcode handler_idx;
|
|
bool src_swap;
|
|
};
|
|
|
|
static const struct sm6_cmp_info *sm6_map_cmp2_op(uint64_t code)
|
|
{
|
|
static const struct sm6_cmp_info cmp_op_table[] =
|
|
{
|
|
[FCMP_FALSE] = {VKD3DSIH_INVALID},
|
|
[FCMP_OEQ] = {VKD3DSIH_EQO},
|
|
[FCMP_OGT] = {VKD3DSIH_LTO, true},
|
|
[FCMP_OGE] = {VKD3DSIH_GEO},
|
|
[FCMP_OLT] = {VKD3DSIH_LTO},
|
|
[FCMP_OLE] = {VKD3DSIH_GEO, true},
|
|
[FCMP_ONE] = {VKD3DSIH_NEO},
|
|
[FCMP_ORD] = {VKD3DSIH_ORD},
|
|
[FCMP_UNO] = {VKD3DSIH_UNO},
|
|
[FCMP_UEQ] = {VKD3DSIH_EQU},
|
|
[FCMP_UGT] = {VKD3DSIH_LTU, true},
|
|
[FCMP_UGE] = {VKD3DSIH_GEU},
|
|
[FCMP_ULT] = {VKD3DSIH_LTU},
|
|
[FCMP_ULE] = {VKD3DSIH_GEU, true},
|
|
[FCMP_UNE] = {VKD3DSIH_NEU},
|
|
[FCMP_TRUE] = {VKD3DSIH_INVALID},
|
|
|
|
[ICMP_EQ] = {VKD3DSIH_IEQ},
|
|
[ICMP_NE] = {VKD3DSIH_INE},
|
|
[ICMP_UGT] = {VKD3DSIH_ULT, true},
|
|
[ICMP_UGE] = {VKD3DSIH_UGE},
|
|
[ICMP_ULT] = {VKD3DSIH_ULT},
|
|
[ICMP_ULE] = {VKD3DSIH_UGE, true},
|
|
[ICMP_SGT] = {VKD3DSIH_ILT, true},
|
|
[ICMP_SGE] = {VKD3DSIH_IGE},
|
|
[ICMP_SLT] = {VKD3DSIH_ILT},
|
|
[ICMP_SLE] = {VKD3DSIH_IGE, true},
|
|
};
|
|
|
|
return (code < ARRAY_SIZE(cmp_op_table)) ? &cmp_op_table[code] : NULL;
|
|
}
|
|
|
|
static void sm6_parser_emit_cmp2(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct vkd3d_shader_instruction *ins, struct sm6_value *dst)
|
|
{
|
|
struct vkd3d_shader_src_param *src_params;
|
|
const struct sm6_type *type_a, *type_b;
|
|
bool is_int, is_fp, silence_warning;
|
|
const struct sm6_cmp_info *cmp;
|
|
const struct sm6_value *a, *b;
|
|
uint64_t code, flags;
|
|
unsigned int i = 0;
|
|
|
|
if (!(dst->type = sm6->bool_type))
|
|
{
|
|
WARN("Bool type not found.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_MODULE,
|
|
"Module does not define a boolean type for comparison results.");
|
|
return;
|
|
}
|
|
|
|
a = sm6_parser_get_value_by_ref(sm6, record, NULL, &i);
|
|
if (!a)
|
|
return;
|
|
b = sm6_parser_get_value_by_ref(sm6, record, a->type, &i);
|
|
if (!b)
|
|
return;
|
|
|
|
if (!dxil_record_validate_operand_count(record, i + 1, i + 2, sm6))
|
|
return;
|
|
|
|
type_a = a->type;
|
|
type_b = b->type;
|
|
is_int = sm6_type_is_bool_i16_i32_i64(type_a);
|
|
is_fp = sm6_type_is_floating_point(type_a);
|
|
|
|
code = record->operands[i++];
|
|
|
|
/* dxcompiler occasionally emits bool not-equal-to-false, which is a no-op. Bool comparisons
|
|
* do not otherwise occur, so deleting these avoids the need for backend support. */
|
|
if (sm6_type_is_bool(type_a) && code == ICMP_NE && sm6_value_is_constant_zero(b))
|
|
{
|
|
ins->opcode = VKD3DSIH_NOP;
|
|
*dst = *a;
|
|
return;
|
|
}
|
|
|
|
if ((!is_int && !is_fp) || is_int != (code >= ICMP_EQ))
|
|
{
|
|
FIXME("Invalid operation %"PRIu64" on type class %u.\n", code, type_a->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Comparison operation %"PRIu64" on type class %u is invalid.", code, type_a->class);
|
|
return;
|
|
}
|
|
|
|
if (type_a != type_b)
|
|
{
|
|
WARN("Type mismatch, type %u width %u vs type %u width %u.\n", type_a->class,
|
|
type_a->u.width, type_b->class, type_b->u.width);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"Type mismatch in comparison operation arguments.");
|
|
}
|
|
|
|
if (!(cmp = sm6_map_cmp2_op(code)) || !cmp->handler_idx || cmp->handler_idx == VKD3DSIH_INVALID)
|
|
{
|
|
FIXME("Unhandled operation %"PRIu64".\n", code);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Comparison operation %"PRIu64" is unhandled.", code);
|
|
return;
|
|
}
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, cmp->handler_idx);
|
|
|
|
flags = (record->operand_count > i) ? record->operands[i] : 0;
|
|
silence_warning = false;
|
|
|
|
if (is_fp)
|
|
{
|
|
if (!(flags & FP_ALLOW_UNSAFE_ALGEBRA))
|
|
ins->flags |= VKD3DSI_PRECISE_X;
|
|
flags &= ~FP_ALLOW_UNSAFE_ALGEBRA;
|
|
/* SPIR-V FPFastMathMode is only available in the Kernel execution model. */
|
|
silence_warning = !(flags & ~(FP_NO_NAN | FP_NO_INF | FP_NO_SIGNED_ZEROS | FP_ALLOW_RECIPROCAL));
|
|
}
|
|
if (flags && silence_warning)
|
|
{
|
|
TRACE("Ignoring fast FP modifier %#"PRIx64".\n", flags);
|
|
}
|
|
else if (flags)
|
|
{
|
|
WARN("Ignoring flags %#"PRIx64".\n", flags);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring flags %#"PRIx64" for a comparison operation.", flags);
|
|
}
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 2, sm6)))
|
|
return;
|
|
src_param_init_from_value(&src_params[0 ^ cmp->src_swap], a);
|
|
src_param_init_from_value(&src_params[1 ^ cmp->src_swap], b);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_cmpxchg(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct vkd3d_shader_instruction *ins, struct sm6_value *dst)
|
|
{
|
|
uint64_t success_ordering, failure_ordering;
|
|
struct vkd3d_shader_dst_param *dst_params;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
const struct sm6_value *ptr, *cmp, *new;
|
|
unsigned int i = 0;
|
|
bool is_volatile;
|
|
uint64_t code;
|
|
|
|
if (!(ptr = sm6_parser_get_value_by_ref(sm6, record, NULL, &i))
|
|
|| !sm6_value_validate_is_pointer_to_i32(ptr, sm6)
|
|
|| !sm6_value_validate_is_backward_ref(ptr, sm6))
|
|
return;
|
|
|
|
if (ptr->u.reg.type != VKD3DSPR_GROUPSHAREDMEM)
|
|
{
|
|
WARN("Register is not groupshared.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"The destination register for a cmpxchg instruction is not groupshared memory.");
|
|
return;
|
|
}
|
|
|
|
if (!(dst->type = sm6_type_get_cmpxchg_result_struct(sm6)))
|
|
{
|
|
WARN("Failed to find result struct.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_MODULE,
|
|
"Module does not define a result struct type for a cmpxchg instruction.");
|
|
return;
|
|
}
|
|
|
|
/* Forward-referenced comparands are stored as value/type pairs, even
|
|
* though in principle we could use the destination type. */
|
|
cmp = sm6_parser_get_value_by_ref(sm6, record, NULL, &i);
|
|
new = sm6_parser_get_value_by_ref(sm6, record, ptr->type->u.pointer.type, &i);
|
|
if (!cmp || !new)
|
|
return;
|
|
|
|
if (!sm6_value_validate_is_i32(cmp, sm6)
|
|
|| !sm6_value_validate_is_i32(new, sm6)
|
|
|| !dxil_record_validate_operand_count(record, i + 3, i + 5, sm6))
|
|
{
|
|
return;
|
|
}
|
|
|
|
is_volatile = record->operands[i++];
|
|
success_ordering = record->operands[i++];
|
|
|
|
if ((code = record->operands[i++]) != 1)
|
|
FIXME("Ignoring synchronisation scope %"PRIu64".\n", code);
|
|
|
|
failure_ordering = (record->operand_count > i) ? record->operands[i++] : success_ordering;
|
|
|
|
/* It's currently not possible to specify an atomic ordering in HLSL, and it defaults to seq_cst. */
|
|
if (success_ordering != ORDERING_SEQCST)
|
|
FIXME("Unhandled success ordering %"PRIu64".\n", success_ordering);
|
|
if (success_ordering != failure_ordering)
|
|
FIXME("Unhandled failure ordering %"PRIu64".\n", failure_ordering);
|
|
|
|
if (record->operand_count > i && record->operands[i])
|
|
FIXME("Ignoring weak cmpxchg.\n");
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_IMM_ATOMIC_CMP_EXCH);
|
|
ins->flags = is_volatile ? VKD3DARF_SEQ_CST | VKD3DARF_VOLATILE : VKD3DARF_SEQ_CST;
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 3, sm6)))
|
|
return;
|
|
src_param_make_constant_uint(&src_params[0], 0);
|
|
src_param_init_from_value(&src_params[1], cmp);
|
|
src_param_init_from_value(&src_params[2], new);
|
|
|
|
if (!(dst_params = instruction_dst_params_alloc(ins, 2, sm6)))
|
|
return;
|
|
register_init_ssa_scalar(&dst_params[0].reg, dst->type, dst, sm6);
|
|
dst_param_init(&dst_params[0]);
|
|
dst_params[1].reg = ptr->u.reg;
|
|
dst_param_init(&dst_params[1]);
|
|
|
|
dst->u.reg = dst_params[0].reg;
|
|
}
|
|
|
|
static void sm6_parser_emit_extractval(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct vkd3d_shader_instruction *ins, struct sm6_value *dst)
|
|
{
|
|
struct vkd3d_shader_src_param *src_param;
|
|
const struct sm6_type *type;
|
|
const struct sm6_value *src;
|
|
unsigned int i = 0;
|
|
uint64_t elem_idx;
|
|
|
|
if (!(src = sm6_parser_get_value_by_ref(sm6, record, NULL, &i)))
|
|
return;
|
|
|
|
if (!dxil_record_validate_operand_min_count(record, i + 1, sm6))
|
|
return;
|
|
|
|
if (record->operand_count > i + 1)
|
|
{
|
|
FIXME("Unhandled multiple indices.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Multiple extractval indices are not supported.");
|
|
return;
|
|
}
|
|
|
|
type = src->type;
|
|
if (!sm6_type_is_aggregate(type))
|
|
{
|
|
WARN("Invalid extraction from non-aggregate.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Source type of an extractval instruction is not an aggregate.");
|
|
return;
|
|
}
|
|
|
|
elem_idx = record->operands[i];
|
|
if (!(type = sm6_type_get_element_type_at_index(type, elem_idx)))
|
|
{
|
|
WARN("Invalid element index %"PRIu64".\n", elem_idx);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Element index %"PRIu64" for an extractval instruction is out of bounds.", elem_idx);
|
|
return;
|
|
}
|
|
if (!sm6_type_is_scalar(type))
|
|
{
|
|
FIXME("Nested extraction is not supported.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Extraction from nested aggregates is not supported.");
|
|
return;
|
|
}
|
|
dst->type = type;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOV);
|
|
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
src_param->reg = src->u.reg;
|
|
src_param_init_scalar(src_param, elem_idx);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static void sm6_parser_emit_gep(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct vkd3d_shader_instruction *ins, struct sm6_value *dst)
|
|
{
|
|
const struct sm6_type *type, *pointee_type;
|
|
unsigned int elem_idx, operand_idx = 2;
|
|
enum bitcode_address_space addr_space;
|
|
const struct sm6_value *elem_value;
|
|
struct vkd3d_shader_register *reg;
|
|
const struct sm6_value *src;
|
|
bool is_in_bounds;
|
|
|
|
if (!dxil_record_validate_operand_min_count(record, 5, sm6)
|
|
|| !(type = sm6_parser_get_type(sm6, record->operands[1]))
|
|
|| !(src = sm6_parser_get_value_by_ref(sm6, record, NULL, &operand_idx))
|
|
|| !sm6_value_validate_is_register(src, sm6)
|
|
|| !sm6_value_validate_is_pointer(src, sm6)
|
|
|| !dxil_record_validate_operand_min_count(record, operand_idx + 2, sm6))
|
|
{
|
|
return;
|
|
}
|
|
|
|
if (src->u.reg.idx_count > 1)
|
|
{
|
|
WARN("Unsupported stacked GEP.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"A GEP instruction on the result of a previous GEP is unsupported.");
|
|
return;
|
|
}
|
|
|
|
is_in_bounds = record->operands[0];
|
|
|
|
if ((pointee_type = src->type->u.pointer.type) != type)
|
|
{
|
|
WARN("Type mismatch, type %u width %u vs type %u width %u.\n", type->class,
|
|
type->u.width, pointee_type->class, pointee_type->u.width);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"Type mismatch in GEP operation arguments.");
|
|
}
|
|
addr_space = src->type->u.pointer.addr_space;
|
|
|
|
if (!(elem_value = sm6_parser_get_value_by_ref(sm6, record, NULL, &operand_idx)))
|
|
return;
|
|
|
|
/* The first index is always zero, to form a simple pointer dereference. */
|
|
if (sm6_value_get_constant_uint(elem_value))
|
|
{
|
|
WARN("Expected constant zero.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"The pointer dereference index for a GEP instruction is not constant zero.");
|
|
return;
|
|
}
|
|
|
|
if (!sm6_type_is_array(pointee_type))
|
|
{
|
|
WARN("Invalid GEP on type class %u.\n", pointee_type->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Source type for index 1 of a GEP instruction is not an array.");
|
|
return;
|
|
}
|
|
|
|
if (!(elem_value = sm6_parser_get_value_by_ref(sm6, record, NULL, &operand_idx)))
|
|
return;
|
|
|
|
/* If indexing is dynamic, just get the type at offset zero. */
|
|
elem_idx = sm6_value_is_constant(elem_value) ? sm6_value_get_constant_uint(elem_value) : 0;
|
|
type = sm6_type_get_element_type_at_index(pointee_type, elem_idx);
|
|
if (!type)
|
|
{
|
|
WARN("Invalid element index %u.\n", elem_idx);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Element index %u for a GEP instruction is out of bounds.", elem_idx);
|
|
return;
|
|
}
|
|
|
|
if (operand_idx < record->operand_count)
|
|
{
|
|
FIXME("Multiple element indices are not implemented.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Multi-dimensional addressing in GEP instructions is not supported.");
|
|
return;
|
|
}
|
|
|
|
if (!(dst->type = sm6_type_get_pointer_to_type(type, addr_space, sm6)))
|
|
{
|
|
WARN("Failed to get pointer type for type %u.\n", type->class);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_MODULE,
|
|
"Module does not define a pointer type for a GEP instruction.");
|
|
return;
|
|
}
|
|
|
|
reg = &dst->u.reg;
|
|
*reg = src->u.reg;
|
|
reg->idx[1].offset = 0;
|
|
register_index_address_init(®->idx[1], elem_value, sm6);
|
|
reg->idx[1].is_in_bounds = is_in_bounds;
|
|
reg->idx_count = 2;
|
|
dst->structure_stride = src->structure_stride;
|
|
|
|
ins->opcode = VKD3DSIH_NOP;
|
|
}
|
|
|
|
static void sm6_parser_emit_load(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct vkd3d_shader_instruction *ins, struct sm6_value *dst)
|
|
{
|
|
const struct sm6_type *elem_type = NULL, *pointee_type;
|
|
unsigned int alignment, operand_count, i = 0;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
const struct sm6_value *ptr;
|
|
uint64_t alignment_code;
|
|
|
|
if (!(ptr = sm6_parser_get_value_by_ref(sm6, record, NULL, &i)))
|
|
return;
|
|
if (!sm6_value_validate_is_register(ptr, sm6)
|
|
|| !sm6_value_validate_is_pointer(ptr, sm6)
|
|
|| !sm6_value_validate_is_backward_ref(ptr, sm6)
|
|
|| !dxil_record_validate_operand_count(record, i + 2, i + 3, sm6))
|
|
return;
|
|
|
|
if (record->operand_count > i + 2 && !(elem_type = sm6_parser_get_type(sm6, record->operands[i++])))
|
|
return;
|
|
|
|
if (!elem_type)
|
|
{
|
|
elem_type = ptr->type->u.pointer.type;
|
|
}
|
|
else if (elem_type != (pointee_type = ptr->type->u.pointer.type))
|
|
{
|
|
WARN("Type mismatch.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"Type mismatch in pointer load arguments.");
|
|
}
|
|
|
|
dst->type = elem_type;
|
|
|
|
if (!sm6_value_validate_is_numeric(dst, sm6))
|
|
return;
|
|
|
|
alignment_code = record->operands[i++];
|
|
if (!bitcode_parse_alignment(alignment_code, &alignment))
|
|
WARN("Invalid alignment %"PRIu64".\n", alignment_code);
|
|
|
|
if (record->operands[i])
|
|
WARN("Ignoring volatile modifier.\n");
|
|
|
|
if (ptr->structure_stride)
|
|
{
|
|
VKD3D_ASSERT(ptr->u.reg.type == VKD3DSPR_GROUPSHAREDMEM);
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_LD_STRUCTURED);
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 3, sm6)))
|
|
return;
|
|
if (ptr->u.reg.idx[1].rel_addr)
|
|
src_params[0] = *ptr->u.reg.idx[1].rel_addr;
|
|
else
|
|
src_param_make_constant_uint(&src_params[0], ptr->u.reg.idx[1].offset);
|
|
/* Struct offset is always zero as there is no struct, just an array. */
|
|
src_param_make_constant_uint(&src_params[1], 0);
|
|
src_param_init_from_value(&src_params[2], ptr);
|
|
src_params[2].reg.alignment = alignment;
|
|
}
|
|
else
|
|
{
|
|
operand_count = 1 + (ptr->u.reg.type == VKD3DSPR_GROUPSHAREDMEM);
|
|
vsir_instruction_init(ins, &sm6->p.location, (operand_count > 1) ? VKD3DSIH_LD_RAW : VKD3DSIH_MOV);
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, operand_count, sm6)))
|
|
return;
|
|
if (operand_count > 1)
|
|
src_param_make_constant_uint(&src_params[0], 0);
|
|
src_param_init_from_value(&src_params[operand_count - 1], ptr);
|
|
src_params[operand_count - 1].reg.alignment = alignment;
|
|
}
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static int phi_incoming_compare(const void *a, const void *b)
|
|
{
|
|
const struct incoming_value *incoming_a = a, *incoming_b = b;
|
|
|
|
return (incoming_a->block > incoming_b->block) - (incoming_a->block < incoming_b->block);
|
|
}
|
|
|
|
static void sm6_parser_emit_phi(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct sm6_function *function, struct sm6_block *code_block, struct vkd3d_shader_instruction *ins,
|
|
struct sm6_value *dst)
|
|
{
|
|
struct incoming_value *incoming;
|
|
const struct sm6_type *type;
|
|
struct sm6_phi *phi;
|
|
unsigned int i, j;
|
|
uint64_t src_idx;
|
|
|
|
if (!(record->operand_count & 1))
|
|
{
|
|
WARN("Invalid operand count %u.\n", record->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Invalid operand count %u for phi instruction.", record->operand_count);
|
|
return;
|
|
}
|
|
|
|
if (!(type = sm6_parser_get_type(sm6, record->operands[0])))
|
|
return;
|
|
if (!sm6_type_is_numeric(type))
|
|
{
|
|
/* dxc doesn't seem to use buffer/resource read return types here. */
|
|
FIXME("Only scalar numeric types are supported.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Result type class %u of a phi instruction is not scalar numeric.", type->class);
|
|
return;
|
|
}
|
|
|
|
dst->type = type;
|
|
register_init_ssa_scalar(&dst->u.reg, type, dst, sm6);
|
|
|
|
if (!(phi = sm6_block_phi_require_space(code_block, sm6)))
|
|
return;
|
|
phi->reg = dst->u.reg;
|
|
phi->incoming_count = record->operand_count / 2u;
|
|
|
|
if (!vkd3d_array_reserve((void **)&phi->incoming, &phi->incoming_capacity, phi->incoming_count,
|
|
sizeof(*phi->incoming)))
|
|
{
|
|
ERR("Failed to allocate phi incoming array.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating a phi incoming array.");
|
|
return;
|
|
}
|
|
incoming = phi->incoming;
|
|
|
|
for (i = 1; i < record->operand_count; i += 2)
|
|
{
|
|
src_idx = sm6->value_count - decode_rotated_signed_value(record->operands[i]);
|
|
/* May be a forward reference. */
|
|
if (src_idx >= sm6->cur_max_value)
|
|
{
|
|
WARN("Invalid value index %"PRIu64".\n", src_idx);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid value index %"PRIu64" for a phi incoming value.", src_idx);
|
|
return;
|
|
}
|
|
|
|
j = i / 2u;
|
|
/* Store the value index in the register for later resolution. */
|
|
incoming[j].reg.idx[0].offset = src_idx;
|
|
incoming[j].block = sm6_function_get_block(function, record->operands[i + 1], sm6);
|
|
}
|
|
|
|
ins->opcode = VKD3DSIH_NOP;
|
|
|
|
qsort(incoming, phi->incoming_count, sizeof(*incoming), phi_incoming_compare);
|
|
|
|
for (i = 1, j = 1; i < phi->incoming_count; ++i)
|
|
{
|
|
if (incoming[i].block != incoming[i - 1].block)
|
|
{
|
|
incoming[j++] = incoming[i];
|
|
continue;
|
|
}
|
|
|
|
if (incoming[i].reg.idx[0].offset != incoming[i - 1].reg.idx[0].offset)
|
|
{
|
|
WARN("PHI conflict.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Two phi incomings have the same block but different values.");
|
|
}
|
|
}
|
|
/* if (j == 1) we should be able to set dst->u.reg to incoming[0].reg, but structurisation
|
|
* may potentially add new incomings. */
|
|
phi->incoming_count = j;
|
|
}
|
|
|
|
static void sm6_parser_emit_ret(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct sm6_block *code_block, struct vkd3d_shader_instruction *ins)
|
|
{
|
|
if (!dxil_record_validate_operand_count(record, 0, 1, sm6))
|
|
return;
|
|
|
|
if (record->operand_count)
|
|
FIXME("Non-void return is not implemented.\n");
|
|
|
|
code_block->terminator.type = TERMINATOR_RET;
|
|
|
|
ins->opcode = VKD3DSIH_NOP;
|
|
}
|
|
|
|
static void sm6_parser_emit_store(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct vkd3d_shader_instruction *ins, struct sm6_value *dst)
|
|
{
|
|
unsigned int i = 0, alignment, operand_count;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_dst_param *dst_param;
|
|
const struct sm6_value *ptr, *src;
|
|
uint64_t alignment_code;
|
|
|
|
if (!(ptr = sm6_parser_get_value_by_ref(sm6, record, NULL, &i))
|
|
|| !sm6_value_validate_is_register(ptr, sm6)
|
|
|| !sm6_value_validate_is_pointer(ptr, sm6)
|
|
|| !sm6_value_validate_is_backward_ref(ptr, sm6))
|
|
{
|
|
return;
|
|
}
|
|
|
|
/* Forward-referenced sources are stored as value/type pairs, even
|
|
* though in principle we could use the destination type. */
|
|
if (!(src = sm6_parser_get_value_by_ref(sm6, record, NULL, &i)))
|
|
return;
|
|
if (!sm6_value_validate_is_numeric(src, sm6))
|
|
return;
|
|
|
|
if (ptr->type->u.pointer.type != src->type)
|
|
{
|
|
WARN("Type mismatch.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"Type mismatch in pointer store arguments.");
|
|
}
|
|
|
|
if (!dxil_record_validate_operand_count(record, i + 2, i + 2, sm6))
|
|
return;
|
|
|
|
alignment_code = record->operands[i++];
|
|
if (!bitcode_parse_alignment(alignment_code, &alignment))
|
|
WARN("Invalid alignment %"PRIu64".\n", alignment_code);
|
|
|
|
if (record->operands[i])
|
|
WARN("Ignoring volatile modifier.\n");
|
|
|
|
if (ptr->structure_stride)
|
|
{
|
|
VKD3D_ASSERT(ptr->u.reg.type == VKD3DSPR_GROUPSHAREDMEM);
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_STORE_STRUCTURED);
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 3, sm6)))
|
|
return;
|
|
if (ptr->u.reg.idx[1].rel_addr)
|
|
src_params[0] = *ptr->u.reg.idx[1].rel_addr;
|
|
else
|
|
src_param_make_constant_uint(&src_params[0], ptr->u.reg.idx[1].offset);
|
|
/* Struct offset is always zero as there is no struct, just an array. */
|
|
src_param_make_constant_uint(&src_params[1], 0);
|
|
src_param_init_from_value(&src_params[2], src);
|
|
}
|
|
else
|
|
{
|
|
operand_count = 1 + (ptr->u.reg.type == VKD3DSPR_GROUPSHAREDMEM);
|
|
vsir_instruction_init(ins, &sm6->p.location, (operand_count > 1) ? VKD3DSIH_STORE_RAW : VKD3DSIH_MOV);
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, operand_count, sm6)))
|
|
return;
|
|
if (operand_count > 1)
|
|
src_param_make_constant_uint(&src_params[0], 0);
|
|
src_param_init_from_value(&src_params[operand_count - 1], src);
|
|
}
|
|
|
|
dst_param = instruction_dst_params_alloc(ins, 1, sm6);
|
|
dst_param_init(dst_param);
|
|
dst_param->reg = ptr->u.reg;
|
|
dst_param->reg.alignment = alignment;
|
|
/* Groupshared stores contain the address in the src params. */
|
|
if (dst_param->reg.type != VKD3DSPR_IDXTEMP)
|
|
dst_param->reg.idx_count = 1;
|
|
}
|
|
|
|
static void sm6_parser_emit_switch(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct sm6_function *function, struct sm6_block *code_block, struct vkd3d_shader_instruction *ins)
|
|
{
|
|
struct sm6_block_terminator *terminator = &code_block->terminator;
|
|
const struct sm6_type *type;
|
|
const struct sm6_value *src;
|
|
unsigned int i = 1, j;
|
|
|
|
if (record->operand_count < 3 || !(record->operand_count & 1))
|
|
{
|
|
WARN("Invalid operand count %u.\n", record->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Invalid operand count %u for a switch instruction.", record->operand_count);
|
|
return;
|
|
}
|
|
|
|
if (!(type = sm6_parser_get_type(sm6, record->operands[0])))
|
|
return;
|
|
|
|
if (!(src = sm6_parser_get_value_by_ref(sm6, record, type, &i))
|
|
|| !sm6_value_validate_is_register(src, sm6))
|
|
return;
|
|
VKD3D_ASSERT(i == 2);
|
|
|
|
if (src->type != type)
|
|
{
|
|
WARN("Type mismatch.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"The type of a switch selector value does not match the selector type.");
|
|
}
|
|
if (!sm6_type_is_integer(type))
|
|
{
|
|
WARN("Selector is not scalar integer.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Selector type class %u of a switch instruction is not scalar integer.", type->class);
|
|
return;
|
|
}
|
|
|
|
terminator->conditional_reg = src->u.reg;
|
|
terminator->type = TERMINATOR_SWITCH;
|
|
|
|
terminator->case_count = record->operand_count / 2u;
|
|
if (!(terminator->cases = vkd3d_calloc(terminator->case_count, sizeof(*terminator->cases))))
|
|
{
|
|
ERR("Failed to allocate case array.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating a switch case array.");
|
|
return;
|
|
}
|
|
|
|
/* Executes 'operand_count / 2' times because operand_count is uneven. */
|
|
for (; i < record->operand_count; i += 2)
|
|
{
|
|
j = i / 2u - 1;
|
|
terminator->cases[j].block = sm6_function_get_block(function, record->operands[i], sm6);
|
|
/* For structurisation it is convenient to store the default in the case array. */
|
|
terminator->cases[j].is_default = !j;
|
|
}
|
|
|
|
for (i = 3; i < record->operand_count; i += 2)
|
|
{
|
|
if (!(src = sm6_parser_get_value_safe(sm6, record->operands[i])))
|
|
return;
|
|
|
|
if (src->type != type)
|
|
{
|
|
WARN("Type mismatch.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"The type of a switch case value does not match the selector type.");
|
|
}
|
|
if (!sm6_value_is_constant(src))
|
|
{
|
|
WARN("Case value is not a constant.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"A switch case value is not a constant.");
|
|
}
|
|
|
|
terminator->cases[i / 2u].value = sm6_value_get_constant_uint64(src);
|
|
}
|
|
|
|
ins->opcode = VKD3DSIH_NOP;
|
|
}
|
|
|
|
static void sm6_parser_emit_vselect(struct sm6_parser *sm6, const struct dxil_record *record,
|
|
struct vkd3d_shader_instruction *ins, struct sm6_value *dst)
|
|
{
|
|
struct vkd3d_shader_src_param *src_params;
|
|
const struct sm6_value *src[3];
|
|
unsigned int i = 0;
|
|
|
|
if (!(src[1] = sm6_parser_get_value_by_ref(sm6, record, NULL, &i))
|
|
|| !(src[2] = sm6_parser_get_value_by_ref(sm6, record, src[1]->type, &i))
|
|
|| !(src[0] = sm6_parser_get_value_by_ref(sm6, record, NULL, &i)))
|
|
{
|
|
return;
|
|
}
|
|
dxil_record_validate_operand_max_count(record, i, sm6);
|
|
|
|
for (i = 0; i < 3; ++i)
|
|
{
|
|
if (!sm6_value_validate_is_register(src[i], sm6))
|
|
return;
|
|
}
|
|
|
|
dst->type = src[1]->type;
|
|
|
|
if (!sm6_value_validate_is_bool(src[0], sm6))
|
|
return;
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_MOVC);
|
|
|
|
if (!(src_params = instruction_src_params_alloc(ins, 3, sm6)))
|
|
return;
|
|
for (i = 0; i < 3; ++i)
|
|
src_param_init_from_value(&src_params[i], src[i]);
|
|
|
|
instruction_dst_param_init_ssa_scalar(ins, sm6);
|
|
}
|
|
|
|
static bool sm6_metadata_value_is_node(const struct sm6_metadata_value *m)
|
|
{
|
|
return m && m->type == VKD3D_METADATA_NODE;
|
|
}
|
|
|
|
static bool sm6_metadata_value_is_value(const struct sm6_metadata_value *m)
|
|
{
|
|
return m && m->type == VKD3D_METADATA_VALUE;
|
|
}
|
|
|
|
static bool sm6_metadata_value_is_string(const struct sm6_metadata_value *m)
|
|
{
|
|
return m && m->type == VKD3D_METADATA_STRING;
|
|
}
|
|
|
|
static bool sm6_metadata_value_is_zero_or_undef(const struct sm6_metadata_value *m)
|
|
{
|
|
return sm6_metadata_value_is_value(m)
|
|
&& (sm6_value_is_undef(m->u.value) || sm6_value_is_constant_zero(m->u.value));
|
|
}
|
|
|
|
static bool sm6_metadata_get_uint_value(const struct sm6_parser *sm6,
|
|
const struct sm6_metadata_value *m, unsigned int *u)
|
|
{
|
|
const struct sm6_value *value;
|
|
|
|
if (!m || m->type != VKD3D_METADATA_VALUE)
|
|
return false;
|
|
|
|
value = m->u.value;
|
|
if (!sm6_value_is_constant(value))
|
|
return false;
|
|
if (!sm6_type_is_integer(value->type))
|
|
return false;
|
|
|
|
*u = register_get_uint_value(&value->u.reg);
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_metadata_get_uint64_value(const struct sm6_parser *sm6,
|
|
const struct sm6_metadata_value *m, uint64_t *u)
|
|
{
|
|
const struct sm6_value *value;
|
|
|
|
if (!m || m->type != VKD3D_METADATA_VALUE)
|
|
return false;
|
|
|
|
value = m->u.value;
|
|
if (!sm6_value_is_constant(value))
|
|
return false;
|
|
if (!sm6_type_is_integer(value->type))
|
|
return false;
|
|
|
|
*u = register_get_uint64_value(&value->u.reg);
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool sm6_metadata_get_float_value(const struct sm6_parser *sm6,
|
|
const struct sm6_metadata_value *m, float *f)
|
|
{
|
|
const struct sm6_value *value;
|
|
|
|
if (!m || m->type != VKD3D_METADATA_VALUE)
|
|
return false;
|
|
|
|
value = m->u.value;
|
|
if (!sm6_value_is_constant(value))
|
|
return false;
|
|
if (!sm6_type_is_floating_point(value->type))
|
|
return false;
|
|
|
|
*f = register_get_float_value(&value->u.reg);
|
|
|
|
return true;
|
|
}
|
|
|
|
static void sm6_parser_metadata_attachment_block_init(struct sm6_parser *sm6, const struct dxil_block *target_block,
|
|
const struct dxil_block *block)
|
|
{
|
|
struct dxil_record *target_record;
|
|
const struct dxil_record *record;
|
|
unsigned int i;
|
|
uint64_t index;
|
|
|
|
for (i = 0; i < block->record_count; ++i)
|
|
{
|
|
record = block->records[i];
|
|
if (record->code != METADATA_ATTACHMENT)
|
|
{
|
|
WARN("Ignoring record with code %u.\n", record->code);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_ATTACHMENT,
|
|
"Ignoring a metadata attachment record with code %u.", record->code);
|
|
continue;
|
|
}
|
|
if (!(record->operand_count & 1))
|
|
{
|
|
WARN("Ignoring function attachment.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_ATTACHMENT,
|
|
"Ignoring a metadata function attachment.");
|
|
continue;
|
|
}
|
|
|
|
index = record->operands[0];
|
|
if (!target_block->record_count || index >= target_block->record_count - 1)
|
|
{
|
|
WARN("Invalid record index %"PRIu64".\n", index);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid record index %"PRIu64" for a metadata attachment.", index);
|
|
continue;
|
|
}
|
|
/* 'index' is an instruction index, but records[0] is DECLAREBLOCKS, not an instruction. */
|
|
target_record = target_block->records[index + 1];
|
|
if (target_record->attachment)
|
|
{
|
|
WARN("Overwriting record attachment.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_ATTACHMENT,
|
|
"The target record for a metadata attachment already has an attachment.");
|
|
}
|
|
target_record->attachment = record;
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_metadata_attachments_init(struct sm6_parser *sm6, const struct dxil_block *block)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < block->child_block_count; ++i)
|
|
{
|
|
if (block->child_blocks[i]->id == METADATA_ATTACHMENT_BLOCK)
|
|
sm6_parser_metadata_attachment_block_init(sm6, block, block->child_blocks[i]);
|
|
}
|
|
}
|
|
|
|
static const struct sm6_metadata_value *sm6_parser_find_metadata_kind(const struct sm6_parser *sm6, uint64_t id)
|
|
{
|
|
unsigned int i, j;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(sm6->metadata_tables); ++i)
|
|
{
|
|
for (j = 0; j < sm6->metadata_tables[i].count; ++j)
|
|
{
|
|
if (sm6->metadata_tables[i].values[j].type == VKD3D_METADATA_KIND
|
|
&& sm6->metadata_tables[i].values[j].u.kind.id == id)
|
|
return &sm6->metadata_tables[i].values[j];
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static const struct sm6_metadata_value *sm6_parser_metadata_get_value(const struct sm6_parser *sm6, uint64_t index)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(sm6->metadata_tables); ++i)
|
|
{
|
|
if (sm6->metadata_tables[i].count > index)
|
|
break;
|
|
index -= sm6->metadata_tables[i].count;
|
|
}
|
|
|
|
return (index < sm6->metadata_tables[i].count) ? &sm6->metadata_tables[i].values[index] : NULL;
|
|
}
|
|
|
|
static bool metadata_node_get_unary_uint(const struct sm6_metadata_node *node, unsigned int *operand,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
if (node->operand_count != 1)
|
|
{
|
|
FIXME("Ignoring node with %u operands.\n", node->operand_count);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_ATTACHMENT,
|
|
"Ignoring metadata attachment node with %u operands; expected unary.", node->operand_count);
|
|
return false;
|
|
}
|
|
if (!sm6_metadata_value_is_value(node->operands[0])
|
|
|| !sm6_metadata_get_uint_value(sm6, node->operands[0], operand))
|
|
{
|
|
WARN("Failed to get operand value.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_ATTACHMENT,
|
|
"Failed to get a metadata attachment operand value; ignoring the attachment.");
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static void metadata_attachment_record_apply(const struct dxil_record *record, enum bitcode_function_code func_code,
|
|
struct vkd3d_shader_instruction *ins, struct sm6_value *dst, struct sm6_parser *sm6)
|
|
{
|
|
static const char *ignored_names[] =
|
|
{
|
|
"alias.scope",
|
|
"dx.controlflow.hints",
|
|
"llvm.loop",
|
|
"noalias",
|
|
"tbaa",
|
|
"range",
|
|
};
|
|
const struct sm6_metadata_node *node;
|
|
const struct sm6_metadata_value *m;
|
|
unsigned int i, j, operand;
|
|
bool ignored = false;
|
|
const char *name;
|
|
|
|
if (record->attachment)
|
|
{
|
|
WARN("Ignoring nested metadata attachment.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_ATTACHMENT,
|
|
"Ignoring a nested metadata attachment.");
|
|
}
|
|
|
|
VKD3D_ASSERT(record->operand_count & 1);
|
|
for (i = 1; i < record->operand_count; i += 2)
|
|
{
|
|
if (!(m = sm6_parser_find_metadata_kind(sm6, record->operands[i])))
|
|
{
|
|
WARN("Failed to find metadata kind %"PRIx64".\n", record->operands[i]);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_ATTACHMENT,
|
|
"Failed to find metadata kind %"PRIx64" for an attachment.", record->operands[i]);
|
|
continue;
|
|
}
|
|
name = m->u.kind.name;
|
|
|
|
m = sm6_parser_metadata_get_value(sm6, record->operands[i + 1]);
|
|
if (!m || !sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Failed to retrieve metadata attachment node.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_ATTACHMENT,
|
|
"Failed to retrieve a metadata attachment node.");
|
|
continue;
|
|
}
|
|
node = m->u.node;
|
|
|
|
if (!strcmp(name, "dx.precise"))
|
|
{
|
|
if (!sm6_value_is_register(dst))
|
|
{
|
|
WARN("Precise value is not a register.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_ATTACHMENT,
|
|
"A value marked as precise is not a register.");
|
|
}
|
|
else if (metadata_node_get_unary_uint(node, &operand, sm6) && operand)
|
|
{
|
|
ins->flags |= sm6_type_is_scalar(dst->type) ? VKD3DSI_PRECISE_X : VKD3DSI_PRECISE_XYZW;
|
|
}
|
|
}
|
|
else if (!strcmp(name, "dx.nonuniform"))
|
|
{
|
|
if (!sm6_value_is_register(dst))
|
|
{
|
|
WARN("Non-uniform value is not a register.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_ATTACHMENT,
|
|
"A value marked as non-uniform is not a register.");
|
|
}
|
|
else if (metadata_node_get_unary_uint(node, &operand, sm6))
|
|
{
|
|
dst->u.reg.non_uniform = !!operand;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
for (j = 0; j < ARRAY_SIZE(ignored_names); ++j)
|
|
if (!strcmp(name, ignored_names[j]))
|
|
break;
|
|
if (j == ARRAY_SIZE(ignored_names))
|
|
{
|
|
WARN("Ignoring metadata attachment '%s'.\n", name);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_ATTACHMENT,
|
|
"Ignoring a metadata attachment named '%s'.", name);
|
|
}
|
|
ignored = true;
|
|
}
|
|
|
|
if (func_code != FUNC_CODE_INST_CALL && !ignored)
|
|
WARN("Metadata attachment target is not a function call.\n");
|
|
}
|
|
}
|
|
|
|
static bool sm6_function_blocks_reserve(struct sm6_function *function, unsigned int reserve)
|
|
{
|
|
if (!vkd3d_array_reserve((void **)&function->blocks, &function->block_capacity,
|
|
reserve, sizeof(*function->blocks)))
|
|
{
|
|
ERR("Failed to allocate code block array.\n");
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static struct sm6_block *sm6_function_create_block(struct sm6_function *function)
|
|
{
|
|
struct sm6_block *block;
|
|
|
|
if (!(block = sm6_block_create()))
|
|
return NULL;
|
|
|
|
function->blocks[function->block_count++] = block;
|
|
/* Set the id to the array index + 1. */
|
|
block->id = function->block_count;
|
|
|
|
return block;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_function_resolve_phi_incomings(const struct sm6_function *function,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
const struct sm6_block *block;
|
|
size_t i, j, block_idx;
|
|
|
|
for (block_idx = 0; block_idx < function->block_count; ++block_idx)
|
|
{
|
|
block = function->blocks[block_idx];
|
|
|
|
for (i = 0; i < block->phi_count; ++i)
|
|
{
|
|
struct sm6_phi *phi = &block->phi[i];
|
|
const struct sm6_value *src;
|
|
|
|
for (j = 0; j < phi->incoming_count; ++j)
|
|
{
|
|
src = &sm6->values[phi->incoming[j].reg.idx[0].offset];
|
|
if (!sm6_value_is_constant(src) && !sm6_value_is_undef(src) && !sm6_value_is_ssa(src))
|
|
{
|
|
FIXME("PHI incoming value is not a constant or SSA register.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"A PHI incoming value is not a constant or SSA register.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
if (src->u.reg.data_type != phi->reg.data_type)
|
|
{
|
|
WARN("Type mismatch.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"The type of a phi incoming value does not match the result type.");
|
|
}
|
|
phi->incoming[j].reg = src->u.reg;
|
|
}
|
|
}
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_function_init(struct sm6_parser *sm6, const struct dxil_block *block,
|
|
struct sm6_function *function)
|
|
{
|
|
struct vsir_program *program = sm6->p.program;
|
|
struct vkd3d_shader_instruction *ins;
|
|
size_t i, block_idx, block_count;
|
|
const struct dxil_record *record;
|
|
const struct sm6_type *fwd_type;
|
|
bool ret_found, is_terminator;
|
|
struct sm6_block *code_block;
|
|
struct sm6_value *dst;
|
|
|
|
if (!(function->declaration = sm6_parser_next_function_definition(sm6)))
|
|
{
|
|
WARN("Failed to find definition to match function body.\n");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (block->record_count < 2)
|
|
{
|
|
/* It should contain at least a block count and a RET instruction. */
|
|
WARN("Invalid function block record count %zu.\n", block->record_count);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
if (block->records[0]->code != FUNC_CODE_DECLAREBLOCKS || !block->records[0]->operand_count
|
|
|| block->records[0]->operands[0] > UINT_MAX)
|
|
{
|
|
WARN("Block count declaration not found or invalid.\n");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (!(block_count = block->records[0]->operands[0]))
|
|
{
|
|
WARN("Function contains no blocks.\n");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (!sm6_function_blocks_reserve(function, block_count))
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
|
|
/* Pre-allocate all blocks to simplify instruction parsing. */
|
|
for (i = 0; i < block_count; ++i)
|
|
{
|
|
if (!sm6_function_create_block(function))
|
|
{
|
|
ERR("Failed to allocate code block.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
}
|
|
function->block_count = block_count;
|
|
code_block = function->blocks[0];
|
|
|
|
sm6->cur_max_value = function->value_count;
|
|
|
|
for (i = 1, block_idx = 0, ret_found = false; i < block->record_count; ++i)
|
|
{
|
|
sm6->p.location.column = i;
|
|
|
|
if (!code_block)
|
|
{
|
|
WARN("Invalid block count %zu.\n", function->block_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND,
|
|
"Invalid block count %zu.", function->block_count);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
/* Some instructions can emit >1 IR instruction, so extra may be used. */
|
|
if (!vkd3d_array_reserve((void **)&code_block->instructions, &code_block->instruction_capacity,
|
|
code_block->instruction_count + MAX_IR_INSTRUCTIONS_PER_DXIL_INSTRUCTION,
|
|
sizeof(*code_block->instructions)))
|
|
{
|
|
ERR("Failed to allocate instructions.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
ins = &code_block->instructions[code_block->instruction_count];
|
|
ins->opcode = VKD3DSIH_INVALID;
|
|
|
|
dst = sm6_parser_get_current_value(sm6);
|
|
fwd_type = dst->type;
|
|
dst->type = NULL;
|
|
dst->value_type = VALUE_TYPE_REG;
|
|
dst->is_back_ref = true;
|
|
is_terminator = false;
|
|
|
|
record = block->records[i];
|
|
switch (record->code)
|
|
{
|
|
case FUNC_CODE_INST_ALLOCA:
|
|
sm6_parser_emit_alloca(sm6, record, ins, dst);
|
|
break;
|
|
case FUNC_CODE_INST_ATOMICRMW:
|
|
{
|
|
struct function_emission_state state = {code_block, ins};
|
|
sm6_parser_emit_atomicrmw(sm6, record, &state, dst);
|
|
program->temp_count = max(program->temp_count, state.temp_idx);
|
|
break;
|
|
}
|
|
case FUNC_CODE_INST_BINOP:
|
|
sm6_parser_emit_binop(sm6, record, ins, dst);
|
|
break;
|
|
case FUNC_CODE_INST_BR:
|
|
sm6_parser_emit_br(sm6, record, function, code_block, ins);
|
|
is_terminator = true;
|
|
break;
|
|
case FUNC_CODE_INST_CALL:
|
|
{
|
|
struct function_emission_state state = {code_block, ins};
|
|
sm6_parser_emit_call(sm6, record, &state, dst);
|
|
program->temp_count = max(program->temp_count, state.temp_idx);
|
|
break;
|
|
}
|
|
case FUNC_CODE_INST_CAST:
|
|
sm6_parser_emit_cast(sm6, record, ins, dst);
|
|
break;
|
|
case FUNC_CODE_INST_CMP2:
|
|
sm6_parser_emit_cmp2(sm6, record, ins, dst);
|
|
break;
|
|
case FUNC_CODE_INST_CMPXCHG:
|
|
sm6_parser_emit_cmpxchg(sm6, record, ins, dst);
|
|
break;
|
|
case FUNC_CODE_INST_EXTRACTVAL:
|
|
sm6_parser_emit_extractval(sm6, record, ins, dst);
|
|
break;
|
|
case FUNC_CODE_INST_GEP:
|
|
sm6_parser_emit_gep(sm6, record, ins, dst);
|
|
break;
|
|
case FUNC_CODE_INST_LOAD:
|
|
sm6_parser_emit_load(sm6, record, ins, dst);
|
|
break;
|
|
case FUNC_CODE_INST_PHI:
|
|
sm6_parser_emit_phi(sm6, record, function, code_block, ins, dst);
|
|
break;
|
|
case FUNC_CODE_INST_RET:
|
|
sm6_parser_emit_ret(sm6, record, code_block, ins);
|
|
is_terminator = true;
|
|
ret_found = true;
|
|
break;
|
|
case FUNC_CODE_INST_STORE:
|
|
sm6_parser_emit_store(sm6, record, ins, dst);
|
|
break;
|
|
case FUNC_CODE_INST_SWITCH:
|
|
sm6_parser_emit_switch(sm6, record, function, code_block, ins);
|
|
is_terminator = true;
|
|
break;
|
|
case FUNC_CODE_INST_VSELECT:
|
|
sm6_parser_emit_vselect(sm6, record, ins, dst);
|
|
break;
|
|
default:
|
|
FIXME("Unhandled dxil instruction %u.\n", record->code);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (sm6->p.failed)
|
|
return VKD3D_ERROR;
|
|
|
|
if (record->attachment)
|
|
metadata_attachment_record_apply(record->attachment, record->code, ins, dst, sm6);
|
|
|
|
if (is_terminator)
|
|
{
|
|
++block_idx;
|
|
code_block = (block_idx < function->block_count) ? function->blocks[block_idx] : NULL;
|
|
}
|
|
if (code_block)
|
|
code_block->instruction_count += ins->opcode != VKD3DSIH_NOP;
|
|
|
|
if (dst->type && fwd_type && dst->type != fwd_type)
|
|
{
|
|
WARN("Type mismatch.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"The type of a result value does not match the type defined by a forward reference.");
|
|
}
|
|
|
|
sm6->value_count += !!dst->type;
|
|
}
|
|
|
|
if (!ret_found)
|
|
{
|
|
WARN("Function contains no RET instruction.\n");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
return sm6_function_resolve_phi_incomings(function, sm6);
|
|
}
|
|
|
|
static void sm6_block_emit_terminator(const struct sm6_block *block, struct sm6_parser *sm6)
|
|
{
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_instruction *ins;
|
|
unsigned int i, count;
|
|
|
|
switch (block->terminator.type)
|
|
{
|
|
case TERMINATOR_UNCOND_BR:
|
|
if (!block->terminator.true_block)
|
|
return;
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_BRANCH);
|
|
if (!(src_params = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
vsir_src_param_init_label(&src_params[0], block->terminator.true_block->id);
|
|
break;
|
|
|
|
case TERMINATOR_COND_BR:
|
|
if (!block->terminator.true_block || !block->terminator.false_block)
|
|
return;
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_BRANCH);
|
|
if (!(src_params = instruction_src_params_alloc(ins, 3, sm6)))
|
|
return;
|
|
src_param_init(&src_params[0]);
|
|
src_params[0].reg = block->terminator.conditional_reg;
|
|
vsir_src_param_init_label(&src_params[1], block->terminator.true_block->id);
|
|
vsir_src_param_init_label(&src_params[2], block->terminator.false_block->id);
|
|
break;
|
|
|
|
case TERMINATOR_SWITCH:
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_SWITCH_MONOLITHIC);
|
|
if (!(src_params = instruction_src_params_alloc(ins, block->terminator.case_count * 2u + 1, sm6)))
|
|
return;
|
|
src_param_init(&src_params[0]);
|
|
src_params[0].reg = block->terminator.conditional_reg;
|
|
/* TODO: emit the merge block id. */
|
|
vsir_src_param_init_label(&src_params[2], 0);
|
|
|
|
for (i = 0, count = 3; i < block->terminator.case_count; ++i)
|
|
{
|
|
const struct terminator_case *switch_case;
|
|
const struct sm6_block *case_block;
|
|
|
|
switch_case = &block->terminator.cases[i];
|
|
if (!(case_block = switch_case->block))
|
|
{
|
|
VKD3D_ASSERT(sm6->p.failed);
|
|
continue;
|
|
}
|
|
if (switch_case->is_default)
|
|
{
|
|
vsir_src_param_init_label(&src_params[1], case_block->id);
|
|
continue;
|
|
}
|
|
|
|
if (src_params[0].reg.data_type == VKD3D_DATA_UINT64)
|
|
{
|
|
vsir_src_param_init(&src_params[count], VKD3DSPR_IMMCONST64, VKD3D_DATA_UINT64, 0);
|
|
src_params[count++].reg.u.immconst_u64[0] = switch_case->value;
|
|
}
|
|
else
|
|
{
|
|
if (switch_case->value > UINT_MAX)
|
|
{
|
|
WARN("Truncating 64-bit constant %"PRIx64".\n", switch_case->value);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"Truncating 64-bit switch case value %"PRIx64" to 32 bits.", switch_case->value);
|
|
}
|
|
vsir_src_param_init(&src_params[count], VKD3DSPR_IMMCONST, VKD3D_DATA_UINT, 0);
|
|
src_params[count++].reg.u.immconst_u32[0] = switch_case->value;
|
|
}
|
|
vsir_src_param_init_label(&src_params[count++], case_block->id);
|
|
}
|
|
|
|
break;
|
|
|
|
case TERMINATOR_RET:
|
|
sm6_parser_add_instruction(sm6, VKD3DSIH_RET);
|
|
break;
|
|
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
static void sm6_block_emit_phi(const struct sm6_block *block, struct sm6_parser *sm6)
|
|
{
|
|
struct vkd3d_shader_instruction *ins;
|
|
unsigned int i, j, incoming_count;
|
|
const struct sm6_phi *src_phi;
|
|
|
|
for (i = 0; i < block->phi_count; ++i)
|
|
{
|
|
struct vkd3d_shader_src_param *src_params;
|
|
struct vkd3d_shader_dst_param *dst_param;
|
|
|
|
src_phi = &block->phi[i];
|
|
incoming_count = src_phi->incoming_count;
|
|
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_PHI);
|
|
if (!(src_params = instruction_src_params_alloc(ins, incoming_count * 2u, sm6)))
|
|
return;
|
|
if (!(dst_param = instruction_dst_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
|
|
for (j = 0; j < incoming_count; ++j)
|
|
{
|
|
const struct sm6_block *incoming_block = src_phi->incoming[j].block;
|
|
unsigned int index = j * 2;
|
|
|
|
src_param_init(&src_params[index]);
|
|
src_params[index].reg = src_phi->incoming[j].reg;
|
|
if (incoming_block)
|
|
vsir_src_param_init_label(&src_params[index + 1], incoming_block->id);
|
|
else
|
|
VKD3D_ASSERT(sm6->p.failed);
|
|
}
|
|
|
|
dst_param_init(dst_param);
|
|
dst_param->reg = src_phi->reg;
|
|
}
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_module_init(struct sm6_parser *sm6, const struct dxil_block *block,
|
|
unsigned int level)
|
|
{
|
|
size_t i, old_value_count = sm6->value_count;
|
|
struct sm6_function *function;
|
|
enum vkd3d_result ret;
|
|
|
|
for (i = 0; i < block->child_block_count; ++i)
|
|
{
|
|
if ((ret = sm6_parser_module_init(sm6, block->child_blocks[i], level + 1)) < 0)
|
|
return ret;
|
|
}
|
|
|
|
sm6->p.location.line = block->id;
|
|
sm6->p.location.column = 0;
|
|
|
|
sm6_parser_metadata_attachments_init(sm6, block);
|
|
|
|
switch (block->id)
|
|
{
|
|
case CONSTANTS_BLOCK:
|
|
/* Level 1 (global) constants are already done in sm6_parser_globals_init(). */
|
|
if (level < 2)
|
|
break;
|
|
function = &sm6->functions[sm6->function_count];
|
|
sm6->cur_max_value = function->value_count;
|
|
return sm6_parser_constants_init(sm6, block);
|
|
|
|
case FUNCTION_BLOCK:
|
|
function = &sm6->functions[sm6->function_count];
|
|
if ((ret = sm6_parser_function_init(sm6, block, function)) < 0)
|
|
return ret;
|
|
/* The value index returns to its previous value after handling a function. It's usually nonzero
|
|
* at the start because of global constants/variables/function declarations. Function constants
|
|
* occur in a child block, so value_count is already saved before they are emitted. */
|
|
memset(&sm6->values[old_value_count], 0, (sm6->value_count - old_value_count) * sizeof(*sm6->values));
|
|
sm6->value_count = old_value_count;
|
|
break;
|
|
|
|
case BLOCKINFO_BLOCK:
|
|
case MODULE_BLOCK:
|
|
case PARAMATTR_BLOCK:
|
|
case PARAMATTR_GROUP_BLOCK:
|
|
case VALUE_SYMTAB_BLOCK:
|
|
case METADATA_BLOCK:
|
|
case METADATA_ATTACHMENT_BLOCK:
|
|
case TYPE_BLOCK:
|
|
break;
|
|
|
|
default:
|
|
FIXME("Unhandled block id %u.\n", block->id);
|
|
break;
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static void sm6_parser_emit_label(struct sm6_parser *sm6, unsigned int label_id)
|
|
{
|
|
struct vkd3d_shader_src_param *src_param;
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_LABEL);
|
|
|
|
if (!(src_param = instruction_src_params_alloc(ins, 1, sm6)))
|
|
return;
|
|
vsir_src_param_init_label(src_param, label_id);
|
|
}
|
|
|
|
static enum vkd3d_result sm6_function_emit_blocks(const struct sm6_function *function, struct sm6_parser *sm6)
|
|
{
|
|
struct vsir_program *program = sm6->p.program;
|
|
unsigned int i;
|
|
|
|
program->block_count = function->block_count;
|
|
|
|
for (i = 0; i < function->block_count; ++i)
|
|
{
|
|
const struct sm6_block *block = function->blocks[i];
|
|
|
|
/* Space for the label and terminator. */
|
|
if (!sm6_parser_require_space(sm6, block->instruction_count + block->phi_count + 2))
|
|
{
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory emitting shader instructions.");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
sm6_parser_emit_label(sm6, block->id);
|
|
sm6_block_emit_phi(block, sm6);
|
|
|
|
memcpy(&program->instructions.elements[program->instructions.count], block->instructions,
|
|
block->instruction_count * sizeof(*block->instructions));
|
|
program->instructions.count += block->instruction_count;
|
|
|
|
sm6_block_emit_terminator(block, sm6);
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static bool sm6_parser_allocate_named_metadata(struct sm6_parser *sm6)
|
|
{
|
|
struct dxil_block *block;
|
|
unsigned int i, j, count;
|
|
|
|
for (i = 0, count = 0; i < sm6->root_block.child_block_count; ++i)
|
|
{
|
|
block = sm6->root_block.child_blocks[i];
|
|
if (block->id != METADATA_BLOCK)
|
|
continue;
|
|
for (j = 0; j < block->record_count; ++j)
|
|
count += block->records[j]->code == METADATA_NAMED_NODE;
|
|
}
|
|
|
|
if (!count)
|
|
return true;
|
|
|
|
return !!(sm6->named_metadata = vkd3d_calloc(count, sizeof(*sm6->named_metadata)));
|
|
}
|
|
|
|
static enum vkd3d_result metadata_value_create_node(struct sm6_metadata_value *m, struct sm6_metadata_table *table,
|
|
unsigned int dst_idx, unsigned int end_count, const struct dxil_record *record, struct sm6_parser *sm6)
|
|
{
|
|
struct sm6_metadata_node *node;
|
|
unsigned int i, offset;
|
|
|
|
m->type = VKD3D_METADATA_NODE;
|
|
if (!(m->value_type = sm6->metadata_type))
|
|
{
|
|
WARN("Metadata type not found.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_METADATA,
|
|
"The type for metadata values was not found.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
if (!(node = vkd3d_malloc(offsetof(struct sm6_metadata_node, operands[record->operand_count]))))
|
|
{
|
|
ERR("Failed to allocate metadata node with %u operands.\n", record->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating a metadata node with %u operands.", record->operand_count);
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
m->u.node = node;
|
|
|
|
node->is_distinct = record->code == METADATA_DISTINCT_NODE;
|
|
|
|
offset = record->code != METADATA_NAMED_NODE;
|
|
|
|
for (i = 0; i < record->operand_count; ++i)
|
|
{
|
|
uint64_t ref;
|
|
|
|
ref = record->operands[i] - offset;
|
|
if (record->operands[i] >= offset && ref >= end_count)
|
|
{
|
|
WARN("Invalid metadata index %"PRIu64".\n", ref);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_METADATA,
|
|
"Metadata index %"PRIu64" is invalid.", ref);
|
|
vkd3d_free(node);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (!node->is_distinct && ref == dst_idx)
|
|
{
|
|
WARN("Metadata self-reference at index %u.\n", dst_idx);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_METADATA,
|
|
"Metadata index %u is self-referencing.", dst_idx);
|
|
vkd3d_free(node);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
node->operands[i] = (record->operands[i] >= offset) ? &table->values[ref] : NULL;
|
|
if (record->code == METADATA_NAMED_NODE && !sm6_metadata_value_is_node(node->operands[i]))
|
|
{
|
|
WARN("Named node operand is not a node.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_METADATA,
|
|
"The operand of a metadata named node is not a node.");
|
|
vkd3d_free(node);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
}
|
|
|
|
node->operand_count = record->operand_count;
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_metadata_init(struct sm6_parser *sm6, const struct dxil_block *block,
|
|
struct sm6_metadata_table *table)
|
|
{
|
|
unsigned int i, count, table_idx, value_idx;
|
|
struct sm6_metadata_value *values, *m;
|
|
const struct dxil_record *record;
|
|
const struct sm6_value *value;
|
|
enum vkd3d_result ret;
|
|
char *name;
|
|
|
|
for (i = 0, count = 0; i < block->record_count; ++i)
|
|
count += block->records[i]->code != METADATA_NAME;
|
|
|
|
if (!(values = vkd3d_calloc(count, sizeof(*values))))
|
|
{
|
|
ERR("Failed to allocate metadata tables.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating metadata tables.");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
table->values = values;
|
|
|
|
for (i = 0, name = NULL; i < block->record_count; ++i)
|
|
{
|
|
record = block->records[i];
|
|
|
|
table_idx = table->count;
|
|
m = &values[table_idx];
|
|
|
|
switch (record->code)
|
|
{
|
|
case METADATA_NAMED_NODE:
|
|
if (!name)
|
|
{
|
|
WARN("Named node has no name.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_METADATA,
|
|
"A metadata named node has no name.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
/* When DXC emits metadata value array reference indices it assumes named nodes
|
|
* are not included in the array. Store named nodes separately. */
|
|
m = &sm6->named_metadata[sm6->named_metadata_count].value;
|
|
sm6->named_metadata[sm6->named_metadata_count].name = name;
|
|
name = NULL;
|
|
|
|
if ((ret = metadata_value_create_node(m, table, UINT_MAX, count, record, sm6)) < 0)
|
|
return ret;
|
|
++sm6->named_metadata_count;
|
|
/* Skip incrementing the table count. */
|
|
continue;
|
|
|
|
case METADATA_DISTINCT_NODE:
|
|
case METADATA_NODE:
|
|
if ((ret = metadata_value_create_node(m, table, table_idx, count, record, sm6)) < 0)
|
|
return ret;
|
|
break;
|
|
|
|
case METADATA_KIND:
|
|
if (!dxil_record_validate_operand_min_count(record, 2, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
m->type = VKD3D_METADATA_KIND;
|
|
m->u.kind.id = record->operands[0];
|
|
if (!(m->u.kind.name = dxil_record_to_string(record, 1, sm6)))
|
|
{
|
|
ERR("Failed to allocate name of a kind.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
break;
|
|
|
|
case METADATA_NAME:
|
|
/* Check the next record to avoid freeing 'name' in all exit paths. */
|
|
if (i + 1 == block->record_count || block->records[i + 1]->code != METADATA_NAMED_NODE)
|
|
{
|
|
WARN("Name is not followed by a named node.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_METADATA,
|
|
"A metadata node name is not followed by a named node.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
/* LLVM allows an empty string here. */
|
|
if (!(name = dxil_record_to_string(record, 0, sm6)))
|
|
{
|
|
ERR("Failed to allocate name.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
continue;
|
|
|
|
case METADATA_STRING:
|
|
/* LLVM allows an empty string here. */
|
|
m->type = VKD3D_METADATA_STRING;
|
|
if (!(m->u.string_value = dxil_record_to_string(record, 0, sm6)))
|
|
{
|
|
ERR("Failed to allocate string.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
break;
|
|
|
|
case METADATA_VALUE:
|
|
if (!dxil_record_validate_operand_count(record, 2, 2, sm6))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
m->type = VKD3D_METADATA_VALUE;
|
|
if (!(m->value_type = sm6_parser_get_type(sm6, record->operands[0])))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
if (record->operands[1] > UINT_MAX)
|
|
WARN("Truncating value index %"PRIu64".\n", record->operands[1]);
|
|
value_idx = record->operands[1];
|
|
if (!(value = sm6_parser_get_value_safe(sm6, value_idx)))
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
|
|
if (!sm6_value_is_constant(value) && !sm6_value_is_undef(value) && !sm6_value_is_icb(value)
|
|
&& !sm6_value_is_function_dcl(value))
|
|
{
|
|
WARN("Value at index %u is not a constant or a function declaration.\n", value_idx);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_METADATA,
|
|
"Metadata value at index %u is not a constant or a function declaration.", value_idx);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
m->u.value = value;
|
|
|
|
if (value->type != m->value_type)
|
|
{
|
|
WARN("Type mismatch.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"The type of a metadata value does not match its referenced value at index %u.", value_idx);
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
FIXME("Unhandled metadata type %u.\n", record->code);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_METADATA,
|
|
"Metadata type %u is unhandled.", record->code);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
++table->count;
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_shader_component_type vkd3d_component_type_from_dxil_component_type(enum dxil_component_type type)
|
|
{
|
|
switch (type)
|
|
{
|
|
case COMPONENT_TYPE_I1:
|
|
return VKD3D_SHADER_COMPONENT_BOOL;
|
|
case COMPONENT_TYPE_I16:
|
|
case COMPONENT_TYPE_I32:
|
|
return VKD3D_SHADER_COMPONENT_INT;
|
|
case COMPONENT_TYPE_U16:
|
|
case COMPONENT_TYPE_U32:
|
|
return VKD3D_SHADER_COMPONENT_UINT;
|
|
case COMPONENT_TYPE_F16:
|
|
case COMPONENT_TYPE_F32:
|
|
case COMPONENT_TYPE_SNORMF32:
|
|
case COMPONENT_TYPE_UNORMF32:
|
|
return VKD3D_SHADER_COMPONENT_FLOAT;
|
|
case COMPONENT_TYPE_F64:
|
|
case COMPONENT_TYPE_SNORMF64:
|
|
case COMPONENT_TYPE_UNORMF64:
|
|
return VKD3D_SHADER_COMPONENT_DOUBLE;
|
|
default:
|
|
FIXME("Unhandled component type %u.\n", type);
|
|
return VKD3D_SHADER_COMPONENT_UINT;
|
|
}
|
|
}
|
|
|
|
static enum vkd3d_shader_minimum_precision minimum_precision_from_dxil_component_type(enum dxil_component_type type)
|
|
{
|
|
switch (type)
|
|
{
|
|
case COMPONENT_TYPE_F16:
|
|
return VKD3D_SHADER_MINIMUM_PRECISION_FLOAT_16;
|
|
case COMPONENT_TYPE_I16:
|
|
return VKD3D_SHADER_MINIMUM_PRECISION_INT_16;
|
|
case COMPONENT_TYPE_U16:
|
|
return VKD3D_SHADER_MINIMUM_PRECISION_UINT_16;
|
|
default:
|
|
return VKD3D_SHADER_MINIMUM_PRECISION_NONE;
|
|
}
|
|
}
|
|
|
|
static const enum vkd3d_shader_sysval_semantic sysval_semantic_table[] =
|
|
{
|
|
[SEMANTIC_KIND_ARBITRARY] = VKD3D_SHADER_SV_NONE,
|
|
[SEMANTIC_KIND_VERTEXID] = VKD3D_SHADER_SV_VERTEX_ID,
|
|
[SEMANTIC_KIND_INSTANCEID] = VKD3D_SHADER_SV_INSTANCE_ID,
|
|
[SEMANTIC_KIND_POSITION] = VKD3D_SHADER_SV_POSITION,
|
|
[SEMANTIC_KIND_CLIPDISTANCE] = VKD3D_SHADER_SV_CLIP_DISTANCE,
|
|
[SEMANTIC_KIND_CULLDISTANCE] = VKD3D_SHADER_SV_CULL_DISTANCE,
|
|
[SEMANTIC_KIND_PRIMITIVEID] = VKD3D_SHADER_SV_PRIMITIVE_ID,
|
|
[SEMANTIC_KIND_ISFRONTFACE] = VKD3D_SHADER_SV_IS_FRONT_FACE,
|
|
[SEMANTIC_KIND_COVERAGE] = VKD3D_SHADER_SV_COVERAGE,
|
|
[SEMANTIC_KIND_TARGET] = VKD3D_SHADER_SV_TARGET,
|
|
[SEMANTIC_KIND_DEPTH] = VKD3D_SHADER_SV_DEPTH,
|
|
[SEMANTIC_KIND_DEPTHLESSEQUAL] = VKD3D_SHADER_SV_DEPTH_LESS_EQUAL,
|
|
[SEMANTIC_KIND_DEPTHGREATEREQUAL] = VKD3D_SHADER_SV_DEPTH_GREATER_EQUAL,
|
|
};
|
|
|
|
static enum vkd3d_shader_sysval_semantic sysval_semantic_from_dxil_semantic_kind(enum dxil_semantic_kind kind,
|
|
enum vkd3d_tessellator_domain domain)
|
|
{
|
|
if (kind == SEMANTIC_KIND_TESSFACTOR)
|
|
{
|
|
switch (domain)
|
|
{
|
|
case VKD3D_TESSELLATOR_DOMAIN_LINE:
|
|
return VKD3D_SHADER_SV_TESS_FACTOR_LINEDET;
|
|
case VKD3D_TESSELLATOR_DOMAIN_TRIANGLE:
|
|
return VKD3D_SHADER_SV_TESS_FACTOR_TRIEDGE;
|
|
case VKD3D_TESSELLATOR_DOMAIN_QUAD:
|
|
return VKD3D_SHADER_SV_TESS_FACTOR_QUADEDGE;
|
|
default:
|
|
/* Error is handled during parsing. */
|
|
return VKD3D_SHADER_SV_TESS_FACTOR_TRIEDGE;
|
|
}
|
|
}
|
|
else if (kind == SEMANTIC_KIND_INSIDETESSFACTOR)
|
|
{
|
|
switch (domain)
|
|
{
|
|
case VKD3D_TESSELLATOR_DOMAIN_LINE:
|
|
return VKD3D_SHADER_SV_TESS_FACTOR_LINEDEN;
|
|
case VKD3D_TESSELLATOR_DOMAIN_TRIANGLE:
|
|
return VKD3D_SHADER_SV_TESS_FACTOR_TRIINT;
|
|
case VKD3D_TESSELLATOR_DOMAIN_QUAD:
|
|
return VKD3D_SHADER_SV_TESS_FACTOR_QUADINT;
|
|
default:
|
|
/* Error is handled during parsing. */
|
|
return VKD3D_SHADER_SV_TESS_FACTOR_TRIEDGE;
|
|
}
|
|
}
|
|
else if (kind < ARRAY_SIZE(sysval_semantic_table))
|
|
{
|
|
return sysval_semantic_table[kind];
|
|
}
|
|
else
|
|
{
|
|
return VKD3D_SHADER_SV_NONE;
|
|
}
|
|
}
|
|
|
|
static const struct sm6_metadata_value *sm6_parser_find_named_metadata(struct sm6_parser *sm6, const char *name)
|
|
{
|
|
const struct sm6_metadata_node *node;
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < sm6->named_metadata_count; ++i)
|
|
{
|
|
if (strcmp(sm6->named_metadata[i].name, name))
|
|
continue;
|
|
|
|
node = sm6->named_metadata[i].value.u.node;
|
|
if (!node->operand_count)
|
|
return NULL;
|
|
if (node->operand_count > 1)
|
|
{
|
|
FIXME("Ignoring %u extra operands for %s.\n", node->operand_count - 1, name);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring %u extra operands for metadata node %s.", node->operand_count - 1, name);
|
|
}
|
|
return node->operands[0];
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static bool sm6_parser_resources_load_register_range(struct sm6_parser *sm6,
|
|
const struct sm6_metadata_node *node, struct vkd3d_shader_register_range *range)
|
|
{
|
|
unsigned int size;
|
|
|
|
if (!sm6_metadata_value_is_value(node->operands[1]))
|
|
{
|
|
WARN("Resource data type is not a value.\n");
|
|
return false;
|
|
}
|
|
if (!sm6_type_is_pointer(node->operands[1]->value_type))
|
|
{
|
|
WARN("Resource type is not a pointer.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_TYPE_MISMATCH,
|
|
"Resource metadata value type is not a pointer.");
|
|
}
|
|
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[3], &range->space))
|
|
{
|
|
WARN("Failed to load register space.\n");
|
|
return false;
|
|
}
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[4], &range->first))
|
|
{
|
|
WARN("Failed to load register first.\n");
|
|
return false;
|
|
}
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[5], &size))
|
|
{
|
|
WARN("Failed to load register range size.\n");
|
|
return false;
|
|
}
|
|
if (!size || (size != UINT_MAX && !vkd3d_bound_range(range->first, size, UINT_MAX)))
|
|
{
|
|
WARN("Invalid register range, first %u, size %u.\n", range->first, size);
|
|
return false;
|
|
}
|
|
range->last = (size == UINT_MAX) ? UINT_MAX : range->first + size - 1;
|
|
|
|
return true;
|
|
}
|
|
|
|
static enum vkd3d_shader_resource_type shader_resource_type_from_dxil_resource_kind(enum dxil_resource_kind kind)
|
|
{
|
|
if (resource_kind_is_texture(kind))
|
|
return kind + 1;
|
|
|
|
switch (kind)
|
|
{
|
|
case RESOURCE_KIND_TYPEDBUFFER:
|
|
case RESOURCE_KIND_RAWBUFFER:
|
|
case RESOURCE_KIND_STRUCTUREDBUFFER:
|
|
return VKD3D_SHADER_RESOURCE_BUFFER;
|
|
default:
|
|
return VKD3D_SHADER_RESOURCE_NONE;
|
|
}
|
|
}
|
|
|
|
static const enum vkd3d_data_type data_type_table[] =
|
|
{
|
|
[COMPONENT_TYPE_INVALID] = VKD3D_DATA_UNUSED,
|
|
[COMPONENT_TYPE_I1] = VKD3D_DATA_UNUSED,
|
|
[COMPONENT_TYPE_I16] = VKD3D_DATA_INT,
|
|
[COMPONENT_TYPE_U16] = VKD3D_DATA_UINT,
|
|
[COMPONENT_TYPE_I32] = VKD3D_DATA_INT,
|
|
[COMPONENT_TYPE_U32] = VKD3D_DATA_UINT,
|
|
[COMPONENT_TYPE_I64] = VKD3D_DATA_UNUSED,
|
|
[COMPONENT_TYPE_U64] = VKD3D_DATA_UNUSED,
|
|
[COMPONENT_TYPE_F16] = VKD3D_DATA_FLOAT,
|
|
[COMPONENT_TYPE_F32] = VKD3D_DATA_FLOAT,
|
|
[COMPONENT_TYPE_F64] = VKD3D_DATA_DOUBLE,
|
|
[COMPONENT_TYPE_SNORMF16] = VKD3D_DATA_SNORM,
|
|
[COMPONENT_TYPE_UNORMF16] = VKD3D_DATA_UNORM,
|
|
[COMPONENT_TYPE_SNORMF32] = VKD3D_DATA_SNORM,
|
|
[COMPONENT_TYPE_UNORMF32] = VKD3D_DATA_UNORM,
|
|
[COMPONENT_TYPE_SNORMF64] = VKD3D_DATA_DOUBLE,
|
|
[COMPONENT_TYPE_UNORMF64] = VKD3D_DATA_DOUBLE,
|
|
[COMPONENT_TYPE_PACKEDS8X32] = VKD3D_DATA_UNUSED,
|
|
[COMPONENT_TYPE_PACKEDU8X32] = VKD3D_DATA_UNUSED,
|
|
};
|
|
|
|
static enum vkd3d_data_type vkd3d_data_type_from_dxil_component_type(enum dxil_component_type type,
|
|
struct sm6_parser *sm6)
|
|
{
|
|
enum vkd3d_data_type data_type;
|
|
|
|
if (type >= ARRAY_SIZE(data_type_table) || (data_type = data_type_table[type]) == VKD3D_DATA_UNUSED)
|
|
{
|
|
FIXME("Unhandled component type %u.\n", type);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Resource descriptor component type %u is unhandled.", type);
|
|
return VKD3D_DATA_FLOAT;
|
|
}
|
|
|
|
return data_type;
|
|
}
|
|
|
|
struct resource_additional_values
|
|
{
|
|
enum vkd3d_data_type data_type;
|
|
unsigned int byte_stride;
|
|
};
|
|
|
|
static bool resources_load_additional_values(struct resource_additional_values *info,
|
|
const struct sm6_metadata_node *node, enum dxil_resource_kind kind, struct sm6_parser *sm6)
|
|
{
|
|
unsigned int i, operand_count, tag, value;
|
|
|
|
info->data_type = VKD3D_DATA_UNUSED;
|
|
info->byte_stride = 0;
|
|
|
|
if (node->operand_count & 1)
|
|
{
|
|
WARN("Operand count is not even.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Operand count for resource descriptor tag/value pairs is not even.");
|
|
}
|
|
operand_count = node->operand_count & ~1u;
|
|
|
|
for (i = 0; i < operand_count; i += 2)
|
|
{
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[i], &tag)
|
|
|| !sm6_metadata_get_uint_value(sm6, node->operands[i + 1], &value))
|
|
{
|
|
WARN("Failed to load tag/value pair at index %u.\n", i);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Resource descriptor tag/value pair at index %u is not an integer pair.", i);
|
|
return false;
|
|
}
|
|
|
|
switch (tag)
|
|
{
|
|
case RESOURCE_TAG_ELEMENT_TYPE:
|
|
if (value && kind != RESOURCE_KIND_TYPEDBUFFER && !resource_kind_is_texture(kind))
|
|
{
|
|
WARN("Invalid type %u for an untyped resource.\n", value);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"An untyped resource has type %u.", value);
|
|
return false;
|
|
}
|
|
info->data_type = vkd3d_data_type_from_dxil_component_type(value, sm6);
|
|
break;
|
|
|
|
case RESOURCE_TAG_ELEMENT_STRIDE:
|
|
if (value && kind != RESOURCE_KIND_STRUCTUREDBUFFER)
|
|
{
|
|
WARN("Invalid stride %u for an unstructured resource.\n", value);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"An unstructured resource has a byte stride.");
|
|
return false;
|
|
}
|
|
info->byte_stride = value;
|
|
break;
|
|
|
|
case RESOURCE_TAG_SAMPLER_FEEDBACK_KIND:
|
|
/* MinMip = 0, MipRegionUsed = 1 */
|
|
FIXME("Unhandled sampler feedback kind %u.\n", value);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Sampler feedback kind %u is unhandled.", value);
|
|
break;
|
|
|
|
case RESOURCE_TAG_ENABLE_ATOMIC_64:
|
|
if (value)
|
|
{
|
|
FIXME("Unsupported 64-bit atomic ops.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"64-bit atomic ops on resources are not supported.");
|
|
}
|
|
break;
|
|
|
|
default:
|
|
FIXME("Unhandled tag %u, value %u.\n", tag, value);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Tag %u for resource descriptor additional value %u is unhandled.", tag, value);
|
|
break;
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static struct vkd3d_shader_resource *sm6_parser_resources_load_common_info(struct sm6_parser *sm6,
|
|
const struct sm6_metadata_value *type_value, bool is_uav, enum dxil_resource_kind kind,
|
|
const struct sm6_metadata_value *m, struct vkd3d_shader_instruction *ins)
|
|
{
|
|
struct resource_additional_values resource_values;
|
|
enum vkd3d_shader_resource_type resource_type;
|
|
unsigned int i;
|
|
|
|
if (!(resource_type = shader_resource_type_from_dxil_resource_kind(kind)))
|
|
{
|
|
FIXME("Unhandled resource kind %u.\n", kind);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Resource kind %u is unhandled.", kind);
|
|
return NULL;
|
|
}
|
|
ins->resource_type = resource_type;
|
|
|
|
if (!m)
|
|
{
|
|
ins->opcode = is_uav ? VKD3DSIH_DCL_UAV_RAW : VKD3DSIH_DCL_RESOURCE_RAW;
|
|
ins->declaration.raw_resource.resource.reg.write_mask = 0;
|
|
return &ins->declaration.raw_resource.resource;
|
|
}
|
|
|
|
if (!sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Resource metadata list is not a node.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Resource descriptor metadata list is not a node.");
|
|
return NULL;
|
|
}
|
|
|
|
if (!resources_load_additional_values(&resource_values, m->u.node, kind, sm6))
|
|
return NULL;
|
|
|
|
if (kind == RESOURCE_KIND_TYPEDBUFFER || resource_kind_is_texture(kind))
|
|
{
|
|
if (resource_values.data_type == VKD3D_DATA_UNUSED)
|
|
{
|
|
WARN("No data type defined.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"A typed resource has no data type.");
|
|
}
|
|
|
|
ins->opcode = is_uav ? VKD3DSIH_DCL_UAV_TYPED : VKD3DSIH_DCL;
|
|
for (i = 0; i < VKD3D_VEC4_SIZE; ++i)
|
|
ins->declaration.semantic.resource_data_type[i] = resource_values.data_type;
|
|
ins->declaration.semantic.resource_type = resource_type;
|
|
ins->declaration.semantic.resource.reg.write_mask = VKD3DSP_WRITEMASK_ALL;
|
|
|
|
return &ins->declaration.semantic.resource;
|
|
}
|
|
else if (kind == RESOURCE_KIND_RAWBUFFER)
|
|
{
|
|
ins->opcode = is_uav ? VKD3DSIH_DCL_UAV_RAW : VKD3DSIH_DCL_RESOURCE_RAW;
|
|
ins->declaration.raw_resource.resource.reg.write_mask = 0;
|
|
|
|
return &ins->declaration.raw_resource.resource;
|
|
}
|
|
else if (kind == RESOURCE_KIND_STRUCTUREDBUFFER)
|
|
{
|
|
ins->opcode = is_uav ? VKD3DSIH_DCL_UAV_STRUCTURED : VKD3DSIH_DCL_RESOURCE_STRUCTURED;
|
|
ins->declaration.structured_resource.byte_stride = resource_values.byte_stride;
|
|
ins->declaration.structured_resource.resource.reg.write_mask = 0;
|
|
|
|
/* TODO: 16-bit resources. */
|
|
if (ins->declaration.structured_resource.byte_stride % 4u)
|
|
{
|
|
WARN("Byte stride %u is not a multiple of 4.\n", ins->declaration.structured_resource.byte_stride);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Structured resource byte stride %u is not a multiple of 4.",
|
|
ins->declaration.structured_resource.byte_stride);
|
|
}
|
|
|
|
return &ins->declaration.structured_resource.resource;
|
|
}
|
|
else
|
|
{
|
|
FIXME("Unhandled resource kind %u.\n", kind);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Resource kind %u is unhandled.", kind);
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void init_resource_declaration(struct vkd3d_shader_resource *resource,
|
|
enum vkd3d_shader_register_type reg_type, enum vkd3d_data_type data_type, unsigned int id,
|
|
const struct vkd3d_shader_register_range *range)
|
|
{
|
|
struct vkd3d_shader_dst_param *param = &resource->reg;
|
|
|
|
param->modifiers = 0;
|
|
param->shift = 0;
|
|
vsir_register_init(¶m->reg, reg_type, data_type, 3);
|
|
param->reg.idx[0].offset = id;
|
|
param->reg.idx[1].offset = range->first;
|
|
param->reg.idx[2].offset = range->last;
|
|
|
|
resource->range = *range;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_resources_load_srv(struct sm6_parser *sm6,
|
|
const struct sm6_metadata_node *node, struct sm6_descriptor_info *d, struct vkd3d_shader_instruction *ins)
|
|
{
|
|
struct vkd3d_shader_resource *resource;
|
|
unsigned int kind;
|
|
|
|
if (node->operand_count < 9)
|
|
{
|
|
WARN("Invalid operand count %u.\n", node->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Invalid operand count %u for an SRV descriptor.", node->operand_count);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
if (node->operand_count > 9)
|
|
{
|
|
WARN("Ignoring %u extra operands.\n", node->operand_count - 9);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring %u extra operands for an SRV descriptor.", node->operand_count - 9);
|
|
}
|
|
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[6], &kind))
|
|
{
|
|
WARN("Failed to load resource type.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"SRV resource type metadata value is not an integer.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_INVALID);
|
|
|
|
if (!(resource = sm6_parser_resources_load_common_info(sm6, node->operands[1], false, kind,
|
|
node->operands[8], ins)))
|
|
{
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
d->resource_type = ins->resource_type;
|
|
d->kind = kind;
|
|
d->reg_type = VKD3DSPR_RESOURCE;
|
|
d->reg_data_type = VKD3D_DATA_UNUSED;
|
|
d->resource_data_type = (ins->opcode == VKD3DSIH_DCL)
|
|
? ins->declaration.semantic.resource_data_type[0] : VKD3D_DATA_UNUSED;
|
|
|
|
init_resource_declaration(resource, VKD3DSPR_RESOURCE, d->reg_data_type, d->id, &d->range);
|
|
|
|
if (resource_kind_is_multisampled(kind))
|
|
{
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[7], &ins->declaration.semantic.sample_count))
|
|
{
|
|
WARN("Failed to load sample count.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"SRV sample count metadata value is not an integer.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
}
|
|
else if (!sm6_metadata_value_is_zero_or_undef(node->operands[7]))
|
|
{
|
|
WARN("Ignoring sample count value.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring an SRV descriptor sample count metadata value which is not constant zero or undefined.");
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_resources_load_uav(struct sm6_parser *sm6,
|
|
const struct sm6_metadata_node *node, struct sm6_descriptor_info *d, struct vkd3d_shader_instruction *ins)
|
|
{
|
|
struct vkd3d_shader_resource *resource;
|
|
unsigned int i, values[4];
|
|
|
|
if (node->operand_count < 11)
|
|
{
|
|
WARN("Invalid operand count %u.\n", node->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Invalid operand count %u for a UAV descriptor.", node->operand_count);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
if (node->operand_count > 11)
|
|
{
|
|
WARN("Ignoring %u extra operands.\n", node->operand_count - 11);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring %u extra operands for a UAV descriptor.", node->operand_count - 11);
|
|
}
|
|
|
|
for (i = 6; i < 10; ++i)
|
|
{
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[i], &values[i - 6]))
|
|
{
|
|
WARN("Failed to load uint value at index %u.\n", i);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"A UAV descriptor operand metadata value is not an integer.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
}
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_INVALID);
|
|
if (values[1])
|
|
ins->flags = VKD3DSUF_GLOBALLY_COHERENT;
|
|
if (values[2])
|
|
ins->flags |= VKD3DSUF_ORDER_PRESERVING_COUNTER;
|
|
if (values[3])
|
|
ins->flags |= VKD3DSUF_RASTERISER_ORDERED_VIEW;
|
|
|
|
if (!(resource = sm6_parser_resources_load_common_info(sm6, node->operands[1], true, values[0],
|
|
node->operands[10], ins)))
|
|
{
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
d->resource_type = ins->resource_type;
|
|
d->kind = values[0];
|
|
d->reg_type = VKD3DSPR_UAV;
|
|
d->reg_data_type = VKD3D_DATA_UNUSED;
|
|
d->resource_data_type = (ins->opcode == VKD3DSIH_DCL_UAV_TYPED)
|
|
? ins->declaration.semantic.resource_data_type[0] : VKD3D_DATA_UNUSED;
|
|
|
|
init_resource_declaration(resource, VKD3DSPR_UAV, d->reg_data_type, d->id, &d->range);
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_resources_load_cbv(struct sm6_parser *sm6,
|
|
const struct sm6_metadata_node *node, struct sm6_descriptor_info *d, struct vkd3d_shader_instruction *ins)
|
|
{
|
|
struct vkd3d_shader_register *reg;
|
|
unsigned int buffer_size;
|
|
|
|
if (node->operand_count < 7)
|
|
{
|
|
WARN("Invalid operand count %u.\n", node->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Invalid operand count %u for a CBV descriptor.", node->operand_count);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
if (node->operand_count > 7 && node->operands[7])
|
|
{
|
|
WARN("Ignoring %u extra operands.\n", node->operand_count - 7);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring %u extra operands for a CBV descriptor.", node->operand_count - 7);
|
|
}
|
|
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[6], &buffer_size))
|
|
{
|
|
WARN("Failed to load buffer size.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Constant buffer size metadata value is not an integer.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_DCL_CONSTANT_BUFFER);
|
|
ins->resource_type = VKD3D_SHADER_RESOURCE_BUFFER;
|
|
ins->declaration.cb.size = buffer_size;
|
|
ins->declaration.cb.src.swizzle = VKD3D_SHADER_NO_SWIZZLE;
|
|
ins->declaration.cb.src.modifiers = VKD3DSPSM_NONE;
|
|
|
|
reg = &ins->declaration.cb.src.reg;
|
|
vsir_register_init(reg, VKD3DSPR_CONSTBUFFER, VKD3D_DATA_FLOAT, 3);
|
|
reg->idx[0].offset = d->id;
|
|
reg->idx[1].offset = d->range.first;
|
|
reg->idx[2].offset = d->range.last;
|
|
|
|
ins->declaration.cb.range = d->range;
|
|
|
|
d->reg_type = VKD3DSPR_CONSTBUFFER;
|
|
d->reg_data_type = VKD3D_DATA_FLOAT;
|
|
d->resource_data_type = VKD3D_DATA_FLOAT;
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_resources_load_sampler(struct sm6_parser *sm6,
|
|
const struct sm6_metadata_node *node, struct sm6_descriptor_info *d, struct vkd3d_shader_instruction *ins)
|
|
{
|
|
struct vkd3d_shader_register *reg;
|
|
unsigned int kind;
|
|
|
|
if (node->operand_count < 7)
|
|
{
|
|
WARN("Invalid operand count %u.\n", node->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Invalid operand count %u for a sampler descriptor.", node->operand_count);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
if (node->operand_count > 7 && node->operands[7])
|
|
{
|
|
WARN("Ignoring %u extra operands.\n", node->operand_count - 7);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring %u extra operands for a sampler descriptor.", node->operand_count - 7);
|
|
}
|
|
|
|
vsir_instruction_init(ins, &sm6->p.location, VKD3DSIH_DCL_SAMPLER);
|
|
ins->resource_type = VKD3D_SHADER_RESOURCE_NONE;
|
|
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[6], &kind))
|
|
{
|
|
WARN("Failed to load sampler mode.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Sampler mode metadata value is not an integer.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
switch (kind)
|
|
{
|
|
case SAMPLER_KIND_DEFAULT:
|
|
break;
|
|
case SAMPLER_KIND_COMPARISON:
|
|
ins->flags = VKD3DSI_SAMPLER_COMPARISON_MODE;
|
|
break;
|
|
default:
|
|
FIXME("Ignoring sampler kind %u.\n", kind);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring sampler kind %u.", kind);
|
|
break;
|
|
}
|
|
|
|
ins->declaration.sampler.src.swizzle = VKD3D_SHADER_NO_SWIZZLE;
|
|
ins->declaration.sampler.src.modifiers = VKD3DSPSM_NONE;
|
|
|
|
reg = &ins->declaration.sampler.src.reg;
|
|
vsir_register_init(reg, VKD3DSPR_SAMPLER, VKD3D_DATA_UNUSED, 3);
|
|
reg->idx[0].offset = d->id;
|
|
reg->idx[1].offset = d->range.first;
|
|
reg->idx[2].offset = d->range.last;
|
|
|
|
ins->declaration.sampler.range = d->range;
|
|
|
|
d->resource_type = ins->resource_type;
|
|
d->kind = RESOURCE_KIND_SAMPLER;
|
|
d->reg_type = VKD3DSPR_SAMPLER;
|
|
d->reg_data_type = VKD3D_DATA_UNUSED;
|
|
d->resource_data_type = VKD3D_DATA_UNUSED;
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_descriptor_type_init(struct sm6_parser *sm6,
|
|
enum vkd3d_shader_descriptor_type type, const struct sm6_metadata_node *descriptor_node)
|
|
{
|
|
struct vkd3d_shader_instruction *ins;
|
|
const struct sm6_metadata_node *node;
|
|
const struct sm6_metadata_value *m;
|
|
struct sm6_descriptor_info *d;
|
|
enum vkd3d_result ret;
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < descriptor_node->operand_count; ++i)
|
|
{
|
|
m = descriptor_node->operands[i];
|
|
if (!sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Resource descriptor is not a node.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Resource descriptor is not a metadata node.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
node = m->u.node;
|
|
if (node->operand_count < 6)
|
|
{
|
|
WARN("Invalid operand count %u.\n", node->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Invalid operand count %u for a descriptor.", node->operand_count);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (!vkd3d_array_reserve((void **)&sm6->descriptors, &sm6->descriptor_capacity,
|
|
sm6->descriptor_count + 1, sizeof(*sm6->descriptors)))
|
|
{
|
|
ERR("Failed to allocate descriptor array.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating the descriptor array.");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
d = &sm6->descriptors[sm6->descriptor_count];
|
|
d->type = type;
|
|
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[0], &d->id))
|
|
{
|
|
WARN("Failed to load resource id.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Resource id metadata value is not an integer.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (!sm6_parser_resources_load_register_range(sm6, node, &d->range))
|
|
{
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Resource register range is invalid.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (!(ins = sm6_parser_require_space(sm6, 1)))
|
|
{
|
|
ERR("Failed to allocate instruction.\n");
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
switch (type)
|
|
{
|
|
case VKD3D_SHADER_DESCRIPTOR_TYPE_CBV:
|
|
if ((ret = sm6_parser_resources_load_cbv(sm6, node, d, ins)) < 0)
|
|
return ret;
|
|
break;
|
|
case VKD3D_SHADER_DESCRIPTOR_TYPE_SRV:
|
|
if ((ret = sm6_parser_resources_load_srv(sm6, node, d, ins)) < 0)
|
|
return ret;
|
|
break;
|
|
case VKD3D_SHADER_DESCRIPTOR_TYPE_UAV:
|
|
if ((ret = sm6_parser_resources_load_uav(sm6, node, d, ins)) < 0)
|
|
return ret;
|
|
break;
|
|
case VKD3D_SHADER_DESCRIPTOR_TYPE_SAMPLER:
|
|
if ((ret = sm6_parser_resources_load_sampler(sm6, node, d, ins)) < 0)
|
|
return ret;
|
|
break;
|
|
default:
|
|
FIXME("Unsupported descriptor type %u.\n", type);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Resource descriptor type %u is unsupported.", type);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
++sm6->descriptor_count;
|
|
++sm6->p.program->instructions.count;
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_resources_init(struct sm6_parser *sm6)
|
|
{
|
|
const struct sm6_metadata_value *m = sm6_parser_find_named_metadata(sm6, "dx.resources");
|
|
enum vkd3d_shader_descriptor_type type;
|
|
const struct sm6_metadata_node *node;
|
|
enum vkd3d_result ret;
|
|
|
|
if (!m)
|
|
return VKD3D_OK;
|
|
|
|
node = m->u.node;
|
|
if (node->operand_count != SHADER_DESCRIPTOR_TYPE_COUNT)
|
|
{
|
|
WARN("Unexpected descriptor type count %u.\n", node->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Descriptor type count %u is invalid.", node->operand_count);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
for (type = 0; type < SHADER_DESCRIPTOR_TYPE_COUNT; ++type)
|
|
{
|
|
if (!(m = node->operands[type]))
|
|
continue;
|
|
|
|
if (!sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Resource list is not a node.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_RESOURCES,
|
|
"Resource list is not a metadata node.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if ((ret = sm6_parser_descriptor_type_init(sm6, type, m->u.node)) < 0)
|
|
return ret;
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static void signature_element_read_additional_element_values(struct signature_element *e,
|
|
const struct sm6_metadata_node *node, struct sm6_parser *sm6)
|
|
{
|
|
unsigned int i, operand_count, value, tag;
|
|
|
|
if (node->operand_count < 11 || !node->operands[10])
|
|
return;
|
|
|
|
if (!sm6_metadata_value_is_node(node->operands[10]))
|
|
{
|
|
WARN("Additional values list is not a node.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"Signature element additional values list is not a metadata node.");
|
|
return;
|
|
}
|
|
|
|
node = node->operands[10]->u.node;
|
|
if (node->operand_count & 1)
|
|
{
|
|
WARN("Operand count is not even.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Operand count for signature element additional tag/value pairs is not even.");
|
|
}
|
|
operand_count = node->operand_count & ~1u;
|
|
|
|
for (i = 0; i < operand_count; i += 2)
|
|
{
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[i], &tag)
|
|
|| !sm6_metadata_get_uint_value(sm6, node->operands[i + 1], &value))
|
|
{
|
|
WARN("Failed to extract tag/value pair.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"Signature element tag/value pair at index %u is not an integer pair.", i);
|
|
continue;
|
|
}
|
|
|
|
switch (tag)
|
|
{
|
|
case ADDITIONAL_TAG_STREAM_INDEX:
|
|
e->stream_index = value;
|
|
break;
|
|
case ADDITIONAL_TAG_RELADDR_MASK:
|
|
/* A mask of components accessed via relative addressing. Seems to replace TPF 'dcl_index_range'. */
|
|
if (value > VKD3DSP_WRITEMASK_ALL)
|
|
{
|
|
WARN("Invalid relative addressed mask %#x.\n", value);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_INVALID_MASK,
|
|
"Mask %#x of relative-addressed components is invalid.", value);
|
|
}
|
|
break;
|
|
case ADDITIONAL_TAG_USED_MASK:
|
|
if (value > VKD3DSP_WRITEMASK_ALL)
|
|
{
|
|
WARN("Invalid used mask %#x.\n", value);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_INVALID_MASK,
|
|
"Mask %#x of used components is invalid.", value);
|
|
value &= VKD3DSP_WRITEMASK_ALL;
|
|
}
|
|
e->used_mask = value;
|
|
break;
|
|
default:
|
|
FIXME("Unhandled tag %u, value %u.\n", tag, value);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Tag %#x for signature element additional value %#x is unhandled.", tag, value);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_read_signature(struct sm6_parser *sm6, const struct sm6_metadata_value *m,
|
|
struct shader_signature *s, enum vkd3d_tessellator_domain tessellator_domain)
|
|
{
|
|
unsigned int i, j, column_count, operand_count, index;
|
|
const struct sm6_metadata_node *node, *element_node;
|
|
struct signature_element *elements, *e;
|
|
unsigned int values[10];
|
|
bool is_register;
|
|
|
|
if (!m)
|
|
return VKD3D_OK;
|
|
|
|
if (!sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Signature element list is not a node.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"Signature element list is not a metadata node.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
node = m->u.node;
|
|
operand_count = node->operand_count;
|
|
|
|
if (!(elements = vkd3d_calloc(operand_count, sizeof(*elements))))
|
|
{
|
|
ERR("Failed to allocate %u signature elements.\n", operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating %u signature elements.", operand_count);
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
for (i = 0; i < operand_count; ++i)
|
|
{
|
|
m = node->operands[i];
|
|
|
|
if (!sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Signature element is not a node.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"Signature element is not a metadata node.");
|
|
goto invalid;
|
|
}
|
|
|
|
element_node = m->u.node;
|
|
if (element_node->operand_count < 10)
|
|
{
|
|
WARN("Invalid operand count %u.\n", element_node->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"Invalid signature element operand count %u.", element_node->operand_count);
|
|
goto invalid;
|
|
}
|
|
if (element_node->operand_count > 11)
|
|
{
|
|
WARN("Ignoring %u extra operands.\n", element_node->operand_count - 11);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring %u extra operands for a signature element.", element_node->operand_count - 11);
|
|
}
|
|
|
|
for (j = 0; j < 10; ++j)
|
|
{
|
|
/* 1 is the semantic name, 4 is semantic index metadata. */
|
|
if (j == 1 || j == 4)
|
|
continue;
|
|
if (!sm6_metadata_get_uint_value(sm6, element_node->operands[j], &values[j]))
|
|
{
|
|
WARN("Failed to load uint value at index %u.\n", j);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"Signature element value at index %u is not an integer.", j);
|
|
goto invalid;
|
|
}
|
|
}
|
|
|
|
e = &elements[i];
|
|
|
|
if (values[0] != i)
|
|
{
|
|
FIXME("Unsupported element id %u not equal to its index %u.\n", values[0], i);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"A non-sequential and non-zero-based element id is not supported.");
|
|
goto invalid;
|
|
}
|
|
|
|
if (!sm6_metadata_value_is_string(element_node->operands[1]))
|
|
{
|
|
WARN("Element name is not a string.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"Signature element name is not a metadata string.");
|
|
goto invalid;
|
|
}
|
|
e->semantic_name = element_node->operands[1]->u.string_value;
|
|
|
|
e->component_type = vkd3d_component_type_from_dxil_component_type(values[2]);
|
|
e->min_precision = minimum_precision_from_dxil_component_type(values[2]);
|
|
|
|
j = values[3];
|
|
e->sysval_semantic = sysval_semantic_from_dxil_semantic_kind(j, tessellator_domain);
|
|
if (j != SEMANTIC_KIND_ARBITRARY && j != SEMANTIC_KIND_TARGET && e->sysval_semantic == VKD3D_SHADER_SV_NONE)
|
|
{
|
|
WARN("Unhandled semantic kind %u.\n", j);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"DXIL semantic kind %u is unhandled.", j);
|
|
goto invalid;
|
|
}
|
|
|
|
if ((e->interpolation_mode = values[5]) >= VKD3DSIM_COUNT)
|
|
{
|
|
WARN("Unhandled interpolation mode %u.\n", e->interpolation_mode);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"Interpolation mode %u is unhandled.", e->interpolation_mode);
|
|
goto invalid;
|
|
}
|
|
|
|
e->register_count = values[6];
|
|
column_count = values[7];
|
|
e->register_index = values[8];
|
|
e->target_location = e->register_index;
|
|
|
|
if ((is_register = e->register_index == UINT_MAX))
|
|
{
|
|
if (register_type_from_dxil_semantic_kind(e->sysval_semantic) == VKD3DSPR_INVALID)
|
|
{
|
|
WARN("Unhandled I/O register semantic kind %u.\n", j);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"DXIL semantic kind %u is unhandled for an I/O register.", j);
|
|
goto invalid;
|
|
}
|
|
}
|
|
else if (e->register_index > MAX_REG_OUTPUT || e->register_count > MAX_REG_OUTPUT - e->register_index)
|
|
{
|
|
WARN("Invalid row start %u with row count %u.\n", e->register_index, e->register_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"A signature element starting row of %u with count %u is invalid.",
|
|
e->register_index, e->register_count);
|
|
goto invalid;
|
|
}
|
|
|
|
index = values[9];
|
|
if (index != UINT8_MAX && (index >= VKD3D_VEC4_SIZE || column_count > VKD3D_VEC4_SIZE - index))
|
|
{
|
|
WARN("Invalid column start %u with count %u.\n", index, column_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"A signature element starting column %u with count %u is invalid.", index, column_count);
|
|
goto invalid;
|
|
}
|
|
|
|
e->mask = vkd3d_write_mask_from_component_count(column_count);
|
|
e->used_mask = e->mask;
|
|
signature_element_read_additional_element_values(e, element_node, sm6);
|
|
|
|
if (index != UINT8_MAX)
|
|
{
|
|
e->mask <<= index;
|
|
e->used_mask <<= index;
|
|
}
|
|
|
|
/* DXIL reads/writes uint for bool I/O. */
|
|
if (e->component_type == VKD3D_SHADER_COMPONENT_BOOL)
|
|
e->component_type = VKD3D_SHADER_COMPONENT_UINT;
|
|
|
|
m = element_node->operands[4];
|
|
if (!sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Semantic index list is not a node.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"Signature element semantic index list is not a metadata node.");
|
|
goto invalid;
|
|
}
|
|
|
|
element_node = m->u.node;
|
|
for (j = 0; j < element_node->operand_count; ++j)
|
|
{
|
|
if (!sm6_metadata_get_uint_value(sm6, element_node->operands[j], &index))
|
|
{
|
|
WARN("Failed to get semantic index for row %u.\n", j);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"Signature element semantic index for row %u is not an integer.", j);
|
|
}
|
|
else if (!j)
|
|
{
|
|
e->semantic_index = index;
|
|
}
|
|
else if (index != e->semantic_index + j)
|
|
{
|
|
WARN("Semantic index %u for row %u is not of an incrementing sequence.\n", index, j);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"Signature element semantic index %u for row %u is not of an incrementing sequence.", index, j);
|
|
}
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < operand_count; ++i)
|
|
{
|
|
if ((elements[i].semantic_name = vkd3d_strdup(elements[i].semantic_name)))
|
|
continue;
|
|
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Failed to allocate signature element semantic name.");
|
|
for (j = 0; j < i; ++j)
|
|
{
|
|
vkd3d_free((void *)elements[j].semantic_name);
|
|
}
|
|
vkd3d_free(elements);
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
shader_signature_cleanup(s);
|
|
s->elements = elements;
|
|
s->element_count = operand_count;
|
|
|
|
return VKD3D_OK;
|
|
|
|
invalid:
|
|
vkd3d_free(elements);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_signatures_init(struct sm6_parser *sm6, const struct sm6_metadata_value *m,
|
|
enum vkd3d_tessellator_domain tessellator_domain)
|
|
{
|
|
struct vsir_program *program = sm6->p.program;
|
|
enum vkd3d_result ret;
|
|
|
|
if (!sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Signature table is not a node.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_SIGNATURE,
|
|
"Signature table is not a metadata node.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (m->u.node->operand_count && (ret = sm6_parser_read_signature(sm6, m->u.node->operands[0],
|
|
&program->input_signature, tessellator_domain)) < 0)
|
|
{
|
|
return ret;
|
|
}
|
|
if (m->u.node->operand_count > 1 && (ret = sm6_parser_read_signature(sm6, m->u.node->operands[1],
|
|
&program->output_signature, tessellator_domain)) < 0)
|
|
{
|
|
return ret;
|
|
}
|
|
if (m->u.node->operand_count > 1 && (ret = sm6_parser_read_signature(sm6, m->u.node->operands[2],
|
|
&program->patch_constant_signature, tessellator_domain)) < 0)
|
|
{
|
|
return ret;
|
|
}
|
|
|
|
sm6_parser_init_input_signature(sm6, &program->input_signature);
|
|
sm6_parser_init_output_signature(sm6, &program->output_signature);
|
|
sm6_parser_init_patch_constant_signature(sm6, &program->patch_constant_signature);
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static void sm6_parser_emit_global_flags(struct sm6_parser *sm6, const struct sm6_metadata_value *m)
|
|
{
|
|
enum vkd3d_shader_global_flags global_flags, mask, rotated_flags;
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
if (!sm6_metadata_get_uint64_value(sm6, m, (uint64_t*)&global_flags))
|
|
{
|
|
WARN("Failed to load global flags.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Global flags metadata value is not an integer.");
|
|
return;
|
|
}
|
|
/* Rotate SKIP_OPTIMIZATION from bit 0 to bit 4 to match vkd3d_shader_global_flags. */
|
|
mask = (VKD3DSGF_SKIP_OPTIMIZATION << 1) - 1;
|
|
rotated_flags = global_flags & mask;
|
|
rotated_flags = (rotated_flags >> 1) | ((rotated_flags & 1) << 4);
|
|
global_flags = (global_flags & ~mask) | rotated_flags;
|
|
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_DCL_GLOBAL_FLAGS);
|
|
ins->declaration.global_flags = global_flags;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_emit_thread_group(struct sm6_parser *sm6, const struct sm6_metadata_value *m)
|
|
{
|
|
struct vkd3d_shader_version *version = &sm6->p.program->shader_version;
|
|
const struct sm6_metadata_node *node;
|
|
struct vkd3d_shader_instruction *ins;
|
|
unsigned int group_sizes[3];
|
|
unsigned int i;
|
|
|
|
if (version->type != VKD3D_SHADER_TYPE_COMPUTE)
|
|
{
|
|
WARN("Shader of type %#x has thread group dimensions.\n", version->type);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Shader has thread group dimensions but is not a compute shader.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (!m || !sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Thread group dimension value is not a node.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Thread group dimension metadata value is not a node.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
node = m->u.node;
|
|
if (node->operand_count != 3)
|
|
{
|
|
WARN("Invalid operand count %u.\n", node->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Thread group dimension operand count %u is invalid.", node->operand_count);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
for (i = 0; i < 3; ++i)
|
|
{
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[i], &group_sizes[i]))
|
|
{
|
|
WARN("Thread group dimension is not an integer value.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Thread group dimension metadata value is not an integer.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
if (!group_sizes[i] || group_sizes[i] > dx_max_thread_group_size[i])
|
|
{
|
|
char dim = "XYZ"[i];
|
|
WARN("Invalid thread group %c dimension %u.\n", dim, group_sizes[i]);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Thread group %c dimension %u is invalid.", dim, group_sizes[i]);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
}
|
|
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_DCL_THREAD_GROUP);
|
|
ins->declaration.thread_group_size.x = group_sizes[0];
|
|
ins->declaration.thread_group_size.y = group_sizes[1];
|
|
ins->declaration.thread_group_size.z = group_sizes[2];
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static void sm6_parser_emit_dcl_count(struct sm6_parser *sm6, enum vkd3d_shader_opcode handler_idx, unsigned int count)
|
|
{
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
ins = sm6_parser_add_instruction(sm6, handler_idx);
|
|
ins->declaration.count = count;
|
|
}
|
|
|
|
static void sm6_parser_emit_dcl_primitive_topology(struct sm6_parser *sm6,
|
|
enum vkd3d_shader_opcode handler_idx, enum vkd3d_primitive_type primitive_type,
|
|
unsigned int patch_vertex_count)
|
|
{
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
ins = sm6_parser_add_instruction(sm6, handler_idx);
|
|
ins->declaration.primitive_type.type = primitive_type;
|
|
ins->declaration.primitive_type.patch_vertex_count = patch_vertex_count;
|
|
}
|
|
|
|
static void sm6_parser_emit_dcl_tessellator_domain(struct sm6_parser *sm6,
|
|
enum vkd3d_tessellator_domain tessellator_domain)
|
|
{
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
if (tessellator_domain == VKD3D_TESSELLATOR_DOMAIN_INVALID || tessellator_domain >= VKD3D_TESSELLATOR_DOMAIN_COUNT)
|
|
{
|
|
WARN("Unhandled domain %u.\n", tessellator_domain);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Domain shader tessellator domain %u is unhandled.", tessellator_domain);
|
|
}
|
|
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_DCL_TESSELLATOR_DOMAIN);
|
|
ins->declaration.tessellator_domain = tessellator_domain;
|
|
}
|
|
|
|
static void sm6_parser_validate_control_point_count(struct sm6_parser *sm6, unsigned int count,
|
|
const char *type)
|
|
{
|
|
if (!count || count > 32)
|
|
{
|
|
WARN("%s control point count %u invalid.\n", type, count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"%s control point count %u is invalid.", type, count);
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_emit_dcl_tessellator_partitioning(struct sm6_parser *sm6,
|
|
enum vkd3d_shader_tessellator_partitioning tessellator_partitioning)
|
|
{
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
if (!tessellator_partitioning || tessellator_partitioning > VKD3D_SHADER_TESSELLATOR_PARTITIONING_FRACTIONAL_EVEN)
|
|
{
|
|
WARN("Unhandled partitioning %u.\n", tessellator_partitioning);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Hull shader tessellator partitioning %u is unhandled.", tessellator_partitioning);
|
|
}
|
|
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_DCL_TESSELLATOR_PARTITIONING);
|
|
ins->declaration.tessellator_partitioning = tessellator_partitioning;
|
|
}
|
|
|
|
static void sm6_parser_emit_dcl_tessellator_output_primitive(struct sm6_parser *sm6,
|
|
enum vkd3d_shader_tessellator_output_primitive primitive)
|
|
{
|
|
struct vkd3d_shader_instruction *ins;
|
|
|
|
if (!primitive || primitive > VKD3D_SHADER_TESSELLATOR_OUTPUT_TRIANGLE_CCW)
|
|
{
|
|
WARN("Unhandled output primitive %u.\n", primitive);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Hull shader tessellator output primitive %u is unhandled.", primitive);
|
|
}
|
|
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_DCL_TESSELLATOR_OUTPUT_PRIMITIVE);
|
|
ins->declaration.tessellator_output_primitive = primitive;
|
|
}
|
|
|
|
static void sm6_parser_emit_dcl_max_tessellation_factor(struct sm6_parser *sm6, struct sm6_metadata_value *m)
|
|
{
|
|
struct vkd3d_shader_instruction *ins;
|
|
float max_tessellation_factor;
|
|
|
|
if (!sm6_metadata_get_float_value(sm6, m, &max_tessellation_factor))
|
|
{
|
|
WARN("Max tess factor property is not a float value.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Hull shader max tessellation factor property operand is not a float.");
|
|
return;
|
|
}
|
|
|
|
/* Exclude non-finite values. */
|
|
if (!(max_tessellation_factor >= 1.0f && max_tessellation_factor <= 64.0f))
|
|
{
|
|
WARN("Invalid max tess factor %f.\n", max_tessellation_factor);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Hull shader max tessellation factor %f is invalid.", max_tessellation_factor);
|
|
}
|
|
|
|
ins = sm6_parser_add_instruction(sm6, VKD3DSIH_DCL_HS_MAX_TESSFACTOR);
|
|
ins->declaration.max_tessellation_factor = max_tessellation_factor;
|
|
}
|
|
|
|
static void sm6_parser_gs_properties_init(struct sm6_parser *sm6, const struct sm6_metadata_value *m)
|
|
{
|
|
enum vkd3d_primitive_type input_primitive = VKD3D_PT_TRIANGLELIST, output_primitive;
|
|
unsigned int i, input_control_point_count = 1, patch_vertex_count = 0;
|
|
const struct sm6_metadata_node *node;
|
|
unsigned int operands[5] = {0};
|
|
|
|
if (!m || !sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Missing or invalid GS properties.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Geometry shader properties node is missing or invalid.");
|
|
return;
|
|
}
|
|
|
|
node = m->u.node;
|
|
if (node->operand_count < ARRAY_SIZE(operands))
|
|
{
|
|
WARN("Invalid operand count %u.\n", node->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Geometry shader properties operand count %u is invalid.", node->operand_count);
|
|
return;
|
|
}
|
|
if (node->operand_count > ARRAY_SIZE(operands))
|
|
{
|
|
WARN("Ignoring %zu extra operands.\n", node->operand_count - ARRAY_SIZE(operands));
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring %zu extra operands for geometry shader properties.",
|
|
node->operand_count - ARRAY_SIZE(operands));
|
|
}
|
|
|
|
for (i = 0; i < node->operand_count; ++i)
|
|
{
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[i], &operands[i]))
|
|
{
|
|
WARN("GS property at index %u is not a uint value.\n", i);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Geometry shader properties operand at index %u is not an integer.", i);
|
|
}
|
|
}
|
|
|
|
switch (i = operands[0])
|
|
{
|
|
case INPUT_PRIMITIVE_POINT:
|
|
input_primitive = VKD3D_PT_POINTLIST;
|
|
input_control_point_count = 1;
|
|
break;
|
|
|
|
case INPUT_PRIMITIVE_LINE:
|
|
input_primitive = VKD3D_PT_LINELIST;
|
|
input_control_point_count = 2;
|
|
break;
|
|
|
|
case INPUT_PRIMITIVE_TRIANGLE:
|
|
input_primitive = VKD3D_PT_TRIANGLELIST;
|
|
input_control_point_count = 3;
|
|
break;
|
|
|
|
case INPUT_PRIMITIVE_LINEWITHADJACENCY:
|
|
input_primitive = VKD3D_PT_LINELIST_ADJ;
|
|
input_control_point_count = 4;
|
|
break;
|
|
|
|
case INPUT_PRIMITIVE_TRIANGLEWITHADJACENY:
|
|
input_primitive = VKD3D_PT_TRIANGLELIST_ADJ;
|
|
input_control_point_count = 6;
|
|
break;
|
|
|
|
default:
|
|
if (i >= INPUT_PRIMITIVE_PATCH1 && i <= INPUT_PRIMITIVE_PATCH32)
|
|
{
|
|
input_primitive = VKD3D_PT_PATCH;
|
|
patch_vertex_count = i - INPUT_PRIMITIVE_PATCH1 + 1;
|
|
break;
|
|
}
|
|
|
|
WARN("Unhandled input primitive %u.\n", i);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Geometry shader input primitive %u is unhandled.", i);
|
|
break;
|
|
}
|
|
|
|
sm6_parser_emit_dcl_primitive_topology(sm6, VKD3DSIH_DCL_INPUT_PRIMITIVE, input_primitive, patch_vertex_count);
|
|
sm6->p.program->input_control_point_count = input_control_point_count;
|
|
|
|
i = operands[1];
|
|
/* Max total scalar count sets an upper limit. We would need to scan outputs to be more precise. */
|
|
if (i > MAX_GS_OUTPUT_TOTAL_SCALARS)
|
|
{
|
|
WARN("GS output vertex count %u invalid.\n", i);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Geometry shader output vertex count %u is invalid.", i);
|
|
}
|
|
sm6_parser_emit_dcl_count(sm6, VKD3DSIH_DCL_VERTICES_OUT, i);
|
|
|
|
if (operands[2] > 1)
|
|
{
|
|
FIXME("Unhandled stream mask %#x.\n", operands[2]);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Geometry shader stream mask %#x is unhandled.", operands[2]);
|
|
}
|
|
|
|
output_primitive = operands[3];
|
|
if (output_primitive == VKD3D_PT_UNDEFINED || output_primitive >= VKD3D_PT_COUNT)
|
|
{
|
|
WARN("Unhandled output primitive %u.\n", output_primitive);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Geometry shader output primitive %u is unhandled.", output_primitive);
|
|
output_primitive = VKD3D_PT_TRIANGLELIST;
|
|
}
|
|
sm6_parser_emit_dcl_primitive_topology(sm6, VKD3DSIH_DCL_OUTPUT_TOPOLOGY, output_primitive, 0);
|
|
|
|
i = operands[4];
|
|
if (!i || i > MAX_GS_INSTANCE_COUNT)
|
|
{
|
|
WARN("GS instance count %u invalid.\n", i);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Geometry shader instance count %u is invalid.", i);
|
|
}
|
|
sm6_parser_emit_dcl_count(sm6, VKD3DSIH_DCL_GS_INSTANCES, i);
|
|
}
|
|
|
|
static enum vkd3d_tessellator_domain sm6_parser_ds_properties_init(struct sm6_parser *sm6,
|
|
const struct sm6_metadata_value *m)
|
|
{
|
|
const struct sm6_metadata_node *node;
|
|
unsigned int operands[2] = {0};
|
|
unsigned int i;
|
|
|
|
if (!m || !sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Missing or invalid DS properties.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Domain shader properties node is missing or invalid.");
|
|
return 0;
|
|
}
|
|
|
|
node = m->u.node;
|
|
if (node->operand_count < ARRAY_SIZE(operands))
|
|
{
|
|
WARN("Invalid operand count %u.\n", node->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Domain shader properties operand count %u is invalid.", node->operand_count);
|
|
return 0;
|
|
}
|
|
if (node->operand_count > ARRAY_SIZE(operands))
|
|
{
|
|
WARN("Ignoring %zu extra operands.\n", node->operand_count - ARRAY_SIZE(operands));
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring %zu extra operands for domain shader properties.",
|
|
node->operand_count - ARRAY_SIZE(operands));
|
|
}
|
|
|
|
for (i = 0; i < node->operand_count; ++i)
|
|
{
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[i], &operands[i]))
|
|
{
|
|
WARN("DS property at index %u is not a uint value.\n", i);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Domain shader properties operand at index %u is not an integer.", i);
|
|
}
|
|
}
|
|
|
|
sm6_parser_emit_dcl_tessellator_domain(sm6, operands[0]);
|
|
sm6_parser_validate_control_point_count(sm6, operands[1], "Domain shader input");
|
|
sm6->p.program->input_control_point_count = operands[1];
|
|
|
|
return operands[0];
|
|
}
|
|
|
|
static enum vkd3d_tessellator_domain sm6_parser_hs_properties_init(struct sm6_parser *sm6,
|
|
const struct sm6_metadata_value *m)
|
|
{
|
|
struct vsir_program *program = sm6->p.program;
|
|
const struct sm6_metadata_node *node;
|
|
unsigned int operands[6] = {0};
|
|
unsigned int i;
|
|
|
|
if (!m || !sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Missing or invalid HS properties.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Hull shader properties node is missing or invalid.");
|
|
return 0;
|
|
}
|
|
|
|
node = m->u.node;
|
|
if (node->operand_count < 7)
|
|
{
|
|
WARN("Invalid operand count %u.\n", node->operand_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_OPERAND_COUNT,
|
|
"Hull shader properties operand count %u is invalid.", node->operand_count);
|
|
return 0;
|
|
}
|
|
if (node->operand_count > 7)
|
|
{
|
|
WARN("Ignoring %u extra operands.\n", node->operand_count - 7);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Ignoring %u extra operands for hull shader properties.", node->operand_count - 7);
|
|
}
|
|
|
|
m = node->operands[0];
|
|
if (!sm6_metadata_value_is_value(m) || !sm6_value_is_function_dcl(m->u.value))
|
|
{
|
|
WARN("Patch constant function node is not a function value.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Hull shader patch constant function node is not a function value.");
|
|
}
|
|
else
|
|
{
|
|
sm6->patch_constant_function = m->u.value->u.function.name;
|
|
}
|
|
|
|
for (i = 1; i < min(node->operand_count, ARRAY_SIZE(operands)); ++i)
|
|
{
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[i], &operands[i]))
|
|
{
|
|
WARN("HS property at index %u is not a uint value.\n", i);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Hull shader properties operand at index %u is not an integer.", i);
|
|
}
|
|
}
|
|
|
|
sm6_parser_validate_control_point_count(sm6, operands[1], "Hull shader input");
|
|
program->input_control_point_count = operands[1];
|
|
sm6_parser_validate_control_point_count(sm6, operands[2], "Hull shader output");
|
|
sm6_parser_emit_dcl_count(sm6, VKD3DSIH_DCL_OUTPUT_CONTROL_POINT_COUNT, operands[2]);
|
|
program->output_control_point_count = operands[2];
|
|
sm6_parser_emit_dcl_tessellator_domain(sm6, operands[3]);
|
|
sm6_parser_emit_dcl_tessellator_partitioning(sm6, operands[4]);
|
|
sm6_parser_emit_dcl_tessellator_output_primitive(sm6, operands[5]);
|
|
sm6_parser_emit_dcl_max_tessellation_factor(sm6, node->operands[6]);
|
|
|
|
return operands[3];
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_entry_point_init(struct sm6_parser *sm6)
|
|
{
|
|
const struct sm6_metadata_value *m = sm6_parser_find_named_metadata(sm6, "dx.entryPoints");
|
|
const struct sm6_metadata_node *node, *entry_node = m ? m->u.node : NULL;
|
|
enum vkd3d_tessellator_domain tessellator_domain = 0;
|
|
unsigned int i, operand_count, tag;
|
|
const struct sm6_value *value;
|
|
enum vkd3d_result ret;
|
|
|
|
if (!entry_node || entry_node->operand_count < 2 || !(m = entry_node->operands[0]))
|
|
{
|
|
WARN("No entry point definition found.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_ENTRY_POINT,
|
|
"No entry point definition found in the metadata.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (m->type != VKD3D_METADATA_VALUE)
|
|
{
|
|
WARN("Entry point definition is not a value.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_ENTRY_POINT,
|
|
"Entry point definition is not a metadata value.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
value = m->u.value;
|
|
if (!sm6_value_is_function_dcl(value))
|
|
{
|
|
WARN("Entry point value is not a function definition.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_ENTRY_POINT,
|
|
"Entry point metadata value does not contain a function definition.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
sm6->entry_point = value->u.function.name;
|
|
if (!sm6_metadata_value_is_string(entry_node->operands[1])
|
|
|| strcmp(sm6->entry_point, entry_node->operands[1]->u.string_value))
|
|
{
|
|
WARN("Entry point function name %s mismatch.\n", sm6->entry_point);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_ENTRY_POINT_MISMATCH,
|
|
"Entry point function name %s does not match the name in metadata.", sm6->entry_point);
|
|
}
|
|
|
|
if (entry_node->operand_count >= 5 && (m = entry_node->operands[4]))
|
|
{
|
|
if (!sm6_metadata_value_is_node(m))
|
|
{
|
|
WARN("Shader properties list is not a node.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Shader properties tag/value list is not a metadata node.");
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
node = m->u.node;
|
|
if (node->operand_count & 1)
|
|
{
|
|
WARN("Operand count is not even.\n");
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_IGNORING_OPERANDS,
|
|
"Operand count for shader properties tag/value pairs is not even.");
|
|
}
|
|
operand_count = node->operand_count & ~1u;
|
|
|
|
for (i = 0; i < operand_count; i += 2)
|
|
{
|
|
if (!sm6_metadata_get_uint_value(sm6, node->operands[i], &tag))
|
|
{
|
|
WARN("Tag is not an integer value.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Shader properties tag at index %u is not an integer.", i);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
switch (tag)
|
|
{
|
|
case SHADER_PROPERTIES_FLAGS:
|
|
sm6_parser_emit_global_flags(sm6, node->operands[i + 1]);
|
|
break;
|
|
case SHADER_PROPERTIES_GEOMETRY:
|
|
sm6_parser_gs_properties_init(sm6, node->operands[i + 1]);
|
|
break;
|
|
case SHADER_PROPERTIES_DOMAIN:
|
|
tessellator_domain = sm6_parser_ds_properties_init(sm6, node->operands[i + 1]);
|
|
break;
|
|
case SHADER_PROPERTIES_HULL:
|
|
tessellator_domain = sm6_parser_hs_properties_init(sm6, node->operands[i + 1]);
|
|
break;
|
|
case SHADER_PROPERTIES_COMPUTE:
|
|
if ((ret = sm6_parser_emit_thread_group(sm6, node->operands[i + 1])) < 0)
|
|
return ret;
|
|
break;
|
|
default:
|
|
FIXME("Unhandled tag %#x.\n", tag);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_PROPERTIES,
|
|
"Shader properties tag %#x is unhandled.", tag);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (entry_node->operand_count >= 3 && (m = entry_node->operands[2])
|
|
&& (ret = sm6_parser_signatures_init(sm6, m, tessellator_domain)) < 0)
|
|
{
|
|
return ret;
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static void sm6_metadata_value_destroy(struct sm6_metadata_value *m)
|
|
{
|
|
switch (m->type)
|
|
{
|
|
case VKD3D_METADATA_NODE:
|
|
vkd3d_free(m->u.node);
|
|
break;
|
|
case VKD3D_METADATA_KIND:
|
|
vkd3d_free(m->u.kind.name);
|
|
break;
|
|
case VKD3D_METADATA_STRING:
|
|
vkd3d_free(m->u.string_value);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void sm6_parser_metadata_cleanup(struct sm6_parser *sm6)
|
|
{
|
|
unsigned int i, j;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(sm6->metadata_tables); ++i)
|
|
{
|
|
for (j = 0; j < sm6->metadata_tables[i].count; ++j)
|
|
sm6_metadata_value_destroy(&sm6->metadata_tables[i].values[j]);
|
|
vkd3d_free(sm6->metadata_tables[i].values);
|
|
}
|
|
for (i = 0; i < sm6->named_metadata_count; ++i)
|
|
{
|
|
sm6_metadata_value_destroy(&sm6->named_metadata[i].value);
|
|
vkd3d_free(sm6->named_metadata[i].name);
|
|
}
|
|
vkd3d_free(sm6->named_metadata);
|
|
}
|
|
|
|
static void sm6_type_table_cleanup(struct sm6_type *types, size_t count)
|
|
{
|
|
size_t i;
|
|
|
|
if (!types)
|
|
return;
|
|
|
|
for (i = 0; i < count; ++i)
|
|
{
|
|
switch (types[i].class)
|
|
{
|
|
case TYPE_CLASS_STRUCT:
|
|
vkd3d_free((void *)types[i].u.struc->name);
|
|
vkd3d_free(types[i].u.struc);
|
|
break;
|
|
case TYPE_CLASS_FUNCTION:
|
|
vkd3d_free(types[i].u.function);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
vkd3d_free(types);
|
|
}
|
|
|
|
static void sm6_symtab_cleanup(struct sm6_symbol *symbols, size_t count)
|
|
{
|
|
size_t i;
|
|
|
|
for (i = 0; i < count; ++i)
|
|
vkd3d_free((void *)symbols[i].name);
|
|
vkd3d_free(symbols);
|
|
}
|
|
|
|
static void sm6_phi_destroy(struct sm6_phi *phi)
|
|
{
|
|
vkd3d_free(phi->incoming);
|
|
}
|
|
|
|
static void sm6_block_destroy(struct sm6_block *block)
|
|
{
|
|
unsigned int i;
|
|
|
|
vkd3d_free(block->instructions);
|
|
for (i = 0; i < block->phi_count; ++i)
|
|
sm6_phi_destroy(&block->phi[i]);
|
|
vkd3d_free(block->phi);
|
|
vkd3d_free(block->terminator.cases);
|
|
vkd3d_free(block);
|
|
}
|
|
|
|
static void sm6_functions_cleanup(struct sm6_function *functions, size_t count)
|
|
{
|
|
size_t i, j;
|
|
|
|
for (i = 0; i < count; ++i)
|
|
{
|
|
for (j = 0; j < functions[i].block_count; ++j)
|
|
sm6_block_destroy(functions[i].blocks[j]);
|
|
vkd3d_free(functions[i].blocks);
|
|
}
|
|
vkd3d_free(functions);
|
|
}
|
|
|
|
static void sm6_parser_cleanup(struct sm6_parser *sm6)
|
|
{
|
|
dxil_block_destroy(&sm6->root_block);
|
|
dxil_global_abbrevs_cleanup(sm6->abbrevs, sm6->abbrev_count);
|
|
sm6_type_table_cleanup(sm6->types, sm6->type_count);
|
|
sm6_symtab_cleanup(sm6->global_symbols, sm6->global_symbol_count);
|
|
sm6_functions_cleanup(sm6->functions, sm6->function_count);
|
|
sm6_parser_metadata_cleanup(sm6);
|
|
vkd3d_free(sm6->descriptors);
|
|
vkd3d_free(sm6->values);
|
|
}
|
|
|
|
static struct sm6_function *sm6_parser_get_function(const struct sm6_parser *sm6, const char *name)
|
|
{
|
|
size_t i;
|
|
for (i = 0; i < sm6->function_count; ++i)
|
|
if (!strcmp(sm6->functions[i].declaration->u.function.name, name))
|
|
return &sm6->functions[i];
|
|
return NULL;
|
|
}
|
|
|
|
static enum vkd3d_result sm6_parser_init(struct sm6_parser *sm6, struct vsir_program *program,
|
|
const struct vkd3d_shader_compile_info *compile_info,
|
|
struct vkd3d_shader_message_context *message_context, struct dxbc_shader_desc *dxbc_desc)
|
|
{
|
|
size_t count, length, function_count, expected_function_count, byte_code_size = dxbc_desc->byte_code_size;
|
|
const struct vkd3d_shader_location location = {.source_name = compile_info->source_name};
|
|
struct shader_signature *patch_constant_signature, *output_signature, *input_signature;
|
|
uint32_t version_token, dxil_version, token_count, magic;
|
|
const uint32_t *byte_code = dxbc_desc->byte_code;
|
|
unsigned int chunk_offset, chunk_size;
|
|
enum bitcode_block_abbreviation abbr;
|
|
struct vkd3d_shader_version version;
|
|
struct dxil_block *block;
|
|
struct sm6_function *fn;
|
|
enum vkd3d_result ret;
|
|
unsigned int i, j;
|
|
|
|
count = byte_code_size / sizeof(*byte_code);
|
|
if (count < 6)
|
|
{
|
|
WARN("Invalid data size %zu.\n", byte_code_size);
|
|
vkd3d_shader_error(message_context, &location, VKD3D_SHADER_ERROR_DXIL_INVALID_SIZE,
|
|
"DXIL chunk size %zu is smaller than the DXIL header size.", byte_code_size);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
version_token = byte_code[0];
|
|
TRACE("Compiler version: 0x%08x.\n", version_token);
|
|
token_count = byte_code[1];
|
|
TRACE("Token count: %u.\n", token_count);
|
|
|
|
if (token_count < 6 || count < token_count)
|
|
{
|
|
WARN("Invalid token count %u (word count %zu).\n", token_count, count);
|
|
vkd3d_shader_error(message_context, &location, VKD3D_SHADER_ERROR_DXIL_INVALID_CHUNK_SIZE,
|
|
"DXIL chunk token count %#x is invalid (word count %zu).", token_count, count);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if (byte_code[2] != TAG_DXIL)
|
|
WARN("Unknown magic number 0x%08x.\n", byte_code[2]);
|
|
|
|
dxil_version = byte_code[3];
|
|
if (dxil_version > 0x102)
|
|
WARN("Unknown DXIL version: 0x%08x.\n", dxil_version);
|
|
else
|
|
TRACE("DXIL version: 0x%08x.\n", dxil_version);
|
|
|
|
chunk_offset = byte_code[4];
|
|
if (chunk_offset < 16 || chunk_offset >= byte_code_size)
|
|
{
|
|
WARN("Invalid bitcode chunk offset %#x (data size %zu).\n", chunk_offset, byte_code_size);
|
|
vkd3d_shader_error(message_context, &location, VKD3D_SHADER_ERROR_DXIL_INVALID_CHUNK_OFFSET,
|
|
"DXIL bitcode chunk has invalid offset %#x (data size %#zx).", chunk_offset, byte_code_size);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
chunk_size = byte_code[5];
|
|
if (chunk_size > byte_code_size - chunk_offset)
|
|
{
|
|
WARN("Invalid bitcode chunk size %#x (data size %zu, chunk offset %#x).\n",
|
|
chunk_size, byte_code_size, chunk_offset);
|
|
vkd3d_shader_error(message_context, &location, VKD3D_SHADER_ERROR_DXIL_INVALID_CHUNK_SIZE,
|
|
"DXIL bitcode chunk has invalid size %#x (data size %#zx, chunk offset %#x).",
|
|
chunk_size, byte_code_size, chunk_offset);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
sm6->start = (const uint32_t *)((const char*)&byte_code[2] + chunk_offset);
|
|
if ((magic = sm6->start[0]) != BITCODE_MAGIC)
|
|
{
|
|
WARN("Unknown magic number 0x%08x.\n", magic);
|
|
vkd3d_shader_warning(message_context, &location, VKD3D_SHADER_WARNING_DXIL_UNKNOWN_MAGIC_NUMBER,
|
|
"DXIL bitcode chunk magic number 0x%08x is not the expected 0x%08x.", magic, BITCODE_MAGIC);
|
|
}
|
|
|
|
sm6->end = &sm6->start[(chunk_size + sizeof(*sm6->start) - 1) / sizeof(*sm6->start)];
|
|
|
|
if ((version.type = version_token >> 16) >= VKD3D_SHADER_TYPE_COUNT)
|
|
{
|
|
FIXME("Unknown shader type %#x.\n", version.type);
|
|
vkd3d_shader_warning(message_context, &location, VKD3D_SHADER_WARNING_DXIL_UNKNOWN_SHADER_TYPE,
|
|
"Unknown shader type %#x.", version.type);
|
|
}
|
|
|
|
version.major = VKD3D_SM6_VERSION_MAJOR(version_token);
|
|
version.minor = VKD3D_SM6_VERSION_MINOR(version_token);
|
|
|
|
if ((abbr = sm6->start[1] & 3) != ENTER_SUBBLOCK)
|
|
{
|
|
WARN("Initial block abbreviation %u is not ENTER_SUBBLOCK.\n", abbr);
|
|
vkd3d_shader_error(message_context, &location, VKD3D_SHADER_ERROR_DXIL_INVALID_BITCODE,
|
|
"DXIL bitcode chunk has invalid initial block abbreviation %u.", abbr);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
/* Estimate instruction count to avoid reallocation in most shaders. */
|
|
count = max(token_count, 400) - 400;
|
|
if (!vsir_program_init(program, compile_info, &version, (count + (count >> 2)) / 2u + 10, VSIR_CF_BLOCKS))
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
vkd3d_shader_parser_init(&sm6->p, program, message_context, compile_info->source_name);
|
|
sm6->ptr = &sm6->start[1];
|
|
sm6->bitpos = 2;
|
|
|
|
input_signature = &program->input_signature;
|
|
output_signature = &program->output_signature;
|
|
patch_constant_signature = &program->patch_constant_signature;
|
|
*input_signature = dxbc_desc->input_signature;
|
|
*output_signature = dxbc_desc->output_signature;
|
|
*patch_constant_signature = dxbc_desc->patch_constant_signature;
|
|
memset(dxbc_desc, 0, sizeof(*dxbc_desc));
|
|
|
|
block = &sm6->root_block;
|
|
if ((ret = dxil_block_init(block, NULL, sm6)) < 0)
|
|
{
|
|
if (ret == VKD3D_ERROR_OUT_OF_MEMORY)
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory parsing DXIL bitcode chunk.");
|
|
else if (ret == VKD3D_ERROR_INVALID_SHADER)
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_BITCODE,
|
|
"DXIL bitcode chunk has invalid bitcode.");
|
|
else
|
|
vkd3d_unreachable();
|
|
goto fail;
|
|
}
|
|
|
|
dxil_global_abbrevs_cleanup(sm6->abbrevs, sm6->abbrev_count);
|
|
sm6->abbrevs = NULL;
|
|
sm6->abbrev_count = 0;
|
|
|
|
length = sm6->ptr - sm6->start - block->start;
|
|
if (length != block->length)
|
|
{
|
|
WARN("Invalid block length %zu; expected %u.\n", length, block->length);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_INVALID_BLOCK_LENGTH,
|
|
"Root block ends with length %zu but indicated length is %u.", length, block->length);
|
|
}
|
|
if (sm6->ptr != sm6->end)
|
|
{
|
|
size_t expected_length = sm6->end - sm6->start;
|
|
length = sm6->ptr - sm6->start;
|
|
WARN("Invalid module length %zu; expected %zu.\n", length, expected_length);
|
|
vkd3d_shader_parser_warning(&sm6->p, VKD3D_SHADER_WARNING_DXIL_INVALID_MODULE_LENGTH,
|
|
"Module ends with length %zu but indicated length is %zu.", length, expected_length);
|
|
}
|
|
|
|
if ((ret = sm6_parser_type_table_init(sm6)) < 0)
|
|
{
|
|
if (ret == VKD3D_ERROR_OUT_OF_MEMORY)
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory parsing DXIL type table.");
|
|
else if (ret == VKD3D_ERROR_INVALID_SHADER)
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_TYPE_TABLE,
|
|
"DXIL type table is invalid.");
|
|
else
|
|
vkd3d_unreachable();
|
|
goto fail;
|
|
}
|
|
|
|
if ((ret = sm6_parser_symtab_init(sm6)) < 0)
|
|
{
|
|
if (ret == VKD3D_ERROR_OUT_OF_MEMORY)
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory parsing DXIL value symbol table.");
|
|
else if (ret == VKD3D_ERROR_INVALID_SHADER)
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_VALUE_SYMTAB,
|
|
"DXIL value symbol table is invalid.");
|
|
else
|
|
vkd3d_unreachable();
|
|
goto fail;
|
|
}
|
|
|
|
if (!(sm6->output_params = vsir_program_get_dst_params(program, output_signature->element_count))
|
|
|| !(sm6->input_params = vsir_program_get_dst_params(program, input_signature->element_count))
|
|
|| !(sm6->patch_constant_params = vsir_program_get_dst_params(program,
|
|
patch_constant_signature->element_count)))
|
|
{
|
|
ERR("Failed to allocate input/output parameters.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating input/output parameters.");
|
|
ret = VKD3D_ERROR_OUT_OF_MEMORY;
|
|
goto fail;
|
|
}
|
|
|
|
function_count = dxil_block_compute_function_count(&sm6->root_block);
|
|
if (!(sm6->functions = vkd3d_calloc(function_count, sizeof(*sm6->functions))))
|
|
{
|
|
ERR("Failed to allocate function array.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating DXIL function array.");
|
|
ret = VKD3D_ERROR_OUT_OF_MEMORY;
|
|
goto fail;
|
|
}
|
|
|
|
if (sm6_parser_compute_max_value_count(sm6, &sm6->root_block, 0) == SIZE_MAX)
|
|
{
|
|
WARN("Value array count overflowed.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_MODULE,
|
|
"Overflow occurred in the DXIL module value count.");
|
|
ret = VKD3D_ERROR_INVALID_SHADER;
|
|
goto fail;
|
|
}
|
|
if (!(sm6->values = vkd3d_calloc(sm6->value_capacity, sizeof(*sm6->values))))
|
|
{
|
|
ERR("Failed to allocate value array.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory allocating DXIL value array.");
|
|
ret = VKD3D_ERROR_OUT_OF_MEMORY;
|
|
goto fail;
|
|
}
|
|
sm6->function_count = 0;
|
|
sm6->ssa_next_id = 1;
|
|
|
|
if ((ret = sm6_parser_globals_init(sm6)) < 0)
|
|
{
|
|
WARN("Failed to load global declarations.\n");
|
|
goto fail;
|
|
}
|
|
|
|
if (!sm6_parser_allocate_named_metadata(sm6))
|
|
{
|
|
ERR("Failed to allocate named metadata array.\n");
|
|
ret = VKD3D_ERROR_OUT_OF_MEMORY;
|
|
goto fail;
|
|
}
|
|
|
|
for (i = 0, j = 0; i < sm6->root_block.child_block_count; ++i)
|
|
{
|
|
block = sm6->root_block.child_blocks[i];
|
|
if (block->id != METADATA_BLOCK)
|
|
continue;
|
|
|
|
if (j == ARRAY_SIZE(sm6->metadata_tables))
|
|
{
|
|
FIXME("Too many metadata tables.\n");
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_METADATA,
|
|
"A metadata table count greater than %zu is unsupported.", ARRAY_SIZE(sm6->metadata_tables));
|
|
ret = VKD3D_ERROR_INVALID_SHADER;
|
|
goto fail;
|
|
}
|
|
|
|
if ((ret = sm6_parser_metadata_init(sm6, block, &sm6->metadata_tables[j++])) < 0)
|
|
goto fail;
|
|
}
|
|
|
|
if ((ret = sm6_parser_entry_point_init(sm6)) < 0)
|
|
goto fail;
|
|
|
|
if ((ret = sm6_parser_resources_init(sm6)) < 0)
|
|
goto fail;
|
|
|
|
if ((ret = sm6_parser_module_init(sm6, &sm6->root_block, 0)) < 0)
|
|
{
|
|
if (ret == VKD3D_ERROR_OUT_OF_MEMORY)
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory parsing DXIL module.");
|
|
else if (ret == VKD3D_ERROR_INVALID_SHADER)
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_MODULE,
|
|
"DXIL module is invalid.");
|
|
goto fail;
|
|
}
|
|
|
|
if (!sm6_parser_require_space(sm6, output_signature->element_count + input_signature->element_count
|
|
+ patch_constant_signature->element_count))
|
|
{
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_OUT_OF_MEMORY,
|
|
"Out of memory emitting shader signature declarations.");
|
|
ret = VKD3D_ERROR_OUT_OF_MEMORY;
|
|
goto fail;
|
|
}
|
|
|
|
program->ssa_count = sm6->ssa_next_id;
|
|
|
|
if (!(fn = sm6_parser_get_function(sm6, sm6->entry_point)))
|
|
{
|
|
WARN("Failed to find entry point %s.\n", sm6->entry_point);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_ENTRY_POINT,
|
|
"The definition of the entry point function '%s' was not found.", sm6->entry_point);
|
|
ret = VKD3D_ERROR_INVALID_SHADER;
|
|
goto fail;
|
|
}
|
|
|
|
if (version.type == VKD3D_SHADER_TYPE_HULL)
|
|
{
|
|
sm6_parser_add_instruction(sm6, VKD3DSIH_HS_CONTROL_POINT_PHASE);
|
|
|
|
if ((ret = sm6_function_emit_blocks(fn, sm6)) < 0)
|
|
goto fail;
|
|
|
|
if (!(fn = sm6_parser_get_function(sm6, sm6->patch_constant_function)))
|
|
{
|
|
WARN("Failed to find patch constant function '%s'.\n", sm6->patch_constant_function);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_MODULE,
|
|
"Failed to find the patch constant function '%s' for a hull shader.",
|
|
sm6->patch_constant_function);
|
|
ret = VKD3D_ERROR_INVALID_SHADER;
|
|
goto fail;
|
|
}
|
|
|
|
sm6_parser_add_instruction(sm6, VKD3DSIH_HS_FORK_PHASE);
|
|
if ((ret = sm6_function_emit_blocks(fn, sm6)) < 0)
|
|
goto fail;
|
|
|
|
expected_function_count = 2;
|
|
}
|
|
else
|
|
{
|
|
if ((ret = sm6_function_emit_blocks(fn, sm6)) < 0)
|
|
goto fail;
|
|
expected_function_count = 1;
|
|
}
|
|
|
|
if (sm6->function_count > expected_function_count)
|
|
{
|
|
FIXME("%zu unhandled functions.\n", sm6->function_count - expected_function_count);
|
|
vkd3d_shader_parser_error(&sm6->p, VKD3D_SHADER_ERROR_DXIL_INVALID_MODULE,
|
|
"%zu functions were not emitted.", sm6->function_count - expected_function_count);
|
|
}
|
|
|
|
dxil_block_destroy(&sm6->root_block);
|
|
|
|
if (sm6->p.failed)
|
|
{
|
|
ret = VKD3D_ERROR_INVALID_SHADER;
|
|
goto fail;
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
|
|
fail:
|
|
sm6_parser_cleanup(sm6);
|
|
vsir_program_cleanup(program);
|
|
return ret;
|
|
}
|
|
|
|
int dxil_parse(const struct vkd3d_shader_compile_info *compile_info, uint64_t config_flags,
|
|
struct vkd3d_shader_message_context *message_context, struct vsir_program *program)
|
|
{
|
|
struct dxbc_shader_desc dxbc_desc = {0};
|
|
struct sm6_parser sm6 = {0};
|
|
uint32_t *byte_code = NULL;
|
|
int ret;
|
|
|
|
MESSAGE("Creating a DXIL parser. This is unsupported; you get to keep all the pieces if it breaks.\n");
|
|
|
|
dxbc_desc.is_dxil = true;
|
|
if ((ret = shader_extract_from_dxbc(&compile_info->source, message_context, compile_info->source_name,
|
|
&dxbc_desc)) < 0)
|
|
{
|
|
WARN("Failed to extract shader, vkd3d result %d.\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
if (((uintptr_t)dxbc_desc.byte_code & (VKD3D_DXBC_CHUNK_ALIGNMENT - 1)))
|
|
{
|
|
/* LLVM bitcode should be 32-bit aligned, but before dxc v1.7.2207 this was not always the case in the DXBC
|
|
* container due to missing padding after signature names. Get an aligned copy to prevent unaligned access. */
|
|
if (!(byte_code = vkd3d_malloc(align(dxbc_desc.byte_code_size, VKD3D_DXBC_CHUNK_ALIGNMENT))))
|
|
{
|
|
ERR("Failed to allocate aligned chunk.\n");
|
|
free_dxbc_shader_desc(&dxbc_desc);
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
|
|
memcpy(byte_code, dxbc_desc.byte_code, dxbc_desc.byte_code_size);
|
|
dxbc_desc.byte_code = byte_code;
|
|
}
|
|
|
|
ret = sm6_parser_init(&sm6, program, compile_info, message_context, &dxbc_desc);
|
|
free_dxbc_shader_desc(&dxbc_desc);
|
|
vkd3d_free(byte_code);
|
|
|
|
if (ret < 0)
|
|
{
|
|
WARN("Failed to parse shader.\n");
|
|
return ret;
|
|
}
|
|
|
|
if ((ret = vkd3d_shader_parser_validate(&sm6.p, config_flags)) < 0)
|
|
{
|
|
WARN("Failed to validate shader after parsing, ret %d.\n", ret);
|
|
|
|
if (TRACE_ON())
|
|
vkd3d_shader_trace(program);
|
|
|
|
sm6_parser_cleanup(&sm6);
|
|
vsir_program_cleanup(program);
|
|
return ret;
|
|
}
|
|
|
|
sm6_parser_cleanup(&sm6);
|
|
|
|
return VKD3D_OK;
|
|
}
|