| /************************************************************************** |
| * |
| * Copyright 2009-2010 VMware, Inc. |
| * All Rights Reserved. |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a |
| * copy of this software and associated documentation files (the |
| * "Software"), to deal in the Software without restriction, including |
| * without limitation the rights to use, copy, modify, merge, publish, |
| * distribute, sub license, and/or sell copies of the Software, and to |
| * permit persons to whom the Software is furnished to do so, subject to |
| * the following conditions: |
| * |
| * The above copyright notice and this permission notice (including the |
| * next paragraph) shall be included in all copies or substantial portions |
| * of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS |
| * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
| * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. |
| * IN NO EVENT SHALL VMWARE, INC AND/OR ITS SUPPLIERS BE LIABLE FOR |
| * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, |
| * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE |
| * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. |
| * |
| **************************************************************************/ |
| |
| |
| #include "pipe/p_screen.h" |
| #include "pipe/p_context.h" |
| #include "pipe/p_state.h" |
| #include "tgsi/tgsi_ureg.h" |
| #include "tgsi/tgsi_build.h" |
| #include "tgsi/tgsi_from_mesa.h" |
| #include "tgsi/tgsi_info.h" |
| #include "tgsi/tgsi_dump.h" |
| #include "tgsi/tgsi_sanity.h" |
| #include "util/u_debug.h" |
| #include "util/u_inlines.h" |
| #include "util/u_memory.h" |
| #include "util/u_math.h" |
| #include "util/u_prim.h" |
| #include "util/u_bitmask.h" |
| #include "GL/gl.h" |
| #include "compiler/shader_info.h" |
| |
| union tgsi_any_token { |
| struct tgsi_header header; |
| struct tgsi_processor processor; |
| struct tgsi_token token; |
| struct tgsi_property prop; |
| struct tgsi_property_data prop_data; |
| struct tgsi_declaration decl; |
| struct tgsi_declaration_range decl_range; |
| struct tgsi_declaration_dimension decl_dim; |
| struct tgsi_declaration_interp decl_interp; |
| struct tgsi_declaration_image decl_image; |
| struct tgsi_declaration_semantic decl_semantic; |
| struct tgsi_declaration_sampler_view decl_sampler_view; |
| struct tgsi_declaration_array array; |
| struct tgsi_immediate imm; |
| union tgsi_immediate_data imm_data; |
| struct tgsi_instruction insn; |
| struct tgsi_instruction_label insn_label; |
| struct tgsi_instruction_texture insn_texture; |
| struct tgsi_instruction_memory insn_memory; |
| struct tgsi_texture_offset insn_texture_offset; |
| struct tgsi_src_register src; |
| struct tgsi_ind_register ind; |
| struct tgsi_dimension dim; |
| struct tgsi_dst_register dst; |
| unsigned value; |
| }; |
| |
| |
| struct ureg_tokens { |
| union tgsi_any_token *tokens; |
| unsigned size; |
| unsigned order; |
| unsigned count; |
| }; |
| |
| #define UREG_MAX_INPUT (4 * PIPE_MAX_SHADER_INPUTS) |
| #define UREG_MAX_SYSTEM_VALUE PIPE_MAX_ATTRIBS |
| #define UREG_MAX_OUTPUT (4 * PIPE_MAX_SHADER_OUTPUTS) |
| #define UREG_MAX_CONSTANT_RANGE 32 |
| #define UREG_MAX_HW_ATOMIC_RANGE 32 |
| #define UREG_MAX_IMMEDIATE 4096 |
| #define UREG_MAX_ADDR 3 |
| #define UREG_MAX_ARRAY_TEMPS 256 |
| |
| struct const_decl { |
| struct { |
| unsigned first; |
| unsigned last; |
| } constant_range[UREG_MAX_CONSTANT_RANGE]; |
| unsigned nr_constant_ranges; |
| }; |
| |
| struct hw_atomic_decl { |
| struct hw_atomic_decl_range { |
| unsigned first; |
| unsigned last; |
| unsigned array_id; |
| } hw_atomic_range[UREG_MAX_HW_ATOMIC_RANGE]; |
| unsigned nr_hw_atomic_ranges; |
| }; |
| |
| #define DOMAIN_DECL 0 |
| #define DOMAIN_INSN 1 |
| |
| struct ureg_program |
| { |
| enum pipe_shader_type processor; |
| bool supports_any_inout_decl_range; |
| int next_shader_processor; |
| |
| struct ureg_input_decl { |
| enum tgsi_semantic semantic_name; |
| unsigned semantic_index; |
| enum tgsi_interpolate_mode interp; |
| unsigned char usage_mask; |
| enum tgsi_interpolate_loc interp_location; |
| unsigned first; |
| unsigned last; |
| unsigned array_id; |
| } input[UREG_MAX_INPUT]; |
| unsigned nr_inputs, nr_input_regs; |
| |
| unsigned vs_inputs[PIPE_MAX_ATTRIBS/32]; |
| |
| struct { |
| enum tgsi_semantic semantic_name; |
| unsigned semantic_index; |
| } system_value[UREG_MAX_SYSTEM_VALUE]; |
| unsigned nr_system_values; |
| |
| struct ureg_output_decl { |
| enum tgsi_semantic semantic_name; |
| unsigned semantic_index; |
| unsigned streams; |
| unsigned usage_mask; /* = TGSI_WRITEMASK_* */ |
| unsigned first; |
| unsigned last; |
| unsigned array_id; |
| boolean invariant; |
| } output[UREG_MAX_OUTPUT]; |
| unsigned nr_outputs, nr_output_regs; |
| |
| struct { |
| union { |
| float f[4]; |
| unsigned u[4]; |
| int i[4]; |
| } value; |
| unsigned nr; |
| unsigned type; |
| } immediate[UREG_MAX_IMMEDIATE]; |
| unsigned nr_immediates; |
| |
| struct ureg_src sampler[PIPE_MAX_SAMPLERS]; |
| unsigned nr_samplers; |
| |
| struct { |
| unsigned index; |
| enum tgsi_texture_type target; |
| enum tgsi_return_type return_type_x; |
| enum tgsi_return_type return_type_y; |
| enum tgsi_return_type return_type_z; |
| enum tgsi_return_type return_type_w; |
| } sampler_view[PIPE_MAX_SHADER_SAMPLER_VIEWS]; |
| unsigned nr_sampler_views; |
| |
| struct { |
| unsigned index; |
| enum tgsi_texture_type target; |
| enum pipe_format format; |
| boolean wr; |
| boolean raw; |
| } image[PIPE_MAX_SHADER_IMAGES]; |
| unsigned nr_images; |
| |
| struct { |
| unsigned index; |
| bool atomic; |
| } buffer[PIPE_MAX_SHADER_BUFFERS]; |
| unsigned nr_buffers; |
| |
| struct util_bitmask *free_temps; |
| struct util_bitmask *local_temps; |
| struct util_bitmask *decl_temps; |
| unsigned nr_temps; |
| |
| unsigned array_temps[UREG_MAX_ARRAY_TEMPS]; |
| unsigned nr_array_temps; |
| |
| struct const_decl const_decls[PIPE_MAX_CONSTANT_BUFFERS]; |
| |
| struct hw_atomic_decl hw_atomic_decls[PIPE_MAX_HW_ATOMIC_BUFFERS]; |
| |
| unsigned properties[TGSI_PROPERTY_COUNT]; |
| |
| unsigned nr_addrs; |
| unsigned nr_instructions; |
| |
| struct ureg_tokens domain[2]; |
| |
| bool use_memory[TGSI_MEMORY_TYPE_COUNT]; |
| |
| bool precise; |
| }; |
| |
| static union tgsi_any_token error_tokens[32]; |
| |
| static void tokens_error( struct ureg_tokens *tokens ) |
| { |
| if (tokens->tokens && tokens->tokens != error_tokens) |
| FREE(tokens->tokens); |
| |
| tokens->tokens = error_tokens; |
| tokens->size = ARRAY_SIZE(error_tokens); |
| tokens->count = 0; |
| } |
| |
| |
| static void tokens_expand( struct ureg_tokens *tokens, |
| unsigned count ) |
| { |
| unsigned old_size = tokens->size * sizeof(unsigned); |
| |
| if (tokens->tokens == error_tokens) { |
| return; |
| } |
| |
| while (tokens->count + count > tokens->size) { |
| tokens->size = (1 << ++tokens->order); |
| } |
| |
| tokens->tokens = REALLOC(tokens->tokens, |
| old_size, |
| tokens->size * sizeof(unsigned)); |
| if (tokens->tokens == NULL) { |
| tokens_error(tokens); |
| } |
| } |
| |
| static void set_bad( struct ureg_program *ureg ) |
| { |
| tokens_error(&ureg->domain[0]); |
| } |
| |
| |
| |
| static union tgsi_any_token *get_tokens( struct ureg_program *ureg, |
| unsigned domain, |
| unsigned count ) |
| { |
| struct ureg_tokens *tokens = &ureg->domain[domain]; |
| union tgsi_any_token *result; |
| |
| if (tokens->count + count > tokens->size) |
| tokens_expand(tokens, count); |
| |
| result = &tokens->tokens[tokens->count]; |
| tokens->count += count; |
| return result; |
| } |
| |
| |
| static union tgsi_any_token *retrieve_token( struct ureg_program *ureg, |
| unsigned domain, |
| unsigned nr ) |
| { |
| if (ureg->domain[domain].tokens == error_tokens) |
| return &error_tokens[0]; |
| |
| return &ureg->domain[domain].tokens[nr]; |
| } |
| |
| |
| void |
| ureg_property(struct ureg_program *ureg, unsigned name, unsigned value) |
| { |
| assert(name < ARRAY_SIZE(ureg->properties)); |
| ureg->properties[name] = value; |
| } |
| |
| struct ureg_src |
| ureg_DECL_fs_input_centroid_layout(struct ureg_program *ureg, |
| enum tgsi_semantic semantic_name, |
| unsigned semantic_index, |
| enum tgsi_interpolate_mode interp_mode, |
| enum tgsi_interpolate_loc interp_location, |
| unsigned index, |
| unsigned usage_mask, |
| unsigned array_id, |
| unsigned array_size) |
| { |
| unsigned i; |
| |
| assert(usage_mask != 0); |
| assert(usage_mask <= TGSI_WRITEMASK_XYZW); |
| |
| for (i = 0; i < ureg->nr_inputs; i++) { |
| if (ureg->input[i].semantic_name == semantic_name && |
| ureg->input[i].semantic_index == semantic_index) { |
| assert(ureg->input[i].interp == interp_mode); |
| assert(ureg->input[i].interp_location == interp_location); |
| if (ureg->input[i].array_id == array_id) { |
| ureg->input[i].usage_mask |= usage_mask; |
| ureg->input[i].last = MAX2(ureg->input[i].last, ureg->input[i].first + array_size - 1); |
| ureg->nr_input_regs = MAX2(ureg->nr_input_regs, ureg->input[i].last + 1); |
| goto out; |
| } |
| assert((ureg->input[i].usage_mask & usage_mask) == 0); |
| } |
| } |
| |
| if (ureg->nr_inputs < UREG_MAX_INPUT) { |
| assert(array_size >= 1); |
| ureg->input[i].semantic_name = semantic_name; |
| ureg->input[i].semantic_index = semantic_index; |
| ureg->input[i].interp = interp_mode; |
| ureg->input[i].interp_location = interp_location; |
| ureg->input[i].first = index; |
| ureg->input[i].last = index + array_size - 1; |
| ureg->input[i].array_id = array_id; |
| ureg->input[i].usage_mask = usage_mask; |
| ureg->nr_input_regs = MAX2(ureg->nr_input_regs, index + array_size); |
| ureg->nr_inputs++; |
| } else { |
| set_bad(ureg); |
| } |
| |
| out: |
| return ureg_src_array_register(TGSI_FILE_INPUT, ureg->input[i].first, |
| array_id); |
| } |
| |
| struct ureg_src |
| ureg_DECL_fs_input_centroid(struct ureg_program *ureg, |
| enum tgsi_semantic semantic_name, |
| unsigned semantic_index, |
| enum tgsi_interpolate_mode interp_mode, |
| enum tgsi_interpolate_loc interp_location, |
| unsigned array_id, |
| unsigned array_size) |
| { |
| return ureg_DECL_fs_input_centroid_layout(ureg, |
| semantic_name, semantic_index, interp_mode, |
| interp_location, |
| ureg->nr_input_regs, TGSI_WRITEMASK_XYZW, array_id, array_size); |
| } |
| |
| |
| struct ureg_src |
| ureg_DECL_vs_input( struct ureg_program *ureg, |
| unsigned index ) |
| { |
| assert(ureg->processor == PIPE_SHADER_VERTEX); |
| assert(index / 32 < ARRAY_SIZE(ureg->vs_inputs)); |
| |
| ureg->vs_inputs[index/32] |= 1 << (index % 32); |
| return ureg_src_register( TGSI_FILE_INPUT, index ); |
| } |
| |
| |
| struct ureg_src |
| ureg_DECL_input_layout(struct ureg_program *ureg, |
| enum tgsi_semantic semantic_name, |
| unsigned semantic_index, |
| unsigned index, |
| unsigned usage_mask, |
| unsigned array_id, |
| unsigned array_size) |
| { |
| return ureg_DECL_fs_input_centroid_layout(ureg, |
| semantic_name, semantic_index, |
| TGSI_INTERPOLATE_CONSTANT, TGSI_INTERPOLATE_LOC_CENTER, |
| index, usage_mask, array_id, array_size); |
| } |
| |
| |
| struct ureg_src |
| ureg_DECL_input(struct ureg_program *ureg, |
| enum tgsi_semantic semantic_name, |
| unsigned semantic_index, |
| unsigned array_id, |
| unsigned array_size) |
| { |
| return ureg_DECL_fs_input_centroid(ureg, semantic_name, semantic_index, |
| TGSI_INTERPOLATE_CONSTANT, |
| TGSI_INTERPOLATE_LOC_CENTER, |
| array_id, array_size); |
| } |
| |
| |
| struct ureg_src |
| ureg_DECL_system_value(struct ureg_program *ureg, |
| enum tgsi_semantic semantic_name, |
| unsigned semantic_index) |
| { |
| unsigned i; |
| |
| for (i = 0; i < ureg->nr_system_values; i++) { |
| if (ureg->system_value[i].semantic_name == semantic_name && |
| ureg->system_value[i].semantic_index == semantic_index) { |
| goto out; |
| } |
| } |
| |
| if (ureg->nr_system_values < UREG_MAX_SYSTEM_VALUE) { |
| ureg->system_value[ureg->nr_system_values].semantic_name = semantic_name; |
| ureg->system_value[ureg->nr_system_values].semantic_index = semantic_index; |
| i = ureg->nr_system_values; |
| ureg->nr_system_values++; |
| } else { |
| set_bad(ureg); |
| } |
| |
| out: |
| return ureg_src_register(TGSI_FILE_SYSTEM_VALUE, i); |
| } |
| |
| |
| struct ureg_dst |
| ureg_DECL_output_layout(struct ureg_program *ureg, |
| enum tgsi_semantic semantic_name, |
| unsigned semantic_index, |
| unsigned streams, |
| unsigned index, |
| unsigned usage_mask, |
| unsigned array_id, |
| unsigned array_size, |
| boolean invariant) |
| { |
| unsigned i; |
| |
| assert(usage_mask != 0); |
| assert(!(streams & 0x03) || (usage_mask & 1)); |
| assert(!(streams & 0x0c) || (usage_mask & 2)); |
| assert(!(streams & 0x30) || (usage_mask & 4)); |
| assert(!(streams & 0xc0) || (usage_mask & 8)); |
| |
| for (i = 0; i < ureg->nr_outputs; i++) { |
| if (ureg->output[i].semantic_name == semantic_name && |
| ureg->output[i].semantic_index == semantic_index) { |
| if (ureg->output[i].array_id == array_id) { |
| ureg->output[i].usage_mask |= usage_mask; |
| ureg->output[i].last = MAX2(ureg->output[i].last, ureg->output[i].first + array_size - 1); |
| ureg->nr_output_regs = MAX2(ureg->nr_output_regs, ureg->output[i].last + 1); |
| goto out; |
| } |
| assert((ureg->output[i].usage_mask & usage_mask) == 0); |
| } |
| } |
| |
| if (ureg->nr_outputs < UREG_MAX_OUTPUT) { |
| ureg->output[i].semantic_name = semantic_name; |
| ureg->output[i].semantic_index = semantic_index; |
| ureg->output[i].usage_mask = usage_mask; |
| ureg->output[i].first = index; |
| ureg->output[i].last = index + array_size - 1; |
| ureg->output[i].array_id = array_id; |
| ureg->output[i].invariant = invariant; |
| ureg->nr_output_regs = MAX2(ureg->nr_output_regs, index + array_size); |
| ureg->nr_outputs++; |
| } |
| else { |
| set_bad( ureg ); |
| i = 0; |
| } |
| |
| out: |
| ureg->output[i].streams |= streams; |
| |
| return ureg_dst_array_register(TGSI_FILE_OUTPUT, ureg->output[i].first, |
| array_id); |
| } |
| |
| |
| struct ureg_dst |
| ureg_DECL_output_masked(struct ureg_program *ureg, |
| unsigned name, |
| unsigned index, |
| unsigned usage_mask, |
| unsigned array_id, |
| unsigned array_size) |
| { |
| return ureg_DECL_output_layout(ureg, name, index, 0, |
| ureg->nr_output_regs, usage_mask, array_id, |
| array_size, FALSE); |
| } |
| |
| |
| struct ureg_dst |
| ureg_DECL_output(struct ureg_program *ureg, |
| enum tgsi_semantic name, |
| unsigned index) |
| { |
| return ureg_DECL_output_masked(ureg, name, index, TGSI_WRITEMASK_XYZW, |
| 0, 1); |
| } |
| |
| struct ureg_dst |
| ureg_DECL_output_array(struct ureg_program *ureg, |
| enum tgsi_semantic semantic_name, |
| unsigned semantic_index, |
| unsigned array_id, |
| unsigned array_size) |
| { |
| return ureg_DECL_output_masked(ureg, semantic_name, semantic_index, |
| TGSI_WRITEMASK_XYZW, |
| array_id, array_size); |
| } |
| |
| |
| /* Returns a new constant register. Keep track of which have been |
| * referred to so that we can emit decls later. |
| * |
| * Constant operands declared with this function must be addressed |
| * with a two-dimensional index. |
| * |
| * There is nothing in this code to bind this constant to any tracked |
| * value or manage any constant_buffer contents -- that's the |
| * resposibility of the calling code. |
| */ |
| void |
| ureg_DECL_constant2D(struct ureg_program *ureg, |
| unsigned first, |
| unsigned last, |
| unsigned index2D) |
| { |
| struct const_decl *decl = &ureg->const_decls[index2D]; |
| |
| assert(index2D < PIPE_MAX_CONSTANT_BUFFERS); |
| |
| if (decl->nr_constant_ranges < UREG_MAX_CONSTANT_RANGE) { |
| uint i = decl->nr_constant_ranges++; |
| |
| decl->constant_range[i].first = first; |
| decl->constant_range[i].last = last; |
| } |
| } |
| |
| |
| /* A one-dimensional, deprecated version of ureg_DECL_constant2D(). |
| * |
| * Constant operands declared with this function must be addressed |
| * with a one-dimensional index. |
| */ |
| struct ureg_src |
| ureg_DECL_constant(struct ureg_program *ureg, |
| unsigned index) |
| { |
| struct const_decl *decl = &ureg->const_decls[0]; |
| unsigned minconst = index, maxconst = index; |
| unsigned i; |
| |
| /* Inside existing range? |
| */ |
| for (i = 0; i < decl->nr_constant_ranges; i++) { |
| if (decl->constant_range[i].first <= index && |
| decl->constant_range[i].last >= index) { |
| goto out; |
| } |
| } |
| |
| /* Extend existing range? |
| */ |
| for (i = 0; i < decl->nr_constant_ranges; i++) { |
| if (decl->constant_range[i].last == index - 1) { |
| decl->constant_range[i].last = index; |
| goto out; |
| } |
| |
| if (decl->constant_range[i].first == index + 1) { |
| decl->constant_range[i].first = index; |
| goto out; |
| } |
| |
| minconst = MIN2(minconst, decl->constant_range[i].first); |
| maxconst = MAX2(maxconst, decl->constant_range[i].last); |
| } |
| |
| /* Create new range? |
| */ |
| if (decl->nr_constant_ranges < UREG_MAX_CONSTANT_RANGE) { |
| i = decl->nr_constant_ranges++; |
| decl->constant_range[i].first = index; |
| decl->constant_range[i].last = index; |
| goto out; |
| } |
| |
| /* Collapse all ranges down to one: |
| */ |
| i = 0; |
| decl->constant_range[0].first = minconst; |
| decl->constant_range[0].last = maxconst; |
| decl->nr_constant_ranges = 1; |
| |
| out: |
| assert(i < decl->nr_constant_ranges); |
| assert(decl->constant_range[i].first <= index); |
| assert(decl->constant_range[i].last >= index); |
| |
| struct ureg_src src = ureg_src_register(TGSI_FILE_CONSTANT, index); |
| return ureg_src_dimension(src, 0); |
| } |
| |
| |
| /* Returns a new hw atomic register. Keep track of which have been |
| * referred to so that we can emit decls later. |
| */ |
| void |
| ureg_DECL_hw_atomic(struct ureg_program *ureg, |
| unsigned first, |
| unsigned last, |
| unsigned buffer_id, |
| unsigned array_id) |
| { |
| struct hw_atomic_decl *decl = &ureg->hw_atomic_decls[buffer_id]; |
| |
| if (decl->nr_hw_atomic_ranges < UREG_MAX_HW_ATOMIC_RANGE) { |
| uint i = decl->nr_hw_atomic_ranges++; |
| |
| decl->hw_atomic_range[i].first = first; |
| decl->hw_atomic_range[i].last = last; |
| decl->hw_atomic_range[i].array_id = array_id; |
| } else { |
| set_bad(ureg); |
| } |
| } |
| |
| static struct ureg_dst alloc_temporary( struct ureg_program *ureg, |
| boolean local ) |
| { |
| unsigned i; |
| |
| /* Look for a released temporary. |
| */ |
| for (i = util_bitmask_get_first_index(ureg->free_temps); |
| i != UTIL_BITMASK_INVALID_INDEX; |
| i = util_bitmask_get_next_index(ureg->free_temps, i + 1)) { |
| if (util_bitmask_get(ureg->local_temps, i) == local) |
| break; |
| } |
| |
| /* Or allocate a new one. |
| */ |
| if (i == UTIL_BITMASK_INVALID_INDEX) { |
| i = ureg->nr_temps++; |
| |
| if (local) |
| util_bitmask_set(ureg->local_temps, i); |
| |
| /* Start a new declaration when the local flag changes */ |
| if (!i || util_bitmask_get(ureg->local_temps, i - 1) != local) |
| util_bitmask_set(ureg->decl_temps, i); |
| } |
| |
| util_bitmask_clear(ureg->free_temps, i); |
| |
| return ureg_dst_register( TGSI_FILE_TEMPORARY, i ); |
| } |
| |
| struct ureg_dst ureg_DECL_temporary( struct ureg_program *ureg ) |
| { |
| return alloc_temporary(ureg, FALSE); |
| } |
| |
| struct ureg_dst ureg_DECL_local_temporary( struct ureg_program *ureg ) |
| { |
| return alloc_temporary(ureg, TRUE); |
| } |
| |
| struct ureg_dst ureg_DECL_array_temporary( struct ureg_program *ureg, |
| unsigned size, |
| boolean local ) |
| { |
| unsigned i = ureg->nr_temps; |
| struct ureg_dst dst = ureg_dst_register( TGSI_FILE_TEMPORARY, i ); |
| |
| if (local) |
| util_bitmask_set(ureg->local_temps, i); |
| |
| /* Always start a new declaration at the start */ |
| util_bitmask_set(ureg->decl_temps, i); |
| |
| ureg->nr_temps += size; |
| |
| /* and also at the end of the array */ |
| util_bitmask_set(ureg->decl_temps, ureg->nr_temps); |
| |
| if (ureg->nr_array_temps < UREG_MAX_ARRAY_TEMPS) { |
| ureg->array_temps[ureg->nr_array_temps++] = i; |
| dst.ArrayID = ureg->nr_array_temps; |
| } |
| |
| return dst; |
| } |
| |
| void ureg_release_temporary( struct ureg_program *ureg, |
| struct ureg_dst tmp ) |
| { |
| if(tmp.File == TGSI_FILE_TEMPORARY) |
| util_bitmask_set(ureg->free_temps, tmp.Index); |
| } |
| |
| |
| /* Allocate a new address register. |
| */ |
| struct ureg_dst ureg_DECL_address( struct ureg_program *ureg ) |
| { |
| if (ureg->nr_addrs < UREG_MAX_ADDR) |
| return ureg_dst_register( TGSI_FILE_ADDRESS, ureg->nr_addrs++ ); |
| |
| assert( 0 ); |
| return ureg_dst_register( TGSI_FILE_ADDRESS, 0 ); |
| } |
| |
| /* Allocate a new sampler. |
| */ |
| struct ureg_src ureg_DECL_sampler( struct ureg_program *ureg, |
| unsigned nr ) |
| { |
| unsigned i; |
| |
| for (i = 0; i < ureg->nr_samplers; i++) |
| if (ureg->sampler[i].Index == (int)nr) |
| return ureg->sampler[i]; |
| |
| if (i < PIPE_MAX_SAMPLERS) { |
| ureg->sampler[i] = ureg_src_register( TGSI_FILE_SAMPLER, nr ); |
| ureg->nr_samplers++; |
| return ureg->sampler[i]; |
| } |
| |
| assert( 0 ); |
| return ureg->sampler[0]; |
| } |
| |
| /* |
| * Allocate a new shader sampler view. |
| */ |
| struct ureg_src |
| ureg_DECL_sampler_view(struct ureg_program *ureg, |
| unsigned index, |
| enum tgsi_texture_type target, |
| enum tgsi_return_type return_type_x, |
| enum tgsi_return_type return_type_y, |
| enum tgsi_return_type return_type_z, |
| enum tgsi_return_type return_type_w) |
| { |
| struct ureg_src reg = ureg_src_register(TGSI_FILE_SAMPLER_VIEW, index); |
| uint i; |
| |
| for (i = 0; i < ureg->nr_sampler_views; i++) { |
| if (ureg->sampler_view[i].index == index) { |
| return reg; |
| } |
| } |
| |
| if (i < PIPE_MAX_SHADER_SAMPLER_VIEWS) { |
| ureg->sampler_view[i].index = index; |
| ureg->sampler_view[i].target = target; |
| ureg->sampler_view[i].return_type_x = return_type_x; |
| ureg->sampler_view[i].return_type_y = return_type_y; |
| ureg->sampler_view[i].return_type_z = return_type_z; |
| ureg->sampler_view[i].return_type_w = return_type_w; |
| ureg->nr_sampler_views++; |
| return reg; |
| } |
| |
| assert(0); |
| return reg; |
| } |
| |
| /* Allocate a new image. |
| */ |
| struct ureg_src |
| ureg_DECL_image(struct ureg_program *ureg, |
| unsigned index, |
| enum tgsi_texture_type target, |
| enum pipe_format format, |
| boolean wr, |
| boolean raw) |
| { |
| struct ureg_src reg = ureg_src_register(TGSI_FILE_IMAGE, index); |
| unsigned i; |
| |
| for (i = 0; i < ureg->nr_images; i++) |
| if (ureg->image[i].index == index) |
| return reg; |
| |
| if (i < PIPE_MAX_SHADER_IMAGES) { |
| ureg->image[i].index = index; |
| ureg->image[i].target = target; |
| ureg->image[i].wr = wr; |
| ureg->image[i].raw = raw; |
| ureg->image[i].format = format; |
| ureg->nr_images++; |
| return reg; |
| } |
| |
| assert(0); |
| return reg; |
| } |
| |
| /* Allocate a new buffer. |
| */ |
| struct ureg_src ureg_DECL_buffer(struct ureg_program *ureg, unsigned nr, |
| bool atomic) |
| { |
| struct ureg_src reg = ureg_src_register(TGSI_FILE_BUFFER, nr); |
| unsigned i; |
| |
| for (i = 0; i < ureg->nr_buffers; i++) |
| if (ureg->buffer[i].index == nr) |
| return reg; |
| |
| if (i < PIPE_MAX_SHADER_BUFFERS) { |
| ureg->buffer[i].index = nr; |
| ureg->buffer[i].atomic = atomic; |
| ureg->nr_buffers++; |
| return reg; |
| } |
| |
| assert(0); |
| return reg; |
| } |
| |
| /* Allocate a memory area. |
| */ |
| struct ureg_src ureg_DECL_memory(struct ureg_program *ureg, |
| unsigned memory_type) |
| { |
| struct ureg_src reg = ureg_src_register(TGSI_FILE_MEMORY, memory_type); |
| |
| ureg->use_memory[memory_type] = true; |
| return reg; |
| } |
| |
| static int |
| match_or_expand_immediate64( const unsigned *v, |
| unsigned nr, |
| unsigned *v2, |
| unsigned *pnr2, |
| unsigned *swizzle ) |
| { |
| unsigned nr2 = *pnr2; |
| unsigned i, j; |
| *swizzle = 0; |
| |
| for (i = 0; i < nr; i += 2) { |
| boolean found = FALSE; |
| |
| for (j = 0; j < nr2 && !found; j += 2) { |
| if (v[i] == v2[j] && v[i + 1] == v2[j + 1]) { |
| *swizzle |= (j << (i * 2)) | ((j + 1) << ((i + 1) * 2)); |
| found = TRUE; |
| } |
| } |
| if (!found) { |
| if ((nr2) >= 4) { |
| return FALSE; |
| } |
| |
| v2[nr2] = v[i]; |
| v2[nr2 + 1] = v[i + 1]; |
| |
| *swizzle |= (nr2 << (i * 2)) | ((nr2 + 1) << ((i + 1) * 2)); |
| nr2 += 2; |
| } |
| } |
| |
| /* Actually expand immediate only when fully succeeded. |
| */ |
| *pnr2 = nr2; |
| return TRUE; |
| } |
| |
| static int |
| match_or_expand_immediate( const unsigned *v, |
| int type, |
| unsigned nr, |
| unsigned *v2, |
| unsigned *pnr2, |
| unsigned *swizzle ) |
| { |
| unsigned nr2 = *pnr2; |
| unsigned i, j; |
| |
| if (type == TGSI_IMM_FLOAT64 || |
| type == TGSI_IMM_UINT64 || |
| type == TGSI_IMM_INT64) |
| return match_or_expand_immediate64(v, nr, v2, pnr2, swizzle); |
| |
| *swizzle = 0; |
| |
| for (i = 0; i < nr; i++) { |
| boolean found = FALSE; |
| |
| for (j = 0; j < nr2 && !found; j++) { |
| if (v[i] == v2[j]) { |
| *swizzle |= j << (i * 2); |
| found = TRUE; |
| } |
| } |
| |
| if (!found) { |
| if (nr2 >= 4) { |
| return FALSE; |
| } |
| |
| v2[nr2] = v[i]; |
| *swizzle |= nr2 << (i * 2); |
| nr2++; |
| } |
| } |
| |
| /* Actually expand immediate only when fully succeeded. |
| */ |
| *pnr2 = nr2; |
| return TRUE; |
| } |
| |
| |
| static struct ureg_src |
| decl_immediate( struct ureg_program *ureg, |
| const unsigned *v, |
| unsigned nr, |
| unsigned type ) |
| { |
| unsigned i, j; |
| unsigned swizzle = 0; |
| |
| /* Could do a first pass where we examine all existing immediates |
| * without expanding. |
| */ |
| |
| for (i = 0; i < ureg->nr_immediates; i++) { |
| if (ureg->immediate[i].type != type) { |
| continue; |
| } |
| if (match_or_expand_immediate(v, |
| type, |
| nr, |
| ureg->immediate[i].value.u, |
| &ureg->immediate[i].nr, |
| &swizzle)) { |
| goto out; |
| } |
| } |
| |
| if (ureg->nr_immediates < UREG_MAX_IMMEDIATE) { |
| i = ureg->nr_immediates++; |
| ureg->immediate[i].type = type; |
| if (match_or_expand_immediate(v, |
| type, |
| nr, |
| ureg->immediate[i].value.u, |
| &ureg->immediate[i].nr, |
| &swizzle)) { |
| goto out; |
| } |
| } |
| |
| set_bad(ureg); |
| |
| out: |
| /* Make sure that all referenced elements are from this immediate. |
| * Has the effect of making size-one immediates into scalars. |
| */ |
| if (type == TGSI_IMM_FLOAT64 || |
| type == TGSI_IMM_UINT64 || |
| type == TGSI_IMM_INT64) { |
| for (j = nr; j < 4; j+=2) { |
| swizzle |= (swizzle & 0xf) << (j * 2); |
| } |
| } else { |
| for (j = nr; j < 4; j++) { |
| swizzle |= (swizzle & 0x3) << (j * 2); |
| } |
| } |
| return ureg_swizzle(ureg_src_register(TGSI_FILE_IMMEDIATE, i), |
| (swizzle >> 0) & 0x3, |
| (swizzle >> 2) & 0x3, |
| (swizzle >> 4) & 0x3, |
| (swizzle >> 6) & 0x3); |
| } |
| |
| |
| struct ureg_src |
| ureg_DECL_immediate( struct ureg_program *ureg, |
| const float *v, |
| unsigned nr ) |
| { |
| union { |
| float f[4]; |
| unsigned u[4]; |
| } fu; |
| unsigned int i; |
| |
| for (i = 0; i < nr; i++) { |
| fu.f[i] = v[i]; |
| } |
| |
| return decl_immediate(ureg, fu.u, nr, TGSI_IMM_FLOAT32); |
| } |
| |
| struct ureg_src |
| ureg_DECL_immediate_f64( struct ureg_program *ureg, |
| const double *v, |
| unsigned nr ) |
| { |
| union { |
| unsigned u[4]; |
| double d[2]; |
| } fu; |
| unsigned int i; |
| |
| assert((nr / 2) < 3); |
| for (i = 0; i < nr / 2; i++) { |
| fu.d[i] = v[i]; |
| } |
| |
| return decl_immediate(ureg, fu.u, nr, TGSI_IMM_FLOAT64); |
| } |
| |
| struct ureg_src |
| ureg_DECL_immediate_uint( struct ureg_program *ureg, |
| const unsigned *v, |
| unsigned nr ) |
| { |
| return decl_immediate(ureg, v, nr, TGSI_IMM_UINT32); |
| } |
| |
| |
| struct ureg_src |
| ureg_DECL_immediate_block_uint( struct ureg_program *ureg, |
| const unsigned *v, |
| unsigned nr ) |
| { |
| uint index; |
| uint i; |
| |
| if (ureg->nr_immediates + (nr + 3) / 4 > UREG_MAX_IMMEDIATE) { |
| set_bad(ureg); |
| return ureg_src_register(TGSI_FILE_IMMEDIATE, 0); |
| } |
| |
| index = ureg->nr_immediates; |
| ureg->nr_immediates += (nr + 3) / 4; |
| |
| for (i = index; i < ureg->nr_immediates; i++) { |
| ureg->immediate[i].type = TGSI_IMM_UINT32; |
| ureg->immediate[i].nr = nr > 4 ? 4 : nr; |
| memcpy(ureg->immediate[i].value.u, |
| &v[(i - index) * 4], |
| ureg->immediate[i].nr * sizeof(uint)); |
| nr -= 4; |
| } |
| |
| return ureg_src_register(TGSI_FILE_IMMEDIATE, index); |
| } |
| |
| |
| struct ureg_src |
| ureg_DECL_immediate_int( struct ureg_program *ureg, |
| const int *v, |
| unsigned nr ) |
| { |
| return decl_immediate(ureg, (const unsigned *)v, nr, TGSI_IMM_INT32); |
| } |
| |
| struct ureg_src |
| ureg_DECL_immediate_uint64( struct ureg_program *ureg, |
| const uint64_t *v, |
| unsigned nr ) |
| { |
| union { |
| unsigned u[4]; |
| uint64_t u64[2]; |
| } fu; |
| unsigned int i; |
| |
| assert((nr / 2) < 3); |
| for (i = 0; i < nr / 2; i++) { |
| fu.u64[i] = v[i]; |
| } |
| |
| return decl_immediate(ureg, fu.u, nr, TGSI_IMM_UINT64); |
| } |
| |
| struct ureg_src |
| ureg_DECL_immediate_int64( struct ureg_program *ureg, |
| const int64_t *v, |
| unsigned nr ) |
| { |
| union { |
| unsigned u[4]; |
| int64_t i64[2]; |
| } fu; |
| unsigned int i; |
| |
| assert((nr / 2) < 3); |
| for (i = 0; i < nr / 2; i++) { |
| fu.i64[i] = v[i]; |
| } |
| |
| return decl_immediate(ureg, fu.u, nr, TGSI_IMM_INT64); |
| } |
| |
| void |
| ureg_emit_src( struct ureg_program *ureg, |
| struct ureg_src src ) |
| { |
| unsigned size = 1 + (src.Indirect ? 1 : 0) + |
| (src.Dimension ? (src.DimIndirect ? 2 : 1) : 0); |
| |
| union tgsi_any_token *out = get_tokens( ureg, DOMAIN_INSN, size ); |
| unsigned n = 0; |
| |
| assert(src.File != TGSI_FILE_NULL); |
| assert(src.File < TGSI_FILE_COUNT); |
| |
| out[n].value = 0; |
| out[n].src.File = src.File; |
| out[n].src.SwizzleX = src.SwizzleX; |
| out[n].src.SwizzleY = src.SwizzleY; |
| out[n].src.SwizzleZ = src.SwizzleZ; |
| out[n].src.SwizzleW = src.SwizzleW; |
| out[n].src.Index = src.Index; |
| out[n].src.Negate = src.Negate; |
| out[0].src.Absolute = src.Absolute; |
| n++; |
| |
| if (src.Indirect) { |
| out[0].src.Indirect = 1; |
| out[n].value = 0; |
| out[n].ind.File = src.IndirectFile; |
| out[n].ind.Swizzle = src.IndirectSwizzle; |
| out[n].ind.Index = src.IndirectIndex; |
| if (!ureg->supports_any_inout_decl_range && |
| (src.File == TGSI_FILE_INPUT || src.File == TGSI_FILE_OUTPUT)) |
| out[n].ind.ArrayID = 0; |
| else |
| out[n].ind.ArrayID = src.ArrayID; |
| n++; |
| } |
| |
| if (src.Dimension) { |
| out[0].src.Dimension = 1; |
| out[n].dim.Dimension = 0; |
| out[n].dim.Padding = 0; |
| if (src.DimIndirect) { |
| out[n].dim.Indirect = 1; |
| out[n].dim.Index = src.DimensionIndex; |
| n++; |
| out[n].value = 0; |
| out[n].ind.File = src.DimIndFile; |
| out[n].ind.Swizzle = src.DimIndSwizzle; |
| out[n].ind.Index = src.DimIndIndex; |
| if (!ureg->supports_any_inout_decl_range && |
| (src.File == TGSI_FILE_INPUT || src.File == TGSI_FILE_OUTPUT)) |
| out[n].ind.ArrayID = 0; |
| else |
| out[n].ind.ArrayID = src.ArrayID; |
| } else { |
| out[n].dim.Indirect = 0; |
| out[n].dim.Index = src.DimensionIndex; |
| } |
| n++; |
| } |
| |
| assert(n == size); |
| } |
| |
| |
| void |
| ureg_emit_dst( struct ureg_program *ureg, |
| struct ureg_dst dst ) |
| { |
| unsigned size = 1 + (dst.Indirect ? 1 : 0) + |
| (dst.Dimension ? (dst.DimIndirect ? 2 : 1) : 0); |
| |
| union tgsi_any_token *out = get_tokens( ureg, DOMAIN_INSN, size ); |
| unsigned n = 0; |
| |
| assert(dst.File != TGSI_FILE_NULL); |
| assert(dst.File != TGSI_FILE_SAMPLER); |
| assert(dst.File != TGSI_FILE_SAMPLER_VIEW); |
| assert(dst.File != TGSI_FILE_IMMEDIATE); |
| assert(dst.File < TGSI_FILE_COUNT); |
| |
| out[n].value = 0; |
| out[n].dst.File = dst.File; |
| out[n].dst.WriteMask = dst.WriteMask; |
| out[n].dst.Indirect = dst.Indirect; |
| out[n].dst.Index = dst.Index; |
| n++; |
| |
| if (dst.Indirect) { |
| out[n].value = 0; |
| out[n].ind.File = dst.IndirectFile; |
| out[n].ind.Swizzle = dst.IndirectSwizzle; |
| out[n].ind.Index = dst.IndirectIndex; |
| if (!ureg->supports_any_inout_decl_range && |
| (dst.File == TGSI_FILE_INPUT || dst.File == TGSI_FILE_OUTPUT)) |
| out[n].ind.ArrayID = 0; |
| else |
| out[n].ind.ArrayID = dst.ArrayID; |
| n++; |
| } |
| |
| if (dst.Dimension) { |
| out[0].dst.Dimension = 1; |
| out[n].dim.Dimension = 0; |
| out[n].dim.Padding = 0; |
| if (dst.DimIndirect) { |
| out[n].dim.Indirect = 1; |
| out[n].dim.Index = dst.DimensionIndex; |
| n++; |
| out[n].value = 0; |
| out[n].ind.File = dst.DimIndFile; |
| out[n].ind.Swizzle = dst.DimIndSwizzle; |
| out[n].ind.Index = dst.DimIndIndex; |
| if (!ureg->supports_any_inout_decl_range && |
| (dst.File == TGSI_FILE_INPUT || dst.File == TGSI_FILE_OUTPUT)) |
| out[n].ind.ArrayID = 0; |
| else |
| out[n].ind.ArrayID = dst.ArrayID; |
| } else { |
| out[n].dim.Indirect = 0; |
| out[n].dim.Index = dst.DimensionIndex; |
| } |
| n++; |
| } |
| |
| assert(n == size); |
| } |
| |
| |
| static void validate( enum tgsi_opcode opcode, |
| unsigned nr_dst, |
| unsigned nr_src ) |
| { |
| #ifndef NDEBUG |
| const struct tgsi_opcode_info *info = tgsi_get_opcode_info( opcode ); |
| assert(info); |
| if (info) { |
| assert(nr_dst == info->num_dst); |
| assert(nr_src == info->num_src); |
| } |
| #endif |
| } |
| |
| struct ureg_emit_insn_result |
| ureg_emit_insn(struct ureg_program *ureg, |
| enum tgsi_opcode opcode, |
| boolean saturate, |
| unsigned precise, |
| unsigned num_dst, |
| unsigned num_src) |
| { |
| union tgsi_any_token *out; |
| uint count = 1; |
| struct ureg_emit_insn_result result; |
| |
| validate( opcode, num_dst, num_src ); |
| |
| out = get_tokens( ureg, DOMAIN_INSN, count ); |
| out[0].insn = tgsi_default_instruction(); |
| out[0].insn.Opcode = opcode; |
| out[0].insn.Saturate = saturate; |
| out[0].insn.Precise = precise || ureg->precise; |
| out[0].insn.NumDstRegs = num_dst; |
| out[0].insn.NumSrcRegs = num_src; |
| |
| result.insn_token = ureg->domain[DOMAIN_INSN].count - count; |
| result.extended_token = result.insn_token; |
| |
| ureg->nr_instructions++; |
| |
| return result; |
| } |
| |
| |
| /** |
| * Emit a label token. |
| * \param label_token returns a token number indicating where the label |
| * needs to be patched later. Later, this value should be passed to the |
| * ureg_fixup_label() function. |
| */ |
| void |
| ureg_emit_label(struct ureg_program *ureg, |
| unsigned extended_token, |
| unsigned *label_token ) |
| { |
| union tgsi_any_token *out, *insn; |
| |
| if (!label_token) |
| return; |
| |
| out = get_tokens( ureg, DOMAIN_INSN, 1 ); |
| out[0].value = 0; |
| |
| insn = retrieve_token( ureg, DOMAIN_INSN, extended_token ); |
| insn->insn.Label = 1; |
| |
| *label_token = ureg->domain[DOMAIN_INSN].count - 1; |
| } |
| |
| /* Will return a number which can be used in a label to point to the |
| * next instruction to be emitted. |
| */ |
| unsigned |
| ureg_get_instruction_number( struct ureg_program *ureg ) |
| { |
| return ureg->nr_instructions; |
| } |
| |
| /* Patch a given label (expressed as a token number) to point to a |
| * given instruction (expressed as an instruction number). |
| */ |
| void |
| ureg_fixup_label(struct ureg_program *ureg, |
| unsigned label_token, |
| unsigned instruction_number ) |
| { |
| union tgsi_any_token *out = retrieve_token( ureg, DOMAIN_INSN, label_token ); |
| |
| out->insn_label.Label = instruction_number; |
| } |
| |
| |
| void |
| ureg_emit_texture(struct ureg_program *ureg, |
| unsigned extended_token, |
| enum tgsi_texture_type target, |
| enum tgsi_return_type return_type, unsigned num_offsets) |
| { |
| union tgsi_any_token *out, *insn; |
| |
| out = get_tokens( ureg, DOMAIN_INSN, 1 ); |
| insn = retrieve_token( ureg, DOMAIN_INSN, extended_token ); |
| |
| insn->insn.Texture = 1; |
| |
| out[0].value = 0; |
| out[0].insn_texture.Texture = target; |
| out[0].insn_texture.NumOffsets = num_offsets; |
| out[0].insn_texture.ReturnType = return_type; |
| } |
| |
| void |
| ureg_emit_texture_offset(struct ureg_program *ureg, |
| const struct tgsi_texture_offset *offset) |
| { |
| union tgsi_any_token *out; |
| |
| out = get_tokens( ureg, DOMAIN_INSN, 1); |
| |
| out[0].value = 0; |
| out[0].insn_texture_offset = *offset; |
| } |
| |
| void |
| ureg_emit_memory(struct ureg_program *ureg, |
| unsigned extended_token, |
| unsigned qualifier, |
| enum tgsi_texture_type texture, |
| enum pipe_format format) |
| { |
| union tgsi_any_token *out, *insn; |
| |
| out = get_tokens( ureg, DOMAIN_INSN, 1 ); |
| insn = retrieve_token( ureg, DOMAIN_INSN, extended_token ); |
| |
| insn->insn.Memory = 1; |
| |
| out[0].value = 0; |
| out[0].insn_memory.Qualifier = qualifier; |
| out[0].insn_memory.Texture = texture; |
| out[0].insn_memory.Format = format; |
| } |
| |
| void |
| ureg_fixup_insn_size(struct ureg_program *ureg, |
| unsigned insn ) |
| { |
| union tgsi_any_token *out = retrieve_token( ureg, DOMAIN_INSN, insn ); |
| |
| assert(out->insn.Type == TGSI_TOKEN_TYPE_INSTRUCTION); |
| out->insn.NrTokens = ureg->domain[DOMAIN_INSN].count - insn - 1; |
| } |
| |
| |
| void |
| ureg_insn(struct ureg_program *ureg, |
| enum tgsi_opcode opcode, |
| const struct ureg_dst *dst, |
| unsigned nr_dst, |
| const struct ureg_src *src, |
| unsigned nr_src, |
| unsigned precise ) |
| { |
| struct ureg_emit_insn_result insn; |
| unsigned i; |
| boolean saturate; |
| |
| if (nr_dst && ureg_dst_is_empty(dst[0])) { |
| return; |
| } |
| |
| saturate = nr_dst ? dst[0].Saturate : FALSE; |
| |
| insn = ureg_emit_insn(ureg, |
| opcode, |
| saturate, |
| precise, |
| nr_dst, |
| nr_src); |
| |
| for (i = 0; i < nr_dst; i++) |
| ureg_emit_dst( ureg, dst[i] ); |
| |
| for (i = 0; i < nr_src; i++) |
| ureg_emit_src( ureg, src[i] ); |
| |
| ureg_fixup_insn_size( ureg, insn.insn_token ); |
| } |
| |
| void |
| ureg_tex_insn(struct ureg_program *ureg, |
| enum tgsi_opcode opcode, |
| const struct ureg_dst *dst, |
| unsigned nr_dst, |
| enum tgsi_texture_type target, |
| enum tgsi_return_type return_type, |
| const struct tgsi_texture_offset *texoffsets, |
| unsigned nr_offset, |
| const struct ureg_src *src, |
| unsigned nr_src ) |
| { |
| struct ureg_emit_insn_result insn; |
| unsigned i; |
| boolean saturate; |
| |
| if (nr_dst && ureg_dst_is_empty(dst[0])) { |
| return; |
| } |
| |
| saturate = nr_dst ? dst[0].Saturate : FALSE; |
| |
| insn = ureg_emit_insn(ureg, |
| opcode, |
| saturate, |
| 0, |
| nr_dst, |
| nr_src); |
| |
| ureg_emit_texture( ureg, insn.extended_token, target, return_type, |
| nr_offset ); |
| |
| for (i = 0; i < nr_offset; i++) |
| ureg_emit_texture_offset( ureg, &texoffsets[i]); |
| |
| for (i = 0; i < nr_dst; i++) |
| ureg_emit_dst( ureg, dst[i] ); |
| |
| for (i = 0; i < nr_src; i++) |
| ureg_emit_src( ureg, src[i] ); |
| |
| ureg_fixup_insn_size( ureg, insn.insn_token ); |
| } |
| |
| |
| void |
| ureg_memory_insn(struct ureg_program *ureg, |
| enum tgsi_opcode opcode, |
| const struct ureg_dst *dst, |
| unsigned nr_dst, |
| const struct ureg_src *src, |
| unsigned nr_src, |
| unsigned qualifier, |
| enum tgsi_texture_type texture, |
| enum pipe_format format) |
| { |
| struct ureg_emit_insn_result insn; |
| unsigned i; |
| |
| insn = ureg_emit_insn(ureg, |
| opcode, |
| FALSE, |
| 0, |
| nr_dst, |
| nr_src); |
| |
| ureg_emit_memory(ureg, insn.extended_token, qualifier, texture, format); |
| |
| for (i = 0; i < nr_dst; i++) |
| ureg_emit_dst(ureg, dst[i]); |
| |
| for (i = 0; i < nr_src; i++) |
| ureg_emit_src(ureg, src[i]); |
| |
| ureg_fixup_insn_size(ureg, insn.insn_token); |
| } |
| |
| |
| static void |
| emit_decl_semantic(struct ureg_program *ureg, |
| unsigned file, |
| unsigned first, |
| unsigned last, |
| enum tgsi_semantic semantic_name, |
| unsigned semantic_index, |
| unsigned streams, |
| unsigned usage_mask, |
| unsigned array_id, |
| boolean invariant) |
| { |
| union tgsi_any_token *out = get_tokens(ureg, DOMAIN_DECL, array_id ? 4 : 3); |
| |
| out[0].value = 0; |
| out[0].decl.Type = TGSI_TOKEN_TYPE_DECLARATION; |
| out[0].decl.NrTokens = 3; |
| out[0].decl.File = file; |
| out[0].decl.UsageMask = usage_mask; |
| out[0].decl.Semantic = 1; |
| out[0].decl.Array = array_id != 0; |
| out[0].decl.Invariant = invariant; |
| |
| out[1].value = 0; |
| out[1].decl_range.First = first; |
| out[1].decl_range.Last = last; |
| |
| out[2].value = 0; |
| out[2].decl_semantic.Name = semantic_name; |
| out[2].decl_semantic.Index = semantic_index; |
| out[2].decl_semantic.StreamX = streams & 3; |
| out[2].decl_semantic.StreamY = (streams >> 2) & 3; |
| out[2].decl_semantic.StreamZ = (streams >> 4) & 3; |
| out[2].decl_semantic.StreamW = (streams >> 6) & 3; |
| |
| if (array_id) { |
| out[3].value = 0; |
| out[3].array.ArrayID = array_id; |
| } |
| } |
| |
| static void |
| emit_decl_atomic_2d(struct ureg_program *ureg, |
| unsigned first, |
| unsigned last, |
| unsigned index2D, |
| unsigned array_id) |
| { |
| union tgsi_any_token *out = get_tokens(ureg, DOMAIN_DECL, array_id ? 4 : 3); |
| |
| out[0].value = 0; |
| out[0].decl.Type = TGSI_TOKEN_TYPE_DECLARATION; |
| out[0].decl.NrTokens = 3; |
| out[0].decl.File = TGSI_FILE_HW_ATOMIC; |
| out[0].decl.UsageMask = TGSI_WRITEMASK_XYZW; |
| out[0].decl.Dimension = 1; |
| out[0].decl.Array = array_id != 0; |
| |
| out[1].value = 0; |
| out[1].decl_range.First = first; |
| out[1].decl_range.Last = last; |
| |
| out[2].value = 0; |
| out[2].decl_dim.Index2D = index2D; |
| |
| if (array_id) { |
| out[3].value = 0; |
| out[3].array.ArrayID = array_id; |
| } |
| } |
| |
| static void |
| emit_decl_fs(struct ureg_program *ureg, |
| unsigned file, |
| unsigned first, |
| unsigned last, |
| enum tgsi_semantic semantic_name, |
| unsigned semantic_index, |
| enum tgsi_interpolate_mode interpolate, |
| enum tgsi_interpolate_loc interpolate_location, |
| unsigned array_id, |
| unsigned usage_mask) |
| { |
| union tgsi_any_token *out = get_tokens(ureg, DOMAIN_DECL, |
| array_id ? 5 : 4); |
| |
| out[0].value = 0; |
| out[0].decl.Type = TGSI_TOKEN_TYPE_DECLARATION; |
| out[0].decl.NrTokens = 4; |
| out[0].decl.File = file; |
| out[0].decl.UsageMask = usage_mask; |
| out[0].decl.Interpolate = 1; |
| out[0].decl.Semantic = 1; |
| out[0].decl.Array = array_id != 0; |
| |
| out[1].value = 0; |
| out[1].decl_range.First = first; |
| out[1].decl_range.Last = last; |
| |
| out[2].value = 0; |
| out[2].decl_interp.Interpolate = interpolate; |
| out[2].decl_interp.Location = interpolate_location; |
| |
| out[3].value = 0; |
| out[3].decl_semantic.Name = semantic_name; |
| out[3].decl_semantic.Index = semantic_index; |
| |
| if (array_id) { |
| out[4].value = 0; |
| out[4].array.ArrayID = array_id; |
| } |
| } |
| |
| static void |
| emit_decl_temps( struct ureg_program *ureg, |
| unsigned first, unsigned last, |
| boolean local, |
| unsigned arrayid ) |
| { |
| union tgsi_any_token *out = get_tokens( ureg, DOMAIN_DECL, |
| arrayid ? 3 : 2 ); |
| |
| out[0].value = 0; |
| out[0].decl.Type = TGSI_TOKEN_TYPE_DECLARATION; |
| out[0].decl.NrTokens = 2; |
| out[0].decl.File = TGSI_FILE_TEMPORARY; |
| out[0].decl.UsageMask = TGSI_WRITEMASK_XYZW; |
| out[0].decl.Local = local; |
| |
| out[1].value = 0; |
| out[1].decl_range.First = first; |
| out[1].decl_range.Last = last; |
| |
| if (arrayid) { |
| out[0].decl.Array = 1; |
| out[2].value = 0; |
| out[2].array.ArrayID = arrayid; |
| } |
| } |
| |
| static void emit_decl_range( struct ureg_program *ureg, |
| unsigned file, |
| unsigned first, |
| unsigned count ) |
| { |
| union tgsi_any_token *out = get_tokens( ureg, DOMAIN_DECL, 2 ); |
| |
| out[0].value = 0; |
| out[0].decl.Type = TGSI_TOKEN_TYPE_DECLARATION; |
| out[0].decl.NrTokens = 2; |
| out[0].decl.File = file; |
| out[0].decl.UsageMask = TGSI_WRITEMASK_XYZW; |
| out[0].decl.Semantic = 0; |
| |
| out[1].value = 0; |
| out[1].decl_range.First = first; |
| out[1].decl_range.Last = first + count - 1; |
| } |
| |
| static void |
| emit_decl_range2D(struct ureg_program *ureg, |
| unsigned file, |
| unsigned first, |
| unsigned last, |
| unsigned index2D) |
| { |
| union tgsi_any_token *out = get_tokens(ureg, DOMAIN_DECL, 3); |
| |
| out[0].value = 0; |
| out[0].decl.Type = TGSI_TOKEN_TYPE_DECLARATION; |
| out[0].decl.NrTokens = 3; |
| out[0].decl.File = file; |
| out[0].decl.UsageMask = TGSI_WRITEMASK_XYZW; |
| out[0].decl.Dimension = 1; |
| |
| out[1].value = 0; |
| out[1].decl_range.First = first; |
| out[1].decl_range.Last = last; |
| |
| out[2].value = 0; |
| out[2].decl_dim.Index2D = index2D; |
| } |
| |
| static void |
| emit_decl_sampler_view(struct ureg_program *ureg, |
| unsigned index, |
| enum tgsi_texture_type target, |
| enum tgsi_return_type return_type_x, |
| enum tgsi_return_type return_type_y, |
| enum tgsi_return_type return_type_z, |
| enum tgsi_return_type return_type_w ) |
| { |
| union tgsi_any_token *out = get_tokens(ureg, DOMAIN_DECL, 3); |
| |
| out[0].value = 0; |
| out[0].decl.Type = TGSI_TOKEN_TYPE_DECLARATION; |
| out[0].decl.NrTokens = 3; |
| out[0].decl.File = TGSI_FILE_SAMPLER_VIEW; |
| out[0].decl.UsageMask = TGSI_WRITEMASK_XYZW; |
| |
| out[1].value = 0; |
| out[1].decl_range.First = index; |
| out[1].decl_range.Last = index; |
| |
| out[2].value = 0; |
| out[2].decl_sampler_view.Resource = target; |
| out[2].decl_sampler_view.ReturnTypeX = return_type_x; |
| out[2].decl_sampler_view.ReturnTypeY = return_type_y; |
| out[2].decl_sampler_view.ReturnTypeZ = return_type_z; |
| out[2].decl_sampler_view.ReturnTypeW = return_type_w; |
| } |
| |
| static void |
| emit_decl_image(struct ureg_program *ureg, |
| unsigned index, |
| enum tgsi_texture_type target, |
| enum pipe_format format, |
| boolean wr, |
| boolean raw) |
| { |
| union tgsi_any_token *out = get_tokens(ureg, DOMAIN_DECL, 3); |
| |
| out[0].value = 0; |
| out[0].decl.Type = TGSI_TOKEN_TYPE_DECLARATION; |
| out[0].decl.NrTokens = 3; |
| out[0].decl.File = TGSI_FILE_IMAGE; |
| out[0].decl.UsageMask = TGSI_WRITEMASK_XYZW; |
| |
| out[1].value = 0; |
| out[1].decl_range.First = index; |
| out[1].decl_range.Last = index; |
| |
| out[2].value = 0; |
| out[2].decl_image.Resource = target; |
| out[2].decl_image.Writable = wr; |
| out[2].decl_image.Raw = raw; |
| out[2].decl_image.Format = format; |
| } |
| |
| static void |
| emit_decl_buffer(struct ureg_program *ureg, |
| unsigned index, |
| bool atomic) |
| { |
| union tgsi_any_token *out = get_tokens(ureg, DOMAIN_DECL, 2); |
| |
| out[0].value = 0; |
| out[0].decl.Type = TGSI_TOKEN_TYPE_DECLARATION; |
| out[0].decl.NrTokens = 2; |
| out[0].decl.File = TGSI_FILE_BUFFER; |
| out[0].decl.UsageMask = TGSI_WRITEMASK_XYZW; |
| out[0].decl.Atomic = atomic; |
| |
| out[1].value = 0; |
| out[1].decl_range.First = index; |
| out[1].decl_range.Last = index; |
| } |
| |
| static void |
| emit_decl_memory(struct ureg_program *ureg, unsigned memory_type) |
| { |
| union tgsi_any_token *out = get_tokens(ureg, DOMAIN_DECL, 2); |
| |
| out[0].value = 0; |
| out[0].decl.Type = TGSI_TOKEN_TYPE_DECLARATION; |
| out[0].decl.NrTokens = 2; |
| out[0].decl.File = TGSI_FILE_MEMORY; |
| out[0].decl.UsageMask = TGSI_WRITEMASK_XYZW; |
| out[0].decl.MemType = memory_type; |
| |
| out[1].value = 0; |
| out[1].decl_range.First = memory_type; |
| out[1].decl_range.Last = memory_type; |
| } |
| |
| static void |
| emit_immediate( struct ureg_program *ureg, |
| const unsigned *v, |
| unsigned type ) |
| { |
| union tgsi_any_token *out = get_tokens( ureg, DOMAIN_DECL, 5 ); |
| |
| out[0].value = 0; |
| out[0].imm.Type = TGSI_TOKEN_TYPE_IMMEDIATE; |
| out[0].imm.NrTokens = 5; |
| out[0].imm.DataType = type; |
| out[0].imm.Padding = 0; |
| |
| out[1].imm_data.Uint = v[0]; |
| out[2].imm_data.Uint = v[1]; |
| out[3].imm_data.Uint = v[2]; |
| out[4].imm_data.Uint = v[3]; |
| } |
| |
| static void |
| emit_property(struct ureg_program *ureg, |
| unsigned name, |
| unsigned data) |
| { |
| union tgsi_any_token *out = get_tokens(ureg, DOMAIN_DECL, 2); |
| |
| out[0].value = 0; |
| out[0].prop.Type = TGSI_TOKEN_TYPE_PROPERTY; |
| out[0].prop.NrTokens = 2; |
| out[0].prop.PropertyName = name; |
| |
| out[1].prop_data.Data = data; |
| } |
| |
| static int |
| input_sort(const void *in_a, const void *in_b) |
| { |
| const struct ureg_input_decl *a = in_a, *b = in_b; |
| |
| return a->first - b->first; |
| } |
| |
| static int |
| output_sort(const void *in_a, const void *in_b) |
| { |
| const struct ureg_output_decl *a = in_a, *b = in_b; |
| |
| return a->first - b->first; |
| } |
| |
| static int |
| atomic_decl_range_sort(const void *in_a, const void *in_b) |
| { |
| const struct hw_atomic_decl_range *a = in_a, *b = in_b; |
| |
| return a->first - b->first; |
| } |
| |
| static void emit_decls( struct ureg_program *ureg ) |
| { |
| unsigned i,j; |
| |
| for (i = 0; i < ARRAY_SIZE(ureg->properties); i++) |
| if (ureg->properties[i] != ~0u) |
| emit_property(ureg, i, ureg->properties[i]); |
| |
| /* While not required by TGSI spec, virglrenderer has a dependency on the |
| * inputs being sorted. |
| */ |
| qsort(ureg->input, ureg->nr_inputs, sizeof(ureg->input[0]), input_sort); |
| |
| if (ureg->processor == PIPE_SHADER_VERTEX) { |
| for (i = 0; i < PIPE_MAX_ATTRIBS; i++) { |
| if (ureg->vs_inputs[i/32] & (1u << (i%32))) { |
| emit_decl_range( ureg, TGSI_FILE_INPUT, i, 1 ); |
| } |
| } |
| } else if (ureg->processor == PIPE_SHADER_FRAGMENT) { |
| if (ureg->supports_any_inout_decl_range) { |
| for (i = 0; i < ureg->nr_inputs; i++) { |
| emit_decl_fs(ureg, |
| TGSI_FILE_INPUT, |
| ureg->input[i].first, |
| ureg->input[i].last, |
| ureg->input[i].semantic_name, |
| ureg->input[i].semantic_index, |
| ureg->input[i].interp, |
| ureg->input[i].interp_location, |
| ureg->input[i].array_id, |
| ureg->input[i].usage_mask); |
| } |
| } |
| else { |
| for (i = 0; i < ureg->nr_inputs; i++) { |
| for (j = ureg->input[i].first; j <= ureg->input[i].last; j++) { |
| emit_decl_fs(ureg, |
| TGSI_FILE_INPUT, |
| j, j, |
| ureg->input[i].semantic_name, |
| ureg->input[i].semantic_index + |
| (j - ureg->input[i].first), |
| ureg->input[i].interp, |
| ureg->input[i].interp_location, 0, |
| ureg->input[i].usage_mask); |
| } |
| } |
| } |
| } else { |
| if (ureg->supports_any_inout_decl_range) { |
| for (i = 0; i < ureg->nr_inputs; i++) { |
| emit_decl_semantic(ureg, |
| TGSI_FILE_INPUT, |
| ureg->input[i].first, |
| ureg->input[i].last, |
| ureg->input[i].semantic_name, |
| ureg->input[i].semantic_index, |
| 0, |
| TGSI_WRITEMASK_XYZW, |
| ureg->input[i].array_id, |
| FALSE); |
| } |
| } |
| else { |
| for (i = 0; i < ureg->nr_inputs; i++) { |
| for (j = ureg->input[i].first; j <= ureg->input[i].last; j++) { |
| emit_decl_semantic(ureg, |
| TGSI_FILE_INPUT, |
| j, j, |
| ureg->input[i].semantic_name, |
| ureg->input[i].semantic_index + |
| (j - ureg->input[i].first), |
| 0, |
| TGSI_WRITEMASK_XYZW, 0, FALSE); |
| } |
| } |
| } |
| } |
| |
| for (i = 0; i < ureg->nr_system_values; i++) { |
| emit_decl_semantic(ureg, |
| TGSI_FILE_SYSTEM_VALUE, |
| i, |
| i, |
| ureg->system_value[i].semantic_name, |
| ureg->system_value[i].semantic_index, |
| 0, |
| TGSI_WRITEMASK_XYZW, 0, FALSE); |
| } |
| |
| /* While not required by TGSI spec, virglrenderer has a dependency on the |
| * outputs being sorted. |
| */ |
| qsort(ureg->output, ureg->nr_outputs, sizeof(ureg->output[0]), output_sort); |
| |
| if (ureg->supports_any_inout_decl_range) { |
| for (i = 0; i < ureg->nr_outputs; i++) { |
| emit_decl_semantic(ureg, |
| TGSI_FILE_OUTPUT, |
| ureg->output[i].first, |
| ureg->output[i].last, |
| ureg->output[i].semantic_name, |
| ureg->output[i].semantic_index, |
| ureg->output[i].streams, |
| ureg->output[i].usage_mask, |
| ureg->output[i].array_id, |
| ureg->output[i].invariant); |
| } |
| } |
| else { |
| for (i = 0; i < ureg->nr_outputs; i++) { |
| for (j = ureg->output[i].first; j <= ureg->output[i].last; j++) { |
| emit_decl_semantic(ureg, |
| TGSI_FILE_OUTPUT, |
| j, j, |
| ureg->output[i].semantic_name, |
| ureg->output[i].semantic_index + |
| (j - ureg->output[i].first), |
| ureg->output[i].streams, |
| ureg->output[i].usage_mask, |
| 0, |
| ureg->output[i].invariant); |
| } |
| } |
| } |
| |
| for (i = 0; i < ureg->nr_samplers; i++) { |
| emit_decl_range( ureg, |
| TGSI_FILE_SAMPLER, |
| ureg->sampler[i].Index, 1 ); |
| } |
| |
| for (i = 0; i < ureg->nr_sampler_views; i++) { |
| emit_decl_sampler_view(ureg, |
| ureg->sampler_view[i].index, |
| ureg->sampler_view[i].target, |
| ureg->sampler_view[i].return_type_x, |
| ureg->sampler_view[i].return_type_y, |
| ureg->sampler_view[i].return_type_z, |
| ureg->sampler_view[i].return_type_w); |
| } |
| |
| for (i = 0; i < ureg->nr_images; i++) { |
| emit_decl_image(ureg, |
| ureg->image[i].index, |
| ureg->image[i].target, |
| ureg->image[i].format, |
| ureg->image[i].wr, |
| ureg->image[i].raw); |
| } |
| |
| for (i = 0; i < ureg->nr_buffers; i++) { |
| emit_decl_buffer(ureg, ureg->buffer[i].index, ureg->buffer[i].atomic); |
| } |
| |
| for (i = 0; i < TGSI_MEMORY_TYPE_COUNT; i++) { |
| if (ureg->use_memory[i]) |
| emit_decl_memory(ureg, i); |
| } |
| |
| for (i = 0; i < PIPE_MAX_CONSTANT_BUFFERS; i++) { |
| struct const_decl *decl = &ureg->const_decls[i]; |
| |
| if (decl->nr_constant_ranges) { |
| uint j; |
| |
| for (j = 0; j < decl->nr_constant_ranges; j++) { |
| emit_decl_range2D(ureg, |
| TGSI_FILE_CONSTANT, |
| decl->constant_range[j].first, |
| decl->constant_range[j].last, |
| i); |
| } |
| } |
| } |
| |
| for (i = 0; i < PIPE_MAX_HW_ATOMIC_BUFFERS; i++) { |
| struct hw_atomic_decl *decl = &ureg->hw_atomic_decls[i]; |
| |
| if (decl->nr_hw_atomic_ranges) { |
| uint j; |
| |
| /* GLSL-to-TGSI generated HW atomic counters in order, and r600 depends |
| * on it. |
| */ |
| qsort(decl->hw_atomic_range, decl->nr_hw_atomic_ranges, sizeof(struct hw_atomic_decl_range), atomic_decl_range_sort); |
| |
| for (j = 0; j < decl->nr_hw_atomic_ranges; j++) { |
| emit_decl_atomic_2d(ureg, |
| decl->hw_atomic_range[j].first, |
| decl->hw_atomic_range[j].last, |
| i, |
| decl->hw_atomic_range[j].array_id); |
| } |
| } |
| } |
| |
| if (ureg->nr_temps) { |
| unsigned array = 0; |
| for (i = 0; i < ureg->nr_temps;) { |
| boolean local = util_bitmask_get(ureg->local_temps, i); |
| unsigned first = i; |
| i = util_bitmask_get_next_index(ureg->decl_temps, i + 1); |
| if (i == UTIL_BITMASK_INVALID_INDEX) |
| i = ureg->nr_temps; |
| |
| if (array < ureg->nr_array_temps && ureg->array_temps[array] == first) |
| emit_decl_temps( ureg, first, i - 1, local, ++array ); |
| else |
| emit_decl_temps( ureg, first, i - 1, local, 0 ); |
| } |
| } |
| |
| if (ureg->nr_addrs) { |
| emit_decl_range( ureg, |
| TGSI_FILE_ADDRESS, |
| 0, ureg->nr_addrs ); |
| } |
| |
| for (i = 0; i < ureg->nr_immediates; i++) { |
| emit_immediate( ureg, |
| ureg->immediate[i].value.u, |
| ureg->immediate[i].type ); |
| } |
| } |
| |
| /* Append the instruction tokens onto the declarations to build a |
| * contiguous stream suitable to send to the driver. |
| */ |
| static void copy_instructions( struct ureg_program *ureg ) |
| { |
| unsigned nr_tokens = ureg->domain[DOMAIN_INSN].count; |
| union tgsi_any_token *out = get_tokens( ureg, |
| DOMAIN_DECL, |
| nr_tokens ); |
| |
| memcpy(out, |
| ureg->domain[DOMAIN_INSN].tokens, |
| nr_tokens * sizeof out[0] ); |
| } |
| |
| |
| static void |
| fixup_header_size(struct ureg_program *ureg) |
| { |
| union tgsi_any_token *out = retrieve_token( ureg, DOMAIN_DECL, 0 ); |
| |
| out->header.BodySize = ureg->domain[DOMAIN_DECL].count - 2; |
| } |
| |
| |
| static void |
| emit_header( struct ureg_program *ureg ) |
| { |
| union tgsi_any_token *out = get_tokens( ureg, DOMAIN_DECL, 2 ); |
| |
| out[0].header.HeaderSize = 2; |
| out[0].header.BodySize = 0; |
| |
| out[1].processor.Processor = ureg->processor; |
| out[1].processor.Padding = 0; |
| } |
| |
| |
| const struct tgsi_token *ureg_finalize( struct ureg_program *ureg ) |
| { |
| const struct tgsi_token *tokens; |
| |
| switch (ureg->processor) { |
| case PIPE_SHADER_VERTEX: |
| case PIPE_SHADER_TESS_EVAL: |
| ureg_property(ureg, TGSI_PROPERTY_NEXT_SHADER, |
| ureg->next_shader_processor == -1 ? |
| PIPE_SHADER_FRAGMENT : |
| ureg->next_shader_processor); |
| break; |
| default: |
| ; /* nothing */ |
| } |
| |
| emit_header( ureg ); |
| emit_decls( ureg ); |
| copy_instructions( ureg ); |
| fixup_header_size( ureg ); |
| |
| if (ureg->domain[0].tokens == error_tokens || |
| ureg->domain[1].tokens == error_tokens) { |
| debug_printf("%s: error in generated shader\n", __FUNCTION__); |
| assert(0); |
| return NULL; |
| } |
| |
| tokens = &ureg->domain[DOMAIN_DECL].tokens[0].token; |
| |
| if (0) { |
| debug_printf("%s: emitted shader %d tokens:\n", __FUNCTION__, |
| ureg->domain[DOMAIN_DECL].count); |
| tgsi_dump( tokens, 0 ); |
| } |
| |
| #if DEBUG |
| /* tgsi_sanity doesn't seem to return if there are too many constants. */ |
| bool too_many_constants = false; |
| for (unsigned i = 0; i < ARRAY_SIZE(ureg->const_decls); i++) { |
| for (unsigned j = 0; j < ureg->const_decls[i].nr_constant_ranges; j++) { |
| if (ureg->const_decls[i].constant_range[j].last > 4096) { |
| too_many_constants = true; |
| break; |
| } |
| } |
| } |
| |
| if (tokens && !too_many_constants && !tgsi_sanity_check(tokens)) { |
| debug_printf("tgsi_ureg.c, sanity check failed on generated tokens:\n"); |
| tgsi_dump(tokens, 0); |
| assert(0); |
| } |
| #endif |
| |
| |
| return tokens; |
| } |
| |
| |
| void *ureg_create_shader( struct ureg_program *ureg, |
| struct pipe_context *pipe, |
| const struct pipe_stream_output_info *so ) |
| { |
| struct pipe_shader_state state = {0}; |
| |
| pipe_shader_state_from_tgsi(&state, ureg_finalize(ureg)); |
| if(!state.tokens) |
| return NULL; |
| |
| if (so) |
| state.stream_output = *so; |
| |
| switch (ureg->processor) { |
| case PIPE_SHADER_VERTEX: |
| return pipe->create_vs_state(pipe, &state); |
| case PIPE_SHADER_TESS_CTRL: |
| return pipe->create_tcs_state(pipe, &state); |
| case PIPE_SHADER_TESS_EVAL: |
| return pipe->create_tes_state(pipe, &state); |
| case PIPE_SHADER_GEOMETRY: |
| return pipe->create_gs_state(pipe, &state); |
| case PIPE_SHADER_FRAGMENT: |
| return pipe->create_fs_state(pipe, &state); |
| default: |
| return NULL; |
| } |
| } |
| |
| |
| const struct tgsi_token *ureg_get_tokens( struct ureg_program *ureg, |
| unsigned *nr_tokens ) |
| { |
| const struct tgsi_token *tokens; |
| |
| ureg_finalize(ureg); |
| |
| tokens = &ureg->domain[DOMAIN_DECL].tokens[0].token; |
| |
| if (nr_tokens) |
| *nr_tokens = ureg->domain[DOMAIN_DECL].count; |
| |
| ureg->domain[DOMAIN_DECL].tokens = NULL; |
| ureg->domain[DOMAIN_DECL].size = 0; |
| ureg->domain[DOMAIN_DECL].order = 0; |
| ureg->domain[DOMAIN_DECL].count = 0; |
| |
| return tokens; |
| } |
| |
| |
| void ureg_free_tokens( const struct tgsi_token *tokens ) |
| { |
| FREE((struct tgsi_token *)tokens); |
| } |
| |
| |
| struct ureg_program * |
| ureg_create(enum pipe_shader_type processor) |
| { |
| return ureg_create_with_screen(processor, NULL); |
| } |
| |
| |
| struct ureg_program * |
| ureg_create_with_screen(enum pipe_shader_type processor, |
| struct pipe_screen *screen) |
| { |
| uint i; |
| struct ureg_program *ureg = CALLOC_STRUCT( ureg_program ); |
| if (!ureg) |
| goto no_ureg; |
| |
| ureg->processor = processor; |
| ureg->supports_any_inout_decl_range = |
| screen && |
| screen->get_shader_param(screen, processor, |
| PIPE_SHADER_CAP_TGSI_ANY_INOUT_DECL_RANGE) != 0; |
| ureg->next_shader_processor = -1; |
| |
| for (i = 0; i < ARRAY_SIZE(ureg->properties); i++) |
| ureg->properties[i] = ~0; |
| |
| ureg->free_temps = util_bitmask_create(); |
| if (ureg->free_temps == NULL) |
| goto no_free_temps; |
| |
| ureg->local_temps = util_bitmask_create(); |
| if (ureg->local_temps == NULL) |
| goto no_local_temps; |
| |
| ureg->decl_temps = util_bitmask_create(); |
| if (ureg->decl_temps == NULL) |
| goto no_decl_temps; |
| |
| return ureg; |
| |
| no_decl_temps: |
| util_bitmask_destroy(ureg->local_temps); |
| no_local_temps: |
| util_bitmask_destroy(ureg->free_temps); |
| no_free_temps: |
| FREE(ureg); |
| no_ureg: |
| return NULL; |
| } |
| |
| |
| void |
| ureg_set_next_shader_processor(struct ureg_program *ureg, unsigned processor) |
| { |
| ureg->next_shader_processor = processor; |
| } |
| |
| |
| unsigned |
| ureg_get_nr_outputs( const struct ureg_program *ureg ) |
| { |
| if (!ureg) |
| return 0; |
| return ureg->nr_outputs; |
| } |
| |
| static void |
| ureg_setup_clipdist_info(struct ureg_program *ureg, |
| const struct shader_info *info) |
| { |
| if (info->clip_distance_array_size) |
| ureg_property(ureg, TGSI_PROPERTY_NUM_CLIPDIST_ENABLED, |
| info->clip_distance_array_size); |
| if (info->cull_distance_array_size) |
| ureg_property(ureg, TGSI_PROPERTY_NUM_CULLDIST_ENABLED, |
| info->cull_distance_array_size); |
| } |
| |
| static void |
| ureg_setup_tess_ctrl_shader(struct ureg_program *ureg, |
| const struct shader_info *info) |
| { |
| ureg_property(ureg, TGSI_PROPERTY_TCS_VERTICES_OUT, |
| info->tess.tcs_vertices_out); |
| } |
| |
| static void |
| ureg_setup_tess_eval_shader(struct ureg_program *ureg, |
| const struct shader_info *info) |
| { |
| ureg_property(ureg, TGSI_PROPERTY_TES_PRIM_MODE, u_tess_prim_from_shader(info->tess._primitive_mode)); |
| |
| STATIC_ASSERT((TESS_SPACING_EQUAL + 1) % 3 == PIPE_TESS_SPACING_EQUAL); |
| STATIC_ASSERT((TESS_SPACING_FRACTIONAL_ODD + 1) % 3 == |
| PIPE_TESS_SPACING_FRACTIONAL_ODD); |
| STATIC_ASSERT((TESS_SPACING_FRACTIONAL_EVEN + 1) % 3 == |
| PIPE_TESS_SPACING_FRACTIONAL_EVEN); |
| |
| ureg_property(ureg, TGSI_PROPERTY_TES_SPACING, |
| (info->tess.spacing + 1) % 3); |
| |
| ureg_property(ureg, TGSI_PROPERTY_TES_VERTEX_ORDER_CW, |
| !info->tess.ccw); |
| ureg_property(ureg, TGSI_PROPERTY_TES_POINT_MODE, |
| info->tess.point_mode); |
| } |
| |
| static void |
| ureg_setup_geometry_shader(struct ureg_program *ureg, |
| const struct shader_info *info) |
| { |
| ureg_property(ureg, TGSI_PROPERTY_GS_INPUT_PRIM, |
| info->gs.input_primitive); |
| ureg_property(ureg, TGSI_PROPERTY_GS_OUTPUT_PRIM, |
| info->gs.output_primitive); |
| ureg_property(ureg, TGSI_PROPERTY_GS_MAX_OUTPUT_VERTICES, |
| info->gs.vertices_out); |
| ureg_property(ureg, TGSI_PROPERTY_GS_INVOCATIONS, |
| info->gs.invocations); |
| } |
| |
| static void |
| ureg_setup_fragment_shader(struct ureg_program *ureg, |
| const struct shader_info *info) |
| { |
| if (info->fs.early_fragment_tests || info->fs.post_depth_coverage) { |
| ureg_property(ureg, TGSI_PROPERTY_FS_EARLY_DEPTH_STENCIL, 1); |
| |
| if (info->fs.post_depth_coverage) |
| ureg_property(ureg, TGSI_PROPERTY_FS_POST_DEPTH_COVERAGE, 1); |
| } |
| |
| if (info->fs.depth_layout != FRAG_DEPTH_LAYOUT_NONE) { |
| switch (info->fs.depth_layout) { |
| case FRAG_DEPTH_LAYOUT_ANY: |
| ureg_property(ureg, TGSI_PROPERTY_FS_DEPTH_LAYOUT, |
| TGSI_FS_DEPTH_LAYOUT_ANY); |
| break; |
| case FRAG_DEPTH_LAYOUT_GREATER: |
| ureg_property(ureg, TGSI_PROPERTY_FS_DEPTH_LAYOUT, |
| TGSI_FS_DEPTH_LAYOUT_GREATER); |
| break; |
| case FRAG_DEPTH_LAYOUT_LESS: |
| ureg_property(ureg, TGSI_PROPERTY_FS_DEPTH_LAYOUT, |
| TGSI_FS_DEPTH_LAYOUT_LESS); |
| break; |
| case FRAG_DEPTH_LAYOUT_UNCHANGED: |
| ureg_property(ureg, TGSI_PROPERTY_FS_DEPTH_LAYOUT, |
| TGSI_FS_DEPTH_LAYOUT_UNCHANGED); |
| break; |
| default: |
| assert(0); |
| } |
| } |
| |
| if (info->fs.advanced_blend_modes) { |
| ureg_property(ureg, TGSI_PROPERTY_FS_BLEND_EQUATION_ADVANCED, |
| info->fs.advanced_blend_modes); |
| } |
| } |
| |
| static void |
| ureg_setup_compute_shader(struct ureg_program *ureg, |
| const struct shader_info *info) |
| { |
| ureg_property(ureg, TGSI_PROPERTY_CS_FIXED_BLOCK_WIDTH, |
| info->workgroup_size[0]); |
| ureg_property(ureg, TGSI_PROPERTY_CS_FIXED_BLOCK_HEIGHT, |
| info->workgroup_size[1]); |
| ureg_property(ureg, TGSI_PROPERTY_CS_FIXED_BLOCK_DEPTH, |
| info->workgroup_size[2]); |
| |
| if (info->shared_size) |
| ureg_DECL_memory(ureg, TGSI_MEMORY_TYPE_SHARED); |
| } |
| |
| void |
| ureg_setup_shader_info(struct ureg_program *ureg, |
| const struct shader_info *info) |
| { |
| if (info->layer_viewport_relative) |
| ureg_property(ureg, TGSI_PROPERTY_LAYER_VIEWPORT_RELATIVE, 1); |
| |
| switch (info->stage) { |
| case MESA_SHADER_VERTEX: |
| ureg_setup_clipdist_info(ureg, info); |
| ureg_set_next_shader_processor(ureg, pipe_shader_type_from_mesa(info->next_stage)); |
| break; |
| case MESA_SHADER_TESS_CTRL: |
| ureg_setup_tess_ctrl_shader(ureg, info); |
| break; |
| case MESA_SHADER_TESS_EVAL: |
| ureg_setup_tess_eval_shader(ureg, info); |
| ureg_setup_clipdist_info(ureg, info); |
| ureg_set_next_shader_processor(ureg, pipe_shader_type_from_mesa(info->next_stage)); |
| break; |
| case MESA_SHADER_GEOMETRY: |
| ureg_setup_geometry_shader(ureg, info); |
| ureg_setup_clipdist_info(ureg, info); |
| break; |
| case MESA_SHADER_FRAGMENT: |
| ureg_setup_fragment_shader(ureg, info); |
| break; |
| case MESA_SHADER_COMPUTE: |
| ureg_setup_compute_shader(ureg, info); |
| break; |
| default: |
| break; |
| } |
| } |
| |
| |
| void ureg_destroy( struct ureg_program *ureg ) |
| { |
| unsigned i; |
| |
| for (i = 0; i < ARRAY_SIZE(ureg->domain); i++) { |
| if (ureg->domain[i].tokens && |
| ureg->domain[i].tokens != error_tokens) |
| FREE(ureg->domain[i].tokens); |
| } |
| |
| util_bitmask_destroy(ureg->free_temps); |
| util_bitmask_destroy(ureg->local_temps); |
| util_bitmask_destroy(ureg->decl_temps); |
| |
| FREE(ureg); |
| } |
| |
| void ureg_set_precise( struct ureg_program *ureg, bool precise ) |
| { |
| ureg->precise = precise; |
| } |