blob: 2df7c53a8f0c8d9dc0d970ab53ceee0e2ca97990 [file] [log] [blame]
/**************************************************************************
*
* Copyright 2010 VMware, Inc.
* All Rights Reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the
* "Software"), to deal in the Software without restriction, including
* without limitation the rights to use, copy, modify, merge, publish,
* distribute, sub license, and/or sell copies of the Software, and to
* permit persons to whom the Software is furnished to do so, subject to
* the following conditions:
*
* The above copyright notice and this permission notice (including the
* next paragraph) shall be included in all copies or substantial portions
* of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
* OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
* IN NO EVENT SHALL VMWARE AND/OR ITS SUPPLIERS BE LIABLE FOR
* ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
*
**************************************************************************/
#include "util/u_math.h"
#include "util/u_memory.h"
#include "util/u_prim.h"
#include "draw/draw_context.h"
#include "draw/draw_gs.h"
#include "draw/draw_tess.h"
#include "draw/draw_vbuf.h"
#include "draw/draw_vertex.h"
#include "draw/draw_pt.h"
#include "draw/draw_prim_assembler.h"
#include "draw/draw_vs.h"
#include "draw/draw_llvm.h"
#include "gallivm/lp_bld_init.h"
#include "gallivm/lp_bld_debug.h"
struct llvm_middle_end {
struct draw_pt_middle_end base;
struct draw_context *draw;
struct pt_emit *emit;
struct pt_so_emit *so_emit;
struct pt_fetch *fetch;
struct pt_post_vs *post_vs;
unsigned vertex_data_offset;
unsigned vertex_size;
enum pipe_prim_type input_prim;
unsigned opt;
struct draw_llvm *llvm;
struct draw_llvm_variant *current_variant;
};
/** cast wrapper */
static inline struct llvm_middle_end *
llvm_middle_end(struct draw_pt_middle_end *middle)
{
return (struct llvm_middle_end *) middle;
}
static void
llvm_middle_end_prepare_gs(struct llvm_middle_end *fpme)
{
struct draw_context *draw = fpme->draw;
struct draw_llvm *llvm = fpme->llvm;
struct draw_geometry_shader *gs = draw->gs.geometry_shader;
struct draw_gs_llvm_variant_list_item *li;
struct llvm_geometry_shader *shader = llvm_geometry_shader(gs);
char store[DRAW_GS_LLVM_MAX_VARIANT_KEY_SIZE];
struct draw_gs_llvm_variant_key *key = draw_gs_llvm_make_variant_key(llvm, store);
/* Search shader's list of variants for the key */
struct draw_gs_llvm_variant *variant = NULL;
LIST_FOR_EACH_ENTRY(li, &shader->variants.list, list) {
if (memcmp(&li->base->key, key, shader->variant_key_size) == 0) {
variant = li->base;
break;
}
}
if (variant) {
/* found the variant, move to head of global list (for LRU) */
list_move_to(&variant->list_item_global.list, &llvm->gs_variants_list.list);
} else {
/* Need to create new variant */
/* First check if we've created too many variants. If so, free
* 3.125% of the LRU to avoid using too much memory.
*/
if (llvm->nr_gs_variants >= DRAW_MAX_SHADER_VARIANTS) {
if (gallivm_debug & GALLIVM_DEBUG_PERF) {
debug_printf("Evicting GS: %u gs variants,\t%u total variants\n",
shader->variants_cached, llvm->nr_gs_variants);
}
/*
* XXX: should we flush here ?
*/
struct draw_gs_llvm_variant_list_item *item;
for (unsigned i = 0; i < DRAW_MAX_SHADER_VARIANTS / 32; i++) {
if (list_is_empty(&llvm->gs_variants_list.list)) {
break;
}
item = list_last_entry(&llvm->gs_variants_list.list,
struct draw_gs_llvm_variant_list_item, list);
assert(item);
assert(item->base);
draw_gs_llvm_destroy_variant(item->base);
}
}
variant = draw_gs_llvm_create_variant(llvm, gs->info.num_outputs, key);
if (variant) {
list_add(&variant->list_item_local.list, &shader->variants.list);
list_add(&variant->list_item_global.list, &llvm->gs_variants_list.list);
llvm->nr_gs_variants++;
shader->variants_cached++;
}
}
gs->current_variant = variant;
}
static void
llvm_middle_end_prepare_tcs(struct llvm_middle_end *fpme)
{
struct draw_context *draw = fpme->draw;
struct draw_llvm *llvm = fpme->llvm;
struct draw_tess_ctrl_shader *tcs = draw->tcs.tess_ctrl_shader;
struct draw_tcs_llvm_variant_list_item *li;
struct llvm_tess_ctrl_shader *shader = llvm_tess_ctrl_shader(tcs);
char store[DRAW_TCS_LLVM_MAX_VARIANT_KEY_SIZE];
const struct draw_tcs_llvm_variant_key *key =
draw_tcs_llvm_make_variant_key(llvm, store);
/* Search shader's list of variants for the key */
struct draw_tcs_llvm_variant *variant = NULL;
LIST_FOR_EACH_ENTRY(li, &shader->variants.list, list) {
if (memcmp(&li->base->key, key, shader->variant_key_size) == 0) {
variant = li->base;
break;
}
}
if (variant) {
/* found the variant, move to head of global list (for LRU) */
list_move_to(&variant->list_item_global.list,
&llvm->tcs_variants_list.list);
} else {
/* Need to create new variant */
/* First check if we've created too many variants. If so, free
* 3.125% of the LRU to avoid using too much memory.
*/
if (llvm->nr_tcs_variants >= DRAW_MAX_SHADER_VARIANTS) {
if (gallivm_debug & GALLIVM_DEBUG_PERF) {
debug_printf("Evicting TCS: %u tcs variants,\t%u total variants\n",
shader->variants_cached, llvm->nr_tcs_variants);
}
/*
* XXX: should we flush here ?
*/
for (unsigned i = 0; i < DRAW_MAX_SHADER_VARIANTS / 32; i++) {
struct draw_tcs_llvm_variant_list_item *item;
if (list_is_empty(&llvm->tcs_variants_list.list)) {
break;
}
item = list_last_entry(&llvm->tcs_variants_list.list,
struct draw_tcs_llvm_variant_list_item, list);
assert(item);
assert(item->base);
draw_tcs_llvm_destroy_variant(item->base);
}
}
variant = draw_tcs_llvm_create_variant(llvm, 0, key);
if (variant) {
list_add(&variant->list_item_local.list, &shader->variants.list);
list_add(&variant->list_item_global.list, &llvm->tcs_variants_list.list);
llvm->nr_tcs_variants++;
shader->variants_cached++;
}
}
tcs->current_variant = variant;
}
static void
llvm_middle_end_prepare_tes(struct llvm_middle_end *fpme)
{
struct draw_context *draw = fpme->draw;
struct draw_llvm *llvm = fpme->llvm;
struct draw_tess_eval_shader *tes = draw->tes.tess_eval_shader;
struct draw_tes_llvm_variant *variant = NULL;
struct draw_tes_llvm_variant_list_item *li;
struct llvm_tess_eval_shader *shader = llvm_tess_eval_shader(tes);
char store[DRAW_TES_LLVM_MAX_VARIANT_KEY_SIZE];
const struct draw_tes_llvm_variant_key *key =
draw_tes_llvm_make_variant_key(llvm, store);
/* Search shader's list of variants for the key */
LIST_FOR_EACH_ENTRY(li, &shader->variants.list, list) {
if (memcmp(&li->base->key, key, shader->variant_key_size) == 0) {
variant = li->base;
break;
}
}
if (variant) {
/* found the variant, move to head of global list (for LRU) */
list_move_to(&variant->list_item_global.list,
&llvm->tes_variants_list.list);
} else {
/* Need to create new variant */
/* First check if we've created too many variants. If so, free
* 3.125% of the LRU to avoid using too much memory.
*/
if (llvm->nr_tes_variants >= DRAW_MAX_SHADER_VARIANTS) {
if (gallivm_debug & GALLIVM_DEBUG_PERF) {
debug_printf("Evicting TES: %u tes variants,\t%u total variants\n",
shader->variants_cached, llvm->nr_tes_variants);
}
/*
* XXX: should we flush here ?
*/
for (unsigned i = 0; i < DRAW_MAX_SHADER_VARIANTS / 32; i++) {
struct draw_tes_llvm_variant_list_item *item;
if (list_is_empty(&llvm->tes_variants_list.list)) {
break;
}
item = list_last_entry(&llvm->tes_variants_list.list,
struct draw_tes_llvm_variant_list_item, list);
assert(item);
assert(item->base);
draw_tes_llvm_destroy_variant(item->base);
}
}
variant = draw_tes_llvm_create_variant(llvm, draw_total_tes_outputs(draw), key);
if (variant) {
list_add(&variant->list_item_local.list, &shader->variants.list);
list_add(&variant->list_item_global.list, &llvm->tes_variants_list.list);
llvm->nr_tes_variants++;
shader->variants_cached++;
}
}
tes->current_variant = variant;
}
/**
* Prepare/validate middle part of the vertex pipeline.
* NOTE: if you change this function, also look at the non-LLVM
* function fetch_pipeline_prepare() for similar changes.
*/
static void
llvm_middle_end_prepare(struct draw_pt_middle_end *middle,
enum pipe_prim_type in_prim,
unsigned opt,
unsigned *max_vertices)
{
struct llvm_middle_end *fpme = llvm_middle_end(middle);
struct draw_context *draw = fpme->draw;
struct draw_llvm *llvm = fpme->llvm;
struct draw_vertex_shader *vs = draw->vs.vertex_shader;
struct draw_geometry_shader *gs = draw->gs.geometry_shader;
struct draw_tess_ctrl_shader *tcs = draw->tcs.tess_ctrl_shader;
struct draw_tess_eval_shader *tes = draw->tes.tess_eval_shader;
const enum pipe_prim_type out_prim =
gs ? gs->output_primitive : tes ? get_tes_output_prim(tes) :
u_assembled_prim(in_prim);
unsigned point_clip = draw->rasterizer->fill_front == PIPE_POLYGON_MODE_POINT ||
out_prim == PIPE_PRIM_POINTS;
fpme->input_prim = in_prim;
fpme->opt = opt;
draw_pt_post_vs_prepare(fpme->post_vs,
draw->clip_xy,
draw->clip_z,
draw->clip_user,
point_clip ? draw->guard_band_points_xy :
draw->guard_band_xy,
draw->bypass_viewport,
draw->rasterizer->clip_halfz,
(draw->vs.edgeflag_output ? TRUE : FALSE));
draw_pt_so_emit_prepare(fpme->so_emit, (gs == NULL && tes == NULL));
if (!(opt & PT_PIPELINE)) {
draw_pt_emit_prepare(fpme->emit, out_prim, max_vertices);
*max_vertices = MAX2(*max_vertices, 4096);
} else {
/* limit max fetches by limiting max_vertices */
*max_vertices = 4096;
}
/* Get the number of float[4] attributes per vertex.
* Note: this must be done after draw_pt_emit_prepare() since that
* can effect the vertex size.
*/
unsigned nr = MAX2(vs->info.num_inputs, draw_total_vs_outputs(draw));
/* Always leave room for the vertex header whether we need it or
* not. It's hard to get rid of it in particular because of the
* viewport code in draw_pt_post_vs.c.
*/
fpme->vertex_size = sizeof(struct vertex_header) + nr * 4 * sizeof(float);
/* return even number */
*max_vertices = *max_vertices & ~1;
/* Find/create the vertex shader variant */
{
struct draw_llvm_variant *variant = NULL;
struct draw_llvm_variant_list_item *li;
struct llvm_vertex_shader *shader = llvm_vertex_shader(vs);
char store[DRAW_LLVM_MAX_VARIANT_KEY_SIZE];
struct draw_llvm_variant_key *key = draw_llvm_make_variant_key(llvm, store);
/* Search shader's list of variants for the key */
LIST_FOR_EACH_ENTRY(li, &shader->variants.list, list) {
if (memcmp(&li->base->key, key, shader->variant_key_size) == 0) {
variant = li->base;
break;
}
}
if (variant) {
/* found the variant, move to head of global list (for LRU) */
list_move_to(&variant->list_item_global.list, &llvm->vs_variants_list.list);
} else {
/* Need to create new variant */
/* First check if we've created too many variants. If so, free
* 3.125% of the LRU to avoid using too much memory.
*/
if (llvm->nr_variants >= DRAW_MAX_SHADER_VARIANTS) {
if (gallivm_debug & GALLIVM_DEBUG_PERF) {
debug_printf("Evicting VS: %u vs variants,\t%u total variants\n",
shader->variants_cached, llvm->nr_variants);
}
/*
* XXX: should we flush here ?
*/
for (unsigned i = 0; i < DRAW_MAX_SHADER_VARIANTS / 32; i++) {
struct draw_llvm_variant_list_item *item;
if (list_is_empty(&llvm->vs_variants_list.list)) {
break;
}
item = list_last_entry(&llvm->vs_variants_list.list,
struct draw_llvm_variant_list_item, list);
assert(item);
assert(item->base);
draw_llvm_destroy_variant(item->base);
}
}
variant = draw_llvm_create_variant(llvm, nr, key);
if (variant) {
list_add(&variant->list_item_local.list, &shader->variants.list);
list_add(&variant->list_item_global.list, &llvm->vs_variants_list.list);
llvm->nr_variants++;
shader->variants_cached++;
}
}
fpme->current_variant = variant;
}
if (gs) {
llvm_middle_end_prepare_gs(fpme);
}
if (tcs) {
llvm_middle_end_prepare_tcs(fpme);
}
if (tes) {
llvm_middle_end_prepare_tes(fpme);
}
}
static unsigned
get_num_consts_robust(struct draw_context *draw, unsigned *sizes, unsigned idx)
{
unsigned const_bytes = sizes[idx];
if (const_bytes < sizeof(float))
return 0;
return DIV_ROUND_UP(const_bytes, draw->constant_buffer_stride);
}
/**
* Bind/update constant buffer pointers, clip planes and viewport dims.
* These are "light weight" parameters which aren't baked into the
* generated code. Updating these items is much cheaper than revalidating
* and rebuilding the generated pipeline code.
*/
static void
llvm_middle_end_bind_parameters(struct draw_pt_middle_end *middle)
{
static const float fake_const_buf[4];
struct llvm_middle_end *fpme = llvm_middle_end(middle);
struct draw_context *draw = fpme->draw;
struct draw_llvm *llvm = fpme->llvm;
unsigned i;
for (i = 0; i < ARRAY_SIZE(llvm->jit_context.constants); ++i) {
/*
* There could be a potential issue with rounding this up, as the
* shader expects 16-byte allocations, the fix is likely to move
* to LOAD intrinsic in the future and remove the vec4 constraint.
*/
int num_consts = get_num_consts_robust(draw, draw->pt.user.vs_constants_size, i);
llvm->jit_context.constants[i].f = draw->pt.user.vs_constants[i];
llvm->jit_context.constants[i].num_elements = num_consts;
if (num_consts == 0) {
llvm->jit_context.constants[i].f = fake_const_buf;
}
}
for (i = 0; i < ARRAY_SIZE(llvm->jit_context.ssbos); ++i) {
int num_ssbos = draw->pt.user.vs_ssbos_size[i];
llvm->jit_context.ssbos[i].u = draw->pt.user.vs_ssbos[i];
llvm->jit_context.ssbos[i].num_elements = num_ssbos;
if (num_ssbos == 0) {
llvm->jit_context.ssbos[i].u = (const uint32_t *)fake_const_buf;
}
}
for (i = 0; i < ARRAY_SIZE(llvm->gs_jit_context.constants); ++i) {
int num_consts = get_num_consts_robust(draw, draw->pt.user.gs_constants_size, i);
llvm->gs_jit_context.constants[i].f = draw->pt.user.gs_constants[i];
llvm->gs_jit_context.constants[i].num_elements = num_consts;
if (num_consts == 0) {
llvm->gs_jit_context.constants[i].f = fake_const_buf;
}
}
for (i = 0; i < ARRAY_SIZE(llvm->gs_jit_context.ssbos); ++i) {
int num_ssbos = draw->pt.user.gs_ssbos_size[i];
llvm->gs_jit_context.ssbos[i].u = draw->pt.user.gs_ssbos[i];
llvm->gs_jit_context.ssbos[i].num_elements = num_ssbos;
if (num_ssbos == 0) {
llvm->gs_jit_context.ssbos[i].u = (const uint32_t *)fake_const_buf;
}
}
for (i = 0; i < ARRAY_SIZE(llvm->tcs_jit_context.constants); ++i) {
int num_consts = get_num_consts_robust(draw, draw->pt.user.tcs_constants_size, i);
llvm->tcs_jit_context.constants[i].f = draw->pt.user.tcs_constants[i];
llvm->tcs_jit_context.constants[i].num_elements = num_consts;
if (num_consts == 0) {
llvm->tcs_jit_context.constants[i].f = fake_const_buf;
}
}
for (i = 0; i < ARRAY_SIZE(llvm->tcs_jit_context.ssbos); ++i) {
int num_ssbos = draw->pt.user.tcs_ssbos_size[i];
llvm->tcs_jit_context.ssbos[i].u = draw->pt.user.tcs_ssbos[i];
llvm->tcs_jit_context.ssbos[i].num_elements = num_ssbos;
if (num_ssbos == 0) {
llvm->tcs_jit_context.ssbos[i].u = (const uint32_t *)fake_const_buf;
}
}
for (i = 0; i < ARRAY_SIZE(llvm->tes_jit_context.constants); ++i) {
int num_consts = get_num_consts_robust(draw, draw->pt.user.tes_constants_size, i);
llvm->tes_jit_context.constants[i].f = draw->pt.user.tes_constants[i];
llvm->tes_jit_context.constants[i].num_elements = num_consts;
if (num_consts == 0) {
llvm->tes_jit_context.constants[i].f = fake_const_buf;
}
}
for (i = 0; i < ARRAY_SIZE(llvm->tes_jit_context.ssbos); ++i) {
int num_ssbos = draw->pt.user.tes_ssbos_size[i];
llvm->tes_jit_context.ssbos[i].u = draw->pt.user.tes_ssbos[i];
llvm->tes_jit_context.ssbos[i].num_elements = num_ssbos;
if (num_ssbos == 0) {
llvm->tes_jit_context.ssbos[i].u = (const uint32_t *)fake_const_buf;
}
}
llvm->jit_context.planes =
(float (*)[DRAW_TOTAL_CLIP_PLANES][4]) draw->pt.user.planes[0];
llvm->gs_jit_context.planes =
(float (*)[DRAW_TOTAL_CLIP_PLANES][4]) draw->pt.user.planes[0];
llvm->jit_context.viewports = draw->viewports;
llvm->gs_jit_context.viewports = draw->viewports;
llvm->jit_context.aniso_filter_table = lp_build_sample_aniso_filter_table();
llvm->gs_jit_context.aniso_filter_table = lp_build_sample_aniso_filter_table();
llvm->tcs_jit_context.aniso_filter_table = lp_build_sample_aniso_filter_table();
llvm->tes_jit_context.aniso_filter_table = lp_build_sample_aniso_filter_table();
}
static void
pipeline(struct llvm_middle_end *llvm,
const struct draw_vertex_info *vert_info,
const struct draw_prim_info *prim_info)
{
if (prim_info->linear)
draw_pipeline_run_linear(llvm->draw, vert_info, prim_info);
else
draw_pipeline_run(llvm->draw, vert_info, prim_info);
}
static void
emit(struct pt_emit *emit,
const struct draw_vertex_info *vert_info,
const struct draw_prim_info *prim_info)
{
if (prim_info->linear)
draw_pt_emit_linear(emit, vert_info, prim_info);
else
draw_pt_emit(emit, vert_info, prim_info);
}
static void
llvm_pipeline_generic(struct draw_pt_middle_end *middle,
const struct draw_fetch_info *fetch_info,
const struct draw_prim_info *in_prim_info)
{
struct llvm_middle_end *fpme = llvm_middle_end(middle);
struct draw_context *draw = fpme->draw;
struct draw_geometry_shader *gshader = draw->gs.geometry_shader;
struct draw_tess_ctrl_shader *tcs_shader = draw->tcs.tess_ctrl_shader;
struct draw_tess_eval_shader *tes_shader = draw->tes.tess_eval_shader;
struct draw_prim_info tcs_prim_info;
struct draw_prim_info tes_prim_info;
struct draw_prim_info gs_prim_info[TGSI_MAX_VERTEX_STREAMS];
struct draw_vertex_info llvm_vert_info;
struct draw_vertex_info tcs_vert_info;
struct draw_vertex_info tes_vert_info;
struct draw_vertex_info *vert_info;
struct draw_prim_info ia_prim_info;
struct draw_vertex_info ia_vert_info;
const struct draw_prim_info *prim_info = in_prim_info;
boolean free_prim_info = FALSE;
unsigned opt = fpme->opt;
boolean clipped = 0;
ushort *tes_elts_out = NULL;
assert(fetch_info->count > 0);
llvm_vert_info.count = fetch_info->count;
llvm_vert_info.vertex_size = fpme->vertex_size;
llvm_vert_info.stride = fpme->vertex_size;
llvm_vert_info.verts = (struct vertex_header *)
MALLOC(fpme->vertex_size *
align(fetch_info->count, lp_native_vector_width / 32) +
DRAW_EXTRA_VERTICES_PADDING);
if (!llvm_vert_info.verts) {
assert(0);
return;
}
if (draw->collect_statistics) {
draw->statistics.ia_vertices += prim_info->count;
if (prim_info->prim == PIPE_PRIM_PATCHES)
draw->statistics.ia_primitives +=
prim_info->count / draw->pt.vertices_per_patch;
else
draw->statistics.ia_primitives +=
u_decomposed_prims_for_vertices(prim_info->prim, prim_info->count);
draw->statistics.vs_invocations += fetch_info->count;
}
{
unsigned start, vertex_id_offset;
const unsigned *elts;
if (fetch_info->linear) {
start = fetch_info->start;
vertex_id_offset = draw->start_index;
elts = NULL;
} else {
start = draw->pt.user.eltMax;
vertex_id_offset = draw->pt.user.eltBias;
elts = fetch_info->elts;
}
/* Run vertex fetch shader */
clipped = fpme->current_variant->jit_func(&fpme->llvm->jit_context,
llvm_vert_info.verts,
draw->pt.user.vbuffer,
fetch_info->count,
start,
fpme->vertex_size,
draw->pt.vertex_buffer,
draw->instance_id,
vertex_id_offset,
draw->start_instance,
elts,
draw->pt.user.drawid,
draw->pt.user.viewid);
/* Finished with fetch and vs */
fetch_info = NULL;
vert_info = &llvm_vert_info;
}
if (opt & PT_SHADE) {
struct draw_vertex_shader *vshader = draw->vs.vertex_shader;
if (tcs_shader) {
draw_tess_ctrl_shader_run(tcs_shader,
draw->pt.user.tcs_constants,
draw->pt.user.tcs_constants_size,
vert_info,
prim_info,
&vshader->info,
&tcs_vert_info,
&tcs_prim_info);
FREE(vert_info->verts);
vert_info = &tcs_vert_info;
prim_info = &tcs_prim_info;
} else if (tes_shader) {
unsigned num_prims = prim_info->count / draw->pt.vertices_per_patch;
tcs_prim_info = *prim_info;
tcs_prim_info.primitive_count = num_prims;
prim_info = &tcs_prim_info;
}
if (tes_shader) {
draw_tess_eval_shader_run(tes_shader,
draw->pt.user.tes_constants,
draw->pt.user.tes_constants_size,
tcs_shader ? tcs_shader->vertices_out : draw->pt.vertices_per_patch,
vert_info,
prim_info,
tcs_shader ? &tcs_shader->info : &vshader->info,
&tes_vert_info,
&tes_prim_info, &tes_elts_out);
FREE(vert_info->verts);
vert_info = &tes_vert_info;
prim_info = &tes_prim_info;
free_prim_info = TRUE;
/*
* pt emit can only handle ushort number of vertices (see
* render->allocate_vertices).
* vsplit guarantees there's never more than 4096, however GS can
* easily blow this up (by a factor of 256 (or even 1024) max).
*/
if (vert_info->count > 65535) {
opt |= PT_PIPELINE;
}
}
}
struct draw_vertex_info gs_vert_info[TGSI_MAX_VERTEX_STREAMS];
memset(&gs_vert_info, 0, sizeof(gs_vert_info));
if ((opt & PT_SHADE) && gshader) {
struct draw_vertex_shader *vshader = draw->vs.vertex_shader;
draw_geometry_shader_run(gshader,
draw->pt.user.gs_constants,
draw->pt.user.gs_constants_size,
vert_info,
prim_info,
tes_shader ? &tes_shader->info : &vshader->info,
gs_vert_info,
gs_prim_info);
FREE(vert_info->verts);
if (free_prim_info) {
FREE(prim_info->primitive_lengths);
FREE(tes_elts_out);
}
vert_info = &gs_vert_info[0];
prim_info = &gs_prim_info[0];
free_prim_info = FALSE;
/*
* pt emit can only handle ushort number of vertices (see
* render->allocate_vertices).
* vsplit guarantees there's never more than 4096, however GS can
* easily blow this up (by a factor of 256 (or even 1024) max).
*/
if (vert_info->count > 65535) {
opt |= PT_PIPELINE;
}
} else {
if (!tes_shader &&
draw_prim_assembler_is_required(draw, prim_info, vert_info)) {
draw_prim_assembler_run(draw, prim_info, vert_info,
&ia_prim_info, &ia_vert_info);
if (ia_vert_info.count) {
FREE(vert_info->verts);
if (free_prim_info) {
FREE(prim_info->primitive_lengths);
FREE(tes_elts_out);
tes_elts_out = NULL;
}
vert_info = &ia_vert_info;
prim_info = &ia_prim_info;
free_prim_info = TRUE;
}
}
}
/* stream output needs to be done before clipping */
draw_pt_so_emit(fpme->so_emit,
gshader ? gshader->num_vertex_streams : 1,
vert_info, prim_info);
if (prim_info->count == 0) {
debug_printf("GS/IA didn't emit any vertices!\n");
} else {
draw_stats_clipper_primitives(draw, prim_info);
/*
* if there's no position, need to stop now, or the latter stages
* will try to access non-existent position output.
*/
if (draw_current_shader_position_output(draw) != -1) {
if ((opt & PT_SHADE) &&
(gshader || tes_shader ||
draw->vs.vertex_shader->info.writes_viewport_index)) {
clipped = draw_pt_post_vs_run(fpme->post_vs, vert_info, prim_info);
}
/* "clipped" also includes non-one edgeflag */
if (clipped) {
opt |= PT_PIPELINE;
}
/* Do we need to run the pipeline? Now will come here if clipped */
if (opt & PT_PIPELINE) {
pipeline(fpme, vert_info, prim_info);
} else {
emit(fpme->emit, vert_info, prim_info);
}
}
}
FREE(vert_info->verts);
if (gshader && gshader->num_vertex_streams > 1)
for (unsigned i = 1; i < gshader->num_vertex_streams; i++)
FREE(gs_vert_info[i].verts);
if (free_prim_info) {
FREE(tes_elts_out);
FREE(prim_info->primitive_lengths);
}
}
static inline enum pipe_prim_type
prim_type(enum pipe_prim_type prim, unsigned flags)
{
if (flags & DRAW_LINE_LOOP_AS_STRIP)
return PIPE_PRIM_LINE_STRIP;
else
return prim;
}
static void
llvm_middle_end_run(struct draw_pt_middle_end *middle,
const unsigned *fetch_elts,
unsigned fetch_count,
const ushort *draw_elts,
unsigned draw_count,
unsigned prim_flags)
{
struct llvm_middle_end *fpme = llvm_middle_end(middle);
struct draw_fetch_info fetch_info;
struct draw_prim_info prim_info;
fetch_info.linear = FALSE;
fetch_info.start = 0;
fetch_info.elts = fetch_elts;
fetch_info.count = fetch_count;
prim_info.linear = FALSE;
prim_info.start = 0;
prim_info.count = draw_count;
prim_info.elts = draw_elts;
prim_info.prim = prim_type(fpme->input_prim, prim_flags);
prim_info.flags = prim_flags;
prim_info.primitive_count = 1;
prim_info.primitive_lengths = &draw_count;
llvm_pipeline_generic(middle, &fetch_info, &prim_info);
}
static void
llvm_middle_end_linear_run(struct draw_pt_middle_end *middle,
unsigned start,
unsigned count,
unsigned prim_flags)
{
struct llvm_middle_end *fpme = llvm_middle_end(middle);
struct draw_fetch_info fetch_info;
struct draw_prim_info prim_info;
fetch_info.linear = TRUE;
fetch_info.start = start;
fetch_info.count = count;
fetch_info.elts = NULL;
prim_info.linear = TRUE;
prim_info.start = 0;
prim_info.count = count;
prim_info.elts = NULL;
prim_info.prim = prim_type(fpme->input_prim, prim_flags);
prim_info.flags = prim_flags;
prim_info.primitive_count = 1;
prim_info.primitive_lengths = &count;
llvm_pipeline_generic(middle, &fetch_info, &prim_info);
}
static boolean
llvm_middle_end_linear_run_elts(struct draw_pt_middle_end *middle,
unsigned start,
unsigned count,
const ushort *draw_elts,
unsigned draw_count,
unsigned prim_flags)
{
struct llvm_middle_end *fpme = llvm_middle_end(middle);
struct draw_fetch_info fetch_info;
struct draw_prim_info prim_info;
fetch_info.linear = TRUE;
fetch_info.start = start;
fetch_info.count = count;
fetch_info.elts = NULL;
prim_info.linear = FALSE;
prim_info.start = 0;
prim_info.count = draw_count;
prim_info.elts = draw_elts;
prim_info.prim = prim_type(fpme->input_prim, prim_flags);
prim_info.flags = prim_flags;
prim_info.primitive_count = 1;
prim_info.primitive_lengths = &draw_count;
llvm_pipeline_generic(middle, &fetch_info, &prim_info);
return TRUE;
}
static void
llvm_middle_end_finish(struct draw_pt_middle_end *middle)
{
/* nothing to do */
}
static void
llvm_middle_end_destroy(struct draw_pt_middle_end *middle)
{
struct llvm_middle_end *fpme = llvm_middle_end(middle);
if (fpme->fetch)
draw_pt_fetch_destroy(fpme->fetch);
if (fpme->emit)
draw_pt_emit_destroy(fpme->emit);
if (fpme->so_emit)
draw_pt_so_emit_destroy(fpme->so_emit);
if (fpme->post_vs)
draw_pt_post_vs_destroy(fpme->post_vs);
FREE(middle);
}
struct draw_pt_middle_end *
draw_pt_fetch_pipeline_or_emit_llvm(struct draw_context *draw)
{
struct llvm_middle_end *fpme = 0;
if (!draw->llvm)
return NULL;
fpme = CALLOC_STRUCT(llvm_middle_end);
if (!fpme)
goto fail;
fpme->base.prepare = llvm_middle_end_prepare;
fpme->base.bind_parameters = llvm_middle_end_bind_parameters;
fpme->base.run = llvm_middle_end_run;
fpme->base.run_linear = llvm_middle_end_linear_run;
fpme->base.run_linear_elts = llvm_middle_end_linear_run_elts;
fpme->base.finish = llvm_middle_end_finish;
fpme->base.destroy = llvm_middle_end_destroy;
fpme->draw = draw;
fpme->fetch = draw_pt_fetch_create(draw);
if (!fpme->fetch)
goto fail;
fpme->post_vs = draw_pt_post_vs_create(draw);
if (!fpme->post_vs)
goto fail;
fpme->emit = draw_pt_emit_create(draw);
if (!fpme->emit)
goto fail;
fpme->so_emit = draw_pt_so_emit_create(draw);
if (!fpme->so_emit)
goto fail;
fpme->llvm = draw->llvm;
if (!fpme->llvm)
goto fail;
fpme->current_variant = NULL;
return &fpme->base;
fail:
if (fpme)
llvm_middle_end_destroy(&fpme->base);
return NULL;
}