/************************************************************************** * * Copyright 2007 Tungsten Graphics, Inc., Cedar Park, Texas. * All Rights Reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sub license, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice (including the * next paragraph) shall be included in all copies or substantial portions * of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. * **************************************************************************/ /** * \file * Build post-transformation, post-clipping vertex buffers and element * lists by hooking into the end of the primitive pipeline and * manipulating the vertex_id field in the vertex headers. * * XXX: work in progress * * \author José Fonseca <jrfonseca@tungstengraphics.com> * \author Keith Whitwell <keith@tungstengraphics.com> */ #include "draw/draw_context.h" #include "draw/draw_vbuf.h" #include "util/u_debug.h" #include "util/u_inlines.h" #include "util/u_math.h" #include "util/u_memory.h" #include "util/u_fifo.h" #include "i915_context.h" #include "i915_reg.h" #include "i915_batch.h" #include "i915_state.h" #define VBUF_MAP_BUFFER /** * Primitive renderer for i915. */ struct i915_vbuf_render { struct vbuf_render base; struct i915_context *i915; /** Vertex size in bytes */ size_t vertex_size; /** Software primitive */ unsigned prim; /** Hardware primitive */ unsigned hwprim; /** Genereate a vertex list */ unsigned fallback; /* Stuff for the vbo */ struct i915_winsys_buffer *vbo; size_t vbo_size; /**< current size of allocated buffer */ size_t vbo_alloc_size; /**< minimum buffer size to allocate */ size_t vbo_hw_offset; /**< offset that we program the hardware with */ size_t vbo_sw_offset; /**< offset that we work with */ size_t vbo_index; /**< index offset to be added to all indices */ void *vbo_ptr; size_t vbo_max_used; size_t vbo_max_index; /**< index offset to be added to all indices */ #ifndef VBUF_MAP_BUFFER size_t map_used_start; size_t map_used_end; size_t map_size; #endif }; /** * Basically a cast wrapper. */ static INLINE struct i915_vbuf_render * i915_vbuf_render(struct vbuf_render *render) { assert(render); return (struct i915_vbuf_render *)render; } /** * If vbo state differs between renderer and context * push state to the context. This function pushes * hw_offset to i915->vbo_offset and vbo to i915->vbo. * * Side effects: * May updates context vbo_offset and vbo fields. */ static void i915_vbuf_update_vbo_state(struct vbuf_render *render) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); struct i915_context *i915 = i915_render->i915; if (i915->vbo != i915_render->vbo || i915->vbo_offset != i915_render->vbo_hw_offset) { i915->vbo = i915_render->vbo; i915->vbo_offset = i915_render->vbo_hw_offset; i915->dirty |= I915_NEW_VBO; } } /** * Callback exported to the draw module. * Returns the current vertex_info. * * Side effects: * If state is dirty update derived state. */ static const struct vertex_info * i915_vbuf_render_get_vertex_info(struct vbuf_render *render) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); struct i915_context *i915 = i915_render->i915; if (i915->dirty) { /* make sure we have up to date vertex layout */ i915_update_derived(i915); } return &i915->current.vertex_info; } /** * Reserve space in the vbo for vertices. * * Side effects: * None. */ static boolean i915_vbuf_render_reserve(struct i915_vbuf_render *i915_render, size_t size) { struct i915_context *i915 = i915_render->i915; if (i915_render->vbo_size < size + i915_render->vbo_sw_offset) return FALSE; if (i915->vbo_flushed) return FALSE; return TRUE; } /** * Allocate a new vbo buffer should there not be enough space for * the requested number of vertices by the draw module. * * Side effects: * Updates hw_offset, sw_offset, index and allocates a new buffer. * Will set i915->vbo to null on buffer allocation. */ static void i915_vbuf_render_new_buf(struct i915_vbuf_render *i915_render, size_t size) { struct i915_context *i915 = i915_render->i915; struct i915_winsys *iws = i915->iws; if (i915_render->vbo) { iws->buffer_unmap(iws, i915_render->vbo); iws->buffer_destroy(iws, i915_render->vbo); /* * XXX If buffers where referenced then this should be done in * update_vbo_state but since they arn't and malloc likes to reuse * memory we need to set it to null */ i915->vbo = NULL; i915_render->vbo = NULL; } i915->vbo_flushed = 0; i915_render->vbo_size = MAX2(size, i915_render->vbo_alloc_size); i915_render->vbo_hw_offset = 0; i915_render->vbo_sw_offset = 0; i915_render->vbo_index = 0; #ifndef VBUF_MAP_BUFFER if (i915_render->vbo_size > i915_render->map_size) { i915_render->map_size = i915_render->vbo_size; FREE(i915_render->vbo_ptr); i915_render->vbo_ptr = MALLOC(i915_render->map_size); } #endif i915_render->vbo = iws->buffer_create(iws, i915_render->vbo_size, I915_NEW_VERTEX); i915_render->vbo_ptr = iws->buffer_map(iws, i915_render->vbo, TRUE); } /** * Callback exported to the draw module. * * Side effects: * Updates hw_offset, sw_offset, index and may allocate * a new buffer. Also updates may update the vbo state * on the i915 context. */ static boolean i915_vbuf_render_allocate_vertices(struct vbuf_render *render, ushort vertex_size, ushort nr_vertices) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); size_t size = (size_t)vertex_size * (size_t)nr_vertices; size_t offset; /* * Align sw_offset with first multiple of vertex size from hw_offset. * Set index to be the multiples from from hw_offset to sw_offset. * i915_vbuf_render_new_buf will reset index, sw_offset, hw_offset * when it allocates a new buffer this is correct. */ { offset = i915_render->vbo_sw_offset - i915_render->vbo_hw_offset; offset = util_align_npot(offset, vertex_size); i915_render->vbo_sw_offset = i915_render->vbo_hw_offset + offset; i915_render->vbo_index = offset / vertex_size; } if (!i915_vbuf_render_reserve(i915_render, size)) i915_vbuf_render_new_buf(i915_render, size); /* * If a new buffer has been alocated sw_offset, * hw_offset & index will be reset by new_buf */ i915_render->vertex_size = vertex_size; i915_vbuf_update_vbo_state(render); if (!i915_render->vbo) return FALSE; return TRUE; } static void * i915_vbuf_render_map_vertices(struct vbuf_render *render) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); struct i915_context *i915 = i915_render->i915; if (i915->vbo_flushed) debug_printf("%s bad vbo flush occured stalling on hw\n", __FUNCTION__); #ifdef VBUF_MAP_BUFFER return (unsigned char *)i915_render->vbo_ptr + i915_render->vbo_sw_offset; #else return (unsigned char *)i915_render->vbo_ptr; #endif } static void i915_vbuf_render_unmap_vertices(struct vbuf_render *render, ushort min_index, ushort max_index) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); struct i915_context *i915 = i915_render->i915; struct i915_winsys *iws = i915->iws; i915_render->vbo_max_index = max_index; i915_render->vbo_max_used = MAX2(i915_render->vbo_max_used, i915_render->vertex_size * (max_index + 1)); #ifdef VBUF_MAP_BUFFER (void)iws; #else i915_render->map_used_start = i915_render->vertex_size * min_index; i915_render->map_used_end = i915_render->vertex_size * (max_index + 1); iws->buffer_write(iws, i915_render->vbo, i915_render->map_used_start + i915_render->vbo_sw_offset, i915_render->map_used_end - i915_render->map_used_start, (unsigned char *)i915_render->vbo_ptr + i915_render->map_used_start); #endif } /** * Ensure that the given max_index given is not larger ushort max. * If it is larger then ushort max it advanced the hw_offset to the * same position in the vbo as sw_offset and set index to zero. * * Side effects: * On failure update hw_offset and index. */ static void i915_vbuf_ensure_index_bounds(struct vbuf_render *render, unsigned max_index) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); if (max_index + i915_render->vbo_index < ((1 << 17) - 1)) return; i915_render->vbo_hw_offset = i915_render->vbo_sw_offset; i915_render->vbo_index = 0; i915_vbuf_update_vbo_state(render); } static void i915_vbuf_render_set_primitive(struct vbuf_render *render, unsigned prim) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); i915_render->prim = prim; switch(prim) { case PIPE_PRIM_POINTS: i915_render->hwprim = PRIM3D_POINTLIST; i915_render->fallback = 0; break; case PIPE_PRIM_LINES: i915_render->hwprim = PRIM3D_LINELIST; i915_render->fallback = 0; break; case PIPE_PRIM_LINE_LOOP: i915_render->hwprim = PRIM3D_LINELIST; i915_render->fallback = PIPE_PRIM_LINE_LOOP; break; case PIPE_PRIM_LINE_STRIP: i915_render->hwprim = PRIM3D_LINESTRIP; i915_render->fallback = 0; break; case PIPE_PRIM_TRIANGLES: i915_render->hwprim = PRIM3D_TRILIST; i915_render->fallback = 0; break; case PIPE_PRIM_TRIANGLE_STRIP: i915_render->hwprim = PRIM3D_TRISTRIP; i915_render->fallback = 0; break; case PIPE_PRIM_TRIANGLE_FAN: i915_render->hwprim = PRIM3D_TRIFAN; i915_render->fallback = 0; break; case PIPE_PRIM_QUADS: i915_render->hwprim = PRIM3D_TRILIST; i915_render->fallback = PIPE_PRIM_QUADS; break; case PIPE_PRIM_QUAD_STRIP: i915_render->hwprim = PRIM3D_TRILIST; i915_render->fallback = PIPE_PRIM_QUAD_STRIP; break; case PIPE_PRIM_POLYGON: i915_render->hwprim = PRIM3D_POLY; i915_render->fallback = 0; break; default: /* FIXME: Actually, can handle a lot more just fine... */ assert(0 && "unexpected prim in i915_vbuf_render_set_primitive()"); } } /** * Used for fallbacks in draw_arrays */ static void draw_arrays_generate_indices(struct vbuf_render *render, unsigned start, uint nr, unsigned type) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); struct i915_context *i915 = i915_render->i915; unsigned i; unsigned end = start + nr + i915_render->vbo_index; start += i915_render->vbo_index; switch(type) { case 0: for (i = start; i+1 < end; i += 2) OUT_BATCH((i+0) | (i+1) << 16); if (i < end) OUT_BATCH(i); break; case PIPE_PRIM_LINE_LOOP: if (nr >= 2) { for (i = start + 1; i < end; i++) OUT_BATCH((i-1) | (i+0) << 16); OUT_BATCH((i-1) | ( start) << 16); } break; case PIPE_PRIM_QUADS: for (i = start; i + 3 < end; i += 4) { OUT_BATCH((i+0) | (i+1) << 16); OUT_BATCH((i+3) | (i+1) << 16); OUT_BATCH((i+2) | (i+3) << 16); } break; case PIPE_PRIM_QUAD_STRIP: for (i = start; i + 3 < end; i += 2) { OUT_BATCH((i+0) | (i+1) << 16); OUT_BATCH((i+3) | (i+2) << 16); OUT_BATCH((i+0) | (i+3) << 16); } break; default: assert(0); } } static unsigned draw_arrays_calc_nr_indices(uint nr, unsigned type) { switch (type) { case 0: return nr; case PIPE_PRIM_LINE_LOOP: if (nr >= 2) return nr * 2; else return 0; case PIPE_PRIM_QUADS: return (nr / 4) * 6; case PIPE_PRIM_QUAD_STRIP: return ((nr - 2) / 2) * 6; default: assert(0); return 0; } } static void draw_arrays_fallback(struct vbuf_render *render, unsigned start, uint nr) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); struct i915_context *i915 = i915_render->i915; unsigned nr_indices; nr_indices = draw_arrays_calc_nr_indices(nr, i915_render->fallback); if (!nr_indices) return; i915_vbuf_ensure_index_bounds(render, start + nr_indices); if (i915->dirty) i915_update_derived(i915); if (i915->hardware_dirty) i915_emit_hardware_state(i915); if (!BEGIN_BATCH(1 + (nr_indices + 1)/2)) { FLUSH_BATCH(NULL); /* Make sure state is re-emitted after a flush: */ i915_emit_hardware_state(i915); i915->vbo_flushed = 1; if (!BEGIN_BATCH(1 + (nr_indices + 1)/2)) { assert(0); goto out; } } OUT_BATCH(_3DPRIMITIVE | PRIM_INDIRECT | i915_render->hwprim | PRIM_INDIRECT_ELTS | nr_indices); draw_arrays_generate_indices(render, start, nr, i915_render->fallback); out: return; } static void i915_vbuf_render_draw_arrays(struct vbuf_render *render, unsigned start, uint nr) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); struct i915_context *i915 = i915_render->i915; if (i915_render->fallback) { draw_arrays_fallback(render, start, nr); return; } i915_vbuf_ensure_index_bounds(render, start + nr); start += i915_render->vbo_index; if (i915->dirty) i915_update_derived(i915); if (i915->hardware_dirty) i915_emit_hardware_state(i915); if (!BEGIN_BATCH(2)) { FLUSH_BATCH(NULL); /* Make sure state is re-emitted after a flush: */ i915_emit_hardware_state(i915); i915->vbo_flushed = 1; if (!BEGIN_BATCH(2)) { assert(0); goto out; } } OUT_BATCH(_3DPRIMITIVE | PRIM_INDIRECT | PRIM_INDIRECT_SEQUENTIAL | i915_render->hwprim | nr); OUT_BATCH(start); /* Beginning vertex index */ out: return; } /** * Used for normal and fallback emitting of indices * If type is zero normal operation assumed. */ static void draw_generate_indices(struct vbuf_render *render, const ushort *indices, uint nr_indices, unsigned type) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); struct i915_context *i915 = i915_render->i915; unsigned i; unsigned o = i915_render->vbo_index; switch(type) { case 0: for (i = 0; i + 1 < nr_indices; i += 2) { OUT_BATCH((o+indices[i]) | (o+indices[i+1]) << 16); } if (i < nr_indices) { OUT_BATCH((o+indices[i])); } break; case PIPE_PRIM_LINE_LOOP: if (nr_indices >= 2) { for (i = 1; i < nr_indices; i++) OUT_BATCH((o+indices[i-1]) | (o+indices[i]) << 16); OUT_BATCH((o+indices[i-1]) | (o+indices[0]) << 16); } break; case PIPE_PRIM_QUADS: for (i = 0; i + 3 < nr_indices; i += 4) { OUT_BATCH((o+indices[i+0]) | (o+indices[i+1]) << 16); OUT_BATCH((o+indices[i+3]) | (o+indices[i+1]) << 16); OUT_BATCH((o+indices[i+2]) | (o+indices[i+3]) << 16); } break; case PIPE_PRIM_QUAD_STRIP: for (i = 0; i + 3 < nr_indices; i += 2) { OUT_BATCH((o+indices[i+0]) | (o+indices[i+1]) << 16); OUT_BATCH((o+indices[i+3]) | (o+indices[i+2]) << 16); OUT_BATCH((o+indices[i+0]) | (o+indices[i+3]) << 16); } break; default: assert(0); break; } } static unsigned draw_calc_nr_indices(uint nr_indices, unsigned type) { switch (type) { case 0: return nr_indices; case PIPE_PRIM_LINE_LOOP: if (nr_indices >= 2) return nr_indices * 2; else return 0; case PIPE_PRIM_QUADS: return (nr_indices / 4) * 6; case PIPE_PRIM_QUAD_STRIP: return ((nr_indices - 2) / 2) * 6; default: assert(0); return 0; } } static void i915_vbuf_render_draw_elements(struct vbuf_render *render, const ushort *indices, uint nr_indices) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); struct i915_context *i915 = i915_render->i915; unsigned save_nr_indices; save_nr_indices = nr_indices; nr_indices = draw_calc_nr_indices(nr_indices, i915_render->fallback); if (!nr_indices) return; i915_vbuf_ensure_index_bounds(render, i915_render->vbo_max_index); if (i915->dirty) i915_update_derived(i915); if (i915->hardware_dirty) i915_emit_hardware_state(i915); if (!BEGIN_BATCH(1 + (nr_indices + 1)/2)) { FLUSH_BATCH(NULL); /* Make sure state is re-emitted after a flush: */ i915_emit_hardware_state(i915); i915->vbo_flushed = 1; if (!BEGIN_BATCH(1 + (nr_indices + 1)/2)) { assert(0); goto out; } } OUT_BATCH(_3DPRIMITIVE | PRIM_INDIRECT | i915_render->hwprim | PRIM_INDIRECT_ELTS | nr_indices); draw_generate_indices(render, indices, save_nr_indices, i915_render->fallback); out: return; } static void i915_vbuf_render_release_vertices(struct vbuf_render *render) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); i915_render->vbo_sw_offset += i915_render->vbo_max_used; i915_render->vbo_max_used = 0; /* * Micro optimization, by calling update here we the offset change * will be picked up on the next pipe_context::draw_*. */ i915_vbuf_update_vbo_state(render); } static void i915_vbuf_render_destroy(struct vbuf_render *render) { struct i915_vbuf_render *i915_render = i915_vbuf_render(render); struct i915_context *i915 = i915_render->i915; struct i915_winsys *iws = i915->iws; if (i915_render->vbo) { i915->vbo = NULL; iws->buffer_unmap(iws, i915_render->vbo); iws->buffer_destroy(iws, i915_render->vbo); } FREE(i915_render); } /** * Create a new primitive render. */ static struct vbuf_render * i915_vbuf_render_create(struct i915_context *i915) { struct i915_vbuf_render *i915_render = CALLOC_STRUCT(i915_vbuf_render); struct i915_winsys *iws = i915->iws; int i; i915_render->i915 = i915; i915_render->base.max_vertex_buffer_bytes = 16*4096; /* NOTE: it must be such that state and vertices indices fit in a single * batch buffer. */ i915_render->base.max_indices = 16*1024; i915_render->base.get_vertex_info = i915_vbuf_render_get_vertex_info; i915_render->base.allocate_vertices = i915_vbuf_render_allocate_vertices; i915_render->base.map_vertices = i915_vbuf_render_map_vertices; i915_render->base.unmap_vertices = i915_vbuf_render_unmap_vertices; i915_render->base.set_primitive = i915_vbuf_render_set_primitive; i915_render->base.draw_elements = i915_vbuf_render_draw_elements; i915_render->base.draw_arrays = i915_vbuf_render_draw_arrays; i915_render->base.release_vertices = i915_vbuf_render_release_vertices; i915_render->base.destroy = i915_vbuf_render_destroy; #ifndef VBUF_MAP_BUFFER i915_render->map_size = 0; i915_render->map_used_start = 0; i915_render->map_used_end = 0; #endif i915_render->vbo = NULL; i915_render->vbo_ptr = NULL; i915_render->vbo_size = 0; i915_render->vbo_hw_offset = 0; i915_render->vbo_sw_offset = 0; i915_render->vbo_alloc_size = i915_render->base.max_vertex_buffer_bytes * 4; #ifdef VBUF_USE_POOL i915_render->pool_used = FALSE; i915_render->pool_buffer_size = i915_render->vbo_alloc_size; i915_render->pool_fifo = u_fifo_create(6); for (i = 0; i < 6; i++) u_fifo_add(i915_render->pool_fifo, iws->buffer_create(iws, i915_render->pool_buffer_size, I915_NEW_VERTEX)); #else (void)i; (void)iws; #endif return &i915_render->base; } /** * Create a new primitive vbuf/render stage. */ struct draw_stage *i915_draw_vbuf_stage(struct i915_context *i915) { struct vbuf_render *render; struct draw_stage *stage; render = i915_vbuf_render_create(i915); if(!render) return NULL; stage = draw_vbuf_stage(i915->draw, render); if(!stage) { render->destroy(render); return NULL; } /** TODO JB: this shouldn't be here */ draw_set_render(i915->draw, render); return stage; }