0,0 → 1,605 |
/************************************************************************** |
* |
* Copyright 2007 Tungsten Graphics, Inc., Cedar Park, Texas. |
* All Rights Reserved. |
* |
* Permission is hereby granted, free of charge, to any person obtaining a |
* copy of this software and associated documentation files (the |
* "Software"), to deal in the Software without restriction, including |
* without limitation the rights to use, copy, modify, merge, publish, |
* distribute, sub license, and/or sell copies of the Software, and to |
* permit persons to whom the Software is furnished to do so, subject to |
* the following conditions: |
* |
* The above copyright notice and this permission notice (including the |
* next paragraph) shall be included in all copies or substantial portions |
* of the Software. |
* |
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS |
* OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. |
* IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR |
* ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, |
* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE |
* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. |
* |
**************************************************************************/ |
|
#ifndef U_INLINES_H |
#define U_INLINES_H |
|
#include "pipe/p_context.h" |
#include "pipe/p_defines.h" |
#include "pipe/p_shader_tokens.h" |
#include "pipe/p_state.h" |
#include "pipe/p_screen.h" |
#include "util/u_debug.h" |
#include "util/u_debug_describe.h" |
#include "util/u_debug_refcnt.h" |
#include "util/u_atomic.h" |
#include "util/u_box.h" |
#include "util/u_math.h" |
|
|
#ifdef __cplusplus |
extern "C" { |
#endif |
|
|
/* |
* Reference counting helper functions. |
*/ |
|
|
static INLINE void |
pipe_reference_init(struct pipe_reference *reference, unsigned count) |
{ |
p_atomic_set(&reference->count, count); |
} |
|
static INLINE boolean |
pipe_is_referenced(struct pipe_reference *reference) |
{ |
return p_atomic_read(&reference->count) != 0; |
} |
|
/** |
* Update reference counting. |
* The old thing pointed to, if any, will be unreferenced. |
* Both 'ptr' and 'reference' may be NULL. |
* \return TRUE if the object's refcount hits zero and should be destroyed. |
*/ |
static INLINE boolean |
pipe_reference_described(struct pipe_reference *ptr, |
struct pipe_reference *reference, |
debug_reference_descriptor get_desc) |
{ |
boolean destroy = FALSE; |
|
if(ptr != reference) { |
/* bump the reference.count first */ |
if (reference) { |
assert(pipe_is_referenced(reference)); |
p_atomic_inc(&reference->count); |
debug_reference(reference, get_desc, 1); |
} |
|
if (ptr) { |
assert(pipe_is_referenced(ptr)); |
if (p_atomic_dec_zero(&ptr->count)) { |
destroy = TRUE; |
} |
debug_reference(ptr, get_desc, -1); |
} |
} |
|
return destroy; |
} |
|
static INLINE boolean |
pipe_reference(struct pipe_reference *ptr, struct pipe_reference *reference) |
{ |
return pipe_reference_described(ptr, reference, |
(debug_reference_descriptor)debug_describe_reference); |
} |
|
static INLINE void |
pipe_surface_reference(struct pipe_surface **ptr, struct pipe_surface *surf) |
{ |
struct pipe_surface *old_surf = *ptr; |
|
if (pipe_reference_described(&(*ptr)->reference, &surf->reference, |
(debug_reference_descriptor)debug_describe_surface)) |
old_surf->context->surface_destroy(old_surf->context, old_surf); |
*ptr = surf; |
} |
|
/** |
* Similar to pipe_surface_reference() but always set the pointer to NULL |
* and pass in an explicit context. The explicit context avoids the problem |
* of using a deleted context's surface_destroy() method when freeing a surface |
* that's shared by multiple contexts. |
*/ |
static INLINE void |
pipe_surface_release(struct pipe_context *pipe, struct pipe_surface **ptr) |
{ |
if (pipe_reference_described(&(*ptr)->reference, NULL, |
(debug_reference_descriptor)debug_describe_surface)) |
pipe->surface_destroy(pipe, *ptr); |
*ptr = NULL; |
} |
|
|
static INLINE void |
pipe_resource_reference(struct pipe_resource **ptr, struct pipe_resource *tex) |
{ |
struct pipe_resource *old_tex = *ptr; |
|
if (pipe_reference_described(&(*ptr)->reference, &tex->reference, |
(debug_reference_descriptor)debug_describe_resource)) |
old_tex->screen->resource_destroy(old_tex->screen, old_tex); |
*ptr = tex; |
} |
|
static INLINE void |
pipe_sampler_view_reference(struct pipe_sampler_view **ptr, struct pipe_sampler_view *view) |
{ |
struct pipe_sampler_view *old_view = *ptr; |
|
if (pipe_reference_described(&(*ptr)->reference, &view->reference, |
(debug_reference_descriptor)debug_describe_sampler_view)) |
old_view->context->sampler_view_destroy(old_view->context, old_view); |
*ptr = view; |
} |
|
/** |
* Similar to pipe_sampler_view_reference() but always set the pointer to |
* NULL and pass in an explicit context. Passing an explicit context is a |
* work-around for fixing a dangling context pointer problem when textures |
* are shared by multiple contexts. XXX fix this someday. |
*/ |
static INLINE void |
pipe_sampler_view_release(struct pipe_context *ctx, |
struct pipe_sampler_view **ptr) |
{ |
struct pipe_sampler_view *old_view = *ptr; |
if (*ptr && (*ptr)->context != ctx) { |
debug_printf_once(("context mis-match in pipe_sampler_view_release()\n")); |
} |
if (pipe_reference_described(&(*ptr)->reference, NULL, |
(debug_reference_descriptor)debug_describe_sampler_view)) { |
ctx->sampler_view_destroy(ctx, old_view); |
} |
*ptr = NULL; |
} |
|
|
static INLINE void |
pipe_so_target_reference(struct pipe_stream_output_target **ptr, |
struct pipe_stream_output_target *target) |
{ |
struct pipe_stream_output_target *old = *ptr; |
|
if (pipe_reference_described(&(*ptr)->reference, &target->reference, |
(debug_reference_descriptor)debug_describe_so_target)) |
old->context->stream_output_target_destroy(old->context, old); |
*ptr = target; |
} |
|
static INLINE void |
pipe_surface_reset(struct pipe_context *ctx, struct pipe_surface* ps, |
struct pipe_resource *pt, unsigned level, unsigned layer) |
{ |
pipe_resource_reference(&ps->texture, pt); |
ps->format = pt->format; |
ps->width = u_minify(pt->width0, level); |
ps->height = u_minify(pt->height0, level); |
ps->u.tex.level = level; |
ps->u.tex.first_layer = ps->u.tex.last_layer = layer; |
ps->context = ctx; |
} |
|
static INLINE void |
pipe_surface_init(struct pipe_context *ctx, struct pipe_surface* ps, |
struct pipe_resource *pt, unsigned level, unsigned layer) |
{ |
ps->texture = 0; |
pipe_reference_init(&ps->reference, 1); |
pipe_surface_reset(ctx, ps, pt, level, layer); |
} |
|
/* Return true if the surfaces are equal. */ |
static INLINE boolean |
pipe_surface_equal(struct pipe_surface *s1, struct pipe_surface *s2) |
{ |
return s1->texture == s2->texture && |
s1->format == s2->format && |
(s1->texture->target != PIPE_BUFFER || |
(s1->u.buf.first_element == s2->u.buf.first_element && |
s1->u.buf.last_element == s2->u.buf.last_element)) && |
(s1->texture->target == PIPE_BUFFER || |
(s1->u.tex.level == s2->u.tex.level && |
s1->u.tex.first_layer == s2->u.tex.first_layer && |
s1->u.tex.last_layer == s2->u.tex.last_layer)); |
} |
|
/* |
* Convenience wrappers for screen buffer functions. |
*/ |
|
static INLINE struct pipe_resource * |
pipe_buffer_create( struct pipe_screen *screen, |
unsigned bind, |
unsigned usage, |
unsigned size ) |
{ |
struct pipe_resource buffer; |
memset(&buffer, 0, sizeof buffer); |
buffer.target = PIPE_BUFFER; |
buffer.format = PIPE_FORMAT_R8_UNORM; /* want TYPELESS or similar */ |
buffer.bind = bind; |
buffer.usage = usage; |
buffer.flags = 0; |
buffer.width0 = size; |
buffer.height0 = 1; |
buffer.depth0 = 1; |
buffer.array_size = 1; |
return screen->resource_create(screen, &buffer); |
} |
|
static INLINE void * |
pipe_buffer_map_range(struct pipe_context *pipe, |
struct pipe_resource *buffer, |
unsigned offset, |
unsigned length, |
unsigned usage, |
struct pipe_transfer **transfer) |
{ |
struct pipe_box box; |
void *map; |
|
assert(offset < buffer->width0); |
assert(offset + length <= buffer->width0); |
assert(length); |
|
u_box_1d(offset, length, &box); |
|
map = pipe->transfer_map(pipe, buffer, 0, usage, &box, transfer); |
if (map == NULL) { |
return NULL; |
} |
|
return map; |
} |
|
|
static INLINE void * |
pipe_buffer_map(struct pipe_context *pipe, |
struct pipe_resource *buffer, |
unsigned usage, |
struct pipe_transfer **transfer) |
{ |
return pipe_buffer_map_range(pipe, buffer, 0, buffer->width0, usage, transfer); |
} |
|
|
static INLINE void |
pipe_buffer_unmap(struct pipe_context *pipe, |
struct pipe_transfer *transfer) |
{ |
pipe->transfer_unmap(pipe, transfer); |
} |
|
static INLINE void |
pipe_buffer_flush_mapped_range(struct pipe_context *pipe, |
struct pipe_transfer *transfer, |
unsigned offset, |
unsigned length) |
{ |
struct pipe_box box; |
int transfer_offset; |
|
assert(length); |
assert(transfer->box.x <= (int) offset); |
assert((int) (offset + length) <= transfer->box.x + transfer->box.width); |
|
/* Match old screen->buffer_flush_mapped_range() behaviour, where |
* offset parameter is relative to the start of the buffer, not the |
* mapped range. |
*/ |
transfer_offset = offset - transfer->box.x; |
|
u_box_1d(transfer_offset, length, &box); |
|
pipe->transfer_flush_region(pipe, transfer, &box); |
} |
|
static INLINE void |
pipe_buffer_write(struct pipe_context *pipe, |
struct pipe_resource *buf, |
unsigned offset, |
unsigned size, |
const void *data) |
{ |
struct pipe_box box; |
unsigned usage = PIPE_TRANSFER_WRITE; |
|
if (offset == 0 && size == buf->width0) { |
usage |= PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE; |
} else { |
usage |= PIPE_TRANSFER_DISCARD_RANGE; |
} |
|
u_box_1d(offset, size, &box); |
|
pipe->transfer_inline_write( pipe, |
buf, |
0, |
usage, |
&box, |
data, |
size, |
0); |
} |
|
/** |
* Special case for writing non-overlapping ranges. |
* |
* We can avoid GPU/CPU synchronization when writing range that has never |
* been written before. |
*/ |
static INLINE void |
pipe_buffer_write_nooverlap(struct pipe_context *pipe, |
struct pipe_resource *buf, |
unsigned offset, unsigned size, |
const void *data) |
{ |
struct pipe_box box; |
|
u_box_1d(offset, size, &box); |
|
pipe->transfer_inline_write(pipe, |
buf, |
0, |
(PIPE_TRANSFER_WRITE | |
PIPE_TRANSFER_UNSYNCHRONIZED), |
&box, |
data, |
0, 0); |
} |
|
static INLINE struct pipe_resource * |
pipe_buffer_create_with_data(struct pipe_context *pipe, |
unsigned bind, |
unsigned usage, |
unsigned size, |
void *ptr) |
{ |
struct pipe_resource *res = pipe_buffer_create(pipe->screen, |
bind, usage, size); |
pipe_buffer_write_nooverlap(pipe, res, 0, size, ptr); |
return res; |
} |
|
static INLINE void |
pipe_buffer_read(struct pipe_context *pipe, |
struct pipe_resource *buf, |
unsigned offset, |
unsigned size, |
void *data) |
{ |
struct pipe_transfer *src_transfer; |
ubyte *map; |
|
map = (ubyte *) pipe_buffer_map_range(pipe, |
buf, |
offset, size, |
PIPE_TRANSFER_READ, |
&src_transfer); |
if (!map) |
return; |
|
memcpy(data, map, size); |
pipe_buffer_unmap(pipe, src_transfer); |
} |
|
static INLINE void * |
pipe_transfer_map(struct pipe_context *context, |
struct pipe_resource *resource, |
unsigned level, unsigned layer, |
enum pipe_transfer_usage usage, |
unsigned x, unsigned y, |
unsigned w, unsigned h, |
struct pipe_transfer **transfer) |
{ |
struct pipe_box box; |
u_box_2d_zslice(x, y, layer, w, h, &box); |
return context->transfer_map(context, |
resource, |
level, |
usage, |
&box, transfer); |
} |
|
static INLINE void * |
pipe_transfer_map_3d(struct pipe_context *context, |
struct pipe_resource *resource, |
unsigned level, |
enum pipe_transfer_usage usage, |
unsigned x, unsigned y, unsigned z, |
unsigned w, unsigned h, unsigned d, |
struct pipe_transfer **transfer) |
{ |
struct pipe_box box; |
u_box_3d(x, y, z, w, h, d, &box); |
return context->transfer_map(context, |
resource, |
level, |
usage, |
&box, transfer); |
} |
|
static INLINE void |
pipe_transfer_unmap( struct pipe_context *context, |
struct pipe_transfer *transfer ) |
{ |
context->transfer_unmap( context, transfer ); |
} |
|
static INLINE void |
pipe_set_constant_buffer(struct pipe_context *pipe, uint shader, uint index, |
struct pipe_resource *buf) |
{ |
if (buf) { |
struct pipe_constant_buffer cb; |
cb.buffer = buf; |
cb.buffer_offset = 0; |
cb.buffer_size = buf->width0; |
cb.user_buffer = NULL; |
pipe->set_constant_buffer(pipe, shader, index, &cb); |
} else { |
pipe->set_constant_buffer(pipe, shader, index, NULL); |
} |
} |
|
|
static INLINE boolean util_get_offset( |
const struct pipe_rasterizer_state *templ, |
unsigned fill_mode) |
{ |
switch(fill_mode) { |
case PIPE_POLYGON_MODE_POINT: |
return templ->offset_point; |
case PIPE_POLYGON_MODE_LINE: |
return templ->offset_line; |
case PIPE_POLYGON_MODE_FILL: |
return templ->offset_tri; |
default: |
assert(0); |
return FALSE; |
} |
} |
|
static INLINE float |
util_get_min_point_size(const struct pipe_rasterizer_state *state) |
{ |
/* The point size should be clamped to this value at the rasterizer stage. |
*/ |
return !state->point_quad_rasterization && |
!state->point_smooth && |
!state->multisample ? 1.0f : 0.0f; |
} |
|
static INLINE void |
util_query_clear_result(union pipe_query_result *result, unsigned type) |
{ |
switch (type) { |
case PIPE_QUERY_OCCLUSION_PREDICATE: |
case PIPE_QUERY_SO_OVERFLOW_PREDICATE: |
case PIPE_QUERY_GPU_FINISHED: |
result->b = FALSE; |
break; |
case PIPE_QUERY_OCCLUSION_COUNTER: |
case PIPE_QUERY_TIMESTAMP: |
case PIPE_QUERY_TIME_ELAPSED: |
case PIPE_QUERY_PRIMITIVES_GENERATED: |
case PIPE_QUERY_PRIMITIVES_EMITTED: |
result->u64 = 0; |
break; |
case PIPE_QUERY_SO_STATISTICS: |
memset(&result->so_statistics, 0, sizeof(result->so_statistics)); |
break; |
case PIPE_QUERY_TIMESTAMP_DISJOINT: |
memset(&result->timestamp_disjoint, 0, sizeof(result->timestamp_disjoint)); |
break; |
case PIPE_QUERY_PIPELINE_STATISTICS: |
memset(&result->pipeline_statistics, 0, sizeof(result->pipeline_statistics)); |
break; |
default: |
memset(result, 0, sizeof(*result)); |
} |
} |
|
/** Convert PIPE_TEXTURE_x to TGSI_TEXTURE_x */ |
static INLINE unsigned |
util_pipe_tex_to_tgsi_tex(enum pipe_texture_target pipe_tex_target, |
unsigned nr_samples) |
{ |
switch (pipe_tex_target) { |
case PIPE_TEXTURE_1D: |
assert(nr_samples <= 1); |
return TGSI_TEXTURE_1D; |
|
case PIPE_TEXTURE_2D: |
return nr_samples > 1 ? TGSI_TEXTURE_2D_MSAA : TGSI_TEXTURE_2D; |
|
case PIPE_TEXTURE_RECT: |
assert(nr_samples <= 1); |
return TGSI_TEXTURE_RECT; |
|
case PIPE_TEXTURE_3D: |
assert(nr_samples <= 1); |
return TGSI_TEXTURE_3D; |
|
case PIPE_TEXTURE_CUBE: |
assert(nr_samples <= 1); |
return TGSI_TEXTURE_CUBE; |
|
case PIPE_TEXTURE_1D_ARRAY: |
assert(nr_samples <= 1); |
return TGSI_TEXTURE_1D_ARRAY; |
|
case PIPE_TEXTURE_2D_ARRAY: |
return nr_samples > 1 ? TGSI_TEXTURE_2D_ARRAY_MSAA : |
TGSI_TEXTURE_2D_ARRAY; |
|
case PIPE_TEXTURE_CUBE_ARRAY: |
return TGSI_TEXTURE_CUBE_ARRAY; |
|
default: |
assert(0 && "unexpected texture target"); |
return TGSI_TEXTURE_UNKNOWN; |
} |
} |
|
|
static INLINE void |
util_copy_constant_buffer(struct pipe_constant_buffer *dst, |
const struct pipe_constant_buffer *src) |
{ |
if (src) { |
pipe_resource_reference(&dst->buffer, src->buffer); |
dst->buffer_offset = src->buffer_offset; |
dst->buffer_size = src->buffer_size; |
dst->user_buffer = src->user_buffer; |
} |
else { |
pipe_resource_reference(&dst->buffer, NULL); |
dst->buffer_offset = 0; |
dst->buffer_size = 0; |
dst->user_buffer = NULL; |
} |
} |
|
static INLINE unsigned |
util_max_layer(const struct pipe_resource *r, unsigned level) |
{ |
switch (r->target) { |
case PIPE_TEXTURE_CUBE: |
return 6 - 1; |
case PIPE_TEXTURE_3D: |
return u_minify(r->depth0, level) - 1; |
case PIPE_TEXTURE_1D_ARRAY: |
case PIPE_TEXTURE_2D_ARRAY: |
case PIPE_TEXTURE_CUBE_ARRAY: |
return r->array_size - 1; |
default: |
return 0; |
} |
} |
|
#ifdef __cplusplus |
} |
#endif |
|
#endif /* U_INLINES_H */ |