/* -*- mode: C; c-file-style: "k&r"; tab-width 4; indent-tabs-mode: t; -*- */ /* * Copyright (C) 2012 Rob Clark * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. * * Authors: * Rob Clark */ #include #include "freedreno_drmif.h" #include "freedreno_priv.h" #include "freedreno_ringbuffer.h" drm_public struct fd_ringbuffer * fd_ringbuffer_new_flags(struct fd_pipe *pipe, uint32_t size, enum fd_ringbuffer_flags flags) { struct fd_ringbuffer *ring; /* we can't really support "growable" rb's in general for * stateobj's since we need a single gpu addr (ie. can't * do the trick of a chain of IB packets): */ if (flags & FD_RINGBUFFER_OBJECT) assert(size); ring = pipe->funcs->ringbuffer_new(pipe, size, flags); if (!ring) return NULL; ring->flags = flags; ring->pipe = pipe; ring->start = ring->funcs->hostptr(ring); ring->end = &(ring->start[ring->size/4]); ring->cur = ring->last_start = ring->start; return ring; } drm_public struct fd_ringbuffer * fd_ringbuffer_new(struct fd_pipe *pipe, uint32_t size) { return fd_ringbuffer_new_flags(pipe, size, 0); } drm_public struct fd_ringbuffer * fd_ringbuffer_new_object(struct fd_pipe *pipe, uint32_t size) { return fd_ringbuffer_new_flags(pipe, size, FD_RINGBUFFER_OBJECT); } drm_public void fd_ringbuffer_del(struct fd_ringbuffer *ring) { if (!atomic_dec_and_test(&ring->refcnt)) return; fd_ringbuffer_reset(ring); ring->funcs->destroy(ring); } drm_public struct fd_ringbuffer * fd_ringbuffer_ref(struct fd_ringbuffer *ring) { STATIC_ASSERT(sizeof(ring->refcnt) <= sizeof(ring->__pad)); atomic_inc(&ring->refcnt); return ring; } /* ringbuffers which are IB targets should set the toplevel rb (ie. * the IB source) as it's parent before emitting reloc's, to ensure * the bookkeeping works out properly. */ drm_public void fd_ringbuffer_set_parent(struct fd_ringbuffer *ring, struct fd_ringbuffer *parent) { /* state objects should not be parented! */ assert(!(ring->flags & FD_RINGBUFFER_OBJECT)); ring->parent = parent; } drm_public void fd_ringbuffer_reset(struct fd_ringbuffer *ring) { uint32_t *start = ring->start; if (ring->pipe->id == FD_PIPE_2D) start = &ring->start[0x140]; ring->cur = ring->last_start = start; if (ring->funcs->reset) ring->funcs->reset(ring); } drm_public int fd_ringbuffer_flush(struct fd_ringbuffer *ring) { return ring->funcs->flush(ring, ring->last_start, -1, NULL); } drm_public int fd_ringbuffer_flush2(struct fd_ringbuffer *ring, int in_fence_fd, int *out_fence_fd) { return ring->funcs->flush(ring, ring->last_start, in_fence_fd, out_fence_fd); } drm_public void fd_ringbuffer_grow(struct fd_ringbuffer *ring, uint32_t ndwords) { assert(ring->funcs->grow); /* unsupported on kgsl */ /* there is an upper bound on IB size, which appears to be 0x100000 */ if (ring->size < 0x100000) ring->size *= 2; ring->funcs->grow(ring, ring->size); ring->start = ring->funcs->hostptr(ring); ring->end = &(ring->start[ring->size/4]); ring->cur = ring->last_start = ring->start; } drm_public uint32_t fd_ringbuffer_timestamp(struct fd_ringbuffer *ring) { return ring->last_timestamp; } drm_public void fd_ringbuffer_reloc(struct fd_ringbuffer *ring, const struct fd_reloc *reloc) { assert(ring->pipe->gpu_id < 500); ring->funcs->emit_reloc(ring, reloc); } drm_public void fd_ringbuffer_reloc2(struct fd_ringbuffer *ring, const struct fd_reloc *reloc) { ring->funcs->emit_reloc(ring, reloc); } drm_public uint32_t fd_ringbuffer_cmd_count(struct fd_ringbuffer *ring) { if (!ring->funcs->cmd_count) return 1; return ring->funcs->cmd_count(ring); } drm_public uint32_t fd_ringbuffer_emit_reloc_ring_full(struct fd_ringbuffer *ring, struct fd_ringbuffer *target, uint32_t cmd_idx) { return ring->funcs->emit_reloc_ring(ring, target, cmd_idx); } drm_public uint32_t fd_ringbuffer_size(struct fd_ringbuffer *ring) { /* only really needed for stateobj ringbuffers, and won't really * do what you expect for growable rb's.. so lets just restrict * this to stateobj's for now: */ assert(ring->flags & FD_RINGBUFFER_OBJECT); return offset_bytes(ring->cur, ring->start); }