/* * Copyright 2012 Red Hat Inc. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR * OTHER DEALINGS IN THE SOFTWARE. * * Authors: Ben Skeggs */ #ifdef HAVE_CONFIG_H #include <config.h> #endif #include <stdio.h> #include <stdlib.h> #include <stdint.h> #include <stdbool.h> #include <string.h> #include <assert.h> #include <errno.h> #include <xf86drm.h> #include <xf86atomic.h> #include "libdrm_lists.h" #include "nouveau_drm.h" #include "nouveau.h" #include "private.h" struct nouveau_pushbuf_krec { struct nouveau_pushbuf_krec *next; struct drm_nouveau_gem_pushbuf_bo buffer[NOUVEAU_GEM_MAX_BUFFERS]; struct drm_nouveau_gem_pushbuf_reloc reloc[NOUVEAU_GEM_MAX_RELOCS]; struct drm_nouveau_gem_pushbuf_push push[NOUVEAU_GEM_MAX_PUSH]; int nr_buffer; int nr_reloc; int nr_push; uint64_t vram_used; uint64_t gart_used; }; struct nouveau_pushbuf_priv { struct nouveau_pushbuf base; struct nouveau_pushbuf_krec *list; struct nouveau_pushbuf_krec *krec; struct nouveau_list bctx_list; struct nouveau_bo *bo; uint32_t type; uint32_t suffix0; uint32_t suffix1; uint32_t *ptr; uint32_t *bgn; int bo_next; int bo_nr; struct nouveau_bo *bos[]; }; static inline struct nouveau_pushbuf_priv * nouveau_pushbuf(struct nouveau_pushbuf *push) { return (struct nouveau_pushbuf_priv *)push; } static int pushbuf_validate(struct nouveau_pushbuf *, bool); static int pushbuf_flush(struct nouveau_pushbuf *); static bool pushbuf_kref_fits(struct nouveau_pushbuf *push, struct nouveau_bo *bo, uint32_t *domains) { struct nouveau_pushbuf_priv *nvpb = nouveau_pushbuf(push); struct nouveau_pushbuf_krec *krec = nvpb->krec; struct nouveau_device *dev = push->client->device; struct nouveau_bo *kbo; struct drm_nouveau_gem_pushbuf_bo *kref; int i; /* VRAM is the only valid domain. GART and VRAM|GART buffers * are all accounted to GART, so if this doesn't fit in VRAM * straight up, a flush is needed. */ if (*domains == NOUVEAU_GEM_DOMAIN_VRAM) { if (krec->vram_used + bo->size > dev->vram_limit) return false; krec->vram_used += bo->size; return true; } /* GART or VRAM|GART buffer. Account both of these buffer types * to GART only for the moment, which simplifies things. If the * buffer can fit already, we're done here. */ if (krec->gart_used + bo->size <= dev->gart_limit) { krec->gart_used += bo->size; return true; } /* Ran out of GART space, if it's a VRAM|GART buffer and it'll * fit into available VRAM, turn it into a VRAM buffer */ if ((*domains & NOUVEAU_GEM_DOMAIN_VRAM) && krec->vram_used + bo->size <= dev->vram_limit) { *domains &= NOUVEAU_GEM_DOMAIN_VRAM; krec->vram_used += bo->size; return true; } /* Still couldn't fit the buffer in anywhere, so as a last resort; * scan the buffer list for VRAM|GART buffers and turn them into * VRAM buffers until we have enough space in GART for this one */ kref = krec->buffer; for (i = 0; i < krec->nr_buffer; i++, kref++) { if (!(kref->valid_domains & NOUVEAU_GEM_DOMAIN_GART)) continue; kbo = (void *)(unsigned long)kref->user_priv; if (!(kref->valid_domains & NOUVEAU_GEM_DOMAIN_VRAM) || krec->vram_used + kbo->size > dev->vram_limit) continue; kref->valid_domains &= NOUVEAU_GEM_DOMAIN_VRAM; krec->gart_used -= kbo->size; krec->vram_used += kbo->size; if (krec->gart_used + bo->size <= dev->gart_limit) { krec->gart_used += bo->size; return true; } } /* Couldn't resolve a placement, need to force a flush */ return false; } static struct drm_nouveau_gem_pushbuf_bo * pushbuf_kref(struct nouveau_pushbuf *push, struct nouveau_bo *bo, uint32_t flags) { struct nouveau_device *dev = push->client->device; struct nouveau_pushbuf_priv *nvpb = nouveau_pushbuf(push); struct nouveau_pushbuf_krec *krec = nvpb->krec; struct nouveau_pushbuf *fpush; struct drm_nouveau_gem_pushbuf_bo *kref; uint32_t domains, domains_wr, domains_rd; domains = 0; if (flags & NOUVEAU_BO_VRAM) domains |= NOUVEAU_GEM_DOMAIN_VRAM; if (flags & NOUVEAU_BO_GART) domains |= NOUVEAU_GEM_DOMAIN_GART; domains_wr = domains * !!(flags & NOUVEAU_BO_WR); domains_rd = domains * !!(flags & NOUVEAU_BO_RD); /* if buffer is referenced on another pushbuf that is owned by the * same client, we need to flush the other pushbuf first to ensure * the correct ordering of commands */ fpush = cli_push_get(push->client, bo); if (fpush && fpush != push) pushbuf_flush(fpush); kref = cli_kref_get(push->client, bo); if (kref) { /* possible conflict in memory types - flush and retry */ if (!(kref->valid_domains & domains)) return NULL; /* VRAM|GART buffer turning into a VRAM buffer. Make sure * it'll fit in VRAM and force a flush if not. */ if ((kref->valid_domains & NOUVEAU_GEM_DOMAIN_GART) && ( domains == NOUVEAU_GEM_DOMAIN_VRAM)) { if (krec->vram_used + bo->size > dev->vram_limit) return NULL; krec->vram_used += bo->size; krec->gart_used -= bo->size; } kref->valid_domains &= domains; kref->write_domains |= domains_wr; kref->read_domains |= domains_rd; } else { if (krec->nr_buffer == NOUVEAU_GEM_MAX_BUFFERS || !pushbuf_kref_fits(push, bo, &domains)) return NULL; kref = &krec->buffer[krec->nr_buffer++]; kref->user_priv = (unsigned long)bo; kref->handle = bo->handle; kref->valid_domains = domains; kref->write_domains = domains_wr; kref->read_domains = domains_rd; kref->presumed.valid = 1; kref->presumed.offset = bo->offset; if (bo->flags & NOUVEAU_BO_VRAM) kref->presumed.domain = NOUVEAU_GEM_DOMAIN_VRAM; else kref->presumed.domain = NOUVEAU_GEM_DOMAIN_GART; cli_kref_set(push->client, bo, kref, push); atomic_inc(&nouveau_bo(bo)->refcnt); } return kref; } static uint32_t pushbuf_krel(struct nouveau_pushbuf *push, struct nouveau_bo *bo, uint32_t data, uint32_t flags, uint32_t vor, uint32_t tor) { struct nouveau_pushbuf_priv *nvpb = nouveau_pushbuf(push); struct nouveau_pushbuf_krec *krec = nvpb->krec; struct drm_nouveau_gem_pushbuf_reloc *krel; struct drm_nouveau_gem_pushbuf_bo *pkref; struct drm_nouveau_gem_pushbuf_bo *bkref; uint32_t reloc = data; pkref = cli_kref_get(push->client, nvpb->bo); bkref = cli_kref_get(push->client, bo); krel = &krec->reloc[krec->nr_reloc++]; assert(pkref); assert(bkref); krel->reloc_bo_index = pkref - krec->buffer; krel->reloc_bo_offset = (push->cur - nvpb->ptr) * 4; krel->bo_index = bkref - krec->buffer; krel->flags = 0; krel->data = data; krel->vor = vor; krel->tor = tor; if (flags & NOUVEAU_BO_LOW) { reloc = (bkref->presumed.offset + data); krel->flags |= NOUVEAU_GEM_RELOC_LOW; } else if (flags & NOUVEAU_BO_HIGH) { reloc = (bkref->presumed.offset + data) >> 32; krel->flags |= NOUVEAU_GEM_RELOC_HIGH; } if (flags & NOUVEAU_BO_OR) { if (bkref->presumed.domain & NOUVEAU_GEM_DOMAIN_VRAM) reloc |= vor; else reloc |= tor; krel->flags |= NOUVEAU_GEM_RELOC_OR; } return reloc; } static void pushbuf_dump(struct nouveau_pushbuf_krec *krec, int krec_id, int chid) { struct drm_nouveau_gem_pushbuf_reloc *krel; struct drm_nouveau_gem_pushbuf_push *kpsh; struct drm_nouveau_gem_pushbuf_bo *kref; struct nouveau_bo *bo; uint32_t *bgn, *end; int i; err("ch%d: krec %d pushes %d bufs %d relocs %d\n", chid, krec_id, krec->nr_push, krec->nr_buffer, krec->nr_reloc); kref = krec->buffer; for (i = 0; i < krec->nr_buffer; i++, kref++) { err("ch%d: buf %08x %08x %08x %08x %08x\n", chid, i, kref->handle, kref->valid_domains, kref->read_domains, kref->write_domains); } krel = krec->reloc; for (i = 0; i < krec->nr_reloc; i++, krel++) { err("ch%d: rel %08x %08x %08x %08x %08x %08x %08x\n", chid, krel->reloc_bo_index, krel->reloc_bo_offset, krel->bo_index, krel->flags, krel->data, krel->vor, krel->tor); } kpsh = krec->push; for (i = 0; i < krec->nr_push; i++, kpsh++) { kref = krec->buffer + kpsh->bo_index; bo = (void *)(unsigned long)kref->user_priv; bgn = (uint32_t *)((char *)bo->map + kpsh->offset); end = bgn + (kpsh->length /4); err("ch%d: psh %08x %010llx %010llx\n", chid, kpsh->bo_index, (unsigned long long)kpsh->offset, (unsigned long long)(kpsh->offset + kpsh->length)); while (bgn < end) err("\t0x%08x\n", *bgn++); } } static int pushbuf_submit(struct nouveau_pushbuf *push, struct nouveau_object *chan) { struct nouveau_pushbuf_priv *nvpb = nouveau_pushbuf(push); struct nouveau_pushbuf_krec *krec = nvpb->list; struct nouveau_device *dev = push->client->device; struct nouveau_drm *drm = nouveau_drm(&dev->object); struct drm_nouveau_gem_pushbuf_bo_presumed *info; struct drm_nouveau_gem_pushbuf_bo *kref; struct drm_nouveau_gem_pushbuf req; struct nouveau_fifo *fifo = chan->data; struct nouveau_bo *bo; int krec_id = 0; int ret = 0, i; if (chan->oclass != NOUVEAU_FIFO_CHANNEL_CLASS) return -EINVAL; if (push->kick_notify) push->kick_notify(push); nouveau_pushbuf_data(push, NULL, 0, 0); while (krec && krec->nr_push) { req.channel = fifo->channel; req.nr_buffers = krec->nr_buffer; req.buffers = (uint64_t)(unsigned long)krec->buffer; req.nr_relocs = krec->nr_reloc; req.nr_push = krec->nr_push; req.relocs = (uint64_t)(unsigned long)krec->reloc; req.push = (uint64_t)(unsigned long)krec->push; req.suffix0 = nvpb->suffix0; req.suffix1 = nvpb->suffix1; req.vram_available = 0; /* for valgrind */ req.gart_available = 0; if (dbg_on(0)) pushbuf_dump(krec, krec_id++, fifo->channel); #ifndef SIMULATE ret = drmCommandWriteRead(drm->fd, DRM_NOUVEAU_GEM_PUSHBUF, &req, sizeof(req)); nvpb->suffix0 = req.suffix0; nvpb->suffix1 = req.suffix1; dev->vram_limit = (req.vram_available * nouveau_device(dev)->vram_limit_percent) / 100; dev->gart_limit = (req.gart_available * nouveau_device(dev)->gart_limit_percent) / 100; #else if (dbg_on(31)) ret = -EINVAL; #endif if (ret) { err("kernel rejected pushbuf: %s\n", strerror(-ret)); pushbuf_dump(krec, krec_id++, fifo->channel); break; } kref = krec->buffer; for (i = 0; i < krec->nr_buffer; i++, kref++) { bo = (void *)(unsigned long)kref->user_priv; info = &kref->presumed; if (!info->valid) { bo->flags &= ~NOUVEAU_BO_APER; if (info->domain == NOUVEAU_GEM_DOMAIN_VRAM) bo->flags |= NOUVEAU_BO_VRAM; else bo->flags |= NOUVEAU_BO_GART; bo->offset = info->offset; } if (kref->write_domains) nouveau_bo(bo)->access |= NOUVEAU_BO_WR; if (kref->read_domains) nouveau_bo(bo)->access |= NOUVEAU_BO_RD; } krec = krec->next; } return ret; } static int pushbuf_flush(struct nouveau_pushbuf *push) { struct nouveau_pushbuf_priv *nvpb = nouveau_pushbuf(push); struct nouveau_pushbuf_krec *krec = nvpb->krec; struct drm_nouveau_gem_pushbuf_bo *kref; struct nouveau_bufctx *bctx, *btmp; struct nouveau_bo *bo; int ret = 0, i; if (push->channel) { ret = pushbuf_submit(push, push->channel); } else { nouveau_pushbuf_data(push, NULL, 0, 0); krec->next = malloc(sizeof(*krec)); nvpb->krec = krec->next; } kref = krec->buffer; for (i = 0; i < krec->nr_buffer; i++, kref++) { bo = (void *)(unsigned long)kref->user_priv; cli_kref_set(push->client, bo, NULL, NULL); if (push->channel) nouveau_bo_ref(NULL, &bo); } krec = nvpb->krec; krec->vram_used = 0; krec->gart_used = 0; krec->nr_buffer = 0; krec->nr_reloc = 0; krec->nr_push = 0; DRMLISTFOREACHENTRYSAFE(bctx, btmp, &nvpb->bctx_list, head) { DRMLISTJOIN(&bctx->current, &bctx->pending); DRMINITLISTHEAD(&bctx->current); DRMLISTDELINIT(&bctx->head); } return ret; } static void pushbuf_refn_fail(struct nouveau_pushbuf *push, int sref, int srel) { struct nouveau_pushbuf_priv *nvpb = nouveau_pushbuf(push); struct nouveau_pushbuf_krec *krec = nvpb->krec; struct drm_nouveau_gem_pushbuf_bo *kref; kref = krec->buffer + sref; while (krec->nr_buffer-- > sref) { struct nouveau_bo *bo = (void *)(unsigned long)kref->user_priv; cli_kref_set(push->client, bo, NULL, NULL); nouveau_bo_ref(NULL, &bo); kref++; } krec->nr_buffer = sref; krec->nr_reloc = srel; } static int pushbuf_refn(struct nouveau_pushbuf *push, bool retry, struct nouveau_pushbuf_refn *refs, int nr) { struct nouveau_pushbuf_priv *nvpb = nouveau_pushbuf(push); struct nouveau_pushbuf_krec *krec = nvpb->krec; struct drm_nouveau_gem_pushbuf_bo *kref; int sref = krec->nr_buffer; int ret = 0, i; for (i = 0; i < nr; i++) { kref = pushbuf_kref(push, refs[i].bo, refs[i].flags); if (!kref) { ret = -ENOSPC; break; } } if (ret) { pushbuf_refn_fail(push, sref, krec->nr_reloc); if (retry) { pushbuf_flush(push); nouveau_pushbuf_space(push, 0, 0, 0); return pushbuf_refn(push, false, refs, nr); } } return ret; } static int pushbuf_validate(struct nouveau_pushbuf *push, bool retry) { struct nouveau_pushbuf_priv *nvpb = nouveau_pushbuf(push); struct nouveau_pushbuf_krec *krec = nvpb->krec; struct drm_nouveau_gem_pushbuf_bo *kref; struct nouveau_bufctx *bctx = push->bufctx; struct nouveau_bufref *bref; int relocs = bctx ? bctx->relocs * 2: 0; int sref, srel, ret; ret = nouveau_pushbuf_space(push, relocs, relocs, 0); if (ret || bctx == NULL) return ret; sref = krec->nr_buffer; srel = krec->nr_reloc; DRMLISTDEL(&bctx->head); DRMLISTADD(&bctx->head, &nvpb->bctx_list); DRMLISTFOREACHENTRY(bref, &bctx->pending, thead) { kref = pushbuf_kref(push, bref->bo, bref->flags); if (!kref) { ret = -ENOSPC; break; } if (bref->packet) { pushbuf_krel(push, bref->bo, bref->packet, 0, 0, 0); *push->cur++ = 0; pushbuf_krel(push, bref->bo, bref->data, bref->flags, bref->vor, bref->tor); *push->cur++ = 0; } } DRMLISTJOIN(&bctx->pending, &bctx->current); DRMINITLISTHEAD(&bctx->pending); if (ret) { pushbuf_refn_fail(push, sref, srel); if (retry) { pushbuf_flush(push); return pushbuf_validate(push, false); } } return ret; } int nouveau_pushbuf_new(struct nouveau_client *client, struct nouveau_object *chan, int nr, uint32_t size, bool immediate, struct nouveau_pushbuf **ppush) { struct nouveau_drm *drm = nouveau_drm(&client->device->object); struct nouveau_fifo *fifo = chan->data; struct nouveau_pushbuf_priv *nvpb; struct nouveau_pushbuf *push; struct drm_nouveau_gem_pushbuf req = {}; int ret; if (chan->oclass != NOUVEAU_FIFO_CHANNEL_CLASS) return -EINVAL; /* nop pushbuf call, to get the current "return to main" sequence * we need to append to the pushbuf on early chipsets */ req.channel = fifo->channel; req.nr_push = 0; ret = drmCommandWriteRead(drm->fd, DRM_NOUVEAU_GEM_PUSHBUF, &req, sizeof(req)); if (ret) return ret; nvpb = calloc(1, sizeof(*nvpb) + nr * sizeof(*nvpb->bos)); if (!nvpb) return -ENOMEM; #ifndef SIMULATE nvpb->suffix0 = req.suffix0; nvpb->suffix1 = req.suffix1; #else nvpb->suffix0 = 0xffffffff; nvpb->suffix1 = 0xffffffff; #endif nvpb->krec = calloc(1, sizeof(*nvpb->krec)); nvpb->list = nvpb->krec; if (!nvpb->krec) { free(nvpb); return -ENOMEM; } push = &nvpb->base; push->client = client; push->channel = immediate ? chan : NULL; push->flags = NOUVEAU_BO_RD; if (fifo->pushbuf & NOUVEAU_GEM_DOMAIN_GART) { push->flags |= NOUVEAU_BO_GART; nvpb->type = NOUVEAU_BO_GART; } else if (fifo->pushbuf & NOUVEAU_GEM_DOMAIN_VRAM) { push->flags |= NOUVEAU_BO_VRAM; nvpb->type = NOUVEAU_BO_VRAM; } nvpb->type |= NOUVEAU_BO_MAP; for (nvpb->bo_nr = 0; nvpb->bo_nr < nr; nvpb->bo_nr++) { ret = nouveau_bo_new(client->device, nvpb->type, 0, size, NULL, &nvpb->bos[nvpb->bo_nr]); if (ret) { nouveau_pushbuf_del(&push); return ret; } } DRMINITLISTHEAD(&nvpb->bctx_list); *ppush = push; return 0; } void nouveau_pushbuf_del(struct nouveau_pushbuf **ppush) { struct nouveau_pushbuf_priv *nvpb = nouveau_pushbuf(*ppush); if (nvpb) { struct drm_nouveau_gem_pushbuf_bo *kref; struct nouveau_pushbuf_krec *krec; while ((krec = nvpb->list)) { kref = krec->buffer; while (krec->nr_buffer--) { unsigned long priv = kref++->user_priv; struct nouveau_bo *bo = (void *)priv; cli_kref_set(nvpb->base.client, bo, NULL, NULL); nouveau_bo_ref(NULL, &bo); } nvpb->list = krec->next; free(krec); } while (nvpb->bo_nr--) nouveau_bo_ref(NULL, &nvpb->bos[nvpb->bo_nr]); nouveau_bo_ref(NULL, &nvpb->bo); free(nvpb); } *ppush = NULL; } struct nouveau_bufctx * nouveau_pushbuf_bufctx(struct nouveau_pushbuf *push, struct nouveau_bufctx *ctx) { struct nouveau_bufctx *prev = push->bufctx; push->bufctx = ctx; return prev; } int nouveau_pushbuf_space(struct nouveau_pushbuf *push, uint32_t dwords, uint32_t relocs, uint32_t pushes) { struct nouveau_pushbuf_priv *nvpb = nouveau_pushbuf(push); struct nouveau_pushbuf_krec *krec = nvpb->krec; struct nouveau_client *client = push->client; struct nouveau_bo *bo = NULL; bool flushed = false; int ret = 0; /* switch to next buffer if insufficient space in the current one */ if (push->cur + dwords >= push->end) { if (nvpb->bo_next < nvpb->bo_nr) { nouveau_bo_ref(nvpb->bos[nvpb->bo_next++], &bo); if (nvpb->bo_next == nvpb->bo_nr && push->channel) nvpb->bo_next = 0; } else { ret = nouveau_bo_new(client->device, nvpb->type, 0, nvpb->bos[0]->size, NULL, &bo); if (ret) return ret; } } /* make sure there's always enough space to queue up the pending * data in the pushbuf proper */ pushes++; /* need to flush if we've run out of space on an immediate pushbuf, * if the new buffer won't fit, or if the kernel push/reloc limits * have been hit */ if ((bo && ( push->channel || !pushbuf_kref(push, bo, push->flags))) || krec->nr_reloc + relocs >= NOUVEAU_GEM_MAX_RELOCS || krec->nr_push + pushes >= NOUVEAU_GEM_MAX_PUSH) { if (nvpb->bo && krec->nr_buffer) pushbuf_flush(push); flushed = true; } /* if necessary, switch to new buffer */ if (bo) { ret = nouveau_bo_map(bo, NOUVEAU_BO_WR, push->client); if (ret) return ret; nouveau_pushbuf_data(push, NULL, 0, 0); nouveau_bo_ref(bo, &nvpb->bo); nouveau_bo_ref(NULL, &bo); nvpb->bgn = nvpb->bo->map; nvpb->ptr = nvpb->bgn; push->cur = nvpb->bgn; push->end = push->cur + (nvpb->bo->size / 4); push->end -= 2 + push->rsvd_kick; /* space for suffix */ } pushbuf_kref(push, nvpb->bo, push->flags); return flushed ? pushbuf_validate(push, false) : 0; } void nouveau_pushbuf_data(struct nouveau_pushbuf *push, struct nouveau_bo *bo, uint64_t offset, uint64_t length) { struct nouveau_pushbuf_priv *nvpb = nouveau_pushbuf(push); struct nouveau_pushbuf_krec *krec = nvpb->krec; struct drm_nouveau_gem_pushbuf_push *kpsh; struct drm_nouveau_gem_pushbuf_bo *kref; if (bo != nvpb->bo && nvpb->bgn != push->cur) { if (nvpb->suffix0 || nvpb->suffix1) { *push->cur++ = nvpb->suffix0; *push->cur++ = nvpb->suffix1; } nouveau_pushbuf_data(push, nvpb->bo, (nvpb->bgn - nvpb->ptr) * 4, (push->cur - nvpb->bgn) * 4); nvpb->bgn = push->cur; } if (bo) { kref = cli_kref_get(push->client, bo); assert(kref); kpsh = &krec->push[krec->nr_push++]; kpsh->bo_index = kref - krec->buffer; kpsh->offset = offset; kpsh->length = length; } } int nouveau_pushbuf_refn(struct nouveau_pushbuf *push, struct nouveau_pushbuf_refn *refs, int nr) { return pushbuf_refn(push, true, refs, nr); } void nouveau_pushbuf_reloc(struct nouveau_pushbuf *push, struct nouveau_bo *bo, uint32_t data, uint32_t flags, uint32_t vor, uint32_t tor) { *push->cur = pushbuf_krel(push, bo, data, flags, vor, tor); push->cur++; } int nouveau_pushbuf_validate(struct nouveau_pushbuf *push) { return pushbuf_validate(push, true); } uint32_t nouveau_pushbuf_refd(struct nouveau_pushbuf *push, struct nouveau_bo *bo) { struct drm_nouveau_gem_pushbuf_bo *kref; uint32_t flags = 0; if (cli_push_get(push->client, bo) == push) { kref = cli_kref_get(push->client, bo); assert(kref); if (kref->read_domains) flags |= NOUVEAU_BO_RD; if (kref->write_domains) flags |= NOUVEAU_BO_WR; } return flags; } int nouveau_pushbuf_kick(struct nouveau_pushbuf *push, struct nouveau_object *chan) { if (!push->channel) return pushbuf_submit(push, chan); pushbuf_flush(push); return pushbuf_validate(push, false); }