431 lines
11 KiB
C
431 lines
11 KiB
C
/*
|
|
* Copyright 2012 Red Hat Inc.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
|
|
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
|
|
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
|
|
* OTHER DEALINGS IN THE SOFTWARE.
|
|
*
|
|
* Authors: Ben Skeggs
|
|
*/
|
|
#include "chan.h"
|
|
|
|
#include <core/client.h>
|
|
#include <core/gpuobj.h>
|
|
#include <core/oproxy.h>
|
|
#include <subdev/mmu.h>
|
|
#include <engine/dma.h>
|
|
|
|
struct nvkm_fifo_chan_object {
|
|
struct nvkm_oproxy oproxy;
|
|
struct nvkm_fifo_chan *chan;
|
|
int hash;
|
|
};
|
|
|
|
static struct nvkm_fifo_engn *
|
|
nvkm_fifo_chan_engn(struct nvkm_fifo_chan *chan, struct nvkm_engine *engine)
|
|
{
|
|
int engi = chan->fifo->func->engine_id(chan->fifo, engine);
|
|
if (engi >= 0)
|
|
return &chan->engn[engi];
|
|
return NULL;
|
|
}
|
|
|
|
static int
|
|
nvkm_fifo_chan_child_fini(struct nvkm_oproxy *base, bool suspend)
|
|
{
|
|
struct nvkm_fifo_chan_object *object =
|
|
container_of(base, typeof(*object), oproxy);
|
|
struct nvkm_engine *engine = object->oproxy.object->engine;
|
|
struct nvkm_fifo_chan *chan = object->chan;
|
|
struct nvkm_fifo_engn *engn = nvkm_fifo_chan_engn(chan, engine);
|
|
const char *name = engine->subdev.name;
|
|
int ret = 0;
|
|
|
|
if (--engn->usecount)
|
|
return 0;
|
|
|
|
if (chan->func->engine_fini) {
|
|
ret = chan->func->engine_fini(chan, engine, suspend);
|
|
if (ret) {
|
|
nvif_error(&chan->object,
|
|
"detach %s failed, %d\n", name, ret);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
if (engn->object) {
|
|
ret = nvkm_object_fini(engn->object, suspend);
|
|
if (ret && suspend)
|
|
return ret;
|
|
}
|
|
|
|
nvif_trace(&chan->object, "detached %s\n", name);
|
|
return ret;
|
|
}
|
|
|
|
static int
|
|
nvkm_fifo_chan_child_init(struct nvkm_oproxy *base)
|
|
{
|
|
struct nvkm_fifo_chan_object *object =
|
|
container_of(base, typeof(*object), oproxy);
|
|
struct nvkm_engine *engine = object->oproxy.object->engine;
|
|
struct nvkm_fifo_chan *chan = object->chan;
|
|
struct nvkm_fifo_engn *engn = nvkm_fifo_chan_engn(chan, engine);
|
|
const char *name = engine->subdev.name;
|
|
int ret;
|
|
|
|
if (engn->usecount++)
|
|
return 0;
|
|
|
|
if (engn->object) {
|
|
ret = nvkm_object_init(engn->object);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
if (chan->func->engine_init) {
|
|
ret = chan->func->engine_init(chan, engine);
|
|
if (ret) {
|
|
nvif_error(&chan->object,
|
|
"attach %s failed, %d\n", name, ret);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
nvif_trace(&chan->object, "attached %s\n", name);
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
nvkm_fifo_chan_child_del(struct nvkm_oproxy *base)
|
|
{
|
|
struct nvkm_fifo_chan_object *object =
|
|
container_of(base, typeof(*object), oproxy);
|
|
struct nvkm_engine *engine = object->oproxy.base.engine;
|
|
struct nvkm_fifo_chan *chan = object->chan;
|
|
struct nvkm_fifo_engn *engn = nvkm_fifo_chan_engn(chan, engine);
|
|
|
|
if (chan->func->object_dtor)
|
|
chan->func->object_dtor(chan, object->hash);
|
|
|
|
if (!--engn->refcount) {
|
|
if (chan->func->engine_dtor)
|
|
chan->func->engine_dtor(chan, engine);
|
|
nvkm_object_del(&engn->object);
|
|
if (chan->vmm)
|
|
atomic_dec(&chan->vmm->engref[engine->subdev.type]);
|
|
}
|
|
}
|
|
|
|
static const struct nvkm_oproxy_func
|
|
nvkm_fifo_chan_child_func = {
|
|
.dtor[0] = nvkm_fifo_chan_child_del,
|
|
.init[0] = nvkm_fifo_chan_child_init,
|
|
.fini[0] = nvkm_fifo_chan_child_fini,
|
|
};
|
|
|
|
static int
|
|
nvkm_fifo_chan_child_new(const struct nvkm_oclass *oclass, void *data, u32 size,
|
|
struct nvkm_object **pobject)
|
|
{
|
|
struct nvkm_engine *engine = oclass->engine;
|
|
struct nvkm_fifo_chan *chan = nvkm_fifo_chan(oclass->parent);
|
|
struct nvkm_fifo_engn *engn = nvkm_fifo_chan_engn(chan, engine);
|
|
struct nvkm_fifo_chan_object *object;
|
|
int ret = 0;
|
|
|
|
if (!(object = kzalloc(sizeof(*object), GFP_KERNEL)))
|
|
return -ENOMEM;
|
|
nvkm_oproxy_ctor(&nvkm_fifo_chan_child_func, oclass, &object->oproxy);
|
|
object->chan = chan;
|
|
*pobject = &object->oproxy.base;
|
|
|
|
if (!engn->refcount++) {
|
|
struct nvkm_oclass cclass = {
|
|
.client = oclass->client,
|
|
.engine = oclass->engine,
|
|
};
|
|
|
|
if (chan->vmm)
|
|
atomic_inc(&chan->vmm->engref[engine->subdev.type]);
|
|
|
|
if (engine->func->fifo.cclass) {
|
|
ret = engine->func->fifo.cclass(chan, &cclass,
|
|
&engn->object);
|
|
} else
|
|
if (engine->func->cclass) {
|
|
ret = nvkm_object_new_(engine->func->cclass, &cclass,
|
|
NULL, 0, &engn->object);
|
|
}
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (chan->func->engine_ctor) {
|
|
ret = chan->func->engine_ctor(chan, oclass->engine,
|
|
engn->object);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
ret = oclass->base.ctor(&(const struct nvkm_oclass) {
|
|
.base = oclass->base,
|
|
.engn = oclass->engn,
|
|
.handle = oclass->handle,
|
|
.object = oclass->object,
|
|
.client = oclass->client,
|
|
.parent = engn->object ?
|
|
engn->object :
|
|
oclass->parent,
|
|
.engine = engine,
|
|
}, data, size, &object->oproxy.object);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (chan->func->object_ctor) {
|
|
object->hash =
|
|
chan->func->object_ctor(chan, object->oproxy.object);
|
|
if (object->hash < 0)
|
|
return object->hash;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
nvkm_fifo_chan_child_get(struct nvkm_object *object, int index,
|
|
struct nvkm_oclass *oclass)
|
|
{
|
|
struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
|
|
struct nvkm_fifo *fifo = chan->fifo;
|
|
struct nvkm_engine *engine;
|
|
u32 engm = chan->engm;
|
|
int engi, ret, c;
|
|
|
|
for (; c = 0, engi = __ffs(engm), engm; engm &= ~(1ULL << engi)) {
|
|
if (!(engine = fifo->func->id_engine(fifo, engi)))
|
|
continue;
|
|
oclass->engine = engine;
|
|
oclass->base.oclass = 0;
|
|
|
|
if (engine->func->fifo.sclass) {
|
|
ret = engine->func->fifo.sclass(oclass, index);
|
|
if (oclass->base.oclass) {
|
|
if (!oclass->base.ctor)
|
|
oclass->base.ctor = nvkm_object_new;
|
|
oclass->ctor = nvkm_fifo_chan_child_new;
|
|
return 0;
|
|
}
|
|
|
|
index -= ret;
|
|
continue;
|
|
}
|
|
|
|
while (engine->func->sclass[c].oclass) {
|
|
if (c++ == index) {
|
|
oclass->base = engine->func->sclass[index];
|
|
if (!oclass->base.ctor)
|
|
oclass->base.ctor = nvkm_object_new;
|
|
oclass->ctor = nvkm_fifo_chan_child_new;
|
|
return 0;
|
|
}
|
|
}
|
|
index -= c;
|
|
}
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
static int
|
|
nvkm_fifo_chan_ntfy(struct nvkm_object *object, u32 type,
|
|
struct nvkm_event **pevent)
|
|
{
|
|
struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
|
|
if (chan->func->ntfy)
|
|
return chan->func->ntfy(chan, type, pevent);
|
|
return -ENODEV;
|
|
}
|
|
|
|
static int
|
|
nvkm_fifo_chan_map(struct nvkm_object *object, void *argv, u32 argc,
|
|
enum nvkm_object_map *type, u64 *addr, u64 *size)
|
|
{
|
|
struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
|
|
*type = NVKM_OBJECT_MAP_IO;
|
|
*addr = chan->addr;
|
|
*size = chan->size;
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
nvkm_fifo_chan_rd32(struct nvkm_object *object, u64 addr, u32 *data)
|
|
{
|
|
struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
|
|
if (unlikely(!chan->user)) {
|
|
chan->user = ioremap(chan->addr, chan->size);
|
|
if (!chan->user)
|
|
return -ENOMEM;
|
|
}
|
|
if (unlikely(addr + 4 > chan->size))
|
|
return -EINVAL;
|
|
*data = ioread32_native(chan->user + addr);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
nvkm_fifo_chan_wr32(struct nvkm_object *object, u64 addr, u32 data)
|
|
{
|
|
struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
|
|
if (unlikely(!chan->user)) {
|
|
chan->user = ioremap(chan->addr, chan->size);
|
|
if (!chan->user)
|
|
return -ENOMEM;
|
|
}
|
|
if (unlikely(addr + 4 > chan->size))
|
|
return -EINVAL;
|
|
iowrite32_native(data, chan->user + addr);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
nvkm_fifo_chan_fini(struct nvkm_object *object, bool suspend)
|
|
{
|
|
struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
|
|
chan->func->fini(chan);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
nvkm_fifo_chan_init(struct nvkm_object *object)
|
|
{
|
|
struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
|
|
chan->func->init(chan);
|
|
return 0;
|
|
}
|
|
|
|
static void *
|
|
nvkm_fifo_chan_dtor(struct nvkm_object *object)
|
|
{
|
|
struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
|
|
struct nvkm_fifo *fifo = chan->fifo;
|
|
void *data = chan->func->dtor(chan);
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&fifo->lock, flags);
|
|
if (!list_empty(&chan->head)) {
|
|
__clear_bit(chan->chid, fifo->mask);
|
|
list_del(&chan->head);
|
|
}
|
|
spin_unlock_irqrestore(&fifo->lock, flags);
|
|
|
|
if (chan->user)
|
|
iounmap(chan->user);
|
|
|
|
if (chan->vmm) {
|
|
nvkm_vmm_part(chan->vmm, chan->inst->memory);
|
|
nvkm_vmm_unref(&chan->vmm);
|
|
}
|
|
|
|
nvkm_gpuobj_del(&chan->push);
|
|
nvkm_gpuobj_del(&chan->inst);
|
|
return data;
|
|
}
|
|
|
|
static const struct nvkm_object_func
|
|
nvkm_fifo_chan_func = {
|
|
.dtor = nvkm_fifo_chan_dtor,
|
|
.init = nvkm_fifo_chan_init,
|
|
.fini = nvkm_fifo_chan_fini,
|
|
.ntfy = nvkm_fifo_chan_ntfy,
|
|
.map = nvkm_fifo_chan_map,
|
|
.rd32 = nvkm_fifo_chan_rd32,
|
|
.wr32 = nvkm_fifo_chan_wr32,
|
|
.sclass = nvkm_fifo_chan_child_get,
|
|
};
|
|
|
|
int
|
|
nvkm_fifo_chan_ctor(const struct nvkm_fifo_chan_func *func,
|
|
struct nvkm_fifo *fifo, u32 size, u32 align, bool zero,
|
|
u64 hvmm, u64 push, u32 engm, int bar, u32 base,
|
|
u32 user, const struct nvkm_oclass *oclass,
|
|
struct nvkm_fifo_chan *chan)
|
|
{
|
|
struct nvkm_client *client = oclass->client;
|
|
struct nvkm_device *device = fifo->engine.subdev.device;
|
|
struct nvkm_dmaobj *dmaobj;
|
|
unsigned long flags;
|
|
int ret;
|
|
|
|
nvkm_object_ctor(&nvkm_fifo_chan_func, oclass, &chan->object);
|
|
chan->func = func;
|
|
chan->fifo = fifo;
|
|
chan->engm = engm;
|
|
INIT_LIST_HEAD(&chan->head);
|
|
|
|
/* instance memory */
|
|
ret = nvkm_gpuobj_new(device, size, align, zero, NULL, &chan->inst);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* allocate push buffer ctxdma instance */
|
|
if (push) {
|
|
dmaobj = nvkm_dmaobj_search(client, push);
|
|
if (IS_ERR(dmaobj))
|
|
return PTR_ERR(dmaobj);
|
|
|
|
ret = nvkm_object_bind(&dmaobj->object, chan->inst, -16,
|
|
&chan->push);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
/* channel address space */
|
|
if (hvmm) {
|
|
struct nvkm_vmm *vmm = nvkm_uvmm_search(client, hvmm);
|
|
if (IS_ERR(vmm))
|
|
return PTR_ERR(vmm);
|
|
|
|
if (vmm->mmu != device->mmu)
|
|
return -EINVAL;
|
|
|
|
ret = nvkm_vmm_join(vmm, chan->inst->memory);
|
|
if (ret)
|
|
return ret;
|
|
|
|
chan->vmm = nvkm_vmm_ref(vmm);
|
|
}
|
|
|
|
/* allocate channel id */
|
|
spin_lock_irqsave(&fifo->lock, flags);
|
|
chan->chid = find_first_zero_bit(fifo->mask, NVKM_FIFO_CHID_NR);
|
|
if (chan->chid >= NVKM_FIFO_CHID_NR) {
|
|
spin_unlock_irqrestore(&fifo->lock, flags);
|
|
return -ENOSPC;
|
|
}
|
|
list_add(&chan->head, &fifo->chan);
|
|
__set_bit(chan->chid, fifo->mask);
|
|
spin_unlock_irqrestore(&fifo->lock, flags);
|
|
|
|
/* determine address of this channel's user registers */
|
|
chan->addr = device->func->resource_addr(device, bar) +
|
|
base + user * chan->chid;
|
|
chan->size = user;
|
|
|
|
nvkm_fifo_cevent(fifo);
|
|
return 0;
|
|
}
|