Loading drivers/gpu/drm/nouveau/include/nvkm/core/subdev.h +0 −3 Original line number Diff line number Diff line Loading @@ -3,9 +3,6 @@ #define __NVKM_SUBDEV_H__ #include <core/device.h> #define nvkm_falcon_cmdq nvkm_msgqueue_queue #define nvkm_falcon_msgq nvkm_msgqueue_queue struct nvkm_subdev { const struct nvkm_subdev_func *func; struct nvkm_device *device; Loading drivers/gpu/drm/nouveau/nvkm/falcon/msgq.c +45 −45 Original line number Diff line number Diff line Loading @@ -23,74 +23,74 @@ #include "qmgr.h" static void msg_queue_open(struct nvkm_msgqueue_queue *queue) nvkm_falcon_msgq_open(struct nvkm_falcon_msgq *msgq) { mutex_lock(&queue->mutex); queue->position = nvkm_falcon_rd32(queue->qmgr->falcon, queue->tail_reg); mutex_lock(&msgq->mutex); msgq->position = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->tail_reg); } static void msg_queue_close(struct nvkm_msgqueue_queue *queue, bool commit) nvkm_falcon_msgq_close(struct nvkm_falcon_msgq *msgq, bool commit) { struct nvkm_falcon *falcon = queue->qmgr->falcon; struct nvkm_falcon *falcon = msgq->qmgr->falcon; if (commit) nvkm_falcon_wr32(falcon, queue->tail_reg, queue->position); nvkm_falcon_wr32(falcon, msgq->tail_reg, msgq->position); mutex_unlock(&queue->mutex); mutex_unlock(&msgq->mutex); } static bool msg_queue_empty(struct nvkm_msgqueue_queue *queue) nvkm_falcon_msgq_empty(struct nvkm_falcon_msgq *msgq) { u32 head = nvkm_falcon_rd32(queue->qmgr->falcon, queue->head_reg); u32 tail = nvkm_falcon_rd32(queue->qmgr->falcon, queue->tail_reg); u32 head = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->head_reg); u32 tail = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->tail_reg); return head == tail; } static int msg_queue_pop(struct nvkm_msgqueue_queue *queue, void *data, u32 size) nvkm_falcon_msgq_pop(struct nvkm_falcon_msgq *msgq, void *data, u32 size) { struct nvkm_falcon *falcon = queue->qmgr->falcon; struct nvkm_falcon *falcon = msgq->qmgr->falcon; u32 head, tail, available; head = nvkm_falcon_rd32(falcon, queue->head_reg); head = nvkm_falcon_rd32(falcon, msgq->head_reg); /* has the buffer looped? */ if (head < queue->position) queue->position = queue->offset; if (head < msgq->position) msgq->position = msgq->offset; tail = queue->position; tail = msgq->position; available = head - tail; if (size > available) { FLCNQ_ERR(queue, "requested %d bytes, but only %d available", FLCNQ_ERR(msgq, "requested %d bytes, but only %d available", size, available); return -EINVAL; } nvkm_falcon_read_dmem(falcon, tail, size, 0, data); queue->position += ALIGN(size, QUEUE_ALIGNMENT); msgq->position += ALIGN(size, QUEUE_ALIGNMENT); return 0; } static int msg_queue_read(struct nvkm_msgqueue_queue *queue, struct nv_falcon_msg *hdr) nvkm_falcon_msgq_read(struct nvkm_falcon_msgq *msgq, struct nv_falcon_msg *hdr) { int ret = 0; msg_queue_open(queue); nvkm_falcon_msgq_open(msgq); if (msg_queue_empty(queue)) if (nvkm_falcon_msgq_empty(msgq)) goto close; ret = msg_queue_pop(queue, hdr, HDR_SIZE); ret = nvkm_falcon_msgq_pop(msgq, hdr, HDR_SIZE); if (ret) { FLCNQ_ERR(queue, "failed to read message header"); FLCNQ_ERR(msgq, "failed to read message header"); goto close; } if (hdr->size > MSG_BUF_SIZE) { FLCNQ_ERR(queue, "message too big, %d bytes", hdr->size); FLCNQ_ERR(msgq, "message too big, %d bytes", hdr->size); ret = -ENOSPC; goto close; } Loading @@ -98,21 +98,21 @@ msg_queue_read(struct nvkm_msgqueue_queue *queue, struct nv_falcon_msg *hdr) if (hdr->size > HDR_SIZE) { u32 read_size = hdr->size - HDR_SIZE; ret = msg_queue_pop(queue, (hdr + 1), read_size); ret = nvkm_falcon_msgq_pop(msgq, (hdr + 1), read_size); if (ret) { FLCNQ_ERR(queue, "failed to read message data"); FLCNQ_ERR(msgq, "failed to read message data"); goto close; } } ret = 1; close: msg_queue_close(queue, (ret >= 0)); nvkm_falcon_msgq_close(msgq, (ret >= 0)); return ret; } static int msgqueue_msg_handle(struct nvkm_falcon_msgq *msgq, struct nv_falcon_msg *hdr) nvkm_falcon_msgq_exec(struct nvkm_falcon_msgq *msgq, struct nv_falcon_msg *hdr) { struct nvkm_falcon_qmgr_seq *seq; Loading @@ -136,6 +136,20 @@ msgqueue_msg_handle(struct nvkm_falcon_msgq *msgq, struct nv_falcon_msg *hdr) return 0; } void nvkm_falcon_msgq_recv(struct nvkm_falcon_msgq *msgq) { /* * We are invoked from a worker thread, so normally we have plenty of * stack space to work with. */ u8 msg_buffer[MSG_BUF_SIZE]; struct nv_falcon_msg *hdr = (void *)msg_buffer; while (nvkm_falcon_msgq_read(msgq, hdr) > 0) nvkm_falcon_msgq_exec(msgq, hdr); } int nvkm_falcon_msgq_recv_initmsg(struct nvkm_falcon_msgq *msgq, void *data, u32 size) Loading @@ -148,31 +162,17 @@ nvkm_falcon_msgq_recv_initmsg(struct nvkm_falcon_msgq *msgq, msgq->tail_reg = falcon->func->msgq.tail; msgq->offset = nvkm_falcon_rd32(falcon, falcon->func->msgq.tail); msg_queue_open(msgq); ret = msg_queue_pop(msgq, data, size); nvkm_falcon_msgq_open(msgq); ret = nvkm_falcon_msgq_pop(msgq, data, size); if (ret == 0 && hdr->size != size) { FLCN_ERR(falcon, "unexpected init message size %d vs %d", hdr->size, size); ret = -EINVAL; } msg_queue_close(msgq, ret == 0); nvkm_falcon_msgq_close(msgq, ret == 0); return ret; } void nvkm_falcon_msgq_recv(struct nvkm_falcon_msgq *queue) { /* * We are invoked from a worker thread, so normally we have plenty of * stack space to work with. */ u8 msg_buffer[MSG_BUF_SIZE]; struct nv_falcon_msg *hdr = (void *)msg_buffer; while (msg_queue_read(queue, hdr) > 0) msgqueue_msg_handle(queue, hdr); } void nvkm_falcon_msgq_init(struct nvkm_falcon_msgq *msgq, u32 index, u32 offset, u32 size) Loading drivers/gpu/drm/nouveau/nvkm/falcon/msgqueue.h +0 −29 Original line number Diff line number Diff line Loading @@ -69,35 +69,6 @@ struct nvkm_msgqueue_func { void (*dtor)(struct nvkm_msgqueue *); }; /** * struct nvkm_msgqueue_queue - information about a command or message queue * * The number of queues is firmware-dependent. All queues must have their * information filled by the init message handler. * * @mutex_lock: to be acquired when the queue is being used * @index: physical queue index * @offset: DMEM offset where this queue begins * @size: size allocated to this queue in DMEM (in bytes) * @position: current write position * @head_reg: address of the HEAD register for this queue * @tail_reg: address of the TAIL register for this queue */ struct nvkm_msgqueue_queue { struct nvkm_falcon_qmgr *qmgr; const char *name; struct mutex mutex; u32 index; u32 offset; u32 size; u32 position; u32 head_reg; u32 tail_reg; struct completion ready; }; /** * struct nvkm_msgqueue - manage a command/message based FW on a falcon * Loading drivers/gpu/drm/nouveau/nvkm/falcon/qmgr.h +26 −0 Original line number Diff line number Diff line Loading @@ -57,6 +57,32 @@ nvkm_falcon_qmgr_seq_acquire(struct nvkm_falcon_qmgr *); void nvkm_falcon_qmgr_seq_release(struct nvkm_falcon_qmgr *, struct nvkm_falcon_qmgr_seq *); struct nvkm_falcon_cmdq { struct nvkm_falcon_qmgr *qmgr; const char *name; struct mutex mutex; struct completion ready; u32 head_reg; u32 tail_reg; u32 offset; u32 size; u32 position; }; struct nvkm_falcon_msgq { struct nvkm_falcon_qmgr *qmgr; const char *name; struct mutex mutex; u32 head_reg; u32 tail_reg; u32 offset; u32 position; }; #define FLCNQ_PRINTK(t,q,f,a...) \ FLCN_PRINTK(t, (q)->qmgr->falcon, "%s: "f, (q)->name, ##a) #define FLCNQ_DBG(q,f,a...) FLCNQ_PRINTK(debug, (q), f, ##a) Loading Loading
drivers/gpu/drm/nouveau/include/nvkm/core/subdev.h +0 −3 Original line number Diff line number Diff line Loading @@ -3,9 +3,6 @@ #define __NVKM_SUBDEV_H__ #include <core/device.h> #define nvkm_falcon_cmdq nvkm_msgqueue_queue #define nvkm_falcon_msgq nvkm_msgqueue_queue struct nvkm_subdev { const struct nvkm_subdev_func *func; struct nvkm_device *device; Loading
drivers/gpu/drm/nouveau/nvkm/falcon/msgq.c +45 −45 Original line number Diff line number Diff line Loading @@ -23,74 +23,74 @@ #include "qmgr.h" static void msg_queue_open(struct nvkm_msgqueue_queue *queue) nvkm_falcon_msgq_open(struct nvkm_falcon_msgq *msgq) { mutex_lock(&queue->mutex); queue->position = nvkm_falcon_rd32(queue->qmgr->falcon, queue->tail_reg); mutex_lock(&msgq->mutex); msgq->position = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->tail_reg); } static void msg_queue_close(struct nvkm_msgqueue_queue *queue, bool commit) nvkm_falcon_msgq_close(struct nvkm_falcon_msgq *msgq, bool commit) { struct nvkm_falcon *falcon = queue->qmgr->falcon; struct nvkm_falcon *falcon = msgq->qmgr->falcon; if (commit) nvkm_falcon_wr32(falcon, queue->tail_reg, queue->position); nvkm_falcon_wr32(falcon, msgq->tail_reg, msgq->position); mutex_unlock(&queue->mutex); mutex_unlock(&msgq->mutex); } static bool msg_queue_empty(struct nvkm_msgqueue_queue *queue) nvkm_falcon_msgq_empty(struct nvkm_falcon_msgq *msgq) { u32 head = nvkm_falcon_rd32(queue->qmgr->falcon, queue->head_reg); u32 tail = nvkm_falcon_rd32(queue->qmgr->falcon, queue->tail_reg); u32 head = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->head_reg); u32 tail = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->tail_reg); return head == tail; } static int msg_queue_pop(struct nvkm_msgqueue_queue *queue, void *data, u32 size) nvkm_falcon_msgq_pop(struct nvkm_falcon_msgq *msgq, void *data, u32 size) { struct nvkm_falcon *falcon = queue->qmgr->falcon; struct nvkm_falcon *falcon = msgq->qmgr->falcon; u32 head, tail, available; head = nvkm_falcon_rd32(falcon, queue->head_reg); head = nvkm_falcon_rd32(falcon, msgq->head_reg); /* has the buffer looped? */ if (head < queue->position) queue->position = queue->offset; if (head < msgq->position) msgq->position = msgq->offset; tail = queue->position; tail = msgq->position; available = head - tail; if (size > available) { FLCNQ_ERR(queue, "requested %d bytes, but only %d available", FLCNQ_ERR(msgq, "requested %d bytes, but only %d available", size, available); return -EINVAL; } nvkm_falcon_read_dmem(falcon, tail, size, 0, data); queue->position += ALIGN(size, QUEUE_ALIGNMENT); msgq->position += ALIGN(size, QUEUE_ALIGNMENT); return 0; } static int msg_queue_read(struct nvkm_msgqueue_queue *queue, struct nv_falcon_msg *hdr) nvkm_falcon_msgq_read(struct nvkm_falcon_msgq *msgq, struct nv_falcon_msg *hdr) { int ret = 0; msg_queue_open(queue); nvkm_falcon_msgq_open(msgq); if (msg_queue_empty(queue)) if (nvkm_falcon_msgq_empty(msgq)) goto close; ret = msg_queue_pop(queue, hdr, HDR_SIZE); ret = nvkm_falcon_msgq_pop(msgq, hdr, HDR_SIZE); if (ret) { FLCNQ_ERR(queue, "failed to read message header"); FLCNQ_ERR(msgq, "failed to read message header"); goto close; } if (hdr->size > MSG_BUF_SIZE) { FLCNQ_ERR(queue, "message too big, %d bytes", hdr->size); FLCNQ_ERR(msgq, "message too big, %d bytes", hdr->size); ret = -ENOSPC; goto close; } Loading @@ -98,21 +98,21 @@ msg_queue_read(struct nvkm_msgqueue_queue *queue, struct nv_falcon_msg *hdr) if (hdr->size > HDR_SIZE) { u32 read_size = hdr->size - HDR_SIZE; ret = msg_queue_pop(queue, (hdr + 1), read_size); ret = nvkm_falcon_msgq_pop(msgq, (hdr + 1), read_size); if (ret) { FLCNQ_ERR(queue, "failed to read message data"); FLCNQ_ERR(msgq, "failed to read message data"); goto close; } } ret = 1; close: msg_queue_close(queue, (ret >= 0)); nvkm_falcon_msgq_close(msgq, (ret >= 0)); return ret; } static int msgqueue_msg_handle(struct nvkm_falcon_msgq *msgq, struct nv_falcon_msg *hdr) nvkm_falcon_msgq_exec(struct nvkm_falcon_msgq *msgq, struct nv_falcon_msg *hdr) { struct nvkm_falcon_qmgr_seq *seq; Loading @@ -136,6 +136,20 @@ msgqueue_msg_handle(struct nvkm_falcon_msgq *msgq, struct nv_falcon_msg *hdr) return 0; } void nvkm_falcon_msgq_recv(struct nvkm_falcon_msgq *msgq) { /* * We are invoked from a worker thread, so normally we have plenty of * stack space to work with. */ u8 msg_buffer[MSG_BUF_SIZE]; struct nv_falcon_msg *hdr = (void *)msg_buffer; while (nvkm_falcon_msgq_read(msgq, hdr) > 0) nvkm_falcon_msgq_exec(msgq, hdr); } int nvkm_falcon_msgq_recv_initmsg(struct nvkm_falcon_msgq *msgq, void *data, u32 size) Loading @@ -148,31 +162,17 @@ nvkm_falcon_msgq_recv_initmsg(struct nvkm_falcon_msgq *msgq, msgq->tail_reg = falcon->func->msgq.tail; msgq->offset = nvkm_falcon_rd32(falcon, falcon->func->msgq.tail); msg_queue_open(msgq); ret = msg_queue_pop(msgq, data, size); nvkm_falcon_msgq_open(msgq); ret = nvkm_falcon_msgq_pop(msgq, data, size); if (ret == 0 && hdr->size != size) { FLCN_ERR(falcon, "unexpected init message size %d vs %d", hdr->size, size); ret = -EINVAL; } msg_queue_close(msgq, ret == 0); nvkm_falcon_msgq_close(msgq, ret == 0); return ret; } void nvkm_falcon_msgq_recv(struct nvkm_falcon_msgq *queue) { /* * We are invoked from a worker thread, so normally we have plenty of * stack space to work with. */ u8 msg_buffer[MSG_BUF_SIZE]; struct nv_falcon_msg *hdr = (void *)msg_buffer; while (msg_queue_read(queue, hdr) > 0) msgqueue_msg_handle(queue, hdr); } void nvkm_falcon_msgq_init(struct nvkm_falcon_msgq *msgq, u32 index, u32 offset, u32 size) Loading
drivers/gpu/drm/nouveau/nvkm/falcon/msgqueue.h +0 −29 Original line number Diff line number Diff line Loading @@ -69,35 +69,6 @@ struct nvkm_msgqueue_func { void (*dtor)(struct nvkm_msgqueue *); }; /** * struct nvkm_msgqueue_queue - information about a command or message queue * * The number of queues is firmware-dependent. All queues must have their * information filled by the init message handler. * * @mutex_lock: to be acquired when the queue is being used * @index: physical queue index * @offset: DMEM offset where this queue begins * @size: size allocated to this queue in DMEM (in bytes) * @position: current write position * @head_reg: address of the HEAD register for this queue * @tail_reg: address of the TAIL register for this queue */ struct nvkm_msgqueue_queue { struct nvkm_falcon_qmgr *qmgr; const char *name; struct mutex mutex; u32 index; u32 offset; u32 size; u32 position; u32 head_reg; u32 tail_reg; struct completion ready; }; /** * struct nvkm_msgqueue - manage a command/message based FW on a falcon * Loading
drivers/gpu/drm/nouveau/nvkm/falcon/qmgr.h +26 −0 Original line number Diff line number Diff line Loading @@ -57,6 +57,32 @@ nvkm_falcon_qmgr_seq_acquire(struct nvkm_falcon_qmgr *); void nvkm_falcon_qmgr_seq_release(struct nvkm_falcon_qmgr *, struct nvkm_falcon_qmgr_seq *); struct nvkm_falcon_cmdq { struct nvkm_falcon_qmgr *qmgr; const char *name; struct mutex mutex; struct completion ready; u32 head_reg; u32 tail_reg; u32 offset; u32 size; u32 position; }; struct nvkm_falcon_msgq { struct nvkm_falcon_qmgr *qmgr; const char *name; struct mutex mutex; u32 head_reg; u32 tail_reg; u32 offset; u32 position; }; #define FLCNQ_PRINTK(t,q,f,a...) \ FLCN_PRINTK(t, (q)->qmgr->falcon, "%s: "f, (q)->name, ##a) #define FLCNQ_DBG(q,f,a...) FLCNQ_PRINTK(debug, (q), f, ##a) Loading