is_eos = (byte_size == 0);
/* Validate args */
- if (!host_buf || !pchnl) {
- status = -EFAULT;
- } else if (is_eos && CHNL_IS_INPUT(pchnl->chnl_mode)) {
- status = -EPERM;
- } else {
- /*
- * Check the channel state: only queue chirp if channel state
- * allows it.
- */
- dw_state = pchnl->dw_state;
- if (dw_state != CHNL_STATEREADY) {
- if (dw_state & CHNL_STATECANCEL)
- status = -ECANCELED;
- else if ((dw_state & CHNL_STATEEOS) &&
- CHNL_IS_OUTPUT(pchnl->chnl_mode))
- status = -EPIPE;
- else
- /* No other possible states left */
- DBC_ASSERT(0);
- }
+ if (!host_buf || !pchnl)
+ return -EFAULT;
+
+ if (is_eos && CHNL_IS_INPUT(pchnl->chnl_mode))
+ return -EPERM;
+
+ /*
+ * Check the channel state: only queue chirp if channel state
+ * allows it.
+ */
+ dw_state = pchnl->dw_state;
+ if (dw_state != CHNL_STATEREADY) {
+ if (dw_state & CHNL_STATECANCEL)
+ return -ECANCELED;
+ if ((dw_state & CHNL_STATEEOS) &&
+ CHNL_IS_OUTPUT(pchnl->chnl_mode))
+ return -EPIPE;
+ /* No other possible states left */
+ DBC_ASSERT(0);
}
dev_obj = dev_get_first();
dev_get_bridge_context(dev_obj, &dev_ctxt);
if (!dev_ctxt)
- status = -EFAULT;
-
- if (status)
- goto func_end;
+ return -EFAULT;
if (pchnl->chnl_type == CHNL_PCPY && pchnl->chnl_id > 1 && host_buf) {
if (!(host_buf < (void *)USERMODE_ADDR)) {
}
/* if addr in user mode, then copy to kernel space */
host_sys_buf = kmalloc(buf_size, GFP_KERNEL);
- if (host_sys_buf == NULL) {
- status = -ENOMEM;
- goto func_end;
- }
+ if (host_sys_buf == NULL)
+ return -ENOMEM;
+
if (CHNL_IS_OUTPUT(pchnl->chnl_mode)) {
status = copy_from_user(host_sys_buf, host_buf,
- buf_size);
+ buf_size);
if (status) {
kfree(host_sys_buf);
host_sys_buf = NULL;
- status = -EFAULT;
- goto func_end;
+ return -EFAULT;
}
}
}
omap_mbox_disable_irq(dev_ctxt->mbox, IRQ_RX);
if (pchnl->chnl_type == CHNL_PCPY) {
/* This is a processor-copy channel. */
- if (!status && CHNL_IS_OUTPUT(pchnl->chnl_mode)) {
+ if (CHNL_IS_OUTPUT(pchnl->chnl_mode)) {
/* Check buffer size on output channels for fit. */
- if (byte_size >
- io_buf_size(pchnl->chnl_mgr_obj->hio_mgr))
+ if (byte_size > io_buf_size(
+ pchnl->chnl_mgr_obj->hio_mgr)) {
status = -EINVAL;
-
- }
- }
- if (!status) {
- /* Get a free chirp: */
- if (!list_empty(&pchnl->free_packets_list)) {
- chnl_packet_obj = list_first_entry(
- &pchnl->free_packets_list,
- struct chnl_irp, link);
- list_del(&chnl_packet_obj->link);
- } else {
- status = -EIO;
+ goto out;
+ }
}
-
}
- if (!status) {
- /* Enqueue the chirp on the chnl's IORequest queue: */
- chnl_packet_obj->host_user_buf = chnl_packet_obj->host_sys_buf =
- host_buf;
- if (pchnl->chnl_type == CHNL_PCPY && pchnl->chnl_id > 1)
- chnl_packet_obj->host_sys_buf = host_sys_buf;
-
- /*
- * Note: for dma chans dw_dsp_addr contains dsp address
- * of SM buffer.
- */
- DBC_ASSERT(chnl_mgr_obj->word_size != 0);
- /* DSP address */
- chnl_packet_obj->dsp_tx_addr =
- dw_dsp_addr / chnl_mgr_obj->word_size;
- chnl_packet_obj->byte_size = byte_size;
- chnl_packet_obj->buf_size = buf_size;
- /* Only valid for output channel */
- chnl_packet_obj->dw_arg = dw_arg;
- chnl_packet_obj->status = (is_eos ? CHNL_IOCSTATEOS :
- CHNL_IOCSTATCOMPLETE);
- list_add_tail(&chnl_packet_obj->link, &pchnl->pio_requests);
- pchnl->cio_reqs++;
- DBC_ASSERT(pchnl->cio_reqs <= pchnl->chnl_packets);
- /*
- * If end of stream, update the channel state to prevent
- * more IOR's.
- */
- if (is_eos)
- pchnl->dw_state |= CHNL_STATEEOS;
- /* Legacy DSM Processor-Copy */
- DBC_ASSERT(pchnl->chnl_type == CHNL_PCPY);
- /* Request IO from the DSP */
- io_request_chnl(chnl_mgr_obj->hio_mgr, pchnl,
- (CHNL_IS_INPUT(pchnl->chnl_mode) ? IO_INPUT :
- IO_OUTPUT), &mb_val);
- sched_dpc = true;
-
- }
+ /* Get a free chirp: */
+ if (list_empty(&pchnl->free_packets_list)) {
+ status = -EIO;
+ goto out;
+ }
+ chnl_packet_obj = list_first_entry(&pchnl->free_packets_list,
+ struct chnl_irp, link);
+ list_del(&chnl_packet_obj->link);
+
+ /* Enqueue the chirp on the chnl's IORequest queue: */
+ chnl_packet_obj->host_user_buf = chnl_packet_obj->host_sys_buf =
+ host_buf;
+ if (pchnl->chnl_type == CHNL_PCPY && pchnl->chnl_id > 1)
+ chnl_packet_obj->host_sys_buf = host_sys_buf;
+
+ /*
+ * Note: for dma chans dw_dsp_addr contains dsp address
+ * of SM buffer.
+ */
+ DBC_ASSERT(chnl_mgr_obj->word_size != 0);
+ /* DSP address */
+ chnl_packet_obj->dsp_tx_addr = dw_dsp_addr / chnl_mgr_obj->word_size;
+ chnl_packet_obj->byte_size = byte_size;
+ chnl_packet_obj->buf_size = buf_size;
+ /* Only valid for output channel */
+ chnl_packet_obj->dw_arg = dw_arg;
+ chnl_packet_obj->status = (is_eos ? CHNL_IOCSTATEOS :
+ CHNL_IOCSTATCOMPLETE);
+ list_add_tail(&chnl_packet_obj->link, &pchnl->pio_requests);
+ pchnl->cio_reqs++;
+ DBC_ASSERT(pchnl->cio_reqs <= pchnl->chnl_packets);
+ /*
+ * If end of stream, update the channel state to prevent
+ * more IOR's.
+ */
+ if (is_eos)
+ pchnl->dw_state |= CHNL_STATEEOS;
+
+ /* Legacy DSM Processor-Copy */
+ DBC_ASSERT(pchnl->chnl_type == CHNL_PCPY);
+ /* Request IO from the DSP */
+ io_request_chnl(chnl_mgr_obj->hio_mgr, pchnl,
+ (CHNL_IS_INPUT(pchnl->chnl_mode) ? IO_INPUT :
+ IO_OUTPUT), &mb_val);
+ sched_dpc = true;
+out:
omap_mbox_enable_irq(dev_ctxt->mbox, IRQ_RX);
spin_unlock_bh(&chnl_mgr_obj->chnl_mgr_lock);
if (mb_val != 0)
if (sched_dpc)
iosm_schedule(chnl_mgr_obj->hio_mgr);
-func_end:
return status;
}
*/
int bridge_chnl_cancel_io(struct chnl_object *chnl_obj)
{
- int status = 0;
struct chnl_object *pchnl = (struct chnl_object *)chnl_obj;
u32 chnl_id = -1;
s8 chnl_mode;
struct chnl_mgr *chnl_mgr_obj = NULL;
/* Check args: */
- if (pchnl && pchnl->chnl_mgr_obj) {
- chnl_id = pchnl->chnl_id;
- chnl_mode = pchnl->chnl_mode;
- chnl_mgr_obj = pchnl->chnl_mgr_obj;
- } else {
- status = -EFAULT;
- }
- if (status)
- goto func_end;
+ if (!pchnl || !pchnl->chnl_mgr_obj)
+ return -EFAULT;
+
+ chnl_id = pchnl->chnl_id;
+ chnl_mode = pchnl->chnl_mode;
+ chnl_mgr_obj = pchnl->chnl_mgr_obj;
/* Mark this channel as cancelled, to prevent further IORequests or
* IORequests or dispatching. */
spin_lock_bh(&chnl_mgr_obj->chnl_mgr_lock);
+
pchnl->dw_state |= CHNL_STATECANCEL;
- if (list_empty(&pchnl->pio_requests))
- goto func_cont;
+
+ if (list_empty(&pchnl->pio_requests)) {
+ spin_unlock_bh(&chnl_mgr_obj->chnl_mgr_lock);
+ return 0;
+ }
if (pchnl->chnl_type == CHNL_PCPY) {
/* Indicate we have no more buffers available for transfer: */
pchnl->cio_reqs--;
DBC_ASSERT(pchnl->cio_reqs >= 0);
}
-func_cont:
+
spin_unlock_bh(&chnl_mgr_obj->chnl_mgr_lock);
-func_end:
- return status;
+
+ return 0;
}
/*
struct chnl_object *pchnl = (struct chnl_object *)chnl_obj;
/* Check args: */
- if (!pchnl) {
- status = -EFAULT;
- goto func_cont;
+ if (!pchnl)
+ return -EFAULT;
+ /* Cancel IO: this ensures no further IO requests or notifications */
+ status = bridge_chnl_cancel_io(chnl_obj);
+ if (status)
+ return status;
+ /* Assert I/O on this channel is now cancelled: Protects from io_dpc */
+ DBC_ASSERT((pchnl->dw_state & CHNL_STATECANCEL));
+ /* Invalidate channel object: Protects from CHNL_GetIOCompletion() */
+ /* Free the slot in the channel manager: */
+ pchnl->chnl_mgr_obj->ap_channel[pchnl->chnl_id] = NULL;
+ spin_lock_bh(&pchnl->chnl_mgr_obj->chnl_mgr_lock);
+ pchnl->chnl_mgr_obj->open_channels -= 1;
+ spin_unlock_bh(&pchnl->chnl_mgr_obj->chnl_mgr_lock);
+ if (pchnl->ntfy_obj) {
+ ntfy_delete(pchnl->ntfy_obj);
+ kfree(pchnl->ntfy_obj);
+ pchnl->ntfy_obj = NULL;
}
- {
- /* Cancel IO: this ensures no further IO requests or
- * notifications. */
- status = bridge_chnl_cancel_io(chnl_obj);
+ /* Reset channel event: (NOTE: user_event freed in user context) */
+ if (pchnl->sync_event) {
+ sync_reset_event(pchnl->sync_event);
+ kfree(pchnl->sync_event);
+ pchnl->sync_event = NULL;
}
-func_cont:
- if (!status) {
- /* Assert I/O on this channel is now cancelled: Protects
- * from io_dpc. */
- DBC_ASSERT((pchnl->dw_state & CHNL_STATECANCEL));
- /* Invalidate channel object: Protects from
- * CHNL_GetIOCompletion(). */
- /* Free the slot in the channel manager: */
- pchnl->chnl_mgr_obj->ap_channel[pchnl->chnl_id] = NULL;
- spin_lock_bh(&pchnl->chnl_mgr_obj->chnl_mgr_lock);
- pchnl->chnl_mgr_obj->open_channels -= 1;
- spin_unlock_bh(&pchnl->chnl_mgr_obj->chnl_mgr_lock);
- if (pchnl->ntfy_obj) {
- ntfy_delete(pchnl->ntfy_obj);
- kfree(pchnl->ntfy_obj);
- pchnl->ntfy_obj = NULL;
- }
- /* Reset channel event: (NOTE: user_event freed in user
- * context.). */
- if (pchnl->sync_event) {
- sync_reset_event(pchnl->sync_event);
- kfree(pchnl->sync_event);
- pchnl->sync_event = NULL;
- }
- /* Free I/O request and I/O completion queues: */
- free_chirp_list(&pchnl->pio_completions);
- pchnl->cio_cs = 0;
+ /* Free I/O request and I/O completion queues: */
+ free_chirp_list(&pchnl->pio_completions);
+ pchnl->cio_cs = 0;
- free_chirp_list(&pchnl->pio_requests);
- pchnl->cio_reqs = 0;
+ free_chirp_list(&pchnl->pio_requests);
+ pchnl->cio_reqs = 0;
- free_chirp_list(&pchnl->free_packets_list);
+ free_chirp_list(&pchnl->free_packets_list);
+
+ /* Release channel object. */
+ kfree(pchnl);
- /* Release channel object. */
- kfree(pchnl);
- pchnl = NULL;
- }
- DBC_ENSURE(status || !pchnl);
return status;
}
int bridge_chnl_get_mgr_info(struct chnl_mgr *hchnl_mgr, u32 ch_id,
struct chnl_mgrinfo *mgr_info)
{
- int status = 0;
struct chnl_mgr *chnl_mgr_obj = (struct chnl_mgr *)hchnl_mgr;
- if (mgr_info != NULL) {
- if (ch_id <= CHNL_MAXCHANNELS) {
- if (hchnl_mgr) {
- /* Return the requested information: */
- mgr_info->chnl_obj =
- chnl_mgr_obj->ap_channel[ch_id];
- mgr_info->open_channels =
- chnl_mgr_obj->open_channels;
- mgr_info->dw_type = chnl_mgr_obj->dw_type;
- /* total # of chnls */
- mgr_info->max_channels =
- chnl_mgr_obj->max_channels;
- } else {
- status = -EFAULT;
- }
- } else {
- status = -ECHRNG;
- }
- } else {
- status = -EFAULT;
- }
+ if (!mgr_info || !hchnl_mgr)
+ return -EFAULT;
- return status;
+ if (ch_id > CHNL_MAXCHANNELS)
+ return -ECHRNG;
+
+ /* Return the requested information: */
+ mgr_info->chnl_obj = chnl_mgr_obj->ap_channel[ch_id];
+ mgr_info->open_channels = chnl_mgr_obj->open_channels;
+ mgr_info->dw_type = chnl_mgr_obj->dw_type;
+ /* total # of chnls */
+ mgr_info->max_channels = chnl_mgr_obj->max_channels;
+
+ return 0;
}
/*
DBC_REQUIRE(pattrs != NULL);
DBC_REQUIRE(hchnl_mgr != NULL);
*chnl = NULL;
+
/* Validate Args: */
- if (pattrs->uio_reqs == 0) {
- status = -EINVAL;
+ if (!pattrs->uio_reqs)
+ return -EINVAL;
+
+ if (!hchnl_mgr)
+ return -EFAULT;
+
+ if (ch_id != CHNL_PICKFREE) {
+ if (ch_id >= chnl_mgr_obj->max_channels)
+ return -ECHRNG;
+ if (chnl_mgr_obj->ap_channel[ch_id] != NULL)
+ return -EALREADY;
} else {
- if (!hchnl_mgr) {
- status = -EFAULT;
- } else {
- if (ch_id != CHNL_PICKFREE) {
- if (ch_id >= chnl_mgr_obj->max_channels)
- status = -ECHRNG;
- else if (chnl_mgr_obj->ap_channel[ch_id] !=
- NULL)
- status = -EALREADY;
- } else {
- /* Check for free channel */
- status =
- search_free_channel(chnl_mgr_obj, &ch_id);
- }
- }
+ /* Check for free channel */
+ status = search_free_channel(chnl_mgr_obj, &ch_id);
+ if (status)
+ return status;
}
- if (status)
- goto func_end;
DBC_ASSERT(ch_id < chnl_mgr_obj->max_channels);
+
/* Create channel object: */
pchnl = kzalloc(sizeof(struct chnl_object), GFP_KERNEL);
- if (!pchnl) {
- status = -ENOMEM;
- goto func_end;
- }
+ if (!pchnl)
+ return -ENOMEM;
+
/* Protect queues from io_dpc: */
pchnl->dw_state = CHNL_STATECANCEL;
+
/* Allocate initial IOR and IOC queues: */
status = create_chirp_list(&pchnl->free_packets_list,
pattrs->uio_reqs);
- if (status) {
- kfree(pchnl);
- goto func_end;
- }
+ if (status)
+ goto out_err;
INIT_LIST_HEAD(&pchnl->pio_requests);
INIT_LIST_HEAD(&pchnl->pio_completions);
pchnl->chnl_packets = pattrs->uio_reqs;
pchnl->cio_cs = 0;
pchnl->cio_reqs = 0;
+
sync_event = kzalloc(sizeof(struct sync_object), GFP_KERNEL);
- if (sync_event)
- sync_init_event(sync_event);
- else
+ if (!sync_event) {
status = -ENOMEM;
-
- if (!status) {
- pchnl->ntfy_obj = kmalloc(sizeof(struct ntfy_object),
- GFP_KERNEL);
- if (pchnl->ntfy_obj)
- ntfy_init(pchnl->ntfy_obj);
- else
- status = -ENOMEM;
+ goto out_err;
}
+ sync_init_event(sync_event);
- if (!status) {
- /* Initialize CHNL object fields: */
- pchnl->chnl_mgr_obj = chnl_mgr_obj;
- pchnl->chnl_id = ch_id;
- pchnl->chnl_mode = chnl_mode;
- pchnl->user_event = sync_event;
- pchnl->sync_event = sync_event;
- /* Get the process handle */
- pchnl->process = current->tgid;
- pchnl->pcb_arg = 0;
- pchnl->bytes_moved = 0;
- /* Default to proc-copy */
- pchnl->chnl_type = CHNL_PCPY;
- }
+ pchnl->ntfy_obj = kmalloc(sizeof(struct ntfy_object), GFP_KERNEL);
+ if (!pchnl->ntfy_obj) {
+ status = -ENOMEM;
+ goto out_err;
+ }
+ ntfy_init(pchnl->ntfy_obj);
+
+ /* Initialize CHNL object fields: */
+ pchnl->chnl_mgr_obj = chnl_mgr_obj;
+ pchnl->chnl_id = ch_id;
+ pchnl->chnl_mode = chnl_mode;
+ pchnl->user_event = sync_event;
+ pchnl->sync_event = sync_event;
+ /* Get the process handle */
+ pchnl->process = current->tgid;
+ pchnl->pcb_arg = 0;
+ pchnl->bytes_moved = 0;
+ /* Default to proc-copy */
+ pchnl->chnl_type = CHNL_PCPY;
+
+ /* Insert channel object in channel manager: */
+ chnl_mgr_obj->ap_channel[pchnl->chnl_id] = pchnl;
+ spin_lock_bh(&chnl_mgr_obj->chnl_mgr_lock);
+ chnl_mgr_obj->open_channels++;
+ spin_unlock_bh(&chnl_mgr_obj->chnl_mgr_lock);
+ /* Return result... */
+ pchnl->dw_state = CHNL_STATEREADY;
+ *chnl = pchnl;
- if (status) {
- /* Free memory */
- free_chirp_list(&pchnl->pio_completions);
- pchnl->cio_cs = 0;
- free_chirp_list(&pchnl->pio_requests);
- free_chirp_list(&pchnl->free_packets_list);
+ return status;
+
+out_err:
+ /* Free memory */
+ free_chirp_list(&pchnl->pio_completions);
+ free_chirp_list(&pchnl->pio_requests);
+ free_chirp_list(&pchnl->free_packets_list);
+
+ if (sync_event)
kfree(sync_event);
- sync_event = NULL;
- if (pchnl->ntfy_obj) {
- ntfy_delete(pchnl->ntfy_obj);
- kfree(pchnl->ntfy_obj);
- pchnl->ntfy_obj = NULL;
- }
- kfree(pchnl);
- } else {
- /* Insert channel object in channel manager: */
- chnl_mgr_obj->ap_channel[pchnl->chnl_id] = pchnl;
- spin_lock_bh(&chnl_mgr_obj->chnl_mgr_lock);
- chnl_mgr_obj->open_channels++;
- spin_unlock_bh(&chnl_mgr_obj->chnl_mgr_lock);
- /* Return result... */
- pchnl->dw_state = CHNL_STATEREADY;
- *chnl = pchnl;
+ if (pchnl->ntfy_obj) {
+ ntfy_delete(pchnl->ntfy_obj);
+ kfree(pchnl->ntfy_obj);
+ pchnl->ntfy_obj = NULL;
}
-func_end:
- DBC_ENSURE((!status && pchnl) || (*chnl == NULL));
+ kfree(pchnl);
+
return status;
}
input_empty = msg_ctr_obj->buf_empty;
num_msgs = msg_ctr_obj->size;
if (input_empty)
- goto func_end;
+ return;
msg_input = pio_mgr->msg_input;
for (i = 0; i < num_msgs; i++) {
/* Read the next message */
addr = (u32) &(((struct msg_dspmsg *)msg_input)->msg.dw_cmd);
msg.msg.dw_cmd =
- read_ext32_bit_dsp_data(pio_mgr->hbridge_context, addr);
+ read_ext32_bit_dsp_data(pio_mgr->hbridge_context, addr);
addr = (u32) &(((struct msg_dspmsg *)msg_input)->msg.dw_arg1);
msg.msg.dw_arg1 =
- read_ext32_bit_dsp_data(pio_mgr->hbridge_context, addr);
+ read_ext32_bit_dsp_data(pio_mgr->hbridge_context, addr);
addr = (u32) &(((struct msg_dspmsg *)msg_input)->msg.dw_arg2);
msg.msg.dw_arg2 =
- read_ext32_bit_dsp_data(pio_mgr->hbridge_context, addr);
+ read_ext32_bit_dsp_data(pio_mgr->hbridge_context, addr);
addr = (u32) &(((struct msg_dspmsg *)msg_input)->msgq_id);
msg.msgq_id =
- read_ext32_bit_dsp_data(pio_mgr->hbridge_context, addr);
+ read_ext32_bit_dsp_data(pio_mgr->hbridge_context, addr);
msg_input += sizeof(struct msg_dspmsg);
/* Determine which queue to put the message in */
dev_dbg(bridge, "input msg: dw_cmd=0x%x dw_arg1=0x%x "
- "dw_arg2=0x%x msgq_id=0x%x\n", msg.msg.dw_cmd,
- msg.msg.dw_arg1, msg.msg.dw_arg2, msg.msgq_id);
+ "dw_arg2=0x%x msgq_id=0x%x\n", msg.msg.dw_cmd,
+ msg.msg.dw_arg1, msg.msg.dw_arg2, msg.msgq_id);
/*
* Interrupt may occur before shared memory and message
* input locations have been set up. If all nodes were
*/
list_for_each_entry(msg_queue_obj, &hmsg_mgr->queue_list,
list_elem) {
- if (msg.msgq_id == msg_queue_obj->msgq_id) {
- /* Found it */
- if (msg.msg.dw_cmd == RMS_EXITACK) {
- /*
- * Call the node exit notification.
- * The exit message does not get
- * queued.
- */
- (*hmsg_mgr->on_exit) ((void *)
- msg_queue_obj->arg,
- msg.msg.dw_arg1);
- break;
- }
+ if (msg.msgq_id != msg_queue_obj->msgq_id)
+ continue;
+ /* Found it */
+ if (msg.msg.dw_cmd == RMS_EXITACK) {
/*
- * Not an exit acknowledgement, queue
- * the message.
+ * Call the node exit notification.
+ * The exit message does not get
+ * queued.
*/
- if (!list_empty(&msg_queue_obj->
- msg_free_list)) {
- pmsg = list_first_entry(
- &msg_queue_obj->msg_free_list,
- struct msg_frame, list_elem);
- list_del(&pmsg->list_elem);
- pmsg->msg_data = msg;
- list_add_tail(&pmsg->list_elem,
- &msg_queue_obj->msg_used_list);
- ntfy_notify
- (msg_queue_obj->ntfy_obj,
- DSP_NODEMESSAGEREADY);
- sync_set_event
- (msg_queue_obj->sync_event);
- } else {
- /*
- * No free frame to copy the
- * message into.
- */
- pr_err("%s: no free msg frames,"
- " discarding msg\n",
- __func__);
- }
+ (*hmsg_mgr->on_exit)(msg_queue_obj->arg,
+ msg.msg.dw_arg1);
break;
}
+ /*
+ * Not an exit acknowledgement, queue
+ * the message.
+ */
+ if (list_empty(&msg_queue_obj->msg_free_list)) {
+ /*
+ * No free frame to copy the
+ * message into.
+ */
+ pr_err("%s: no free msg frames,"
+ " discarding msg\n",
+ __func__);
+ break;
+ }
+
+ pmsg = list_first_entry(&msg_queue_obj->msg_free_list,
+ struct msg_frame, list_elem);
+ list_del(&pmsg->list_elem);
+ pmsg->msg_data = msg;
+ list_add_tail(&pmsg->list_elem,
+ &msg_queue_obj->msg_used_list);
+ ntfy_notify(msg_queue_obj->ntfy_obj,
+ DSP_NODEMESSAGEREADY);
+ sync_set_event(msg_queue_obj->sync_event);
}
}
/* Set the post SWI flag */
msg_ctr_obj->post_swi = true;
sm_interrupt_dsp(pio_mgr->hbridge_context, MBX_PCPY_CLASS);
}
-func_end:
- return;
}
/*
{
u32 num_msgs = 0;
u32 i;
- u8 *msg_output;
+ struct msg_dspmsg *msg_output;
struct msg_frame *pmsg;
struct msg_ctrl *msg_ctr_obj;
- u32 output_empty;
u32 val;
u32 addr;
msg_ctr_obj = pio_mgr->msg_output_ctrl;
/* Check if output has been cleared */
- output_empty = msg_ctr_obj->buf_empty;
- if (output_empty) {
- num_msgs = (hmsg_mgr->msgs_pending > hmsg_mgr->max_msgs) ?
- hmsg_mgr->max_msgs : hmsg_mgr->msgs_pending;
- msg_output = pio_mgr->msg_output;
- /* Copy num_msgs messages into shared memory */
- for (i = 0; i < num_msgs; i++) {
- if (!list_empty(&hmsg_mgr->msg_used_list)) {
- pmsg = list_first_entry(
- &hmsg_mgr->msg_used_list,
- struct msg_frame, list_elem);
- list_del(&pmsg->list_elem);
- val = (pmsg->msg_data).msgq_id;
- addr = (u32) &(((struct msg_dspmsg *)
- msg_output)->msgq_id);
- write_ext32_bit_dsp_data(
- pio_mgr->hbridge_context, addr, val);
- val = (pmsg->msg_data).msg.dw_cmd;
- addr = (u32) &((((struct msg_dspmsg *)
- msg_output)->msg).dw_cmd);
- write_ext32_bit_dsp_data(
- pio_mgr->hbridge_context, addr, val);
- val = (pmsg->msg_data).msg.dw_arg1;
- addr = (u32) &((((struct msg_dspmsg *)
- msg_output)->msg).dw_arg1);
- write_ext32_bit_dsp_data(
- pio_mgr->hbridge_context, addr, val);
- val = (pmsg->msg_data).msg.dw_arg2;
- addr = (u32) &((((struct msg_dspmsg *)
- msg_output)->msg).dw_arg2);
- write_ext32_bit_dsp_data(
- pio_mgr->hbridge_context, addr, val);
- msg_output += sizeof(struct msg_dspmsg);
- list_add_tail(&pmsg->list_elem,
- &hmsg_mgr->msg_free_list);
- sync_set_event(hmsg_mgr->sync_event);
- }
- }
+ if (!msg_ctr_obj->buf_empty)
+ return;
+
+ num_msgs = (hmsg_mgr->msgs_pending > hmsg_mgr->max_msgs) ?
+ hmsg_mgr->max_msgs : hmsg_mgr->msgs_pending;
+ msg_output = (struct msg_dspmsg *) pio_mgr->msg_output;
+
+ /* Copy num_msgs messages into shared memory */
+ for (i = 0; i < num_msgs; i++) {
+ if (list_empty(&hmsg_mgr->msg_used_list))
+ continue;
+
+ pmsg = list_first_entry(&hmsg_mgr->msg_used_list,
+ struct msg_frame, list_elem);
+ list_del(&pmsg->list_elem);
+
+ val = (pmsg->msg_data).msgq_id;
+ addr = (u32) &msg_output->msgq_id;
+ write_ext32_bit_dsp_data(pio_mgr->hbridge_context, addr, val);
+
+ val = (pmsg->msg_data).msg.dw_cmd;
+ addr = (u32) &msg_output->msg.dw_cmd;
+ write_ext32_bit_dsp_data(pio_mgr->hbridge_context, addr, val);
+
+ val = (pmsg->msg_data).msg.dw_arg1;
+ addr = (u32) &msg_output->msg.dw_arg1;
+ write_ext32_bit_dsp_data(pio_mgr->hbridge_context, addr, val);
+
+ val = (pmsg->msg_data).msg.dw_arg2;
+ addr = (u32) &msg_output->msg.dw_arg2;
+ write_ext32_bit_dsp_data(pio_mgr->hbridge_context, addr, val);
- if (num_msgs > 0) {
- hmsg_mgr->msgs_pending -= num_msgs;
+ msg_output++;
+ list_add_tail(&pmsg->list_elem, &hmsg_mgr->msg_free_list);
+ sync_set_event(hmsg_mgr->sync_event);
+ }
+
+ if (num_msgs > 0) {
+ hmsg_mgr->msgs_pending -= num_msgs;
#if _CHNL_WORDSIZE == 2
- /*
- * Access can be different SM access word size
- * (e.g. 16/32 bit words)
- */
- msg_ctr_obj->size = (u16) num_msgs;
+ /*
+ * Access can be different SM access word size
+ * (e.g. 16/32 bit words)
+ */
+ msg_ctr_obj->size = (u16) num_msgs;
#else
- msg_ctr_obj->size = num_msgs;
+ msg_ctr_obj->size = num_msgs;
#endif
- msg_ctr_obj->buf_empty = false;
- /* Set the post SWI flag */
- msg_ctr_obj->post_swi = true;
- /* Tell the DSP we have written the output. */
- sm_interrupt_dsp(pio_mgr->hbridge_context,
- MBX_PCPY_CLASS);
- }
+ msg_ctr_obj->buf_empty = false;
+ /* Set the post SWI flag */
+ msg_ctr_obj->post_swi = true;
+ /* Tell the DSP we have written the output. */
+ sm_interrupt_dsp(pio_mgr->hbridge_context, MBX_PCPY_CLASS);
}
}
struct io_mgr *hio_mgr;
int status = 0;
- if (!msg_man || !msg_callback || !hdev_obj) {
- status = -EFAULT;
- goto func_end;
- }
+ if (!msg_man || !msg_callback || !hdev_obj)
+ return -EFAULT;
+
dev_get_io_mgr(hdev_obj, &hio_mgr);
- if (!hio_mgr) {
- status = -EFAULT;
- goto func_end;
- }
+ if (!hio_mgr)
+ return -EFAULT;
+
*msg_man = NULL;
/* Allocate msg_ctrl manager object */
msg_mgr_obj = kzalloc(sizeof(struct msg_mgr), GFP_KERNEL);
+ if (!msg_mgr_obj)
+ return -ENOMEM;
- if (msg_mgr_obj) {
- msg_mgr_obj->on_exit = msg_callback;
- msg_mgr_obj->hio_mgr = hio_mgr;
- /* List of MSG_QUEUEs */
- INIT_LIST_HEAD(&msg_mgr_obj->queue_list);
- /* Queues of message frames for messages to the DSP. Message
- * frames will only be added to the free queue when a
- * msg_queue object is created. */
- INIT_LIST_HEAD(&msg_mgr_obj->msg_free_list);
- INIT_LIST_HEAD(&msg_mgr_obj->msg_used_list);
- spin_lock_init(&msg_mgr_obj->msg_mgr_lock);
-
- /* Create an event to be used by bridge_msg_put() in waiting
- * for an available free frame from the message manager. */
- msg_mgr_obj->sync_event =
- kzalloc(sizeof(struct sync_object), GFP_KERNEL);
- if (!msg_mgr_obj->sync_event)
- status = -ENOMEM;
- else
- sync_init_event(msg_mgr_obj->sync_event);
-
- if (!status)
- *msg_man = msg_mgr_obj;
- else
- delete_msg_mgr(msg_mgr_obj);
-
- } else {
- status = -ENOMEM;
+ msg_mgr_obj->on_exit = msg_callback;
+ msg_mgr_obj->hio_mgr = hio_mgr;
+ /* List of MSG_QUEUEs */
+ INIT_LIST_HEAD(&msg_mgr_obj->queue_list);
+ /*
+ * Queues of message frames for messages to the DSP. Message
+ * frames will only be added to the free queue when a
+ * msg_queue object is created.
+ */
+ INIT_LIST_HEAD(&msg_mgr_obj->msg_free_list);
+ INIT_LIST_HEAD(&msg_mgr_obj->msg_used_list);
+ spin_lock_init(&msg_mgr_obj->msg_mgr_lock);
+
+ /*
+ * Create an event to be used by bridge_msg_put() in waiting
+ * for an available free frame from the message manager.
+ */
+ msg_mgr_obj->sync_event =
+ kzalloc(sizeof(struct sync_object), GFP_KERNEL);
+ if (!msg_mgr_obj->sync_event) {
+ kfree(msg_mgr_obj);
+ return -ENOMEM;
}
-func_end:
+ sync_init_event(msg_mgr_obj->sync_event);
+
+ *msg_man = msg_mgr_obj;
+
return status;
}
* Create a msg_queue for sending/receiving messages to/from a node
* on the DSP.
*/
-int bridge_msg_create_queue(struct msg_mgr *hmsg_mgr,
- struct msg_queue **msgq,
+int bridge_msg_create_queue(struct msg_mgr *hmsg_mgr, struct msg_queue **msgq,
u32 msgq_id, u32 max_msgs, void *arg)
{
u32 i;
struct msg_queue *msg_q;
int status = 0;
- if (!hmsg_mgr || msgq == NULL) {
- status = -EFAULT;
- goto func_end;
- }
+ if (!hmsg_mgr || msgq == NULL)
+ return -EFAULT;
*msgq = NULL;
/* Allocate msg_queue object */
msg_q = kzalloc(sizeof(struct msg_queue), GFP_KERNEL);
- if (!msg_q) {
- status = -ENOMEM;
- goto func_end;
- }
+ if (!msg_q)
+ return -ENOMEM;
+
msg_q->max_msgs = max_msgs;
msg_q->hmsg_mgr = hmsg_mgr;
msg_q->arg = arg; /* Node handle */
/* Create event that will be signalled when a message from
* the DSP is available. */
- if (!status) {
- msg_q->sync_event = kzalloc(sizeof(struct sync_object),
- GFP_KERNEL);
- if (msg_q->sync_event)
- sync_init_event(msg_q->sync_event);
- else
- status = -ENOMEM;
+ msg_q->sync_event = kzalloc(sizeof(struct sync_object), GFP_KERNEL);
+ if (!msg_q->sync_event) {
+ status = -ENOMEM;
+ goto out_err;
+
}
+ sync_init_event(msg_q->sync_event);
/* Create a notification list for message ready notification. */
- if (!status) {
- msg_q->ntfy_obj = kmalloc(sizeof(struct ntfy_object),
- GFP_KERNEL);
- if (msg_q->ntfy_obj)
- ntfy_init(msg_q->ntfy_obj);
- else
- status = -ENOMEM;
+ msg_q->ntfy_obj = kmalloc(sizeof(struct ntfy_object), GFP_KERNEL);
+ if (!msg_q->ntfy_obj) {
+ status = -ENOMEM;
+ goto out_err;
}
+ ntfy_init(msg_q->ntfy_obj);
/* Create events that will be used to synchronize cleanup
* when the object is deleted. sync_done will be set to
* unblock threads in MSG_Put() or MSG_Get(). sync_done_ack
* will be set by the unblocked thread to signal that it
* is unblocked and will no longer reference the object. */
- if (!status) {
- msg_q->sync_done = kzalloc(sizeof(struct sync_object),
- GFP_KERNEL);
- if (msg_q->sync_done)
- sync_init_event(msg_q->sync_done);
- else
- status = -ENOMEM;
+ msg_q->sync_done = kzalloc(sizeof(struct sync_object), GFP_KERNEL);
+ if (!msg_q->sync_done) {
+ status = -ENOMEM;
+ goto out_err;
}
+ sync_init_event(msg_q->sync_done);
- if (!status) {
- msg_q->sync_done_ack = kzalloc(sizeof(struct sync_object),
- GFP_KERNEL);
- if (msg_q->sync_done_ack)
- sync_init_event(msg_q->sync_done_ack);
- else
- status = -ENOMEM;
+ msg_q->sync_done_ack = kzalloc(sizeof(struct sync_object), GFP_KERNEL);
+ if (!msg_q->sync_done_ack) {
+ status = -ENOMEM;
+ goto out_err;
}
+ sync_init_event(msg_q->sync_done_ack);
- if (!status) {
- /* Enter critical section */
- spin_lock_bh(&hmsg_mgr->msg_mgr_lock);
- /* Initialize message frames and put in appropriate queues */
- for (i = 0; i < max_msgs && !status; i++) {
- status = add_new_msg(&hmsg_mgr->msg_free_list);
- if (!status) {
- num_allocated++;
- status = add_new_msg(&msg_q->msg_free_list);
- }
- }
- if (status) {
- /* Stay inside CS to prevent others from taking any
- * of the newly allocated message frames. */
- delete_msg_queue(msg_q, num_allocated);
- } else {
- list_add_tail(&msg_q->list_elem,
- &hmsg_mgr->queue_list);
- *msgq = msg_q;
- /* Signal that free frames are now available */
- if (!list_empty(&hmsg_mgr->msg_free_list))
- sync_set_event(hmsg_mgr->sync_event);
-
+ /* Enter critical section */
+ spin_lock_bh(&hmsg_mgr->msg_mgr_lock);
+ /* Initialize message frames and put in appropriate queues */
+ for (i = 0; i < max_msgs && !status; i++) {
+ status = add_new_msg(&hmsg_mgr->msg_free_list);
+ if (!status) {
+ num_allocated++;
+ status = add_new_msg(&msg_q->msg_free_list);
}
- /* Exit critical section */
+ }
+ if (status) {
spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
- } else {
- delete_msg_queue(msg_q, 0);
+ goto out_err;
}
-func_end:
+
+ list_add_tail(&msg_q->list_elem, &hmsg_mgr->queue_list);
+ *msgq = msg_q;
+ /* Signal that free frames are now available */
+ if (!list_empty(&hmsg_mgr->msg_free_list))
+ sync_set_event(hmsg_mgr->sync_event);
+
+ /* Exit critical section */
+ spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
+
+ return 0;
+out_err:
+ delete_msg_queue(msg_q, num_allocated);
return status;
}
u32 io_msg_pend;
if (!msg_queue_obj || !msg_queue_obj->hmsg_mgr)
- goto func_end;
+ return;
hmsg_mgr = msg_queue_obj->hmsg_mgr;
msg_queue_obj->done = true;
delete_msg_queue(msg_queue_obj, msg_queue_obj->max_msgs);
if (list_empty(&hmsg_mgr->msg_free_list))
sync_reset_event(hmsg_mgr->sync_event);
-
spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
-func_end:
- return;
}
/*
{
struct msg_frame *msg_frame_obj;
struct msg_mgr *hmsg_mgr;
- bool got_msg = false;
struct sync_object *syncs[2];
u32 index;
int status = 0;
- if (!msg_queue_obj || pmsg == NULL) {
- status = -ENOMEM;
- goto func_end;
- }
+ if (!msg_queue_obj || pmsg == NULL)
+ return -ENOMEM;
hmsg_mgr = msg_queue_obj->hmsg_mgr;
- /* Enter critical section */
spin_lock_bh(&hmsg_mgr->msg_mgr_lock);
/* If a message is already there, get it */
if (!list_empty(&msg_queue_obj->msg_used_list)) {
&msg_queue_obj->msg_free_list);
if (list_empty(&msg_queue_obj->msg_used_list))
sync_reset_event(msg_queue_obj->sync_event);
+ spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
+ return 0;
+ }
- got_msg = true;
- } else {
- if (msg_queue_obj->done)
- status = -EPERM;
- else
- msg_queue_obj->io_msg_pend++;
-
+ if (msg_queue_obj->done) {
+ spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
+ return -EPERM;
}
- /* Exit critical section */
+ msg_queue_obj->io_msg_pend++;
spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
- if (!status && !got_msg) {
- /* Wait til message is available, timeout, or done. We don't
- * have to schedule the DPC, since the DSP will send messages
- * when they are available. */
- syncs[0] = msg_queue_obj->sync_event;
- syncs[1] = msg_queue_obj->sync_done;
- status = sync_wait_on_multiple_events(syncs, 2, utimeout,
- &index);
- /* Enter critical section */
- spin_lock_bh(&hmsg_mgr->msg_mgr_lock);
- if (msg_queue_obj->done) {
- msg_queue_obj->io_msg_pend--;
- /* Exit critical section */
- spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
- /* Signal that we're not going to access msg_queue_obj
- * anymore, so it can be deleted. */
- (void)sync_set_event(msg_queue_obj->sync_done_ack);
- status = -EPERM;
- } else {
- if (!status && !list_empty(&msg_queue_obj->
- msg_used_list)) {
- /* Get msg from used list */
- msg_frame_obj = list_first_entry(
- &msg_queue_obj->msg_used_list,
- struct msg_frame, list_elem);
- list_del(&msg_frame_obj->list_elem);
- /* Copy message into pmsg and put frame on the
- * free list */
- *pmsg = msg_frame_obj->msg_data.msg;
- list_add_tail(&msg_frame_obj->list_elem,
- &msg_queue_obj->msg_free_list);
- }
- msg_queue_obj->io_msg_pend--;
- /* Reset the event if there are still queued messages */
- if (!list_empty(&msg_queue_obj->msg_used_list))
- sync_set_event(msg_queue_obj->sync_event);
-
- /* Exit critical section */
- spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
- }
+
+ /*
+ * Wait til message is available, timeout, or done. We don't
+ * have to schedule the DPC, since the DSP will send messages
+ * when they are available.
+ */
+ syncs[0] = msg_queue_obj->sync_event;
+ syncs[1] = msg_queue_obj->sync_done;
+ status = sync_wait_on_multiple_events(syncs, 2, utimeout, &index);
+
+ spin_lock_bh(&hmsg_mgr->msg_mgr_lock);
+ if (msg_queue_obj->done) {
+ msg_queue_obj->io_msg_pend--;
+ spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
+ /*
+ * Signal that we're not going to access msg_queue_obj
+ * anymore, so it can be deleted.
+ */
+ sync_set_event(msg_queue_obj->sync_done_ack);
+ return -EPERM;
}
-func_end:
+ if (!status && !list_empty(&msg_queue_obj->msg_used_list)) {
+ /* Get msg from used list */
+ msg_frame_obj = list_first_entry(&msg_queue_obj->msg_used_list,
+ struct msg_frame, list_elem);
+ list_del(&msg_frame_obj->list_elem);
+ /* Copy message into pmsg and put frame on the free list */
+ *pmsg = msg_frame_obj->msg_data.msg;
+ list_add_tail(&msg_frame_obj->list_elem,
+ &msg_queue_obj->msg_free_list);
+ }
+ msg_queue_obj->io_msg_pend--;
+ /* Reset the event if there are still queued messages */
+ if (!list_empty(&msg_queue_obj->msg_used_list))
+ sync_set_event(msg_queue_obj->sync_event);
+
+ spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
+
return status;
}
{
struct msg_frame *msg_frame_obj;
struct msg_mgr *hmsg_mgr;
- bool put_msg = false;
struct sync_object *syncs[2];
u32 index;
- int status = 0;
+ int status;
+
+ if (!msg_queue_obj || !pmsg || !msg_queue_obj->hmsg_mgr)
+ return -EFAULT;
- if (!msg_queue_obj || !pmsg || !msg_queue_obj->hmsg_mgr) {
- status = -ENOMEM;
- goto func_end;
- }
hmsg_mgr = msg_queue_obj->hmsg_mgr;
spin_lock_bh(&hmsg_mgr->msg_mgr_lock);
list_add_tail(&msg_frame_obj->list_elem,
&hmsg_mgr->msg_used_list);
hmsg_mgr->msgs_pending++;
- put_msg = true;
+
if (list_empty(&hmsg_mgr->msg_free_list))
sync_reset_event(hmsg_mgr->sync_event);
spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
/* Schedule a DPC, to do the actual data transfer: */
iosm_schedule(hmsg_mgr->hio_mgr);
- } else {
- if (msg_queue_obj->done)
- status = -EPERM;
- else
- msg_queue_obj->io_msg_pend++;
+ return 0;
+ }
+ if (msg_queue_obj->done) {
spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
+ return -EPERM;
}
- if (!status && !put_msg) {
- /* Wait til a free message frame is available, timeout,
- * or done */
- syncs[0] = hmsg_mgr->sync_event;
- syncs[1] = msg_queue_obj->sync_done;
- status = sync_wait_on_multiple_events(syncs, 2, utimeout,
- &index);
- if (status)
- goto func_end;
- /* Enter critical section */
- spin_lock_bh(&hmsg_mgr->msg_mgr_lock);
- if (msg_queue_obj->done) {
- msg_queue_obj->io_msg_pend--;
- /* Exit critical section */
- spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
- /* Signal that we're not going to access msg_queue_obj
- * anymore, so it can be deleted. */
- (void)sync_set_event(msg_queue_obj->sync_done_ack);
- status = -EPERM;
- } else {
- if (list_empty(&hmsg_mgr->msg_free_list)) {
- status = -EFAULT;
- goto func_cont;
- }
- /* Get msg from free list */
- msg_frame_obj = list_first_entry(
- &hmsg_mgr->msg_free_list,
- struct msg_frame, list_elem);
- list_del(&msg_frame_obj->list_elem);
- /*
- * Copy message into pmsg and put frame on the
- * used list.
- */
- msg_frame_obj->msg_data.msg = *pmsg;
- msg_frame_obj->msg_data.msgq_id =
- msg_queue_obj->msgq_id;
- list_add_tail(&msg_frame_obj->list_elem,
- &hmsg_mgr->msg_used_list);
- hmsg_mgr->msgs_pending++;
- /*
- * Schedule a DPC, to do the actual
- * data transfer.
- */
- iosm_schedule(hmsg_mgr->hio_mgr);
-
- msg_queue_obj->io_msg_pend--;
- /* Reset event if there are still frames available */
- if (!list_empty(&hmsg_mgr->msg_free_list))
- sync_set_event(hmsg_mgr->sync_event);
-func_cont:
- /* Exit critical section */
- spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
- }
+ msg_queue_obj->io_msg_pend++;
+
+ spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
+
+ /* Wait til a free message frame is available, timeout, or done */
+ syncs[0] = hmsg_mgr->sync_event;
+ syncs[1] = msg_queue_obj->sync_done;
+ status = sync_wait_on_multiple_events(syncs, 2, utimeout, &index);
+ if (status)
+ return status;
+
+ /* Enter critical section */
+ spin_lock_bh(&hmsg_mgr->msg_mgr_lock);
+ if (msg_queue_obj->done) {
+ msg_queue_obj->io_msg_pend--;
+ /* Exit critical section */
+ spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
+ /*
+ * Signal that we're not going to access msg_queue_obj
+ * anymore, so it can be deleted.
+ */
+ sync_set_event(msg_queue_obj->sync_done_ack);
+ return -EPERM;
}
-func_end:
- return status;
+
+ if (list_empty(&hmsg_mgr->msg_free_list)) {
+ spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
+ return -EFAULT;
+ }
+
+ /* Get msg from free list */
+ msg_frame_obj = list_first_entry(&hmsg_mgr->msg_free_list,
+ struct msg_frame, list_elem);
+ /*
+ * Copy message into pmsg and put frame on the
+ * used list.
+ */
+ list_del(&msg_frame_obj->list_elem);
+ msg_frame_obj->msg_data.msg = *pmsg;
+ msg_frame_obj->msg_data.msgq_id = msg_queue_obj->msgq_id;
+ list_add_tail(&msg_frame_obj->list_elem, &hmsg_mgr->msg_used_list);
+ hmsg_mgr->msgs_pending++;
+ /*
+ * Schedule a DPC, to do the actual
+ * data transfer.
+ */
+ iosm_schedule(hmsg_mgr->hio_mgr);
+
+ msg_queue_obj->io_msg_pend--;
+ /* Reset event if there are still frames available */
+ if (!list_empty(&hmsg_mgr->msg_free_list))
+ sync_set_event(hmsg_mgr->sync_event);
+
+ /* Exit critical section */
+ spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
+
+ return 0;
}
/*
static int add_new_msg(struct list_head *msg_list)
{
struct msg_frame *pmsg;
- int status = 0;
pmsg = kzalloc(sizeof(struct msg_frame), GFP_ATOMIC);
- if (pmsg != NULL) {
- list_add_tail(&pmsg->list_elem, msg_list);
- } else {
- status = -ENOMEM;
- }
+ if (!pmsg)
+ return -ENOMEM;
- return status;
+ list_add_tail(&pmsg->list_elem, msg_list);
+
+ return 0;
}
/*
static void delete_msg_mgr(struct msg_mgr *hmsg_mgr)
{
if (!hmsg_mgr)
- goto func_end;
+ return;
/* FIXME: free elements from queue_list? */
free_msg_list(&hmsg_mgr->msg_free_list);
free_msg_list(&hmsg_mgr->msg_used_list);
-
kfree(hmsg_mgr->sync_event);
-
kfree(hmsg_mgr);
-func_end:
- return;
}
/*
u32 i;
if (!msg_queue_obj || !msg_queue_obj->hmsg_mgr)
- goto func_end;
+ return;
hmsg_mgr = msg_queue_obj->hmsg_mgr;
kfree(msg_queue_obj->sync_done_ack);
kfree(msg_queue_obj);
-func_end:
- return;
-
}
/*
struct msg_frame *pmsg, *tmp;
if (!msg_list)
- goto func_end;
+ return;
list_for_each_entry_safe(pmsg, tmp, msg_list, list_elem) {
list_del(&pmsg->list_elem);
kfree(pmsg);
}
-
-func_end:
- return;
}