/* Function interface to Bridge driver */
struct bridge_drv_interface *intf_fxns;
- struct io_mgr *hio_mgr; /* IO manager */
+ struct io_mgr *iomgr; /* IO manager */
struct list_head queue_list; /* List of MSG_QUEUEs */
spinlock_t msg_mgr_lock; /* For critical sections */
/* Signalled when MsgFrame is available */
* which may cause timeouts and/or failure offunction sync_wait_on_event.
* This invariant condition is:
*
- * list_empty(&pchnl->pio_completions) ==> pchnl->sync_event is reset
+ * list_empty(&pchnl->io_completions) ==> pchnl->sync_event is reset
* and
- * !list_empty(&pchnl->pio_completions) ==> pchnl->sync_event is set.
+ * !list_empty(&pchnl->io_completions) ==> pchnl->sync_event is set.
*/
#include <linux/types.h>
if (CHNL_IS_OUTPUT(pchnl->chnl_mode)) {
/* Check buffer size on output channels for fit. */
if (byte_size > io_buf_size(
- pchnl->chnl_mgr_obj->hio_mgr)) {
+ pchnl->chnl_mgr_obj->iomgr)) {
status = -EINVAL;
goto out;
}
chnl_packet_obj->arg = dw_arg;
chnl_packet_obj->status = (is_eos ? CHNL_IOCSTATEOS :
CHNL_IOCSTATCOMPLETE);
- list_add_tail(&chnl_packet_obj->link, &pchnl->pio_requests);
+ list_add_tail(&chnl_packet_obj->link, &pchnl->io_requests);
pchnl->cio_reqs++;
DBC_ASSERT(pchnl->cio_reqs <= pchnl->chnl_packets);
/*
/* Legacy DSM Processor-Copy */
DBC_ASSERT(pchnl->chnl_type == CHNL_PCPY);
/* Request IO from the DSP */
- io_request_chnl(chnl_mgr_obj->hio_mgr, pchnl,
+ io_request_chnl(chnl_mgr_obj->iomgr, pchnl,
(CHNL_IS_INPUT(pchnl->chnl_mode) ? IO_INPUT :
IO_OUTPUT), &mb_val);
sched_dpc = true;
/* Schedule a DPC, to do the actual data transfer */
if (sched_dpc)
- iosm_schedule(chnl_mgr_obj->hio_mgr);
+ iosm_schedule(chnl_mgr_obj->iomgr);
return status;
}
pchnl->state |= CHNL_STATECANCEL;
- if (list_empty(&pchnl->pio_requests)) {
+ if (list_empty(&pchnl->io_requests)) {
spin_unlock_bh(&chnl_mgr_obj->chnl_mgr_lock);
return 0;
}
if (pchnl->chnl_type == CHNL_PCPY) {
/* Indicate we have no more buffers available for transfer: */
if (CHNL_IS_INPUT(pchnl->chnl_mode)) {
- io_cancel_chnl(chnl_mgr_obj->hio_mgr, chnl_id);
+ io_cancel_chnl(chnl_mgr_obj->iomgr, chnl_id);
} else {
/* Record that we no longer have output buffers
* available: */
}
}
/* Move all IOR's to IOC queue: */
- list_for_each_entry_safe(chirp, tmp, &pchnl->pio_requests, link) {
+ list_for_each_entry_safe(chirp, tmp, &pchnl->io_requests, link) {
list_del(&chirp->link);
chirp->byte_size = 0;
chirp->status |= CHNL_IOCSTATCANCEL;
- list_add_tail(&chirp->link, &pchnl->pio_completions);
+ list_add_tail(&chirp->link, &pchnl->io_completions);
pchnl->cio_cs++;
pchnl->cio_reqs--;
DBC_ASSERT(pchnl->cio_reqs >= 0);
DBC_ASSERT((pchnl->state & CHNL_STATECANCEL));
/* Invalidate channel object: Protects from CHNL_GetIOCompletion() */
/* Free the slot in the channel manager: */
- pchnl->chnl_mgr_obj->ap_channel[pchnl->chnl_id] = NULL;
+ pchnl->chnl_mgr_obj->channels[pchnl->chnl_id] = NULL;
spin_lock_bh(&pchnl->chnl_mgr_obj->chnl_mgr_lock);
pchnl->chnl_mgr_obj->open_channels -= 1;
spin_unlock_bh(&pchnl->chnl_mgr_obj->chnl_mgr_lock);
pchnl->sync_event = NULL;
}
/* Free I/O request and I/O completion queues: */
- free_chirp_list(&pchnl->pio_completions);
+ free_chirp_list(&pchnl->io_completions);
pchnl->cio_cs = 0;
- free_chirp_list(&pchnl->pio_requests);
+ free_chirp_list(&pchnl->io_requests);
pchnl->cio_reqs = 0;
free_chirp_list(&pchnl->free_packets_list);
DBC_ASSERT(mgr_attrts->max_channels == CHNL_MAXCHANNELS);
max_channels = CHNL_MAXCHANNELS + CHNL_MAXCHANNELS * CHNL_PCPY;
/* Create array of channels */
- chnl_mgr_obj->ap_channel = kzalloc(sizeof(struct chnl_object *)
+ chnl_mgr_obj->channels = kzalloc(sizeof(struct chnl_object *)
* max_channels, GFP_KERNEL);
- if (chnl_mgr_obj->ap_channel) {
+ if (chnl_mgr_obj->channels) {
/* Initialize chnl_mgr object */
chnl_mgr_obj->type = CHNL_TYPESM;
chnl_mgr_obj->word_size = mgr_attrts->word_size;
for (chnl_id = 0; chnl_id < chnl_mgr_obj->max_channels;
chnl_id++) {
status =
- bridge_chnl_close(chnl_mgr_obj->ap_channel
+ bridge_chnl_close(chnl_mgr_obj->channels
[chnl_id]);
if (status)
dev_dbg(bridge, "%s: Error status 0x%x\n",
}
/* Free channel manager object: */
- kfree(chnl_mgr_obj->ap_channel);
+ kfree(chnl_mgr_obj->channels);
/* Set hchnl_mgr to NULL in device object. */
dev_set_chnl_mgr(chnl_mgr_obj->dev_obj, NULL);
&& (pchnl->chnl_type == CHNL_PCPY)) {
/* Wait for IO completions, up to the specified
* timeout: */
- while (!list_empty(&pchnl->pio_requests) && !status) {
+ while (!list_empty(&pchnl->io_requests) && !status) {
status = bridge_chnl_get_ioc(chnl_obj,
timeout, &chnl_ioc_obj);
if (status)
pchnl->state &= ~CHNL_STATECANCEL;
}
}
- DBC_ENSURE(status || list_empty(&pchnl->pio_requests));
+ DBC_ENSURE(status || list_empty(&pchnl->io_requests));
return status;
}
if (!chan_ioc || !pchnl) {
status = -EFAULT;
} else if (timeout == CHNL_IOCNOWAIT) {
- if (list_empty(&pchnl->pio_completions))
+ if (list_empty(&pchnl->io_completions))
status = -EREMOTEIO;
}
ioc.status = CHNL_IOCSTATCOMPLETE;
if (timeout !=
- CHNL_IOCNOWAIT && list_empty(&pchnl->pio_completions)) {
+ CHNL_IOCNOWAIT && list_empty(&pchnl->io_completions)) {
if (timeout == CHNL_IOCINFINITE)
timeout = SYNC_INFINITE;
* fails due to unkown causes. */
/* Even though Wait failed, there may be something in
* the Q: */
- if (list_empty(&pchnl->pio_completions)) {
+ if (list_empty(&pchnl->io_completions)) {
ioc.status |= CHNL_IOCSTATCANCEL;
dequeue_ioc = false;
}
omap_mbox_disable_irq(dev_ctxt->mbox, IRQ_RX);
if (dequeue_ioc) {
/* Dequeue IOC and set chan_ioc; */
- DBC_ASSERT(!list_empty(&pchnl->pio_completions));
- chnl_packet_obj = list_first_entry(&pchnl->pio_completions,
+ DBC_ASSERT(!list_empty(&pchnl->io_completions));
+ chnl_packet_obj = list_first_entry(&pchnl->io_completions,
struct chnl_irp, link);
list_del(&chnl_packet_obj->link);
/* Update chan_ioc from channel state and chirp: */
ioc.buf_size = 0;
}
/* Ensure invariant: If any IOC's are queued for this channel... */
- if (!list_empty(&pchnl->pio_completions)) {
+ if (!list_empty(&pchnl->io_completions)) {
/* Since DSPStream_Reclaim() does not take a timeout
* parameter, we pass the stream's timeout value to
* bridge_chnl_get_ioc. We cannot determine whether or not
return -ECHRNG;
/* Return the requested information: */
- mgr_info->chnl_obj = chnl_mgr_obj->ap_channel[ch_id];
+ mgr_info->chnl_obj = chnl_mgr_obj->channels[ch_id];
mgr_info->open_channels = chnl_mgr_obj->open_channels;
mgr_info->type = chnl_mgr_obj->type;
/* total # of chnls */
if (ch_id != CHNL_PICKFREE) {
if (ch_id >= chnl_mgr_obj->max_channels)
return -ECHRNG;
- if (chnl_mgr_obj->ap_channel[ch_id] != NULL)
+ if (chnl_mgr_obj->channels[ch_id] != NULL)
return -EALREADY;
} else {
/* Check for free channel */
if (status)
goto out_err;
- INIT_LIST_HEAD(&pchnl->pio_requests);
- INIT_LIST_HEAD(&pchnl->pio_completions);
+ INIT_LIST_HEAD(&pchnl->io_requests);
+ INIT_LIST_HEAD(&pchnl->io_completions);
pchnl->chnl_packets = pattrs->uio_reqs;
pchnl->cio_cs = 0;
pchnl->chnl_type = CHNL_PCPY;
/* Insert channel object in channel manager: */
- chnl_mgr_obj->ap_channel[pchnl->chnl_id] = pchnl;
+ chnl_mgr_obj->channels[pchnl->chnl_id] = pchnl;
spin_lock_bh(&chnl_mgr_obj->chnl_mgr_lock);
chnl_mgr_obj->open_channels++;
spin_unlock_bh(&chnl_mgr_obj->chnl_mgr_lock);
out_err:
/* Free memory */
- free_chirp_list(&pchnl->pio_completions);
- free_chirp_list(&pchnl->pio_requests);
+ free_chirp_list(&pchnl->io_completions);
+ free_chirp_list(&pchnl->io_requests);
free_chirp_list(&pchnl->free_packets_list);
if (sync_event)
DBC_REQUIRE(chnl_mgr_obj);
for (i = 0; i < chnl_mgr_obj->max_channels; i++) {
- if (chnl_mgr_obj->ap_channel[i] == NULL) {
+ if (chnl_mgr_obj->channels[i] == NULL) {
status = 0;
*chnl = i;
break;
*io_man = NULL;
dev_get_chnl_mgr(hdev_obj, &hchnl_mgr);
- if (!hchnl_mgr || hchnl_mgr->hio_mgr)
+ if (!hchnl_mgr || hchnl_mgr->iomgr)
return -EFAULT;
/*
}
/* Return IO manager object to caller... */
- hchnl_mgr->hio_mgr = pio_mgr;
+ hchnl_mgr->iomgr = pio_mgr;
*io_man = pio_mgr;
return 0;
DBC_ASSERT(chnl_id);
goto func_end;
}
- pchnl = chnl_mgr_obj->ap_channel[chnl_id];
+ pchnl = chnl_mgr_obj->channels[chnl_id];
if ((pchnl != NULL) && CHNL_IS_INPUT(pchnl->chnl_mode)) {
if ((pchnl->state & ~CHNL_STATEEOS) == CHNL_STATEREADY) {
/* Get the I/O request, and attempt a transfer */
- if (!list_empty(&pchnl->pio_requests)) {
+ if (!list_empty(&pchnl->io_requests)) {
if (!pchnl->cio_reqs)
goto func_end;
chnl_packet_obj = list_first_entry(
- &pchnl->pio_requests,
+ &pchnl->io_requests,
struct chnl_irp, link);
list_del(&chnl_packet_obj->link);
pchnl->cio_reqs--;
DSP_STREAMDONE);
}
/* Tell DSP if no more I/O buffers available */
- if (list_empty(&pchnl->pio_requests))
+ if (list_empty(&pchnl->io_requests))
set_chnl_free(sm, pchnl->chnl_id);
clear_chnl = true;
notify_client = true;
* signalled by the only IO completion list consumer:
* bridge_chnl_get_ioc().
*/
- signal_event = list_empty(&pchnl->pio_completions);
+ signal_event = list_empty(&pchnl->io_completions);
/* Enqueue the IO completion info for the client */
- list_add_tail(&chnl_packet_obj->link, &pchnl->pio_completions);
+ list_add_tail(&chnl_packet_obj->link, &pchnl->io_completions);
pchnl->cio_cs++;
if (pchnl->cio_cs > pchnl->chnl_packets)
if (chnl_id == OUTPUTNOTREADY)
goto func_end;
- pchnl = chnl_mgr_obj->ap_channel[chnl_id];
- if (!pchnl || list_empty(&pchnl->pio_requests)) {
+ pchnl = chnl_mgr_obj->channels[chnl_id];
+ if (!pchnl || list_empty(&pchnl->io_requests)) {
/* Shouldn't get here */
goto func_end;
}
goto func_end;
/* Get the I/O request, and attempt a transfer */
- chnl_packet_obj = list_first_entry(&pchnl->pio_requests,
+ chnl_packet_obj = list_first_entry(&pchnl->io_requests,
struct chnl_irp, link);
list_del(&chnl_packet_obj->link);
pchnl->cio_reqs--;
/* Record fact that no more I/O buffers available */
- if (list_empty(&pchnl->pio_requests))
+ if (list_empty(&pchnl->io_requests))
chnl_mgr_obj->output_mask &= ~(1 << chnl_id);
/* Transfer buffer to DSP side */
return -ENOMEM;
msg_mgr_obj->on_exit = msg_callback;
- msg_mgr_obj->hio_mgr = hio_mgr;
+ msg_mgr_obj->iomgr = hio_mgr;
/* List of MSG_QUEUEs */
INIT_LIST_HEAD(&msg_mgr_obj->queue_list);
/*
/* Release critical section before scheduling DPC */
spin_unlock_bh(&hmsg_mgr->msg_mgr_lock);
/* Schedule a DPC, to do the actual data transfer: */
- iosm_schedule(hmsg_mgr->hio_mgr);
+ iosm_schedule(hmsg_mgr->iomgr);
return 0;
}
* Schedule a DPC, to do the actual
* data transfer.
*/
- iosm_schedule(hmsg_mgr->hio_mgr);
+ iosm_schedule(hmsg_mgr->iomgr);
msg_queue_obj->io_msg_pend--;
/* Reset event if there are still frames available */
struct chnl_mgr {
/* Function interface to Bridge driver */
struct bridge_drv_interface *intf_fxns;
- struct io_mgr *hio_mgr; /* IO manager */
+ struct io_mgr *iomgr; /* IO manager */
/* Device this board represents */
struct dev_object *dev_obj;
u32 word_size; /* Size in bytes of DSP word */
u8 max_channels; /* Total number of channels */
u8 open_channels; /* Total number of open channels */
- struct chnl_object **ap_channel; /* Array of channels */
+ struct chnl_object **channels; /* Array of channels */
u8 type; /* Type of channel class library */
/* If no shm syms, return for CHNL_Open */
int chnl_open_status;
struct sync_object *sync_event;
u32 process; /* Process which created this channel */
u32 cb_arg; /* Argument to use with callback */
- struct list_head pio_requests; /* List of IOR's to driver */
+ struct list_head io_requests; /* List of IOR's to driver */
s32 cio_cs; /* Number of IOC's in queue */
s32 cio_reqs; /* Number of IORequests in queue */
s32 chnl_packets; /* Initial number of free Irps */
/* List of IOC's from driver */
- struct list_head pio_completions;
+ struct list_head io_completions;
struct list_head free_packets_list; /* List of free Irps */
struct ntfy_object *ntfy_obj;
u32 bytes_moved; /* Total number of bytes transfered */
/* MGR Module */
struct {
u32 node_id;
- struct dsp_ndbprops __user *pndb_props;
- u32 undb_props_size;
- u32 __user *pu_num_nodes;
+ struct dsp_ndbprops __user *ndb_props;
+ u32 ndb_props_size;
+ u32 __user *num_nodes;
} args_mgr_enumnode_info;
struct {
u32 processor_id;
struct dsp_processorinfo __user *processor_info;
u32 processor_info_size;
- u32 __user *pu_num_procs;
+ u32 __user *num_procs;
} args_mgr_enumproc_info;
struct {
struct dsp_uuid *uuid_obj;
enum dsp_dcdobjtype obj_type;
- char *psz_path_name;
+ char *sz_path_name;
} args_mgr_registerobject;
struct {
struct {
struct dsp_notification __user *__user *anotifications;
u32 count;
- u32 __user *pu_index;
+ u32 __user *index;
u32 timeout;
} args_mgr_wait;
struct {
void *processor;
u32 cmd;
- struct dsp_cbdata __user *pargs;
+ struct dsp_cbdata __user *args;
} args_proc_ctrl;
struct {
void *processor;
void *__user *node_tab;
u32 node_tab_size;
- u32 __user *pu_num_nodes;
- u32 __user *pu_allocated;
+ u32 __user *num_nodes;
+ u32 __user *allocated;
} args_proc_enumnode_info;
struct {
struct {
void *processor;
u8 __user *buf;
- u8 __user *psize;
+ u8 __user *size;
u32 max_size;
} args_proc_gettrace;
void *processor;
u32 event_mask;
u32 notify_type;
- struct dsp_notification __user *hnotification;
+ struct dsp_notification __user *notification;
} args_proc_register_notify;
struct {
void *processor;
u32 size;
- void *__user *pp_rsv_addr;
+ void *__user *rsv_addr;
} args_proc_rsvmem;
struct {
void *processor;
u32 size;
- void *prsv_addr;
+ void *rsv_addr;
} args_proc_unrsvmem;
struct {
void *processor;
- void *pmpu_addr;
+ void *mpu_addr;
u32 size;
void *req_addr;
- void *__user *pp_map_addr;
- u32 ul_map_attr;
+ void *__user *map_addr;
+ u32 map_attr;
} args_proc_mapmem;
struct {
struct {
void *processor;
- void *pmpu_addr;
+ void *mpu_addr;
u32 size;
u32 dir;
} args_proc_dma;
struct {
void *processor;
- void *pmpu_addr;
+ void *mpu_addr;
u32 size;
u32 ul_flags;
} args_proc_flushmemory;
struct {
void *processor;
- void *pmpu_addr;
+ void *mpu_addr;
u32 size;
} args_proc_invalidatememory;
struct {
void *processor;
struct dsp_uuid __user *node_id_ptr;
- struct dsp_cbdata __user *pargs;
+ struct dsp_cbdata __user *args;
struct dsp_nodeattrin __user *attr_in;
void *__user *ph_node;
} args_node_allocate;
void *node;
u32 event_mask;
u32 notify_type;
- struct dsp_notification __user *hnotification;
+ struct dsp_notification __user *notification;
} args_node_registernotify;
struct {
void *stream;
u32 event_mask;
u32 notify_type;
- struct dsp_notification __user *hnotification;
+ struct dsp_notification __user *notification;
} args_strm_registernotify;
struct {
struct chnl_mgr *chnl_mgr; /* Channel manager. */
struct deh_mgr *deh_mgr; /* DEH manager. */
struct msg_mgr *msg_mgr; /* Message manager. */
- struct io_mgr *hio_mgr; /* IO manager (CHNL, msg_ctrl) */
+ struct io_mgr *iomgr; /* IO manager (CHNL, msg_ctrl) */
struct cmm_object *cmm_mgr; /* SM memory manager. */
struct dmm_object *dmm_mgr; /* Dynamic memory manager. */
u32 word_size; /* DSP word size: quick access. */
(struct dev_object *)dev_obj, NULL);
/* Only create IO manager if we have a channel manager */
if (!status && dev_obj->chnl_mgr) {
- status = io_create(&dev_obj->hio_mgr, dev_obj,
+ status = io_create(&dev_obj->iomgr, dev_obj,
&io_mgr_attrs);
}
/* Only create DEH manager if we have an IO manager */
}
/* Free the io, channel, and message managers for this board: */
- if (dev_obj->hio_mgr) {
- io_destroy(dev_obj->hio_mgr);
- dev_obj->hio_mgr = NULL;
+ if (dev_obj->iomgr) {
+ io_destroy(dev_obj->iomgr);
+ dev_obj->iomgr = NULL;
}
if (dev_obj->chnl_mgr) {
chnl_destroy(dev_obj->chnl_mgr);
DBC_REQUIRE(hdev_obj);
if (hdev_obj) {
- *io_man = hdev_obj->hio_mgr;
+ *io_man = hdev_obj->iomgr;
} else {
*io_man = NULL;
status = -EFAULT;
u8 *pndb_props;
u32 num_nodes;
int status = 0;
- u32 size = args->args_mgr_enumnode_info.undb_props_size;
+ u32 size = args->args_mgr_enumnode_info.ndb_props_size;
if (size < sizeof(struct dsp_ndbprops))
return -EINVAL;
(struct dsp_ndbprops *)pndb_props, size,
&num_nodes);
}
- CP_TO_USR(args->args_mgr_enumnode_info.pndb_props, pndb_props, status,
+ CP_TO_USR(args->args_mgr_enumnode_info.ndb_props, pndb_props, status,
size);
- CP_TO_USR(args->args_mgr_enumnode_info.pu_num_nodes, &num_nodes, status,
+ CP_TO_USR(args->args_mgr_enumnode_info.num_nodes, &num_nodes, status,
1);
kfree(pndb_props);
}
CP_TO_USR(args->args_mgr_enumproc_info.processor_info, processor_info,
status, size);
- CP_TO_USR(args->args_mgr_enumproc_info.pu_num_procs, &num_procs,
+ CP_TO_USR(args->args_mgr_enumproc_info.num_procs, &num_procs,
status, 1);
kfree(processor_info);
goto func_end;
/* path_size is increased by 1 to accommodate NULL */
path_size = strlen_user((char *)
- args->args_mgr_registerobject.psz_path_name) +
+ args->args_mgr_registerobject.sz_path_name) +
1;
psz_path_name = kmalloc(path_size, GFP_KERNEL);
if (!psz_path_name) {
}
ret = strncpy_from_user(psz_path_name,
(char *)args->args_mgr_registerobject.
- psz_path_name, path_size);
+ sz_path_name, path_size);
if (!ret) {
status = -EFAULT;
goto func_end;
args->args_mgr_wait.
timeout);
}
- CP_TO_USR(args->args_mgr_wait.pu_index, &index, status, 1);
+ CP_TO_USR(args->args_mgr_wait.index, &index, status, 1);
return status;
}
u32 procwrap_ctrl(union trapped_args *args, void *pr_ctxt)
{
u32 cb_data_size, __user * psize = (u32 __user *)
- args->args_proc_ctrl.pargs;
+ args->args_proc_ctrl.args;
u8 *pargs = NULL;
int status = 0;
void *hprocessor = ((struct process_context *)pr_ctxt)->processor;
goto func_end;
}
- CP_FM_USR(pargs, args->args_proc_ctrl.pargs, status,
+ CP_FM_USR(pargs, args->args_proc_ctrl.args, status,
cb_data_size);
}
if (!status) {
(struct dsp_cbdata *)pargs);
}
- /* CP_TO_USR(args->args_proc_ctrl.pargs, pargs, status, 1); */
+ /* CP_TO_USR(args->args_proc_ctrl.args, pargs, status, 1); */
kfree(pargs);
func_end:
return status;
&num_nodes, &alloc_cnt);
CP_TO_USR(args->args_proc_enumnode_info.node_tab, node_tab, status,
num_nodes);
- CP_TO_USR(args->args_proc_enumnode_info.pu_num_nodes, &num_nodes,
+ CP_TO_USR(args->args_proc_enumnode_info.num_nodes, &num_nodes,
status, 1);
- CP_TO_USR(args->args_proc_enumnode_info.pu_allocated, &alloc_cnt,
+ CP_TO_USR(args->args_proc_enumnode_info.allocated, &alloc_cnt,
status, 1);
return status;
}
return -EINVAL;
status = proc_end_dma(pr_ctxt,
- args->args_proc_dma.pmpu_addr,
+ args->args_proc_dma.mpu_addr,
args->args_proc_dma.size,
args->args_proc_dma.dir);
return status;
return -EINVAL;
status = proc_begin_dma(pr_ctxt,
- args->args_proc_dma.pmpu_addr,
+ args->args_proc_dma.mpu_addr,
args->args_proc_dma.size,
args->args_proc_dma.dir);
return status;
return -EINVAL;
status = proc_flush_memory(pr_ctxt,
- args->args_proc_flushmemory.pmpu_addr,
+ args->args_proc_flushmemory.mpu_addr,
args->args_proc_flushmemory.size,
args->args_proc_flushmemory.ul_flags);
return status;
status =
proc_invalidate_memory(pr_ctxt,
- args->args_proc_invalidatememory.pmpu_addr,
+ args->args_proc_invalidatememory.mpu_addr,
args->args_proc_invalidatememory.size);
return status;
}
return -EINVAL;
status = proc_map(args->args_proc_mapmem.processor,
- args->args_proc_mapmem.pmpu_addr,
+ args->args_proc_mapmem.mpu_addr,
args->args_proc_mapmem.size,
args->args_proc_mapmem.req_addr, &map_addr,
- args->args_proc_mapmem.ul_map_attr, pr_ctxt);
+ args->args_proc_mapmem.map_attr, pr_ctxt);
if (!status) {
- if (put_user(map_addr, args->args_proc_mapmem.pp_map_addr)) {
+ if (put_user(map_addr, args->args_proc_mapmem.map_addr)) {
status = -EINVAL;
proc_un_map(hprocessor, map_addr, pr_ctxt);
}
args->args_proc_register_notify.event_mask,
args->args_proc_register_notify.notify_type,
¬ification);
- CP_TO_USR(args->args_proc_register_notify.hnotification, ¬ification,
+ CP_TO_USR(args->args_proc_register_notify.notification, ¬ification,
status, 1);
return status;
}
args->args_proc_rsvmem.size, &prsv_addr,
pr_ctxt);
if (!status) {
- if (put_user(prsv_addr, args->args_proc_rsvmem.pp_rsv_addr)) {
+ if (put_user(prsv_addr, args->args_proc_rsvmem.rsv_addr)) {
status = -EINVAL;
proc_un_reserve_memory(args->args_proc_rsvmem.
processor, prsv_addr, pr_ctxt);
void *hprocessor = ((struct process_context *)pr_ctxt)->processor;
status = proc_un_reserve_memory(hprocessor,
- args->args_proc_unrsvmem.prsv_addr,
+ args->args_proc_unrsvmem.rsv_addr,
pr_ctxt);
return status;
}
int status = 0;
struct dsp_uuid node_uuid;
u32 cb_data_size = 0;
- u32 __user *psize = (u32 __user *) args->args_node_allocate.pargs;
+ u32 __user *psize = (u32 __user *) args->args_node_allocate.args;
u8 *pargs = NULL;
struct dsp_nodeattrin proc_attr_in, *attr_in = NULL;
struct node_res_object *node_res;
status = -ENOMEM;
}
- CP_FM_USR(pargs, args->args_node_allocate.pargs, status,
+ CP_FM_USR(pargs, args->args_node_allocate.args, status,
cb_data_size);
}
CP_FM_USR(&node_uuid, args->args_node_allocate.node_id_ptr, status, 1);
if (!args->args_proc_register_notify.event_mask)
CP_FM_USR(¬ification,
- args->args_proc_register_notify.hnotification,
+ args->args_proc_register_notify.notification,
status, 1);
status = node_register_notify(node_res->node,
args->args_node_registernotify.event_mask,
args->args_node_registernotify.
notify_type, ¬ification);
- CP_TO_USR(args->args_node_registernotify.hnotification, ¬ification,
+ CP_TO_USR(args->args_node_registernotify.notification, ¬ification,
status, 1);
return status;
}
args->args_strm_registernotify.event_mask,
args->args_strm_registernotify.
notify_type, ¬ification);
- CP_TO_USR(args->args_strm_registernotify.hnotification, ¬ification,
+ CP_TO_USR(args->args_strm_registernotify.notification, ¬ification,
status, 1);
return status;
struct node_createargs create_args; /* Args for node create func */
nodeenv node_env; /* Environment returned by RMS */
struct dcd_genericobj dcd_props; /* Node properties from DCD */
- struct dsp_cbdata *pargs; /* Optional args to pass to node */
+ struct dsp_cbdata *args; /* Optional args to pass to node */
struct ntfy_object *ntfy_obj; /* Manages registered notifications */
char *str_dev_name; /* device name, if device node */
struct sync_object *sync_done; /* Synchronize node_terminate */