/* Check the wwan ips if it is valid with Channel as input. */ staticint ipc_imem_check_wwan_ips(struct ipc_mem_channel *chnl)
{ if (chnl) return chnl->ctype == IPC_CTYPE_WWAN &&
chnl->if_id == IPC_MEM_MUX_IP_CH_IF_ID; returnfalse;
}
/* This timer handler will retry DL buff allocation if a pipe has no free buf * and gives doorbell if TD is available
*/ staticint ipc_imem_tq_td_alloc_timer(struct iosm_imem *ipc_imem, int arg, void *msg, size_t size)
{ bool new_buffers_available = false; bool retry_allocation = false; int i;
for (i = 0; i < IPC_MEM_MAX_CHANNELS; i++) { struct ipc_pipe *pipe = &ipc_imem->channels[i].dl_pipe;
if (!pipe->is_open || pipe->nr_of_queued_entries > 0) continue;
while (ipc_imem_dl_skb_alloc(ipc_imem, pipe))
new_buffers_available = true;
if (pipe->nr_of_queued_entries == 0)
retry_allocation = true;
}
if (new_buffers_available)
ipc_protocol_doorbell_trigger(ipc_imem->ipc_protocol,
IPC_HP_DL_PROCESS);
if (retry_allocation) {
ipc_imem->hrtimer_period =
ktime_set(0, IPC_TD_ALLOC_TIMER_PERIOD_MS * 1000 * 1000ULL); if (!hrtimer_active(&ipc_imem->td_alloc_timer))
hrtimer_start(&ipc_imem->td_alloc_timer,
ipc_imem->hrtimer_period,
HRTIMER_MODE_REL);
} return 0;
}
staticenum hrtimer_restart ipc_imem_td_alloc_timer_cb(struct hrtimer *hr_timer)
{ struct iosm_imem *ipc_imem =
container_of(hr_timer, struct iosm_imem, td_alloc_timer); /* Post an async tasklet event to trigger HP update Doorbell */
ipc_task_queue_send_task(ipc_imem, ipc_imem_tq_td_alloc_timer, 0, NULL,
0, false); return HRTIMER_NORESTART;
}
/* Fast update timer tasklet handler to trigger HP update */ staticint ipc_imem_tq_fast_update_timer_cb(struct iosm_imem *ipc_imem, int arg, void *msg, size_t size)
{
ipc_protocol_doorbell_trigger(ipc_imem->ipc_protocol,
IPC_HP_FAST_TD_UPD_TMR);
return 0;
}
staticenum hrtimer_restart
ipc_imem_fast_update_timer_cb(struct hrtimer *hr_timer)
{ struct iosm_imem *ipc_imem =
container_of(hr_timer, struct iosm_imem, fast_update_timer); /* Post an async tasklet event to trigger HP update Doorbell */
ipc_task_queue_send_task(ipc_imem, ipc_imem_tq_fast_update_timer_cb, 0,
NULL, 0, false); return HRTIMER_NORESTART;
}
/** * ipc_imem_td_update_timer_start - Starts the TD Update Timer if not started. * @ipc_imem: Pointer to imem data-struct
*/ void ipc_imem_td_update_timer_start(struct iosm_imem *ipc_imem)
{ /* Use the TD update timer only in the runtime phase */ if (!ipc_imem->enter_runtime || ipc_imem->td_update_timer_suspended) { /* trigger the doorbell irq on CP directly. */
ipc_protocol_doorbell_trigger(ipc_imem->ipc_protocol,
IPC_HP_TD_UPD_TMR_START); return;
}
if (!hrtimer_active(&ipc_imem->tdupdate_timer)) {
ipc_imem->hrtimer_period =
ktime_set(0, TD_UPDATE_DEFAULT_TIMEOUT_USEC * 1000ULL); if (!hrtimer_active(&ipc_imem->tdupdate_timer))
hrtimer_start(&ipc_imem->tdupdate_timer,
ipc_imem->hrtimer_period,
HRTIMER_MODE_REL);
}
}
void ipc_imem_hrtimer_stop(struct hrtimer *hr_timer)
{ if (hrtimer_active(hr_timer))
hrtimer_cancel(hr_timer);
}
/** * ipc_imem_adb_timer_start - Starts the adb Timer if not starting. * @ipc_imem: Pointer to imem data-struct
*/ void ipc_imem_adb_timer_start(struct iosm_imem *ipc_imem)
{ if (!hrtimer_active(&ipc_imem->adb_timer)) {
ipc_imem->hrtimer_period =
ktime_set(0, IOSM_AGGR_MUX_ADB_FINISH_TIMEOUT_NSEC);
hrtimer_start(&ipc_imem->adb_timer,
ipc_imem->hrtimer_period,
HRTIMER_MODE_REL);
}
}
/* Analyze the uplink pipe of all active channels. */ for (i = 0; i < ipc_imem->nr_of_channels; i++) {
channel = &ipc_imem->channels[i];
if (channel->state != IMEM_CHANNEL_ACTIVE) continue;
pipe = &channel->ul_pipe;
/* Get the reference to the skbuf accumulator list. */
ul_list = &channel->ul_list;
/* Fill the transfer descriptor with the uplink buffer info. */ if (!ipc_imem_check_wwan_ips(channel)) {
hpda_ctrl_pending |=
ipc_protocol_ul_td_send(ipc_imem->ipc_protocol,
pipe, ul_list);
} else {
hpda_pending |=
ipc_protocol_ul_td_send(ipc_imem->ipc_protocol,
pipe, ul_list);
}
}
/* forced HP update needed for non data channels */ if (hpda_ctrl_pending) {
hpda_pending = false;
ipc_protocol_doorbell_trigger(ipc_imem->ipc_protocol,
IPC_HP_UL_WRITE_TD);
}
return hpda_pending;
}
void ipc_imem_ipc_init_check(struct iosm_imem *ipc_imem)
{ int timeout = IPC_MODEM_BOOT_TIMEOUT;
/* Trigger the CP interrupt to enter the init state. */
ipc_doorbell_fire(ipc_imem->pcie, IPC_DOORBELL_IRQ_IPC,
IPC_MEM_DEVICE_IPC_INIT); /* Wait for the CP update. */ do { if (ipc_mmio_get_ipc_state(ipc_imem->mmio) ==
ipc_imem->ipc_requested_state) { /* Prepare the MMIO space */
ipc_mmio_config(ipc_imem->mmio);
/* Trigger the CP irq to enter the running state. */
ipc_imem->ipc_requested_state =
IPC_MEM_DEVICE_IPC_RUNNING;
ipc_doorbell_fire(ipc_imem->pcie, IPC_DOORBELL_IRQ_IPC,
IPC_MEM_DEVICE_IPC_RUNNING);
return;
}
msleep(20);
} while (--timeout);
/* timeout */
dev_err(ipc_imem->dev, "%s: ipc_status(%d) ne. IPC_MEM_DEVICE_IPC_INIT",
ipc_imem_phase_get_string(ipc_imem->phase),
ipc_mmio_get_ipc_state(ipc_imem->mmio));
/* Analyze the packet type and distribute it. */ staticvoid ipc_imem_dl_skb_process(struct iosm_imem *ipc_imem, struct ipc_pipe *pipe, struct sk_buff *skb)
{
u16 port_id;
if (!skb) return;
/* An AT/control or IP packet is expected. */ switch (pipe->channel->ctype) { case IPC_CTYPE_CTRL:
port_id = pipe->channel->channel_id;
ipc_pcie_addr_unmap(ipc_imem->pcie, IPC_CB(skb)->len,
IPC_CB(skb)->mapping,
IPC_CB(skb)->direction); if (port_id == IPC_MEM_CTRL_CHL_ID_7)
ipc_imem_sys_devlink_notify_rx(ipc_imem->ipc_devlink,
skb); elseif (ipc_is_trace_channel(ipc_imem, port_id))
ipc_trace_port_rx(ipc_imem, skb); else
wwan_port_rx(ipc_imem->ipc_port[port_id]->iosm_port,
skb); break;
case IPC_CTYPE_WWAN: if (pipe->channel->if_id == IPC_MEM_MUX_IP_CH_IF_ID)
ipc_mux_dl_decode(ipc_imem->mux, skb); break; default:
dev_err(ipc_imem->dev, "Invalid channel type"); break;
}
}
/* Process the downlink data and pass them to the char or net layer. */ staticvoid ipc_imem_dl_pipe_process(struct iosm_imem *ipc_imem, struct ipc_pipe *pipe)
{
s32 cnt = 0, processed_td_cnt = 0; struct ipc_mem_channel *channel;
u32 head = 0, tail = 0; bool processed = false; struct sk_buff *skb;
/* Seek for pipes with pending DL data. */ while (cnt--) {
skb = ipc_protocol_dl_td_process(ipc_imem->ipc_protocol, pipe);
/* Analyze the packet type and distribute it. */
ipc_imem_dl_skb_process(ipc_imem, pipe, skb);
}
/* try to allocate new empty DL SKbs from head..tail - 1*/ while (ipc_imem_dl_skb_alloc(ipc_imem, pipe))
processed = true;
if (processed && !ipc_imem_check_wwan_ips(channel)) { /* Force HP update for non IP channels */
ipc_protocol_doorbell_trigger(ipc_imem->ipc_protocol,
IPC_HP_DL_PROCESS);
processed = false;
/* If Fast Update timer is already running then stop */
ipc_imem_hrtimer_stop(&ipc_imem->fast_update_timer);
}
/* Any control channel process will get immediate HP update. * Start Fast update timer only for IP channel if all the TDs were * used in last process.
*/ if (processed && (processed_td_cnt == pipe->nr_of_entries - 1)) {
ipc_imem->hrtimer_period =
ktime_set(0, FORCE_UPDATE_DEFAULT_TIMEOUT_USEC * 1000ULL);
hrtimer_start(&ipc_imem->fast_update_timer,
ipc_imem->hrtimer_period, HRTIMER_MODE_REL);
}
if (ipc_imem->app_notify_dl_pend)
complete(&ipc_imem->dl_pend_sem);
}
/* Free UL buffers. */ while (cnt--) {
skb = ipc_protocol_ul_td_process(ipc_imem->ipc_protocol, pipe);
if (!skb) continue;
/* If the user app was suspended in uplink direction - blocking * write, resume it.
*/ if (IPC_CB(skb)->op_type == UL_USR_OP_BLOCKED)
complete(&channel->ul_sem);
/* Free the skbuf element. */ if (IPC_CB(skb)->op_type == UL_MUX_OP_ADB) { if (channel->if_id == IPC_MEM_MUX_IP_CH_IF_ID)
ipc_mux_ul_encoded_process(ipc_imem->mux, skb); else
dev_err(ipc_imem->dev, "OP Type is UL_MUX, unknown if_id %d",
channel->if_id);
} else {
ipc_pcie_kfree_skb(ipc_imem->pcie, skb);
}
}
/* Trace channel stats for IP UL pipe. */ if (ipc_imem_check_wwan_ips(pipe->channel))
ipc_mux_check_n_restart_tx(ipc_imem->mux);
if (ipc_imem->app_notify_ul_pend)
complete(&ipc_imem->ul_pend_sem);
}
/* Execute the UL bundle timer actions, generating the doorbell irq. */ staticint ipc_imem_tq_td_update_timer_cb(struct iosm_imem *ipc_imem, int arg, void *msg, size_t size)
{
ipc_protocol_doorbell_trigger(ipc_imem->ipc_protocol,
IPC_HP_TD_UPD_TMR); return 0;
}
/* Consider link power management in the runtime phase. */ staticvoid ipc_imem_slp_control_exec(struct iosm_imem *ipc_imem)
{ /* link will go down, Test pending UL packets.*/ if (ipc_protocol_pm_dev_sleep_handle(ipc_imem->ipc_protocol) &&
hrtimer_active(&ipc_imem->tdupdate_timer)) { /* Generate the doorbell irq. */
ipc_imem_tq_td_update_timer_cb(ipc_imem, 0, NULL, 0); /* Stop the TD update timer. */
ipc_imem_hrtimer_stop(&ipc_imem->tdupdate_timer); /* Stop the fast update timer. */
ipc_imem_hrtimer_stop(&ipc_imem->fast_update_timer);
}
}
/* Execute startup timer and wait for delayed start (e.g. NAND) */ staticint ipc_imem_tq_startup_timer_cb(struct iosm_imem *ipc_imem, int arg, void *msg, size_t size)
{ /* Update & check the current operation phase. */ if (ipc_imem_phase_update(ipc_imem) != IPC_P_RUN) return -EIO;
if (ipc_mmio_get_ipc_state(ipc_imem->mmio) ==
IPC_MEM_DEVICE_IPC_UNINIT) {
ipc_imem->ipc_requested_state = IPC_MEM_DEVICE_IPC_INIT;
/* Callback to send the modem ready uevent */ staticint ipc_imem_send_mdm_rdy_cb(struct iosm_imem *ipc_imem, int arg, void *msg, size_t size)
{ enum ipc_mem_exec_stage exec_stage =
ipc_imem_get_exec_stage_buffered(ipc_imem);
if (exec_stage == IPC_MEM_EXEC_STAGE_RUN)
ipc_uevent_send(ipc_imem->dev, UEVENT_MDM_READY);
return 0;
}
/* This function is executed in a task context via an ipc_worker object, * as the creation or removal of device can't be done from tasklet.
*/ staticvoid ipc_imem_run_state_worker(struct work_struct *instance)
{ struct ipc_chnl_cfg chnl_cfg_port = { 0 }; struct ipc_mux_config mux_cfg; struct iosm_imem *ipc_imem;
u8 ctrl_chl_idx = 0; int ret;
if (ipc_imem->ipc_status ==
IPC_MEM_DEVICE_IPC_RUNNING) {
schedule_work(&ipc_imem->run_state_worker);
}
}
/* Consider power management in the runtime phase. */
ipc_imem_slp_control_exec(ipc_imem); break; /* Continue with skbuf processing. */
/* Unexpected phases. */ case IPC_P_OFF: case IPC_P_OFF_REQ:
dev_err(ipc_imem->dev, "confused phase %s",
ipc_imem_phase_get_string(phase)); return;
case IPC_P_PSI: if (old_phase != IPC_P_ROM) break;
fallthrough; /* On CP the PSI phase is already active. */
case IPC_P_ROM: /* Before CP ROM driver starts the PSI image, it sets * the exit_code field on the doorbell scratchpad and * triggers the irq.
*/
ipc_imem_rom_irq_exec(ipc_imem); return;
default: break;
}
/* process message ring */
ipc_protocol_msg_process(ipc_imem, irq);
/* process all open pipes */ for (i = 0; i < IPC_MEM_MAX_CHANNELS; i++) { struct ipc_pipe *ul_pipe = &ipc_imem->channels[i].ul_pipe; struct ipc_pipe *dl_pipe = &ipc_imem->channels[i].dl_pipe;
if (dl_pipe->nr_of_queued_entries == 0)
retry_allocation = true;
}
if (ul_pipe->is_open)
ipc_imem_ul_pipe_process(ipc_imem, ul_pipe);
}
/* Try to generate new ADB or ADGH. */ if (ipc_mux_ul_data_encode(ipc_imem->mux)) {
ipc_imem_td_update_timer_start(ipc_imem); if (ipc_imem->mux->protocol == MUX_AGGREGATION)
ipc_imem_adb_timer_start(ipc_imem);
}
/* Continue the send procedure with accumulated SIO or NETIF packets. * Reset the debounce flags.
*/
ul_pending |= ipc_imem_ul_write_td(ipc_imem);
/* if UL data is pending restart TD update timer */ if (ul_pending) {
ipc_imem->hrtimer_period =
ktime_set(0, TD_UPDATE_DEFAULT_TIMEOUT_USEC * 1000ULL); if (!hrtimer_active(&ipc_imem->tdupdate_timer))
hrtimer_start(&ipc_imem->tdupdate_timer,
ipc_imem->hrtimer_period,
HRTIMER_MODE_REL);
}
/* If CP has executed the transition * from IPC_INIT to IPC_RUNNING in the PSI * phase, wake up the flash app to open the pipes.
*/ if ((phase == IPC_P_PSI || phase == IPC_P_EBL) &&
ipc_imem->ipc_requested_state == IPC_MEM_DEVICE_IPC_RUNNING &&
ipc_mmio_get_ipc_state(ipc_imem->mmio) ==
IPC_MEM_DEVICE_IPC_RUNNING) {
complete(&ipc_imem->ipc_devlink->devlink_sio.channel->ul_sem);
}
/* Reset the expected CP state. */
ipc_imem->ipc_requested_state = IPC_MEM_DEVICE_IPC_DONT_CARE;
if (retry_allocation) {
ipc_imem->hrtimer_period =
ktime_set(0, IPC_TD_ALLOC_TIMER_PERIOD_MS * 1000 * 1000ULL); if (!hrtimer_active(&ipc_imem->td_alloc_timer))
hrtimer_start(&ipc_imem->td_alloc_timer,
ipc_imem->hrtimer_period,
HRTIMER_MODE_REL);
}
}
/* Callback by tasklet for handling interrupt events. */ staticint ipc_imem_tq_irq_cb(struct iosm_imem *ipc_imem, int arg, void *msg,
size_t size)
{
ipc_imem_handle_irq(ipc_imem, arg);
return 0;
}
void ipc_imem_ul_send(struct iosm_imem *ipc_imem)
{ /* start doorbell irq delay timer if UL is pending */ if (ipc_imem_ul_write_td(ipc_imem))
ipc_imem_td_update_timer_start(ipc_imem);
}
/* Check the execution stage and update the AP phase */ staticenum ipc_phase ipc_imem_phase_update_check(struct iosm_imem *ipc_imem, enum ipc_mem_exec_stage stage)
{ switch (stage) { case IPC_MEM_EXEC_STAGE_BOOT: if (ipc_imem->phase != IPC_P_ROM) { /* Send this event only once */
ipc_uevent_send(ipc_imem->dev, UEVENT_ROM_READY);
}
ipc_imem->phase = IPC_P_ROM; break;
case IPC_MEM_EXEC_STAGE_PSI:
ipc_imem->phase = IPC_P_PSI; break;
case IPC_MEM_EXEC_STAGE_EBL:
ipc_imem->phase = IPC_P_EBL; break;
case IPC_MEM_EXEC_STAGE_RUN: if (ipc_imem->phase != IPC_P_RUN &&
ipc_imem->ipc_status == IPC_MEM_DEVICE_IPC_RUNNING) {
ipc_uevent_send(ipc_imem->dev, UEVENT_MDM_READY);
}
ipc_imem->phase = IPC_P_RUN; break;
case IPC_MEM_EXEC_STAGE_CRASH: if (ipc_imem->phase != IPC_P_CRASH)
ipc_uevent_send(ipc_imem->dev, UEVENT_CRASH);
ipc_imem->phase = IPC_P_CRASH; break;
case IPC_MEM_EXEC_STAGE_CD_READY: if (ipc_imem->phase != IPC_P_CD_READY)
ipc_uevent_send(ipc_imem->dev, UEVENT_CD_READY);
ipc_imem->phase = IPC_P_CD_READY; break;
default: /* unknown exec stage: * assume that link is down and send info to listeners
*/
ipc_uevent_send(ipc_imem->dev, UEVENT_CD_READY_LINK_DOWN); break;
}
return ipc_imem->phase;
}
/* Send msg to device to open pipe */ staticbool ipc_imem_pipe_open(struct iosm_imem *ipc_imem, struct ipc_pipe *pipe)
{ union ipc_msg_prep_args prep_args = {
.pipe_open.pipe = pipe,
};
if (ipc_protocol_msg_send(ipc_imem->ipc_protocol,
IPC_MSG_PREP_PIPE_OPEN, &prep_args) == 0)
pipe->is_open = true;
return pipe->is_open;
}
/* Allocates the TDs for the given pipe along with firing HP update DB. */ staticint ipc_imem_tq_pipe_td_alloc(struct iosm_imem *ipc_imem, int arg, void *msg, size_t size)
{ struct ipc_pipe *dl_pipe = msg; bool processed = false; int i;
for (i = 0; i < dl_pipe->nr_of_entries - 1; i++)
processed |= ipc_imem_dl_skb_alloc(ipc_imem, dl_pipe);
/* Trigger the doorbell irq to inform CP that new downlink buffers are * available.
*/ if (processed)
ipc_protocol_doorbell_trigger(ipc_imem->ipc_protocol, arg);
/* Get the CP execution state and map it to the AP phase. */ enum ipc_phase ipc_imem_phase_update(struct iosm_imem *ipc_imem)
{ enum ipc_mem_exec_stage exec_stage =
ipc_imem_get_exec_stage_buffered(ipc_imem); /* If the CP stage is undef, return the internal precalculated phase. */ return ipc_imem->phase == IPC_P_OFF_REQ ?
ipc_imem->phase :
ipc_imem_phase_update_check(ipc_imem, exec_stage);
}
void ipc_imem_channel_close(struct iosm_imem *ipc_imem, int channel_id)
{ struct ipc_mem_channel *channel;
if (channel_id < 0 || channel_id >= ipc_imem->nr_of_channels) {
dev_err(ipc_imem->dev, "invalid channel id %d", channel_id); return;
}
channel = &ipc_imem->channels[channel_id];
if (channel->state == IMEM_CHANNEL_FREE) {
dev_err(ipc_imem->dev, "ch[%d]: invalid channel state %d",
channel_id, channel->state); return;
}
/* Free only the channel id in the CP power off mode. */ if (channel->state == IMEM_CHANNEL_RESERVED) /* Release only the channel id. */ goto channel_free;
if (ipc_imem->phase == IPC_P_RUN) {
ipc_imem_pipe_close(ipc_imem, &channel->ul_pipe);
ipc_imem_pipe_close(ipc_imem, &channel->dl_pipe);
}
int ipc_imem_channel_alloc(struct iosm_imem *ipc_imem, int index, enum ipc_ctype ctype)
{ struct ipc_mem_channel *channel; int i;
/* Find channel of given type/index */ for (i = 0; i < ipc_imem->nr_of_channels; i++) {
channel = &ipc_imem->channels[i]; if (channel->ctype == ctype && channel->index == index) break;
}
if (i >= ipc_imem->nr_of_channels) {
dev_dbg(ipc_imem->dev, "no channel definition for index=%d ctype=%d", index,
ctype); return -ECHRNG;
}
if (ipc_imem->channels[i].state != IMEM_CHANNEL_FREE) {
dev_dbg(ipc_imem->dev, "channel is in use"); return -EBUSY;
}
if (channel->ctype == IPC_CTYPE_WWAN &&
index == IPC_MEM_MUX_IP_CH_IF_ID)
channel->if_id = index;
/* Send IPC protocol uninit to the modem when Link is active. */ staticvoid ipc_imem_device_ipc_uninit(struct iosm_imem *ipc_imem)
{ int timeout = IPC_MODEM_UNINIT_TIMEOUT_MS; enum ipc_mem_device_ipc_state ipc_state;
/* When PCIe link is up set IPC_UNINIT * of the modem otherwise ignore it when PCIe link down happens.
*/ if (ipc_pcie_check_data_link_active(ipc_imem->pcie)) { /* set modem to UNINIT * (in case we want to reload the AP driver without resetting * the modem)
*/
ipc_doorbell_fire(ipc_imem->pcie, IPC_DOORBELL_IRQ_IPC,
IPC_MEM_DEVICE_IPC_UNINIT);
ipc_state = ipc_mmio_get_ipc_state(ipc_imem->mmio);
/* Wait for maximum 30ms to allow the Modem to uninitialize the * protocol.
*/ while ((ipc_state <= IPC_MEM_DEVICE_IPC_DONT_CARE) &&
(ipc_state != IPC_MEM_DEVICE_IPC_UNINIT) &&
(timeout > 0)) {
usleep_range(1000, 1250);
timeout--;
ipc_state = ipc_mmio_get_ipc_state(ipc_imem->mmio);
}
}
}
/* After CP has unblocked the PCIe link, save the start address of the doorbell * scratchpad and prepare the shared memory region. If the flashing to RAM * procedure shall be executed, copy the chip information from the doorbell * scratchtpad to the application buffer and wake up the flash app.
*/ staticint ipc_imem_config(struct iosm_imem *ipc_imem)
{ enum ipc_phase phase;
/* Initialize the semaphore for the blocking read UL/DL transfer. */
init_completion(&ipc_imem->ul_pend_sem);
/* Either CP shall be in the power off or power on phase. */ switch (phase) { case IPC_P_ROM:
ipc_imem->hrtimer_period = ktime_set(0, 1000 * 1000 * 1000ULL); /* poll execution stage (for delayed start, e.g. NAND) */ if (!hrtimer_active(&ipc_imem->startup_timer))
hrtimer_start(&ipc_imem->startup_timer,
ipc_imem->hrtimer_period,
HRTIMER_MODE_REL); return 0;
case IPC_P_PSI: case IPC_P_EBL: case IPC_P_RUN: /* The initial IPC state is IPC_MEM_DEVICE_IPC_UNINIT. */
ipc_imem->ipc_requested_state = IPC_MEM_DEVICE_IPC_UNINIT;
/* Verify the exepected initial state. */ if (ipc_imem->ipc_requested_state ==
ipc_mmio_get_ipc_state(ipc_imem->mmio)) {
ipc_imem_ipc_init_check(ipc_imem);
return 0;
}
dev_err(ipc_imem->dev, "ipc_status(%d) != IPC_MEM_DEVICE_IPC_UNINIT",
ipc_mmio_get_ipc_state(ipc_imem->mmio)); break; case IPC_P_CRASH: case IPC_P_CD_READY:
dev_dbg(ipc_imem->dev, "Modem is in phase %d, reset Modem to collect CD",
phase); return 0; default:
dev_err(ipc_imem->dev, "unexpected operation phase %d", phase); break;
}
/* Verify the CP execution state, copy the chip info, * change the execution phase to ROM
*/ staticint ipc_imem_devlink_trigger_chip_info_cb(struct iosm_imem *ipc_imem, int arg, void *msg,
size_t msgsize)
{ enum ipc_mem_exec_stage stage; struct sk_buff *skb; int rc = -EINVAL;
size_t size;
/* Test the CP execution state. */
stage = ipc_mmio_get_exec_stage(ipc_imem->mmio); if (stage != IPC_MEM_EXEC_STAGE_BOOT) {
dev_err(ipc_imem->dev, "Execution_stage: expected BOOT, received = %X", stage); goto trigger_chip_info_fail;
} /* Allocate a new sk buf for the chip info. */
size = ipc_imem->mmio->chip_info_size; if (size > IOSM_CHIP_INFO_SIZE_MAX) goto trigger_chip_info_fail;
skb = ipc_pcie_alloc_local_skb(ipc_imem->pcie, GFP_ATOMIC, size); if (!skb) {
dev_err(ipc_imem->dev, "exhausted skbuf kernel DL memory");
rc = -ENOMEM; goto trigger_chip_info_fail;
} /* Copy the chip info characters into the ipc_skb. */
ipc_mmio_copy_chip_info(ipc_imem->mmio, skb_put(skb, size), size); /* First change to the ROM boot phase. */
dev_dbg(ipc_imem->dev, "execution_stage[%X] eq. BOOT", stage);
ipc_imem->phase = ipc_imem_phase_update(ipc_imem);
ipc_imem_sys_devlink_notify_rx(ipc_imem->ipc_devlink, skb);
rc = 0;
trigger_chip_info_fail: return rc;
}
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.