Loading drivers/bus/mhi/core/internal.h +1 −1 Original line number Diff line number Diff line Loading @@ -682,7 +682,7 @@ void mhi_rddm_prepare(struct mhi_controller *mhi_cntrl, struct image_info *img_info); void mhi_fw_load_handler(struct mhi_controller *mhi_cntrl); int mhi_prepare_channel(struct mhi_controller *mhi_cntrl, struct mhi_chan *mhi_chan); struct mhi_chan *mhi_chan, unsigned int flags); int mhi_init_chan_ctxt(struct mhi_controller *mhi_cntrl, struct mhi_chan *mhi_chan); void mhi_deinit_chan_ctxt(struct mhi_controller *mhi_cntrl, Loading drivers/bus/mhi/core/main.c +6 −3 Original line number Diff line number Diff line Loading @@ -1430,7 +1430,7 @@ static void mhi_unprepare_channel(struct mhi_controller *mhi_cntrl, } int mhi_prepare_channel(struct mhi_controller *mhi_cntrl, struct mhi_chan *mhi_chan) struct mhi_chan *mhi_chan, unsigned int flags) { int ret = 0; struct device *dev = &mhi_chan->mhi_dev->dev; Loading @@ -1455,6 +1455,9 @@ int mhi_prepare_channel(struct mhi_controller *mhi_cntrl, if (ret) goto error_pm_state; if (mhi_chan->dir == DMA_FROM_DEVICE) mhi_chan->pre_alloc = !!(flags & MHI_CH_INBOUND_ALLOC_BUFS); /* Pre-allocate buffer for xfer ring */ if (mhi_chan->pre_alloc) { int nr_el = get_nr_avail_ring_elements(mhi_cntrl, Loading Loading @@ -1610,7 +1613,7 @@ void mhi_reset_chan(struct mhi_controller *mhi_cntrl, struct mhi_chan *mhi_chan) } /* Move channel to start state */ int mhi_prepare_for_transfer(struct mhi_device *mhi_dev) int mhi_prepare_for_transfer(struct mhi_device *mhi_dev, unsigned int flags) { int ret, dir; struct mhi_controller *mhi_cntrl = mhi_dev->mhi_cntrl; Loading @@ -1621,7 +1624,7 @@ int mhi_prepare_for_transfer(struct mhi_device *mhi_dev) if (!mhi_chan) continue; ret = mhi_prepare_channel(mhi_cntrl, mhi_chan); ret = mhi_prepare_channel(mhi_cntrl, mhi_chan, flags); if (ret) goto error_open_chan; } Loading drivers/net/mhi/net.c +1 −1 Original line number Diff line number Diff line Loading @@ -335,7 +335,7 @@ static int mhi_net_newlink(void *ctxt, struct net_device *ndev, u32 if_id, u64_stats_init(&mhi_netdev->stats.tx_syncp); /* Start MHI channels */ err = mhi_prepare_for_transfer(mhi_dev); err = mhi_prepare_for_transfer(mhi_dev, 0); if (err) goto out_err; Loading drivers/net/wwan/mhi_wwan_ctrl.c +1 −1 Original line number Diff line number Diff line Loading @@ -110,7 +110,7 @@ static int mhi_wwan_ctrl_start(struct wwan_port *port) int ret; /* Start mhi device's channel(s) */ ret = mhi_prepare_for_transfer(mhiwwan->mhi_dev); ret = mhi_prepare_for_transfer(mhiwwan->mhi_dev, 0); if (ret) return ret; Loading include/linux/mhi.h +6 −1 Original line number Diff line number Diff line Loading @@ -719,8 +719,13 @@ void mhi_device_put(struct mhi_device *mhi_dev); * host and device execution environments match and * channels are in a DISABLED state. * @mhi_dev: Device associated with the channels * @flags: MHI channel flags */ int mhi_prepare_for_transfer(struct mhi_device *mhi_dev); int mhi_prepare_for_transfer(struct mhi_device *mhi_dev, unsigned int flags); /* Automatically allocate and queue inbound buffers */ #define MHI_CH_INBOUND_ALLOC_BUFS BIT(0) /** * mhi_unprepare_from_transfer - Reset UL and DL channels for data transfer. Loading Loading
drivers/bus/mhi/core/internal.h +1 −1 Original line number Diff line number Diff line Loading @@ -682,7 +682,7 @@ void mhi_rddm_prepare(struct mhi_controller *mhi_cntrl, struct image_info *img_info); void mhi_fw_load_handler(struct mhi_controller *mhi_cntrl); int mhi_prepare_channel(struct mhi_controller *mhi_cntrl, struct mhi_chan *mhi_chan); struct mhi_chan *mhi_chan, unsigned int flags); int mhi_init_chan_ctxt(struct mhi_controller *mhi_cntrl, struct mhi_chan *mhi_chan); void mhi_deinit_chan_ctxt(struct mhi_controller *mhi_cntrl, Loading
drivers/bus/mhi/core/main.c +6 −3 Original line number Diff line number Diff line Loading @@ -1430,7 +1430,7 @@ static void mhi_unprepare_channel(struct mhi_controller *mhi_cntrl, } int mhi_prepare_channel(struct mhi_controller *mhi_cntrl, struct mhi_chan *mhi_chan) struct mhi_chan *mhi_chan, unsigned int flags) { int ret = 0; struct device *dev = &mhi_chan->mhi_dev->dev; Loading @@ -1455,6 +1455,9 @@ int mhi_prepare_channel(struct mhi_controller *mhi_cntrl, if (ret) goto error_pm_state; if (mhi_chan->dir == DMA_FROM_DEVICE) mhi_chan->pre_alloc = !!(flags & MHI_CH_INBOUND_ALLOC_BUFS); /* Pre-allocate buffer for xfer ring */ if (mhi_chan->pre_alloc) { int nr_el = get_nr_avail_ring_elements(mhi_cntrl, Loading Loading @@ -1610,7 +1613,7 @@ void mhi_reset_chan(struct mhi_controller *mhi_cntrl, struct mhi_chan *mhi_chan) } /* Move channel to start state */ int mhi_prepare_for_transfer(struct mhi_device *mhi_dev) int mhi_prepare_for_transfer(struct mhi_device *mhi_dev, unsigned int flags) { int ret, dir; struct mhi_controller *mhi_cntrl = mhi_dev->mhi_cntrl; Loading @@ -1621,7 +1624,7 @@ int mhi_prepare_for_transfer(struct mhi_device *mhi_dev) if (!mhi_chan) continue; ret = mhi_prepare_channel(mhi_cntrl, mhi_chan); ret = mhi_prepare_channel(mhi_cntrl, mhi_chan, flags); if (ret) goto error_open_chan; } Loading
drivers/net/mhi/net.c +1 −1 Original line number Diff line number Diff line Loading @@ -335,7 +335,7 @@ static int mhi_net_newlink(void *ctxt, struct net_device *ndev, u32 if_id, u64_stats_init(&mhi_netdev->stats.tx_syncp); /* Start MHI channels */ err = mhi_prepare_for_transfer(mhi_dev); err = mhi_prepare_for_transfer(mhi_dev, 0); if (err) goto out_err; Loading
drivers/net/wwan/mhi_wwan_ctrl.c +1 −1 Original line number Diff line number Diff line Loading @@ -110,7 +110,7 @@ static int mhi_wwan_ctrl_start(struct wwan_port *port) int ret; /* Start mhi device's channel(s) */ ret = mhi_prepare_for_transfer(mhiwwan->mhi_dev); ret = mhi_prepare_for_transfer(mhiwwan->mhi_dev, 0); if (ret) return ret; Loading
include/linux/mhi.h +6 −1 Original line number Diff line number Diff line Loading @@ -719,8 +719,13 @@ void mhi_device_put(struct mhi_device *mhi_dev); * host and device execution environments match and * channels are in a DISABLED state. * @mhi_dev: Device associated with the channels * @flags: MHI channel flags */ int mhi_prepare_for_transfer(struct mhi_device *mhi_dev); int mhi_prepare_for_transfer(struct mhi_device *mhi_dev, unsigned int flags); /* Automatically allocate and queue inbound buffers */ #define MHI_CH_INBOUND_ALLOC_BUFS BIT(0) /** * mhi_unprepare_from_transfer - Reset UL and DL channels for data transfer. Loading