Loading drivers/spi/spi-axi-spi-engine.c +0 −1 Original line number Diff line number Diff line Loading @@ -525,7 +525,6 @@ static int spi_engine_probe(struct platform_device *pdev) if (ret) goto err_ref_clk_disable; master->dev.parent = &pdev->dev; master->dev.of_node = pdev->dev.of_node; master->mode_bits = SPI_CPOL | SPI_CPHA | SPI_3WIRE; master->bits_per_word_mask = SPI_BPW_MASK(8); Loading drivers/spi/spi-bcm53xx.c +76 −2 Original line number Diff line number Diff line Loading @@ -10,6 +10,7 @@ #include "spi-bcm53xx.h" #define BCM53XXSPI_MAX_SPI_BAUD 13500000 /* 216 MHz? */ #define BCM53XXSPI_FLASH_WINDOW SZ_32M /* The longest observed required wait was 19 ms */ #define BCM53XXSPI_SPE_TIMEOUT_MS 80 Loading @@ -17,8 +18,10 @@ struct bcm53xxspi { struct bcma_device *core; struct spi_master *master; void __iomem *mmio_base; size_t read_offset; bool bspi; /* Boot SPI mode with memory mapping */ }; static inline u32 bcm53xxspi_read(struct bcm53xxspi *b53spi, u16 offset) Loading @@ -32,6 +35,50 @@ static inline void bcm53xxspi_write(struct bcm53xxspi *b53spi, u16 offset, bcma_write32(b53spi->core, offset, value); } static void bcm53xxspi_disable_bspi(struct bcm53xxspi *b53spi) { struct device *dev = &b53spi->core->dev; unsigned long deadline; u32 tmp; if (!b53spi->bspi) return; tmp = bcm53xxspi_read(b53spi, B53SPI_BSPI_MAST_N_BOOT_CTRL); if (tmp & 0x1) return; deadline = jiffies + usecs_to_jiffies(200); do { tmp = bcm53xxspi_read(b53spi, B53SPI_BSPI_BUSY_STATUS); if (!(tmp & 0x1)) { bcm53xxspi_write(b53spi, B53SPI_BSPI_MAST_N_BOOT_CTRL, 0x1); ndelay(200); b53spi->bspi = false; return; } udelay(1); } while (!time_after_eq(jiffies, deadline)); dev_warn(dev, "Timeout disabling BSPI\n"); } static void bcm53xxspi_enable_bspi(struct bcm53xxspi *b53spi) { u32 tmp; if (b53spi->bspi) return; tmp = bcm53xxspi_read(b53spi, B53SPI_BSPI_MAST_N_BOOT_CTRL); if (!(tmp & 0x1)) return; bcm53xxspi_write(b53spi, B53SPI_BSPI_MAST_N_BOOT_CTRL, 0x0); b53spi->bspi = true; } static inline unsigned int bcm53xxspi_calc_timeout(size_t len) { /* Do some magic calculation based on length and buad. Add 10% and 1. */ Loading Loading @@ -176,6 +223,8 @@ static int bcm53xxspi_transfer_one(struct spi_master *master, u8 *buf; size_t left; bcm53xxspi_disable_bspi(b53spi); if (t->tx_buf) { buf = (u8 *)t->tx_buf; left = t->len; Loading Loading @@ -206,6 +255,22 @@ static int bcm53xxspi_transfer_one(struct spi_master *master, return 0; } static int bcm53xxspi_flash_read(struct spi_device *spi, struct spi_flash_read_message *msg) { struct bcm53xxspi *b53spi = spi_master_get_devdata(spi->master); int ret = 0; if (msg->from + msg->len > BCM53XXSPI_FLASH_WINDOW) return -EINVAL; bcm53xxspi_enable_bspi(b53spi); memcpy_fromio(msg->buf, b53spi->mmio_base + msg->from, msg->len); msg->retlen = msg->len; return ret; } /************************************************** * BCMA **************************************************/ Loading @@ -222,6 +287,7 @@ MODULE_DEVICE_TABLE(bcma, bcm53xxspi_bcma_tbl); static int bcm53xxspi_bcma_probe(struct bcma_device *core) { struct device *dev = &core->dev; struct bcm53xxspi *b53spi; struct spi_master *master; int err; Loading @@ -231,7 +297,7 @@ static int bcm53xxspi_bcma_probe(struct bcma_device *core) return -ENOTSUPP; } master = spi_alloc_master(&core->dev, sizeof(*b53spi)); master = spi_alloc_master(dev, sizeof(*b53spi)); if (!master) return -ENOMEM; Loading @@ -239,11 +305,19 @@ static int bcm53xxspi_bcma_probe(struct bcma_device *core) b53spi->master = master; b53spi->core = core; if (core->addr_s[0]) b53spi->mmio_base = devm_ioremap(dev, core->addr_s[0], BCM53XXSPI_FLASH_WINDOW); b53spi->bspi = true; bcm53xxspi_disable_bspi(b53spi); master->transfer_one = bcm53xxspi_transfer_one; if (b53spi->mmio_base) master->spi_flash_read = bcm53xxspi_flash_read; bcma_set_drvdata(core, b53spi); err = devm_spi_register_master(&core->dev, master); err = devm_spi_register_master(dev, master); if (err) { spi_master_put(master); bcma_set_drvdata(core, NULL); Loading drivers/spi/spi-cadence.c +138 −106 Original line number Diff line number Diff line Loading @@ -19,44 +19,46 @@ #include <linux/of_irq.h> #include <linux/of_address.h> #include <linux/platform_device.h> #include <linux/pm_runtime.h> #include <linux/spi/spi.h> /* Name of this driver */ #define CDNS_SPI_NAME "cdns-spi" /* Register offset definitions */ #define CDNS_SPI_CR_OFFSET 0x00 /* Configuration Register, RW */ #define CDNS_SPI_ISR_OFFSET 0x04 /* Interrupt Status Register, RO */ #define CDNS_SPI_IER_OFFSET 0x08 /* Interrupt Enable Register, WO */ #define CDNS_SPI_IDR_OFFSET 0x0c /* Interrupt Disable Register, WO */ #define CDNS_SPI_IMR_OFFSET 0x10 /* Interrupt Enabled Mask Register, RO */ #define CDNS_SPI_ER_OFFSET 0x14 /* Enable/Disable Register, RW */ #define CDNS_SPI_DR_OFFSET 0x18 /* Delay Register, RW */ #define CDNS_SPI_TXD_OFFSET 0x1C /* Data Transmit Register, WO */ #define CDNS_SPI_RXD_OFFSET 0x20 /* Data Receive Register, RO */ #define CDNS_SPI_SICR_OFFSET 0x24 /* Slave Idle Count Register, RW */ #define CDNS_SPI_THLD_OFFSET 0x28 /* Transmit FIFO Watermark Register,RW */ #define CDNS_SPI_CR 0x00 /* Configuration Register, RW */ #define CDNS_SPI_ISR 0x04 /* Interrupt Status Register, RO */ #define CDNS_SPI_IER 0x08 /* Interrupt Enable Register, WO */ #define CDNS_SPI_IDR 0x0c /* Interrupt Disable Register, WO */ #define CDNS_SPI_IMR 0x10 /* Interrupt Enabled Mask Register, RO */ #define CDNS_SPI_ER 0x14 /* Enable/Disable Register, RW */ #define CDNS_SPI_DR 0x18 /* Delay Register, RW */ #define CDNS_SPI_TXD 0x1C /* Data Transmit Register, WO */ #define CDNS_SPI_RXD 0x20 /* Data Receive Register, RO */ #define CDNS_SPI_SICR 0x24 /* Slave Idle Count Register, RW */ #define CDNS_SPI_THLD 0x28 /* Transmit FIFO Watermark Register,RW */ #define SPI_AUTOSUSPEND_TIMEOUT 3000 /* * SPI Configuration Register bit Masks * * This register contains various control bits that affect the operation * of the SPI controller */ #define CDNS_SPI_CR_MANSTRT_MASK 0x00010000 /* Manual TX Start */ #define CDNS_SPI_CR_CPHA_MASK 0x00000004 /* Clock Phase Control */ #define CDNS_SPI_CR_CPOL_MASK 0x00000002 /* Clock Polarity Control */ #define CDNS_SPI_CR_SSCTRL_MASK 0x00003C00 /* Slave Select Mask */ #define CDNS_SPI_CR_PERI_SEL_MASK 0x00000200 /* Peripheral Select Decode */ #define CDNS_SPI_CR_BAUD_DIV_MASK 0x00000038 /* Baud Rate Divisor Mask */ #define CDNS_SPI_CR_MSTREN_MASK 0x00000001 /* Master Enable Mask */ #define CDNS_SPI_CR_MANSTRTEN_MASK 0x00008000 /* Manual TX Enable Mask */ #define CDNS_SPI_CR_SSFORCE_MASK 0x00004000 /* Manual SS Enable Mask */ #define CDNS_SPI_CR_BAUD_DIV_4_MASK 0x00000008 /* Default Baud Div Mask */ #define CDNS_SPI_CR_DEFAULT_MASK (CDNS_SPI_CR_MSTREN_MASK | \ CDNS_SPI_CR_SSCTRL_MASK | \ CDNS_SPI_CR_SSFORCE_MASK | \ CDNS_SPI_CR_BAUD_DIV_4_MASK) #define CDNS_SPI_CR_MANSTRT 0x00010000 /* Manual TX Start */ #define CDNS_SPI_CR_CPHA 0x00000004 /* Clock Phase Control */ #define CDNS_SPI_CR_CPOL 0x00000002 /* Clock Polarity Control */ #define CDNS_SPI_CR_SSCTRL 0x00003C00 /* Slave Select Mask */ #define CDNS_SPI_CR_PERI_SEL 0x00000200 /* Peripheral Select Decode */ #define CDNS_SPI_CR_BAUD_DIV 0x00000038 /* Baud Rate Divisor Mask */ #define CDNS_SPI_CR_MSTREN 0x00000001 /* Master Enable Mask */ #define CDNS_SPI_CR_MANSTRTEN 0x00008000 /* Manual TX Enable Mask */ #define CDNS_SPI_CR_SSFORCE 0x00004000 /* Manual SS Enable Mask */ #define CDNS_SPI_CR_BAUD_DIV_4 0x00000008 /* Default Baud Div Mask */ #define CDNS_SPI_CR_DEFAULT (CDNS_SPI_CR_MSTREN | \ CDNS_SPI_CR_SSCTRL | \ CDNS_SPI_CR_SSFORCE | \ CDNS_SPI_CR_BAUD_DIV_4) /* * SPI Configuration Register - Baud rate and slave select Loading @@ -77,21 +79,21 @@ * All the four interrupt registers (Status/Mask/Enable/Disable) have the same * bit definitions. */ #define CDNS_SPI_IXR_TXOW_MASK 0x00000004 /* SPI TX FIFO Overwater */ #define CDNS_SPI_IXR_MODF_MASK 0x00000002 /* SPI Mode Fault */ #define CDNS_SPI_IXR_RXNEMTY_MASK 0x00000010 /* SPI RX FIFO Not Empty */ #define CDNS_SPI_IXR_DEFAULT_MASK (CDNS_SPI_IXR_TXOW_MASK | \ CDNS_SPI_IXR_MODF_MASK) #define CDNS_SPI_IXR_TXFULL_MASK 0x00000008 /* SPI TX Full */ #define CDNS_SPI_IXR_ALL_MASK 0x0000007F /* SPI all interrupts */ #define CDNS_SPI_IXR_TXOW 0x00000004 /* SPI TX FIFO Overwater */ #define CDNS_SPI_IXR_MODF 0x00000002 /* SPI Mode Fault */ #define CDNS_SPI_IXR_RXNEMTY 0x00000010 /* SPI RX FIFO Not Empty */ #define CDNS_SPI_IXR_DEFAULT (CDNS_SPI_IXR_TXOW | \ CDNS_SPI_IXR_MODF) #define CDNS_SPI_IXR_TXFULL 0x00000008 /* SPI TX Full */ #define CDNS_SPI_IXR_ALL 0x0000007F /* SPI all interrupts */ /* * SPI Enable Register bit Masks * * This register is used to enable or disable the SPI controller */ #define CDNS_SPI_ER_ENABLE_MASK 0x00000001 /* SPI Enable Bit Mask */ #define CDNS_SPI_ER_DISABLE_MASK 0x0 /* SPI Disable Bit Mask */ #define CDNS_SPI_ER_ENABLE 0x00000001 /* SPI Enable Bit Mask */ #define CDNS_SPI_ER_DISABLE 0x0 /* SPI Disable Bit Mask */ /* SPI FIFO depth in bytes */ #define CDNS_SPI_FIFO_DEPTH 128 Loading Loading @@ -149,56 +151,51 @@ static inline void cdns_spi_write(struct cdns_spi *xspi, u32 offset, u32 val) */ static void cdns_spi_init_hw(struct cdns_spi *xspi) { u32 ctrl_reg = CDNS_SPI_CR_DEFAULT_MASK; u32 ctrl_reg = CDNS_SPI_CR_DEFAULT; if (xspi->is_decoded_cs) ctrl_reg |= CDNS_SPI_CR_PERI_SEL_MASK; ctrl_reg |= CDNS_SPI_CR_PERI_SEL; cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_DISABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_IDR_OFFSET, CDNS_SPI_IXR_ALL_MASK); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_DISABLE); cdns_spi_write(xspi, CDNS_SPI_IDR, CDNS_SPI_IXR_ALL); /* Clear the RX FIFO */ while (cdns_spi_read(xspi, CDNS_SPI_ISR_OFFSET) & CDNS_SPI_IXR_RXNEMTY_MASK) cdns_spi_read(xspi, CDNS_SPI_RXD_OFFSET); while (cdns_spi_read(xspi, CDNS_SPI_ISR) & CDNS_SPI_IXR_RXNEMTY) cdns_spi_read(xspi, CDNS_SPI_RXD); cdns_spi_write(xspi, CDNS_SPI_ISR_OFFSET, CDNS_SPI_IXR_ALL_MASK); cdns_spi_write(xspi, CDNS_SPI_CR_OFFSET, ctrl_reg); cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_ENABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_ISR, CDNS_SPI_IXR_ALL); cdns_spi_write(xspi, CDNS_SPI_CR, ctrl_reg); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_ENABLE); } /** * cdns_spi_chipselect - Select or deselect the chip select line * @spi: Pointer to the spi_device structure * @is_on: Select(0) or deselect (1) the chip select line * @is_high: Select(0) or deselect (1) the chip select line */ static void cdns_spi_chipselect(struct spi_device *spi, bool is_high) { struct cdns_spi *xspi = spi_master_get_devdata(spi->master); u32 ctrl_reg; ctrl_reg = cdns_spi_read(xspi, CDNS_SPI_CR_OFFSET); ctrl_reg = cdns_spi_read(xspi, CDNS_SPI_CR); if (is_high) { /* Deselect the slave */ ctrl_reg |= CDNS_SPI_CR_SSCTRL_MASK; ctrl_reg |= CDNS_SPI_CR_SSCTRL; } else { /* Select the slave */ ctrl_reg &= ~CDNS_SPI_CR_SSCTRL_MASK; ctrl_reg &= ~CDNS_SPI_CR_SSCTRL; if (!(xspi->is_decoded_cs)) ctrl_reg |= ((~(CDNS_SPI_SS0 << spi->chip_select)) << CDNS_SPI_SS_SHIFT) & CDNS_SPI_CR_SSCTRL_MASK; CDNS_SPI_CR_SSCTRL; else ctrl_reg |= (spi->chip_select << CDNS_SPI_SS_SHIFT) & CDNS_SPI_CR_SSCTRL_MASK; CDNS_SPI_CR_SSCTRL; } cdns_spi_write(xspi, CDNS_SPI_CR_OFFSET, ctrl_reg); cdns_spi_write(xspi, CDNS_SPI_CR, ctrl_reg); } /** Loading @@ -212,14 +209,15 @@ static void cdns_spi_config_clock_mode(struct spi_device *spi) struct cdns_spi *xspi = spi_master_get_devdata(spi->master); u32 ctrl_reg, new_ctrl_reg; new_ctrl_reg = ctrl_reg = cdns_spi_read(xspi, CDNS_SPI_CR_OFFSET); new_ctrl_reg = cdns_spi_read(xspi, CDNS_SPI_CR); ctrl_reg = new_ctrl_reg; /* Set the SPI clock phase and clock polarity */ new_ctrl_reg &= ~(CDNS_SPI_CR_CPHA_MASK | CDNS_SPI_CR_CPOL_MASK); new_ctrl_reg &= ~(CDNS_SPI_CR_CPHA | CDNS_SPI_CR_CPOL); if (spi->mode & SPI_CPHA) new_ctrl_reg |= CDNS_SPI_CR_CPHA_MASK; new_ctrl_reg |= CDNS_SPI_CR_CPHA; if (spi->mode & SPI_CPOL) new_ctrl_reg |= CDNS_SPI_CR_CPOL_MASK; new_ctrl_reg |= CDNS_SPI_CR_CPOL; if (new_ctrl_reg != ctrl_reg) { /* Loading @@ -228,11 +226,9 @@ static void cdns_spi_config_clock_mode(struct spi_device *spi) * polarity as it will cause the SPI slave to see spurious clock * transitions. To workaround the issue toggle the ER register. */ cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_DISABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_CR_OFFSET, new_ctrl_reg); cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_ENABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_DISABLE); cdns_spi_write(xspi, CDNS_SPI_CR, new_ctrl_reg); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_ENABLE); } } Loading @@ -259,7 +255,7 @@ static void cdns_spi_config_clock_freq(struct spi_device *spi, frequency = clk_get_rate(xspi->ref_clk); ctrl_reg = cdns_spi_read(xspi, CDNS_SPI_CR_OFFSET); ctrl_reg = cdns_spi_read(xspi, CDNS_SPI_CR); /* Set the clock frequency */ if (xspi->speed_hz != transfer->speed_hz) { Loading @@ -269,12 +265,12 @@ static void cdns_spi_config_clock_freq(struct spi_device *spi, (frequency / (2 << baud_rate_val)) > transfer->speed_hz) baud_rate_val++; ctrl_reg &= ~CDNS_SPI_CR_BAUD_DIV_MASK; ctrl_reg &= ~CDNS_SPI_CR_BAUD_DIV; ctrl_reg |= baud_rate_val << CDNS_SPI_BAUD_DIV_SHIFT; xspi->speed_hz = frequency / (2 << baud_rate_val); } cdns_spi_write(xspi, CDNS_SPI_CR_OFFSET, ctrl_reg); cdns_spi_write(xspi, CDNS_SPI_CR, ctrl_reg); } /** Loading Loading @@ -313,10 +309,9 @@ static void cdns_spi_fill_tx_fifo(struct cdns_spi *xspi) while ((trans_cnt < CDNS_SPI_FIFO_DEPTH) && (xspi->tx_bytes > 0)) { if (xspi->txbuf) cdns_spi_write(xspi, CDNS_SPI_TXD_OFFSET, *xspi->txbuf++); cdns_spi_write(xspi, CDNS_SPI_TXD, *xspi->txbuf++); else cdns_spi_write(xspi, CDNS_SPI_TXD_OFFSET, 0); cdns_spi_write(xspi, CDNS_SPI_TXD, 0); xspi->tx_bytes--; trans_cnt++; Loading Loading @@ -344,19 +339,18 @@ static irqreturn_t cdns_spi_irq(int irq, void *dev_id) u32 intr_status, status; status = IRQ_NONE; intr_status = cdns_spi_read(xspi, CDNS_SPI_ISR_OFFSET); cdns_spi_write(xspi, CDNS_SPI_ISR_OFFSET, intr_status); intr_status = cdns_spi_read(xspi, CDNS_SPI_ISR); cdns_spi_write(xspi, CDNS_SPI_ISR, intr_status); if (intr_status & CDNS_SPI_IXR_MODF_MASK) { if (intr_status & CDNS_SPI_IXR_MODF) { /* Indicate that transfer is completed, the SPI subsystem will * identify the error as the remaining bytes to be * transferred is non-zero */ cdns_spi_write(xspi, CDNS_SPI_IDR_OFFSET, CDNS_SPI_IXR_DEFAULT_MASK); cdns_spi_write(xspi, CDNS_SPI_IDR, CDNS_SPI_IXR_DEFAULT); spi_finalize_current_transfer(master); status = IRQ_HANDLED; } else if (intr_status & CDNS_SPI_IXR_TXOW_MASK) { } else if (intr_status & CDNS_SPI_IXR_TXOW) { unsigned long trans_cnt; trans_cnt = xspi->rx_bytes - xspi->tx_bytes; Loading @@ -365,7 +359,7 @@ static irqreturn_t cdns_spi_irq(int irq, void *dev_id) while (trans_cnt) { u8 data; data = cdns_spi_read(xspi, CDNS_SPI_RXD_OFFSET); data = cdns_spi_read(xspi, CDNS_SPI_RXD); if (xspi->rxbuf) *xspi->rxbuf++ = data; Loading @@ -378,8 +372,8 @@ static irqreturn_t cdns_spi_irq(int irq, void *dev_id) cdns_spi_fill_tx_fifo(xspi); } else { /* Transfer is completed */ cdns_spi_write(xspi, CDNS_SPI_IDR_OFFSET, CDNS_SPI_IXR_DEFAULT_MASK); cdns_spi_write(xspi, CDNS_SPI_IDR, CDNS_SPI_IXR_DEFAULT); spi_finalize_current_transfer(master); } status = IRQ_HANDLED; Loading @@ -387,6 +381,7 @@ static irqreturn_t cdns_spi_irq(int irq, void *dev_id) return status; } static int cdns_prepare_message(struct spi_master *master, struct spi_message *msg) { Loading Loading @@ -421,8 +416,7 @@ static int cdns_transfer_one(struct spi_master *master, cdns_spi_fill_tx_fifo(xspi); cdns_spi_write(xspi, CDNS_SPI_IER_OFFSET, CDNS_SPI_IXR_DEFAULT_MASK); cdns_spi_write(xspi, CDNS_SPI_IER, CDNS_SPI_IXR_DEFAULT); return transfer->len; } Loading @@ -439,8 +433,7 @@ static int cdns_prepare_transfer_hardware(struct spi_master *master) { struct cdns_spi *xspi = spi_master_get_devdata(master); cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_ENABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_ENABLE); return 0; } Loading @@ -458,8 +451,7 @@ static int cdns_unprepare_transfer_hardware(struct spi_master *master) { struct cdns_spi *xspi = spi_master_get_devdata(master); cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_DISABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_DISABLE); return 0; } Loading @@ -481,7 +473,7 @@ static int cdns_spi_probe(struct platform_device *pdev) u32 num_cs; master = spi_alloc_master(&pdev->dev, sizeof(*xspi)); if (master == NULL) if (!master) return -ENOMEM; xspi = spi_master_get_devdata(master); Loading Loading @@ -521,6 +513,11 @@ static int cdns_spi_probe(struct platform_device *pdev) goto clk_dis_apb; } pm_runtime_enable(&pdev->dev); pm_runtime_use_autosuspend(&pdev->dev); pm_runtime_set_autosuspend_delay(&pdev->dev, SPI_AUTOSUSPEND_TIMEOUT); pm_runtime_set_active(&pdev->dev); ret = of_property_read_u32(pdev->dev.of_node, "num-cs", &num_cs); if (ret < 0) master->num_chipselect = CDNS_SPI_DEFAULT_NUM_CS; Loading @@ -535,11 +532,14 @@ static int cdns_spi_probe(struct platform_device *pdev) /* SPI controller initializations */ cdns_spi_init_hw(xspi); pm_runtime_mark_last_busy(&pdev->dev); pm_runtime_put_autosuspend(&pdev->dev); irq = platform_get_irq(pdev, 0); if (irq <= 0) { ret = -ENXIO; dev_err(&pdev->dev, "irq number is invalid\n"); goto remove_master; goto clk_dis_all; } ret = devm_request_irq(&pdev->dev, irq, cdns_spi_irq, Loading @@ -547,7 +547,7 @@ static int cdns_spi_probe(struct platform_device *pdev) if (ret != 0) { ret = -ENXIO; dev_err(&pdev->dev, "request_irq failed\n"); goto remove_master; goto clk_dis_all; } master->prepare_transfer_hardware = cdns_prepare_transfer_hardware; Loading @@ -555,6 +555,7 @@ static int cdns_spi_probe(struct platform_device *pdev) master->transfer_one = cdns_transfer_one; master->unprepare_transfer_hardware = cdns_unprepare_transfer_hardware; master->set_cs = cdns_spi_chipselect; master->auto_runtime_pm = true; master->mode_bits = SPI_CPOL | SPI_CPHA; /* Set to default valid value */ Loading @@ -572,6 +573,8 @@ static int cdns_spi_probe(struct platform_device *pdev) return ret; clk_dis_all: pm_runtime_set_suspended(&pdev->dev); pm_runtime_disable(&pdev->dev); clk_disable_unprepare(xspi->ref_clk); clk_dis_apb: clk_disable_unprepare(xspi->pclk); Loading @@ -595,11 +598,12 @@ static int cdns_spi_remove(struct platform_device *pdev) struct spi_master *master = platform_get_drvdata(pdev); struct cdns_spi *xspi = spi_master_get_devdata(master); cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_DISABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_DISABLE); clk_disable_unprepare(xspi->ref_clk); clk_disable_unprepare(xspi->pclk); pm_runtime_set_suspended(&pdev->dev); pm_runtime_disable(&pdev->dev); spi_unregister_master(master); Loading @@ -613,21 +617,14 @@ static int cdns_spi_remove(struct platform_device *pdev) * This function disables the SPI controller and * changes the driver state to "suspend" * * Return: Always 0 * Return: 0 on success and error value on error */ static int __maybe_unused cdns_spi_suspend(struct device *dev) { struct platform_device *pdev = to_platform_device(dev); struct spi_master *master = platform_get_drvdata(pdev); struct cdns_spi *xspi = spi_master_get_devdata(master); spi_master_suspend(master); clk_disable_unprepare(xspi->ref_clk); clk_disable_unprepare(xspi->pclk); return 0; return spi_master_suspend(master); } /** Loading @@ -642,8 +639,23 @@ static int __maybe_unused cdns_spi_resume(struct device *dev) { struct platform_device *pdev = to_platform_device(dev); struct spi_master *master = platform_get_drvdata(pdev); return spi_master_resume(master); } /** * cdns_spi_runtime_resume - Runtime resume method for the SPI driver * @dev: Address of the platform_device structure * * This function enables the clocks * * Return: 0 on success and error value on error */ static int __maybe_unused cnds_runtime_resume(struct device *dev) { struct spi_master *master = dev_get_drvdata(dev); struct cdns_spi *xspi = spi_master_get_devdata(master); int ret = 0; int ret; ret = clk_prepare_enable(xspi->pclk); if (ret) { Loading @@ -657,13 +669,33 @@ static int __maybe_unused cdns_spi_resume(struct device *dev) clk_disable(xspi->pclk); return ret; } spi_master_resume(master); return 0; } /** * cdns_spi_runtime_suspend - Runtime suspend method for the SPI driver * @dev: Address of the platform_device structure * * This function disables the clocks * * Return: Always 0 */ static int __maybe_unused cnds_runtime_suspend(struct device *dev) { struct spi_master *master = dev_get_drvdata(dev); struct cdns_spi *xspi = spi_master_get_devdata(master); clk_disable_unprepare(xspi->ref_clk); clk_disable_unprepare(xspi->pclk); return 0; } static SIMPLE_DEV_PM_OPS(cdns_spi_dev_pm_ops, cdns_spi_suspend, cdns_spi_resume); static const struct dev_pm_ops cdns_spi_dev_pm_ops = { SET_RUNTIME_PM_OPS(cnds_runtime_suspend, cnds_runtime_resume, NULL) SET_SYSTEM_SLEEP_PM_OPS(cdns_spi_suspend, cdns_spi_resume) }; static const struct of_device_id cdns_spi_of_match[] = { { .compatible = "xlnx,zynq-spi-r1p6" }, Loading drivers/spi/spi-davinci.c +24 −52 Original line number Diff line number Diff line Loading @@ -23,7 +23,6 @@ #include <linux/clk.h> #include <linux/dmaengine.h> #include <linux/dma-mapping.h> #include <linux/edma.h> #include <linux/of.h> #include <linux/of_device.h> #include <linux/of_gpio.h> Loading @@ -33,8 +32,6 @@ #include <linux/platform_data/spi-davinci.h> #define SPI_NO_RESOURCE ((resource_size_t)-1) #define CS_DEFAULT 0xFF #define SPIFMT_PHASE_MASK BIT(16) Loading Loading @@ -130,8 +127,6 @@ struct davinci_spi { struct dma_chan *dma_rx; struct dma_chan *dma_tx; int dma_rx_chnum; int dma_tx_chnum; struct davinci_spi_platform_data pdata; Loading Loading @@ -797,35 +792,19 @@ static irqreturn_t davinci_spi_irq(s32 irq, void *data) static int davinci_spi_request_dma(struct davinci_spi *dspi) { dma_cap_mask_t mask; struct device *sdev = dspi->bitbang.master->dev.parent; int r; dma_cap_zero(mask); dma_cap_set(DMA_SLAVE, mask); dspi->dma_rx = dma_request_channel(mask, edma_filter_fn, &dspi->dma_rx_chnum); if (!dspi->dma_rx) { dev_err(sdev, "request RX DMA channel failed\n"); r = -ENODEV; goto rx_dma_failed; } dspi->dma_rx = dma_request_chan(sdev, "rx"); if (IS_ERR(dspi->dma_rx)) return PTR_ERR(dspi->dma_rx); dspi->dma_tx = dma_request_channel(mask, edma_filter_fn, &dspi->dma_tx_chnum); if (!dspi->dma_tx) { dev_err(sdev, "request TX DMA channel failed\n"); r = -ENODEV; goto tx_dma_failed; dspi->dma_tx = dma_request_chan(sdev, "tx"); if (IS_ERR(dspi->dma_tx)) { dma_release_channel(dspi->dma_rx); return PTR_ERR(dspi->dma_tx); } return 0; tx_dma_failed: dma_release_channel(dspi->dma_rx); rx_dma_failed: return r; } #if defined(CONFIG_OF) Loading Loading @@ -936,8 +915,6 @@ static int davinci_spi_probe(struct platform_device *pdev) struct davinci_spi *dspi; struct davinci_spi_platform_data *pdata; struct resource *r; resource_size_t dma_rx_chan = SPI_NO_RESOURCE; resource_size_t dma_tx_chan = SPI_NO_RESOURCE; int ret = 0; u32 spipc0; Loading Loading @@ -1044,27 +1021,15 @@ static int davinci_spi_probe(struct platform_device *pdev) } } r = platform_get_resource(pdev, IORESOURCE_DMA, 0); if (r) dma_rx_chan = r->start; r = platform_get_resource(pdev, IORESOURCE_DMA, 1); if (r) dma_tx_chan = r->start; dspi->bitbang.txrx_bufs = davinci_spi_bufs; if (dma_rx_chan != SPI_NO_RESOURCE && dma_tx_chan != SPI_NO_RESOURCE) { dspi->dma_rx_chnum = dma_rx_chan; dspi->dma_tx_chnum = dma_tx_chan; ret = davinci_spi_request_dma(dspi); if (ret) if (ret == -EPROBE_DEFER) { goto free_clk; dev_info(&pdev->dev, "DMA: supported\n"); dev_info(&pdev->dev, "DMA: RX channel: %pa, TX channel: %pa, event queue: %d\n", &dma_rx_chan, &dma_tx_chan, pdata->dma_event_q); } else if (ret) { dev_info(&pdev->dev, "DMA is not supported (%d)\n", ret); dspi->dma_rx = NULL; dspi->dma_tx = NULL; } dspi->get_rx = davinci_spi_rx_buf_u8; Loading Loading @@ -1102,8 +1067,10 @@ static int davinci_spi_probe(struct platform_device *pdev) return ret; free_dma: if (dspi->dma_rx) { dma_release_channel(dspi->dma_rx); dma_release_channel(dspi->dma_tx); } free_clk: clk_disable_unprepare(dspi->clk); free_master: Loading Loading @@ -1134,6 +1101,11 @@ static int davinci_spi_remove(struct platform_device *pdev) clk_disable_unprepare(dspi->clk); spi_master_put(master); if (dspi->dma_rx) { dma_release_channel(dspi->dma_rx); dma_release_channel(dspi->dma_tx); } return 0; } Loading drivers/spi/spi-dln2.c +2 −0 Original line number Diff line number Diff line Loading @@ -683,6 +683,7 @@ static int dln2_spi_probe(struct platform_device *pdev) struct spi_master *master; struct dln2_spi *dln2; struct dln2_platform_data *pdata = dev_get_platdata(&pdev->dev); struct device *dev = &pdev->dev; int ret; master = spi_alloc_master(&pdev->dev, sizeof(*dln2)); Loading @@ -700,6 +701,7 @@ static int dln2_spi_probe(struct platform_device *pdev) } dln2->master = master; dln2->master->dev.of_node = dev->of_node; dln2->pdev = pdev; dln2->port = pdata->port; /* cs/mode can never be 0xff, so the first transfer will set them */ Loading Loading
drivers/spi/spi-axi-spi-engine.c +0 −1 Original line number Diff line number Diff line Loading @@ -525,7 +525,6 @@ static int spi_engine_probe(struct platform_device *pdev) if (ret) goto err_ref_clk_disable; master->dev.parent = &pdev->dev; master->dev.of_node = pdev->dev.of_node; master->mode_bits = SPI_CPOL | SPI_CPHA | SPI_3WIRE; master->bits_per_word_mask = SPI_BPW_MASK(8); Loading
drivers/spi/spi-bcm53xx.c +76 −2 Original line number Diff line number Diff line Loading @@ -10,6 +10,7 @@ #include "spi-bcm53xx.h" #define BCM53XXSPI_MAX_SPI_BAUD 13500000 /* 216 MHz? */ #define BCM53XXSPI_FLASH_WINDOW SZ_32M /* The longest observed required wait was 19 ms */ #define BCM53XXSPI_SPE_TIMEOUT_MS 80 Loading @@ -17,8 +18,10 @@ struct bcm53xxspi { struct bcma_device *core; struct spi_master *master; void __iomem *mmio_base; size_t read_offset; bool bspi; /* Boot SPI mode with memory mapping */ }; static inline u32 bcm53xxspi_read(struct bcm53xxspi *b53spi, u16 offset) Loading @@ -32,6 +35,50 @@ static inline void bcm53xxspi_write(struct bcm53xxspi *b53spi, u16 offset, bcma_write32(b53spi->core, offset, value); } static void bcm53xxspi_disable_bspi(struct bcm53xxspi *b53spi) { struct device *dev = &b53spi->core->dev; unsigned long deadline; u32 tmp; if (!b53spi->bspi) return; tmp = bcm53xxspi_read(b53spi, B53SPI_BSPI_MAST_N_BOOT_CTRL); if (tmp & 0x1) return; deadline = jiffies + usecs_to_jiffies(200); do { tmp = bcm53xxspi_read(b53spi, B53SPI_BSPI_BUSY_STATUS); if (!(tmp & 0x1)) { bcm53xxspi_write(b53spi, B53SPI_BSPI_MAST_N_BOOT_CTRL, 0x1); ndelay(200); b53spi->bspi = false; return; } udelay(1); } while (!time_after_eq(jiffies, deadline)); dev_warn(dev, "Timeout disabling BSPI\n"); } static void bcm53xxspi_enable_bspi(struct bcm53xxspi *b53spi) { u32 tmp; if (b53spi->bspi) return; tmp = bcm53xxspi_read(b53spi, B53SPI_BSPI_MAST_N_BOOT_CTRL); if (!(tmp & 0x1)) return; bcm53xxspi_write(b53spi, B53SPI_BSPI_MAST_N_BOOT_CTRL, 0x0); b53spi->bspi = true; } static inline unsigned int bcm53xxspi_calc_timeout(size_t len) { /* Do some magic calculation based on length and buad. Add 10% and 1. */ Loading Loading @@ -176,6 +223,8 @@ static int bcm53xxspi_transfer_one(struct spi_master *master, u8 *buf; size_t left; bcm53xxspi_disable_bspi(b53spi); if (t->tx_buf) { buf = (u8 *)t->tx_buf; left = t->len; Loading Loading @@ -206,6 +255,22 @@ static int bcm53xxspi_transfer_one(struct spi_master *master, return 0; } static int bcm53xxspi_flash_read(struct spi_device *spi, struct spi_flash_read_message *msg) { struct bcm53xxspi *b53spi = spi_master_get_devdata(spi->master); int ret = 0; if (msg->from + msg->len > BCM53XXSPI_FLASH_WINDOW) return -EINVAL; bcm53xxspi_enable_bspi(b53spi); memcpy_fromio(msg->buf, b53spi->mmio_base + msg->from, msg->len); msg->retlen = msg->len; return ret; } /************************************************** * BCMA **************************************************/ Loading @@ -222,6 +287,7 @@ MODULE_DEVICE_TABLE(bcma, bcm53xxspi_bcma_tbl); static int bcm53xxspi_bcma_probe(struct bcma_device *core) { struct device *dev = &core->dev; struct bcm53xxspi *b53spi; struct spi_master *master; int err; Loading @@ -231,7 +297,7 @@ static int bcm53xxspi_bcma_probe(struct bcma_device *core) return -ENOTSUPP; } master = spi_alloc_master(&core->dev, sizeof(*b53spi)); master = spi_alloc_master(dev, sizeof(*b53spi)); if (!master) return -ENOMEM; Loading @@ -239,11 +305,19 @@ static int bcm53xxspi_bcma_probe(struct bcma_device *core) b53spi->master = master; b53spi->core = core; if (core->addr_s[0]) b53spi->mmio_base = devm_ioremap(dev, core->addr_s[0], BCM53XXSPI_FLASH_WINDOW); b53spi->bspi = true; bcm53xxspi_disable_bspi(b53spi); master->transfer_one = bcm53xxspi_transfer_one; if (b53spi->mmio_base) master->spi_flash_read = bcm53xxspi_flash_read; bcma_set_drvdata(core, b53spi); err = devm_spi_register_master(&core->dev, master); err = devm_spi_register_master(dev, master); if (err) { spi_master_put(master); bcma_set_drvdata(core, NULL); Loading
drivers/spi/spi-cadence.c +138 −106 Original line number Diff line number Diff line Loading @@ -19,44 +19,46 @@ #include <linux/of_irq.h> #include <linux/of_address.h> #include <linux/platform_device.h> #include <linux/pm_runtime.h> #include <linux/spi/spi.h> /* Name of this driver */ #define CDNS_SPI_NAME "cdns-spi" /* Register offset definitions */ #define CDNS_SPI_CR_OFFSET 0x00 /* Configuration Register, RW */ #define CDNS_SPI_ISR_OFFSET 0x04 /* Interrupt Status Register, RO */ #define CDNS_SPI_IER_OFFSET 0x08 /* Interrupt Enable Register, WO */ #define CDNS_SPI_IDR_OFFSET 0x0c /* Interrupt Disable Register, WO */ #define CDNS_SPI_IMR_OFFSET 0x10 /* Interrupt Enabled Mask Register, RO */ #define CDNS_SPI_ER_OFFSET 0x14 /* Enable/Disable Register, RW */ #define CDNS_SPI_DR_OFFSET 0x18 /* Delay Register, RW */ #define CDNS_SPI_TXD_OFFSET 0x1C /* Data Transmit Register, WO */ #define CDNS_SPI_RXD_OFFSET 0x20 /* Data Receive Register, RO */ #define CDNS_SPI_SICR_OFFSET 0x24 /* Slave Idle Count Register, RW */ #define CDNS_SPI_THLD_OFFSET 0x28 /* Transmit FIFO Watermark Register,RW */ #define CDNS_SPI_CR 0x00 /* Configuration Register, RW */ #define CDNS_SPI_ISR 0x04 /* Interrupt Status Register, RO */ #define CDNS_SPI_IER 0x08 /* Interrupt Enable Register, WO */ #define CDNS_SPI_IDR 0x0c /* Interrupt Disable Register, WO */ #define CDNS_SPI_IMR 0x10 /* Interrupt Enabled Mask Register, RO */ #define CDNS_SPI_ER 0x14 /* Enable/Disable Register, RW */ #define CDNS_SPI_DR 0x18 /* Delay Register, RW */ #define CDNS_SPI_TXD 0x1C /* Data Transmit Register, WO */ #define CDNS_SPI_RXD 0x20 /* Data Receive Register, RO */ #define CDNS_SPI_SICR 0x24 /* Slave Idle Count Register, RW */ #define CDNS_SPI_THLD 0x28 /* Transmit FIFO Watermark Register,RW */ #define SPI_AUTOSUSPEND_TIMEOUT 3000 /* * SPI Configuration Register bit Masks * * This register contains various control bits that affect the operation * of the SPI controller */ #define CDNS_SPI_CR_MANSTRT_MASK 0x00010000 /* Manual TX Start */ #define CDNS_SPI_CR_CPHA_MASK 0x00000004 /* Clock Phase Control */ #define CDNS_SPI_CR_CPOL_MASK 0x00000002 /* Clock Polarity Control */ #define CDNS_SPI_CR_SSCTRL_MASK 0x00003C00 /* Slave Select Mask */ #define CDNS_SPI_CR_PERI_SEL_MASK 0x00000200 /* Peripheral Select Decode */ #define CDNS_SPI_CR_BAUD_DIV_MASK 0x00000038 /* Baud Rate Divisor Mask */ #define CDNS_SPI_CR_MSTREN_MASK 0x00000001 /* Master Enable Mask */ #define CDNS_SPI_CR_MANSTRTEN_MASK 0x00008000 /* Manual TX Enable Mask */ #define CDNS_SPI_CR_SSFORCE_MASK 0x00004000 /* Manual SS Enable Mask */ #define CDNS_SPI_CR_BAUD_DIV_4_MASK 0x00000008 /* Default Baud Div Mask */ #define CDNS_SPI_CR_DEFAULT_MASK (CDNS_SPI_CR_MSTREN_MASK | \ CDNS_SPI_CR_SSCTRL_MASK | \ CDNS_SPI_CR_SSFORCE_MASK | \ CDNS_SPI_CR_BAUD_DIV_4_MASK) #define CDNS_SPI_CR_MANSTRT 0x00010000 /* Manual TX Start */ #define CDNS_SPI_CR_CPHA 0x00000004 /* Clock Phase Control */ #define CDNS_SPI_CR_CPOL 0x00000002 /* Clock Polarity Control */ #define CDNS_SPI_CR_SSCTRL 0x00003C00 /* Slave Select Mask */ #define CDNS_SPI_CR_PERI_SEL 0x00000200 /* Peripheral Select Decode */ #define CDNS_SPI_CR_BAUD_DIV 0x00000038 /* Baud Rate Divisor Mask */ #define CDNS_SPI_CR_MSTREN 0x00000001 /* Master Enable Mask */ #define CDNS_SPI_CR_MANSTRTEN 0x00008000 /* Manual TX Enable Mask */ #define CDNS_SPI_CR_SSFORCE 0x00004000 /* Manual SS Enable Mask */ #define CDNS_SPI_CR_BAUD_DIV_4 0x00000008 /* Default Baud Div Mask */ #define CDNS_SPI_CR_DEFAULT (CDNS_SPI_CR_MSTREN | \ CDNS_SPI_CR_SSCTRL | \ CDNS_SPI_CR_SSFORCE | \ CDNS_SPI_CR_BAUD_DIV_4) /* * SPI Configuration Register - Baud rate and slave select Loading @@ -77,21 +79,21 @@ * All the four interrupt registers (Status/Mask/Enable/Disable) have the same * bit definitions. */ #define CDNS_SPI_IXR_TXOW_MASK 0x00000004 /* SPI TX FIFO Overwater */ #define CDNS_SPI_IXR_MODF_MASK 0x00000002 /* SPI Mode Fault */ #define CDNS_SPI_IXR_RXNEMTY_MASK 0x00000010 /* SPI RX FIFO Not Empty */ #define CDNS_SPI_IXR_DEFAULT_MASK (CDNS_SPI_IXR_TXOW_MASK | \ CDNS_SPI_IXR_MODF_MASK) #define CDNS_SPI_IXR_TXFULL_MASK 0x00000008 /* SPI TX Full */ #define CDNS_SPI_IXR_ALL_MASK 0x0000007F /* SPI all interrupts */ #define CDNS_SPI_IXR_TXOW 0x00000004 /* SPI TX FIFO Overwater */ #define CDNS_SPI_IXR_MODF 0x00000002 /* SPI Mode Fault */ #define CDNS_SPI_IXR_RXNEMTY 0x00000010 /* SPI RX FIFO Not Empty */ #define CDNS_SPI_IXR_DEFAULT (CDNS_SPI_IXR_TXOW | \ CDNS_SPI_IXR_MODF) #define CDNS_SPI_IXR_TXFULL 0x00000008 /* SPI TX Full */ #define CDNS_SPI_IXR_ALL 0x0000007F /* SPI all interrupts */ /* * SPI Enable Register bit Masks * * This register is used to enable or disable the SPI controller */ #define CDNS_SPI_ER_ENABLE_MASK 0x00000001 /* SPI Enable Bit Mask */ #define CDNS_SPI_ER_DISABLE_MASK 0x0 /* SPI Disable Bit Mask */ #define CDNS_SPI_ER_ENABLE 0x00000001 /* SPI Enable Bit Mask */ #define CDNS_SPI_ER_DISABLE 0x0 /* SPI Disable Bit Mask */ /* SPI FIFO depth in bytes */ #define CDNS_SPI_FIFO_DEPTH 128 Loading Loading @@ -149,56 +151,51 @@ static inline void cdns_spi_write(struct cdns_spi *xspi, u32 offset, u32 val) */ static void cdns_spi_init_hw(struct cdns_spi *xspi) { u32 ctrl_reg = CDNS_SPI_CR_DEFAULT_MASK; u32 ctrl_reg = CDNS_SPI_CR_DEFAULT; if (xspi->is_decoded_cs) ctrl_reg |= CDNS_SPI_CR_PERI_SEL_MASK; ctrl_reg |= CDNS_SPI_CR_PERI_SEL; cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_DISABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_IDR_OFFSET, CDNS_SPI_IXR_ALL_MASK); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_DISABLE); cdns_spi_write(xspi, CDNS_SPI_IDR, CDNS_SPI_IXR_ALL); /* Clear the RX FIFO */ while (cdns_spi_read(xspi, CDNS_SPI_ISR_OFFSET) & CDNS_SPI_IXR_RXNEMTY_MASK) cdns_spi_read(xspi, CDNS_SPI_RXD_OFFSET); while (cdns_spi_read(xspi, CDNS_SPI_ISR) & CDNS_SPI_IXR_RXNEMTY) cdns_spi_read(xspi, CDNS_SPI_RXD); cdns_spi_write(xspi, CDNS_SPI_ISR_OFFSET, CDNS_SPI_IXR_ALL_MASK); cdns_spi_write(xspi, CDNS_SPI_CR_OFFSET, ctrl_reg); cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_ENABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_ISR, CDNS_SPI_IXR_ALL); cdns_spi_write(xspi, CDNS_SPI_CR, ctrl_reg); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_ENABLE); } /** * cdns_spi_chipselect - Select or deselect the chip select line * @spi: Pointer to the spi_device structure * @is_on: Select(0) or deselect (1) the chip select line * @is_high: Select(0) or deselect (1) the chip select line */ static void cdns_spi_chipselect(struct spi_device *spi, bool is_high) { struct cdns_spi *xspi = spi_master_get_devdata(spi->master); u32 ctrl_reg; ctrl_reg = cdns_spi_read(xspi, CDNS_SPI_CR_OFFSET); ctrl_reg = cdns_spi_read(xspi, CDNS_SPI_CR); if (is_high) { /* Deselect the slave */ ctrl_reg |= CDNS_SPI_CR_SSCTRL_MASK; ctrl_reg |= CDNS_SPI_CR_SSCTRL; } else { /* Select the slave */ ctrl_reg &= ~CDNS_SPI_CR_SSCTRL_MASK; ctrl_reg &= ~CDNS_SPI_CR_SSCTRL; if (!(xspi->is_decoded_cs)) ctrl_reg |= ((~(CDNS_SPI_SS0 << spi->chip_select)) << CDNS_SPI_SS_SHIFT) & CDNS_SPI_CR_SSCTRL_MASK; CDNS_SPI_CR_SSCTRL; else ctrl_reg |= (spi->chip_select << CDNS_SPI_SS_SHIFT) & CDNS_SPI_CR_SSCTRL_MASK; CDNS_SPI_CR_SSCTRL; } cdns_spi_write(xspi, CDNS_SPI_CR_OFFSET, ctrl_reg); cdns_spi_write(xspi, CDNS_SPI_CR, ctrl_reg); } /** Loading @@ -212,14 +209,15 @@ static void cdns_spi_config_clock_mode(struct spi_device *spi) struct cdns_spi *xspi = spi_master_get_devdata(spi->master); u32 ctrl_reg, new_ctrl_reg; new_ctrl_reg = ctrl_reg = cdns_spi_read(xspi, CDNS_SPI_CR_OFFSET); new_ctrl_reg = cdns_spi_read(xspi, CDNS_SPI_CR); ctrl_reg = new_ctrl_reg; /* Set the SPI clock phase and clock polarity */ new_ctrl_reg &= ~(CDNS_SPI_CR_CPHA_MASK | CDNS_SPI_CR_CPOL_MASK); new_ctrl_reg &= ~(CDNS_SPI_CR_CPHA | CDNS_SPI_CR_CPOL); if (spi->mode & SPI_CPHA) new_ctrl_reg |= CDNS_SPI_CR_CPHA_MASK; new_ctrl_reg |= CDNS_SPI_CR_CPHA; if (spi->mode & SPI_CPOL) new_ctrl_reg |= CDNS_SPI_CR_CPOL_MASK; new_ctrl_reg |= CDNS_SPI_CR_CPOL; if (new_ctrl_reg != ctrl_reg) { /* Loading @@ -228,11 +226,9 @@ static void cdns_spi_config_clock_mode(struct spi_device *spi) * polarity as it will cause the SPI slave to see spurious clock * transitions. To workaround the issue toggle the ER register. */ cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_DISABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_CR_OFFSET, new_ctrl_reg); cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_ENABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_DISABLE); cdns_spi_write(xspi, CDNS_SPI_CR, new_ctrl_reg); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_ENABLE); } } Loading @@ -259,7 +255,7 @@ static void cdns_spi_config_clock_freq(struct spi_device *spi, frequency = clk_get_rate(xspi->ref_clk); ctrl_reg = cdns_spi_read(xspi, CDNS_SPI_CR_OFFSET); ctrl_reg = cdns_spi_read(xspi, CDNS_SPI_CR); /* Set the clock frequency */ if (xspi->speed_hz != transfer->speed_hz) { Loading @@ -269,12 +265,12 @@ static void cdns_spi_config_clock_freq(struct spi_device *spi, (frequency / (2 << baud_rate_val)) > transfer->speed_hz) baud_rate_val++; ctrl_reg &= ~CDNS_SPI_CR_BAUD_DIV_MASK; ctrl_reg &= ~CDNS_SPI_CR_BAUD_DIV; ctrl_reg |= baud_rate_val << CDNS_SPI_BAUD_DIV_SHIFT; xspi->speed_hz = frequency / (2 << baud_rate_val); } cdns_spi_write(xspi, CDNS_SPI_CR_OFFSET, ctrl_reg); cdns_spi_write(xspi, CDNS_SPI_CR, ctrl_reg); } /** Loading Loading @@ -313,10 +309,9 @@ static void cdns_spi_fill_tx_fifo(struct cdns_spi *xspi) while ((trans_cnt < CDNS_SPI_FIFO_DEPTH) && (xspi->tx_bytes > 0)) { if (xspi->txbuf) cdns_spi_write(xspi, CDNS_SPI_TXD_OFFSET, *xspi->txbuf++); cdns_spi_write(xspi, CDNS_SPI_TXD, *xspi->txbuf++); else cdns_spi_write(xspi, CDNS_SPI_TXD_OFFSET, 0); cdns_spi_write(xspi, CDNS_SPI_TXD, 0); xspi->tx_bytes--; trans_cnt++; Loading Loading @@ -344,19 +339,18 @@ static irqreturn_t cdns_spi_irq(int irq, void *dev_id) u32 intr_status, status; status = IRQ_NONE; intr_status = cdns_spi_read(xspi, CDNS_SPI_ISR_OFFSET); cdns_spi_write(xspi, CDNS_SPI_ISR_OFFSET, intr_status); intr_status = cdns_spi_read(xspi, CDNS_SPI_ISR); cdns_spi_write(xspi, CDNS_SPI_ISR, intr_status); if (intr_status & CDNS_SPI_IXR_MODF_MASK) { if (intr_status & CDNS_SPI_IXR_MODF) { /* Indicate that transfer is completed, the SPI subsystem will * identify the error as the remaining bytes to be * transferred is non-zero */ cdns_spi_write(xspi, CDNS_SPI_IDR_OFFSET, CDNS_SPI_IXR_DEFAULT_MASK); cdns_spi_write(xspi, CDNS_SPI_IDR, CDNS_SPI_IXR_DEFAULT); spi_finalize_current_transfer(master); status = IRQ_HANDLED; } else if (intr_status & CDNS_SPI_IXR_TXOW_MASK) { } else if (intr_status & CDNS_SPI_IXR_TXOW) { unsigned long trans_cnt; trans_cnt = xspi->rx_bytes - xspi->tx_bytes; Loading @@ -365,7 +359,7 @@ static irqreturn_t cdns_spi_irq(int irq, void *dev_id) while (trans_cnt) { u8 data; data = cdns_spi_read(xspi, CDNS_SPI_RXD_OFFSET); data = cdns_spi_read(xspi, CDNS_SPI_RXD); if (xspi->rxbuf) *xspi->rxbuf++ = data; Loading @@ -378,8 +372,8 @@ static irqreturn_t cdns_spi_irq(int irq, void *dev_id) cdns_spi_fill_tx_fifo(xspi); } else { /* Transfer is completed */ cdns_spi_write(xspi, CDNS_SPI_IDR_OFFSET, CDNS_SPI_IXR_DEFAULT_MASK); cdns_spi_write(xspi, CDNS_SPI_IDR, CDNS_SPI_IXR_DEFAULT); spi_finalize_current_transfer(master); } status = IRQ_HANDLED; Loading @@ -387,6 +381,7 @@ static irqreturn_t cdns_spi_irq(int irq, void *dev_id) return status; } static int cdns_prepare_message(struct spi_master *master, struct spi_message *msg) { Loading Loading @@ -421,8 +416,7 @@ static int cdns_transfer_one(struct spi_master *master, cdns_spi_fill_tx_fifo(xspi); cdns_spi_write(xspi, CDNS_SPI_IER_OFFSET, CDNS_SPI_IXR_DEFAULT_MASK); cdns_spi_write(xspi, CDNS_SPI_IER, CDNS_SPI_IXR_DEFAULT); return transfer->len; } Loading @@ -439,8 +433,7 @@ static int cdns_prepare_transfer_hardware(struct spi_master *master) { struct cdns_spi *xspi = spi_master_get_devdata(master); cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_ENABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_ENABLE); return 0; } Loading @@ -458,8 +451,7 @@ static int cdns_unprepare_transfer_hardware(struct spi_master *master) { struct cdns_spi *xspi = spi_master_get_devdata(master); cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_DISABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_DISABLE); return 0; } Loading @@ -481,7 +473,7 @@ static int cdns_spi_probe(struct platform_device *pdev) u32 num_cs; master = spi_alloc_master(&pdev->dev, sizeof(*xspi)); if (master == NULL) if (!master) return -ENOMEM; xspi = spi_master_get_devdata(master); Loading Loading @@ -521,6 +513,11 @@ static int cdns_spi_probe(struct platform_device *pdev) goto clk_dis_apb; } pm_runtime_enable(&pdev->dev); pm_runtime_use_autosuspend(&pdev->dev); pm_runtime_set_autosuspend_delay(&pdev->dev, SPI_AUTOSUSPEND_TIMEOUT); pm_runtime_set_active(&pdev->dev); ret = of_property_read_u32(pdev->dev.of_node, "num-cs", &num_cs); if (ret < 0) master->num_chipselect = CDNS_SPI_DEFAULT_NUM_CS; Loading @@ -535,11 +532,14 @@ static int cdns_spi_probe(struct platform_device *pdev) /* SPI controller initializations */ cdns_spi_init_hw(xspi); pm_runtime_mark_last_busy(&pdev->dev); pm_runtime_put_autosuspend(&pdev->dev); irq = platform_get_irq(pdev, 0); if (irq <= 0) { ret = -ENXIO; dev_err(&pdev->dev, "irq number is invalid\n"); goto remove_master; goto clk_dis_all; } ret = devm_request_irq(&pdev->dev, irq, cdns_spi_irq, Loading @@ -547,7 +547,7 @@ static int cdns_spi_probe(struct platform_device *pdev) if (ret != 0) { ret = -ENXIO; dev_err(&pdev->dev, "request_irq failed\n"); goto remove_master; goto clk_dis_all; } master->prepare_transfer_hardware = cdns_prepare_transfer_hardware; Loading @@ -555,6 +555,7 @@ static int cdns_spi_probe(struct platform_device *pdev) master->transfer_one = cdns_transfer_one; master->unprepare_transfer_hardware = cdns_unprepare_transfer_hardware; master->set_cs = cdns_spi_chipselect; master->auto_runtime_pm = true; master->mode_bits = SPI_CPOL | SPI_CPHA; /* Set to default valid value */ Loading @@ -572,6 +573,8 @@ static int cdns_spi_probe(struct platform_device *pdev) return ret; clk_dis_all: pm_runtime_set_suspended(&pdev->dev); pm_runtime_disable(&pdev->dev); clk_disable_unprepare(xspi->ref_clk); clk_dis_apb: clk_disable_unprepare(xspi->pclk); Loading @@ -595,11 +598,12 @@ static int cdns_spi_remove(struct platform_device *pdev) struct spi_master *master = platform_get_drvdata(pdev); struct cdns_spi *xspi = spi_master_get_devdata(master); cdns_spi_write(xspi, CDNS_SPI_ER_OFFSET, CDNS_SPI_ER_DISABLE_MASK); cdns_spi_write(xspi, CDNS_SPI_ER, CDNS_SPI_ER_DISABLE); clk_disable_unprepare(xspi->ref_clk); clk_disable_unprepare(xspi->pclk); pm_runtime_set_suspended(&pdev->dev); pm_runtime_disable(&pdev->dev); spi_unregister_master(master); Loading @@ -613,21 +617,14 @@ static int cdns_spi_remove(struct platform_device *pdev) * This function disables the SPI controller and * changes the driver state to "suspend" * * Return: Always 0 * Return: 0 on success and error value on error */ static int __maybe_unused cdns_spi_suspend(struct device *dev) { struct platform_device *pdev = to_platform_device(dev); struct spi_master *master = platform_get_drvdata(pdev); struct cdns_spi *xspi = spi_master_get_devdata(master); spi_master_suspend(master); clk_disable_unprepare(xspi->ref_clk); clk_disable_unprepare(xspi->pclk); return 0; return spi_master_suspend(master); } /** Loading @@ -642,8 +639,23 @@ static int __maybe_unused cdns_spi_resume(struct device *dev) { struct platform_device *pdev = to_platform_device(dev); struct spi_master *master = platform_get_drvdata(pdev); return spi_master_resume(master); } /** * cdns_spi_runtime_resume - Runtime resume method for the SPI driver * @dev: Address of the platform_device structure * * This function enables the clocks * * Return: 0 on success and error value on error */ static int __maybe_unused cnds_runtime_resume(struct device *dev) { struct spi_master *master = dev_get_drvdata(dev); struct cdns_spi *xspi = spi_master_get_devdata(master); int ret = 0; int ret; ret = clk_prepare_enable(xspi->pclk); if (ret) { Loading @@ -657,13 +669,33 @@ static int __maybe_unused cdns_spi_resume(struct device *dev) clk_disable(xspi->pclk); return ret; } spi_master_resume(master); return 0; } /** * cdns_spi_runtime_suspend - Runtime suspend method for the SPI driver * @dev: Address of the platform_device structure * * This function disables the clocks * * Return: Always 0 */ static int __maybe_unused cnds_runtime_suspend(struct device *dev) { struct spi_master *master = dev_get_drvdata(dev); struct cdns_spi *xspi = spi_master_get_devdata(master); clk_disable_unprepare(xspi->ref_clk); clk_disable_unprepare(xspi->pclk); return 0; } static SIMPLE_DEV_PM_OPS(cdns_spi_dev_pm_ops, cdns_spi_suspend, cdns_spi_resume); static const struct dev_pm_ops cdns_spi_dev_pm_ops = { SET_RUNTIME_PM_OPS(cnds_runtime_suspend, cnds_runtime_resume, NULL) SET_SYSTEM_SLEEP_PM_OPS(cdns_spi_suspend, cdns_spi_resume) }; static const struct of_device_id cdns_spi_of_match[] = { { .compatible = "xlnx,zynq-spi-r1p6" }, Loading
drivers/spi/spi-davinci.c +24 −52 Original line number Diff line number Diff line Loading @@ -23,7 +23,6 @@ #include <linux/clk.h> #include <linux/dmaengine.h> #include <linux/dma-mapping.h> #include <linux/edma.h> #include <linux/of.h> #include <linux/of_device.h> #include <linux/of_gpio.h> Loading @@ -33,8 +32,6 @@ #include <linux/platform_data/spi-davinci.h> #define SPI_NO_RESOURCE ((resource_size_t)-1) #define CS_DEFAULT 0xFF #define SPIFMT_PHASE_MASK BIT(16) Loading Loading @@ -130,8 +127,6 @@ struct davinci_spi { struct dma_chan *dma_rx; struct dma_chan *dma_tx; int dma_rx_chnum; int dma_tx_chnum; struct davinci_spi_platform_data pdata; Loading Loading @@ -797,35 +792,19 @@ static irqreturn_t davinci_spi_irq(s32 irq, void *data) static int davinci_spi_request_dma(struct davinci_spi *dspi) { dma_cap_mask_t mask; struct device *sdev = dspi->bitbang.master->dev.parent; int r; dma_cap_zero(mask); dma_cap_set(DMA_SLAVE, mask); dspi->dma_rx = dma_request_channel(mask, edma_filter_fn, &dspi->dma_rx_chnum); if (!dspi->dma_rx) { dev_err(sdev, "request RX DMA channel failed\n"); r = -ENODEV; goto rx_dma_failed; } dspi->dma_rx = dma_request_chan(sdev, "rx"); if (IS_ERR(dspi->dma_rx)) return PTR_ERR(dspi->dma_rx); dspi->dma_tx = dma_request_channel(mask, edma_filter_fn, &dspi->dma_tx_chnum); if (!dspi->dma_tx) { dev_err(sdev, "request TX DMA channel failed\n"); r = -ENODEV; goto tx_dma_failed; dspi->dma_tx = dma_request_chan(sdev, "tx"); if (IS_ERR(dspi->dma_tx)) { dma_release_channel(dspi->dma_rx); return PTR_ERR(dspi->dma_tx); } return 0; tx_dma_failed: dma_release_channel(dspi->dma_rx); rx_dma_failed: return r; } #if defined(CONFIG_OF) Loading Loading @@ -936,8 +915,6 @@ static int davinci_spi_probe(struct platform_device *pdev) struct davinci_spi *dspi; struct davinci_spi_platform_data *pdata; struct resource *r; resource_size_t dma_rx_chan = SPI_NO_RESOURCE; resource_size_t dma_tx_chan = SPI_NO_RESOURCE; int ret = 0; u32 spipc0; Loading Loading @@ -1044,27 +1021,15 @@ static int davinci_spi_probe(struct platform_device *pdev) } } r = platform_get_resource(pdev, IORESOURCE_DMA, 0); if (r) dma_rx_chan = r->start; r = platform_get_resource(pdev, IORESOURCE_DMA, 1); if (r) dma_tx_chan = r->start; dspi->bitbang.txrx_bufs = davinci_spi_bufs; if (dma_rx_chan != SPI_NO_RESOURCE && dma_tx_chan != SPI_NO_RESOURCE) { dspi->dma_rx_chnum = dma_rx_chan; dspi->dma_tx_chnum = dma_tx_chan; ret = davinci_spi_request_dma(dspi); if (ret) if (ret == -EPROBE_DEFER) { goto free_clk; dev_info(&pdev->dev, "DMA: supported\n"); dev_info(&pdev->dev, "DMA: RX channel: %pa, TX channel: %pa, event queue: %d\n", &dma_rx_chan, &dma_tx_chan, pdata->dma_event_q); } else if (ret) { dev_info(&pdev->dev, "DMA is not supported (%d)\n", ret); dspi->dma_rx = NULL; dspi->dma_tx = NULL; } dspi->get_rx = davinci_spi_rx_buf_u8; Loading Loading @@ -1102,8 +1067,10 @@ static int davinci_spi_probe(struct platform_device *pdev) return ret; free_dma: if (dspi->dma_rx) { dma_release_channel(dspi->dma_rx); dma_release_channel(dspi->dma_tx); } free_clk: clk_disable_unprepare(dspi->clk); free_master: Loading Loading @@ -1134,6 +1101,11 @@ static int davinci_spi_remove(struct platform_device *pdev) clk_disable_unprepare(dspi->clk); spi_master_put(master); if (dspi->dma_rx) { dma_release_channel(dspi->dma_rx); dma_release_channel(dspi->dma_tx); } return 0; } Loading
drivers/spi/spi-dln2.c +2 −0 Original line number Diff line number Diff line Loading @@ -683,6 +683,7 @@ static int dln2_spi_probe(struct platform_device *pdev) struct spi_master *master; struct dln2_spi *dln2; struct dln2_platform_data *pdata = dev_get_platdata(&pdev->dev); struct device *dev = &pdev->dev; int ret; master = spi_alloc_master(&pdev->dev, sizeof(*dln2)); Loading @@ -700,6 +701,7 @@ static int dln2_spi_probe(struct platform_device *pdev) } dln2->master = master; dln2->master->dev.of_node = dev->of_node; dln2->pdev = pdev; dln2->port = pdata->port; /* cs/mode can never be 0xff, so the first transfer will set them */ Loading