@@ -14,6 +14,8 @@
#include <linux/clk.h>
#include <linux/delay.h>
#include <linux/device.h>
+#include <linux/dmaengine.h>
+#include <linux/dma-mapping.h>
#include <linux/interrupt.h>
#include <linux/io.h>
#include <linux/module.h>
@@ -50,6 +52,12 @@
#define SUNXI_FIFO_STA_TF_CNT_MASK 0x7f
#define SUNXI_FIFO_STA_TF_CNT_BITS 16
+static int wait_for_dma = 1;
+module_param(wait_for_dma, int, 0644);
+MODULE_PARM_DESC(wait_for_dma,
+ "When acquiring a DMA channel returns EDEFER return and let kernel defer spi master probe.\n"
+ "Non-DMA operation is used otherwise (defaults to wait for DMA driver to load).");
+
enum SPI_SUNXI_TYPE {
SPI_SUN4I = 1,
SPI_SUN6I,
@@ -61,6 +69,7 @@ enum SUNXI_REG_ENUM {
SUNXI_TFR_CTL_REG,
SUNXI_INT_CTL_REG,
SUNXI_INT_STA_REG,
+ SUNXI_DMA_CTL_REG,
SUNXI_WAIT_REG,
SUNXI_CLK_CTL_REG,
SUNXI_BURST_CNT_REG,
@@ -79,6 +88,7 @@ static int sun4i_regmap[SUNXI_NUM_REGS] = {
/* SUNXI_TFR_CTL_REG */ 0x08,
/* SUNXI_INT_CTL_REG */ 0x0c,
/* SUNXI_INT_STA_REG */ 0x10,
+/* SUNXI_DMA_CTL_REG */ 0x14,
/* SUNXI_WAIT_REG */ 0x18,
/* SUNXI_CLK_CTL_REG */ 0x1c,
/* SUNXI_BURST_CNT_REG */ 0x20,
@@ -93,6 +103,7 @@ static int sun6i_regmap[SUNXI_NUM_REGS] = {
/* SUNXI_TFR_CTL_REG */ 0x08,
/* SUNXI_INT_CTL_REG */ 0x10,
/* SUNXI_INT_STA_REG */ 0x14,
+/* SUNXI_DMA_CTL_REG */ -1,
/* SUNXI_WAIT_REG */ 0x20,
/* SUNXI_CLK_CTL_REG */ 0x24,
/* SUNXI_BURST_CNT_REG */ 0x30,
@@ -110,6 +121,7 @@ enum SUNXI_BITMAP_ENUM {
SUNXI_TFR_CTL_CPHA,
SUNXI_TFR_CTL_CPOL,
SUNXI_TFR_CTL_CS_ACTIVE_LOW,
+ SUNXI_CTL_DMA_DEDICATED,
SUNXI_TFR_CTL_FBS,
SUNXI_CTL_TF_RST,
SUNXI_CTL_RF_RST,
@@ -121,6 +133,9 @@ enum SUNXI_BITMAP_ENUM {
SUNXI_TFR_CTL_CS_LEVEL,
SUNXI_CTL_TP,
SUNXI_INT_CTL_TC,
+ SUNXI_CTL_DMA_RF_READY,
+ SUNXI_CTL_DMA_TF_NOT_FULL,
+ SUNXI_CTL_DMA_TF_HALF,
SUNXI_BITMAP_SIZE
};
@@ -130,6 +145,7 @@ static int sun4i_bitmap[SUNXI_BITMAP_SIZE] = {
/* SUNXI_TFR_CTL_CPHA */ BIT(2),
/* SUNXI_TFR_CTL_CPOL */ BIT(3),
/* SUNXI_TFR_CTL_CS_ACTIVE_LOW */ BIT(4),
+/* SUNXI_CTL_DMA_DEDICATED */ BIT(5),
/* SUNXI_TFR_CTL_FBS */ BIT(6),
/* SUNXI_CTL_TF_RST */ BIT(8),
/* SUNXI_CTL_RF_RST */ BIT(9),
@@ -141,6 +157,9 @@ static int sun4i_bitmap[SUNXI_BITMAP_SIZE] = {
/* SUNXI_TFR_CTL_CS_LEVEL */ BIT(17),
/* SUNXI_CTL_TP */ BIT(18),
/* SUNXI_INT_CTL_TC */ BIT(16),
+/* SUNXI_CTL_DMA_RF_READY */ BIT(0),
+/* SUNXI_CTL_DMA_TF_NOT_FULL */ BIT(10),
+/* SUNXI_CTL_DMA_TF_HALF */ BIT(9),
};
static int sun6i_bitmap[SUNXI_BITMAP_SIZE] = {
@@ -149,6 +168,12 @@ static int sun6i_bitmap[SUNXI_BITMAP_SIZE] = {
/* SUNXI_TFR_CTL_CPHA */ BIT(0),
/* SUNXI_TFR_CTL_CPOL */ BIT(1),
/* SUNXI_TFR_CTL_CS_ACTIVE_LOW */ BIT(2),
+/*
+ * Bit 9 is listed as dedicated dma control for rx.
+ * There is no dedicated dma control bit listed for tx and bit 25
+ * on the logical position is listed as unused.
+ */
+/* SUNXI_CTL_DMA_DEDICATED */ BIT(9)|BIT(25),
/* SUNXI_TFR_CTL_FBS */ BIT(12),
/* SUNXI_CTL_TF_RST */ BIT(31),
/* SUNXI_CTL_RF_RST */ BIT(15),
@@ -160,6 +185,15 @@ static int sun6i_bitmap[SUNXI_BITMAP_SIZE] = {
/* SUNXI_TFR_CTL_CS_LEVEL */ BIT(7),
/* SUNXI_CTL_TP */ BIT(7),
/* SUNXI_INT_CTL_TC */ BIT(12),
+/*
+ * On sun4i there are separate bits enabling request on different fifo levels.
+ * On sun6i there is a level field and enable bit which enables request on that
+ * FIFO level. Only one level is ever used so just pack the relevant bits as
+ * one constant.
+ */
+/* SUNXI_CTL_DMA_RF_READY */ BIT(0)|BIT(8),
+/* SUNXI_CTL_DMA_TF_NOT_FULL */ (0x7f << 16)|BIT(24),
+/* SUNXI_CTL_DMA_TF_HALF */ BIT(23)|BIT(24),
};
struct sunxi_spi {
@@ -207,6 +241,20 @@ static inline u32 sspi_bits(struct sunxi_spi *sspi,
return (*sspi->bitmap)[name];
}
+static inline void sunxi_spi_set(struct sunxi_spi *sspi, u32 reg, u32 value)
+{
+ u32 orig = sunxi_spi_read(sspi, reg);
+
+ sunxi_spi_write(sspi, reg, orig | value);
+}
+
+static inline void sunxi_spi_unset(struct sunxi_spi *sspi, u32 reg, u32 value)
+{
+ u32 orig = sunxi_spi_read(sspi, reg);
+
+ sunxi_spi_write(sspi, reg, orig & ~value);
+}
+
static inline void sunxi_spi_drain_fifo(struct sunxi_spi *sspi, int len)
{
u32 reg, cnt;
@@ -243,6 +291,15 @@ static inline void sunxi_spi_fill_fifo(struct sunxi_spi *sspi, int len)
}
}
+static bool sunxi_spi_can_dma(struct spi_master *master,
+ struct spi_device *spi,
+ struct spi_transfer *tfr)
+{
+ struct sunxi_spi *sspi = spi_master_get_devdata(master);
+
+ return tfr->len >= sspi->fifo_depth;
+}
+
static void sunxi_spi_set_cs(struct spi_device *spi, bool enable)
{
struct sunxi_spi *sspi = spi_master_get_devdata(spi->master);
@@ -284,6 +341,8 @@ static size_t sunxi_spi_max_transfer_size(struct spi_device *spi)
struct spi_master *master = spi->master;
struct sunxi_spi *sspi = spi_master_get_devdata(master);
+ if (master->can_dma)
+ return SUNXI_CNT_MASK;
return sspi->fifo_depth - 1;
}
@@ -292,22 +351,27 @@ static int sunxi_spi_transfer_one(struct spi_master *master,
struct spi_transfer *tfr)
{
struct sunxi_spi *sspi = spi_master_get_devdata(master);
+ struct dma_async_tx_descriptor *desc_tx = NULL, *desc_rx = NULL;
unsigned int mclk_rate, div, timeout;
unsigned int start, end, tx_time;
unsigned int tx_len = 0;
int ret = 0;
- u32 reg;
+ u32 reg, trigger = 0;
+
+ if (!master->can_dma) {
+ /* We don't support transfer larger than the FIFO */
+ if (tfr->len > sspi->fifo_depth)
+ return -EMSGSIZE;
+ /*
+ * Filling the FIFO fully causes timeout for some reason
+ * at least on spi2 on A10s
+ */
+ if ((sspi->type == SPI_SUN4I) &&
+ tfr->tx_buf && tfr->len >= sspi->fifo_depth)
+ return -EMSGSIZE;
+ }
- /* We don't support transfer larger than the FIFO */
- if (tfr->len > sspi->fifo_depth)
- return -EMSGSIZE;
-
- /*
- * Filling the FIFO fully causes timeout for some reason
- * at least on spi2 on A10s
- */
- if ((sspi->type == SPI_SUN4I) &&
- tfr->tx_buf && tfr->len >= sspi->fifo_depth)
+ if (tfr->len > SUNXI_CNT_MASK)
return -EMSGSIZE;
reinit_completion(&sspi->done);
@@ -405,17 +469,81 @@ static int sunxi_spi_transfer_one(struct spi_master *master,
sunxi_spi_write(sspi, SUNXI_BURST_CTL_CNT_REG,
SUNXI_BURST_CTL_CNT_STC(tx_len));
- /* Fill the TX FIFO */
- sunxi_spi_fill_fifo(sspi, sspi->fifo_depth);
+ /* Setup transfer buffers */
+ if (sunxi_spi_can_dma(master, spi, tfr)) {
+ dev_dbg(&sspi->master->dev, "Using DMA mode for transfer\n");
+
+ if (sspi->tx_buf) {
+ desc_tx = dmaengine_prep_slave_sg(master->dma_tx,
+ tfr->tx_sg.sgl, tfr->tx_sg.nents,
+ DMA_TO_DEVICE,
+ DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
+ if (!desc_tx) {
+ dev_err(&sspi->master->dev,
+ "Couldn't prepare dma slave\n");
+ ret = -EIO;
+ goto out;
+ }
+
+ if (sspi->type == SPI_SUN4I)
+ trigger |= sspi_bits(sspi, SUNXI_CTL_DMA_TF_NOT_FULL);
+ else
+ trigger |= sspi_bits(sspi, SUNXI_CTL_DMA_TF_HALF);
+
+ dmaengine_submit(desc_tx);
+ dma_async_issue_pending(master->dma_tx);
+ }
+
+ if (sspi->rx_buf) {
+ desc_rx = dmaengine_prep_slave_sg(master->dma_rx,
+ tfr->rx_sg.sgl, tfr->rx_sg.nents,
+ DMA_FROM_DEVICE,
+ DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
+ if (!desc_rx) {
+ dev_err(&sspi->master->dev,
+ "Couldn't prepare dma slave\n");
+ ret = -EIO;
+ goto out;
+ }
+
+ trigger |= sspi_bits(sspi, SUNXI_CTL_DMA_RF_READY);
+
+ dmaengine_submit(desc_rx);
+ dma_async_issue_pending(master->dma_rx);
+ }
+
+ /* Enable Dedicated DMA requests */
+ if (sspi->type == SPI_SUN4I) {
+ sunxi_spi_set(sspi, SUNXI_TFR_CTL_REG,
+ sspi_bits(sspi, SUNXI_CTL_DMA_DEDICATED));
+ sunxi_spi_write(sspi, SUNXI_DMA_CTL_REG, trigger);
+ } else {
+ trigger |= sspi_bits(sspi, SUNXI_CTL_DMA_DEDICATED);
+ sunxi_spi_write(sspi, SUNXI_FIFO_CTL_REG, trigger);
+ }
+ } else {
+ dev_dbg(&sspi->master->dev, "Using PIO mode for transfer\n");
+
+ /* Disable DMA requests */
+ if (sspi->type == SPI_SUN4I) {
+ sunxi_spi_unset(sspi, SUNXI_TFR_CTL_REG,
+ sspi_bits(sspi, SUNXI_CTL_DMA_DEDICATED));
+ sunxi_spi_write(sspi, SUNXI_DMA_CTL_REG, 0);
+ } else {
+ sunxi_spi_write(sspi, SUNXI_FIFO_CTL_REG, 0);
+ }
+
+ /* Fill the TX FIFO */
+ sunxi_spi_fill_fifo(sspi, sspi->fifo_depth);
+ }
/* Enable the interrupts */
sunxi_spi_write(sspi, SUNXI_INT_CTL_REG,
sspi_bits(sspi, SUNXI_INT_CTL_TC));
/* Start the transfer */
- reg = sunxi_spi_read(sspi, SUNXI_TFR_CTL_REG);
- sunxi_spi_write(sspi, SUNXI_TFR_CTL_REG,
- reg | sspi_bits(sspi, SUNXI_TFR_CTL_XCH));
+ sunxi_spi_set(sspi, SUNXI_TFR_CTL_REG,
+ sspi_bits(sspi, SUNXI_TFR_CTL_XCH));
tx_time = max(tfr->len * 8 * 2 / (tfr->speed_hz / 1000), 100U);
start = jiffies;
@@ -431,9 +559,23 @@ static int sunxi_spi_transfer_one(struct spi_master *master,
goto out;
}
+out:
+ if (ret < 0 && sunxi_spi_can_dma(master, spi, tfr)) {
+ dev_dbg(&master->dev, "DMA channel teardown");
+ if (sspi->tx_buf)
+ dmaengine_terminate_sync(master->dma_tx);
+ if (sspi->rx_buf)
+ dmaengine_terminate_sync(master->dma_rx);
+ }
+
+ /*
+ * By this time either the transfer has ended and we have data in the
+ * FIFO buffer from a PIO RX transfer or the buffer is empty
+ * or something has failed.
+ * Empty the buffer either way to avoid leaving garbage around.
+ */
sunxi_spi_drain_fifo(sspi, sspi->fifo_depth);
-out:
sunxi_spi_write(sspi, SUNXI_INT_CTL_REG, 0);
return ret;
@@ -515,6 +657,7 @@ static int sunxi_spi_runtime_suspend(struct device *dev)
static int sunxi_spi_probe(struct platform_device *pdev)
{
+ struct dma_slave_config dma_sconfig;
struct spi_master *master;
struct sunxi_spi *sspi;
struct resource *res;
@@ -625,6 +768,54 @@ static int sunxi_spi_probe(struct platform_device *pdev)
}
}
+ master->dma_tx = dma_request_slave_channel_reason(&pdev->dev, "tx");
+ if (IS_ERR(master->dma_tx)) {
+ dev_err(&pdev->dev, "Unable to acquire DMA channel TX\n");
+ ret = PTR_ERR(master->dma_tx);
+ goto err_dma_chan;
+ }
+
+ dma_sconfig.direction = DMA_MEM_TO_DEV;
+ dma_sconfig.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
+ dma_sconfig.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
+ dma_sconfig.dst_addr = res->start + sspi_reg(sspi, SUNXI_TXDATA_REG);
+ dma_sconfig.src_maxburst = 1;
+ dma_sconfig.dst_maxburst = 1;
+
+ ret = dmaengine_slave_config(master->dma_tx, &dma_sconfig);
+ if (ret) {
+ dev_err(&pdev->dev, "Unable to configure TX DMA slave\n");
+ goto err_tx_dma_release;
+ }
+
+ master->dma_rx = dma_request_slave_channel_reason(&pdev->dev, "rx");
+ if (IS_ERR(master->dma_rx)) {
+ dev_err(&pdev->dev, "Unable to acquire DMA channel RX\n");
+ ret = PTR_ERR(master->dma_rx);
+ goto err_tx_dma_release;
+ }
+
+ dma_sconfig.direction = DMA_DEV_TO_MEM;
+ dma_sconfig.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
+ dma_sconfig.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
+ dma_sconfig.src_addr = res->start + sspi_reg(sspi, SUNXI_RXDATA_REG);
+ dma_sconfig.src_maxburst = 1;
+ dma_sconfig.dst_maxburst = 1;
+
+ ret = dmaengine_slave_config(master->dma_rx, &dma_sconfig);
+ if (ret) {
+ dev_err(&pdev->dev, "Unable to configure RX DMA slave\n");
+ goto err_rx_dma_release;
+ }
+
+ /*
+ * This is a bit dodgy. If you set can_dma then map_msg in spi.c
+ * apparently dereferences your dma channels if non-NULL even if your
+ * can_dma never returns true (and crashes if the channel is an error
+ * pointer). So just don't set can_dma unless both channels are valid.
+ */
+ master->can_dma = sunxi_spi_can_dma;
+wakeup:
/*
* This wake-up/shutdown pattern is to be able to have the
* device woken up, even if runtime_pm is disabled
@@ -665,18 +856,40 @@ static int sunxi_spi_probe(struct platform_device *pdev)
return 0;
+err_rx_dma_release:
+ dma_release_channel(master->dma_rx);
+err_tx_dma_release:
+ dma_release_channel(master->dma_tx);
+err_dma_chan:
+ master->dma_tx = NULL;
+ master->dma_rx = NULL;
+ if ((ret == -EPROBE_DEFER) && wait_for_dma)
+ goto err_free_master;
+ goto wakeup;
+
err_pm_disable:
pm_runtime_disable(&pdev->dev);
sunxi_spi_runtime_suspend(&pdev->dev);
err_free_master:
+ if (master->can_dma) {
+ dma_release_channel(master->dma_rx);
+ dma_release_channel(master->dma_tx);
+ }
spi_master_put(master);
return ret;
}
static int sunxi_spi_remove(struct platform_device *pdev)
{
+ struct spi_master *master = platform_get_drvdata(pdev);
+
pm_runtime_disable(&pdev->dev);
+ if (master->can_dma) {
+ dma_release_channel(master->dma_rx);
+ dma_release_channel(master->dma_tx);
+ }
+
return 0;
}