diff mbox

[v3,6/6] spi: sun6i: add DMA transfers support

Message ID 20180403154449.2443-7-ssuloev@orpaltech.com (mailing list archive)
State New, archived
Headers show

Commit Message

Sergey Suloev April 3, 2018, 3:44 p.m. UTC
DMA transfers are now available for sun6i and sun8i SoCs.
The DMA mode is used automatically as soon as requested
transfer length is more than FIFO length.

Changes in v3:
1) Debug log enhancements.

Signed-off-by: Sergey Suloev <ssuloev@orpaltech.com>
---
 drivers/spi/spi-sun6i.c | 331 ++++++++++++++++++++++++++++++++++++++++++------
 1 file changed, 294 insertions(+), 37 deletions(-)

Comments

Maxime Ripard April 4, 2018, 7 a.m. UTC | #1
On Tue, Apr 03, 2018 at 06:44:49PM +0300, Sergey Suloev wrote:
> DMA transfers are now available for sun6i and sun8i SoCs.
> The DMA mode is used automatically as soon as requested
> transfer length is more than FIFO length.
> 
> Changes in v3:
> 1) Debug log enhancements.
> 
> Signed-off-by: Sergey Suloev <ssuloev@orpaltech.com>
> ---
>  drivers/spi/spi-sun6i.c | 331 ++++++++++++++++++++++++++++++++++++++++++------
>  1 file changed, 294 insertions(+), 37 deletions(-)
> 
> diff --git a/drivers/spi/spi-sun6i.c b/drivers/spi/spi-sun6i.c
> index 2fa9d6e..7f41871 100644
> --- a/drivers/spi/spi-sun6i.c
> +++ b/drivers/spi/spi-sun6i.c
> @@ -14,6 +14,8 @@
>  #include <linux/clk.h>
>  #include <linux/delay.h>
>  #include <linux/device.h>
> +#include <linux/dmaengine.h>
> +#include <linux/dma-mapping.h>
>  #include <linux/interrupt.h>
>  #include <linux/io.h>
>  #include <linux/module.h>
> @@ -55,17 +57,20 @@
>  
>  #define SUN6I_FIFO_CTL_REG		0x18
>  #define SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_MASK	0xff
> -#define SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_BITS	0
> +#define SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_POS	0

Why are you changing that name

> +#define SUN6I_FIFO_CTL_RF_DRQ_EN		BIT(8)
>  #define SUN6I_FIFO_CTL_RF_RST			BIT(15)
>  #define SUN6I_FIFO_CTL_TF_ERQ_TRIG_LEVEL_MASK	0xff
> -#define SUN6I_FIFO_CTL_TF_ERQ_TRIG_LEVEL_BITS	16
> +#define SUN6I_FIFO_CTL_TF_ERQ_TRIG_LEVEL_POS	16

Ditto.

> +#define SUN6I_FIFO_CTL_TF_DRQ_EN		BIT(24)
>  #define SUN6I_FIFO_CTL_TF_RST			BIT(31)
> +#define SUN6I_FIFO_CTL_DMA_DEDICATE		BIT(9)|BIT(25)
>  
>  #define SUN6I_FIFO_STA_REG		0x1c
>  #define SUN6I_FIFO_STA_RF_CNT_MASK		0x7f
> -#define SUN6I_FIFO_STA_RF_CNT_BITS		0
> +#define SUN6I_FIFO_STA_RF_CNT_POS		0

Ditto.

>  #define SUN6I_FIFO_STA_TF_CNT_MASK		0x7f
> -#define SUN6I_FIFO_STA_TF_CNT_BITS		16
> +#define SUN6I_FIFO_STA_TF_CNT_POS		16

Ditto.

>  
>  #define SUN6I_CLK_CTL_REG		0x24
>  #define SUN6I_CLK_CTL_CDR2_MASK			0xff
> @@ -135,7 +140,7 @@ static inline u32 sun6i_spi_get_tx_fifo_count(struct sun6i_spi *sspi)
>  {
>  	u32 reg = sun6i_spi_read(sspi, SUN6I_FIFO_STA_REG);
>  
> -	reg >>= SUN6I_FIFO_STA_TF_CNT_BITS;
> +	reg >>= SUN6I_FIFO_STA_TF_CNT_POS;
>  
>  	return reg & SUN6I_FIFO_STA_TF_CNT_MASK;
>  }
> @@ -148,7 +153,7 @@ static inline void sun6i_spi_drain_fifo(struct sun6i_spi *sspi, int len)
>  	/* See how much data is available */
>  	reg = sun6i_spi_read(sspi, SUN6I_FIFO_STA_REG);
>  	reg &= SUN6I_FIFO_STA_RF_CNT_MASK;
> -	cnt = reg >> SUN6I_FIFO_STA_RF_CNT_BITS;
> +	cnt = reg >> SUN6I_FIFO_STA_RF_CNT_POS;
>  
>  	if (len > cnt)
>  		len = cnt;
> @@ -177,6 +182,15 @@ static inline void sun6i_spi_fill_fifo(struct sun6i_spi *sspi, int len)
>  	}
>  }
>  
> +static bool sun6i_spi_can_dma(struct spi_master *master,
> +			      struct spi_device *spi,
> +			      struct spi_transfer *tfr)
> +{
> +	struct sun6i_spi *sspi = spi_master_get_devdata(master);
> +
> +	return tfr->len > sspi->fifo_depth;
> +}
> +
>  static void sun6i_spi_set_cs(struct spi_device *spi, bool enable)
>  {
>  	struct sun6i_spi *sspi = spi_master_get_devdata(spi->master);
> @@ -208,6 +222,9 @@ static size_t sun6i_spi_max_transfer_size(struct spi_device *spi)
>  	struct spi_master *master = spi->master;
>  	struct sun6i_spi *sspi = spi_master_get_devdata(master);
>  
> +	if (master->can_dma)
> +		return SUN6I_MAX_XFER_SIZE;
> +
>  	return sspi->fifo_depth;
>  }
>  
> @@ -268,16 +285,187 @@ static int sun6i_spi_wait_for_transfer(struct spi_device *spi,
>  	return 0;
>  }
>  
> +static void sun6i_spi_dma_callback(void *param)
> +{
> +	struct spi_master *master = param;
> +
> +	dev_dbg(&master->dev, "DMA transfer complete\n");
> +	spi_finalize_current_transfer(master);
> +}
> +
> +static int sun6i_spi_dmap_prep_tx(struct spi_master *master,
> +				  struct spi_transfer *tfr,
> +				  dma_cookie_t *cookie)
> +{
> +	struct dma_async_tx_descriptor *chan_desc = NULL;
> +
> +	chan_desc = dmaengine_prep_slave_sg(master->dma_tx,
> +					    tfr->tx_sg.sgl, tfr->tx_sg.nents,
> +					    DMA_TO_DEVICE,
> +					    DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
> +	if (!chan_desc) {
> +		dev_err(&master->dev,
> +			"Couldn't prepare TX DMA slave\n");
> +		return -EIO;
> +	}
> +
> +	chan_desc->callback = sun6i_spi_dma_callback;
> +	chan_desc->callback_param = master;
> +
> +	*cookie = dmaengine_submit(chan_desc);
> +	dma_async_issue_pending(master->dma_tx);
> +
> +	return 0;
> +}
> +
> +static int sun6i_spi_dmap_prep_rx(struct spi_master *master,
> +				  struct spi_transfer *tfr,
> +				  dma_cookie_t *cookie)
> +{
> +	struct dma_async_tx_descriptor *chan_desc = NULL;
> +
> +	chan_desc = dmaengine_prep_slave_sg(master->dma_rx,
> +					    tfr->rx_sg.sgl, tfr->rx_sg.nents,
> +					    DMA_FROM_DEVICE,
> +					    DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
> +	if (!chan_desc) {
> +		dev_err(&master->dev,
> +			"Couldn't prepare RX DMA slave\n");
> +		return -EIO;
> +	}
> +
> +	chan_desc->callback = sun6i_spi_dma_callback;
> +	chan_desc->callback_param = master;
> +
> +	*cookie = dmaengine_submit(chan_desc);
> +	dma_async_issue_pending(master->dma_rx);
> +
> +	return 0;
> +}
> +
> +static int sun6i_spi_transfer_one_dma(struct spi_device *spi,
> +				      struct spi_transfer *tfr)
> +{
> +	struct spi_master *master = spi->master;
> +	struct sun6i_spi *sspi = spi_master_get_devdata(master);
> +	dma_cookie_t tx_cookie = 0,rx_cookie = 0;
> +	enum dma_status status;
> +	int ret;
> +	u32 reg, trig_level = 0;
> +
> +	dev_dbg(&master->dev, "Using DMA mode for transfer\n");
> +
> +	/* Disable interrupts */
> +	sun6i_spi_write(sspi, SUN6I_INT_CTL_REG, 0);
> +
> +	reg = sun6i_spi_read(sspi, SUN6I_FIFO_CTL_REG);
> +
> +	if (sspi->tx_buf) {
> +		ret = sun6i_spi_dmap_prep_tx(master, tfr, &tx_cookie);
> +		if (ret)
> +			goto out;
> +
> +		reg |= SUN6I_FIFO_CTL_TF_DRQ_EN;
> +
> +		trig_level = sspi->fifo_depth;
> +		reg &= ~SUN6I_FIFO_CTL_TF_ERQ_TRIG_LEVEL_MASK;
> +		reg |= (trig_level << SUN6I_FIFO_CTL_TF_ERQ_TRIG_LEVEL_POS);
> +	}
> +
> +	if (sspi->rx_buf) {
> +		ret = sun6i_spi_dmap_prep_rx(master, tfr, &rx_cookie);
> +		if (ret)
> +			goto out;
> +
> +		reg |= SUN6I_FIFO_CTL_RF_DRQ_EN;
> +
> +		trig_level = 1;
> +		reg &= ~SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_MASK;
> +		reg |= (trig_level << SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_POS);
> +	}
> +
> +	/* Enable Dedicated DMA requests */
> +	sun6i_spi_write(sspi, SUN6I_FIFO_CTL_REG,
> +			reg | SUN6I_FIFO_CTL_DMA_DEDICATE);
> +
> +	/* Start transfer */
> +	sun6i_spi_set(sspi, SUN6I_TFR_CTL_REG, SUN6I_TFR_CTL_XCH);
> +
> +	ret = sun6i_spi_wait_for_transfer(spi, tfr);
> +	if (ret)
> +		goto out;
> +
> +	if (sspi->tx_buf && (status = dma_async_is_tx_complete(master->dma_tx,
> +			tx_cookie, NULL, NULL))) {
> +		dev_warn(&master->dev,
> +			"DMA returned completion status of: %s\n",
> +			status == DMA_ERROR ? "error" : "in progress");
> +	}
> +	if (sspi->rx_buf && (status = dma_async_is_tx_complete(master->dma_rx,
> +			rx_cookie, NULL, NULL))) {
> +		dev_warn(&master->dev,
> +			"DMA returned completion status of: %s\n",
> +			status == DMA_ERROR ? "error" : "in progress");
> +	}

I'm not sure why that would be needed.

> +out:
> +	if (ret) {
> +		dev_dbg(&master->dev, "DMA channel teardown\n");
> +		if (sspi->tx_buf)
> +			dmaengine_terminate_sync(master->dma_tx);
> +		if (sspi->rx_buf)
> +			dmaengine_terminate_sync(master->dma_rx);
> +	}

That can be addressed through additional labels.

> +	sun6i_spi_drain_fifo(sspi, sspi->fifo_depth);

Why are you doing a PIO access here?

> +	return ret;
> +}
> +
> +static int sun6i_spi_transfer_one_pio(struct spi_device *spi,
> +				      struct spi_transfer *tfr)
> +{
> +	struct spi_master *master = spi->master;
> +	struct sun6i_spi *sspi = spi_master_get_devdata(master);
> +	unsigned int trig_level;
> +	int ret;
> +
> +	/* Disable DMA requests */
> +	sun6i_spi_write(sspi, SUN6I_FIFO_CTL_REG, 0);
> +
> +	/*
> +	 * Setup FIFO interrupt trigger level
> +	 * Here we choose 3/4 of the full fifo depth, as it's the hardcoded
> +	 * value used in old generation of Allwinner SPI controller.
> +	 * (See spi-sun4i.c)
> +	 */
> +	trig_level = sspi->fifo_depth / 4 * 3;
> +	sun6i_spi_write(sspi, SUN6I_FIFO_CTL_REG,
> +			(trig_level << SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_POS));
> +
> +	sun6i_spi_fill_fifo(sspi, sspi->fifo_depth);
> +
> +	/* Enable the interrupts */
> +	sun6i_spi_set(sspi, SUN6I_INT_CTL_REG, SUN6I_INT_CTL_TC |
> +					       SUN6I_INT_CTL_RF_RDY);
> +
> +	/* Start transfer */
> +	sun6i_spi_set(sspi, SUN6I_TFR_CTL_REG, SUN6I_TFR_CTL_XCH);
> +
> +	ret = sun6i_spi_wait_for_transfer(spi, tfr);
> +
> +	sun6i_spi_write(sspi, SUN6I_INT_CTL_REG, 0);
> +
> +	return ret;
> +}
> +
>  static int sun6i_spi_transfer_one(struct spi_master *master,
>  				  struct spi_device *spi,
>  				  struct spi_transfer *tfr)
>  {
>  	struct sun6i_spi *sspi = spi_master_get_devdata(master);
> -	unsigned int mclk_rate, div, timeout;
> -	unsigned int start, end, tx_time;
> -	unsigned int trig_level;
> +	unsigned int mclk_rate, div;
>  	unsigned int tx_len = 0;
> -	int ret = 0;
>  	u32 reg;
>  
>  	/* A zero length transfer never finishes if programmed
> @@ -285,10 +473,15 @@ static int sun6i_spi_transfer_one(struct spi_master *master,
>  	if (!tfr->len)
>  		return 0;
>  
> -	/* Don't support transfer larger than the FIFO */
> -	if (tfr->len > sspi->fifo_depth)
> +	if (tfr->len > SUN6I_MAX_XFER_SIZE)
>  		return -EMSGSIZE;
>  
> +	if (!master->can_dma) {
> +		/* Don't support transfer larger than the FIFO */
> +		if (tfr->len > sspi->fifo_depth)
> +			return -EMSGSIZE;
> +	}
> +
>  	sspi->tx_buf = tfr->tx_buf;
>  	sspi->rx_buf = tfr->rx_buf;
>  	sspi->len = tfr->len;
> @@ -300,16 +493,6 @@ static int sun6i_spi_transfer_one(struct spi_master *master,
>  	sun6i_spi_write(sspi, SUN6I_FIFO_CTL_REG,
>  			SUN6I_FIFO_CTL_RF_RST | SUN6I_FIFO_CTL_TF_RST);
>  
> -	/*
> -	 * Setup FIFO interrupt trigger level
> -	 * Here we choose 3/4 of the full fifo depth, as it's the hardcoded
> -	 * value used in old generation of Allwinner SPI controller.
> -	 * (See spi-sun4i.c)
> -	 */
> -	trig_level = sspi->fifo_depth / 4 * 3;
> -	sun6i_spi_write(sspi, SUN6I_FIFO_CTL_REG,
> -			(trig_level << SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_BITS));
> -
>  	/*
>  	 * If it's a TX only transfer, we don't want to fill the RX
>  	 * FIFO with bogus data
> @@ -364,22 +547,10 @@ static int sun6i_spi_transfer_one(struct spi_master *master,
>  	sun6i_spi_write(sspi, SUN6I_BURST_CTL_CNT_REG,
>  			SUN6I_BURST_CTL_CNT_STC(tx_len));
>  
> -	/* Fill the TX FIFO */
> -	sun6i_spi_fill_fifo(sspi, sspi->fifo_depth);
> -
> -	/* Enable the interrupts */
> -	sun6i_spi_enable_interrupt(sspi, SUN6I_INT_CTL_TC |
> -					 SUN6I_INT_CTL_RF_RDY);
> -
> -	/* Start the transfer */
> -	sun6i_spi_set(sspi, SUN6I_TFR_CTL_REG, SUN6I_TFR_CTL_XCH);
> -
> -	/* Wait for completion */
> -	ret = sun6i_spi_wait_for_transfer(spi, tfr);
> -
> -	sun6i_spi_write(sspi, SUN6I_INT_CTL_REG, 0);
> +	if (sun6i_spi_can_dma(master, spi, tfr))
> +		return sun6i_spi_transfer_one_dma(spi, tfr);
>  
> -	return ret;
> +	return sun6i_spi_transfer_one_pio(spi, tfr);

You should really split the function creation here into a separate patch

>  }
>  
>  static irqreturn_t sun6i_spi_handler(int irq, void *dev_id)
> @@ -409,6 +580,77 @@ static irqreturn_t sun6i_spi_handler(int irq, void *dev_id)
>  	return IRQ_NONE;
>  }
>  
> +static int sun6i_spi_dma_setup(struct platform_device *pdev,
> +			       struct resource *res)
> +{
> +	struct spi_master *master = platform_get_drvdata(pdev);
> +	struct sun6i_spi *sspi = spi_master_get_devdata(master);
> +	struct dma_slave_config dma_sconf;
> +	int ret;
> +
> +	master->dma_tx = dma_request_slave_channel_reason(&pdev->dev, "tx");
> +	if (IS_ERR(master->dma_tx)) {
> +		dev_err(&pdev->dev, "Unable to acquire DMA TX channel\n");
> +		ret = PTR_ERR(master->dma_tx);
> +		goto out;
> +	}
> +
> +	dma_sconf.direction = DMA_MEM_TO_DEV;
> +	dma_sconf.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
> +	dma_sconf.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
> +	dma_sconf.dst_addr = res->start + SUN6I_TXDATA_REG;
> +	dma_sconf.dst_maxburst = 1;
> +	dma_sconf.src_maxburst = 1;
> +
> +	ret = dmaengine_slave_config(master->dma_tx, &dma_sconf);
> +	if (ret) {
> +		dev_err(&pdev->dev, "Unable to configure DMA TX slave\n");
> +		goto err_rel_tx;
> +	}
> +
> +	master->dma_rx = dma_request_slave_channel_reason(&pdev->dev, "rx");
> +	if (IS_ERR(master->dma_rx)) {
> +		dev_err(&pdev->dev, "Unable to acquire DMA RX channel\n");
> +		ret = PTR_ERR(master->dma_rx);
> +		goto err_rel_tx;
> +	}
> +
> +	dma_sconf.direction = DMA_DEV_TO_MEM;
> +	dma_sconf.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
> +	dma_sconf.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
> +	dma_sconf.src_addr = res->start + SUN6I_RXDATA_REG;
> +	dma_sconf.src_maxburst = 1;
> +	dma_sconf.dst_maxburst = 1;
> +
> +	ret = dmaengine_slave_config(master->dma_rx, &dma_sconf);
> +	if (ret) {
> +		dev_err(&pdev->dev, "Unable to configure DMA RX slave\n");
> +		goto err_rel_rx;
> +	}
> +
> +	/* don't set can_dma unless both channels are valid*/
> +	master->can_dma = sun6i_spi_can_dma;
> +
> +	return 0;
> +
> +err_rel_rx:
> +	dma_release_channel(master->dma_rx);
> +err_rel_tx:
> +	dma_release_channel(master->dma_tx);
> +out:
> +	master->dma_tx = NULL;
> +	master->dma_rx = NULL;
> +	return ret;
> +}
> +
> +static void sun6i_spi_dma_release(struct spi_master *master)
> +{
> +	if (master->can_dma) {
> +		dma_release_channel(master->dma_rx);
> +		dma_release_channel(master->dma_tx);
> +	}
> +}
> +
>  static int sun6i_spi_runtime_resume(struct device *dev)
>  {
>  	struct spi_master *master = dev_get_drvdata(dev);
> @@ -530,6 +772,16 @@ static int sun6i_spi_probe(struct platform_device *pdev)
>  		goto err_free_master;
>  	}
>  
> +	ret = sun6i_spi_dma_setup(pdev, res);
> +	if (ret) {
> +		if (ret == -EPROBE_DEFER) {
> +			/* wait for the dma driver to load */

There's no need for that comment.

Maxime
diff mbox

Patch

diff --git a/drivers/spi/spi-sun6i.c b/drivers/spi/spi-sun6i.c
index 2fa9d6e..7f41871 100644
--- a/drivers/spi/spi-sun6i.c
+++ b/drivers/spi/spi-sun6i.c
@@ -14,6 +14,8 @@ 
 #include <linux/clk.h>
 #include <linux/delay.h>
 #include <linux/device.h>
+#include <linux/dmaengine.h>
+#include <linux/dma-mapping.h>
 #include <linux/interrupt.h>
 #include <linux/io.h>
 #include <linux/module.h>
@@ -55,17 +57,20 @@ 
 
 #define SUN6I_FIFO_CTL_REG		0x18
 #define SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_MASK	0xff
-#define SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_BITS	0
+#define SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_POS	0
+#define SUN6I_FIFO_CTL_RF_DRQ_EN		BIT(8)
 #define SUN6I_FIFO_CTL_RF_RST			BIT(15)
 #define SUN6I_FIFO_CTL_TF_ERQ_TRIG_LEVEL_MASK	0xff
-#define SUN6I_FIFO_CTL_TF_ERQ_TRIG_LEVEL_BITS	16
+#define SUN6I_FIFO_CTL_TF_ERQ_TRIG_LEVEL_POS	16
+#define SUN6I_FIFO_CTL_TF_DRQ_EN		BIT(24)
 #define SUN6I_FIFO_CTL_TF_RST			BIT(31)
+#define SUN6I_FIFO_CTL_DMA_DEDICATE		BIT(9)|BIT(25)
 
 #define SUN6I_FIFO_STA_REG		0x1c
 #define SUN6I_FIFO_STA_RF_CNT_MASK		0x7f
-#define SUN6I_FIFO_STA_RF_CNT_BITS		0
+#define SUN6I_FIFO_STA_RF_CNT_POS		0
 #define SUN6I_FIFO_STA_TF_CNT_MASK		0x7f
-#define SUN6I_FIFO_STA_TF_CNT_BITS		16
+#define SUN6I_FIFO_STA_TF_CNT_POS		16
 
 #define SUN6I_CLK_CTL_REG		0x24
 #define SUN6I_CLK_CTL_CDR2_MASK			0xff
@@ -135,7 +140,7 @@  static inline u32 sun6i_spi_get_tx_fifo_count(struct sun6i_spi *sspi)
 {
 	u32 reg = sun6i_spi_read(sspi, SUN6I_FIFO_STA_REG);
 
-	reg >>= SUN6I_FIFO_STA_TF_CNT_BITS;
+	reg >>= SUN6I_FIFO_STA_TF_CNT_POS;
 
 	return reg & SUN6I_FIFO_STA_TF_CNT_MASK;
 }
@@ -148,7 +153,7 @@  static inline void sun6i_spi_drain_fifo(struct sun6i_spi *sspi, int len)
 	/* See how much data is available */
 	reg = sun6i_spi_read(sspi, SUN6I_FIFO_STA_REG);
 	reg &= SUN6I_FIFO_STA_RF_CNT_MASK;
-	cnt = reg >> SUN6I_FIFO_STA_RF_CNT_BITS;
+	cnt = reg >> SUN6I_FIFO_STA_RF_CNT_POS;
 
 	if (len > cnt)
 		len = cnt;
@@ -177,6 +182,15 @@  static inline void sun6i_spi_fill_fifo(struct sun6i_spi *sspi, int len)
 	}
 }
 
+static bool sun6i_spi_can_dma(struct spi_master *master,
+			      struct spi_device *spi,
+			      struct spi_transfer *tfr)
+{
+	struct sun6i_spi *sspi = spi_master_get_devdata(master);
+
+	return tfr->len > sspi->fifo_depth;
+}
+
 static void sun6i_spi_set_cs(struct spi_device *spi, bool enable)
 {
 	struct sun6i_spi *sspi = spi_master_get_devdata(spi->master);
@@ -208,6 +222,9 @@  static size_t sun6i_spi_max_transfer_size(struct spi_device *spi)
 	struct spi_master *master = spi->master;
 	struct sun6i_spi *sspi = spi_master_get_devdata(master);
 
+	if (master->can_dma)
+		return SUN6I_MAX_XFER_SIZE;
+
 	return sspi->fifo_depth;
 }
 
@@ -268,16 +285,187 @@  static int sun6i_spi_wait_for_transfer(struct spi_device *spi,
 	return 0;
 }
 
+static void sun6i_spi_dma_callback(void *param)
+{
+	struct spi_master *master = param;
+
+	dev_dbg(&master->dev, "DMA transfer complete\n");
+	spi_finalize_current_transfer(master);
+}
+
+static int sun6i_spi_dmap_prep_tx(struct spi_master *master,
+				  struct spi_transfer *tfr,
+				  dma_cookie_t *cookie)
+{
+	struct dma_async_tx_descriptor *chan_desc = NULL;
+
+	chan_desc = dmaengine_prep_slave_sg(master->dma_tx,
+					    tfr->tx_sg.sgl, tfr->tx_sg.nents,
+					    DMA_TO_DEVICE,
+					    DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
+	if (!chan_desc) {
+		dev_err(&master->dev,
+			"Couldn't prepare TX DMA slave\n");
+		return -EIO;
+	}
+
+	chan_desc->callback = sun6i_spi_dma_callback;
+	chan_desc->callback_param = master;
+
+	*cookie = dmaengine_submit(chan_desc);
+	dma_async_issue_pending(master->dma_tx);
+
+	return 0;
+}
+
+static int sun6i_spi_dmap_prep_rx(struct spi_master *master,
+				  struct spi_transfer *tfr,
+				  dma_cookie_t *cookie)
+{
+	struct dma_async_tx_descriptor *chan_desc = NULL;
+
+	chan_desc = dmaengine_prep_slave_sg(master->dma_rx,
+					    tfr->rx_sg.sgl, tfr->rx_sg.nents,
+					    DMA_FROM_DEVICE,
+					    DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
+	if (!chan_desc) {
+		dev_err(&master->dev,
+			"Couldn't prepare RX DMA slave\n");
+		return -EIO;
+	}
+
+	chan_desc->callback = sun6i_spi_dma_callback;
+	chan_desc->callback_param = master;
+
+	*cookie = dmaengine_submit(chan_desc);
+	dma_async_issue_pending(master->dma_rx);
+
+	return 0;
+}
+
+static int sun6i_spi_transfer_one_dma(struct spi_device *spi,
+				      struct spi_transfer *tfr)
+{
+	struct spi_master *master = spi->master;
+	struct sun6i_spi *sspi = spi_master_get_devdata(master);
+	dma_cookie_t tx_cookie = 0,rx_cookie = 0;
+	enum dma_status status;
+	int ret;
+	u32 reg, trig_level = 0;
+
+	dev_dbg(&master->dev, "Using DMA mode for transfer\n");
+
+	/* Disable interrupts */
+	sun6i_spi_write(sspi, SUN6I_INT_CTL_REG, 0);
+
+	reg = sun6i_spi_read(sspi, SUN6I_FIFO_CTL_REG);
+
+	if (sspi->tx_buf) {
+		ret = sun6i_spi_dmap_prep_tx(master, tfr, &tx_cookie);
+		if (ret)
+			goto out;
+
+		reg |= SUN6I_FIFO_CTL_TF_DRQ_EN;
+
+		trig_level = sspi->fifo_depth;
+		reg &= ~SUN6I_FIFO_CTL_TF_ERQ_TRIG_LEVEL_MASK;
+		reg |= (trig_level << SUN6I_FIFO_CTL_TF_ERQ_TRIG_LEVEL_POS);
+	}
+
+	if (sspi->rx_buf) {
+		ret = sun6i_spi_dmap_prep_rx(master, tfr, &rx_cookie);
+		if (ret)
+			goto out;
+
+		reg |= SUN6I_FIFO_CTL_RF_DRQ_EN;
+
+		trig_level = 1;
+		reg &= ~SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_MASK;
+		reg |= (trig_level << SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_POS);
+	}
+
+	/* Enable Dedicated DMA requests */
+	sun6i_spi_write(sspi, SUN6I_FIFO_CTL_REG,
+			reg | SUN6I_FIFO_CTL_DMA_DEDICATE);
+
+	/* Start transfer */
+	sun6i_spi_set(sspi, SUN6I_TFR_CTL_REG, SUN6I_TFR_CTL_XCH);
+
+	ret = sun6i_spi_wait_for_transfer(spi, tfr);
+	if (ret)
+		goto out;
+
+	if (sspi->tx_buf && (status = dma_async_is_tx_complete(master->dma_tx,
+			tx_cookie, NULL, NULL))) {
+		dev_warn(&master->dev,
+			"DMA returned completion status of: %s\n",
+			status == DMA_ERROR ? "error" : "in progress");
+	}
+	if (sspi->rx_buf && (status = dma_async_is_tx_complete(master->dma_rx,
+			rx_cookie, NULL, NULL))) {
+		dev_warn(&master->dev,
+			"DMA returned completion status of: %s\n",
+			status == DMA_ERROR ? "error" : "in progress");
+	}
+
+out:
+	if (ret) {
+		dev_dbg(&master->dev, "DMA channel teardown\n");
+		if (sspi->tx_buf)
+			dmaengine_terminate_sync(master->dma_tx);
+		if (sspi->rx_buf)
+			dmaengine_terminate_sync(master->dma_rx);
+	}
+
+	sun6i_spi_drain_fifo(sspi, sspi->fifo_depth);
+
+	return ret;
+}
+
+static int sun6i_spi_transfer_one_pio(struct spi_device *spi,
+				      struct spi_transfer *tfr)
+{
+	struct spi_master *master = spi->master;
+	struct sun6i_spi *sspi = spi_master_get_devdata(master);
+	unsigned int trig_level;
+	int ret;
+
+	/* Disable DMA requests */
+	sun6i_spi_write(sspi, SUN6I_FIFO_CTL_REG, 0);
+
+	/*
+	 * Setup FIFO interrupt trigger level
+	 * Here we choose 3/4 of the full fifo depth, as it's the hardcoded
+	 * value used in old generation of Allwinner SPI controller.
+	 * (See spi-sun4i.c)
+	 */
+	trig_level = sspi->fifo_depth / 4 * 3;
+	sun6i_spi_write(sspi, SUN6I_FIFO_CTL_REG,
+			(trig_level << SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_POS));
+
+	sun6i_spi_fill_fifo(sspi, sspi->fifo_depth);
+
+	/* Enable the interrupts */
+	sun6i_spi_set(sspi, SUN6I_INT_CTL_REG, SUN6I_INT_CTL_TC |
+					       SUN6I_INT_CTL_RF_RDY);
+
+	/* Start transfer */
+	sun6i_spi_set(sspi, SUN6I_TFR_CTL_REG, SUN6I_TFR_CTL_XCH);
+
+	ret = sun6i_spi_wait_for_transfer(spi, tfr);
+
+	sun6i_spi_write(sspi, SUN6I_INT_CTL_REG, 0);
+
+	return ret;
+}
+
 static int sun6i_spi_transfer_one(struct spi_master *master,
 				  struct spi_device *spi,
 				  struct spi_transfer *tfr)
 {
 	struct sun6i_spi *sspi = spi_master_get_devdata(master);
-	unsigned int mclk_rate, div, timeout;
-	unsigned int start, end, tx_time;
-	unsigned int trig_level;
+	unsigned int mclk_rate, div;
 	unsigned int tx_len = 0;
-	int ret = 0;
 	u32 reg;
 
 	/* A zero length transfer never finishes if programmed
@@ -285,10 +473,15 @@  static int sun6i_spi_transfer_one(struct spi_master *master,
 	if (!tfr->len)
 		return 0;
 
-	/* Don't support transfer larger than the FIFO */
-	if (tfr->len > sspi->fifo_depth)
+	if (tfr->len > SUN6I_MAX_XFER_SIZE)
 		return -EMSGSIZE;
 
+	if (!master->can_dma) {
+		/* Don't support transfer larger than the FIFO */
+		if (tfr->len > sspi->fifo_depth)
+			return -EMSGSIZE;
+	}
+
 	sspi->tx_buf = tfr->tx_buf;
 	sspi->rx_buf = tfr->rx_buf;
 	sspi->len = tfr->len;
@@ -300,16 +493,6 @@  static int sun6i_spi_transfer_one(struct spi_master *master,
 	sun6i_spi_write(sspi, SUN6I_FIFO_CTL_REG,
 			SUN6I_FIFO_CTL_RF_RST | SUN6I_FIFO_CTL_TF_RST);
 
-	/*
-	 * Setup FIFO interrupt trigger level
-	 * Here we choose 3/4 of the full fifo depth, as it's the hardcoded
-	 * value used in old generation of Allwinner SPI controller.
-	 * (See spi-sun4i.c)
-	 */
-	trig_level = sspi->fifo_depth / 4 * 3;
-	sun6i_spi_write(sspi, SUN6I_FIFO_CTL_REG,
-			(trig_level << SUN6I_FIFO_CTL_RF_RDY_TRIG_LEVEL_BITS));
-
 	/*
 	 * If it's a TX only transfer, we don't want to fill the RX
 	 * FIFO with bogus data
@@ -364,22 +547,10 @@  static int sun6i_spi_transfer_one(struct spi_master *master,
 	sun6i_spi_write(sspi, SUN6I_BURST_CTL_CNT_REG,
 			SUN6I_BURST_CTL_CNT_STC(tx_len));
 
-	/* Fill the TX FIFO */
-	sun6i_spi_fill_fifo(sspi, sspi->fifo_depth);
-
-	/* Enable the interrupts */
-	sun6i_spi_enable_interrupt(sspi, SUN6I_INT_CTL_TC |
-					 SUN6I_INT_CTL_RF_RDY);
-
-	/* Start the transfer */
-	sun6i_spi_set(sspi, SUN6I_TFR_CTL_REG, SUN6I_TFR_CTL_XCH);
-
-	/* Wait for completion */
-	ret = sun6i_spi_wait_for_transfer(spi, tfr);
-
-	sun6i_spi_write(sspi, SUN6I_INT_CTL_REG, 0);
+	if (sun6i_spi_can_dma(master, spi, tfr))
+		return sun6i_spi_transfer_one_dma(spi, tfr);
 
-	return ret;
+	return sun6i_spi_transfer_one_pio(spi, tfr);
 }
 
 static irqreturn_t sun6i_spi_handler(int irq, void *dev_id)
@@ -409,6 +580,77 @@  static irqreturn_t sun6i_spi_handler(int irq, void *dev_id)
 	return IRQ_NONE;
 }
 
+static int sun6i_spi_dma_setup(struct platform_device *pdev,
+			       struct resource *res)
+{
+	struct spi_master *master = platform_get_drvdata(pdev);
+	struct sun6i_spi *sspi = spi_master_get_devdata(master);
+	struct dma_slave_config dma_sconf;
+	int ret;
+
+	master->dma_tx = dma_request_slave_channel_reason(&pdev->dev, "tx");
+	if (IS_ERR(master->dma_tx)) {
+		dev_err(&pdev->dev, "Unable to acquire DMA TX channel\n");
+		ret = PTR_ERR(master->dma_tx);
+		goto out;
+	}
+
+	dma_sconf.direction = DMA_MEM_TO_DEV;
+	dma_sconf.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
+	dma_sconf.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
+	dma_sconf.dst_addr = res->start + SUN6I_TXDATA_REG;
+	dma_sconf.dst_maxburst = 1;
+	dma_sconf.src_maxburst = 1;
+
+	ret = dmaengine_slave_config(master->dma_tx, &dma_sconf);
+	if (ret) {
+		dev_err(&pdev->dev, "Unable to configure DMA TX slave\n");
+		goto err_rel_tx;
+	}
+
+	master->dma_rx = dma_request_slave_channel_reason(&pdev->dev, "rx");
+	if (IS_ERR(master->dma_rx)) {
+		dev_err(&pdev->dev, "Unable to acquire DMA RX channel\n");
+		ret = PTR_ERR(master->dma_rx);
+		goto err_rel_tx;
+	}
+
+	dma_sconf.direction = DMA_DEV_TO_MEM;
+	dma_sconf.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
+	dma_sconf.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
+	dma_sconf.src_addr = res->start + SUN6I_RXDATA_REG;
+	dma_sconf.src_maxburst = 1;
+	dma_sconf.dst_maxburst = 1;
+
+	ret = dmaengine_slave_config(master->dma_rx, &dma_sconf);
+	if (ret) {
+		dev_err(&pdev->dev, "Unable to configure DMA RX slave\n");
+		goto err_rel_rx;
+	}
+
+	/* don't set can_dma unless both channels are valid*/
+	master->can_dma = sun6i_spi_can_dma;
+
+	return 0;
+
+err_rel_rx:
+	dma_release_channel(master->dma_rx);
+err_rel_tx:
+	dma_release_channel(master->dma_tx);
+out:
+	master->dma_tx = NULL;
+	master->dma_rx = NULL;
+	return ret;
+}
+
+static void sun6i_spi_dma_release(struct spi_master *master)
+{
+	if (master->can_dma) {
+		dma_release_channel(master->dma_rx);
+		dma_release_channel(master->dma_tx);
+	}
+}
+
 static int sun6i_spi_runtime_resume(struct device *dev)
 {
 	struct spi_master *master = dev_get_drvdata(dev);
@@ -530,6 +772,16 @@  static int sun6i_spi_probe(struct platform_device *pdev)
 		goto err_free_master;
 	}
 
+	ret = sun6i_spi_dma_setup(pdev, res);
+	if (ret) {
+		if (ret == -EPROBE_DEFER) {
+			/* wait for the dma driver to load */
+			goto err_free_master;
+		}
+		dev_warn(&pdev->dev,
+			"Unable to setup DMA channels: DMA transfers disabled\n");
+	}
+
 	/*
 	 * This wake-up/shutdown pattern is to be able to have the
 	 * device woken up, even if runtime_pm is disabled
@@ -556,14 +808,19 @@  err_pm_disable:
 	pm_runtime_disable(&pdev->dev);
 	sun6i_spi_runtime_suspend(&pdev->dev);
 err_free_master:
+	sun6i_spi_dma_release(master);
 	spi_master_put(master);
 	return ret;
 }
 
 static int sun6i_spi_remove(struct platform_device *pdev)
 {
+	struct spi_master *master = platform_get_drvdata(pdev);
+
 	pm_runtime_force_suspend(&pdev->dev);
 
+	sun6i_spi_dma_release(master);
+
 	return 0;
 }