Message ID | 1393628200-12317-1-git-send-email-santosh.shilimkar@ti.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Friday 28 February 2014 17:56:40 Santosh Shilimkar wrote: > The Packet DMA driver sets up the dma channels and flows for the > QMSS(Queue Manager SubSystem) who triggers the actual data movements > across clients using destination queues. Every client modules like > NETCP(Network Coprocessor), SRIO(Serial Rapid IO) and CRYPTO > Engines has its own instance of packet dma hardware. QMSS has also > an internal packet DMA module which is used as an infrastructure > DMA with zero copy. > > Patch adds DMAEngine driver for Keystone Packet DMA hardware. > The specifics on the device tree bindings for Packet DMA can be > found in: > Documentation/devicetree/bindings/dma/keystone-pktdma.txt > > The driver implements the configuration functions using standard DMAEngine > apis. The data movement is managed by QMSS device driver. The description of this sounds a bit like the x-gene queue manager/tag manager, rather than a traditional DMA engine. For that driver, we decided to use the mailbox subsystem instead of the DMA subsystem. Can you explain what kind of data is actually getting transferred in by your hardware here? You say the DMA is performed by the QMSS, so do you use the pktdma driver to drive the data transfers of the QMSS or do you just use it for flow control by passing short (a few bytes) messages and also have to program the pktdma? > + > +bool dma_keystone_filter_fn(struct dma_chan *chan, void *param) > +{ > + if (chan->device->dev->driver == &keystone_dma_driver.driver) { > + struct keystone_dma_chan *kdma_chan = from_achan(chan); > + unsigned chan_id = *(u32 *)param & KEYSTONE_DMA_CHAN_ID_MASK; > + unsigned chan_type = *(u32 *)param >> KEYSTONE_DMA_TYPE_SHIFT; > + > + if (chan_type == KEYSTONE_DMA_TX_CHAN_TYPE && > + kdma_chan->direction == DMA_MEM_TO_DEV) > + return chan_id == kdma_chan->channel; > + > + if (chan_type == KEYSTONE_DMA_RX_FLOW_TYPE && > + kdma_chan->direction == DMA_DEV_TO_MEM) > + return chan_id == kdma_chan->flow; > + } > + return false; > +} > +EXPORT_SYMBOL_GPL(dma_keystone_filter_fn); The filter function should be removed here and replaced with a simpler custom xlate function calling dma_get_slave_chan() on the matching channel. > diff --git a/include/linux/keystone-pktdma.h b/include/linux/keystone-pktdma.h > new file mode 100644 > index 0000000..e6a66c8 > --- /dev/null > +++ b/include/linux/keystone-pktdma.h > @@ -0,0 +1,168 @@ A DMA engine driver should not have a public API. Please move all the contents of the two header files into the driver itself to ensure none of this is visible to slave drivers. If it turns out that you use declarations from this header in slave drivers at the moment, please explain what they are so we can discuss alternative solutions. Arnd
On Friday 28 February 2014 06:14 PM, Arnd Bergmann wrote: > On Friday 28 February 2014 17:56:40 Santosh Shilimkar wrote: >> The Packet DMA driver sets up the dma channels and flows for the >> QMSS(Queue Manager SubSystem) who triggers the actual data movements >> across clients using destination queues. Every client modules like >> NETCP(Network Coprocessor), SRIO(Serial Rapid IO) and CRYPTO >> Engines has its own instance of packet dma hardware. QMSS has also >> an internal packet DMA module which is used as an infrastructure >> DMA with zero copy. >> >> Patch adds DMAEngine driver for Keystone Packet DMA hardware. >> The specifics on the device tree bindings for Packet DMA can be >> found in: >> Documentation/devicetree/bindings/dma/keystone-pktdma.txt >> >> The driver implements the configuration functions using standard DMAEngine >> apis. The data movement is managed by QMSS device driver. > > The description of this sounds a bit like the x-gene queue manager/tag > manager, rather than a traditional DMA engine. For that driver, we > decided to use the mailbox subsystem instead of the DMA subsystem. > > Can you explain what kind of data is actually getting transferred > in by your hardware here? You say the DMA is performed by the QMSS, > so do you use the pktdma driver to drive the data transfers of the > QMSS or do you just use it for flow control by passing short (a few > bytes) messages and also have to program the pktdma? > I don't know what x-gene stuff is, but the keystone packet dma is hardware DMA controller(per client subystem) which has all the traditional DMA controller properties. The packet DMA needs to be setup just like any slave dma configuration. The actual data transfers are triggered by operations on Queue Manager SubSystem. I just posted the QMSS driver as well which explains how the system work. >> + >> +bool dma_keystone_filter_fn(struct dma_chan *chan, void *param) >> +{ >> + if (chan->device->dev->driver == &keystone_dma_driver.driver) { >> + struct keystone_dma_chan *kdma_chan = from_achan(chan); >> + unsigned chan_id = *(u32 *)param & KEYSTONE_DMA_CHAN_ID_MASK; >> + unsigned chan_type = *(u32 *)param >> KEYSTONE_DMA_TYPE_SHIFT; >> + >> + if (chan_type == KEYSTONE_DMA_TX_CHAN_TYPE && >> + kdma_chan->direction == DMA_MEM_TO_DEV) >> + return chan_id == kdma_chan->channel; >> + >> + if (chan_type == KEYSTONE_DMA_RX_FLOW_TYPE && >> + kdma_chan->direction == DMA_DEV_TO_MEM) >> + return chan_id == kdma_chan->flow; >> + } >> + return false; >> +} >> +EXPORT_SYMBOL_GPL(dma_keystone_filter_fn); > > The filter function should be removed here and replaced with a simpler > custom xlate function calling dma_get_slave_chan() on the matching > channel. > Can you please expand your comment please ? I though the filter function is the one should implement for specific channel allocations. >> diff --git a/include/linux/keystone-pktdma.h b/include/linux/keystone-pktdma.h >> new file mode 100644 >> index 0000000..e6a66c8 >> --- /dev/null >> +++ b/include/linux/keystone-pktdma.h >> @@ -0,0 +1,168 @@ > > A DMA engine driver should not have a public API. Please move all the > contents of the two header files into the driver itself to ensure none > of this is visible to slave drivers. If it turns out that you use > declarations from this header in slave drivers at the moment, please > explain what they are so we can discuss alternative solutions. > Yes the slave drivers make use of these. The dt-binding header is created because some the macro's are shared between DTS and code. keystone-pktdma.h carriers information about the packet dma config which is done by slave drivers. It also contains the dma descriptor format and its associate macro's used by slave drivers. Since the hardware perspective, the descriptor represent the packet format dma expects and hence its left in the dma header header. Regards, Santosh
On Fri, 2014-02-28 at 17:56 -0500, Santosh Shilimkar wrote: > From: Sandeep Nair <sandeep_n@ti.com> > > The Packet DMA driver sets up the dma channels and flows for the > QMSS(Queue Manager SubSystem) who triggers the actual data movements > across clients using destination queues. Every client modules like > NETCP(Network Coprocessor), SRIO(Serial Rapid IO) and CRYPTO > Engines has its own instance of packet dma hardware. QMSS has also > an internal packet DMA module which is used as an infrastructure > DMA with zero copy. > > Patch adds DMAEngine driver for Keystone Packet DMA hardware. > The specifics on the device tree bindings for Packet DMA can be > found in: > Documentation/devicetree/bindings/dma/keystone-pktdma.txt > > The driver implements the configuration functions using standard DMAEngine > apis. The data movement is managed by QMSS device driver. > Few comments below. > Cc: Vinod Koul <vinod.koul@intel.com> > Cc: Russell King <linux@arm.linux.org.uk> > Cc: Grant Likely <grant.likely@linaro.org> > Cc: Rob Herring <robh+dt@kernel.org> > Cc: Mark Rutland <mark.rutland@arm.com> > Signed-off-by: Sandeep Nair <sandeep_n@ti.com> > Signed-off-by: Santosh Shilimkar <santosh.shilimkar@ti.com> > --- > .../devicetree/bindings/dma/keystone-pktdma.txt | 72 ++ > drivers/dma/Kconfig | 8 + > drivers/dma/Makefile | 1 + > drivers/dma/keystone-pktdma.c | 795 ++++++++++++++++++++ > include/dt-bindings/dma/keystone.h | 33 + > include/linux/keystone-pktdma.h | 168 +++++ > 6 files changed, 1077 insertions(+) > create mode 100644 Documentation/devicetree/bindings/dma/keystone-pktdma.txt > create mode 100644 drivers/dma/keystone-pktdma.c > create mode 100644 include/dt-bindings/dma/keystone.h > create mode 100644 include/linux/keystone-pktdma.h > > diff --git a/Documentation/devicetree/bindings/dma/keystone-pktdma.txt b/Documentation/devicetree/bindings/dma/keystone-pktdma.txt > new file mode 100644 > index 0000000..ea061d5 > --- /dev/null > +++ b/Documentation/devicetree/bindings/dma/keystone-pktdma.txt > @@ -0,0 +1,72 @@ > +Keystone Packet DMA Controller > + > +This document explains the device tree bindings for the packet dma > +on keystone devices. The the Network coprocessor, Cypto engines > +and the SRIO on Keystone devices all have their own packet dma modules. > +Each individual packet dma has a certain number of RX channels, > +RX flows and TX channels. Each instance of the packet DMA is being > +initialized through device specific bindings. > + > +* DMA controller > + > +Required properties: > + - compatible: Should be "ti,keystone-pktdma" > + - reg: Should contain register location and length of the following pktdma > + register regions. The value for "reg-names" property of the respective > + region is specified in parenthesis. > + - Global control register region (global). > + - Tx DMA channel configuration register region (txchan). > + - Rx DMA channel configuration register region (rxchan). > + - Tx DMA channel Scheduler configuration register region (txsched). > + - Rx DMA flow configuration register region (rxflow). > + - reg-names: Names for the above regions. The name to be used is specified in > + the above description of "reg" property. > + - qm-base-address: Base address of the logical queue managers for pktdma. > + - #dma-cells: Has to be 1. Keystone-pktdma doesn't support anything else. > + > +Optional properties: > + - enable-all: Enable all DMA channels. > + - loop-back: To loopback Tx streaming I/F to Rx streaming I/F. Used for > + infrastructure transfers. > + - rx-retry-timeout: Number of pktdma cycles to wait before retry on buffer > + starvation. > + > +Example: > + netcp-dma: pktdma@2004000 { > + compatible = "ti,keystone-pktdma"; > + reg = <0x2004000 0x100>, > + <0x2004400 0x120>, > + <0x2004800 0x300>, > + <0x2004c00 0x120>, > + <0x2005000 0x400>; > + reg-names = "global", "txchan", "rxchan", "txsched", > + "rxflow"; > + qm-base-address = <0x23a80000 0x23a90000 > + 0x23aa0000 0x23ab0000>; > + #dma-cells = <1>; > + /* enable-all; */ > + rx-retry-timeout = <3500>; > + /* loop-back; */ > + }; > + > + > +* DMA client > + > +Required properties: > +- dmas: One DMA request specifier consisting of a phandle to the DMA controller > + followed by the integer specifying the channel identifier. The channel > + identifier is encoded as follows: > + - bits 7-0: Tx DMA channel number or the Rx flow number. > + - bits 31-24: Channel type. 0xff for Tx DMA channel & 0xfe for Rx flow. > +- dma-names: List of string identifiers for the DMA requests. > + > +Example: > + > + netcp: netcp@2090000 { > + ... > + dmas = <&netcpdma KEYSTONE_DMA_RX_FLOW(22)>, > + <&netcpdma KEYSTONE_DMA_RX_FLOW(23)>, > + <&netcpdma KEYSTONE_DMA_TX_CHAN(8)>; > + dma-names = "netrx0", "netrx1", "nettx"; > + ... > + }; > diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig > index 9bed1a2..722b99a 100644 > --- a/drivers/dma/Kconfig > +++ b/drivers/dma/Kconfig > @@ -350,6 +350,14 @@ config MOXART_DMA > help > Enable support for the MOXA ART SoC DMA controller. > > +config KEYSTONE_PKTDMA > + tristate "TI Keystone Packet DMA support" > + depends on ARCH_KEYSTONE > + select DMA_ENGINE > + help > + Enable support for the Packet DMA engine on Texas Instruments' > + Keystone family of devices. > + > config DMA_ENGINE > bool > > diff --git a/drivers/dma/Makefile b/drivers/dma/Makefile > index a029d0f4..6d69c6d 100644 > --- a/drivers/dma/Makefile > +++ b/drivers/dma/Makefile > @@ -44,3 +44,4 @@ obj-$(CONFIG_DMA_JZ4740) += dma-jz4740.o > obj-$(CONFIG_TI_CPPI41) += cppi41.o > obj-$(CONFIG_K3_DMA) += k3dma.o > obj-$(CONFIG_MOXART_DMA) += moxart-dma.o > +obj-$(CONFIG_KEYSTONE_PKTDMA) += keystone-pktdma.o > diff --git a/drivers/dma/keystone-pktdma.c b/drivers/dma/keystone-pktdma.c > new file mode 100644 > index 0000000..b3f77e5 > --- /dev/null > +++ b/drivers/dma/keystone-pktdma.c > @@ -0,0 +1,795 @@ > +/* > + * Copyright (C) 2014 Texas Instruments Incorporated > + * Authors: Sandeep Nair <sandeep_n@ti.com> > + * Cyril Chemparathy <cyril@ti.com> > + * Santosh Shilimkar <santosh.shilimkar@ti.com> > + * > + * This program is free software; you can redistribute it and/or > + * modify it under the terms of the GNU General Public License as > + * published by the Free Software Foundation version 2. > + * > + * This program is distributed "as is" WITHOUT ANY WARRANTY of any > + * kind, whether express or implied; without even the implied warranty > + * of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > + * GNU General Public License for more details. > + */ > + > +#include <linux/io.h> > +#include <linux/sched.h> > +#include <linux/module.h> > +#include <linux/dma-direction.h> > +#include <linux/dmaengine.h> > +#include <linux/interrupt.h> > +#include <linux/of_dma.h> > +#include <linux/of_address.h> > +#include <linux/platform_device.h> > +#include <linux/keystone-pktdma.h> > +#include <linux/pm_runtime.h> > +#include <dt-bindings/dma/keystone.h> > + > +#define BITS(x) (BIT(x) - 1) > +#define REG_MASK 0xffffffff > + > +#define DMA_LOOPBACK BIT(31) > +#define DMA_ENABLE BIT(31) > +#define DMA_TEARDOWN BIT(30) > + > +#define DMA_TX_FILT_PSWORDS BIT(29) > +#define DMA_TX_FILT_EINFO BIT(30) > +#define DMA_TX_PRIO_SHIFT 0 > +#define DMA_RX_PRIO_SHIFT 16 > +#define DMA_PRIO_MASK BITS(3) > +#define DMA_PRIO_DEFAULT 0 > +#define DMA_RX_TIMEOUT_DEFAULT 17500 /* cycles */ > +#define DMA_RX_TIMEOUT_MASK BITS(16) > +#define DMA_RX_TIMEOUT_SHIFT 0 > + > +#define CHAN_HAS_EPIB BIT(30) > +#define CHAN_HAS_PSINFO BIT(29) > +#define CHAN_ERR_RETRY BIT(28) > +#define CHAN_PSINFO_AT_SOP BIT(25) > +#define CHAN_SOP_OFF_SHIFT 16 > +#define CHAN_SOP_OFF_MASK BITS(9) > +#define DESC_TYPE_SHIFT 26 > +#define DESC_TYPE_MASK BITS(2) > + > +/* > + * QMGR & QNUM together make up 14 bits with QMGR as the 2 MSb's in the logical > + * navigator cloud mapping scheme. > + * using the 14bit physical queue numbers directly maps into this scheme. > + */ > +#define CHAN_QNUM_MASK BITS(14) > +#define DMA_MAX_QMS 4 > +#define DMA_TIMEOUT 1000 /* msecs */ > + > +struct reg_global { > + u32 revision; > + u32 perf_control; > + u32 emulation_control; > + u32 priority_control; > + u32 qm_base_address[4]; > +}; > + > +struct reg_chan { > + u32 control; > + u32 mode; > + u32 __rsvd[6]; > +}; > + > +struct reg_tx_sched { > + u32 prio; > +}; > + > +struct reg_rx_flow { > + u32 control; > + u32 tags; > + u32 tag_sel; > + u32 fdq_sel[2]; > + u32 thresh[3]; > +}; > + > +#define BUILD_CHECK_REGS() \ > + do { \ > + BUILD_BUG_ON(sizeof(struct reg_global) != 32); \ > + BUILD_BUG_ON(sizeof(struct reg_chan) != 32); \ > + BUILD_BUG_ON(sizeof(struct reg_rx_flow) != 32); \ > + BUILD_BUG_ON(sizeof(struct reg_tx_sched) != 4); \ > + } while (0) > + > +enum keystone_chan_state { > + /* stable states */ > + CHAN_STATE_OPENED, > + CHAN_STATE_CLOSED, > +}; > + > +struct keystone_dma_device { > + struct dma_device engine; > + bool loopback, enable_all; > + unsigned tx_priority, rx_priority, rx_timeout; > + unsigned logical_queue_managers; > + unsigned qm_base_address[DMA_MAX_QMS]; > + struct reg_global __iomem *reg_global; > + struct reg_chan __iomem *reg_tx_chan; > + struct reg_rx_flow __iomem *reg_rx_flow; > + struct reg_chan __iomem *reg_rx_chan; > + struct reg_tx_sched __iomem *reg_tx_sched; > + unsigned max_rx_chan, max_tx_chan; > + unsigned max_rx_flow; > + atomic_t in_use; > +}; > +#define to_dma(dma) (&(dma)->engine) > +#define dma_dev(dma) ((dma)->engine.dev) > + > +struct keystone_dma_chan { > + struct dma_chan achan; > + enum dma_transfer_direction direction; > + atomic_t state; > + struct keystone_dma_device *dma; > + > + /* registers */ > + struct reg_chan __iomem *reg_chan; > + struct reg_tx_sched __iomem *reg_tx_sched; > + struct reg_rx_flow __iomem *reg_rx_flow; > + > + /* configuration stuff */ > + unsigned channel, flow; > +}; > +#define from_achan(ch) container_of(ch, struct keystone_dma_chan, achan) > +#define to_achan(ch) (&(ch)->achan) > +#define chan_dev(ch) (&to_achan(ch)->dev->device) > +#define chan_num(ch) ((ch->direction == DMA_MEM_TO_DEV) ? \ > + ch->channel : ch->flow) > +#define chan_vdbg(ch, format, arg...) \ > + dev_vdbg(chan_dev(ch), format, ##arg); > + > +/** > + * dev_to_dma_chan - convert a device pointer to the its sysfs container object > + * @dev - device node > + */ > +static inline struct dma_chan *dev_to_dma_chan(struct device *dev) > +{ > + struct dma_chan_dev *chan_dev; > + > + chan_dev = container_of(dev, typeof(*chan_dev), device); > + return chan_dev->chan; > +} > + > +static inline enum keystone_chan_state > +chan_get_state(struct keystone_dma_chan *chan) > +{ > + return atomic_read(&chan->state); > +} > + > +static int chan_start(struct keystone_dma_chan *chan, > + struct dma_keystone_cfg *cfg) > +{ > + u32 v = 0; > + > + if ((chan->direction == DMA_MEM_TO_DEV) && chan->reg_chan) { > + if (cfg->u.tx.filt_pswords) > + v |= DMA_TX_FILT_PSWORDS; > + if (cfg->u.tx.filt_einfo) > + v |= DMA_TX_FILT_EINFO; > + writel_relaxed(v, &chan->reg_chan->mode); > + writel_relaxed(DMA_ENABLE, &chan->reg_chan->control); > + } > + > + if (chan->reg_tx_sched) > + writel_relaxed(cfg->u.tx.priority, &chan->reg_tx_sched->prio); > + > + if (chan->reg_rx_flow) { > + v = 0; > + > + if (cfg->u.rx.einfo_present) > + v |= CHAN_HAS_EPIB; > + if (cfg->u.rx.psinfo_present) > + v |= CHAN_HAS_PSINFO; > + if (cfg->u.rx.err_mode == DMA_RETRY) > + v |= CHAN_ERR_RETRY; > + v |= (cfg->u.rx.desc_type & DESC_TYPE_MASK) << DESC_TYPE_SHIFT; > + if (cfg->u.rx.psinfo_at_sop) > + v |= CHAN_PSINFO_AT_SOP; > + v |= (cfg->u.rx.sop_offset & CHAN_SOP_OFF_MASK) > + << CHAN_SOP_OFF_SHIFT; > + v |= cfg->u.rx.dst_q & CHAN_QNUM_MASK; > + > + writel_relaxed(v, &chan->reg_rx_flow->control); > + writel_relaxed(0, &chan->reg_rx_flow->tags); > + writel_relaxed(0, &chan->reg_rx_flow->tag_sel); > + > + v = cfg->u.rx.fdq[0] << 16; > + v |= cfg->u.rx.fdq[1] & CHAN_QNUM_MASK; > + writel_relaxed(v, &chan->reg_rx_flow->fdq_sel[0]); > + > + v = cfg->u.rx.fdq[2] << 16; > + v |= cfg->u.rx.fdq[3] & CHAN_QNUM_MASK; > + writel_relaxed(v, &chan->reg_rx_flow->fdq_sel[1]); > + > + writel_relaxed(0, &chan->reg_rx_flow->thresh[0]); > + writel_relaxed(0, &chan->reg_rx_flow->thresh[1]); > + writel_relaxed(0, &chan->reg_rx_flow->thresh[2]); > + } > + > + return 0; > +} > + > +static int chan_teardown(struct keystone_dma_chan *chan) > +{ > + unsigned long end, value; > + > + if (!chan->reg_chan) > + return 0; > + > + /* indicate teardown */ > + writel_relaxed(DMA_TEARDOWN, &chan->reg_chan->control); > + > + /* wait for the dma to shut itself down */ > + end = jiffies + msecs_to_jiffies(DMA_TIMEOUT); > + do { > + value = readl_relaxed(&chan->reg_chan->control); > + if ((value & DMA_ENABLE) == 0) > + break; > + } while (time_after(end, jiffies)); > + > + if (readl_relaxed(&chan->reg_chan->control) & DMA_ENABLE) { > + dev_err(chan_dev(chan), "timeout waiting for teardown\n"); > + return -ETIMEDOUT; > + } > + > + return 0; > +} > + > +static void chan_stop(struct keystone_dma_chan *chan) > +{ > + if (chan->reg_rx_flow) { > + /* first detach fdqs, starve out the flow */ > + writel_relaxed(0, &chan->reg_rx_flow->fdq_sel[0]); > + writel_relaxed(0, &chan->reg_rx_flow->fdq_sel[1]); > + writel_relaxed(0, &chan->reg_rx_flow->thresh[0]); > + writel_relaxed(0, &chan->reg_rx_flow->thresh[1]); > + writel_relaxed(0, &chan->reg_rx_flow->thresh[2]); > + } > + > + /* teardown the dma channel */ > + chan_teardown(chan); > + > + /* then disconnect the completion side */ > + if (chan->reg_rx_flow) { > + writel_relaxed(0, &chan->reg_rx_flow->control); > + writel_relaxed(0, &chan->reg_rx_flow->tags); > + writel_relaxed(0, &chan->reg_rx_flow->tag_sel); > + } > + chan_vdbg(chan, "channel stopped\n"); > +} > + > +static void keystone_dma_hw_init(struct keystone_dma_device *dma) > +{ > + unsigned v; > + int i; > + > + v = dma->loopback ? DMA_LOOPBACK : 0; > + writel_relaxed(v, &dma->reg_global->emulation_control); > + > + v = readl_relaxed(&dma->reg_global->perf_control); > + v |= ((dma->rx_timeout & DMA_RX_TIMEOUT_MASK) << DMA_RX_TIMEOUT_SHIFT); > + writel_relaxed(v, &dma->reg_global->perf_control); > + > + v = ((dma->tx_priority << DMA_TX_PRIO_SHIFT) | > + (dma->rx_priority << DMA_RX_PRIO_SHIFT)); > + > + writel_relaxed(v, &dma->reg_global->priority_control); > + > + if (dma->enable_all) { > + for (i = 0; i < dma->max_tx_chan; i++) { > + writel_relaxed(0, &dma->reg_tx_chan[i].mode); > + writel_relaxed(DMA_ENABLE, > + &dma->reg_tx_chan[i].control); > + } > + } > + > + /* Always enable all Rx channels. Rx paths are managed using flows */ > + for (i = 0; i < dma->max_rx_chan; i++) > + writel_relaxed(DMA_ENABLE, &dma->reg_rx_chan[i].control); > + > + for (i = 0; i < dma->logical_queue_managers; i++) > + writel_relaxed(dma->qm_base_address[i], > + &dma->reg_global->qm_base_address[i]); > +} > + > +static void keystone_dma_hw_destroy(struct keystone_dma_device *dma) > +{ > + int i; > + unsigned v; > + > + v = ~DMA_ENABLE & REG_MASK; > + > + for (i = 0; i < dma->max_rx_chan; i++) > + writel_relaxed(v, &dma->reg_rx_chan[i].control); > + > + for (i = 0; i < dma->max_tx_chan; i++) > + writel_relaxed(v, &dma->reg_tx_chan[i].control); > +} > + > +static int chan_init(struct dma_chan *achan) > +{ > + struct keystone_dma_chan *chan = from_achan(achan); > + struct keystone_dma_device *dma = chan->dma; > + > + chan_vdbg(chan, "initializing %s channel\n", > + chan->direction == DMA_MEM_TO_DEV ? "transmit" : > + chan->direction == DMA_DEV_TO_MEM ? "receive" : > + "unknown"); > + > + if (chan->direction != DMA_MEM_TO_DEV && > + chan->direction != DMA_DEV_TO_MEM) { is_slave_direction(). > + dev_err(chan_dev(chan), "bad direction\n"); > + return -EINVAL; > + } > + > + atomic_set(&chan->state, CHAN_STATE_OPENED); > + > + if (atomic_inc_return(&dma->in_use) <= 1) > + keystone_dma_hw_init(dma); > + > + return 0; > +} > + > +static void chan_destroy(struct dma_chan *achan) > +{ > + struct keystone_dma_chan *chan = from_achan(achan); > + struct keystone_dma_device *dma = chan->dma; > + > + if (chan_get_state(chan) == CHAN_STATE_CLOSED) > + return; > + > + chan_vdbg(chan, "destroying channel\n"); > + chan_stop(chan); > + atomic_set(&chan->state, CHAN_STATE_CLOSED); > + if (atomic_dec_return(&dma->in_use) <= 0) > + keystone_dma_hw_destroy(dma); > + chan_vdbg(chan, "channel destroyed\n"); > +} > + > +static int chan_keystone_config(struct keystone_dma_chan *chan, > + struct dma_keystone_cfg *cfg) > +{ > + if (chan_get_state(chan) != CHAN_STATE_OPENED) > + return -ENODEV; > + > + if (cfg->sl_cfg.direction != chan->direction) > + return -EINVAL; I think slave_config.direction is kinda deprecated. > + > + return chan_start(chan, cfg); > +} > + > +static int chan_control(struct dma_chan *achan, enum dma_ctrl_cmd cmd, > + unsigned long arg) > +{ > + struct keystone_dma_chan *chan = from_achan(achan); > + struct dma_keystone_cfg *keystone_config; > + struct dma_slave_config *dma_cfg; > + int ret; > + > + switch (cmd) { > + case DMA_SLAVE_CONFIG: > + dma_cfg = (struct dma_slave_config *)arg; > + keystone_config = keystone_cfg_from_slave_config(dma_cfg); > + ret = chan_keystone_config(chan, keystone_config); > + break; > + > + default: > + ret = -ENOTSUPP; > + break; > + } > + return ret; You may avoid 'ret' variable at all. Just return value explicitly. > +} > + > +static void __iomem *pktdma_get_regs( > + struct keystone_dma_device *dma, const char *name, > + resource_size_t *_size) > +{ > + struct device *dev = dma_dev(dma); > + struct device_node *node = dev->of_node; > + resource_size_t size; > + struct resource res; > + void __iomem *regs; > + int i; > + > + i = of_property_match_string(node, "reg-names", name); > + if (of_address_to_resource(node, i, &res)) { > + dev_err(dev, "could not find %s resource(index %d)\n", name, i); > + return NULL; > + } > + size = resource_size(&res); > + > + regs = of_iomap(node, i); > + if (!regs) { > + dev_err(dev, "could not map %s resource (index %d)\n", name, i); > + return NULL; > + } > + > + dev_dbg(dev, "index: %d, res:%s, size:%x, phys:%x, virt:%p\n", > + i, name, (unsigned int)size, (unsigned int)res.start, regs); I think here is better to use %pR to print resource. > + > + if (_size) > + *_size = size; > + > + return regs; > +} > + > +static int pktdma_init_rx_chan(struct keystone_dma_chan *chan, > + struct device_node *node, > + u32 flow) > +{ > + struct keystone_dma_device *dma = chan->dma; > + struct device *dev = dma_dev(chan->dma); > + > + chan->flow = flow; > + chan->reg_rx_flow = dma->reg_rx_flow + flow; > + dev_dbg(dev, "rx flow(%d) (%p)\n", chan->flow, chan->reg_rx_flow); > + > + return 0; > +} > + > +static int pktdma_init_tx_chan(struct keystone_dma_chan *chan, > + struct device_node *node, > + u32 channel) > +{ > + struct keystone_dma_device *dma = chan->dma; > + struct device *dev = dma_dev(chan->dma); > + > + chan->channel = channel; > + chan->reg_chan = dma->reg_tx_chan + channel; > + chan->reg_tx_sched = dma->reg_tx_sched + channel; > + dev_dbg(dev, "tx channel(%d) (%p)\n", chan->channel, chan->reg_chan); > + > + return 0; > +} > + > +static int pktdma_init_chan(struct keystone_dma_device *dma, > + struct device_node *node, > + enum dma_transfer_direction dir, > + unsigned chan_num) > +{ > + struct device *dev = dma_dev(dma); > + struct keystone_dma_chan *chan; > + struct dma_chan *achan; > + int ret = -EINVAL; > + > + chan = devm_kzalloc(dev, sizeof(*chan), GFP_KERNEL); > + if (!chan) > + return -ENOMEM; > + > + achan = to_achan(chan); > + achan->device = &dma->engine; > + chan->dma = dma; > + chan->direction = DMA_NONE; > + atomic_set(&chan->state, CHAN_STATE_OPENED); > + > + if (dir == DMA_MEM_TO_DEV) { > + chan->direction = dir; > + ret = pktdma_init_tx_chan(chan, node, chan_num); > + } else if (dir == DMA_DEV_TO_MEM) { > + chan->direction = dir; > + ret = pktdma_init_rx_chan(chan, node, chan_num); > + } else { > + dev_err(dev, "channel(%d) direction unknown\n", chan_num); > + } > + > + if (ret < 0) > + goto fail; > + > + list_add_tail(&to_achan(chan)->device_node, &to_dma(dma)->channels); > + return 0; > + > +fail: > + devm_kfree(dev, chan); > + return ret; Below you skip channels which failed to initialize. Is it real use case? > +} > + > +/* dummy function: feature not supported */ > +static enum dma_status chan_xfer_status(struct dma_chan *achan, > + dma_cookie_t cookie, > + struct dma_tx_state *txstate) > +{ > + WARN(1, "xfer status not supported\n"); > + return DMA_ERROR; > +} > + > +/* dummy function: feature not supported */ > +static void chan_issue_pending(struct dma_chan *chan) > +{ > + WARN(1, "issue pending not supported\n"); > +} > + > +static ssize_t keystone_dma_show_chan_num(struct device *dev, > + struct device_attribute *attr, char *buf) > +{ > + struct dma_chan *achan = dev_to_dma_chan(dev); > + struct keystone_dma_chan *chan = from_achan(achan); > + > + return scnprintf(buf, PAGE_SIZE, "%u\n", chan->channel); > +} > + > +static ssize_t keystone_dma_show_flow(struct device *dev, > + struct device_attribute *attr, char *buf) > +{ > + struct dma_chan *achan = dev_to_dma_chan(dev); > + struct keystone_dma_chan *chan = from_achan(achan); > + > + return scnprintf(buf, PAGE_SIZE, "%u\n", chan->flow); > +} > + > +static DEVICE_ATTR(chan_num, S_IRUSR, keystone_dma_show_chan_num, NULL); > +static DEVICE_ATTR(rx_flow, S_IRUSR, keystone_dma_show_flow, NULL); > + > +static void keystone_dma_destroy_attr(struct keystone_dma_device *dma) > +{ > + struct dma_device *engine = to_dma(dma); > + struct keystone_dma_chan *chan; > + struct dma_chan *achan; > + struct device *dev; > + > + list_for_each_entry(achan, &engine->channels, device_node) { > + chan = from_achan(achan); > + dev = chan_dev(chan); > + > + /* remove sysfs entries */ > + if (chan->direction == DMA_MEM_TO_DEV) > + device_remove_file(dev, &dev_attr_chan_num); > + else > + device_remove_file(dev, &dev_attr_rx_flow); > + } > +} > + > +static int keystone_dma_setup_attr(struct keystone_dma_device *dma) > +{ > + struct dma_device *engine = to_dma(dma); > + struct keystone_dma_chan *chan; > + struct dma_chan *achan; > + struct device *dev; > + int status = 0; > + > + list_for_each_entry(achan, &engine->channels, device_node) { > + chan = from_achan(achan); > + dev = chan_dev(chan); > + > + /* add sysfs entries */ > + if (chan->direction == DMA_MEM_TO_DEV) { > + status = device_create_file(dev, &dev_attr_chan_num); > + if (status) > + dev_warn(dev, > + "Couldn't create sysfs chan_num\n"); > + } else { > + status = device_create_file(dev, &dev_attr_rx_flow); > + if (status) > + dev_warn(dev, > + "Couldn't create sysfs rx_flow\n"); > + } > + } > + return status; > +} > + > +static struct of_dma_filter_info keystone_dma_info = { > + .filter_fn = dma_keystone_filter_fn, > +}; > + > +static int keystone_dma_probe(struct platform_device *pdev) > +{ > + unsigned max_tx_chan, max_rx_chan, max_rx_flow, max_tx_sched; > + struct device_node *node = pdev->dev.of_node; > + struct keystone_dma_device *dma; > + struct device_node *chan; > + int ret, len, num_chan = 0; > + struct dma_device *engine; > + resource_size_t size; > + u32 timeout; > + u32 i; > + > + if (!node) { > + dev_err(&pdev->dev, "could not find device info\n"); > + return -EINVAL; > + } > + > + dma = devm_kzalloc(&pdev->dev, sizeof(*dma), GFP_KERNEL); > + if (!dma) { > + dev_err(&pdev->dev, "could not allocate driver mem\n"); > + return -ENOMEM; > + } > + > + engine = to_dma(dma); > + engine->dev = &pdev->dev; > + platform_set_drvdata(pdev, dma); > + > + pm_runtime_enable(&pdev->dev); > + ret = pm_runtime_get_sync(&pdev->dev); > + if (ret < 0) { > + dev_err(&pdev->dev, "unable to enable pktdma, err %d\n", ret); > + return ret; > + } > + > + dma->reg_global = pktdma_get_regs(dma, "global", &size); > + if (!dma->reg_global) > + return -ENODEV; > + if (size < sizeof(struct reg_global)) { > + dev_err(dma_dev(dma), "bad size %pa for global regs\n", &size); > + return -ENODEV; > + } > + > + dma->reg_tx_chan = pktdma_get_regs(dma, "txchan", &size); > + if (!dma->reg_tx_chan) > + return -ENODEV; > + > + max_tx_chan = size / sizeof(struct reg_chan); > + dma->reg_rx_chan = pktdma_get_regs(dma, "rxchan", &size); > + if (!dma->reg_rx_chan) > + return -ENODEV; > + > + max_rx_chan = size / sizeof(struct reg_chan); > + dma->reg_tx_sched = pktdma_get_regs(dma, "txsched", &size); > + if (!dma->reg_tx_sched) > + return -ENODEV; > + > + max_tx_sched = size / sizeof(struct reg_tx_sched); > + dma->reg_rx_flow = pktdma_get_regs(dma, "rxflow", &size); > + if (!dma->reg_rx_flow) > + return -ENODEV; > + > + max_rx_flow = size / sizeof(struct reg_rx_flow); > + dma->rx_priority = DMA_PRIO_DEFAULT; > + dma->tx_priority = DMA_PRIO_DEFAULT; > + > + dma->enable_all = (of_get_property(node, "enable-all", NULL) != NULL); > + dma->loopback = (of_get_property(node, "loop-back", NULL) != NULL); > + > + ret = of_property_read_u32(node, "rx-retry-timeout", &timeout); > + if (ret < 0) { > + dev_dbg(&pdev->dev, "unspecified rx timeout using value %d\n", > + DMA_RX_TIMEOUT_DEFAULT); > + timeout = DMA_RX_TIMEOUT_DEFAULT; > + } > + dma->rx_timeout = timeout; > + > + if (!of_find_property(node, "qm-base-address", &len)) { > + dev_err(&pdev->dev, "unspecified qm-base-address\n"); > + return -ENODEV; > + } > + > + dma->logical_queue_managers = len / sizeof(u32); > + if (dma->logical_queue_managers > DMA_MAX_QMS) { > + dev_warn(&pdev->dev, "too many queue mgrs(>%d) rest ignored\n", mgrs -> managers ? Since the string too long, put it on separate line, but don't break. > + dma->logical_queue_managers); > + dma->logical_queue_managers = DMA_MAX_QMS; > + } > + > + ret = of_property_read_u32_array(node, "qm-base-address", > + dma->qm_base_address, > + dma->logical_queue_managers); > + if (ret) { > + dev_err(&pdev->dev, "invalid qm-base-address\n"); > + return ret; > + } > + > + dma->max_rx_chan = max_rx_chan; > + dma->max_rx_flow = max_rx_flow; > + dma->max_tx_chan = min(max_tx_chan, max_tx_sched); > + > + atomic_set(&dma->in_use, 0); > + > + INIT_LIST_HEAD(&engine->channels); > + > + for (i = 0; i < dma->max_tx_chan; i++) { > + if (pktdma_init_chan(dma, chan, DMA_MEM_TO_DEV, i) >= 0) > + num_chan++; > + } > + > + for (i = 0; i < dma->max_rx_chan; i++) { > + if (pktdma_init_chan(dma, chan, DMA_DEV_TO_MEM, i) >= 0) > + num_chan++; > + } > + > + if (list_empty(&engine->channels)) { > + dev_err(dma_dev(dma), "no valid channels\n"); > + return -ENODEV; > + } > + > + dma_cap_set(DMA_SLAVE, engine->cap_mask); > + engine->device_alloc_chan_resources = chan_init; > + engine->device_free_chan_resources = chan_destroy; > + engine->device_control = chan_control; > + engine->device_tx_status = chan_xfer_status; > + engine->device_issue_pending = chan_issue_pending; > + > + ret = dma_async_device_register(engine); > + if (ret) { > + dev_err(&pdev->dev, "unable to register dma engine\n"); > + return ret; > + } > + > + keystone_dma_info.dma_cap = engine->cap_mask; > + ret = of_dma_controller_register(node, of_dma_simple_xlate, > + &keystone_dma_info); > + if (ret) { > + dev_err(&pdev->dev, "Failed to register dma controller\n"); > + dma_async_device_unregister(engine); > + return ret; > + } > + > + ret = keystone_dma_setup_attr(dma); > + if (ret) { > + dev_err(&pdev->dev, "unable to setup device attr\n"); > + return ret; > + } > + > + dev_info(dma_dev(dma), "registered %d logical channels, flows %d, tx chans: %d, rx chans: %d%s\n", > + num_chan, dma->max_rx_flow, dma->max_tx_chan, > + dma->max_rx_chan, dma->loopback ? ", loopback" : ""); > + > + return 0; > +} > + > +static int keystone_dma_remove(struct platform_device *pdev) > +{ > + struct keystone_dma_device *dma = platform_get_drvdata(pdev); > + struct dma_device *engine = to_dma(dma); > + > + pm_runtime_put_sync(&pdev->dev); > + pm_runtime_disable(&pdev->dev); > + > + keystone_dma_destroy_attr(dma); > + dma_async_device_unregister(engine); > + > + return 0; > +} > + > +static struct of_device_id of_match[] = { > + { .compatible = "ti,keystone-pktdma", }, > + {}, > +}; > + > +MODULE_DEVICE_TABLE(of, of_match); > + > +static struct platform_driver keystone_dma_driver = { > + .probe = keystone_dma_probe, > + .remove = keystone_dma_remove, > + .driver = { > + .name = "keystone-pktdma", > + .owner = THIS_MODULE, > + .of_match_table = of_match, > + }, > +}; > + > +bool dma_keystone_filter_fn(struct dma_chan *chan, void *param) > +{ > + if (chan->device->dev->driver == &keystone_dma_driver.driver) { > + struct keystone_dma_chan *kdma_chan = from_achan(chan); > + unsigned chan_id = *(u32 *)param & KEYSTONE_DMA_CHAN_ID_MASK; > + unsigned chan_type = *(u32 *)param >> KEYSTONE_DMA_TYPE_SHIFT; > + > + if (chan_type == KEYSTONE_DMA_TX_CHAN_TYPE && > + kdma_chan->direction == DMA_MEM_TO_DEV) > + return chan_id == kdma_chan->channel; > + > + if (chan_type == KEYSTONE_DMA_RX_FLOW_TYPE && > + kdma_chan->direction == DMA_DEV_TO_MEM) > + return chan_id == kdma_chan->flow; > + } > + return false; > +} > +EXPORT_SYMBOL_GPL(dma_keystone_filter_fn); > + > +static int __init keystone_dma_init(void) > +{ > + BUILD_CHECK_REGS(); Could you move this somewhere else and use module_platform_driver() instead of init/exit? > + return platform_driver_register(&keystone_dma_driver); > +} > +module_init(keystone_dma_init); > + > +static void __exit keystone_dma_exit(void) > +{ > + platform_driver_unregister(&keystone_dma_driver); > +} > +module_exit(keystone_dma_exit); > + > +MODULE_LICENSE("GPL v2"); > +MODULE_DESCRIPTION("TI Keystone Packet DMA driver"); > diff --git a/include/dt-bindings/dma/keystone.h b/include/dt-bindings/dma/keystone.h > new file mode 100644 > index 0000000..abe029e > --- /dev/null > +++ b/include/dt-bindings/dma/keystone.h > @@ -0,0 +1,33 @@ > +/* > + * Copyright (C) 2014 Texas Instruments Incorporated > + * Authors: Sandeep Nair <sandeep_n@ti.com> > + * Santosh Shilimkar <santosh.shilimkar@ti.com> > + * > + * This program is free software; you can redistribute it and/or > + * modify it under the terms of the GNU General Public License as > + * published by the Free Software Foundation version 2. > + * > + * This program is distributed "as is" WITHOUT ANY WARRANTY of any > + * kind, whether express or implied; without even the implied warranty > + * of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > + * GNU General Public License for more details. > + */ > + > +#ifndef __DT_BINDINGS_KEYSTONE_DMA_H__ > +#define __DT_BINDINGS_KEYSTONE_DMA_H__ > + > +#define KEYSTONE_DMA_CHAN_ID_MASK (0xff) > +#define KEYSTONE_DMA_TYPE_SHIFT (24) > +#define KEYSTONE_DMA_TX_CHAN_TYPE (0xff) > +#define KEYSTONE_DMA_RX_FLOW_TYPE (0xfe) > + > +#define KEYSTONE_DMA_TX_CHAN(id) ((id & KEYSTONE_DMA_CHAN_ID_MASK) | \ > + (KEYSTONE_DMA_TX_CHAN_TYPE << \ > + KEYSTONE_DMA_TYPE_SHIFT)) > + > +#define KEYSTONE_DMA_RX_FLOW(id) ((id & KEYSTONE_DMA_CHAN_ID_MASK) | \ > + (KEYSTONE_DMA_RX_FLOW_TYPE << \ > + KEYSTONE_DMA_TYPE_SHIFT)) > + > + > +#endif /* __DT_BINDINGS_KEYSTONE_DMA_H_ */ > diff --git a/include/linux/keystone-pktdma.h b/include/linux/keystone-pktdma.h > new file mode 100644 > index 0000000..e6a66c8 > --- /dev/null > +++ b/include/linux/keystone-pktdma.h > @@ -0,0 +1,168 @@ > +/* > + * Copyright (C) 2014 Texas Instruments Incorporated > + * Authors: Sandeep Nair <sandeep_n@ti.com > + * Cyril Chemparathy <cyril@ti.com > + Santosh Shilimkar <santosh.shilimkar@ti.com> > + * > + * This program is free software; you can redistribute it and/or > + * modify it under the terms of the GNU General Public License as > + * published by the Free Software Foundation version 2. > + * > + * This program is distributed "as is" WITHOUT ANY WARRANTY of any > + * kind, whether express or implied; without even the implied warranty > + * of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > + * GNU General Public License for more details. > + */ > + > +#ifndef __KEYSTONE_PKTDMA_H__ > +#define __KEYSTONE_PKTDMA_H__ > + > +#include <linux/dmaengine.h> > + > +/* > + * PKTDMA descriptor manipulation macros for host packet descriptor > + */ > +#define MASK(x) (BIT(x) - 1) > +#define DMA_KEYSTONE_DESC_PKT_LEN_MASK MASK(22) > +#define DMA_KEYSTONE_DESC_PKT_LEN_SHIFT 0 > +#define DMA_KEYSTONE_DESC_PS_INFO_IN_SOP BIT(22) > +#define DMA_KEYSTONE_DESC_PS_INFO_IN_DESC 0 > +#define DMA_KEYSTONE_DESC_TAG_MASK MASK(8) > +#define DMA_KEYSTONE_DESC_SAG_HI_SHIFT 24 > +#define DMA_KEYSTONE_DESC_STAG_LO_SHIFT 16 > +#define DMA_KEYSTONE_DESC_DTAG_HI_SHIFT 8 > +#define DMA_KEYSTONE_DESC_DTAG_LO_SHIFT 0 > +#define DMA_KEYSTONE_DESC_HAS_EPIB BIT(31) > +#define DMA_KEYSTONE_DESC_NO_EPIB 0 > +#define DMA_KEYSTONE_DESC_PSLEN_SHIFT 24 > +#define DMA_KEYSTONE_DESC_PSLEN_MASK MASK(6) > +#define DMA_KEYSTONE_DESC_ERR_FLAG_SHIFT 20 > +#define DMA_KEYSTONE_DESC_ERR_FLAG_MASK MASK(4) > +#define DMA_KEYSTONE_DESC_PSFLAG_SHIFT 16 > +#define DMA_KEYSTONE_DESC_PSFLAG_MASK MASK(4) > +#define DMA_KEYSTONE_DESC_RETQ_SHIFT 0 > +#define DMA_KEYSTONE_DESC_RETQ_MASK MASK(14) > +#define DMA_KEYSTONE_DESC_BUF_LEN_MASK MASK(22) > + > +#define DMA_KEYSTONE_NUM_EPIB_WORDS 4 > +#define DMA_KEYSTONE_NUM_PS_WORDS 16 > +#define DMA_KEYSTONE_FDQ_PER_CHAN 4 > + > +/* Tx channel scheduling priority */ > +enum dma_keystone_tx_priority { > + DMA_PRIO_HIGH = 0, > + DMA_PRIO_MED_H, > + DMA_PRIO_MED_L, > + DMA_PRIO_LOW > +}; > + > +/* Rx channel error handling mode during buffer starvation */ > +enum dma_keystone_rx_err_mode { > + DMA_DROP = 0, > + DMA_RETRY > +}; > + > +/* Rx flow size threshold configuration */ > +enum dma_keystone_rx_thresholds { > + DMA_THRESH_NONE = 0, > + DMA_THRESH_0 = 1, > + DMA_THRESH_0_1 = 3, > + DMA_THRESH_0_1_2 = 7 > +}; > + > +/* Descriptor type */ > +enum dma_keystone_desc_type { > + DMA_DESC_HOST = 0, > + DMA_DESC_MONOLITHIC = 2 > +}; > + > +/** > + * struct dma_keystone_tx_cfg: Tx channel configuration > + * @filt_einfo: Filter extended packet info > + * @filt_pswords: Filter PS words present > + * @dma_keystone_tx_priority: Tx channel scheduling priority > + */ > +struct dma_keystone_tx_cfg { > + bool filt_einfo; > + bool filt_pswords; > + enum dma_keystone_tx_priority priority; > +}; > + > +/** > + * struct dma_keystone_rx_cfg: Rx flow configuration > + * @einfo_present: Extended packet info present > + * @psinfo_present: PS words present > + * @dma_keystone_rx_err_mode: Error during buffer starvation > + * @dma_keystone_desc_type: Host or Monolithic desc > + * @psinfo_at_sop: PS word located at start of packet > + * @sop_offset: Start of packet offset > + * @dst_q: Destination queue for a given flow > + * @thresh: Rx flow size threshold > + * @fdq[]: Free desc Queue array > + * @sz_thresh0: RX packet size threshold 0 > + * @sz_thresh1: RX packet size threshold 1 > + * @sz_thresh2: RX packet size threshold 2 > + */ > +struct dma_keystone_rx_cfg { > + bool einfo_present; > + bool psinfo_present; > + enum dma_keystone_rx_err_mode err_mode; > + enum dma_keystone_desc_type desc_type; > + bool psinfo_at_sop; > + unsigned int sop_offset; > + unsigned int dst_q; > + enum dma_keystone_rx_thresholds thresh; > + unsigned int fdq[DMA_KEYSTONE_FDQ_PER_CHAN]; > + unsigned int sz_thresh0; > + unsigned int sz_thresh1; > + unsigned int sz_thresh2; > +}; > + > +/** > + * struct dma_keystone_cfg: Pktdma channel configuration > + * @sl_cfg: Slave configuration > + * @tx: Tx channel configuration > + * @rx: Rx flow configuration > + */ > +struct dma_keystone_cfg { > + struct dma_slave_config sl_cfg; > + union { > + struct dma_keystone_tx_cfg tx; > + struct dma_keystone_rx_cfg rx; > + } u; > +}; > + > +/** > + * struct dma_keystone_desc: Host packet descriptor layout > + * @desc_info: Descriptor information like id, type, length > + * @tag_info: Flow tag info written in during RX > + * @packet_info: Queue Manager, policy, flags etc > + * @buff_len: Buffer length in bytes > + * @buff: Buffer pointer > + * @next_desc: For chaining the descriptors > + * @orig_len: length since 'buff_len' can be overwritten > + * @orig_buff: buff pointer since 'buff' can be overwritten > + * @epib: Extended packet info block > + * @psdata: Protocol specific > + */ > +struct dma_keystone_desc { > + u32 desc_info; > + u32 tag_info; > + u32 packet_info; > + u32 buff_len; > + u32 buff; > + u32 next_desc; > + u32 orig_len; > + u32 orig_buff; > + u32 epib[DMA_KEYSTONE_NUM_EPIB_WORDS]; > + u32 psdata[DMA_KEYSTONE_NUM_PS_WORDS]; > + u32 pad[4]; > +} ____cacheline_aligned; > + > +#define keystone_cfg_to_slave_config(keystone_cfg) (&keystone_cfg.sl_cfg) > +#define keystone_cfg_from_slave_config(dma_cfg) container_of(dma_cfg, \ > + struct dma_keystone_cfg, sl_cfg) > +/* Keystone PKTDMA filter function */ > +bool dma_keystone_filter_fn(struct dma_chan *chan, void *param); > + > +#endif /* __KEYSTONE_PKTDMA_H__ */
On Saturday 01 March 2014 07:44 AM, Santosh Shilimkar wrote: > On Friday 28 February 2014 06:14 PM, Arnd Bergmann wrote: >> On Friday 28 February 2014 17:56:40 Santosh Shilimkar wrote: >>> The Packet DMA driver sets up the dma channels and flows for the >>> QMSS(Queue Manager SubSystem) who triggers the actual data movements >>> across clients using destination queues. Every client modules like >>> NETCP(Network Coprocessor), SRIO(Serial Rapid IO) and CRYPTO >>> Engines has its own instance of packet dma hardware. QMSS has also >>> an internal packet DMA module which is used as an infrastructure >>> DMA with zero copy. >>> >>> Patch adds DMAEngine driver for Keystone Packet DMA hardware. >>> The specifics on the device tree bindings for Packet DMA can be >>> found in: >>> Documentation/devicetree/bindings/dma/keystone-pktdma.txt >>> >>> The driver implements the configuration functions using standard DMAEngine >>> apis. The data movement is managed by QMSS device driver. >> >> The description of this sounds a bit like the x-gene queue manager/tag >> manager, rather than a traditional DMA engine. For that driver, we >> decided to use the mailbox subsystem instead of the DMA subsystem. >> >> Can you explain what kind of data is actually getting transferred >> in by your hardware here? You say the DMA is performed by the QMSS, >> so do you use the pktdma driver to drive the data transfers of the >> QMSS or do you just use it for flow control by passing short (a few >> bytes) messages and also have to program the pktdma? >> > I don't know what x-gene stuff is, but the keystone packet dma > is hardware DMA controller(per client subystem) which has all > the traditional DMA controller properties. The packet DMA needs > to be setup just like any slave dma configuration. The actual > data transfers are triggered by operations on Queue Manager > SubSystem. > > I just posted the QMSS driver as well which explains how the system > work. > Updating the thread for record since me and Arnd had some more chat at the Linaro Connect to extend the above description. Will try to add some more information with updated patchset after aligning with DMAEngine folks. >>> + >>> +bool dma_keystone_filter_fn(struct dma_chan *chan, void *param) >>> +{ >>> + if (chan->device->dev->driver == &keystone_dma_driver.driver) { >>> + struct keystone_dma_chan *kdma_chan = from_achan(chan); >>> + unsigned chan_id = *(u32 *)param & KEYSTONE_DMA_CHAN_ID_MASK; >>> + unsigned chan_type = *(u32 *)param >> KEYSTONE_DMA_TYPE_SHIFT; >>> + >>> + if (chan_type == KEYSTONE_DMA_TX_CHAN_TYPE && >>> + kdma_chan->direction == DMA_MEM_TO_DEV) >>> + return chan_id == kdma_chan->channel; >>> + >>> + if (chan_type == KEYSTONE_DMA_RX_FLOW_TYPE && >>> + kdma_chan->direction == DMA_DEV_TO_MEM) >>> + return chan_id == kdma_chan->flow; >>> + } >>> + return false; >>> +} >>> +EXPORT_SYMBOL_GPL(dma_keystone_filter_fn); >> >> The filter function should be removed here and replaced with a simpler >> custom xlate function calling dma_get_slave_chan() on the matching >> channel. >> > Can you please expand your comment please ? I though the filter function > is the one should implement for specific channel allocations. > I got this clarified thanks to Arnd. We can get rid of the filter function and use the xlate funtion. >>> diff --git a/include/linux/keystone-pktdma.h b/include/linux/keystone-pktdma.h >>> new file mode 100644 >>> index 0000000..e6a66c8 >>> --- /dev/null >>> +++ b/include/linux/keystone-pktdma.h >>> @@ -0,0 +1,168 @@ >> >> A DMA engine driver should not have a public API. Please move all the >> contents of the two header files into the driver itself to ensure none >> of this is visible to slave drivers. If it turns out that you use >> declarations from this header in slave drivers at the moment, please >> explain what they are so we can discuss alternative solutions. >> > Yes the slave drivers make use of these. The dt-binding header is created > because some the macro's are shared between DTS and code. > > keystone-pktdma.h carriers information about the packet dma config > which is done by slave drivers. It also contains the dma descriptor > format and its associate macro's used by slave drivers. Since > the hardware perspective, the descriptor represent the packet > format dma expects and hence its left in the dma header header. > Also i will remove the descriptor related fields in next version. Regards, Santosh
On Fri, Feb 28, 2014 at 05:56:40PM -0500, Santosh Shilimkar wrote: > From: Sandeep Nair <sandeep_n@ti.com> > > The Packet DMA driver sets up the dma channels and flows for the > QMSS(Queue Manager SubSystem) who triggers the actual data movements > across clients using destination queues. Every client modules like > NETCP(Network Coprocessor), SRIO(Serial Rapid IO) and CRYPTO > Engines has its own instance of packet dma hardware. QMSS has also > an internal packet DMA module which is used as an infrastructure > DMA with zero copy. > > Patch adds DMAEngine driver for Keystone Packet DMA hardware. > The specifics on the device tree bindings for Packet DMA can be > found in: > Documentation/devicetree/bindings/dma/keystone-pktdma.txt > > The driver implements the configuration functions using standard DMAEngine > apis. The data movement is managed by QMSS device driver. Pls use subsystem appropriate name so here would have been dmaengine: ... So i am still missing stuff like prepare calls, irq, descriptor management to call this a dmaengine driver. I guess you need to explain a bit more why the data movement is handled by some other driver and not by this one few more comments inline > > Cc: Vinod Koul <vinod.koul@intel.com> > Cc: Russell King <linux@arm.linux.org.uk> > Cc: Grant Likely <grant.likely@linaro.org> > Cc: Rob Herring <robh+dt@kernel.org> > Cc: Mark Rutland <mark.rutland@arm.com> > Signed-off-by: Sandeep Nair <sandeep_n@ti.com> > Signed-off-by: Santosh Shilimkar <santosh.shilimkar@ti.com> > --- > .../devicetree/bindings/dma/keystone-pktdma.txt | 72 ++ > drivers/dma/Kconfig | 8 + > drivers/dma/Makefile | 1 + > drivers/dma/keystone-pktdma.c | 795 ++++++++++++++++++++ > include/dt-bindings/dma/keystone.h | 33 + > include/linux/keystone-pktdma.h | 168 +++++ > 6 files changed, 1077 insertions(+) > create mode 100644 Documentation/devicetree/bindings/dma/keystone-pktdma.txt > create mode 100644 drivers/dma/keystone-pktdma.c > create mode 100644 include/dt-bindings/dma/keystone.h > create mode 100644 include/linux/keystone-pktdma.h > > diff --git a/Documentation/devicetree/bindings/dma/keystone-pktdma.txt b/Documentation/devicetree/bindings/dma/keystone-pktdma.txt > new file mode 100644 > index 0000000..ea061d5 > --- /dev/null > +++ b/Documentation/devicetree/bindings/dma/keystone-pktdma.txt > @@ -0,0 +1,72 @@ > +Keystone Packet DMA Controller > + > +This document explains the device tree bindings for the packet dma > +on keystone devices. The the Network coprocessor, Cypto engines > +and the SRIO on Keystone devices all have their own packet dma modules. > +Each individual packet dma has a certain number of RX channels, > +RX flows and TX channels. Each instance of the packet DMA is being > +initialized through device specific bindings. > + > +* DMA controller > + > +Required properties: > + - compatible: Should be "ti,keystone-pktdma" > + - reg: Should contain register location and length of the following pktdma > + register regions. The value for "reg-names" property of the respective > + region is specified in parenthesis. > + - Global control register region (global). > + - Tx DMA channel configuration register region (txchan). > + - Rx DMA channel configuration register region (rxchan). > + - Tx DMA channel Scheduler configuration register region (txsched). > + - Rx DMA flow configuration register region (rxflow). > + - reg-names: Names for the above regions. The name to be used is specified in > + the above description of "reg" property. > + - qm-base-address: Base address of the logical queue managers for pktdma. > + - #dma-cells: Has to be 1. Keystone-pktdma doesn't support anything else. > + > +Optional properties: > + - enable-all: Enable all DMA channels. > + - loop-back: To loopback Tx streaming I/F to Rx streaming I/F. Used for > + infrastructure transfers. > + - rx-retry-timeout: Number of pktdma cycles to wait before retry on buffer > + starvation. > + > +Example: > + netcp-dma: pktdma@2004000 { > + compatible = "ti,keystone-pktdma"; > + reg = <0x2004000 0x100>, > + <0x2004400 0x120>, > + <0x2004800 0x300>, > + <0x2004c00 0x120>, > + <0x2005000 0x400>; > + reg-names = "global", "txchan", "rxchan", "txsched", > + "rxflow"; > + qm-base-address = <0x23a80000 0x23a90000 > + 0x23aa0000 0x23ab0000>; > + #dma-cells = <1>; > + /* enable-all; */ > + rx-retry-timeout = <3500>; > + /* loop-back; */ > + }; > + > + > +* DMA client > + > +Required properties: > +- dmas: One DMA request specifier consisting of a phandle to the DMA controller > + followed by the integer specifying the channel identifier. The channel > + identifier is encoded as follows: > + - bits 7-0: Tx DMA channel number or the Rx flow number. > + - bits 31-24: Channel type. 0xff for Tx DMA channel & 0xfe for Rx flow. > +- dma-names: List of string identifiers for the DMA requests. > + > +Example: > + > + netcp: netcp@2090000 { > + ... > + dmas = <&netcpdma KEYSTONE_DMA_RX_FLOW(22)>, > + <&netcpdma KEYSTONE_DMA_RX_FLOW(23)>, > + <&netcpdma KEYSTONE_DMA_TX_CHAN(8)>; > + dma-names = "netrx0", "netrx1", "nettx"; > + ... > + }; Can you pls separate the binding to separate patch and also this needs ack from DT folks > diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig > index 9bed1a2..722b99a 100644 > --- a/drivers/dma/Kconfig > +++ b/drivers/dma/Kconfig > @@ -350,6 +350,14 @@ config MOXART_DMA > help > Enable support for the MOXA ART SoC DMA controller. > > +config KEYSTONE_PKTDMA > + tristate "TI Keystone Packet DMA support" > + depends on ARCH_KEYSTONE > + select DMA_ENGINE > + help > + Enable support for the Packet DMA engine on Texas Instruments' > + Keystone family of devices. > + > config DMA_ENGINE > bool > > diff --git a/drivers/dma/Makefile b/drivers/dma/Makefile > index a029d0f4..6d69c6d 100644 > --- a/drivers/dma/Makefile > +++ b/drivers/dma/Makefile > @@ -44,3 +44,4 @@ obj-$(CONFIG_DMA_JZ4740) += dma-jz4740.o > obj-$(CONFIG_TI_CPPI41) += cppi41.o > obj-$(CONFIG_K3_DMA) += k3dma.o > obj-$(CONFIG_MOXART_DMA) += moxart-dma.o > +obj-$(CONFIG_KEYSTONE_PKTDMA) += keystone-pktdma.o > diff --git a/drivers/dma/keystone-pktdma.c b/drivers/dma/keystone-pktdma.c > new file mode 100644 > index 0000000..b3f77e5 > --- /dev/null > +++ b/drivers/dma/keystone-pktdma.c > @@ -0,0 +1,795 @@ > +/* > + * Copyright (C) 2014 Texas Instruments Incorporated > + * Authors: Sandeep Nair <sandeep_n@ti.com> > + * Cyril Chemparathy <cyril@ti.com> > + * Santosh Shilimkar <santosh.shilimkar@ti.com> > + * > + * This program is free software; you can redistribute it and/or > + * modify it under the terms of the GNU General Public License as > + * published by the Free Software Foundation version 2. > + * > + * This program is distributed "as is" WITHOUT ANY WARRANTY of any > + * kind, whether express or implied; without even the implied warranty > + * of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > + * GNU General Public License for more details. > + */ > + > +#include <linux/io.h> > +#include <linux/sched.h> > +#include <linux/module.h> > +#include <linux/dma-direction.h> > +#include <linux/dmaengine.h> > +#include <linux/interrupt.h> > +#include <linux/of_dma.h> > +#include <linux/of_address.h> > +#include <linux/platform_device.h> > +#include <linux/keystone-pktdma.h> > +#include <linux/pm_runtime.h> > +#include <dt-bindings/dma/keystone.h> > + > +#define BITS(x) (BIT(x) - 1) this might get confusing, perhaps a better name could be given? > +#define REG_MASK 0xffffffff > + > +#define DMA_LOOPBACK BIT(31) > +#define DMA_ENABLE BIT(31) > +#define DMA_TEARDOWN BIT(30) > + > +#define DMA_TX_FILT_PSWORDS BIT(29) > +#define DMA_TX_FILT_EINFO BIT(30) > +#define DMA_TX_PRIO_SHIFT 0 > +#define DMA_RX_PRIO_SHIFT 16 > +#define DMA_PRIO_MASK BITS(3) > +#define DMA_PRIO_DEFAULT 0 > +#define DMA_RX_TIMEOUT_DEFAULT 17500 /* cycles */ > +#define DMA_RX_TIMEOUT_MASK BITS(16) > +#define DMA_RX_TIMEOUT_SHIFT 0 > + > +#define CHAN_HAS_EPIB BIT(30) > +#define CHAN_HAS_PSINFO BIT(29) > +#define CHAN_ERR_RETRY BIT(28) > +#define CHAN_PSINFO_AT_SOP BIT(25) > +#define CHAN_SOP_OFF_SHIFT 16 > +#define CHAN_SOP_OFF_MASK BITS(9) > +#define DESC_TYPE_SHIFT 26 > +#define DESC_TYPE_MASK BITS(2) > + > +/* > + * QMGR & QNUM together make up 14 bits with QMGR as the 2 MSb's in the logical > + * navigator cloud mapping scheme. > + * using the 14bit physical queue numbers directly maps into this scheme. > + */ > +#define CHAN_QNUM_MASK BITS(14) > +#define DMA_MAX_QMS 4 > +#define DMA_TIMEOUT 1000 /* msecs */ > + > +struct reg_global { > + u32 revision; > + u32 perf_control; > + u32 emulation_control; > + u32 priority_control; > + u32 qm_base_address[4]; > +}; > + > +struct reg_chan { > + u32 control; > + u32 mode; > + u32 __rsvd[6]; > +}; > + > +struct reg_tx_sched { > + u32 prio; > +}; > + > +struct reg_rx_flow { > + u32 control; > + u32 tags; > + u32 tag_sel; > + u32 fdq_sel[2]; > + u32 thresh[3]; > +}; > + > +#define BUILD_CHECK_REGS() \ > + do { \ > + BUILD_BUG_ON(sizeof(struct reg_global) != 32); \ > + BUILD_BUG_ON(sizeof(struct reg_chan) != 32); \ > + BUILD_BUG_ON(sizeof(struct reg_rx_flow) != 32); \ > + BUILD_BUG_ON(sizeof(struct reg_tx_sched) != 4); \ > + } while (0) why is this required, do you want to use __packed__ to ensure right size? > + > +enum keystone_chan_state { > + /* stable states */ > + CHAN_STATE_OPENED, > + CHAN_STATE_CLOSED, > +}; > + > +struct keystone_dma_device { > + struct dma_device engine; > + bool loopback, enable_all; > + unsigned tx_priority, rx_priority, rx_timeout; > + unsigned logical_queue_managers; > + unsigned qm_base_address[DMA_MAX_QMS]; > + struct reg_global __iomem *reg_global; > + struct reg_chan __iomem *reg_tx_chan; > + struct reg_rx_flow __iomem *reg_rx_flow; > + struct reg_chan __iomem *reg_rx_chan; > + struct reg_tx_sched __iomem *reg_tx_sched; > + unsigned max_rx_chan, max_tx_chan; > + unsigned max_rx_flow; > + atomic_t in_use; > +}; > +#define to_dma(dma) (&(dma)->engine) > +#define dma_dev(dma) ((dma)->engine.dev) > + > +struct keystone_dma_chan { > + struct dma_chan achan; > + enum dma_transfer_direction direction; > + atomic_t state; > + struct keystone_dma_device *dma; > + > + /* registers */ > + struct reg_chan __iomem *reg_chan; > + struct reg_tx_sched __iomem *reg_tx_sched; > + struct reg_rx_flow __iomem *reg_rx_flow; > + > + /* configuration stuff */ > + unsigned channel, flow; > +}; > +#define from_achan(ch) container_of(ch, struct keystone_dma_chan, achan) > +#define to_achan(ch) (&(ch)->achan) > +#define chan_dev(ch) (&to_achan(ch)->dev->device) > +#define chan_num(ch) ((ch->direction == DMA_MEM_TO_DEV) ? \ > + ch->channel : ch->flow) > +#define chan_vdbg(ch, format, arg...) \ > + dev_vdbg(chan_dev(ch), format, ##arg); > + > +/** > + * dev_to_dma_chan - convert a device pointer to the its sysfs container object > + * @dev - device node > + */ > +static inline struct dma_chan *dev_to_dma_chan(struct device *dev) > +{ > + struct dma_chan_dev *chan_dev; > + > + chan_dev = container_of(dev, typeof(*chan_dev), device); > + return chan_dev->chan; > +} > + > +static inline enum keystone_chan_state > +chan_get_state(struct keystone_dma_chan *chan) > +{ > + return atomic_read(&chan->state); > +} > + > +static int chan_start(struct keystone_dma_chan *chan, > + struct dma_keystone_cfg *cfg) > +{ > + u32 v = 0; > + > + if ((chan->direction == DMA_MEM_TO_DEV) && chan->reg_chan) { why second check? > + if (cfg->u.tx.filt_pswords) > + v |= DMA_TX_FILT_PSWORDS; > + if (cfg->u.tx.filt_einfo) > + v |= DMA_TX_FILT_EINFO; > + writel_relaxed(v, &chan->reg_chan->mode); > + writel_relaxed(DMA_ENABLE, &chan->reg_chan->control); > + } no else for DMA_DEV_TO_MEM? > + > + if (chan->reg_tx_sched) > + writel_relaxed(cfg->u.tx.priority, &chan->reg_tx_sched->prio); > + > + if (chan->reg_rx_flow) { > + v = 0; > + > + if (cfg->u.rx.einfo_present) > + v |= CHAN_HAS_EPIB; > + if (cfg->u.rx.psinfo_present) > + v |= CHAN_HAS_PSINFO; > + if (cfg->u.rx.err_mode == DMA_RETRY) > + v |= CHAN_ERR_RETRY; > + v |= (cfg->u.rx.desc_type & DESC_TYPE_MASK) << DESC_TYPE_SHIFT; > + if (cfg->u.rx.psinfo_at_sop) > + v |= CHAN_PSINFO_AT_SOP; > + v |= (cfg->u.rx.sop_offset & CHAN_SOP_OFF_MASK) > + << CHAN_SOP_OFF_SHIFT; > + v |= cfg->u.rx.dst_q & CHAN_QNUM_MASK; > + > + writel_relaxed(v, &chan->reg_rx_flow->control); > + writel_relaxed(0, &chan->reg_rx_flow->tags); > + writel_relaxed(0, &chan->reg_rx_flow->tag_sel); > + > + v = cfg->u.rx.fdq[0] << 16; > + v |= cfg->u.rx.fdq[1] & CHAN_QNUM_MASK; > + writel_relaxed(v, &chan->reg_rx_flow->fdq_sel[0]); > + > + v = cfg->u.rx.fdq[2] << 16; > + v |= cfg->u.rx.fdq[3] & CHAN_QNUM_MASK; > + writel_relaxed(v, &chan->reg_rx_flow->fdq_sel[1]); > + > + writel_relaxed(0, &chan->reg_rx_flow->thresh[0]); > + writel_relaxed(0, &chan->reg_rx_flow->thresh[1]); > + writel_relaxed(0, &chan->reg_rx_flow->thresh[2]); > + } > + > + return 0; > +} > + > +static int chan_teardown(struct keystone_dma_chan *chan) > +{ > + unsigned long end, value; > + > + if (!chan->reg_chan) > + return 0; > + > + /* indicate teardown */ > + writel_relaxed(DMA_TEARDOWN, &chan->reg_chan->control); > + > + /* wait for the dma to shut itself down */ > + end = jiffies + msecs_to_jiffies(DMA_TIMEOUT); > + do { > + value = readl_relaxed(&chan->reg_chan->control); > + if ((value & DMA_ENABLE) == 0) > + break; > + } while (time_after(end, jiffies)); > + > + if (readl_relaxed(&chan->reg_chan->control) & DMA_ENABLE) { > + dev_err(chan_dev(chan), "timeout waiting for teardown\n"); > + return -ETIMEDOUT; > + } > + > + return 0; > +} > + > +static void chan_stop(struct keystone_dma_chan *chan) > +{ > + if (chan->reg_rx_flow) { > + /* first detach fdqs, starve out the flow */ > + writel_relaxed(0, &chan->reg_rx_flow->fdq_sel[0]); > + writel_relaxed(0, &chan->reg_rx_flow->fdq_sel[1]); > + writel_relaxed(0, &chan->reg_rx_flow->thresh[0]); > + writel_relaxed(0, &chan->reg_rx_flow->thresh[1]); > + writel_relaxed(0, &chan->reg_rx_flow->thresh[2]); > + } > + > + /* teardown the dma channel */ > + chan_teardown(chan); > + > + /* then disconnect the completion side */ > + if (chan->reg_rx_flow) { > + writel_relaxed(0, &chan->reg_rx_flow->control); > + writel_relaxed(0, &chan->reg_rx_flow->tags); > + writel_relaxed(0, &chan->reg_rx_flow->tag_sel); > + } > + chan_vdbg(chan, "channel stopped\n"); > +} > + > +static void keystone_dma_hw_init(struct keystone_dma_device *dma) > +{ > + unsigned v; > + int i; > + > + v = dma->loopback ? DMA_LOOPBACK : 0; > + writel_relaxed(v, &dma->reg_global->emulation_control); > + > + v = readl_relaxed(&dma->reg_global->perf_control); > + v |= ((dma->rx_timeout & DMA_RX_TIMEOUT_MASK) << DMA_RX_TIMEOUT_SHIFT); > + writel_relaxed(v, &dma->reg_global->perf_control); > + > + v = ((dma->tx_priority << DMA_TX_PRIO_SHIFT) | > + (dma->rx_priority << DMA_RX_PRIO_SHIFT)); > + > + writel_relaxed(v, &dma->reg_global->priority_control); > + > + if (dma->enable_all) { > + for (i = 0; i < dma->max_tx_chan; i++) { > + writel_relaxed(0, &dma->reg_tx_chan[i].mode); > + writel_relaxed(DMA_ENABLE, > + &dma->reg_tx_chan[i].control); > + } > + } > + > + /* Always enable all Rx channels. Rx paths are managed using flows */ > + for (i = 0; i < dma->max_rx_chan; i++) > + writel_relaxed(DMA_ENABLE, &dma->reg_rx_chan[i].control); > + > + for (i = 0; i < dma->logical_queue_managers; i++) > + writel_relaxed(dma->qm_base_address[i], > + &dma->reg_global->qm_base_address[i]); > +} > + > +static void keystone_dma_hw_destroy(struct keystone_dma_device *dma) > +{ > + int i; > + unsigned v; > + > + v = ~DMA_ENABLE & REG_MASK; > + > + for (i = 0; i < dma->max_rx_chan; i++) > + writel_relaxed(v, &dma->reg_rx_chan[i].control); > + > + for (i = 0; i < dma->max_tx_chan; i++) > + writel_relaxed(v, &dma->reg_tx_chan[i].control); > +} > + > +static int chan_init(struct dma_chan *achan) > +{ > + struct keystone_dma_chan *chan = from_achan(achan); > + struct keystone_dma_device *dma = chan->dma; > + > + chan_vdbg(chan, "initializing %s channel\n", > + chan->direction == DMA_MEM_TO_DEV ? "transmit" : > + chan->direction == DMA_DEV_TO_MEM ? "receive" : > + "unknown"); > + > + if (chan->direction != DMA_MEM_TO_DEV && > + chan->direction != DMA_DEV_TO_MEM) { > + dev_err(chan_dev(chan), "bad direction\n"); > + return -EINVAL; > + } is_slave_direction() pls > + > + atomic_set(&chan->state, CHAN_STATE_OPENED); > + > + if (atomic_inc_return(&dma->in_use) <= 1) > + keystone_dma_hw_init(dma); > + > + return 0; > +} > + > +static void chan_destroy(struct dma_chan *achan) > +{ > + struct keystone_dma_chan *chan = from_achan(achan); > + struct keystone_dma_device *dma = chan->dma; > + > + if (chan_get_state(chan) == CHAN_STATE_CLOSED) > + return; > + > + chan_vdbg(chan, "destroying channel\n"); > + chan_stop(chan); > + atomic_set(&chan->state, CHAN_STATE_CLOSED); > + if (atomic_dec_return(&dma->in_use) <= 0) > + keystone_dma_hw_destroy(dma); > + chan_vdbg(chan, "channel destroyed\n"); > +} > + > +static int chan_keystone_config(struct keystone_dma_chan *chan, > + struct dma_keystone_cfg *cfg) > +{ > + if (chan_get_state(chan) != CHAN_STATE_OPENED) > + return -ENODEV; > + > + if (cfg->sl_cfg.direction != chan->direction) > + return -EINVAL; direction is deprecated... > + > + return chan_start(chan, cfg); > +} > + > +static int chan_control(struct dma_chan *achan, enum dma_ctrl_cmd cmd, > + unsigned long arg) > +{ > + struct keystone_dma_chan *chan = from_achan(achan); > + struct dma_keystone_cfg *keystone_config; > + struct dma_slave_config *dma_cfg; > + int ret; > + > + switch (cmd) { > + case DMA_SLAVE_CONFIG: > + dma_cfg = (struct dma_slave_config *)arg; > + keystone_config = keystone_cfg_from_slave_config(dma_cfg); > + ret = chan_keystone_config(chan, keystone_config); > + break; > + > + default: > + ret = -ENOTSUPP; > + break; > + } > + return ret; > +} > + > +static void __iomem *pktdma_get_regs( > + struct keystone_dma_device *dma, const char *name, > + resource_size_t *_size) > +{ > + struct device *dev = dma_dev(dma); > + struct device_node *node = dev->of_node; > + resource_size_t size; > + struct resource res; > + void __iomem *regs; > + int i; > + > + i = of_property_match_string(node, "reg-names", name); > + if (of_address_to_resource(node, i, &res)) { > + dev_err(dev, "could not find %s resource(index %d)\n", name, i); > + return NULL; > + } > + size = resource_size(&res); > + > + regs = of_iomap(node, i); > + if (!regs) { > + dev_err(dev, "could not map %s resource (index %d)\n", name, i); > + return NULL; > + } > + > + dev_dbg(dev, "index: %d, res:%s, size:%x, phys:%x, virt:%p\n", > + i, name, (unsigned int)size, (unsigned int)res.start, regs); > + > + if (_size) > + *_size = size; > + > + return regs; > +} > + > +static int pktdma_init_rx_chan(struct keystone_dma_chan *chan, > + struct device_node *node, > + u32 flow) > +{ > + struct keystone_dma_device *dma = chan->dma; > + struct device *dev = dma_dev(chan->dma); > + > + chan->flow = flow; > + chan->reg_rx_flow = dma->reg_rx_flow + flow; > + dev_dbg(dev, "rx flow(%d) (%p)\n", chan->flow, chan->reg_rx_flow); > + > + return 0; > +} > + > +static int pktdma_init_tx_chan(struct keystone_dma_chan *chan, > + struct device_node *node, > + u32 channel) > +{ > + struct keystone_dma_device *dma = chan->dma; > + struct device *dev = dma_dev(chan->dma); > + > + chan->channel = channel; > + chan->reg_chan = dma->reg_tx_chan + channel; > + chan->reg_tx_sched = dma->reg_tx_sched + channel; > + dev_dbg(dev, "tx channel(%d) (%p)\n", chan->channel, chan->reg_chan); > + > + return 0; > +} > + > +static int pktdma_init_chan(struct keystone_dma_device *dma, > + struct device_node *node, > + enum dma_transfer_direction dir, > + unsigned chan_num) > +{ > + struct device *dev = dma_dev(dma); > + struct keystone_dma_chan *chan; > + struct dma_chan *achan; > + int ret = -EINVAL; > + > + chan = devm_kzalloc(dev, sizeof(*chan), GFP_KERNEL); > + if (!chan) > + return -ENOMEM; > + > + achan = to_achan(chan); > + achan->device = &dma->engine; > + chan->dma = dma; > + chan->direction = DMA_NONE; > + atomic_set(&chan->state, CHAN_STATE_OPENED); > + > + if (dir == DMA_MEM_TO_DEV) { > + chan->direction = dir; > + ret = pktdma_init_tx_chan(chan, node, chan_num); > + } else if (dir == DMA_DEV_TO_MEM) { > + chan->direction = dir; > + ret = pktdma_init_rx_chan(chan, node, chan_num); > + } else { > + dev_err(dev, "channel(%d) direction unknown\n", chan_num); > + } > + > + if (ret < 0) > + goto fail; > + > + list_add_tail(&to_achan(chan)->device_node, &to_dma(dma)->channels); > + return 0; > + > +fail: > + devm_kfree(dev, chan); ?? > + return ret; > +} > + > +/* dummy function: feature not supported */ > +static enum dma_status chan_xfer_status(struct dma_chan *achan, > + dma_cookie_t cookie, > + struct dma_tx_state *txstate) > +{ > + WARN(1, "xfer status not supported\n"); > + return DMA_ERROR; > +} > + > +/* dummy function: feature not supported */ > +static void chan_issue_pending(struct dma_chan *chan) > +{ > + WARN(1, "issue pending not supported\n"); > +} Supporting status is okay but not issue_pending. This breaks use of dmaengine API. What we expect is that user will do channel allocation, prepare a descriptor, then submit it. Submit doesnt start the transaction, this call does! So we need implementation here! > + > +static ssize_t keystone_dma_show_chan_num(struct device *dev, > + struct device_attribute *attr, char *buf) > +{ > + struct dma_chan *achan = dev_to_dma_chan(dev); > + struct keystone_dma_chan *chan = from_achan(achan); > + > + return scnprintf(buf, PAGE_SIZE, "%u\n", chan->channel); > +} ??? > + > +static ssize_t keystone_dma_show_flow(struct device *dev, > + struct device_attribute *attr, char *buf) > +{ > + struct dma_chan *achan = dev_to_dma_chan(dev); > + struct keystone_dma_chan *chan = from_achan(achan); > + > + return scnprintf(buf, PAGE_SIZE, "%u\n", chan->flow); > +} > + > +static DEVICE_ATTR(chan_num, S_IRUSR, keystone_dma_show_chan_num, NULL); > +static DEVICE_ATTR(rx_flow, S_IRUSR, keystone_dma_show_flow, NULL); okay why do we need these?
On Tuesday 11 March 2014 06:23 PM, Vinod Koul wrote: > On Fri, Feb 28, 2014 at 05:56:40PM -0500, Santosh Shilimkar wrote: >> From: Sandeep Nair <sandeep_n@ti.com> >> >> The Packet DMA driver sets up the dma channels and flows for the >> QMSS(Queue Manager SubSystem) who triggers the actual data movements >> across clients using destination queues. Every client modules like >> NETCP(Network Coprocessor), SRIO(Serial Rapid IO) and CRYPTO >> Engines has its own instance of packet dma hardware. QMSS has also >> an internal packet DMA module which is used as an infrastructure >> DMA with zero copy. >> >> Patch adds DMAEngine driver for Keystone Packet DMA hardware. >> The specifics on the device tree bindings for Packet DMA can be >> found in: >> Documentation/devicetree/bindings/dma/keystone-pktdma.txt >> >> The driver implements the configuration functions using standard DMAEngine >> apis. The data movement is managed by QMSS device driver. > Pls use subsystem appropriate name so here would have been dmaengine: ... > > So i am still missing stuff like prepare calls, irq, descriptor management to > call this a dmaengine driver. > > I guess you need to explain a bit more why the data movement is handled by some > other driver and not by this one > To expand above statement, Packet DMA hardware blocks on Keystone SOCs are DMAEngines. QMSS is centralised subsystem manages multiple functionalities including triggering dma transfers, descriptor management and completion irqs. There are separate instance of packet DMA hardware block per client device. We program the DMA hardware to allocate channels and flows. So the packet DMA resouces like dma channels and dma flows are configured and managed through the DMAEngine driver. Thats why we implement only device_alloc_chan_resources, device_free_chan_resources and device_control DMAEngine APIs. > few more comments inline > >> >> Cc: Vinod Koul <vinod.koul@intel.com> >> Cc: Russell King <linux@arm.linux.org.uk> >> Cc: Grant Likely <grant.likely@linaro.org> >> Cc: Rob Herring <robh+dt@kernel.org> >> Cc: Mark Rutland <mark.rutland@arm.com> >> Signed-off-by: Sandeep Nair <sandeep_n@ti.com> >> Signed-off-by: Santosh Shilimkar <santosh.shilimkar@ti.com> >> --- >> .../devicetree/bindings/dma/keystone-pktdma.txt | 72 ++ >> drivers/dma/Kconfig | 8 + >> drivers/dma/Makefile | 1 + >> drivers/dma/keystone-pktdma.c | 795 ++++++++++++++++++++ >> include/dt-bindings/dma/keystone.h | 33 + >> include/linux/keystone-pktdma.h | 168 +++++ >> 6 files changed, 1077 insertions(+) >> create mode 100644 Documentation/devicetree/bindings/dma/keystone-pktdma.txt >> create mode 100644 drivers/dma/keystone-pktdma.c >> create mode 100644 include/dt-bindings/dma/keystone.h >> create mode 100644 include/linux/keystone-pktdma.h >> >> diff --git a/Documentation/devicetree/bindings/dma/keystone-pktdma.txt b/Documentation/devicetree/bindings/dma/keystone-pktdma.txt >> new file mode 100644 >> index 0000000..ea061d5 >> --- /dev/null >> +++ b/Documentation/devicetree/bindings/dma/keystone-pktdma.txt >> @@ -0,0 +1,72 @@ >> +Keystone Packet DMA Controller >> + >> +This document explains the device tree bindings for the packet dma >> +on keystone devices. The the Network coprocessor, Cypto engines >> +and the SRIO on Keystone devices all have their own packet dma modules. >> +Each individual packet dma has a certain number of RX channels, >> +RX flows and TX channels. Each instance of the packet DMA is being >> +initialized through device specific bindings. >> + >> +* DMA controller >> + >> +Required properties: >> + - compatible: Should be "ti,keystone-pktdma" >> + - reg: Should contain register location and length of the following pktdma >> + register regions. The value for "reg-names" property of the respective >> + region is specified in parenthesis. >> + - Global control register region (global). >> + - Tx DMA channel configuration register region (txchan). >> + - Rx DMA channel configuration register region (rxchan). >> + - Tx DMA channel Scheduler configuration register region (txsched). >> + - Rx DMA flow configuration register region (rxflow). >> + - reg-names: Names for the above regions. The name to be used is specified in >> + the above description of "reg" property. >> + - qm-base-address: Base address of the logical queue managers for pktdma. >> + - #dma-cells: Has to be 1. Keystone-pktdma doesn't support anything else. >> + >> +Optional properties: >> + - enable-all: Enable all DMA channels. >> + - loop-back: To loopback Tx streaming I/F to Rx streaming I/F. Used for >> + infrastructure transfers. >> + - rx-retry-timeout: Number of pktdma cycles to wait before retry on buffer >> + starvation. >> + >> +Example: >> + netcp-dma: pktdma@2004000 { >> + compatible = "ti,keystone-pktdma"; >> + reg = <0x2004000 0x100>, >> + <0x2004400 0x120>, >> + <0x2004800 0x300>, >> + <0x2004c00 0x120>, >> + <0x2005000 0x400>; >> + reg-names = "global", "txchan", "rxchan", "txsched", >> + "rxflow"; >> + qm-base-address = <0x23a80000 0x23a90000 >> + 0x23aa0000 0x23ab0000>; >> + #dma-cells = <1>; >> + /* enable-all; */ >> + rx-retry-timeout = <3500>; >> + /* loop-back; */ >> + }; >> + >> + >> +* DMA client >> + >> +Required properties: >> +- dmas: One DMA request specifier consisting of a phandle to the DMA controller >> + followed by the integer specifying the channel identifier. The channel >> + identifier is encoded as follows: >> + - bits 7-0: Tx DMA channel number or the Rx flow number. >> + - bits 31-24: Channel type. 0xff for Tx DMA channel & 0xfe for Rx flow. >> +- dma-names: List of string identifiers for the DMA requests. >> + >> +Example: >> + >> + netcp: netcp@2090000 { >> + ... >> + dmas = <&netcpdma KEYSTONE_DMA_RX_FLOW(22)>, >> + <&netcpdma KEYSTONE_DMA_RX_FLOW(23)>, >> + <&netcpdma KEYSTONE_DMA_TX_CHAN(8)>; >> + dma-names = "netrx0", "netrx1", "nettx"; >> + ... >> + }; > Can you pls separate the binding to separate patch and also this needs ack from DT > folks > Will do that in next version. DT folks have been CC'ed on the patch. Will get their ack on DT bindings. [..] >> diff --git a/drivers/dma/keystone-pktdma.c b/drivers/dma/keystone-pktdma.c >> new file mode 100644 >> index 0000000..b3f77e5 >> --- /dev/null >> +++ b/drivers/dma/keystone-pktdma.c >> @@ -0,0 +1,795 @@ >> +/* >> + * Copyright (C) 2014 Texas Instruments Incorporated >> + * Authors: Sandeep Nair <sandeep_n@ti.com> >> + * Cyril Chemparathy <cyril@ti.com> >> + * Santosh Shilimkar <santosh.shilimkar@ti.com> >> + * >> + * This program is free software; you can redistribute it and/or >> + * modify it under the terms of the GNU General Public License as >> + * published by the Free Software Foundation version 2. >> + * >> + * This program is distributed "as is" WITHOUT ANY WARRANTY of any >> + * kind, whether express or implied; without even the implied warranty >> + * of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the >> + * GNU General Public License for more details. >> + */ >> + >> +#include <linux/io.h> >> +#include <linux/sched.h> >> +#include <linux/module.h> >> +#include <linux/dma-direction.h> >> +#include <linux/dmaengine.h> >> +#include <linux/interrupt.h> >> +#include <linux/of_dma.h> >> +#include <linux/of_address.h> >> +#include <linux/platform_device.h> >> +#include <linux/keystone-pktdma.h> >> +#include <linux/pm_runtime.h> >> +#include <dt-bindings/dma/keystone.h> >> + >> +#define BITS(x) (BIT(x) - 1) > this might get confusing, perhaps a better name could be given? > Would "BIT_MASK" be ok with you ? >> +#define REG_MASK 0xffffffff >> + >> +#define DMA_LOOPBACK BIT(31) >> +#define DMA_ENABLE BIT(31) >> +#define DMA_TEARDOWN BIT(30) >> + >> +#define DMA_TX_FILT_PSWORDS BIT(29) >> +#define DMA_TX_FILT_EINFO BIT(30) >> +#define DMA_TX_PRIO_SHIFT 0 >> +#define DMA_RX_PRIO_SHIFT 16 >> +#define DMA_PRIO_MASK BITS(3) >> +#define DMA_PRIO_DEFAULT 0 >> +#define DMA_RX_TIMEOUT_DEFAULT 17500 /* cycles */ >> +#define DMA_RX_TIMEOUT_MASK BITS(16) >> +#define DMA_RX_TIMEOUT_SHIFT 0 >> + >> +#define CHAN_HAS_EPIB BIT(30) >> +#define CHAN_HAS_PSINFO BIT(29) >> +#define CHAN_ERR_RETRY BIT(28) >> +#define CHAN_PSINFO_AT_SOP BIT(25) >> +#define CHAN_SOP_OFF_SHIFT 16 >> +#define CHAN_SOP_OFF_MASK BITS(9) >> +#define DESC_TYPE_SHIFT 26 >> +#define DESC_TYPE_MASK BITS(2) >> + >> +/* >> + * QMGR & QNUM together make up 14 bits with QMGR as the 2 MSb's in the logical >> + * navigator cloud mapping scheme. >> + * using the 14bit physical queue numbers directly maps into this scheme. >> + */ >> +#define CHAN_QNUM_MASK BITS(14) >> +#define DMA_MAX_QMS 4 >> +#define DMA_TIMEOUT 1000 /* msecs */ >> + >> +struct reg_global { >> + u32 revision; >> + u32 perf_control; >> + u32 emulation_control; >> + u32 priority_control; >> + u32 qm_base_address[4]; >> +}; >> + >> +struct reg_chan { >> + u32 control; >> + u32 mode; >> + u32 __rsvd[6]; >> +}; >> + >> +struct reg_tx_sched { >> + u32 prio; >> +}; >> + >> +struct reg_rx_flow { >> + u32 control; >> + u32 tags; >> + u32 tag_sel; >> + u32 fdq_sel[2]; >> + u32 thresh[3]; >> +}; >> + >> +#define BUILD_CHECK_REGS() \ >> + do { \ >> + BUILD_BUG_ON(sizeof(struct reg_global) != 32); \ >> + BUILD_BUG_ON(sizeof(struct reg_chan) != 32); \ >> + BUILD_BUG_ON(sizeof(struct reg_rx_flow) != 32); \ >> + BUILD_BUG_ON(sizeof(struct reg_tx_sched) != 4); \ >> + } while (0) > why is this required, do you want to use __packed__ to ensure right size? > This is just to ensure the register sanity. We should use __packed__ as well. We can take the BUILD_CHECK_REGS() out if you don't prefer it. >> + >> +enum keystone_chan_state { >> + /* stable states */ >> + CHAN_STATE_OPENED, >> + CHAN_STATE_CLOSED, >> +}; >> + >> +struct keystone_dma_device { >> + struct dma_device engine; >> + bool loopback, enable_all; >> + unsigned tx_priority, rx_priority, rx_timeout; >> + unsigned logical_queue_managers; >> + unsigned qm_base_address[DMA_MAX_QMS]; >> + struct reg_global __iomem *reg_global; >> + struct reg_chan __iomem *reg_tx_chan; >> + struct reg_rx_flow __iomem *reg_rx_flow; >> + struct reg_chan __iomem *reg_rx_chan; >> + struct reg_tx_sched __iomem *reg_tx_sched; >> + unsigned max_rx_chan, max_tx_chan; >> + unsigned max_rx_flow; >> + atomic_t in_use; >> +}; >> +#define to_dma(dma) (&(dma)->engine) >> +#define dma_dev(dma) ((dma)->engine.dev) >> + >> +struct keystone_dma_chan { >> + struct dma_chan achan; >> + enum dma_transfer_direction direction; >> + atomic_t state; >> + struct keystone_dma_device *dma; >> + >> + /* registers */ >> + struct reg_chan __iomem *reg_chan; >> + struct reg_tx_sched __iomem *reg_tx_sched; >> + struct reg_rx_flow __iomem *reg_rx_flow; >> + >> + /* configuration stuff */ >> + unsigned channel, flow; >> +}; >> +#define from_achan(ch) container_of(ch, struct keystone_dma_chan, achan) >> +#define to_achan(ch) (&(ch)->achan) >> +#define chan_dev(ch) (&to_achan(ch)->dev->device) >> +#define chan_num(ch) ((ch->direction == DMA_MEM_TO_DEV) ? \ >> + ch->channel : ch->flow) >> +#define chan_vdbg(ch, format, arg...) \ >> + dev_vdbg(chan_dev(ch), format, ##arg); >> + >> +/** >> + * dev_to_dma_chan - convert a device pointer to the its sysfs container object >> + * @dev - device node >> + */ >> +static inline struct dma_chan *dev_to_dma_chan(struct device *dev) >> +{ >> + struct dma_chan_dev *chan_dev; >> + >> + chan_dev = container_of(dev, typeof(*chan_dev), device); >> + return chan_dev->chan; >> +} >> + >> +static inline enum keystone_chan_state >> +chan_get_state(struct keystone_dma_chan *chan) >> +{ >> + return atomic_read(&chan->state); >> +} >> + >> +static int chan_start(struct keystone_dma_chan *chan, >> + struct dma_keystone_cfg *cfg) >> +{ >> + u32 v = 0; >> + >> + if ((chan->direction == DMA_MEM_TO_DEV) && chan->reg_chan) { > why second check? In case we to enable all channels with 'enable_all', then we can skip per channel configuration. Second check is for that case. >> + if (cfg->u.tx.filt_pswords) >> + v |= DMA_TX_FILT_PSWORDS; >> + if (cfg->u.tx.filt_einfo) >> + v |= DMA_TX_FILT_EINFO; >> + writel_relaxed(v, &chan->reg_chan->mode); >> + writel_relaxed(DMA_ENABLE, &chan->reg_chan->control); >> + } > no else for DMA_DEV_TO_MEM? 'chan->reg_rx_flow' case below takes care of "DMA_DEV_TO_MEM" case. Since it was explicit, the additional direction check isn't used. >> + if (chan->reg_tx_sched) >> + writel_relaxed(cfg->u.tx.priority, &chan->reg_tx_sched->prio); >> + >> + if (chan->reg_rx_flow) { >> + v = 0; >> + >> + if (cfg->u.rx.einfo_present) >> + v |= CHAN_HAS_EPIB; >> + if (cfg->u.rx.psinfo_present) >> + v |= CHAN_HAS_PSINFO; >> + if (cfg->u.rx.err_mode == DMA_RETRY) >> + v |= CHAN_ERR_RETRY; >> + v |= (cfg->u.rx.desc_type & DESC_TYPE_MASK) << DESC_TYPE_SHIFT; >> + if (cfg->u.rx.psinfo_at_sop) >> + v |= CHAN_PSINFO_AT_SOP; >> + v |= (cfg->u.rx.sop_offset & CHAN_SOP_OFF_MASK) >> + << CHAN_SOP_OFF_SHIFT; >> + v |= cfg->u.rx.dst_q & CHAN_QNUM_MASK; >> + >> + writel_relaxed(v, &chan->reg_rx_flow->control); >> + writel_relaxed(0, &chan->reg_rx_flow->tags); >> + writel_relaxed(0, &chan->reg_rx_flow->tag_sel); >> + >> + v = cfg->u.rx.fdq[0] << 16; >> + v |= cfg->u.rx.fdq[1] & CHAN_QNUM_MASK; >> + writel_relaxed(v, &chan->reg_rx_flow->fdq_sel[0]); >> + >> + v = cfg->u.rx.fdq[2] << 16; >> + v |= cfg->u.rx.fdq[3] & CHAN_QNUM_MASK; >> + writel_relaxed(v, &chan->reg_rx_flow->fdq_sel[1]); >> + >> + writel_relaxed(0, &chan->reg_rx_flow->thresh[0]); >> + writel_relaxed(0, &chan->reg_rx_flow->thresh[1]); >> + writel_relaxed(0, &chan->reg_rx_flow->thresh[2]); >> + } >> + >> + return 0; >> +} >> + [...] >> +static int chan_init(struct dma_chan *achan) >> +{ >> + struct keystone_dma_chan *chan = from_achan(achan); >> + struct keystone_dma_device *dma = chan->dma; >> + >> + chan_vdbg(chan, "initializing %s channel\n", >> + chan->direction == DMA_MEM_TO_DEV ? "transmit" : >> + chan->direction == DMA_DEV_TO_MEM ? "receive" : >> + "unknown"); >> + >> + if (chan->direction != DMA_MEM_TO_DEV && >> + chan->direction != DMA_DEV_TO_MEM) { >> + dev_err(chan_dev(chan), "bad direction\n"); >> + return -EINVAL; >> + } > is_slave_direction() pls > ok. [..] >> +static int chan_keystone_config(struct keystone_dma_chan *chan, >> + struct dma_keystone_cfg *cfg) >> +{ >> + if (chan_get_state(chan) != CHAN_STATE_OPENED) >> + return -ENODEV; >> + >> + if (cfg->sl_cfg.direction != chan->direction) >> + return -EINVAL; > direction is deprecated... ok will remove that. The above check is suprefluous any ways. [..] >> +static int pktdma_init_chan(struct keystone_dma_device *dma, >> + struct device_node *node, >> + enum dma_transfer_direction dir, >> + unsigned chan_num) >> +{ >> + struct device *dev = dma_dev(dma); >> + struct keystone_dma_chan *chan; >> + struct dma_chan *achan; >> + int ret = -EINVAL; >> + >> + chan = devm_kzalloc(dev, sizeof(*chan), GFP_KERNEL); >> + if (!chan) >> + return -ENOMEM; >> + >> + achan = to_achan(chan); >> + achan->device = &dma->engine; >> + chan->dma = dma; >> + chan->direction = DMA_NONE; >> + atomic_set(&chan->state, CHAN_STATE_OPENED); >> + >> + if (dir == DMA_MEM_TO_DEV) { >> + chan->direction = dir; >> + ret = pktdma_init_tx_chan(chan, node, chan_num); >> + } else if (dir == DMA_DEV_TO_MEM) { >> + chan->direction = dir; >> + ret = pktdma_init_rx_chan(chan, node, chan_num); >> + } else { >> + dev_err(dev, "channel(%d) direction unknown\n", chan_num); >> + } >> + >> + if (ret < 0) >> + goto fail; >> + >> + list_add_tail(&to_achan(chan)->device_node, &to_dma(dma)->channels); >> + return 0; >> + >> +fail: >> + devm_kfree(dev, chan); > ?? Missed out. Will removed that. >> +/* dummy function: feature not supported */ >> +static enum dma_status chan_xfer_status(struct dma_chan *achan, >> + dma_cookie_t cookie, >> + struct dma_tx_state *txstate) >> +{ >> + WARN(1, "xfer status not supported\n"); >> + return DMA_ERROR; >> +} >> + >> +/* dummy function: feature not supported */ >> +static void chan_issue_pending(struct dma_chan *chan) >> +{ >> + WARN(1, "issue pending not supported\n"); >> +} > Supporting status is okay but not issue_pending. This breaks use of dmaengine > API. What we expect is that user will do channel allocation, prepare a > descriptor, then submit it. Submit doesnt start the transaction, this call does! > So we need implementation here! > As explained in the begining, the data movement prep, trigger and completion is managed by QMSS and hence we don't have any implementation here.. Thats how the hardware is designed. >> + >> +static ssize_t keystone_dma_show_chan_num(struct device *dev, >> + struct device_attribute *attr, char *buf) >> +{ >> + struct dma_chan *achan = dev_to_dma_chan(dev); >> + struct keystone_dma_chan *chan = from_achan(achan); >> + >> + return scnprintf(buf, PAGE_SIZE, "%u\n", chan->channel); >> +} > ??? > >> + >> +static ssize_t keystone_dma_show_flow(struct device *dev, >> + struct device_attribute *attr, char *buf) >> +{ >> + struct dma_chan *achan = dev_to_dma_chan(dev); >> + struct keystone_dma_chan *chan = from_achan(achan); >> + >> + return scnprintf(buf, PAGE_SIZE, "%u\n", chan->flow); >> +} >> + >> +static DEVICE_ATTR(chan_num, S_IRUSR, keystone_dma_show_chan_num, NULL); >> +static DEVICE_ATTR(rx_flow, S_IRUSR, keystone_dma_show_flow, NULL); > okay why do we need these? > The sysfs was added just to see runtime which DMA resources have been used. No major need as such. Regards, Santosh
On Wed, Mar 12, 2014 at 03:50:32AM +0800, Santosh Shilimkar wrote: > On Tuesday 11 March 2014 06:23 PM, Vinod Koul wrote: > > On Fri, Feb 28, 2014 at 05:56:40PM -0500, Santosh Shilimkar wrote: > >> From: Sandeep Nair <sandeep_n@ti.com> > >> > >> The Packet DMA driver sets up the dma channels and flows for the > >> QMSS(Queue Manager SubSystem) who triggers the actual data movements > >> across clients using destination queues. Every client modules like > >> NETCP(Network Coprocessor), SRIO(Serial Rapid IO) and CRYPTO > >> Engines has its own instance of packet dma hardware. QMSS has also > >> an internal packet DMA module which is used as an infrastructure > >> DMA with zero copy. > >> > >> Patch adds DMAEngine driver for Keystone Packet DMA hardware. > >> The specifics on the device tree bindings for Packet DMA can be > >> found in: > >> Documentation/devicetree/bindings/dma/keystone-pktdma.txt > >> > >> The driver implements the configuration functions using standard DMAEngine > >> apis. The data movement is managed by QMSS device driver. > > Pls use subsystem appropriate name so here would have been dmaengine: ... > > > > So i am still missing stuff like prepare calls, irq, descriptor management to > > call this a dmaengine driver. > > > > I guess you need to explain a bit more why the data movement is handled by some > > other driver and not by this one > > > To expand above statement, Packet DMA hardware blocks on Keystone SOCs > are DMAEngines. QMSS is centralised subsystem manages multiple functionalities > including triggering dma transfers, descriptor management and completion > irqs. There are separate instance of packet DMA hardware block per client > device. We program the DMA hardware to allocate channels and flows. So > the packet DMA resouces like dma channels and dma flows are configured > and managed through the DMAEngine driver. Thats why we implement only > device_alloc_chan_resources, device_free_chan_resources and device_control > DMAEngine APIs. Sorry am bit lost. If its a dmaengine then you still need to program the transfers, how does that part work? > >> +#define BITS(x) (BIT(x) - 1) > > this might get confusing, perhaps a better name could be given? > > > Would "BIT_MASK" be ok with you ? something which would imply its x -1, am not really good with name so no suggestions :) > >> + > >> +#define BUILD_CHECK_REGS() \ > >> + do { \ > >> + BUILD_BUG_ON(sizeof(struct reg_global) != 32); \ > >> + BUILD_BUG_ON(sizeof(struct reg_chan) != 32); \ > >> + BUILD_BUG_ON(sizeof(struct reg_rx_flow) != 32); \ > >> + BUILD_BUG_ON(sizeof(struct reg_tx_sched) != 4); \ > >> + } while (0) > > why is this required, do you want to use __packed__ to ensure right size? > > > This is just to ensure the register sanity. We should use __packed__ as > well. We can take the BUILD_CHECK_REGS() out if you don't prefer it. putting packed ensures so no need for this check.
On Thursday 13 March 2014 12:00 AM, Vinod Koul wrote: > On Wed, Mar 12, 2014 at 03:50:32AM +0800, Santosh Shilimkar wrote: >> On Tuesday 11 March 2014 06:23 PM, Vinod Koul wrote: >>> On Fri, Feb 28, 2014 at 05:56:40PM -0500, Santosh Shilimkar wrote: >>>> From: Sandeep Nair <sandeep_n@ti.com> >>>> >>>> The Packet DMA driver sets up the dma channels and flows for the >>>> QMSS(Queue Manager SubSystem) who triggers the actual data movements >>>> across clients using destination queues. Every client modules like >>>> NETCP(Network Coprocessor), SRIO(Serial Rapid IO) and CRYPTO >>>> Engines has its own instance of packet dma hardware. QMSS has also >>>> an internal packet DMA module which is used as an infrastructure >>>> DMA with zero copy. >>>> >>>> Patch adds DMAEngine driver for Keystone Packet DMA hardware. >>>> The specifics on the device tree bindings for Packet DMA can be >>>> found in: >>>> Documentation/devicetree/bindings/dma/keystone-pktdma.txt >>>> >>>> The driver implements the configuration functions using standard DMAEngine >>>> apis. The data movement is managed by QMSS device driver. >>> Pls use subsystem appropriate name so here would have been dmaengine: ... >>> >>> So i am still missing stuff like prepare calls, irq, descriptor management to >>> call this a dmaengine driver. >>> >>> I guess you need to explain a bit more why the data movement is handled by some >>> other driver and not by this one >>> >> To expand above statement, Packet DMA hardware blocks on Keystone SOCs >> are DMAEngines. QMSS is centralised subsystem manages multiple functionalities >> including triggering dma transfers, descriptor management and completion >> irqs. There are separate instance of packet DMA hardware block per client >> device. We program the DMA hardware to allocate channels and flows. So >> the packet DMA resouces like dma channels and dma flows are configured >> and managed through the DMAEngine driver. Thats why we implement only >> device_alloc_chan_resources, device_free_chan_resources and device_control >> DMAEngine APIs. > Sorry am bit lost. If its a dmaengine then you still need to program the > transfers, how does that part work? > To simplify this bit more, you can think of this as DMA channels, flows are allocated and DMA channels are enabled by DMA engine and they remains enabled always as long as the channel in use. Enablling dma channel actually don't start the DMA transfer but just sets up the connection/pipe with peripheral and memory and vice a versa. All the descriptor management, triggering, sending completion interrupt or hardware signal to DMAEngine all managed by centralised QMSS. Actual copy of data is still done by DMA hardware but its completely transparent to software. DMAEngine hardware takes care of that in the backyard. >>>> +#define BITS(x) (BIT(x) - 1) >>> this might get confusing, perhaps a better name could be given? >>> >> Would "BIT_MASK" be ok with you ? > something which would imply its x -1, am not really good with name so no > suggestions :) > me too. BIT_INVERT_MASK() ? >>>> + >>>> +#define BUILD_CHECK_REGS() \ >>>> + do { \ >>>> + BUILD_BUG_ON(sizeof(struct reg_global) != 32); \ >>>> + BUILD_BUG_ON(sizeof(struct reg_chan) != 32); \ >>>> + BUILD_BUG_ON(sizeof(struct reg_rx_flow) != 32); \ >>>> + BUILD_BUG_ON(sizeof(struct reg_tx_sched) != 4); \ >>>> + } while (0) >>> why is this required, do you want to use __packed__ to ensure right size? >>> >> This is just to ensure the register sanity. We should use __packed__ as >> well. We can take the BUILD_CHECK_REGS() out if you don't prefer it. > putting packed ensures so no need for this check. > OK
On Thu, Mar 13, 2014 at 05:16:52AM +0800, Santosh Shilimkar wrote: > On Thursday 13 March 2014 12:00 AM, Vinod Koul wrote: > > On Wed, Mar 12, 2014 at 03:50:32AM +0800, Santosh Shilimkar wrote: > >> On Tuesday 11 March 2014 06:23 PM, Vinod Koul wrote: > >>> On Fri, Feb 28, 2014 at 05:56:40PM -0500, Santosh Shilimkar wrote: > >>>> From: Sandeep Nair <sandeep_n@ti.com> > >>>> > >>>> The Packet DMA driver sets up the dma channels and flows for the > >>>> QMSS(Queue Manager SubSystem) who triggers the actual data movements > >>>> across clients using destination queues. Every client modules like > >>>> NETCP(Network Coprocessor), SRIO(Serial Rapid IO) and CRYPTO > >>>> Engines has its own instance of packet dma hardware. QMSS has also > >>>> an internal packet DMA module which is used as an infrastructure > >>>> DMA with zero copy. > >>>> > >>>> Patch adds DMAEngine driver for Keystone Packet DMA hardware. > >>>> The specifics on the device tree bindings for Packet DMA can be > >>>> found in: > >>>> Documentation/devicetree/bindings/dma/keystone-pktdma.txt > >>>> > >>>> The driver implements the configuration functions using standard DMAEngine > >>>> apis. The data movement is managed by QMSS device driver. > >>> Pls use subsystem appropriate name so here would have been dmaengine: ... > >>> > >>> So i am still missing stuff like prepare calls, irq, descriptor management to > >>> call this a dmaengine driver. > >>> > >>> I guess you need to explain a bit more why the data movement is handled by some > >>> other driver and not by this one > >>> > >> To expand above statement, Packet DMA hardware blocks on Keystone SOCs > >> are DMAEngines. QMSS is centralised subsystem manages multiple functionalities > >> including triggering dma transfers, descriptor management and completion > >> irqs. There are separate instance of packet DMA hardware block per client > >> device. We program the DMA hardware to allocate channels and flows. So > >> the packet DMA resouces like dma channels and dma flows are configured > >> and managed through the DMAEngine driver. Thats why we implement only > >> device_alloc_chan_resources, device_free_chan_resources and device_control > >> DMAEngine APIs. > > Sorry am bit lost. If its a dmaengine then you still need to program the > > transfers, how does that part work? > > > To simplify this bit more, you can think of this as DMA channels, flows > are allocated and DMA channels are enabled by DMA engine and they remains > enabled always as long as the channel in use. Enablling dma channel > actually don't start the DMA transfer but just sets up the connection/pipe > with peripheral and memory and vice a versa. > > All the descriptor management, triggering, sending completion interrupt or > hardware signal to DMAEngine all managed by centralised QMSS. > > Actual copy of data is still done by DMA hardware but its completely > transparent to software. DMAEngine hardware takes care of that in the > backyard. So you will use the dmaengine just for setting up the controller. Not for actual transfers. Those would be governed by the QMSS, right? This means that someone expecting to use dmaengine API will get confused about this and doing part (alloc) thru dmaengine and rest (transfers) using some other API. This brings to me the design approach, does it really make sense creating dmaengine driver for this when we are not fully complying to the API
On Monday 17 March 2014 12:42 AM, Vinod Koul wrote: > On Thu, Mar 13, 2014 at 05:16:52AM +0800, Santosh Shilimkar wrote: >> On Thursday 13 March 2014 12:00 AM, Vinod Koul wrote: >>> On Wed, Mar 12, 2014 at 03:50:32AM +0800, Santosh Shilimkar wrote: >>>> On Tuesday 11 March 2014 06:23 PM, Vinod Koul wrote: >>>>> On Fri, Feb 28, 2014 at 05:56:40PM -0500, Santosh Shilimkar wrote: >>>>>> From: Sandeep Nair <sandeep_n@ti.com> >>>>>> >>>>>> The Packet DMA driver sets up the dma channels and flows for the >>>>>> QMSS(Queue Manager SubSystem) who triggers the actual data movements >>>>>> across clients using destination queues. Every client modules like >>>>>> NETCP(Network Coprocessor), SRIO(Serial Rapid IO) and CRYPTO >>>>>> Engines has its own instance of packet dma hardware. QMSS has also >>>>>> an internal packet DMA module which is used as an infrastructure >>>>>> DMA with zero copy. >>>>>> >>>>>> Patch adds DMAEngine driver for Keystone Packet DMA hardware. >>>>>> The specifics on the device tree bindings for Packet DMA can be >>>>>> found in: >>>>>> Documentation/devicetree/bindings/dma/keystone-pktdma.txt >>>>>> >>>>>> The driver implements the configuration functions using standard DMAEngine >>>>>> apis. The data movement is managed by QMSS device driver. >>>>> Pls use subsystem appropriate name so here would have been dmaengine: ... >>>>> >>>>> So i am still missing stuff like prepare calls, irq, descriptor management to >>>>> call this a dmaengine driver. >>>>> >>>>> I guess you need to explain a bit more why the data movement is handled by some >>>>> other driver and not by this one >>>>> >>>> To expand above statement, Packet DMA hardware blocks on Keystone SOCs >>>> are DMAEngines. QMSS is centralised subsystem manages multiple functionalities >>>> including triggering dma transfers, descriptor management and completion >>>> irqs. There are separate instance of packet DMA hardware block per client >>>> device. We program the DMA hardware to allocate channels and flows. So >>>> the packet DMA resouces like dma channels and dma flows are configured >>>> and managed through the DMAEngine driver. Thats why we implement only >>>> device_alloc_chan_resources, device_free_chan_resources and device_control >>>> DMAEngine APIs. >>> Sorry am bit lost. If its a dmaengine then you still need to program the >>> transfers, how does that part work? >>> >> To simplify this bit more, you can think of this as DMA channels, flows >> are allocated and DMA channels are enabled by DMA engine and they remains >> enabled always as long as the channel in use. Enablling dma channel >> actually don't start the DMA transfer but just sets up the connection/pipe >> with peripheral and memory and vice a versa. >> >> All the descriptor management, triggering, sending completion interrupt or >> hardware signal to DMAEngine all managed by centralised QMSS. >> >> Actual copy of data is still done by DMA hardware but its completely >> transparent to software. DMAEngine hardware takes care of that in the >> backyard. > So you will use the dmaengine just for setting up the controller. Not for actual > transfers. Those would be governed by the QMSS, right? > Correct. > This means that someone expecting to use dmaengine API will get confused about > this and doing part (alloc) thru dmaengine and rest (transfers) using some other > API. This brings to me the design approach, does it really make sense creating > dmaengine driver for this when we are not fully complying to the API > Thats fair. The rationale behind usage of DMEngine was that its the closest available subsystem which can be leveraged for this hardware. We can pretty much use all the standard DMAEngine device tree parsing as well as the config API to setup DMAs. I think you made your stand clear, just to confirm, you don't prefer this driver to be a DMAEngine driver considering it doesn't fully complying to the APIs. We could document the deviation of 'transfer' handling to avoid any confusion. Let me know your final thoughts and thanks a lot for the review. Regards, Santosh
On Mon, Mar 17, 2014 at 03:37:47PM -0400, Santosh Shilimkar wrote: > >> To simplify this bit more, you can think of this as DMA channels, flows > >> are allocated and DMA channels are enabled by DMA engine and they remains > >> enabled always as long as the channel in use. Enablling dma channel > >> actually don't start the DMA transfer but just sets up the connection/pipe > >> with peripheral and memory and vice a versa. > >> > >> All the descriptor management, triggering, sending completion interrupt or > >> hardware signal to DMAEngine all managed by centralised QMSS. > >> > >> Actual copy of data is still done by DMA hardware but its completely > >> transparent to software. DMAEngine hardware takes care of that in the > >> backyard. > > So you will use the dmaengine just for setting up the controller. Not for actual > > transfers. Those would be governed by the QMSS, right? > > > Correct. > > > This means that someone expecting to use dmaengine API will get confused about > > this and doing part (alloc) thru dmaengine and rest (transfers) using some other > > API. This brings to me the design approach, does it really make sense creating > > dmaengine driver for this when we are not fully complying to the API > > > Thats fair. The rationale behind usage of DMEngine was that its the closest > available subsystem which can be leveraged for this hardware. We can > pretty much use all the standard DMAEngine device tree parsing as well as > the config API to setup DMAs. > > I think you made your stand clear, just to confirm, you don't prefer this > driver to be a DMAEngine driver considering it doesn't fully complying to > the APIs. We could document the deviation of 'transfer' handling to avoid > any confusion. Yup, a user will just get confused as the driver doenst conform the dmaengine API. Unless someone comes up witha strong argument on why it should be dmaengine driver and what befits we see form such a model, i would like a damengine driver to comply to standard API and usage.
On Tuesday 18 March 2014 20:54:44 Vinod Koul wrote: > On Mon, Mar 17, 2014 at 03:37:47PM -0400, Santosh Shilimkar wrote: > > >> To simplify this bit more, you can think of this as DMA channels, flows > > >> are allocated and DMA channels are enabled by DMA engine and they remains > > >> enabled always as long as the channel in use. Enablling dma channel > > >> actually don't start the DMA transfer but just sets up the connection/pipe > > >> with peripheral and memory and vice a versa. > > >> > > >> All the descriptor management, triggering, sending completion interrupt or > > >> hardware signal to DMAEngine all managed by centralised QMSS. > > >> > > >> Actual copy of data is still done by DMA hardware but its completely > > >> transparent to software. DMAEngine hardware takes care of that in the > > >> backyard. > > > So you will use the dmaengine just for setting up the controller. Not for actual > > > transfers. Those would be governed by the QMSS, right? > > > > > Correct. > > > > > This means that someone expecting to use dmaengine API will get confused about > > > this and doing part (alloc) thru dmaengine and rest (transfers) using some other > > > API. This brings to me the design approach, does it really make sense creating > > > dmaengine driver for this when we are not fully complying to the API > > > > > Thats fair. The rationale behind usage of DMEngine was that its the closest > > available subsystem which can be leveraged for this hardware. We can > > pretty much use all the standard DMAEngine device tree parsing as well as > > the config API to setup DMAs. > > > > I think you made your stand clear, just to confirm, you don't prefer this > > driver to be a DMAEngine driver considering it doesn't fully complying to > > the APIs. We could document the deviation of 'transfer' handling to avoid > > any confusion. > Yup, a user will just get confused as the driver doenst conform the dmaengine > API. Unless someone comes up witha strong argument on why it should be > dmaengine driver and what befits we see form such a model, i would like a > damengine driver to comply to standard API and usage. I think it would be possible to turn the QMSS driver into a library and have the packet DMA code use the proper dmaengine API by calling into that code. The main user of packet DMA (the ethernet driver) would however still have to call into QMSS directly, so I'm not sure if it's worth the effort. Arnd
On Tue, Mar 18, 2014 at 04:38:25PM +0100, Arnd Bergmann wrote: > On Tuesday 18 March 2014 20:54:44 Vinod Koul wrote: > > On Mon, Mar 17, 2014 at 03:37:47PM -0400, Santosh Shilimkar wrote: > > > >> To simplify this bit more, you can think of this as DMA channels, flows > > > >> are allocated and DMA channels are enabled by DMA engine and they remains > > > >> enabled always as long as the channel in use. Enablling dma channel > > > >> actually don't start the DMA transfer but just sets up the connection/pipe > > > >> with peripheral and memory and vice a versa. > > > >> > > > >> All the descriptor management, triggering, sending completion interrupt or > > > >> hardware signal to DMAEngine all managed by centralised QMSS. > > > >> > > > >> Actual copy of data is still done by DMA hardware but its completely > > > >> transparent to software. DMAEngine hardware takes care of that in the > > > >> backyard. > > > > So you will use the dmaengine just for setting up the controller. Not for actual > > > > transfers. Those would be governed by the QMSS, right? > > > > > > > Correct. > > > > > > > This means that someone expecting to use dmaengine API will get confused about > > > > this and doing part (alloc) thru dmaengine and rest (transfers) using some other > > > > API. This brings to me the design approach, does it really make sense creating > > > > dmaengine driver for this when we are not fully complying to the API > > > > > > > Thats fair. The rationale behind usage of DMEngine was that its the closest > > > available subsystem which can be leveraged for this hardware. We can > > > pretty much use all the standard DMAEngine device tree parsing as well as > > > the config API to setup DMAs. > > > > > > I think you made your stand clear, just to confirm, you don't prefer this > > > driver to be a DMAEngine driver considering it doesn't fully complying to > > > the APIs. We could document the deviation of 'transfer' handling to avoid > > > any confusion. > > Yup, a user will just get confused as the driver doenst conform the dmaengine > > API. Unless someone comes up witha strong argument on why it should be > > dmaengine driver and what befits we see form such a model, i would like a > > damengine driver to comply to standard API and usage. > > I think it would be possible to turn the QMSS driver into a library and have > the packet DMA code use the proper dmaengine API by calling into that code. > > The main user of packet DMA (the ethernet driver) would however still have > to call into QMSS directly, so I'm not sure if it's worth the effort. Wouldn't that make clients use a standard API and also help in transaction management by using existing infrastructure
On Tuesday 18 March 2014 11:24 AM, Vinod Koul wrote: > On Mon, Mar 17, 2014 at 03:37:47PM -0400, Santosh Shilimkar wrote: >>>> To simplify this bit more, you can think of this as DMA channels, flows >>>> are allocated and DMA channels are enabled by DMA engine and they remains >>>> enabled always as long as the channel in use. Enablling dma channel >>>> actually don't start the DMA transfer but just sets up the connection/pipe >>>> with peripheral and memory and vice a versa. >>>> >>>> All the descriptor management, triggering, sending completion interrupt or >>>> hardware signal to DMAEngine all managed by centralised QMSS. >>>> >>>> Actual copy of data is still done by DMA hardware but its completely >>>> transparent to software. DMAEngine hardware takes care of that in the >>>> backyard. >>> So you will use the dmaengine just for setting up the controller. Not for actual >>> transfers. Those would be governed by the QMSS, right? >>> >> Correct. >> >>> This means that someone expecting to use dmaengine API will get confused about >>> this and doing part (alloc) thru dmaengine and rest (transfers) using some other >>> API. This brings to me the design approach, does it really make sense creating >>> dmaengine driver for this when we are not fully complying to the API >>> >> Thats fair. The rationale behind usage of DMEngine was that its the closest >> available subsystem which can be leveraged for this hardware. We can >> pretty much use all the standard DMAEngine device tree parsing as well as >> the config API to setup DMAs. >> >> I think you made your stand clear, just to confirm, you don't prefer this >> driver to be a DMAEngine driver considering it doesn't fully complying to >> the APIs. We could document the deviation of 'transfer' handling to avoid >> any confusion. > Yup, a user will just get confused as the driver doenst conform the dmaengine > API. Unless someone comes up witha strong argument on why it should be > dmaengine driver and what befits we see form such a model, i would like a > damengine driver to comply to standard API and usage. > OK thanks !! Regards, Santosh
On Tuesday 18 March 2014 11:38 AM, Arnd Bergmann wrote: > On Tuesday 18 March 2014 20:54:44 Vinod Koul wrote: >> On Mon, Mar 17, 2014 at 03:37:47PM -0400, Santosh Shilimkar wrote: >>>>> To simplify this bit more, you can think of this as DMA channels, flows >>>>> are allocated and DMA channels are enabled by DMA engine and they remains >>>>> enabled always as long as the channel in use. Enablling dma channel >>>>> actually don't start the DMA transfer but just sets up the connection/pipe >>>>> with peripheral and memory and vice a versa. >>>>> >>>>> All the descriptor management, triggering, sending completion interrupt or >>>>> hardware signal to DMAEngine all managed by centralised QMSS. >>>>> >>>>> Actual copy of data is still done by DMA hardware but its completely >>>>> transparent to software. DMAEngine hardware takes care of that in the >>>>> backyard. >>>> So you will use the dmaengine just for setting up the controller. Not for actual >>>> transfers. Those would be governed by the QMSS, right? >>>> >>> Correct. >>> >>>> This means that someone expecting to use dmaengine API will get confused about >>>> this and doing part (alloc) thru dmaengine and rest (transfers) using some other >>>> API. This brings to me the design approach, does it really make sense creating >>>> dmaengine driver for this when we are not fully complying to the API >>>> >>> Thats fair. The rationale behind usage of DMEngine was that its the closest >>> available subsystem which can be leveraged for this hardware. We can >>> pretty much use all the standard DMAEngine device tree parsing as well as >>> the config API to setup DMAs. >>> >>> I think you made your stand clear, just to confirm, you don't prefer this >>> driver to be a DMAEngine driver considering it doesn't fully complying to >>> the APIs. We could document the deviation of 'transfer' handling to avoid >>> any confusion. >> Yup, a user will just get confused as the driver doenst conform the dmaengine >> API. Unless someone comes up witha strong argument on why it should be >> dmaengine driver and what befits we see form such a model, i would like a >> damengine driver to comply to standard API and usage. > > I think it would be possible to turn the QMSS driver into a library and have > the packet DMA code use the proper dmaengine API by calling into that code. > > The main user of packet DMA (the ethernet driver) would however still have > to call into QMSS directly, so I'm not sure if it's worth the effort. > Its not. Am going to move this driver along with QMSS which is one of the options we discussed. Regards, Santosh
On Tuesday 18 March 2014, Santosh Shilimkar wrote: > > On Tuesday 18 March 2014 11:38 AM, Arnd Bergmann wrote: > > On Tuesday 18 March 2014 20:54:44 Vinod Koul wrote: > >> On Mon, Mar 17, 2014 at 03:37:47PM -0400, Santosh Shilimkar wrote: > >>>>> To simplify this bit more, you can think of this as DMA channels, flows > >>>>> are allocated and DMA channels are enabled by DMA engine and they remains > >>>>> enabled always as long as the channel in use. Enablling dma channel > >>>>> actually don't start the DMA transfer but just sets up the connection/pipe > >>>>> with peripheral and memory and vice a versa. > >>>>> > >>>>> All the descriptor management, triggering, sending completion interrupt or > >>>>> hardware signal to DMAEngine all managed by centralised QMSS. > >>>>> > >>>>> Actual copy of data is still done by DMA hardware but its completely > >>>>> transparent to software. DMAEngine hardware takes care of that in the > >>>>> backyard. > >>>> So you will use the dmaengine just for setting up the controller. Not for actual > >>>> transfers. Those would be governed by the QMSS, right? > >>>> > >>> Correct. > >>> > >>>> This means that someone expecting to use dmaengine API will get confused about > >>>> this and doing part (alloc) thru dmaengine and rest (transfers) using some other > >>>> API. This brings to me the design approach, does it really make sense creating > >>>> dmaengine driver for this when we are not fully complying to the API > >>>> > >>> Thats fair. The rationale behind usage of DMEngine was that its the closest > >>> available subsystem which can be leveraged for this hardware. We can > >>> pretty much use all the standard DMAEngine device tree parsing as well as > >>> the config API to setup DMAs. > >>> > >>> I think you made your stand clear, just to confirm, you don't prefer this > >>> driver to be a DMAEngine driver considering it doesn't fully complying to > >>> the APIs. We could document the deviation of 'transfer' handling to avoid > >>> any confusion. > >> Yup, a user will just get confused as the driver doenst conform the dmaengine > >> API. Unless someone comes up witha strong argument on why it should be > >> dmaengine driver and what befits we see form such a model, i would like a > >> damengine driver to comply to standard API and usage. > > > > I think it would be possible to turn the QMSS driver into a library and have > > the packet DMA code use the proper dmaengine API by calling into that code. > > > > The main user of packet DMA (the ethernet driver) would however still have > > to call into QMSS directly, so I'm not sure if it's worth the effort. > > > Its not. Am going to move this driver along with QMSS which is one > of the options we discussed. Ok, fair enough. Looking forward to the patches. Arnd
diff --git a/Documentation/devicetree/bindings/dma/keystone-pktdma.txt b/Documentation/devicetree/bindings/dma/keystone-pktdma.txt new file mode 100644 index 0000000..ea061d5 --- /dev/null +++ b/Documentation/devicetree/bindings/dma/keystone-pktdma.txt @@ -0,0 +1,72 @@ +Keystone Packet DMA Controller + +This document explains the device tree bindings for the packet dma +on keystone devices. The the Network coprocessor, Cypto engines +and the SRIO on Keystone devices all have their own packet dma modules. +Each individual packet dma has a certain number of RX channels, +RX flows and TX channels. Each instance of the packet DMA is being +initialized through device specific bindings. + +* DMA controller + +Required properties: + - compatible: Should be "ti,keystone-pktdma" + - reg: Should contain register location and length of the following pktdma + register regions. The value for "reg-names" property of the respective + region is specified in parenthesis. + - Global control register region (global). + - Tx DMA channel configuration register region (txchan). + - Rx DMA channel configuration register region (rxchan). + - Tx DMA channel Scheduler configuration register region (txsched). + - Rx DMA flow configuration register region (rxflow). + - reg-names: Names for the above regions. The name to be used is specified in + the above description of "reg" property. + - qm-base-address: Base address of the logical queue managers for pktdma. + - #dma-cells: Has to be 1. Keystone-pktdma doesn't support anything else. + +Optional properties: + - enable-all: Enable all DMA channels. + - loop-back: To loopback Tx streaming I/F to Rx streaming I/F. Used for + infrastructure transfers. + - rx-retry-timeout: Number of pktdma cycles to wait before retry on buffer + starvation. + +Example: + netcp-dma: pktdma@2004000 { + compatible = "ti,keystone-pktdma"; + reg = <0x2004000 0x100>, + <0x2004400 0x120>, + <0x2004800 0x300>, + <0x2004c00 0x120>, + <0x2005000 0x400>; + reg-names = "global", "txchan", "rxchan", "txsched", + "rxflow"; + qm-base-address = <0x23a80000 0x23a90000 + 0x23aa0000 0x23ab0000>; + #dma-cells = <1>; + /* enable-all; */ + rx-retry-timeout = <3500>; + /* loop-back; */ + }; + + +* DMA client + +Required properties: +- dmas: One DMA request specifier consisting of a phandle to the DMA controller + followed by the integer specifying the channel identifier. The channel + identifier is encoded as follows: + - bits 7-0: Tx DMA channel number or the Rx flow number. + - bits 31-24: Channel type. 0xff for Tx DMA channel & 0xfe for Rx flow. +- dma-names: List of string identifiers for the DMA requests. + +Example: + + netcp: netcp@2090000 { + ... + dmas = <&netcpdma KEYSTONE_DMA_RX_FLOW(22)>, + <&netcpdma KEYSTONE_DMA_RX_FLOW(23)>, + <&netcpdma KEYSTONE_DMA_TX_CHAN(8)>; + dma-names = "netrx0", "netrx1", "nettx"; + ... + }; diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig index 9bed1a2..722b99a 100644 --- a/drivers/dma/Kconfig +++ b/drivers/dma/Kconfig @@ -350,6 +350,14 @@ config MOXART_DMA help Enable support for the MOXA ART SoC DMA controller. +config KEYSTONE_PKTDMA + tristate "TI Keystone Packet DMA support" + depends on ARCH_KEYSTONE + select DMA_ENGINE + help + Enable support for the Packet DMA engine on Texas Instruments' + Keystone family of devices. + config DMA_ENGINE bool diff --git a/drivers/dma/Makefile b/drivers/dma/Makefile index a029d0f4..6d69c6d 100644 --- a/drivers/dma/Makefile +++ b/drivers/dma/Makefile @@ -44,3 +44,4 @@ obj-$(CONFIG_DMA_JZ4740) += dma-jz4740.o obj-$(CONFIG_TI_CPPI41) += cppi41.o obj-$(CONFIG_K3_DMA) += k3dma.o obj-$(CONFIG_MOXART_DMA) += moxart-dma.o +obj-$(CONFIG_KEYSTONE_PKTDMA) += keystone-pktdma.o diff --git a/drivers/dma/keystone-pktdma.c b/drivers/dma/keystone-pktdma.c new file mode 100644 index 0000000..b3f77e5 --- /dev/null +++ b/drivers/dma/keystone-pktdma.c @@ -0,0 +1,795 @@ +/* + * Copyright (C) 2014 Texas Instruments Incorporated + * Authors: Sandeep Nair <sandeep_n@ti.com> + * Cyril Chemparathy <cyril@ti.com> + * Santosh Shilimkar <santosh.shilimkar@ti.com> + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License as + * published by the Free Software Foundation version 2. + * + * This program is distributed "as is" WITHOUT ANY WARRANTY of any + * kind, whether express or implied; without even the implied warranty + * of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + */ + +#include <linux/io.h> +#include <linux/sched.h> +#include <linux/module.h> +#include <linux/dma-direction.h> +#include <linux/dmaengine.h> +#include <linux/interrupt.h> +#include <linux/of_dma.h> +#include <linux/of_address.h> +#include <linux/platform_device.h> +#include <linux/keystone-pktdma.h> +#include <linux/pm_runtime.h> +#include <dt-bindings/dma/keystone.h> + +#define BITS(x) (BIT(x) - 1) +#define REG_MASK 0xffffffff + +#define DMA_LOOPBACK BIT(31) +#define DMA_ENABLE BIT(31) +#define DMA_TEARDOWN BIT(30) + +#define DMA_TX_FILT_PSWORDS BIT(29) +#define DMA_TX_FILT_EINFO BIT(30) +#define DMA_TX_PRIO_SHIFT 0 +#define DMA_RX_PRIO_SHIFT 16 +#define DMA_PRIO_MASK BITS(3) +#define DMA_PRIO_DEFAULT 0 +#define DMA_RX_TIMEOUT_DEFAULT 17500 /* cycles */ +#define DMA_RX_TIMEOUT_MASK BITS(16) +#define DMA_RX_TIMEOUT_SHIFT 0 + +#define CHAN_HAS_EPIB BIT(30) +#define CHAN_HAS_PSINFO BIT(29) +#define CHAN_ERR_RETRY BIT(28) +#define CHAN_PSINFO_AT_SOP BIT(25) +#define CHAN_SOP_OFF_SHIFT 16 +#define CHAN_SOP_OFF_MASK BITS(9) +#define DESC_TYPE_SHIFT 26 +#define DESC_TYPE_MASK BITS(2) + +/* + * QMGR & QNUM together make up 14 bits with QMGR as the 2 MSb's in the logical + * navigator cloud mapping scheme. + * using the 14bit physical queue numbers directly maps into this scheme. + */ +#define CHAN_QNUM_MASK BITS(14) +#define DMA_MAX_QMS 4 +#define DMA_TIMEOUT 1000 /* msecs */ + +struct reg_global { + u32 revision; + u32 perf_control; + u32 emulation_control; + u32 priority_control; + u32 qm_base_address[4]; +}; + +struct reg_chan { + u32 control; + u32 mode; + u32 __rsvd[6]; +}; + +struct reg_tx_sched { + u32 prio; +}; + +struct reg_rx_flow { + u32 control; + u32 tags; + u32 tag_sel; + u32 fdq_sel[2]; + u32 thresh[3]; +}; + +#define BUILD_CHECK_REGS() \ + do { \ + BUILD_BUG_ON(sizeof(struct reg_global) != 32); \ + BUILD_BUG_ON(sizeof(struct reg_chan) != 32); \ + BUILD_BUG_ON(sizeof(struct reg_rx_flow) != 32); \ + BUILD_BUG_ON(sizeof(struct reg_tx_sched) != 4); \ + } while (0) + +enum keystone_chan_state { + /* stable states */ + CHAN_STATE_OPENED, + CHAN_STATE_CLOSED, +}; + +struct keystone_dma_device { + struct dma_device engine; + bool loopback, enable_all; + unsigned tx_priority, rx_priority, rx_timeout; + unsigned logical_queue_managers; + unsigned qm_base_address[DMA_MAX_QMS]; + struct reg_global __iomem *reg_global; + struct reg_chan __iomem *reg_tx_chan; + struct reg_rx_flow __iomem *reg_rx_flow; + struct reg_chan __iomem *reg_rx_chan; + struct reg_tx_sched __iomem *reg_tx_sched; + unsigned max_rx_chan, max_tx_chan; + unsigned max_rx_flow; + atomic_t in_use; +}; +#define to_dma(dma) (&(dma)->engine) +#define dma_dev(dma) ((dma)->engine.dev) + +struct keystone_dma_chan { + struct dma_chan achan; + enum dma_transfer_direction direction; + atomic_t state; + struct keystone_dma_device *dma; + + /* registers */ + struct reg_chan __iomem *reg_chan; + struct reg_tx_sched __iomem *reg_tx_sched; + struct reg_rx_flow __iomem *reg_rx_flow; + + /* configuration stuff */ + unsigned channel, flow; +}; +#define from_achan(ch) container_of(ch, struct keystone_dma_chan, achan) +#define to_achan(ch) (&(ch)->achan) +#define chan_dev(ch) (&to_achan(ch)->dev->device) +#define chan_num(ch) ((ch->direction == DMA_MEM_TO_DEV) ? \ + ch->channel : ch->flow) +#define chan_vdbg(ch, format, arg...) \ + dev_vdbg(chan_dev(ch), format, ##arg); + +/** + * dev_to_dma_chan - convert a device pointer to the its sysfs container object + * @dev - device node + */ +static inline struct dma_chan *dev_to_dma_chan(struct device *dev) +{ + struct dma_chan_dev *chan_dev; + + chan_dev = container_of(dev, typeof(*chan_dev), device); + return chan_dev->chan; +} + +static inline enum keystone_chan_state +chan_get_state(struct keystone_dma_chan *chan) +{ + return atomic_read(&chan->state); +} + +static int chan_start(struct keystone_dma_chan *chan, + struct dma_keystone_cfg *cfg) +{ + u32 v = 0; + + if ((chan->direction == DMA_MEM_TO_DEV) && chan->reg_chan) { + if (cfg->u.tx.filt_pswords) + v |= DMA_TX_FILT_PSWORDS; + if (cfg->u.tx.filt_einfo) + v |= DMA_TX_FILT_EINFO; + writel_relaxed(v, &chan->reg_chan->mode); + writel_relaxed(DMA_ENABLE, &chan->reg_chan->control); + } + + if (chan->reg_tx_sched) + writel_relaxed(cfg->u.tx.priority, &chan->reg_tx_sched->prio); + + if (chan->reg_rx_flow) { + v = 0; + + if (cfg->u.rx.einfo_present) + v |= CHAN_HAS_EPIB; + if (cfg->u.rx.psinfo_present) + v |= CHAN_HAS_PSINFO; + if (cfg->u.rx.err_mode == DMA_RETRY) + v |= CHAN_ERR_RETRY; + v |= (cfg->u.rx.desc_type & DESC_TYPE_MASK) << DESC_TYPE_SHIFT; + if (cfg->u.rx.psinfo_at_sop) + v |= CHAN_PSINFO_AT_SOP; + v |= (cfg->u.rx.sop_offset & CHAN_SOP_OFF_MASK) + << CHAN_SOP_OFF_SHIFT; + v |= cfg->u.rx.dst_q & CHAN_QNUM_MASK; + + writel_relaxed(v, &chan->reg_rx_flow->control); + writel_relaxed(0, &chan->reg_rx_flow->tags); + writel_relaxed(0, &chan->reg_rx_flow->tag_sel); + + v = cfg->u.rx.fdq[0] << 16; + v |= cfg->u.rx.fdq[1] & CHAN_QNUM_MASK; + writel_relaxed(v, &chan->reg_rx_flow->fdq_sel[0]); + + v = cfg->u.rx.fdq[2] << 16; + v |= cfg->u.rx.fdq[3] & CHAN_QNUM_MASK; + writel_relaxed(v, &chan->reg_rx_flow->fdq_sel[1]); + + writel_relaxed(0, &chan->reg_rx_flow->thresh[0]); + writel_relaxed(0, &chan->reg_rx_flow->thresh[1]); + writel_relaxed(0, &chan->reg_rx_flow->thresh[2]); + } + + return 0; +} + +static int chan_teardown(struct keystone_dma_chan *chan) +{ + unsigned long end, value; + + if (!chan->reg_chan) + return 0; + + /* indicate teardown */ + writel_relaxed(DMA_TEARDOWN, &chan->reg_chan->control); + + /* wait for the dma to shut itself down */ + end = jiffies + msecs_to_jiffies(DMA_TIMEOUT); + do { + value = readl_relaxed(&chan->reg_chan->control); + if ((value & DMA_ENABLE) == 0) + break; + } while (time_after(end, jiffies)); + + if (readl_relaxed(&chan->reg_chan->control) & DMA_ENABLE) { + dev_err(chan_dev(chan), "timeout waiting for teardown\n"); + return -ETIMEDOUT; + } + + return 0; +} + +static void chan_stop(struct keystone_dma_chan *chan) +{ + if (chan->reg_rx_flow) { + /* first detach fdqs, starve out the flow */ + writel_relaxed(0, &chan->reg_rx_flow->fdq_sel[0]); + writel_relaxed(0, &chan->reg_rx_flow->fdq_sel[1]); + writel_relaxed(0, &chan->reg_rx_flow->thresh[0]); + writel_relaxed(0, &chan->reg_rx_flow->thresh[1]); + writel_relaxed(0, &chan->reg_rx_flow->thresh[2]); + } + + /* teardown the dma channel */ + chan_teardown(chan); + + /* then disconnect the completion side */ + if (chan->reg_rx_flow) { + writel_relaxed(0, &chan->reg_rx_flow->control); + writel_relaxed(0, &chan->reg_rx_flow->tags); + writel_relaxed(0, &chan->reg_rx_flow->tag_sel); + } + chan_vdbg(chan, "channel stopped\n"); +} + +static void keystone_dma_hw_init(struct keystone_dma_device *dma) +{ + unsigned v; + int i; + + v = dma->loopback ? DMA_LOOPBACK : 0; + writel_relaxed(v, &dma->reg_global->emulation_control); + + v = readl_relaxed(&dma->reg_global->perf_control); + v |= ((dma->rx_timeout & DMA_RX_TIMEOUT_MASK) << DMA_RX_TIMEOUT_SHIFT); + writel_relaxed(v, &dma->reg_global->perf_control); + + v = ((dma->tx_priority << DMA_TX_PRIO_SHIFT) | + (dma->rx_priority << DMA_RX_PRIO_SHIFT)); + + writel_relaxed(v, &dma->reg_global->priority_control); + + if (dma->enable_all) { + for (i = 0; i < dma->max_tx_chan; i++) { + writel_relaxed(0, &dma->reg_tx_chan[i].mode); + writel_relaxed(DMA_ENABLE, + &dma->reg_tx_chan[i].control); + } + } + + /* Always enable all Rx channels. Rx paths are managed using flows */ + for (i = 0; i < dma->max_rx_chan; i++) + writel_relaxed(DMA_ENABLE, &dma->reg_rx_chan[i].control); + + for (i = 0; i < dma->logical_queue_managers; i++) + writel_relaxed(dma->qm_base_address[i], + &dma->reg_global->qm_base_address[i]); +} + +static void keystone_dma_hw_destroy(struct keystone_dma_device *dma) +{ + int i; + unsigned v; + + v = ~DMA_ENABLE & REG_MASK; + + for (i = 0; i < dma->max_rx_chan; i++) + writel_relaxed(v, &dma->reg_rx_chan[i].control); + + for (i = 0; i < dma->max_tx_chan; i++) + writel_relaxed(v, &dma->reg_tx_chan[i].control); +} + +static int chan_init(struct dma_chan *achan) +{ + struct keystone_dma_chan *chan = from_achan(achan); + struct keystone_dma_device *dma = chan->dma; + + chan_vdbg(chan, "initializing %s channel\n", + chan->direction == DMA_MEM_TO_DEV ? "transmit" : + chan->direction == DMA_DEV_TO_MEM ? "receive" : + "unknown"); + + if (chan->direction != DMA_MEM_TO_DEV && + chan->direction != DMA_DEV_TO_MEM) { + dev_err(chan_dev(chan), "bad direction\n"); + return -EINVAL; + } + + atomic_set(&chan->state, CHAN_STATE_OPENED); + + if (atomic_inc_return(&dma->in_use) <= 1) + keystone_dma_hw_init(dma); + + return 0; +} + +static void chan_destroy(struct dma_chan *achan) +{ + struct keystone_dma_chan *chan = from_achan(achan); + struct keystone_dma_device *dma = chan->dma; + + if (chan_get_state(chan) == CHAN_STATE_CLOSED) + return; + + chan_vdbg(chan, "destroying channel\n"); + chan_stop(chan); + atomic_set(&chan->state, CHAN_STATE_CLOSED); + if (atomic_dec_return(&dma->in_use) <= 0) + keystone_dma_hw_destroy(dma); + chan_vdbg(chan, "channel destroyed\n"); +} + +static int chan_keystone_config(struct keystone_dma_chan *chan, + struct dma_keystone_cfg *cfg) +{ + if (chan_get_state(chan) != CHAN_STATE_OPENED) + return -ENODEV; + + if (cfg->sl_cfg.direction != chan->direction) + return -EINVAL; + + return chan_start(chan, cfg); +} + +static int chan_control(struct dma_chan *achan, enum dma_ctrl_cmd cmd, + unsigned long arg) +{ + struct keystone_dma_chan *chan = from_achan(achan); + struct dma_keystone_cfg *keystone_config; + struct dma_slave_config *dma_cfg; + int ret; + + switch (cmd) { + case DMA_SLAVE_CONFIG: + dma_cfg = (struct dma_slave_config *)arg; + keystone_config = keystone_cfg_from_slave_config(dma_cfg); + ret = chan_keystone_config(chan, keystone_config); + break; + + default: + ret = -ENOTSUPP; + break; + } + return ret; +} + +static void __iomem *pktdma_get_regs( + struct keystone_dma_device *dma, const char *name, + resource_size_t *_size) +{ + struct device *dev = dma_dev(dma); + struct device_node *node = dev->of_node; + resource_size_t size; + struct resource res; + void __iomem *regs; + int i; + + i = of_property_match_string(node, "reg-names", name); + if (of_address_to_resource(node, i, &res)) { + dev_err(dev, "could not find %s resource(index %d)\n", name, i); + return NULL; + } + size = resource_size(&res); + + regs = of_iomap(node, i); + if (!regs) { + dev_err(dev, "could not map %s resource (index %d)\n", name, i); + return NULL; + } + + dev_dbg(dev, "index: %d, res:%s, size:%x, phys:%x, virt:%p\n", + i, name, (unsigned int)size, (unsigned int)res.start, regs); + + if (_size) + *_size = size; + + return regs; +} + +static int pktdma_init_rx_chan(struct keystone_dma_chan *chan, + struct device_node *node, + u32 flow) +{ + struct keystone_dma_device *dma = chan->dma; + struct device *dev = dma_dev(chan->dma); + + chan->flow = flow; + chan->reg_rx_flow = dma->reg_rx_flow + flow; + dev_dbg(dev, "rx flow(%d) (%p)\n", chan->flow, chan->reg_rx_flow); + + return 0; +} + +static int pktdma_init_tx_chan(struct keystone_dma_chan *chan, + struct device_node *node, + u32 channel) +{ + struct keystone_dma_device *dma = chan->dma; + struct device *dev = dma_dev(chan->dma); + + chan->channel = channel; + chan->reg_chan = dma->reg_tx_chan + channel; + chan->reg_tx_sched = dma->reg_tx_sched + channel; + dev_dbg(dev, "tx channel(%d) (%p)\n", chan->channel, chan->reg_chan); + + return 0; +} + +static int pktdma_init_chan(struct keystone_dma_device *dma, + struct device_node *node, + enum dma_transfer_direction dir, + unsigned chan_num) +{ + struct device *dev = dma_dev(dma); + struct keystone_dma_chan *chan; + struct dma_chan *achan; + int ret = -EINVAL; + + chan = devm_kzalloc(dev, sizeof(*chan), GFP_KERNEL); + if (!chan) + return -ENOMEM; + + achan = to_achan(chan); + achan->device = &dma->engine; + chan->dma = dma; + chan->direction = DMA_NONE; + atomic_set(&chan->state, CHAN_STATE_OPENED); + + if (dir == DMA_MEM_TO_DEV) { + chan->direction = dir; + ret = pktdma_init_tx_chan(chan, node, chan_num); + } else if (dir == DMA_DEV_TO_MEM) { + chan->direction = dir; + ret = pktdma_init_rx_chan(chan, node, chan_num); + } else { + dev_err(dev, "channel(%d) direction unknown\n", chan_num); + } + + if (ret < 0) + goto fail; + + list_add_tail(&to_achan(chan)->device_node, &to_dma(dma)->channels); + return 0; + +fail: + devm_kfree(dev, chan); + return ret; +} + +/* dummy function: feature not supported */ +static enum dma_status chan_xfer_status(struct dma_chan *achan, + dma_cookie_t cookie, + struct dma_tx_state *txstate) +{ + WARN(1, "xfer status not supported\n"); + return DMA_ERROR; +} + +/* dummy function: feature not supported */ +static void chan_issue_pending(struct dma_chan *chan) +{ + WARN(1, "issue pending not supported\n"); +} + +static ssize_t keystone_dma_show_chan_num(struct device *dev, + struct device_attribute *attr, char *buf) +{ + struct dma_chan *achan = dev_to_dma_chan(dev); + struct keystone_dma_chan *chan = from_achan(achan); + + return scnprintf(buf, PAGE_SIZE, "%u\n", chan->channel); +} + +static ssize_t keystone_dma_show_flow(struct device *dev, + struct device_attribute *attr, char *buf) +{ + struct dma_chan *achan = dev_to_dma_chan(dev); + struct keystone_dma_chan *chan = from_achan(achan); + + return scnprintf(buf, PAGE_SIZE, "%u\n", chan->flow); +} + +static DEVICE_ATTR(chan_num, S_IRUSR, keystone_dma_show_chan_num, NULL); +static DEVICE_ATTR(rx_flow, S_IRUSR, keystone_dma_show_flow, NULL); + +static void keystone_dma_destroy_attr(struct keystone_dma_device *dma) +{ + struct dma_device *engine = to_dma(dma); + struct keystone_dma_chan *chan; + struct dma_chan *achan; + struct device *dev; + + list_for_each_entry(achan, &engine->channels, device_node) { + chan = from_achan(achan); + dev = chan_dev(chan); + + /* remove sysfs entries */ + if (chan->direction == DMA_MEM_TO_DEV) + device_remove_file(dev, &dev_attr_chan_num); + else + device_remove_file(dev, &dev_attr_rx_flow); + } +} + +static int keystone_dma_setup_attr(struct keystone_dma_device *dma) +{ + struct dma_device *engine = to_dma(dma); + struct keystone_dma_chan *chan; + struct dma_chan *achan; + struct device *dev; + int status = 0; + + list_for_each_entry(achan, &engine->channels, device_node) { + chan = from_achan(achan); + dev = chan_dev(chan); + + /* add sysfs entries */ + if (chan->direction == DMA_MEM_TO_DEV) { + status = device_create_file(dev, &dev_attr_chan_num); + if (status) + dev_warn(dev, + "Couldn't create sysfs chan_num\n"); + } else { + status = device_create_file(dev, &dev_attr_rx_flow); + if (status) + dev_warn(dev, + "Couldn't create sysfs rx_flow\n"); + } + } + return status; +} + +static struct of_dma_filter_info keystone_dma_info = { + .filter_fn = dma_keystone_filter_fn, +}; + +static int keystone_dma_probe(struct platform_device *pdev) +{ + unsigned max_tx_chan, max_rx_chan, max_rx_flow, max_tx_sched; + struct device_node *node = pdev->dev.of_node; + struct keystone_dma_device *dma; + struct device_node *chan; + int ret, len, num_chan = 0; + struct dma_device *engine; + resource_size_t size; + u32 timeout; + u32 i; + + if (!node) { + dev_err(&pdev->dev, "could not find device info\n"); + return -EINVAL; + } + + dma = devm_kzalloc(&pdev->dev, sizeof(*dma), GFP_KERNEL); + if (!dma) { + dev_err(&pdev->dev, "could not allocate driver mem\n"); + return -ENOMEM; + } + + engine = to_dma(dma); + engine->dev = &pdev->dev; + platform_set_drvdata(pdev, dma); + + pm_runtime_enable(&pdev->dev); + ret = pm_runtime_get_sync(&pdev->dev); + if (ret < 0) { + dev_err(&pdev->dev, "unable to enable pktdma, err %d\n", ret); + return ret; + } + + dma->reg_global = pktdma_get_regs(dma, "global", &size); + if (!dma->reg_global) + return -ENODEV; + if (size < sizeof(struct reg_global)) { + dev_err(dma_dev(dma), "bad size %pa for global regs\n", &size); + return -ENODEV; + } + + dma->reg_tx_chan = pktdma_get_regs(dma, "txchan", &size); + if (!dma->reg_tx_chan) + return -ENODEV; + + max_tx_chan = size / sizeof(struct reg_chan); + dma->reg_rx_chan = pktdma_get_regs(dma, "rxchan", &size); + if (!dma->reg_rx_chan) + return -ENODEV; + + max_rx_chan = size / sizeof(struct reg_chan); + dma->reg_tx_sched = pktdma_get_regs(dma, "txsched", &size); + if (!dma->reg_tx_sched) + return -ENODEV; + + max_tx_sched = size / sizeof(struct reg_tx_sched); + dma->reg_rx_flow = pktdma_get_regs(dma, "rxflow", &size); + if (!dma->reg_rx_flow) + return -ENODEV; + + max_rx_flow = size / sizeof(struct reg_rx_flow); + dma->rx_priority = DMA_PRIO_DEFAULT; + dma->tx_priority = DMA_PRIO_DEFAULT; + + dma->enable_all = (of_get_property(node, "enable-all", NULL) != NULL); + dma->loopback = (of_get_property(node, "loop-back", NULL) != NULL); + + ret = of_property_read_u32(node, "rx-retry-timeout", &timeout); + if (ret < 0) { + dev_dbg(&pdev->dev, "unspecified rx timeout using value %d\n", + DMA_RX_TIMEOUT_DEFAULT); + timeout = DMA_RX_TIMEOUT_DEFAULT; + } + dma->rx_timeout = timeout; + + if (!of_find_property(node, "qm-base-address", &len)) { + dev_err(&pdev->dev, "unspecified qm-base-address\n"); + return -ENODEV; + } + + dma->logical_queue_managers = len / sizeof(u32); + if (dma->logical_queue_managers > DMA_MAX_QMS) { + dev_warn(&pdev->dev, "too many queue mgrs(>%d) rest ignored\n", + dma->logical_queue_managers); + dma->logical_queue_managers = DMA_MAX_QMS; + } + + ret = of_property_read_u32_array(node, "qm-base-address", + dma->qm_base_address, + dma->logical_queue_managers); + if (ret) { + dev_err(&pdev->dev, "invalid qm-base-address\n"); + return ret; + } + + dma->max_rx_chan = max_rx_chan; + dma->max_rx_flow = max_rx_flow; + dma->max_tx_chan = min(max_tx_chan, max_tx_sched); + + atomic_set(&dma->in_use, 0); + + INIT_LIST_HEAD(&engine->channels); + + for (i = 0; i < dma->max_tx_chan; i++) { + if (pktdma_init_chan(dma, chan, DMA_MEM_TO_DEV, i) >= 0) + num_chan++; + } + + for (i = 0; i < dma->max_rx_chan; i++) { + if (pktdma_init_chan(dma, chan, DMA_DEV_TO_MEM, i) >= 0) + num_chan++; + } + + if (list_empty(&engine->channels)) { + dev_err(dma_dev(dma), "no valid channels\n"); + return -ENODEV; + } + + dma_cap_set(DMA_SLAVE, engine->cap_mask); + engine->device_alloc_chan_resources = chan_init; + engine->device_free_chan_resources = chan_destroy; + engine->device_control = chan_control; + engine->device_tx_status = chan_xfer_status; + engine->device_issue_pending = chan_issue_pending; + + ret = dma_async_device_register(engine); + if (ret) { + dev_err(&pdev->dev, "unable to register dma engine\n"); + return ret; + } + + keystone_dma_info.dma_cap = engine->cap_mask; + ret = of_dma_controller_register(node, of_dma_simple_xlate, + &keystone_dma_info); + if (ret) { + dev_err(&pdev->dev, "Failed to register dma controller\n"); + dma_async_device_unregister(engine); + return ret; + } + + ret = keystone_dma_setup_attr(dma); + if (ret) { + dev_err(&pdev->dev, "unable to setup device attr\n"); + return ret; + } + + dev_info(dma_dev(dma), "registered %d logical channels, flows %d, tx chans: %d, rx chans: %d%s\n", + num_chan, dma->max_rx_flow, dma->max_tx_chan, + dma->max_rx_chan, dma->loopback ? ", loopback" : ""); + + return 0; +} + +static int keystone_dma_remove(struct platform_device *pdev) +{ + struct keystone_dma_device *dma = platform_get_drvdata(pdev); + struct dma_device *engine = to_dma(dma); + + pm_runtime_put_sync(&pdev->dev); + pm_runtime_disable(&pdev->dev); + + keystone_dma_destroy_attr(dma); + dma_async_device_unregister(engine); + + return 0; +} + +static struct of_device_id of_match[] = { + { .compatible = "ti,keystone-pktdma", }, + {}, +}; + +MODULE_DEVICE_TABLE(of, of_match); + +static struct platform_driver keystone_dma_driver = { + .probe = keystone_dma_probe, + .remove = keystone_dma_remove, + .driver = { + .name = "keystone-pktdma", + .owner = THIS_MODULE, + .of_match_table = of_match, + }, +}; + +bool dma_keystone_filter_fn(struct dma_chan *chan, void *param) +{ + if (chan->device->dev->driver == &keystone_dma_driver.driver) { + struct keystone_dma_chan *kdma_chan = from_achan(chan); + unsigned chan_id = *(u32 *)param & KEYSTONE_DMA_CHAN_ID_MASK; + unsigned chan_type = *(u32 *)param >> KEYSTONE_DMA_TYPE_SHIFT; + + if (chan_type == KEYSTONE_DMA_TX_CHAN_TYPE && + kdma_chan->direction == DMA_MEM_TO_DEV) + return chan_id == kdma_chan->channel; + + if (chan_type == KEYSTONE_DMA_RX_FLOW_TYPE && + kdma_chan->direction == DMA_DEV_TO_MEM) + return chan_id == kdma_chan->flow; + } + return false; +} +EXPORT_SYMBOL_GPL(dma_keystone_filter_fn); + +static int __init keystone_dma_init(void) +{ + BUILD_CHECK_REGS(); + return platform_driver_register(&keystone_dma_driver); +} +module_init(keystone_dma_init); + +static void __exit keystone_dma_exit(void) +{ + platform_driver_unregister(&keystone_dma_driver); +} +module_exit(keystone_dma_exit); + +MODULE_LICENSE("GPL v2"); +MODULE_DESCRIPTION("TI Keystone Packet DMA driver"); diff --git a/include/dt-bindings/dma/keystone.h b/include/dt-bindings/dma/keystone.h new file mode 100644 index 0000000..abe029e --- /dev/null +++ b/include/dt-bindings/dma/keystone.h @@ -0,0 +1,33 @@ +/* + * Copyright (C) 2014 Texas Instruments Incorporated + * Authors: Sandeep Nair <sandeep_n@ti.com> + * Santosh Shilimkar <santosh.shilimkar@ti.com> + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License as + * published by the Free Software Foundation version 2. + * + * This program is distributed "as is" WITHOUT ANY WARRANTY of any + * kind, whether express or implied; without even the implied warranty + * of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + */ + +#ifndef __DT_BINDINGS_KEYSTONE_DMA_H__ +#define __DT_BINDINGS_KEYSTONE_DMA_H__ + +#define KEYSTONE_DMA_CHAN_ID_MASK (0xff) +#define KEYSTONE_DMA_TYPE_SHIFT (24) +#define KEYSTONE_DMA_TX_CHAN_TYPE (0xff) +#define KEYSTONE_DMA_RX_FLOW_TYPE (0xfe) + +#define KEYSTONE_DMA_TX_CHAN(id) ((id & KEYSTONE_DMA_CHAN_ID_MASK) | \ + (KEYSTONE_DMA_TX_CHAN_TYPE << \ + KEYSTONE_DMA_TYPE_SHIFT)) + +#define KEYSTONE_DMA_RX_FLOW(id) ((id & KEYSTONE_DMA_CHAN_ID_MASK) | \ + (KEYSTONE_DMA_RX_FLOW_TYPE << \ + KEYSTONE_DMA_TYPE_SHIFT)) + + +#endif /* __DT_BINDINGS_KEYSTONE_DMA_H_ */ diff --git a/include/linux/keystone-pktdma.h b/include/linux/keystone-pktdma.h new file mode 100644 index 0000000..e6a66c8 --- /dev/null +++ b/include/linux/keystone-pktdma.h @@ -0,0 +1,168 @@ +/* + * Copyright (C) 2014 Texas Instruments Incorporated + * Authors: Sandeep Nair <sandeep_n@ti.com + * Cyril Chemparathy <cyril@ti.com + Santosh Shilimkar <santosh.shilimkar@ti.com> + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License as + * published by the Free Software Foundation version 2. + * + * This program is distributed "as is" WITHOUT ANY WARRANTY of any + * kind, whether express or implied; without even the implied warranty + * of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + */ + +#ifndef __KEYSTONE_PKTDMA_H__ +#define __KEYSTONE_PKTDMA_H__ + +#include <linux/dmaengine.h> + +/* + * PKTDMA descriptor manipulation macros for host packet descriptor + */ +#define MASK(x) (BIT(x) - 1) +#define DMA_KEYSTONE_DESC_PKT_LEN_MASK MASK(22) +#define DMA_KEYSTONE_DESC_PKT_LEN_SHIFT 0 +#define DMA_KEYSTONE_DESC_PS_INFO_IN_SOP BIT(22) +#define DMA_KEYSTONE_DESC_PS_INFO_IN_DESC 0 +#define DMA_KEYSTONE_DESC_TAG_MASK MASK(8) +#define DMA_KEYSTONE_DESC_SAG_HI_SHIFT 24 +#define DMA_KEYSTONE_DESC_STAG_LO_SHIFT 16 +#define DMA_KEYSTONE_DESC_DTAG_HI_SHIFT 8 +#define DMA_KEYSTONE_DESC_DTAG_LO_SHIFT 0 +#define DMA_KEYSTONE_DESC_HAS_EPIB BIT(31) +#define DMA_KEYSTONE_DESC_NO_EPIB 0 +#define DMA_KEYSTONE_DESC_PSLEN_SHIFT 24 +#define DMA_KEYSTONE_DESC_PSLEN_MASK MASK(6) +#define DMA_KEYSTONE_DESC_ERR_FLAG_SHIFT 20 +#define DMA_KEYSTONE_DESC_ERR_FLAG_MASK MASK(4) +#define DMA_KEYSTONE_DESC_PSFLAG_SHIFT 16 +#define DMA_KEYSTONE_DESC_PSFLAG_MASK MASK(4) +#define DMA_KEYSTONE_DESC_RETQ_SHIFT 0 +#define DMA_KEYSTONE_DESC_RETQ_MASK MASK(14) +#define DMA_KEYSTONE_DESC_BUF_LEN_MASK MASK(22) + +#define DMA_KEYSTONE_NUM_EPIB_WORDS 4 +#define DMA_KEYSTONE_NUM_PS_WORDS 16 +#define DMA_KEYSTONE_FDQ_PER_CHAN 4 + +/* Tx channel scheduling priority */ +enum dma_keystone_tx_priority { + DMA_PRIO_HIGH = 0, + DMA_PRIO_MED_H, + DMA_PRIO_MED_L, + DMA_PRIO_LOW +}; + +/* Rx channel error handling mode during buffer starvation */ +enum dma_keystone_rx_err_mode { + DMA_DROP = 0, + DMA_RETRY +}; + +/* Rx flow size threshold configuration */ +enum dma_keystone_rx_thresholds { + DMA_THRESH_NONE = 0, + DMA_THRESH_0 = 1, + DMA_THRESH_0_1 = 3, + DMA_THRESH_0_1_2 = 7 +}; + +/* Descriptor type */ +enum dma_keystone_desc_type { + DMA_DESC_HOST = 0, + DMA_DESC_MONOLITHIC = 2 +}; + +/** + * struct dma_keystone_tx_cfg: Tx channel configuration + * @filt_einfo: Filter extended packet info + * @filt_pswords: Filter PS words present + * @dma_keystone_tx_priority: Tx channel scheduling priority + */ +struct dma_keystone_tx_cfg { + bool filt_einfo; + bool filt_pswords; + enum dma_keystone_tx_priority priority; +}; + +/** + * struct dma_keystone_rx_cfg: Rx flow configuration + * @einfo_present: Extended packet info present + * @psinfo_present: PS words present + * @dma_keystone_rx_err_mode: Error during buffer starvation + * @dma_keystone_desc_type: Host or Monolithic desc + * @psinfo_at_sop: PS word located at start of packet + * @sop_offset: Start of packet offset + * @dst_q: Destination queue for a given flow + * @thresh: Rx flow size threshold + * @fdq[]: Free desc Queue array + * @sz_thresh0: RX packet size threshold 0 + * @sz_thresh1: RX packet size threshold 1 + * @sz_thresh2: RX packet size threshold 2 + */ +struct dma_keystone_rx_cfg { + bool einfo_present; + bool psinfo_present; + enum dma_keystone_rx_err_mode err_mode; + enum dma_keystone_desc_type desc_type; + bool psinfo_at_sop; + unsigned int sop_offset; + unsigned int dst_q; + enum dma_keystone_rx_thresholds thresh; + unsigned int fdq[DMA_KEYSTONE_FDQ_PER_CHAN]; + unsigned int sz_thresh0; + unsigned int sz_thresh1; + unsigned int sz_thresh2; +}; + +/** + * struct dma_keystone_cfg: Pktdma channel configuration + * @sl_cfg: Slave configuration + * @tx: Tx channel configuration + * @rx: Rx flow configuration + */ +struct dma_keystone_cfg { + struct dma_slave_config sl_cfg; + union { + struct dma_keystone_tx_cfg tx; + struct dma_keystone_rx_cfg rx; + } u; +}; + +/** + * struct dma_keystone_desc: Host packet descriptor layout + * @desc_info: Descriptor information like id, type, length + * @tag_info: Flow tag info written in during RX + * @packet_info: Queue Manager, policy, flags etc + * @buff_len: Buffer length in bytes + * @buff: Buffer pointer + * @next_desc: For chaining the descriptors + * @orig_len: length since 'buff_len' can be overwritten + * @orig_buff: buff pointer since 'buff' can be overwritten + * @epib: Extended packet info block + * @psdata: Protocol specific + */ +struct dma_keystone_desc { + u32 desc_info; + u32 tag_info; + u32 packet_info; + u32 buff_len; + u32 buff; + u32 next_desc; + u32 orig_len; + u32 orig_buff; + u32 epib[DMA_KEYSTONE_NUM_EPIB_WORDS]; + u32 psdata[DMA_KEYSTONE_NUM_PS_WORDS]; + u32 pad[4]; +} ____cacheline_aligned; + +#define keystone_cfg_to_slave_config(keystone_cfg) (&keystone_cfg.sl_cfg) +#define keystone_cfg_from_slave_config(dma_cfg) container_of(dma_cfg, \ + struct dma_keystone_cfg, sl_cfg) +/* Keystone PKTDMA filter function */ +bool dma_keystone_filter_fn(struct dma_chan *chan, void *param); + +#endif /* __KEYSTONE_PKTDMA_H__ */