On Sun, Jan 12, 2014 at 1:19 PM, Arnd Bergmann <a...@arndb.de> wrote: > On Friday 10 January 2014, Ravi Patel wrote: > >> Do you want any further clarification or document related to QMTM. >> We want to make sure everyone is on same page, understand and >> conclude upon that QMTM is a device and and not a bus or a dma >> engine. > > I have a much better understanding now, but there are still a few open > questions from my side. Let me try to explain in my own words what I > think is the relevant information (part of this is still guessing). > It took me a while to figure out what it does from your description, > and then some more time to see what it's actually good for (as > opposed to adding complexity). > > Please confirm or correct the individual statements in this > description: > > The QMTM serves as a relay for short (a few bytes) messages between > the OS software and various slave hardware blocks on the SoC. > The messages are typically but not always DMA descriptors used by the > slave device used for starting bus master transactions by the slave, > or for notifying sofware about the competion of a DMA transaction. > > The message format is specific to the slave device and the QMTM > only understands the common header of the message. > > OS software sees the messages in cache-coherent memory and does > not require any cache flushes or MMIO access for inbound messages > and only a single posted MMIO write for outbound messages. > > The queues are likely designed to be per-thread and don't require > software-side locking. > > For outbound messages, the QMTM is the bus master of a device-to-device > DMA transaction that gets started once a message is queued and the > device has signaled that it is ready for receiving it. The QMTM needs > to know the bus address of the device as well as a slave ID for > the signal pin. > For inbound messages, the QMTM slave initiates a busmaster transaction > and needs to know the bus address of its QMTM port, while the QMTM > needs to know only the slave ID that is associated with the queue. > > In addition to those hardware properties, the QMTM driver needs to > set up a memory buffer for the message queue as seen by the CPU, > and needs tell the QMTM the location as well as some other > properties such as the message length.
Your description is correct upto this point. > For inbound messages, the QMTM serves a similar purpose as an MSI > controller, ensuring that inbound DMA data has arrived in RAM > before an interrupt is delivered to the CPU and thereby avoiding > the need for an expensive MMIO read to serialize the DMA. For inbound messages, slave device generates message on a completion of a inbound DMA operation or any relevant operation targeted to the CPU. The QMTM's role is to just trigger an interrupt to CPU when there is a new message arrived from a slave device. QMTM doesn't know what the message was for. It is upto the upper layer drivers to decide how to process this message. > The resources managed by the QMTM are both SoC-global (e.g. bus > bandwidth) and slave specific (e.g. ethernet bandwith or buffer space). > Global resource management is performed to prevent one slave > device from monopolizing the system or preventing other slaves > from making forward progress. > Examples for local resource management (I had to think about this > a long time, but probably some of these are wrong) would be > * balancing between multiple non-busmaster devices connected to > a dma-engine > * distributing incoming ethernet data to the available CPUs based on > a flow classifier in the MAC, e.g. by IOV MAC address, VLAN tag > or even individual TCP connection depending on the NIC's capabilities. > * 802.1p flow control for incoming ethernet data based on the amount > of data queued up between the MAC and the driver > * interrupt mitigation for both inbound data and outbound completion, > by delaying the IRQ to the OS until multiple messages have arrived > or a queue specific amount of time has passed. > * controlling the amount of outbound buffer space per flow to minimize > buffer-bloat between an ethernet driver and the NIC hardware. > * reordering data from outbound flows based on priority. > > This is basically my current interpretation, I hope I got at least > some of it right this time ;-) You have got them right. Although we have taken Ethernet examples here, most of the local resource management apply to other slave devices also. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/