On 08/23/2017 06:00 PM, Vinod Koul wrote:
> On Tue, Aug 22, 2017 at 05:59:26PM +0200, Pierre Yves MORDRET wrote:
>>
>>
>>>
>>> #define STM32_MDMA_SHIFT(n)         (ffs(n) - 1))
>>> #define STM32_MDMA_SET(n, mask)             ((n) << STM32_MDMA_SHIFT(mask))
>>> #define STM32_MDMA_GET(n, mask)             (((n) && mask) >> 
>>> STM32_MDMA_SHIFT(mask))
>>>
>>> Basically, u extract the shift using the mask value and ffs helping out, so
>>> no need to define these and reduce chances of coding errors...
>>>
>>
>> OK.
>> but I would prefer if you don't mind
> 
> hmmm, I don't have a very strong opinion, so okay. But from programming PoV
> it reduces human errors..
> 
>> #define STM32_MDMA_SET(n, mask)              (((n) << 
>> STM32_MDMA_SHIFT(mask)) & mask)
>>

I agree with your proposal :) I just said I prefer for the setter
#define STM32_MDMA_SET(n, mask)         (((n) << STM32_MDMA_SHIFT(mask)) & mask)
instead of
#define STM32_MDMA_SET(n, mask)         ((n) << STM32_MDMA_SHIFT(mask))

>>>> +static int stm32_mdma_get_width(struct stm32_mdma_chan *chan,
>>>> +                          enum dma_slave_buswidth width)
>>>> +{
>>>> +  switch (width) {
>>>> +  case DMA_SLAVE_BUSWIDTH_1_BYTE:
>>>> +  case DMA_SLAVE_BUSWIDTH_2_BYTES:
>>>> +  case DMA_SLAVE_BUSWIDTH_4_BYTES:
>>>> +  case DMA_SLAVE_BUSWIDTH_8_BYTES:
>>>> +          return ffs(width) - 1;
>>>> +  default:
>>>> +          dev_err(chan2dev(chan), "Dma bus width not supported\n");
>>>
>>> please log the width here, helps in debug...
>>>
>>
>> Hum.. just a dev_dbg to log the actual width or within the dev_err ?
> 
> latter pls
> 

OK

>>>> +static struct dma_async_tx_descriptor *
>>>> +stm32_mdma_prep_dma_cyclic(struct dma_chan *c, dma_addr_t buf_addr,
>>>> +                     size_t buf_len, size_t period_len,
>>>> +                     enum dma_transfer_direction direction,
>>>> +                     unsigned long flags)
>>>> +{
>>>> +  struct stm32_mdma_chan *chan = to_stm32_mdma_chan(c);
>>>> +  struct stm32_mdma_device *dmadev = stm32_mdma_get_dev(chan);
>>>> +  struct dma_slave_config *dma_config = &chan->dma_config;
>>>> +  struct stm32_mdma_desc *desc;
>>>> +  dma_addr_t src_addr, dst_addr;
>>>> +  u32 ccr, ctcr, ctbr, count;
>>>> +  int i, ret;
>>>> +
>>>> +  if (!buf_len || !period_len || period_len > STM32_MDMA_MAX_BLOCK_LEN) {
>>>> +          dev_err(chan2dev(chan), "Invalid buffer/period len\n");
>>>> +          return NULL;
>>>> +  }
>>>> +
>>>> +  if (buf_len % period_len) {
>>>> +          dev_err(chan2dev(chan), "buf_len not multiple of period_len\n");
>>>> +          return NULL;
>>>> +  }
>>>> +
>>>> +  /*
>>>> +   * We allow to take more number of requests till DMA is
>>>> +   * not started. The driver will loop over all requests.
>>>> +   * Once DMA is started then new requests can be queued only after
>>>> +   * terminating the DMA.
>>>> +   */
>>>> +  if (chan->busy) {
>>>> +          dev_err(chan2dev(chan), "Request not allowed when dma busy\n");
>>>> +          return NULL;
>>>> +  }
>>>
>>> is that a HW restriction? Once a txn is completed can't we submit
>>> subsequent txn..? Can you explain this part please.
>>>
>>
>> Driver can prepare any request Slave SG, Memcpy or Cyclic. But if the 
>> channel is
>> busy to complete a DMA transfer, the request will be put in pending list. 
>> This
>> is only when the DMA transfer is going to be completed the next descriptor is
>> going to be processed and started.
>> However for cyclic this is different since when cyclic is ignited the channel
>> will be busy until its termination. This is why we forbid any DMA preparation
>> for this channel.
>> Nonetheless I believe we have a flaw here since we have to forbid
>> Slave/Memcpy/Cyclic whether a cyclic request is on-going.
> 
> But you are not submitting a txn to HW. The prepare_xxx operation prepares a
> descriptor which is pushed to pending queue on submit and further pushed to
> hw on queue move or issue_pending()
> 
> So here we should ideally accept the request.
> 
> After you finish memcpy you can submit a memcpy right...?
> 

Correct TXn is not submitted to HW only descriptors are prepared. They are going
to be pushed to HW on issue_pending if not channel not running. Otherwise the
next pending request is pushed to HW on DMA completion.
So yes I can queue a memcpy can be submitted during a memcpy (or Slave SG). The
mempy will be pushed to HW after the completion of the preceding.

The only drawback is cyclic. If a cyclic DMA operation is running, no other
requests (cyclic or SG or memcpy) can be granted or queued.

>>
>>
>>>> +  ret = device_property_read_u32(&pdev->dev, "dma-requests",
>>>> +                                 &nr_requests);
>>>> +  if (ret) {
>>>> +          nr_requests = STM32_MDMA_MAX_REQUESTS;
>>>> +          dev_warn(&pdev->dev, "MDMA defaulting on %i request lines\n",
>>>> +                   nr_requests);
>>>> +  }
>>>> +
>>>> +  count = of_property_count_u32_elems(of_node, "st,ahb-addr-masks");
>>>
>>> We dont have device_property_xxx for this?
>>
>> Sorry no. Well didn't figure out one though.
> 
> we do :) the array property with NULL argument returns the size of array..
> 
> int device_property_read_u32_array(struct device *dev, const char *propname,
>                                  u32 *val, size_t nval)
> 
> Documentation says:
>  Return: number of values if @val was %NULL,
> 

Dho. ok :)

>>
>>>
>>>> +  if (count < 0)
>>>> +          count = 0;
>>>> +
>>>> +  dmadev = devm_kzalloc(&pdev->dev, sizeof(*dmadev) + sizeof(u32) * count,
>>>> +                        GFP_KERNEL);
>>>> +  if (!dmadev)
>>>> +          return -ENOMEM;
>>>> +
>>>> +  dmadev->nr_channels = nr_channels;
>>>> +  dmadev->nr_requests = nr_requests;
>>>> +  of_property_read_u32_array(of_node, "st,ahb-addr-masks",
>>>> +                             dmadev->ahb_addr_masks,
>>>> +                             count);
>>>
>>> i know we have an device api for array reads :)
>>> and I think that helps in former case..
>>>
>>
>> Correct :) device_property_read_u32_array
> 
> yes..
> 

Thanks
Py

Reply via email to