Hi,
Tested for calculating values for 300 columns. Analyser takes around 4
minutes to generate the plan. Is this normal?




Regards,
Madhukara Phatak
http://datamantra.io/

On Tue, May 19, 2015 at 4:35 PM, madhu phatak <phatak....@gmail.com> wrote:

> Hi,
> I am using spark 1.3.1
>
>
>
>
> Regards,
> Madhukara Phatak
> http://datamantra.io/
>
> On Tue, May 19, 2015 at 4:34 PM, Wangfei (X) <wangf...@huawei.com> wrote:
>
>>  And which version are you using
>>
>> 发自我的 iPhone
>>
>> 在 2015年5月19日,18:29,"ayan guha" <guha.a...@gmail.com> 写道:
>>
>>   can you kindly share your code?
>>
>> On Tue, May 19, 2015 at 8:04 PM, madhu phatak <phatak....@gmail.com>
>> wrote:
>>
>>> Hi,
>>> I  am trying run spark sql aggregation on a file with 26k columns. No of
>>> rows is very small. I am running into issue that spark is taking huge
>>> amount of time to parse the sql and create a logical plan. Even if i have
>>> just one row, it's taking more than 1 hour just to get pass the parsing.
>>> Any idea how to optimize in these kind of scenarios?
>>>
>>>
>>>  Regards,
>>>  Madhukara Phatak
>>> http://datamantra.io/
>>>
>>
>>
>>
>>  --
>> Best Regards,
>> Ayan Guha
>>
>>
>

Reply via email to