Hi Chetan,

You can use

spark-submit showDF.py | hadoop fs -put - showDF.txt

showDF.py:

from pyspark.sql import SparkSession


spark = SparkSession.builder.appName("Write stdout").getOrCreate()

spark.sparkContext.setLogLevel("OFF")


spark.table("<yourdf>").show(100,truncate=false)

But is there any specific reason you want to write it to hdfs? Is this for
human consumption?

Regards,
Nuthan

On Sat, Apr 13, 2019 at 6:41 PM Chetan Khatri <chetan.opensou...@gmail.com>
wrote:

> Hello Users,
>
> In spark when I have a DataFrame and do  .show(100) the output which gets
> printed, I wants to save as it is content to txt file in HDFS.
>
> How can I do this?
>
> Thanks
>


-- 
Nuthan Reddy
Sigmoid Analytics

-- 
*Disclaimer*: This is not a mass e-mail and my intention here is purely 
from a business perspective, and not to spam or encroach your privacy. I am 
writing with a specific agenda to build a personal business connection. 
Being a reputed and genuine organization, Sigmoid respects the digital 
security of every prospect and tries to comply with GDPR and other regional 
laws. Please let us know if you feel otherwise and we will rectify the 
misunderstanding and adhere to comply in the future. In case we have missed 
any of the compliance, it is completely unintentional.

Reply via email to