Hello Yasir,

You need to check your 'PYTHONPATH' environment variable.

For windows, If I do a "pip install", the package is installed in
"lib\site-packages" under the python folder. If I "print (sys.path)", I see
"lib\site-packages" as one of the entries, and I can expect "import
<package>" to work.

Find the installation location of 'findspark' and add it to the PYTHONPATH,
you can do that inside your script as well like:

import sys
sys.path.append('X:\PathTo\findspark\module')

Hope it works.

Regards,
Akchhaya Sharma


On Fri, Apr 10, 2020 at 4:35 PM Yasir Elgohary <ygoh...@gmail.com> wrote:

> Peace dear all,
>
> I hope you all are well and healthy...
>
> I am brand new to Spark/Hadoop. My env. is: Windows 7 with
> Jupyter/Anaconda and Spark/Hadoop all installed on my laptop. How can I run
> the following without errors:
>
> import findspark
> findspark.init()
> findspark.find()
> from pyspark.sql import SparkSession
>
> This is the error msg. I get:
>
> ModuleNotFoundError: No module named 'findspark'
>
>
> It seems I missing something for Spark to run well with Jupyter/Anaconda on 
> Windows 7.
>
>
> Cheers
>
>
>
>
>
> Cheers
>

Reply via email to