Hello Yasir, You need to check your 'PYTHONPATH' environment variable.
For windows, If I do a "pip install", the package is installed in "lib\site-packages" under the python folder. If I "print (sys.path)", I see "lib\site-packages" as one of the entries, and I can expect "import <package>" to work. Find the installation location of 'findspark' and add it to the PYTHONPATH, you can do that inside your script as well like: import sys sys.path.append('X:\PathTo\findspark\module') Hope it works. Regards, Akchhaya Sharma On Fri, Apr 10, 2020 at 4:35 PM Yasir Elgohary <ygoh...@gmail.com> wrote: > Peace dear all, > > I hope you all are well and healthy... > > I am brand new to Spark/Hadoop. My env. is: Windows 7 with > Jupyter/Anaconda and Spark/Hadoop all installed on my laptop. How can I run > the following without errors: > > import findspark > findspark.init() > findspark.find() > from pyspark.sql import SparkSession > > This is the error msg. I get: > > ModuleNotFoundError: No module named 'findspark' > > > It seems I missing something for Spark to run well with Jupyter/Anaconda on > Windows 7. > > > Cheers > > > > > > Cheers >