I ran the following code
spark.sparkContext.list_packages()
on spark 3.4.1 and i get below error
An error was encountered:
AttributeError
[Traceback (most recent call last):
, File "/tmp/spark-3d66c08a-08a3-4d4e-9fdf-45853f65e03d/shell_wrapper.py",
line 113, in exec
self._exec_then_eval(code)
, File "/tmp/spark-3d66c08a-08a3-4d4e-9fdf-45853f65e03d/shell_wrapper.py",
line 106, in _exec_then_eval
exec(compile(last, '<string>', 'single'), self.globals)
, File "<string>", line 1, in <module>
, AttributeError: 'SparkContext' object has no attribute 'list_packages'
]
Is list_packages and install_pypi_package available for vanilla spark or only
available for AWS services?
Thank you