Check your hive-site.xml
What is this set to
<property>
<name>hive.metastore.uris</name>
<value>thrift://rhes75:9083</value>
<description>Thrift URI for the remote metastore. Used by metastore
client to connect to remote metastore.</description>
</property>
if you find any S3-related configurations (like fs.defaultFS pointing to
S3) in hive-site.xml, that could be the cause of the problem. Remove or
correct those settings. The S3 configuration should be handled entirely
within your Spark application's configuration.
HTH
Dr Mich Talebzadeh,
Architect | Data Science | Financial Crime | Forensic Analysis | GDPR
view my Linkedin profile
<https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
On Thu, 30 Jan 2025 at 13:15, Márkus Andor Rudolf <[email protected]>
wrote:
> Hi Hive team,
>
> I'm encountering an issue where Hive Metastore (HMS), running as a Spark
> sidecar container, is attempting to read/write to S3 even though it's only
> being used as a metastore. Since HMS is solely functioning as a metadata
> service in this setup, these S3 operations seem unnecessary.
>
> Can someone help explain:
>
> 1. How to completely disable S3 access for HMS?
> 2. Are there specific configuration parameters I should modify?
> 3. Is there any potential impact on functionality if S3 access is
> disabled?
>
>
> Current setup:
>
> - HMS running as a Spark sidecar container
> - Only using HMS for metadata storage
> - No direct data storage requirements for HMS
>
>
> Thanks in advance for your help.
>
> Best regards,
> Andor Markus
>