2MD opened a new issue, #12131:
URL: https://github.com/apache/iceberg/issues/12131
### Apache Iceberg version
1.7.1 (latest release)
### Query engine
Spark
### Please describe the bug 🐞
```
val icebergVersion = "1.7.1"
val sparkVersion = "3.3.2"
lazy val icebergDependencies = Seq(
"org.apache.iceberg" %% "iceberg-spark-runtime-3.3" % icebergVersion,
"org.scala-lang.modules" % "scala-collection-compat_2.12" % "2.12.0",
"org.apache.iceberg" % "iceberg-aws-bundle" % icebergVersion,
"org.apache.iceberg" % "iceberg-hive-metastore" % icebergVersion
)
"org.apache.hive" % "hive-metastore" % "3.1.3" % Provided,
"org.apache.spark" %% "spark-hive" % sparkVersion % Provided,
("org.apache.hive" % "hive-exec" % "3.1.3").classifier("core") % Provided,
("org.apache.iceberg" % "iceberg-hive-metastore" % icebergVersion %
IntegrationTest).classifier("tests")
```
I put hive-schema-3.1.0.derby.sql in my resources folder
project/
├── src/
│ ├── it/
│ │ ├── resources/
│ │ │ └── hive-schema-3.1.0.derby.sql
I have some test
```
var metastore: TestHiveMetastore = _
var uris: String = _
implicit var spark: SparkSession = _
override def beforeAll(): Unit = {
super.beforeAll()
metastore = new TestHiveMetastore()
metastore.start()
uris = metastore.hiveConf().get("hive.metastore.uris")
spark = SparkSession.builder
.master("local[*]")
.config("spark.hadoop.hive.metastore.uris", uris)
.config("spark.sql.legacy.parquet.nanosAsLong", "false")
.config("spark.sql.extensions",
"org.apache.iceberg.spark.extensions.IcebergSparkSessionExtensions")
.config(s"spark.sql.catalog.$icebergCatalog",
"org.apache.iceberg.spark.SparkCatalog")
.config(s"spark.sql.catalog.$icebergCatalog.type", "hive")
.config("spark.hadoop.fs.s3a.endpoint", minioContainer.getHostAddress)
.config("spark.hadoop.fs.s3a.access.key",
minioContainer.getMinioAccessValue)
.config("spark.hadoop.fs.s3a.secret.key",
minioContainer.getMinioAccessValue)
.config("spark.hadoop.fs.s3a.path.style.access", "true")
.config("spark.hadoop.fs.s3a.impl",
"org.apache.hadoop.fs.s3a.S3AFileSystem")
.config("spark.sql.iceberg.check-ordering", "false")
.config("write.metadata.delete-after-commit.enabled", "false")
.config("spark.sql.iceberg.handle-timestamp-without-timezone", "true")
.enableHiveSupport()
.getOrCreate()
s3.createBucket(bucket)
}
```
I run my test : > sbt "IntegrationTest/testOnly"
```
Caused by: java.lang.NullPointerException
[error] at java.base/java.io.Reader.(Reader.java:167)
[error] at java.base/java.io.InputStreamReader.(InputStreamReader.java:72)
[error] at
org.apache.iceberg.hive.TestHiveMetastore.setupMetastoreDB(TestHiveMetastore.java:281)
[error] at
org.apache.iceberg.hive.TestHiveMetastore.(TestHiveMetastore.java:106)
[error] at
ru.samokat.datalake.loader.BaseLoadSuite.beforeAll(BaseLoadSuite.scala:49)
[error] at
ru.samokat.datalake.loader.BaseLoadSuite.beforeAll$(BaseLoadSuite.scala:46)
```
He can't find hive-schema-3.1.0.derby.sql because in code loaded file across
SystemClassLoader
### Willingness to contribute
- [x] I can contribute a fix for this bug independently
- [ ] I would be willing to contribute a fix for this bug with guidance from
the Iceberg community
- [ ] I cannot contribute a fix for this bug at this time
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]