This is an automated email from the ASF dual-hosted git repository. ngangam pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/hive.git
commit 0b9a4f4304d4b7c7a6ab77fad1b1bae26cb7eede Author: Naveen Gangam <ngan...@cloudera.com> AuthorDate: Tue Nov 17 00:42:47 2020 -0500 HIVE-24396: Build failure in itests due to unimplemented interface methods --- .../hcatalog/listener/DummyRawStoreFailEvent.java | 27 ++++++++++++++++++++++ .../ql/ddl/database/create/CreateDatabaseDesc.java | 3 +-- .../database/create/CreateDatabaseOperation.java | 3 +-- 3 files changed, 29 insertions(+), 4 deletions(-) diff --git a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/DummyRawStoreFailEvent.java b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/DummyRawStoreFailEvent.java index f93a3c7..2d7ad24 100644 --- a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/DummyRawStoreFailEvent.java +++ b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/DummyRawStoreFailEvent.java @@ -51,6 +51,7 @@ import org.apache.hadoop.hive.metastore.api.AlreadyExistsException; import org.apache.hadoop.hive.metastore.api.ColumnStatistics; import org.apache.hadoop.hive.metastore.api.CreationMetadata; import org.apache.hadoop.hive.metastore.api.CurrentNotificationEventId; +import org.apache.hadoop.hive.metastore.api.DataConnector; import org.apache.hadoop.hive.metastore.api.Database; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.metastore.api.FileMetadataExprType; @@ -245,6 +246,32 @@ public class DummyRawStoreFailEvent implements RawStore, Configurable { } @Override + public List<String> getAllDataConnectors() throws MetaException { + return objectStore.getAllDataConnectors(); + } + + @Override + public DataConnector getDataConnector(String connectorName) throws NoSuchObjectException { + return objectStore.getDataConnector(connectorName); + } + + @Override + public boolean alterDataConnector(String connectorName, DataConnector connector) + throws MetaException, NoSuchObjectException { + return objectStore.alterDataConnector(connectorName, connector); + } + + @Override + public boolean dropDataConnector(String connector) throws MetaException, NoSuchObjectException { + return objectStore.dropDataConnector(connector); + } + + @Override + public void createDataConnector(DataConnector connector) throws MetaException, InvalidObjectException { + objectStore.createDataConnector(connector); + } + + @Override public boolean createType(Type type) { return objectStore.createType(type); } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/ddl/database/create/CreateDatabaseDesc.java b/ql/src/java/org/apache/hadoop/hive/ql/ddl/database/create/CreateDatabaseDesc.java index 45df31c..1590133 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/ddl/database/create/CreateDatabaseDesc.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/ddl/database/create/CreateDatabaseDesc.java @@ -34,7 +34,6 @@ import javax.xml.crypto.Data; @Explain(displayName = "Create Database", explainLevels = { Level.USER, Level.DEFAULT, Level.EXTENDED }) public class CreateDatabaseDesc implements DDLDesc, Serializable { private static final long serialVersionUID = 1L; - public static final String REMOTEDB_LOCATION = "REMOTE_LOCATION".intern(); private final String databaseName; private final String comment; @@ -57,9 +56,9 @@ public class CreateDatabaseDesc implements DDLDesc, Serializable { this.comment = comment; if (dbtype != null && dbtype.equalsIgnoreCase("REMOTE")) { this.dbType = DatabaseType.REMOTE; - this.locationUri = REMOTEDB_LOCATION; // this is non-null in the HMSDB this.connectorName = connectorName; this.remoteDbName = remoteDbName; + this.locationUri = null; this.managedLocationUri = null; } else { this.dbType = DatabaseType.NATIVE; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/ddl/database/create/CreateDatabaseOperation.java b/ql/src/java/org/apache/hadoop/hive/ql/ddl/database/create/CreateDatabaseOperation.java index 97e318d..def742d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/ddl/database/create/CreateDatabaseOperation.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/ddl/database/create/CreateDatabaseOperation.java @@ -54,15 +54,14 @@ public class CreateDatabaseOperation extends DDLOperation<CreateDatabaseDesc> { if (desc.getManagedLocationUri() != null) { database.setManagedLocationUri(desc.getManagedLocationUri()); } - makeLocationQualified(database); if (database.getLocationUri().equalsIgnoreCase(database.getManagedLocationUri())) { throw new HiveException("Managed and external locations for database cannot be the same"); } } else { - database.setLocationUri(CreateDatabaseDesc.REMOTEDB_LOCATION); database.setConnector_name(desc.getConnectorName()); database.setRemote_dbname(desc.getRemoteDbName()); } + makeLocationQualified(database); context.getDb().createDatabase(database, desc.getIfNotExists()); } catch (AlreadyExistsException ex) { //it would be better if AlreadyExistsException had an errorCode field....