[jira] [Commented] (FLINK-10556) Integration with Apache Hive
[ https://issues.apache.org/jira/browse/FLINK-10556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16986612#comment-16986612 ] Bowen Li commented on FLINK-10556: -- [~lzljs3620320] I think it can be closed now > Integration with Apache Hive > > > Key: FLINK-10556 > URL: https://issues.apache.org/jira/browse/FLINK-10556 > Project: Flink > Issue Type: New Feature > Components: Connectors / Hive, Table SQL / Ecosystem >Affects Versions: 1.6.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang >Priority: Major > Attachments: Proposal_ Integrate Flink with Hive Ecosystem.pdf > > > This is an umbrella JIRA tracking all enhancement and issues related to > integrating Flink with Hive ecosystem. This is an outcome of a discussion in > the community, and thanks go to everyone that provided feedback and interest. > Specifically, we'd like to see the following features and capabilities > immediately in Flink: > # Metadata interoperability > # Data interoperability > # Data type compatibility > # Hive UDF support > # DDL/DML/Query language compatibility > For a longer term, we'd also like to add or improve: > # Compatible SQL service, client tools, JDBC/ODBC drivers > # Better task failure tolerance and task scheduling > # Support other user customizations in Hive (storage handlers, serdes, etc). > I will provide more details regarding the proposal in a doc shortly. Design > doc, if deemed necessary, will be provided in each related sub tasks under > this JIRA. > Feedback and contributions are greatly welcome! -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-10556) Integration with Apache Hive
[ https://issues.apache.org/jira/browse/FLINK-10556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16986552#comment-16986552 ] Jingsong Lee commented on FLINK-10556: -- Hi [~xuefuz] [~phoenixjiangnan], we have already done many major work on hive integration, shall we close this ticket now? > Integration with Apache Hive > > > Key: FLINK-10556 > URL: https://issues.apache.org/jira/browse/FLINK-10556 > Project: Flink > Issue Type: New Feature > Components: Connectors / Hive, Table SQL / Ecosystem >Affects Versions: 1.6.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang >Priority: Major > Attachments: Proposal_ Integrate Flink with Hive Ecosystem.pdf > > > This is an umbrella JIRA tracking all enhancement and issues related to > integrating Flink with Hive ecosystem. This is an outcome of a discussion in > the community, and thanks go to everyone that provided feedback and interest. > Specifically, we'd like to see the following features and capabilities > immediately in Flink: > # Metadata interoperability > # Data interoperability > # Data type compatibility > # Hive UDF support > # DDL/DML/Query language compatibility > For a longer term, we'd also like to add or improve: > # Compatible SQL service, client tools, JDBC/ODBC drivers > # Better task failure tolerance and task scheduling > # Support other user customizations in Hive (storage handlers, serdes, etc). > I will provide more details regarding the proposal in a doc shortly. Design > doc, if deemed necessary, will be provided in each related sub tasks under > this JIRA. > Feedback and contributions are greatly welcome! -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-10556) Integration with Apache Hive
[ https://issues.apache.org/jira/browse/FLINK-10556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16670873#comment-16670873 ] Xuefu Zhang commented on FLINK-10556: - The Google doc for the proposal can be found [here|https://docs.google.com/document/d/1kcYxnekm0j8pOrLDSdW9tNSlam7mtmFtCzBOZMTUYOs/edit?usp=sharing]. Relevant design docs will be published under child issues of this JIRA (FLINK-10744 and FLINK-10729). > Integration with Apache Hive > > > Key: FLINK-10556 > URL: https://issues.apache.org/jira/browse/FLINK-10556 > Project: Flink > Issue Type: New Feature > Components: Batch Connectors and Input/Output Formats, SQL Client, > Table API SQL >Affects Versions: 1.6.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang >Priority: Major > Attachments: Proposal_ Integrate Flink with Hive Ecosystem.pdf > > > This is an umbrella JIRA tracking all enhancement and issues related to > integrating Flink with Hive ecosystem. This is an outcome of a discussion in > the community, and thanks go to everyone that provided feedback and interest. > Specifically, we'd like to see the following features and capabilities > immediately in Flink: > # Metadata interoperability > # Data interoperability > # Data type compatibility > # Hive UDF support > # DDL/DML/Query language compatibility > For a longer term, we'd also like to add or improve: > # Compatible SQL service, client tools, JDBC/ODBC drivers > # Better task failure tolerance and task scheduling > # Support other user customizations in Hive (storage handlers, serdes, etc). > I will provide more details regarding the proposal in a doc shortly. Design > doc, if deemed necessary, will be provided in each related sub tasks under > this JIRA. > Feedback and contributions are greatly welcome! -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-10556) Integration with Apache Hive
[ https://issues.apache.org/jira/browse/FLINK-10556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16670683#comment-16670683 ] Bowen Li commented on FLINK-10556: -- I agree. It's not urgent to consider for now, and it can be figured out along the way. Thanks for sharing a google doc version! it's easier for everyone to comment. > Integration with Apache Hive > > > Key: FLINK-10556 > URL: https://issues.apache.org/jira/browse/FLINK-10556 > Project: Flink > Issue Type: New Feature > Components: Batch Connectors and Input/Output Formats, SQL Client, > Table API SQL >Affects Versions: 1.6.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang >Priority: Major > Attachments: Flink-Hive Metastore Connectivity Design.pdf, Proposal_ > Integrate Flink with Hive Ecosystem.pdf > > > This is an umbrella JIRA tracking all enhancement and issues related to > integrating Flink with Hive ecosystem. This is an outcome of a discussion in > the community, and thanks go to everyone that provided feedback and interest. > Specifically, we'd like to see the following features and capabilities > immediately in Flink: > # Metadata interoperability > # Data interoperability > # Data type compatibility > # Hive UDF support > # DDL/DML/Query language compatibility > For a longer term, we'd also like to add or improve: > # Compatible SQL service, client tools, JDBC/ODBC drivers > # Better task failure tolerance and task scheduling > # Support other user customizations in Hive (storage handlers, serdes, etc). > I will provide more details regarding the proposal in a doc shortly. Design > doc, if deemed necessary, will be provided in each related sub tasks under > this JIRA. > Feedback and contributions are greatly welcome! -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-10556) Integration with Apache Hive
[ https://issues.apache.org/jira/browse/FLINK-10556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16670609#comment-16670609 ] Xuefu Zhang commented on FLINK-10556: - [~phoenixjiangnan] Thanks for your feedback. Re: {{FlinkYamlCatalog}} idea, while I'm not entirely sure of the need, as I can image it's hard to implement or use. I think a migration path might be sufficient for now. However, if there is a strong request coming in the future, the current design can certainly accommodate that at least. Also, there is a link above of the design doc. Please feel free to share your input over there. > Integration with Apache Hive > > > Key: FLINK-10556 > URL: https://issues.apache.org/jira/browse/FLINK-10556 > Project: Flink > Issue Type: New Feature > Components: Batch Connectors and Input/Output Formats, SQL Client, > Table API SQL >Affects Versions: 1.6.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang >Priority: Major > Attachments: Flink-Hive Metastore Connectivity Design.pdf, Proposal_ > Integrate Flink with Hive Ecosystem.pdf > > > This is an umbrella JIRA tracking all enhancement and issues related to > integrating Flink with Hive ecosystem. This is an outcome of a discussion in > the community, and thanks go to everyone that provided feedback and interest. > Specifically, we'd like to see the following features and capabilities > immediately in Flink: > # Metadata interoperability > # Data interoperability > # Data type compatibility > # Hive UDF support > # DDL/DML/Query language compatibility > For a longer term, we'd also like to add or improve: > # Compatible SQL service, client tools, JDBC/ODBC drivers > # Better task failure tolerance and task scheduling > # Support other user customizations in Hive (storage handlers, serdes, etc). > I will provide more details regarding the proposal in a doc shortly. Design > doc, if deemed necessary, will be provided in each related sub tasks under > this JIRA. > Feedback and contributions are greatly welcome! -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-10556) Integration with Apache Hive
[ https://issues.apache.org/jira/browse/FLINK-10556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16670599#comment-16670599 ] Xuefu Zhang commented on FLINK-10556: - Google doc for the design can be found at https://docs.google.com/document/d/1SkppRD_rE3uOKSN-LuZCqn4f7dz0zW5aa6T_hBZq5_o/edit?usp=sharing. Feedback is greatly appreciated. > Integration with Apache Hive > > > Key: FLINK-10556 > URL: https://issues.apache.org/jira/browse/FLINK-10556 > Project: Flink > Issue Type: New Feature > Components: Batch Connectors and Input/Output Formats, SQL Client, > Table API SQL >Affects Versions: 1.6.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang >Priority: Major > Attachments: Flink-Hive Metastore Connectivity Design.pdf, Proposal_ > Integrate Flink with Hive Ecosystem.pdf > > > This is an umbrella JIRA tracking all enhancement and issues related to > integrating Flink with Hive ecosystem. This is an outcome of a discussion in > the community, and thanks go to everyone that provided feedback and interest. > Specifically, we'd like to see the following features and capabilities > immediately in Flink: > # Metadata interoperability > # Data interoperability > # Data type compatibility > # Hive UDF support > # DDL/DML/Query language compatibility > For a longer term, we'd also like to add or improve: > # Compatible SQL service, client tools, JDBC/ODBC drivers > # Better task failure tolerance and task scheduling > # Support other user customizations in Hive (storage handlers, serdes, etc). > I will provide more details regarding the proposal in a doc shortly. Design > doc, if deemed necessary, will be provided in each related sub tasks under > this JIRA. > Feedback and contributions are greatly welcome! -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-10556) Integration with Apache Hive
[ https://issues.apache.org/jira/browse/FLINK-10556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16669607#comment-16669607 ] Bowen Li commented on FLINK-10556: -- sorry that I put the wrong jira number in one of my PRs and got linked here... Thanks, Xuefu! The design doc looks very good to me! One small thing caught my eyes is external catalog's compatibility towards current YAML config file used by SQL Client, as mentioned in section `Additional Notes`. The parsing logic of YAML config file currently is on a different code path than catalog's, and the diverge will create maintenance issue. Providing a migration utility, as suggested in design doc, is one way to solve it, while we still need to maintain the utility tool. In the long run, I'd suggest unifying the code paths of YAML config file and catalog by developing something like a {{FlinkYamlCatalog}}. It will make both maintenance and migration a lot easier. What do you think? > Integration with Apache Hive > > > Key: FLINK-10556 > URL: https://issues.apache.org/jira/browse/FLINK-10556 > Project: Flink > Issue Type: New Feature > Components: Batch Connectors and Input/Output Formats, SQL Client, > Table API SQL >Affects Versions: 1.6.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang >Priority: Major > Labels: pull-request-available > Attachments: Flink-Hive Metastore Connectivity Design.pdf, Proposal_ > Integrate Flink with Hive Ecosystem.pdf > > > This is an umbrella JIRA tracking all enhancement and issues related to > integrating Flink with Hive ecosystem. This is an outcome of a discussion in > the community, and thanks go to everyone that provided feedback and interest. > Specifically, we'd like to see the following features and capabilities > immediately in Flink: > # Metadata interoperability > # Data interoperability > # Data type compatibility > # Hive UDF support > # DDL/DML/Query language compatibility > For a longer term, we'd also like to add or improve: > # Compatible SQL service, client tools, JDBC/ODBC drivers > # Better task failure tolerance and task scheduling > # Support other user customizations in Hive (storage handlers, serdes, etc). > I will provide more details regarding the proposal in a doc shortly. Design > doc, if deemed necessary, will be provided in each related sub tasks under > this JIRA. > Feedback and contributions are greatly welcome! -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-10556) Integration with Apache Hive
[ https://issues.apache.org/jira/browse/FLINK-10556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16669582#comment-16669582 ] ASF GitHub Bot commented on FLINK-10556: bowenli86 closed pull request #6969: [FLINK-10556][Table API & SQL]Add APIs to ExternalCatalog, CrudExternalCatalog and InMemoryCrudExternalCatalog for views and UDFs URL: https://github.com/apache/flink/pull/6969 This is a PR merged from a forked repository. As GitHub hides the original diff on merge, it is displayed below for the sake of provenance: As this is a foreign pull request (from a fork), the diff is supplied below (as it won't show otherwise due to GitHub magic): diff --git a/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/api/exceptions.scala b/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/api/exceptions.scala index 7fc7de50e07..0f1c80df544 100644 --- a/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/api/exceptions.scala +++ b/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/api/exceptions.scala @@ -107,6 +107,70 @@ case class CatalogAlreadyExistException( def this(catalog: String) = this(catalog, null) } +/** + * Exception for operation on a nonexistent view + * + * @param catalog catalog name + * @param view view name + * @param cause the cause + */ +case class ViewNotExistException( + catalog: String, + view: String, + cause: Throwable) + extends RuntimeException(s"View $view does not exist.", cause) { + + def this(catalog: String, view: String) = this(catalog, view, null) +} + +/** + * Exception for adding an already existent view + * + * @param catalog catalog name + * @param view view name + * @param cause the cause + */ +case class ViewAlreadyExistException( + catalog: String, + view: String, + cause: Throwable) + extends RuntimeException(s"View $view already exists.", cause) { + + def this(catalog: String, view: String) = this(catalog, view, null) +} + +/** + * Exception for operation on a nonexistent function + * + * @param catalog catalog name + * @param function function name + * @param cause the cause + */ +case class FunctionNotExistException( + catalog: String, + function: String, + cause: Throwable) + extends RuntimeException(s"Function $function does not exist.", cause) { + + def this(catalog: String, function: String) = this(catalog, function, null) +} + +/** + * Exception for adding an already existent function + * + * @param catalog catalog name + * @param function function name + * @param cause the cause + */ +case class FunctionAlreadyExistException( + catalog: String, + function: String, + cause: Throwable) + extends RuntimeException(s"Function $function already exists.", cause) { + + def this(catalog: String, function: String) = this(catalog, function, null) +} + /** * Exception for not finding a [[TableFactory]] for the given properties. * diff --git a/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/catalog/CrudExternalCatalog.scala b/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/catalog/CrudExternalCatalog.scala index 4db9497a712..02bdbed21fc 100644 --- a/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/catalog/CrudExternalCatalog.scala +++ b/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/catalog/CrudExternalCatalog.scala @@ -19,9 +19,11 @@ package org.apache.flink.table.catalog import org.apache.flink.table.api._ +import org.apache.flink.table.functions.UserDefinedFunction /** - * The CrudExternalCatalog provides methods to create, drop, and alter (sub-)catalogs or tables. + * The CrudExternalCatalog provides methods to create, drop, and alter (sub-)catalogs, tables, + * views and UDFs. */ trait CrudExternalCatalog extends ExternalCatalog { @@ -103,4 +105,86 @@ trait CrudExternalCatalog extends ExternalCatalog { @throws[CatalogNotExistException] def alterSubCatalog(name: String, catalog: ExternalCatalog, ignoreIfNotExists: Boolean): Unit + /** +* Adds a view to this catalog. +* +* @param viewName The name of the view to add. +* @param view The view to add. +* @param ignoreIfExists Flag to specify behavior if a view with the given name already exists: +* if set to false, throw an exception, +* if set to true, nothing happens. +* @throws ViewAlreadyExistException thrown if view already exists and ignoreIfExists is false +*/ + @throws[ViewAlreadyExistException] + def createView(viewName: String, view: String, ignoreIfExists: Boolean): Unit + + /** +* Deletes a view from this catalog. +* +* @param viewName Name of the view to delete. +* @param ignoreIfNotExists Flag to specify behavior if the view does not exist:
[jira] [Commented] (FLINK-10556) Integration with Apache Hive
[ https://issues.apache.org/jira/browse/FLINK-10556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16669576#comment-16669576 ] ASF GitHub Bot commented on FLINK-10556: bowenli86 opened a new pull request #6969: [FLINK-10556][Table API & SQL]Add APIs to ExternalCatalog, CrudExternalCatalog and InMemoryCrudExternalCatalog for views and UDFs URL: https://github.com/apache/flink/pull/6969 ## What is the purpose of the change Currently Flink's external catalog have APIs for tables only. However, views and UDFs are also common objects in a catalog. Adding initial APIs and in-memory implementations for views and UDFs to external catalog. These APIs are required when we store Flink views and UDFs in an external persistent storage. These APIs will evolve as we make progress in Flink-Hive integration. ## Brief change log - added initial APIs for views and UDFs in `ExternalCatalog` and `CrudExternalCatalog` - added in-memory implementations in `InMemoryCrudExternalCatalog` - added relevant tests ## Verifying this change This change added tests and can be verified as follows: - Added unit tests in `InMemoryExternalCatalogTest` ## Does this pull request potentially affect one of the following parts: none ## Documentation - Does this pull request introduce a new feature? (yes ) - If yes, how is the feature documented? (JavaDocs) This is an automated message from the Apache Git Service. To respond to the message, please log on GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org > Integration with Apache Hive > > > Key: FLINK-10556 > URL: https://issues.apache.org/jira/browse/FLINK-10556 > Project: Flink > Issue Type: New Feature > Components: Batch Connectors and Input/Output Formats, SQL Client, > Table API SQL >Affects Versions: 1.6.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang >Priority: Major > Labels: pull-request-available > Attachments: Flink-Hive Metastore Connectivity Design.pdf, Proposal_ > Integrate Flink with Hive Ecosystem.pdf > > > This is an umbrella JIRA tracking all enhancement and issues related to > integrating Flink with Hive ecosystem. This is an outcome of a discussion in > the community, and thanks go to everyone that provided feedback and interest. > Specifically, we'd like to see the following features and capabilities > immediately in Flink: > # Metadata interoperability > # Data interoperability > # Data type compatibility > # Hive UDF support > # DDL/DML/Query language compatibility > For a longer term, we'd also like to add or improve: > # Compatible SQL service, client tools, JDBC/ODBC drivers > # Better task failure tolerance and task scheduling > # Support other user customizations in Hive (storage handlers, serdes, etc). > I will provide more details regarding the proposal in a doc shortly. Design > doc, if deemed necessary, will be provided in each related sub tasks under > this JIRA. > Feedback and contributions are greatly welcome! -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-10556) Integration with Apache Hive
[ https://issues.apache.org/jira/browse/FLINK-10556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16669164#comment-16669164 ] Xuefu Zhang commented on FLINK-10556: - I have attached the design doc on Hive metastore integration. It covers using Hive as a schema registry as well as a metadata connector for Hive. Please review and share your input. Thanks. > Integration with Apache Hive > > > Key: FLINK-10556 > URL: https://issues.apache.org/jira/browse/FLINK-10556 > Project: Flink > Issue Type: New Feature > Components: Batch Connectors and Input/Output Formats, SQL Client, > Table API SQL >Affects Versions: 1.6.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang >Priority: Major > Attachments: Flink-Hive Metastore Connectivity Design.pdf, Proposal_ > Integrate Flink with Hive Ecosystem.pdf > > > This is an umbrella JIRA tracking all enhancement and issues related to > integrating Flink with Hive ecosystem. This is an outcome of a discussion in > the community, and thanks go to everyone that provided feedback and interest. > Specifically, we'd like to see the following features and capabilities > immediately in Flink: > # Metadata interoperability > # Data interoperability > # Data type compatibility > # Hive UDF support > # DDL/DML/Query language compatibility > For a longer term, we'd also like to add or improve: > # Compatible SQL service, client tools, JDBC/ODBC drivers > # Better task failure tolerance and task scheduling > # Support other user customizations in Hive (storage handlers, serdes, etc). > I will provide more details regarding the proposal in a doc shortly. Design > doc, if deemed necessary, will be provided in each related sub tasks under > this JIRA. > Feedback and contributions are greatly welcome! -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-10556) Integration with Apache Hive
[ https://issues.apache.org/jira/browse/FLINK-10556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16667607#comment-16667607 ] Bowen Li commented on FLINK-10556: -- FLINK-10686 will create a Java-based flink-table-common module that Flink-Hive integration may depend on. Once FLINK-10686 is done, the work of Flink-Hive integration may be written completely in Java. background: the community has decided to move away from Scala as we've learned Scala is hard to maintain. > Integration with Apache Hive > > > Key: FLINK-10556 > URL: https://issues.apache.org/jira/browse/FLINK-10556 > Project: Flink > Issue Type: New Feature > Components: Batch Connectors and Input/Output Formats, SQL Client, > Table API SQL >Affects Versions: 1.6.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang >Priority: Major > Attachments: Proposal_ Integrate Flink with Hive Ecosystem.pdf > > > This is an umbrella JIRA tracking all enhancement and issues related to > integrating Flink with Hive ecosystem. This is an outcome of a discussion in > the community, and thanks go to everyone that provided feedback and interest. > Specifically, we'd like to see the following features and capabilities > immediately in Flink: > # Metadata interoperability > # Data interoperability > # Data type compatibility > # Hive UDF support > # DDL/DML/Query language compatibility > For a longer term, we'd also like to add or improve: > # Compatible SQL service, client tools, JDBC/ODBC drivers > # Better task failure tolerance and task scheduling > # Support other user customizations in Hive (storage handlers, serdes, etc). > I will provide more details regarding the proposal in a doc shortly. Design > doc, if deemed necessary, will be provided in each related sub tasks under > this JIRA. > Feedback and contributions are greatly welcome! -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-10556) Integration with Apache Hive
[ https://issues.apache.org/jira/browse/FLINK-10556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16662551#comment-16662551 ] Xuefu Zhang commented on FLINK-10556: - The proposal is attached. > Integration with Apache Hive > > > Key: FLINK-10556 > URL: https://issues.apache.org/jira/browse/FLINK-10556 > Project: Flink > Issue Type: New Feature > Components: Batch Connectors and Input/Output Formats, SQL Client, > Table API SQL >Affects Versions: 1.6.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang >Priority: Major > Attachments: Proposal_ Integrate Flink with Hive Ecosystem.pdf > > > This is an umbrella JIRA tracking all enhancement and issues related to > integrating Flink with Hive ecosystem. This is an outcome of a discussion in > the community, and thanks go to everyone that provided feedback and interest. > Specifically, we'd like to see the following features and capabilities > immediately in Flink: > # Metadata interoperability > # Data interoperability > # Data type compatibility > # Hive UDF support > # DDL/DML/Query language compatibility > For a longer term, we'd also like to add or improve: > # Compatible SQL service, client tools, JDBC/ODBC drivers > # Better task failure tolerance and task scheduling > # Support other user customizations in Hive (storage handlers, serdes, etc). > I will provide more details regarding the proposal in a doc shortly. Design > doc, if deemed necessary, will be provided in each related sub tasks under > this JIRA. > Feedback and contributions are greatly welcome! -- This message was sent by Atlassian JIRA (v7.6.3#76005)