[ https://issues.apache.org/jira/browse/FLINK-19025?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17184918#comment-17184918 ]
Rui Li commented on FLINK-19025: -------------------------------- Hey [~McClone], are you writing streaming data into hive orc table? If so, there's indeed a known issue which has been fixed in FLINK-18659. You should be able to run this use case if you apply that patch, and set {{table.exec.hive.fallback-mapred-writer=true}}, which is the default setting. If you hit the issue when writing batch data into hive, please provide the stacktrace of the exception. > table sql write orc file but hive2.1.1 can not read > --------------------------------------------------- > > Key: FLINK-19025 > URL: https://issues.apache.org/jira/browse/FLINK-19025 > Project: Flink > Issue Type: Bug > Components: Connectors / ORC > Affects Versions: 1.11.0 > Reporter: McClone > Priority: Major > > table sql write orc file but hive2.1.1 create external table can not read > data.Because flink use orc-core-1.5.6.jar but hive 2.1.1 use his own orcfile > jar. -- This message was sent by Atlassian Jira (v8.3.4#803005)