[ 
https://issues.apache.org/jira/browse/FLINK-17384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

jackylau updated FLINK-17384:
-----------------------------
    Comment: was deleted

(was: Hi [~liyu],  I have committed my code, but the log is below, which is not 
relevant with my code

2020-05-14T13:46:32.9352627Z [ERROR] Failures: 
 2020-05-14T13:46:32.9361371Z [ERROR] 
KafkaProducerExactlyOnceITCase>KafkaProducerTestBase.testExactlyOnceRegularSink:309->KafkaProducerTestBase.testExactlyOnce:370
 Test failed: Job execution failed

ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-surefire-plugin:2.22.1:test (end-to-end-tests) 
on project flink-metrics-availability-test: Unable to generate classpath: 
org.apache.maven.artifact.resolver.ArtifactResolutionException: Could not 
transfer artifact org.apache.maven.surefire:surefire-grouper:jar:2.22.1 from/to 
alicloud-mvn-mirror 
([http://mavenmirror.alicloud.dak8s.net:8888/repository/maven-central/):] Entry 
[id:18][route:{}->http://mavenmirror.alicloud.dak8s.net:8888][state:null] has 
not been leased from this pool.

 

How to solve it , and why that happends. how to make the 
[flinkbot|https://github.com/flinkbot] rerun azure)

> support read hbase conf dir from flink.conf just like hadoop_conf
> -----------------------------------------------------------------
>
>                 Key: FLINK-17384
>                 URL: https://issues.apache.org/jira/browse/FLINK-17384
>             Project: Flink
>          Issue Type: Bug
>          Components: Connectors / HBase, Deployment / Scripts
>    Affects Versions: 1.10.0
>            Reporter: jackylau
>            Assignee: jackylau
>            Priority: Major
>              Labels: pull-request-available
>             Fix For: 1.11.0
>
>
> hi all:
> when user interacts with hbase should do 2 things when using sql
>  # export HBASE_CONF_DIR
>  # add hbase libs to flink_lib(because the hbase connnector doesn't have 
> client's( and others) jar)
> i think it needs to optimise it.
> for 1) we should support read hbase conf dir from flink.conf just like 
> hadoop_conf in  config.sh
> for 2) we should support HBASE_CLASSPATH in  config.sh. In case of jar 
> conflicts such as guava , we also should support flink-hbase-shaded just like 
> hadoop does



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to