This is an automated email from the ASF dual-hosted git repository.

martijnvisser pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/flink.git


The following commit(s) were added to refs/heads/master by this push:
     new 1e3e644a242 [FLINK-24940][docs] Correct usage about how to create Hive 
catalog via Flink SQL CLI. This closes #17829
1e3e644a242 is described below

commit 1e3e644a24202434749f9a1b51fda87af6d0054a
Author: yuxia Luo <luoyu...@alibaba-inc.com>
AuthorDate: Thu Nov 18 20:17:01 2021 +0800

    [FLINK-24940][docs] Correct usage about how to create Hive catalog via 
Flink SQL CLI. This closes #17829
---
 .../docs/connectors/table/hive/hive_catalog.md     | 34 ++++++++++------------
 .../docs/connectors/table/hive/hive_catalog.md     | 34 ++++++++++------------
 2 files changed, 30 insertions(+), 38 deletions(-)

diff --git a/docs/content.zh/docs/connectors/table/hive/hive_catalog.md 
b/docs/content.zh/docs/connectors/table/hive/hive_catalog.md
index 6cc9e3041b2..dc2e461fd7f 100644
--- a/docs/content.zh/docs/connectors/table/hive/hive_catalog.md
+++ b/docs/content.zh/docs/connectors/table/hive/hive_catalog.md
@@ -137,28 +137,23 @@ Time taken: 0.028 seconds, Fetched: 0 row(s)
 ```
 
 
-#### step 2: configure Flink cluster and SQL CLI
-
-Add all Hive dependencies to `/lib` dir in Flink distribution, and modify SQL 
CLI's yaml config file `sql-cli-defaults.yaml` as following:
-
-```yaml
-
-execution:
-    type: streaming
-    ...
-    current-catalog: myhive  # set the HiveCatalog as the current catalog of 
the session
-    current-database: mydatabase
-    
-catalogs:
-   - name: myhive
-     type: hive
-     hive-conf-dir: /opt/hive-conf  # contains hive-site.xml
+#### step 2: start SQL Client, and create a Hive catalog with Flink SQL DDL
+
+Add all Hive dependencies to `/lib` dir in Flink distribution, and create a 
Hive catalog in Flink SQL CLI as following:
+
+```bash
+
+Flink SQL> CREATE CATALOG myhive WITH (
+  'type' = 'hive',
+  'hive-conf-dir' = '/opt/hive-conf'
+);
+
 ```
 
 
 #### step 3: set up a Kafka cluster
 
-Bootstrap a local Kafka 2.3.0 cluster with a topic named "test", and produce 
some simple data to the topic as tuple of name and age.
+Bootstrap a local Kafka cluster with a topic named "test", and produce some 
simple data to the topic as tuple of name and age.
 
 ```bash
 
@@ -180,11 +175,12 @@ john,21
 ```
 
 
-#### step 4: start SQL Client, and create a Kafka table with Flink SQL DDL
+#### step 4: create a Kafka table with Flink SQL DDL
 
-Start Flink SQL Client, create a simple Kafka 2.3.0 table via DDL, and verify 
its schema.
+Create a simple Kafka table with Flink SQL DDL, and verify its schema.
 
 ```bash
+Flink SQL> USE CATALOG myhive;
 
 Flink SQL> CREATE TABLE mykafka (name String, age Int) WITH (
    'connector.type' = 'kafka',
diff --git a/docs/content/docs/connectors/table/hive/hive_catalog.md 
b/docs/content/docs/connectors/table/hive/hive_catalog.md
index 90c1aebf057..932e18fcc0d 100644
--- a/docs/content/docs/connectors/table/hive/hive_catalog.md
+++ b/docs/content/docs/connectors/table/hive/hive_catalog.md
@@ -137,28 +137,23 @@ Time taken: 0.028 seconds, Fetched: 0 row(s)
 ```
 
 
-#### step 2: configure Flink cluster and SQL CLI
-
-Add all Hive dependencies to `/lib` dir in Flink distribution, and modify SQL 
CLI's yaml config file `sql-cli-defaults.yaml` as following:
-
-```yaml
-
-execution:
-    type: streaming
-    ...
-    current-catalog: myhive  # set the HiveCatalog as the current catalog of 
the session
-    current-database: mydatabase
-    
-catalogs:
-   - name: myhive
-     type: hive
-     hive-conf-dir: /opt/hive-conf  # contains hive-site.xml
+#### step 2: start SQL Client, and create a Hive catalog with Flink SQL DDL
+
+Add all Hive dependencies to `/lib` dir in Flink distribution, and create a 
Hive catalog in Flink SQL CLI as following:
+
+```bash
+
+Flink SQL> CREATE CATALOG myhive WITH (
+  'type' = 'hive',
+  'hive-conf-dir' = '/opt/hive-conf'
+);
+
 ```
 
 
 #### step 3: set up a Kafka cluster
 
-Bootstrap a local Kafka 2.3.0 cluster with a topic named "test", and produce 
some simple data to the topic as tuple of name and age.
+Bootstrap a local Kafka cluster with a topic named "test", and produce some 
simple data to the topic as tuple of name and age.
 
 ```bash
 
@@ -180,11 +175,12 @@ john,21
 ```
 
 
-#### step 4: start SQL Client, and create a Kafka table with Flink SQL DDL
+#### step 4: create a Kafka table with Flink SQL DDL
 
-Start Flink SQL Client, create a simple Kafka 2.3.0 table via DDL, and verify 
its schema.
+Create a simple Kafka table with Flink SQL DDL, and verify its schema.
 
 ```bash
+Flink SQL> USE CATALOG myhive;
 
 Flink SQL> CREATE TABLE mykafka (name String, age Int) WITH (
    'connector.type' = 'kafka',

Reply via email to