Re: Flink performance

2024-03-12 Thread Robin Moffatt via user
It would be useful if you shared what you've found already, or could give a
bit more detail about what it is that you're looking for. Numbers on their
own don't really tell a full picture.

Here are a few links that I found for you that might help:

*
https://flink.apache.org/2022/01/04/how-we-improved-scheduler-performance-for-large-scale-jobs-part-one/
* https://shopify.engineering/optimizing-apache-flink-applications-tips
* https://github.com/apache/flink-benchmarks
*
https://github.com/aws-samples/amazon-kinesis-data-analytics-flink-benchmarking-utility

-- 
Robin Moffatt
Principal DevEx Engineer, Decodable



On Tue, 12 Mar 2024 at 06:59, Kamal Mittal via user 
wrote:

> Hello Community,
>
>
>
> Please share info. for below query.
>
>
>
> Rgds,
>
> Kamal
>
>
>
> *From:* Kamal Mittal via user 
> *Sent:* Monday, March 11, 2024 1:18 PM
> *To:* user@flink.apache.org
> *Subject:* Flink performance
>
>
>
> Hello,
>
>
>
> Can you please point me to documentation if any such available where flink
> talks about or documented performance numbers w.r.t certain use cases?
>
>
>
> Rgds,
>
> Kamal
>


Re: Re: Running Flink SQL in production

2024-03-08 Thread Robin Moffatt via user
That makes sense, thank you. I found FLIP-316 [1] and will keep an eye on
it too.

Thanks, Robin.

[1]
https://cwiki.apache.org/confluence/display/FLINK/FLIP-316%3A+Support+application+mode+for+SQL+Gateway


On Fri, 8 Mar 2024 at 13:56, Zhanghao Chen 
wrote:

> Hi Robin,
>
> It's better to use Application mode [1] for mission-critical long-running
> SQL jobs as it provides better isolation, you can utilize the table API to
> package a jar as suggested by Feng to do so. Neither SQL client nor SQL
> gateway supports submitting SQL in Application mode for now, but there's
> some on-going effort [2]. Hopefully, it would be much easier to do so in
> the future.
>
> [1]
> https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/overview/#application-mode
>
> [2] https://issues.apache.org/jira/browse/FLINK-26541
>
> Best,
> Zhanghao Chen
> --
> *From:* Feng Jin 
> *Sent:* Friday, March 8, 2024 9:46
> *To:* Xuyang 
> *Cc:* Robin Moffatt ; user@flink.apache.org <
> user@flink.apache.org>
> *Subject:* Re: Re: Running Flink SQL in production
>
> Hi,
>
> If you need to use Flink SQL in a production environment, I think it would
> be better to use the Table API [1] and package it into a jar.
> Then submit the jar to the cluster environment.
>
> [1]
> https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/dev/table/common/#sql
>
> Best,
> Feng
>
> On Thu, Mar 7, 2024 at 9:56 PM Xuyang  wrote:
>
> Hi.
> Hmm, if I'm mistaken, please correct me. Using a SQL client might not be
> very convenient for those who need to verify the
> results of submissions, such as checking for exceptions related to
> submission failures, and so on.
>
>
> --
> Best!
> Xuyang
>
>
> 在 2024-03-07 17:32:07,"Robin Moffatt"  写道:
>
> Thanks for the reply.
> In terms of production, my thinking is you'll have your SQL in a file
> under code control. Whether that SQL ends up getting submitted via an
> invocation of SQL Client with -f or via REST API seems moot. WDYT?
>
>
>
> On Thu, 7 Mar 2024 at 01:53, Xuyang  wrote:
>
> Hi, IMO, both the SQL Client and the Restful API can provide connections
> to the SQL Gateway service for submitting jobs. A slight difference is that
> the SQL Client also offers a command-line visual interface for users to
> view results.
> In your production scenes, placing the SQL to be submitted into a file and
> then using the '-f' command in SQL Client to submit the file sounds a bit
> roundabout. You can just use the Restful API to submit them directly?
>
>
> --
> Best!
> Xuyang
>
>
> At 2024-03-07 04:11:01, "Robin Moffatt via user" 
> wrote:
>
> I'm reading the deployment guide[1] and wanted to check my understanding.
> For deploying a SQL job into production, would the pattern be to write the
> SQL in a file that's under source control, and pass that file as an
> argument to SQL Client with -f argument (as in this docs example[2])?
> Or script a call to the SQL Gateway's REST API?
>
> Are there pros and cons to each approach?
>
> thanks, Robin
>
> [1]:
> https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/deployment/overview/
> [2]:
> https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/dev/table/sqlclient/#execute-sql-files
>
>


Running Flink SQL in production

2024-03-06 Thread Robin Moffatt via user
I'm reading the deployment guide[1] and wanted to check my understanding.
For deploying a SQL job into production, would the pattern be to write the
SQL in a file that's under source control, and pass that file as an
argument to SQL Client with -f argument (as in this docs example[2])?
Or script a call to the SQL Gateway's REST API?

Are there pros and cons to each approach?

thanks, Robin

[1]:
https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/deployment/overview/
[2]:
https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/dev/table/sqlclient/#execute-sql-files


[Meta question] Sharing blog posts

2024-02-16 Thread Robin Moffatt via user
Hi,
I have a netiquette question - is it ok to share blog posts here that
are specific to Apache Flink and not vendor-focussed?

thanks,

Robin.


DESCRIBE CATALOG not available?

2024-01-26 Thread Robin Moffatt via user
In FLIP-69 [1] it looks like DESCRIBE CATALOG was added, but when I
try it from SQL Client in 1.18.1 it doesn't seem to work

Flink SQL> SHOW CATALOGS;
+-+
|catalog name |
+-+
|   c_new |
| default_catalog |
+-+
2 rows in set

Flink SQL> DESCRIBE CATALOG default_catalog;
[ERROR] Could not execute SQL statement. Reason:
org.apache.calcite.sql.validate.SqlValidatorException: Column
'default_catalog' not found in any table

Poking around I found the JIRA that implemented it [2] and can see the
code present today [3]

Any ideas?

Thanks, Robin.

[1] 
https://cwiki.apache.org/confluence/display/FLINK/FLIP-69%3A+Flink+SQL+DDL+Enhancement
[2] https://issues.apache.org/jira/browse/FLINK-14689
[3] 
https://github.com/apache/flink/blob/master/flink-table/flink-sql-parser/src/main/java/org/apache/flink/sql/parser/dql/SqlDescribeCatalog.java