went on the wrong direction of what you
> needed.
> Best
> -P.
>
> [1] https://openpyxl.readthedocs.io/en/stable/
> [2]
> https://github.com/apache/beam/blob/master/sdks/python/apache_beam/io/fileio_test.py#L128-L148
>
> On Mon, Apr 15, 2019 at 12:47 PM Henrique Molina <
&
Hello
I would like to use best practices from Apache Beams to read Xlsx. however
I found examples only related with Cs extension.
someone there is sample using ParDo to Collect all columns and sheets from
Excel xlsx ?
Afterwards I will put into google Big query.
Thanks & Regards
Hi Shrikant,
Pay attention in your parameter --output=gs://test-bucket*/c *\
Your configuration is indicate directory /c , not */b*
So, check in your Storage GCP if exist this directory : /c (
-output=gs://test-bucket/c )
and check :
gsutil ls gs://test-bucket*/c*
Cheers
Carlos Molina
On Thu,
Hi folks,
I using query select * from VIEW_*1* after View_*2*, on database , and
next step is collect rows and export to CSV.
I actual in this point:
PCollection> view1 = p.apply(JdbcIO.>read()
.withDataSourceConfiguration(JdbcIO.DataSourceConfiguration.create(
"org.postgresql.Driver","jdbc:pos
;
> ByteArrayOutputStream download = new ByteArrayOutputStream();
> boolean result = ftpClient.retrieveFile(f.getName(), download);
> String destinationPath = saveCSV(download, f.getName()); // save
> CSV in Storage Google cloud
> outputReceiver.output(destinationPath)
Dear all,
I Using FTP Client to download some file dynamically , and the file is csv.
( it is working fine)
And the next step I need to open the files, and read lines
Somebody could help me using the good practices in this approach ?
I using Java > Google DataFlow > apache beam 2.9.0
PCollection
Hi Folks ,
I'm newbie in Beam, but I looking for some way to read an File stored at
FTP
First of all, I could create ParDo, using FTPClient (Commons-net) and
access returning an Byte[] of File *.csv.
second ParDO create the csv
third PardDo using the TextIO to read lines
Somebody could share