It sounds like you want to sequentially read a file, sequentially process
the records and sequentially write them. The best way to do this is likely
without using Beam, just write some Java or Python code using standard file
APIs (use Beam's FileSystem APIs if you need to access data on a non-local
filesystem).

On Tue, Aug 21, 2018 at 7:11 AM [email protected] <[email protected]>
wrote:

> Hi
>
> I have to process a big file and call several Pardo's to do some
> transformations.  Records in file dont have any unique key.
>
> Lets say file 'testfile' has 1 million records.
>
> After processing , I want to generate only one output file same as my
> input 'testfile' and also i have a requirement to write those 1 million
> records in same order (after applying some Pardo's)
>
> What is best way to do it
>
> Thanks
> Aniruddh
>
>
>
>

Reply via email to