It sounds like you want to sequentially read a file, sequentially process the records and sequentially write them. The best way to do this is likely without using Beam, just write some Java or Python code using standard file APIs (use Beam's FileSystem APIs if you need to access data on a non-local filesystem).
On Tue, Aug 21, 2018 at 7:11 AM [email protected] <[email protected]> wrote: > Hi > > I have to process a big file and call several Pardo's to do some > transformations. Records in file dont have any unique key. > > Lets say file 'testfile' has 1 million records. > > After processing , I want to generate only one output file same as my > input 'testfile' and also i have a requirement to write those 1 million > records in same order (after applying some Pardo's) > > What is best way to do it > > Thanks > Aniruddh > > > >
