I have written a reporting program.It works fine when the number of records in the input are around 100,000.Our input file has more that 1,300,000 records.There are duplicates records corresponding to a key.Is there any way to split the file so that we can always include all the records for the last key and the next split should start with next key.
A 2 ---------> 100,000 records
When we do a split by 100,000
Output of split should be is
There is no limit on the number of records that we can have for a particular key.
The input file may have up-to 1,300,000 record.So, I need to split them in 13 parts.The problem is that there are duplicates records corresponding to a key.If I directly split them by 100,000 then I might miss some of the data for the last key and it will again included in next file.
RECFM=FB and LRECL = 213.
There is only one Rule :- Splitted file should have all the records for the last key that is on 100,000 .................1,300,000 record and it should not be included in next split.