Problem with reading multiple files
Moderators: chulett, rschirm, roy
Problem with reading multiple files
Hi,
I created a job where I am using * to input multiple files. I created 6 text files for the input, and each file has 2 providers. For some reason, the output dataset is splitting the output right down the middle. Each file has 44 rows, times 6, is 264. I input a sequence number (1-22) for each provider and a tracking number starting at 5000 (that increases per provider instance). So with 6 files of 2 providers each, the tracking number should be 5000-5011. However, the output is 5000-5005 and then 5000-5005... splitting it right down the middle. Is there a way to make this not happen? Make the output file go straight through without a split?
As a side note, whenever I add another file, totaling 7, the job fails. Can the input seq file only take 6 files at once?
I created a job where I am using * to input multiple files. I created 6 text files for the input, and each file has 2 providers. For some reason, the output dataset is splitting the output right down the middle. Each file has 44 rows, times 6, is 264. I input a sequence number (1-22) for each provider and a tracking number starting at 5000 (that increases per provider instance). So with 6 files of 2 providers each, the tracking number should be 5000-5011. However, the output is 5000-5005 and then 5000-5005... splitting it right down the middle. Is there a way to make this not happen? Make the output file go straight through without a split?
As a side note, whenever I add another file, totaling 7, the job fails. Can the input seq file only take 6 files at once?
SCOTTDun
I changed the job to run on 1 node because I have a smaller amount of data per incoming file. Then I used an environment variable to increase character length for record delimiter to 100000000. This works for now but I need a better way to do it for when there are millions of clients per file.
The job was aborting after more than 6 input files were being loaded so I used the parameter to increase character length before record delimiter.
The job was aborting after more than 6 input files were being loaded so I used the parameter to increase character length before record delimiter.
SCOTTDun
I don't understand the question. What do you mean by "run"? You can certainly use the Execute Command stage to return a list of all filenames that match that wildcard pattern, is that what you are after?ScottDun wrote:how can I run multiple files in the command execute stage in a sequence job
-craig
"You can never have too many knives" -- Logan Nine Fingers
"You can never have too many knives" -- Logan Nine Fingers