cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Parsing large files

ash42
New Contributor III

Hey all,

I am trying to tread a File (260+Mb) and parse it and and then insert the data to a Table. And when i m using SQL Bulk load The csv parser always seems to be stuckt at output 4100 docs

image

and when i m using normal insert table snap , this entire process taking 20-25mins

File format sample :
image

CSV Parser

image

PS: it works fine and quick with smaller size files

3 REPLIES 3

bojanvelevski
Valued Contributor

The difference between Insert and Bulk is that Insert writes down every record separately, and Bulk does it in chunks, so when a first chunk is ready, Bulk loads that chunk into the DB. Maybe thatโ€™s the hold you are facing, and you didnโ€™t wait enough? I am also curious about the join, and whatโ€™s going on in there, because Bulk Load will result in only one object on output, and thatโ€™s not the original. Also, whatโ€™s the snap after the CSV Parser?

Definitely the lock is not on the CSV parser, so you can try and start disconnecting the snaps one after the other to check whereโ€™s your problem.

ash42
New Contributor III

Thanks,seems like it was problem with, the reason I used join with tail snap is because i do not want the pipeline to continue executing untill the insert operation is fully completed, So used tail snap to act as a waiting snap

this is the updated pipeline
image

bojanvelevski
Valued Contributor

This might, and probably will result in proceeding with the pipeline even before the data is loaded into the table. The Join is one of the solutions to achieve what you need, but you need to be careful while configuring. Combine both, the new setup + Join (1on1)

image

That will result in having one object on output, it will wait both inputs to complete and will move the execution forward.