Hi,
 
We've got to load large pipe-delimited files. When loading these into a SQL 
Server DB by using Rhino ETL (relying upon FileHelpers), is it mandatory to 
provide a record class?
We have to load files to different tables which have dozens of columns - it 
might take us a whole day to generate them. I guess we can write a small 
tool to generate the record classes out of the SQL Server tables.
 
Another approach would be to write an IDataReader wrapper for a FileStream 
and the pass it on to a SqlBulkCopy.
 
SqlBulkCopy does require column mappings as well but it does allow column 
ordinals - that's easy.
 
Any ideas/suggestions?
 
Thanks.

-- 
You received this message because you are subscribed to the Google Groups 
"Rhino Tools Dev" group.
To view this discussion on the web visit 
https://groups.google.com/d/msg/rhino-tools-dev/-/IfCZWSwU1HsJ.
To post to this group, send email to [email protected].
To unsubscribe from this group, send email to 
[email protected].
For more options, visit this group at 
http://groups.google.com/group/rhino-tools-dev?hl=en.

Reply via email to