Tracy Rahmlow wrote: > > We attempted to reorg a table and data corruption resulted. We have isolated > the issue and currently have a tar open with Oracle. But basically, when we > export the table with a buffer=10485760 the import process corrupts the data. > The row count matches, however some fields that initially were null now contain > data (there were some other odd data issues in addition). No errors were > produced during either the export or import process. When we used a > buffer=65536 the table data was correct. This is reproducable. Has anybody > seen this before? (AIX 4.3.3 / Oracle 8.1.7.3) Can the buffer size be set too > high?
This is just an idea based on past problems with the OCIs, but array size (as in 'array fetch') is limited to 32K (smells of signed short index). Your buffer size indirectly specifies the number of rows you return with a single batch (something in the order of magnitude of buffer / avg_row_len). It is not impossible, if your rows are short, that your theoretical array size is past the limit and that Oracle doesn't handle it properly. -- Regards, Stephane Faroult Oriole Software -- Please see the official ORACLE-L FAQ: http://www.orafaq.net -- Author: Stephane Faroult INET: [EMAIL PROTECTED] Fat City Network Services -- 858-538-5051 http://www.fatcity.com San Diego, California -- Mailing list and web hosting services --------------------------------------------------------------------- To REMOVE yourself from this mailing list, send an E-Mail message to: [EMAIL PROTECTED] (note EXACT spelling of 'ListGuru') and in the message BODY, include a line containing: UNSUB ORACLE-L (or the name of mailing list you want to be removed from). You may also send the HELP command for other information (like subscribing).