[
https://issues.apache.org/jira/browse/SQOOP-3463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Toan Nguyen updated SQOOP-3463:
-------------------------------
Description:
It's my script
" sqoop-import --connect $DATASOURCE --username $USERNAME --password $PASSWORD
-driver com.mysql.jdbc.Driver --query "SELECT * FROM sales_order_item WHERE
item_id > 0 AND \$CONDITIONS LIMIT $ITEM_ID, 10000" --target-dir
/user/raw/magento/$MERCHANT_ID/sales_order_item -m 8 --split-by item_id
--fields-terminated-by "|" --merge-key item_id --hive-import --hive-table
raw_magento_$MERCHANT_ID.sales_order_item --verbose --direct – "
And after completed process, i got only 9992 records. The happen is same with n
mapper, when i increase n, it will lost n records.
Everything will be OK with only 1 mapper. But if i want to import large
records, i have to use more mapper. So what should i do now? Please support me.
Thanks in advance
was:
It's my script
" sqoop-import --connect $DATASOURCE --username $USERNAME --password $PASSWORD
--driver com.mysql.jdbc.Driver \
--query "SELECT * FROM sales_order_item WHERE item_id > 0 AND \$CONDITIONS
LIMIT $ITEM_ID, 10000" \
--target-dir /user/raw/magento/$MERCHANT_ID/sales_order_item -m 8 \
--split-by item_id \
--fields-terminated-by "|" \
--merge-key item_id \
--hive-import \
--hive-table raw_magento_$MERCHANT_ID.sales_order_item \
--verbose \
--direct – "
And after completed process, i got only 9992 records. The happen is same with n
mapper, when i increase n, it will lost n records.
Everything will be OK with only 1 mapper. But if i want to import large
records, i have to use more mapper. So what should i do now? Please support me.
Thanks in advance
> Sqoop import from MYSQL to Hive fails when increase mapper
> ----------------------------------------------------------
>
> Key: SQOOP-3463
> URL: https://issues.apache.org/jira/browse/SQOOP-3463
> Project: Sqoop
> Issue Type: Bug
> Reporter: Toan Nguyen
> Priority: Major
>
> It's my script
> " sqoop-import --connect $DATASOURCE --username $USERNAME --password
> $PASSWORD -driver com.mysql.jdbc.Driver --query "SELECT * FROM
> sales_order_item WHERE item_id > 0 AND \$CONDITIONS LIMIT $ITEM_ID, 10000"
> --target-dir /user/raw/magento/$MERCHANT_ID/sales_order_item -m 8 --split-by
> item_id --fields-terminated-by "|" --merge-key item_id --hive-import
> --hive-table raw_magento_$MERCHANT_ID.sales_order_item --verbose --direct – "
> And after completed process, i got only 9992 records. The happen is same with
> n mapper, when i increase n, it will lost n records.
> Everything will be OK with only 1 mapper. But if i want to import large
> records, i have to use more mapper. So what should i do now? Please support
> me. Thanks in advance
--
This message was sent by Atlassian Jira
(v8.3.4#803005)