Greetings All!
I'm using v1.2 of sqoop and I'm having trouble getting the --query command line
parameter working. I can do the moral equivalent of the following using the
--table parameter sucessfully.
Any pointers would be greatly appreciated.
Thanx,
Dave
>sqoop import --connect
>"jdbc:sqlserver://FooBar;database=AdventureWorks2008R2;user=foo;password=bar"
--query "SELECT * FROM AWBuildVersion WHERE $CONDITIONS" --split-by
SystemInformationID --target-dir /tmp/Data/Dave1
Missing argument for option: e
usage: sqoop import [GENERIC-ARGS] [TOOL-ARGS]
Common arguments:
--connect <jdbc-uri> Specify JDBC connect
string
--connection-manager <conn-manager-class-name> Specify connection
manager class name
--driver <class-name> Manually specify JDBC
driver class to use
--hadoop-home <dir> Override $HADOOP_HOME
--help Print usage
instructions
-P Read password from
console
--password <password> Set authentication
password
--username <username> Set authentication
username
--verbose Print more
information while
working
Import control arguments:
--append Imports data in append mode
--as-sequencefile Imports data to SequenceFiles
--as-textfile Imports data as plain text (default)
--columns <col,col,col...> Columns to import from table
--direct Use direct import fast path
--direct-split-size <n> Split the input stream every 'n' bytes
when importing in direct mode
-e,--query <statement> Import results of SQL 'statement'
--inline-lob-limit <n> Set the maximum size for an inline LOB
-m,--num-mappers <n> Use 'n' map tasks to import in parallel
--split-by <column-name> Column of the table used to split work
units
--table <table-name> Table to read
--target-dir <dir> HDFS plain table destination
--warehouse-dir <dir> HDFS parent for table destination
--where <where clause> WHERE clause to use during import
-z,--compress Enable compression
Incremental import arguments:
--check-column <column> Source column to check for incremental
change
--incremental <import-type> Define an incremental import of type
'append' or 'lastmodified'
--last-value <value> Last imported value in the incremental
check column
Output line formatting arguments:
--enclosed-by <char> Sets a required field enclosing
character
--escaped-by <char> Sets the escape character
--fields-terminated-by <char> Sets the field separator character
--lines-terminated-by <char> Sets the end-of-line character
--mysql-delimiters Uses MySQL's default delimiter set:
fields: , lines: \n escaped-by: \
optionally-enclosed-by: '
--optionally-enclosed-by <char> Sets a field enclosing character
Input parsing arguments:
--input-enclosed-by <char> Sets a required field encloser
--input-escaped-by <char> Sets the input escape
character
--input-fields-terminated-by <char> Sets the input field separator
--input-lines-terminated-by <char> Sets the input end-of-line
char
--input-optionally-enclosed-by <char> Sets a field enclosing
character
Hive arguments:
--hive-home <dir> Override $HIVE_HOME
--hive-import Import tables into Hive (Uses Hive's
default delimiters if none are set.)
--hive-overwrite Overwrite existing data in the Hive table
--hive-table <table-name> Sets the table name to use when importing
to hive
HBase arguments:
--column-family <family> Sets the target column family for the
import
--hbase-create-table If specified, create missing HBase tables
--hbase-row-key <col> Specifies which input column to use as the
row key
--hbase-table <table> Import to <table> in HBase
Code generation arguments:
--bindir <dir> Output directory for
compiled objects
--class-name <name> Sets the generated
class name. This
overrides
--package-name. When
combined with
--jar-file, sets the
input class.
--input-null-non-string <input-null-non-string> Input null
non-string
representation
--input-null-string <input-null-string> Input null string
representation
--jar-file <file> Disable code
generation; use
specified jar
--null-non-string <null-non-string> Null non-string
representation
--null-string <null-string> Null string
representation
--outdir <dir> Output directory for
generated code
--package-name <name> Put auto-generated
classes in this
package
Generic Hadoop command-line arguments:
(must preceed any tool-specific arguments)
Generic options supported are
-conf <configuration file> specify an application configuration file
-D <property=value> use value for given property
-fs <local|namenode:port> specify a namenode
-jt <local|jobtracker:port> specify a job tracker
-files <comma separated list of files> specify comma separated files to be
copied to the map reduce cluster
-libjars <comma separated list of jars> specify comma separated jar files to
include in the classpath.
-archives <comma separated list of archives> specify comma separated
archives to be unarchived on the compute machines.
The general command line syntax is
bin/hadoop command [genericOptions] [commandOptions]
At minimum, you must specify --connect and --table
Arguments to mysqldump and other subprograms may be supplied
after a '--' on the command line.
>