Hi,
I got the following error when running some pig script,

Error initializing attempt_201201031543_0083_m_000000_1:
java.lang.IllegalArgumentException: Wrong FS:
hdfs://10.2.0.135:54310/app/datastore/hadoop-hadoop/mapred/staging/hadoop/.staging/job_201201031543_0083/job.xml,
expected: hdfs://dev-hadoop-01.***.com:54310
        at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:410)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.checkPath(DistributedFileSystem.java:106)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.getPathName(DistributedFileSystem.java:162)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:542)
        at 
org.apache.hadoop.mapred.TaskTracker.localizeJobConfFile(TaskTracker.java:1280)
        at 
org.apache.hadoop.mapred.TaskTracker.initializeJob(TaskTracker.java:1174)
        at 
org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:1098)
        at 
org.apache.hadoop.mapred.TaskTracker.startNewTask(TaskTracker.java:2271)
        at 
org.apache.hadoop.mapred.TaskTracker$TaskLauncher.run(TaskTracker.java:2235)

It seems some ips in config files need to be changed to hostnames. Any hints?

Shawn

Reply via email to