[jira] Updated: (HADOOP-423) file paths are not normalized

2006-09-28 Thread Wendy Chien (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-423?page=all ] Wendy Chien updated HADOOP-423: --- Status: Patch Available (was: Open) > file paths are not normalized > - > > Key: HADOOP-423 > URL: ht

[jira] Updated: (HADOOP-423) file paths are not normalized

2006-09-28 Thread Wendy Chien (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-423?page=all ] Wendy Chien updated HADOOP-423: --- Attachment: (was: pathnorm2.patch) > file paths are not normalized > - > > Key: HADOOP-423 > URL:

[jira] Updated: (HADOOP-423) file paths are not normalized

2006-09-28 Thread Wendy Chien (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-423?page=all ] Wendy Chien updated HADOOP-423: --- Attachment: pathnorm3.patch > file paths are not normalized > - > > Key: HADOOP-423 > URL: http://issu

[jira] Commented: (HADOOP-423) file paths are not normalized

2006-09-28 Thread Doug Cutting (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-423?page=comments#action_12438586 ] Doug Cutting commented on HADOOP-423: - > I don't understand where you don't want to call Path.toString(), in DFSClient? You're right. I mistakenly assumed tha

[jira] Commented: (HADOOP-423) file paths are not normalized

2006-09-28 Thread Wendy Chien (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-423?page=comments#action_12438585 ] Wendy Chien commented on HADOOP-423: You're right, I've changed the logic in the namenode check. I don't understand where you don't want to call Path.toString(

[jira] Commented: (HADOOP-489) Seperating user logs from system logs in map reduce

2006-09-28 Thread Doug Cutting (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-489?page=comments#action_12438584 ] Doug Cutting commented on HADOOP-489: - +1 These sound like good changes. They will permit folks to get 'tail -f'-like log output while jobs are running, to co

[jira] Commented: (HADOOP-489) Seperating user logs from system logs in map reduce

2006-09-28 Thread Owen O'Malley (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-489?page=comments#action_12438579 ] Owen O'Malley commented on HADOOP-489: -- In terms of managing the user logs in the task tracker, there would be 2 settings: 1. mapred.task.log.size.max

[jira] Updated: (HADOOP-519) HDFS File API should be extended to include positional read

2006-09-28 Thread Milind Bhandarkar (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-519?page=all ] Milind Bhandarkar updated HADOOP-519: - Attachment: pread.patch This is the new patch that fixes the problems in earlier patch mentioned by Doug. > HDFS File API should be extended to inc

[jira] Updated: (HADOOP-519) HDFS File API should be extended to include positional read

2006-09-28 Thread Milind Bhandarkar (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-519?page=all ] Milind Bhandarkar updated HADOOP-519: - Status: Patch Available (was: Open) Patch submitted. > HDFS File API should be extended to include positional read > ---

[jira] Updated: (HADOOP-519) HDFS File API should be extended to include positional read

2006-09-28 Thread Milind Bhandarkar (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-519?page=all ] Milind Bhandarkar updated HADOOP-519: - Attachment: (was: pread.patch) > HDFS File API should be extended to include positional read > ---

[jira] Commented: (HADOOP-489) Seperating user logs from system logs in map reduce

2006-09-28 Thread Owen O'Malley (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-489?page=comments#action_12438557 ] Owen O'Malley commented on HADOOP-489: -- I'd like to have: 1. A jsp on the task trackers that let's me fetch stdout/stderr from the Tasks with urls like: htt

[jira] Assigned: (HADOOP-549) NullPointerException in TaskReport's serialization code

2006-09-28 Thread Owen O'Malley (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-549?page=all ] Owen O'Malley reassigned HADOOP-549: Assignee: Owen O'Malley > NullPointerException in TaskReport's serialization code > --- > >

[jira] Updated: (HADOOP-557) task tracker cannot find mapoutput files

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-557?page=all ] Sameer Paranjpye updated HADOOP-557: Component/s: mapred Description: I ran a large job on our cluster over the weekend. At some point, some map tasks were re-run successfully. However

[jira] Updated: (HADOOP-466) Startup scripts will not start instances of Hadoop daemons w/different configs w/o setting separate PID directories

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-466?page=all ] Sameer Paranjpye updated HADOOP-466: Component/s: conf > Startup scripts will not start instances of Hadoop daemons w/different > configs w/o setting separate PID directories > ---

[jira] Updated: (HADOOP-543) Error to open job files

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-543?page=all ] Sameer Paranjpye updated HADOOP-543: Component/s: mapred Description: I was running a faily large job on Hadoop release 0.6.2. The job failed because a lot of map tasks failed with foll

[jira] Updated: (HADOOP-439) Streaming does not work for text data if the records don't fit in a short UTF8 [2^16/3 characters]

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-439?page=all ] Sameer Paranjpye updated HADOOP-439: Component/s: contrib/streaming > Streaming does not work for text data if the records don't fit in a short > UTF8 [2^16/3 characters] > ---

[jira] Updated: (HADOOP-230) improve syntax of the hadoop dfs command

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-230?page=all ] Sameer Paranjpye updated HADOOP-230: Component/s: dfs > improve syntax of the hadoop dfs command > > > Key: HADOOP-230 >

[jira] Commented: (HADOOP-550) Text constructure can throw exception

2006-09-28 Thread Doug Cutting (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-550?page=comments#action_12438547 ] Doug Cutting commented on HADOOP-550: - Two minor nits: 1. Instead of ignoring the CharacterCodingException that should never be thrown, it would be better to

[jira] Updated: (HADOOP-555) Tasks should inherit some of the server's environment

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-555?page=all ] Sameer Paranjpye updated HADOOP-555: Component/s: mapred > Tasks should inherit some of the server's environment > - > >

[jira] Updated: (HADOOP-513) IllegalStateException is thrown by TaskTracker

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-513?page=all ] Sameer Paranjpye updated HADOOP-513: Component/s: mapred > IllegalStateException is thrown by TaskTracker > -- > > Key: HADOOP-51

[jira] Updated: (HADOOP-564) we should use hdfs:// in all API URIs

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-564?page=all ] Sameer Paranjpye updated HADOOP-564: Component/s: dfs > we should use hdfs:// in all API URIs > - > > Key: HADOOP-564 > U

[jira] Updated: (HADOOP-563) DFS client should try to re-new lease if it gets a lease expiration exception when it adds a block to a file

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-563?page=all ] Sameer Paranjpye updated HADOOP-563: Component/s: dfs Description: In the current DFS client implementation, there is one thread responsible for renewing leases. If for whatever reason

[jira] Updated: (HADOOP-535) back to back testing of codecs

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-535?page=all ] Sameer Paranjpye updated HADOOP-535: Component/s: io > back to back testing of codecs > -- > > Key: HADOOP-535 > URL: http://issu

[jira] Updated: (HADOOP-485) allow a different comparator for grouping keys in calls to reduce

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-485?page=all ] Sameer Paranjpye updated HADOOP-485: Component/s: mapred > allow a different comparator for grouping keys in calls to reduce > --

[jira] Assigned: (HADOOP-489) Seperating user logs from system logs in map reduce

2006-09-28 Thread Owen O'Malley (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-489?page=all ] Owen O'Malley reassigned HADOOP-489: Assignee: Owen O'Malley (was: Mahadev konar) > Seperating user logs from system logs in map reduce > --

[jira] Updated: (HADOOP-429) Periodically move blocks from full nodes to those with space

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-429?page=all ] Sameer Paranjpye updated HADOOP-429: Component/s: dfs > Periodically move blocks from full nodes to those with space > - > >

[jira] Updated: (HADOOP-552) getMapOutput doesn't reliably detect errors and throw to the caller

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-552?page=all ] Sameer Paranjpye updated HADOOP-552: Component/s: mapred > getMapOutput doesn't reliably detect errors and throw to the caller >

[jira] Updated: (HADOOP-544) Replace the job, tip and task ids with objects.

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-544?page=all ] Sameer Paranjpye updated HADOOP-544: Component/s: mapred > Replace the job, tip and task ids with objects. > --- > > Key: HADOOP-

[jira] Updated: (HADOOP-500) Datanode should scan blocks continuously to detect bad blocks / CRC errors

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-500?page=all ] Sameer Paranjpye updated HADOOP-500: Component/s: dfs > Datanode should scan blocks continuously to detect bad blocks / CRC errors >

[jira] Updated: (HADOOP-511) mapred.reduce.tasks not used

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-511?page=all ] Sameer Paranjpye updated HADOOP-511: Component/s: mapred > mapred.reduce.tasks not used > > > Key: HADOOP-511 > URL: http://issu

[jira] Updated: (HADOOP-475) The value iterator to reduce function should be clonable

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-475?page=all ] Sameer Paranjpye updated HADOOP-475: Component/s: mapred Description: In the current framework, when the user implements the reduce method of Reducer class, the user can only iterate

[jira] Updated: (HADOOP-448) DistributedFileSystem uses the wrong user.name to set the working directory.

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-448?page=all ] Sameer Paranjpye updated HADOOP-448: Component/s: dfs > DistributedFileSystem uses the wrong user.name to set the working directory. > --

[jira] Updated: (HADOOP-445) Parallel data/socket writing for DFSOutputStream

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-445?page=all ] Sameer Paranjpye updated HADOOP-445: Component/s: dfs > Parallel data/socket writing for DFSOutputStream > > > Key: HADOOP-4

[jira] Updated: (HADOOP-379) provide progress feedback while the reducer is sorting

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-379?page=all ] Sameer Paranjpye updated HADOOP-379: Component/s: mapred > provide progress feedback while the reducer is sorting > -- > >

[jira] Updated: (HADOOP-308) Task Tracker does not handle the case of read only local dir case correctly

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-308?page=all ] Sameer Paranjpye updated HADOOP-308: Component/s: mapred Description: In case that the local dir is not writable on a node, the tasks on the node will fail as expected, with an except

[jira] Updated: (HADOOP-357) hadoop doesn't handle 0 reduces

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-357?page=all ] Sameer Paranjpye updated HADOOP-357: Component/s: mapred > hadoop doesn't handle 0 reduces > --- > > Key: HADOOP-357 > URL: http:

[jira] Updated: (HADOOP-338) the number of maps in the JobConf does not match reality

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-338?page=all ] Sameer Paranjpye updated HADOOP-338: Component/s: mapred > the number of maps in the JobConf does not match reality > > >

[jira] Updated: (HADOOP-333) we should have some checks that the sort benchmark generates correct outputs

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-333?page=all ] Sameer Paranjpye updated HADOOP-333: Component/s: mapred > we should have some checks that the sort benchmark generates correct outputs > ---

[jira] Updated: (HADOOP-249) Improving Map -> Reduce performance and Task JVM reuse

2006-09-28 Thread Sameer Paranjpye (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-249?page=all ] Sameer Paranjpye updated HADOOP-249: Component/s: mapred > Improving Map -> Reduce performance and Task JVM reuse > -- > >

[jira] Updated: (HADOOP-550) Text constructure can throw exception

2006-09-28 Thread Hairong Kuang (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-550?page=all ] Hairong Kuang updated HADOOP-550: - Status: Patch Available (was: Open) Fix Version/s: 0.7.0 Affects Version/s: 0.6.2 > Text constructure can throw exception > --

[jira] Updated: (HADOOP-550) Text constructure can throw exception

2006-09-28 Thread Hairong Kuang (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-550?page=all ] Hairong Kuang updated HADOOP-550: - Attachment: text.patch > Text constructure can throw exception > - > > Key: HADOOP-550 > U

[jira] Commented: (HADOOP-550) Text constructure can throw exception

2006-09-28 Thread Hairong Kuang (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-550?page=comments#action_12438541 ] Hairong Kuang commented on HADOOP-550: -- Thanks for your comments, Addison. Currently Text is the default clas for map/reduce text input files, in which record

[jira] Updated: (HADOOP-560) tasks should have a "killed" state

2006-09-28 Thread Doug Cutting (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-560?page=all ] Doug Cutting updated HADOOP-560: Status: Resolved (was: Patch Available) Resolution: Fixed I just committed this. Thanks, Owen! > tasks should have a "killed" state > ---

Re: Creating splits/tasks at the client

2006-09-28 Thread Bryan A. P. Pendleton
I'm largely at fault for the "user code running in the JobTracker" that exists. I support this change - but, I might reformulate it. Why not make this a sort of special Job? It can even be formulated roughly like this: input -> map(Job,FilePath) -> reduce(Job,FileSplits) -> SchedulableJob It mi

[jira] Commented: (HADOOP-239) job tracker WI drops jobs after 24 hours

2006-09-28 Thread Doug Cutting (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-239?page=comments#action_12438516 ] Doug Cutting commented on HADOOP-239: - > does this sound reasonable? Yes, it sounds great to me! Thanks! > job tracker WI drops jobs after 24 hours > ---

Re: Creating splits/tasks at the client

2006-09-28 Thread Doug Cutting
Benjamin Reed wrote: One of the things that bothers me about the JobTracker is that it is running user code when it creates the FileSplits. In the long term this puts the JobTracker JVM at risk due to errors in the user code. JVM's are supposed to be able to do this kind of stuff securely. Sti

[jira] Commented: (HADOOP-538) Implement a nio's 'direct buffer' based wrapper over zlib to improve performance of java.util.zip.{De|In}flater as a 'custom codec'

2006-09-28 Thread Arun C Murthy (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-538?page=comments#action_12438503 ] Arun C Murthy commented on HADOOP-538: -- Sounds like shipping libhadoop-linux-i386.so along should be a sweet-spot between performance and maintainence/release

[jira] Updated: (HADOOP-423) file paths are not normalized

2006-09-28 Thread Doug Cutting (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-423?page=all ] Doug Cutting updated HADOOP-423: Status: Open (was: Patch Available) Fix Version/s: 0.7.0 I think the check in the NameNode isn't quite right. We should never permit "." or ".." a

[jira] Commented: (HADOOP-538) Implement a nio's 'direct buffer' based wrapper over zlib to improve performance of java.util.zip.{De|In}flater as a 'custom codec'

2006-09-28 Thread Owen O'Malley (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-538?page=comments#action_12438492 ] Owen O'Malley commented on HADOOP-538: -- I was mostly putting in DF since it had been mentioned and I didn't think the NativeCode class should be buried up in

[jira] Commented: (HADOOP-239) job tracker WI drops jobs after 24 hours

2006-09-28 Thread Sanjay Dahiya (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-239?page=comments#action_12438487 ] Sanjay Dahiya commented on HADOOP-239: -- Here are some new changes ( will submit patch in a while ) Moved code to Java 5 Replaced all data attributes in JobI

[jira] Commented: (HADOOP-538) Implement a nio's 'direct buffer' based wrapper over zlib to improve performance of java.util.zip.{De|In}flater as a 'custom codec'

2006-09-28 Thread Doug Cutting (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-538?page=comments#action_12438483 ] Doug Cutting commented on HADOOP-538: - Let me refine that a bit: if we ship with any pre-built binaries, I think it should be libhadoop-linux-i386.so only. Al

[jira] Commented: (HADOOP-538) Implement a nio's 'direct buffer' based wrapper over zlib to improve performance of java.util.zip.{De|In}flater as a 'custom codec'

2006-09-28 Thread Doug Cutting (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-538?page=comments#action_12438481 ] Doug Cutting commented on HADOOP-538: - > we need a class that is responsible for trying to load libhadoop.so and > having fall back replacements +1 I questio

[jira] Updated: (HADOOP-560) tasks should have a "killed" state

2006-09-28 Thread Owen O'Malley (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-560?page=all ] Owen O'Malley updated HADOOP-560: - Status: Patch Available (was: Open) > tasks should have a "killed" state > -- > > Key: HADOOP-560 >

[jira] Updated: (HADOOP-560) tasks should have a "killed" state

2006-09-28 Thread Owen O'Malley (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-560?page=all ] Owen O'Malley updated HADOOP-560: - Attachment: kill-state-2.patch I fixed another couple of things: 1. my previous patch tickled a previous bug that was causing jobs to not complete when the

[jira] Commented: (HADOOP-538) Implement a nio's 'direct buffer' based wrapper over zlib to improve performance of java.util.zip.{De|In}flater as a 'custom codec'

2006-09-28 Thread Owen O'Malley (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-538?page=comments#action_12438478 ] Owen O'Malley commented on HADOOP-538: -- Ok, I think we need a class that is responsible for trying to load libhadoop.so and having fall back replacements. pa

Re: [jira] Created: (HADOOP-553) DataNode and NameNode main() should catch and report exceptions.

2006-09-28 Thread Arkady Borkovsky
Should this be true (main() calls and reports exceptions) for all system components? Especially for user-facing ones like hadoop client? On Sep 20, 2006, at 3:31 PM, Konstantin Shvachko (JIRA) wrote: DataNode and NameNode main() should catch and report exceptions.

[jira] Commented: (HADOOP-538) Implement a nio's 'direct buffer' based wrapper over zlib to improve performance of java.util.zip.{De|In}flater as a 'custom codec'

2006-09-28 Thread Andrzej Bialecki (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-538?page=comments#action_12438395 ] Andrzej Bialecki commented on HADOOP-538: -- If we are to permit use of a native libhadoop there is at least one more thing that could be optimized - the a

[jira] Commented: (HADOOP-538) Implement a nio's 'direct buffer' based wrapper over zlib to improve performance of java.util.zip.{De|In}flater as a 'custom codec'

2006-09-28 Thread Arun C Murthy (JIRA)
[ http://issues.apache.org/jira/browse/HADOOP-538?page=comments#action_12438384 ] Arun C Murthy commented on HADOOP-538: -- Good points Doug; thanks! Some thoughts... I like the 'fallback' mechanism, it should ease hadoop adoption for people

Creating splits/tasks at the client

2006-09-28 Thread Benjamin Reed
One of the things that bothers me about the JobTracker is that it is running user code when it creates the FileSplits. In the long term this puts the JobTracker JVM at risk due to errors in the user code. The JobTracker uses the InputFormat to create a set of tasks that it then schedules. The task