Build failed in Jenkins: Hadoop-0.20.204-Build #4

2011-07-16 Thread Apache Jenkins Server
See https://builds.apache.org/job/Hadoop-0.20.204-Build/4/changes

Changes:

[omalley] HADOOP-7459. Remove jdk-1.6.0 dependency check from rpm. (omalley)

--
[...truncated 8398 lines...]
 [exec] checking for ld used by gcc... /usr/bin/ld
 [exec] checking if the linker (/usr/bin/ld) is GNU ld... yes
 [exec] checking for /usr/bin/ld option to reload object files... -r
 [exec] checking for BSD-compatible nm... /usr/bin/nm -B
 [exec] checking whether ln -s works... yes
 [exec] checking how to recognise dependent libraries... pass_all
 [exec] checking dlfcn.h usability... yes
 [exec] checking dlfcn.h presence... yes
 [exec] checking for dlfcn.h... yes
 [exec] checking how to run the C++ preprocessor... g++ -E
 [exec] checking for g77... no
 [exec] checking for xlf... no
 [exec] checking for f77... no
 [exec] checking for frt... no
 [exec] checking for pgf77... no
 [exec] checking for cf77... no
 [exec] checking for fort77... no
 [exec] checking for fl32... no
 [exec] checking for af77... no
 [exec] checking for xlf90... no
 [exec] checking for f90... no
 [exec] checking for pgf90... no
 [exec] checking for pghpf... no
 [exec] checking for epcf90... no
 [exec] checking for gfortran... no
 [exec] checking for g95... no
 [exec] checking for xlf95... no
 [exec] checking for f95... no
 [exec] checking for fort... no
 [exec] checking for ifort... no
 [exec] checking for ifc... no
 [exec] checking for efc... no
 [exec] checking for pgf95... no
 [exec] checking for lf95... no
 [exec] checking for ftn... no
 [exec] checking whether we are using the GNU Fortran 77 compiler... no
 [exec] checking whether  accepts -g... no
 [exec] checking the maximum length of command line arguments... 32768
 [exec] checking command to parse /usr/bin/nm -B output from gcc object... 
ok
 [exec] checking for objdir... .libs
 [exec] checking for ar... ar
 [exec] checking for ranlib... ranlib
 [exec] checking for strip... strip
 [exec] checking if gcc static flag  works... yes
 [exec] checking if gcc supports -fno-rtti -fno-exceptions... no
 [exec] checking for gcc option to produce PIC... -fPIC
 [exec] checking if gcc PIC flag -fPIC works... yes
 [exec] checking if gcc supports -c -o file.o... yes
 [exec] checking whether the gcc linker (/usr/bin/ld -m elf_x86_64) 
supports shared libraries... yes
 [exec] checking whether -lc should be explicitly linked in... no
 [exec] checking dynamic linker characteristics... GNU/Linux ld.so
 [exec] checking how to hardcode library paths into programs... immediate
 [exec] checking whether stripping libraries is possible... yes
 [exec] checking if libtool supports shared libraries... yes
 [exec] checking whether to build shared libraries... yes
 [exec] checking whether to build static libraries... yes
 [exec] configure: creating libtool
 [exec] appending configuration tag CXX to libtool
 [exec] checking for ld used by g++... /usr/bin/ld -m elf_x86_64
 [exec] checking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes
 [exec] checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) 
supports shared libraries... yes
 [exec] checking for g++ option to produce PIC... -fPIC
 [exec] checking if g++ PIC flag -fPIC works... yes
 [exec] checking if g++ supports -c -o file.o... yes
 [exec] checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) 
supports shared libraries... yes
 [exec] checking dynamic linker characteristics... GNU/Linux ld.so
 [exec] checking how to hardcode library paths into programs... immediate
 [exec] checking whether stripping libraries is possible... yes
 [exec] appending configuration tag F77 to libtool
 [exec] checking for unistd.h... (cached) yes
 [exec] checking for stdbool.h that conforms to C99... yes
 [exec] checking for _Bool... no
 [exec] checking for an ANSI C-conforming const... yes
 [exec] checking for off_t... yes
 [exec] checking for size_t... yes
 [exec] checking whether strerror_r is declared... yes
 [exec] checking for strerror_r... yes
 [exec] checking whether strerror_r returns char *... yes
 [exec] checking for mkdir... yes
 [exec] checking for uname... yes
 [exec] configure: creating ./config.status
 [exec] config.status: creating Makefile
 [exec] config.status: creating impl/config.h
 [exec] config.status: impl/config.h is unchanged
 [exec] config.status: executing depfiles commands

compile-c++-utils:
 [exec] make[1]: Entering directory 
`https://builds.apache.org/job/Hadoop-0.20.204-Build/ws/trunk/build/c++-build/Linux-i386-32/utils'
 [exec] test -z 
https://builds.apache.org/job/Hadoop-0.20.204-Build/ws/trunk/build/c++/Linux-i386-32/lib;
 || mkdir -p -- 

[jira] [Resolved] (HADOOP-3291) Add StackWritable and QueueWritable classes

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-3291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-3291.
-

Resolution: Not A Problem

I feel these are better suited as 3rd party packages, or under projects like 
Mahout where they may be utilized as a utility class.

However, feel free to reopen if you feel they add good value to Hadoop common 
itself.

 Add StackWritable and QueueWritable classes
 ---

 Key: HADOOP-3291
 URL: https://issues.apache.org/jira/browse/HADOOP-3291
 Project: Hadoop Common
  Issue Type: New Feature
  Components: io
 Environment: All
Reporter: Dennis Kubes
Assignee: Dennis Kubes
 Attachments: HADOOP-3291-1-20080421.patch


 Adds Writable classes for FIFO Queu and LIFO Stack data structures.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-6342) Create a script to squash a common, hdfs, and mapreduce tarball into a single hadoop tarball

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6342?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-6342.
-

   Resolution: Not A Problem
Fix Version/s: (was: 0.21.1)
   0.22.0

Looks like this was not marked as resolved post Tom's comment earlier, of 
https://issues.apache.org/jira/browse/HADOOP-6846 having fixed it in 0.22

 Create a script to squash a common, hdfs, and mapreduce tarball into a single 
 hadoop tarball
 

 Key: HADOOP-6342
 URL: https://issues.apache.org/jira/browse/HADOOP-6342
 Project: Hadoop Common
  Issue Type: New Feature
  Components: build
Reporter: Owen O'Malley
Assignee: Owen O'Malley
Priority: Blocker
 Fix For: 0.22.0

 Attachments: HADOOP-6342.2.patch, HADOOP-6342.patch, h-6342.patch, 
 tar-munge, tar-munge


 It would be convenient for the transition if we had a script to take a set of 
 common, hdfs, and mapreduce tarballs and merge them into a single tarball. 
 This is intended just to help users who don't want to transition to split 
 projects for deployment immediately.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-516) Eclipse-based GUI: DFS explorer and basic Map/Reduce job launcher

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-516?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-516.


Resolution: Fixed

The plugin seems to have integrated this feature already. Should this still be 
open, please reopen it.

 Eclipse-based GUI: DFS explorer and basic Map/Reduce job launcher
 -

 Key: HADOOP-516
 URL: https://issues.apache.org/jira/browse/HADOOP-516
 Project: Hadoop Common
  Issue Type: New Feature
 Environment: Eclipse 3.2
 JDK 1.5
Reporter: Frédéric Bertin
 Attachments: hdfsExplorer.zip, hdfsExplorer2.zip


 to increase productivity in our current project (which makes a heavy use of 
 Hadoop), we wrote a small Eclipse-based GUI application which basically 
 consists in 2 views:
*  a HDFS explorer adapted from Eclipse filesystem explorer example.
  For now, it includes the following features:
  o classical tree-based browsing interface, with directory
content being detailed in a 3 columns table (file name, file
size, file type)
  o refresh button
  o delete file or directory (with confirm dialog): select files
in the tree or table and click the Delete button
  o rename file or directory: simple click on the file in the
table, type the new name and validate
  o open file with system editor: select the file in the table
and click Open button (works on Windows, not on Linux)
  o internal drag  drop
  o external drag  drop from the local filesystem to the HDFS
(the opposite doesn't work)
* a MapReduce *very* simple job launcher:
  o select the job XML configuration file
  o run the job
  o kill the job
  o visualize map and reduce progress with progress bars
  o open a browser on the Hadoop job tracker web interface 
 INSTALLATION NOTES:
  - Eclipse 3.2
  - JDK 1.5
  - import the archive in Eclipse
  - copy your hadoop conf file (hadoop-default.xml in src folder) - this 
 step should be moved in the GUI later
  - right-click on the project and Run As - Eclipse Application
  - enjoy...

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-323) IO Exception at LocalFileSystem.renameRaw, when running Nutch nightly builds (0.8-dev).

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-323?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-323.


Resolution: Invalid

Was fixed long time ago, but wasn't closed. Doesn't apply today.

I run LJRunner and it never really complains in any run about things like these.

Closing as Invalid (now).

 IO Exception at LocalFileSystem.renameRaw, when running Nutch nightly builds 
 (0.8-dev).
 ---

 Key: HADOOP-323
 URL: https://issues.apache.org/jira/browse/HADOOP-323
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Affects Versions: 0.3.2
 Environment: Windows XP + CygWin
Reporter: KuroSaka TeruHiko

 IO Exception at LocalFileSystem.renameRaw, when running Nutch nightly builds 
 (0.8-dev).
 Please see the deatil descriptions in:
 http://issues.apache.org/jira/browse/NUTCH-266
 Not knowing how to reclassify an existing bug, I am opening this new bug 
 under Hadoop.
 The version number is 0.3.3 but because I don't see it in the jira list, I 
 chose the closest matching version.  The Nutch-with-GUI build was running 
 with hadoop-0.2 but stopped running, exhibiting the same symptom with other 
 nightly builds, when switched to use hadoop-0.3.3.
 I checked fs as component but this bug could also be caused by the order in 
 which jobs are scheduled, I suspect.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-466) Startup scripts will not start instances of Hadoop daemons w/different configs w/o setting separate PID directories

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-466.


   Resolution: Fixed
Fix Version/s: 0.20.0

This problem indeed exists if one doesn't use the HADOOP_IDENT_STRING, but 
that's a valid workaround than adding a dependency on md5sum and the like (or 
do we already use it?).

I think this may be resolved as fixed with the availability of 
HADOOP_IDENT_STRING to workaround with.

Workaround (tested to work in 0.20.2):

{code}
# To start a second DN on same machine, with separated config.
HADOOP_IDENT_STRING=$USER-DN2 hadoop-daemon.sh --config /conf/dn2 start datanode
HADOOP_IDENT_STRING=$USER-DN2 hadoop-daemon.sh --config /conf/dn2 stop datanode
# These manage the PIDs as well, and will not complain that stuff is already 
running.
{code}

 Startup scripts will not start instances of Hadoop daemons w/different 
 configs w/o setting separate PID directories
 ---

 Key: HADOOP-466
 URL: https://issues.apache.org/jira/browse/HADOOP-466
 Project: Hadoop Common
  Issue Type: Improvement
  Components: conf
Affects Versions: 0.5.0
Reporter: Vetle Roeim
 Fix For: 0.20.0

 Attachments: hadoop-466.diff


 Configuration directories can be specified by either setting HADOOP_CONF_DIR 
 or using the --config command line option. However, the hadoop-daemon.sh 
 script will not start the daemons unless the PID directory is separate for 
 each configuration.
 The issue is that the code for generating PID filenames is not dependent on 
 the configuration directory. While the PID directory can be changed in 
 hadoop-env.sh, it seems a little unnecessary to have this restriction.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-378) too many files are distributed to slaves nodes

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-378?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-378.


Resolution: Incomplete

(This one might have gone stale over time, and perhaps does not apply anymore)

This one needs to explain more about what its talking about - deployment of 
Hadoop via HADOOP_MASTER which uses rsync (and thereby does not pass out 
incrementals)?

This would be resolved by better packaging, which is gonna be addressed in 
future via Bigtop and the like, hopefully. A package for examples, a package 
for test, a package for clients. With mavenized builds, this is very easy to do 
and is on its way.

Resolving as Incomplete, needs more desc.

 too many files are distributed to slaves nodes
 --

 Key: HADOOP-378
 URL: https://issues.apache.org/jira/browse/HADOOP-378
 Project: Hadoop Common
  Issue Type: Bug
  Components: conf
Reporter: Yoram Arnon
Priority: Trivial

 currently all the svn tree is sync'ed to the slaves on startup, excluding 
 only .svn files.
 that includes the example jar file, sources, and other files that are not 
 required.
 It would  be better to pick and choose the files that get sync'ed. It will 
 improve sync times on startup, and can prevent some name conflicts, since all 
 the jar files are also included in the classpath by the hadoop script.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-684) it would be nice to be able to view log and output files w/in the browser

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-684?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-684.


Resolution: Invalid

All log files are plain text files openable by a ton of tools out there, even 
browsers today, and the DFS does carry a browser along in the web UI for the 
DFS files.

Logger levels can also be set and logs can be queried via the /logs and 
/logLevel areas of any web UI under Hadoop.

Resolving as Invalid (now).

 it would be nice to be able to view log and output files w/in the browser
 -

 Key: HADOOP-684
 URL: https://issues.apache.org/jira/browse/HADOOP-684
 Project: Hadoop Common
  Issue Type: Improvement
  Components: scripts
 Environment: osx, ubuntu 6.10b
Reporter: James Todd
Priority: Minor

 it would be nice to be able to view the node logs and output files from w/in 
 the browser vs saving the file and opening an external application.
 one way to achieve this is to simply add .txt as a suffix to the relevant 
 files, eg:
 Index: bin/hadoop-daemon.sh
 ===
 --- bin/hadoop-daemon.sh(revision 468719)
 +++ bin/hadoop-daemon.sh(working copy)
 @@ -50,9 +50,9 @@
  fi
  
  # some variables
 -export HADOOP_LOGFILE=hadoop-$HADOOP_IDENT_STRING-$command-`hostname`.log
 +export HADOOP_LOGFILE=hadoop-$HADOOP_IDENT_STRING-$command-`hostname`.log.txt
  export HADOOP_ROOT_LOGGER=INFO,DRFA
 -log=$HADOOP_LOG_DIR/hadoop-$HADOOP_IDENT_STRING-$command-`hostname`.out
 +log=$HADOOP_LOG_DIR/hadoop-$HADOOP_IDENT_STRING-$command-`hostname`.out.txt
  pid=$HADOOP_PID_DIR/hadoop-$HADOOP_IDENT_STRING-$command.pid
  
  # Set default scheduling priority
 it has been suggested that perhaps the content type of the log files could be 
 specified, thereby allowing for wider/richer support.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-888) SequenceFile constructors should not accept a FileSystem parameter

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-888?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-888.


Resolution: Duplicate

This has been done by HADOOP-6856 (FS addition is now a deprecated addition 
with the revamp, Options-style builder). Resolving as duplicate.

 SequenceFile constructors should not accept a FileSystem parameter
 --

 Key: HADOOP-888
 URL: https://issues.apache.org/jira/browse/HADOOP-888
 Project: Hadoop Common
  Issue Type: Improvement
  Components: io
Affects Versions: 0.10.1
Reporter: Doug Cutting
Priority: Minor

 The SequenceFile constructors accept a FileSystem, Path and Configuration.  
 The FileSystem is redundant, as the combination of the Configuration and the 
 Path determine the FileSystem.  These methods should thus be deprecated and 
 replaced by FileSystem-less versions.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-960) Incorrect number of map tasks when there are multiple input files

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-960?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-960.


Resolution: Invalid

The ability to specify mapred.map.tasks is going away with the new API of MR. 
The only _right_ way to control splits is to have your own InputFormat that 
does it the way you need it to. The default way has worked for many (being 
local-data sensitive, as long as such information is available, but also split 
size tunable), and can also be asked to process whole files with a very simple 
subclass/configuration.

Resolving as invalid (now, and onwards) since InputFormat#getSplits(…) is not 
going anywhere, and can do what you want it to.

Regd. record num splits, MR now has NLineInputFormat as well, which indeed 
opens and reads through the file.

 Incorrect number of map tasks when there are multiple input files
 -

 Key: HADOOP-960
 URL: https://issues.apache.org/jira/browse/HADOOP-960
 Project: Hadoop Common
  Issue Type: Improvement
  Components: documentation
Affects Versions: 0.10.1
Reporter: Andrew McNabb
Priority: Minor

 This problem happens with hadoop-streaming and possibly elsewhere.  If there 
 are 5 input files, it will create 130 map tasks, even if 
 mapred.map.tasks=128.  The number of map tasks is incorrectly set to a 
 multiple of the number of files.  (I wrote a much more complete bug report, 
 but Jira lost it when it had an error, so I'm not in the mood to write it all 
 again)

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1295) hadoop-config.sh resolving symlinks leads to errors

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1295?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1295.
-

Resolution: Not A Problem

As described by Doug in the last comment here, this can be avoided if 
HADOOP_HOME is explicitly set, which ought to be easy to do when you are 
managing your versions.

Feel free to reopen if there was something else, of course.

 hadoop-config.sh resolving symlinks leads to errors 
 

 Key: HADOOP-1295
 URL: https://issues.apache.org/jira/browse/HADOOP-1295
 Project: Hadoop Common
  Issue Type: Improvement
  Components: scripts
 Environment: RHEL3
Reporter: Lee Faris
Priority: Minor

 My company uses a versioned deployment system where the final results are 
 symlinked. 
 For example:The hadoop package would be located at this location on all boxes
 /apollo/env/Hadoop/
 This is a symlink generated by the system.  The hardlink can look like this:
 box1: /apollo/env/Hadoop - /apollo/version/Hadoop-4456
 box2: /apollo/env/Hadoop - /apollo/version/Hadoop-10039445
 This piece of script from hadoop-config.sh resolves symlinks into hard links:
 while [ -h $this ]; do
   ls=`ls -ld $this`
   link=`expr $ls : '.*- \(.*\)$'`
   if expr $link : '.*/.*'  /dev/null; then
 this=$link
   else
 this=`dirname $this`/$link
   fi
 done
 I am not sure why this is done.  Commenting out the code makes things work 
 for our system.  
 I assume that was put in for a reason.  Is there a solution for the original 
 need for this code to that can work with our use case?

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1713) Create a utility to convert binary (sequence and compressed) files to strings.

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1713?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1713.
-

Resolution: Fixed

The {{hadoop fs -text}} shell utility does just this, with detection.

Looks like it was added after '07, and this appears to be a stale ticket now. 
Marking this as 'fixed'.

 Create a utility to convert binary (sequence and compressed) files to strings.
 --

 Key: HADOOP-1713
 URL: https://issues.apache.org/jira/browse/HADOOP-1713
 Project: Hadoop Common
  Issue Type: New Feature
  Components: io
Reporter: Owen O'Malley

 It would be nice to have a utility that looked at the first N bytes of a file 
 and picked a decoder for it into Strings. It could then be hooked up to 
 bin/hadoop fs cat and the web/ui to textify sequence and compressed files.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-907) Unable to set Replication factor on SequenceFile

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-907?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-907.


Resolution: Fixed

Resolved by https://issues.apache.org/jira/browse/HADOOP-6856 by using the 
ReplicationOption

 Unable to set Replication factor on SequenceFile
 

 Key: HADOOP-907
 URL: https://issues.apache.org/jira/browse/HADOOP-907
 Project: Hadoop Common
  Issue Type: Bug
  Components: io
Affects Versions: 0.10.1
 Environment: all environments
Reporter: Doug Judd
Priority: Minor

 There does not seem to be a way to set the replication factor for a 
 SequenceFile when creating one.  The API doesn't provide a way to pass it in 
 and a call to FileSystem.setReplication() does not have any effect if called 
 immediately after a call to SequenceFile.createWriter()

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1555) Fix a few more FindBugs issues

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1555?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1555.
-

Resolution: Cannot Reproduce

Running:
{code}
ant findbugs -Dfindbugs.home=/path/to/findbugs
{code}

Tells me that there does not seem to be any findbugs on common trunk. Marking 
as cannot reproduce (now).

This was opened back in '07. A new JIRA can be filed for current findbugs.

 Fix a few more FindBugs issues
 --

 Key: HADOOP-1555
 URL: https://issues.apache.org/jira/browse/HADOOP-1555
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Affects Versions: 0.13.0
Reporter: Albert Strasheim
 Attachments: findbugsv0.patch


 I'm attaching a patch to fix a few more FindBugs issues. Most of these fixes 
 are relatively minor. I also went ahead and fixed some issues in the tests.
 This patch includes some fixes for inconsistent synchronization reported by 
 FindBugs. However, FindBugs still reports some cases of inconsistent 
 synchronization that I'm not quite sure how to fix.
 Also, in ChecksumFileSystem.copyToLocalFile, there is an instanceof 
 ChecksumFileSystem check that FindBugs says will always return true. I'm not 
 quite sure what was intended here.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1371) Remove phased file system

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1371?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1371.
-

Resolution: Not A Problem

PhasedFileSystem was phased-out in 0.17 apparently.

JIRA is a no-issue now. Marking as 'Not a problem' (anymore).

 Remove phased file system
 -

 Key: HADOOP-1371
 URL: https://issues.apache.org/jira/browse/HADOOP-1371
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs
Reporter: Owen O'Malley
Assignee: Arun C Murthy

 It was deprecated in version 13 and should be removed in 14. There is a 
 little bit of cleanup in streaming.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1684) Release notes point to browsable JIRA

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1684?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1684.
-

Resolution: Not A Problem

Release notes are now being compiled and an editorial pass is made over them 
for every release as well. I believe these are at least 70% user friendly, 
having once been a beginner and read the release notes from 0.19 - 0.20. The 
process is still done in the same way since, I believe.

I do not think this JIRA is valid anymore (may be cause of time, it was opened 
in '07 when nutch hosted hadoop's links).

Thus, Not-a-problem (anymore).

 Release notes point to browsable JIRA
 -

 Key: HADOOP-1684
 URL: https://issues.apache.org/jira/browse/HADOOP-1684
 Project: Hadoop Common
  Issue Type: Improvement
  Components: documentation
Reporter: Marco Nicosia

 Currently, the web pages on lucene.apache.org/hadoop link release notes as 
 a pointer to JIRA. 
 http://issues.apache.org/jira/browse/HADOOP?report=com.atlassian.jira.plugin.system.project:changelog-panel
 The problem with this is that JIRA's really not very user-friendly. Most of 
 the Hadoop JIRA tickets describe the underlying cause, not the user exposed 
 symptoms, so users who are nsidering a new release (say, looking for API 
 changes) can't make head or tails of half of the tickets without being 
 familiar with Hadoop internals.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-397) c/c++ record io library does not use autoconf

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-397?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-397.


Resolution: Won't Fix

Resolving as Won't Fix, since the whole recordio component is now deprecated in 
favor of Avro (and technically ought to be removed in 0.22/0.23). Please see 
https://issues.apache.org/jira/browse/HADOOP-6155

 c/c++ record io library does not use autoconf
 -

 Key: HADOOP-397
 URL: https://issues.apache.org/jira/browse/HADOOP-397
 Project: Hadoop Common
  Issue Type: Improvement
  Components: record
Reporter: Owen O'Malley
Assignee: Vivek Ratan

 It would be more convenient, if the C/C++ portion of record io used autoconf 
 to configure the location of the required libraries and generate make files.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1222) Record IO C++ binding: buffer type not handled correctly

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1222?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1222.
-

Resolution: Won't Fix

Resolving as Won't Fix, since the whole recordio component is now deprecated in 
favor of Avro (and technically ought to be removed in 0.22/0.23).

Please see https://issues.apache.org/jira/browse/HADOOP-6155

 Record IO C++ binding: buffer type not handled correctly
 

 Key: HADOOP-1222
 URL: https://issues.apache.org/jira/browse/HADOOP-1222
 Project: Hadoop Common
  Issue Type: Bug
  Components: record
Reporter: David Bowen
 Attachments: test.cc


  
 I added this code to the test, which currently only tests 
 serialization/deserialization of an empty buffer.
std::string b = r1.getBufferVal();
 static char buffer[] = {0, 1, 2, 3, 4, 5};
 for (int i = 0; i  6; i++) {
   b.push_back(buffer[i]);
 }
 The csv test fails.  The generated file looks like this.
 T,102,4567,99344109427290,3.145000,1.523400,',# 0 1 2 3 4 5 0 1 2 3 4 
 5,v{},m{}
 The xml test passes, but the data in the xml file is wrong:
 valuestring000102030405000102030405000102030405/string/value

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1223) Record IO C++ binding: non-empty vector of strings does not work

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1223.
-

Resolution: Won't Fix

Resolving as Won't Fix, since the whole recordio component is now deprecated in 
favor of Avro (and technically ought to be removed in 0.22/0.23).

Please see https://issues.apache.org/jira/browse/HADOOP-6155

 Record IO C++ binding: non-empty vector of strings does not work
 

 Key: HADOOP-1223
 URL: https://issues.apache.org/jira/browse/HADOOP-1223
 Project: Hadoop Common
  Issue Type: Bug
  Components: record
Reporter: David Bowen
 Attachments: test.cc


 It works in the binary case, but not in CSV or XML.
 Here is the code to put some strings in the vector.
 std::vectorstd::string v = r1.getVectorVal();
 v.push_back(hello);
 v.push_back(world);
 In the CSV file, the strings appear twice, for some reason.  In the XML file 
 they appear three times.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1277) The class generated by Hadoop Record rcc should provide a static method to return the DDL string

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1277?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1277.
-

Resolution: Won't Fix

Resolving as Won't Fix, since the whole recordio component is now deprecated in 
favor of Avro (and technically ought to be removed in 0.22/0.23).

Please see https://issues.apache.org/jira/browse/HADOOP-6155

 The class generated by Hadoop Record rcc should provide a static method to 
 return the DDL string
 

 Key: HADOOP-1277
 URL: https://issues.apache.org/jira/browse/HADOOP-1277
 Project: Hadoop Common
  Issue Type: New Feature
  Components: record
Reporter: Runping Qi

 The method will look like:
 public static string getDDL();
 With this class, when a map/reduce job write out sequence file swith such a 
 generated class as its value class, the job
 can also save the DDL of the class into a file.
 With such a file around, we can implement a record reader that can generate 
 the required class on demand, thus, can read 
 a sequence file of Hadoop Records without having the class a priori.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1225) Record IO class should provide a toString(String charset) method

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1225.
-

Resolution: Won't Fix

Resolving as Won't Fix, since the whole recordio component is now deprecated in 
favor of Avro (and technically ought to be removed in 0.22/0.23).

Please see https://issues.apache.org/jira/browse/HADOOP-6155

 Record IO class should provide a toString(String charset) method
 

 Key: HADOOP-1225
 URL: https://issues.apache.org/jira/browse/HADOOP-1225
 Project: Hadoop Common
  Issue Type: Improvement
  Components: record
Reporter: Runping Qi
Assignee: Sameer Paranjpye

 Currently, the toString() function returns the csv format serialized form of 
 the record object.
 Unfortunately, all the fields of Buffer type are serialized into hex string. 
 Although this is a loss less conversion, it is not 
 the most convenient form, when perople use Buffer to store international 
 texts. With
 a new function toString(String charset) , the user can pass a charset to 
 indicate the desired way to convert a Buffer to a String.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1095) Provide ByteStreams in C++ version of record I/O

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1095?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1095.
-

Resolution: Won't Fix

Resolving as Won't Fix, since the whole recordio component is now deprecated in 
favor of Avro (and technically ought to be removed in 0.22/0.23).

Please see https://issues.apache.org/jira/browse/HADOOP-6155

 Provide ByteStreams in C++ version of record I/O
 

 Key: HADOOP-1095
 URL: https://issues.apache.org/jira/browse/HADOOP-1095
 Project: Hadoop Common
  Issue Type: Improvement
  Components: record
Affects Versions: 0.12.0
 Environment: All
Reporter: Milind Bhandarkar
Assignee: Vivek Ratan

  Implement ByteInStream and ByteOutStream for C++ runtime, as they will be 
 needed for using Hadoop Record I/O with forthcoming C++ MapReduce framework 
 (currently, only FileStreams are provided.) 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1227) Record IO C++ binding: cannot write more than one record to an XML stream and read them back

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1227?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1227.
-

Resolution: Won't Fix

Resolving as Won't Fix, since the whole recordio component is now deprecated in 
favor of Avro (and technically ought to be removed in 0.22/0.23).

Please see https://issues.apache.org/jira/browse/HADOOP-6155

 Record IO C++ binding: cannot write more than one record to an XML stream and 
 read them back
 

 Key: HADOOP-1227
 URL: https://issues.apache.org/jira/browse/HADOOP-1227
 Project: Hadoop Common
  Issue Type: Bug
  Components: record
Reporter: David Bowen

 I tried just writing the same record twice and then reading it back twice, 
 and got a segmentation fault.
 This works fine in the binary and csv cases.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-6712) librecordio support for xerces 3, eliminate compiler warnings and the (optional) ability to compile in the source directory

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6712?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-6712.
-

Resolution: Won't Fix

Resolving as Won't Fix, since the whole recordio component is now deprecated in 
favor of Avro (and technically ought to be removed in 0.22/0.23).

Please see https://issues.apache.org/jira/browse/HADOOP-6155

Do reopen if you wish to maintain recordio over 0.20.x branches. Although Avro 
works with 0.20 as well.

 librecordio support for xerces 3, eliminate compiler warnings and the 
 (optional) ability to compile in the source directory
 ---

 Key: HADOOP-6712
 URL: https://issues.apache.org/jira/browse/HADOOP-6712
 Project: Hadoop Common
  Issue Type: Bug
  Components: record
 Environment: 64-bit linux w/gcc 4.4.3 w/xerces 3
Reporter: John Plevyak
 Attachments: librecordio-jp-v1.patch


 I don't know if this code is current supported, but since it is in the tree 
 here are some fixes:
 1. support for xerces 3.X as well as 2.X
 the patch checks XERCES_VERSION_MAJOR and I have tested on 3.X but before
committing, someone should retest on 2.X
 2. gcc 4.4.3 on 64-bit complains about using %lld with int64_t.  Casting
 to 'long long int' solves the issue
 3. since there is currently no ant target, check if LIBRECORDIO_BUILD_DIR
 is undefined and if so assume '.' to support compiling in the source 
 directory
 This should not effect normal compilation if/when an ant target is 
 created.
 patch attached

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1916) FSShell put or CopyFromLocal incorrectly treats .

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1916?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1916.
-

Resolution: Fixed

Fixed by a redesign on trunk.

Please https://issues.apache.org/jira/browse/HADOOP-7176 for the umbrella of 
changes leading to.

 FSShell put or CopyFromLocal incorrectly treats .
 ---

 Key: HADOOP-1916
 URL: https://issues.apache.org/jira/browse/HADOOP-1916
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Affects Versions: 0.14.1
Reporter: Konstantin Shvachko
Assignee: Chris Douglas
 Attachments: 1916.patch


 The following dfs shell command
 {code}
 bin/hadoop dfs -put README.txt .
 {code}
 results in creating a file /user/user name with the contents of README.txt.
 A correct behavior would be creating a directory and a file in it: 
 /user/user name/README.txt
 The put command works correctly if /user/user name already exists.
 So the following sequence of command leads to the desired result:
 {code}
 bin/hadoop dfs -mkdir .
 bin/hadoop dfs -put README.txt .
 {code}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1919) Add option to allow Binding Jetty to localhost

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1919.
-

Resolution: Not A Problem

Not a problem.

Can be doable by the previous comment, with logical repercussions.

 Add option to allow Binding Jetty to localhost
 --

 Key: HADOOP-1919
 URL: https://issues.apache.org/jira/browse/HADOOP-1919
 Project: Hadoop Common
  Issue Type: New Feature
Affects Versions: 0.14.0
Reporter: Thurman Turner
Priority: Minor

 We would like a configurable option to have Jetty bound to the loopback 
 address of the machine so that the dfs-browser is not accessible from outside 
 the host.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HADOOP-1994) Variable names generated by Record I/O should not clash with user fields

2011-07-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1994?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HADOOP-1994.
-

Resolution: Won't Fix

Record IO has been deprecated with the advent of Avro.

Please see HADOOP-6155

Resolving as Won't Fix.

 Variable names generated by Record I/O should not clash with user fields
 

 Key: HADOOP-1994
 URL: https://issues.apache.org/jira/browse/HADOOP-1994
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Vivek Ratan
Assignee: Vivek Ratan

 The code (Java and C++) spit out by the Record I/O compiler contains 
 variables. We need to make sure these variable names don't clash with names 
 used by users in the DDL, otherwise the generated code will not compile. 
 Variable names such as 'a', 'peer', etc, are used. We need better names. For 
 example, if I have a DDL of the form
 {code}
 class s1 {
   int a;
   boolean peer;
   int a_;
 }
 {code}
 Both the Java and C++ code will not compile. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira