Hi all
not sure if this is discussed before, but does the merging for ORC works:
alter table ... partition() concatenate for ORC
, tried several times never been successful and some google search show
some failure encountered by others as well. is this best way to merge small
ORC files??
many
if i've not mistaken the -Dhbase.import.version=0.98 should be put as
-Dhbase.import.version=0.94
On Thu, Jan 8, 2015 at 5:05 PM, dylan...@163.com dylan...@163.com wrote:
Hi,
How to migrate hbase table from hbase-0.94 to hbase-0.98, which both
belong to different hadoop clusters.
I had
Hi,
In mapreduce with reduce output format of NullWritable,
AvroValuemyavroclass
Is it possible for me to get the reduce output without the string
key:none and value , i.e. only the content of myavroclass itself??
Thanks!
what i get now:
{u'value': {u'MYCOLUMN1': 'Hello', u'MYCOLUMN2':
Configuration conf = getConf();
conf.setLong(mapreduce.input.fileinputformat.split.maxsize,1000);
// u can set this to some small value (in bytes) to ensure your file will
split to multiple mappers , provided the format is not un-splitable format
like .snappy.
On Tue, Jul 8, 2014 at 7:32
is there any setting that can set on run time of job for maximum mapper
concurrently running ?
i know there is a jobtracker level parameter that can be set, but that will
be global parameter for every job. Is it possible to set per job ?
Hi
some questions on yarn+fairscheduler:
1. is the maxMaps and maxReduces in allocations.xml actually work? notice
it is not working in my setup.
2. my job always fail with diagnostics such as: Task
task_1400033851458_4824_m_06 failed 4 times .
is it possible that this is due to preempted too
, Du Lam delim123...@gmail.com wrote:
Hi
some questions on yarn+fairscheduler:
1. is the maxMaps and maxReduces in allocations.xml actually work?
notice it
is not working in my setup.
2. my job always fail with diagnostics such as: Task
task_1400033851458_4824_m_06 failed 4 times