Re: editing etc hosts files of a cluster

2009-10-19 Thread Last-chance Architect

DNS ;)

Ramesh.Ramasamy wrote:

Hi,

I have a cluster setup with 3 nodes, and I'm adding hostname details (in
/etc/hosts) manually in each node. Seems it is not an effective approach.
How this scenario is handled in big clusters?

Is there any simple of way to add the hostname details in all the nodes by
editing a single entry/file/script? 


Thanks and Regards,
Ramesh




--
***
The 'Last-Chance' Architect
www.galatea.com
(US) +1 303 731 3116
(UK) +44 20 8144 4367
***


Re: Datanode Throwing NoRouteToHostException

2009-10-18 Thread Last-chance Architect
].getLocalPort() 


returned 50070
2009-10-17 17:05:40,520 INFO org.apache.hadoop.http.HttpServer: Jetty bound
to port 50070
2009-10-17 17:05:40,520 INFO org.mortbay.log: jetty-6.1.14
2009-10-17 17:05:46,612 INFO org.mortbay.log: Started
selectchannelconnec...@0.0.0.0:50070
2009-10-17 17:05:46,612 INFO
org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at:
0.0.0.0:50070
2009-10-17 17:05:46,613 INFO org.apache.hadoop.ipc.Server: IPC Server
Responder: starting
2009-10-17 17:05:46,615 INFO org.apache.hadoop.ipc.Server: IPC Server
listener on 54310: starting
2009-10-17 17:05:46,629 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 0 on 54310: starting
2009-10-17 17:05:46,634 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 1 on 54310: starting
2009-10-17 17:05:46,636 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 2 on 54310: starting
2009-10-17 17:05:46,637 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 4 on 54310: starting
2009-10-17 17:05:46,641 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 6 on 54310: starting
2009-10-17 17:05:46,642 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 7 on 54310: starting
2009-10-17 17:05:46,643 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 8 on 54310: starting
2009-10-17 17:05:46,643 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 3 on 54310: starting
2009-10-17 17:05:46,900 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 5 on 54310: starting
2009-10-17 17:05:46,916 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 9 on 54310: starting
2009-10-17 17:11:10,967 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
192.168.1.125
2009-10-17 17:11:10,968 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of transactions:
0 Total time for transactions(ms): 0Number of 

transactions batched in Syncs: 0 Number of syncs: 0 SyncTimes(ms): 0 
2009-10-17 17:11:11,518 INFO

org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll FSImage from
192.168.1.125
2009-10-17 17:11:11,519 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of transactions:
0 Total time for transactions(ms): 0Number of 

transactions batched in Syncs: 0 Number of syncs: 1 SyncTimes(ms): 2 



   


--
***
The 'Last-Chance' Architect
www.galatea.com
(US) +1 303 731 3116
(UK) +44 20 8144 4367
***


Re: Hadoop Developer Needed

2009-10-15 Thread Last-chance Architect

Hi Alex,

I'm a senior Java/J2EE developer/architect. I am American, but live 
abroad. I've been using Hadoop for the past year, but have many years 
experience in the field of large-scale data analysis, warehousing, etc.


Let me know what you think.

Cheers,

Lajos

alevin wrote:
Overview 


The SQL Data Migration Specialist plays a crucial role in converting new
Client's data onto Brilig's service platforms. We are looking for a talented
and energetic full-time freelance programmer to work both remotely and
onsite at our midtown Manhattan location. The Specialist will work with our
clients' technical teams to determine the optimal formats and requirements
to create files for subsequent import into a Brilig remote database during
the implementation process. This process involves extracting, scrubbing,
combining, transforming, validating and importing large data tables into
final data sets suitable for loading into Brilig's defined databases. The
Specialist will be responsible for creating/editing the database structure
and writing of all import scripts and programs. The ability to work on
multiple projects simultaneously while meeting tight deadlines is critical.
This project will last for 3 months but may be extended or may eventually
lead to a full time position in our fun and exciting startup. Must be able
to travel to client meetings and work independently. 

Responsibilities: 


- Subject Matter Expert on software tools used in the entire data migration
process from extraction to validation and load 
- Design, develop and execute quality data movement processes that are
consistent, repeatable and scalable 
- Streamline testing, audit and validation processes through data scrubbing
routines and presentation of audit reports prior to load 
- Roll out newly developed processes via documentation and training 
- Maintain and manage a template library of executed solutions to leverage
against future opportunities 
- Identify, clarify, and resolve issues and risks, escalating them as needed 
- Build and nourish strong business relationships with external clients 

Please include: 

- Salary Requirements 
- Availability 

Experience 

- At least 3-5 years experience in the development of java applications 
- Use of XML and other protocols for data exchange between systems 
- SQL database design and implementation 
- Experience with Eclipse, Maven, and SVN a plus 
- Experience with Htable and Hadoop a big plus 
- Excellent communication skills with both technical and non-technical
colleagues 
- Upper management and client facing skills 
- Interest in keeping up with technology advances 

PLEASE NOTE: 
US citizens and Green Card Holders and those authorized to work in the US
only. We are unable to sponsor or transfer H-1B candidates. 

Contact: 

Alex Levin, COO 
Brilig 
ale...@brilig.com 





--
***
The 'Last-Chance' Architect
www.galatea.com
(US) +1 303 731 3116
(UK) +44 20 8144 4367
***


Re: Hadoop Developer Needed

2009-10-15 Thread Last-chance Architect

Whoops guys, sorry. Hit the reply too soon ;)

Lajos


Last-chance Architect wrote:

Hi Alex,

I'm a senior Java/J2EE developer/architect. I am American, but live 
abroad. I've been using Hadoop for the past year, but have many years 
experience in the field of large-scale data analysis, warehousing, etc.


Let me know what you think.

Cheers,

Lajos

alevin wrote:

Overview
The SQL Data Migration Specialist plays a crucial role in converting new
Client's data onto Brilig's service platforms. We are looking for a 
talented

and energetic full-time freelance programmer to work both remotely and
onsite at our midtown Manhattan location. The Specialist will work 
with our
clients' technical teams to determine the optimal formats and 
requirements
to create files for subsequent import into a Brilig remote database 
during

the implementation process. This process involves extracting, scrubbing,
combining, transforming, validating and importing large data tables into
final data sets suitable for loading into Brilig's defined databases. The
Specialist will be responsible for creating/editing the database 
structure

and writing of all import scripts and programs. The ability to work on
multiple projects simultaneously while meeting tight deadlines is 
critical.

This project will last for 3 months but may be extended or may eventually
lead to a full time position in our fun and exciting startup. Must be 
able

to travel to client meetings and work independently.
Responsibilities:
- Subject Matter Expert on software tools used in the entire data 
migration
process from extraction to validation and load - Design, develop and 
execute quality data movement processes that are
consistent, repeatable and scalable - Streamline testing, audit and 
validation processes through data scrubbing
routines and presentation of audit reports prior to load - Roll out 
newly developed processes via documentation and training - Maintain 
and manage a template library of executed solutions to leverage
against future opportunities - Identify, clarify, and resolve issues 
and risks, escalating them as needed - Build and nourish strong 
business relationships with external clients

Please include:
- Salary Requirements - Availability
Experience
- At least 3-5 years experience in the development of java 
applications - Use of XML and other protocols for data exchange 
between systems - SQL database design and implementation - Experience 
with Eclipse, Maven, and SVN a plus - Experience with Htable and 
Hadoop a big plus - Excellent communication skills with both technical 
and non-technical
colleagues - Upper management and client facing skills - Interest in 
keeping up with technology advances
PLEASE NOTE: US citizens and Green Card Holders and those authorized 
to work in the US

only. We are unable to sponsor or transfer H-1B candidates.
Contact:
Alex Levin, COO Brilig ale...@brilig.com





--
***
The 'Last-Chance' Architect
www.galatea.com
(US) +1 303 731 3116
(UK) +44 20 8144 4367
***


Re: Map Recude code doubt

2009-10-15 Thread Last-chance Architect

Shwitzu,

Why can't you just use query thru a Filesystem object and find the file 
you want?


Lajos


shwitzu wrote:

Hello All,

I was wondering if our map reduce code can just return the location of the
file? Or place the actual file in a given output directory by searching
based on a keyword.

Let me make myself clear

If there are 100 image files in my HDFS and I want to extract one image
file. If I give a keyword which matches the name of a file in my HDFS will
my mapper and  reducer code be able to locate that file and put back the 
original file in a given location??


Please let me know if you have need more information.

Thanks


--
***
The 'Last-Chance' Architect
www.galatea.com
(US) +1 303 731 3116
(UK) +44 20 8144 4367
***