Can anyone help me on this issue. I have an account on the cluster and I cannot go and start server on each server process on each tasktracker.
Akhil akhil1988 wrote: > > Hi All, > > I am porting a machine learning application on Hadoop using MapReduce. The > architecture of the application goes like this: > 1. run a number of server processes which take around 2-3 minutes to start > and then remain as daemon waiting for a client to call for a connection. > During the startup these server processes get trained on the trainng > dataset. > > 2. A client is then run which connects to servers and process or test any > data that it wants to. The client is basically our job, which we will be > converted to the mapreduce model of hadoop. > > Now, since each server takes a good amount of time to start, needless to > say that we want each of these server processes to be pre-running on all > the tasktrackers(all nodes) so that when a mapreduce(client) task come to > that node, the servers are already running and the client just uses them > for its purpose. The server process keeps on running waiting for another > map task that may be assigned to that node. > > > That means, a server process is started on each node once and it waits for > a connection by a client. When clients( implemeted as map reduce) come to > that node they connect to the server, do they their processing and > leave(or finish). > > Can you please tell me how should I go about starting the server on each > node. If I am not clear, please ask any questions. Any help in this regard > will be greatly appreciated. > > Thank You! > Akhil > > -- View this message in context: http://www.nabble.com/Implementing-CLient-Server-architecture-using-MapReduce-tp23916757p23928505.html Sent from the Hadoop core-user mailing list archive at Nabble.com.