Re: Yay for 1.0.0! EC2 Still has problems.
I am also running into "modules/mod_authn_alias.so" issue on r3.8xlarge when launched cluster with ./spark-ec2; so ganglia is not accessible. From the posts it seems that Patrick suggested using Ubuntu 12.04. Can you please provide name of AMI that can be used with -a flag that will not have this issue? - I am running script with "--spark-git-repo=https://github.com/apache/spark";, which I assume should deploy the latest code. - I have been able to launch cluster on m2.4xlarge, where ganglia works. - From what I understand we are not supposed to use any random AMI??; it will be helpful to publish list of AMIS that people use with different instances. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Yay-for-1-0-0-EC2-Still-has-problems-tp6578p9307.html Sent from the Apache Spark User List mailing list archive at Nabble.com.
Re: Yay for 1.0.0! EC2 Still has problems.
I am also running into "modules/mod_authn_alias.so" issue on r3.8xlarge when launched cluster with ./spark-ec2; so ganglia is not accessible. From the posts it seems that Patrick suggested using Ubuntu 12.04. Can you please provide name of AMI that can be used with -a flag that will not have this issue? - I am running script with "--spark-git-repo=https://github.com/apache/spark";, which I assume should deploy the latest code. - I have been able to launch cluster on m2.4xlarge, where ganglia works. - From what I understand we are not supposed to use any random AMI??; it will be helpful to publish list of AMIS that people use with different instances. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Yay-for-1-0-0-EC2-Still-has-problems-tp6578p9306.html Sent from the Apache Spark User List mailing list archive at Nabble.com.
Re: Yay for 1.0.0! EC2 Still has problems.
On Wed, Jun 4, 2014 at 9:35 AM, Jeremy Lee wrote: > Oh, I went back to m1.large while those issues get sorted out. Random side note, Amazon is deprecating the m1 instances in favor of m3 instances, which have SSDs and more ECUs than their m1 counterparts. m3.2xlarge has 30GB of RAM and may be a good-enough substitution for the r3 instances for you for the time being. Nick
Re: Yay for 1.0.0! EC2 Still has problems.
On Wed, Jun 4, 2014 at 12:31 PM, Matei Zaharia wrote: > Ah, sorry to hear you had more problems. Some thoughts on them: > There will always be more problems, 'tis the nature of coding. :-) I try not to bother the list until I've smacked my head against them for a few hours, so it's only the "most confusing" stuff I pour out here. I'm actually progressing pretty well. > (the streaming.Twitter ones especially) depend on there being a > "/mnt/spark" and "/mnt2/spark" directory (I think for java tempfiles?) and > those don't seem to exist out-of-the-box. > > I think this is a side-effect of the r3 instances not having those drives > mounted. Our setup script would normally create these directories. What was > the error? > > Oh, I went back to m1.large while those issues get sorted out. I decided I had enough problems without messing with that too. (seriously, why does Amazon do these things? It's like they _try_ to make the instances incompatible.) I forget the exact error, but it traced through createTempFile and it was fairly clear about the directory being missing. Things like "bin/run-example SparkPi" worked fine, but I'll bet twitter4j creates temp files, so "bin/run-example streaming.TwitterPopularTags" broke. What did you change log4j.properties to? It should be changed to say > log4j.rootCategory=WARN, console but maybe another log4j.properties is > somehow arriving on the classpath. This is definitely a common problem so > we need to add some explicit docs on it. > I seem to have this sorted out, don't ask me how. Once again I was probably editing things on the cluster master when I should have been editing the cluster controller, or vice versa. But, yeah, many of the examples just get lost in a sea of DAG INFO messages. > Are you going through http://spark.apache.org/docs/latest/quick-start.html? > You should be able to do just sbt package. Once you do that you don’t need > to deploy your application’s JAR to the cluster, just pass it to > spark-submit and it will automatically be sent over. > Ah, that answers another question I just asked elsewhere... Yup, I re-read pretty much every documentation page daily. And I'm making my way through every video. > > Meanwhile I'm learning scala... Great Turing's Ghost, it's the dream > language we've theorized about for years! I hadn't realized! > > Indeed, glad you’re enjoying it. > "Enjoying", not yet alas, I'm sure I'll get there. But I do understand the implications of a mixed functional-imperative language with closures and lambdas. That is serious voodoo. -- Jeremy Lee BCompSci(Hons) The Unorthodox Engineers
Re: Yay for 1.0.0! EC2 Still has problems.
Ah, sorry to hear you had more problems. Some thoughts on them: > Thanks for that, Matei! I'll look at that once I get a spare moment. :-) > > If you like, I'll keep documenting my newbie problems and frustrations... > perhaps it might make things easier for others. > > Another issue I seem to have found (now that I can get small clusters up): > some of the examples (the streaming.Twitter ones especially) depend on there > being a "/mnt/spark" and "/mnt2/spark" directory (I think for java > tempfiles?) and those don't seem to exist out-of-the-box. I have to create > those directories and use "copy-dir" to get them to the workers before those > examples run. I think this is a side-effect of the r3 instances not having those drives mounted. Our setup script would normally create these directories. What was the error? > Much of the the last two days for me have been about failing to get any of my > own code to work, except for in spark-shell. (which is very nice, btw) > > At first I tried editing the examples, because I took the documentation > literally when it said "Finally, Spark includes several samples in the > examples directory (Scala, Java, Python). You can run them as follows:" but > of course didn't realize editing them is pointless because while the source > is there, the code is actually pulled from a .jar elsewhere. Doh. (so obvious > in hindsight) > > I couldn't even turn down the voluminous INFO messages to WARNs, no matter > how many conf/log4j.properties files I edited or copy-dir'd. I'm sure there's > a trick to that I'm not getting. What did you change log4j.properties to? It should be changed to say log4j.rootCategory=WARN, console but maybe another log4j.properties is somehow arriving on the classpath. This is definitely a common problem so we need to add some explicit docs on it. > Even trying to build SimpleApp I've run into the problem that all the > documentation says to use "sbt/sbt assemble", but sbt doesn't seem to be in > the 1.0.0 pre-built packages that I downloaded. Are you going through http://spark.apache.org/docs/latest/quick-start.html? You should be able to do just sbt package. Once you do that you don’t need to deploy your application’s JAR to the cluster, just pass it to spark-submit and it will automatically be sent over. > Ah... yes.. there it is in the source package. I suppose that means that in > order to deploy any new code to the cluster, I've got to rebuild from source > on my "cluster controller". OK, I never liked that Amazon Linux AMI anyway. > I'm going to start from scratch again with an Ubuntu 12.04 instance, > hopefully that will be more auspicious... > > Meanwhile I'm learning scala... Great Turing's Ghost, it's the dream language > we've theorized about for years! I hadn't realized! Indeed, glad you’re enjoying it. Matei > > > > On Mon, Jun 2, 2014 at 12:05 PM, Matei Zaharia > wrote: > FYI, I opened https://issues.apache.org/jira/browse/SPARK-1990 to track this. > Matei > > > On Jun 1, 2014, at 6:14 PM, Jeremy Lee wrote: > >> Sort of.. there were two separate issues, but both related to AWS.. >> >> I've sorted the confusion about the Master/Worker AMI ... use the version >> chosen by the scripts. (and use the right instance type so the script can >> choose wisely) >> >> But yes, one also needs a "launch machine" to kick off the cluster, and for >> that I _also_ was using an Amazon instance... (made sense.. I have a team >> that will needs to do things as well, not just me) and I was just pointing >> out that if you use the "most recommended by Amazon" AMI (for your free >> micro instance, for example) you get python 2.6 and the ec2 scripts fail. >> >> That merely needs a line in the documentation saying "use Ubuntu for your >> cluster controller, not Amazon Linux" or somesuch. But yeah, for a newbie, >> it was hard working out when to use "default" or "custom" AMIs for various >> parts of the setup. >> >> >> On Mon, Jun 2, 2014 at 4:01 AM, Patrick Wendell wrote: >> Hey just to clarify this - my understanding is that the poster >> (Jeremey) was using a custom AMI to *launch* spark-ec2. I normally >> launch spark-ec2 from my laptop. And he was looking for an AMI that >> had a high enough version of python. >> >> Spark-ec2 itself has a flag "-a" that allows you to give a specific >> AMI. This flag is just an internal tool that we use for testing when >> we spin new AMI's. Users can't set that to an arbitrary AMI because we >> tightly control things like the Java and OS versions, libraries, etc. >> >> >> On Sun, Jun 1, 2014 at 12:51 AM, Jeremy Lee >> wrote: >> > *sigh* OK, I figured it out. (Thank you Nick, for the hint) >> > >> > "m1.large" works, (I swear I tested that earlier and had similar issues... >> > ) >> > >> > It was my obsession with starting "r3.*large" instances. Clearly I hadn't >> > patched the script in all the places.. which I think caused it to default >> > to >>
Re: Yay for 1.0.0! EC2 Still has problems.
Thanks for that, Matei! I'll look at that once I get a spare moment. :-) If you like, I'll keep documenting my newbie problems and frustrations... perhaps it might make things easier for others. Another issue I seem to have found (now that I can get small clusters up): some of the examples (the streaming.Twitter ones especially) depend on there being a "/mnt/spark" and "/mnt2/spark" directory (I think for java tempfiles?) and those don't seem to exist out-of-the-box. I have to create those directories and use "copy-dir" to get them to the workers before those examples run. Much of the the last two days for me have been about failing to get any of my own code to work, except for in spark-shell. (which is very nice, btw) At first I tried editing the examples, because I took the documentation literally when it said "Finally, Spark includes several samples in the examples directory (Scala, Java, Python). You can run them as follows:" but of course didn't realize editing them is pointless because while the source is there, the code is actually pulled from a .jar elsewhere. Doh. (so obvious in hindsight) I couldn't even turn down the voluminous INFO messages to WARNs, no matter how many conf/log4j.properties files I edited or copy-dir'd. I'm sure there's a trick to that I'm not getting. Even trying to build SimpleApp I've run into the problem that all the documentation says to use "sbt/sbt assemble", but sbt doesn't seem to be in the 1.0.0 pre-built packages that I downloaded. Ah... yes.. there it is in the source package. I suppose that means that in order to deploy any new code to the cluster, I've got to rebuild from source on my "cluster controller". OK, I never liked that Amazon Linux AMI anyway. I'm going to start from scratch again with an Ubuntu 12.04 instance, hopefully that will be more auspicious... Meanwhile I'm learning scala... Great Turing's Ghost, it's the dream language we've theorized about for years! I hadn't realized! On Mon, Jun 2, 2014 at 12:05 PM, Matei Zaharia wrote: > FYI, I opened https://issues.apache.org/jira/browse/SPARK-1990 to track > this. > > Matei > > > On Jun 1, 2014, at 6:14 PM, Jeremy Lee > wrote: > > Sort of.. there were two separate issues, but both related to AWS.. > > I've sorted the confusion about the Master/Worker AMI ... use the version > chosen by the scripts. (and use the right instance type so the script can > choose wisely) > > But yes, one also needs a "launch machine" to kick off the cluster, and > for that I _also_ was using an Amazon instance... (made sense.. I have a > team that will needs to do things as well, not just me) and I was just > pointing out that if you use the "most recommended by Amazon" AMI (for your > free micro instance, for example) you get python 2.6 and the ec2 scripts > fail. > > That merely needs a line in the documentation saying "use Ubuntu for your > cluster controller, not Amazon Linux" or somesuch. But yeah, for a newbie, > it was hard working out when to use "default" or "custom" AMIs for various > parts of the setup. > > > On Mon, Jun 2, 2014 at 4:01 AM, Patrick Wendell > wrote: > >> Hey just to clarify this - my understanding is that the poster >> (Jeremey) was using a custom AMI to *launch* spark-ec2. I normally >> launch spark-ec2 from my laptop. And he was looking for an AMI that >> had a high enough version of python. >> >> Spark-ec2 itself has a flag "-a" that allows you to give a specific >> AMI. This flag is just an internal tool that we use for testing when >> we spin new AMI's. Users can't set that to an arbitrary AMI because we >> tightly control things like the Java and OS versions, libraries, etc. >> >> >> On Sun, Jun 1, 2014 at 12:51 AM, Jeremy Lee >> wrote: >> > *sigh* OK, I figured it out. (Thank you Nick, for the hint) >> > >> > "m1.large" works, (I swear I tested that earlier and had similar >> issues... ) >> > >> > It was my obsession with starting "r3.*large" instances. Clearly I >> hadn't >> > patched the script in all the places.. which I think caused it to >> default to >> > the Amazon AMI. I'll have to take a closer look at the code and see if I >> > can't fix it correctly, because I really, really do want nodes with 2x >> the >> > CPU and 4x the memory for the same low spot price. :-) >> > >> > I've got a cluster up now, at least. Time for the fun stuff... >> > >> > Thanks everyone for the help! >> > >> > >> > >> > On Sun, Jun 1, 2014 at 5:19 PM, Nicholas Chammas >> > wrote: >> >> >> >> If you are explicitly specifying the AMI in your invocation of >> spark-ec2, >> >> may I suggest simply removing any explicit mention of AMI from your >> >> invocation? spark-ec2 automatically selects an appropriate AMI based >> on the >> >> specified instance type. >> >> >> >> 2014년 6월 1일 일요일, Nicholas Chammas님이 작성한 >> 메시지: >> >> >> >>> Could you post how exactly you are invoking spark-ec2? And are you >> having >> >>> trouble just with r3 instances, or with any instance type? >> >>> >> >>> 2014년 6월 1일 일요
Re: Yay for 1.0.0! EC2 Still has problems.
FYI, I opened https://issues.apache.org/jira/browse/SPARK-1990 to track this. Matei On Jun 1, 2014, at 6:14 PM, Jeremy Lee wrote: > Sort of.. there were two separate issues, but both related to AWS.. > > I've sorted the confusion about the Master/Worker AMI ... use the version > chosen by the scripts. (and use the right instance type so the script can > choose wisely) > > But yes, one also needs a "launch machine" to kick off the cluster, and for > that I _also_ was using an Amazon instance... (made sense.. I have a team > that will needs to do things as well, not just me) and I was just pointing > out that if you use the "most recommended by Amazon" AMI (for your free micro > instance, for example) you get python 2.6 and the ec2 scripts fail. > > That merely needs a line in the documentation saying "use Ubuntu for your > cluster controller, not Amazon Linux" or somesuch. But yeah, for a newbie, it > was hard working out when to use "default" or "custom" AMIs for various parts > of the setup. > > > On Mon, Jun 2, 2014 at 4:01 AM, Patrick Wendell wrote: > Hey just to clarify this - my understanding is that the poster > (Jeremey) was using a custom AMI to *launch* spark-ec2. I normally > launch spark-ec2 from my laptop. And he was looking for an AMI that > had a high enough version of python. > > Spark-ec2 itself has a flag "-a" that allows you to give a specific > AMI. This flag is just an internal tool that we use for testing when > we spin new AMI's. Users can't set that to an arbitrary AMI because we > tightly control things like the Java and OS versions, libraries, etc. > > > On Sun, Jun 1, 2014 at 12:51 AM, Jeremy Lee > wrote: > > *sigh* OK, I figured it out. (Thank you Nick, for the hint) > > > > "m1.large" works, (I swear I tested that earlier and had similar issues... ) > > > > It was my obsession with starting "r3.*large" instances. Clearly I hadn't > > patched the script in all the places.. which I think caused it to default to > > the Amazon AMI. I'll have to take a closer look at the code and see if I > > can't fix it correctly, because I really, really do want nodes with 2x the > > CPU and 4x the memory for the same low spot price. :-) > > > > I've got a cluster up now, at least. Time for the fun stuff... > > > > Thanks everyone for the help! > > > > > > > > On Sun, Jun 1, 2014 at 5:19 PM, Nicholas Chammas > > wrote: > >> > >> If you are explicitly specifying the AMI in your invocation of spark-ec2, > >> may I suggest simply removing any explicit mention of AMI from your > >> invocation? spark-ec2 automatically selects an appropriate AMI based on the > >> specified instance type. > >> > >> 2014년 6월 1일 일요일, Nicholas Chammas님이 작성한 메시지: > >> > >>> Could you post how exactly you are invoking spark-ec2? And are you having > >>> trouble just with r3 instances, or with any instance type? > >>> > >>> 2014년 6월 1일 일요일, Jeremy Lee님이 작성한 메시지: > >>> > >>> It's been another day of spinning up dead clusters... > >>> > >>> I thought I'd finally worked out what everyone else knew - don't use the > >>> default AMI - but I've now run through all of the "official" quick-start > >>> linux releases and I'm none the wiser: > >>> > >>> Amazon Linux AMI 2014.03.1 - ami-7aba833f (64-bit) > >>> Provisions servers, connects, installs, but the webserver on the master > >>> will not start > >>> > >>> Red Hat Enterprise Linux 6.5 (HVM) - ami-5cdce419 > >>> Spot instance requests are not supported for this AMI. > >>> > >>> SuSE Linux Enterprise Server 11 sp3 (HVM) - ami-1a88bb5f > >>> Not tested - costs 10x more for spot instances, not economically viable. > >>> > >>> Ubuntu Server 14.04 LTS (HVM) - ami-f64f77b3 > >>> Provisions servers, but "git" is not pre-installed, so the cluster setup > >>> fails. > >>> > >>> Amazon Linux AMI (HVM) 2014.03.1 - ami-5aba831f > >>> Provisions servers, but "git" is not pre-installed, so the cluster setup > >>> fails. > > > > > > > > > > -- > > Jeremy Lee BCompSci(Hons) > > The Unorthodox Engineers > > > > -- > Jeremy Lee BCompSci(Hons) > The Unorthodox Engineers
Re: Yay for 1.0.0! EC2 Still has problems.
Sort of.. there were two separate issues, but both related to AWS.. I've sorted the confusion about the Master/Worker AMI ... use the version chosen by the scripts. (and use the right instance type so the script can choose wisely) But yes, one also needs a "launch machine" to kick off the cluster, and for that I _also_ was using an Amazon instance... (made sense.. I have a team that will needs to do things as well, not just me) and I was just pointing out that if you use the "most recommended by Amazon" AMI (for your free micro instance, for example) you get python 2.6 and the ec2 scripts fail. That merely needs a line in the documentation saying "use Ubuntu for your cluster controller, not Amazon Linux" or somesuch. But yeah, for a newbie, it was hard working out when to use "default" or "custom" AMIs for various parts of the setup. On Mon, Jun 2, 2014 at 4:01 AM, Patrick Wendell wrote: > Hey just to clarify this - my understanding is that the poster > (Jeremey) was using a custom AMI to *launch* spark-ec2. I normally > launch spark-ec2 from my laptop. And he was looking for an AMI that > had a high enough version of python. > > Spark-ec2 itself has a flag "-a" that allows you to give a specific > AMI. This flag is just an internal tool that we use for testing when > we spin new AMI's. Users can't set that to an arbitrary AMI because we > tightly control things like the Java and OS versions, libraries, etc. > > > On Sun, Jun 1, 2014 at 12:51 AM, Jeremy Lee > wrote: > > *sigh* OK, I figured it out. (Thank you Nick, for the hint) > > > > "m1.large" works, (I swear I tested that earlier and had similar > issues... ) > > > > It was my obsession with starting "r3.*large" instances. Clearly I hadn't > > patched the script in all the places.. which I think caused it to > default to > > the Amazon AMI. I'll have to take a closer look at the code and see if I > > can't fix it correctly, because I really, really do want nodes with 2x > the > > CPU and 4x the memory for the same low spot price. :-) > > > > I've got a cluster up now, at least. Time for the fun stuff... > > > > Thanks everyone for the help! > > > > > > > > On Sun, Jun 1, 2014 at 5:19 PM, Nicholas Chammas > > wrote: > >> > >> If you are explicitly specifying the AMI in your invocation of > spark-ec2, > >> may I suggest simply removing any explicit mention of AMI from your > >> invocation? spark-ec2 automatically selects an appropriate AMI based on > the > >> specified instance type. > >> > >> 2014년 6월 1일 일요일, Nicholas Chammas님이 작성한 > 메시지: > >> > >>> Could you post how exactly you are invoking spark-ec2? And are you > having > >>> trouble just with r3 instances, or with any instance type? > >>> > >>> 2014년 6월 1일 일요일, Jeremy Lee님이 작성한 메시지: > >>> > >>> It's been another day of spinning up dead clusters... > >>> > >>> I thought I'd finally worked out what everyone else knew - don't use > the > >>> default AMI - but I've now run through all of the "official" > quick-start > >>> linux releases and I'm none the wiser: > >>> > >>> Amazon Linux AMI 2014.03.1 - ami-7aba833f (64-bit) > >>> Provisions servers, connects, installs, but the webserver on the master > >>> will not start > >>> > >>> Red Hat Enterprise Linux 6.5 (HVM) - ami-5cdce419 > >>> Spot instance requests are not supported for this AMI. > >>> > >>> SuSE Linux Enterprise Server 11 sp3 (HVM) - ami-1a88bb5f > >>> Not tested - costs 10x more for spot instances, not economically > viable. > >>> > >>> Ubuntu Server 14.04 LTS (HVM) - ami-f64f77b3 > >>> Provisions servers, but "git" is not pre-installed, so the cluster > setup > >>> fails. > >>> > >>> Amazon Linux AMI (HVM) 2014.03.1 - ami-5aba831f > >>> Provisions servers, but "git" is not pre-installed, so the cluster > setup > >>> fails. > > > > > > > > > > -- > > Jeremy Lee BCompSci(Hons) > > The Unorthodox Engineers > -- Jeremy Lee BCompSci(Hons) The Unorthodox Engineers
Re: Yay for 1.0.0! EC2 Still has problems.
More specifically with the -a flag, you *can* set your own AMI, but you’ll need to base it off ours. This is because spark-ec2 assumes that some packages (e.g. java, Python 2.6) are already available on the AMI. Matei On Jun 1, 2014, at 11:01 AM, Patrick Wendell wrote: > Hey just to clarify this - my understanding is that the poster > (Jeremey) was using a custom AMI to *launch* spark-ec2. I normally > launch spark-ec2 from my laptop. And he was looking for an AMI that > had a high enough version of python. > > Spark-ec2 itself has a flag "-a" that allows you to give a specific > AMI. This flag is just an internal tool that we use for testing when > we spin new AMI's. Users can't set that to an arbitrary AMI because we > tightly control things like the Java and OS versions, libraries, etc. > > > On Sun, Jun 1, 2014 at 12:51 AM, Jeremy Lee > wrote: >> *sigh* OK, I figured it out. (Thank you Nick, for the hint) >> >> "m1.large" works, (I swear I tested that earlier and had similar issues... ) >> >> It was my obsession with starting "r3.*large" instances. Clearly I hadn't >> patched the script in all the places.. which I think caused it to default to >> the Amazon AMI. I'll have to take a closer look at the code and see if I >> can't fix it correctly, because I really, really do want nodes with 2x the >> CPU and 4x the memory for the same low spot price. :-) >> >> I've got a cluster up now, at least. Time for the fun stuff... >> >> Thanks everyone for the help! >> >> >> >> On Sun, Jun 1, 2014 at 5:19 PM, Nicholas Chammas >> wrote: >>> >>> If you are explicitly specifying the AMI in your invocation of spark-ec2, >>> may I suggest simply removing any explicit mention of AMI from your >>> invocation? spark-ec2 automatically selects an appropriate AMI based on the >>> specified instance type. >>> >>> 2014년 6월 1일 일요일, Nicholas Chammas님이 작성한 메시지: >>> Could you post how exactly you are invoking spark-ec2? And are you having trouble just with r3 instances, or with any instance type? 2014년 6월 1일 일요일, Jeremy Lee님이 작성한 메시지: It's been another day of spinning up dead clusters... I thought I'd finally worked out what everyone else knew - don't use the default AMI - but I've now run through all of the "official" quick-start linux releases and I'm none the wiser: Amazon Linux AMI 2014.03.1 - ami-7aba833f (64-bit) Provisions servers, connects, installs, but the webserver on the master will not start Red Hat Enterprise Linux 6.5 (HVM) - ami-5cdce419 Spot instance requests are not supported for this AMI. SuSE Linux Enterprise Server 11 sp3 (HVM) - ami-1a88bb5f Not tested - costs 10x more for spot instances, not economically viable. Ubuntu Server 14.04 LTS (HVM) - ami-f64f77b3 Provisions servers, but "git" is not pre-installed, so the cluster setup fails. Amazon Linux AMI (HVM) 2014.03.1 - ami-5aba831f Provisions servers, but "git" is not pre-installed, so the cluster setup fails. >> >> >> >> >> -- >> Jeremy Lee BCompSci(Hons) >> The Unorthodox Engineers
Re: Yay for 1.0.0! EC2 Still has problems.
Ah yes, looking back at the first email in the thread, indeed that was the case. For the record, I too launch clusters from my laptop, where I have Python 2.7 installed. On Sun, Jun 1, 2014 at 2:01 PM, Patrick Wendell wrote: > Hey just to clarify this - my understanding is that the poster > (Jeremey) was using a custom AMI to *launch* spark-ec2. I normally > launch spark-ec2 from my laptop. And he was looking for an AMI that > had a high enough version of python. > > Spark-ec2 itself has a flag "-a" that allows you to give a specific > AMI. This flag is just an internal tool that we use for testing when > we spin new AMI's. Users can't set that to an arbitrary AMI because we > tightly control things like the Java and OS versions, libraries, etc. > > > On Sun, Jun 1, 2014 at 12:51 AM, Jeremy Lee > wrote: > > *sigh* OK, I figured it out. (Thank you Nick, for the hint) > > > > "m1.large" works, (I swear I tested that earlier and had similar > issues... ) > > > > It was my obsession with starting "r3.*large" instances. Clearly I hadn't > > patched the script in all the places.. which I think caused it to > default to > > the Amazon AMI. I'll have to take a closer look at the code and see if I > > can't fix it correctly, because I really, really do want nodes with 2x > the > > CPU and 4x the memory for the same low spot price. :-) > > > > I've got a cluster up now, at least. Time for the fun stuff... > > > > Thanks everyone for the help! > > > > > > > > On Sun, Jun 1, 2014 at 5:19 PM, Nicholas Chammas > > wrote: > >> > >> If you are explicitly specifying the AMI in your invocation of > spark-ec2, > >> may I suggest simply removing any explicit mention of AMI from your > >> invocation? spark-ec2 automatically selects an appropriate AMI based on > the > >> specified instance type. > >> > >> 2014년 6월 1일 일요일, Nicholas Chammas님이 작성한 > 메시지: > >> > >>> Could you post how exactly you are invoking spark-ec2? And are you > having > >>> trouble just with r3 instances, or with any instance type? > >>> > >>> 2014년 6월 1일 일요일, Jeremy Lee님이 작성한 메시지: > >>> > >>> It's been another day of spinning up dead clusters... > >>> > >>> I thought I'd finally worked out what everyone else knew - don't use > the > >>> default AMI - but I've now run through all of the "official" > quick-start > >>> linux releases and I'm none the wiser: > >>> > >>> Amazon Linux AMI 2014.03.1 - ami-7aba833f (64-bit) > >>> Provisions servers, connects, installs, but the webserver on the master > >>> will not start > >>> > >>> Red Hat Enterprise Linux 6.5 (HVM) - ami-5cdce419 > >>> Spot instance requests are not supported for this AMI. > >>> > >>> SuSE Linux Enterprise Server 11 sp3 (HVM) - ami-1a88bb5f > >>> Not tested - costs 10x more for spot instances, not economically > viable. > >>> > >>> Ubuntu Server 14.04 LTS (HVM) - ami-f64f77b3 > >>> Provisions servers, but "git" is not pre-installed, so the cluster > setup > >>> fails. > >>> > >>> Amazon Linux AMI (HVM) 2014.03.1 - ami-5aba831f > >>> Provisions servers, but "git" is not pre-installed, so the cluster > setup > >>> fails. > > > > > > > > > > -- > > Jeremy Lee BCompSci(Hons) > > The Unorthodox Engineers >
Re: Yay for 1.0.0! EC2 Still has problems.
Hey just to clarify this - my understanding is that the poster (Jeremey) was using a custom AMI to *launch* spark-ec2. I normally launch spark-ec2 from my laptop. And he was looking for an AMI that had a high enough version of python. Spark-ec2 itself has a flag "-a" that allows you to give a specific AMI. This flag is just an internal tool that we use for testing when we spin new AMI's. Users can't set that to an arbitrary AMI because we tightly control things like the Java and OS versions, libraries, etc. On Sun, Jun 1, 2014 at 12:51 AM, Jeremy Lee wrote: > *sigh* OK, I figured it out. (Thank you Nick, for the hint) > > "m1.large" works, (I swear I tested that earlier and had similar issues... ) > > It was my obsession with starting "r3.*large" instances. Clearly I hadn't > patched the script in all the places.. which I think caused it to default to > the Amazon AMI. I'll have to take a closer look at the code and see if I > can't fix it correctly, because I really, really do want nodes with 2x the > CPU and 4x the memory for the same low spot price. :-) > > I've got a cluster up now, at least. Time for the fun stuff... > > Thanks everyone for the help! > > > > On Sun, Jun 1, 2014 at 5:19 PM, Nicholas Chammas > wrote: >> >> If you are explicitly specifying the AMI in your invocation of spark-ec2, >> may I suggest simply removing any explicit mention of AMI from your >> invocation? spark-ec2 automatically selects an appropriate AMI based on the >> specified instance type. >> >> 2014년 6월 1일 일요일, Nicholas Chammas님이 작성한 메시지: >> >>> Could you post how exactly you are invoking spark-ec2? And are you having >>> trouble just with r3 instances, or with any instance type? >>> >>> 2014년 6월 1일 일요일, Jeremy Lee님이 작성한 메시지: >>> >>> It's been another day of spinning up dead clusters... >>> >>> I thought I'd finally worked out what everyone else knew - don't use the >>> default AMI - but I've now run through all of the "official" quick-start >>> linux releases and I'm none the wiser: >>> >>> Amazon Linux AMI 2014.03.1 - ami-7aba833f (64-bit) >>> Provisions servers, connects, installs, but the webserver on the master >>> will not start >>> >>> Red Hat Enterprise Linux 6.5 (HVM) - ami-5cdce419 >>> Spot instance requests are not supported for this AMI. >>> >>> SuSE Linux Enterprise Server 11 sp3 (HVM) - ami-1a88bb5f >>> Not tested - costs 10x more for spot instances, not economically viable. >>> >>> Ubuntu Server 14.04 LTS (HVM) - ami-f64f77b3 >>> Provisions servers, but "git" is not pre-installed, so the cluster setup >>> fails. >>> >>> Amazon Linux AMI (HVM) 2014.03.1 - ami-5aba831f >>> Provisions servers, but "git" is not pre-installed, so the cluster setup >>> fails. > > > > > -- > Jeremy Lee BCompSci(Hons) > The Unorthodox Engineers
Re: Yay for 1.0.0! EC2 Still has problems.
*sigh* OK, I figured it out. (Thank you Nick, for the hint) "m1.large" works, (I swear I tested that earlier and had similar issues... ) It was my obsession with starting "r3.*large" instances. Clearly I hadn't patched the script in all the places.. which I think caused it to default to the Amazon AMI. I'll have to take a closer look at the code and see if I can't fix it correctly, because I really, really do want nodes with 2x the CPU and 4x the memory for the same low spot price. :-) I've got a cluster up now, at least. Time for the fun stuff... Thanks everyone for the help! On Sun, Jun 1, 2014 at 5:19 PM, Nicholas Chammas wrote: > If you are explicitly specifying the AMI in your invocation of spark-ec2, > may I suggest simply removing any explicit mention of AMI from your > invocation? spark-ec2 automatically selects an appropriate AMI based on > the specified instance type. > > 2014년 6월 1일 일요일, Nicholas Chammas님이 작성한 메시지: > > Could you post how exactly you are invoking spark-ec2? And are you having >> trouble just with r3 instances, or with any instance type? >> >> 2014년 6월 1일 일요일, Jeremy Lee님이 작성한 메시지: >> >> It's been another day of spinning up dead clusters... >> >> I thought I'd finally worked out what everyone else knew - don't use the >> default AMI - but I've now run through all of the "official" quick-start >> linux releases and I'm none the wiser: >> >> Amazon Linux AMI 2014.03.1 - ami-7aba833f (64-bit) >> Provisions servers, connects, installs, but the webserver on the master >> will not start >> >> Red Hat Enterprise Linux 6.5 (HVM) - ami-5cdce419 >> Spot instance requests are not supported for this AMI. >> >> SuSE Linux Enterprise Server 11 sp3 (HVM) - ami-1a88bb5f >> Not tested - costs 10x more for spot instances, not economically viable. >> >> Ubuntu Server 14.04 LTS (HVM) - ami-f64f77b3 >> Provisions servers, but "git" is not pre-installed, so the cluster setup >> fails. >> >> Amazon Linux AMI (HVM) 2014.03.1 - ami-5aba831f >> Provisions servers, but "git" is not pre-installed, so the cluster setup >> fails. >> >> -- Jeremy Lee BCompSci(Hons) The Unorthodox Engineers
Re: Yay for 1.0.0! EC2 Still has problems.
If you are explicitly specifying the AMI in your invocation of spark-ec2, may I suggest simply removing any explicit mention of AMI from your invocation? spark-ec2 automatically selects an appropriate AMI based on the specified instance type. 2014년 6월 1일 일요일, Nicholas Chammas님이 작성한 메시지: > Could you post how exactly you are invoking spark-ec2? And are you having > trouble just with r3 instances, or with any instance type? > > 2014년 6월 1일 일요일, Jeremy Lee >님이 작성한 > 메시지: > > It's been another day of spinning up dead clusters... > > I thought I'd finally worked out what everyone else knew - don't use the > default AMI - but I've now run through all of the "official" quick-start > linux releases and I'm none the wiser: > > Amazon Linux AMI 2014.03.1 - ami-7aba833f (64-bit) > Provisions servers, connects, installs, but the webserver on the master > will not start > > Red Hat Enterprise Linux 6.5 (HVM) - ami-5cdce419 > Spot instance requests are not supported for this AMI. > > SuSE Linux Enterprise Server 11 sp3 (HVM) - ami-1a88bb5f > Not tested - costs 10x more for spot instances, not economically viable. > > Ubuntu Server 14.04 LTS (HVM) - ami-f64f77b3 > Provisions servers, but "git" is not pre-installed, so the cluster setup > fails. > > Amazon Linux AMI (HVM) 2014.03.1 - ami-5aba831f > Provisions servers, but "git" is not pre-installed, so the cluster setup > fails. > >
Re: Yay for 1.0.0! EC2 Still has problems.
Could you post how exactly you are invoking spark-ec2? And are you having trouble just with r3 instances, or with any instance type? 2014년 6월 1일 일요일, Jeremy Lee님이 작성한 메시지: > It's been another day of spinning up dead clusters... > > I thought I'd finally worked out what everyone else knew - don't use the > default AMI - but I've now run through all of the "official" quick-start > linux releases and I'm none the wiser: > > Amazon Linux AMI 2014.03.1 - ami-7aba833f (64-bit) > Provisions servers, connects, installs, but the webserver on the master > will not start > > Red Hat Enterprise Linux 6.5 (HVM) - ami-5cdce419 > Spot instance requests are not supported for this AMI. > > SuSE Linux Enterprise Server 11 sp3 (HVM) - ami-1a88bb5f > Not tested - costs 10x more for spot instances, not economically viable. > > Ubuntu Server 14.04 LTS (HVM) - ami-f64f77b3 > Provisions servers, but "git" is not pre-installed, so the cluster setup > fails. > > Amazon Linux AMI (HVM) 2014.03.1 - ami-5aba831f > Provisions servers, but "git" is not pre-installed, so the cluster setup > fails. > > Have I missed something? What AMI's are people using? I've just gone back > through the archives, and I'm seeing a lot of "I can't get EC2 to work" and > not a single "My EC2 has post-install issues", > > The quickstart page says "...can have a spark cluster up and running in > five minutes." But it's been three days for me so far. I'm about to bite > the bullet and start building my own AMI's from scratch... if anyone can > save me from that, I'd be most grateful. > > -- > Jeremy Lee BCompSci(Hons) > The Unorthodox Engineers >
Re: Yay for 1.0.0! EC2 Still has problems.
It's been another day of spinning up dead clusters... I thought I'd finally worked out what everyone else knew - don't use the default AMI - but I've now run through all of the "official" quick-start linux releases and I'm none the wiser: Amazon Linux AMI 2014.03.1 - ami-7aba833f (64-bit) Provisions servers, connects, installs, but the webserver on the master will not start Red Hat Enterprise Linux 6.5 (HVM) - ami-5cdce419 Spot instance requests are not supported for this AMI. SuSE Linux Enterprise Server 11 sp3 (HVM) - ami-1a88bb5f Not tested - costs 10x more for spot instances, not economically viable. Ubuntu Server 14.04 LTS (HVM) - ami-f64f77b3 Provisions servers, but "git" is not pre-installed, so the cluster setup fails. Amazon Linux AMI (HVM) 2014.03.1 - ami-5aba831f Provisions servers, but "git" is not pre-installed, so the cluster setup fails. Have I missed something? What AMI's are people using? I've just gone back through the archives, and I'm seeing a lot of "I can't get EC2 to work" and not a single "My EC2 has post-install issues", The quickstart page says "...can have a spark cluster up and running in five minutes." But it's been three days for me so far. I'm about to bite the bullet and start building my own AMI's from scratch... if anyone can save me from that, I'd be most grateful. -- Jeremy Lee BCompSci(Hons) The Unorthodox Engineers
Re: Yay for 1.0.0! EC2 Still has problems.
Oh, sorry, I forgot to add: here are the extra lines in my spark_ec2.py @205 "r3.large":"hvm", "r3.xlarge": "hvm", "r3.2xlarge": "hvm", "r3.4xlarge": "hvm", "r3.8xlarge": "hvm" Clearly a masterpiece of hacking. :-) I haven't tested all of them. The r3 set seems to act like i2. On Sun, Jun 1, 2014 at 12:45 AM, Jeremy Lee wrote: > Hi there, Patrick. Thanks for the reply... > > It wouldn't surprise me that AWS Ubuntu has Python 2.7. Ubuntu is cool > like that. :-) > > Alas, the Amazon Linux AMI (2014.03.1) does not, and it's the very first > one on the recommended instance list. (Ubuntu is #4, after Amazon, RedHat, > SUSE) So, users such as myself who deliberately pick the "Most Amazon-ish > obvious first choice" find they picked the wrong one. > > But that's trivial compared to the failure of the cluster to come up, > apparently due to the master's http configuration. Any help on that would > be much appreciated... it's giving me serious grief. > > > > On Sat, May 31, 2014 at 1:37 PM, Patrick Wendell > wrote: > >> Hi Jeremy, >> >> That's interesting, I don't think anyone has ever reported an issue >> running these scripts due to Python incompatibility, but they may require >> Python 2.7+. I regularly run them from the AWS Ubuntu 12.04 AMI... that >> might be a good place to start. But if there is a straightforward way to >> make them compatible with 2.6 we should do that. >> >> For r3.large, we can add that to the script. It's a newer type. Any >> interest in contributing this? >> >> - Patrick >> >> On May 30, 2014 5:08 AM, "Jeremy Lee" >> wrote: >> >>> >>> Hi there! I'm relatively new to the list, so sorry if this is a repeat: >>> >>> I just wanted to mention there are still problems with the EC2 scripts. >>> Basically, they don't work. >>> >>> First, if you run the scripts on Amazon's own suggested version of >>> linux, they break because amazon installs Python2.6.9, and the scripts use >>> a couple of Python2.7 commands. I have to "sudo yum install python27", and >>> then edit the spark-ec2 shell script to use that specific version. >>> Annoying, but minor. >>> >>> (the base "python" command isn't upgraded to 2.7 on many systems, >>> apparently because it would break yum) >>> >>> The second minor problem is that the script doesn't know about the >>> "r3.large" servers... also easily fixed by adding to the spark_ec2.py >>> script. Minor, >>> >>> The big problem is that after the EC2 cluster is provisioned, installed, >>> set up, and everything, it fails to start up the webserver on the master. >>> Here's the tail of the log: >>> >>> Starting GANGLIA gmond:[ OK ] >>> Shutting down GANGLIA gmond: [FAILED] >>> Starting GANGLIA gmond:[ OK ] >>> Connection to ec2-54-183-82-48.us-west-1.compute.amazonaws.com closed. >>> Shutting down GANGLIA gmond: [FAILED] >>> Starting GANGLIA gmond:[ OK ] >>> Connection to ec2-54-183-82-24.us-west-1.compute.amazonaws.com closed. >>> Shutting down GANGLIA gmetad: [FAILED] >>> Starting GANGLIA gmetad: [ OK ] >>> Stopping httpd:[FAILED] >>> Starting httpd: httpd: Syntax error on line 153 of >>> /etc/httpd/conf/httpd.conf: Cannot load modules/mod_authn_alias.so into >>> server: /etc/httpd/modules/mod_authn_alias.so: cannot open shared object >>> file: No such file or directory >>>[FAILED] >>> >>> Basically, the AMI you have chosen does not seem to have a "full" >>> install of apache, and is missing several modules that are referred to in >>> the httpd.conf file that is installed. The full list of missing modules is: >>> >>> authn_alias_module modules/mod_authn_alias.so >>> authn_default_module modules/mod_authn_default.so >>> authz_default_module modules/mod_authz_default.so >>> ldap_module modules/mod_ldap.so >>> authnz_ldap_module modules/mod_authnz_ldap.so >>> disk_cache_module modules/mod_disk_cache.so >>> >>> Alas, even if these modules are commented out, the server still fails to >>> start. >>> >>> root@ip-172-31-11-193 ~]$ service httpd start >>> Starting httpd: AH00534: httpd: Configuration error: No MPM loaded. >>> >>> That means Spark 1.0.0 clusters on EC2 are Dead-On-Arrival when run >>> according to the instructions. Sorry. >>> >>> Any suggestions on how to proceed? I'll keep trying to fix the >>> webserver, but (a) changes to httpd.conf get blown away by "resume", and >>> (b) anything I do has to be redone every time I provision another cluster. >>> Ugh. >>> >>> -- >>> Jeremy Lee BCompSci(Hons) >>> The Unorthodox Engineers >>> >> >> >> >> >> > > > -- > Jeremy Lee BCompSci(Hons) > The Unorthodox Engineers > -- Jeremy Lee BCompSci(Hons) The Unorthodox Engineers
Re: Yay for 1.0.0! EC2 Still has problems.
Hi there, Patrick. Thanks for the reply... It wouldn't surprise me that AWS Ubuntu has Python 2.7. Ubuntu is cool like that. :-) Alas, the Amazon Linux AMI (2014.03.1) does not, and it's the very first one on the recommended instance list. (Ubuntu is #4, after Amazon, RedHat, SUSE) So, users such as myself who deliberately pick the "Most Amazon-ish obvious first choice" find they picked the wrong one. But that's trivial compared to the failure of the cluster to come up, apparently due to the master's http configuration. Any help on that would be much appreciated... it's giving me serious grief. On Sat, May 31, 2014 at 1:37 PM, Patrick Wendell wrote: > Hi Jeremy, > > That's interesting, I don't think anyone has ever reported an issue > running these scripts due to Python incompatibility, but they may require > Python 2.7+. I regularly run them from the AWS Ubuntu 12.04 AMI... that > might be a good place to start. But if there is a straightforward way to > make them compatible with 2.6 we should do that. > > For r3.large, we can add that to the script. It's a newer type. Any > interest in contributing this? > > - Patrick > > On May 30, 2014 5:08 AM, "Jeremy Lee" > wrote: > >> >> Hi there! I'm relatively new to the list, so sorry if this is a repeat: >> >> I just wanted to mention there are still problems with the EC2 scripts. >> Basically, they don't work. >> >> First, if you run the scripts on Amazon's own suggested version of linux, >> they break because amazon installs Python2.6.9, and the scripts use a >> couple of Python2.7 commands. I have to "sudo yum install python27", and >> then edit the spark-ec2 shell script to use that specific version. >> Annoying, but minor. >> >> (the base "python" command isn't upgraded to 2.7 on many systems, >> apparently because it would break yum) >> >> The second minor problem is that the script doesn't know about the >> "r3.large" servers... also easily fixed by adding to the spark_ec2.py >> script. Minor, >> >> The big problem is that after the EC2 cluster is provisioned, installed, >> set up, and everything, it fails to start up the webserver on the master. >> Here's the tail of the log: >> >> Starting GANGLIA gmond:[ OK ] >> Shutting down GANGLIA gmond: [FAILED] >> Starting GANGLIA gmond:[ OK ] >> Connection to ec2-54-183-82-48.us-west-1.compute.amazonaws.com closed. >> Shutting down GANGLIA gmond: [FAILED] >> Starting GANGLIA gmond:[ OK ] >> Connection to ec2-54-183-82-24.us-west-1.compute.amazonaws.com closed. >> Shutting down GANGLIA gmetad: [FAILED] >> Starting GANGLIA gmetad: [ OK ] >> Stopping httpd:[FAILED] >> Starting httpd: httpd: Syntax error on line 153 of >> /etc/httpd/conf/httpd.conf: Cannot load modules/mod_authn_alias.so into >> server: /etc/httpd/modules/mod_authn_alias.so: cannot open shared object >> file: No such file or directory >>[FAILED] >> >> Basically, the AMI you have chosen does not seem to have a "full" install >> of apache, and is missing several modules that are referred to in the >> httpd.conf file that is installed. The full list of missing modules is: >> >> authn_alias_module modules/mod_authn_alias.so >> authn_default_module modules/mod_authn_default.so >> authz_default_module modules/mod_authz_default.so >> ldap_module modules/mod_ldap.so >> authnz_ldap_module modules/mod_authnz_ldap.so >> disk_cache_module modules/mod_disk_cache.so >> >> Alas, even if these modules are commented out, the server still fails to >> start. >> >> root@ip-172-31-11-193 ~]$ service httpd start >> Starting httpd: AH00534: httpd: Configuration error: No MPM loaded. >> >> That means Spark 1.0.0 clusters on EC2 are Dead-On-Arrival when run >> according to the instructions. Sorry. >> >> Any suggestions on how to proceed? I'll keep trying to fix the webserver, >> but (a) changes to httpd.conf get blown away by "resume", and (b) anything >> I do has to be redone every time I provision another cluster. Ugh. >> >> -- >> Jeremy Lee BCompSci(Hons) >> The Unorthodox Engineers >> > > > > > -- Jeremy Lee BCompSci(Hons) The Unorthodox Engineers
Re: Yay for 1.0.0! EC2 Still has problems.
We are migrating our scripts to r3. Thr lineage is in spark-ec2 would be happy to migrate those too. Having trouble with ganglia setup currently :) Regards Mayur On 31 May 2014 09:07, "Patrick Wendell" wrote: > Hi Jeremy, > > That's interesting, I don't think anyone has ever reported an issue > running these scripts due to Python incompatibility, but they may require > Python 2.7+. I regularly run them from the AWS Ubuntu 12.04 AMI... that > might be a good place to start. But if there is a straightforward way to > make them compatible with 2.6 we should do that. > > For r3.large, we can add that to the script. It's a newer type. Any > interest in contributing this? > > - Patrick > > On May 30, 2014 5:08 AM, "Jeremy Lee" > wrote: > >> >> Hi there! I'm relatively new to the list, so sorry if this is a repeat: >> >> I just wanted to mention there are still problems with the EC2 scripts. >> Basically, they don't work. >> >> First, if you run the scripts on Amazon's own suggested version of linux, >> they break because amazon installs Python2.6.9, and the scripts use a >> couple of Python2.7 commands. I have to "sudo yum install python27", and >> then edit the spark-ec2 shell script to use that specific version. >> Annoying, but minor. >> >> (the base "python" command isn't upgraded to 2.7 on many systems, >> apparently because it would break yum) >> >> The second minor problem is that the script doesn't know about the >> "r3.large" servers... also easily fixed by adding to the spark_ec2.py >> script. Minor, >> >> The big problem is that after the EC2 cluster is provisioned, installed, >> set up, and everything, it fails to start up the webserver on the master. >> Here's the tail of the log: >> >> Starting GANGLIA gmond:[ OK ] >> Shutting down GANGLIA gmond: [FAILED] >> Starting GANGLIA gmond:[ OK ] >> Connection to ec2-54-183-82-48.us-west-1.compute.amazonaws.com closed. >> Shutting down GANGLIA gmond: [FAILED] >> Starting GANGLIA gmond:[ OK ] >> Connection to ec2-54-183-82-24.us-west-1.compute.amazonaws.com closed. >> Shutting down GANGLIA gmetad: [FAILED] >> Starting GANGLIA gmetad: [ OK ] >> Stopping httpd:[FAILED] >> Starting httpd: httpd: Syntax error on line 153 of >> /etc/httpd/conf/httpd.conf: Cannot load modules/mod_authn_alias.so into >> server: /etc/httpd/modules/mod_authn_alias.so: cannot open shared object >> file: No such file or directory >>[FAILED] >> >> Basically, the AMI you have chosen does not seem to have a "full" install >> of apache, and is missing several modules that are referred to in the >> httpd.conf file that is installed. The full list of missing modules is: >> >> authn_alias_module modules/mod_authn_alias.so >> authn_default_module modules/mod_authn_default.so >> authz_default_module modules/mod_authz_default.so >> ldap_module modules/mod_ldap.so >> authnz_ldap_module modules/mod_authnz_ldap.so >> disk_cache_module modules/mod_disk_cache.so >> >> Alas, even if these modules are commented out, the server still fails to >> start. >> >> root@ip-172-31-11-193 ~]$ service httpd start >> Starting httpd: AH00534: httpd: Configuration error: No MPM loaded. >> >> That means Spark 1.0.0 clusters on EC2 are Dead-On-Arrival when run >> according to the instructions. Sorry. >> >> Any suggestions on how to proceed? I'll keep trying to fix the webserver, >> but (a) changes to httpd.conf get blown away by "resume", and (b) anything >> I do has to be redone every time I provision another cluster. Ugh. >> >> -- >> Jeremy Lee BCompSci(Hons) >> The Unorthodox Engineers >> > > > > >
Re: Yay for 1.0.0! EC2 Still has problems.
Hi Jeremy, That's interesting, I don't think anyone has ever reported an issue running these scripts due to Python incompatibility, but they may require Python 2.7+. I regularly run them from the AWS Ubuntu 12.04 AMI... that might be a good place to start. But if there is a straightforward way to make them compatible with 2.6 we should do that. For r3.large, we can add that to the script. It's a newer type. Any interest in contributing this? - Patrick On May 30, 2014 5:08 AM, "Jeremy Lee" wrote: > > Hi there! I'm relatively new to the list, so sorry if this is a repeat: > > I just wanted to mention there are still problems with the EC2 scripts. > Basically, they don't work. > > First, if you run the scripts on Amazon's own suggested version of linux, > they break because amazon installs Python2.6.9, and the scripts use a > couple of Python2.7 commands. I have to "sudo yum install python27", and > then edit the spark-ec2 shell script to use that specific version. > Annoying, but minor. > > (the base "python" command isn't upgraded to 2.7 on many systems, > apparently because it would break yum) > > The second minor problem is that the script doesn't know about the > "r3.large" servers... also easily fixed by adding to the spark_ec2.py > script. Minor, > > The big problem is that after the EC2 cluster is provisioned, installed, > set up, and everything, it fails to start up the webserver on the master. > Here's the tail of the log: > > Starting GANGLIA gmond:[ OK ] > Shutting down GANGLIA gmond: [FAILED] > Starting GANGLIA gmond:[ OK ] > Connection to ec2-54-183-82-48.us-west-1.compute.amazonaws.com closed. > Shutting down GANGLIA gmond: [FAILED] > Starting GANGLIA gmond:[ OK ] > Connection to ec2-54-183-82-24.us-west-1.compute.amazonaws.com closed. > Shutting down GANGLIA gmetad: [FAILED] > Starting GANGLIA gmetad: [ OK ] > Stopping httpd:[FAILED] > Starting httpd: httpd: Syntax error on line 153 of > /etc/httpd/conf/httpd.conf: Cannot load modules/mod_authn_alias.so into > server: /etc/httpd/modules/mod_authn_alias.so: cannot open shared object > file: No such file or directory >[FAILED] > > Basically, the AMI you have chosen does not seem to have a "full" install > of apache, and is missing several modules that are referred to in the > httpd.conf file that is installed. The full list of missing modules is: > > authn_alias_module modules/mod_authn_alias.so > authn_default_module modules/mod_authn_default.so > authz_default_module modules/mod_authz_default.so > ldap_module modules/mod_ldap.so > authnz_ldap_module modules/mod_authnz_ldap.so > disk_cache_module modules/mod_disk_cache.so > > Alas, even if these modules are commented out, the server still fails to > start. > > root@ip-172-31-11-193 ~]$ service httpd start > Starting httpd: AH00534: httpd: Configuration error: No MPM loaded. > > That means Spark 1.0.0 clusters on EC2 are Dead-On-Arrival when run > according to the instructions. Sorry. > > Any suggestions on how to proceed? I'll keep trying to fix the webserver, > but (a) changes to httpd.conf get blown away by "resume", and (b) anything > I do has to be redone every time I provision another cluster. Ugh. > > -- > Jeremy Lee BCompSci(Hons) > The Unorthodox Engineers >