Re: ZuriHac 2019 - GHC Track

2019-06-10 Thread Simon Hafner
I have a ~ 30 minute talk which covers my GHC proposal (NoToplevelFieldSelectors), as well as parts of the renamer. I could hold that at any point, if there's still time slots over. Am Di., 28. Mai 2019 um 23:37 Uhr schrieb Ben Gamari : > Andreas Herrmann writes: > > > Dear GHC devs, > > > I've

Fwd: Saving large textfile

2016-04-24 Thread Simon Hafner
2016-04-24 13:38 GMT+02:00 Stefan Falk : > sc.parallelize(cfile.toString() > .split("\n"), 1) Try `sc.textFile(pathToFile)` instead. >java.io.IOException: Broken pipe >at sun.nio.ch.FileDispatcherImpl.write0(Native Method) >at

Re: StreamCorruptedException during deserialization

2016-03-29 Thread Simon Hafner
2016-03-29 11:25 GMT+02:00 Robert Schmidtke : > Is there a meaningful way for me to find out what exactly is going wrong > here? Any help and hints are greatly appreciated! Maybe a version mismatch between the jars on the cluster?

Re: Output is being stored on the clusters (slaves).

2016-03-24 Thread Simon Hafner
2016-03-24 11:09 GMT+01:00 Shishir Anshuman : > I am using two Slaves to run the ALS algorithm. I am saving the predictions > in a textfile using : > saveAsTextFile(path) > > The predictions is getting stored on the slaves but I want the predictions > to be saved

Re: No active SparkContext

2016-03-24 Thread Simon Hafner
2016-03-24 9:54 GMT+01:00 Max Schmidt : > we're using with the java-api (1.6.0) a ScheduledExecutor that continuously > executes a SparkJob to a standalone cluster. I'd recommend Scala. > After each job we close the JavaSparkContext and create a new one. Why do that? You can

Re: Installing Spark on Mac

2016-03-04 Thread Simon Hafner
I'd try `brew install spark` or `apache-spark` and see where that gets you. https://github.com/Homebrew/homebrew 2016-03-04 21:18 GMT+01:00 Aida : > Hi all, > > I am a complete novice and was wondering whether anyone would be willing to > provide me with a step by step

Re: Running synchronized JRI code

2016-02-15 Thread Simon Hafner
2016-02-15 14:02 GMT+01:00 Sun, Rui : > On computation, RRDD launches one R process for each partition, so there > won't be thread-safe issue > > Could you give more details on your new environment? Running on EC2, I start the executors via /usr/bin/R CMD javareconf -e

Re: Running synchronized JRI code

2016-02-15 Thread Simon Hafner
2016-02-15 4:35 GMT+01:00 Sun, Rui : > Yes, JRI loads an R dynamic library into the executor JVM, which faces > thread-safe issue when there are multiple task threads within the executor. > > I am thinking if the demand like yours (calling R code in RDD > transformations) is

Running synchronized JRI code

2016-02-14 Thread Simon Hafner
Hello I'm currently running R code in an executor via JRI. Because R is single-threaded, any call to R needs to be wrapped in a `synchronized`. Now I can use a bit more than one core per executor, which is undesirable. Is there a way to tell spark that this specific application (or even specific

Re: Serializing DataSets

2016-01-19 Thread Simon Hafner
The occasional type error if the casting goes wrong for whatever reason. 2016-01-19 1:22 GMT+08:00 Michael Armbrust <mich...@databricks.com>: > What error? > > On Mon, Jan 18, 2016 at 9:01 AM, Simon Hafner <reactorm...@gmail.com> wrote: >> >> And for deseriali

Re: Serializing DataSets

2016-01-18 Thread Simon Hafner
toDF()). We'll likely be combining the classes > in Spark 2.0 to remove this awkwardness. > > On Tue, Jan 12, 2016 at 11:20 PM, Simon Hafner <reactorm...@gmail.com> > wrote: >> >> What's the proper way to write DataSets to disk

[jira] [Commented] (SPARK-12677) Lazy file discovery for parquet

2016-01-13 Thread Simon Hafner (JIRA)
[ https://issues.apache.org/jira/browse/SPARK-12677?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15096594#comment-15096594 ] Simon Hafner commented on SPARK-12677: -- What would be the gain? The application would crash

Serializing DataSets

2016-01-12 Thread Simon Hafner
What's the proper way to write DataSets to disk? Convert them to a DataFrame and use the writers there? - To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org

Re: Compiling spark 1.5.1 fails with scala.reflect.internal.Types$TypeError: bad symbolic reference.

2015-12-16 Thread Simon Hafner
, how did you resolved the problem. > > On Fri, Oct 16, 2015 at 9:54 AM, Simon Hafner <reactorm...@gmail.com> wrote: >> >> Fresh clone of spark 1.5.1, java version "1.7.0_85" >> >> build/mvn -Pyarn -Phadoop-2.4 -Dhadoop.version=2.4.0 -DskipTest

[jira] [Commented] (SPARK-11539) Debian packaging

2015-11-05 Thread Simon Hafner (JIRA)
[ https://issues.apache.org/jira/browse/SPARK-11539?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14992587#comment-14992587 ] Simon Hafner commented on SPARK-11539: -- sbt-native-packager makes it slightly easier. > Deb

[jira] [Created] (SPARK-11539) Debian packaging

2015-11-05 Thread Simon Hafner (JIRA)
Simon Hafner created SPARK-11539: Summary: Debian packaging Key: SPARK-11539 URL: https://issues.apache.org/jira/browse/SPARK-11539 Project: Spark Issue Type: New Feature

[jira] [Issue Comment Deleted] (SPARK-11539) Debian packaging

2015-11-05 Thread Simon Hafner (JIRA)
[ https://issues.apache.org/jira/browse/SPARK-11539?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Simon Hafner updated SPARK-11539: - Comment: was deleted (was: sbt-native-packager makes it slightly easier.) > Debian packag

Fwd: Where does mllib's .save method save a model to?

2015-11-03 Thread Simon Hafner
2015-11-03 20:26 GMT+01:00 xenocyon : > I want to save an mllib model to disk, and am trying the model.save > operation as described in > http://spark.apache.org/docs/latest/mllib-collaborative-filtering.html#examples: > > model.save(sc, "myModelPath") > > But after running

Fwd: collect() local faster than 4 node cluster

2015-11-03 Thread Simon Hafner
2015-11-03 20:07 GMT+01:00 Sebastian Kuepers : > Hey, > > with collect() RDDs elements are send as a list back to the driver. > > If have a 4 node cluster (based on Mesos) in a datacenter and I have my > local dev machine. > > I work with a small 200MB

Re: Support Ordering on UserDefinedType

2015-11-03 Thread Simon Hafner
2015-11-03 23:20 GMT+01:00 Ionized : > TypeUtils.getInterpretedOrdering currently only supports AtomicType and > StructType. Is it possible to add support for UserDefinedType as well? Yes, make a PR to spark.

[jira] [Created] (SPARK-11268) Non-daemon startup scripts

2015-10-22 Thread Simon Hafner (JIRA)
Simon Hafner created SPARK-11268: Summary: Non-daemon startup scripts Key: SPARK-11268 URL: https://issues.apache.org/jira/browse/SPARK-11268 Project: Spark Issue Type: Improvement

Compiling spark 1.5.1 fails with scala.reflect.internal.Types$TypeError: bad symbolic reference.

2015-10-16 Thread Simon Hafner
Fresh clone of spark 1.5.1, java version "1.7.0_85" build/mvn -Pyarn -Phadoop-2.4 -Dhadoop.version=2.4.0 -DskipTests clean package [error] bad symbolic reference. A signature in WebUI.class refers to term eclipse [error] in package org which is not available. [error] It may be completely missing

udaf with multiple return values in spark 1.5.0

2015-09-06 Thread Simon Hafner
Hi everyone is it possible to return multiple values with an udaf defined in spark 1.5.0? The documentation [1] mentions abstract def dataType: DataType The DataType of the returned value of this UserDefinedAggregateFunction. so it's only possible to return a single value. Should I use

[jira] [Created] (SPARK-10053) SparkR isn't exporting lapply

2015-08-17 Thread Simon Hafner (JIRA)
Simon Hafner created SPARK-10053: Summary: SparkR isn't exporting lapply Key: SPARK-10053 URL: https://issues.apache.org/jira/browse/SPARK-10053 Project: Spark Issue Type: Bug

[jira] [Created] (SPARK-8821) The ec2 script doesn't run on python 3 with an utf8 env

2015-07-03 Thread Simon Hafner (JIRA)
Simon Hafner created SPARK-8821: --- Summary: The ec2 script doesn't run on python 3 with an utf8 env Key: SPARK-8821 URL: https://issues.apache.org/jira/browse/SPARK-8821 Project: Spark Issue

Re: DUCC doesn't use all available machines

2014-11-30 Thread Simon Hafner
2014-11-30 7:25 GMT-06:00 Eddie Epstein eaepst...@gmail.com: On Sat, Nov 29, 2014 at 4:46 PM, Simon Hafner reactorm...@gmail.com wrote: I've thrown some numbers at it (doubling each) and it's running at comfortable 125 procs. However, at about 6.1k of 6.5k items, the procs drop down to 30

Re: DUCC org.apache.uima.util.InvalidXMLException and no logs

2014-11-28 Thread Simon Hafner
2014-11-27 11:44 GMT-06:00 Eddie Epstein eaepst...@gmail.com: Those are the only two log files? Should be a ducc.log (probably with no more info than on the console), and either one or both of the job driver logfiles: jd.out.log and jobid-JD-jdnode-jdpid.log. If for some reason the job driver

Ducc: Rename failed

2014-11-28 Thread Simon Hafner
When running DUCC in cluster mode, I get Rename failed. The file mentioned in the error message exists in the txt.processed/ directory. The mount is via nfs (rw,sync,insecure). org.apache.uima.resource.ResourceProcessException: Received Exception In Message From Service on Queue:ducc.jd.queue.75

Re: Ducc: Rename failed

2014-11-28 Thread Simon Hafner
2014-11-28 14:18 GMT-06:00 Eddie Epstein eaepst...@gmail.com: To debug, please add the following option to the job submission: --all_in_one local This will run all the code in a single process on the machine doing the submit. Hopefully the log file and/or console will be more informative.

DUCC org.apache.uima.util.InvalidXMLException and no logs

2014-11-26 Thread Simon Hafner
When launching the Raw Text example application, it doesn't load with the following error: [ducc@ip-10-0-0-164 analysis]$ MyAppDir=$PWD MyInputDir=$PWD/txt MyOutputDir=$PWD/txt.processed ~/ducc_install/bin/ducc_submit -f DuccRawTextSpec.job Job 50 submitted id:50 location:5991@ip-10-0-0-164 id:50

wholeTextFiles on 20 nodes

2014-11-23 Thread Simon Hafner
I have 20 nodes via EC2 and an application that reads the data via wholeTextFiles. I've tried to copy the data into hadoop via copyFromLocal, and I get 14/11/24 02:00:07 INFO hdfs.DFSClient: Exception in createBlockOutputStream 172.31.2.209:50010 java.io.IOException: Bad connect ack with

Re: DUCC 1.1.0- How to Run two DUCC version on same machines with different user

2014-11-17 Thread Simon Hafner
2014-11-17 0:00 GMT-06:00 reshu.agarwal reshu.agar...@orkash.com: I want to run two DUCC version i.e. 1.0.0 and 1.1.0 on same machines with different user. Can this be possible? Yes, that should be possible. You'll have to make sure there are no port conflicts, I'd guess the ActiveMQ port is

Re: DUCC stuck at WaitingForResources on an Amazon Linux

2014-11-13 Thread Simon Hafner
look there for reasons the resources are not being allocated? Eddie On Wed, Nov 12, 2014 at 4:07 PM, Simon Hafner reactorm...@gmail.com wrote: 4 shares total, 2 in use. 2014-11-12 5:06 GMT-06:00 Lou DeGenaro lou.degen...@gmail.com: Try looking at your DUCC's web server. On the System

DUCC stuck at WaitingForResources on an Amazon Linux

2014-11-12 Thread Simon Hafner
I've set up DUCC according to https://cwiki.apache.org/confluence/display/UIMA/DUCC ducc_install/bin/ducc_submit -f ducc_install/examples/simple/1.job the job is stuck at WaitingForResources. 12 Nov 2014 10:37:30,175 INFO Agent.LinuxNodeMetricsProcessor - process N/A ... Agent

Re: DUCC stuck at WaitingForResources on an Amazon Linux

2014-11-12 Thread Simon Hafner
4 shares total, 2 in use. 2014-11-12 5:06 GMT-06:00 Lou DeGenaro lou.degen...@gmail.com: Try looking at your DUCC's web server. On the System - Machines page do you see any shares not inuse? Lou. On Wed, Nov 12, 2014 at 5:51 AM, Simon Hafner reactorm...@gmail.com wrote: I've set up DUCC

log4j logging control via sbt

2014-11-05 Thread Simon Hafner
I've tried to set the log4j logger to warn only via log4j properties file in cat src/test/resources/log4j.properties log4j.logger.org.apache.spark=WARN or in sbt via javaOptions += -Dlog4j.logger.org.apache.spark=WARN But the logger still gives me INFO messages to stdout when I run my tests

Spark with HLists

2014-10-29 Thread Simon Hafner
I tried using shapeless HLists as data storage for data inside spark. Unsurprisingly, it failed. The deserialization isn't well-defined because of all the implicits used by shapeless. How could I make it work? Sample Code: /* SimpleApp.scala */ import org.apache.spark.SparkContext import

[kde] Add certificates to Kopete

2014-04-15 Thread Simon Hafner
Hello how do I add root certificates to kopete? Add added ca-cert to both kleopatra and the system ca-certificates, but kopete still complains about an invalid SSL certificate from the server due to invalid root certificate. Cheers, Simon ___

[opensc-devel] pam_p11 (without pin) and ssh (with pin) on one card

2012-09-25 Thread Simon Hafner
Hey y'all I have an ePass2003, and I'd like to use it for pam_p11 and ssh. The pam_p11 key should be usable without a pin, or can I provide the pin by using the password field? I'd like to know which paths are possible. The other object stored is an ssh key secured by a pin. My problem is now

More specific logging than xev

2012-08-15 Thread Simon Hafner
not switch desktop. `wmctrl -s 1` gives the same message ClientMessage event, serial 26, synthetic YES, window 0xab, message_type 0x13f (_NET_CURRENT_DESKTOP), format 32 but it switches desktops. How can I debug that further? Greetings Simon Hafner ___ wm

Re: Contextual modes for operators.

2012-08-15 Thread Simon Hafner
2012/8/15 Andreas Liljeqvist bon...@gmail.com: I am not really familiar with elisp though, what are the equivalence of multimethods as found in clojure for example? According to #emacs, there is eieio, if I you associate multimethods with dynamic dispatch. cW stopping in the case of

Re: [Voyage-linux] Can't boot voyage linux 0.8.5

2012-08-14 Thread Simon Hafner
liquid liquidee@... writes: I am trying to install voyage linux 0.8.5 on an Alix 3d2 board via pxe boot. I can see that it loads the image properly, but the boot process hangs on Switching to clocksource tsc. [...] I've got the same problem with voyage 0.8.5, I'd appreciate help.

[Moses-support] extract translation table

2012-06-03 Thread Simon Hafner
Hi all is there a nice way to get the top 100 translations? I'm trying to compare two languages on character ngram level, to find common edit paths. The idea is to train moses for that pair and then extract the most common ngram pairs. Is this even possible or are they normalized based on their

Re: lucene-core-3.3.0 not optimizing

2011-12-02 Thread Simon Hafner
On 02.12.2011, at 04:54, KARTHIK SHIVAKUMAR wrote: Hi Spec O/s win os 7 Jdk : 1.6.0_29 Lucene lucene-core-3.3.0 Finally after Indexing successfully ,Why this Code does not optimize ( sample code ) INDEX_WRITER.optimize(100); INDEX_WRITER.commit();

Re: [rspec-users] RSpec newbie ran rspec 2.4 on WinXP, got HOME undefined

2011-01-10 Thread Simon Hafner
On Monday 10 January 2011 22.37:46 RichardOnRails wrote: Hi, I'm running WinXP-Pro/SP3 Ruby 1.8.6 K:/_Utilities/ruby186-26_rc2/ruby/lib/ruby/gems/1.8/gems/rspec- core-2.4.0/lib/rspec/core/configuration_options.rb:9:couldn't find HOME environment -- expanding `~' (ArgumentError) I