or a seperate RDD for sparql operations ala SchemaRDD .. operators for sparql can be defined thr.. not a bad idea :)
Mayur Rustagi Ph: +1 (760) 203 3257 http://www.sigmoidanalytics.com @mayur_rustagi <https://twitter.com/mayur_rustagi> On Fri, Jun 20, 2014 at 3:56 PM, andy petrella <andy.petre...@gmail.com> wrote: > Maybe some SPARQL features in Shark, then ? > > aℕdy ℙetrella > about.me/noootsab > [image: aℕdy ℙetrella on about.me] > > <http://about.me/noootsab> > > > On Fri, Jun 20, 2014 at 9:45 PM, Mayur Rustagi <mayur.rust...@gmail.com> > wrote: > >> You are looking to create Shark operators for RDF? Since Shark backend is >> shifting to SparkSQL it would be slightly hard but much better effort would >> be to shift Gremlin to Spark (though a much beefier one :) ) >> >> Mayur Rustagi >> Ph: +1 (760) 203 3257 >> http://www.sigmoidanalytics.com >> @mayur_rustagi <https://twitter.com/mayur_rustagi> >> >> >> >> On Fri, Jun 20, 2014 at 3:39 PM, andy petrella <andy.petre...@gmail.com> >> wrote: >> >>> For RDF, may GraphX be particularly approriated? >>> >>> aℕdy ℙetrella >>> about.me/noootsab >>> [image: aℕdy ℙetrella on about.me] >>> >>> <http://about.me/noootsab> >>> >>> >>> On Thu, Jun 19, 2014 at 4:49 PM, Flavio Pompermaier < >>> pomperma...@okkam.it> wrote: >>> >>>> Hi guys, >>>> >>>> I'm analyzing the possibility to use Spark to analyze RDF files and >>>> define reusable Shark operators on them (custom filtering, transforming, >>>> aggregating, etc). Is that possible? Any hint? >>>> >>>> Best, >>>> Flavio >>>> >>> >>> >> >