RE: Support/Consulting companies

2016-08-19 Thread Huang, Roger
http://thelastpickle.com/


From: Roxy Ubi [mailto:roxy...@gmail.com]
Sent: Friday, August 19, 2016 2:02 PM
To: user@cassandra.apache.org
Subject: Support/Consulting companies

Howdy,
I'm looking for a list of support or consulting companies that provide 
contracting services related to Cassandra.  Is there a comprehensive list 
somewhere?  Alternatively could you folks tell me who you use?
Thanks in advance for any replies!
Roxy


RE: Validation of Data after data migration from RDBMS to Cassandra

2015-08-06 Thread Huang, Roger
Surbhi,
A general note, Cassandra tables are not normalized as in an RDBMS schema, so 
it’s not straight migration.
Best practice is to design your queries and then design your Cassandra schema 
to support those queries.
Patrick McFadin has a series on C* data modeling.
http://docs.datastax.com/en/cassandra/2.1
Roger



From: Surbhi Gupta [mailto:surbhi.gupt...@gmail.com]
Sent: Thursday, August 06, 2015 11:31 AM
To: user@cassandra.apache.org
Subject: Validation of Data after data migration from RDBMS to Cassandra

Hi,

We have to migrate the data from Oracle/mysql to Cassandra.
I wanted to understand, if we have any tool/utilitiy which can help in 
validation the data after the data migration to Cassandra.

Thanks
Surbhi


RE: Validation of Data after data migration from RDBMS to Cassandra

2015-08-06 Thread Huang, Roger
Surbhi
Sorry I meant to send the CQL link 
http://docs.datastax.com/en/cql/3.1/cql/cql_intro_c.html

Roger

From: Huang, Roger [mailto:rohu...@visa.com]
Sent: Thursday, August 06, 2015 11:39 AM
To: user@cassandra.apache.org
Subject: RE: Validation of Data after data migration from RDBMS to Cassandra

Surbhi,
A general note, Cassandra tables are not normalized as in an RDBMS schema, so 
it’s not straight migration.
Best practice is to design your queries and then design your Cassandra schema 
to support those queries.
Patrick McFadin has a series on C* data modeling.
http://docs.datastax.com/en/cassandra/2.1
Roger



From: Surbhi Gupta [mailto:surbhi.gupt...@gmail.com]
Sent: Thursday, August 06, 2015 11:31 AM
To: user@cassandra.apache.orgmailto:user@cassandra.apache.org
Subject: Validation of Data after data migration from RDBMS to Cassandra

Hi,

We have to migrate the data from Oracle/mysql to Cassandra.
I wanted to understand, if we have any tool/utilitiy which can help in 
validation the data after the data migration to Cassandra.

Thanks
Surbhi


RE: How to share large resources like dictionaries while processing data with Spark ?

2015-06-04 Thread Huang, Roger
Is the dictionary read-only?
Did you look at 
http://spark.apache.org/docs/latest/programming-guide.html#broadcast-variables ?


-Original Message-
From: dgoldenberg [mailto:dgoldenberg...@gmail.com] 
Sent: Thursday, June 04, 2015 4:50 PM
To: user@spark.apache.org
Subject: How to share large resources like dictionaries while processing data 
with Spark ?

We have some pipelines defined where sometimes we need to load potentially 
large resources such as dictionaries.

What would be the best strategy for sharing such resources among the 
transformations/actions within a consumer?  Can they be shared somehow across 
the RDD's?

I'm looking for a way to load such a resource once into the cluster memory and 
have it be available throughout the lifecycle of a consumer...

Thanks.



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/How-to-share-large-resources-like-dictionaries-while-processing-data-with-Spark-tp23162.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional 
commands, e-mail: user-h...@spark.apache.org


-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



RE: STORM UI -- trident topology continuously emitting ..

2015-05-21 Thread Huang, Roger
Have you looked at
http://www.sematext.com/spm/index.html


From: Hadi Sotudeh [mailto:hadisotudeh1...@gmail.com]
Sent: Thursday, May 21, 2015 9:53 AM
To: user
Cc: Nikhil Singh
Subject: Re: STORM UI -- trident topology continuously emitting ..

Is there any other monitoring tools for storm?


RE: Spark on Cassandra

2015-04-29 Thread Huang, Roger
http://planetcassandra.org/getting-started-with-apache-spark-and-cassandra/
http://planetcassandra.org/blog/holy-momentum-batman-spark-and-cassandra-circa-2015-w-datastax-connector-and-java/
https://github.com/datastax/spark-cassandra-connector



From: Cody Koeninger [mailto:c...@koeninger.org]
Sent: Wednesday, April 29, 2015 12:15 PM
To: Matthew Johnson
Cc: user@spark.apache.org
Subject: Re: Spark on Cassandra

Hadoop version doesn't matter if you're just using cassandra.

On Wed, Apr 29, 2015 at 12:08 PM, Matthew Johnson 
matt.john...@algomi.commailto:matt.john...@algomi.com wrote:
Hi all,

I am new to Spark, but excited to use it with our Cassandra cluster. I have 
read in a few places that Spark can interact directly with Cassandra now, so I 
decided to download it and have a play – I am happy to run it in standalone 
cluster mode initially. When I go to download it 
(http://spark.apache.org/downloads.html) I see a bunch of pre-built versions 
for Hadoop and MapR, but no mention of Cassandra – if I am running it in 
standalone cluster mode, does it matter which pre-built package I download? 
Would all of them work? Or do I have to build it myself from source with some 
special config for Cassandra?

Thanks!
Matt



RE: Spark + Mahout

2014-09-18 Thread Huang, Roger
Taka,
Have you considered using Spark’s own MLlib k-means clustering?
http://spark.apache.org/docs/latest/mllib-clustering.html

Roger

From: Daniel Takabayashi [mailto:takabaya...@scanboo.com.br]
Sent: Thursday, September 18, 2014 1:50 PM
To: user@spark.apache.org
Subject: Spark + Mahout

Hi guys,

Is possible to run a mahout kmeans throws spark infrastructure?


Thanks,
taka (Brazil)


FW: Indexing a term into separate Lucene indexes

2014-06-20 Thread Huang, Roger

If I have documents with a person and his email address: 
u...@domain.commailto:u...@domain.com

How can I configure Solr (4.6) so that the email address source field is 
indexed as

-  the user part of the address (e.g., user) is in Lucene index X

-  the domain part of the address (e.g., domain.com) is in a separate 
Lucene index Y

I would like to be able search as follows:

-  Find all people whose email addresses have user part = userXyz

-  Find all people whose email addresses have domain part = 
domainABC.com

-  Find the person with exact email address = 
user...@domainabc.commailto:user...@domainabc.com

Would I use a copyField declaration in my schema?
http://wiki.apache.org/solr/SchemaXml#Copy_Fields

Thanks!


RE: Indexing a term into separate Lucene indexes

2014-06-20 Thread Huang, Roger
Shawn,
Thanks for your response.
Due to security requirements, I do need the name and domain parts of the email 
address stored in separate Lucene indexes.
How do you recommend doing this?  What are the challenges?
Once the name and domain parts of the email address are in different Lucene 
indexes, would I need to modify my  Solr search string?
Thanks,
Roger


-Original Message-
From: Shawn Heisey [mailto:s...@elyograg.org] 
Sent: Friday, June 20, 2014 10:19 AM
To: solr-user@lucene.apache.org
Subject: Re: Indexing a term into separate Lucene indexes

On 6/19/2014 4:51 PM, Huang, Roger wrote:
 If I have documents with a person and his email address: 
 u...@domain.commailto:u...@domain.com

 How can I configure Solr (4.6) so that the email address source field 
 is indexed as

 -  the user part of the address (e.g., user) is in Lucene index X

 -  the domain part of the address (e.g., domain.com) is in a 
 separate Lucene index Y

 I would like to be able search as follows:

 -  Find all people whose email addresses have user part = userXyz

 -  Find all people whose email addresses have domain part = 
 domainABC.com

 -  Find the person with exact email address = user...@domainabc.com

 Would I use a copyField declaration in my schema?
 http://wiki.apache.org/solr/SchemaXml#Copy_Fields

I don't think you actually want the data to end up in entirely different 
indexes.  Although it is possible to search more than one separate index, 
that's very likely NOT what you want to do, and it comes with its own 
challenges.  What you most likely want is to put this data into different 
fields within the same index.

You'll need to write custom code to accomplish this, especially if you need the 
stored data to contain only the parts rather than the complete email address.  
A copyField can get the data to additional fields, but I'm not aware of 
anything built-in to the schema that can trim the unwanted information from the 
new fields, and even if there is, any stored data will be the original data for 
all three fields.  It's up to you whether this custom code is in a user 
application that does your indexing or in a custom update processor that you 
load as a plugin to Solr itself.  Extending whatever user application you are 
already using for indexing is very likely to be a lot easier.

Thanks,
Shawn



Indexing a term into separate Lucene indexes

2014-06-19 Thread Huang, Roger
If I have documents with a person and his email address: 
u...@domain.commailto:u...@domain.com

How can I configure Solr (4.6) so that the email address source field is 
indexed as

-  the user part of the address (e.g., user) is in Lucene index X

-  the domain part of the address (e.g., domain.com) is in a separate 
Lucene index Y

I would like to be able search as follows:

-  Find all people whose email addresses have user part = userXyz

-  Find all people whose email addresses have domain part = 
domainABC.com

-  Find the person with exact email address = user...@domainabc.com

Would I use a copyField declaration in my schema?
http://wiki.apache.org/solr/SchemaXml#Copy_Fields

Thanks!


RE: Doubts on Apache Storm

2014-05-06 Thread Huang, Roger
Milind,
Have you looked at http://storm.incubator.apache.org/documentation/Home.html 
http://storm.incubator.apache.org/documentation/Creating-a-new-Storm-project.html
 
Roger

-Original Message-
From: milind.pa...@polarisft.com [mailto:milind.pa...@polarisft.com] 
Sent: Tuesday, May 06, 2014 8:55 AM
To: user@storm.incubator.apache.org
Subject: Fw: Doubts on Apache Storm


Hi All,

Thanks for clarifying the doubts.

Can I request if there is any doc that demonstrate the deployment steps, pseudo 
code, FAQ, etc. (Basically basic building blocks)

Regards
Milind Patil | Intellect Liquidity cash management 8SWS 031 | Silver metropolis 
| Western express highway | Goregaon (East) | Mumbai | 400 063. INDIA
Board: 91-22-67801500  | 91-22-42029200 | Ext: 1734 Mobile: +91 9920612360
| mail id: milind.pa...@polarisft.com
- Forwarded by Milind Patil/BSC37/polarisFT on 05/06/2014 07:24 PM
-

From:   Marc Vaillant vaill...@animetrics.com
To: user@storm.incubator.apache.org
Date:   05/06/2014 07:12 PM
Subject:Re: Doubts on Apache Storm



On Tue, May 06, 2014 at 03:21:13PM +0530, milind.pa...@polarisft.com wrote:

 Hi,

Is Nimbus mandatory for storm? (Our development env is neither using
Nimbus nor any other cloud environment)

I think you might have misunderstood nimbus.  It is a daemon that is part of 
storm, *not* nimbus from the nimbus project.


 (I am new to Apache storm, it would really help me if any basic 
 document available on Apache storm)

 Regards
 Milind Patil | Intellect Liquidity cash management 8SWS 031 | Silver 
 metropolis | Western express highway | Goregaon (East)
|
 Mumbai | 400 063. INDIA
 Board: 91-22-67801500  | 91-22-42029200 | Ext: 1734 Mobile: +91
9920612360
 | mail id: milind.pa...@polarisft.com



 This e-Mail may contain proprietary and confidential information and 
 is
sent for the intended recipient(s) only.  If by an addressing or transmission 
error this mail has been misdirected to you, you are requested to delete this 
mail immediately. You are also hereby notified that any use, any form of 
reproduction, dissemination, copying, disclosure, modification, distribution 
and/or publication of this e-mail message, contents or its attachment other 
than by its intended recipient/s is strictly prohibited.

 Visit us at http://www.polarisFT.com




This e-Mail may contain proprietary and confidential information and is sent 
for the intended recipient(s) only.  If by an addressing or transmission error 
this mail has been misdirected to you, you are requested to delete this mail 
immediately. You are also hereby notified that any use, any form of 
reproduction, dissemination, copying, disclosure, modification, distribution 
and/or publication of this e-mail message, contents or its attachment other 
than by its intended recipient/s is strictly prohibited.

Visit us at http://www.polarisFT.com



RE: Is there a good reference book for storm users?

2014-05-05 Thread Huang, Roger
Quinton Anderson’s book is good, and includes ML examples.
http://www.packtpub.com/storm-realtime-processing-cookbook/book
-Roger

From: Jarek Śmiejczak [mailto:j...@smiejczak.net]
Sent: Sunday, May 04, 2014 6:01 AM
To: user@storm.incubator.apache.org
Subject: Re: Is there a good reference book for storm users?

Two weeks ago i've bought: http://www.manning.com/sallen/ and i'm really 
surprised about contents. Everything is explained in really non-boring and 
professional manner. I think it's worth to buy this book even before final 
release.

On Sun, May 4, 2014 at 9:23 AM, Sajith 
sajith...@gmail.commailto:sajith...@gmail.com wrote:
Thank you very much Milinda and Eugene

On Sun, May 4, 2014 at 8:23 AM, Eugene 
edvor...@gmail.commailto:edvor...@gmail.com wrote:
I second that. Taylor and Brian book good for advanced users. A lot of 
information about Trident. Very interesting use cases and written very well 
with step-by-step instructions.


On Sat, May 3, 2014 at 6:41 PM, Milinda Pathirage 
mpath...@umail.iu.edumailto:mpath...@umail.iu.edu wrote:
I think Taylor's book [1] is good. I bought that, but still couldn't
read it. But according to table of content it looks promising.

Thanks
Milinda


[1] 
http://www.amazon.com/Storm-Blueprints-Distributed-Real-time-Computation/dp/178216829X/ref=sr_1_5?ie=UTF8qid=1399156748sr=8-5keywords=storm

On Sat, May 3, 2014 at 3:36 PM, Sajith 
sajith...@gmail.commailto:sajith...@gmail.com wrote:
 Hi all,

 Can someone suggest me an up-to date comprehensive reference book on storm
 for advanced users?

 Thanks,
 Sajith.


--
Milinda Pathirage

PhD Student | Research Assistant
School of Informatics and Computing | Data to Insight Center
Indiana University

twitter: milindalakmal
skype: milinda.pathirage
blog: http://milinda.pathirage.org



--
Eugene Dvorkin
Software Engineer
New York City Storm User Group - organizer
WebMD
email: edvor...@gmail.commailto:edvor...@gmail.com
phone: 646-250-9649tel:646-250-9649
eugenedvorkin.comhttp://eugenedvorkin.com
Connect with me on:
LinkedInhttp://www.linkedin.com/in/eugenedvorkin 
Twitterhttp://twitter.com/edvorkin




--
jarek@reijutsu:~$ fortune
You have Egyptian flu: you're going to be a mummy.
jarek@reijutsu:~$ fortune
You now have Asian Flu.


RE: Spark Training

2014-05-01 Thread Huang, Roger
If you're in the Bay Area, the Spark Summit would be a great source of 
information.
http://spark-summit.org/2014
-Roger

From: Nicholas Chammas [mailto:nicholas.cham...@gmail.com]
Sent: Thursday, May 01, 2014 10:12 AM
To: u...@spark.incubator.apache.org
Subject: Spark Training

There are many freely-available resources for the enterprising individual to 
use if they want to Spark up their life.

For others, some structured training is in order. Say I want everyone from my 
department at my company to get something like the AMP 
Camphttp://ampcamp.berkeley.edu/ experience, perhaps on-site.

What are my options for that?

Databricks doesn't have a contact page, so I figured this would be the next 
best place to ask.

Nick



View this message in context: Spark 
Traininghttp://apache-spark-user-list.1001560.n3.nabble.com/Spark-Training-tp5166.html
Sent from the Apache Spark User List mailing list 
archivehttp://apache-spark-user-list.1001560.n3.nabble.com/ at Nabble.com.


RE: Can Storm write an Aggregate Record to Postgres or SQL Server?

2014-04-08 Thread Huang, Roger
Neil,
Take a look at using tick tuples
http://nathanmarz.github.io/storm/doc/backtype/storm/Config.html#TOPOLOGY_TICK_TUPLE_FREQ_SECS

and the Storm RDBMS bolt  
https://github.com/nathanmarz/storm-contrib/tree/master/storm-rdbms
-Roger

From: Neil Carroll [mailto:carroll_n...@hotmail.com]
Sent: Tuesday, April 08, 2014 1:42 PM
To: user@storm.incubator.apache.org
Subject: Can Storm write an Aggregate Record to Postgres or SQL Server?

I'm new to Storm and want to use it to aggregate log data over a 5 minute 
period and write aggregate records (for each transaction type) into a DCMS (SQL 
or Postgres). I believe Storm can do this - and is there sample code available?
Thanks

Neil




RE: [DISCUSS] Pulling Contrib Modules into Apache

2014-02-26 Thread Huang, Roger
Bobby,
I vote to include both storm-yarn and storm-deploy.
Roger


-Original Message-
From: Brian O'Neill [mailto:boneil...@gmail.com] On Behalf Of Brian O'Neill
Sent: Wednesday, February 26, 2014 3:39 PM
To: dev@storm.incubator.apache.org
Cc: u...@storm.incubator.apache.org
Subject: Re: [DISCUSS] Pulling Contrib Modules into Apache


Bobby,

FWIW, I¹d love to see storm-yarn inside.  I think we could definitely make 
things easier on the end-user if they were more cohesive.

e.g. Imagine if we had ³storm launch yarn² inside of $storm/bin that would 
kickoff a storm-yarn launch, with whatever version was built.  It would likely 
simplify the ³create-tarball² and storm-yarn getStormConfig process as well.

-brian

---
Brian O'Neill
Chief Technology Officer

Health Market Science
The Science of Better Results
2700 Horizon Drive € King of Prussia, PA € 19406
M: 215.588.6024 € @boneill42 http://www.twitter.com/boneill42  € 
healthmarketscience.com

This information transmitted in this email message is for the intended 
recipient only and may contain confidential and/or privileged material. If you 
received this email in error and are not the intended recipient, or the person 
responsible to deliver it to the intended recipient, please contact the sender 
at the email above and delete this email and any attachments and destroy any 
copies thereof. Any review, retransmission, dissemination, copying or other use 
of, or taking any action in reliance upon, this information by persons or 
entities other than the intended recipient is strictly prohibited.
 






On 2/26/14, 4:25 PM, Bobby Evans ev...@yahoo-inc.com wrote:

I totally agree and I am +1 on bringing these spout/trident pieces in, 
assuming there are committers to support them.

I am also curious about how people feel about pulling in other projects 
like storm-starter, storm-deploy, storm-mesos, and storm-yarn?

Storm-starter in my option seems more like documentation and it would 
be nice to pull in so that it stays up to date with storm itself, just 
like the documentation.

The others are more of ways to run storm in different environments.  
They seem like there could be a lot of coupling between them and storm 
as storm evolves, and they kind of fit with integrate storm with 
*Technology X*² except X in this case is a compute environment instead 
of a data source or store. But then again we also just shot down a 
request to create juju charms for storm.

‹Bobby

From: P. Taylor Goetz ptgo...@gmail.commailto:ptgo...@gmail.com
Reply-To: 
dev@storm.incubator.apache.orgmailto:dev@storm.incubator.apache.org
Date: Wednesday, February 26, 2014 at 1:21 PM
To: 
dev@storm.incubator.apache.orgmailto:dev@storm.incubator.apache.org
Cc: 
u...@storm.incubator.apache.orgmailto:u...@storm.incubator.apache.org
u...@storm.incubator.apache.orgmailto:u...@storm.incubator.apache.org

Subject: Re: [DISCUSS] Pulling Contrib Modules into Apache

Thanks for the feedback Bobby.

To clarify, I¹m mainly talking about spout/bolt/trident state 
implementations that integrate storm with *Technology X*, where 
*Technology X* is not a fundamental part of storm.

Examples would be technologies that are part of or related to the 
Hadoop/Big Data ecosystem and enable the Lamda Architecture, e.g.: 
Kafka, HDFS, HBase, Cassandra, etc.

The idea behind having one or more Storm committers act as a ³sponsor² 
is to make sure new additions are done carefully and with good reason. 
To add a new module, it would require committer/PPMC consensus, and 
assignment of one or more sponsors. Part of a sponsor¹s job would be to 
ensure that a module is maintained, which would require enough 
familiarity with the code so support it long term. If a new module was 
proposed, but no committers were willing to act as a sponsor, it would 
not be added.

It would be the Committers¹/PPMC¹s responsibly to make sure things 
didn¹t get out of hand, and to do something about it if it does.

Here¹s an old Hadoop JIRA thread [1] discussing the addition of Hive as 
a contrib module, similar to what happened with HBase as Bobby pointed out.
Some interesting points are brought up. The difference here is that 
both HBase and Hive were pretty big codebases relative to Hadoop. With 
spout/bolt/state implementations I doubt we¹d see anything along that 
scale.

- Taylor

[1] https://issues.apache.org/jira/browse/HADOOP-3601


On Feb 26, 2014, at 12:35 PM, Bobby Evans 
ev...@yahoo-inc.commailto:ev...@yahoo-inc.com wrote:

I can see a lot of value in having a distribution of storm that comes 
with batteries included, everything is tested together and you know it 
works.  But I don¹t see much long term developer benefit in building 
them all together.  If there is strong coupling between storm and these 
external projects so that they break when storm changes then we need to 
understand the coupling and decide if we want to reduce that coupling 
by stabilizing APIs, improving version numbering and release process, 

RE: [DISCUSS] Pulling Contrib Modules into Apache

2014-02-26 Thread Huang, Roger
Bobby,
I vote to include both storm-yarn and storm-deploy.
Roger


-Original Message-
From: Brian O'Neill [mailto:boneil...@gmail.com] On Behalf Of Brian O'Neill
Sent: Wednesday, February 26, 2014 3:39 PM
To: d...@storm.incubator.apache.org
Cc: user@storm.incubator.apache.org
Subject: Re: [DISCUSS] Pulling Contrib Modules into Apache


Bobby,

FWIW, I¹d love to see storm-yarn inside.  I think we could definitely make 
things easier on the end-user if they were more cohesive.

e.g. Imagine if we had ³storm launch yarn² inside of $storm/bin that would 
kickoff a storm-yarn launch, with whatever version was built.  It would likely 
simplify the ³create-tarball² and storm-yarn getStormConfig process as well.

-brian

---
Brian O'Neill
Chief Technology Officer

Health Market Science
The Science of Better Results
2700 Horizon Drive € King of Prussia, PA € 19406
M: 215.588.6024 € @boneill42 http://www.twitter.com/boneill42  € 
healthmarketscience.com

This information transmitted in this email message is for the intended 
recipient only and may contain confidential and/or privileged material. If you 
received this email in error and are not the intended recipient, or the person 
responsible to deliver it to the intended recipient, please contact the sender 
at the email above and delete this email and any attachments and destroy any 
copies thereof. Any review, retransmission, dissemination, copying or other use 
of, or taking any action in reliance upon, this information by persons or 
entities other than the intended recipient is strictly prohibited.
 






On 2/26/14, 4:25 PM, Bobby Evans ev...@yahoo-inc.com wrote:

I totally agree and I am +1 on bringing these spout/trident pieces in, 
assuming there are committers to support them.

I am also curious about how people feel about pulling in other projects 
like storm-starter, storm-deploy, storm-mesos, and storm-yarn?

Storm-starter in my option seems more like documentation and it would 
be nice to pull in so that it stays up to date with storm itself, just 
like the documentation.

The others are more of ways to run storm in different environments.  
They seem like there could be a lot of coupling between them and storm 
as storm evolves, and they kind of fit with integrate storm with 
*Technology X*² except X in this case is a compute environment instead 
of a data source or store. But then again we also just shot down a 
request to create juju charms for storm.

‹Bobby

From: P. Taylor Goetz ptgo...@gmail.commailto:ptgo...@gmail.com
Reply-To: 
d...@storm.incubator.apache.orgmailto:d...@storm.incubator.apache.org
Date: Wednesday, February 26, 2014 at 1:21 PM
To: 
d...@storm.incubator.apache.orgmailto:d...@storm.incubator.apache.org
Cc: 
user@storm.incubator.apache.orgmailto:user@storm.incubator.apache.org
user@storm.incubator.apache.orgmailto:user@storm.incubator.apache.org

Subject: Re: [DISCUSS] Pulling Contrib Modules into Apache

Thanks for the feedback Bobby.

To clarify, I¹m mainly talking about spout/bolt/trident state 
implementations that integrate storm with *Technology X*, where 
*Technology X* is not a fundamental part of storm.

Examples would be technologies that are part of or related to the 
Hadoop/Big Data ecosystem and enable the Lamda Architecture, e.g.: 
Kafka, HDFS, HBase, Cassandra, etc.

The idea behind having one or more Storm committers act as a ³sponsor² 
is to make sure new additions are done carefully and with good reason. 
To add a new module, it would require committer/PPMC consensus, and 
assignment of one or more sponsors. Part of a sponsor¹s job would be to 
ensure that a module is maintained, which would require enough 
familiarity with the code so support it long term. If a new module was 
proposed, but no committers were willing to act as a sponsor, it would 
not be added.

It would be the Committers¹/PPMC¹s responsibly to make sure things 
didn¹t get out of hand, and to do something about it if it does.

Here¹s an old Hadoop JIRA thread [1] discussing the addition of Hive as 
a contrib module, similar to what happened with HBase as Bobby pointed out.
Some interesting points are brought up. The difference here is that 
both HBase and Hive were pretty big codebases relative to Hadoop. With 
spout/bolt/state implementations I doubt we¹d see anything along that 
scale.

- Taylor

[1] https://issues.apache.org/jira/browse/HADOOP-3601


On Feb 26, 2014, at 12:35 PM, Bobby Evans 
ev...@yahoo-inc.commailto:ev...@yahoo-inc.com wrote:

I can see a lot of value in having a distribution of storm that comes 
with batteries included, everything is tested together and you know it 
works.  But I don¹t see much long term developer benefit in building 
them all together.  If there is strong coupling between storm and these 
external projects so that they break when storm changes then we need to 
understand the coupling and decide if we want to reduce that coupling 
by stabilizing APIs, improving version numbering and release 

RE: Problem inserting set when query contains IF NOT EXISTS.

2014-01-13 Thread Huang, Roger
Validimir,
Thanks what version of Cassandra?
-Roger


From: Vladimir Prudnikov [mailto:v.prudni...@gmail.com]
Sent: Monday, January 13, 2014 11:57 AM
To: user
Subject: Problem inserting set when query contains IF NOT EXISTS.

Hi all,
I've spend a lot of time finding a bug in system, but it turns out that the 
problem is in Cassandra.

Here is how to reproduce.

=
CREATE KEYSPACE IF NOT EXISTS test_set WITH REPLICATION = { 'class' : 
'SimpleStrategy', 'replication_factor' : 1 };
USE test_set;

CREATE TABLE IF NOT EXISTS user (
key timeuuid PRIMARY KEY,
username text,
email text,
first_name text,
last_name text,
features settext,
) WITH caching='all';

INSERT INTO user(key,username,email,first_name,last_name,features) VALUES 
(now(),'ainsttp0ess2kiphu2pe1bbrle','l3b7brn6jp9e8s0mmsr7ae5...@mmcm4jf9a9g9b95c053ksbsi18.commailto:l3b7brn6jp9e8s0mmsr7ae5...@mmcm4jf9a9g9b95c053ksbsi18.com','gqh9ekmv6vc9nf1ce8eo3rjcdd','fmg92158br9ddivoj59417q514',{'i6v8i4a5gpnris5chjibllqf0','480m4c1obiq61ilii1g7rm0v17','50kovlifrtrtqihnvmbefaeacl'})
 IF NOT EXISTS;

select * from user;
==

The problem here is that user.features is null instead of set of 3 strings.
If you remove `IF NOT EXISTS` it executes correctly and set of string will be 
inserted.

I don't see any problem with the queries, seems to be the problem with C*.

--
Vladimir Prudnikov


JMeter HTTP Request equivalent of curl -XGET -d

2013-07-11 Thread Huang, Roger
Hi
In JMeter how can I configure HTTP Request to make the equivalent of an HTTP 
GET and pass in a JSON document with the -d (data) flag?

curl -XGET 'http://localhost:9200/vis/userDemo/_search?pretty=true' -d '{
 query : {
 query_string : {
 default_field : firstname,
 query: Joe~
 }
 }
 }'

I would also like to make an HTTP POST and pass in a JSON document with -d.
Thanks,
R


-
To unsubscribe, e-mail: user-unsubscr...@jmeter.apache.org
For additional commands, e-mail: user-h...@jmeter.apache.org



RE: How to specify an expression for a Mojo annotation parameter in Maven 3?

2013-01-17 Thread Huang, Roger
Hi Stephen,
Can you elaborate on your response?
I'm trying to figure out the Mojo annotations so that the field 
deliverableSettingsBaseDir is injected with an expression of:
 value of the basedir property concatenated with a constant value 
/settings.xconfig.

Thanks
Roger



-Original Message-
From: Stephen Connolly [mailto:stephen.alan.conno...@gmail.com] 
Sent: Thursday, January 17, 2013 11:04 AM
To: Maven Users List
Subject: Re: How to specify an expression for a Mojo annotation parameter in 
Maven 3?

That is really a default value not an expression


On 17 January 2013 17:01, Huang, Roger roghu...@paypal.com wrote:


 Is there a way to specify an expression for the Mojo annotation 
 @parameter in Maven 3?

 I have a Mojo, GenerateConfigMojo.java, with property 
 deliverableSettings of type java.io.File.
 Here is the original GenerateConfigMojo.java, built w/ Maven 2.2.1:

 /**
 * Goal which touches a timestamp file.
 *
  * @goal generate-config
 * @phase generate-resources
 */
 public class GenerateConfigMojo extends AbstractMojo {

 snip
 /**
 * @parameter expression=${basedir}/settings.xconfig
 * @required
 */
 private File deliverableSettings;

 When I build this with Maven 3.0.4, I get this WARNING:

 [WARNING]
 com.paypal.maven.plugins.deliverable.GenerateConfigMojo#deliverableSettings:^M
 [WARNING]   The syntax^M
 [WARNING] @parameter expression=${property}^M
 [WARNING]   is deprecated, please use^M
 [WARNING] @parameter property=property^M
 [WARNING]   instead.^M

 My current solution is to set the expression value in the Java code.
 I changed the Java 5 annotation to

 /**
 * Goal which touches a timestamp file.
 *
  * @goal generate-config
 * @phase generate-resources
 */
 public class GenerateConfigMojo extends AbstractMojo {

 snip
 /**
 * @parameter property=basedir
 * @required
 */
 private File deliverableSettingsBaseDir;

 snip

this.deliverableSettings = new File(
 this.deliverableSettingsBaseDir.getPath() + /settings.xconfig);

 Is there a way to still specify the expression via annotations?

 Thanks,
 Roger



-
To unsubscribe, e-mail: users-unsubscr...@maven.apache.org
For additional commands, e-mail: users-h...@maven.apache.org



RE: documentation on PlayOrm released

2012-11-07 Thread Huang, Roger
Dean,
What's the URL?
-Roger


-Original Message-
From: Hiller, Dean [mailto:dean.hil...@nrel.gov] 
Sent: Wednesday, November 07, 2012 7:43 AM
To: user@cassandra.apache.org
Subject: documentation on PlayOrm released

The first set of documentation on PlayOrm is now released.  It is also still 
growing as we have a dedicated person working on more documentation.  Check it 
out when you have a chance.

Later,
Dean