RE: Support/Consulting companies
http://thelastpickle.com/ From: Roxy Ubi [mailto:roxy...@gmail.com] Sent: Friday, August 19, 2016 2:02 PM To: user@cassandra.apache.org Subject: Support/Consulting companies Howdy, I'm looking for a list of support or consulting companies that provide contracting services related to Cassandra. Is there a comprehensive list somewhere? Alternatively could you folks tell me who you use? Thanks in advance for any replies! Roxy
RE: Validation of Data after data migration from RDBMS to Cassandra
Surbhi, A general note, Cassandra tables are not normalized as in an RDBMS schema, so it’s not straight migration. Best practice is to design your queries and then design your Cassandra schema to support those queries. Patrick McFadin has a series on C* data modeling. http://docs.datastax.com/en/cassandra/2.1 Roger From: Surbhi Gupta [mailto:surbhi.gupt...@gmail.com] Sent: Thursday, August 06, 2015 11:31 AM To: user@cassandra.apache.org Subject: Validation of Data after data migration from RDBMS to Cassandra Hi, We have to migrate the data from Oracle/mysql to Cassandra. I wanted to understand, if we have any tool/utilitiy which can help in validation the data after the data migration to Cassandra. Thanks Surbhi
RE: Validation of Data after data migration from RDBMS to Cassandra
Surbhi Sorry I meant to send the CQL link http://docs.datastax.com/en/cql/3.1/cql/cql_intro_c.html Roger From: Huang, Roger [mailto:rohu...@visa.com] Sent: Thursday, August 06, 2015 11:39 AM To: user@cassandra.apache.org Subject: RE: Validation of Data after data migration from RDBMS to Cassandra Surbhi, A general note, Cassandra tables are not normalized as in an RDBMS schema, so it’s not straight migration. Best practice is to design your queries and then design your Cassandra schema to support those queries. Patrick McFadin has a series on C* data modeling. http://docs.datastax.com/en/cassandra/2.1 Roger From: Surbhi Gupta [mailto:surbhi.gupt...@gmail.com] Sent: Thursday, August 06, 2015 11:31 AM To: user@cassandra.apache.orgmailto:user@cassandra.apache.org Subject: Validation of Data after data migration from RDBMS to Cassandra Hi, We have to migrate the data from Oracle/mysql to Cassandra. I wanted to understand, if we have any tool/utilitiy which can help in validation the data after the data migration to Cassandra. Thanks Surbhi
RE: How to share large resources like dictionaries while processing data with Spark ?
Is the dictionary read-only? Did you look at http://spark.apache.org/docs/latest/programming-guide.html#broadcast-variables ? -Original Message- From: dgoldenberg [mailto:dgoldenberg...@gmail.com] Sent: Thursday, June 04, 2015 4:50 PM To: user@spark.apache.org Subject: How to share large resources like dictionaries while processing data with Spark ? We have some pipelines defined where sometimes we need to load potentially large resources such as dictionaries. What would be the best strategy for sharing such resources among the transformations/actions within a consumer? Can they be shared somehow across the RDD's? I'm looking for a way to load such a resource once into the cluster memory and have it be available throughout the lifecycle of a consumer... Thanks. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/How-to-share-large-resources-like-dictionaries-while-processing-data-with-Spark-tp23162.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org - To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org
RE: STORM UI -- trident topology continuously emitting ..
Have you looked at http://www.sematext.com/spm/index.html From: Hadi Sotudeh [mailto:hadisotudeh1...@gmail.com] Sent: Thursday, May 21, 2015 9:53 AM To: user Cc: Nikhil Singh Subject: Re: STORM UI -- trident topology continuously emitting .. Is there any other monitoring tools for storm?
RE: Spark on Cassandra
http://planetcassandra.org/getting-started-with-apache-spark-and-cassandra/ http://planetcassandra.org/blog/holy-momentum-batman-spark-and-cassandra-circa-2015-w-datastax-connector-and-java/ https://github.com/datastax/spark-cassandra-connector From: Cody Koeninger [mailto:c...@koeninger.org] Sent: Wednesday, April 29, 2015 12:15 PM To: Matthew Johnson Cc: user@spark.apache.org Subject: Re: Spark on Cassandra Hadoop version doesn't matter if you're just using cassandra. On Wed, Apr 29, 2015 at 12:08 PM, Matthew Johnson matt.john...@algomi.commailto:matt.john...@algomi.com wrote: Hi all, I am new to Spark, but excited to use it with our Cassandra cluster. I have read in a few places that Spark can interact directly with Cassandra now, so I decided to download it and have a play – I am happy to run it in standalone cluster mode initially. When I go to download it (http://spark.apache.org/downloads.html) I see a bunch of pre-built versions for Hadoop and MapR, but no mention of Cassandra – if I am running it in standalone cluster mode, does it matter which pre-built package I download? Would all of them work? Or do I have to build it myself from source with some special config for Cassandra? Thanks! Matt
RE: Spark + Mahout
Taka, Have you considered using Spark’s own MLlib k-means clustering? http://spark.apache.org/docs/latest/mllib-clustering.html Roger From: Daniel Takabayashi [mailto:takabaya...@scanboo.com.br] Sent: Thursday, September 18, 2014 1:50 PM To: user@spark.apache.org Subject: Spark + Mahout Hi guys, Is possible to run a mahout kmeans throws spark infrastructure? Thanks, taka (Brazil)
FW: Indexing a term into separate Lucene indexes
If I have documents with a person and his email address: u...@domain.commailto:u...@domain.com How can I configure Solr (4.6) so that the email address source field is indexed as - the user part of the address (e.g., user) is in Lucene index X - the domain part of the address (e.g., domain.com) is in a separate Lucene index Y I would like to be able search as follows: - Find all people whose email addresses have user part = userXyz - Find all people whose email addresses have domain part = domainABC.com - Find the person with exact email address = user...@domainabc.commailto:user...@domainabc.com Would I use a copyField declaration in my schema? http://wiki.apache.org/solr/SchemaXml#Copy_Fields Thanks!
RE: Indexing a term into separate Lucene indexes
Shawn, Thanks for your response. Due to security requirements, I do need the name and domain parts of the email address stored in separate Lucene indexes. How do you recommend doing this? What are the challenges? Once the name and domain parts of the email address are in different Lucene indexes, would I need to modify my Solr search string? Thanks, Roger -Original Message- From: Shawn Heisey [mailto:s...@elyograg.org] Sent: Friday, June 20, 2014 10:19 AM To: solr-user@lucene.apache.org Subject: Re: Indexing a term into separate Lucene indexes On 6/19/2014 4:51 PM, Huang, Roger wrote: If I have documents with a person and his email address: u...@domain.commailto:u...@domain.com How can I configure Solr (4.6) so that the email address source field is indexed as - the user part of the address (e.g., user) is in Lucene index X - the domain part of the address (e.g., domain.com) is in a separate Lucene index Y I would like to be able search as follows: - Find all people whose email addresses have user part = userXyz - Find all people whose email addresses have domain part = domainABC.com - Find the person with exact email address = user...@domainabc.com Would I use a copyField declaration in my schema? http://wiki.apache.org/solr/SchemaXml#Copy_Fields I don't think you actually want the data to end up in entirely different indexes. Although it is possible to search more than one separate index, that's very likely NOT what you want to do, and it comes with its own challenges. What you most likely want is to put this data into different fields within the same index. You'll need to write custom code to accomplish this, especially if you need the stored data to contain only the parts rather than the complete email address. A copyField can get the data to additional fields, but I'm not aware of anything built-in to the schema that can trim the unwanted information from the new fields, and even if there is, any stored data will be the original data for all three fields. It's up to you whether this custom code is in a user application that does your indexing or in a custom update processor that you load as a plugin to Solr itself. Extending whatever user application you are already using for indexing is very likely to be a lot easier. Thanks, Shawn
Indexing a term into separate Lucene indexes
If I have documents with a person and his email address: u...@domain.commailto:u...@domain.com How can I configure Solr (4.6) so that the email address source field is indexed as - the user part of the address (e.g., user) is in Lucene index X - the domain part of the address (e.g., domain.com) is in a separate Lucene index Y I would like to be able search as follows: - Find all people whose email addresses have user part = userXyz - Find all people whose email addresses have domain part = domainABC.com - Find the person with exact email address = user...@domainabc.com Would I use a copyField declaration in my schema? http://wiki.apache.org/solr/SchemaXml#Copy_Fields Thanks!
RE: Doubts on Apache Storm
Milind, Have you looked at http://storm.incubator.apache.org/documentation/Home.html http://storm.incubator.apache.org/documentation/Creating-a-new-Storm-project.html Roger -Original Message- From: milind.pa...@polarisft.com [mailto:milind.pa...@polarisft.com] Sent: Tuesday, May 06, 2014 8:55 AM To: user@storm.incubator.apache.org Subject: Fw: Doubts on Apache Storm Hi All, Thanks for clarifying the doubts. Can I request if there is any doc that demonstrate the deployment steps, pseudo code, FAQ, etc. (Basically basic building blocks) Regards Milind Patil | Intellect Liquidity cash management 8SWS 031 | Silver metropolis | Western express highway | Goregaon (East) | Mumbai | 400 063. INDIA Board: 91-22-67801500 | 91-22-42029200 | Ext: 1734 Mobile: +91 9920612360 | mail id: milind.pa...@polarisft.com - Forwarded by Milind Patil/BSC37/polarisFT on 05/06/2014 07:24 PM - From: Marc Vaillant vaill...@animetrics.com To: user@storm.incubator.apache.org Date: 05/06/2014 07:12 PM Subject:Re: Doubts on Apache Storm On Tue, May 06, 2014 at 03:21:13PM +0530, milind.pa...@polarisft.com wrote: Hi, Is Nimbus mandatory for storm? (Our development env is neither using Nimbus nor any other cloud environment) I think you might have misunderstood nimbus. It is a daemon that is part of storm, *not* nimbus from the nimbus project. (I am new to Apache storm, it would really help me if any basic document available on Apache storm) Regards Milind Patil | Intellect Liquidity cash management 8SWS 031 | Silver metropolis | Western express highway | Goregaon (East) | Mumbai | 400 063. INDIA Board: 91-22-67801500 | 91-22-42029200 | Ext: 1734 Mobile: +91 9920612360 | mail id: milind.pa...@polarisft.com This e-Mail may contain proprietary and confidential information and is sent for the intended recipient(s) only. If by an addressing or transmission error this mail has been misdirected to you, you are requested to delete this mail immediately. You are also hereby notified that any use, any form of reproduction, dissemination, copying, disclosure, modification, distribution and/or publication of this e-mail message, contents or its attachment other than by its intended recipient/s is strictly prohibited. Visit us at http://www.polarisFT.com This e-Mail may contain proprietary and confidential information and is sent for the intended recipient(s) only. If by an addressing or transmission error this mail has been misdirected to you, you are requested to delete this mail immediately. You are also hereby notified that any use, any form of reproduction, dissemination, copying, disclosure, modification, distribution and/or publication of this e-mail message, contents or its attachment other than by its intended recipient/s is strictly prohibited. Visit us at http://www.polarisFT.com
RE: Is there a good reference book for storm users?
Quinton Anderson’s book is good, and includes ML examples. http://www.packtpub.com/storm-realtime-processing-cookbook/book -Roger From: Jarek Śmiejczak [mailto:j...@smiejczak.net] Sent: Sunday, May 04, 2014 6:01 AM To: user@storm.incubator.apache.org Subject: Re: Is there a good reference book for storm users? Two weeks ago i've bought: http://www.manning.com/sallen/ and i'm really surprised about contents. Everything is explained in really non-boring and professional manner. I think it's worth to buy this book even before final release. On Sun, May 4, 2014 at 9:23 AM, Sajith sajith...@gmail.commailto:sajith...@gmail.com wrote: Thank you very much Milinda and Eugene On Sun, May 4, 2014 at 8:23 AM, Eugene edvor...@gmail.commailto:edvor...@gmail.com wrote: I second that. Taylor and Brian book good for advanced users. A lot of information about Trident. Very interesting use cases and written very well with step-by-step instructions. On Sat, May 3, 2014 at 6:41 PM, Milinda Pathirage mpath...@umail.iu.edumailto:mpath...@umail.iu.edu wrote: I think Taylor's book [1] is good. I bought that, but still couldn't read it. But according to table of content it looks promising. Thanks Milinda [1] http://www.amazon.com/Storm-Blueprints-Distributed-Real-time-Computation/dp/178216829X/ref=sr_1_5?ie=UTF8qid=1399156748sr=8-5keywords=storm On Sat, May 3, 2014 at 3:36 PM, Sajith sajith...@gmail.commailto:sajith...@gmail.com wrote: Hi all, Can someone suggest me an up-to date comprehensive reference book on storm for advanced users? Thanks, Sajith. -- Milinda Pathirage PhD Student | Research Assistant School of Informatics and Computing | Data to Insight Center Indiana University twitter: milindalakmal skype: milinda.pathirage blog: http://milinda.pathirage.org -- Eugene Dvorkin Software Engineer New York City Storm User Group - organizer WebMD email: edvor...@gmail.commailto:edvor...@gmail.com phone: 646-250-9649tel:646-250-9649 eugenedvorkin.comhttp://eugenedvorkin.com Connect with me on: LinkedInhttp://www.linkedin.com/in/eugenedvorkin Twitterhttp://twitter.com/edvorkin -- jarek@reijutsu:~$ fortune You have Egyptian flu: you're going to be a mummy. jarek@reijutsu:~$ fortune You now have Asian Flu.
RE: Spark Training
If you're in the Bay Area, the Spark Summit would be a great source of information. http://spark-summit.org/2014 -Roger From: Nicholas Chammas [mailto:nicholas.cham...@gmail.com] Sent: Thursday, May 01, 2014 10:12 AM To: u...@spark.incubator.apache.org Subject: Spark Training There are many freely-available resources for the enterprising individual to use if they want to Spark up their life. For others, some structured training is in order. Say I want everyone from my department at my company to get something like the AMP Camphttp://ampcamp.berkeley.edu/ experience, perhaps on-site. What are my options for that? Databricks doesn't have a contact page, so I figured this would be the next best place to ask. Nick View this message in context: Spark Traininghttp://apache-spark-user-list.1001560.n3.nabble.com/Spark-Training-tp5166.html Sent from the Apache Spark User List mailing list archivehttp://apache-spark-user-list.1001560.n3.nabble.com/ at Nabble.com.
RE: Can Storm write an Aggregate Record to Postgres or SQL Server?
Neil, Take a look at using tick tuples http://nathanmarz.github.io/storm/doc/backtype/storm/Config.html#TOPOLOGY_TICK_TUPLE_FREQ_SECS and the Storm RDBMS bolt https://github.com/nathanmarz/storm-contrib/tree/master/storm-rdbms -Roger From: Neil Carroll [mailto:carroll_n...@hotmail.com] Sent: Tuesday, April 08, 2014 1:42 PM To: user@storm.incubator.apache.org Subject: Can Storm write an Aggregate Record to Postgres or SQL Server? I'm new to Storm and want to use it to aggregate log data over a 5 minute period and write aggregate records (for each transaction type) into a DCMS (SQL or Postgres). I believe Storm can do this - and is there sample code available? Thanks Neil
RE: [DISCUSS] Pulling Contrib Modules into Apache
Bobby, I vote to include both storm-yarn and storm-deploy. Roger -Original Message- From: Brian O'Neill [mailto:boneil...@gmail.com] On Behalf Of Brian O'Neill Sent: Wednesday, February 26, 2014 3:39 PM To: dev@storm.incubator.apache.org Cc: u...@storm.incubator.apache.org Subject: Re: [DISCUSS] Pulling Contrib Modules into Apache Bobby, FWIW, I¹d love to see storm-yarn inside. I think we could definitely make things easier on the end-user if they were more cohesive. e.g. Imagine if we had ³storm launch yarn² inside of $storm/bin that would kickoff a storm-yarn launch, with whatever version was built. It would likely simplify the ³create-tarball² and storm-yarn getStormConfig process as well. -brian --- Brian O'Neill Chief Technology Officer Health Market Science The Science of Better Results 2700 Horizon Drive € King of Prussia, PA € 19406 M: 215.588.6024 € @boneill42 http://www.twitter.com/boneill42 € healthmarketscience.com This information transmitted in this email message is for the intended recipient only and may contain confidential and/or privileged material. If you received this email in error and are not the intended recipient, or the person responsible to deliver it to the intended recipient, please contact the sender at the email above and delete this email and any attachments and destroy any copies thereof. Any review, retransmission, dissemination, copying or other use of, or taking any action in reliance upon, this information by persons or entities other than the intended recipient is strictly prohibited. On 2/26/14, 4:25 PM, Bobby Evans ev...@yahoo-inc.com wrote: I totally agree and I am +1 on bringing these spout/trident pieces in, assuming there are committers to support them. I am also curious about how people feel about pulling in other projects like storm-starter, storm-deploy, storm-mesos, and storm-yarn? Storm-starter in my option seems more like documentation and it would be nice to pull in so that it stays up to date with storm itself, just like the documentation. The others are more of ways to run storm in different environments. They seem like there could be a lot of coupling between them and storm as storm evolves, and they kind of fit with integrate storm with *Technology X*² except X in this case is a compute environment instead of a data source or store. But then again we also just shot down a request to create juju charms for storm. ‹Bobby From: P. Taylor Goetz ptgo...@gmail.commailto:ptgo...@gmail.com Reply-To: dev@storm.incubator.apache.orgmailto:dev@storm.incubator.apache.org Date: Wednesday, February 26, 2014 at 1:21 PM To: dev@storm.incubator.apache.orgmailto:dev@storm.incubator.apache.org Cc: u...@storm.incubator.apache.orgmailto:u...@storm.incubator.apache.org u...@storm.incubator.apache.orgmailto:u...@storm.incubator.apache.org Subject: Re: [DISCUSS] Pulling Contrib Modules into Apache Thanks for the feedback Bobby. To clarify, I¹m mainly talking about spout/bolt/trident state implementations that integrate storm with *Technology X*, where *Technology X* is not a fundamental part of storm. Examples would be technologies that are part of or related to the Hadoop/Big Data ecosystem and enable the Lamda Architecture, e.g.: Kafka, HDFS, HBase, Cassandra, etc. The idea behind having one or more Storm committers act as a ³sponsor² is to make sure new additions are done carefully and with good reason. To add a new module, it would require committer/PPMC consensus, and assignment of one or more sponsors. Part of a sponsor¹s job would be to ensure that a module is maintained, which would require enough familiarity with the code so support it long term. If a new module was proposed, but no committers were willing to act as a sponsor, it would not be added. It would be the Committers¹/PPMC¹s responsibly to make sure things didn¹t get out of hand, and to do something about it if it does. Here¹s an old Hadoop JIRA thread [1] discussing the addition of Hive as a contrib module, similar to what happened with HBase as Bobby pointed out. Some interesting points are brought up. The difference here is that both HBase and Hive were pretty big codebases relative to Hadoop. With spout/bolt/state implementations I doubt we¹d see anything along that scale. - Taylor [1] https://issues.apache.org/jira/browse/HADOOP-3601 On Feb 26, 2014, at 12:35 PM, Bobby Evans ev...@yahoo-inc.commailto:ev...@yahoo-inc.com wrote: I can see a lot of value in having a distribution of storm that comes with batteries included, everything is tested together and you know it works. But I don¹t see much long term developer benefit in building them all together. If there is strong coupling between storm and these external projects so that they break when storm changes then we need to understand the coupling and decide if we want to reduce that coupling by stabilizing APIs, improving version numbering and release process,
RE: [DISCUSS] Pulling Contrib Modules into Apache
Bobby, I vote to include both storm-yarn and storm-deploy. Roger -Original Message- From: Brian O'Neill [mailto:boneil...@gmail.com] On Behalf Of Brian O'Neill Sent: Wednesday, February 26, 2014 3:39 PM To: d...@storm.incubator.apache.org Cc: user@storm.incubator.apache.org Subject: Re: [DISCUSS] Pulling Contrib Modules into Apache Bobby, FWIW, I¹d love to see storm-yarn inside. I think we could definitely make things easier on the end-user if they were more cohesive. e.g. Imagine if we had ³storm launch yarn² inside of $storm/bin that would kickoff a storm-yarn launch, with whatever version was built. It would likely simplify the ³create-tarball² and storm-yarn getStormConfig process as well. -brian --- Brian O'Neill Chief Technology Officer Health Market Science The Science of Better Results 2700 Horizon Drive € King of Prussia, PA € 19406 M: 215.588.6024 € @boneill42 http://www.twitter.com/boneill42 € healthmarketscience.com This information transmitted in this email message is for the intended recipient only and may contain confidential and/or privileged material. If you received this email in error and are not the intended recipient, or the person responsible to deliver it to the intended recipient, please contact the sender at the email above and delete this email and any attachments and destroy any copies thereof. Any review, retransmission, dissemination, copying or other use of, or taking any action in reliance upon, this information by persons or entities other than the intended recipient is strictly prohibited. On 2/26/14, 4:25 PM, Bobby Evans ev...@yahoo-inc.com wrote: I totally agree and I am +1 on bringing these spout/trident pieces in, assuming there are committers to support them. I am also curious about how people feel about pulling in other projects like storm-starter, storm-deploy, storm-mesos, and storm-yarn? Storm-starter in my option seems more like documentation and it would be nice to pull in so that it stays up to date with storm itself, just like the documentation. The others are more of ways to run storm in different environments. They seem like there could be a lot of coupling between them and storm as storm evolves, and they kind of fit with integrate storm with *Technology X*² except X in this case is a compute environment instead of a data source or store. But then again we also just shot down a request to create juju charms for storm. ‹Bobby From: P. Taylor Goetz ptgo...@gmail.commailto:ptgo...@gmail.com Reply-To: d...@storm.incubator.apache.orgmailto:d...@storm.incubator.apache.org Date: Wednesday, February 26, 2014 at 1:21 PM To: d...@storm.incubator.apache.orgmailto:d...@storm.incubator.apache.org Cc: user@storm.incubator.apache.orgmailto:user@storm.incubator.apache.org user@storm.incubator.apache.orgmailto:user@storm.incubator.apache.org Subject: Re: [DISCUSS] Pulling Contrib Modules into Apache Thanks for the feedback Bobby. To clarify, I¹m mainly talking about spout/bolt/trident state implementations that integrate storm with *Technology X*, where *Technology X* is not a fundamental part of storm. Examples would be technologies that are part of or related to the Hadoop/Big Data ecosystem and enable the Lamda Architecture, e.g.: Kafka, HDFS, HBase, Cassandra, etc. The idea behind having one or more Storm committers act as a ³sponsor² is to make sure new additions are done carefully and with good reason. To add a new module, it would require committer/PPMC consensus, and assignment of one or more sponsors. Part of a sponsor¹s job would be to ensure that a module is maintained, which would require enough familiarity with the code so support it long term. If a new module was proposed, but no committers were willing to act as a sponsor, it would not be added. It would be the Committers¹/PPMC¹s responsibly to make sure things didn¹t get out of hand, and to do something about it if it does. Here¹s an old Hadoop JIRA thread [1] discussing the addition of Hive as a contrib module, similar to what happened with HBase as Bobby pointed out. Some interesting points are brought up. The difference here is that both HBase and Hive were pretty big codebases relative to Hadoop. With spout/bolt/state implementations I doubt we¹d see anything along that scale. - Taylor [1] https://issues.apache.org/jira/browse/HADOOP-3601 On Feb 26, 2014, at 12:35 PM, Bobby Evans ev...@yahoo-inc.commailto:ev...@yahoo-inc.com wrote: I can see a lot of value in having a distribution of storm that comes with batteries included, everything is tested together and you know it works. But I don¹t see much long term developer benefit in building them all together. If there is strong coupling between storm and these external projects so that they break when storm changes then we need to understand the coupling and decide if we want to reduce that coupling by stabilizing APIs, improving version numbering and release
RE: Problem inserting set when query contains IF NOT EXISTS.
Validimir, Thanks what version of Cassandra? -Roger From: Vladimir Prudnikov [mailto:v.prudni...@gmail.com] Sent: Monday, January 13, 2014 11:57 AM To: user Subject: Problem inserting set when query contains IF NOT EXISTS. Hi all, I've spend a lot of time finding a bug in system, but it turns out that the problem is in Cassandra. Here is how to reproduce. = CREATE KEYSPACE IF NOT EXISTS test_set WITH REPLICATION = { 'class' : 'SimpleStrategy', 'replication_factor' : 1 }; USE test_set; CREATE TABLE IF NOT EXISTS user ( key timeuuid PRIMARY KEY, username text, email text, first_name text, last_name text, features settext, ) WITH caching='all'; INSERT INTO user(key,username,email,first_name,last_name,features) VALUES (now(),'ainsttp0ess2kiphu2pe1bbrle','l3b7brn6jp9e8s0mmsr7ae5...@mmcm4jf9a9g9b95c053ksbsi18.commailto:l3b7brn6jp9e8s0mmsr7ae5...@mmcm4jf9a9g9b95c053ksbsi18.com','gqh9ekmv6vc9nf1ce8eo3rjcdd','fmg92158br9ddivoj59417q514',{'i6v8i4a5gpnris5chjibllqf0','480m4c1obiq61ilii1g7rm0v17','50kovlifrtrtqihnvmbefaeacl'}) IF NOT EXISTS; select * from user; == The problem here is that user.features is null instead of set of 3 strings. If you remove `IF NOT EXISTS` it executes correctly and set of string will be inserted. I don't see any problem with the queries, seems to be the problem with C*. -- Vladimir Prudnikov
JMeter HTTP Request equivalent of curl -XGET -d
Hi In JMeter how can I configure HTTP Request to make the equivalent of an HTTP GET and pass in a JSON document with the -d (data) flag? curl -XGET 'http://localhost:9200/vis/userDemo/_search?pretty=true' -d '{ query : { query_string : { default_field : firstname, query: Joe~ } } }' I would also like to make an HTTP POST and pass in a JSON document with -d. Thanks, R - To unsubscribe, e-mail: user-unsubscr...@jmeter.apache.org For additional commands, e-mail: user-h...@jmeter.apache.org
RE: How to specify an expression for a Mojo annotation parameter in Maven 3?
Hi Stephen, Can you elaborate on your response? I'm trying to figure out the Mojo annotations so that the field deliverableSettingsBaseDir is injected with an expression of: value of the basedir property concatenated with a constant value /settings.xconfig. Thanks Roger -Original Message- From: Stephen Connolly [mailto:stephen.alan.conno...@gmail.com] Sent: Thursday, January 17, 2013 11:04 AM To: Maven Users List Subject: Re: How to specify an expression for a Mojo annotation parameter in Maven 3? That is really a default value not an expression On 17 January 2013 17:01, Huang, Roger roghu...@paypal.com wrote: Is there a way to specify an expression for the Mojo annotation @parameter in Maven 3? I have a Mojo, GenerateConfigMojo.java, with property deliverableSettings of type java.io.File. Here is the original GenerateConfigMojo.java, built w/ Maven 2.2.1: /** * Goal which touches a timestamp file. * * @goal generate-config * @phase generate-resources */ public class GenerateConfigMojo extends AbstractMojo { snip /** * @parameter expression=${basedir}/settings.xconfig * @required */ private File deliverableSettings; When I build this with Maven 3.0.4, I get this WARNING: [WARNING] com.paypal.maven.plugins.deliverable.GenerateConfigMojo#deliverableSettings:^M [WARNING] The syntax^M [WARNING] @parameter expression=${property}^M [WARNING] is deprecated, please use^M [WARNING] @parameter property=property^M [WARNING] instead.^M My current solution is to set the expression value in the Java code. I changed the Java 5 annotation to /** * Goal which touches a timestamp file. * * @goal generate-config * @phase generate-resources */ public class GenerateConfigMojo extends AbstractMojo { snip /** * @parameter property=basedir * @required */ private File deliverableSettingsBaseDir; snip this.deliverableSettings = new File( this.deliverableSettingsBaseDir.getPath() + /settings.xconfig); Is there a way to still specify the expression via annotations? Thanks, Roger - To unsubscribe, e-mail: users-unsubscr...@maven.apache.org For additional commands, e-mail: users-h...@maven.apache.org
RE: documentation on PlayOrm released
Dean, What's the URL? -Roger -Original Message- From: Hiller, Dean [mailto:dean.hil...@nrel.gov] Sent: Wednesday, November 07, 2012 7:43 AM To: user@cassandra.apache.org Subject: documentation on PlayOrm released The first set of documentation on PlayOrm is now released. It is also still growing as we have a dedicated person working on more documentation. Check it out when you have a chance. Later, Dean