[ https://issues.apache.org/jira/browse/SOLR-10317?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16106364#comment-16106364 ]
Ishan Chattopadhyaya commented on SOLR-10317: --------------------------------------------- {quote} > You need to be specific in terms of steps the user needs to take. https://github.com/viveknarang/lucene-solr/tree/SolrNightlyBenchmarks/dev-tools/SolrNightBenchmarks#steps-to-launch The steps have always existed on this file! Do you find any steps missing? {quote} What existed and exists now is "The system has the ability to check and download required data files from the source, as specified in the properties file". Again, why would a user care about the system's ability to download files? What he needs to know is, how to invoke that ability. Which "properties file", which property in that properties file needs to be adjusted? How to enable/disable this ability etc.? {quote} I don't think there is any complexity here. Checking the existence of these files is already an essential step. The files are only downloaded when missing and essentially a wget is done to download files please see code below. If you still insist I will create a shell script for this. {quote} Checking for existence is good. But automatically downloading it is unnecessary complexity. "essentially a wget is done to download files" <--- why the f*** would you want to invoke a shell command from within Java, when you can have the user/admin do it before launching your service? Downloading files requires a specialized software, and the benchmarking suite should not be concerned with that. Resuming partial downloads, multi-threaded downloading, etc. are special features that the downloader should take care of. I think the business of downloading files should be left for the user to deal with. {quote} > Secondly, "please use the link provided below" is equivalent to telling > the user to "go f*** yourself" The link has been provided as an option and is not a required step. When the user has configured the system to automatically download required files, there is no need for the user to manually download files. {quote} That link is useless. How will the user download all the files from that link? {quote} > Why are you calling it "test data files"? Please suggest what you want these files to be referred as. {quote} Perhaps "Benchmarking data files" or "Data files". > Solr Nightly Benchmarks > ----------------------- > > Key: SOLR-10317 > URL: https://issues.apache.org/jira/browse/SOLR-10317 > Project: Solr > Issue Type: Task > Reporter: Ishan Chattopadhyaya > Labels: gsoc2017, mentor > Attachments: changes-lucene-20160907.json, > changes-solr-20160907.json, managed-schema, > Narang-Vivek-SOLR-10317-Solr-Nightly-Benchmarks.docx, > Narang-Vivek-SOLR-10317-Solr-Nightly-Benchmarks-FINAL-PROPOSAL.pdf, > SOLR-10317.patch, SOLR-10317.patch, solrconfig.xml > > > Solr needs nightly benchmarks reporting. Similar Lucene benchmarks can be > found here, https://home.apache.org/~mikemccand/lucenebench/. > Preferably, we need: > # A suite of benchmarks that build Solr from a commit point, start Solr > nodes, both in SolrCloud and standalone mode, and record timing information > of various operations like indexing, querying, faceting, grouping, > replication etc. > # It should be possible to run them either as an independent suite or as a > Jenkins job, and we should be able to report timings as graphs (Jenkins has > some charting plugins). > # The code should eventually be integrated in the Solr codebase, so that it > never goes out of date. > There is some prior work / discussion: > # https://github.com/shalinmangar/solr-perf-tools (Shalin) > # https://github.com/chatman/solr-upgrade-tests/blob/master/BENCHMARKS.md > (Ishan/Vivek) > # SOLR-2646 & SOLR-9863 (Mark Miller) > # https://home.apache.org/~mikemccand/lucenebench/ (Mike McCandless) > # https://github.com/lucidworks/solr-scale-tk (Tim Potter) > There is support for building, starting, indexing/querying and stopping Solr > in some of these frameworks above. However, the benchmarks run are very > limited. Any of these can be a starting point, or a new framework can as well > be used. The motivation is to be able to cover every functionality of Solr > with a corresponding benchmark that is run every night. > Proposing this as a GSoC 2017 project. I'm willing to mentor, and I'm sure > [~shalinmangar] and [~markrmil...@gmail.com] would help here. -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org