[ https://issues.apache.org/jira/browse/SOLR-10317?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16106232#comment-16106232 ]
Ishan Chattopadhyaya edited comment on SOLR-10317 at 7/29/17 8:15 PM: ---------------------------------------------------------------------- I saw this in the section for running last n days of commits. {code} * While there is no feature implemented (yet), that supports automatic crunching of performance data for commits for last (n) days, there is still a way to do this. * Prepare and arrange (in order) a list of commit hashes for the (n) days, either one for each day or all of the commits for (n) days. * Now create empty files with name (no extension) EXACTLY equal to commit hashes in the sub-folder /data/commit_queue/ found in the directory as configured in the properties file for benchmarkAppDirectory parameter. The hashes will be picked in the order of their creation (FIFO). * Finally, in the section "Steps to run" follow all the steps but for step 8 above, use parameter --from-queue instead of --latest-commit {code} Just use a shell script like this to do print last 30 days of commits (the last commit of every day): {code} # Print the last commit of every day, for past 30 days, in the order of most recent day first for n in {1..30}; do date=`date --date="$n days ago" +%Y-%m-%d`; git log --after="$date 00:00" --before="$date 23:59"|head -1|cut -f 2 -d " "; done {code} bq. Steps to run for N days are added in the README file. Yes, currently, it has to be done manually but I will try to add a new feature so this could be done using parameters very soon. No need to add a parameter. Shell scripting can be your friend. All the steps that you mentioned can be automated using the above script; just pass the commit to the --commit parameter. Kindly update the steps in the README asap. was (Author: ichattopadhyaya): I saw this in the section for running last n days of commits. {code} * While there is no feature implemented (yet), that supports automatic crunching of performance data for commits for last (n) days, there is still a way to do this. * Prepare and arrange (in order) a list of commit hashes for the (n) days, either one for each day or all of the commits for (n) days. * Now create empty files with name (no extension) EXACTLY equal to commit hashes in the sub-folder /data/commit_queue/ found in the directory as configured in the properties file for benchmarkAppDirectory parameter. The hashes will be picked in the order of their creation (FIFO). * Finally, in the section "Steps to run" follow all the steps but for step 8 above, use parameter --from-queue instead of --latest-commit {code} Just use a shell script like this to do this: {code} for n in {1..30}; do date=`date --date="$n days ago" +%Y-%m-%d`; git log --after="$date 00:00" --before="$date 23:59"|head -1|cut -f 2 -d " "; done {code} bq. Steps to run for N days are added in the README file. Yes, currently, it has to be done manually but I will try to add a new feature so this could be done using parameters very soon. No need to add a parameter. Shell scripting can be your friend. All the steps that you mentioned can be automated using the above script. Kindly update the steps asap. > Solr Nightly Benchmarks > ----------------------- > > Key: SOLR-10317 > URL: https://issues.apache.org/jira/browse/SOLR-10317 > Project: Solr > Issue Type: Task > Reporter: Ishan Chattopadhyaya > Labels: gsoc2017, mentor > Attachments: changes-lucene-20160907.json, > changes-solr-20160907.json, managed-schema, > Narang-Vivek-SOLR-10317-Solr-Nightly-Benchmarks.docx, > Narang-Vivek-SOLR-10317-Solr-Nightly-Benchmarks-FINAL-PROPOSAL.pdf, > SOLR-10317.patch, SOLR-10317.patch, solrconfig.xml > > > Solr needs nightly benchmarks reporting. Similar Lucene benchmarks can be > found here, https://home.apache.org/~mikemccand/lucenebench/. > Preferably, we need: > # A suite of benchmarks that build Solr from a commit point, start Solr > nodes, both in SolrCloud and standalone mode, and record timing information > of various operations like indexing, querying, faceting, grouping, > replication etc. > # It should be possible to run them either as an independent suite or as a > Jenkins job, and we should be able to report timings as graphs (Jenkins has > some charting plugins). > # The code should eventually be integrated in the Solr codebase, so that it > never goes out of date. > There is some prior work / discussion: > # https://github.com/shalinmangar/solr-perf-tools (Shalin) > # https://github.com/chatman/solr-upgrade-tests/blob/master/BENCHMARKS.md > (Ishan/Vivek) > # SOLR-2646 & SOLR-9863 (Mark Miller) > # https://home.apache.org/~mikemccand/lucenebench/ (Mike McCandless) > # https://github.com/lucidworks/solr-scale-tk (Tim Potter) > There is support for building, starting, indexing/querying and stopping Solr > in some of these frameworks above. However, the benchmarks run are very > limited. Any of these can be a starting point, or a new framework can as well > be used. The motivation is to be able to cover every functionality of Solr > with a corresponding benchmark that is run every night. > Proposing this as a GSoC 2017 project. I'm willing to mentor, and I'm sure > [~shalinmangar] and [~markrmil...@gmail.com] would help here. -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org