http://www.mediawiki.org/wiki/Special:Code/MediaWiki/89903
Revision: 89903 Author: rfaulk Date: 2011-06-12 01:45:00 +0000 (Sun, 12 Jun 2011) Log Message: ----------- Modification that emits the query name whenever a query is executed some new fields added to the squid log record table loader Modified Paths: -------------- trunk/fundraiser-statistics/fundraiser-scripts/classes/DataLoader.py Modified: trunk/fundraiser-statistics/fundraiser-scripts/classes/DataLoader.py =================================================================== --- trunk/fundraiser-statistics/fundraiser-scripts/classes/DataLoader.py 2011-06-12 01:42:26 UTC (rev 89902) +++ trunk/fundraiser-statistics/fundraiser-scripts/classes/DataLoader.py 2011-06-12 01:45:00 UTC (rev 89903) @@ -1,7 +1,7 @@ """ This module provides access to the datasource and enables querying. The class -atructure defined has DataLoader as the base which outlines the basic members +structure defined has DataLoader as the base which outlines the basic members and functionality. This interface is extended for interaction with specific sources of data. @@ -228,6 +228,7 @@ self.init_db() query_name = self.get_sql_filename_for_query() + print >> sys.stdout, 'Using query: ' + query_name metrics = Hlp.AutoVivification() times = Hlp.AutoVivification() @@ -255,7 +256,7 @@ metric_index = QD.get_metric_index(query_name, metric_name) time_index = QD.get_time_index(query_name) - + """ Compose the data for each separate donor pipeline artifact """ try: # err_msg = sql_stmnt @@ -324,6 +325,11 @@ for key in metrics.keys(): metrics_new = list() for i in range(len(metrics[key])): + + """ Change null values to 0 """ + if metrics[key][i] == None or metrics[key][i] == 'NULL': + metrics[key][i] = 0 + metrics_new.append(float(metrics[key][i])) metrics[key] = metrics_new @@ -343,6 +349,7 @@ def combine_rows(self): query_name = self.get_sql_filename_for_query() + print >> sys.stdout, 'Using query: ' + query_name col_types = self._data_handler_.get_col_types(self._query_type_) key_index = QD.get_key_index(query_name) @@ -361,7 +368,7 @@ key = row[key_index] try: - data_dict[key] == None + data_dict[key] == None # check for a Index Error except KeyError as e: data_dict[key] = dict() @@ -370,6 +377,10 @@ col_type = col_types[i] field = row[i] + """ Change null values to 0 """ + if field == None or field == 'NULL': + field = 0.0 + if col_type == self._data_handler_._COLTYPE_RATE_: try: @@ -737,10 +748,11 @@ end_time = params['end_time'] query_name = self.get_sql_filename_for_query() + print >> sys.stdout, 'Using query: ' + query_name """ Load the SQL File & Format """ filename = self._sql_path_+ query_name + '.sql' - sql_stmnt = Hlp.read_sql(filename) + sql_stmnt = Hlp.read_sql(filename) sql_stmnt = QD.format_query(query_name, sql_stmnt, [start_time, end_time]) """ Get Indexes into Query """ @@ -789,6 +801,7 @@ end_time = params['end_time'] query_name = self.get_sql_filename_for_query() + print >> sys.stdout, 'Using query: ' + query_name """ Load the SQL File & Format """ filename = self._sql_path_+ query_name + '.sql' @@ -1250,7 +1263,8 @@ def get_all_rows_unique_start_time(self): select_stmnt = 'select type, log_copy_time, start_time, end_time, log_completion_pct, total_rows from ' \ - '(select max(log_copy_time) as max_copy_time from squid_log_record group by start_time) as temp join squid_log_record on max_copy_time = squid_log_record.log_copy_time' + '(select type as temp_type, max(log_copy_time) as max_copy_time from squid_log_record group by type, start_time) as temp join ' \ + 'squid_log_record on (max_copy_time = squid_log_record.log_copy_time and temp_type = type)' self.init_db() @@ -1260,7 +1274,7 @@ except: self._db_.rollback() self.close_db() - print >> sys.stderr, 'Could not execute: ' + update_stmnt + print >> sys.stderr, 'Could not execute: ' + select_stmnt return -1 else: self.close_db() _______________________________________________ MediaWiki-CVS mailing list MediaWiki-CVS@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/mediawiki-cvs