Mercurial > public > src > rhodecode
diff pylons_app/lib/celerylib/tasks.py @ 537:2256c78afe53 celery
implemented basic autoupdating statistics fetched from database
author | Marcin Kuzminski <marcin@python-works.com> |
---|---|
date | Wed, 22 Sep 2010 04:30:36 +0200 |
parents | b12ea84fb906 |
children | fb0c3af6031b |
line wrap: on
line diff
--- a/pylons_app/lib/celerylib/tasks.py Tue Sep 21 15:36:46 2010 +0200 +++ b/pylons_app/lib/celerylib/tasks.py Wed Sep 22 04:30:36 2010 +0200 @@ -1,7 +1,6 @@ from celery.decorators import task from celery.task.sets import subtask from celeryconfig import PYLONS_CONFIG as config -from datetime import datetime, timedelta from pylons.i18n.translation import _ from pylons_app.lib.celerylib import run_task from pylons_app.lib.helpers import person @@ -10,7 +9,6 @@ from operator import itemgetter from vcs.backends.hg import MercurialRepository from time import mktime -import calendar import traceback import json @@ -83,94 +81,132 @@ return 'LockHeld' @task -def get_commits_stats(repo): +def get_commits_stats(repo_name, ts_min_y, ts_max_y): + author_key_cleaner = lambda k: person(k).replace('"', "") #for js data compatibilty + + from pylons_app.model.db import Statistics, Repository log = get_commits_stats.get_logger() - aggregate = OrderedDict() - overview_aggregate = OrderedDict() + commits_by_day_author_aggregate = {} + commits_by_day_aggregate = {} repos_path = get_hg_ui_settings()['paths_root_path'].replace('*', '') - repo = MercurialRepository(repos_path + repo) - #graph range - td = datetime.today() + timedelta(days=1) - y, m, d = td.year, td.month, td.day + repo = MercurialRepository(repos_path + repo_name) + + skip_date_limit = True + parse_limit = 500 #limit for single task changeset parsing + last_rev = 0 + last_cs = None + timegetter = itemgetter('time') + + sa = get_session() - ts_min_y = mktime((y - 1, (td - timedelta(days=calendar.mdays[m])).month, - d, 0, 0, 0, 0, 0, 0,)) - ts_min_m = mktime((y, (td - timedelta(days=calendar.mdays[m])).month, - d, 0, 0, 0, 0, 0, 0,)) + dbrepo = sa.query(Repository)\ + .filter(Repository.repo_name == repo_name).scalar() + cur_stats = sa.query(Statistics)\ + .filter(Statistics.repository == dbrepo).scalar() + if cur_stats: + last_rev = cur_stats.stat_on_revision - ts_max_y = mktime((y, m, d, 0, 0, 0, 0, 0, 0,)) - skip_date_limit = True + if last_rev == repo.revisions[-1]: + #pass silently without any work + return True - def author_key_cleaner(k): - k = person(k) - k = k.replace('"', "") #for js data compatibilty - return k - - for cs in repo[:200]:#added limit 200 until fix #29 is made + if cur_stats: + commits_by_day_aggregate = OrderedDict( + json.loads( + cur_stats.commit_activity_combined)) + commits_by_day_author_aggregate = json.loads(cur_stats.commit_activity) + + for cnt, rev in enumerate(repo.revisions[last_rev:]): + last_cs = cs = repo.get_changeset(rev) k = '%s-%s-%s' % (cs.date.timetuple()[0], cs.date.timetuple()[1], cs.date.timetuple()[2]) timetupple = [int(x) for x in k.split('-')] timetupple.extend([0 for _ in xrange(6)]) k = mktime(timetupple) - if aggregate.has_key(author_key_cleaner(cs.author)): - if aggregate[author_key_cleaner(cs.author)].has_key(k): - aggregate[author_key_cleaner(cs.author)][k]["commits"] += 1 - aggregate[author_key_cleaner(cs.author)][k]["added"] += len(cs.added) - aggregate[author_key_cleaner(cs.author)][k]["changed"] += len(cs.changed) - aggregate[author_key_cleaner(cs.author)][k]["removed"] += len(cs.removed) + if commits_by_day_author_aggregate.has_key(author_key_cleaner(cs.author)): + try: + l = [timegetter(x) for x in commits_by_day_author_aggregate\ + [author_key_cleaner(cs.author)]['data']] + time_pos = l.index(k) + except ValueError: + time_pos = False + + if time_pos >= 0 and time_pos is not False: + + datadict = commits_by_day_author_aggregate\ + [author_key_cleaner(cs.author)]['data'][time_pos] + + datadict["commits"] += 1 + datadict["added"] += len(cs.added) + datadict["changed"] += len(cs.changed) + datadict["removed"] += len(cs.removed) + #print datadict else: - #aggregate[author_key_cleaner(cs.author)].update(dates_range) + #print 'ELSE !!!!' if k >= ts_min_y and k <= ts_max_y or skip_date_limit: - aggregate[author_key_cleaner(cs.author)][k] = {} - aggregate[author_key_cleaner(cs.author)][k]["commits"] = 1 - aggregate[author_key_cleaner(cs.author)][k]["added"] = len(cs.added) - aggregate[author_key_cleaner(cs.author)][k]["changed"] = len(cs.changed) - aggregate[author_key_cleaner(cs.author)][k]["removed"] = len(cs.removed) + + datadict = {"time":k, + "commits":1, + "added":len(cs.added), + "changed":len(cs.changed), + "removed":len(cs.removed), + } + commits_by_day_author_aggregate\ + [author_key_cleaner(cs.author)]['data'].append(datadict) else: + #print k, 'nokey ADDING' if k >= ts_min_y and k <= ts_max_y or skip_date_limit: - aggregate[author_key_cleaner(cs.author)] = OrderedDict() - #aggregate[author_key_cleaner(cs.author)].update(dates_range) - aggregate[author_key_cleaner(cs.author)][k] = {} - aggregate[author_key_cleaner(cs.author)][k]["commits"] = 1 - aggregate[author_key_cleaner(cs.author)][k]["added"] = len(cs.added) - aggregate[author_key_cleaner(cs.author)][k]["changed"] = len(cs.changed) - aggregate[author_key_cleaner(cs.author)][k]["removed"] = len(cs.removed) + commits_by_day_author_aggregate[author_key_cleaner(cs.author)] = { + "label":author_key_cleaner(cs.author), + "data":[{"time":k, + "commits":1, + "added":len(cs.added), + "changed":len(cs.changed), + "removed":len(cs.removed), + }], + "schema":["commits"], + } - - if overview_aggregate.has_key(k): - overview_aggregate[k] += 1 +# #gather all data by day + if commits_by_day_aggregate.has_key(k): + commits_by_day_aggregate[k] += 1 else: - overview_aggregate[k] = 1 - + commits_by_day_aggregate[k] = 1 + + if cnt >= parse_limit: + #don't fetch to much data since we can freeze application + break + overview_data = [] - for k, v in overview_aggregate.items(): + for k, v in commits_by_day_aggregate.items(): overview_data.append([k, v]) overview_data = sorted(overview_data, key=itemgetter(0)) - data = {} - for author in aggregate: - commit_data = sorted([{"time":x, - "commits":aggregate[author][x]['commits'], - "added":aggregate[author][x]['added'], - "changed":aggregate[author][x]['changed'], - "removed":aggregate[author][x]['removed'], - } for x in aggregate[author]], - key=itemgetter('time')) - data[author] = {"label":author, - "data":commit_data, - "schema":["commits"] - } - - if not data: - data[author_key_cleaner(repo.contact)] = { + if not commits_by_day_author_aggregate: + commits_by_day_author_aggregate[author_key_cleaner(repo.contact)] = { "label":author_key_cleaner(repo.contact), "data":[0, 1], "schema":["commits"], } - - return (ts_min_m, ts_max_y, json.dumps(data), json.dumps(overview_data)) + + stats = cur_stats if cur_stats else Statistics() + stats.commit_activity = json.dumps(commits_by_day_author_aggregate) + stats.commit_activity_combined = json.dumps(overview_data) + stats.repository = dbrepo + stats.stat_on_revision = last_cs.revision + stats.languages = json.dumps({'_TOTAL_':0, '':0}) + + try: + sa.add(stats) + sa.commit() + except: + log.error(traceback.format_exc()) + sa.rollback() + return False + + return True @task def reset_user_password(user_email): @@ -184,10 +220,11 @@ user = sa.query(User).filter(User.email == user_email).scalar() new_passwd = auth.PasswordGenerator().gen_password(8, auth.PasswordGenerator.ALPHABETS_BIG_SMALL) - user.password = auth.get_crypt_password(new_passwd) - sa.add(user) - sa.commit() - log.info('change password for %s', user_email) + if user: + user.password = auth.get_crypt_password(new_passwd) + sa.add(user) + sa.commit() + log.info('change password for %s', user_email) if new_passwd is None: raise Exception('unable to generate new password')