-
Notifications
You must be signed in to change notification settings - Fork 6
/
main.py
166 lines (143 loc) · 9.03 KB
/
main.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
import argparse
import json
import logging as log
import os
from time import time as ts
import dateparser
import yaml
from typing import Dict
from szz.ag_szz import AGSZZ
from szz.aszz.a_szz import ASZZ
from szz.b_szz import BaseSZZ
from szz.util.check_requirements import check_requirements
from szz.dfszz.df_szz import DFSZZ
from szz.l_szz import LSZZ
from szz.ma_szz import MASZZ, DetectLineMoved
from szz.r_szz import RSZZ
from szz.ra_szz import RASZZ
from szz.pd_szz import PyDrillerSZZ
from szz.common.issue_date import parse_issue_date
from pathlib import Path
import random
log.basicConfig(level=log.INFO, format='%(asctime)s :: %(funcName)s - %(levelname)s :: %(message)s')
log.getLogger('pydriller').setLevel(log.WARNING)
def main(input_json: str, out_json: str, conf: Dict, repos_dir: str):
with open(input_json, 'r') as in_file:
bugfix_commits = json.loads(in_file.read())
tot = len(bugfix_commits)
for i, commit in enumerate(bugfix_commits):
bug_inducing_commits = set()
repo_name = commit['repo_name']
repo_url = f'https://test:[email protected]/{repo_name}.git' # using test:test as git login to skip private repos during clone
fix_commit = commit['fix_commit_hash']
log.info(f'{i + 1} of {tot}: {repo_name} {fix_commit}')
issue_date = None
if conf.get('issue_date_filter', None):
issue_date = parse_issue_date(commit)
szz_name = conf['szz_name']
if szz_name == 'b':
b_szz = BaseSZZ(repo_full_name=repo_name, repo_url=repo_url, repos_dir=repos_dir)
imp_files = b_szz.get_impacted_files(fix_commit_hash=fix_commit, file_ext_to_parse=conf.get('file_ext_to_parse'), only_deleted_lines=True)
bug_inducing_commits = b_szz.find_bic(fix_commit_hash=fix_commit,
impacted_files=imp_files,
issue_date_filter=conf.get('issue_date_filter'),
issue_date=issue_date)
elif szz_name == 'ag':
ag_szz = AGSZZ(repo_full_name=repo_name, repo_url=repo_url, repos_dir=repos_dir)
imp_files = ag_szz.get_impacted_files(fix_commit_hash=fix_commit, file_ext_to_parse=conf.get('file_ext_to_parse'), only_deleted_lines=True)
bug_inducing_commits = ag_szz.find_bic(fix_commit_hash=fix_commit,
impacted_files=imp_files,
max_change_size=conf.get('max_change_size'),
issue_date_filter=conf.get('issue_date_filter'),
issue_date=issue_date)
elif szz_name == 'ma':
ma_szz = MASZZ(repo_full_name=repo_name, repo_url=repo_url, repos_dir=repos_dir)
imp_files = ma_szz.get_impacted_files(fix_commit_hash=fix_commit, file_ext_to_parse=conf.get('file_ext_to_parse'), only_deleted_lines=True)
bug_inducing_commits = ma_szz.find_bic(fix_commit_hash=fix_commit,
impacted_files=imp_files,
max_change_size=conf.get('max_change_size'),
detect_move_from_other_files=DetectLineMoved(conf.get('detect_move_from_other_files')),
issue_date_filter=conf.get('issue_date_filter'),
issue_date=issue_date,
filter_revert_commits=conf.get('filter_revert_commits', False))
elif szz_name == 'r':
r_szz = RSZZ(repo_full_name=repo_name, repo_url=repo_url, repos_dir=repos_dir)
imp_files = r_szz.get_impacted_files(fix_commit_hash=fix_commit, file_ext_to_parse=conf.get('file_ext_to_parse'), only_deleted_lines=True)
bug_inducing_commits = r_szz.find_bic(fix_commit_hash=fix_commit,
impacted_files=imp_files,
max_change_size=conf.get('max_change_size'),
detect_move_from_other_files=DetectLineMoved(conf.get('detect_move_from_other_files')),
issue_date_filter=conf.get('issue_date_filter'),
issue_date=issue_date,
filter_revert_commits=conf.get('filter_revert_commits', False))
elif szz_name == 'l':
l_szz = LSZZ(repo_full_name=repo_name, repo_url=repo_url, repos_dir=repos_dir)
imp_files = l_szz.get_impacted_files(fix_commit_hash=fix_commit, file_ext_to_parse=conf.get('file_ext_to_parse'), only_deleted_lines=True)
bug_inducing_commits = l_szz.find_bic(fix_commit_hash=fix_commit,
impacted_files=imp_files,
max_change_size=conf.get('max_change_size'),
detect_move_from_other_files=DetectLineMoved(conf.get('detect_move_from_other_files')),
issue_date_filter=conf.get('issue_date_filter'),
issue_date=issue_date,
filter_revert_commits=conf.get('filter_revert_commits', False))
elif szz_name == 'ra':
ra_szz = RASZZ(repo_full_name=repo_name, repo_url=repo_url, repos_dir=repos_dir)
imp_files = ra_szz.get_impacted_files(fix_commit_hash=fix_commit, file_ext_to_parse=conf.get('file_ext_to_parse'), only_deleted_lines=True)
bug_inducing_commits = ra_szz.find_bic(fix_commit_hash=fix_commit,
impacted_files=imp_files,
max_change_size=conf.get('max_change_size'),
detect_move_from_other_files=DetectLineMoved(conf.get('detect_move_from_other_files')),
issue_date_filter=conf.get('issue_date_filter'),
issue_date=issue_date,
filter_revert_commits=conf.get('filter_revert_commits', False))
elif szz_name == 'pd':
pd_szz = PyDrillerSZZ(repo_full_name=repo_name, repo_url=repo_url, repos_dir=repos_dir)
imp_files = pd_szz.get_impacted_files(fix_commit_hash=fix_commit, file_ext_to_parse=conf.get('file_ext_to_parse'), only_deleted_lines=True)
bug_inducing_commits = pd_szz.find_bic(fix_commit_hash=fix_commit,
impacted_files=imp_files,
issue_date_filter=conf.get('issue_date_filter'),
issue_date=issue_date)
elif szz_name == 'a':
a_szz = ASZZ(repo_full_name=repo_name, repo_url=repo_url, repos_dir=repos_dir)
bug_inducing_commits = a_szz.start(fix_commit_hash=fix_commit, commit_issue_date=issue_date, **conf)
elif szz_name == 'df':
df_szz = DFSZZ(repo_full_name=repo_name, repo_url=repo_url, repos_dir=repos_dir)
bug_inducing_commits = df_szz.start(fix_commit_hash=fix_commit, commit_issue_date=issue_date, **conf)
else:
log.info(f'SZZ implementation not found: {szz_name}')
exit(-3)
log.info(f"result: {bug_inducing_commits}")
bugfix_commits[i]["inducing_commit_hash"] = [bic.hexsha for bic in bug_inducing_commits if bic]
if os.path.exists(out_json):
out_json = out_json.replace('.json', f'.{random.randint(1, 99)}.json')
with open(out_json, 'w') as out:
json.dump(bugfix_commits, out)
log.info(f"results saved in {out_json}")
log.info("+++ DONE +++")
if __name__ == "__main__":
check_requirements()
parser = argparse.ArgumentParser(description='USAGE: python main.py <bugfix_commits.json> <conf_file path> <repos_directory(optional)>\n* If <repos_directory> is not set, pyszz will download each repository')
parser.add_argument('input_json', type=str, help='/path/to/bug-fixes.json')
parser.add_argument('conf_file', type=str, help='/path/to/configuration-file.yml')
parser.add_argument('repos_dir', type=str, nargs='?', help='/path/to/repo-directory')
args = parser.parse_args()
if not os.path.isfile(args.input_json):
log.error('invalid input json')
exit(-2)
if not os.path.isfile(args.conf_file):
log.error('invalid conf file')
exit(-2)
with open(args.conf_file, 'r') as f:
conf = yaml.safe_load(f)
log.info(f"parsed conf yml '{args.conf_file}': {conf}")
szz_name = conf['szz_name']
out_dir = 'out'
if not os.path.isdir(out_dir):
os.makedirs(out_dir)
conf_file_name = Path(args.conf_file).name.split('.')[0]
out_json = os.path.join(out_dir, f'bic_{conf_file_name}_{int(ts())}.json')
if not szz_name:
log.error('The configuration file does not define the SZZ name. Please, fix.')
exit(-3)
log.info(f'Launching {szz_name}-szz')
main(args.input_json, out_json, conf, args.repos_dir)