report.py 18 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399
  1. #
  2. # Metrix++, Copyright 2009-2019, Metrix++ Project
  3. # Link: https://github.com/metrixplusplus/metrixplusplus
  4. #
  5. # This file is a part of Metrix++ Tool.
  6. #
  7. import logging
  8. import re
  9. import pytablewriter
  10. import mpp.api
  11. import mpp.utils
  12. import mpp.cout
  13. DIGIT_COUNT = 8
  14. class Plugin(mpp.api.Plugin, mpp.api.IConfigurable, mpp.api.IRunable):
  15. MODE_NEW = 0x01
  16. MODE_TREND = 0x03
  17. MODE_TOUCHED = 0x07
  18. MODE_ALL = 0x15
  19. def declare_configuration(self, parser):
  20. self.parser = parser
  21. parser.add_option("--output-dir", "--od", default='./metrixpp/',
  22. help="Set the output folder. [default: %default].")
  23. parser.add_option("--format", "--ft", default='txt', choices=['txt', 'md', 'html', 'rst', 'latex'],
  24. help="Format of the output data. "
  25. "Possible values are 'txt', 'md', 'html' or 'rst' [default: %default]")
  26. parser.add_option("--disable-suppressions", "--ds", action="store_true", default=False,
  27. help = "If not set (none), all suppressions are ignored"
  28. " and associated warnings are printed. [default: %default]")
  29. parser.add_option("--warn-mode", "--wm", default='all', choices=['new', 'trend', 'touched', 'all'],
  30. help="Defines the warnings mode. "
  31. "'all' - all warnings active, "
  32. "'new' - warnings for new regions/files only, "
  33. "'trend' - warnings for new regions/files and for bad trend of modified regions/files, "
  34. "'touched' - warnings for new and modified regions/files "
  35. "[default: %default]")
  36. parser.add_option("--min-limit", "--min", action="multiopt",
  37. help="A threshold per 'namespace:field' metric in order to select regions, "
  38. "which have got metric value less than the specified limit. "
  39. "This option can be specified multiple times, if it is necessary to apply several limits. "
  40. "Should be in the format: <namespace>:<field>:<limit-value>[:region_type[,region_type]], for example: "
  41. "'std.code.lines:comments:1', or 'std.code.lines:comments:1:function,class'. "
  42. "Region types is optional specifier, and if not defined the limit is applied to regions of all types.")
  43. parser.add_option("--max-limit", "--max", action="multiopt",
  44. help="A threshold per 'namespace:field' metric in order to select regions, "
  45. "which have got metric value more than the specified limit. "
  46. "This option can be specified multiple times, if it is necessary to apply several limits. "
  47. "Should be in the format: <namespace>:<field>:<limit-value>[:region_type[,region_type]], for example: "
  48. "'std.code.complexity:cyclomatic:7', or 'std.code.complexity:maxdepth:5:function'. "
  49. "Region types is optional specifier, and if not defined the limit is applied to regions of all types.")
  50. def configure(self, options):
  51. self.out_dir = options.__dict__['output_dir']
  52. self.out_format = options.__dict__['format']
  53. self.no_suppress = options.__dict__['disable_suppressions']
  54. if options.__dict__['warn_mode'] == 'new':
  55. self.mode = self.MODE_NEW
  56. elif options.__dict__['warn_mode'] == 'trend':
  57. self.mode = self.MODE_TREND
  58. elif options.__dict__['warn_mode'] == 'touched':
  59. self.mode = self.MODE_TOUCHED
  60. elif options.__dict__['warn_mode'] == 'all':
  61. self.mode = self.MODE_ALL
  62. if self.mode != self.MODE_ALL and options.__dict__['db_file_prev'] == None:
  63. self.parser.error("option --warn-mode: The mode '" + options.__dict__['warn_mode'] + "' requires '--db-file-prev' option set")
  64. class Limit(object):
  65. def __init__(self, limit_type, limit, namespace, field, db_filter, region_types, original):
  66. self.type = limit_type
  67. self.limit = limit
  68. self.namespace = namespace
  69. self.field = field
  70. self.filter = db_filter
  71. self.region_types = region_types
  72. self.original = original
  73. def __repr__(self):
  74. return "'{0}:{1}' {2} {3} [applied to '{4}' region type(s)]".format(
  75. self.namespace, self.field, self.filter[1], self.limit,
  76. mpp.api.Region.T().to_str(self.region_types))
  77. self.limits = []
  78. pattern = re.compile(r'''([^:]+)[:]([^:]+)[:]([-+]?[0-9]+(?:[.][0-9]+)?)(?:[:](.+))?''')
  79. if options.__dict__['max_limit'] != None:
  80. for each in options.__dict__['max_limit']:
  81. match = re.match(pattern, each)
  82. if match == None:
  83. self.parser.error("option --max-limit: Invalid format: " + each)
  84. region_types = 0x00
  85. if match.group(4) != None:
  86. for region_type in match.group(4).split(','):
  87. region_type = region_type.strip()
  88. group_id = mpp.api.Region.T().from_str(region_type)
  89. if group_id == None:
  90. self.parser.error(
  91. "option --max-limit: uknown region type (allowed: global, class, struct, namespace, function, interface, any): " + region_type)
  92. region_types |= group_id
  93. else:
  94. region_types = mpp.api.Region.T().ANY
  95. limit = Limit("max", float(match.group(3)), match.group(1), match.group(2),
  96. (match.group(2), '>', float(match.group(3))), region_types, each)
  97. self.limits.append(limit)
  98. if options.__dict__['min_limit'] != None:
  99. for each in options.__dict__['min_limit']:
  100. match = re.match(pattern, each)
  101. if match == None:
  102. self.parser.error("option --min-limit: Invalid format: " + each)
  103. region_types = 0x00
  104. if match.group(4) != None:
  105. for region_type in match.group(4).split(','):
  106. region_type = region_type.strip()
  107. group_id = mpp.api.Region.T().from_str(region_type)
  108. if group_id == None:
  109. self.parser.error(
  110. "option --max-limit: uknown region type (allowed: global, class, struct, namespace, function, interface, any): " + region_type)
  111. region_types |= group_id
  112. else:
  113. region_types = mpp.api.Region.T().ANY
  114. limit = Limit("min", float(match.group(3)), match.group(1), match.group(2),
  115. (match.group(2), '<', float(match.group(3))), region_types, each)
  116. self.limits.append(limit)
  117. def initialize(self):
  118. super(Plugin, self).initialize()
  119. db_loader = self.get_plugin('mpp.dbf').get_loader()
  120. self._verify_namespaces(db_loader.iterate_namespace_names())
  121. for each in db_loader.iterate_namespace_names():
  122. self._verify_fields(each, db_loader.get_namespace(each).iterate_field_names())
  123. def _verify_namespaces(self, valid_namespaces):
  124. valid = []
  125. for each in valid_namespaces:
  126. valid.append(each)
  127. for each in self.limits:
  128. if each.namespace not in valid:
  129. self.parser.error("option --{0}-limit: metric '{1}:{2}' is not available in the database file.".
  130. format(each.type, each.namespace, each.field))
  131. def _verify_fields(self, namespace, valid_fields):
  132. valid = []
  133. for each in valid_fields:
  134. valid.append(each)
  135. for each in self.limits:
  136. if each.namespace == namespace:
  137. if each.field not in valid:
  138. self.parser.error("option --{0}-limit: metric '{1}:{2}' is not available in the database file.".
  139. format(each.type, each.namespace, each.field))
  140. def iterate_limits(self):
  141. for each in self.limits:
  142. yield each
  143. def is_mode_matched(self, limit, value, diff, is_modified):
  144. if is_modified == None:
  145. # means new region, True in all modes
  146. return True
  147. if self.mode == self.MODE_ALL:
  148. return True
  149. if self.mode == self.MODE_TOUCHED and is_modified == True:
  150. return True
  151. if self.mode == self.MODE_TREND and is_modified == True:
  152. if limit < value and diff > 0:
  153. return True
  154. if limit > value and diff < 0:
  155. return True
  156. return False
  157. def run(self, args):
  158. return main(self, args)
  159. def run(self, args):
  160. return main(self, args)
  161. def loadSubdirs(loader, path, subdirs, subfiles):
  162. aggregated_data = loader.load_aggregated_data(path)
  163. if not aggregated_data:
  164. return subdirs, subfiles
  165. for subfile in aggregated_data.get_subfiles():
  166. subfiles.append(aggregated_data.path + "/" + subfile)
  167. for subdir in aggregated_data.get_subdirs():
  168. subdir = aggregated_data.path + "/" + subdir
  169. subdirs.append(subdir)
  170. subdirs, subfiles = loadSubdirs(loader, subdir, subdirs, subfiles)
  171. return subdirs, subfiles
  172. def main(plugin, args):
  173. exit_code = 0
  174. data = {"fileMetrixList" : [],
  175. "regionMetrixList" : [],
  176. "files" : []}
  177. loader_prev = plugin.get_plugin('mpp.dbf').get_loader_prev()
  178. loader = plugin.get_plugin('mpp.dbf').get_loader()
  179. paths = None
  180. if len(args) == 0:
  181. subdirs, paths = loadSubdirs(loader, ".", [], [])
  182. else:
  183. paths = args
  184. # Try to optimise iterative change scans
  185. modified_file_ids = None
  186. if plugin.mode != plugin.MODE_ALL:
  187. modified_file_ids = get_list_of_modified_files(loader, loader_prev)
  188. for path in paths:
  189. path = mpp.utils.preprocess_path(path)
  190. aggregated_data = loader.load_aggregated_data(path)
  191. file_data = loader.load_file_data(path)
  192. for key in aggregated_data.data:
  193. if not key in data["fileMetrixList"]:
  194. data["fileMetrixList"].append(key)
  195. file = {"path" : path,
  196. "file_id" : file_data.file_id,
  197. "regions" : [],
  198. "data" : aggregated_data.data}
  199. data["files"].append(file)
  200. for reg in file_data.iterate_regions():
  201. region = {"name" : reg.name,
  202. "region_id" : reg.region_id,
  203. "line_begin" : reg.line_begin,
  204. "data" : reg.get_data_tree()}
  205. file["regions"].append(region)
  206. for key in region["data"]:
  207. if not key in data["regionMetrixList"]:
  208. data["regionMetrixList"].append(key)
  209. for limit in plugin.iterate_limits():
  210. warns_count = 0
  211. logging.info("Applying limit: " + str(limit))
  212. filters = [limit.filter]
  213. if modified_file_ids != None:
  214. filters.append(('file_id', 'IN', modified_file_ids))
  215. sort_by = None
  216. limit_by = None
  217. selected_data = loader.load_selected_data(limit.namespace,
  218. fields=[limit.field],
  219. path=path,
  220. filters=filters,
  221. sort_by=sort_by,
  222. limit_by=limit_by)
  223. if selected_data == None:
  224. mpp.utils.report_bad_path(path)
  225. exit_code += 1
  226. continue
  227. is_modified = None
  228. diff = None
  229. file_data = loader.load_file_data(select_data.get_path())
  230. file_data_prev = loader_prev.load_file_data(select_data.get_path())
  231. if file_data_prev != None:
  232. if file_data.get_checksum() == file_data_prev.get_checksum():
  233. diff = 0
  234. is_modified = False
  235. else:
  236. matcher = mpp.utils.FileRegionsMatcher(file_data, file_data_prev)
  237. prev_id = matcher.get_prev_id(select_data.get_region().get_id())
  238. if matcher.is_matched(select_data.get_region().get_id()):
  239. if matcher.is_modified(select_data.get_region().get_id()):
  240. is_modified = True
  241. else:
  242. is_modified = False
  243. diff = mpp.api.DiffData(select_data,
  244. file_data_prev.get_region(prev_id)).get_data(limit.namespace,
  245. limit.field)
  246. if (plugin.is_mode_matched(limit.limit,
  247. select_data.get_data(limit.namespace, limit.field),
  248. diff,
  249. is_modified) == False):
  250. continue
  251. is_sup = is_metric_suppressed(limit.namespace, limit.field, loader, select_data)
  252. # add a warning flag to the metric
  253. for file in data["files"]:
  254. if file["path"] == select_data.get_path():
  255. for region in file["regions"]:
  256. if region["region_id"] == select_data.region_id:
  257. for metric in region["data"]:
  258. if metric == limit.namespace:
  259. region["data"][metric]["warning"] = True
  260. region["data"][metric]["suppressed"] = is_sup
  261. if is_sup == True and plugin.no_suppress == False:
  262. continue
  263. warns_count += 1
  264. exit_code += 1
  265. writer = pytablewriter.SpaceAlignedTableWriter()
  266. writer.headers = ["file"] + data["fileMetrixList"]
  267. matrix = [];
  268. for file in data["files"]:
  269. line = []
  270. line.append(file["path"])
  271. for metric in data["fileMetrixList"]:
  272. if metric in file["data"]:
  273. for value in file["data"][metric].values():
  274. values = []
  275. values.append(value["min"])
  276. values.append(value["max"])
  277. values.append(value["avg"])
  278. values.append(value["total"])
  279. values.append(value["count"])
  280. line.append(values)
  281. break
  282. else:
  283. line.append("---")
  284. matrix.append(line)
  285. writer.table_name = file["path"]
  286. writer.value_matrix = matrix
  287. writer.write_table()
  288. writer = pytablewriter.SpaceAlignedTableWriter()
  289. writer.headers = ["line", "name"] + data["regionMetrixList"]
  290. matrix = [];
  291. for file in data["files"]:
  292. for region in file["regions"]:
  293. line = []
  294. line.append(str(region["line_begin"]))
  295. line.append(region["name"])
  296. for metric in data["regionMetrixList"]:
  297. if metric in region["data"]:
  298. for value in region["data"][metric].values():
  299. line.append(str(value))
  300. break
  301. else:
  302. line.append("---")
  303. matrix.append(line)
  304. writer.table_name = file["path"]
  305. writer.value_matrix = matrix
  306. writer.write_table()
  307. return exit_code
  308. def get_list_of_modified_files(loader, loader_prev):
  309. logging.info("Identifying changed files...")
  310. old_files_map = {}
  311. for each in loader_prev.iterate_file_data():
  312. old_files_map[each.get_path()] = each.get_checksum()
  313. if len(old_files_map) == 0:
  314. return None
  315. modified_file_ids = []
  316. for each in loader.iterate_file_data():
  317. if len(modified_file_ids) > 1000: # If more than 1000 files changed, skip optimisation
  318. return None
  319. if (each.get_path() not in list(old_files_map.keys())) or old_files_map[
  320. each.get_path()] != each.get_checksum():
  321. modified_file_ids.append(str(each.get_id()))
  322. old_files_map = None
  323. if len(modified_file_ids) != 0:
  324. modified_file_ids = " , ".join(modified_file_ids)
  325. modified_file_ids = "(" + modified_file_ids + ")"
  326. return modified_file_ids
  327. return None
  328. def is_metric_suppressed(metric_namespace, metric_field, loader, select_data):
  329. data = loader.load_file_data(select_data.get_path())
  330. if select_data.get_region() != None:
  331. data = data.get_region(select_data.get_region().get_id())
  332. sup_data = data.get_data('std.suppress', 'list')
  333. else:
  334. sup_data = data.get_data('std.suppress.file', 'list')
  335. if sup_data != None and sup_data.find('[' + metric_namespace + ':' + metric_field + ']') != -1:
  336. return True
  337. return False