view.py 36 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702
  1. #
  2. # Metrix++, Copyright 2009-2019, Metrix++ Project
  3. # Link: https://github.com/metrixplusplus/metrixplusplus
  4. #
  5. # This file is a part of Metrix++ Tool.
  6. #
  7. import logging
  8. import sys
  9. from metrixpp.mpp import api
  10. from metrixpp.mpp import utils
  11. from metrixpp.mpp import cout
  12. from metrixpp.mpp import promout
  13. DIGIT_COUNT = 8
  14. class Plugin(api.Plugin, api.IConfigurable, api.IRunable):
  15. MODE_NEW = 0x01
  16. MODE_TOUCHED = 0x03
  17. MODE_ALL = 0x07
  18. def declare_configuration(self, parser):
  19. self.parser = parser
  20. parser.add_option("--format", "--ft", default='txt', choices=['txt', 'xml', 'python', 'prometheus'],
  21. help="Format of the output data. "
  22. "Possible values are 'xml', 'txt', 'python' or 'prometheus' [default: %default]")
  23. parser.add_option("--nest-regions", "--nr", action="store_true", default=False,
  24. help="If the option is set (True), data for regions is exported in the form of a tree. "
  25. "Otherwise, all regions are exported in plain list. [default: %default]")
  26. parser.add_option("--max-distribution-rows", "--mdr", type=int, default=20,
  27. help="Maximum number of rows in distribution tables. "
  28. "If it is set to 0, the tool does not optimize the size of distribution tables [default: %default]")
  29. parser.add_option("--scope-mode", "--sm", default='all', choices=['new', 'touched', 'all'],
  30. help="Defines the analysis scope mode. "
  31. "'all' - all available regions and files are taken into account, "
  32. "'new' - only new regions and files are taken into account, "
  33. "'touched' - only new and modified regions and files are taken into account. "
  34. "Modes 'new' and 'touched' may require more time for processing than mode 'all' "
  35. "[default: %default]")
  36. def configure(self, options):
  37. self.out_format = options.__dict__['format']
  38. self.nest_regions = options.__dict__['nest_regions']
  39. self.dist_columns = options.__dict__['max_distribution_rows']
  40. if options.__dict__['scope_mode'] == 'new':
  41. self.mode = self.MODE_NEW
  42. elif options.__dict__['scope_mode'] == 'touched':
  43. self.mode = self.MODE_TOUCHED
  44. elif options.__dict__['scope_mode'] == 'all':
  45. self.mode = self.MODE_ALL
  46. if self.mode != self.MODE_ALL and options.__dict__['db_file_prev'] == None:
  47. self.parser.error("option --scope-mode: The mode '" + options.__dict__['scope_mode'] + "' requires '--db-file-prev' option set")
  48. def run(self, args):
  49. loader_prev = self.get_plugin('metrixpp.mpp.dbf').get_loader_prev()
  50. loader = self.get_plugin('metrixpp.mpp.dbf').get_loader()
  51. paths = None
  52. if len(args) == 0:
  53. paths = [""]
  54. else:
  55. paths = args
  56. (result, exit_code) = export_to_str(self.out_format,
  57. paths,
  58. loader,
  59. loader_prev,
  60. self.nest_regions,
  61. self.dist_columns,
  62. self.mode)
  63. print(result)
  64. return exit_code
  65. def export_to_str(out_format, paths, loader, loader_prev, nest_regions, dist_columns, mode):
  66. exit_code = 0
  67. result = ""
  68. if out_format == 'xml':
  69. result += "<view>\n"
  70. elif out_format == 'python':
  71. result += "{'view': ["
  72. for (ind, path) in enumerate(paths):
  73. path = utils.preprocess_path(path)
  74. aggregated_data, aggregated_data_prev = load_aggregated_data_with_mode(loader, loader_prev, path , mode)
  75. aggregated_data_tree = {}
  76. subdirs = []
  77. subfiles = []
  78. if aggregated_data != None:
  79. aggregated_data_tree = aggregated_data.get_data_tree()
  80. subdirs = sorted(aggregated_data.get_subdirs())
  81. subfiles = sorted(aggregated_data.get_subfiles())
  82. else:
  83. utils.report_bad_path(path)
  84. exit_code += 1
  85. aggregated_data_tree = append_suppressions(path, aggregated_data_tree, loader, mode)
  86. if aggregated_data_prev != None:
  87. aggregated_data_prev_tree = aggregated_data_prev.get_data_tree()
  88. aggregated_data_prev_tree = append_suppressions(path, aggregated_data_prev_tree, loader_prev, mode)
  89. aggregated_data_tree = append_diff(aggregated_data_tree,
  90. aggregated_data_prev_tree)
  91. aggregated_data_tree = compress_dist(aggregated_data_tree, dist_columns)
  92. file_data = loader.load_file_data(path)
  93. file_data_tree = {}
  94. if file_data != None:
  95. file_data_tree = file_data.get_data_tree()
  96. file_data_prev = loader_prev.load_file_data(path)
  97. append_regions(file_data_tree, file_data, file_data_prev, nest_regions)
  98. data = {"info": {"path": path, "id": ind + 1},
  99. "aggregated-data": aggregated_data_tree,
  100. "file-data": file_data_tree,
  101. "subdirs": subdirs,
  102. "subfiles": subfiles}
  103. if out_format == 'txt':
  104. cout_txt(data, loader)
  105. elif out_format == 'prometheus':
  106. cout_prom(data, loader)
  107. elif out_format == 'xml':
  108. result += utils.serialize_to_xml(data, root_name = "data", digitCount = DIGIT_COUNT) + "\n"
  109. elif out_format == 'python':
  110. postfix = ""
  111. if ind < len(paths) - 1:
  112. postfix = ", "
  113. result += utils.serialize_to_python(data, root_name = "data") + postfix
  114. if out_format == 'xml':
  115. result += "</view>"
  116. elif out_format == 'python':
  117. result += "]}"
  118. return (result, exit_code)
  119. def load_aggregated_data_with_mode(loader, loader_prev, path, mode):
  120. if mode == Plugin.MODE_ALL:
  121. aggregated_data = loader.load_aggregated_data(path)
  122. aggregated_data_prev = loader_prev.load_aggregated_data(path)
  123. else:
  124. assert(mode == Plugin.MODE_NEW or mode == Plugin.MODE_TOUCHED)
  125. class AggregatedFilteredData(api.AggregatedData):
  126. def __init__(self, loader, path):
  127. super(AggregatedFilteredData, self).__init__(loader, path)
  128. self.in_processing_mode = True
  129. for name in loader.iterate_namespace_names():
  130. namespace = loader.get_namespace(name)
  131. for field in namespace.iterate_field_names():
  132. if namespace.get_field_python_type(field) == str:
  133. # skip string type fields
  134. continue
  135. self.set_data(name, field, {
  136. 'count': 0,
  137. 'nonzero': namespace.is_field_non_zero(field),
  138. 'min': None,
  139. 'max': None,
  140. 'total': 0.0,
  141. 'avg': None,
  142. 'distribution-bars': {},
  143. 'sup': 0
  144. })
  145. def get_data_tree(self, namespaces=None):
  146. self.in_processing_mode = False
  147. # need to convert distribution map to a list and calculate average
  148. for name in loader.iterate_namespace_names():
  149. namespace = loader.get_namespace(name)
  150. for field in namespace.iterate_field_names():
  151. if namespace.get_field_python_type(field) == str:
  152. # skip string type fields
  153. continue
  154. data = self.get_data(name, field)
  155. bars_list = []
  156. for metric_value in sorted(data['distribution-bars'].keys()):
  157. bars_list.append({'metric': metric_value,
  158. 'count': data['distribution-bars'][metric_value],
  159. 'ratio': ((float(data['distribution-bars'][metric_value]) /
  160. float(data['count'])))})
  161. data['distribution-bars'] = bars_list
  162. if data['count'] != 0:
  163. data['avg'] = float(data['total']) / float(data['count'])
  164. self.set_data(name, field, data)
  165. return super(AggregatedFilteredData, self).get_data_tree(namespaces=namespaces)
  166. def _append_data(self, orig_data):
  167. # flag to protect ourselves from getting incomplete data
  168. # the workflow in this tool: append data first and after get it using get_data_tree()
  169. assert(self.in_processing_mode == True)
  170. sup_data = orig_data.get_data('std.suppress', 'list')
  171. data = orig_data.get_data_tree()
  172. for namespace in list(data.keys()):
  173. for field in list(data[namespace].keys()):
  174. aggr_data = self.get_data(namespace, field)
  175. metric_value = data[namespace][field]
  176. if isinstance(metric_value, str):
  177. # skip string type fields
  178. continue
  179. if aggr_data['min'] == None or aggr_data['min'] > metric_value:
  180. aggr_data['min'] = metric_value
  181. if aggr_data['max'] == None or aggr_data['max'] < metric_value:
  182. aggr_data['max'] = metric_value
  183. aggr_data['count'] += 1
  184. aggr_data['total'] += metric_value
  185. # average is calculated later on get_data_tree
  186. if metric_value not in list(aggr_data['distribution-bars'].keys()):
  187. aggr_data['distribution-bars'][metric_value] = 0
  188. aggr_data['distribution-bars'][metric_value] += 1
  189. if sup_data != None:
  190. if sup_data.find('[{0}:{1}]'.format(namespace, field)) != -1:
  191. aggr_data['sup'] += 1
  192. self.set_data(namespace, field, aggr_data)
  193. def _append_file_data(self, file_data):
  194. self._append_data(file_data)
  195. for region in file_data.iterate_regions():
  196. self._append_data(region)
  197. result = AggregatedFilteredData(loader, path)
  198. result_prev = AggregatedFilteredData(loader_prev, path)
  199. prev_file_ids = set()
  200. file_data_iterator = loader.iterate_file_data(path)
  201. if file_data_iterator != None:
  202. for file_data in file_data_iterator:
  203. file_data_prev = loader_prev.load_file_data(file_data.get_path())
  204. if file_data_prev != None:
  205. prev_file_ids.add(file_data_prev.get_id())
  206. if (file_data_prev == None and (mode == Plugin.MODE_NEW or mode == Plugin.MODE_TOUCHED)):
  207. # new file and required mode matched
  208. logging.info("Processing: " + file_data.get_path() + " [new]")
  209. result._append_file_data(file_data)
  210. elif (file_data.get_checksum() != file_data_prev.get_checksum()):
  211. # modified file and required mode matched
  212. logging.info("Processing: " + file_data.get_path() + " [modified]")
  213. # append file data without appending regions...
  214. if (mode == Plugin.MODE_TOUCHED):
  215. # if required mode matched
  216. result._append_data(file_data)
  217. result_prev._append_data(file_data_prev)
  218. # process regions separately
  219. matcher = utils.FileRegionsMatcher(file_data, file_data_prev)
  220. prev_reg_ids = set()
  221. for region in file_data.iterate_regions():
  222. prev_id = matcher.get_prev_id(region.get_id())
  223. if prev_id != None:
  224. prev_reg_ids.add(prev_id)
  225. if (matcher.is_matched(region.get_id()) == False and
  226. (mode == Plugin.MODE_NEW or mode == Plugin.MODE_TOUCHED)):
  227. # new region
  228. logging.debug("Processing region: " + region.get_name() + " [new]")
  229. result._append_data(region)
  230. elif matcher.is_modified(region.get_id()) and mode == Plugin.MODE_TOUCHED:
  231. # modified region
  232. logging.debug("Processing region: " + region.get_name() + " [modified]")
  233. result._append_data(region)
  234. result_prev._append_data(file_data_prev.get_region(prev_id))
  235. if mode == Plugin.MODE_TOUCHED:
  236. for region_prev in file_data_prev.iterate_regions():
  237. if region_prev.get_id() not in prev_reg_ids:
  238. # deleted region
  239. logging.debug("Processing region: " + region_prev.get_name() + " [deleted]")
  240. result_prev._append_data(region_prev)
  241. if mode == Plugin.MODE_TOUCHED:
  242. file_data_prev_iterator = loader_prev.iterate_file_data(path)
  243. if file_data_prev_iterator != None:
  244. for file_data_prev in file_data_prev_iterator:
  245. if file_data_prev.get_id() not in prev_file_ids:
  246. # deleted file and required mode matched
  247. logging.info("Processing: " + file_data_prev.get_path() + " [deleted]")
  248. result_prev._append_file_data(file_data_prev)
  249. return (result, result_prev)
  250. return (aggregated_data, aggregated_data_prev)
  251. def append_regions(file_data_tree, file_data, file_data_prev, nest_regions):
  252. regions_matcher = None
  253. if file_data_prev != None:
  254. file_data_tree = append_diff(file_data_tree,
  255. file_data_prev.get_data_tree())
  256. regions_matcher = utils.FileRegionsMatcher(file_data, file_data_prev)
  257. if nest_regions == False:
  258. regions = []
  259. for region in file_data.iterate_regions():
  260. region_data_tree = region.get_data_tree()
  261. is_modified = None
  262. if regions_matcher != None and regions_matcher.is_matched(region.get_id()):
  263. region_data_prev = file_data_prev.get_region(regions_matcher.get_prev_id(region.get_id()))
  264. region_data_tree = append_diff(region_data_tree,
  265. region_data_prev.get_data_tree())
  266. is_modified = regions_matcher.is_modified(region.get_id())
  267. regions.append({"info": {"name" : region.name,
  268. 'type': api.Region.T().to_str(region.get_type()),
  269. 'modified': is_modified,
  270. 'cursor' : region.cursor,
  271. 'line_begin': region.line_begin,
  272. 'line_end': region.line_end,
  273. 'offset_begin': region.begin,
  274. 'offset_end': region.end},
  275. "data": region_data_tree})
  276. file_data_tree['regions'] = regions
  277. else:
  278. def append_rec(region_id, file_data_tree, file_data, file_data_prev):
  279. region = file_data.get_region(region_id)
  280. region_data_tree = region.get_data_tree()
  281. is_modified = None
  282. if regions_matcher != None and regions_matcher.is_matched(region.get_id()):
  283. region_data_prev = file_data_prev.get_region(regions_matcher.get_prev_id(region.get_id()))
  284. region_data_tree = append_diff(region_data_tree,
  285. region_data_prev.get_data_tree())
  286. is_modified = regions_matcher.is_modified(region.get_id())
  287. result = {"info": {"name" : region.name,
  288. 'type' : api.Region.T().to_str(region.get_type()),
  289. 'modified': is_modified,
  290. 'cursor' : region.cursor,
  291. 'line_begin': region.line_begin,
  292. 'line_end': region.line_end,
  293. 'offset_begin': region.begin,
  294. 'offset_end': region.end},
  295. "data": region_data_tree,
  296. "subregions": []}
  297. for sub_id in file_data.get_region(region_id).iterate_subregion_ids():
  298. result['subregions'].append(append_rec(sub_id, file_data_tree, file_data, file_data_prev))
  299. return result
  300. file_data_tree['regions'] = []
  301. file_data_tree['regions'].append(append_rec(1, file_data_tree, file_data, file_data_prev))
  302. def append_diff(main_tree, prev_tree):
  303. assert(main_tree != None)
  304. assert(prev_tree != None)
  305. for name in list(main_tree.keys()):
  306. if name not in list(prev_tree.keys()):
  307. continue
  308. for field in list(main_tree[name].keys()):
  309. if field not in list(prev_tree[name].keys()):
  310. continue
  311. if isinstance(main_tree[name][field], dict) and isinstance(prev_tree[name][field], dict):
  312. diff = {}
  313. for key in list(main_tree[name][field].keys()):
  314. if key not in list(prev_tree[name][field].keys()):
  315. continue
  316. main_val = main_tree[name][field][key]
  317. prev_val = prev_tree[name][field][key]
  318. if main_val == None:
  319. main_val = 0
  320. if prev_val == None:
  321. prev_val = 0
  322. if isinstance(main_val, list) and isinstance(prev_val, list):
  323. main_tree[name][field][key] = append_diff_list(main_val, prev_val)
  324. else:
  325. diff[key] = main_val - prev_val
  326. main_tree[name][field]['__diff__'] = diff
  327. elif (not isinstance(main_tree[name][field], dict)) and (not isinstance(prev_tree[name][field], dict)):
  328. if '__diff__' not in main_tree[name]:
  329. main_tree[name]['__diff__'] = {}
  330. main_tree[name]['__diff__'][field] = main_tree[name][field] - prev_tree[name][field]
  331. return main_tree
  332. def append_diff_list(main_list, prev_list):
  333. merged_list = {}
  334. for bar in main_list:
  335. merged_list[bar['metric']] = {'count': bar['count'], '__diff__':bar['count'], 'ratio': bar['ratio']}
  336. for bar in prev_list:
  337. if bar['metric'] in list(merged_list.keys()):
  338. merged_list[bar['metric']]['__diff__'] = \
  339. merged_list[bar['metric']]['count'] - bar['count']
  340. else:
  341. merged_list[bar['metric']] = {'count': 0, '__diff__':-bar['count'], 'ratio': 0}
  342. result = []
  343. for metric in sorted(merged_list.keys()):
  344. result.append({'metric':metric,
  345. 'count':merged_list[metric]['count'],
  346. 'ratio':merged_list[metric]['ratio'],
  347. '__diff__':merged_list[metric]['__diff__']})
  348. return result
  349. def append_suppressions(path, data, loader, mode):
  350. if mode == Plugin.MODE_ALL:
  351. # in other modes, suppressions are appended during data loading
  352. for namespace in list(data.keys()):
  353. for field in list(data[namespace].keys()):
  354. selected_data = loader.load_selected_data('std.suppress',
  355. fields = ['list'],
  356. path=path,
  357. filters = [('list', 'LIKE', '%[{0}:{1}]%'.format(namespace, field))])
  358. if selected_data == None:
  359. data[namespace][field]['sup'] = 0
  360. else:
  361. count = 0
  362. for each in selected_data:
  363. each = each # used
  364. count += 1
  365. data[namespace][field]['sup'] = count
  366. return data
  367. def compress_dist(data, columns):
  368. if columns == 0:
  369. return data
  370. for namespace in list(data.keys()):
  371. for field in list(data[namespace].keys()):
  372. metric_data = data[namespace][field]
  373. distr = metric_data['distribution-bars']
  374. columns = float(columns) # to trigger floating calculations
  375. if metric_data['count'] == 0:
  376. continue
  377. new_dist = []
  378. remaining_count = metric_data['count']
  379. next_consume = None
  380. next_bar = None
  381. max_count = -sys.maxsize - 1
  382. min_count = sys.maxsize
  383. sum_ratio = 0
  384. for (ind, bar) in enumerate(distr):
  385. if next_bar == None:
  386. # start new bar
  387. next_bar = {'count': bar['count'],
  388. 'ratio': bar['ratio'],
  389. 'metric_s': bar['metric'],
  390. 'metric_f': bar['metric']}
  391. if '__diff__' in list(bar.keys()):
  392. next_bar['__diff__'] = bar['__diff__']
  393. next_consume = int(remaining_count/ (columns - len(new_dist)))
  394. else:
  395. # merge to existing bar
  396. next_bar['count'] += bar['count']
  397. next_bar['ratio'] += bar['ratio']
  398. next_bar['metric_f'] = bar['metric']
  399. if '__diff__' in list(bar.keys()):
  400. next_bar['__diff__'] += bar['__diff__']
  401. next_consume -= bar['count']
  402. if (next_consume <= 0 # consumed enough
  403. or (ind + 1) == len(distr)): # or the last bar
  404. # append to new distribution
  405. if isinstance(next_bar['metric_s'], float):
  406. next_bar['metric_s'] = "{0:.4f}".format(next_bar['metric_s'])
  407. next_bar['metric_f'] = "{0:.4f}".format(next_bar['metric_f'])
  408. else:
  409. next_bar['metric_s'] = str(next_bar['metric_s'])
  410. next_bar['metric_f'] = str(next_bar['metric_f'])
  411. if next_bar['metric_s'] == next_bar['metric_f']:
  412. next_bar['metric'] = next_bar['metric_s']
  413. else:
  414. next_bar['metric'] = next_bar['metric_s'] + "-" + next_bar['metric_f']
  415. del next_bar['metric_s']
  416. del next_bar['metric_f']
  417. new_dist.append(next_bar)
  418. sum_ratio += next_bar['ratio']
  419. if max_count < next_bar['count']:
  420. max_count = next_bar['count']
  421. if min_count > next_bar['count'] and next_bar['count'] != 0:
  422. min_count = next_bar['count']
  423. remaining_count -= next_bar['count']
  424. next_bar = None
  425. # check that consumed all
  426. assert((ind + 1) != len(distr) or remaining_count == 0)
  427. if remaining_count == 0:
  428. break
  429. if (float(max_count - min_count) / metric_data['count'] < 0.05 and
  430. metric_data['count'] > columns and
  431. len(new_dist) > 1):
  432. # trick here:
  433. # if all bars are even in the new distribution AND
  434. # there are many items in the distribution (> max distribution rows),
  435. # it is better to do linear compression instead
  436. new_dist = []
  437. step = int(round(float(metric_data['max'] - metric_data['min']) / columns))
  438. next_end_limit = metric_data['min']
  439. next_bar = None
  440. for (ind, bar) in enumerate(distr):
  441. if next_bar == None:
  442. # start new bar
  443. next_bar = {'count': bar['count'],
  444. 'ratio': bar['ratio'],
  445. 'metric_s': next_end_limit,
  446. 'metric_f': bar['metric']}
  447. if '__diff__' in list(bar.keys()):
  448. next_bar['__diff__'] = bar['__diff__']
  449. next_end_limit += step
  450. else:
  451. # merge to existing bar
  452. next_bar['count'] += bar['count']
  453. next_bar['ratio'] += bar['ratio']
  454. next_bar['metric_f'] = bar['metric']
  455. if '__diff__' in list(bar.keys()):
  456. next_bar['__diff__'] += bar['__diff__']
  457. if (next_bar['metric_f'] >= next_end_limit # consumed enough
  458. or (ind + 1) == len(distr)): # or the last bar
  459. if (ind + 1) != len(distr):
  460. next_bar['metric_f'] = next_end_limit
  461. # append to new distribution
  462. if isinstance(next_bar['metric_s'], float):
  463. next_bar['metric_s'] = "{0:.4f}".format(next_bar['metric_s'])
  464. next_bar['metric_f'] = "{0:.4f}".format(next_bar['metric_f'])
  465. else:
  466. next_bar['metric_s'] = str(next_bar['metric_s'])
  467. next_bar['metric_f'] = str(next_bar['metric_f'])
  468. next_bar['metric'] = next_bar['metric_s'] + "-" + next_bar['metric_f']
  469. del next_bar['metric_s']
  470. del next_bar['metric_f']
  471. new_dist.append(next_bar)
  472. next_bar = None
  473. data[namespace][field]['distribution-bars'] = new_dist
  474. return data
  475. def cout_txt_regions(path, regions, indent = 0):
  476. for region in regions:
  477. details = [
  478. ('Region name', region['info']['name']),
  479. ('Region type', region['info']['type']),
  480. ('Offsets', str(region['info']['offset_begin']) + "-" + str(region['info']['offset_end'])),
  481. ('Line numbers', str(region['info']['line_begin']) + "-" + str(region['info']['line_end'])),
  482. ('Modified', str(region['info']['modified']))
  483. ]
  484. for namespace in sorted(list(region['data'].keys())):
  485. diff_data = {}
  486. if '__diff__' in list(region['data'][namespace].keys()):
  487. diff_data = region['data'][namespace]['__diff__']
  488. for field in sorted(list(region['data'][namespace].keys())):
  489. diff_str = ""
  490. if field == '__diff__':
  491. continue
  492. if field in list(diff_data.keys()):
  493. diff_str = " [" + ("+" if diff_data[field] >= 0 else "") + "{0:.3g}".format(diff_data[field]) + "]"
  494. details.append((namespace + ":" + field, str(region['data'][namespace][field]) + diff_str))
  495. cout.notify(path,
  496. region['info']['cursor'],
  497. cout.SEVERITY_INFO,
  498. "Metrics per '" + region['info']['name']+ "' region",
  499. details,
  500. indent=indent)
  501. if 'subregions' in list(region.keys()):
  502. cout_txt_regions(path, region['subregions'], indent=indent+1)
  503. def cout_txt(data, loader):
  504. details = []
  505. for key in list(data['file-data'].keys()):
  506. if key == 'regions':
  507. cout_txt_regions(data['info']['path'], data['file-data'][key])
  508. else:
  509. namespace = key
  510. diff_data = {}
  511. if '__diff__' in list(data['file-data'][namespace].keys()):
  512. diff_data = data['file-data'][namespace]['__diff__']
  513. for field in sorted(list(data['file-data'][namespace].keys())):
  514. diff_str = ""
  515. if field == '__diff__':
  516. continue
  517. if field in list(diff_data.keys()):
  518. diff_str = " [" + ("+" if diff_data[field] >= 0 else "") + str(diff_data[field]) + "]"
  519. details.append((namespace + ":" + field, str(data['file-data'][namespace][field]) + diff_str))
  520. if len(details) > 0:
  521. cout.notify(data['info']['path'],
  522. 0,
  523. cout.SEVERITY_INFO,
  524. "Metrics per file",
  525. details)
  526. attr_map = {'total': 'Total',
  527. 'avg': 'Average',
  528. 'min': 'Minimum',
  529. 'max': 'Maximum',
  530. }
  531. for namespace in sorted(list(data['aggregated-data'].keys())):
  532. for field in sorted(list(data['aggregated-data'][namespace].keys())):
  533. details = []
  534. diff_data = {}
  535. if '__diff__' in list(data['aggregated-data'][namespace][field].keys()):
  536. diff_data = data['aggregated-data'][namespace][field]['__diff__']
  537. for attr in ['avg', 'min', 'max', 'total']:
  538. diff_str = ""
  539. if attr in list(diff_data.keys()):
  540. if isinstance(diff_data[attr], float):
  541. diff_str = " [" + ("+" if diff_data[attr] >= 0 else "") + str(round(diff_data[attr], DIGIT_COUNT)) + "]"
  542. else:
  543. diff_str = " [" + ("+" if diff_data[attr] >= 0 else "") + str(diff_data[attr]) + "]"
  544. if attr == 'avg' and data['aggregated-data'][namespace][field]['nonzero'] == True:
  545. diff_str += " (excluding zero metric values)"
  546. if isinstance(data['aggregated-data'][namespace][field][attr], float):
  547. # round the data to reach same results on platforms with different precision
  548. details.append((attr_map[attr], str(round(data['aggregated-data'][namespace][field][attr], DIGIT_COUNT)) + diff_str))
  549. else:
  550. details.append((attr_map[attr], str(data['aggregated-data'][namespace][field][attr]) + diff_str))
  551. measured = data['aggregated-data'][namespace][field]['count']
  552. if 'count' in list(diff_data.keys()):
  553. diff_str = ' [{0:{1}}]'.format(diff_data['count'], '+' if diff_data['count'] >= 0 else '')
  554. sup_diff_str = ""
  555. if 'sup' in list(diff_data.keys()):
  556. sup_diff_str = ' [{0:{1}}]'.format(diff_data['sup'], '+' if diff_data['sup'] >= 0 else '')
  557. elem_name = 'regions'
  558. if loader.get_namespace(namespace).are_regions_supported() == False:
  559. elem_name = 'files'
  560. details.append(('Distribution',
  561. '{0}{1} {2} in total (including {3}{4} suppressed)'.format(measured,
  562. diff_str,
  563. elem_name,
  564. data['aggregated-data'][namespace][field]['sup'],
  565. sup_diff_str)))
  566. details.append((' Metric value', 'Ratio : R-sum : Number of ' + elem_name))
  567. count_str_len = len(str(measured))
  568. sum_ratio = 0
  569. for bar in data['aggregated-data'][namespace][field]['distribution-bars']:
  570. sum_ratio += bar['ratio']
  571. diff_str = ""
  572. if '__diff__' in list(bar.keys()):
  573. if bar['__diff__'] >= 0:
  574. diff_str = ' [+{0:<{1}}]'.format(bar['__diff__'], count_str_len)
  575. else:
  576. diff_str = ' [{0:<{1}}]'.format(bar['__diff__'], count_str_len+1)
  577. if isinstance(bar['metric'], float):
  578. metric_str = "{0:.4f}".format(bar['metric'])
  579. else:
  580. metric_str = str(bar['metric'])
  581. metric_str = (" " * (cout.DETAILS_OFFSET - len(metric_str) - 1)) + metric_str
  582. count_str = str(bar['count'])
  583. count_str = ((" " * (count_str_len - len(count_str))) + count_str + diff_str + "\t")
  584. details.append((metric_str,
  585. "{0:.3f}".format(bar['ratio']) + " : " + "{0:.3f}".format(sum_ratio) + " : " +
  586. count_str + ('|' * int(bar['ratio']*100))))
  587. cout.notify(data['info']['path'],
  588. '', # no line number
  589. cout.SEVERITY_INFO,
  590. "Overall metrics for '" + namespace + ":" + field + "' metric",
  591. details)
  592. details = []
  593. for each in sorted(data['subdirs']):
  594. details.append(('Directory', each))
  595. for each in sorted(data['subfiles']):
  596. details.append(('File', each))
  597. if len(details) > 0:
  598. cout.notify(data['info']['path'],
  599. '', # no line number
  600. cout.SEVERITY_INFO,
  601. "Directory content:",
  602. details)
  603. def cout_prom_regions(path, regions, indent = 0):
  604. for region in regions:
  605. details = []
  606. for namespace in sorted(list(region['data'].keys())):
  607. diff_data = {}
  608. if '__diff__' in list(region['data'][namespace].keys()):
  609. diff_data = region['data'][namespace]['__diff__']
  610. for field in sorted(list(region['data'][namespace].keys())):
  611. diff_str = ""
  612. if field == '__diff__':
  613. continue
  614. if field in list(diff_data.keys()):
  615. diff_str = " [" + ("+" if diff_data[field] >= 0 else "") + str(diff_data[field]) + "]"
  616. details.append((namespace + ":" + field, str(region['data'][namespace][field]) + diff_str))
  617. promout.notify(path = path,
  618. region = region['info']['name'],
  619. metric = "",
  620. details = details)
  621. if 'subregions' in list(region.keys()):
  622. cout_txt_regions(path, region['subregions'], indent=indent+1)
  623. def cout_prom(data, loader):
  624. for key in list(data['file-data'].keys()):
  625. if key == 'regions':
  626. cout_prom_regions(data['info']['path'], data['file-data'][key])
  627. for namespace in sorted(list(data['aggregated-data'].keys())):
  628. for field in sorted(list(data['aggregated-data'][namespace].keys())):
  629. details = []
  630. for attr in ['avg', 'min', 'max', 'total']:
  631. if isinstance(data['aggregated-data'][namespace][field][attr], float):
  632. # round the data to reach same results on platforms with different precision
  633. details.append((attr, str(round(data['aggregated-data'][namespace][field][attr], DIGIT_COUNT))))
  634. else:
  635. details.append((attr, str(data['aggregated-data'][namespace][field][attr])))
  636. promout.notify(path = data['info']['path'],
  637. metric = namespace + "." + field,
  638. details = details)