view.py 9.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224
  1. #
  2. # Metrix++, Copyright 2009-2013, Metrix++ Project
  3. # Link: http://metrixplusplus.sourceforge.net
  4. #
  5. # This file is a part of Metrix++ Tool.
  6. #
  7. # Metrix++ is free software: you can redistribute it and/or modify
  8. # it under the terms of the GNU General Public License as published by
  9. # the Free Software Foundation, version 3 of the License.
  10. #
  11. # Metrix++ is distributed in the hope that it will be useful,
  12. # but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  14. # GNU General Public License for more details.
  15. #
  16. # You should have received a copy of the GNU General Public License
  17. # along with Metrix++. If not, see <http://www.gnu.org/licenses/>.
  18. #
  19. import mpp.log
  20. import mpp.dbf
  21. import mpp.utils
  22. import mpp.cmdparser
  23. import mpp.api
  24. class Tool(mpp.api.ITool):
  25. def run(self, tool_args):
  26. return main(tool_args)
  27. def main(tool_args):
  28. log_plugin = mpp.log.Plugin()
  29. db_plugin = mpp.dbf.Plugin()
  30. parser = mpp.cmdparser.MultiOptionParser(usage="Usage: %prog view [options] -- [path 1] ... [path N]")
  31. log_plugin.declare_configuration(parser)
  32. db_plugin.declare_configuration(parser)
  33. parser.add_option("--format", "--ft", default='xml', choices=['txt', 'xml', 'python'], help="Format of the output data. "
  34. "Possible values are 'xml', 'txt' or 'python' [default: %default]")
  35. parser.add_option("--nest-regions", "--nr", action="store_true", default=False,
  36. help="If the option is set (True), data for regions is exported in the form of a tree. "
  37. "Otherwise, all regions are exported in plain list. [default: %default]")
  38. (options, args) = parser.parse_args(tool_args)
  39. log_plugin.configure(options)
  40. db_plugin.configure(options)
  41. out_format = options.__dict__['format']
  42. nest_regions = options.__dict__['nest_regions']
  43. log_plugin.initialize()
  44. db_plugin.initialize()
  45. loader_prev = db_plugin.get_loader_prev()
  46. loader = db_plugin.get_loader()
  47. # Check for versions consistency
  48. if db_plugin.dbfile_prev != None:
  49. mpp.utils.check_db_metadata(loader, loader_prev)
  50. paths = None
  51. if len(args) == 0:
  52. paths = [""]
  53. else:
  54. paths = args
  55. (result, exit_code) = export_to_str(out_format, paths, loader, loader_prev, nest_regions)
  56. print result
  57. return exit_code
  58. def export_to_str(out_format, paths, loader, loader_prev, nest_regions):
  59. exit_code = 0
  60. result = ""
  61. if out_format == 'txt':
  62. result += "=" * 80 + "\n" + "Export" + "\n" + "_" * 80 + "\n\n"
  63. elif out_format == 'xml':
  64. result += "<export>\n"
  65. elif out_format == 'python':
  66. result += "{'export': ["
  67. for (ind, path) in enumerate(paths):
  68. path = mpp.utils.preprocess_path(path)
  69. aggregated_data = loader.load_aggregated_data(path)
  70. aggregated_data_tree = {}
  71. subdirs = []
  72. subfiles = []
  73. if aggregated_data != None:
  74. aggregated_data_tree = aggregated_data.get_data_tree()
  75. subdirs = aggregated_data.get_subdirs()
  76. subfiles = aggregated_data.get_subfiles()
  77. else:
  78. mpp.utils.report_bad_path(path)
  79. exit_code += 1
  80. aggregated_data_prev = loader_prev.load_aggregated_data(path)
  81. if aggregated_data_prev != None:
  82. aggregated_data_tree = append_diff(aggregated_data_tree,
  83. aggregated_data_prev.get_data_tree())
  84. file_data = loader.load_file_data(path)
  85. file_data_tree = {}
  86. if file_data != None:
  87. file_data_tree = file_data.get_data_tree()
  88. file_data_prev = loader_prev.load_file_data(path)
  89. append_regions(file_data_tree, file_data, file_data_prev, nest_regions)
  90. data = {"info": {"path": path, "id": ind + 1},
  91. "aggregated-data": aggregated_data_tree,
  92. "file-data": file_data_tree,
  93. "subdirs": subdirs,
  94. "subfiles": subfiles}
  95. if out_format == 'txt':
  96. result += mpp.utils.serialize_to_txt(data, root_name = "data") + "\n"
  97. elif out_format == 'xml':
  98. result += mpp.utils.serialize_to_xml(data, root_name = "data") + "\n"
  99. elif out_format == 'python':
  100. postfix = ""
  101. if ind < len(paths) - 1:
  102. postfix = ", "
  103. result += mpp.utils.serialize_to_python(data, root_name = "data") + postfix
  104. if out_format == 'txt':
  105. result += "\n"
  106. elif out_format == 'xml':
  107. result += "</export>"
  108. elif out_format == 'python':
  109. result += "]}"
  110. return (result, exit_code)
  111. def append_regions(file_data_tree, file_data, file_data_prev, nest_regions):
  112. regions_matcher = None
  113. if file_data_prev != None:
  114. file_data_tree = append_diff(file_data_tree,
  115. file_data_prev.get_data_tree())
  116. regions_matcher = mpp.utils.FileRegionsMatcher(file_data, file_data_prev)
  117. if nest_regions == False:
  118. regions = []
  119. for region in file_data.iterate_regions():
  120. region_data_tree = region.get_data_tree()
  121. if regions_matcher != None and regions_matcher.is_matched(region.get_id()):
  122. region_data_prev = file_data_prev.get_region(regions_matcher.get_prev_id(region.get_id()))
  123. region_data_tree = append_diff(region_data_tree,
  124. region_data_prev.get_data_tree())
  125. regions.append({"info": {"name" : region.name,
  126. 'type' : file_data.get_region_types()().to_str(region.get_type()),
  127. "cursor" : region.cursor,
  128. 'line_begin': region.line_begin,
  129. 'line_end': region.line_end,
  130. 'offset_begin': region.begin,
  131. 'offset_end': region.end},
  132. "data": region_data_tree})
  133. file_data_tree['regions'] = regions
  134. else:
  135. def append_rec(region_id, file_data_tree, file_data, file_data_prev):
  136. region = file_data.get_region(region_id)
  137. region_data_tree = region.get_data_tree()
  138. if regions_matcher != None and regions_matcher.is_matched(region.get_id()):
  139. region_data_prev = file_data_prev.get_region(regions_matcher.get_prev_id(region.get_id()))
  140. region_data_tree = append_diff(region_data_tree,
  141. region_data_prev.get_data_tree())
  142. result = {"info": {"name" : region.name,
  143. 'type' : file_data.get_region_types()().to_str(region.get_type()),
  144. "cursor" : region.cursor,
  145. 'line_begin': region.line_begin,
  146. 'line_end': region.line_end,
  147. 'offset_begin': region.begin,
  148. 'offset_end': region.end},
  149. "data": region_data_tree,
  150. "subregions": []}
  151. for sub_id in file_data.get_region(region_id).iterate_subregion_ids():
  152. result['subregions'].append(append_rec(sub_id, file_data_tree, file_data, file_data_prev))
  153. return result
  154. file_data_tree['regions'] = []
  155. file_data_tree['regions'].append(append_rec(1, file_data_tree, file_data, file_data_prev))
  156. def append_diff(main_tree, prev_tree):
  157. assert(main_tree != None)
  158. assert(prev_tree != None)
  159. for name in main_tree.keys():
  160. if name not in prev_tree.keys():
  161. continue
  162. for field in main_tree[name].keys():
  163. if field not in prev_tree[name].keys():
  164. continue
  165. if isinstance(main_tree[name][field], dict) and isinstance(prev_tree[name][field], dict):
  166. diff = {}
  167. for key in main_tree[name][field].keys():
  168. if key not in prev_tree[name][field].keys():
  169. continue
  170. main_val = main_tree[name][field][key]
  171. prev_val = prev_tree[name][field][key]
  172. if main_val == None:
  173. main_val = 0
  174. if prev_val == None:
  175. prev_val = 0
  176. if isinstance(main_val, list) and isinstance(prev_val, list):
  177. main_tree[name][field][key] = append_diff_list(main_val, prev_val)
  178. else:
  179. diff[key] = main_val - prev_val
  180. main_tree[name][field]['__diff__'] = diff
  181. elif (not isinstance(main_tree[name][field], dict)) and (not isinstance(prev_tree[name][field], dict)):
  182. if '__diff__' not in main_tree[name]:
  183. main_tree[name]['__diff__'] = {}
  184. main_tree[name]['__diff__'][field] = main_tree[name][field] - prev_tree[name][field]
  185. return main_tree
  186. def append_diff_list(main_list, prev_list):
  187. merged_list = {}
  188. for bar in main_list:
  189. merged_list[bar['metric']] = {'count': bar['count'], '__diff__':0}
  190. for bar in prev_list:
  191. if bar['metric'] in merged_list.keys():
  192. merged_list[bar['metric']]['__diff__'] = \
  193. merged_list[bar['metric']]['count'] - bar['count']
  194. else:
  195. merged_list[bar['metric']] = {'count': 0, '__diff__':-bar['count']}
  196. result = []
  197. for metric in sorted(merged_list.keys()):
  198. result.append({'metric':metric, 'count':merged_list[metric]['count'], '__diff__':merged_list[metric]['__diff__']})
  199. return result