api.py 48 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186
  1. #
  2. # Metrix++, Copyright 2009-2013, Metrix++ Project
  3. # Link: http://metrixplusplus.sourceforge.net
  4. #
  5. # This file is a part of Metrix++ Tool.
  6. #
  7. # Metrix++ is free software: you can redistribute it and/or modify
  8. # it under the terms of the GNU General Public License as published by
  9. # the Free Software Foundation, version 3 of the License.
  10. #
  11. # Metrix++ is distributed in the hope that it will be useful,
  12. # but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  14. # GNU General Public License for more details.
  15. #
  16. # You should have received a copy of the GNU General Public License
  17. # along with Metrix++. If not, see <http://www.gnu.org/licenses/>.
  18. #
  19. import os.path
  20. import sys
  21. import mpp.internal.dbwrap
  22. import mpp.internal.api_impl
  23. ##############################################################################
  24. #
  25. #
  26. #
  27. ##############################################################################
  28. class Data(object):
  29. def __init__(self):
  30. self.data = {}
  31. def get_data(self, namespace, field):
  32. if namespace not in self.data.keys():
  33. return None
  34. if field not in self.data[namespace].keys():
  35. return None
  36. return self.data[namespace][field]
  37. def set_data(self, namespace, field, value):
  38. if namespace not in self.data:
  39. self.data[namespace] = {}
  40. self.data[namespace][field] = value
  41. def iterate_namespaces(self):
  42. for namespace in self.data.keys():
  43. yield namespace
  44. def iterate_fields(self, namespace):
  45. for field in self.data[namespace].keys():
  46. yield (field, self.data[namespace][field])
  47. def get_data_tree(self, namespaces=None):
  48. return self.data
  49. def __repr__(self):
  50. return object.__repr__(self) + " with data " + self.data.__repr__()
  51. class LoadableData(Data):
  52. def __init__(self, loader, file_id, region_id):
  53. Data.__init__(self)
  54. self.loader = loader
  55. self.file_id = file_id
  56. self.region_id = region_id
  57. self.loaded_namespaces = []
  58. self.changed_namespaces = []
  59. def load_namespace(self, namespace):
  60. namespace_obj = self.loader.get_namespace(namespace)
  61. if namespace_obj == None:
  62. return
  63. regions_supported = namespace_obj.are_regions_supported()
  64. if ((self.region_id == None and regions_supported == True) or
  65. (self.region_id != None and regions_supported == False)):
  66. return
  67. row = self.loader.db.get_row(namespace, self.file_id, self.region_id)
  68. if row == None:
  69. return
  70. for column_name in row.keys():
  71. try:
  72. packager = namespace_obj._get_field_packager(column_name)
  73. except mpp.internal.api_impl.PackagerError:
  74. continue
  75. if row[column_name] == None:
  76. continue
  77. Data.set_data(self, namespace, column_name, packager.unpack(row[column_name]))
  78. def set_data(self, namespace, field, value):
  79. if namespace not in self.changed_namespaces:
  80. self.changed_namespaces.append(namespace)
  81. return Data.set_data(self, namespace, field, value)
  82. def get_data(self, namespace, field):
  83. if namespace not in self.loaded_namespaces:
  84. self.loaded_namespaces.append(namespace)
  85. self.load_namespace(namespace)
  86. return Data.get_data(self, namespace, field)
  87. def is_namespace_updated(self, namespace):
  88. return namespace in self.changed_namespaces
  89. def is_namespace_loaded(self, namespace):
  90. return namespace in self.loaded_namespaces
  91. def get_data_tree(self, namespaces=None):
  92. if namespaces == None:
  93. namespaces = self.loader.iterate_namespace_names()
  94. for each in namespaces:
  95. self.load_namespace(each)
  96. return Data.get_data_tree(self)
  97. class Region(LoadableData):
  98. class T(object):
  99. NONE = 0x00
  100. GLOBAL = 0x01
  101. CLASS = 0x02
  102. STRUCT = 0x04
  103. NAMESPACE = 0x08
  104. FUNCTION = 0x10
  105. INTERFACE = 0x20
  106. ANY = 0xFF
  107. def to_str(self, group):
  108. if group == self.NONE:
  109. return "none"
  110. elif group == self.GLOBAL:
  111. return "global"
  112. elif group == self.CLASS:
  113. return "class"
  114. elif group == self.STRUCT:
  115. return "struct"
  116. elif group == self.NAMESPACE:
  117. return "namespace"
  118. elif group == self.FUNCTION:
  119. return "function"
  120. elif group == self.INTERFACE:
  121. return "interface"
  122. else:
  123. assert(False)
  124. def __init__(self, loader, file_id, region_id, region_name, offset_begin, offset_end, line_begin, line_end, cursor_line, group, checksum):
  125. LoadableData.__init__(self, loader, file_id, region_id)
  126. self.name = region_name
  127. self.begin = offset_begin
  128. self.end = offset_end
  129. self.line_begin = line_begin
  130. self.line_end = line_end
  131. self.cursor = cursor_line
  132. self.group = group
  133. self.checksum = checksum
  134. self.children = []
  135. def get_id(self):
  136. return self.region_id
  137. def get_name(self):
  138. return self.name
  139. def get_offset_begin(self):
  140. return self.begin
  141. def get_offset_end(self):
  142. return self.end
  143. def get_line_begin(self):
  144. return self.line_begin
  145. def get_line_end(self):
  146. return self.line_end
  147. def get_cursor(self):
  148. return self.cursor
  149. def get_type(self):
  150. return self.group
  151. def get_checksum(self):
  152. return self.checksum
  153. def iterate_subregion_ids(self):
  154. return self.children
  155. def _register_subregion_id(self, child_id):
  156. self.children.append(child_id)
  157. class Marker(object):
  158. class T(object):
  159. NONE = 0x00
  160. COMMENT = 0x01
  161. STRING = 0x02
  162. PREPROCESSOR = 0x04
  163. CODE = 0x08
  164. ANY = 0xFF
  165. def to_str(self, group):
  166. if group == self.NONE:
  167. return "none"
  168. elif group == self.COMMENT:
  169. return "comment"
  170. elif group == self.STRING:
  171. return "string"
  172. elif group == self.PREPROCESSOR:
  173. return "preprocessor"
  174. elif group == self.CODE:
  175. return "code"
  176. else:
  177. assert(False)
  178. def __init__(self, offset_begin, offset_end, group):
  179. self.begin = offset_begin
  180. self.end = offset_end
  181. self.group = group
  182. def get_offset_begin(self):
  183. return self.begin
  184. def get_offset_end(self):
  185. return self.end
  186. def get_type(self):
  187. return self.group
  188. class FileData(LoadableData):
  189. def __init__(self, loader, path, file_id, checksum, content):
  190. LoadableData.__init__(self, loader, file_id, None)
  191. self.path = path
  192. self.checksum = checksum
  193. self.content = content
  194. self.regions = None
  195. self.markers = None
  196. self.loader = loader
  197. self.loading_tmp = []
  198. def get_id(self):
  199. return self.file_id
  200. def get_path(self):
  201. return self.path
  202. def get_checksum(self):
  203. return self.checksum
  204. def get_content(self):
  205. return self.content
  206. def _internal_append_region(self, region):
  207. # here we apply some magic - we rely on special ordering of coming regions,
  208. # which is supported by code parsers
  209. prev_id = None
  210. while True:
  211. if len(self.loading_tmp) == 0:
  212. break
  213. prev_id = self.loading_tmp.pop()
  214. if self.get_region(prev_id).get_offset_end() > region.get_offset_begin():
  215. self.loading_tmp.append(prev_id) # return back
  216. break
  217. self.loading_tmp.append(region.get_id())
  218. if prev_id != None:
  219. self.get_region(prev_id)._register_subregion_id(region.get_id())
  220. self.regions.append(region)
  221. def load_regions(self):
  222. if self.regions == None:
  223. self.regions = []
  224. for each in self.loader.db.iterate_regions(self.get_id()):
  225. self._internal_append_region(Region(self.loader,
  226. self.get_id(),
  227. each.region_id,
  228. each.name,
  229. each.begin,
  230. each.end,
  231. each.line_begin,
  232. each.line_end,
  233. each.cursor,
  234. each.group,
  235. each.checksum))
  236. assert(len(self.regions) == each.region_id)
  237. def add_region(self, region_name, offset_begin, offset_end, line_begin, line_end, cursor_line, group, checksum):
  238. if self.regions == None:
  239. self.regions = [] # do not load in time of collection
  240. new_id = len(self.regions) + 1
  241. self._internal_append_region(Region(self.loader, self.get_id(), new_id, region_name, offset_begin, offset_end, line_begin, line_end, cursor_line, group, checksum))
  242. self.loader.db.create_region(self.file_id, new_id, region_name, offset_begin, offset_end, line_begin, line_end, cursor_line, group, checksum)
  243. return new_id
  244. def get_region(self, region_id):
  245. self.load_regions()
  246. return self.regions[region_id - 1]
  247. def iterate_regions(self, filter_group = Region.T.ANY):
  248. self.load_regions()
  249. for each in self.regions:
  250. if each.group & filter_group:
  251. yield each
  252. def are_regions_loaded(self):
  253. return self.regions != None
  254. def load_markers(self):
  255. if self.markers == None:
  256. self.markers = []
  257. for each in self.loader.db.iterate_markers(self.get_id()):
  258. self.markers.append(Marker(each.begin, each.end, each.group))
  259. def add_marker(self, offset_begin, offset_end, group):
  260. if self.markers == None:
  261. self.markers = [] # do not load in time of collection
  262. self.markers.append(Marker(offset_begin, offset_end, group))
  263. self.loader.db.create_marker(self.file_id, offset_begin, offset_end, group)
  264. def iterate_markers(self, filter_group = Marker.T.ANY,
  265. region_id = None, exclude_children = True, merge = False):
  266. self.load_markers()
  267. # merged markers
  268. if merge == True:
  269. next_marker = None
  270. for marker in self.iterate_markers(filter_group, region_id, exclude_children, merge = False):
  271. if next_marker != None:
  272. if next_marker.get_offset_end() == marker.get_offset_begin():
  273. # sequential markers
  274. next_marker = Marker(next_marker.get_offset_begin(),
  275. marker.get_offset_end(),
  276. next_marker.get_type() | marker.get_type())
  277. else:
  278. yield next_marker
  279. next_marker = None
  280. if next_marker == None:
  281. next_marker = Marker(marker.get_offset_begin(),
  282. marker.get_offset_end(),
  283. marker.get_type())
  284. if next_marker != None:
  285. yield next_marker
  286. # all markers per file
  287. elif region_id == None:
  288. next_code_marker_start = 0
  289. for marker in self.markers:
  290. if Marker.T.CODE & filter_group and next_code_marker_start < marker.get_offset_begin():
  291. yield Marker(next_code_marker_start, marker.get_offset_begin(), Marker.T.CODE)
  292. if marker.group & filter_group:
  293. yield marker
  294. next_code_marker_start = marker.get_offset_end()
  295. if Marker.T.CODE & filter_group and next_code_marker_start < len(self.get_content()):
  296. yield Marker(next_code_marker_start, len(self.get_content()), Marker.T.CODE)
  297. # markers per region
  298. else:
  299. region = self.get_region(region_id)
  300. if region != None:
  301. # code parsers and database know about non-code markers
  302. # clients want to iterate code as markers as well
  303. # so, we embed code markers in run-time
  304. class CodeMarker(Marker):
  305. pass
  306. # cache markers for all regions if it does not exist
  307. if hasattr(region, '_markers_list') == False:
  308. # subroutine to populate _markers_list attribute
  309. # _markers_list does include code markers
  310. def cache_markers_list_rec(data, region_id, marker_start_ind, next_code_marker_start):
  311. region = data.get_region(region_id)
  312. region._markers_list = []
  313. region._first_marker_ind = marker_start_ind
  314. #next_code_marker_start = region.get_offset_begin()
  315. for sub_id in region.iterate_subregion_ids():
  316. subregion = data.get_region(sub_id)
  317. # cache all markers before the subregion
  318. while len(data.markers) > marker_start_ind and \
  319. subregion.get_offset_begin() > data.markers[marker_start_ind].get_offset_begin():
  320. if next_code_marker_start < data.markers[marker_start_ind].get_offset_begin():
  321. # append code markers coming before non-code marker
  322. region._markers_list.append(CodeMarker(next_code_marker_start,
  323. data.markers[marker_start_ind].get_offset_begin(),
  324. Marker.T.CODE))
  325. next_code_marker_start = data.markers[marker_start_ind].get_offset_end()
  326. region._markers_list.append(marker_start_ind)
  327. marker_start_ind += 1
  328. # cache all code markers before the subregion but after the last marker
  329. if next_code_marker_start < subregion.get_offset_begin():
  330. region._markers_list.append(CodeMarker(next_code_marker_start,
  331. subregion.get_offset_begin(),
  332. Marker.T.CODE))
  333. next_code_marker_start = subregion.get_offset_begin()
  334. # here is the recursive call for all sub-regions
  335. (marker_start_ind, next_code_marker_start) = cache_markers_list_rec(data,
  336. sub_id,
  337. marker_start_ind,
  338. next_code_marker_start)
  339. # cache all markers after the last subregion
  340. while len(data.markers) > marker_start_ind and \
  341. region.get_offset_end() > data.markers[marker_start_ind].get_offset_begin():
  342. # append code markers coming before non-code marker
  343. if next_code_marker_start < data.markers[marker_start_ind].get_offset_begin():
  344. region._markers_list.append(CodeMarker(next_code_marker_start,
  345. data.markers[marker_start_ind].get_offset_begin(),
  346. Marker.T.CODE))
  347. next_code_marker_start = data.markers[marker_start_ind].get_offset_end()
  348. region._markers_list.append(marker_start_ind)
  349. marker_start_ind += 1
  350. # cache the last code segment after the last marker
  351. if next_code_marker_start < region.get_offset_end():
  352. region._markers_list.append(CodeMarker(next_code_marker_start,
  353. region.get_offset_end(),
  354. Marker.T.CODE))
  355. next_code_marker_start = region.get_offset_end()
  356. # return the starting point for the next call of this function
  357. return (marker_start_ind, next_code_marker_start)
  358. # append markers list to all regions recursively
  359. (next_marker_pos, next_code_marker_start) = cache_markers_list_rec(self, 1, 0, 0)
  360. assert(next_marker_pos == len(self.markers))
  361. # excluding subregions
  362. if exclude_children == True:
  363. for marker_ind in region._markers_list:
  364. if isinstance(marker_ind, int):
  365. marker = self.markers[marker_ind]
  366. else:
  367. marker = marker_ind # CodeMarker
  368. if marker.group & filter_group:
  369. yield marker
  370. # including subregions
  371. else:
  372. next_code_marker_start = region.get_offset_begin()
  373. for marker in self.markers[region._first_marker_ind:]:
  374. if marker.get_offset_begin() >= region.get_offset_end():
  375. break
  376. if region.get_offset_begin() > marker.get_offset_begin():
  377. continue
  378. if Marker.T.CODE & filter_group and next_code_marker_start < marker.get_offset_begin():
  379. yield Marker(next_code_marker_start, marker.get_offset_begin(), Marker.T.CODE)
  380. if marker.group & filter_group:
  381. yield marker
  382. next_code_marker_start = marker.get_offset_end()
  383. if Marker.T.CODE & filter_group and next_code_marker_start < region.get_offset_end():
  384. yield Marker(next_code_marker_start, region.get_offset_end(), Marker.T.CODE)
  385. def are_markers_loaded(self):
  386. return self.markers != None
  387. def __repr__(self):
  388. return Data.__repr__(self) + " and regions " + self.regions.__repr__()
  389. class AggregatedData(Data):
  390. def __init__(self, loader, path):
  391. Data.__init__(self)
  392. self.path = path
  393. self.loader = loader
  394. self.subdirs = None
  395. self.subfiles = None
  396. def get_subdirs(self):
  397. if self.subdirs != None:
  398. return self.subdirs
  399. self.subdirs = []
  400. if self.path != None:
  401. for subdir in self.loader.db.iterate_dircontent(self.path, include_subdirs = True, include_subfiles = False):
  402. self.subdirs.append(subdir)
  403. return self.subdirs
  404. def get_subfiles(self):
  405. if self.subfiles != None:
  406. return self.subfiles
  407. self.subfiles = []
  408. if self.path != None:
  409. for subfile in self.loader.db.iterate_dircontent(self.path, include_subdirs = False, include_subfiles = True):
  410. self.subfiles.append(subfile)
  411. return self.subfiles
  412. class SelectData(Data):
  413. def __init__(self, loader, path, file_id, region_id):
  414. Data.__init__(self)
  415. self.loader = loader
  416. self.path = path
  417. self.file_id = file_id
  418. self.region_id = region_id
  419. self.region = None
  420. def get_path(self):
  421. return self.path
  422. def get_region(self):
  423. if self.region == None and self.region_id != None:
  424. row = self.loader.db.get_region(self.file_id, self.region_id)
  425. if row != None:
  426. self.region = Region(self.loader,
  427. self.file_id,
  428. self.region_id,
  429. row.name,
  430. row.begin,
  431. row.end,
  432. row.line_begin,
  433. row.line_end,
  434. row.cursor,
  435. row.group,
  436. row.checksum)
  437. return self.region
  438. class DiffData(Data):
  439. def __init__(self, new_data, old_data):
  440. Data.__init__(self)
  441. self.new_data = new_data
  442. self.old_data = old_data
  443. def get_data(self, namespace, field):
  444. new_data = self.new_data.get_data(namespace, field)
  445. old_data = self.old_data.get_data(namespace, field)
  446. if new_data == None:
  447. return None
  448. if old_data == None:
  449. # non_zero fields has got zero value by default if missed
  450. # the data can be also unavailable,
  451. # because previous collection does not include that
  452. # but external tools (like limit.py) should warn about this,
  453. # using list of registered database properties
  454. old_data = 0
  455. return new_data - old_data
  456. ####################################
  457. # Loader
  458. ####################################
  459. class Namespace(object):
  460. class NamespaceError(Exception):
  461. def __init__(self, namespace, reason):
  462. Exception.__init__(self, "Namespace '"
  463. + namespace
  464. + "': '"
  465. + reason
  466. + "'")
  467. class FieldError(Exception):
  468. def __init__(self, field, reason):
  469. Exception.__init__(self, "Field '"
  470. + field
  471. + "': '"
  472. + reason
  473. + "'")
  474. def __init__(self, db_handle, name, support_regions = False, version='1.0'):
  475. if not isinstance(name, str):
  476. raise Namespace.NamespaceError(name, "name not a string")
  477. self.name = name
  478. self.support_regions = support_regions
  479. self.fields = {}
  480. self.db = db_handle
  481. if self.db.check_table(name) == False:
  482. self.db.create_table(name, support_regions, version)
  483. else:
  484. for column in self.db.iterate_columns(name):
  485. self.add_field(column.name,
  486. mpp.internal.api_impl.PackagerFactory().get_python_type(column.sql_type),
  487. non_zero=column.non_zero)
  488. def get_name(self):
  489. return self.name
  490. def are_regions_supported(self):
  491. return self.support_regions
  492. def add_field(self, field_name, python_type, non_zero=False):
  493. if not isinstance(field_name, str):
  494. raise Namespace.FieldError(field_name, "field_name not a string")
  495. packager = mpp.internal.api_impl.PackagerFactory().create(python_type, non_zero)
  496. if field_name in self.fields.keys():
  497. raise Namespace.FieldError(field_name, "double used")
  498. self.fields[field_name] = packager
  499. if self.db.check_column(self.get_name(), field_name) == False:
  500. # - False if cloned
  501. # - True if created
  502. return self.db.create_column(self.name, field_name, packager.get_sql_type(), non_zero=non_zero)
  503. return None # if double request
  504. def iterate_field_names(self):
  505. for name in self.fields.keys():
  506. yield name
  507. def check_field(self, field_name):
  508. try:
  509. self._get_field_packager(field_name)
  510. except mpp.internal.api_impl.PackagerError:
  511. return False
  512. return True
  513. def get_field_sql_type(self, field_name):
  514. try:
  515. return self._get_field_packager(field_name).get_sql_type()
  516. except mpp.internal.api_impl.PackagerError:
  517. raise Namespace.FieldError(field_name, 'does not exist')
  518. def get_field_python_type(self, field_name):
  519. try:
  520. return self._get_field_packager(field_name).get_python_type()
  521. except mpp.internal.api_impl.PackagerError:
  522. raise Namespace.FieldError(field_name, 'does not exist')
  523. def is_field_non_zero(self, field_name):
  524. try:
  525. return self._get_field_packager(field_name).is_non_zero()
  526. except mpp.internal.api_impl.PackagerError:
  527. raise Namespace.FieldError(field_name, 'does not exist')
  528. def _get_field_packager(self, field_name):
  529. if field_name in self.fields.keys():
  530. return self.fields[field_name]
  531. else:
  532. raise mpp.internal.api_impl.PackagerError("unknown field " + field_name + " requested")
  533. class Loader(object):
  534. def __init__(self):
  535. self.namespaces = {}
  536. self.db = None
  537. self.last_file_data = None # for performance boost reasons
  538. def create_database(self, dbfile, previous_db = None):
  539. self.db = mpp.internal.dbwrap.Database()
  540. try:
  541. self.db.create(dbfile, clone_from=previous_db)
  542. except:
  543. return False
  544. return True
  545. def open_database(self, dbfile, read_only = True):
  546. self.db = mpp.internal.dbwrap.Database()
  547. if os.path.exists(dbfile) == False:
  548. return False
  549. try:
  550. self.db.connect(dbfile, read_only=read_only)
  551. except:
  552. return False
  553. for table in self.db.iterate_tables():
  554. self.create_namespace(table.name, table.support_regions)
  555. return True
  556. def set_property(self, property_name, value):
  557. if self.db == None:
  558. return None
  559. return self.db.set_property(property_name, str(value))
  560. def get_property(self, property_name):
  561. if self.db == None:
  562. return None
  563. return self.db.get_property(property_name)
  564. def iterate_properties(self):
  565. if self.db == None:
  566. return None
  567. return self.db.iterate_properties()
  568. def create_namespace(self, name, support_regions = False, version='1.0'):
  569. if self.db == None:
  570. return None
  571. if name in self.namespaces.keys():
  572. raise Namespace.NamespaceError(name, "double used")
  573. new_namespace = Namespace(self.db, name, support_regions, version)
  574. self.namespaces[name] = new_namespace
  575. return new_namespace
  576. def iterate_namespace_names(self):
  577. for name in self.namespaces.keys():
  578. yield name
  579. def get_namespace(self, name):
  580. if name in self.namespaces.keys():
  581. return self.namespaces[name]
  582. else:
  583. return None
  584. def create_file_data(self, path, checksum, content):
  585. if self.db == None:
  586. return None
  587. (new_id, is_updated) = self.db.create_file(path, checksum)
  588. result = FileData(self, path, new_id, checksum, content)
  589. self.last_file_data = result
  590. return (result, is_updated)
  591. def load_file_data(self, path):
  592. if self.db == None:
  593. return None
  594. if self.last_file_data != None and self.last_file_data.get_path() == path:
  595. return self.last_file_data
  596. data = self.db.get_file(path)
  597. if data == None:
  598. return None
  599. result = FileData(self, data.path, data.id, data.checksum, None)
  600. self.last_file_data = result
  601. return result
  602. class DataNotPackable(Exception):
  603. def __init__(self, namespace, field, value, packager, extra_message):
  604. Exception.__init__(self, "Data '"
  605. + str(value)
  606. + "' of type "
  607. + str(value.__class__)
  608. + " referred by '"
  609. + namespace
  610. + "=>"
  611. + field
  612. + "' is not packable by registered packager '"
  613. + str(packager.__class__)
  614. + "': " + extra_message)
  615. def save_file_data(self, file_data):
  616. if self.db == None:
  617. return None
  618. class DataIterator(object):
  619. def iterate_packed_values(self, data, namespace, support_regions = False):
  620. for each in data.iterate_fields(namespace):
  621. space = self.loader.get_namespace(namespace)
  622. if space == None:
  623. raise Loader.DataNotPackable(namespace, each[0], each[1], None, "The namespace has not been found")
  624. try:
  625. packager = space._get_field_packager(each[0])
  626. except mpp.internal.api_impl.PackagerError:
  627. raise Loader.DataNotPackable(namespace, each[0], each[1], None, "The field has not been found")
  628. if space.support_regions != support_regions:
  629. raise Loader.DataNotPackable(namespace, each[0], each[1], packager, "Incompatible support for regions")
  630. try:
  631. packed_data = packager.pack(each[1])
  632. if packed_data == None:
  633. continue
  634. except mpp.internal.api_impl.PackagerError:
  635. raise Loader.DataNotPackable(namespace, each[0], each[1], packager, "Packager raised exception")
  636. yield (each[0], packed_data)
  637. def __init__(self, loader, data, namespace, support_regions = False):
  638. self.loader = loader
  639. self.iterator = self.iterate_packed_values(data, namespace, support_regions)
  640. def __iter__(self):
  641. return self.iterator
  642. for namespace in file_data.iterate_namespaces():
  643. if file_data.is_namespace_updated(namespace) == False:
  644. continue
  645. self.db.add_row(namespace,
  646. file_data.get_id(),
  647. None,
  648. DataIterator(self, file_data, namespace))
  649. if file_data.are_regions_loaded():
  650. for region in file_data.iterate_regions():
  651. for namespace in region.iterate_namespaces():
  652. if region.is_namespace_updated(namespace) == False:
  653. continue
  654. self.db.add_row(namespace,
  655. file_data.get_id(),
  656. region.get_id(),
  657. DataIterator(self, region, namespace, support_regions = True))
  658. def iterate_file_data(self, path = None, path_like_filter = "%"):
  659. if self.db == None:
  660. return None
  661. final_path_like = path_like_filter
  662. if path != None:
  663. if self.db.check_dir(path) == False and self.db.check_file(path) == False:
  664. return None
  665. final_path_like = path + path_like_filter
  666. class FileDataIterator(object):
  667. def iterate_file_data(self, loader, final_path_like):
  668. for data in loader.db.iterate_files(path_like=final_path_like):
  669. yield FileData(loader, data.path, data.id, data.checksum, None)
  670. def __init__(self, loader, final_path_like):
  671. self.iterator = self.iterate_file_data(loader, final_path_like)
  672. def __iter__(self):
  673. return self.iterator
  674. if self.db == None:
  675. return None
  676. return FileDataIterator(self, final_path_like)
  677. def load_aggregated_data(self, path = None, path_like_filter = "%", namespaces = None):
  678. if self.db == None:
  679. return None
  680. final_path_like = path_like_filter
  681. if path != None:
  682. if self.db.check_dir(path) == False and self.db.check_file(path) == False:
  683. return None
  684. final_path_like = path + path_like_filter
  685. if namespaces == None:
  686. namespaces = self.namespaces.keys()
  687. result = AggregatedData(self, path)
  688. for name in namespaces:
  689. namespace = self.get_namespace(name)
  690. data = self.db.aggregate_rows(name, path_like = final_path_like)
  691. for field in data.keys():
  692. if namespace.get_field_python_type(field) == str:
  693. continue
  694. data[field]['nonzero'] = namespace.is_field_non_zero(field)
  695. distribution = self.db.count_rows(name, path_like = final_path_like, group_by_column = field)
  696. data[field]['distribution-bars'] = []
  697. for each in distribution:
  698. if each[0] == None:
  699. continue
  700. assert(float(data[field]['count'] != 0))
  701. data[field]['distribution-bars'].append({'metric': each[0],
  702. 'count': each[1],
  703. 'ratio': (float(each[1]) / float(data[field]['count']))})
  704. result.set_data(name, field, data[field])
  705. return result
  706. def load_selected_data(self, namespace, fields = None, path = None, path_like_filter = "%", filters = [],
  707. sort_by = None, limit_by = None):
  708. if self.db == None:
  709. return None
  710. final_path_like = path_like_filter
  711. if path != None:
  712. if self.db.check_dir(path) == False and self.db.check_file(path) == False:
  713. return None
  714. final_path_like = path + path_like_filter
  715. namespace_obj = self.get_namespace(namespace)
  716. if namespace_obj == None:
  717. return None
  718. class SelectDataIterator(object):
  719. def iterate_selected_values(self, loader, namespace_obj, final_path_like, fields, filters, sort_by, limit_by):
  720. for row in loader.db.select_rows(namespace_obj.get_name(), path_like=final_path_like, filters=filters,
  721. order_by=sort_by, limit_by=limit_by):
  722. region_id = None
  723. if namespace_obj.are_regions_supported() == True:
  724. region_id = row['region_id']
  725. data = SelectData(loader, row['path'], row['id'], region_id)
  726. field_names = fields
  727. if fields == None:
  728. field_names = namespace_obj.iterate_field_names()
  729. for field in field_names:
  730. data.set_data(namespace, field, row[field])
  731. yield data
  732. def __init__(self, loader, namespace_obj, final_path_like, fields, filters, sort_by, limit_by):
  733. self.iterator = self.iterate_selected_values(loader, namespace_obj, final_path_like, fields, filters, sort_by, limit_by)
  734. def __iter__(self):
  735. return self.iterator
  736. return SelectDataIterator(self, namespace_obj, final_path_like, fields, filters, sort_by, limit_by)
  737. class BasePlugin(object):
  738. def initialize(self):
  739. pass
  740. def terminate(self):
  741. pass
  742. def set_name(self, name):
  743. self.name = name
  744. def get_name(self):
  745. if hasattr(self, 'name') == False:
  746. return None
  747. return self.name
  748. def set_version(self, version):
  749. self.version = version
  750. def get_version(self):
  751. if hasattr(self, 'version') == False:
  752. return None
  753. return self.version
  754. def _set_plugin_loader(self, loader):
  755. self.plugin_loader = loader
  756. def _get_plugin_loader(self):
  757. if hasattr(self, 'plugin_loader') == False:
  758. return None
  759. return self.plugin_loader
  760. def get_plugin(self, plugin_name):
  761. return self._get_plugin_loader().get_plugin(plugin_name)
  762. def get_action(self):
  763. return self._get_plugin_loader().get_action()
  764. class Plugin(BasePlugin):
  765. class Field(object):
  766. def __init__(self, name, ftype, non_zero=False):
  767. self.name = name
  768. self.type = ftype
  769. self.non_zero = non_zero
  770. class Property(object):
  771. def __init__(self, name, value):
  772. self.name = name
  773. self.value = value
  774. def initialize(self, namespace=None, support_regions=True, fields=[], properties=[]):
  775. super(Plugin, self).initialize()
  776. if hasattr(self, 'is_updated') == False:
  777. self.is_updated = False # original initialization
  778. db_loader = self.get_plugin('mpp.dbf').get_loader()
  779. if namespace == None:
  780. namespace = self.get_name()
  781. if (len(fields) != 0 or len(properties) != 0):
  782. prev_version = db_loader.set_property(self.get_name() + ":version", self.get_version())
  783. if prev_version != self.get_version():
  784. self.is_updated = True
  785. for prop in properties:
  786. assert(prop.name != 'version')
  787. prev_prop = db_loader.set_property(self.get_name() + ":" + prop.name, prop.value)
  788. if prev_prop != prop.value:
  789. self.is_updated = True
  790. if len(fields) != 0:
  791. namespace_obj = db_loader.create_namespace(namespace,
  792. support_regions=support_regions,
  793. version=self.get_version())
  794. for field in fields:
  795. is_created = namespace_obj.add_field(field.name, field.type, non_zero=field.non_zero)
  796. assert(is_created != None)
  797. # if field is created (not cloned from the previous db),
  798. # mark the plug-in as updated in order to trigger full rescan
  799. self.is_updated = self.is_updated or is_created
  800. class MetricPluginMixin(object):
  801. class AliasError(Exception):
  802. def __init__(self, alias):
  803. Exception.__init__(self, "Unknown pattern alias: " + str(alias))
  804. class PlainCounter(object):
  805. def __init__(self, plugin, alias, data, region):
  806. self.plugin = plugin
  807. self.alias = alias
  808. self.data = data
  809. self.region = region
  810. self.result = 0
  811. def count(self, marker, pattern_to_search):
  812. self.result += len(pattern_to_search.findall(self.data.get_content(),
  813. marker.get_offset_begin(),
  814. marker.get_offset_end()))
  815. def get_result(self):
  816. return self.result
  817. class IterIncrementCounter(PlainCounter):
  818. def count(self, marker, pattern_to_search):
  819. self.marker = marker
  820. self.pattern_to_search = pattern_to_search
  821. for match in pattern_to_search.finditer(self.data.get_content(),
  822. marker.get_offset_begin(),
  823. marker.get_offset_end()):
  824. self.result += self.increment(match)
  825. def increment(self, match):
  826. return 1
  827. class IterAssignCounter(PlainCounter):
  828. def count(self, marker, pattern_to_search):
  829. self.marker = marker
  830. self.pattern_to_search = pattern_to_search
  831. for match in pattern_to_search.finditer(self.data.get_content(),
  832. marker.get_offset_begin(),
  833. marker.get_offset_end()):
  834. self.result = self.assign(match)
  835. def assign(self, match):
  836. return self.result
  837. def declare_metric(self, is_active, field,
  838. pattern_to_search_or_map_of_patterns,
  839. marker_type_mask=Marker.T.ANY,
  840. region_type_mask=Region.T.ANY,
  841. exclude_subregions=True,
  842. merge_markers=False):
  843. if hasattr(self, '_fields') == False:
  844. self._fields = {}
  845. if isinstance(pattern_to_search_or_map_of_patterns, dict):
  846. map_of_patterns = pattern_to_search_or_map_of_patterns
  847. else:
  848. map_of_patterns = {'*': pattern_to_search_or_map_of_patterns}
  849. # client may suply with pattern or pair of pattern + counter class
  850. for key in map_of_patterns.keys():
  851. if isinstance(map_of_patterns[key], tuple) == False:
  852. # if it is not a pair, create a pair using default counter class
  853. map_of_patterns[key] = (map_of_patterns[key],
  854. MetricPluginMixin.PlainCounter)
  855. if is_active == True:
  856. self._fields[field.name] = (field,
  857. marker_type_mask,
  858. exclude_subregions,
  859. merge_markers,
  860. map_of_patterns,
  861. region_type_mask)
  862. def is_active(self, metric_name = None):
  863. if metric_name == None:
  864. return (len(self._fields.keys()) > 0)
  865. return (metric_name in self._fields.keys())
  866. def get_fields(self):
  867. result = []
  868. for key in self._fields.keys():
  869. result.append(self._fields[key][0])
  870. return result
  871. def callback(self, parent, data, is_updated):
  872. # count if metric is enabled,
  873. # and (optimization for the case of iterative rescan:)
  874. # if file is updated or this plugin's settings are updated
  875. is_updated = is_updated or self.is_updated
  876. if is_updated == True:
  877. for field in self.get_fields():
  878. self.count_if_active(field.name, data, alias=parent.get_name())
  879. # if parent, notify children
  880. if isinstance(self, Parent):
  881. self.notify_children(data, is_updated)
  882. def count_if_active(self, metric_name, data, namespace=None, alias='*'):
  883. if self.is_active(metric_name) == False:
  884. return
  885. if namespace == None:
  886. namespace = self.get_name()
  887. field_data = self._fields[metric_name]
  888. if alias not in field_data[4].keys():
  889. if '*' not in field_data[4].keys():
  890. raise self.AliasError(alias)
  891. else:
  892. alias = '*'
  893. (pattern_to_search, counter_class) = field_data[4][alias]
  894. for region in data.iterate_regions(filter_group=field_data[5]):
  895. counter = counter_class(self, alias, data, region)
  896. for marker in data.iterate_markers(
  897. filter_group = field_data[1],
  898. region_id = region.get_id(),
  899. exclude_children = field_data[2],
  900. merge=field_data[3]):
  901. counter.count(marker, pattern_to_search)
  902. count = counter.get_result()
  903. if count != 0 or field_data[0].non_zero == False:
  904. region.set_data(namespace, metric_name, count)
  905. class InterfaceNotImplemented(Exception):
  906. def __init__(self, obj):
  907. Exception.__init__(self, "Method '"
  908. + sys._getframe(1).f_code.co_name
  909. + "' has not been implemented for "
  910. + str(obj.__class__))
  911. class IConfigurable(object):
  912. def configure(self, options):
  913. raise InterfaceNotImplemented(self)
  914. def declare_configuration(self, optparser):
  915. raise InterfaceNotImplemented(self)
  916. class IRunable(object):
  917. def run(self, args):
  918. raise InterfaceNotImplemented(self)
  919. class IParser(object):
  920. def process(self, parent, data, is_updated):
  921. raise InterfaceNotImplemented(self)
  922. class ICode(object):
  923. pass
  924. class CallbackNotImplemented(Exception):
  925. def __init__(self, obj, callback_name):
  926. Exception.__init__(self, "Callback '"
  927. + callback_name
  928. + "' has not been implemented for "
  929. + str(obj.__class__))
  930. class Child(object):
  931. def notify(self, parent, callback_name, *args):
  932. if hasattr(self, callback_name) == False:
  933. raise CallbackNotImplemented(self, callback_name)
  934. self.__getattribute__(callback_name)(parent, *args)
  935. def subscribe_by_parents_name(self, parent_name, callback_name='callback'):
  936. self.get_plugin(parent_name).subscribe(self, callback_name)
  937. def subscribe_by_parents_names(self, parent_names, callback_name='callback'):
  938. for parent_name in parent_names:
  939. self.get_plugin(parent_name).subscribe(self, callback_name)
  940. def subscribe_by_parents_interface(self, interface, callback_name='callback'):
  941. for plugin in self._get_plugin_loader().iterate_plugins():
  942. if isinstance(plugin, interface):
  943. plugin.subscribe(self, callback_name)
  944. class Parent(object):
  945. def init_Parent(self):
  946. if hasattr(self, 'children') == False:
  947. self.children = []
  948. def subscribe(self, obj, callback_name):
  949. self.init_Parent()
  950. if (isinstance(obj, Child) == False):
  951. raise TypeError()
  952. self.children.append((obj,callback_name))
  953. def unsubscribe(self, obj, callback_name):
  954. self.init_Parent()
  955. self.children.remove((obj, callback_name))
  956. def notify_children(self, *args):
  957. self.init_Parent()
  958. for child in self.children:
  959. child[0].notify(self, child[1], *args)
  960. def iterate_children(self):
  961. self.init_Parent()
  962. for child in self.children:
  963. yield child