api.py 48 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227
  1. #
  2. # Metrix++, Copyright 2009-2013, Metrix++ Project
  3. # Link: http://metrixplusplus.sourceforge.net
  4. #
  5. # This file is a part of Metrix++ Tool.
  6. #
  7. # Metrix++ is free software: you can redistribute it and/or modify
  8. # it under the terms of the GNU General Public License as published by
  9. # the Free Software Foundation, version 3 of the License.
  10. #
  11. # Metrix++ is distributed in the hope that it will be useful,
  12. # but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  14. # GNU General Public License for more details.
  15. #
  16. # You should have received a copy of the GNU General Public License
  17. # along with Metrix++. If not, see <http://www.gnu.org/licenses/>.
  18. #
  19. import logging
  20. import os.path
  21. import mpp.internal.dbwrap
  22. ##############################################################################
  23. #
  24. #
  25. #
  26. ##############################################################################
  27. class Data(object):
  28. def __init__(self):
  29. self.data = {}
  30. def get_data(self, namespace, field):
  31. if namespace not in self.data.keys():
  32. return None
  33. if field not in self.data[namespace].keys():
  34. return None
  35. return self.data[namespace][field]
  36. def set_data(self, namespace, field, value):
  37. if namespace not in self.data:
  38. self.data[namespace] = {}
  39. self.data[namespace][field] = value
  40. def iterate_namespaces(self):
  41. for namespace in self.data.keys():
  42. yield namespace
  43. def iterate_fields(self, namespace):
  44. for field in self.data[namespace].keys():
  45. yield (field, self.data[namespace][field])
  46. def get_data_tree(self, namespaces=None):
  47. return self.data
  48. def __repr__(self):
  49. return object.__repr__(self) + " with data " + self.data.__repr__()
  50. class LoadableData(Data):
  51. def __init__(self, loader, file_id, region_id):
  52. Data.__init__(self)
  53. self.loader = loader
  54. self.file_id = file_id
  55. self.region_id = region_id
  56. self.loaded_namespaces = []
  57. self.changed_namespaces = []
  58. def load_namespace(self, namespace):
  59. try:
  60. row = self.loader.db.get_row(namespace, self.file_id, self.region_id)
  61. except Exception:
  62. logging.debug("No data in the database for namespace: " + namespace)
  63. return
  64. if row == None:
  65. return
  66. for column_name in row.keys():
  67. packager = self.loader.get_namespace(namespace).get_field_packager(column_name)
  68. if packager == None:
  69. continue
  70. if row[column_name] == None:
  71. continue
  72. Data.set_data(self, namespace, column_name, packager.unpack(row[column_name]))
  73. def set_data(self, namespace, field, value):
  74. if namespace not in self.changed_namespaces:
  75. self.changed_namespaces.append(namespace)
  76. return Data.set_data(self, namespace, field, value)
  77. def get_data(self, namespace, field):
  78. if namespace not in self.loaded_namespaces:
  79. self.loaded_namespaces.append(namespace)
  80. self.load_namespace(namespace)
  81. return Data.get_data(self, namespace, field)
  82. def is_namespace_updated(self, namespace):
  83. return namespace in self.changed_namespaces
  84. def is_namespace_loaded(self, namespace):
  85. return namespace in self.loaded_namespaces
  86. def get_data_tree(self, namespaces=None):
  87. if namespaces == None:
  88. namespaces = self.loader.iterate_namespace_names()
  89. for each in namespaces:
  90. self.load_namespace(each)
  91. return Data.get_data_tree(self)
  92. class Region(LoadableData):
  93. class T(object):
  94. NONE = 0x00
  95. GLOBAL = 0x01
  96. CLASS = 0x02
  97. STRUCT = 0x04
  98. NAMESPACE = 0x08
  99. FUNCTION = 0x10
  100. INTERFACE = 0x20
  101. ANY = 0xFF
  102. def to_str(self, group):
  103. if group == self.NONE:
  104. return "none"
  105. elif group == self.GLOBAL:
  106. return "global"
  107. elif group == self.CLASS:
  108. return "class"
  109. elif group == self.STRUCT:
  110. return "struct"
  111. elif group == self.NAMESPACE:
  112. return "namespace"
  113. elif group == self.FUNCTION:
  114. return "function"
  115. elif group == self.INTERFACE:
  116. return "interface"
  117. else:
  118. assert(False)
  119. def __init__(self, loader, file_id, region_id, region_name, offset_begin, offset_end, line_begin, line_end, cursor_line, group, checksum):
  120. LoadableData.__init__(self, loader, file_id, region_id)
  121. self.name = region_name
  122. self.begin = offset_begin
  123. self.end = offset_end
  124. self.line_begin = line_begin
  125. self.line_end = line_end
  126. self.cursor = cursor_line
  127. self.group = group
  128. self.checksum = checksum
  129. self.children = []
  130. def get_id(self):
  131. return self.region_id
  132. def get_name(self):
  133. return self.name
  134. def get_offset_begin(self):
  135. return self.begin
  136. def get_offset_end(self):
  137. return self.end
  138. def get_line_begin(self):
  139. return self.line_begin
  140. def get_line_end(self):
  141. return self.line_end
  142. def get_cursor(self):
  143. return self.cursor
  144. def get_type(self):
  145. return self.group
  146. def get_checksum(self):
  147. return self.checksum
  148. def register_subregion_id(self, child_id):
  149. self.children.append(child_id)
  150. def iterate_subregion_ids(self):
  151. return self.children
  152. class Marker(object):
  153. class T(object):
  154. NONE = 0x00
  155. COMMENT = 0x01
  156. STRING = 0x02
  157. PREPROCESSOR = 0x04
  158. CODE = 0x08
  159. ALL_EXCEPT_CODE = 0x07
  160. ANY = 0xFF
  161. def to_str(self, group):
  162. if group == self.NONE:
  163. return "none"
  164. elif group == self.COMMENT:
  165. return "comment"
  166. elif group == self.STRING:
  167. return "string"
  168. elif group == self.PREPROCESSOR:
  169. return "preprocessor"
  170. elif group == self.CODE:
  171. return "code"
  172. else:
  173. assert(False)
  174. def __init__(self, offset_begin, offset_end, group):
  175. self.begin = offset_begin
  176. self.end = offset_end
  177. self.group = group
  178. def get_offset_begin(self):
  179. return self.begin
  180. def get_offset_end(self):
  181. return self.end
  182. def get_type(self):
  183. return self.group
  184. class FileData(LoadableData):
  185. def __init__(self, loader, path, file_id, checksum, content):
  186. LoadableData.__init__(self, loader, file_id, None)
  187. self.path = path
  188. self.checksum = checksum
  189. self.content = content
  190. self.regions = None
  191. self.markers = None
  192. self.loader = loader
  193. self.loading_tmp = []
  194. def get_id(self):
  195. return self.file_id
  196. def get_path(self):
  197. return self.path
  198. def get_checksum(self):
  199. return self.checksum
  200. def get_content(self):
  201. return self.content
  202. def _internal_append_region(self, region):
  203. # here we apply some magic - we rely on special ordering of coming regions,
  204. # which is supported by code parsers
  205. prev_id = None
  206. while True:
  207. if len(self.loading_tmp) == 0:
  208. break
  209. prev_id = self.loading_tmp.pop()
  210. if self.get_region(prev_id).get_offset_end() > region.get_offset_begin():
  211. self.loading_tmp.append(prev_id) # return back
  212. break
  213. self.loading_tmp.append(region.get_id())
  214. if prev_id != None:
  215. self.get_region(prev_id).register_subregion_id(region.get_id())
  216. self.regions.append(region)
  217. def load_regions(self):
  218. if self.regions == None:
  219. self.regions = []
  220. for each in self.loader.db.iterate_regions(self.get_id()):
  221. self._internal_append_region(Region(self.loader,
  222. self.get_id(),
  223. each.region_id,
  224. each.name,
  225. each.begin,
  226. each.end,
  227. each.line_begin,
  228. each.line_end,
  229. each.cursor,
  230. each.group,
  231. each.checksum))
  232. assert(len(self.regions) == each.region_id)
  233. def add_region(self, region_name, offset_begin, offset_end, line_begin, line_end, cursor_line, group, checksum):
  234. if self.regions == None:
  235. self.regions = [] # do not load in time of collection
  236. new_id = len(self.regions) + 1
  237. self._internal_append_region(Region(self.loader, self.get_id(), new_id, region_name, offset_begin, offset_end, line_begin, line_end, cursor_line, group, checksum))
  238. self.loader.db.create_region(self.file_id, new_id, region_name, offset_begin, offset_end, line_begin, line_end, cursor_line, group, checksum)
  239. return new_id
  240. def get_region(self, region_id):
  241. self.load_regions()
  242. return self.regions[region_id - 1]
  243. def get_region_types(self):
  244. return Region.T
  245. def iterate_regions(self, filter_group = Region.T.ANY):
  246. self.load_regions()
  247. for each in self.regions:
  248. if each.group & filter_group:
  249. yield each
  250. def are_regions_loaded(self):
  251. return self.regions != None
  252. def load_markers(self):
  253. if self.markers == None:
  254. self.markers = []
  255. for each in self.loader.db.iterate_markers(self.get_id()):
  256. self.markers.append(Marker(each.begin, each.end, each.group))
  257. def add_marker(self, offset_begin, offset_end, group):
  258. if self.markers == None:
  259. self.markers = [] # do not load in time of collection
  260. self.markers.append(Marker(offset_begin, offset_end, group))
  261. self.loader.db.create_marker(self.file_id, offset_begin, offset_end, group)
  262. def iterate_markers(self, filter_group = Marker.T.ANY,
  263. region_id = None, exclude_children = True, merge = False):
  264. self.load_markers()
  265. # merged markers
  266. if merge == True:
  267. next_marker = None
  268. for marker in self.iterate_markers(filter_group, region_id, exclude_children, merge = False):
  269. if next_marker != None:
  270. if next_marker.get_offset_end() == marker.get_offset_begin():
  271. # sequential markers
  272. next_marker = Marker(next_marker.get_offset_begin(),
  273. marker.get_offset_end(),
  274. next_marker.get_type() | marker.get_type())
  275. else:
  276. yield next_marker
  277. next_marker = None
  278. if next_marker == None:
  279. next_marker = Marker(marker.get_offset_begin(),
  280. marker.get_offset_end(),
  281. marker.get_type())
  282. if next_marker != None:
  283. yield next_marker
  284. # all markers per file
  285. elif region_id == None:
  286. next_code_marker_start = 0
  287. for marker in self.markers:
  288. if Marker.T.CODE & filter_group and next_code_marker_start < marker.get_offset_begin():
  289. yield Marker(next_code_marker_start, marker.get_offset_begin(), Marker.T.CODE)
  290. if marker.group & filter_group:
  291. yield marker
  292. next_code_marker_start = marker.get_offset_end()
  293. if Marker.T.CODE & filter_group and next_code_marker_start < len(self.get_content()):
  294. yield Marker(next_code_marker_start, len(self.get_content()), Marker.T.CODE)
  295. # markers per region
  296. else:
  297. region = self.get_region(region_id)
  298. if region != None:
  299. # code parsers and database know about non-code markers
  300. # clients want to iterate code as markers as well
  301. # so, we embed code markers in run-time
  302. class CodeMarker(Marker):
  303. pass
  304. # cache markers for all regions if it does not exist
  305. if hasattr(region, '_markers_list') == False:
  306. # subroutine to populate _markers_list attribute
  307. # _markers_list does include code markers
  308. def cache_markers_list_rec(data, region_id, marker_start_ind, next_code_marker_start):
  309. region = data.get_region(region_id)
  310. region._markers_list = []
  311. region._first_marker_ind = marker_start_ind
  312. #next_code_marker_start = region.get_offset_begin()
  313. for sub_id in region.iterate_subregion_ids():
  314. subregion = data.get_region(sub_id)
  315. # cache all markers before the subregion
  316. while len(data.markers) > marker_start_ind and \
  317. subregion.get_offset_begin() > data.markers[marker_start_ind].get_offset_begin():
  318. if next_code_marker_start < data.markers[marker_start_ind].get_offset_begin():
  319. # append code markers coming before non-code marker
  320. region._markers_list.append(CodeMarker(next_code_marker_start,
  321. data.markers[marker_start_ind].get_offset_begin(),
  322. Marker.T.CODE))
  323. next_code_marker_start = data.markers[marker_start_ind].get_offset_end()
  324. region._markers_list.append(marker_start_ind)
  325. marker_start_ind += 1
  326. # cache all code markers before the subregion but after the last marker
  327. if next_code_marker_start < subregion.get_offset_begin():
  328. region._markers_list.append(CodeMarker(next_code_marker_start,
  329. subregion.get_offset_begin(),
  330. Marker.T.CODE))
  331. next_code_marker_start = subregion.get_offset_begin()
  332. # here is the recursive call for all sub-regions
  333. (marker_start_ind, next_code_marker_start) = cache_markers_list_rec(data,
  334. sub_id,
  335. marker_start_ind,
  336. next_code_marker_start)
  337. # cache all markers after the last subregion
  338. while len(data.markers) > marker_start_ind and \
  339. region.get_offset_end() > data.markers[marker_start_ind].get_offset_begin():
  340. # append code markers coming before non-code marker
  341. if next_code_marker_start < data.markers[marker_start_ind].get_offset_begin():
  342. region._markers_list.append(CodeMarker(next_code_marker_start,
  343. data.markers[marker_start_ind].get_offset_begin(),
  344. Marker.T.CODE))
  345. next_code_marker_start = data.markers[marker_start_ind].get_offset_end()
  346. region._markers_list.append(marker_start_ind)
  347. marker_start_ind += 1
  348. # cache the last code segment after the last marker
  349. if next_code_marker_start < region.get_offset_end():
  350. region._markers_list.append(CodeMarker(next_code_marker_start,
  351. region.get_offset_end(),
  352. Marker.T.CODE))
  353. next_code_marker_start = region.get_offset_end()
  354. # return the starting point for the next call of this function
  355. return (marker_start_ind, next_code_marker_start)
  356. # append markers list to all regions recursively
  357. (next_marker_pos, next_code_marker_start) = cache_markers_list_rec(self, 1, 0, 0)
  358. assert(next_marker_pos == len(self.markers))
  359. # excluding subregions
  360. if exclude_children == True:
  361. for marker_ind in region._markers_list:
  362. if isinstance(marker_ind, int):
  363. marker = self.markers[marker_ind]
  364. else:
  365. marker = marker_ind # CodeMarker
  366. if marker.group & filter_group:
  367. yield marker
  368. # including subregions
  369. else:
  370. next_code_marker_start = region.get_offset_begin()
  371. for marker in self.markers[region._first_marker_ind:]:
  372. if marker.get_offset_begin() >= region.get_offset_end():
  373. break
  374. if region.get_offset_begin() > marker.get_offset_begin():
  375. continue
  376. if Marker.T.CODE & filter_group and next_code_marker_start < marker.get_offset_begin():
  377. yield Marker(next_code_marker_start, marker.get_offset_begin(), Marker.T.CODE)
  378. if marker.group & filter_group:
  379. yield marker
  380. next_code_marker_start = marker.get_offset_end()
  381. if Marker.T.CODE & filter_group and next_code_marker_start < region.get_offset_end():
  382. yield Marker(next_code_marker_start, region.get_offset_end(), Marker.T.CODE)
  383. def get_marker_types(self):
  384. return Marker.T
  385. def are_markers_loaded(self):
  386. return self.markers != None
  387. def __repr__(self):
  388. return Data.__repr__(self) + " and regions " + self.regions.__repr__()
  389. class AggregatedData(Data):
  390. def __init__(self, loader, path):
  391. Data.__init__(self)
  392. self.path = path
  393. self.loader = loader
  394. self.subdirs = None
  395. self.subfiles = None
  396. def get_subdirs(self):
  397. if self.subdirs != None:
  398. return self.subdirs
  399. self.subdirs = []
  400. if self.path != None:
  401. for subdir in self.loader.db.iterate_dircontent(self.path, include_subdirs = True, include_subfiles = False):
  402. self.subdirs.append(subdir)
  403. return self.subdirs
  404. def get_subfiles(self):
  405. if self.subfiles != None:
  406. return self.subfiles
  407. self.subfiles = []
  408. if self.path != None:
  409. for subfile in self.loader.db.iterate_dircontent(self.path, include_subdirs = False, include_subfiles = True):
  410. self.subfiles.append(subfile)
  411. return self.subfiles
  412. class SelectData(Data):
  413. def __init__(self, loader, path, file_id, region_id):
  414. Data.__init__(self)
  415. self.loader = loader
  416. self.path = path
  417. self.file_id = file_id
  418. self.region_id = region_id
  419. self.region = None
  420. def get_path(self):
  421. return self.path
  422. def get_region(self):
  423. if self.region == None and self.region_id != None:
  424. row = self.loader.db.get_region(self.file_id, self.region_id)
  425. if row != None:
  426. self.region = Region(self.loader,
  427. self.file_id,
  428. self.region_id,
  429. row.name,
  430. row.begin,
  431. row.end,
  432. row.line_begin,
  433. row.line_end,
  434. row.cursor,
  435. row.group,
  436. row.checksum)
  437. return self.region
  438. class DiffData(Data):
  439. def __init__(self, new_data, old_data):
  440. Data.__init__(self)
  441. self.new_data = new_data
  442. self.old_data = old_data
  443. def get_data(self, namespace, field):
  444. new_data = self.new_data.get_data(namespace, field)
  445. old_data = self.old_data.get_data(namespace, field)
  446. if new_data == None:
  447. return None
  448. if old_data == None:
  449. # non_zero fields has got zero value by default if missed
  450. # the data can be also unavailable,
  451. # because previous collection does not include that
  452. # but external tools (like limit.py) should warn about this,
  453. # using list of registered database properties
  454. old_data = 0
  455. return new_data - old_data
  456. ####################################
  457. # Packager Interface
  458. ####################################
  459. class PackagerError(Exception):
  460. def __init__(self):
  461. Exception.__init__(self, "Failed to pack or unpack.")
  462. class PackagerFactory(object):
  463. def create(self, python_type, non_zero):
  464. if python_type == None:
  465. return PackagerFactory.SkipPackager()
  466. if python_type == int:
  467. if non_zero == False:
  468. return PackagerFactory.IntPackager()
  469. else:
  470. return PackagerFactory.IntNonZeroPackager()
  471. if python_type == float and non_zero == False:
  472. return PackagerFactory.FloatPackager()
  473. if python_type == str:
  474. return PackagerFactory.StringPackager()
  475. class PackagerFactoryError(Exception):
  476. def __init__(self, python_type):
  477. Exception.__init__(self, "Python type '" + str(python_type) + "' is not supported by the factory.")
  478. raise PackagerFactoryError(python_type)
  479. def get_python_type(self, sql_type):
  480. if sql_type == "integer":
  481. return int
  482. if sql_type == "real":
  483. return float
  484. if sql_type == "text":
  485. return str
  486. class PackagerFactoryError(Exception):
  487. def __init__(self, sql_type):
  488. Exception.__init__(self, "SQL type '" + str(sql_type) + "' is not supported by the factory.")
  489. raise PackagerFactoryError(sql_type)
  490. class IPackager(object):
  491. def pack(self, unpacked_data):
  492. raise mpp.api.InterfaceNotImplemented(self)
  493. def unpack(self, packed_data):
  494. raise mpp.api.InterfaceNotImplemented(self)
  495. def get_sql_type(self):
  496. raise mpp.api.InterfaceNotImplemented(self)
  497. def get_python_type(self):
  498. raise mpp.api.InterfaceNotImplemented(self)
  499. def is_non_zero(self):
  500. return False
  501. class IntPackager(IPackager):
  502. def pack(self, unpacked_data):
  503. if not isinstance(unpacked_data, int):
  504. raise PackagerError()
  505. return str(unpacked_data)
  506. def unpack(self, packed_data):
  507. try:
  508. return int(packed_data)
  509. except ValueError:
  510. raise PackagerError()
  511. def get_sql_type(self):
  512. return "integer"
  513. def get_python_type(self):
  514. return int
  515. class IntNonZeroPackager(IntPackager):
  516. def pack(self, unpacked_data):
  517. if unpacked_data == 0:
  518. raise PackagerError()
  519. return PackagerFactory.IntPackager.pack(self, unpacked_data)
  520. def is_non_zero(self):
  521. return True
  522. class FloatPackager(IPackager):
  523. def pack(self, unpacked_data):
  524. if not isinstance(unpacked_data, float):
  525. raise PackagerError()
  526. return str(unpacked_data)
  527. def unpack(self, packed_data):
  528. try:
  529. return float(packed_data)
  530. except ValueError:
  531. raise PackagerError()
  532. def get_sql_type(self):
  533. return "real"
  534. def get_python_type(self):
  535. return float
  536. class FloatNonZeroPackager(FloatPackager):
  537. def pack(self, unpacked_data):
  538. if unpacked_data == 0:
  539. raise PackagerError()
  540. return PackagerFactory.FloatPackager.pack(self, unpacked_data)
  541. def is_non_zero(self):
  542. return True
  543. class StringPackager(IPackager):
  544. def pack(self, unpacked_data):
  545. if not isinstance(unpacked_data, str):
  546. raise PackagerError()
  547. return str(unpacked_data)
  548. def unpack(self, packed_data):
  549. try:
  550. return str(packed_data)
  551. except ValueError:
  552. raise PackagerError()
  553. def get_sql_type(self):
  554. return "text"
  555. def get_python_type(self):
  556. return str
  557. class SkipPackager(IPackager):
  558. def pack(self, unpacked_data):
  559. return None
  560. def unpack(self, packed_data):
  561. return None
  562. def get_sql_type(self):
  563. return None
  564. def get_python_type(self):
  565. return None
  566. ####################################
  567. # Loader
  568. ####################################
  569. class NamespaceError(Exception):
  570. def __init__(self, namespace, reason):
  571. Exception.__init__(self, "Namespace '"
  572. + namespace
  573. + "': '"
  574. + reason
  575. + "'")
  576. class FieldError(Exception):
  577. def __init__(self, field, reason):
  578. Exception.__init__(self, "Field '"
  579. + field
  580. + "': '"
  581. + reason
  582. + "'")
  583. class Namespace(object):
  584. def __init__(self, db_handle, name, support_regions = False, version='1.0'):
  585. if not isinstance(name, str):
  586. raise NamespaceError(name, "name not a string")
  587. self.name = name
  588. self.support_regions = support_regions
  589. self.fields = {}
  590. self.db = db_handle
  591. if self.db.check_table(name) == False:
  592. self.db.create_table(name, support_regions, version)
  593. else:
  594. for column in self.db.iterate_columns(name):
  595. self.add_field(column.name, PackagerFactory().get_python_type(column.sql_type), non_zero=column.non_zero)
  596. def get_name(self):
  597. return self.name
  598. def are_regions_supported(self):
  599. return self.support_regions
  600. def add_field(self, field_name, python_type, non_zero=False):
  601. if not isinstance(field_name, str):
  602. raise FieldError(field_name, "field_name not a string")
  603. packager = PackagerFactory().create(python_type, non_zero)
  604. if field_name in self.fields.keys():
  605. raise FieldError(field_name, "double used")
  606. self.fields[field_name] = packager
  607. if self.db.check_column(self.get_name(), field_name) == False:
  608. # - False if cloned
  609. # - True if created
  610. return self.db.create_column(self.name, field_name, packager.get_sql_type(), non_zero=non_zero)
  611. return None # if double request
  612. def iterate_field_names(self):
  613. for name in self.fields.keys():
  614. yield name
  615. def get_field_packager(self, field_name):
  616. if field_name in self.fields.keys():
  617. return self.fields[field_name]
  618. else:
  619. return None
  620. def get_field_sql_type(self, field_name):
  621. return self.get_field_packager(field_name).get_sql_type()
  622. def get_field_python_type(self, field_name):
  623. return self.get_field_packager(field_name).get_python_type()
  624. class DataNotPackable(Exception):
  625. def __init__(self, namespace, field, value, packager, extra_message):
  626. Exception.__init__(self, "Data '"
  627. + str(value)
  628. + "' of type "
  629. + str(value.__class__)
  630. + " referred by '"
  631. + namespace
  632. + "=>"
  633. + field
  634. + "' is not packable by registered packager '"
  635. + str(packager.__class__)
  636. + "': " + extra_message)
  637. class Loader(object):
  638. def __init__(self):
  639. self.namespaces = {}
  640. self.db = None
  641. self.last_file_data = None # for performance boost reasons
  642. def create_database(self, dbfile, previous_db = None):
  643. self.db = mpp.internal.dbwrap.Database()
  644. try:
  645. self.db.create(dbfile, clone_from=previous_db)
  646. except:
  647. return False
  648. return True
  649. def open_database(self, dbfile, read_only = True):
  650. self.db = mpp.internal.dbwrap.Database()
  651. if os.path.exists(dbfile) == False:
  652. return False
  653. try:
  654. self.db.connect(dbfile, read_only=read_only)
  655. except:
  656. return False
  657. for table in self.db.iterate_tables():
  658. self.create_namespace(table.name, table.support_regions)
  659. return True
  660. def set_property(self, property_name, value):
  661. if self.db == None:
  662. return None
  663. return self.db.set_property(property_name, value)
  664. def get_property(self, property_name):
  665. if self.db == None:
  666. return None
  667. return self.db.get_property(property_name)
  668. def iterate_properties(self):
  669. if self.db == None:
  670. return None
  671. return self.db.iterate_properties()
  672. def create_namespace(self, name, support_regions = False, version='1.0'):
  673. if self.db == None:
  674. return None
  675. if name in self.namespaces.keys():
  676. raise NamespaceError(name, "double used")
  677. new_namespace = Namespace(self.db, name, support_regions, version)
  678. self.namespaces[name] = new_namespace
  679. return new_namespace
  680. def iterate_namespace_names(self):
  681. for name in self.namespaces.keys():
  682. yield name
  683. def get_namespace(self, name):
  684. if name in self.namespaces.keys():
  685. return self.namespaces[name]
  686. else:
  687. return None
  688. def create_file_data(self, path, checksum, content):
  689. if self.db == None:
  690. return None
  691. (new_id, is_updated) = self.db.create_file(path, checksum)
  692. result = FileData(self, path, new_id, checksum, content)
  693. self.last_file_data = result
  694. return (result, is_updated)
  695. def load_file_data(self, path):
  696. if self.db == None:
  697. return None
  698. if self.last_file_data != None and self.last_file_data.get_path() == path:
  699. return self.last_file_data
  700. data = self.db.get_file(path)
  701. if data == None:
  702. return None
  703. result = FileData(self, data.path, data.id, data.checksum, None)
  704. self.last_file_data = result
  705. return result
  706. def save_file_data(self, file_data):
  707. if self.db == None:
  708. return None
  709. class DataIterator(object):
  710. def iterate_packed_values(self, data, namespace, support_regions = False):
  711. for each in data.iterate_fields(namespace):
  712. space = self.loader.get_namespace(namespace)
  713. if space == None:
  714. raise DataNotPackable(namespace, each[0], each[1], None, "The namespace has not been found")
  715. packager = space.get_field_packager(each[0])
  716. if packager == None:
  717. raise DataNotPackable(namespace, each[0], each[1], None, "The field has not been found")
  718. if space.support_regions != support_regions:
  719. raise DataNotPackable(namespace, each[0], each[1], packager, "Incompatible support for regions")
  720. try:
  721. packed_data = packager.pack(each[1])
  722. if packed_data == None:
  723. continue
  724. except PackagerError:
  725. raise DataNotPackable(namespace, each[0], each[1], packager, "Packager raised exception")
  726. yield (each[0], packed_data)
  727. def __init__(self, loader, data, namespace, support_regions = False):
  728. self.loader = loader
  729. self.iterator = self.iterate_packed_values(data, namespace, support_regions)
  730. def __iter__(self):
  731. return self.iterator
  732. for namespace in file_data.iterate_namespaces():
  733. if file_data.is_namespace_updated(namespace) == False:
  734. continue
  735. self.db.add_row(namespace,
  736. file_data.get_id(),
  737. None,
  738. DataIterator(self, file_data, namespace))
  739. if file_data.are_regions_loaded():
  740. for region in file_data.iterate_regions():
  741. for namespace in region.iterate_namespaces():
  742. if region.is_namespace_updated(namespace) == False:
  743. continue
  744. self.db.add_row(namespace,
  745. file_data.get_id(),
  746. region.get_id(),
  747. DataIterator(self, region, namespace, support_regions = True))
  748. def iterate_file_data(self, path = None, path_like_filter = "%"):
  749. if self.db == None:
  750. return None
  751. final_path_like = path_like_filter
  752. if path != None:
  753. if self.db.check_dir(path) == False and self.db.check_file(path) == False:
  754. return None
  755. final_path_like = path + path_like_filter
  756. class FileDataIterator(object):
  757. def iterate_file_data(self, loader, final_path_like):
  758. for data in loader.db.iterate_files(path_like=final_path_like):
  759. yield FileData(loader, data.path, data.id, data.checksum, None)
  760. def __init__(self, loader, final_path_like):
  761. self.iterator = self.iterate_file_data(loader, final_path_like)
  762. def __iter__(self):
  763. return self.iterator
  764. if self.db == None:
  765. return None
  766. return FileDataIterator(self, final_path_like)
  767. def load_aggregated_data(self, path = None, path_like_filter = "%", namespaces = None):
  768. if self.db == None:
  769. return None
  770. final_path_like = path_like_filter
  771. if path != None:
  772. if self.db.check_dir(path) == False and self.db.check_file(path) == False:
  773. return None
  774. final_path_like = path + path_like_filter
  775. if namespaces == None:
  776. namespaces = self.namespaces.keys()
  777. result = AggregatedData(self, path)
  778. for name in namespaces:
  779. namespace = self.get_namespace(name)
  780. data = self.db.aggregate_rows(name, path_like = final_path_like)
  781. for field in data.keys():
  782. if namespace.get_field_packager(field).get_python_type() == str:
  783. continue
  784. if namespace.get_field_packager(field).is_non_zero() == True:
  785. data[field]['min'] = None
  786. data[field]['avg'] = None
  787. distribution = self.db.count_rows(name, path_like = final_path_like, group_by_column = field)
  788. data[field]['distribution-bars'] = []
  789. for each in distribution:
  790. if each[0] == None:
  791. continue
  792. assert(float(data[field]['count'] != 0))
  793. data[field]['distribution-bars'].append({'metric': each[0],
  794. 'count': each[1],
  795. 'ratio': round((float(each[1]) / float(data[field]['count'])), 4)})
  796. result.set_data(name, field, data[field])
  797. return result
  798. def load_selected_data(self, namespace, fields = None, path = None, path_like_filter = "%", filters = [],
  799. sort_by = None, limit_by = None):
  800. if self.db == None:
  801. return None
  802. final_path_like = path_like_filter
  803. if path != None:
  804. if self.db.check_dir(path) == False and self.db.check_file(path) == False:
  805. return None
  806. final_path_like = path + path_like_filter
  807. namespace_obj = self.get_namespace(namespace)
  808. if namespace_obj == None:
  809. return None
  810. class SelectDataIterator(object):
  811. def iterate_selected_values(self, loader, namespace_obj, final_path_like, fields, filters, sort_by, limit_by):
  812. for row in loader.db.select_rows(namespace_obj.get_name(), path_like=final_path_like, filters=filters,
  813. order_by=sort_by, limit_by=limit_by):
  814. region_id = None
  815. if namespace_obj.are_regions_supported() == True:
  816. region_id = row['region_id']
  817. data = SelectData(loader, row['path'], row['id'], region_id)
  818. field_names = fields
  819. if fields == None:
  820. field_names = namespace_obj.iterate_field_names()
  821. for field in field_names:
  822. data.set_data(namespace, field, row[field])
  823. yield data
  824. def __init__(self, loader, namespace_obj, final_path_like, fields, filters, sort_by, limit_by):
  825. self.iterator = self.iterate_selected_values(loader, namespace_obj, final_path_like, fields, filters, sort_by, limit_by)
  826. def __iter__(self):
  827. return self.iterator
  828. return SelectDataIterator(self, namespace_obj, final_path_like, fields, filters, sort_by, limit_by)
  829. class BasePlugin(object):
  830. def initialize(self):
  831. pass
  832. def terminate(self):
  833. pass
  834. def set_name(self, name):
  835. self.name = name
  836. def get_name(self):
  837. if hasattr(self, 'name') == False:
  838. return None
  839. return self.name
  840. def set_version(self, version):
  841. self.version = version
  842. def get_version(self):
  843. if hasattr(self, 'version') == False:
  844. return None
  845. return self.version
  846. def set_plugin_loader(self, loader):
  847. self.plugin_loader = loader
  848. def get_plugin_loader(self):
  849. if hasattr(self, 'plugin_loader') == False:
  850. return None
  851. return self.plugin_loader
  852. class Plugin(BasePlugin):
  853. class Field(object):
  854. def __init__(self, name, ftype, non_zero=False):
  855. self.name = name
  856. self.type = ftype
  857. self.non_zero = non_zero
  858. class Property(object):
  859. def __init__(self, name, value):
  860. self.name = name
  861. self.value = value
  862. def initialize(self, namespace=None, support_regions=True, fields=[], properties=[]):
  863. super(Plugin, self).initialize()
  864. if hasattr(self, 'is_updated') == False:
  865. self.is_updated = False # original initialization
  866. db_loader = self.get_plugin_loader().get_plugin('mpp.dbf').get_loader()
  867. if namespace == None:
  868. namespace = self.get_name()
  869. if (len(fields) != 0 or len(properties) != 0):
  870. prev_version = db_loader.set_property(self.get_name() + ":version", self.get_version())
  871. if prev_version != self.get_version():
  872. self.is_updated = True
  873. for prop in properties:
  874. assert(prop.name != 'version')
  875. prev_prop = db_loader.set_property(self.get_name() + ":" + prop.name, prop.value)
  876. if prev_prop != prop.value:
  877. self.is_updated = True
  878. if len(fields) != 0:
  879. namespace_obj = db_loader.create_namespace(namespace,
  880. support_regions=support_regions,
  881. version=self.get_version())
  882. for field in fields:
  883. is_created = namespace_obj.add_field(field.name, field.type, non_zero=field.non_zero)
  884. assert(is_created != None)
  885. # if field is created (not cloned from the previous db),
  886. # mark the plug-in as updated in order to trigger full rescan
  887. self.is_updated = self.is_updated or is_created
  888. class SimpleMetricMixin(object):
  889. class AliasError(Exception):
  890. def __init__(self, alias):
  891. Exception.__init__(self, "Unknown alias: " + str(alias))
  892. def declare_metric(self, is_active, field,
  893. pattern_to_search_or_map_of_patterns,
  894. marker_type_mask=Marker.T.ANY,
  895. region_type_mask=Region.T.ANY,
  896. exclude_subregions=True,
  897. merge_markers=False):
  898. if hasattr(self, '_fields') == False:
  899. self._fields = {}
  900. if isinstance(pattern_to_search_or_map_of_patterns, dict):
  901. map_of_patterns = pattern_to_search_or_map_of_patterns
  902. else:
  903. map_of_patterns = {'default': pattern_to_search_or_map_of_patterns}
  904. if is_active == True:
  905. self._fields[field.name] = (field,
  906. marker_type_mask,
  907. exclude_subregions,
  908. merge_markers,
  909. map_of_patterns,
  910. region_type_mask)
  911. def is_active(self, metric_name = None):
  912. if metric_name == None:
  913. return (len(self._fields.keys()) > 0)
  914. return (metric_name in self._fields.keys())
  915. def get_fields(self):
  916. result = []
  917. for key in self._fields.keys():
  918. result.append(self._fields[key][0])
  919. return result
  920. def count_if_active(self, metric_name, data, namespace=None, alias='default'):
  921. if self.is_active(metric_name) == False:
  922. return
  923. if namespace == None:
  924. namespace = self.get_name()
  925. field_data = self._fields[metric_name]
  926. text = data.get_content()
  927. if alias not in field_data[4].keys():
  928. raise self.AliasError(alias)
  929. pattern_to_search = field_data[4][alias]
  930. for region in data.iterate_regions(filter_group=field_data[5]):
  931. count = 0
  932. for marker in data.iterate_markers(
  933. filter_group = field_data[1],
  934. region_id = region.get_id(),
  935. exclude_children = field_data[2],
  936. merge=field_data[3]):
  937. count += len(pattern_to_search.findall(text, marker.get_offset_begin(), marker.get_offset_end()))
  938. region.set_data(namespace, metric_name, count)
  939. class InterfaceNotImplemented(Exception):
  940. def __init__(self, obj):
  941. import sys
  942. Exception.__init__(self, "Method '"
  943. + sys._getframe(1).f_code.co_name
  944. + "' has not been implemented for "
  945. + str(obj.__class__))
  946. class IConfigurable(object):
  947. def configure(self, options):
  948. raise InterfaceNotImplemented(self)
  949. def declare_configuration(self, optparser):
  950. raise InterfaceNotImplemented(self)
  951. class IRunable(object):
  952. def run(self, args):
  953. raise InterfaceNotImplemented(self)
  954. class IParser(object):
  955. def process(self, parent, data, is_updated):
  956. raise InterfaceNotImplemented(self)
  957. class ICode(object):
  958. pass
  959. class CallbackNotImplemented(Exception):
  960. def __init__(self, obj, callback_name):
  961. Exception.__init__(self, "Callback '"
  962. + callback_name
  963. + "' has not been implemented for "
  964. + str(obj.__class__))
  965. class Child(object):
  966. def notify(self, parent, callback_name, *args):
  967. if hasattr(self, callback_name) == False:
  968. raise CallbackNotImplemented(self, callback_name)
  969. self.__getattribute__(callback_name)(parent, *args)
  970. def subscribe_by_parents_name(self, parent_name, callback_name='callback'):
  971. self.get_plugin_loader().get_plugin(parent_name).subscribe(self, callback_name)
  972. def subscribe_by_parents_interface(self, interface, callback_name='callback'):
  973. for plugin in self.get_plugin_loader().iterate_plugins():
  974. if isinstance(plugin, interface):
  975. plugin.subscribe(self, callback_name)
  976. class Parent(object):
  977. def init_Parent(self):
  978. if hasattr(self, 'children') == False:
  979. self.children = []
  980. def subscribe(self, obj, callback_name):
  981. self.init_Parent()
  982. if (isinstance(obj, Child) == False):
  983. raise TypeError()
  984. self.children.append((obj,callback_name))
  985. def unsubscribe(self, obj, callback_name):
  986. self.init_Parent()
  987. self.children.remove((obj, callback_name))
  988. def notify_children(self, *args):
  989. self.init_Parent()
  990. for child in self.children:
  991. child[0].notify(self, child[1], *args)
  992. def iterate_children(self):
  993. self.init_Parent()
  994. for child in self.children:
  995. yield child