api.py 48 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229
  1. #
  2. # Metrix++, Copyright 2009-2013, Metrix++ Project
  3. # Link: http://metrixplusplus.sourceforge.net
  4. #
  5. # This file is a part of Metrix++ Tool.
  6. #
  7. # Metrix++ is free software: you can redistribute it and/or modify
  8. # it under the terms of the GNU General Public License as published by
  9. # the Free Software Foundation, version 3 of the License.
  10. #
  11. # Metrix++ is distributed in the hope that it will be useful,
  12. # but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  14. # GNU General Public License for more details.
  15. #
  16. # You should have received a copy of the GNU General Public License
  17. # along with Metrix++. If not, see <http://www.gnu.org/licenses/>.
  18. #
  19. import os.path
  20. import mpp.internal.dbwrap
  21. ##############################################################################
  22. #
  23. #
  24. #
  25. ##############################################################################
  26. class Data(object):
  27. def __init__(self):
  28. self.data = {}
  29. def get_data(self, namespace, field):
  30. if namespace not in self.data.keys():
  31. return None
  32. if field not in self.data[namespace].keys():
  33. return None
  34. return self.data[namespace][field]
  35. def set_data(self, namespace, field, value):
  36. if namespace not in self.data:
  37. self.data[namespace] = {}
  38. self.data[namespace][field] = value
  39. def iterate_namespaces(self):
  40. for namespace in self.data.keys():
  41. yield namespace
  42. def iterate_fields(self, namespace):
  43. for field in self.data[namespace].keys():
  44. yield (field, self.data[namespace][field])
  45. def get_data_tree(self, namespaces=None):
  46. return self.data
  47. def __repr__(self):
  48. return object.__repr__(self) + " with data " + self.data.__repr__()
  49. class LoadableData(Data):
  50. def __init__(self, loader, file_id, region_id):
  51. Data.__init__(self)
  52. self.loader = loader
  53. self.file_id = file_id
  54. self.region_id = region_id
  55. self.loaded_namespaces = []
  56. self.changed_namespaces = []
  57. def load_namespace(self, namespace):
  58. namespace_obj = self.loader.get_namespace(namespace)
  59. if namespace_obj == None:
  60. return
  61. regions_supported = namespace_obj.are_regions_supported()
  62. if ((self.region_id == None and regions_supported == True) or
  63. (self.region_id != None and regions_supported == False)):
  64. return
  65. row = self.loader.db.get_row(namespace, self.file_id, self.region_id)
  66. if row == None:
  67. return
  68. for column_name in row.keys():
  69. packager = namespace_obj.get_field_packager(column_name)
  70. if packager == None:
  71. continue
  72. if row[column_name] == None:
  73. continue
  74. Data.set_data(self, namespace, column_name, packager.unpack(row[column_name]))
  75. def set_data(self, namespace, field, value):
  76. if namespace not in self.changed_namespaces:
  77. self.changed_namespaces.append(namespace)
  78. return Data.set_data(self, namespace, field, value)
  79. def get_data(self, namespace, field):
  80. if namespace not in self.loaded_namespaces:
  81. self.loaded_namespaces.append(namespace)
  82. self.load_namespace(namespace)
  83. return Data.get_data(self, namespace, field)
  84. def is_namespace_updated(self, namespace):
  85. return namespace in self.changed_namespaces
  86. def is_namespace_loaded(self, namespace):
  87. return namespace in self.loaded_namespaces
  88. def get_data_tree(self, namespaces=None):
  89. if namespaces == None:
  90. namespaces = self.loader.iterate_namespace_names()
  91. for each in namespaces:
  92. self.load_namespace(each)
  93. return Data.get_data_tree(self)
  94. class Region(LoadableData):
  95. class T(object):
  96. NONE = 0x00
  97. GLOBAL = 0x01
  98. CLASS = 0x02
  99. STRUCT = 0x04
  100. NAMESPACE = 0x08
  101. FUNCTION = 0x10
  102. INTERFACE = 0x20
  103. ANY = 0xFF
  104. def to_str(self, group):
  105. if group == self.NONE:
  106. return "none"
  107. elif group == self.GLOBAL:
  108. return "global"
  109. elif group == self.CLASS:
  110. return "class"
  111. elif group == self.STRUCT:
  112. return "struct"
  113. elif group == self.NAMESPACE:
  114. return "namespace"
  115. elif group == self.FUNCTION:
  116. return "function"
  117. elif group == self.INTERFACE:
  118. return "interface"
  119. else:
  120. assert(False)
  121. def __init__(self, loader, file_id, region_id, region_name, offset_begin, offset_end, line_begin, line_end, cursor_line, group, checksum):
  122. LoadableData.__init__(self, loader, file_id, region_id)
  123. self.name = region_name
  124. self.begin = offset_begin
  125. self.end = offset_end
  126. self.line_begin = line_begin
  127. self.line_end = line_end
  128. self.cursor = cursor_line
  129. self.group = group
  130. self.checksum = checksum
  131. self.children = []
  132. def get_id(self):
  133. return self.region_id
  134. def get_name(self):
  135. return self.name
  136. def get_offset_begin(self):
  137. return self.begin
  138. def get_offset_end(self):
  139. return self.end
  140. def get_line_begin(self):
  141. return self.line_begin
  142. def get_line_end(self):
  143. return self.line_end
  144. def get_cursor(self):
  145. return self.cursor
  146. def get_type(self):
  147. return self.group
  148. def get_checksum(self):
  149. return self.checksum
  150. def register_subregion_id(self, child_id):
  151. self.children.append(child_id)
  152. def iterate_subregion_ids(self):
  153. return self.children
  154. class Marker(object):
  155. class T(object):
  156. NONE = 0x00
  157. COMMENT = 0x01
  158. STRING = 0x02
  159. PREPROCESSOR = 0x04
  160. CODE = 0x08
  161. ALL_EXCEPT_CODE = 0x07
  162. ANY = 0xFF
  163. def to_str(self, group):
  164. if group == self.NONE:
  165. return "none"
  166. elif group == self.COMMENT:
  167. return "comment"
  168. elif group == self.STRING:
  169. return "string"
  170. elif group == self.PREPROCESSOR:
  171. return "preprocessor"
  172. elif group == self.CODE:
  173. return "code"
  174. else:
  175. assert(False)
  176. def __init__(self, offset_begin, offset_end, group):
  177. self.begin = offset_begin
  178. self.end = offset_end
  179. self.group = group
  180. def get_offset_begin(self):
  181. return self.begin
  182. def get_offset_end(self):
  183. return self.end
  184. def get_type(self):
  185. return self.group
  186. class FileData(LoadableData):
  187. def __init__(self, loader, path, file_id, checksum, content):
  188. LoadableData.__init__(self, loader, file_id, None)
  189. self.path = path
  190. self.checksum = checksum
  191. self.content = content
  192. self.regions = None
  193. self.markers = None
  194. self.loader = loader
  195. self.loading_tmp = []
  196. def get_id(self):
  197. return self.file_id
  198. def get_path(self):
  199. return self.path
  200. def get_checksum(self):
  201. return self.checksum
  202. def get_content(self):
  203. return self.content
  204. def _internal_append_region(self, region):
  205. # here we apply some magic - we rely on special ordering of coming regions,
  206. # which is supported by code parsers
  207. prev_id = None
  208. while True:
  209. if len(self.loading_tmp) == 0:
  210. break
  211. prev_id = self.loading_tmp.pop()
  212. if self.get_region(prev_id).get_offset_end() > region.get_offset_begin():
  213. self.loading_tmp.append(prev_id) # return back
  214. break
  215. self.loading_tmp.append(region.get_id())
  216. if prev_id != None:
  217. self.get_region(prev_id).register_subregion_id(region.get_id())
  218. self.regions.append(region)
  219. def load_regions(self):
  220. if self.regions == None:
  221. self.regions = []
  222. for each in self.loader.db.iterate_regions(self.get_id()):
  223. self._internal_append_region(Region(self.loader,
  224. self.get_id(),
  225. each.region_id,
  226. each.name,
  227. each.begin,
  228. each.end,
  229. each.line_begin,
  230. each.line_end,
  231. each.cursor,
  232. each.group,
  233. each.checksum))
  234. assert(len(self.regions) == each.region_id)
  235. def add_region(self, region_name, offset_begin, offset_end, line_begin, line_end, cursor_line, group, checksum):
  236. if self.regions == None:
  237. self.regions = [] # do not load in time of collection
  238. new_id = len(self.regions) + 1
  239. self._internal_append_region(Region(self.loader, self.get_id(), new_id, region_name, offset_begin, offset_end, line_begin, line_end, cursor_line, group, checksum))
  240. self.loader.db.create_region(self.file_id, new_id, region_name, offset_begin, offset_end, line_begin, line_end, cursor_line, group, checksum)
  241. return new_id
  242. def get_region(self, region_id):
  243. self.load_regions()
  244. return self.regions[region_id - 1]
  245. def get_region_types(self):
  246. return Region.T
  247. def iterate_regions(self, filter_group = Region.T.ANY):
  248. self.load_regions()
  249. for each in self.regions:
  250. if each.group & filter_group:
  251. yield each
  252. def are_regions_loaded(self):
  253. return self.regions != None
  254. def load_markers(self):
  255. if self.markers == None:
  256. self.markers = []
  257. for each in self.loader.db.iterate_markers(self.get_id()):
  258. self.markers.append(Marker(each.begin, each.end, each.group))
  259. def add_marker(self, offset_begin, offset_end, group):
  260. if self.markers == None:
  261. self.markers = [] # do not load in time of collection
  262. self.markers.append(Marker(offset_begin, offset_end, group))
  263. self.loader.db.create_marker(self.file_id, offset_begin, offset_end, group)
  264. def iterate_markers(self, filter_group = Marker.T.ANY,
  265. region_id = None, exclude_children = True, merge = False):
  266. self.load_markers()
  267. # merged markers
  268. if merge == True:
  269. next_marker = None
  270. for marker in self.iterate_markers(filter_group, region_id, exclude_children, merge = False):
  271. if next_marker != None:
  272. if next_marker.get_offset_end() == marker.get_offset_begin():
  273. # sequential markers
  274. next_marker = Marker(next_marker.get_offset_begin(),
  275. marker.get_offset_end(),
  276. next_marker.get_type() | marker.get_type())
  277. else:
  278. yield next_marker
  279. next_marker = None
  280. if next_marker == None:
  281. next_marker = Marker(marker.get_offset_begin(),
  282. marker.get_offset_end(),
  283. marker.get_type())
  284. if next_marker != None:
  285. yield next_marker
  286. # all markers per file
  287. elif region_id == None:
  288. next_code_marker_start = 0
  289. for marker in self.markers:
  290. if Marker.T.CODE & filter_group and next_code_marker_start < marker.get_offset_begin():
  291. yield Marker(next_code_marker_start, marker.get_offset_begin(), Marker.T.CODE)
  292. if marker.group & filter_group:
  293. yield marker
  294. next_code_marker_start = marker.get_offset_end()
  295. if Marker.T.CODE & filter_group and next_code_marker_start < len(self.get_content()):
  296. yield Marker(next_code_marker_start, len(self.get_content()), Marker.T.CODE)
  297. # markers per region
  298. else:
  299. region = self.get_region(region_id)
  300. if region != None:
  301. # code parsers and database know about non-code markers
  302. # clients want to iterate code as markers as well
  303. # so, we embed code markers in run-time
  304. class CodeMarker(Marker):
  305. pass
  306. # cache markers for all regions if it does not exist
  307. if hasattr(region, '_markers_list') == False:
  308. # subroutine to populate _markers_list attribute
  309. # _markers_list does include code markers
  310. def cache_markers_list_rec(data, region_id, marker_start_ind, next_code_marker_start):
  311. region = data.get_region(region_id)
  312. region._markers_list = []
  313. region._first_marker_ind = marker_start_ind
  314. #next_code_marker_start = region.get_offset_begin()
  315. for sub_id in region.iterate_subregion_ids():
  316. subregion = data.get_region(sub_id)
  317. # cache all markers before the subregion
  318. while len(data.markers) > marker_start_ind and \
  319. subregion.get_offset_begin() > data.markers[marker_start_ind].get_offset_begin():
  320. if next_code_marker_start < data.markers[marker_start_ind].get_offset_begin():
  321. # append code markers coming before non-code marker
  322. region._markers_list.append(CodeMarker(next_code_marker_start,
  323. data.markers[marker_start_ind].get_offset_begin(),
  324. Marker.T.CODE))
  325. next_code_marker_start = data.markers[marker_start_ind].get_offset_end()
  326. region._markers_list.append(marker_start_ind)
  327. marker_start_ind += 1
  328. # cache all code markers before the subregion but after the last marker
  329. if next_code_marker_start < subregion.get_offset_begin():
  330. region._markers_list.append(CodeMarker(next_code_marker_start,
  331. subregion.get_offset_begin(),
  332. Marker.T.CODE))
  333. next_code_marker_start = subregion.get_offset_begin()
  334. # here is the recursive call for all sub-regions
  335. (marker_start_ind, next_code_marker_start) = cache_markers_list_rec(data,
  336. sub_id,
  337. marker_start_ind,
  338. next_code_marker_start)
  339. # cache all markers after the last subregion
  340. while len(data.markers) > marker_start_ind and \
  341. region.get_offset_end() > data.markers[marker_start_ind].get_offset_begin():
  342. # append code markers coming before non-code marker
  343. if next_code_marker_start < data.markers[marker_start_ind].get_offset_begin():
  344. region._markers_list.append(CodeMarker(next_code_marker_start,
  345. data.markers[marker_start_ind].get_offset_begin(),
  346. Marker.T.CODE))
  347. next_code_marker_start = data.markers[marker_start_ind].get_offset_end()
  348. region._markers_list.append(marker_start_ind)
  349. marker_start_ind += 1
  350. # cache the last code segment after the last marker
  351. if next_code_marker_start < region.get_offset_end():
  352. region._markers_list.append(CodeMarker(next_code_marker_start,
  353. region.get_offset_end(),
  354. Marker.T.CODE))
  355. next_code_marker_start = region.get_offset_end()
  356. # return the starting point for the next call of this function
  357. return (marker_start_ind, next_code_marker_start)
  358. # append markers list to all regions recursively
  359. (next_marker_pos, next_code_marker_start) = cache_markers_list_rec(self, 1, 0, 0)
  360. assert(next_marker_pos == len(self.markers))
  361. # excluding subregions
  362. if exclude_children == True:
  363. for marker_ind in region._markers_list:
  364. if isinstance(marker_ind, int):
  365. marker = self.markers[marker_ind]
  366. else:
  367. marker = marker_ind # CodeMarker
  368. if marker.group & filter_group:
  369. yield marker
  370. # including subregions
  371. else:
  372. next_code_marker_start = region.get_offset_begin()
  373. for marker in self.markers[region._first_marker_ind:]:
  374. if marker.get_offset_begin() >= region.get_offset_end():
  375. break
  376. if region.get_offset_begin() > marker.get_offset_begin():
  377. continue
  378. if Marker.T.CODE & filter_group and next_code_marker_start < marker.get_offset_begin():
  379. yield Marker(next_code_marker_start, marker.get_offset_begin(), Marker.T.CODE)
  380. if marker.group & filter_group:
  381. yield marker
  382. next_code_marker_start = marker.get_offset_end()
  383. if Marker.T.CODE & filter_group and next_code_marker_start < region.get_offset_end():
  384. yield Marker(next_code_marker_start, region.get_offset_end(), Marker.T.CODE)
  385. def get_marker_types(self):
  386. return Marker.T
  387. def are_markers_loaded(self):
  388. return self.markers != None
  389. def __repr__(self):
  390. return Data.__repr__(self) + " and regions " + self.regions.__repr__()
  391. class AggregatedData(Data):
  392. def __init__(self, loader, path):
  393. Data.__init__(self)
  394. self.path = path
  395. self.loader = loader
  396. self.subdirs = None
  397. self.subfiles = None
  398. def get_subdirs(self):
  399. if self.subdirs != None:
  400. return self.subdirs
  401. self.subdirs = []
  402. if self.path != None:
  403. for subdir in self.loader.db.iterate_dircontent(self.path, include_subdirs = True, include_subfiles = False):
  404. self.subdirs.append(subdir)
  405. return self.subdirs
  406. def get_subfiles(self):
  407. if self.subfiles != None:
  408. return self.subfiles
  409. self.subfiles = []
  410. if self.path != None:
  411. for subfile in self.loader.db.iterate_dircontent(self.path, include_subdirs = False, include_subfiles = True):
  412. self.subfiles.append(subfile)
  413. return self.subfiles
  414. class SelectData(Data):
  415. def __init__(self, loader, path, file_id, region_id):
  416. Data.__init__(self)
  417. self.loader = loader
  418. self.path = path
  419. self.file_id = file_id
  420. self.region_id = region_id
  421. self.region = None
  422. def get_path(self):
  423. return self.path
  424. def get_region(self):
  425. if self.region == None and self.region_id != None:
  426. row = self.loader.db.get_region(self.file_id, self.region_id)
  427. if row != None:
  428. self.region = Region(self.loader,
  429. self.file_id,
  430. self.region_id,
  431. row.name,
  432. row.begin,
  433. row.end,
  434. row.line_begin,
  435. row.line_end,
  436. row.cursor,
  437. row.group,
  438. row.checksum)
  439. return self.region
  440. class DiffData(Data):
  441. def __init__(self, new_data, old_data):
  442. Data.__init__(self)
  443. self.new_data = new_data
  444. self.old_data = old_data
  445. def get_data(self, namespace, field):
  446. new_data = self.new_data.get_data(namespace, field)
  447. old_data = self.old_data.get_data(namespace, field)
  448. if new_data == None:
  449. return None
  450. if old_data == None:
  451. # non_zero fields has got zero value by default if missed
  452. # the data can be also unavailable,
  453. # because previous collection does not include that
  454. # but external tools (like limit.py) should warn about this,
  455. # using list of registered database properties
  456. old_data = 0
  457. return new_data - old_data
  458. ####################################
  459. # Packager Interface
  460. ####################################
  461. class PackagerError(Exception):
  462. def __init__(self):
  463. Exception.__init__(self, "Failed to pack or unpack.")
  464. class PackagerFactory(object):
  465. def create(self, python_type, non_zero):
  466. if python_type == None:
  467. return PackagerFactory.SkipPackager()
  468. if python_type == int:
  469. if non_zero == False:
  470. return PackagerFactory.IntPackager()
  471. else:
  472. return PackagerFactory.IntNonZeroPackager()
  473. if python_type == float and non_zero == False:
  474. return PackagerFactory.FloatPackager()
  475. if python_type == str:
  476. return PackagerFactory.StringPackager()
  477. class PackagerFactoryError(Exception):
  478. def __init__(self, python_type):
  479. Exception.__init__(self, "Python type '" + str(python_type) + "' is not supported by the factory.")
  480. raise PackagerFactoryError(python_type)
  481. def get_python_type(self, sql_type):
  482. if sql_type == "integer":
  483. return int
  484. if sql_type == "real":
  485. return float
  486. if sql_type == "text":
  487. return str
  488. class PackagerFactoryError(Exception):
  489. def __init__(self, sql_type):
  490. Exception.__init__(self, "SQL type '" + str(sql_type) + "' is not supported by the factory.")
  491. raise PackagerFactoryError(sql_type)
  492. class IPackager(object):
  493. def pack(self, unpacked_data):
  494. raise mpp.api.InterfaceNotImplemented(self)
  495. def unpack(self, packed_data):
  496. raise mpp.api.InterfaceNotImplemented(self)
  497. def get_sql_type(self):
  498. raise mpp.api.InterfaceNotImplemented(self)
  499. def get_python_type(self):
  500. raise mpp.api.InterfaceNotImplemented(self)
  501. def is_non_zero(self):
  502. return False
  503. class IntPackager(IPackager):
  504. def pack(self, unpacked_data):
  505. if not isinstance(unpacked_data, int):
  506. raise PackagerError()
  507. return str(unpacked_data)
  508. def unpack(self, packed_data):
  509. try:
  510. return int(packed_data)
  511. except ValueError:
  512. raise PackagerError()
  513. def get_sql_type(self):
  514. return "integer"
  515. def get_python_type(self):
  516. return int
  517. class IntNonZeroPackager(IntPackager):
  518. def pack(self, unpacked_data):
  519. if unpacked_data == 0:
  520. raise PackagerError()
  521. return PackagerFactory.IntPackager.pack(self, unpacked_data)
  522. def is_non_zero(self):
  523. return True
  524. class FloatPackager(IPackager):
  525. def pack(self, unpacked_data):
  526. if not isinstance(unpacked_data, float):
  527. raise PackagerError()
  528. return str(unpacked_data)
  529. def unpack(self, packed_data):
  530. try:
  531. return float(packed_data)
  532. except ValueError:
  533. raise PackagerError()
  534. def get_sql_type(self):
  535. return "real"
  536. def get_python_type(self):
  537. return float
  538. class FloatNonZeroPackager(FloatPackager):
  539. def pack(self, unpacked_data):
  540. if unpacked_data == 0:
  541. raise PackagerError()
  542. return PackagerFactory.FloatPackager.pack(self, unpacked_data)
  543. def is_non_zero(self):
  544. return True
  545. class StringPackager(IPackager):
  546. def pack(self, unpacked_data):
  547. if not isinstance(unpacked_data, str):
  548. raise PackagerError()
  549. return str(unpacked_data)
  550. def unpack(self, packed_data):
  551. try:
  552. return str(packed_data)
  553. except ValueError:
  554. raise PackagerError()
  555. def get_sql_type(self):
  556. return "text"
  557. def get_python_type(self):
  558. return str
  559. class SkipPackager(IPackager):
  560. def pack(self, unpacked_data):
  561. return None
  562. def unpack(self, packed_data):
  563. return None
  564. def get_sql_type(self):
  565. return None
  566. def get_python_type(self):
  567. return None
  568. ####################################
  569. # Loader
  570. ####################################
  571. class NamespaceError(Exception):
  572. def __init__(self, namespace, reason):
  573. Exception.__init__(self, "Namespace '"
  574. + namespace
  575. + "': '"
  576. + reason
  577. + "'")
  578. class FieldError(Exception):
  579. def __init__(self, field, reason):
  580. Exception.__init__(self, "Field '"
  581. + field
  582. + "': '"
  583. + reason
  584. + "'")
  585. class Namespace(object):
  586. def __init__(self, db_handle, name, support_regions = False, version='1.0'):
  587. if not isinstance(name, str):
  588. raise NamespaceError(name, "name not a string")
  589. self.name = name
  590. self.support_regions = support_regions
  591. self.fields = {}
  592. self.db = db_handle
  593. if self.db.check_table(name) == False:
  594. self.db.create_table(name, support_regions, version)
  595. else:
  596. for column in self.db.iterate_columns(name):
  597. self.add_field(column.name, PackagerFactory().get_python_type(column.sql_type), non_zero=column.non_zero)
  598. def get_name(self):
  599. return self.name
  600. def are_regions_supported(self):
  601. return self.support_regions
  602. def add_field(self, field_name, python_type, non_zero=False):
  603. if not isinstance(field_name, str):
  604. raise FieldError(field_name, "field_name not a string")
  605. packager = PackagerFactory().create(python_type, non_zero)
  606. if field_name in self.fields.keys():
  607. raise FieldError(field_name, "double used")
  608. self.fields[field_name] = packager
  609. if self.db.check_column(self.get_name(), field_name) == False:
  610. # - False if cloned
  611. # - True if created
  612. return self.db.create_column(self.name, field_name, packager.get_sql_type(), non_zero=non_zero)
  613. return None # if double request
  614. def iterate_field_names(self):
  615. for name in self.fields.keys():
  616. yield name
  617. def get_field_packager(self, field_name):
  618. if field_name in self.fields.keys():
  619. return self.fields[field_name]
  620. else:
  621. return None
  622. def get_field_sql_type(self, field_name):
  623. return self.get_field_packager(field_name).get_sql_type()
  624. def get_field_python_type(self, field_name):
  625. return self.get_field_packager(field_name).get_python_type()
  626. class DataNotPackable(Exception):
  627. def __init__(self, namespace, field, value, packager, extra_message):
  628. Exception.__init__(self, "Data '"
  629. + str(value)
  630. + "' of type "
  631. + str(value.__class__)
  632. + " referred by '"
  633. + namespace
  634. + "=>"
  635. + field
  636. + "' is not packable by registered packager '"
  637. + str(packager.__class__)
  638. + "': " + extra_message)
  639. class Loader(object):
  640. def __init__(self):
  641. self.namespaces = {}
  642. self.db = None
  643. self.last_file_data = None # for performance boost reasons
  644. def create_database(self, dbfile, previous_db = None):
  645. self.db = mpp.internal.dbwrap.Database()
  646. try:
  647. self.db.create(dbfile, clone_from=previous_db)
  648. except:
  649. return False
  650. return True
  651. def open_database(self, dbfile, read_only = True):
  652. self.db = mpp.internal.dbwrap.Database()
  653. if os.path.exists(dbfile) == False:
  654. return False
  655. try:
  656. self.db.connect(dbfile, read_only=read_only)
  657. except:
  658. return False
  659. for table in self.db.iterate_tables():
  660. self.create_namespace(table.name, table.support_regions)
  661. return True
  662. def set_property(self, property_name, value):
  663. if self.db == None:
  664. return None
  665. return self.db.set_property(property_name, value)
  666. def get_property(self, property_name):
  667. if self.db == None:
  668. return None
  669. return self.db.get_property(property_name)
  670. def iterate_properties(self):
  671. if self.db == None:
  672. return None
  673. return self.db.iterate_properties()
  674. def create_namespace(self, name, support_regions = False, version='1.0'):
  675. if self.db == None:
  676. return None
  677. if name in self.namespaces.keys():
  678. raise NamespaceError(name, "double used")
  679. new_namespace = Namespace(self.db, name, support_regions, version)
  680. self.namespaces[name] = new_namespace
  681. return new_namespace
  682. def iterate_namespace_names(self):
  683. for name in self.namespaces.keys():
  684. yield name
  685. def get_namespace(self, name):
  686. if name in self.namespaces.keys():
  687. return self.namespaces[name]
  688. else:
  689. return None
  690. def create_file_data(self, path, checksum, content):
  691. if self.db == None:
  692. return None
  693. (new_id, is_updated) = self.db.create_file(path, checksum)
  694. result = FileData(self, path, new_id, checksum, content)
  695. self.last_file_data = result
  696. return (result, is_updated)
  697. def load_file_data(self, path):
  698. if self.db == None:
  699. return None
  700. if self.last_file_data != None and self.last_file_data.get_path() == path:
  701. return self.last_file_data
  702. data = self.db.get_file(path)
  703. if data == None:
  704. return None
  705. result = FileData(self, data.path, data.id, data.checksum, None)
  706. self.last_file_data = result
  707. return result
  708. def save_file_data(self, file_data):
  709. if self.db == None:
  710. return None
  711. class DataIterator(object):
  712. def iterate_packed_values(self, data, namespace, support_regions = False):
  713. for each in data.iterate_fields(namespace):
  714. space = self.loader.get_namespace(namespace)
  715. if space == None:
  716. raise DataNotPackable(namespace, each[0], each[1], None, "The namespace has not been found")
  717. packager = space.get_field_packager(each[0])
  718. if packager == None:
  719. raise DataNotPackable(namespace, each[0], each[1], None, "The field has not been found")
  720. if space.support_regions != support_regions:
  721. raise DataNotPackable(namespace, each[0], each[1], packager, "Incompatible support for regions")
  722. try:
  723. packed_data = packager.pack(each[1])
  724. if packed_data == None:
  725. continue
  726. except PackagerError:
  727. raise DataNotPackable(namespace, each[0], each[1], packager, "Packager raised exception")
  728. yield (each[0], packed_data)
  729. def __init__(self, loader, data, namespace, support_regions = False):
  730. self.loader = loader
  731. self.iterator = self.iterate_packed_values(data, namespace, support_regions)
  732. def __iter__(self):
  733. return self.iterator
  734. for namespace in file_data.iterate_namespaces():
  735. if file_data.is_namespace_updated(namespace) == False:
  736. continue
  737. self.db.add_row(namespace,
  738. file_data.get_id(),
  739. None,
  740. DataIterator(self, file_data, namespace))
  741. if file_data.are_regions_loaded():
  742. for region in file_data.iterate_regions():
  743. for namespace in region.iterate_namespaces():
  744. if region.is_namespace_updated(namespace) == False:
  745. continue
  746. self.db.add_row(namespace,
  747. file_data.get_id(),
  748. region.get_id(),
  749. DataIterator(self, region, namespace, support_regions = True))
  750. def iterate_file_data(self, path = None, path_like_filter = "%"):
  751. if self.db == None:
  752. return None
  753. final_path_like = path_like_filter
  754. if path != None:
  755. if self.db.check_dir(path) == False and self.db.check_file(path) == False:
  756. return None
  757. final_path_like = path + path_like_filter
  758. class FileDataIterator(object):
  759. def iterate_file_data(self, loader, final_path_like):
  760. for data in loader.db.iterate_files(path_like=final_path_like):
  761. yield FileData(loader, data.path, data.id, data.checksum, None)
  762. def __init__(self, loader, final_path_like):
  763. self.iterator = self.iterate_file_data(loader, final_path_like)
  764. def __iter__(self):
  765. return self.iterator
  766. if self.db == None:
  767. return None
  768. return FileDataIterator(self, final_path_like)
  769. def load_aggregated_data(self, path = None, path_like_filter = "%", namespaces = None):
  770. if self.db == None:
  771. return None
  772. final_path_like = path_like_filter
  773. if path != None:
  774. if self.db.check_dir(path) == False and self.db.check_file(path) == False:
  775. return None
  776. final_path_like = path + path_like_filter
  777. if namespaces == None:
  778. namespaces = self.namespaces.keys()
  779. result = AggregatedData(self, path)
  780. for name in namespaces:
  781. namespace = self.get_namespace(name)
  782. data = self.db.aggregate_rows(name, path_like = final_path_like)
  783. for field in data.keys():
  784. if namespace.get_field_packager(field).get_python_type() == str:
  785. continue
  786. if namespace.get_field_packager(field).is_non_zero() == True:
  787. data[field]['min'] = None
  788. data[field]['avg'] = None
  789. distribution = self.db.count_rows(name, path_like = final_path_like, group_by_column = field)
  790. data[field]['distribution-bars'] = []
  791. for each in distribution:
  792. if each[0] == None:
  793. continue
  794. assert(float(data[field]['count'] != 0))
  795. data[field]['distribution-bars'].append({'metric': each[0],
  796. 'count': each[1],
  797. 'ratio': (float(each[1]) / float(data[field]['count']))})
  798. result.set_data(name, field, data[field])
  799. return result
  800. def load_selected_data(self, namespace, fields = None, path = None, path_like_filter = "%", filters = [],
  801. sort_by = None, limit_by = None):
  802. if self.db == None:
  803. return None
  804. final_path_like = path_like_filter
  805. if path != None:
  806. if self.db.check_dir(path) == False and self.db.check_file(path) == False:
  807. return None
  808. final_path_like = path + path_like_filter
  809. namespace_obj = self.get_namespace(namespace)
  810. if namespace_obj == None:
  811. return None
  812. class SelectDataIterator(object):
  813. def iterate_selected_values(self, loader, namespace_obj, final_path_like, fields, filters, sort_by, limit_by):
  814. for row in loader.db.select_rows(namespace_obj.get_name(), path_like=final_path_like, filters=filters,
  815. order_by=sort_by, limit_by=limit_by):
  816. region_id = None
  817. if namespace_obj.are_regions_supported() == True:
  818. region_id = row['region_id']
  819. data = SelectData(loader, row['path'], row['id'], region_id)
  820. field_names = fields
  821. if fields == None:
  822. field_names = namespace_obj.iterate_field_names()
  823. for field in field_names:
  824. data.set_data(namespace, field, row[field])
  825. yield data
  826. def __init__(self, loader, namespace_obj, final_path_like, fields, filters, sort_by, limit_by):
  827. self.iterator = self.iterate_selected_values(loader, namespace_obj, final_path_like, fields, filters, sort_by, limit_by)
  828. def __iter__(self):
  829. return self.iterator
  830. return SelectDataIterator(self, namespace_obj, final_path_like, fields, filters, sort_by, limit_by)
  831. class BasePlugin(object):
  832. def initialize(self):
  833. pass
  834. def terminate(self):
  835. pass
  836. def set_name(self, name):
  837. self.name = name
  838. def get_name(self):
  839. if hasattr(self, 'name') == False:
  840. return None
  841. return self.name
  842. def set_version(self, version):
  843. self.version = version
  844. def get_version(self):
  845. if hasattr(self, 'version') == False:
  846. return None
  847. return self.version
  848. def set_plugin_loader(self, loader):
  849. self.plugin_loader = loader
  850. def get_plugin_loader(self):
  851. if hasattr(self, 'plugin_loader') == False:
  852. return None
  853. return self.plugin_loader
  854. class Plugin(BasePlugin):
  855. class Field(object):
  856. def __init__(self, name, ftype, non_zero=False):
  857. self.name = name
  858. self.type = ftype
  859. self.non_zero = non_zero
  860. class Property(object):
  861. def __init__(self, name, value):
  862. self.name = name
  863. self.value = value
  864. def initialize(self, namespace=None, support_regions=True, fields=[], properties=[]):
  865. super(Plugin, self).initialize()
  866. if hasattr(self, 'is_updated') == False:
  867. self.is_updated = False # original initialization
  868. db_loader = self.get_plugin_loader().get_plugin('mpp.dbf').get_loader()
  869. if namespace == None:
  870. namespace = self.get_name()
  871. if (len(fields) != 0 or len(properties) != 0):
  872. prev_version = db_loader.set_property(self.get_name() + ":version", self.get_version())
  873. if prev_version != self.get_version():
  874. self.is_updated = True
  875. for prop in properties:
  876. assert(prop.name != 'version')
  877. prev_prop = db_loader.set_property(self.get_name() + ":" + prop.name, prop.value)
  878. if prev_prop != prop.value:
  879. self.is_updated = True
  880. if len(fields) != 0:
  881. namespace_obj = db_loader.create_namespace(namespace,
  882. support_regions=support_regions,
  883. version=self.get_version())
  884. for field in fields:
  885. is_created = namespace_obj.add_field(field.name, field.type, non_zero=field.non_zero)
  886. assert(is_created != None)
  887. # if field is created (not cloned from the previous db),
  888. # mark the plug-in as updated in order to trigger full rescan
  889. self.is_updated = self.is_updated or is_created
  890. class SimpleMetricMixin(object):
  891. class AliasError(Exception):
  892. def __init__(self, alias):
  893. Exception.__init__(self, "Unknown alias: " + str(alias))
  894. def declare_metric(self, is_active, field,
  895. pattern_to_search_or_map_of_patterns,
  896. marker_type_mask=Marker.T.ANY,
  897. region_type_mask=Region.T.ANY,
  898. exclude_subregions=True,
  899. merge_markers=False):
  900. if hasattr(self, '_fields') == False:
  901. self._fields = {}
  902. if isinstance(pattern_to_search_or_map_of_patterns, dict):
  903. map_of_patterns = pattern_to_search_or_map_of_patterns
  904. else:
  905. map_of_patterns = {'default': pattern_to_search_or_map_of_patterns}
  906. if is_active == True:
  907. self._fields[field.name] = (field,
  908. marker_type_mask,
  909. exclude_subregions,
  910. merge_markers,
  911. map_of_patterns,
  912. region_type_mask)
  913. def is_active(self, metric_name = None):
  914. if metric_name == None:
  915. return (len(self._fields.keys()) > 0)
  916. return (metric_name in self._fields.keys())
  917. def get_fields(self):
  918. result = []
  919. for key in self._fields.keys():
  920. result.append(self._fields[key][0])
  921. return result
  922. def count_if_active(self, metric_name, data, namespace=None, alias='default'):
  923. if self.is_active(metric_name) == False:
  924. return
  925. if namespace == None:
  926. namespace = self.get_name()
  927. field_data = self._fields[metric_name]
  928. text = data.get_content()
  929. if alias not in field_data[4].keys():
  930. raise self.AliasError(alias)
  931. pattern_to_search = field_data[4][alias]
  932. for region in data.iterate_regions(filter_group=field_data[5]):
  933. count = 0
  934. for marker in data.iterate_markers(
  935. filter_group = field_data[1],
  936. region_id = region.get_id(),
  937. exclude_children = field_data[2],
  938. merge=field_data[3]):
  939. count += len(pattern_to_search.findall(text, marker.get_offset_begin(), marker.get_offset_end()))
  940. region.set_data(namespace, metric_name, count)
  941. class InterfaceNotImplemented(Exception):
  942. def __init__(self, obj):
  943. import sys
  944. Exception.__init__(self, "Method '"
  945. + sys._getframe(1).f_code.co_name
  946. + "' has not been implemented for "
  947. + str(obj.__class__))
  948. class IConfigurable(object):
  949. def configure(self, options):
  950. raise InterfaceNotImplemented(self)
  951. def declare_configuration(self, optparser):
  952. raise InterfaceNotImplemented(self)
  953. class IRunable(object):
  954. def run(self, args):
  955. raise InterfaceNotImplemented(self)
  956. class IParser(object):
  957. def process(self, parent, data, is_updated):
  958. raise InterfaceNotImplemented(self)
  959. class ICode(object):
  960. pass
  961. class CallbackNotImplemented(Exception):
  962. def __init__(self, obj, callback_name):
  963. Exception.__init__(self, "Callback '"
  964. + callback_name
  965. + "' has not been implemented for "
  966. + str(obj.__class__))
  967. class Child(object):
  968. def notify(self, parent, callback_name, *args):
  969. if hasattr(self, callback_name) == False:
  970. raise CallbackNotImplemented(self, callback_name)
  971. self.__getattribute__(callback_name)(parent, *args)
  972. def subscribe_by_parents_name(self, parent_name, callback_name='callback'):
  973. self.get_plugin_loader().get_plugin(parent_name).subscribe(self, callback_name)
  974. def subscribe_by_parents_interface(self, interface, callback_name='callback'):
  975. for plugin in self.get_plugin_loader().iterate_plugins():
  976. if isinstance(plugin, interface):
  977. plugin.subscribe(self, callback_name)
  978. class Parent(object):
  979. def init_Parent(self):
  980. if hasattr(self, 'children') == False:
  981. self.children = []
  982. def subscribe(self, obj, callback_name):
  983. self.init_Parent()
  984. if (isinstance(obj, Child) == False):
  985. raise TypeError()
  986. self.children.append((obj,callback_name))
  987. def unsubscribe(self, obj, callback_name):
  988. self.init_Parent()
  989. self.children.remove((obj, callback_name))
  990. def notify_children(self, *args):
  991. self.init_Parent()
  992. for child in self.children:
  993. child[0].notify(self, child[1], *args)
  994. def iterate_children(self):
  995. self.init_Parent()
  996. for child in self.children:
  997. yield child