bcp_log.py 2.9 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677
  1. import re
  2. from dataclasses import dataclass
  3. from pathlib import Path
  4. from datetime import datetime
  5. @dataclass
  6. class BulkcopyResult:
  7. filename: str
  8. timestamp: datetime
  9. imported: int = 0
  10. exported: int = 0
  11. ignored: int = 0
  12. import_duration: float = 0.0
  13. export_duration: float = 0.0
  14. @property
  15. def missing(self) -> int:
  16. return self.exported - self.imported - self.ignored
  17. def to_csv(self):
  18. return (f"{self.filename};{self.timestamp.strftime('%d.%m.%Y %H:%M:%S')};"
  19. + f"{self.exported};{self.imported};{self.ignored};{self.missing};"
  20. + f"{self.export_duration};{self.import_duration}")
  21. def __str__(self) -> str:
  22. return "\n".join([f'Filename: {self.filename}',
  23. f"Last run: {self.timestamp.strftime('%d.%m.%Y %H:%M')}",
  24. '',
  25. f'Exported: {self.exported:>7}',
  26. f'Imported: {self.imported:>7}',
  27. f'Ignored: {self.ignored:>7}',
  28. f'Missing: {self.missing:>7}',
  29. '',
  30. f'Duration: {self.export_duration:>11} s',
  31. f' {self.import_duration:>11} s'])
  32. def check_logfiles(prefix, base_dir):
  33. ts = datetime.fromtimestamp(Path(f"{base_dir}/{prefix}.in.log").stat().st_mtime)
  34. result = BulkcopyResult(filename=prefix, timestamp=ts)
  35. with open(f"{base_dir}/{prefix}.in.log", 'r') as frh:
  36. result.ignored = len(frh.readlines())
  37. # info output of export
  38. with open(f"{base_dir}/{prefix}.bcp1.log", 'r', encoding='cp850', errors='ignore') as frh:
  39. raw_logs = frh.read()
  40. match = re.search(r'(\d+) Zeilen kopiert.', raw_logs)
  41. result.exported = int(match.group(1)) if match else 0
  42. match2 = re.search(r'Zeit .* gesamt: (\d+)', raw_logs)
  43. result.export_duration = int(match2.group(1)) / 1000 if match2 else 0
  44. # info output of import
  45. with open(f"{base_dir}/{prefix}.bcp2.log", 'r', encoding='cp850', errors='ignore') as frh:
  46. raw_logs = frh.read()
  47. match = re.search(r'(\d+) Zeilen kopiert.', raw_logs)
  48. result.imported = int(match.group(1)) if match else 0
  49. match2 = re.search(r'Zeit .* gesamt: (\d+)', raw_logs)
  50. result.import_duration = int(match2.group(1)) / 1000 if match2 else 0
  51. return result
  52. def check_directory(base_dir):
  53. res = []
  54. for filename in Path(base_dir).glob('*.bcp1.log'):
  55. stem = filename.name[:-9]
  56. res.append(check_logfiles(stem, base_dir).to_csv())
  57. with open(base_dir + '/info.log', 'w') as fwh:
  58. fwh.write('filename;timestamp;imported;exported;ignored;import_duration;export_duration\n')
  59. fwh.write('\n'.join(res))
  60. if __name__ == '__main__':
  61. check_directory('/home/robert/projekte/python/dbtools/SQL/temp')
  62. # check_logfiles('ORDER_LINE_1')