|
@@ -1,51 +1,52 @@
|
|
|
# from itertools import chain
|
|
|
import json
|
|
|
+import re
|
|
|
+from collections import defaultdict
|
|
|
from pathlib import Path
|
|
|
+
|
|
|
import pandas as pd
|
|
|
from bs4 import BeautifulSoup
|
|
|
-from collections import defaultdict
|
|
|
-import re
|
|
|
-import sys
|
|
|
from config import Config
|
|
|
|
|
|
|
|
|
def get_path_info(base_dir):
|
|
|
- path_info_df = pd.read_csv(base_dir + '/logs/status/path_info.csv', sep=';',
|
|
|
- encoding='latin-1', converters={'process': str})
|
|
|
- path_info_df.rename(columns={'name': 'filename'}, inplace=True)
|
|
|
- path_info_df['filename'] = path_info_df['filename'].str.lower()
|
|
|
- path_info_df.set_index('filename', inplace=True, drop=False)
|
|
|
- return path_info_df.to_dict(orient='index')
|
|
|
+ path_info_df = pd.read_csv(
|
|
|
+ base_dir + "/logs/status/path_info.csv", sep=";", encoding="latin-1", converters={"process": str}
|
|
|
+ )
|
|
|
+ path_info_df.rename(columns={"name": "filename"}, inplace=True)
|
|
|
+ path_info_df["filename"] = path_info_df["filename"].str.lower()
|
|
|
+ path_info_df.set_index("filename", inplace=True, drop=False)
|
|
|
+ return path_info_df.to_dict(orient="index")
|
|
|
|
|
|
|
|
|
def get_cubes(base_dir, cfg: Config, path_info, cubes_models):
|
|
|
- ver_files = [k for k in path_info.keys() if re.search(r'\\cubes\\.*\.ver', k)]
|
|
|
+ ver_files = [k for k in path_info.keys() if re.search(r"\\cubes\\.*\.ver", k)]
|
|
|
cubes = {}
|
|
|
for file in ver_files:
|
|
|
- match = re.search(r'\\cubes\\(.*)__\d+\.ver$', file)
|
|
|
+ match = re.search(r"\\cubes\\(.*)__\d+\.ver$", file)
|
|
|
cube_name = match[1]
|
|
|
cube_subdir = f"{file[:-4]}\\{cube_name}.mdc"
|
|
|
cube_out = f"{cfg.system_dir}\\cube_out\\{cube_name}.mdc"
|
|
|
|
|
|
cubes[cube_name] = {
|
|
|
- 'deployed_mdc': path_info[cube_subdir],
|
|
|
- 'cube_out_mdc': path_info[cube_out],
|
|
|
- 'model': cubes_models[cube_name],
|
|
|
- 'errors': []
|
|
|
+ "deployed_mdc": path_info[cube_subdir],
|
|
|
+ "cube_out_mdc": path_info[cube_out],
|
|
|
+ "model": cubes_models[cube_name],
|
|
|
+ "errors": [],
|
|
|
}
|
|
|
return cubes
|
|
|
|
|
|
|
|
|
def get_models(base_dir, cfg: Config, path_info, fm_sources):
|
|
|
models = {}
|
|
|
- for file in Path(base_dir + '\\config\\models').glob('*.log'):
|
|
|
- with open(file, 'r') as frh:
|
|
|
- model_infos = frh.read().lower().replace('"', '').replace(',', '').split('--')
|
|
|
- datasources = model_infos[1].split('\n')
|
|
|
- datasources = [d for d in datasources if d != '']
|
|
|
+ for file in Path(base_dir + "\\config\\models").glob("*.log"):
|
|
|
+ with open(file, "r") as frh:
|
|
|
+ model_infos = frh.read().lower().replace('"', "").replace(",", "").split("--")
|
|
|
+ datasources = model_infos[1].split("\n")
|
|
|
+ datasources = [d for d in datasources if d != ""]
|
|
|
fm_src = []
|
|
|
fm_package = None
|
|
|
- if '[' in datasources[0]:
|
|
|
+ if "[" in datasources[0]:
|
|
|
fm_package = datasources.pop(0).upper()
|
|
|
fm_src = datasources
|
|
|
datasources = []
|
|
@@ -54,43 +55,43 @@ def get_models(base_dir, cfg: Config, path_info, fm_sources):
|
|
|
datasources.extend(fm_sources[src])
|
|
|
datasources = sorted(set(datasources))
|
|
|
|
|
|
- cube = re.search(r'\\cube_out\\(.*)\.mdc', model_infos[0])[1]
|
|
|
+ cube = re.search(r"\\cube_out\\(.*)\.mdc", model_infos[0])[1]
|
|
|
models[file.name[:-8].lower()] = {
|
|
|
- 'framework_manager': 'J' if fm_package else 'N',
|
|
|
- 'fm_package': fm_package,
|
|
|
- 'fm_sources': fm_src,
|
|
|
- 'datasources': datasources,
|
|
|
- 'cube': cube,
|
|
|
- 'logfile_tasks': path_info.get(cfg.portal_dir + '\\tasks\\logs\\' + file.name[:-8] + '.log', {'mtime': '0'}),
|
|
|
- 'logfile_system': path_info.get(cfg.system_dir + '\\logs\\' + file.name[:-8] + '.log', {'mtime': '0'}),
|
|
|
- 'model_file': path_info.get(cfg.system_dir + '\\models\\' + file.name[:-4], {'mtime': '0'}),
|
|
|
- 'model_file_filled': path_info.get(cfg.system_dir + '\\models\\gefuellt\\' + file.name[:-4], {'mtime': '0'}),
|
|
|
+ "framework_manager": "J" if fm_package else "N",
|
|
|
+ "fm_package": fm_package,
|
|
|
+ "fm_sources": fm_src,
|
|
|
+ "datasources": datasources,
|
|
|
+ "cube": cube,
|
|
|
+ "logfile_tasks": path_info.get(
|
|
|
+ cfg.portal_dir + "\\tasks\\logs\\" + file.name[:-8] + ".log", {"mtime": "0"}
|
|
|
+ ),
|
|
|
+ "logfile_system": path_info.get(cfg.system_dir + "\\logs\\" + file.name[:-8] + ".log", {"mtime": "0"}),
|
|
|
+ "model_file": path_info.get(cfg.system_dir + "\\models\\" + file.name[:-4], {"mtime": "0"}),
|
|
|
+ "model_file_filled": path_info.get(
|
|
|
+ cfg.system_dir + "\\models\\gefuellt\\" + file.name[:-4], {"mtime": "0"}
|
|
|
+ ),
|
|
|
}
|
|
|
return models
|
|
|
|
|
|
|
|
|
def get_database_info(base_dir, cfg: Config):
|
|
|
- db_info_df = pd.read_csv(base_dir + '/logs/status/db_info.csv', sep=';',
|
|
|
- encoding='latin-1')
|
|
|
- db_info_df = db_info_df[db_info_df['DatabaseName'] == 'GC']
|
|
|
- db_info_df['table'] = db_info_df['TableName'].str.lower()
|
|
|
- db_info_df.set_index('table', inplace=True)
|
|
|
- return db_info_df.to_dict(orient='index')
|
|
|
+ db_info_df = pd.read_csv(base_dir + "/logs/status/db_info.csv", sep=";", encoding="latin-1")
|
|
|
+ db_info_df = db_info_df[db_info_df["DatabaseName"] == "GC"]
|
|
|
+ db_info_df["table"] = db_info_df["TableName"].str.lower()
|
|
|
+ db_info_df.set_index("table", inplace=True)
|
|
|
+ return db_info_df.to_dict(orient="index")
|
|
|
|
|
|
|
|
|
def get_fm_sources(base_dir, cfg):
|
|
|
- bs = BeautifulSoup(open(base_dir + '\\config\\fm\\model.xml', 'r'), 'xml')
|
|
|
+ bs = BeautifulSoup(open(base_dir + "\\config\\fm\\model.xml", "r"), "xml")
|
|
|
sources = defaultdict(list)
|
|
|
- for item in bs.find_all('queryItem'):
|
|
|
- p = item.parent.parent.find('name').string
|
|
|
- if item.parent.parent.name == 'folder':
|
|
|
- p = item.parent.parent.parent.find('name').string
|
|
|
- parent = "[{0}].[{1}]".format(
|
|
|
- p,
|
|
|
- item.parent.find('name').string
|
|
|
- )
|
|
|
- src = ''
|
|
|
- exp = ''
|
|
|
+ for item in bs.find_all("queryItem"):
|
|
|
+ p = item.parent.parent.find("name").string
|
|
|
+ if item.parent.parent.name == "folder":
|
|
|
+ p = item.parent.parent.parent.find("name").string
|
|
|
+ parent = "[{0}].[{1}]".format(p, item.parent.find("name").string)
|
|
|
+ src = ""
|
|
|
+ exp = ""
|
|
|
if item.expression:
|
|
|
if item.expression.refobj:
|
|
|
src = item.expression.refobj.string
|
|
@@ -98,26 +99,26 @@ def get_fm_sources(base_dir, cfg):
|
|
|
exp = item.expression.string
|
|
|
elif item.externalName:
|
|
|
exp = item.externalName.string
|
|
|
- sources[parent].append((item.find('name').string, src, exp))
|
|
|
+ sources[parent].append((item.find("name").string, src, exp))
|
|
|
interface = {}
|
|
|
for k, fields in sources.items():
|
|
|
- if '[Schnittstelle]' not in k:
|
|
|
+ if "[Schnittstelle]" not in k:
|
|
|
continue
|
|
|
- key = k.split('.')[-1][1:-1].lower()
|
|
|
+ key = k.split(".")[-1][1:-1].lower()
|
|
|
links = []
|
|
|
for field in fields:
|
|
|
- links.append(follow_links(sources, field, ''))
|
|
|
- interface[key] = sorted(list(set([re.search(r'\.\[(.*)\]$', e)[1].lower() for e in links if '[Import]' in e])))
|
|
|
+ links.append(follow_links(sources, field, ""))
|
|
|
+ interface[key] = sorted(list(set([re.search(r"\.\[(.*)\]$", e)[1].lower() for e in links if "[Import]" in e])))
|
|
|
|
|
|
return interface
|
|
|
|
|
|
|
|
|
def follow_links(sources, field, value):
|
|
|
- if field[1] == '':
|
|
|
+ if field[1] == "":
|
|
|
if field[2] == field[0]:
|
|
|
return value
|
|
|
return field[2]
|
|
|
- match = re.search(r'(\[.*\]\.\[.*\])\.\[(.*)\]', field[1])
|
|
|
+ match = re.search(r"(\[.*\]\.\[.*\])\.\[(.*)\]", field[1])
|
|
|
key1 = match[1]
|
|
|
val1 = match[2]
|
|
|
if key1 in sources:
|
|
@@ -129,35 +130,43 @@ def follow_links(sources, field, value):
|
|
|
|
|
|
|
|
|
def get_datasources(base_dir, cfg, path_info):
|
|
|
- all_datasources = set([re.search(r'\\iqd\\.*\\(.*)\.imr', k)[1]
|
|
|
- for k in path_info.keys()
|
|
|
- if re.search(r'\\iqd\\.*\\.*\.imr', k)])
|
|
|
+ all_datasources = set(
|
|
|
+ [re.search(r"\\iqd\\.*\\(.*)\.imr", k)[1] for k in path_info.keys() if re.search(r"\\iqd\\.*\\.*\.imr", k)]
|
|
|
+ )
|
|
|
datasources = {}
|
|
|
for ds in all_datasources:
|
|
|
- ds_search = f'\\{ds}.imr'
|
|
|
- imr_files = [k for k in path_info.keys()
|
|
|
- if re.search(r'\\iqd\\.*\.imr', k) and ds_search in k
|
|
|
- and 'austausch' not in k]
|
|
|
+ ds_search = f"\\{ds}.imr"
|
|
|
+ imr_files = [
|
|
|
+ k for k in path_info.keys() if re.search(r"\\iqd\\.*\.imr", k) and ds_search in k and "austausch" not in k
|
|
|
+ ]
|
|
|
if len(imr_files) == 0:
|
|
|
- imr_file = '0.imr'
|
|
|
+ imr_file = "0.imr"
|
|
|
else:
|
|
|
imr_file = imr_files.pop(0)
|
|
|
datasources[ds] = {
|
|
|
- 'imr_file': path_info.get(imr_file, {'mtime': '0'}),
|
|
|
- 'iqd_file': path_info.get(imr_file[:-4] + '.iqd', {'mtime': '0'}),
|
|
|
- 'csv_file': path_info.get(cfg.system_dir + '\\export\\' + ds + '.csv', {'mtime': '0'}),
|
|
|
- 'csv_file_iqd_folder': path_info.get(imr_file[:-4] + '.csv', {'mtime': '0'}),
|
|
|
- 'duplicates': imr_files
|
|
|
+ "imr_file": path_info.get(imr_file, {"mtime": "0"}),
|
|
|
+ "iqd_file": path_info.get(imr_file[:-4] + ".iqd", {"mtime": "0"}),
|
|
|
+ "csv_file": path_info.get(cfg.system_dir + "\\export\\" + ds + ".csv", {"mtime": "0"}),
|
|
|
+ "csv_file_iqd_folder": path_info.get(imr_file[:-4] + ".csv", {"mtime": "0"}),
|
|
|
+ "duplicates": imr_files,
|
|
|
}
|
|
|
return datasources
|
|
|
|
|
|
|
|
|
def cubes_to_models(models):
|
|
|
- models_sort = sorted([(v.get('logfile_tasks')['mtime'],
|
|
|
- v.get('logfile_system')['mtime'],
|
|
|
- v.get('model_file_filled')['mtime'],
|
|
|
- v.get('model_file')['mtime'],
|
|
|
- k, v['cube']) for k, v in models.items()])
|
|
|
+ models_sort = sorted(
|
|
|
+ [
|
|
|
+ (
|
|
|
+ v.get("logfile_tasks")["mtime"],
|
|
|
+ v.get("logfile_system")["mtime"],
|
|
|
+ v.get("model_file_filled")["mtime"],
|
|
|
+ v.get("model_file")["mtime"],
|
|
|
+ k,
|
|
|
+ v["cube"],
|
|
|
+ )
|
|
|
+ for k, v in models.items()
|
|
|
+ ]
|
|
|
+ )
|
|
|
result = {}
|
|
|
for m in models_sort:
|
|
|
result[m[5]] = m[4]
|
|
@@ -165,8 +174,8 @@ def cubes_to_models(models):
|
|
|
|
|
|
|
|
|
def main():
|
|
|
- base_dir = 'app/temp/unzipped/loeffler-c11_2022-07-08_112547'
|
|
|
- cfg = Config(str(Path(base_dir + '\\gaps.ini').absolute()))
|
|
|
+ base_dir = "app/temp/unzipped/loeffler-c11_2022-07-08_112547"
|
|
|
+ cfg = Config(str(Path(base_dir + "\\gaps.ini").absolute()))
|
|
|
# Dateiliste
|
|
|
path_info = get_path_info(base_dir)
|
|
|
# random_bat_file = [k for k in path_info.keys() if re.search(r'\\Tasks\\.*\.bat', k)][0]
|
|
@@ -176,16 +185,16 @@ def main():
|
|
|
result = {}
|
|
|
|
|
|
# Modelle und Datenquellen
|
|
|
- result['fm_sources'] = get_fm_sources(base_dir, cfg)
|
|
|
- result['models'] = get_models(base_dir, cfg, path_info, result['fm_sources'])
|
|
|
- result['database'] = get_database_info(base_dir, cfg)
|
|
|
- result['datasources'] = get_datasources(base_dir, cfg, path_info)
|
|
|
- cubes_models = cubes_to_models(result['models'])
|
|
|
- result['cubes'] = get_cubes(base_dir, cfg, path_info, cubes_models)
|
|
|
+ result["fm_sources"] = get_fm_sources(base_dir, cfg)
|
|
|
+ result["models"] = get_models(base_dir, cfg, path_info, result["fm_sources"])
|
|
|
+ result["database"] = get_database_info(base_dir, cfg)
|
|
|
+ result["datasources"] = get_datasources(base_dir, cfg, path_info)
|
|
|
+ cubes_models = cubes_to_models(result["models"])
|
|
|
+ result["cubes"] = get_cubes(base_dir, cfg, path_info, cubes_models)
|
|
|
# Cubes aktuell?
|
|
|
# Rest aktuell?
|
|
|
- json.dump(result, open('app/temp/logs/export.json', 'w'), indent=2)
|
|
|
+ json.dump(result, open("app/temp/logs/export.json", "w"), indent=2)
|
|
|
|
|
|
|
|
|
-if __name__ == '__main__':
|
|
|
+if __name__ == "__main__":
|
|
|
main()
|