Add data_analysis/src/data_analysis/cli.py
This commit is contained in:
parent
90e5b03840
commit
5466a06f4e
1 changed files with 95 additions and 0 deletions
95
data_analysis/src/data_analysis/cli.py
Normal file
95
data_analysis/src/data_analysis/cli.py
Normal file
|
|
@ -0,0 +1,95 @@
|
||||||
|
import argparse
|
||||||
|
import json
|
||||||
|
import sys
|
||||||
|
from pathlib import Path
|
||||||
|
import logging
|
||||||
|
from typing import List
|
||||||
|
|
||||||
|
import pandas as pd # third party dependency
|
||||||
|
|
||||||
|
from data_analysis.core import analyze_data, AnalysisResult
|
||||||
|
from data_analysis.io_utils import LogEntry
|
||||||
|
|
||||||
|
logging.basicConfig(level=logging.INFO, format='[%(levelname)s] %(message)s')
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def parse_args() -> argparse.Namespace:
|
||||||
|
parser = argparse.ArgumentParser(
|
||||||
|
description="Analyse der Donau-Rover Sensordaten für Muster und Anomalien."
|
||||||
|
)
|
||||||
|
parser.add_argument(
|
||||||
|
"--input",
|
||||||
|
required=True,
|
||||||
|
help="Pfad zur JSON-Datei mit Rover-Logdaten."
|
||||||
|
)
|
||||||
|
parser.add_argument(
|
||||||
|
"--output",
|
||||||
|
required=False,
|
||||||
|
help="Pfad zur Zielausgabedatei für die Analyseergebnisse."
|
||||||
|
)
|
||||||
|
return parser.parse_args()
|
||||||
|
|
||||||
|
|
||||||
|
def load_log_entries(input_path: Path) -> List[LogEntry]:
|
||||||
|
if not input_path.exists():
|
||||||
|
raise FileNotFoundError(f"Eingabedatei nicht gefunden: {input_path}")
|
||||||
|
|
||||||
|
try:
|
||||||
|
df = pd.read_json(input_path)
|
||||||
|
except ValueError as e:
|
||||||
|
raise ValueError(f"Fehler beim Einlesen von JSON: {e}")
|
||||||
|
|
||||||
|
required_columns = {"t", "Lx", "dB", "Temp", "Inference"}
|
||||||
|
if not required_columns.issubset(df.columns):
|
||||||
|
missing = required_columns - set(df.columns)
|
||||||
|
raise ValueError(f"Fehlende Felder in Eingabedatei: {missing}")
|
||||||
|
|
||||||
|
entries: List[LogEntry] = []
|
||||||
|
for _, row in df.iterrows():
|
||||||
|
try:
|
||||||
|
entry = LogEntry(
|
||||||
|
timestamp=row["t"],
|
||||||
|
luminosity=int(row["Lx"]),
|
||||||
|
sound_level=float(row["dB"]),
|
||||||
|
temperature=float(row["Temp"]),
|
||||||
|
inference=float(row["Inference"])
|
||||||
|
)
|
||||||
|
entries.append(entry)
|
||||||
|
except (TypeError, ValueError) as e:
|
||||||
|
logger.warning(f"Überspringe ungültigen Datensatz: {e}")
|
||||||
|
if not entries:
|
||||||
|
raise ValueError("Keine gültigen LogEinträge geladen.")
|
||||||
|
return entries
|
||||||
|
|
||||||
|
|
||||||
|
def save_analysis_result(result: AnalysisResult, output_path: Path) -> None:
|
||||||
|
output_path.parent.mkdir(parents=True, exist_ok=True)
|
||||||
|
result_dict = {
|
||||||
|
"significant_patterns": result.significant_patterns,
|
||||||
|
"anomaly_events": result.anomaly_events,
|
||||||
|
}
|
||||||
|
with output_path.open("w", encoding="utf-8") as f:
|
||||||
|
json.dump(result_dict, f, indent=2, ensure_ascii=False)
|
||||||
|
logger.info(f"Analyseergebnisse gespeichert unter: {output_path}")
|
||||||
|
|
||||||
|
|
||||||
|
def main() -> None:
|
||||||
|
args = parse_args()
|
||||||
|
input_path = Path(args.input)
|
||||||
|
output_path = Path(args.output) if args.output else Path("output/analysis_result.json")
|
||||||
|
|
||||||
|
try:
|
||||||
|
log_entries = load_log_entries(input_path)
|
||||||
|
result = analyze_data(log_entries)
|
||||||
|
|
||||||
|
assert isinstance(result, AnalysisResult), "Analyseergebnis ist ungültig."
|
||||||
|
save_analysis_result(result, output_path)
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
logger.error(f"Fehler während der Analyse: {e}")
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
||||||
Loading…
Reference in a new issue