mirror of
https://github.com/Akkudoktor-EOS/EOS.git
synced 2026-04-11 16:26:20 +00:00
Adapters for Home Assistant and NodeRED integration are added. Akkudoktor-EOS can now be run as Home Assistant add-on and standalone. As Home Assistant add-on EOS uses ingress to fully integrate the EOSdash dashboard in Home Assistant. The fix includes several bug fixes that are not directly related to the adapter implementation but are necessary to keep EOS running properly and to test and document the changes. * fix: development version scheme The development versioning scheme is adaptet to fit to docker and home assistant expectations. The new scheme is x.y.z and x.y.z.dev<hash>. Hash is only digits as expected by home assistant. Development version is appended by .dev as expected by docker. * fix: use mean value in interval on resampling for array When downsampling data use the mean value of all values within the new sampling interval. * fix: default battery ev soc and appliance wh Make the genetic simulation return default values for the battery SoC, electric vehicle SoC and appliance load if these assets are not used. * fix: import json string Strip outer quotes from JSON strings on import to be compliant to json.loads() expectation. * fix: default interval definition for import data Default interval must be defined in lowercase human definition to be accepted by pendulum. * fix: clearoutside schema change * feat: add adapters for integrations Adapters for Home Assistant and NodeRED integration are added. Akkudoktor-EOS can now be run as Home Assistant add-on and standalone. As Home Assistant add-on EOS uses ingress to fully integrate the EOSdash dashboard in Home Assistant. * feat: allow eos to be started with root permissions and drop priviledges Home assistant starts all add-ons with root permissions. Eos now drops root permissions if an applicable user is defined by paramter --run_as_user. The docker image defines the user eos to be used. * feat: make eos supervise and monitor EOSdash Eos now not only starts EOSdash but also monitors EOSdash during runtime and restarts EOSdash on fault. EOSdash logging is captured by EOS and forwarded to the EOS log to provide better visibility. * feat: add duration to string conversion Make to_duration to also return the duration as string on request. * chore: Use info logging to report missing optimization parameters In parameter preparation for automatic optimization an error was logged for missing paramters. Log is now down using the info level. * chore: make EOSdash use the EOS data directory for file import/ export EOSdash use the EOS data directory for file import/ export by default. This allows to use the configuration import/ export function also within docker images. * chore: improve EOSdash config tab display Improve display of JSON code and add more forms for config value update. * chore: make docker image file system layout similar to home assistant Only use /data directory for persistent data. This is handled as a docker volume. The /data volume is mapped to ~/.local/share/net.akkudoktor.eos if using docker compose. * chore: add home assistant add-on development environment Add VSCode devcontainer and task definition for home assistant add-on development. * chore: improve documentation
246 lines
8.6 KiB
Python
246 lines
8.6 KiB
Python
"""Utility for configuring Loguru loggers."""
|
|
|
|
import json
|
|
import logging as pylogging
|
|
import os
|
|
import re
|
|
import sys
|
|
from pathlib import Path
|
|
from types import FrameType
|
|
from typing import Any, List, Optional
|
|
|
|
import pendulum
|
|
from loguru import logger
|
|
|
|
from akkudoktoreos.core.logabc import LOGGING_LEVELS
|
|
|
|
|
|
class InterceptHandler(pylogging.Handler):
|
|
"""A logging handler that redirects standard Python logging messages to Loguru.
|
|
|
|
This handler ensures consistency between the `logging` module and Loguru by intercepting
|
|
logs sent to the standard logging system and re-emitting them through Loguru with proper
|
|
formatting and context (including exception info and call depth).
|
|
|
|
Attributes:
|
|
loglevel_mapping (dict): Mapping from standard logging levels to Loguru level names.
|
|
"""
|
|
|
|
loglevel_mapping: dict[int, str] = {
|
|
50: "CRITICAL",
|
|
40: "ERROR",
|
|
30: "WARNING",
|
|
20: "INFO",
|
|
10: "DEBUG",
|
|
5: "TRACE",
|
|
0: "NOTSET",
|
|
}
|
|
|
|
def emit(self, record: pylogging.LogRecord) -> None:
|
|
"""Emits a logging record by forwarding it to Loguru with preserved metadata.
|
|
|
|
Args:
|
|
record (logging.LogRecord): A record object containing log message and metadata.
|
|
"""
|
|
# Skip DEBUG logs from matplotlib - very noisy
|
|
if record.name.startswith("matplotlib") and record.levelno <= pylogging.DEBUG:
|
|
return
|
|
|
|
try:
|
|
level = logger.level(record.levelname).name
|
|
except AttributeError:
|
|
level = self.loglevel_mapping.get(record.levelno, "INFO")
|
|
|
|
frame: Optional[FrameType] = pylogging.currentframe()
|
|
depth: int = 2
|
|
while frame and frame.f_code.co_filename == pylogging.__file__:
|
|
frame = frame.f_back
|
|
depth += 1
|
|
|
|
log = logger.bind(request_id="app")
|
|
log.opt(depth=depth, exception=record.exc_info).log(level, record.getMessage())
|
|
|
|
|
|
console_handler_id = None
|
|
file_handler_id = None
|
|
|
|
|
|
def logging_track_config(config_eos: Any, path: str, old_value: Any, value: Any) -> None:
|
|
"""Track logging config changes."""
|
|
global console_handler_id, file_handler_id
|
|
|
|
if not path.startswith("logging"):
|
|
raise ValueError(f"Logging shall not track '{path}'")
|
|
|
|
if not config_eos.logging.console_level:
|
|
# No value given - check environment value - may also be None
|
|
config_eos.logging.console_level = os.getenv("EOS_LOGGING__LEVEL")
|
|
if not config_eos.logging.file_level:
|
|
# No value given - check environment value - may also be None
|
|
config_eos.logging.file_level = os.getenv("EOS_LOGGING__LEVEL")
|
|
|
|
# Remove handlers
|
|
if console_handler_id:
|
|
try:
|
|
logger.remove(console_handler_id)
|
|
except Exception as e:
|
|
logger.debug("Exception on logger.remove: {}", e, exc_info=True)
|
|
console_handler_id = None
|
|
if file_handler_id:
|
|
try:
|
|
logger.remove(file_handler_id)
|
|
except Exception as e:
|
|
logger.debug("Exception on logger.remove: {}", e, exc_info=True)
|
|
file_handler_id = None
|
|
|
|
# Create handlers with new configuration
|
|
# Always add console handler
|
|
if config_eos.logging.console_level not in LOGGING_LEVELS:
|
|
logger.error(
|
|
f"Invalid console log level '{config_eos.logging.console_level} - forced to INFO'."
|
|
)
|
|
config_eos.logging.console_level = "INFO"
|
|
|
|
console_handler_id = logger.add(
|
|
sys.stderr,
|
|
enqueue=True,
|
|
backtrace=True,
|
|
level=config_eos.logging.console_level,
|
|
# format=_console_format
|
|
)
|
|
|
|
# Add file handler
|
|
if config_eos.logging.file_level and config_eos.logging.file_path:
|
|
if config_eos.logging.file_level not in LOGGING_LEVELS:
|
|
logger.error(
|
|
f"Invalid file log level '{config_eos.logging.console_level}' - forced to INFO."
|
|
)
|
|
config_eos.logging.file_level = "INFO"
|
|
|
|
file_handler_id = logger.add(
|
|
sink=config_eos.logging.file_path,
|
|
rotation="100 MB",
|
|
retention="3 days",
|
|
enqueue=True,
|
|
backtrace=True,
|
|
level=config_eos.logging.file_level,
|
|
serialize=True, # JSON dict formatting
|
|
# format=_file_format
|
|
)
|
|
|
|
# Redirect standard logging to Loguru
|
|
pylogging.basicConfig(handlers=[InterceptHandler()], level=0)
|
|
# Redirect uvicorn and fastapi logging to Loguru
|
|
pylogging.getLogger("uvicorn.access").handlers = [InterceptHandler()]
|
|
for pylogger_name in ["uvicorn", "uvicorn.error", "fastapi"]:
|
|
pylogger = pylogging.getLogger(pylogger_name)
|
|
pylogger.handlers = [InterceptHandler()]
|
|
pylogger.propagate = False
|
|
|
|
logger.info(
|
|
f"Logger reconfigured - console: {config_eos.logging.console_level}, file: {config_eos.logging.file_level}."
|
|
)
|
|
|
|
|
|
def read_file_log(
|
|
log_path: Path,
|
|
limit: int = 100,
|
|
level: Optional[str] = None,
|
|
contains: Optional[str] = None,
|
|
regex: Optional[str] = None,
|
|
from_time: Optional[str] = None,
|
|
to_time: Optional[str] = None,
|
|
tail: bool = False,
|
|
) -> List[dict]:
|
|
"""Read and filter structured log entries from a JSON-formatted log file.
|
|
|
|
Args:
|
|
log_path (Path): Path to the JSON-formatted log file.
|
|
limit (int, optional): Maximum number of log entries to return. Defaults to 100.
|
|
level (Optional[str], optional): Filter logs by log level (e.g., "INFO", "ERROR"). Defaults to None.
|
|
contains (Optional[str], optional): Filter logs that contain this substring in their message. Case-insensitive. Defaults to None.
|
|
regex (Optional[str], optional): Filter logs whose message matches this regular expression. Defaults to None.
|
|
from_time (Optional[str], optional): ISO 8601 datetime string to filter logs not earlier than this time. Defaults to None.
|
|
to_time (Optional[str], optional): ISO 8601 datetime string to filter logs not later than this time. Defaults to None.
|
|
tail (bool, optional): If True, read the last lines of the file (like `tail -n`). Defaults to False.
|
|
|
|
Returns:
|
|
List[dict]: A list of filtered log entries as dictionaries.
|
|
|
|
Raises:
|
|
FileNotFoundError: If the log file does not exist.
|
|
ValueError: If the datetime strings are invalid or improperly formatted.
|
|
Exception: For other unforeseen I/O or parsing errors.
|
|
"""
|
|
if not log_path.exists():
|
|
raise FileNotFoundError("Log file not found")
|
|
|
|
try:
|
|
from_dt = pendulum.parse(from_time) if from_time else None
|
|
to_dt = pendulum.parse(to_time) if to_time else None
|
|
except Exception as e:
|
|
raise ValueError(f"Invalid date/time format: {e}")
|
|
|
|
regex_pattern = re.compile(regex) if regex else None
|
|
|
|
def matches_filters(log: dict) -> bool:
|
|
if level and log.get("level", {}).get("name") != level.upper():
|
|
return False
|
|
if contains and contains.lower() not in log.get("message", "").lower():
|
|
return False
|
|
if regex_pattern and not regex_pattern.search(log.get("message", "")):
|
|
return False
|
|
if from_dt or to_dt:
|
|
try:
|
|
log_time = pendulum.parse(log["time"])
|
|
except Exception:
|
|
return False
|
|
if from_dt and log_time < from_dt:
|
|
return False
|
|
if to_dt and log_time > to_dt:
|
|
return False
|
|
return True
|
|
|
|
matched_logs = []
|
|
lines: list[str] = []
|
|
|
|
if tail:
|
|
with log_path.open("rb") as f:
|
|
f.seek(0, 2)
|
|
end = f.tell()
|
|
buffer = bytearray()
|
|
pointer = end
|
|
|
|
while pointer > 0 and len(lines) < limit * 5:
|
|
pointer -= 1
|
|
f.seek(pointer)
|
|
byte = f.read(1)
|
|
if byte == b"\n":
|
|
if buffer:
|
|
line = buffer[::-1].decode("utf-8", errors="ignore")
|
|
lines.append(line)
|
|
buffer.clear()
|
|
else:
|
|
buffer.append(byte[0])
|
|
if buffer:
|
|
line = buffer[::-1].decode("utf-8", errors="ignore")
|
|
lines.append(line)
|
|
lines = lines[::-1]
|
|
else:
|
|
with log_path.open("r", encoding="utf-8", newline=None) as f_txt:
|
|
lines = f_txt.readlines()
|
|
|
|
for line in lines:
|
|
if not line.strip():
|
|
continue
|
|
try:
|
|
log = json.loads(line)
|
|
except json.JSONDecodeError:
|
|
continue
|
|
if matches_filters(log):
|
|
matched_logs.append(log)
|
|
if len(matched_logs) >= limit:
|
|
break
|
|
|
|
return matched_logs
|