diff --git a/.gitignore b/.gitignore index 4dcea9df..b367ad8d 100644 --- a/.gitignore +++ b/.gitignore @@ -1,7 +1,5 @@ *~ -tests/__pycache__/ -management/__pycache__/ -tools/__pycache__/ +**/__pycache__/ externals/ .env .vagrant diff --git a/conf/miabldap-capture.service b/conf/miabldap-capture.service new file mode 100644 index 00000000..98542171 --- /dev/null +++ b/conf/miabldap-capture.service @@ -0,0 +1,11 @@ +[Unit] +Description=MIAB-LDAP log capture daemon +ConditionPathExists=/etc/mailinabox.conf + +[Service] +Type=simple +ExecStart=/usr/bin/python3 %BIN%/management/reporting/capture/capture.py +ExecReload=/bin/kill -HUP $MAINPID + +[Install] +WantedBy=multi-user.target diff --git a/ehdd/postinstall.sh b/ehdd/postinstall.sh index fe0d5394..b630ee70 100755 --- a/ehdd/postinstall.sh +++ b/ehdd/postinstall.sh @@ -14,6 +14,7 @@ if [ -e "$EHDD_IMG" ]; then systemctl disable php7.2-fpm systemctl disable mailinabox systemctl disable fail2ban + systemctl disable miabldap-capture #systemctl disable nsd [ -x /usr/sbin/slapd ] && systemctl disable slapd diff --git a/ehdd/startup.sh b/ehdd/startup.sh index ad4bd6dd..42d707ff 100755 --- a/ehdd/startup.sh +++ b/ehdd/startup.sh @@ -12,5 +12,6 @@ if [ -s /etc/mailinabox.conf ]; then systemctl link -f $(pwd)/conf/mailinabox.service systemctl start fail2ban systemctl restart mailinabox + systemctl start miabldap-capture fi diff --git a/management/backup.py b/management/backup.py index ac60fc5d..5a155512 100755 --- a/management/backup.py +++ b/management/backup.py @@ -252,6 +252,7 @@ def perform_backup(full_backup): service_command("postfix", "stop", quit=True) service_command("dovecot", "stop", quit=True) service_command("slapd", "stop", quit=True) + service_command("miabldap-capture", "stop", quit=True) # Execute a pre-backup script that copies files outside the homedir. # Run as the STORAGE_USER user, not as root. Pass our settings in @@ -281,6 +282,7 @@ def perform_backup(full_backup): get_env(env)) finally: # Start services again. + service_command("miabldap-capture", "start", quit=False) service_command("slapd", "start", quit=False) service_command("dovecot", "start", quit=False) service_command("postfix", "start", quit=False) diff --git a/management/daemon.py b/management/daemon.py index 363f4b56..45827018 100755 --- a/management/daemon.py +++ b/management/daemon.py @@ -696,6 +696,16 @@ def log_failed_login(request): # APP +from daemon_logger import add_python_logging +add_python_logging(app) + +from daemon_ui_common import add_ui_common +add_ui_common(app) + +from daemon_reports import add_reports +add_reports(app, env, authorized_personnel_only) + + if __name__ == '__main__': if "DEBUG" in os.environ: # Turn on Flask debugging. diff --git a/management/daemon_logger.py b/management/daemon_logger.py new file mode 100644 index 00000000..3a3eb1ae --- /dev/null +++ b/management/daemon_logger.py @@ -0,0 +1,111 @@ +# -*- indent-tabs-mode: t; tab-width: 4; python-indent-offset: 4; -*- +import logging +import flask + +# setup our root logger + +# keep a separate logger from app.logger, which only logs WARNING and +# above, doesn't include the module name, or authentication +# details + +class textcolor: + DANGER = '\033[31m' + WARN = '\033[93m' + SUCCESS = '\033[32m' + BOLD = '\033[1m' + FADED= '\033[37m' + RESET = '\033[0m' + + + +class AuthLogFormatter(logging.Formatter): + def __init__(self): + fmt='%(name)s:%(lineno)d(%(username)s/%(client)s): %(levelname)s[%(thread)d]: %(color)s%(message)s%(color_reset)s' + super(AuthLogFormatter, self).__init__(fmt=fmt) + + +# +# when logging, a "client" (oauth client) and/or an explicit +# "username" (in the case the user in question is not logged in but +# you want the username to appear in the logs) may be provided in the +# log.xxx() call as the last argument. eg: +# +# log.warning('login attempt failed', { 'username': email }) +# + +class AuthLogFilter(logging.Filter): + def __init__(self, color_output, get_session_username_function): + self.color_output = color_output + self.get_session_username = get_session_username_function + super(AuthLogFilter, self).__init__() + + ''' add `username` and `client` context info the the LogRecord ''' + def filter(self, record): + record.color = '' + if self.color_output: + if record.levelno == logging.DEBUG: + record.color=textcolor.FADED + elif record.levelno == logging.INFO: + record.color=textcolor.BOLD + elif record.levelno == logging.WARNING: + record.color=textcolor.WARN + elif record.levelno in [logging.ERROR, logging.CRITICAL]: + record.color=textcolor.DANGER + + record.color_reset = textcolor.RESET if record.color else '' + record.client = '-' + record.username = '-' + record.thread = record.thread % 10000 + + opts = None + args_len = len(record.args) + if type(record.args) == dict: + opts = record.args + record.args = () + elif args_len>0 and type(record.args[args_len-1]) == dict: + opts = record.args[args_len-1] + record.args = record.args[0:args_len-1] + + if opts: + record.client = opts.get('client', '-') + record.username = opts.get('username', '-') + + if record.username == '-': + try: + record.username = self.get_session_username() + except (RuntimeError, KeyError): + # not in an HTTP request context or not logged in + pass + + return True + +def get_session_username(): + if flask.request and hasattr(flask.request, 'user_email'): + # this is an admin panel login via "authorized_personnel_only" + return flask.request.user_email + + # otherwise, this may be a user session login + return flask.session['user_id'] + + +def add_python_logging(app): + # log to stdout in development mode + if app.debug: + log_level = logging.DEBUG + log_handler = logging.StreamHandler() + + # log to syslog in production mode + else: + import utils + log_level = logging.INFO + log_handler = utils.create_syslog_handler() + + logging.basicConfig(level=log_level, handlers=[]) + log_handler.setLevel(log_level) + log_handler.addFilter(AuthLogFilter( + app.debug, + get_session_username + )) + log_handler.setFormatter(AuthLogFormatter()) + log = logging.getLogger('') + log.addHandler(log_handler) diff --git a/management/daemon_reports.py b/management/daemon_reports.py new file mode 100644 index 00000000..26927900 --- /dev/null +++ b/management/daemon_reports.py @@ -0,0 +1,216 @@ +# -*- indent-tabs-mode: t; tab-width: 4; python-indent-offset: 4; -*- + +import os +import logging +import json +import datetime +import time +import subprocess + +log = logging.getLogger(__name__) + +from flask import request, Response, session, redirect, jsonify, send_from_directory +from functools import wraps +from reporting.capture.db.SqliteConnFactory import SqliteConnFactory +import reporting.uidata as uidata + +from mailconfig import get_mail_users + + +def add_reports(app, env, authorized_personnel_only): + '''call this function to add reporting/sem endpoints + + `app` is a Flask instance + `env` is the Mail-in-a-Box LDAP environment + `authorized_personnel_only` is a flask wrapper from daemon.py + ensuring only authenticated admins can access the endpoint + + ''' + + CAPTURE_STORAGE_ROOT = os.environ.get( + 'CAPTURE_STORAGE_ROOT', + os.path.join(env['STORAGE_ROOT'], 'reporting') + ) + sqlite_file = os.path.join(CAPTURE_STORAGE_ROOT, 'capture.sqlite') + db_conn_factory = SqliteConnFactory(sqlite_file) + + # UI support + ui_dir = os.path.join(os.path.dirname(app.template_folder), 'reporting/ui') + def send_ui_file(filename): + return send_from_directory(ui_dir, filename) + + @app.route("/reports/ui/", methods=['GET']) + def get_reporting_ui_file(filename): + return send_ui_file(filename) + + @app.route('/reports') + def reporting_redir(): + return redirect('/reports/') + + @app.route('/reports/', methods=['GET']) + def reporting_main(): + return send_ui_file('index.html') + + + # Decorator to unwrap json payloads. It returns the json as a dict + # in named argument 'payload' + def json_payload(func): + @wraps(func) + def wrapper(*args, **kwargs): + try: + log.debug('payload:%s', request.data) + payload = json.loads(request.data) + return func(*args, payload=payload, **kwargs) + except json.decoder.JSONDecodeError as e: + log.warning('Bad request: data:%s ex:%s', request.data, e) + return ("Bad request", 400) + return wrapper + + @app.route('/reports/uidata/messages-sent', methods=['POST']) + @authorized_personnel_only + @json_payload + def get_data_chart_messages_sent(payload): + conn = db_conn_factory.connect() + try: + return jsonify(uidata.messages_sent(conn, payload)) + except uidata.InvalidArgsError as e: + return ('invalid request', 400) + finally: + db_conn_factory.close(conn) + + @app.route('/reports/uidata/messages-received', methods=['POST']) + @authorized_personnel_only + @json_payload + def get_data_chart_messages_received(payload): + conn = db_conn_factory.connect() + try: + return jsonify(uidata.messages_received(conn, payload)) + except uidata.InvalidArgsError as e: + return ('invalid request', 400) + finally: + db_conn_factory.close(conn) + + @app.route('/reports/uidata/user-activity', methods=['POST']) + @authorized_personnel_only + @json_payload + def get_data_user_activity(payload): + conn = db_conn_factory.connect() + try: + return jsonify(uidata.user_activity(conn, payload)) + except uidata.InvalidArgsError as e: + return ('invalid request', 400) + finally: + db_conn_factory.close(conn) + + @app.route('/reports/uidata/flagged-connections', methods=['POST']) + @authorized_personnel_only + @json_payload + def get_data_flagged_connections(payload): + conn = db_conn_factory.connect() + try: + return jsonify(uidata.flagged_connections(conn, payload)) + except uidata.InvalidArgsError as e: + return ('invalid request', 400) + finally: + db_conn_factory.close(conn) + + @app.route('/reports/uidata/remote-sender-activity', methods=['POST']) + @authorized_personnel_only + @json_payload + def get_data_remote_sender_activity(payload): + conn = db_conn_factory.connect() + try: + return jsonify(uidata.remote_sender_activity(conn, payload)) + except uidata.InvalidArgsError as e: + return ('invalid request', 400) + finally: + db_conn_factory.close(conn) + + @app.route('/reports/uidata/user-list', methods=['GET']) + @authorized_personnel_only + def get_data_user_list(): + return jsonify(get_mail_users(env, as_map=False)) + + @app.route('/reports/uidata/select-list-suggestions', methods=['POST']) + @authorized_personnel_only + @json_payload + def suggest(payload): + conn = db_conn_factory.connect() + try: + return jsonify(uidata.select_list_suggestions(conn, payload)) + except uidata.InvalidArgsError as e: + return ('invalid request', 400) + finally: + db_conn_factory.close(conn) + + @app.route('/reports/capture/config', methods=['GET']) + @authorized_personnel_only + def get_capture_config(): + try: + with open("/var/run/mailinabox/runtime_config.json") as fp: + return Response(fp.read(), mimetype="text/json") + except FileNotFoundError: + return jsonify({ 'status':'error', 'reason':'not running' }) + + @app.route('/reports/capture/config', methods=['POST']) + @authorized_personnel_only + @json_payload + def save_capture_config(payload): + try: + with open("/var/run/mailinabox/runtime_config.json") as fp: + loc = json.loads(fp.read()).get('from', { 'type':'unknown' }) + except FileNotFoundError: + return ('service is not running', 403) + + # loc: { type:'file', location:'' } + if loc.get('type') != 'file': + return ('storage type is %s' % loc.get('type'), 403) + + log.warning('overwriting config file %s', loc['location']) + + # remove runtime-config extra fields that don't belong in + # the user config + if 'from' in payload: + del payload['from'] + + with open(loc['location'], "w") as fp: + fp.write(json.dumps(payload, indent=4)) + + r = subprocess.run(["systemctl", "reload", "miabldap-capture"]) + if r.returncode != 0: + log.warning('systemctl reload faild for miabldap-capture: code=%s', r.returncode) + else: + # wait a sec for daemon to pick up new config + # TODO: monitor runtime config for mtime change + time.sleep(1) + # invalidate stats cache. if prune policy changed, the stats + # may be invalid + uidata.clear_cache() + + return ("ok", 200) + + + @app.route('/reports/capture/service/status', methods=['GET']) + @authorized_personnel_only + def get_capture_status(): + service = "miabldap-capture.service" + + if not os.path.exists("/etc/systemd/system/" + service): + return jsonify([ 'not installed', 'not installed' ]) + + r1 = subprocess.run(["systemctl", "is-active", "--quiet", service ]) + r2 = subprocess.run(["systemctl", "is-enabled", "--quiet", service ]) + + return jsonify([ + 'running' if r1.returncode == 0 else 'stopped', + 'enabled' if r2.returncode == 0 else 'disabled' + ]) + + @app.route('/reports/capture/db/stats', methods=['GET']) + @authorized_personnel_only + def get_db_stats(): + conn = db_conn_factory.connect() + try: + return jsonify(uidata.capture_db_stats(conn)) + finally: + db_conn_factory.close(conn) diff --git a/management/daemon_ui_common.py b/management/daemon_ui_common.py new file mode 100644 index 00000000..73e09bd8 --- /dev/null +++ b/management/daemon_ui_common.py @@ -0,0 +1,31 @@ +# -*- indent-tabs-mode: t; tab-width: 4; python-indent-offset: 4; -*- + +import os +import logging +import json +import datetime + +# setup our root logger - named so oauth/*.py files become children +log = logging.getLogger(__name__) + +from flask import request, session, redirect, jsonify, send_from_directory + + + +def add_ui_common(app): + ''' + call this function to add an endpoint that delivers common ui files + + `app` is a Flask instance + ''' + + # UI support + ui_dir = os.path.join(os.path.dirname(app.template_folder), 'ui-common') + def send_ui_file(filename): + return send_from_directory(ui_dir, filename) + + @app.route("/ui-common/", methods=['GET']) + def get_common_ui_file(filename): + return send_ui_file(filename) + + diff --git a/management/reporting/capture/.gitignore b/management/reporting/capture/.gitignore new file mode 100644 index 00000000..fcd595d6 --- /dev/null +++ b/management/reporting/capture/.gitignore @@ -0,0 +1,2 @@ +tests/ +run.sh diff --git a/management/reporting/capture/capture.py b/management/reporting/capture/capture.py new file mode 100755 index 00000000..47b56e0f --- /dev/null +++ b/management/reporting/capture/capture.py @@ -0,0 +1,300 @@ +#!/usr/bin/env python3 +from logs.TailFile import TailFile +from mail.InboundMailLogHandler import InboundMailLogHandler +from logs.ReadPositionStoreInFile import ReadPositionStoreInFile +from db.SqliteConnFactory import SqliteConnFactory +from db.SqliteEventStore import SqliteEventStore +from db.Pruner import Pruner +from util.env import load_env_vars_from_file + +import time +import logging, logging.handlers +import json +import re +import os +import sys +import signal + +log = logging.getLogger(__name__) + +if os.path.exists('/etc/mailinabox.conf'): + env = load_env_vars_from_file('/etc/mailinabox.conf') +else: + env = { 'STORAGE_ROOT': os.environ.get('STORAGE_ROOT', os.getcwd()) } + +CAPTURE_STORAGE_ROOT = os.path.join(env['STORAGE_ROOT'], 'reporting') + + +# default configuration, if not specified or +# CAPTURE_STORAGE_ROOT/config.json does not exist +config_default = { + 'default_config': True, + 'capture': True, + 'prune_policy': { + 'frequency_min': 2400, + 'older_than_days': 30 + }, + 'drop_disposition': { + 'failed_login_attempt': True, + 'suspected_scanner': False, + 'reject': True + } +} + +config_default_file = os.path.join(CAPTURE_STORAGE_ROOT, 'config.json') + + +# options + +options = { + 'daemon': True, + 'log_level': logging.WARNING, + 'log_file': "/var/log/mail.log", + 'pos_file': "/var/lib/mailinabox/capture-pos.json", + 'sqlite_file': os.path.join(CAPTURE_STORAGE_ROOT, 'capture.sqlite'), + 'working_dir': "/var/run/mailinabox", + 'config': config_default, + '_config_file': False, # absolute path if "config" was from a file + '_runtime_config_file': "runtime_config.json" # relative to working_dir +} + + +def read_config_file(file, throw=False): + try: + with open(file) as fp: + newconfig = json.loads(fp.read()) + newconfig['from'] = { 'type':'file', 'location':file } + return newconfig + except FileNotFoundError as e: + if not throw: + return False + raise e + +def write_config(tofile, config): + d = os.path.dirname(tofile) + if d and not os.path.exists(d): + os.mkdir(d, mode=0o770) + with open(tofile, "w") as fp: + fp.write(json.dumps(config)) + +def usage(): + print('usage: %s [options]' % sys.argv[0]) + sys.exit(1) + +def process_cmdline(options): + argi = 1 + marg = 0 + while argi < len(sys.argv): + arg = sys.argv[argi] + have_next = ( argi+1 < len(sys.argv) ) + + if arg=='-d': + options['daemon'] = False + + elif arg=='-loglevel' and have_next: + argi += 1 + arg = sys.argv[argi].lower() + if arg=='info': + options['log_level'] = logging.INFO + elif arg=='warning': + options['log_level'] = logging.WARNING + elif arg=='debug': + options['log_level'] = logging.DEBUG + elif arg=='error': + options['log_level'] = logging.ERROR + else: + sys.stderr.write('unknown log level "%s"\n' % sys.argv[argi]) + + elif arg=='-config' and have_next: + argi += 1 + arg = sys.argv[argi] + try: + if arg.startswith('{'): + options['config'] = json.loads(arg) + else: + newconfig = read_config_file(arg, throw=True) + options['config'] = newconfig + options['_config_file'] = arg + except Exception as e: + if options['daemon']: log.exception(e) + raise e + + elif arg=='-logfile' and have_next: + argi+=1 + options['log_file'] = sys.argv[argi] + + elif arg=='-posfile' and have_next: + argi+=1 + options['pos_file'] = sys.argv[argi] + + elif arg=='-sqlitefile' and have_next: + argi+=1 + options['sqlite_file'] = sys.argv[argi] + + elif arg.startswith('-'): + usage() + + else: + if marg==0: + options['log_file'] = arg + elif marg==1: + options['pos_file'] = arg + elif marg==2: + options['sqlite_file'] = arg + else: + usage() + marg += 1 + argi += 1 + + +def set_working_dir(working_dir): + try: + if not os.path.exists(working_dir): + os.mkdir(working_dir, mode=0o770) + os.chdir(working_dir) + except Exception as e: + log.exception(e) + raise e + +def close_stdio(): + sys.stdout.close() + sys.stderr.close() + sys.stdin.close() + + +# if config.json exists in the default location start with that +# instead of `config_default`. config can still be changed with the +# command line argument "-config" + +newconfig = read_config_file(config_default_file, throw=False) +if newconfig: + options['config'] = newconfig + options['_config_file'] = config_default_file + +# process command line +process_cmdline(options) + + +# init logging & set working directory +if options['daemon']: + logging.basicConfig( + level = options['log_level'], + handlers= [ + logging.handlers.SysLogHandler(address='/dev/log') + ], + format = 'miabldap/capture %(message)s' + ) + log.warning('MIAB-LDAP capture/SEM daemon starting: wd=%s; log_file=%s; pos_file=%s; db=%s', + options['working_dir'], + options['log_file'], + options['pos_file'], + options['sqlite_file'] + ) + close_stdio() + set_working_dir(options['working_dir']) + +else: + logging.basicConfig(level=options['log_level']) + log.info('starting: log_file=%s; pos_file=%s; db=%s', + options['log_file'], + options['pos_file'], + options['sqlite_file'] + ) + + +# save runtime config +write_config(options['_runtime_config_file'], options['config']) + + +# start modules +log.info('config: %s', options['config']) +try: + db_conn_factory = SqliteConnFactory( + options['sqlite_file'] + ) + event_store = SqliteEventStore( + db_conn_factory + ) + position_store = ReadPositionStoreInFile( + options['pos_file'] + ) + mail_tail = TailFile( + options['log_file'], + position_store + ) + inbound_mail_handler = InboundMailLogHandler( + event_store, + capture_enabled = options['config'].get('capture',True), + drop_disposition = options['config'].get('drop_disposition') + ) + mail_tail.add_handler(inbound_mail_handler) + pruner = Pruner( + db_conn_factory, + policy=options['config']['prune_policy'] + ) + pruner.add_prunable(event_store) + +except Exception as e: + if options['daemon']: log.exception(e) + raise e + + +# termination handler for graceful shutdowns +def terminate(sig, stack): + if sig == signal.SIGTERM: + log.warning("shutting down due to SIGTERM") + log.debug("stopping mail_tail") + mail_tail.stop() + log.debug("stopping pruner") + pruner.stop() + log.debug("stopping position_store") + position_store.stop() + log.debug("stopping event_store") + event_store.stop() + try: + os.remove(options['_runtime_config_file']) + except Exception: + pass + log.info("stopped") + +# reload settings handler +def reload(sig, stack): + # if the default config (`config_default`) is in use, check to see + # if a default config.json (`config_default_file`) now exists, and + # if so, use that + if options['config'].get('default_config', False) and os.path.exists(config_default_file): + options['config']['default_config'] = False + options['_config_file'] = config_default_file + + log.info('%s records are in-progress', + inbound_mail_handler.get_inprogress_count()) + + if options['_config_file']: + log.info('reloading %s', options['_config_file']) + try: + newconfig = read_config_file(options['_config_file'], throw=True) + pruner.set_policy( + newconfig['prune_policy'] + ) + inbound_mail_handler.set_capture_enabled( + newconfig.get('capture', True) + ) + inbound_mail_handler.update_drop_disposition( + newconfig.get('drop_disposition', {}) + ) + write_config(options['_runtime_config_file'], newconfig) + except Exception as e: + if options['daemon']: + log.exception(e) + else: + raise e + +signal.signal(signal.SIGTERM, terminate) +signal.signal(signal.SIGINT, terminate) +signal.signal(signal.SIGHUP, reload) + + +# monitor and capture +mail_tail.start() +mail_tail.join() + diff --git a/management/reporting/capture/db/DatabaseConnectionFactory.py b/management/reporting/capture/db/DatabaseConnectionFactory.py new file mode 100644 index 00000000..c4324d2c --- /dev/null +++ b/management/reporting/capture/db/DatabaseConnectionFactory.py @@ -0,0 +1,10 @@ + +class DatabaseConnectionFactory(object): + def connect(self): + raise NotImplementedError() + + def close(self, conn): + raise NotImplementedError() + + + diff --git a/management/reporting/capture/db/EventStore.py b/management/reporting/capture/db/EventStore.py new file mode 100644 index 00000000..f97e8056 --- /dev/null +++ b/management/reporting/capture/db/EventStore.py @@ -0,0 +1,122 @@ +# -*- indent-tabs-mode: t; tab-width: 4; python-indent-offset: 4; -*- + +import threading +import queue +import logging +from .Prunable import Prunable + +log = logging.getLogger(__name__) + + +'''subclass this and override: + + write_rec() + read_rec() + +to provide storage for event "records" + +EventStore is thread safe and uses a single thread to write all +records. + +''' + +class EventStore(Prunable): + def __init__(self, db_conn_factory): + self.db_conn_factory = db_conn_factory + # we'll have a single thread do all the writing to the database + #self.queue = queue.SimpleQueue() # available in Python 3.7+ + self.queue = queue.Queue() + self.interrupt = threading.Event() + self.rec_added = threading.Event() + self.have_event = threading.Event() + self.t = threading.Thread( + target=self._bg_writer, + name="EventStore", + daemon=True + ) + self.max_queue_size = 100000 + self.t.start() + + def connect(self): + return self.db_conn_factory.connect() + + def close(self, conn): + self.db_conn_factory.close(conn) + + def write_rec(self, conn, type, rec): + '''write a "rec" of the given "type" to the database. The subclass + must know how to do that. "type" is a string identifier of the + subclass's choosing. Users of this class should call store() + and not this function, which will queue the request and a + thread managed by this class will call this function. + + ''' + raise NotImplementedError() + + def read_rec(self, conn, type, args): + '''read from the database''' + raise NotImplementedError() + + def prune(self, conn): + raise NotImplementedError() + + def store(self, type, rec): + self.queue.put({ + 'type': type, + 'rec': rec + }) + self.rec_added.set() + self.have_event.set() + + def stop(self): + self.interrupt.set() + self.have_event.set() + self.t.join() + + def __del__(self): + log.debug('EventStore __del__') + self.interrupt.set() + self.have_event.set() + + def _pop(self): + try: + return self.queue.get(block=False) + except queue.Empty: + return None + + def _bg_writer(self): + log.debug('start EventStore thread') + conn = self.connect() + try: + while not self.interrupt.is_set() or not self.queue.empty(): + item = self._pop() + if item: + try: + self.write_rec(conn, item['type'], item['rec']) + except Exception as e: + log.exception(e) + retry_count = item.get('retry_count', 0) + if self.interrupt.is_set(): + log.warning('interrupted, dropping record: %s',item) + elif retry_count > 2: + log.warning('giving up after %s attempts, dropping record: %s', retry_count, item) + elif self.queue.qsize() >= self.max_queue_size: + log.warning('queue full, dropping record: %s', item) + else: + item['retry_count'] = retry_count + 1 + self.queue.put(item) + # wait for another record to prevent immediate retry + if not self.interrupt.is_set(): + self.have_event.wait() + self.rec_added.clear() + self.have_event.clear() + self.queue.task_done() # remove for SimpleQueue + + else: + self.have_event.wait() + self.rec_added.clear() + self.have_event.clear() + + finally: + self.close(conn) + diff --git a/management/reporting/capture/db/Prunable.py b/management/reporting/capture/db/Prunable.py new file mode 100644 index 00000000..f28f9a05 --- /dev/null +++ b/management/reporting/capture/db/Prunable.py @@ -0,0 +1,7 @@ + + +class Prunable(object): + def prune(self, conn, policy): + raise NotImplementedError() + + diff --git a/management/reporting/capture/db/Pruner.py b/management/reporting/capture/db/Pruner.py new file mode 100644 index 00000000..063b29ae --- /dev/null +++ b/management/reporting/capture/db/Pruner.py @@ -0,0 +1,79 @@ + +import threading +import logging + +log = logging.getLogger(__name__) + + +class Pruner(object): + '''periodically calls the prune() method of registered Prunable + objects + + ''' + def __init__(self, db_conn_factory, policy={ + 'older_than_days': 7, + 'frequency_min': 60 + }): + self.db_conn_factory = db_conn_factory + self.policy = policy + self.prunables = [] + self.interrupt = threading.Event() + self._new_thread() + self.t.start() + + + def _new_thread(self): + self.interrupt.clear() + self.t = threading.Thread( + target=self._bg_pruner, + name="Pruner", + daemon=True + ) + + def add_prunable(self, inst): + self.prunables.append(inst) + + def set_policy(self, policy): + self.stop() + self.policy = policy + # a new thread object must be created or Python(<3.8?) throws + # RuntimeError("threads can only be started once") + self._new_thread() + self.t.start() + + def stop(self, do_join=True): + self.interrupt.set() + if do_join: + self.t.join() + + def connect(self): + return self.db_conn_factory.connect() + + def close(self, conn): + self.db_conn_factory.close(conn) + + def __del__(self): + self.stop(do_join=False) + + def _bg_pruner(self): + conn = self.connect() + + def do_prune(): + for prunable in self.prunables: + if not self.interrupt.is_set(): + try: + prunable.prune(conn, self.policy) + except Exception as e: + log.exception(e) + + try: + # prune right-off + do_prune() + + while not self.interrupt.is_set(): + # wait until interrupted or it's time to prune + if self.interrupt.wait(self.policy['frequency_min'] * 60) is not True: + do_prune() + + finally: + self.close(conn) diff --git a/management/reporting/capture/db/SqliteConnFactory.py b/management/reporting/capture/db/SqliteConnFactory.py new file mode 100644 index 00000000..44833063 --- /dev/null +++ b/management/reporting/capture/db/SqliteConnFactory.py @@ -0,0 +1,52 @@ +import os, stat +import sqlite3 +import logging +import threading + +from .DatabaseConnectionFactory import DatabaseConnectionFactory + +log = logging.getLogger(__name__) + + +class SqliteConnFactory(DatabaseConnectionFactory): + def __init__(self, db_path): + super(SqliteConnFactory, self).__init__() + log.debug('factory for %s', db_path) + self.db_path = db_path + self.db_basename = os.path.basename(db_path) + self.ensure_exists() + + def ensure_exists(self): + # create the parent directory and set its permissions + parent = os.path.dirname(self.db_path) + if parent != '' and not os.path.exists(parent): + os.makedirs(parent) + os.chmod(parent, + stat.S_IRWXU | + stat.S_IRGRP | + stat.S_IXGRP | + stat.S_IROTH | + stat.S_IXOTH + ) + + # if the database is new, create an empty file and set file + # permissions + if not os.path.exists(self.db_path): + log.debug('creating empty database: %s', self.db_basename) + with open(self.db_path, 'w') as fp: + pass + + os.chmod(self.db_path, + stat.S_IRUSR | + stat.S_IWUSR + ) + + def connect(self): + log.debug('opening database %s', self.db_basename) + conn = sqlite3.connect(self.db_path) + conn.row_factory = sqlite3.Row + return conn + + def close(self, conn): + log.debug('closing database %s', self.db_basename) + conn.close() diff --git a/management/reporting/capture/db/SqliteEventStore.py b/management/reporting/capture/db/SqliteEventStore.py new file mode 100644 index 00000000..61c54b82 --- /dev/null +++ b/management/reporting/capture/db/SqliteEventStore.py @@ -0,0 +1,368 @@ +# -*- indent-tabs-mode: t; tab-width: 4; python-indent-offset: 4; -*- + +import sqlite3 +import os, stat +import logging +import json +import datetime +from .EventStore import EventStore + +log = logging.getLogger(__name__) + +# +# schema +# +mta_conn_fields = [ + 'service', + 'service_tid', + 'connect_time', + 'disconnect_time', + 'remote_host', + 'remote_ip', + 'sasl_method', + 'sasl_username', + 'remote_auth_success', + 'remote_auth_attempts', + 'remote_used_starttls', + 'disposition', +] + +mta_accept_fields = [ + 'mta_conn_id', + 'queue_time', + 'queue_remove_time', + 'subsystems', +# 'spf_tid', + 'spf_result', + 'spf_reason', + 'postfix_msg_id', + 'message_id', + 'dkim_result', + 'dkim_reason', + 'dmarc_result', + 'dmarc_reason', + 'envelope_from', + 'message_size', + 'message_nrcpt', + 'accept_status', + 'failure_info', + 'failure_category', +] + +mta_delivery_fields = [ + 'mta_accept_id', + 'service', +# 'service_tid', + 'rcpt_to', +# 'postgrey_tid', + 'postgrey_result', + 'postgrey_reason', + 'postgrey_delay', +# 'spam_tid', + 'spam_result', + 'spam_score', + 'relay', + 'status', + 'delay', + 'delivery_connection', + 'delivery_connection_info', + 'delivery_info', + 'failure_category', +] + + +db_info_create_table_stmt = "CREATE TABLE IF NOT EXISTS db_info(id INTEGER PRIMARY KEY AUTOINCREMENT, key TEXT NOT NULL, value TEXT NOT NULL)" + +schema_updates = [ + # update 0 + [ + # three "mta" tables having a one-to-many-to-many relationship: + # mta_connection(1) -> mta_accept(0:N) -> mta_delivery(0:N) + # + + "CREATE TABLE mta_connection(\ + mta_conn_id INTEGER PRIMARY KEY AUTOINCREMENT,\ + service TEXT NOT NULL, /* 'smtpd', 'submission' or 'pickup' */\ + service_tid TEXT NOT NULL,\ + connect_time TEXT NOT NULL,\ + disconnect_time TEXT,\ + remote_host TEXT COLLATE NOCASE,\ + remote_ip TEXT COLLATE NOCASE,\ + sasl_method TEXT, /* sasl: submission service only */\ + sasl_username TEXT COLLATE NOCASE,\ + remote_auth_success INTEGER, /* count of successes */\ + remote_auth_attempts INTEGER, /* count of attempts */\ + remote_used_starttls INTEGER, /* 1 if STARTTLS used */\ + disposition TEXT /* 'normal','scanner','login_attempt',etc */\ + )", + + "CREATE INDEX idx_mta_connection_connect_time ON mta_connection(connect_time, sasl_username COLLATE NOCASE)", + + + "CREATE TABLE mta_accept(\ + mta_accept_id INTEGER PRIMARY KEY AUTOINCREMENT,\ + mta_conn_id INTEGER,\ + queue_time TEXT,\ + queue_remove_time TEXT,\ + subsystems TEXT,\ + /*spf_tid TEXT,*/\ + spf_result TEXT,\ + spf_reason TEXT,\ + postfix_msg_id TEXT,\ + message_id TEXT,\ + dkim_result TEXT,\ + dkim_reason TEXT,\ + dmarc_result TEXT,\ + dmarc_reason TEXT,\ + envelope_from TEXT COLLATE NOCASE,\ + message_size INTEGER,\ + message_nrcpt INTEGER,\ + accept_status TEXT, /* 'accept','greylist','spf-reject',others... */\ + failure_info TEXT, /* details from mta or subsystems */\ + failure_category TEXT,\ + FOREIGN KEY(mta_conn_id) REFERENCES mta_connection(mta_conn_id) ON DELETE RESTRICT\ + )", + + "CREATE TABLE mta_delivery(\ + mta_delivery_id INTEGER PRIMARY KEY AUTOINCREMENT,\ + mta_accept_id INTEGER,\ + service TEXT, /* 'lmtp' or 'smtp' */\ + /*service_tid TEXT,*/\ + rcpt_to TEXT COLLATE NOCASE, /* email addr */\ + /*postgrey_tid TEXT,*/\ + postgrey_result TEXT,\ + postgrey_reason TEXT,\ + postgrey_delay NUMBER,\ + /*spam_tid TEXT,*/ /* spam: lmtp only */\ + spam_result TEXT, /* 'clean' or 'spam' */\ + spam_score NUMBER, /* eg: 2.10 */\ + relay TEXT, /* hostname[IP]:port */\ + status TEXT, /* 'sent', 'bounce', 'reject', etc */\ + delay NUMBER, /* fractional seconds, 'sent' status only */\ + delivery_connection TEXT, /* 'trusted' or 'untrusted' */\ + delivery_connection_info TEXT, /* details on TLS connection */\ + delivery_info TEXT, /* details from the remote mta */\ + failure_category TEXT,\ + FOREIGN KEY(mta_accept_id) REFERENCES mta_accept(mta_accept_id) ON DELETE RESTRICT\ + )", + + "CREATE INDEX idx_mta_delivery_rcpt_to ON mta_delivery(rcpt_to COLLATE NOCASE)", + + "CREATE TABLE state_cache(\ + state_cache_id INTEGER PRIMARY KEY AUTOINCREMENT,\ + owner_id INTEGER NOT NULL,\ + state TEXT\ + )", + + "INSERT INTO db_info (key,value) VALUES ('schema_version', '0')" + ] +] + + + +class SqliteEventStore(EventStore): + + def __init__(self, db_conn_factory): + super(SqliteEventStore, self).__init__(db_conn_factory) + self.update_schema() + + def update_schema(self): + ''' update the schema to the latest version + + ''' + c = None + conn = None + try: + conn = self.connect() + c = conn.cursor() + c.execute(db_info_create_table_stmt) + conn.commit() + c.execute("SELECT value from db_info WHERE key='schema_version'") + v = c.fetchone() + if v is None: + v = -1 + else: + v = int(v[0]) + for idx in range(v+1, len(schema_updates)): + log.info('updating database to v%s', idx) + for stmt in schema_updates[idx]: + try: + c.execute(stmt) + except Exception as e: + log.error('problem with sql statement at version=%s error="%s" stmt="%s"' % (idx, e, stmt)) + raise e + + conn.commit() + + finally: + if c: c.close(); c=None + if conn: self.close(conn); conn=None + + + + def write_rec(self, conn, type, rec): + if type=='inbound_mail': + #log.debug('wrote inbound_mail record') + self.write_inbound_mail(conn, rec) + elif type=='state': + ''' rec: { + owner_id: int, + state: list + } + ''' + self.write_state(conn, rec) + else: + raise ValueError('type "%s" not implemented' % type) + + + def _insert(self, table, fields): + insert = 'INSERT INTO ' + table + ' (' + \ + ",".join(fields) + \ + ') VALUES (' + \ + "?,"*(len(fields)-1) + \ + '?)' + return insert + + def _values(self, fields, data_dict): + values = [] + for field in fields: + if field in data_dict: + values.append(data_dict[field]) + data_dict.pop(field) + else: + values.append(None) + + for field in data_dict: + if type(data_dict[field]) != list and not field.startswith('_') and not field.endswith('_tid'): + log.warning('unused field: %s', field) + return values + + + def write_inbound_mail(self, conn, rec): + c = None + try: + c = conn.cursor() + + # mta_connection + insert = self._insert('mta_connection', mta_conn_fields) + values = self._values(mta_conn_fields, rec) + #log.debug('INSERT: %s VALUES: %s REC=%s', insert, values, rec) + c.execute(insert, values) + conn_id = c.lastrowid + + accept_insert = self._insert('mta_accept', mta_accept_fields) + delivery_insert = self._insert('mta_delivery', mta_delivery_fields) + for accept in rec.get('mta_accept', []): + accept['mta_conn_id'] = conn_id + values = self._values(mta_accept_fields, accept) + c.execute(accept_insert, values) + accept_id = c.lastrowid + + for delivery in accept.get('mta_delivery', []): + delivery['mta_accept_id'] = accept_id + values = self._values(mta_delivery_fields, delivery) + c.execute(delivery_insert, values) + + conn.commit() + + except sqlite3.Error as e: + conn.rollback() + raise e + + finally: + if c: c.close(); c=None + + + + def write_state(self, conn, rec): + c = None + try: + c = conn.cursor() + + owner_id = rec['owner_id'] + insert = 'INSERT INTO state_cache (owner_id, state) VALUES (?, ?)' + for item in rec['state']: + item_json = json.dumps(item) + c.execute(insert, (owner_id, item_json)) + + conn.commit() + + except sqlite3.Error as e: + conn.rollback() + raise e + + finally: + if c: c.close(); c=None + + + def read_rec(self, conn, type, args): + if type=='state': + return self.read_state( + conn, + args['owner_id'], + args.get('clear',False) + ) + else: + raise ValueError('type "%s" not implemented' % type) + + def read_state(self, conn, owner_id, clear): + c = None + state = [] + try: + c = conn.cursor() + select = 'SELECT state FROM state_cache WHERE owner_id=? ORDER BY state_cache_id' + for row in c.execute(select, (owner_id,)): + state.append(json.loads(row[0])) + + if clear: + delete = 'DELETE FROM state_cache WHERE owner_id=?' + c.execute(delete, (owner_id,)) + conn.commit() + + finally: + if c: c.close(); c=None + + return state + + def prune(self, conn, policy): + older_than_days = datetime.timedelta(days=policy['older_than_days']) + if older_than_days.days <= 0: + return + now = datetime.datetime.now(datetime.timezone.utc) + d = (now - older_than_days) + dstr = d.isoformat(sep=' ', timespec='seconds') + + c = None + try: + c = conn.cursor() + deletes = [ + 'DELETE FROM mta_delivery WHERE mta_accept_id IN (\ + SELECT mta_accept.mta_accept_id FROM mta_accept\ + JOIN mta_connection ON mta_connection.mta_conn_id = mta_accept.mta_conn_id\ + WHERE connect_time < ?)', + + 'DELETE FROM mta_accept WHERE mta_accept_id IN (\ + SELECT mta_accept.mta_accept_id FROM mta_accept\ + JOIN mta_connection ON mta_connection.mta_conn_id = mta_accept.mta_conn_id\ + WHERE connect_time < ?)', + + 'DELETE FROM mta_connection WHERE connect_time < ?' + ] + + counts = [] + for delete in deletes: + c.execute(delete, (dstr,)) + counts.append(str(c.rowcount)) + conn.commit() + counts.reverse() + log.info("pruned %s rows", "/".join(counts)) + + except sqlite3.Error as e: + conn.rollback() + raise e + + finally: + if c: c.close() + + + diff --git a/management/reporting/capture/db/__init__.py b/management/reporting/capture/db/__init__.py new file mode 100644 index 00000000..e69de29b diff --git a/management/reporting/capture/logs/DateParser.py b/management/reporting/capture/logs/DateParser.py new file mode 100644 index 00000000..3e321c13 --- /dev/null +++ b/management/reporting/capture/logs/DateParser.py @@ -0,0 +1,33 @@ +import datetime +import pytz + +rsyslog_traditional_regexp = '^(.{15})' + +with open('/etc/timezone') as fp: + timezone_id = fp.read().strip() + + +def rsyslog_traditional(str): + # Handles the default timestamp in rsyslog + # (RSYSLOG_TraditionalFileFormat) + # + # eg: "Dec 6 06:25:04" (always 15 characters) + # + # the date string is in local time + # + d = datetime.datetime.strptime(str, '%b %d %H:%M:%S') + + # since the log date has no year, use the current year + today = datetime.date.today() + year = today.year + if d.month == 12 and today.month == 1: + year -= 1 + d = d.replace(year=year) + + # convert to UTC + if timezone_id == 'Etc/UTC': + return d + local_tz = pytz.timezone(timezone_id) + return local_tz.localize(d, is_dst=None).astimezone(pytz.utc) + + diff --git a/management/reporting/capture/logs/ReadLineHandler.py b/management/reporting/capture/logs/ReadLineHandler.py new file mode 100644 index 00000000..a0b629f9 --- /dev/null +++ b/management/reporting/capture/logs/ReadLineHandler.py @@ -0,0 +1,15 @@ + +'''subclass this and override methods to handle log output''' +class ReadLineHandler(object): + def handle(self, line): + ''' handle a single line of output ''' + raise NotImplementedError() + + def end_of_callbacks(self, thread): + '''called when no more output will be sent to handle(). override this + method to save state, or perform cleanup during this + callback + + ''' + pass + diff --git a/management/reporting/capture/logs/ReadPositionStore.py b/management/reporting/capture/logs/ReadPositionStore.py new file mode 100644 index 00000000..cbd6a4ec --- /dev/null +++ b/management/reporting/capture/logs/ReadPositionStore.py @@ -0,0 +1,26 @@ +'''subclass this and override all methods to persist the position of +the log file that has been processed so far. + +this enables the log monitor to pick up where it left off + +a single FilePositionStore can safely be used with multiple +LogMonitor instances + +''' +class ReadPositionStore(object): + def get(self, log_file, inode): + '''return the offset from the start of the file of the last + position saved for log_file having the given inode, or zero if + no position is currently saved + + ''' + raise NotImplementedError() + + def save(self, log_file, inode, offset): + '''save the current position''' + raise NotImplementedError() + + def clear(self, log_file): + '''remove all entries for `log_file`''' + raise NotImplementedError() + diff --git a/management/reporting/capture/logs/ReadPositionStoreInFile.py b/management/reporting/capture/logs/ReadPositionStoreInFile.py new file mode 100644 index 00000000..15f4c786 --- /dev/null +++ b/management/reporting/capture/logs/ReadPositionStoreInFile.py @@ -0,0 +1,84 @@ +from .ReadPositionStore import ReadPositionStore + +import threading +import json +import os +import logging + +log = logging.getLogger(__name__) + + +class ReadPositionStoreInFile(ReadPositionStore): + def __init__(self, output_file): + self.output_file = output_file + self.changed = False + self.lock = threading.Lock() + self.interrupt = threading.Event() + + if os.path.exists(output_file): + with open(output_file, "r", encoding="utf-8") as fp: + self.db = json.loads(fp.read()) + else: + self.db = {} + + self.t = threading.Thread( + target=self._persist_bg, + name="ReadPositionStoreInFile", + daemon=True + ) + self.t.start() + + def __del__(self): + log.debug('ReadPositionStoreInFile __del__') + self.interrupt.set() + + def stop(self): + self.interrupt.set() + self.t.join() + + def get(self, file, inode): + with self.lock: + if file in self.db and str(inode) in self.db[file]: + return self.db[file][str(inode)] + return 0 + + def save(self, file, inode, pos): + with self.lock: + if not file in self.db: + self.db[file] = { str(inode):pos } + else: + self.db[file][str(inode)] = pos + self.changed = True + + def clear(self, file): + with self.lock: + self.db[file] = {} + self.changed = True + + + def persist(self): + if self.changed: + try: + with open(self.output_file, "w") as fp: + with self.lock: + json_str = json.dumps(self.db) + self.changed = False + + try: + fp.write(json_str) + except Exception as e: + with self.lock: + self.changed = True + log.error(e) + + except Exception as e: + log.error(e) + + + def _persist_bg(self): + while not self.interrupt.is_set(): + # wait 60 seconds before persisting + self.interrupt.wait(60) + # even if interrupted, persist one final time + self.persist() + diff --git a/management/reporting/capture/logs/TailFile.py b/management/reporting/capture/logs/TailFile.py new file mode 100644 index 00000000..6882c51c --- /dev/null +++ b/management/reporting/capture/logs/TailFile.py @@ -0,0 +1,160 @@ +import threading +import os +import logging +import stat + +from .ReadLineHandler import ReadLineHandler + +log = logging.getLogger(__name__) + +'''Spawn a thread to "tail" a log file. For each line read, provided +callbacks do something with the output. Callbacks must be a subclass +of ReadLineHandler. + +''' + +class TailFile(threading.Thread): + def __init__(self, log_file, store=None): + ''' log_file - the log file to monitor + store - a ReadPositionStore instance + ''' + self.log_file = log_file + self.store = store + + self.fp = None + self.inode = None + self.callbacks = [] + self.interrupt = threading.Event() + + name=f'{__name__}-{os.path.basename(log_file)}' + log.debug('init thread: %s', name) + super(TailFile, self).__init__(name=name, daemon=True) + + def stop(self, do_join=True): + log.debug('TailFile stopping') + self.interrupt.set() + # close must be called to unblock the thread fp.readline() call + self._close() + if do_join: + self.join() + + def __del__(self): + self.stop(do_join=False) + + def add_handler(self, fn): + assert self.is_alive() == False + self.callbacks.append(fn) + + def clear_callbacks(self): + assert self.is_alive() == False + self.callbacks = [] + + def _open(self): + self._close() + self.inode = os.stat(self.log_file)[stat.ST_INO] + self.fp = open( + self.log_file, + "r", + encoding="utf-8", + errors="backslashreplace" + ) + + def _close(self): + if self.fp is not None: + self.fp.close() + self.fp = None + + def _is_rotated(self): + try: + return os.stat(self.log_file)[stat.ST_INO] != self.inode + except FileNotFoundError: + return False + + def _issue_callbacks(self, line): + for cb in self.callbacks: + if isinstance(cb, ReadLineHandler): + cb.handle(line) + else: + cb(line) + + def _notify_end_of_callbacks(self): + for cb in self.callbacks: + if isinstance(cb, ReadLineHandler): + cb.end_of_callbacks(self) + + def _restore_read_position(self): + if self.fp is None: + return + + if self.store is None: + self.fp.seek( + 0, + os.SEEK_END + ) + else: + pos = self.store.get(self.log_file, self.inode) + size = os.stat(self.log_file)[stat.ST_SIZE] + if size < pos: + log.debug("truncated: %s" % self.log_file) + self.fp.seek(0, os.SEEK_SET) + else: + # if pos>size here, the seek call succeeds and returns + # 'pos', but future reads will fail + self.fp.seek(pos, os.SEEK_SET) + + def run(self): + self.interrupt.clear() + + # initial open - wait until file exists + while not self.interrupt.is_set() and self.fp is None: + try: + self._open() + except FileNotFoundError: + log.debug('log file "%s" not found, waiting...', self.log_file) + self.interrupt.wait(2) + continue + + # restore reading position + self._restore_read_position() + + while not self.interrupt.is_set(): + try: + line = self.fp.readline() # blocking + if line=='': + log.debug('got EOF') + # EOF - check if file was rotated + if self._is_rotated(): + log.debug('rotated') + self._open() + if self.store is not None: + self.store.clear(self.log_file) + + # if not rotated, sleep + else: + self.interrupt.wait(1) + + else: + # save position and call all callbacks + if self.store is not None: + self.store.save( + self.log_file, + self.inode, + self.fp.tell() + ) + self._issue_callbacks(line) + + except Exception as e: + log.exception(e) + if self.interrupt.wait(1) is not True: + if self._is_rotated(): + self._open() + + + self._close() + + try: + self._notify_end_of_callbacks() + except Exception as e: + log.exception(e) + + diff --git a/management/reporting/capture/logs/__init__.py b/management/reporting/capture/logs/__init__.py new file mode 100644 index 00000000..e69de29b diff --git a/management/reporting/capture/mail/InboundMailLogHandler.py b/management/reporting/capture/mail/InboundMailLogHandler.py new file mode 100644 index 00000000..5dea3662 --- /dev/null +++ b/management/reporting/capture/mail/InboundMailLogHandler.py @@ -0,0 +1,1465 @@ +import logging +import re +import datetime +import traceback +import ipaddress +import threading +from logs.ReadLineHandler import ReadLineHandler +import logs.DateParser +from db.EventStore import EventStore +from util.DictQuery import DictQuery +from util.safe import (safe_int, safe_append, safe_del) +from .PostfixLogParser import PostfixLogParser + +log = logging.getLogger(__name__) + + +STATE_CACHE_OWNER_ID = 1 + + +class InboundMailLogHandler(ReadLineHandler): + ''' + ''' + def __init__(self, record_store, + date_regexp = logs.DateParser.rsyslog_traditional_regexp, + date_parser_fn = logs.DateParser.rsyslog_traditional, + capture_enabled = True, + drop_disposition = None + ): + ''' EventStore instance for persisting "records" ''' + self.record_store = record_store + self.set_capture_enabled(capture_enabled) + + # A "record" is composed by parsing all the syslog output from + # the activity generated by the MTA (postfix) from a single + # remote connection. Once a full history of the connection, + # including delivery of any messages is complete, the record + # is written to the record_store. + # + # `recs` is an array holding incomplete, in-progress + # "records". This array has the following format: + # + # (For convenience, it's easier to refer to the table column + # names found in SqliteEventStore for the dict member names that + # are used here since they're all visible in one place.) + # + # [{ + # ... fields of the mta_connection table ... + # mta_accept: [ + # { + # ... fields of the mta_accept table ... + # mta_delivery: [ + # { + # ... fields of the mta_delivery table ... + # }, + # ... + # ], + # }, + # ... + # ], + # }] + # + # IMPORTANT: + # + # No methods in this class are safe to call by any thread + # other than the caller of handle(), unless marked as + # thread-safe. + # + + # our in-progress record queue is a simple list + self.recs = self.get_cached_state(clear=True) + + # maximum size of the in-progress record queue + self.current_inprogress_recs = len(self.recs) + self.max_inprogress_recs = 100 + + # records that have these dispositions will be dropped (not + # recorded in the record store + self.drop_disposition_lock = threading.Lock() + self.drop_disposition = { + 'failed_login_attempt': False, + 'suspected_scanner': False, + 'reject': False + } + self.update_drop_disposition(drop_disposition) + + # regular expression that matches a syslog date (always anchored) + self.date_regexp = date_regexp + if date_regexp.startswith('^'): + self.date_regexp = date_regexp[1:] + + # function that parses the syslog date + self.date_parser_fn = date_parser_fn + + + # 1. 1a. postfix/smtpd[13698]: connect from host.tld[1.2.3.4] + # 1=date + # 2=service ("submission/smptd" or "smtpd") + # 3=service_tid + # 4=remote_host + # 5=remote_ip + self.re_connect_from = re.compile('^' + self.date_regexp + ' [^ ]+ postfix/(submission/smtpd|smtpd)\[(\d+)\]: connect from ([^\[]+)\[([^\]]+)\]') + + # 1b. Dec 6 07:01:39 mail postfix/pickup[7853]: A684B1F787: uid=0 from= + # 1=date + # 2=service ("pickup") + # 3=service_tid + # 4=postfix_msg_id + self.re_local_pickup = re.compile('^' + self.date_regexp + ' [^ ]+ postfix/(pickup)\[(\d+)\]: ([A-F0-9]+): ') + + # 2. policyd-spf[13703]: prepend Received-SPF: Pass (mailfrom) identity=mailfrom; client-ip=1.2.3.4 helo=host.tld; envelope-from=alice@post.com; receiver= + # 1=spf_tid + # 2=spf_result + # 3=spf_reason "(mailfrom)" + self.re_spf_1 = re.compile('policyd-spf\[(\d+)\]: prepend Received-SPF: ([^ ]+) (\([^\)]+\)){0,1}') + + # 2a. policyd-spf[26231]: 550 5.7.23 Message rejected due to: Receiver policy for SPF Softfail. Please see http://www.openspf.net/Why?s=mfrom;id=test@google.com;ip=1.2.3.4;r= + # 1=spf_tid + # 2=spf_reason + self.re_spf_2 = re.compile('policyd-spf\[(\d+)\]: (\d\d\d \d+\.\d+\.\d+ Message rejected [^;]+)') + + # 3. Dec 9 14:46:57 mail postgrey[879]: action=greylist, reason=new, client_name=host.tld, client_address=1.2.3.4/32, sender=alice@post.com, recipient=mia@myhost.com + # 3a. Dec 6 18:31:28 mail postgrey[879]: 0E98D1F787: action=pass, reason=triplet found, client_name=host.tld, client_address=1.2.3.4/32, sender=alice@post.com, recipient=mia@myhost.com + # 1=postgrey_tid + # 2=postfix_msg_id (re-1 only) + self.re_postgrey_1 = re.compile('postgrey\[(\d+)\]: ([A-F0-9]+):') + self.re_postgrey_2 = re.compile('postgrey\[(\d+)\]:') + + # 4b. postfix/submission/smtpd[THREAD-ID]: NOQUEUE: reject: RCPT from unknown[1.2.3.4]: 550 5.7.23 : Recipient address rejected: Message rejected due to: Receiver policy for SPF Softfail. Please see http://www.openspf.net/Why?s=mfrom;id=test@google.com;ip=10.0.2.15;r=; from= to= proto=ESMTP helo= + # 4c. postfix/smtpd[THREAD-ID]: "NOQUEUE: reject: ....: Recipient address rejected: Greylisted, seehttp://postgrey.../; ..." + # 1=service ("submission/smtpd" or "smtpd") + # 2=service_tid + # 3=accept_status ("reject") + self.re_postfix_noqueue = re.compile('postfix/(submission/smtpd|smtpd)\[(\d+)\]: NOQUEUE: ([^:]+): ') + + # 4. postfix/smtpd[THREAD-ID]: "POSTFIX-MSG-ID" (eg: "DD95A1F796"): client=DNS[IP] + # 4a. postfix/submission/smtpd[THREAD-ID]: POSTFIX-MSG-ID: client=DNS[IP], sasl_method=LOGIN, sasl_username=mia@myhost.com + # 1=service ("submission/smtpd" or "smtpd") + # 2=service_tid + # 3=postfix_msg_id + self.re_postfix_msg_id = re.compile('postfix/(submission/smtpd|smtpd)\[(\d+)\]: ([A-F0-9]+):') + + # 5. Dec 10 06:48:48 mail postfix/cleanup[7435]: 031AF20076: message-id=<20201210114848.031AF20076@myhost.com> + # 1=postfix_msg_id + # 2=message_id + self.re_postfix_message_id = re.compile('postfix/cleanup\[\d+\]: ([A-F0-9]+): message-id=(<[^>]*>)') + + # 6. opendkim: POSTFIX-MSG-ID: + # Dec 6 08:21:33 mail opendkim[6267]: DD95A1F796: s=pf2014 d=github.com SSL + # SSL: + # source: https://sourceforge.net/p/opendkim/git/ci/master/tree/opendkim/opendkim.c + # function: dkimf_log_ssl_errors(), output can be: + # s=pf2014 d=github.com SSL + # SSL + # if is empty, no error + + # 1=postfix_msg_id + # 2=verification detail + # 3=error-msg + self.re_opendkim_ssl = re.compile('opendkim\[\d+\]: ([A-F0-9]+):(.*) SSL ?(.*)') + # 1=postfix_msg_id + # 2=error-msg + self.re_opendkim_error = re.compile('opendkim\[\d+\]: ([A-F0-9]+): (.*)') + + # 7. opendmarc: POSTFIX-MSG-ID: result: [pass/fail] + # Dec 6 08:21:33 mail opendmarc[729]: DD95A1F796 ignoring Authentication-Results at 18 from mx.google.com + # Dec 6 08:21:33 mail opendmarc[729]: DD95A1F796: github.com pass + # Dec 6 13:46:30 mail opendmarc[729]: 0EA8F1FB12: domain.edu fail + # 1=postfix_msg_id + # 2=domain + # 3="pass","none","fail" + self.re_opendmarc_result = re.compile('opendmarc\[\d+\]: ([A-F0-9]+): ([^ ]+) ([^\s]+)') + + + # 13. postfix/qmgr: POSTFIX-MSG-ID: "removed" + # Dec 11 08:30:15 mail postfix/qmgr[9021]: C01F71F787: removed + # 1=date + # 2=postfix_msg_id + self.re_queue_removed = re.compile('^' + self.date_regexp + ' [^ ]+ postfix/qmgr\[\d+\]: ([A-F0-9]+): removed') + + # 8. postfix/qmgr: POSTFIX-MSG-ID: from=user@tld, size=N, nrcpt=1 (queue active) + # 1=date + # 2=postfix_msg_id + # 3=all comma-separated key-value pairs (must be split) + self.re_queue_added = re.compile('^' + self.date_regexp + ' [^ ]+ postfix/qmgr\[\d+\]: ([A-F0-9]+): (?!removed)') + + # 11. spampd: "clean message |(unknown) (SCORE/MAX-SCORE) from for in 1.51s, N bytes" + # 11(a) spampd: "identified spam |(unknown) (5.12/5.00) from for in 1.51s, N bytes" + # 1=spam_tid + # 2="clean message" | "identified spam" (other?) + # 3=message_id ("(unknown)" if message id is "<>") + # 4=spam_score + # 5=envelope_from + # 6=rcpt_to + self.re_spampd = re.compile('spampd\[(\d+)\]: (clean message|identified spam) (<[^>]*>|\(unknown\)) \((-?\d+\.\d+)/\d+\.\d+\) from <([^>]+)> for <([^>]+)>') + + + # 10. postfix/smtpd[THREAD-ID]: disconnect-from: [starttls=1,auth=0] + # 10a. postfix/submission/smtpd[THREAD-ID]: disconnect-from: [starttls=1,auth=1] + # 1=date + # 2=service ("submission/smptd" or "smtpd") + # 3=service_tid + # 4=remote_host + # 5=remote_ip + self.re_disconnect = re.compile('^' + self.date_regexp + ' [^ ]+ postfix/(submission/smtpd|smtpd)\[(\d+)\]: disconnect from ([^\[]+)\[([^\]]+)\]') + + + # postfix/smtp[18333]: Trusted TLS connection established to mx01.mail.icloud.com[17.57.154.23]:25: TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits) + # postfix/smtp[566]: Untrusted TLS connection established to xyz.tld[1.2.3.4]:25: TLSv1.2 with cipher AES128-GCM-SHA256 (128/128 bits) + # postfix/smtp[15125]: Verified TLS connection established to mx1.comcast.net[96.114.157.80]:25: TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits) + # 1=service ("smtp") + # 2=service_tid + # 3=delivery_connection ("trusted", "untrusted", "verified") + # 4=tls details + self.re_pre_delivery = re.compile('postfix/(smtp)\[(\d+)\]: ([^ ]+) (TLS connection established.*)') + + # 12. postfix/lmtp: POSTFIX-MSG-ID: to=, relay=127.0.0.1[127.0.0.1]:10025, delay=4.7, delays=1/0.01/0.01/3.7, dsn=2.0.0, status=sent (250 2.0.0 YB5nM1eS01+lSgAAlWWVsw Saved) + # 12a. postfix/lmtp: POSTFIX_MSG-ID: to=user@tld, status=bounced (host...said...550 5.1.1 User doesn't exist ....) + # 12b. postfix/smtp[32052]: A493B1FAF1: to=, relay=mx.where.net[64.65.66.104]:25, delay=1.2, delays=0.65/0.06/0.4/0.09, dsn=2.0.0, status=sent (250 2.0.0 OK 7E/38-26906-CDC5DCF5): None + + # 1=system ("lmtp" or "smtp") + # 2=system_tid + # 3=postfix_msg_id + self.re_delivery = re.compile('postfix/(lmtp|smtp)\[(\d+)\]: ([A-F0-9]+): ') + + # 13. postfix/bounce: POSTFIX-MSG-ID: "sender non-delivery notification" + + # key=postfix_msg_id, value=reference to a record in `recs` + self.index_postfix_msg_id = {} + + # deferred delivery settings indexed by delivery service tid + # key=service_tid value={ settings:{} } + self.dds_by_tid = {} + + + + def set_capture_enabled(self, capture_enabled): + ''' thread-safe ''' + self.capture_enabled = capture_enabled + + def update_drop_disposition(self, drop_disposition): + ''' thread-safe ''' + with self.drop_disposition_lock: + self.drop_disposition.update(drop_disposition) + + def get_inprogress_count(self): + ''' thread-safe ''' + return self.current_inprogress_recs + + + def datetime_as_str(self, d): + # iso-8601 time friendly to sqlite3 + timestamp = d.isoformat(sep=' ', timespec='seconds') + # strip the utc offset from the iso format (ie. remove "+00:00") + idx = timestamp.find('+00:00') + if idx>0: + timestamp = timestamp[0:idx] + return timestamp + + def parse_date(self, str): + # we're expecting UTC times from date_parser() + d = self.date_parser_fn(str) + return self.datetime_as_str(d) + + def get_cached_state(self, clear=True): + conn = None + try: + # obtain the cached records from the record store + conn = self.record_store.connect() + recs = self.record_store.read_rec(conn, 'state', { + "owner_id": STATE_CACHE_OWNER_ID, + "clear": clear + }) + log.info('read %s incomplete records from cache', len(recs)) + + # eliminate stale records - "stale" should be longer than + # the "give-up" time for postfix (4-5 days) + stale = datetime.timedelta(days=7) + cutoff = self.datetime_as_str( + datetime.datetime.now(datetime.timezone.utc) - stale + ) + newlist = [ rec for rec in recs if rec['connect_time'] >= cutoff ] + if len(newlist) < len(recs): + log.warning('dropping %s stale incomplete records', + len(recs) - len(newlist)) + return newlist + finally: + if conn: self.record_store.close(conn) + + def save_state(self): + log.info('saving state to cache: %s records', len(self.recs)) + conn = None + try: + conn = self.record_store.connect() + self.record_store.write_rec(conn, 'state', { + 'owner_id': STATE_CACHE_OWNER_ID, + 'state': self.recs + }) + finally: + if conn: self.record_store.close(conn) + + + def add_new_connection(self, mta_conn): + ''' queue a mta_connection record ''' + threshold = self.max_inprogress_recs + ( len(self.recs) * 0.05 ) + if len(self.recs) > threshold: + backoff = len(self.recs) - self.max_inprogress_recs + int( self.max_inprogress_recs * 0.10 ) + log.warning('dropping %s old records', backoff) + self.recs = self.recs[min(len(self.recs),backoff):] + + self.recs.append(mta_conn) + return mta_conn + + def remove_connection(self, mta_conn): + ''' remove a mta_connection record from queue ''' + for mta_accept in mta_conn.get('mta_accept', []): + # remove reference in index_postfix_msg_id + postfix_msg_id = mta_accept.get('postfix_msg_id') + safe_del(self.index_postfix_msg_id, postfix_msg_id) + + # remove collected pre-delivery data + log.debug(mta_accept.get('_delete_dds_tids')) + for service_tid in mta_accept.get('_delete_dds_tids', {}): + safe_del(self.dds_by_tid, service_tid) + + self.recs.remove(mta_conn) + + def failure_category(self, reason, default_value): + if not reason: + return default_value + if "Greylisted" in reason: + return 'greylisted' + if 'SPF' in reason: + return 'spf' + if 'Sender address' in reason: + return 'sender_rejected' + if 'Relay' in reason: + return 'relay' + if 'spamhaus' in reason: + return 'spamhaus' + if 'Service unavailable' in reason: + return 'service_unavailable' + log.debug('no failure_category for: "%s"', reason) + return default_value + + + def append_mta_accept(self, mta_conn, vals=None): + ''' associate a mta_accept record with a connection ''' + mta_accept = vals or {} + safe_append(mta_conn, 'mta_accept', mta_accept) + return mta_accept + + + def find_by(self, mta_conn_q, mta_accept_q, auto_add=False, debug=False): + '''find records using field-matching queries + + If mta_accept_q is None: return a list of mta_conn matching query + `mta_conn_q` + + If `mta_accept_q` is given: return a list of + mta_conn,mta_accept pairs matching queries `mta_conn_q` and + `mta_accept_q` respectively. + + If there is no mta_accept record matching `mta_accept_q` and + `mta_accept_q['autoset']` is true and there is an mta_accept record + having no key as specified in mta_accept_q['key'], then + automatically add one and return that mta_accept record. + + If `auto_add` is true, and there is no matching mta_accept + record matching `mta_accept_q`, and no acceptable record + exists in the case where `autoset` is true, then add a new + mta_accept record with the key/value pair of `mta_accept_q` to + the most recent mta_conn. + + ''' + + if debug: + log.debug('mta_accept_q: %s', mta_accept_q) + + # find all candidate recs with matching mta_conn_q, ordered by most + # recent last + candidates = DictQuery.find(self.recs, mta_conn_q, reverse=False) + if len(candidates)==0: + if debug: log.debug('no candidates') + return [] + + elif not candidates[0]['exact']: + # there were no exact matches. apply autosets to the best + # match requiring the fewest number of autosets (index 0) + if debug: log.debug('no exact candidates') + DictQuery.autoset(candidates[0]) + candidates[0]['exact'] = True + candidates = [ candidates[0] ] + + else: + # there is at least one exact match - remove all non-exact + # candidates + candidates = [ + candidate for candidate in candidates if candidate['exact'] + ] + + if mta_accept_q is None: + return [ (candidate['item'], None) for candidate in candidates ] + elif type(mta_accept_q) is not list: + mta_accept_q = [ mta_accept_q ] + + count_of_nonoptional_accept_q_items = 0 + for q in mta_accept_q: + if not q.get('autoset') and not q.get('optional'): + count_of_nonoptional_accept_q_items += 1 + + if debug: + log.debug('candidates: qty=%s', len(candidates)) + + # for each candidate, issue the mta_accept_q query and assign + # a rank to each candidate. keep track of the best `mta_accept` + # record match + idx = 0 + while idx0: + if accept_candidates[0]['exact']: + # exact match + if debug: + log.debug('exact match accept=%s',accept_candidates[0]) + candidate['best_accept'] = accept_candidates[0] + candidate['best_rank'] = '00000.{0:08d}'.format(idx) + + else: + candidate['best_accept'] = accept_candidates[0] + candidate['best_rank'] = '{0:05d}.{1:08d}'.format( + len(accept_candidates[0]['autoset_list']), + idx + ) + elif 'mta_accept' not in candidate['item'] and \ + count_of_nonoptional_accept_q_items>0: + # for auto-add: when no candidate has any mta_accept + # matches, prefer a candidate that failed to meet the + # query requirements because there was nothing to + # query, versus adding another mta_accept record to an + # existing list that didn't meet the query + # requirements + candidate['best_rank'] = '99998.{0:08d}'.format(idx) + else: + candidate['best_rank'] = '99999.{0:08d}'.format(idx) + + if debug: + log.debug('candidate %s result: %s', idx, candidate) + + idx+=1 + + # sort the candidates by least # autoset's required + candidates.sort(key=lambda x: x['best_rank']) + + if 'best_accept' in candidates[0]: + # at least one match was successful. if the best candidate + # wasn't exact, apply autosets + if not candidates[0]['best_accept']['exact']: + DictQuery.autoset(candidates[0]['best_accept']) + if debug: + log.debug('best match (auto-set) accept=%s',candidates[0]['best_accept']['item']) + return [ (candidates[0]['item'], candidates[0]['best_accept']['item']) ] + # otherwise return all exact matches + else: + rtn = [] + for candidate in candidates: + if not 'best_accept' in candidate: + break + best_accept = candidate['best_accept'] + if not best_accept['exact']: + break + rtn.append( (candidate['item'], best_accept['item']) ) + if debug: + log.debug('best matches (exact)=%s', rtn) + return rtn + + # if autoset is not possible, and there are no exact matches, + # add a new accept record to the highest ranked candidate + if auto_add: + if debug: log.debug('auto-add new accept') + mta_conn = candidates[0]['item'] + v = {} + for q in mta_accept_q: + v[q['key']] = q['value'] + mta_accept = self.append_mta_accept(mta_conn, v) + return [ (mta_conn, mta_accept) ] + + if debug: log.debug("no matches") + return [] + + + def find_first(self, *args, **kwargs): + '''find the "best" result and return it - find_by() returns the list + ordered, with the first being the "best" + + ''' + r = self.find_by(*args, **kwargs) + if len(r)==0: + return (None, None) + return r[0] + + def index_by_postfix_msg_id(self, mta_conn, mta_accept): + postfix_msg_id = mta_accept['postfix_msg_id'] + assert postfix_msg_id not in self.index_postfix_msg_id + self.index_postfix_msg_id[postfix_msg_id] = { + 'mta_conn': mta_conn, + 'mta_accept': mta_accept + } + + def find_by_postfix_msg_id(self, postfix_msg_id): + '''postfix message id's are unique and we maintain a separate index + for them to give constant-time lookups since many log entries report + this id + + ''' + if postfix_msg_id in self.index_postfix_msg_id: + cache_val = self.index_postfix_msg_id[postfix_msg_id] + return cache_val['mta_conn'], cache_val['mta_accept'] + else: + msg = 'postfix_msg_id "%s" not cached' % postfix_msg_id + # if log.isEnabledFor(logging.DEBUG): + # msg = [ msg + '\n' ] + traceback.format_stack() + # log.warning("".join(msg)) + # else: + log.warning(msg) + return None, None + + def defer_delivery_settings_by_tid(self, service_tid, settings): + '''select messages from smtp outbound delivery come in before knowing + when recipient(s) they apply to. for instance, TLS details + when making a connection to a remote relay. this function will + defer those settings until find_delivery() can assign them to + the recipient(s) involved by that smtp process + + ''' + if service_tid not in self.dds_by_tid: + self.dds_by_tid[service_tid] = { + "settings": settings + } + else: + self.dds_by_tid[service_tid]['settings'].update(settings) + + def defer_delivery_settings_by_rcpt(self, mta_accept, rcpt_to, subsystem, settings): + '''some recipient-oriented log entries, in particular from + SpamAssassin and Postgrey, don't have enough information to + unambiguously select a connection it applies to, and therefore + can match multiple active connections. Use this function to + defer the matchup until the ambiguity can be + resoloved. find_delivery() will do this. + + ''' + if '_dds' not in mta_accept: + mta_accept['_dds'] = {} + dds = mta_accept['_dds'] + if rcpt_to not in dds: + dds[rcpt_to] = { 'subsystems':[], 'settings':{} } + dds[rcpt_to]['settings'].update(settings) + dds[rcpt_to]['subsystems'].append(subsystem) + + def find_delivery(self, mta_accept, rcpt_to, service_tid=None, auto_add=False): + if 'mta_delivery' not in mta_accept: + if auto_add: + mta_accept['mta_delivery'] = [] + else: + return None + + rcpt_to = rcpt_to.lower() + for delivery in mta_accept['mta_delivery']: + if 'rcpt_to' in delivery and rcpt_to == delivery['rcpt_to'].lower(): + return delivery + + if auto_add: + delivery = { + 'rcpt_to': rcpt_to + } + if '_dds' in mta_accept and rcpt_to in mta_accept['_dds']: + dds = mta_accept['_dds'][rcpt_to] + for subsystem in dds['subsystems']: + self.add_subsystem(mta_accept, subsystem) + delivery.update(dds['settings']) + del mta_accept['_dds'][rcpt_to] + + if service_tid is not None and service_tid in self.dds_by_tid: + dds = self.dds_by_tid[service_tid] + delivery.update(dds['settings']) + # save the tid for cleanup during remove_connection() + if '_delete_dds_tids' not in mta_accept: + mta_accept['_delete_dds_tids'] = { } + mta_accept['_delete_dds_tids'][service_tid] = True + + mta_accept['mta_delivery'].append(delivery) + return delivery + + + def add_subsystem(self, mta_accept, subsystem): + if 'subsystems' not in mta_accept: + mta_accept['subsystems'] = subsystem + #elif subsystem not in mta_accept['subsystems']: + else: + mta_accept['subsystems'] += ','+subsystem + + + def match_connect(self, line): + # 1. 1a. postfix/smtpd[13698]: connect from host.tld[1.2.3.4] + m = self.re_connect_from.search(line) + if m: + mta_conn = { + "connect_time": self.parse_date(m.group(1)), # "YYYY-MM-DD HH:MM:SS" + "service": "smtpd" if m.group(2)=="smtpd" else "submission", + "service_tid": m.group(3), + "remote_host": m.group(4), + "remote_ip": m.group(5) + } + self.add_new_connection(mta_conn) + return { 'mta_conn': mta_conn } + + def match_local_pickup(self, line): + # 1b. Dec 6 07:01:39 mail postfix/pickup[7853]: A684B1F787: uid=0 from= + # 1=date + # 2=service ("pickup") + # 3=service_tid + # 4=postfix_msg_id + m = self.re_local_pickup.search(line) + if m: + mta_conn = { + "connect_time": self.parse_date(m.group(1)), + "disconnect_time": self.parse_date(m.group(1)), + "service": m.group(2), + "service_tid": m.group(3), + "remote_host": "localhost", + "remote_ip": "127.0.0.1" + } + mta_accept = { + "postfix_msg_id": m.group(4), + } + self.append_mta_accept(mta_conn, mta_accept) + self.add_new_connection(mta_conn) + self.index_by_postfix_msg_id(mta_conn, mta_accept) + return { 'mta_conn': mta_conn, 'mta_accept': mta_accept } + + + def match_policyd_spf(self, line): + v = None + client_ip = None + envelope_from = None + # 2. policyd-spf[13703]: prepend Received-SPF: Pass (mailfrom) identity=mailfrom; client-ip=1.2.3.4; helo=host.tld; envelope-from=alice@post.com; receiver= + m = self.re_spf_1.search(line) + if m: + v = { + "spf_tid": m.group(1), + "spf_result": m.group(2), + "spf_reason": PostfixLogParser.strip_brackets( + m.group(3), + bracket_l='(', + bracket_r=')' + ) + } + pairs = line[m.end():].split(';') + for str in pairs: + pair = str.strip().split('=', 1) + if len(pair)==2: + if pair[0]=='client-ip': + client_ip=pair[1] + elif pair[0]=='envelope-from': + envelope_from = pair[1] + + else: + # 2a. policyd-spf[26231]: 550 5.7.23 Message rejected due to: Receiver policy for SPF Softfail. Please see http://www.openspf.net/Why?s=mfrom;id=test@google.com;ip=1.2.3.4;r= + m = self.re_spf_2.search(line) + if m: + v = { + "spf_tid": m.group(1), + "spf_result": "reject", + "spf_reason": m.group(2), + } + pairs = line[m.end():].split(';') + for str in pairs: + # note: 'id' and 'ip' are part of the url, but we need + # to match records somehow... + pair = str.strip().split('=', 1) + if len(pair)==2: + if pair[0]=='ip': + client_ip = pair[1] + elif pair[0]=='id': + envelope_from = pair[1] + + if v: + mta_conn_q = [{ + 'key':'remote_ip', 'value':client_ip, + 'ignorecase': True + }] + mta_accept_q = [ + { 'key': 'envelope_from', 'value': envelope_from, + 'ignorecase': True, 'autoset': True }, + { 'key': 'spf_result', 'value': None } + ] + mta_conn, mta_accept = self.find_first( + mta_conn_q, + mta_accept_q, + auto_add=True + ) + if mta_accept: + mta_accept.update(v) + self.add_subsystem(mta_accept, "policyd-spf") + return { 'mta_conn':mta_conn, 'mta_accept':mta_accept } + return True + + + def match_postgrey(self, line): + # Dec 9 14:46:57 mail postgrey[879]: action=greylist, reason=new, client_name=host.tld, client_address=1.2.3.4/32, sender=alice@post.com, recipient=mia@myhost.com + # 3. postgrey: "client_address=1.2.3.4/32, sender=alice@post.com" [action="pass|greylist", reason="client whitelist|triplet found|new", delay=x-seconds] + # 3a. Dec 6 18:31:28 mail postgrey[879]: 0E98D1F787: action=pass, reason=triplet found, client_name=host.tld, client_address=1.2.3.4/32, sender=alice@post.com, recipient=mia@myhost.com + # 1=postgrey_tid + # 2=postfix_msg_id (re-1 only) + + v = None + client_ip = None + envelope_from = None + rcpt_to = None + postfix_msg_id = None + m = self.re_postgrey_1.search(line) + if not m: m = self.re_postgrey_2.search(line) + + if m: + v = { + 'postgrey_tid': m.group(1) + } + try: + postfix_msg_id = m.group(2) + except IndexError: + pass + + pairs = line[m.end():].split(',') + for str in pairs: + pair = str.strip().split('=', 1) + if len(pair)==2: + if pair[0]=='action': + v['postgrey_result'] = pair[1] + elif pair[0]=='reason': + v['postgrey_reason'] = pair[1] + elif pair[0]=='sender': + envelope_from = pair[1] + elif pair[0]=='recipient': + rcpt_to = pair[1] + elif pair[0]=='client_address': + client_ip=pair[1] + # normalize the ipv6 address + # postfix: 2607:f8b0:4864:20::32b + # postgrey: 2607:F8B0:4864:20:0:0:0:32B/128 + idx = client_ip.find('/') + if idx>=0: + client_ip = client_ip[0:idx] + if ':' in client_ip: + addr = ipaddress.ip_address(client_ip) + client_ip = addr.__str__() + + elif pair[0]=='delay': + v['postgrey_delay'] = pair[1] + + if v: + matches = [] + if postfix_msg_id: + mta_conn, mta_accept = self.find_by_postfix_msg_id( + postfix_msg_id + ) + if mta_accept: + matches.append( (mta_conn, mta_accept) ) + + if len(matches)==0: + mta_conn_q = [{ + 'key':'remote_ip', 'value':client_ip, + 'ignorecase': True + }] + mta_accept_q = [ + { 'key': 'envelope_from', 'value': envelope_from, + 'ignorecase': True, 'autoset': True } + ] + + matches = self.find_by(mta_conn_q, mta_accept_q, auto_add=True) + + if len(matches)>0: + log.debug('MATCHES(%s): %s', len(matches), matches) + auto_add = ( len(matches)==1 ) + for mta_conn, mta_accept in matches: + mta_delivery = self.find_delivery( + mta_accept, + rcpt_to, + auto_add=auto_add + ) + if mta_delivery: + mta_delivery.update(v) + log.debug('DELIVERY(postgrey): %s', mta_accept) + self.add_subsystem(mta_accept, "postgrey") + return { + 'mta_conn': mta_conn, + 'mta_accept': mta_accept, + 'mta_delivery': mta_delivery + } + + # ambiguous: two or more active connections sending a + # message with the exact same FROM address! defer + # matching until another find_delivery(auto_add=True) + # is called elsewhere (lmtp) + for mta_conn, mta_accept in matches: + self.defer_delivery_settings_by_rcpt( + mta_accept, + rcpt_to, + 'postgrey', + v + ) + return { 'deferred': True } + + return True + + + def match_postfix_msg_id(self, line): + m = self.re_postfix_noqueue.search(line) + if m: + # 4b. postfix/submission/smtpd[THREAD-ID]: NOQUEUE: reject: RCPT from unknown[1.2.3.4]: 550 5.7.23 : Recipient address rejected: Message rejected due to: Receiver policy for SPF Softfail. Please see http://www.openspf.net/Why?s=mfrom;id=test@google.com;ip=1.2.3.4;r=; from= to= proto=ESMTP helo= + # 4c. postfix/smtpd[THREAD-ID]: "NOQUEUE: reject: ....: Recipient address rejected: Greylisted, seehttp://postgrey.../; ..." + # 1=service ("submission/smtpd" or "smtpd") + # 2=service_tid + # 3=accept_status ("reject") + service_tid = m.group(2) + reason = line[m.end():].rstrip() + envelope_from = None + idx = reason.find('; ') # note the space + if idx>=0: + for pair in PostfixLogParser.SplitList(reason[idx+2:], delim=' '): + if pair['name'] == 'from': + envelope_from = pair['value'] + break + reason=reason[0:idx] + + v = { + 'postfix_msg_id': 'NOQUEUE', + 'accept_status': m.group(3), + 'failure_info': reason, + 'failure_category':self.failure_category(reason,'postfix_other') + } + + mta_conn_q = [{ 'key':'service_tid', 'value':service_tid }] + mta_accept_q = [ + { 'key': 'envelope_from', 'value': envelope_from, + 'ignorecase': True, 'autoset': True }, + { 'key': 'postfix_msg_id', 'value': None } + ] + mta_conn, mta_accept = self.find_first( + mta_conn_q, + mta_accept_q, + auto_add=True + ) + if mta_accept: + mta_accept.update(v) + return { 'mta_conn': mta_conn, 'mta_accept': mta_accept } + return True + + m = self.re_postfix_msg_id.search(line) + if m: + # 4. postfix/smtpd[THREAD-ID]: "POSTFIX-MSG-ID" (eg: "DD95A1F796"): client=DNS[IP] + # 4a. postfix/submission/smtpd[THREAD-ID]: POSTFIX-MSG-ID: client=DNS[IP], sasl_method=LOGIN, sasl_username=mia@myhost.com + # 1=service ("submission/smtpd" or "smtpd") + # 2=service_tid + # 3=postfix_msg_id + service_tid = m.group(2) + postfix_msg_id = m.group(3) + remote_host = None + remote_ip = None + v = { + } + for pair in PostfixLogParser.SplitList(line[m.end():]): + if pair['name']=='sasl_method': + v['sasl_method'] = pair['value'].strip() + elif pair['name']=='sasl_username': + v['sasl_username'] = pair['value'].strip() + elif pair['name']=='client': + remote_host, remote_ip = \ + PostfixLogParser.split_host(pair['value']) + + mta_conn_q = [ + { 'key': 'service_tid', 'value': service_tid }, + { 'key': 'remote_ip', 'value': remote_ip } + ] + mta_accept_q = { 'key': 'postfix_msg_id', 'value': None } + mta_conn, mta_accept = self.find_first( + mta_conn_q, + mta_accept_q, + auto_add=True, + debug=False + ) + + if mta_accept: + mta_conn.update(v) + mta_accept.update({ + 'postfix_msg_id': postfix_msg_id + }) + self.index_by_postfix_msg_id( + mta_conn, + mta_accept + ) + return { 'mta_conn':mta_conn, 'mta_accept':mta_accept } + return True + + + def match_message_id(self, line): + # 5. Dec 10 06:48:48 mail postfix/cleanup[7435]: 031AF20076: message-id=<20201210114848.031AF20076@myhost.com> + m = self.re_postfix_message_id.search(line) + if m: + postfix_msg_id = m.group(1) + v = { + 'message_id': m.group(2) + } + mta_conn, mta_accept = self.find_by_postfix_msg_id(postfix_msg_id) + # auto-add ? + if mta_accept: + mta_accept.update(v) + return { 'mta_conn':mta_conn, 'mta_accept':mta_accept } + return True + + + + def match_opendkim(self, line): + # 1=postfix_msg_id + # 2=verification detail + # 3=error-msg + m = self.re_opendkim_ssl.search(line) + if m: + postfix_msg_id = m.group(1) + err = m.group(3).strip() + if err != '': + v = { + 'dkim_result': 'error', + 'dkim_reason': err + } + else: + v = { + 'dkim_result': 'pass', + 'dkim_reason': m.group(2).strip() + } + mta_conn, mta_accept = self.find_by_postfix_msg_id(postfix_msg_id) + if mta_accept: + mta_accept.update(v) + self.add_subsystem(mta_accept, "dkim") + return { 'mta_conn': mta_conn, 'mta_accept': mta_accept } + return True + + # 1=postfix_msg_id + # 2=error-msg + m = self.re_opendkim_error.search(line) + if m: + postfix_msg_id = m.group(1) + err = m.group(2).strip() + v = { + 'dkim_result': 'error', + 'dkim_reason': err + } + mta_conn, mta_accept = self.find_by_postfix_msg_id(postfix_msg_id) + if mta_accept: + mta_accept.update(v) + self.add_subsystem(mta_accept, "dkim") + return { 'mta_conn': mta_conn, 'mta_accept': mta_accept } + return True + + + def match_opendmarc(self, line): + # 1=postfix_msg_id + # 2=domain + # 3="pass","none","fail" + m = self.re_opendmarc_result.search(line) + if m: + postfix_msg_id = m.group(1) + v = { + 'dmarc_result': m.group(3), + 'dmarc_reason': m.group(2) + } + mta_conn, mta_accept = self.find_by_postfix_msg_id(postfix_msg_id) + if mta_accept: + mta_accept.update(v) + self.add_subsystem(mta_accept, "dmarc") + return { 'mta_conn': mta_conn, 'mta_accept': mta_accept } + return True + + + + def match_postfix_queue_removed(self, line): + # 13. postfix/qmgr: POSTFIX-MSG-ID: "removed" + # 1=date + # 2=postfix_msg_id + m = self.re_queue_removed.search(line) + if m: + postfix_msg_id = m.group(2) + v = { + 'queue_remove_time': self.parse_date(m.group(1)) + } + mta_conn, mta_accept = self.find_by_postfix_msg_id(postfix_msg_id) + if mta_accept: + mta_accept.update(v) + return { 'mta_conn': mta_conn, 'mta_accept': mta_accept } + return True + + + def match_postfix_queue_added(self, line): + # 8. postfix/qmgr: POSTFIX-MSG-ID: from=user@tld, size=N, nrcpt=1 (queue active) + # 1=date + # 2=postfix_msg_id + m = self.re_queue_added.search(line) + if m: + postfix_msg_id = m.group(2) + v = { + 'queue_time': self.parse_date(m.group(1)), + 'accept_status': 'queued', + } + envelope_from = None + for pair in PostfixLogParser.SplitList(line[m.end():]): + if pair['name']=='size': + v['message_size'] = safe_int(pair['value']) + elif pair['name']=='nrcpt': + v['message_nrcpt'] = safe_int(pair['value']) + elif pair['name']=='from': + envelope_from = pair['value'] + + mta_conn, mta_accept = self.find_by_postfix_msg_id(postfix_msg_id) + if mta_accept: + if envelope_from and 'envelope_from' not in mta_accept: + v['envelope_from'] = envelope_from + mta_accept.update(v) + return { 'mta_conn': mta_conn, 'mta_accept': mta_accept } + return True + + + def match_spampd(self, line): + # 11. spampd: "clean message |(unknown) (SCORE/MAX-SCORE) from for in 1.51s, N bytes" + # 11(a) spampd: "identified spam |(unknown) (5.12/5.00) from for in 1.51s, N bytes" + # 1=spam_tid + # 2="clean message" | "identified spam" (other?) + # 3=message_id ("(unknown)" if message id is "<>") + # 4=spam_score + # 5=envelope_from + # 6=rcpt_to + m = self.re_spampd.search(line) + if m: + message_id = m.group(3) + from_email = m.group(5) + rcpt_to = m.group(6) + spam_result = m.group(2) + + if message_id == '(unknown)': + message_id = '<>' + + if spam_result == 'clean message': + spam_result = 'clean' + elif spam_result == 'identified spam': + spam_result = 'spam' + + v = { + 'spam_tid': m.group(1), + 'spam_result': spam_result, + 'spam_score': m.group(4) + } + + mta_accept_q = [ + { 'key':'message_id', 'value':message_id }, + { 'key':'envelope_from', 'value':from_email, + 'ignorecase': True }, + ] + matches = self.find_by('*', mta_accept_q, debug=False) + + if len(matches)==0 and message_id=='<>': + # not sure why this happens - the message has a valid + # message-id reported by postfix, but spampd doesn't + # see it + mta_accept_q = [ + { 'key':'envelope_from', 'value':from_email, + 'ignorecase': True }, + ] + matches = self.find_by('*', mta_accept_q, debug=False) + if len(matches)>1: + # ambiguous - can't match it + matches = [] + + + if len(matches)>0: + #log.debug('MATCHES(%s): %s', len(matches), matches) + auto_add = ( len(matches)==1 ) + for mta_conn, mta_accept in matches: + mta_delivery = self.find_delivery( + mta_accept, + rcpt_to, + auto_add=auto_add + ) + if mta_delivery: + mta_delivery.update(v) + log.debug('DELIVERY(spam): %s', mta_accept) + self.add_subsystem(mta_accept, "spam") + return { + 'mta_conn': mta_conn, + 'mta_accept': mta_accept, + 'mta_delivery': mta_delivery + } + + # ambiguous: two or more active connections sending a + # message with the exact same message-id from the + # exact same FROM address! defer matching until + # another find_delivery(auto_add=True) is called + # elsewhere (lmtp) + for mta_conn, mta_accept in matches: + self.defer_delivery_settings_by_rcpt( + mta_accept, + rcpt_to, + 'spam', + v + ) + return { 'deferred': True } + + return True + + + + def match_disconnect(self, line): + # disconnect from unknown[1.2.3.4] ehlo=1 auth=0/1 quit=1 commands=2/3 + # 1=date + # 2=service ("submission/smptd" or "smtpd") + # 3=service_tid + # 4=remote_host + # 5=remote_ip + m = self.re_disconnect.search(line) + if m: + service_tid = m.group(3) + remote_ip = m.group(5) + v = { + 'disconnect_time': self.parse_date(m.group(1)), + 'remote_auth_success': 0, + 'remote_auth_attempts': 0, + 'remote_used_starttls': 0 + } + pairs = line[m.end():].split(' ') + for str in pairs: + pair = str.strip().split('=', 1) + if len(pair)==2: + if pair[0]=='auth': + idx = pair[1].find('/') + if idx>=0: + v['remote_auth_success'] = safe_int(pair[1][0:idx]) + v['remote_auth_attempts'] = safe_int(pair[1][idx+1:]) + else: + v['remote_auth_success'] = safe_int(pair[1]) + v['remote_auth_attempts'] = safe_int(pair[1]) + + elif pair[0]=='starttls': + v['remote_used_starttls'] = safe_int(pair[1]) + + + mta_conn_q = [ + { 'key': 'service_tid', 'value': service_tid }, + { 'key': 'remote_ip', 'value': remote_ip } + ] + mta_conn, mta_accept = self.find_first(mta_conn_q, None) + if mta_conn: + mta_conn.update(v) + return { 'mta_conn': mta_conn } + return True + + + def match_pre_delivery(self, line): + # postfix/smtp[18333]: Trusted TLS connection established to mx01.mail.icloud.com[17.57.154.23]:25: TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits) + # postfix/smtp[566]: Untrusted TLS connection established to host.tld[1.2.3.4]:25: TLSv1.2 with cipher AES128-GCM-SHA256 (128/128 bits) + # 1=service ("smtp") + # 2=service_tid + # 3=delivery_connection ("Trusted", "Untrusted", "Verified") + # 4=tls details + m = self.re_pre_delivery.search(line) + if m: + service_tid = m.group(2) + delivery_connection = m.group(3).lower() + self.defer_delivery_settings_by_tid(service_tid, { + 'delivery_connection': delivery_connection, + 'delivery_connection_info': m.group(3) + ' ' + m.group(4) + }) + return { "deferred": True } + + + def match_delivery(self, line): + # 12. postfix/lmtp: POSTFIX-MSG-ID: to=, relay=127.0.0.1[127.0.0.1]:10025, delay=4.7, delays=1/0.01/0.01/3.7, dsn=2.0.0, status=sent (250 2.0.0 YB5nM1eS01+lSgAAlWWVsw Saved) + # 12a. postfix/lmtp: POSTFIX_MSG-ID: to=user@tld, status=bounced (host...said...550 5.1.1 User doesn't exist ....) + # 12b. postfix/smtp[32052]: A493B1FAF1: to=, relay=mx.post.com[1.2.3.4]:25, delay=1.2, delays=0.65/0.06/0.4/0.09, dsn=2.0.0, status=sent (250 2.0.0 OK 7E/38-26906-CDC5DCF5): None + # 12c. postfix/smtp[21816]: BD1D31FB12: host mx2.comcast.net[2001:558:fe21:2a::6] refused to talk to me: 554 resimta-ch2-18v.sys.comcast.net resimta-ch2-18v.sys.comcast.net 2600:3c02::f03c:92ff:febb:192f found on one or more DNSBLs, see http://postmaster.comcast.net/smtp-error-codes.php#BL000001 + # 1=system ("lmtp" or "smtp") + # 2=system_tid + # 3=postfix_msg_id + m = self.re_delivery.search(line) + if m: + service = m.group(1) + service_tid = m.group(2) + postfix_msg_id = m.group(3) + mta_conn, mta_accept = self.find_by_postfix_msg_id(postfix_msg_id) + if not mta_conn: + return True + + if 'status=' not in line: + # temporary error: postfix will keep trying + # 12c + reason = line[m.end():].strip() + category = self.failure_category(reason, 'temporary_error') + if 'failure_info' in mta_accept: + mta_accept['failure_info'] += '\n' + reason + if mta_accept['failure_category'] != category: + mta_accept['failure_category'] = 'multiple' + else: + v = { + 'failure_info': reason, + 'failure_category': category + } + mta_accept.update(v) + return { 'mta_conn': mta_conn, 'mta_accept': mta_accept } + + + # 12, 12a, 12b + detail = PostfixLogParser.SplitList(line[m.end():]).asDict() + if 'to' not in detail: + return True + + mta_delivery = self.find_delivery( + mta_accept, + detail['to']['value'], + service_tid=service_tid, + auto_add=True + ) + mta_delivery['service'] = service + mta_delivery['service_tid'] = service_tid + log.debug('DELIVERY(accept): %s', mta_accept) + + if 'status' in detail: + result = detail['status']['value'].strip() + comment = detail['status'].get('comment') + mta_delivery['status'] = result + if result == 'sent': + safe_del(mta_delivery, 'failure_category') + + else: + mta_delivery['failure_category'] = \ + self.failure_category(comment, service + "_other") + + if comment: + if mta_delivery.get('delivery_info'): + mta_delivery['delivery_info'] += ("\n" + comment) + else: + mta_delivery['delivery_info'] = comment + + if 'delay' in detail: + mta_delivery['delay'] = detail['delay']['value'].strip() + + if 'relay' in detail: + mta_delivery['relay'] = detail['relay']['value'].strip() + + self.add_subsystem(mta_accept, service) + return { 'mta_conn':mta_conn, 'mta_accept':mta_accept, 'mta_delivery':mta_delivery } + + + def store(self, mta_conn): + def all_rejects(mta_accept): + if not mta_accept: return False + all = True + for accept in mta_accept: + if accept.get('accept_status') != 'reject': # or accept.get('failure_category') == 'greylisted': + all = False + break + return all + + if 'disposition' not in mta_conn: + if 'queue_time' not in mta_conn and \ + mta_conn.get('remote_auth_success') == 0 and \ + mta_conn.get('remote_auth_attempts', 0) > 0: + mta_conn.update({ + 'disposition': 'failed_login_attempt', + }) + + elif 'mta_accept' not in mta_conn and \ + mta_conn.get('remote_auth_success') == 0 and \ + mta_conn.get('remote_auth_attempts') == 0: + mta_conn.update({ + 'disposition': 'suspected_scanner', + }) + + elif all_rejects(mta_conn.get('mta_accept')): + mta_conn.update({ + 'disposition': 'reject' + }) + elif mta_conn.get('remote_used_starttls',0)==0 and \ + mta_conn.get('remote_ip') != '127.0.0.1': + mta_conn.update({ + 'disposition': 'insecure' + }) + else: + mta_conn.update({ + 'disposition': 'ok', + }) + + drop = False + with self.drop_disposition_lock: + drop = self.drop_disposition.get(mta_conn['disposition'], False) + + if not drop: + log.debug('store: %s', mta_conn) + try: + self.record_store.store('inbound_mail', mta_conn) + except Exception as e: + log.exception(e) + + self.remove_connection(mta_conn) + + + def log_match(self, match_str, match_result, line): + if match_result is True: + log.info('%s [unmatched]: %s', match_str, line) + + elif match_result: + if match_result.get('deferred', False): + log.debug('%s [deferred]: %s', match_str, line) + + elif 'mta_conn' in match_result: + log.debug('%s: %s: %s', match_str, line, match_result['mta_conn']) + else: + log.error('no mta_conn in match_result: ', match_result) + else: + log.debug('%s: %s', match_str, line) + + + def test_end_of_rec(self, match_result): + if not match_result or match_result is True or match_result.get('deferred', False): + return False + return self.end_of_rec(match_result['mta_conn']) + + def end_of_rec(self, mta_conn): + '''a client must be disconnected and all accepted messages removed + from queue for the record to be "complete" + + ''' + if 'disconnect_time' not in mta_conn: + return False + + nothing_queued = True + for mta_accept in mta_conn.get('mta_accept',[]): + if 'postfix_msg_id' in mta_accept and \ + mta_accept['postfix_msg_id'] != 'NOQUEUE' and \ + 'queue_remove_time' not in mta_accept: + nothing_queued = False + break + + return nothing_queued + + + def handle(self, line): + '''overrides ReadLineHandler method + + This function is called by the main log reading thread in + TailFile. All additional log reading is blocked until this + function completes. + + The storage engine (`record_store`, a SqliteEventStore + instance) does not block, so this function will return before + the record is saved to disk. + + IMPORTANT: + + The data structures and methods in this class are not thread + safe. It is not okay to call any of them when the instance is + registered with TailFile. + + ''' + if not self.capture_enabled: + return + + self.current_inprogress_recs = len(self.recs) + + log.debug('recs in progress: %s, dds_by_tid=%s', + len(self.recs), + len(self.dds_by_tid) + ) + match = self.match_connect(line) + if match: + self.log_match('connect', match, line) + return + + match = self.match_local_pickup(line) + if match: + self.log_match('local_pickup', match, line) + return + + match = self.match_policyd_spf(line) + if match: + self.log_match('policyd_spf', match, line) + return + + match = self.match_postgrey(line) + if match: + self.log_match('postgrey', match, line) + return + + match = self.match_postfix_msg_id(line) + if match: + self.log_match('postfix_msg_id', match, line) + return + + match = self.match_message_id(line) + if match: + self.log_match('message_id', match, line) + return + + match = self.match_opendkim(line) + if match: + self.log_match('opendkim', match, line) + return + + match = self.match_opendmarc(line) + if match: + self.log_match('opendmarc', match, line) + return + + match = self.match_postfix_queue_added(line) + if match: + self.log_match('queue added', match, line) + return + + match = self.match_spampd(line) + if match: + self.log_match('spam', match, line) + return + + match = self.match_disconnect(line) + if match: + self.log_match('disconnect', match, line) + if self.test_end_of_rec(match): + # we're done - not queued and disconnected ... save it + self.store(match['mta_conn']) + return + + match = self.match_pre_delivery(line) + if match: + self.log_match('pre-delivery', match, line) + return + + match = self.match_delivery(line) + if match: + self.log_match('delivery', match, line) + return + + match = self.match_postfix_queue_removed(line) + if match: + self.log_match('queue removed', match, line) + if self.test_end_of_rec(match): + # we're done - not queued and disconnected ... save it + self.store(match['mta_conn']) + return + + self.log_match('IGNORED', None, line) + + + def end_of_callbacks(self, thread): + '''overrides ReadLineHandler method + + save incomplete records so we can pick up where we left off + + ''' + self.save_state() + diff --git a/management/reporting/capture/mail/PostfixLogParser.py b/management/reporting/capture/mail/PostfixLogParser.py new file mode 100644 index 00000000..1b8ed699 --- /dev/null +++ b/management/reporting/capture/mail/PostfixLogParser.py @@ -0,0 +1,125 @@ + +class PostfixLogParser(object): + + @staticmethod + def split_host(str): + ''' split string in form HOST[IP] and return HOST and IP ''' + ip_start = str.find('[') + ip_end = -1 + if ip_start>=0: + ip_end = str.find(']', ip_start) + if ip_start<0 or ip_end<0: + return str, str + return str[0:ip_start], str[ip_start+1:ip_end] + + @staticmethod + def strip_brackets(str, bracket_l='<', bracket_r='>'): + # strip enclosing '<>' + if len(str)>=2 and str[0]==bracket_l and str[-1]==bracket_r: + return str[1:-1] + return str + + + class SplitList(object): + ''' split a postfix name=value list. For example: + + "delay=4.7, to=, status=sent (250 2.0.0 YB5nM1eS01+lSgAAlWWVsw Saved)" + + returns: { + "delay": { + "name": "delay", + "value": "4.7" + }, + "to": { + "name": "to", + "value": "alice@post.com" + }, + "status": { + "name": "status", + "value": "sent", + "comment": "250 2.0.0 YB5nM1eS01+lSgAAlWWVsw Saved" + } + } + + ''' + def __init__(self, str, delim=',', strip_brackets=True): + self.str = str + self.delim = delim + self.strip_brackets = True + self.pos = 0 + + def asDict(self): + d = {} + for pair in self: + d[pair['name']] = pair + return d + + def __iter__(self): + self.pos = 0 + return self + + def __next__(self): + if self.pos >= len(self.str): + raise StopIteration + + # name + eq = self.str.find('=', self.pos) + if eq<0: + self.pos = len(self.str) + raise StopIteration + + name = self.str[self.pos:eq].strip() + + # value and comment + self.pos = eq+1 + value = [] + comment = [] + + while self.pos < len(self.str): + c = self.str[self.pos] + self.pos += 1 + + if c=='<': + idx = self.str.find('>', self.pos) + if idx>=0: + value.append(self.str[self.pos-1:idx+1]) + self.pos = idx+1 + continue + + if c=='(': + # parens may be nested... + open_count = 1 + begin = self.pos + while self.pos < len(self.str) and open_count>0: + c = self.str[self.pos] + self.pos += 1 + if c=='(': + open_count += 1 + elif c==')': + open_count -= 1 + if open_count == 0: + comment.append(self.str[begin:self.pos-1]) + else: + comment.append(self.str[begin:len(self.str)]) + continue + + if c==self.delim: + break + + begin = self.pos-1 + while self.pos < len(self.str): + lookahead = self.str[self.pos] + if lookahead in [self.delim,'<','(']: + break + self.pos += 1 + + value.append(self.str[begin:self.pos]) + + if self.strip_brackets and len(value)==1: + value[0] = PostfixLogParser.strip_brackets(value[0]) + + return { + 'name': name, + 'value': ''.join(value), + 'comment': None if len(comment)==0 else '; '.join(comment) + } diff --git a/management/reporting/capture/mail/__init__.py b/management/reporting/capture/mail/__init__.py new file mode 100644 index 00000000..e69de29b diff --git a/management/reporting/capture/util/DictQuery.py b/management/reporting/capture/util/DictQuery.py new file mode 100644 index 00000000..ccbf1be5 --- /dev/null +++ b/management/reporting/capture/util/DictQuery.py @@ -0,0 +1,108 @@ + + +class DictQuery(object): + + @staticmethod + def find(data_list, q_list, return_first_exact=False, reverse=False): + '''find items in list `data_list` using the query specified in + `q_list` (a list of dicts). + + side-effects: + q_list is modified ('_val' is added) + + ''' + if data_list is None: + if return_first_exact: + return None + else: + return [] + + if type(q_list) is not list: + q_list = [ q_list ] + + # set _val to value.lower() if ignorecase is True + for q in q_list: + if q=='*': continue + ignorecase = q.get('ignorecase', False) + match_val = q['value'] + if ignorecase and match_val is not None: + match_val = match_val.lower() + q['_val'] = match_val + + # find all matches + matches = [] + direction = -1 if reverse else 1 + idx = len(data_list)-1 if reverse else 0 + while (reverse and idx>=0) or (not reverse and idx'+ + ''+ + ''+ + '' + , + + data: function() { + return { + stats: null, + stats_time: null, + row_counts: {} + }; + }, + + created: function() { + this.getStats(); + }, + + methods: { + getStats: function() { + axios.get('/reports/capture/db/stats') + .then(response => { + this.stats = response.data; + this.stats_time = Date.now(); + + // convert dates + var parser = d3.utcParse(this.stats.date_parse_format); + [ 'min', 'max' ].forEach( k => { + var d = parser(this.stats.mta_connect.connect_time[k]); + this.stats.mta_connect.connect_time[k] = d; + this.stats.mta_connect.connect_time[k+'_str'] = + d==null ? '-' : DateFormatter.dt_long(d); + }); + + // make a small bvTable of row counts + this.row_counts = { + items: [], + fields: [ 'name', 'count', 'percent' ], + field_types: [ + { type:'text/plain', label:'Disposition' }, + 'number/plain', + { type: 'number/percent', label:'Pct', places:1 }, + ], + }; + BvTable.setFieldDefinitions( + this.row_counts.fields, + this.row_counts.field_types + ); + this.row_counts.fields[0].formatter = (v, key, item) => { + return new ConnectionDisposition(v).short_desc + }; + this.row_counts.fields[0].tdClass = 'text-capitalize'; + + + const total = this.stats.mta_connect.count; + for (var name in this.stats.mta_connect.disposition) + { + const count = + this.stats.mta_connect.disposition[name].count; + this.row_counts.items.push({ + name: name, + count: count, + percent: count / total + }); + } + this.row_counts.items.sort((a,b) => { + return a.count > b.count ? -1 : + a.count < b.count ? 1 : 0; + }) + this.row_counts.items.push({ + name:'Total', + count:this.stats.mta_connect.count, + percent:1, + '_rowVariant': 'primary' + }); + + + }) + .catch(error => { + this.$root.handleError(error); + }); + }, + } +}); diff --git a/management/reporting/ui/chart-multi-line-timeseries.js b/management/reporting/ui/chart-multi-line-timeseries.js new file mode 100644 index 00000000..cb71eb36 --- /dev/null +++ b/management/reporting/ui/chart-multi-line-timeseries.js @@ -0,0 +1,206 @@ +Vue.component('chart-multi-line-timeseries', { + props: { + chart_data: { type:Object, required:false }, /* TimeseriesData */ + width: { type:Number, default: ChartPrefs.default_width }, + height: { type:Number, default: ChartPrefs.default_height }, + }, + + render: function(ce) { + return ChartVue.create_svg(ce, [0, 0, this.width, this.height]); + }, + + data: function() { + return { + tsdata: this.chart_data, + margin: { + top: ChartPrefs.axis_font_size, + bottom: ChartPrefs.axis_font_size * 2, + left: ChartPrefs.axis_font_size *3, + right: ChartPrefs.axis_font_size + }, + xscale: null, + yscale: null, + colors: ChartPrefs.line_colors + }; + }, + + watch: { + 'chart_data': function(newval) { + this.tsdata = newval; + this.draw(); + } + }, + + mounted: function() { + this.draw(); + }, + + methods: { + + draw: function() { + if (! this.tsdata) { + return; + } + + const svg = d3.select(this.$el); + svg.selectAll("g").remove(); + + if (this.tsdata.dates.length == 0) { + // no data ... + svg.append("g") + .append("text") + .attr("font-family", ChartPrefs.default_font_family) + .attr("font-size", ChartPrefs.label_font_size) + .attr("text-anchor", "middle") + .attr("x", this.width/2) + .attr("y", this.height/2) + .text("no data"); + } + + this.xscale = d3.scaleUtc() + .domain(d3.extent(this.tsdata.dates)) + .nice() + .range([this.margin.left, this.width - this.margin.right]) + + this.yscale = d3.scaleLinear() + .domain([ + d3.min(this.tsdata.series, d => d3.min(d.values)), + d3.max(this.tsdata.series, d => d3.max(d.values)) + ]) + .nice() + .range([this.height - this.margin.bottom, this.margin.top]) + + svg.append("g") + .call(this.xAxis.bind(this)) + .attr("font-size", ChartPrefs.axis_font_size); + + svg.append("g") + .call(this.yAxis.bind(this)) + .attr("font-size", ChartPrefs.axis_font_size); + + const line = d3.line() + .defined(d => !isNaN(d)) + .x((d, i) => this.xscale(this.tsdata.dates[i])) + .y(d => this.yscale(d)); + + const path = svg.append("g") + .attr("fill", "none") + .attr("stroke-width", 1.5) + .attr("stroke-linejoin", "round") + .attr("stroke-linecap", "round") + .selectAll("path") + .data(this.tsdata.series) + .join("path") + .style("mix-blend-mode", "multiply") + .style("stroke", (d, i) => this.colors[i]) + .attr("d", d => line(d.values)) + ; + + svg.call(this.hover.bind(this), path); + }, + + xAxis: function(g) { + var x = g.attr( + 'transform', + `translate(0, ${this.height - this.margin.bottom})` + ).call( + d3.axisBottom(this.xscale) + .ticks(this.width / 80) + .tickSizeOuter(0) + ); + return x; + }, + + yAxis: function(g) { + var y = g.attr( + "transform", + `translate(${this.margin.left},0)` + ).call( + d3.axisLeft(this.yscale) + .ticks(this.height/50) + ).call( + g => g.select(".domain").remove() + ).call( + g => ChartVue.add_yAxisLegend(g, this.tsdata, this.colors) + ); + return y; + }, + + hover: function(svg, path) { + if ("ontouchstart" in document) svg + .style("-webkit-tap-highlight-color", "transparent") + .on("touchmove", moved.bind(this)) + .on("touchstart", entered) + .on("touchend", left) + else svg + .on("mousemove", moved.bind(this)) + .on("mouseenter", entered) + .on("mouseleave", left); + + const dot = svg.append("g") + .attr("display", "none"); + + dot.append("circle") + .attr("r", 2.5); + + dot.append("text") + .attr("font-family", ChartPrefs.default_font_family) + .attr("font-size", ChartPrefs.default_font_size) + .attr("text-anchor", "middle") + .attr("y", -8); + + function moved(event) { + if (!event) event = d3.event; + event.preventDefault(); + var pointer; + if (d3.pointer) + pointer = d3.pointer(event, svg.node()); + else + pointer = d3.mouse(svg.node()); + const xvalue = this.xscale.invert(pointer[0]); // date + const yvalue = this.yscale.invert(pointer[1]); // number + //const i = d3.bisectCenter(this.tsdata.dates, xvalue); // index + const i = d3.bisect(this.tsdata.dates, xvalue); // index + if (i >= this.tsdata.dates.length) return; + // closest series + var closest = null; + for (var sidx=0; sidx= s.values.length) { + dot.attr("display", "none"); + return; + } + else { + dot.attr("display", null); + path.attr("stroke", d => d === s ? null : "#ddd") + .filter(d => d === s).raise(); + dot.attr( + "transform", + `translate(${this.xscale(this.tsdata.dates[i])},${this.yscale(s.values[i])})` + ); + dot.select("text").text(`${this.tsdata.formatDateTimeShort(this.tsdata.dates[i])} (${NumberFormatter.format(s.values[i])})`); + } + } + + function entered() { + path.style("mix-blend-mode", null).attr("stroke", "#ddd"); + dot.attr("display", null); + } + + function left() { + path.style("mix-blend-mode", "multiply").attr("stroke", null); + dot.attr("display", "none"); + } + } + } +}); + + diff --git a/management/reporting/ui/chart-pie.js b/management/reporting/ui/chart-pie.js new file mode 100644 index 00000000..ae771e42 --- /dev/null +++ b/management/reporting/ui/chart-pie.js @@ -0,0 +1,165 @@ +Vue.component('chart-pie', { + /* + * chart_data: [ + * { name: 'name', value: value }, + * ... + * ] + * + * if prop `labels` is false, a legend is shown instead of + * labeling each pie slice + */ + props: { + chart_data: Array, + formatter: { type: Function, default: NumberFormatter.format }, + name_formatter: Function, + labels: { type:Boolean, default: true }, + width: { type:Number, default: ChartPrefs.default_width }, + height: { type:Number, default: ChartPrefs.default_height }, + }, + + render: function(ce) { + var svg = ChartVue.create_svg(ce, [ + -this.width/2, -this.height/2, this.width, this.height + ]); + if (this.labels) { + return svg; + } + + /* +
+
+ {{d.value_str}} {{d.name}} +
+
+ */ + var legend_children = []; + this.legend.forEach(d => { + var span = ce('span', { attrs: { + 'class': 'd-inline-block text-right pr-1 mr-1 rounded', + 'style': `width:5em; background-color:${d.color}` + }}, this.formatter(d.value)); + + legend_children.push(ce('div', [ span, d.name ])); + }); + + var div_legend = ce('div', { attrs: { + 'class': 'ml-1 mt-2' + }}, legend_children); + + return ce('div', { attrs: { + 'class': "d-flex align-items-start" + }}, [ svg, div_legend ]); + + }, + + computed: { + legend: function() { + if (this.labels) { + return null; + } + + var legend = []; + if (this.chart_data) { + this.chart_data.forEach((d,i) => { + legend.push({ + name: this.name_formatter ? + this.name_formatter(d.name) : d.name, + value: d.value, + color: this.colors[i % this.colors.length] + }); + }); + } + legend.sort((a,b) => { + return a.value > b.value ? -11 : + a.value < b.value ? 1 : 0; + }); + return legend; + } + }, + + data: function() { + return { + chdata: this.chart_data, + colors: this.colors || ChartPrefs.colors, + }; + }, + + watch: { + 'chart_data': function(newval) { + this.chdata = newval; + this.draw(); + } + }, + + mounted: function() { + this.draw(); + }, + + methods: { + + draw: function() { + if (! this.chdata) return; + + var svg = d3.select(this.$el); + if (! this.labels) svg = svg.select('svg'); + svg.selectAll("g").remove(); + + var chdata = this.chdata; + var nodata = false; + if (d3.sum(this.chdata, d => d.value) == 0) { + // no data + chdata = [{ name:'no data', value:100 }] + nodata = true; + } + + const pie = d3.pie().sort(null).value(d => d.value); + const arcs = pie(chdata); + const arc = d3.arc() + .innerRadius(0) + .outerRadius(Math.min(this.width, this.height) / 2 - 1); + + var radius = Math.min(this.width, this.height) / 2; + if (chdata.length == 1) + radius *= 0.1; + else if (chdata.length <= 3) + radius *= 0.65; + else if (chdata.length <= 6) + radius *= 0.7; + else + radius *= 0.8; + arcLabel = d3.arc().innerRadius(radius).outerRadius(radius); + + svg.append("g") + .attr("stroke", "white") + .selectAll("path") + .data(arcs) + .join("path") + .attr("fill", (d,i) => this.colors[i % this.colors.length]) + .attr("d", arc) + .append("title") + .text(d => `${d.data.name}: ${this.formatter(d.data.value)}`); + + if (this.labels) { + svg.append("g") + .attr("font-family", ChartPrefs.default_font_family) + .attr("font-size", ChartPrefs.label_font_size) + .attr("text-anchor", "middle") + .selectAll("text") + .data(arcs) + .join("text") + .attr("transform", d => `translate(${arcLabel.centroid(d)})`) + .call(text => text.append("tspan") + .attr("y", "-0.4em") + .attr("font-weight", "bold") + .text(d => d.data.name)) + .call(text => text.filter(d => (d.endAngle - d.startAngle) > 0.25).append("tspan") + .attr("x", 0) + .attr("y", "0.7em") + .attr("fill-opacity", 0.7) + .text(d => nodata ? null : this.formatter(d.data.value))); + } + } + + }, + +}); diff --git a/management/reporting/ui/chart-stacked-bar-timeseries.js b/management/reporting/ui/chart-stacked-bar-timeseries.js new file mode 100644 index 00000000..ea78d4a8 --- /dev/null +++ b/management/reporting/ui/chart-stacked-bar-timeseries.js @@ -0,0 +1,225 @@ +/* + stacked bar chart +*/ + +Vue.component('chart-stacked-bar-timeseries', { + props: { + chart_data: { type:Object, required:false }, /* TimeseriesData */ + width: { type:Number, default: ChartPrefs.default_width }, + height: { type:Number, default: ChartPrefs.default_height }, + }, + + render: function(ce) { + return ChartVue.create_svg(ce, [0, 0, this.width, this.height]); + }, + + data: function() { + return { + tsdata: null, + stacked: null, + margin: { + top: ChartPrefs.axis_font_size, + bottom: ChartPrefs.axis_font_size*2, + left: ChartPrefs.axis_font_size*3, + right: ChartPrefs.axis_font_size + }, + xscale: null, + yscale: null, + colors: ChartPrefs.colors, /* array of colors */ + }; + }, + + mounted: function() { + if (this.chart_data) { + this.stack(this.chart_data); + this.draw(); + } + }, + + watch: { + 'chart_data': function(newv, oldv) { + this.stack(newv); + this.draw(); + } + }, + + methods: { + stack: function(data) { + /* "stack" the data using d3.stack() */ + // 1. reorganize into the format stack() wants -- an + // array of objects, with each object having a key of + // 'date', plus one for each series + var stacker_input = data.dates.map((d, i) => { + var array_el = { date: d } + data.series.forEach(s => { + array_el[s.name] = s.values[i]; + }) + return array_el; + }); + + // 2. call d3.stack() to get the stacking function, which + // creates yet another version of the series data, + // reformatted for more easily creating stacked bars. + // + // It returns a new array (see the d3 docs): + // [ + // [ /* series 1 */ + // [ Number, Number, data: { date: Date } ], + // [ ... ], ... + // ], + // [ /* series 2 */ + // [ Number, Number, data: { date: Date } ], + // [ ... ], ... + // ], + // ... + // ] + // + var stacker = d3.stack() + .keys(data.series.map(s => s.name)) + .order(d3.stackOrderNone) + .offset(d3.stackOffsetNone); + + // 3. store the data + this.tsdata = data; + this.stacked = stacker(stacker_input); + }, + + + draw: function() { + const svg = d3.select(this.$el); + svg.selectAll("g").remove(); + + if (this.tsdata.dates.length == 0) { + // no data ... + svg.append("g") + .append("text") + .attr("font-family", ChartPrefs.default_font_family) + .attr("font-size", ChartPrefs.label_font_size) + .attr("text-anchor", "middle") + .attr("x", this.width/2) + .attr("y", this.height/2) + .text("no data"); + } + + this.xscale = d3.scaleUtc() + .domain(d3.extent(this.tsdata.dates)) + .nice() + .range([this.margin.left, this.width - this.margin.right]) + + var barwidth = this.tsdata.barwidth(this.xscale, 1); + var padding = barwidth / 2; + + this.yscale = d3.scaleLinear() + .domain([ + 0, + d3.sum(this.tsdata.series, s => d3.max(s.values)) + ]) + .range([ + this.height - this.margin.bottom, + this.margin.top, + ]); + + svg.append("g") + .call(this.xAxis.bind(this, padding)) + .attr("font-size", ChartPrefs.axis_font_size); + + svg.append("g") + .call(this.yAxis.bind(this)) + .attr("font-size", ChartPrefs.axis_font_size); + + + for (var s_idx=0; s_idx this.xscale(d.data.date) - barwidth/2 + padding) + .attr("y", d => this.yscale(d[1])) + .attr("height", d => this.yscale(d[0]) - this.yscale(d[1])) + .attr("width", barwidth) + .call( hover.bind(this) ) + + // .append("title") + // .text(d => `${this.tsdata.series[s_idx].name}: ${NumberFormatter.format(d.data[this.tsdata.series[s_idx].name])}`) + ; + } + + var hovinfo = svg.append("g"); + + function hover(rect) { + if ("ontouchstart" in document) rect + .style("-webkit-tap-highlight-color", "transparent") + .on("touchstart", entered.bind(this)) + .on("touchend", left) + else rect + .on("mouseenter", entered.bind(this)) + .on("mouseleave", left); + + function entered(event, d) { + var rect = d3.select(event.target) + .attr("fill", "#ccc"); + var d = rect.datum(); + var s_idx = d3.select(rect.node().parentNode).datum(); + var s_name = this.tsdata.series[s_idx].name; + var v = d.data[s_name]; + var x = Number(rect.attr('x')) + barwidth/2; + + hovinfo.attr( + "transform", + `translate( ${x}, ${rect.attr('y')} )`) + .append('text') + .attr("font-family", ChartPrefs.default_font_family) + .attr("font-size", ChartPrefs.default_font_size) + .attr("text-anchor", "middle") + .attr("y", -3) + .text(`${this.tsdata.formatDateTimeShort(d.data.date)}`); + hovinfo.append("text") + .attr("font-family", ChartPrefs.default_font_family) + .attr("font-size", ChartPrefs.default_font_size) + .attr("text-anchor", "middle") + .attr("y", -3 - ChartPrefs.default_font_size) + .text(`${s_name} (${NumberFormatter.format(v)})`); + + } + + function left(event) { + d3.select(event.target).attr("fill", null); + hovinfo.selectAll("text").remove(); + } + } + }, + + xAxis: function(padding, g) { + var x = g.attr( + 'transform', + `translate(${padding}, ${this.height - this.margin.bottom})` + ).call( + d3.axisBottom(this.xscale) + .ticks(this.width / 80) + .tickSizeOuter(0) + ); + return x; + }, + + yAxis: function(g) { + var y = g.attr( + "transform", + `translate(${this.margin.left},0)` + ).call( + d3.axisLeft(this.yscale) + .ticks(this.height/50) + ).call(g => + g.select(".domain").remove() + ).call(g => { + ChartVue.add_yAxisLegend(g, this.tsdata, this.colors); + }); + + return y; + }, + + } +}); + + diff --git a/management/reporting/ui/chart-table.js b/management/reporting/ui/chart-table.js new file mode 100644 index 00000000..b2f7ca2d --- /dev/null +++ b/management/reporting/ui/chart-table.js @@ -0,0 +1,45 @@ +Vue.component('chart-table', { + props: { + items: Array, + fields: Array, + caption: String + }, + + /* */ + render: function(ce) { + var scopedSlots= { + 'table-caption': props => + ce('span', { class: { 'text-nowrap':true }}, this.caption) + }; + if (this.$scopedSlots) { + for (var slotName in this.$scopedSlots) { + scopedSlots[slotName] = this.$scopedSlots[slotName]; + } + } + var table = ce('b-table-lite', { + props: { + 'striped': true, + 'small': true, + 'fields': this.fields_x, + 'items': this.items, + 'caption-top': true + }, + scopedSlots: scopedSlots + }); + + return table; + }, + + computed: { + fields_x: function() { + if (this.items.length == 0) { + return [{ + key: 'no data', + thClass: 'text-nowrap' + }]; + } + return this.fields; + } + } + +}); diff --git a/management/reporting/ui/chart.readme.txt b/management/reporting/ui/chart.readme.txt new file mode 100644 index 00000000..f695809e --- /dev/null +++ b/management/reporting/ui/chart.readme.txt @@ -0,0 +1,20 @@ +/* +most charts have been adapted from ones at Obserable, +which had the following license: + +https://observablehq.com/@d3/multi-line-chart + +Copyright 2018–2020 Observable, Inc. + +Permission to use, copy, modify, and/or distribute this software for any +purpose with or without fee is hereby granted, provided that the above +copyright notice and this permission notice appear in all copies. + +THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES +WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF +MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR +ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES +WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN +ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF +OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. +*/ diff --git a/management/reporting/ui/charting.js b/management/reporting/ui/charting.js new file mode 100644 index 00000000..98618f95 --- /dev/null +++ b/management/reporting/ui/charting.js @@ -0,0 +1,972 @@ + +class ChartPrefs { + static get colors() { + // see: https://github.com/d3/d3-scale-chromatic + return d3.schemeSet2; + } + + static get line_colors() { + // see: https://github.com/d3/d3-scale-chromatic + return d3.schemeCategory10; + } + + static get default_width() { + return 600; + } + + static get default_height() { + return 400; + } + + static get axis_font_size() { + return 12; + } + + static get default_font_size() { + return 10; + } + + static get label_font_size() { + return 12; + } + + static get default_font_family() { + return "sans-serif"; + } + + static get locales() { + return "en"; + } +}; + + +class DateFormatter { + /* + * date and time + */ + static dt_long(d, options) { + let opt = { + weekday: 'long', + year: 'numeric', + month: 'long', + day: 'numeric' + }; + Object.assign(opt, options); + return d.toLocaleString(ChartPrefs.locales, opt); + } + static dt_short(d, options) { + return d.toLocaleString(ChartPrefs.locales, options); + } + + /* + * date + */ + static d_long(d, options) { + let opt = { + weekday: 'long', + year: 'numeric', + month: 'long', + day: 'numeric' + }; + Object.assign(opt, options); + return d.toLocaleDateString(ChartPrefs.locales, opt); + } + static d_short(d, options) { + return d.toLocaleDateString(ChartPrefs.locales, options); + } + + /* + * time + */ + static t_long(d, options) { + return d.toLocaleTimeString(ChartPrefs.locales, options); + } + static t_short(d, options) { + return d.toLocaleTimeString(ChartPrefs.locales, options); + } + static t_span(d, unit) { + // `d` is milliseconds + // `unit` is desired max precision output unit (eg 's') + unit = unit || 's'; + const cvt = [{ + ms: (24 * 60 * 60 * 1000), + ushort: 'd', + ulong: 'day' + }, { + ms: (60 * 60 * 1000), + ushort: 'h', + ulong: 'hour' + }, { + ms: (60 * 1000), + ushort: 'm', + ulong: 'minute' + }, { + ms: 1000, + ushort: 's', + ulong: 'second' + }, { + ms: 1, + ushort: 'ms', + ulong: 'milliseconds' + }]; + + var first = false; + var remainder = d; + var out = []; + var done = false; + cvt.forEach( c => { + if (done) return; + var amt = Math.floor( remainder / c.ms ); + remainder = remainder % c.ms; + if (first || amt > 0) { + first = true; + out.push(amt + c.ushort); + } + if (unit == c.ushort || unit == c.ulong) { + done = true; + } + }); + return out.join(' '); + } + + /* + * universal "YYYY-MM-DD HH:MM:SS" formats + */ + static ymd(d) { + const ye = d.getFullYear(); + const mo = '0'+(d.getMonth() + 1); + const da = '0'+d.getDate(); + return `${ye}-${mo.substr(mo.length-2)}-${da.substr(da.length-2)}`; + } + + static ymd_utc(d) { + const ye = d.getUTCFullYear(); + const mo = '0'+(d.getUTCMonth() + 1); + const da = '0'+d.getUTCDate(); + return `${ye}-${mo.substr(mo.length-2)}-${da.substr(da.length-2)}`; + } + + static ymdhms(d) { + const ho = '0'+d.getHours(); + const mi = '0'+d.getMinutes(); + const se = '0'+d.getSeconds(); + return `${DateFormatter.ymd(d)} ${ho.substr(ho.length-2)}:${mi.substr(mi.length-2)}:${se.substr(se.length-2)}`; + } + + static ymdhms_utc(d) { + const ho = '0'+d.getUTCHours(); + const mi = '0'+d.getUTCMinutes(); + const se = '0'+d.getUTCSeconds(); + return `${DateFormatter.ymd_utc(d)} ${ho.substr(ho.length-2)}:${mi.substr(mi.length-2)}:${se.substr(se.length-2)}`; + } +}; + + +class DateRange { + /* + * ranges + */ + static ytd() { + var s = new Date(); + s.setMonth(0); + s.setDate(1); + s.setHours(0); + s.setMinutes(0); + s.setSeconds(0); + s.setMilliseconds(0); + return [ s, new Date() ]; + } + static ytd_as_ymd() { + return DateRange.ytd().map(d => DateFormatter.ymd(d)); + } + + static mtd() { + var s = new Date(); + s.setDate(1); + s.setHours(0); + s.setMinutes(0); + s.setSeconds(0); + s.setMilliseconds(0); + return [ s, new Date() ]; + } + static mtd_as_ymd() { + return DateRange.mtd().map(d => DateFormatter.ymd(d)); + } + + static wtd() { + var s = new Date(); + var offset = s.getDay() * (24 * 60 * 60 * 1000); + s.setTime(s.getTime() - offset); + s.setHours(0); + s.setMinutes(0); + s.setSeconds(0); + s.setMilliseconds(0); + return [ s, new Date() ]; + } + static wtd_as_ymd() { + return DateRange.wtd().map(d => DateFormatter.ymd(d)); + } + + static rangeFromType(type) { + if (type == 'wtd') + return DateRange.wtd(); + else if (type == 'mtd') + return DateRange.mtd(); + else if (type == 'ytd') + return DateRange.ytd(); + return null; + } +}; + + +class NumberFormatter { + static format(v) { + return isNaN(v) || v===null ? "N/A" : v.toLocaleString(ChartPrefs.locales); + } + + static decimalFormat(v, places, style) { + if (isNaN(v) || v===null) return "N/A"; + if (places === undefined || isNaN(places)) places = 1; + if (style === undefined || typeof style != 'string') style = 'decimal'; + var options = { + style: style, + minimumFractionDigits: places + }; + v = v.toLocaleString(ChartPrefs.locales, options); + return v; + } + + static percentFormat(v, places) { + if (places === undefined || isNaN(places)) places = 0; + return NumberFormatter.decimalFormat(v, places, 'percent'); + } + + static humanFormat(v, places) { + if (isNaN(v) || v===null) return "N/A"; + if (places === undefined || isNaN(places)) places = 1; + const options = { + style: 'unit', + minimumFractionDigits: places, + unit: 'byte' + }; + var xunit = ''; + const f = Math.pow(10, places); + if (v >= NumberFormatter.tb) { + v = Math.round(v / NumberFormatter.tb * f) / f; + options.unit='terabyte'; + xunit = 'T'; + } + else if (v >= NumberFormatter.gb) { + v = Math.round(v / NumberFormatter.gb * f) / f; + options.unit='gigabyte'; + xunit = 'G'; + } + else if (v >= NumberFormatter.mb) { + v = Math.round(v / NumberFormatter.mb * f) / f; + options.unit='megabyte'; + xunit = 'M'; + } + else if (v >= NumberFormatter.kb) { + v = Math.round(v / NumberFormatter.kb * f) / f; + options.unit='kilobyte'; + xunit = 'K'; + } + else { + options.minimumFractionDigits = 0; + places = 0; + } + try { + return v.toLocaleString(ChartPrefs.locales, options); + } catch(e) { + if (e instanceof RangeError) { + // probably "invalid unit" + return NumberFormatter.decimalFormat(v, places) + xunit; + } + } + } +}; + +// define static constants in NumberFormatter +['kb','mb','gb','tb'].forEach((unit,idx) => { + Object.defineProperty(NumberFormatter, unit, { + value: Math.pow(1024, idx+1), + writable: false, + enumerable: false, + configurable: false + }); +}); + + +class BvTable { + constructor(data, opt) { + opt = opt || {}; + Object.assign(this, data); + if (!this.items || !this.fields || !this.field_types) { + throw new AssertionError(); + } + + BvTable.arraysToObjects(this.items, this.fields); + BvTable.setFieldDefinitions(this.fields, this.field_types); + + if (opt._showDetails) { + // _showDetails must be set to make it reactive + this.items.forEach(item => { + item._showDetails = false; + }) + } + } + + field_index_of(key) { + for (var i=0; i=0) return this.fields[i]; + return this.x_fields && !only_showing ? this.x_fields[key] : null; + } + + combine_fields(names, name2, formatter) { + // combine field(s) `names` into `name2`, then remove + // `names`. use `formatter` as the formatter function + // for the new combined field. + // + // if name2 is not given, just remove all `names` fields + // + // removed fields are placed into this.x_fields array + if (typeof names == 'string') names = [ names ] + var idx2 = name2 ? this.field_index_of(name2) : -1; + if (! this.x_fields) this.x_fields = {}; + + names.forEach(name1 => { + var idx1 = this.field_index_of(name1); + if (idx1 < 0) return; + this.x_fields[name1] = this.fields[idx1]; + this.fields.splice(idx1, 1); + if (idx2>idx1) --idx2; + }); + + if (idx2 < 0) return null; + + this.fields[idx2].formatter = formatter; + return this.fields[idx2]; + } + + + static arraysToObjects(items, fields) { + /* + * convert array-of-arrays `items` to an array of objects + * suitable for a items (rows of the table). + * + * `items` is modified in-place + * + * `fields` is an array of strings, which will become the keys + * of each new object. the length of each array in `items` + * must match the length of `fields` and the indexes must + * correspond. + * + * the primary purpose is to allow the data provider (server) + * to send something like: + * + * { "items": [ + * [ "alice", 10.6, 200, "top-10" ], + * .... + * ], + * "fields": [ "name", "x", "y", "label" ] + * } + * + * instead of: + * + * { "items": [ + * { "name":"a", "x":10.6, "y":200, "label":"top-10" }, + * ... + * ], + * "fields": [ "name", "x", "y", "label" ] + * } + * + * which requires much more bandwidth + * + */ + if (items.length > 0 && !Array.isArray(items[0])) + { + // already converted + return; + } + for (var i=0; i { + o[field] = items[i][idx]; + }); + items[i] = o; + } + } + + static setFieldDefinitions(fields, types) { + /* + * change elements of array `fields` to bootstrap-vue table + * field (column) definitions + * + * `fields` is an array of field names or existing field + * definitions to update. `types` is a correponding array + * having the type of each field which will cause one or more + * of the following properties to be set on each field: + * 'tdClass', 'thClass', 'label', and 'formatter' + */ + for (var i=0; i0) { + // optional format, eg "text/email" + var s = ft.type.split('/'); + ft.type = s[0]; + ft.subtype = s.length > 1 ? s[1] : null; + } + + if (ft.label !== undefined) { + field.label = ft.label; + } + + if (ft.type == 'decimal') { + Object.assign(ft, { + type: 'number', + subtype: 'decimal' + }); + } + + if (ft.type == 'text') { + // as-is + } + else if (ft.type == 'number') { + if (ft.subtype == 'plain' || + ft.subtype == 'decimal' && isNaN(ft.places) + ) + { + Object.assign( + field, + BvTableField.numberFieldDefinition() + ); + } + + else if (ft.subtype == 'size') { + Object.assign( + field, + BvTableField.sizeFieldDefinition() + ); + } + + else if (ft.subtype == 'decimal') { + Object.assign( + field, + BvTableField.decimalFieldDefinition(ft.places) + ); + } + + else if (ft.subtype == 'percent') { + Object.assign( + field, + BvTableField.percentFieldDefinition(ft.places) + ); + } + } + else if (ft.type == 'datetime') { + Object.assign( + field, + BvTableField.datetimeFieldDefinition(ft.showas || 'short', ft.format) + ); + } + else if (ft.type == 'time' && ft.subtype == 'span') { + Object.assign( + field, + BvTableField.timespanFieldDefinition(ft.unit || 'ms') + ); + } + + } + + + static numberFieldDefinition() { + // field definition for a localized numeric value. + // eg: "5,001". for additional attributes, see: + // https://bootstrap-vue.org/docs/components/table#field-definition-reference + return { + formatter: NumberFormatter.format, + tdClass: 'text-right', + thClass: 'text-right' + }; + } + + static sizeFieldDefinition(decimal_places) { + // field definition for a localized numeric value in + // human readable format. eg: "5.1K". `decimal_places` is + // optional, which defaults to 1 + return { + formatter: value => + NumberFormatter.humanFormat(value, decimal_places), + tdClass: 'text-right text-nowrap', + thClass: 'text-right' + }; + } + + static datetimeFieldDefinition(variant, format) { + // if the formatter is passed string (utc) dates, convert them + // to a native Date objects using the format in `format`. eg: + // "%Y-%m-%d %H:%M:%S". + // + // `variant` can be "long" (default) or "short" + var parser = (format ? d3.utcParse(format) : null); + if (variant === 'short') { + return { + formatter: v => + DateFormatter.dt_short(parser ? parser(v) : v) + }; + } + else { + return { + formatter: v => + DateFormatter.dt_long(parser ? parser(v) : v) + }; + } + } + + static timespanFieldDefinition(unit, output_unit) { + var factor = 1; + if (unit == 's') factor = 1000; + return { + formatter: v => DateFormatter.t_span(v * factor, output_unit) + }; + } + + static decimalFieldDefinition(decimal_places) { + return { + formatter: value => + NumberFormatter.decimalFormat(value, decimal_places), + tdClass: 'text-right', + thClass: 'text-right' + }; + } + + static percentFieldDefinition(decimal_places) { + return { + formatter: value => + NumberFormatter.percentFormat(value, decimal_places), + tdClass: 'text-right', + thClass: 'text-right' + }; + } + + add_cls(cls, to_what) { + if (Array.isArray(this[to_what])) { + this[to_what].push(cls); + } + else if (this[to_what] !== undefined) { + this[to_what] = [ this[to_what], cls ]; + } + else { + this[to_what] = cls; + } + } + + add_tdClass(cls) { + this.add_cls(cls, 'tdClass'); + } + +}; + + +class MailBvTable extends BvTable { + flag(key, fn) { + var field = this.get_field(key, true); + if (!field) return; + field.add_tdClass(fn); + } + + flag_fields(tdClass) { + // flag on certain cell values by setting _flagged in each + // "flagged" item during its tdClass callback function. add + // `tdClass` to the rendered value + + tdClass = tdClass || 'text-danger'; + + this.flag('accept_status', (v, key, item) => { + if (v === 'reject') { + item._flagged = true; + return tdClass; + } + }); + + this.flag('relay', (v, key, item) => { + if (item.delivery_connection == 'untrusted') { + item._flagged = true; + return tdClass; + } + }); + + this.flag('status', (v, key, item) => { + if (v != 'sent') { + item._flagged = true; + return tdClass; + } + }); + + this.flag('spam_result', (v, key, item) => { + if (item.spam_result && v != 'clean') { + item._flagged = true; + return tdClass; + } + }); + + this.flag('spf_result', (v, key, item) => { + if (v == 'Fail' ||v == 'Softfail') { + item._flagged = true; + return tdClass; + } + }); + + this.flag('dkim_result', (v, key, item) => { + if (item.dkim_result && v != 'pass') { + item._flagged = true; + return tdClass; + } + }); + + this.flag('dmarc_result', (v, key, item) => { + if (v == 'fail') { + item._flagged = true; + return tdClass; + } + }); + + this.flag('postgrey_result', (v, key, item) => { + if (item.postgrey_result && v != 'pass') { + item._flagged = true; + return tdClass; + } + }); + + this.flag('disposition', (v, key, item) => { + if (item.disposition != 'ok') { + item._flagged = true; + return tdClass; + } + }); + + return this; + } + + apply_rowVariant_grouping(variant, group_fn) { + // there is 1 row for each recipient of a message + // - give all rows of the same message the same + // color + // + // variant is a bootstrap variant like "primary" + // + // group_fn is a callback receiving an item (row of data) and + // the item index and should return null if the item is not + // showing or return the group value + var last_group = -1; + var count = 0; + for (var idx=0; idx < this.items.length; idx++) + { + const item = this.items[idx]; + const group = group_fn(item, idx); + if (group === null || group === undefined) continue + + if (group != last_group) { + ++count; + last_group = group; + } + item._rowVariant = count % 2 == 0 ? variant : ''; + } + } +} + + +class ChartVue { + + static svg_attrs(viewBox) { + var attrs = { + width: viewBox[2], + height: viewBox[3], + viewBox: viewBox.join(' '), + style: 'overflow: visible', + xmlns: 'http://www.w3.org/2000/svg' + }; + return attrs; + } + + static create_svg(create_fn, viewBox, children) { + var svg = create_fn('svg', { + attrs: ChartVue.svg_attrs(viewBox), + children + }); + return svg; + } + + static add_yAxisLegend(g, data, colors) { + //var gtick = g.select(".tick:last-of-type").append("g"); + const h = ChartPrefs.axis_font_size; + var gtick = g.append("g") + .attr('transform', + `translate(0, ${h * data.series.length})`); + + gtick.selectAll('rect') + .data(data.series) + .join('rect') + .attr('x', 3) + .attr('y', (d, i) => -h + i*h) + .attr('width', h) + .attr('height', h) + .attr('fill', (d, i) => colors[i]); + gtick.selectAll('text') + .data(data.series) + .join('text') + .attr('x', h + 6) + .attr('y', (d, i) => i*h ) + .attr("text-anchor", "start") + .attr("font-weight", "bold") + .attr("fill", 'currentColor') + .text(d => d.name); + return g; + } +}; + + + +/* + * Timeseries data layout: { + * y: 'description', + * binsize: Number, // size in minutes, + * date_parse_format: '%Y-%m-%d', + * dates: [ 'YYYY-MM-DD HH:MM:SS', ... ], + * series: [ + * { + * id: 'id', + * name: 'series 1 desc', + * values: [ Number, .... ] + * }, + * { + * id: 'id', + * name: 'series 2 desc' + * values: [ ... ], + * }, + * ... + * ] + * } + */ + +class TimeseriesData { + constructor(data) { + Object.assign(this, data); + this.convert_dates(); + } + + get_series(id) { + for (var i=0; i desired[id] = true); + this.series.forEach(s => { + if (desired[s.id]) dataview.series.push(s); + }); + return new TimeseriesData(dataview); + } + + binsizeWithUnit() { + // normalize binsize (which is a time span in minutes) + const days = Math.floor(this.binsize / (24 * 60)); + const hours = Math.floor( (this.binsize - days*24*60) / 60 ); + const mins = this.binsize - days*24*60 - hours*60; + if (days == 0 && hours == 0) { + return { + unit: 'minute', + value: mins + }; + } + if (days == 0) { + return { + unit: 'hour', + value: hours + }; + } + return { + unit: 'day', + value: days + }; + } + + binsizeTimespan() { + /* return the binsize timespan in seconds */ + return this.binsize * 60; + } + + static binsizeOfRange(range) { + // target 100-120 datapoints + const target = 100; + const tolerance = 0.2; // 20% + + if (typeof range[0] == 'string') { + var parser = d3.utcParse('%Y-%m-%d %H:%M:%S'); + range = range.map(parser); + } + + const span_min = Math.ceil( + (range[1].getTime() - range[0].getTime()) / (1000*60*target) + ); + const bin_days = Math.floor(span_min / (24*60)); + const bin_hours = Math.floor((span_min - bin_days*24*60) / 60); + if (bin_days >= 1) { + return bin_days * 24 * 60 + + (bin_hours > (24 * tolerance) ? bin_hours*60: 0); + } + + const bin_mins = span_min - bin_days*24*60 - bin_hours*60; + if (bin_hours >= 1) { + return bin_hours * 60 + + (bin_mins > (60 * tolerance) ? bin_mins: 0 ); + } + return bin_mins; + } + + barwidth(xscale, barspacing) { + /* get the width of a bar in a bar chart */ + var start = this.range[0]; + var end = this.range[1]; + var bins = (end.getTime() - start.getTime()) / (1000 * this.binsizeTimespan()); + return Math.max(1, (xscale.range()[1] - xscale.range()[0])/bins - (barspacing || 0)); + } + + formatDateTimeLong(d) { + var options = { hour: 'numeric' }; + var b = this.binsizeWithUnit(); + if (b.unit === 'minute') { + options.minute = 'numeric'; + return DateFormatter.dt_long(d, options); + } + if (b.unit === 'hour') { + return DateFormatter.dt_long(d, options); + } + if (b.unit === 'day') { + return DateFormatter.d_long(d); + } + throw new Error(`Unknown binsize unit: ${b.unit}`); + } + + formatDateTimeShort(d) { + var options = { + year: 'numeric', + month: 'numeric', + day: 'numeric', + weekday: undefined + }; + var b = this.binsizeWithUnit(); + if (b.unit === 'minute') { + Object.assign(options, { + hour: 'numeric', + minute: 'numeric' + }); + return DateFormatter.dt_long(d, options); + } + if (b.unit === 'hour') { + options.hour = 'numeric'; + return DateFormatter.dt_long(d, options); + } + if (b.unit === 'day') { + return DateFormatter.d_short(d); + } + throw new Error(`Unknown binsize unit: ${b.unit}`); + } + + + convert_dates() { + // all dates from the server are UTC strings + // convert to Date objects + if (this.dates.length > 0 && typeof this.dates[0] == 'string') + { + var parser = d3.utcParse(this.date_parse_format); + this.dates = this.dates.map(parser); + } + if (this.range.length > 0 && typeof this.range[0] == 'string') + { + var parser = d3.utcParse(this.range_parse_format); + this.range = this.range.map(parser); + } + } +}; + + +class ConnectionDisposition { + constructor(disposition) { + const data = { + 'failed_login_attempt': { + short_desc: 'failed login attempt', + }, + 'insecure': { + short_desc: 'insecure connection' + }, + 'ok': { + short_desc: 'normal, secure connection' + }, + 'reject': { + short_desc: 'mail attempt rejected' + }, + 'suspected_scanner': { + short_desc: 'suspected scanner' + } + }; + this.disposition = disposition; + this.info = data[disposition]; + if (! this.info) { + this.info = { + short_desc: disposition.replace('_',' ') + } + } + } + + get short_desc() { + return this.info.short_desc; + } + + static formatter(disposition) { + return new ConnectionDisposition(disposition).short_desc; + } +}; diff --git a/management/reporting/ui/date-range-picker.js b/management/reporting/ui/date-range-picker.js new file mode 100644 index 00000000..9fd1f918 --- /dev/null +++ b/management/reporting/ui/date-range-picker.js @@ -0,0 +1,165 @@ +Vue.component('date-range-picker', { + props: { + start_range: [ String, Array ], // "ytd", "mtd", "wtd", or [start, end] where start and end are strings in format YYYY-MM-DD in localti + recall_id: String, // save / recall from localStorage + }, + template: '
'+ + '
Date range:
'+ + '
From:
' + + '
To:
' + + '
' + , + data: function() { + var range_type = null; + var range = null; + var default_range_type = 'mtd'; + const recall_id_prefix = 'date-range-picker/'; + + var v = null; + if (typeof this.start_range === 'string') { + if (this.start_range.substring(0,1) == '-') { + default_range_type = this.start_range.substring(1); + } + else { + v = this.validate_input_range(this.start_range); + } + } + else { + v = this.validate_input_range(this.start_range); + } + + if (v) { + // handles explicit valid "range-type", [ start, end ] + range_type = v.range_type; + range = v.range; + } + else if (this.recall_id) { + const id = recall_id_prefix+this.recall_id; + try { + var v = JSON.parse(localStorage.getItem(id)); + range = v.range; + range_type = v.range_type; + } catch(e) { + // pass + console.error(e); + console.log(localStorage.getItem(id)); + } + } + + if (!range) { + range_type = default_range_type; + range = DateRange.rangeFromType(range_type) + .map(DateFormatter.ymd); + } + + return { + recall_id_full:(this.recall_id ? + recall_id_prefix + this.recall_id : + null), + range: range, + range_type: range_type, + options: [ + { value:'wtd', text:'Week-to-date' }, + { value:'mtd', text:'Month-to-date' }, + { value:'ytd', text:'Year-to-date' }, + { value:'custom', text:'Custom' } + ], + } + }, + + created: function() { + this.notify_change(true); + }, + + watch: { + 'range': function() { + this.notify_change(); + } + }, + + methods: { + + validate_input_range: function(range) { + // if range is a string it's a range_type (eg "ytd") + // othersize its an 2-element array [start, end] of dates + // in YYYY-MM-DD (localtime) format + if (typeof range == 'string') { + var dates = DateRange.rangeFromType(range) + .map(DateFormatter.ymd); + if (! range) return null; + return { range:dates, range_type:range }; + } + else if (range.length == 2) { + var parser = d3.timeParse('%Y-%m-%d'); + if (! parser(range[0]) || !parser(range[1])) + return null; + return { range, range_type:'custom' }; + } + else { + return null; + } + }, + + set_range: function(range) { + // if range is a string it's a range_type (eg "ytd") + // othersize its an 2-element array [start, end] of dates + // in YYYY-MM-DD (localtime) format + var v = this.validate_input_range(range); + if (!v) return false; + this.range = v.range; + this.range_type = v.range_type; + this.notify_change(); + return true; + }, + + notify_change: function(init) { + var parser = d3.timeParse('%Y-%m-%d'); + + var end_utc = new Date(); + end_utc.setTime( + parser(this.range[1]).getTime() + (24 * 60 * 60 * 1000) + ); + var range_utc = [ + DateFormatter.ymdhms_utc(parser(this.range[0])), + DateFormatter.ymdhms_utc(end_utc) + ]; + + this.$emit('change', { + // localtime "YYYY-MM-DD" format - exactly what came + // from the ui + range: this.range, + + // convert localtime to utc, include hours. add 1 day + // to end so that range_utc encompasses times >=start + // and + + + Activity - MiaB-LDAP + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + diff --git a/management/reporting/ui/index.js b/management/reporting/ui/index.js new file mode 100644 index 00000000..d47061fb --- /dev/null +++ b/management/reporting/ui/index.js @@ -0,0 +1,75 @@ +/* + * reports index page + */ + + + +const app = { + router: new VueRouter({ + routes: [ + { path: '/', component: Vue.component('page-reports-main') }, + { path: '/settings', component: Vue.component('page-settings') }, + { path: '/:panel', component: Vue.component('page-reports-main') }, + ], + scrollBehavior: function(to, from, savedPosition) { + if (savedPosition) { + return savedPosition + } + }, + }), + + components: { + 'page-settings': Vue.component('page-settings'), + 'page-reports-main': Vue.component('page-reports-main'), + }, + + data: { + me: null, + }, + + mounted: function() { + this.getMe(); + }, + + methods: { + getMe: function() { + axios.get('me').then(response => { + this.me = new Me(response.data); + }).catch(error => { + this.handleError(error); + }); + }, + + handleError: function(error) { + if (error instanceof AuthenticationError) { + console.log(error); + window.location = '/admin'; + return; + } + + console.error(error); + if (error instanceof ReferenceError) { + // uncaught coding bug, ignore + return; + } + if (error.status && error.reason) + { + // axios + error = error.reason; + } + this.$nextTick(() => {alert(''+error) }); + } + } +}; + + + +function init_app() { + init_axios_interceptors(); + + UserSettings.load().then(settings => { + new Vue(app).$mount('#app'); + }).catch(error => { + alert('' + error); + }); +} diff --git a/management/reporting/ui/page-reports-main.html b/management/reporting/ui/page-reports-main.html new file mode 100644 index 00000000..f2f16560 --- /dev/null +++ b/management/reporting/ui/page-reports-main.html @@ -0,0 +1,94 @@ + + + + + + + + + + +
+ +
+
+ + + Choose + + Messages sent + + Messages received + + User activity + + Remote sender activity + + Notable connections + + + + + + + + + + + + + + + + + + + + + + + + +
diff --git a/management/reporting/ui/page-reports-main.js b/management/reporting/ui/page-reports-main.js new file mode 100644 index 00000000..d9d52707 --- /dev/null +++ b/management/reporting/ui/page-reports-main.js @@ -0,0 +1,110 @@ +Vue.component('page-reports-main', function(resolve, reject) { + axios.get('reports/ui/page-reports-main.html').then((response) => { resolve({ + + template: response.data, + + components: { + 'page-layout': Vue.component('page-layout'), + 'reports-page-header': Vue.component('reports-page-header'), + 'date-range-picker': Vue.component('date-range-picker'), + 'panel-messages-sent': Vue.component('panel-messages-sent'), + 'panel-messages-received': Vue.component('panel-messages-received'), + 'panel-flagged-connections': Vue.component('panel-flagged-connections'), + 'panel-user-activity': Vue.component('panel-user-activity'), + }, + + data: function() { + return { + // page-header loading spinner + loading: 0, + + // panels + panel: this.$route.params.panel || '', + + // date picker - the range, if in the route, is set + // via date_change(), called during date picker + // create() + range_utc: null, // Array(2): Date objects (UTC) + range: null, // Array(2): YYYY-MM-DD (localtime) + range_type: null, // String: "custom","ytd","mtd", etc + }; + }, + + beforeRouteUpdate: function(to, from, next) { + //console.log(`page route update: to=${JSON.stringify({path:to.path, query:to.query})} .... from=${JSON.stringify({path:from.path, query:from.query})}`); + this.panel = to.params.panel; + // note: the range is already set in the class data + // + // 1. at component start - the current range is extracted + // from $route by get_route_range() and passed to the + // date picker as a prop, which subsequently + // $emits('change') during creation where date_change() + // updates the class data. + // + // 2. during user interaction - the date picker + // $emits('change') where date_change() updates the + // class data. + next(); + }, + + methods: { + get_start_range: function(to, default_range) { + if (to.query.range_type) { + return to.query.range_type; + } + else if (to.query.start && to.query.end) { + // start and end must be YYYY-MM-DD (localtime) + return [ + to.query.start, + to.query.end + ]; + } + else { + return default_range; + } + }, + + get_binsize: function() { + if (! this.range_utc) return 0; + return TimeseriesData.binsizeOfRange(this.range_utc); + }, + + date_change: function(evt) { + // date picker 'change' event + this.range_type = evt.range_type; + this.range_utc = evt.range_utc; + this.range = evt.range; + var route = this.get_route(this.panel); + if (! evt.init) { + this.$router.replace(route); + } + }, + + get_route: function(panel, ex_query) { + // return vue-router route to `panel` + // eg: "/?start=YYYY-MM-DD&end=YYYY-MM-DD" + // + // additional panel query elements should be set in + // the panel's activate method + var route = { path: panel }; + if (this.range_type != 'custom') { + route.query = { + range_type: this.range_type + }; + } + else { + route.query = { + start: this.range[0], + end: this.range[1] + }; + } + Object.assign(route.query, ex_query); + return route; + }, + + } + })}).catch((e) => { + reject(e); + }); + +}); diff --git a/management/reporting/ui/page-settings.html b/management/reporting/ui/page-settings.html new file mode 100644 index 00000000..cb454a46 --- /dev/null +++ b/management/reporting/ui/page-settings.html @@ -0,0 +1,72 @@ + + + + +
+
+
Settings
+ Back to reports +
+ + + + + UI settings + + + +
+
Table data row limit
+ +
+ {{row_limit_error}} +
+
+
+
+ + + + + Capture daemon + + + +

+ {{status[0]}} + {{status[1]}} + {{ capture_config.capture ? 'capturing' : 'paused' }} +

+

(systemd service "miabldap-capture")

+ + + + Capture enabled + Warning: when capture is disabled, the daemon will no longer record log activity +
+
Delete database records older than
+ +
days
+
+
+ (a value of zero preserves all records) +
+ + Ignore failed login attempts + Ignore suspected scanner activity + Ignore rejected mail attempts + +
+ +
+ Commit changes and update server +
+ +
+
+ +
+ +
diff --git a/management/reporting/ui/page-settings.js b/management/reporting/ui/page-settings.js new file mode 100644 index 00000000..93de87b6 --- /dev/null +++ b/management/reporting/ui/page-settings.js @@ -0,0 +1,123 @@ +Vue.component('page-settings', function(resolve, reject) { + axios.get('reports/ui/page-settings.html').then((response) => { resolve({ + + template: response.data, + + components: { + 'page-layout': Vue.component('page-layout'), + 'reports-page-header': Vue.component('reports-page-header'), + }, + + data: function() { + return { + from_route: null, + loading: 0, + + // server status and config + capture_config: null, + config_changed: false, + status: null, + + // capture config models that require processing + // before the value is valid for `capture_config`, or + // the `capture_config` value is used by multiple + // elements (eg. one showing current state) + capture: true, + older_than_days: '', + + // user settings + row_limit: '' + UserSettings.get().row_limit, + row_limit_error: '' + }; + }, + + beforeRouteEnter: function(to, from, next) { + next(vm => { + vm.from_route = from; + }); + }, + + created: function() { + this.loadData(); + }, + + methods: { + is_running: function() { + return this.status[0] == 'running'; + }, + + status_variant: function(status) { + if (status == 'running') return 'success'; + if (status == 'enabled') return 'success'; + if (status == 'disabled') return 'warning'; + if (status === true) return 'success'; + return 'danger' + }, + + loadData: function() { + this.loading += 1; + Promise.all([ + CaptureConfig.get(), + axios.get('/reports/capture/service/status') + ]).then(responses => { + this.capture_config = responses[0]; + if (this.capture_config.status !== 'error') { + this.older_than_days = '' + + this.capture_config.prune_policy.older_than_days; + this.capture = this.capture_config.capture; + } + this.status = responses[1].data; + this.config_changed = false; + }).catch(error => { + this.$root.handleError(error); + }).finally(() => { + this.loading -= 1; + }); + }, + + update_user_settings: function() { + if (this.row_limit == '') { + this.row_limit_error = 'not valid'; + return; + } + try { + const s = UserSettings.get(); + s.row_limit = Number(this.row_limit); + } catch(e) { + this.row_limit_error = e.message; + } + }, + + config_changed_if: function(v, range_min, range_max, current_value) { + v = Number(v); + if (range_min !== null && v < range_min) return; + if (range_max !== null && v > range_max) return; + if (current_value !== null && v == current_value) return; + this.config_changed = true; + }, + + save_capture_config: function() { + this.loading+=1; + var newconfig = Object.assign({}, this.capture_config); + this.capture_config.prune_policy.older_than_days = + Number(this.older_than_days); + newconfig.capture = this.capture; + axios.post('/reports/capture/config', newconfig) + .then(response => { + this.loadData(); + }) + .catch(error => { + this.$root.handleError(error); + }) + .finally(() => { + this.loading-=1; + }); + } + } + + })}).catch((e) => { + reject(e); + }); +}); + + diff --git a/management/reporting/ui/panel-flagged-connections.html b/management/reporting/ui/panel-flagged-connections.html new file mode 100644 index 00000000..652db1c5 --- /dev/null +++ b/management/reporting/ui/panel-flagged-connections.html @@ -0,0 +1,87 @@ +
+
+
+ Connections by disposition + + +
+ + + + + + + +
+ + + + + + + + + +
+ +
+
+ Mail delivery rejects by category + + +
+ + + + +
+ +
+ +
+ diff --git a/management/reporting/ui/panel-flagged-connections.js b/management/reporting/ui/panel-flagged-connections.js new file mode 100644 index 00000000..8621ca44 --- /dev/null +++ b/management/reporting/ui/panel-flagged-connections.js @@ -0,0 +1,133 @@ +Vue.component('panel-flagged-connections', function(resolve, reject) { + axios.get('reports/ui/panel-flagged-connections.html').then((response) => { resolve({ + + template: response.data, + + props: { + date_range: Array, // YYYY-MM-DD strings (UTC) + binsize: Number, // for timeseries charts, in minutes + user_link: Object, // a route + remote_sender_email_link: Object, // a route + remote_sender_server_link: Object, // a route + width: { type:Number, default: ChartPrefs.default_width }, + height: { type:Number, default: ChartPrefs.default_height }, + }, + + components: { + 'chart-multi-line-timeseries': Vue.component('chart-multi-line-timeseries'), + 'chart-stacked-bar-timeseries': Vue.component('chart-stacked-bar-timeseries'), + 'chart-pie': Vue.component('chart-pie'), + 'chart-table': Vue.component('chart-table'), + }, + + computed: { + radius_pie: function() { + return this.height / 5; + }, + linechart_height: function() { + return this.height / 2; + } + }, + + data: function() { + return { + data_date_range: null, + colors: ChartPrefs.colors, + failed_logins: null, // TimeseriesData + suspected_scanners: null, // TimeseriesData + connections_by_disposition: null, // pie chart data + disposition_formatter: ConnectionDisposition.formatter, + reject_by_failure_category: null, // pie chart data + top_hosts_rejected: null, // table + insecure_inbound: null, // table + insecure_outbound: null, // table + }; + }, + + activated: function() { + // see if props changed when deactive + if (this.date_range && this.date_range !== this.data_date_range) + this.getChartData(); + }, + + watch: { + // watch props for changes + 'date_range': function() { + this.getChartData(); + } + }, + + methods: { + link_to_user: function(user_id, tab) { + // add user=user_id to the user_link route + var r = Object.assign({}, this.user_link); + r.query = Object.assign({}, this.user_link.query); + r.query.user = user_id; + r.query.tab = tab; + return r; + }, + link_to_remote_sender_email: function(email) { + // add email=email to the remote_sender_email route + var r = Object.assign({}, this.remote_sender_email_link); + r.query = Object.assign({}, this.remote_sender_email_link.query); + r.query.email = email; + return r; + }, + link_to_remote_sender_server: function(server) { + // add server=server to the remote_sender_server route + var r = Object.assign({}, this.remote_sender_server_link); + r.query = Object.assign({}, this.remote_sender_server_link.query); + r.query.server = server; + return r; + }, + + getChartData: function() { + this.$emit('loading', 1); + axios.post('reports/uidata/flagged-connections', { + 'start': this.date_range[0], + 'end': this.date_range[1], + 'binsize': this.binsize, + }).then(response => { + this.data_date_range = this.date_range; + + // line charts + var ts = new TimeseriesData(response.data.flagged); + this.failed_logins = + ts.dataView(['failed_login_attempt']); + this.suspected_scanners = + ts.dataView(['suspected_scanner']); + + // pie chart for connections by disposition + this.connections_by_disposition = + response.data.connections_by_disposition; + + // pie chart for reject by failure_category + this.reject_by_failure_category = + response.data.reject_by_failure_category; + + // table of top 10 hosts rejected by failure_category + this.top_hosts_rejected = + new BvTable(response.data.top_hosts_rejected); + + // insecure connections tables + this.insecure_inbound + = new BvTable(response.data.insecure_inbound); + this.insecure_outbound + = new BvTable(response.data.insecure_outbound); + + }).catch(error => { + this.$root.handleError(error); + }).finally(() => { + this.$emit('loading', -1); + }); + }, + + } + + + })}).catch((e) => { + reject(e); + }); + +}); + diff --git a/management/reporting/ui/panel-messages-received.html b/management/reporting/ui/panel-messages-received.html new file mode 100644 index 00000000..7961063e --- /dev/null +++ b/management/reporting/ui/panel-messages-received.html @@ -0,0 +1,64 @@ +
+
+ + + + +
+ + + + + + + +
+ +
+ + + + + + + +
+ +
+
diff --git a/management/reporting/ui/panel-messages-received.js b/management/reporting/ui/panel-messages-received.js new file mode 100644 index 00000000..3be49f17 --- /dev/null +++ b/management/reporting/ui/panel-messages-received.js @@ -0,0 +1,116 @@ +/* + set of charts/tables showing messages received from internet servers + +*/ + +Vue.component('panel-messages-received', function(resolve, reject) { + axios.get('reports/ui/panel-messages-received.html').then((response) => { resolve({ + + template: response.data, + + props: { + date_range: Array, + binsize: Number, + user_link: Object, + remote_sender_email_link: Object, + remote_sender_server_link: Object, + width: { type:Number, default: ChartPrefs.default_width }, + height: { type:Number, default: ChartPrefs.default_height }, + }, + + components: { + 'chart-multi-line-timeseries': Vue.component('chart-multi-line-timeseries'), + // 'chart-stacked-bar-timeseries': Vue.component('chart-stacked-bar-timeseries'), + // 'chart-pie': Vue.component('chart-pie'), + 'chart-table': Vue.component('chart-table'), + }, + + data: function() { + return { + data_date_range: null, + data_received: null, + top_senders_by_count: null, + top_senders_by_size: null, + top_hosts_by_spam_score: null, + top_user_receiving_spam: null, + }; + }, + + computed: { + }, + + activated: function() { + // see if props changed when deactive + if (this.date_range && this.date_range !== this.data_date_range) + this.getChartData(); + }, + + watch: { + // watch props for changes + 'date_range': function() { + this.getChartData(); + } + }, + + methods: { + link_to_user: function(user_id) { + // add user=user_id to the user_link route + var r = Object.assign({}, this.user_link); + r.query = Object.assign({}, this.user_link.query); + r.query.user = user_id; + return r; + }, + link_to_remote_sender_email: function(email) { + // add email=email to the remote_sender_email route + var r = Object.assign({}, this.remote_sender_email_link); + r.query = Object.assign({}, this.remote_sender_email_link.query); + r.query.email = email; + return r; + }, + link_to_remote_sender_server: function(server) { + // add server=server to the remote_sender_server route + var r = Object.assign({}, this.remote_sender_server_link); + r.query = Object.assign({}, this.remote_sender_server_link.query); + r.query.server = server; + return r; + }, + + getChartData: function() { + this.$emit('loading', 1); + axios.post('reports/uidata/messages-received', { + 'start': this.date_range[0], + 'end': this.date_range[1], + 'binsize': this.binsize, + }).then(response => { + this.data_date_range = this.date_range; + var ts = new TimeseriesData(response.data.ts_received); + this.data_received = ts; + + [ 'top_senders_by_count', + 'top_senders_by_size', + 'top_hosts_by_spam_score', + 'top_user_receiving_spam' + ].forEach(item => { + this[item] = response.data[item]; + BvTable.setFieldDefinitions( + this[item].fields, + this[item].field_types + ); + }); + + }).catch(error => { + this.$root.handleError(error); + }).finally(() => { + this.$emit('loading', -1); + }); + }, + + } + + + })}).catch((e) => { + reject(e); + }); + +}); + diff --git a/management/reporting/ui/panel-messages-sent.html b/management/reporting/ui/panel-messages-sent.html new file mode 100644 index 00000000..5e2bbc6a --- /dev/null +++ b/management/reporting/ui/panel-messages-sent.html @@ -0,0 +1,50 @@ +
+
+ + + +
+ + + + + + + +
+ + + + + + +
+
diff --git a/management/reporting/ui/panel-messages-sent.js b/management/reporting/ui/panel-messages-sent.js new file mode 100644 index 00000000..f9d55f14 --- /dev/null +++ b/management/reporting/ui/panel-messages-sent.js @@ -0,0 +1,132 @@ +/* + set of charts/tables showing messages sent by local users + - number of messages sent over time + - delivered locally + - delivered remotely + - top senders + + emits: + 'loading' event=number +*/ + +Vue.component('panel-messages-sent', function(resolve, reject) { + axios.get('reports/ui/panel-messages-sent.html').then((response) => { resolve({ + + template: response.data, + + props: { + date_range: Array, // YYYY-MM-DD strings (UTC) + binsize: Number, // for timeseries charts, in minutes + // to enable clickable users, specify the route in + // user_link. the user_id will be added to the + // route.query as user=user_id. If set to 'true', the + // current route will be used. + user_link: Object, + width: { type:Number, default: ChartPrefs.default_width }, + height: { type:Number, default: ChartPrefs.default_height }, + }, + + components: { + 'chart-multi-line-timeseries': Vue.component('chart-multi-line-timeseries'), + 'chart-stacked-bar-timeseries': Vue.component('chart-stacked-bar-timeseries'), + 'chart-pie': Vue.component('chart-pie'), + 'chart-table': Vue.component('chart-table'), + }, + + data: function() { + return { + data_date_range: null, + data_sent: null, + data_recip: null, + data_recip_pie: null, + top_senders_by_count: null, + top_senders_by_size: null, + }; + }, + + computed: { + height_sent: function() { + return this.height / 2; + }, + + height_recip: function() { + return this.height / 2; + }, + + radius_recip_pie: function() { + return this.height /5; + }, + }, + + activated: function() { + // see if props changed when deactive + if (this.date_range && this.date_range !== this.data_date_range) + this.getChartData(); + }, + + watch: { + // watch props for changes + 'date_range': function() { + this.getChartData(); + } + }, + + methods: { + link_to_user: function(user_id) { + // add user=user_id to the user_link route + var r = Object.assign({}, this.user_link); + r.query = Object.assign({}, this.user_link.query); + r.query.user = user_id; + return r; + }, + + getChartData: function() { + this.$emit('loading', 1); + axios.post('reports/uidata/messages-sent', { + 'start': this.date_range[0], + 'end': this.date_range[1], + 'binsize': this.binsize, + }).then(response => { + this.data_date_range = this.date_range; + var ts = new TimeseriesData(response.data.ts_sent); + this.data_sent = ts.dataView(['sent']); + this.data_recip = ts.dataView(['local','remote']) + + this.data_recip_pie = [{ + name:'local', + value:d3.sum(ts.get_series('local').values) + }, { + name:'remote', + value:d3.sum(ts.get_series('remote').values) + }]; + + this.top_senders_by_count = + response.data.top_senders_by_count; + BvTable.setFieldDefinitions( + this.top_senders_by_count.fields, + this.top_senders_by_count.field_types + ); + + this.top_senders_by_size = + response.data.top_senders_by_size; + BvTable.setFieldDefinitions( + this.top_senders_by_size.fields, + this.top_senders_by_size.field_types + ); + + }).catch(error => { + this.$root.handleError(error); + }).finally(() => { + this.$emit('loading', -1); + }); + }, + + } + + + })}).catch((e) => { + reject(e); + }); + +}); + diff --git a/management/reporting/ui/panel-remote-sender-activity.html b/management/reporting/ui/panel-remote-sender-activity.html new file mode 100644 index 00000000..4877379a --- /dev/null +++ b/management/reporting/ui/panel-remote-sender-activity.html @@ -0,0 +1,74 @@ +
+ + + +
Too many results - the server returned only a limited set.
+ + +
+ + + + + + +
+ Email + Server +
+ + + + + Search + + + * Tables limited to {{ get_row_limit() }} rows + Flagged only +
+ + + + + + + + + + + +
diff --git a/management/reporting/ui/panel-remote-sender-activity.js b/management/reporting/ui/panel-remote-sender-activity.js new file mode 100644 index 00000000..8a8c32fc --- /dev/null +++ b/management/reporting/ui/panel-remote-sender-activity.js @@ -0,0 +1,247 @@ +/* + details on the activity of a remote sender (envelope from) +*/ + +Vue.component('panel-remote-sender-activity', function(resolve, reject) { + axios.get('reports/ui/panel-remote-sender-activity.html').then((response) => { resolve({ + + template: response.data, + + props: { + date_range: Array, // YYYY-MM-DD strings (UTC) + }, + + data: function() { + const usersetting_prefix = 'panel-rsa-'; + const sender_type = this.$route.query.email ? 'email' : + ( this.$route.query.server ? 'server' : 'email' ); + + return { + email: this.$route.query.email || '', /* v-model */ + server: this.$route.query.server || '', /* v-model */ + sender_type: sender_type, /* "email" or "server" only */ + + tab_index: 0, /* v-model */ + + show_only_flagged: false, + show_only_flagged_filter: null, + + data_sender: null, /* sender for active table data */ + data_sender_type: null, /* "email" or "server" */ + data_date_range: null, /* date range for active table data */ + + activity: null, /* table data */ + disposition_formatter: ConnectionDisposition.formatter, + + /* recent list */ + set_prefix: usersetting_prefix, + recent_senders: UserSettings.get() + .get_recent_list(usersetting_prefix + sender_type), + + /* suggestions (from server) */ + select_list: { suggestions: [] } + }; + }, + + activated: function() { + const new_email = this.$route.query.email; + const new_server = this.$route.query.server; + const new_sender_type = new_email ? 'email' : + ( new_server ? 'server' : null ); + + var load = false; + if (new_sender_type && + new_sender_type != this.sender_type) + { + this.sender_type = new_sender_type; + load = true; + } + if (this.sender_type == 'email' && + new_email && + new_email != this.email) + { + this.email = new_email; + this.getChartData(); + return; + } + if (this.sender_type == 'server' && + new_server && + new_server != this.server) + { + this.server = new_server; + this.getChartData(); + return; + } + + // see if props changed when deactive + if (load || this.date_range && + this.date_range !== this.data_date_range) + { + this.getChartData(); + } + else + { + // ensure the route query contains the sender + this.update_route(); + } + }, + + watch: { + // watch props for changes + 'date_range': function() { + this.getChartData(); + } + }, + + methods: { + update_recent_list: function() { + this.recent_senders = UserSettings.get() + .get_recent_list(this.set_prefix + this.sender_type); + }, + + update_route: function() { + // ensure the route contains query element + // "email=" or "server=" + // for the loaded data + if (this.data_sender && this.data_sender !== this.$route.query[this.sender_type]) { + var route = Object.assign({}, this.$route); + route.query = Object.assign({}, this.$route.query); + delete route.query.sender; + delete route.query.email; + route.query[this.sender_type] = this.data_sender; + this.$router.replace(route); + } + }, + + change_sender: function() { + axios.post('/reports/uidata/select-list-suggestions', { + type: this.sender_type == 'email' ? + 'envelope_from' : 'remote_host', + query: this.sender_type == 'email' ? + this.email.trim() : this.server.trim(), + start_date: this.date_range[0], + end_date: this.date_range[1] + }).then(response => { + if (response.data.exact) { + this.getChartData(); + } + else { + this.select_list = response.data; + this.$refs.suggest_modal.show() + } + }).catch(error => { + this.$root.handleError(error); + }); + }, + + choose_suggestion: function(suggestion) { + this[this.sender_type] = suggestion; + this.getChartData(); + this.$refs.suggest_modal.hide(); + }, + + combine_fields: function() { + // remove these fields... + this.activity + .combine_fields([ + 'sent_id', + 'sasl_username', + 'spam_score', + 'dkim_reason', + 'dmarc_reason', + 'postgrey_reason', + 'postgrey_delay', + 'category', + 'failure_info', + ]); + }, + + get_row_limit: function() { + return UserSettings.get().row_limit; + }, + + update_activity_rowVariant: function() { + // there is 1 row for each recipient of a message + // - give all rows of the same message the same + // color + this.activity.apply_rowVariant_grouping('info', (item, idx) => { + if (this.show_only_flagged && !item._flagged) + return null; + return item.sent_id; + }); + }, + + show_only_flagged_change: function() { + // 'change' event callback for checkbox + this.update_activity_rowVariant(); + // trigger BV to filter or not filter via + // reactive `show_only_flagged_filter` + this.show_only_flagged_filter= + (this.show_only_flagged ? 'yes' : null ); + }, + + table_filter_cb: function(item, filter) { + // when filter is non-null, this is called by BV for + // each row to determine whether it will be filtered + // (false) or included in the output (true) + return item._flagged; + }, + + getChartData: function() { + if (!this.date_range || !this[this.sender_type]) { + return; + } + + this.$emit('loading', 1); + axios.post('reports/uidata/remote-sender-activity', { + row_limit: this.get_row_limit(), + sender: this[this.sender_type].trim(), + sender_type: this.sender_type, + start_date: this.date_range[0], + end_date: this.date_range[1] + + }).then(response => { + this.data_sender = this[this.sender_type].trim(); + this.data_sender_type = this.sender_type; + this.data_date_range = this.date_range; + this.update_route(); + this.recent_senders = UserSettings.get() + .add_to_recent_list( + this.set_prefix + this.sender_type, + this[this.sender_type] + ); + this.show_only_flagged = false; + this.show_only_flagged_filter = null; + + /* setup table data */ + this.activity = + new MailBvTable(response.data.activity, { + _showDetails: true + }); + this.combine_fields(); + this.activity + .flag_fields() + .get_field('connect_time') + .add_tdClass('text-nowrap'); + this.update_activity_rowVariant(); + + }).catch(error => { + this.$root.handleError(error); + + }).finally(() => { + this.$emit('loading', -1); + }); + + }, + + row_clicked: function(item, index, event) { + item._showDetails = ! item._showDetails; + }, + + } + + })}).catch((e) => { + reject(e); + }); + +}); diff --git a/management/reporting/ui/panel-user-activity.html b/management/reporting/ui/panel-user-activity.html new file mode 100644 index 00000000..390b0933 --- /dev/null +++ b/management/reporting/ui/panel-user-activity.html @@ -0,0 +1,75 @@ +
+ + + + + + + + + + Change user + + + * Tables limited to {{ get_row_limit() }} rows + Flagged only + + + + + + + + + + + + + + + + + + + +
diff --git a/management/reporting/ui/panel-user-activity.js b/management/reporting/ui/panel-user-activity.js new file mode 100644 index 00000000..3f6c8dfb --- /dev/null +++ b/management/reporting/ui/panel-user-activity.js @@ -0,0 +1,260 @@ +/* + details on the activity of a user +*/ + +Vue.component('panel-user-activity', function(resolve, reject) { + axios.get('reports/ui/panel-user-activity.html').then((response) => { resolve({ + + template: response.data, + + props: { + date_range: Array, // YYYY-MM-DD strings (UTC) + }, + + components: { + 'wbr-text': Vue.component('wbr-text'), + }, + + data: function() { + var start_tab = this.$route.query.tab ? + Number(this.$route.query.tab) : + 0; + return { + user_id: this.$route.query.user || '', /* v-model */ + tab_index: start_tab, /* v-model */ + show_only_flagged: false, + show_only_flagged_filter: null, + data_user_id: null, /* user_id for active table data */ + data_date_range: null, /* date range for active table data */ + sent_mail: null, + received_mail: null, + all_users: [], + disposition_formatter: ConnectionDisposition.formatter, + }; + }, + + activated: function() { + const new_tab = Number(this.$route.query.tab); + const new_user = this.$route.query.user; + + if (new_user && new_user != this.user_id) { + this.user_id = new_user; + this.getChartData(isNaN(new_tab) ? 0 : new_tab); + return; + } + + // first time activated... + if (this.all_users.length == 0) + this.getChartData(new_tab); + + // see if props changed when deactive + else if (this.date_range && this.date_range !== this.data_date_range) + this.getChartData(new_tab); + else { + // ensure the route query contains "user=" + if (!isNaN(new_tab)) this.tab_index = new_tab; + this.update_route(); + } + + }, + + watch: { + // watch props for changes + 'date_range': function() { + this.getChartData(); + } + }, + + methods: { + update_route: function() { + // ensure the route contains query element + // "user=" for the loaded data + if (this.data_user_id && this.data_user_id !== this.$route.query.user) { + var route = Object.assign({}, this.$route); + route.query = Object.assign({}, this.$route.query); + route.query.user=this.data_user_id; + this.$router.replace(route); + } + }, + + change_user: function() { + this.getChartData(0); + }, + + combine_sent_mail_fields: function() { + // remove these fields... + this.sent_mail.combine_fields([ + 'sent_id', + 'spam_score', + 'delivery_info', + 'delivery_connection_info', + ]); + + // combine fields 'envelope_from' and 'rcpt_to' + this.sent_mail.combine_fields( + 'envelope_from', + 'rcpt_to', + (v, key, item) => { + if (item.envelope_from == this.data_user_id) + return v; + return `${v} (FROM: ${item.envelope_from})`; + }); + + // combine fields 'relay', 'delivery_connection' + this.sent_mail.combine_fields( + 'delivery_connection', + 'relay', + (v, key, item) => { + if (item.service == 'lmtp') + return ''; + var s = v.split('[', 1); + // remove the ip address + v = s[0]; + if (!item.delivery_connection || + item.delivery_connection == 'trusted' || + item.delivery_connection == 'verified') + { + return v; + } + return `${v}: ${item.delivery_connection}`; + }); + + }, + + combine_received_mail_fields: function() { + // remove these fields + this.received_mail.combine_fields([ + 'dkim_reason', + 'dmarc_reason', + 'postgrey_reason', + 'postgrey_delay', + 'spam_score' + + ]); + // combine fields 'envelope_from' and 'sasl_username' + var f = this.received_mail.combine_fields( + 'sasl_username', + 'envelope_from', + (v, key, item) => { + if (!item.sasl_username || item.envelope_from == item.sasl_username) + return v; + return `${v} (${item.sasl_username})`; + }); + f.label = 'Evelope From (user)'; + }, + + get_row_limit: function() { + return UserSettings.get().row_limit; + }, + + update_sent_mail_rowVariant: function() { + // there is 1 row for each recipient of a message + // - give all rows of the same message the same + // color + this.sent_mail.apply_rowVariant_grouping('info', item => { + if (this.show_only_flagged && !item._flagged) + return null; + return item.sent_id; + }); + }, + + show_only_flagged_change: function() { + // 'change' event callback for checkbox + this.update_sent_mail_rowVariant(); + // trigger BV to filter or not filter via + // reactive `show_only_flagged_filter` + this.show_only_flagged_filter= + (this.show_only_flagged ? 'yes' : null ); + }, + + + table_filter_cb: function(item, filter) { + // when filter is non-null, called by BV for each row + // to determine whether it will be filtered (false) or + // included in the output (true) + return item._flagged; + }, + + + getChartData: function(switch_to_tab) { + if (this.all_users.length == 0) { + this.$emit('loading', 1); + axios.get('reports/uidata/user-list').then(response => { + this.all_users = response.data; + + }).catch(error => { + this.$root.handleError(error); + + }).finally(() => { + this.$emit('loading', -1); + }); + } + + + if (!this.date_range || !this.user_id) { + return; + } + + this.$emit('loading', 1); + const promise = axios.post('reports/uidata/user-activity', { + row_limit: this.get_row_limit(), + user_id: this.user_id.trim(), + start_date: this.date_range[0], + end_date: this.date_range[1] + + }).then(response => { + + this.data_user_id = this.user_id.trim(); + this.data_date_range = this.date_range; + if (!isNaN(switch_to_tab)) + this.tab_index = switch_to_tab; + this.update_route(); + this.$emit('change', this.data_user_id); + this.show_only_flagged = false; + this.show_only_flagged_filter = null; + + /* setup sent_mail */ + this.sent_mail = new MailBvTable( + response.data.sent_mail, { + _showDetails: true + }); + this.combine_sent_mail_fields(); + this.sent_mail + .flag_fields() + .get_field('connect_time') + .add_tdClass('text-nowrap'); + this.update_sent_mail_rowVariant(); + + + /* setup received_mail */ + this.received_mail = new MailBvTable( + response.data.received_mail, { + _showDetails: true + }); + this.combine_received_mail_fields(); + this.received_mail + .flag_fields() + .get_field('connect_time') + .add_tdClass('text-nowrap'); + + }).catch(error => { + this.$root.handleError(error); + + }).finally(() => { + this.$emit('loading', -1); + }); + + return promise; + }, + + row_clicked: function(item, index, event) { + item._showDetails = ! item._showDetails; + }, + + } + + })}).catch((e) => { + reject(e); + }); + +}); diff --git a/management/reporting/ui/reports-page-header.js b/management/reporting/ui/reports-page-header.js new file mode 100644 index 00000000..538b6fc6 --- /dev/null +++ b/management/reporting/ui/reports-page-header.js @@ -0,0 +1,24 @@ +Vue.component('reports-page-header', { + props: { + loading_counter: { type:Number, required:true }, + }, + + components: { + 'page-header': Vue.component('page-header'), + }, + + template: + ''+ + ''+ + '' + , + +}); diff --git a/management/reporting/ui/settings.js b/management/reporting/ui/settings.js new file mode 100644 index 00000000..179dc4c4 --- /dev/null +++ b/management/reporting/ui/settings.js @@ -0,0 +1,94 @@ +window.miabldap = window.miabldap || {}; + +class CaptureConfig { + static get() { + return axios.get('/reports/capture/config').then(response => { + var cc = new CaptureConfig(); + Object.assign(cc, response.data); + return cc; + }); + } +}; + + +class UserSettings { + static load() { + if (window.miabldap.user_settings) { + return Promise.resolve(window.miabldap.user_settings); + } + + var s = new UserSettings(); + var json = localStorage.getItem('user_settings'); + if (json) { + s.data = JSON.parse(json); + } + else { + s.data = { + row_limit: 1000 + }; + } + window.miabldap.user_settings = s; + return Promise.resolve(s); + } + + static get() { + return window.miabldap.user_settings; + } + + save() { + var json = JSON.stringify(this.data); + localStorage.setItem('user_settings', json); + } + + _add_recent(list, val) { + var found = -1; + list.forEach((str, idx) => { + if (str.toLowerCase() == val.toLowerCase()) { + found = idx; + } + }); + if (found >= 0) { + // move it to the top + list.splice(found, 1); + } + list.unshift(val); + while (list.length > 10) list.pop(); + } + + + /* row limit */ + get row_limit() { + return this.data.row_limit; + } + + set row_limit(v) { + v = Number(v); + if (isNaN(v)) { + throw new ValueError("invalid") + } + else if (v < 5) { + throw new ValueError("minimum 5") + } + this.data.row_limit = v; + this.save(); + return v; + } + + get_recent_list(name) { + return this.data['recent_' + name]; + } + + add_to_recent_list(name, value) { + const dataname = 'recent_' + name; + var v = this.data[dataname]; + if (! v) { + this.data[dataname] = [ value ]; + this.save(); + return this.data[dataname]; + } + this._add_recent(v, value); + this.data[dataname] = v; + this.save(); + return v; + } +}; diff --git a/management/reporting/ui/wbr-text.js b/management/reporting/ui/wbr-text.js new file mode 100644 index 00000000..e661681a --- /dev/null +++ b/management/reporting/ui/wbr-text.js @@ -0,0 +1,48 @@ +/* + * This component adds elements after all characters given by + * `break_chars` in the given text. + * + * enables the browser to wrap long text at those points + * (without it the browser will only wrap at space and hyphen). + * + * Additionally, if `text_break_threshold` is greater than 0 and there + * is a segment of text that exceeds that length, the bootstrap css + * class "text-break" will be added to the , which causes the + * browser to wrap at any character of the text. + */ + +Vue.component('wbr-text', { + props: { + text: { type:String, required: true }, + break_chars: { type:String, default:'@_.,:+=' }, + text_break_threshold: { type:Number, default:0 }, + }, + + render: function(ce) { + var children = []; + var start=-1; + var idx=0; + var longest=-1; + while (idx < this.text.length) { + if (this.break_chars.indexOf(this.text[idx]) != -1) { + var sliver = this.text.substring(start+1, idx+1); + longest = Math.max(longest, sliver.length); + children.push(sliver); + children.push(ce('wbr')); + start=idx; + } + idx++; + } + + if (start < this.text.length-1) { + var sliver = this.text.substring(start+1); + longest = Math.max(longest, sliver.length); + children.push(sliver); + } + + var data = { }; + if (this.text_break_threshold>0 && longest>this.text_break_threshold) + data['class'] = { 'text-break': true }; + return ce('span', data, children); + } +}); diff --git a/management/reporting/uidata/DictCache.py b/management/reporting/uidata/DictCache.py new file mode 100644 index 00000000..0952792e --- /dev/null +++ b/management/reporting/uidata/DictCache.py @@ -0,0 +1,33 @@ +import datetime +import threading + +# +# thread-safe dict cache +# + +class DictCache(object): + def __init__(self, valid_for): + '''`valid_for` must be a datetime.timedelta object indicating how long + a cache item is valid + + ''' + self.obj = None + self.time = None + self.valid_for = valid_for + self.guard = threading.Lock() + + def get(self): + now = datetime.datetime.now() + with self.guard: + if self.obj and (now - self.time) <= self.valid_for: + return self.obj.copy() + + def set(self, obj): + with self.guard: + self.obj = obj.copy() + self.time = datetime.datetime.now() + + def reset(self): + with self.guard: + self.obj = None + self.time = None diff --git a/management/reporting/uidata/Timeseries.py b/management/reporting/uidata/Timeseries.py new file mode 100644 index 00000000..b6730f61 --- /dev/null +++ b/management/reporting/uidata/Timeseries.py @@ -0,0 +1,119 @@ +import datetime +import bisect + +class Timeseries(object): + def __init__(self, desc, start_date, end_date, binsize): + # start_date: 'YYYY-MM-DD' or 'YYYY-MM-DD HH:MM:SS' + # start: 'YYYY-MM-DD HH:MM:SS' + self.start = self.full_datetime_str(start_date, False) + + # end_date: 'YYYY-MM-DD' or 'YYYY-MM-DD HH:MM:SS' + # end: 'YYYY-MM-DD HH:MM:SS' + self.end = self.full_datetime_str(end_date, True) + + # binsize: integer in minutes + self.binsize = binsize + + # timefmt is a format string for sqlite strftime() that puts a + # sqlite datetime into a "bin" date + self.timefmt='%Y-%m-%d' + + # parsefmt is a date parser string to be used to re-interpret + # "bin" grouping dates (data.dates) to native dates + parsefmt='%Y-%m-%d' + + b = self.binsizeWithUnit() + + if b['unit'] == 'hour': + self.timefmt+=' %H:00:00' + parsefmt+=' %H:%M:%S' + elif b['unit'] == 'minute': + self.timefmt+=' %H:%M:00' + parsefmt+=' %H:%M:%S' + + self.dates = [] # dates must be "bin" date strings + self.series = [] + + self.data = { + 'range': [ self.start, self.end ], + 'range_parse_format': '%Y-%m-%d %H:%M:%S', + 'binsize': self.binsize, + 'date_parse_format': parsefmt, + 'y': desc, + 'dates': self.dates, + 'series': self.series + } + + def full_datetime_str(self, date_str, next_day): + if ':' in date_str: + return date_str + elif not next_day: + return date_str + " 00:00:00" + else: + d = datetime.datetime.strptime(date_str, '%Y-%m-%d') + d = d + datetime.timedelta(days=1) + return d.strftime('%Y-%m-%d 00:00:00') + + def binsizeWithUnit(self): + # normalize binsize (which is a time span in minutes) + days = int(self.binsize / (24 * 60)) + hours = int((self.binsize - days*24*60) / 60 ) + mins = self.binsize - days*24*60 - hours*60 + if days == 0 and hours == 0: + return { + 'unit': 'minute', + 'value': mins + } + + if days == 0: + return { + 'unit': 'hour', + 'value': hours + } + + return { + 'unit': 'day', + 'value': days + } + + + def append_date(self, date_str): + '''date_str should be a "bin" date - that is a date formatted with + self.timefmt. + + 1. it should be greater than the previous bin so that the date + list remains sorted + + 2. d3js does not require that all dates be added for a + timespan if there is no data for the bin + + ''' + self.dates.append(date_str) + + def insert_date(self, date_str): + '''adds bin date if it does not exist and returns the new index. if + the date already exists, returns the existing index. + + ''' + i = bisect.bisect_right(self.dates, date_str) + if i == len(self.dates): + self.dates.append(date_str) + return i + if self.dates[i] == date_str: + return i + self.dates.insert(i, date_str) + return i + + def add_series(self, id, name): + s = { + 'id': id, + 'name': name, + 'values': [] + } + self.series.append(s) + return s + + + def asDict(self): + return self.data + diff --git a/management/reporting/uidata/__init__.py b/management/reporting/uidata/__init__.py new file mode 100644 index 00000000..0afa0c46 --- /dev/null +++ b/management/reporting/uidata/__init__.py @@ -0,0 +1,9 @@ +from .exceptions import (InvalidArgsError) +from .select_list_suggestions import select_list_suggestions +from .messages_sent import messages_sent +from .messages_received import messages_received +from .user_activity import user_activity +from .remote_sender_activity import remote_sender_activity +from .flagged_connections import flagged_connections +from .capture_db_stats import capture_db_stats +from .capture_db_stats import clear_cache diff --git a/management/reporting/uidata/capture_db_stats.py b/management/reporting/uidata/capture_db_stats.py new file mode 100644 index 00000000..a4df3a0a --- /dev/null +++ b/management/reporting/uidata/capture_db_stats.py @@ -0,0 +1,50 @@ +import datetime +from .DictCache import DictCache + +# +# because of the table scan (select_2 below), cache stats for 5 +# minutes +# +last_stats = DictCache(datetime.timedelta(minutes=5)) + +def clear_cache(): + last_stats.reset() + + +def capture_db_stats(conn): + + stats = last_stats.get() + if stats: + return stats + + select_1 = 'SELECT min(connect_time) AS `min`, max(connect_time) AS `max`, count(*) AS `count` FROM mta_connection' + + # table scan + select_2 = 'SELECT disposition, count(*) AS `count` FROM mta_connection GROUP BY disposition' + + c = conn.cursor() + stats = { + # all times are in this format: "YYYY-MM-DD HH:MM:SS" (utc) + 'date_parse_format': '%Y-%m-%d %H:%M:%S' + } + try: + row = c.execute(select_1).fetchone() + stats['mta_connect'] = { + 'connect_time': { + 'min': row['min'], + 'max': row['max'], # YYYY-MM-DD HH:MM:SS (utc) + }, + 'count': row['count'], + 'disposition': {} + } + + for row in c.execute(select_2): + stats['mta_connect']['disposition'][row['disposition']] = { + 'count': row['count'] + } + + finally: + c.close() + + last_stats.set(stats) + return stats diff --git a/management/reporting/uidata/exceptions.py b/management/reporting/uidata/exceptions.py new file mode 100644 index 00000000..b2288348 --- /dev/null +++ b/management/reporting/uidata/exceptions.py @@ -0,0 +1,5 @@ +class MiabLdapError(Exception): + pass + +class InvalidArgsError(MiabLdapError): + pass diff --git a/management/reporting/uidata/flagged_connections.1.sql b/management/reporting/uidata/flagged_connections.1.sql new file mode 100644 index 00000000..398bbfb6 --- /dev/null +++ b/management/reporting/uidata/flagged_connections.1.sql @@ -0,0 +1,14 @@ +-- +-- returns count of failed_login_attempt in each 'bin', which is the +-- connection time rounded (as defined by {timefmt}) +-- +SELECT + strftime('{timefmt}',connect_time) AS `bin`, + count(*) AS `count` +FROM mta_connection +WHERE + disposition='failed_login_attempt' AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY strftime('{timefmt}',connect_time) +ORDER BY connect_time diff --git a/management/reporting/uidata/flagged_connections.2.sql b/management/reporting/uidata/flagged_connections.2.sql new file mode 100644 index 00000000..fb99a7f0 --- /dev/null +++ b/management/reporting/uidata/flagged_connections.2.sql @@ -0,0 +1,14 @@ +-- +-- returns count of suspected_scanner in each 'bin', which is the +-- connection time rounded (as defined by {timefmt}) +-- +SELECT + strftime('{timefmt}',connect_time) AS `bin`, + count(*) AS `count` +FROM mta_connection +WHERE + disposition='suspected_scanner' AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY strftime('{timefmt}',connect_time) +ORDER BY connect_time diff --git a/management/reporting/uidata/flagged_connections.3.sql b/management/reporting/uidata/flagged_connections.3.sql new file mode 100644 index 00000000..b8a6a185 --- /dev/null +++ b/management/reporting/uidata/flagged_connections.3.sql @@ -0,0 +1,8 @@ +SELECT failure_category, count(*) AS `count` +FROM mta_connection +JOIN mta_accept ON mta_accept.mta_conn_id = mta_connection.mta_conn_id +WHERE + disposition='reject' AND + connect_time >=:start_date AND + connect_time <:end_date +GROUP BY failure_category diff --git a/management/reporting/uidata/flagged_connections.4.sql b/management/reporting/uidata/flagged_connections.4.sql new file mode 100644 index 00000000..9f15ac7f --- /dev/null +++ b/management/reporting/uidata/flagged_connections.4.sql @@ -0,0 +1,14 @@ +-- +-- top 10 servers getting rejected by category +-- +SELECT CASE WHEN remote_host='unknown' THEN remote_ip ELSE remote_host END AS `remote_host`, mta_accept.failure_category AS `category`, count(*) AS `count` +FROM mta_connection +JOIN mta_accept ON mta_accept.mta_conn_id = mta_connection.mta_conn_id +WHERE + mta_connection.service='smtpd' AND + accept_status = 'reject' AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY CASE WHEN remote_host='unknown' THEN remote_ip ELSE remote_host END, mta_accept.failure_category +ORDER BY count(*) DESC +LIMIT 10 diff --git a/management/reporting/uidata/flagged_connections.5.sql b/management/reporting/uidata/flagged_connections.5.sql new file mode 100644 index 00000000..37761ffe --- /dev/null +++ b/management/reporting/uidata/flagged_connections.5.sql @@ -0,0 +1,12 @@ +-- +-- inbound mail using an insecure connection (no use of STARTTLS) +-- +SELECT mta_connection.service AS `service`, sasl_username, envelope_from, rcpt_to, count(*) AS `count` +FROM mta_connection +LEFT JOIN mta_accept ON mta_accept.mta_conn_id = mta_connection.mta_conn_id +LEFT JOIN mta_delivery ON mta_delivery.mta_accept_id = mta_accept.mta_accept_id +WHERE + disposition = 'insecure' AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY mta_connection.service, sasl_username, envelope_from, rcpt_to diff --git a/management/reporting/uidata/flagged_connections.6.sql b/management/reporting/uidata/flagged_connections.6.sql new file mode 100644 index 00000000..3da22b47 --- /dev/null +++ b/management/reporting/uidata/flagged_connections.6.sql @@ -0,0 +1,12 @@ +-- +-- outbound mail using an insecure connection (low grade encryption) +-- +SELECT mta_delivery.service AS `service`, sasl_username, envelope_from, rcpt_to, count(*) AS `count` +FROM mta_connection +LEFT JOIN mta_accept ON mta_accept.mta_conn_id = mta_connection.mta_conn_id +LEFT JOIN mta_delivery ON mta_delivery.mta_accept_id = mta_accept.mta_accept_id +WHERE + delivery_connection = 'untrusted' AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY mta_connection.service, sasl_username, envelope_from, rcpt_to diff --git a/management/reporting/uidata/flagged_connections.py b/management/reporting/uidata/flagged_connections.py new file mode 100644 index 00000000..240952f9 --- /dev/null +++ b/management/reporting/uidata/flagged_connections.py @@ -0,0 +1,142 @@ +from .Timeseries import Timeseries +from .exceptions import InvalidArgsError +from .top import select_top + +with open(__file__.replace('.py','.1.sql')) as fp: + select_1 = fp.read() + +with open(__file__.replace('.py','.2.sql')) as fp: + select_2 = fp.read() + +with open(__file__.replace('.py','.3.sql')) as fp: + select_3 = fp.read() + +with open(__file__.replace('.py','.4.sql')) as fp: + select_4 = fp.read() + +with open(__file__.replace('.py','.5.sql')) as fp: + select_5 = fp.read() + +with open(__file__.replace('.py','.6.sql')) as fp: + select_6 = fp.read() + + +def flagged_connections(conn, args): + try: + ts = Timeseries( + "Failed login attempts and suspected scanners over time", + args['start'], + args['end'], + args['binsize'] + ) + except KeyError: + raise InvalidArgsError() + + c = conn.cursor() + + # pie chart for "connections by disposition" + select = 'SELECT disposition, count(*) AS `count` FROM mta_connection WHERE connect_time>=:start_date AND connect_time<:end_date GROUP BY disposition' + connections_by_disposition = [] + for row in c.execute(select, {'start_date':ts.start, 'end_date':ts.end}): + connections_by_disposition.append({ + 'name': row[0], + 'value': row[1] + }) + + # timeseries = failed logins count + s_failed_login = ts.add_series('failed_login_attempt', 'failed login attempts') + for row in c.execute(select_1.format(timefmt=ts.timefmt), { + 'start_date': ts.start, + 'end_date': ts.end + }): + ts.append_date(row['bin']) + s_failed_login['values'].append(row['count']) + + # timeseries = suspected scanners count + s_scanner = ts.add_series('suspected_scanner', 'connections by suspected scanners') + for row in c.execute(select_2.format(timefmt=ts.timefmt), { + 'start_date': ts.start, + 'end_date': ts.end + }): + ts.insert_date(row['bin']) + s_scanner['values'].append(row['count']) + + + # pie chart for "disposition=='reject' grouped by failure_category" + reject_by_failure_category = [] + for row in c.execute(select_3, { + 'start_date': ts.start, + 'end_date': ts.end + }): + reject_by_failure_category.append({ + 'name': row[0], + 'value': row[1] + }) + + # top 10 servers rejected by category + top_hosts_rejected = select_top( + c, + select_4, + ts.start, + ts.end, + "Top servers rejected by category", + [ 'remote_host', 'category', 'count' ], + [ 'text/hostname', 'text/plain', 'number/plain' ] + ) + + # insecure inbound connections - no limit + insecure_inbound = select_top( + c, + select_5, + ts.start, + ts.end, + "Insecure inbound connections (no use of STARTTLS)", + [ + 'service', + 'sasl_username', + 'envelope_from', + 'rcpt_to', + 'count' + ], + [ + 'text/plain', # service + 'text/plain', # sasl_username + 'text/email', # envelope_from + { 'type':'text/email', 'label':'Recipient' }, # rcpt_to + 'number/plain', # count + ] + ) + + # insecure outbound connections - no limit + insecure_outbound = select_top( + c, + select_6, + ts.start, + ts.end, + "Insecure outbound connections (low grade encryption)", + [ + 'service', + 'sasl_username', + 'envelope_from', + 'rcpt_to', + 'count' + ], + [ + 'text/plain', # service + 'text/plain', # sasl_username + 'text/email', # envelope_from + { 'type':'text/email', 'label':'Recipient' }, # rcpt_to + 'number/plain', # count + ] + ) + + + + return { + 'connections_by_disposition': connections_by_disposition, + 'flagged': ts.asDict(), + 'reject_by_failure_category': reject_by_failure_category, + 'top_hosts_rejected': top_hosts_rejected, + 'insecure_inbound': insecure_inbound, + 'insecure_outbound': insecure_outbound, + } diff --git a/management/reporting/uidata/messages_received.1.sql b/management/reporting/uidata/messages_received.1.sql new file mode 100644 index 00000000..382158a8 --- /dev/null +++ b/management/reporting/uidata/messages_received.1.sql @@ -0,0 +1,15 @@ +-- +-- returns count of messages received by smtpd in each 'bin', which is +-- the connection time rounded (as defined by {timefmt}) +-- +SELECT + strftime('{timefmt}',connect_time) AS `bin`, + count(*) AS `count` +FROM mta_accept +JOIN mta_connection ON mta_connection.mta_conn_id = mta_accept.mta_conn_id +WHERE + mta_connection.service = 'smtpd' AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY strftime('{timefmt}',connect_time) +ORDER BY connect_time diff --git a/management/reporting/uidata/messages_received.2.sql b/management/reporting/uidata/messages_received.2.sql new file mode 100644 index 00000000..8149afaa --- /dev/null +++ b/management/reporting/uidata/messages_received.2.sql @@ -0,0 +1,14 @@ +-- +-- top 10 senders (envelope_from) by message count +-- +SELECT count(mta_accept_id) AS `count`, envelope_from AS `email` +FROM mta_connection +JOIN mta_accept on mta_accept.mta_conn_id = mta_connection.mta_conn_id +WHERE + mta_connection.service = "smtpd" AND + accept_status != 'reject' AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY envelope_from +ORDER BY count(mta_accept_id) DESC +LIMIT 10 diff --git a/management/reporting/uidata/messages_received.3.sql b/management/reporting/uidata/messages_received.3.sql new file mode 100644 index 00000000..834c88c1 --- /dev/null +++ b/management/reporting/uidata/messages_received.3.sql @@ -0,0 +1,12 @@ +-- +-- top 10 senders (envelope_from) by message size +-- +SELECT sum(message_size) AS `size`, envelope_from AS `email` +FROM mta_connection +JOIN mta_accept on mta_accept.mta_conn_id = mta_connection.mta_conn_id +WHERE mta_connection.service = "smtpd" AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY envelope_from +ORDER BY sum(message_size) DESC +LIMIT 10 diff --git a/management/reporting/uidata/messages_received.4.sql b/management/reporting/uidata/messages_received.4.sql new file mode 100644 index 00000000..96c0621f --- /dev/null +++ b/management/reporting/uidata/messages_received.4.sql @@ -0,0 +1,13 @@ +-- +-- top 10 remote servers/domains (remote hosts) by average spam score +-- +SELECT CASE WHEN remote_host='unknown' THEN remote_ip ELSE remote_host END AS `remote_host`, avg(spam_score) AS avg_spam_score FROM mta_connection +JOIN mta_accept ON mta_accept.mta_conn_id = mta_connection.mta_conn_id +JOIN mta_delivery ON mta_accept.mta_accept_id = mta_delivery.mta_accept_id +WHERE mta_connection.service='smtpd' AND + spam_score IS NOT NULL AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY CASE WHEN remote_host='unknown' THEN remote_ip ELSE remote_host END +ORDER BY avg(spam_score) DESC +LIMIT 10 diff --git a/management/reporting/uidata/messages_received.5.sql b/management/reporting/uidata/messages_received.5.sql new file mode 100644 index 00000000..e069af33 --- /dev/null +++ b/management/reporting/uidata/messages_received.5.sql @@ -0,0 +1,12 @@ +-- +-- top 10 users receiving the most spam +-- +SELECT rcpt_to, count(*) AS count FROM mta_delivery +JOIN mta_accept ON mta_accept.mta_accept_id = mta_delivery.mta_accept_id +JOIN mta_connection ON mta_accept.mta_conn_id = mta_connection.mta_conn_id +WHERE spam_result='spam' AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY rcpt_to +ORDER BY count(*) DESC +LIMIT 10 diff --git a/management/reporting/uidata/messages_received.py b/management/reporting/uidata/messages_received.py new file mode 100644 index 00000000..84d68d37 --- /dev/null +++ b/management/reporting/uidata/messages_received.py @@ -0,0 +1,108 @@ +from .Timeseries import Timeseries +from .exceptions import InvalidArgsError +from .top import select_top + +with open(__file__.replace('.py','.1.sql')) as fp: + select_1 = fp.read() + +with open(__file__.replace('.py','.2.sql')) as fp: + select_2 = fp.read() + +with open(__file__.replace('.py','.3.sql')) as fp: + select_3 = fp.read() + +with open(__file__.replace('.py','.4.sql')) as fp: + select_4 = fp.read() + +with open(__file__.replace('.py','.5.sql')) as fp: + select_5 = fp.read() + + + +def messages_received(conn, args): + ''' + messages recived from the internet + + ''' + try: + ts = Timeseries( + "Messages received from the internet", + args['start'], + args['end'], + args['binsize'] + ) + except KeyError: + raise InvalidArgsError() + + s_received = ts.add_series('received', 'messages received') + + c = conn.cursor() + try: + for row in c.execute(select_1.format(timefmt=ts.timefmt), { + 'start_date':ts.start, + 'end_date':ts.end + }): + ts.append_date(row['bin']) + s_received['values'].append(row['count']) + + + # top 10 senders (envelope_from) by message count + top_senders_by_count = select_top( + c, + select_2, + ts.start, + ts.end, + "Top 10 senders by count", + [ 'email', 'count' ], + [ 'text/email', 'number/plain' ] + ) + + # top 10 senders (envelope_from) by message size + top_senders_by_size = select_top( + c, + select_3, + ts.start, + ts.end, + "Top 10 senders by size", + [ 'email', 'size' ], + [ 'text/email', 'number/size' ] + ) + + # top 10 remote servers/domains (remote hosts) by average spam score + top_hosts_by_spam_score = select_top( + c, + select_4, + ts.start, + ts.end, + "Top servers by average spam score", + [ 'remote_host', 'avg_spam_score' ], + [ 'text/hostname', { 'type':'decimal', 'places':2} ] + ) + + # top 10 users receiving the most spam + top_user_receiving_spam = select_top( + c, + select_5, + ts.start, + ts.end, + "Top 10 users receiving spam", + [ + 'rcpt_to', + 'count' + ], + [ + { 'type': 'text', 'subtype':'email', 'label':'User' }, + 'number/plain' + ] + ) + + finally: + c.close() + + return { + 'top_senders_by_count': top_senders_by_count, + 'top_senders_by_size': top_senders_by_size, + 'top_hosts_by_spam_score': top_hosts_by_spam_score, + 'top_user_receiving_spam': top_user_receiving_spam, + 'ts_received': ts.asDict(), + } diff --git a/management/reporting/uidata/messages_sent.1.sql b/management/reporting/uidata/messages_sent.1.sql new file mode 100644 index 00000000..eb11fc20 --- /dev/null +++ b/management/reporting/uidata/messages_sent.1.sql @@ -0,0 +1,16 @@ +-- +-- returns count of sent messages in each 'bin', which is the connection +-- time rounded (as defined by {timefmt}) +-- +SELECT + strftime('{timefmt}',connect_time) AS `bin`, + count(*) AS `sent_count` +FROM mta_accept +JOIN mta_connection ON mta_connection.mta_conn_id = mta_accept.mta_conn_id +JOIN mta_delivery ON mta_delivery.mta_accept_id = mta_accept.mta_accept_id +WHERE + (mta_connection.service = 'submission' OR mta_connection.service = 'pickup') AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY strftime('{timefmt}',connect_time) +ORDER BY connect_time diff --git a/management/reporting/uidata/messages_sent.2.sql b/management/reporting/uidata/messages_sent.2.sql new file mode 100644 index 00000000..de7b6584 --- /dev/null +++ b/management/reporting/uidata/messages_sent.2.sql @@ -0,0 +1,18 @@ +-- +-- returns count of sent messages delivered in each 'bin'/delivery +-- service combination. the bin is the connection time rounded (as +-- defined by {timefmt}) +-- +SELECT + strftime('{timefmt}',connect_time) AS `bin`, + mta_delivery.service AS `delivery_service`, + count(*) AS `delivery_count` +FROM mta_accept +JOIN mta_connection ON mta_connection.mta_conn_id = mta_accept.mta_conn_id +JOIN mta_delivery ON mta_delivery.mta_accept_id = mta_accept.mta_accept_id +WHERE + (mta_connection.service = 'submission' OR mta_connection.service = 'pickup') AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY strftime('{timefmt}',connect_time), mta_delivery.service +ORDER BY connect_time diff --git a/management/reporting/uidata/messages_sent.3.sql b/management/reporting/uidata/messages_sent.3.sql new file mode 100644 index 00000000..b1acda3a --- /dev/null +++ b/management/reporting/uidata/messages_sent.3.sql @@ -0,0 +1,12 @@ +-- +-- top 10 senders by message count +-- +select count(mta_accept_id) as count, sasl_username as username +from mta_connection +join mta_accept on mta_accept.mta_conn_id = mta_connection.mta_conn_id +where mta_connection.service = "submission" AND + connect_time >= :start_date AND + connect_time < :end_date +group by sasl_username +order by count(mta_accept_id) DESC +limit 10 diff --git a/management/reporting/uidata/messages_sent.4.sql b/management/reporting/uidata/messages_sent.4.sql new file mode 100644 index 00000000..44e98d38 --- /dev/null +++ b/management/reporting/uidata/messages_sent.4.sql @@ -0,0 +1,12 @@ +-- +-- top 10 senders by message size +-- +SELECT sum(message_size) AS message_size_total, sasl_username AS username +FROM mta_connection +JOIN mta_accept ON mta_accept.mta_conn_id = mta_connection.mta_conn_id +WHERE mta_connection.service = "submission" AND + connect_time >= :start_date AND + connect_time < :end_date +GROUP BY sasl_username +ORDER BY sum(message_size) DESC +LIMIT 10 diff --git a/management/reporting/uidata/messages_sent.py b/management/reporting/uidata/messages_sent.py new file mode 100644 index 00000000..0892b1a8 --- /dev/null +++ b/management/reporting/uidata/messages_sent.py @@ -0,0 +1,113 @@ +from .Timeseries import Timeseries +from .exceptions import InvalidArgsError + +with open(__file__.replace('.py','.1.sql')) as fp: + select_1 = fp.read() + +with open(__file__.replace('.py','.2.sql')) as fp: + select_2 = fp.read() + +with open(__file__.replace('.py','.3.sql')) as fp: + select_3 = fp.read() + +with open(__file__.replace('.py','.4.sql')) as fp: + select_4 = fp.read() + + +def messages_sent(conn, args): + ''' + messages sent by local users + - delivered locally & remotely + + ''' + try: + ts = Timeseries( + "Messages sent by users", + args['start'], + args['end'], + args['binsize'] + ) + except KeyError: + raise InvalidArgsError() + + s_sent = ts.add_series('sent', 'messages sent') + s_local = ts.add_series('local', 'local recipients') + s_remote = ts.add_series('remote', 'remote recipients') + + c = conn.cursor() + try: + for row in c.execute(select_1.format(timefmt=ts.timefmt), { + 'start_date':ts.start, + 'end_date':ts.end + }): + ts.dates.append(row['bin']) + s_sent['values'].append(row['sent_count']) + + date_idx = -1 + + # the returned bins are the same as select_1 because the + # querie's WHERE and JOINs are the same + for row in c.execute(select_2.format(timefmt=ts.timefmt), { + 'start_date':ts.start, + 'end_date':ts.end + }): + if date_idx>=0 and ts.dates[date_idx] == row['bin']: + if row['delivery_service']=='smtp': + s_remote['values'][-1] = row['delivery_count'] + elif row['delivery_service']=='lmtp': + s_local['values'][-1] = row['delivery_count'] + + else: + date_idx += 1 + if date_idx >= len(ts.dates): + break + if row['delivery_service']=='smtp': + s_remote['values'].append(row['delivery_count']) + s_local['values'].append(0) + elif row['delivery_service']=='lmtp': + s_remote['values'].append(0) + s_local['values'].append(row['delivery_count']) + + + top_senders1 = { + 'start': ts.start, + 'end': ts.end, + 'y': 'Top 10 users by count', + 'fields': ['user','count'], + 'field_types': ['text/email','number/plain'], + 'items': [] + } + for row in c.execute(select_3, { + 'start_date':ts.start, + 'end_date':ts.end + }): + top_senders1['items'].append({ + 'user': row['username'], + 'count': row['count'] + }) + + top_senders2 = { + 'start': ts.start, + 'end': ts.end, + 'y': 'Top 10 users by size', + 'fields': ['user','size'], + 'field_types': ['text/email','number/size'], + 'items': [] + } + for row in c.execute(select_4, { + 'start_date':ts.start, + 'end_date':ts.end + }): + top_senders2['items'].append({ + 'user': row['username'], + 'size': row['message_size_total'] + }) + + finally: + c.close() + + return { + 'top_senders_by_count': top_senders1, + 'top_senders_by_size': top_senders2, + 'ts_sent': ts.asDict(), + } diff --git a/management/reporting/uidata/remote_sender_activity.1.sql b/management/reporting/uidata/remote_sender_activity.1.sql new file mode 100644 index 00000000..94c8e448 --- /dev/null +++ b/management/reporting/uidata/remote_sender_activity.1.sql @@ -0,0 +1,19 @@ +-- +-- details on remote senders +-- query: envelope_from +-- +SELECT +-- mta_connection +connect_time, mta_connection.service AS `service`, sasl_username, disposition, +-- mta_accept +mta_accept.mta_accept_id AS mta_accept_id, spf_result, dkim_result, dkim_reason, dmarc_result, dmarc_reason, accept_status, failure_info, mta_accept.failure_category AS `category`, +-- mta_delivery +rcpt_to, postgrey_result, postgrey_reason, postgrey_delay, spam_score, spam_result, message_size +FROM mta_connection +JOIN mta_accept ON mta_accept.mta_conn_id = mta_connection.mta_conn_id +LEFT JOIN mta_delivery ON mta_accept.mta_accept_id = mta_delivery.mta_accept_id +WHERE + envelope_from = :envelope_from AND + connect_time >= :start_date AND + connect_time < :end_date +ORDER BY connect_time diff --git a/management/reporting/uidata/remote_sender_activity.2.sql b/management/reporting/uidata/remote_sender_activity.2.sql new file mode 100644 index 00000000..800ecae7 --- /dev/null +++ b/management/reporting/uidata/remote_sender_activity.2.sql @@ -0,0 +1,20 @@ +-- +-- details on remote sender host +-- query: remote_host or remote_ip +-- +SELECT +-- mta_connection +connect_time, disposition, +-- mta_accept +mta_accept.mta_accept_id AS mta_accept_id, spf_result, dkim_result, dkim_reason, dmarc_result, dmarc_reason, accept_status, failure_info, mta_accept.failure_category AS `category`, envelope_from, +-- mta_delivery +rcpt_to, postgrey_result, postgrey_reason, postgrey_delay, spam_score, spam_result, message_size +FROM mta_connection +LEFT JOIN mta_accept ON mta_accept.mta_conn_id = mta_connection.mta_conn_id +LEFT JOIN mta_delivery ON mta_accept.mta_accept_id = mta_delivery.mta_accept_id +WHERE + (remote_host = :remote_host OR remote_ip = :remote_host) AND + mta_connection.service = 'smtpd' AND + connect_time >= :start_date AND + connect_time < :end_date +ORDER BY connect_time diff --git a/management/reporting/uidata/remote_sender_activity.py b/management/reporting/uidata/remote_sender_activity.py new file mode 100644 index 00000000..f5f061ab --- /dev/null +++ b/management/reporting/uidata/remote_sender_activity.py @@ -0,0 +1,180 @@ +from .Timeseries import Timeseries +from .exceptions import InvalidArgsError + + +with open(__file__.replace('.py','.1.sql')) as fp: + select_1 = fp.read() + +with open(__file__.replace('.py','.2.sql')) as fp: + select_2 = fp.read() + + +def remote_sender_activity(conn, args): + ''' + details on remote senders (envelope from) + ''' + + try: + sender = args['sender'] + sender_type = args['sender_type'] + + if sender_type not in ['email', 'server']: + raise InvalidArgsError() + + # use Timeseries to get a normalized start/end range + ts = Timeseries( + 'Remote sender activity', + args['start_date'], + args['end_date'], + 0 + ) + except KeyError: + raise InvalidArgsError() + + # limit results + try: + limit = 'LIMIT ' + str(int(args.get('row_limit', 1000))); + except ValueError: + limit = 'LIMIT 1000' + + c = conn.cursor() + + if sender_type == 'email': + select = select_1 + fields = [ + # mta_connection + 'connect_time', + 'service', + 'sasl_username', + # mta_delivery + 'rcpt_to', + # mta_accept + 'disposition', + 'accept_status', + 'spf_result', + 'dkim_result', + 'dkim_reason', + 'dmarc_result', + 'dmarc_reason', + 'failure_info', + 'category', # failure_category + # mta_delivery + 'postgrey_result', + 'postgrey_reason', + 'postgrey_delay', + 'spam_score', + 'spam_result', + 'message_size', + 'sent_id', # must be last + ] + field_types = [ + { 'type':'datetime', 'format': '%Y-%m-%d %H:%M:%S' },# connect_time + 'text/plain', # service + 'text/plain', # sasl_username + { 'type':'text/email', 'label':'Recipient' }, # rcpt_to + 'text/plain', # disposition + 'text/plain', # accept_status + 'text/plain', # spf_result + 'text/plain', # dkim_result + 'text/plain', # dkim_reason + 'text/plain', # dmarc_result + 'text/plain', # dmarc_reason + 'text/plain', # failure_info + 'text/plain', # category (mta_accept.failure_category) + 'text/plain', # postgrey_result + 'text/plain', # postgrey_reason + { 'type':'time/span', 'unit':'s' }, # postgrey_delay + { 'type':'decimal', 'places':2 }, # spam_score + 'text/plain', # spam_result + 'number/size', # message_size + 'number/plain', # sent_id - must be last + ] + select_args = { + 'envelope_from': sender, + 'start_date': ts.start, + 'end_date': ts.end + } + + elif sender_type == 'server': + select = select_2 + fields = [ + # mta_connection + 'connect_time', + # mta_accept + 'envelope_from', + # mta_delivery + 'rcpt_to', + 'disposition', + # mta_accept + 'accept_status', + 'spf_result', + 'dkim_result', + 'dkim_reason', + 'dmarc_result', + 'dmarc_reason', + 'failure_info', + 'category', # failure_category + # mta_delivery + 'postgrey_result', + 'postgrey_reason', + 'postgrey_delay', + 'spam_score', + 'spam_result', + 'message_size', + 'sent_id', # must be last + ] + field_types = [ + { 'type':'datetime', 'format': '%Y-%m-%d %H:%M:%S' },# connect_time + { 'type':'text/email', 'label':'From' }, # envelope_from + { 'type':'text/email', 'label':'Recipient' }, # rcpt_to + 'text/plain', # disposition + 'text/plain', # accept_status + 'text/plain', # spf_result + 'text/plain', # dkim_result + 'text/plain', # dkim_reason + 'text/plain', # dmarc_result + 'text/plain', # dmarc_reason + 'text/plain', # failure_info + 'text/plain', # category (mta_accept.failure_category) + 'text/plain', # postgrey_result + 'text/plain', # postgrey_reason + { 'type':'time/span', 'unit':'s' }, # postgrey_delay + { 'type':'decimal', 'places':2 }, # spam_score + 'text/plain', # spam_result + 'number/size', # message_size + 'number/plain', # sent_id - must be last + ] + select_args = { + 'remote_host': sender, + 'start_date': ts.start, + 'end_date': ts.end + } + + + activity = { + 'start': ts.start, + 'end': ts.end, + 'y': 'Remote sender activity', + 'fields': fields, + 'field_types': field_types, + 'items': [], + 'unique_sends': 0 + } + last_mta_accept_id = -1 + sent_id = 0 + for row in c.execute(select + limit, select_args): + v = [] + for key in activity['fields']: + if key != 'sent_id': + v.append(row[key]) + if row['mta_accept_id'] is None or last_mta_accept_id != row['mta_accept_id']: + activity['unique_sends'] += 1 + last_mta_accept_id = row['mta_accept_id'] + sent_id += 1 + v.append(sent_id) + activity['items'].append(v) + + + return { + 'activity': activity, + } diff --git a/management/reporting/uidata/select_list_suggestions.py b/management/reporting/uidata/select_list_suggestions.py new file mode 100644 index 00000000..3b53a099 --- /dev/null +++ b/management/reporting/uidata/select_list_suggestions.py @@ -0,0 +1,115 @@ +from .Timeseries import Timeseries +from .exceptions import InvalidArgsError +import logging + +log = logging.getLogger(__name__) + + +def select_list_suggestions(conn, args): + + try: + query_type = args['type'] + query = args['query'].strip() + ts = None + if 'start_date' in args: + # use Timeseries to get a normalized start/end range + ts = Timeseries( + 'select list suggestions', + args['start_date'], + args['end_date'], + 0 + ) + except KeyError: + raise InvalidArgsError() + + # escape query with backslash for fuzzy match (LIKE) + query_escaped = query.replace("\\", "\\\\").replace("%","\\%").replace("_","\\_") + limit = 100 + + queries = { + 'remote_host': { + 'select': "DISTINCT CASE WHEN remote_host='unknown' THEN remote_ip ELSE remote_host END", + 'from': "mta_connection", + 'join': {}, + 'order_by': "remote_host", + 'where_exact': [ "(remote_host = ? OR remote_ip = ?)" ], + 'args_exact': [ query, query ], + 'where_fuzzy': [ "(remote_host LIKE ? ESCAPE '\\' OR remote_ip LIKE ? ESCAPE '\\')" ], + 'args_fuzzy': [ '%'+query_escaped+'%', query_escaped+'%' ] + }, + 'rcpt_to': { + 'select': "DISTINCT rcpt_to", + 'from': 'mta_delivery', + 'join': {}, + 'order_by': "rcpt_to", + 'where_exact': [ "rcpt_to = ?" ], + 'args_exact': [ query, ], + 'where_fuzzy': [ "rcpt_to LIKE ? ESCAPE '\\'" ], + 'args_fuzzy': [ '%'+query_escaped+'%' ] + }, + 'envelope_from': { + 'select': "DISTINCT envelope_from", + 'from': "mta_accept", + 'join': {}, + 'order_by': 'envelope_from', + 'where_exact': [ "envelope_from = ?" ], + 'args_exact': [ query, ], + 'where_fuzzy': [ "envelope_from LIKE ? ESCAPE '\\'" ], + 'args_fuzzy': [ '%'+query_escaped+'%' ] + }, + } + + q = queries.get(query_type) + if not q: + raise InvalidArgError() + + if ts: + q['where_exact'] += [ 'connect_time>=?', 'connect_time=?', 'connect_time=limit + } + + finally: + c.close() + diff --git a/management/reporting/uidata/top.py b/management/reporting/uidata/top.py new file mode 100644 index 00000000..a3cab34b --- /dev/null +++ b/management/reporting/uidata/top.py @@ -0,0 +1,32 @@ +def select_top(c, select, start, end, y, fields, field_types): + '''`c` is a cursor + + `select` is the select query `start` and `end` are the range in +the format YYYY-MM-DD HH:MM:SS and the select query must have +substitutes 'start_date' and 'end_date'. + + `y` is a description of the dataset + + `fields` are all fields to select by name + + `field_types` are the corresponding field types the caller will + need to render the data visuals + ''' + + top = { + 'start': start, + 'end': end, + 'y': y, + 'fields': fields, + 'field_types': field_types, + 'items': [] + } + for row in c.execute(select, { + 'start_date':start, + 'end_date':end + }): + v = {} + for key in fields: + v[key] = row[key] + top['items'].append(v) + return top diff --git a/management/reporting/uidata/user_activity.1.sql b/management/reporting/uidata/user_activity.1.sql new file mode 100644 index 00000000..66fe7e97 --- /dev/null +++ b/management/reporting/uidata/user_activity.1.sql @@ -0,0 +1,17 @@ +-- +-- details on user sent mail +-- +SELECT +-- mta_connection +connect_time, sasl_method, +-- mta_accept +mta_accept.mta_accept_id AS mta_accept_id, envelope_from, +-- mta_delivery +mta_delivery.service AS service, rcpt_to, spam_score, spam_result, message_size, status, relay, delivery_info, delivery_connection, delivery_connection_info +FROM mta_accept +JOIN mta_connection ON mta_accept.mta_conn_id = mta_connection.mta_conn_id +JOIN mta_delivery ON mta_accept.mta_accept_id = mta_delivery.mta_accept_id +WHERE sasl_username = :user_id AND + connect_time >= :start_date AND + connect_time < :end_date +ORDER BY connect_time, mta_accept.mta_accept_id diff --git a/management/reporting/uidata/user_activity.2.sql b/management/reporting/uidata/user_activity.2.sql new file mode 100644 index 00000000..f1d05877 --- /dev/null +++ b/management/reporting/uidata/user_activity.2.sql @@ -0,0 +1,17 @@ +-- +-- details on user received mail +-- +SELECT +-- mta_connection +connect_time, mta_connection.service AS service, sasl_username, disposition, +-- mta_accept +envelope_from, spf_result, dkim_result, dkim_reason, dmarc_result, dmarc_reason, +-- mta_delivery +postgrey_result, postgrey_reason, postgrey_delay, spam_score, spam_result, message_size +FROM mta_accept +JOIN mta_connection ON mta_accept.mta_conn_id = mta_connection.mta_conn_id +JOIN mta_delivery ON mta_accept.mta_accept_id = mta_delivery.mta_accept_id +WHERE rcpt_to = :user_id AND + connect_time >= :start_date AND + connect_time < :end_date +ORDER BY connect_time diff --git a/management/reporting/uidata/user_activity.py b/management/reporting/uidata/user_activity.py new file mode 100644 index 00000000..091ef186 --- /dev/null +++ b/management/reporting/uidata/user_activity.py @@ -0,0 +1,167 @@ +from .Timeseries import Timeseries +from .exceptions import InvalidArgsError + +with open(__file__.replace('.py','.1.sql')) as fp: + select_1 = fp.read() + +with open(__file__.replace('.py','.2.sql')) as fp: + select_2 = fp.read() + + +def user_activity(conn, args): + ''' + details on user activity + ''' + try: + user_id = args['user_id'] + + # use Timeseries to get a normalized start/end range + ts = Timeseries( + 'User activity', + args['start_date'], + args['end_date'], + 0 + ) + except KeyError: + raise InvalidArgsError() + + # limit results + try: + limit = 'LIMIT ' + str(int(args.get('row_limit', 1000))); + except ValueError: + limit = 'LIMIT 1000' + + # + # sent mail by user + # + c = conn.cursor() + + sent_mail = { + 'start': ts.start, + 'end': ts.end, + 'y': 'Sent mail', + 'fields': [ + # mta_connection + 'connect_time', + 'sasl_method', + # mta_accept + 'envelope_from', + # mta_delivery + 'rcpt_to', + 'service', + 'spam_score', + 'spam_result', + 'message_size', + 'status', + 'relay', + 'delivery_info', + 'delivery_connection', + 'delivery_connection_info', + 'sent_id', # must be last + ], + 'field_types': [ + { 'type':'datetime', 'format': '%Y-%m-%d %H:%M:%S' },# connect_time + 'text/plain', # sasl_method + 'text/email', # envelope_from + { 'type':'text/email', 'label':'Recipient' }, # rcpt_to + 'text/plain', # mta_delivery.service + { 'type':'decimal', 'places':2 }, # spam_score + 'text/plain', # spam_result + 'number/size', # message_size + 'text/plain', # status + 'text/hostname', # relay + 'text/plain', # delivery_info + 'text/plain', # delivery_connection + 'text/plain', # delivery_connection_info + 'number/plain', # sent_id - must be last + ], + 'items': [], + 'unique_sends': 0 + } + last_mta_accept_id = -1 + sent_id = 0 + for row in c.execute(select_1 + limit, { + 'user_id': user_id, + 'start_date': ts.start, + 'end_date': ts.end + }): + v = [] + for key in sent_mail['fields']: + if key != 'sent_id': + v.append(row[key]) + if last_mta_accept_id != row['mta_accept_id']: + sent_mail['unique_sends'] += 1 + last_mta_accept_id = row['mta_accept_id'] + sent_id += 1 + v.append(sent_id) + sent_mail['items'].append(v) + + + # + # received mail by user + # + + received_mail = { + 'start': ts.start, + 'end': ts.end, + 'y': 'Sent mail', + 'fields': [ + # mta_connection + 'connect_time', + 'service', + 'sasl_username', + + # mta_accept + 'envelope_from', + 'disposition', + 'spf_result', + 'dkim_result', + 'dkim_reason', + 'dmarc_result', + 'dmarc_reason', + + # mta_delivery + 'postgrey_result', + 'postgrey_reason', + 'postgrey_delay', + 'spam_score', + 'spam_result', + 'message_size', + ], + 'field_types': [ + { 'type':'datetime', 'format': '%Y-%m-%d %H:%M:%S' },# connect_time + 'text/plain', # mta_connection.service + 'text/email', # sasl_username + 'text/email', # envelope_from + 'text/plain', # disposition + 'text/plain', # spf_result + 'text/plain', # dkim_result + 'text/plain', # dkim_result + 'text/plain', # dmarc_result + 'text/plain', # dmarc_result + 'text/plain', # postgrey_result + 'text/plain', # postgrey_reason + { 'type':'time/span', 'unit':'s' }, # postgrey_delay + { 'type':'decimal', 'places':2 }, # spam_score + 'text/plain', # spam_result + 'number/size', # message_size + ], + 'items': [] + } + + for row in c.execute(select_2 + limit, { + 'user_id': user_id, + 'start_date': ts.start, + 'end_date': ts.end + }): + v = [] + for key in received_mail['fields']: + v.append(row[key]) + received_mail['items'].append(v) + + + + return { + 'sent_mail': sent_mail, + 'received_mail': received_mail + } diff --git a/management/templates/index.html b/management/templates/index.html index 12f6ad8e..35be0616 100644 --- a/management/templates/index.html +++ b/management/templates/index.html @@ -109,6 +109,7 @@
  • Contacts/Calendar
  • Web
  • +
  • Activity