Skip to main content

A Python daemon and library for collecting, enriching, storing and forwarding network events from Suricata and NFStream.

Project description

PTS logo

Mongoose

Collect, enrich, store and forward Suricata alerts and network flows.

License: GPL v3

Website | Documentation | GitHub | Support

Mongoose is a Python daemon and library for collecting, enriching, storing and forwarding network events from Suricata and NFStream.

It runs a thread-safe pub-sub pipeline (ProcessingQueue) where collectors publish raw events to topics and subscribers — enricher, SQLite store, forwarders — consume them concurrently.

Key features

  • Modular collectors: Suricata EVE (alerts and netflow via Unix socket), NFStream (live packet capture from a network interface).
  • Automatic enrichment: traffic direction (inbound / outbound / local), Community ID calculation, reverse DNS hostname lookup, event type classification, and flow risk scoring via a configurable severity cache.
  • GeoIP enrichment: MaxMind (GeoLite2-ASN, GeoLite2-City, GeoLite2-Country) and IP66 databases, with daily automatic database updates.
  • Pluggable forwarders: local file output, HTTP(S) webhooks (immediate, bulk or periodic modes with retry logic and multiple authentication methods), and Discord (rich embed formatting).
  • Topic filtering: forwarders can be scoped to specific topics and filtered by event attributes.
  • Drop-in webhook configuration: new webhook forwarders can be added at runtime by dropping a YAML file into a watched directory — no restart required.
  • SQLite storage: enriched events are persisted with configurable history pruning by record count or age.
  • Sharded LRU cache: thread-safe severity cache with optional TTL used for flow risk scoring.
  • Singleton engine: a single Engine instance manages the full component lifecycle (start, stop, reload).
  • Systemd and PiRogue integration: the CLI daemon supports sd_notify and reads the isolated interface from pirogue-admin-client when available.

Pipeline topics

Events flow through the following pub-sub topics:

Topic Description
network-dpi Raw DPI flows from NFStream
network-alert Raw Suricata alerts
network-flow Raw Suricata netflow records
enriched-network-dpi Enriched DPI flows
enriched-network-alert Enriched Suricata alerts
enriched-network-flow Enriched netflow records

Installation

Install in a virtual environment and editable mode for development:

python -m venv .venv && source .venv/bin/activate
pip install -e .

CLI usage

The package installs a mongoosed daemon entry-point:

# show top-level help
mongoosed --help

# run with a configuration file
mongoosed --config /etc/mongoose/mongoose.yaml

# override the network interface used by NFStream
mongoosed --config mongoose.yaml --interface eth0

# set logging verbosity
mongoosed --config mongoose.yaml --logging-level DEBUG

Python library usage

import time
from mongoose.core.engine import Engine

engine = Engine("config.yaml")
engine.start()
time.sleep(6)
engine.stop()

Configuration

Configured via a YAML file. All keys live under a top-level configuration key.

configuration:
  collector:
    suricata:
      socket_path: "/run/suricata.socket"  # Suricata Unix socket
      collect_alerts: true                 # collect Suricata alerts
      collect_netflow: false               # collect Suricata netflow records
      enable: true

    nf_stream:
      interface: "eth0"          # network interface for live capture
      active_timeout: 120        # seconds before an active flow expires
      enable: false

  enrichment:
    geoip:
      source: "ip66"             # "ip66" (default) or "maxmind"
      enable: true

  forwarder:
    webhooks:
      - url: "https://hooks.example.com/ingest"
        auth_type: "bearer"      # none | basic | bearer | header
        auth_token: "${WEBHOOK_AUTH_TOKEN}"
        verify_ssl: true
        retry_count: 3
        retry_delay: 5.0
        timeout: 10.0
        mode: "immediate"        # immediate | bulk | periodic
        bulk_size: 10
        periodic_interval: 5.0
        periodic_rate: 10
        topics:
          - "enriched-network-dpi"
          - "enriched-network-alert"
        enable: true

  database_path: "mongoose.db"

  history:
    max_duration_days: 14  # keep records for at most 14 days
    max_records: null      # optional hard cap on row count per table
    enable: true

  cache:
    severity:
      max_size: 1024       # maximum entries in the severity LRU cache
      ttl_seconds: null    # optional TTL; null means entries never expire
      enable: true

  extra_configuration_dir: "/var/lib/mongoose"

Drop-in webhook configuration

Place a YAML file matching the WebhookForwarderConfiguration schema inside <extra_configuration_dir>/webhook.d/. The engine watches this directory and activates new forwarders when a file is created, and deactivates them when it is deleted.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pirogue_mongoose-1.0.2.tar.gz (38.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

pirogue_mongoose-1.0.2-py3-none-any.whl (56.4 kB view details)

Uploaded Python 3

File details

Details for the file pirogue_mongoose-1.0.2.tar.gz.

File metadata

  • Download URL: pirogue_mongoose-1.0.2.tar.gz
  • Upload date:
  • Size: 38.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: uv/0.7.16

File hashes

Hashes for pirogue_mongoose-1.0.2.tar.gz
Algorithm Hash digest
SHA256 e19c9d1ef0d97f5458c69e493186396bf38e2a3eb800d8909137b9f40fc5920c
MD5 2a2761752029a793dc11c5e9eec00b68
BLAKE2b-256 2badecbe15167f2f1b7454b2d6667d77bf9476515d162927f9967457d410dbf0

See more details on using hashes here.

File details

Details for the file pirogue_mongoose-1.0.2-py3-none-any.whl.

File metadata

File hashes

Hashes for pirogue_mongoose-1.0.2-py3-none-any.whl
Algorithm Hash digest
SHA256 c209d67d0775a3bfa0799bbf1e68322a146a35d07057a487bf80ecd50e8ad9e9
MD5 d808933daa623d735fdd3c8adf837b89
BLAKE2b-256 9d75878afe740c503f79c17d47a9de68f69030de9362c2a92b4554ac7e18edd1

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page