Skip to main content

Add logging to Code Ocean capsules

Project description

AIND Logging Standards

This repository outlines the logging standards for AIND, enabling logs that are consistent, searchable, and impactful. This document serves a guideline and includes topics such as best practices and observability integration.

Standards

Best Practices

Observability Integration

-----------------------------------------------------

:star: Standards

Structured Logging

Log messages should be structured in JSON format. If saving logs to a file, the logs should be structured as JSON Lines (JSONL).

Example structured logs in a file:

{"timestamp": "2026-02-09T21:19:24.061270Z", "level": "INFO", "message": "Starting program"}
{"timestamp": "2026-02-09T21:19:24.061270Z", "level": "INFO", "message": "Connecting to database"}
{"timestamp": "2026-02-09T21:19:24.061774Z", "level": "INFO", "message": "Successfully retrieved data"}
{"timestamp": "2026-02-09T21:19:24.061774Z", "level": "INFO", "message": "Closing program"}

Levels

Levels should be used to organize logs with different severity. The following leveling scheme is based on the python standard logging library. When using levels in logs, the string should be used.

The table below provides the string used for a level, the number value associated with that level, and a description of what events/scenarios should be associated with that level.

Level String Level Number Description
DEBUG 10 Detailed information that is usually useful only to developers
INFO 20 Significant business events
WARNING 30 Events that may indicate future problems
ERROR 40 Errors that affect a specific operation
CRITICAL 50 Errors that affect entire program

Fields/Labels

At a minimum, projects should contain the following labels: timestamp, level, and message.

The table below contains a list of common fields/labels used in AIND projects. These fields are not mandatory to add to a log message, but when applicable, they should be used with the exact label name and the expected format. The table describes each field's format, purpose, and provides examples.

Field/Label Required Description Format Examples
timestamp Time the log was created ISO8601 UTC
level Log level See level standards DEBUG, INFO, WARNING, ERROR, CRITICAl
message String message about event String "Error connecting to database... retrying"
lineno Source line number where log call was issued Integer 1, 20, 39
process Process ID (from OS) String 1231
processName Process name String python.exe
thread Thread ID (from OS) String 12913
threadName Thread name String python.exe
hostname Hostname of the machine running software String w10dt700140
software_name The software generating the logs String waterlog, aind-data-schema
software_version Version of the software generating the logs String 1.0.4.dev0
rig_id SIPE's naming convention for rigs/machines. Commonly found as an environment variable called aibs_comp_id on the machine SIPE's rig naming convention frg_1_a, ephys_9_acq
instrument_id Scientific Computing's convention for instruments. Definition can be found here String MESO1, EPHYS1
acquisition_name unique name of session defined as <subject_id>_<acquisition-date>_<acquisition-time> String 123456_2026-01-01_23-00-00
subject_id Subject ID (Unique identifier for the subject of data acquisition) String 123456
specimen_id An identifier for a tissue removed from a subject and should include the subject_id in the name String Full brain: 123456. Sectioned brain: 123456_001, ..., 123456_00N
pipeline_name Name of pipeline processing, transforming and/or QC'ing raw data String codeocean pipeline, airflow DAG
process_name Name of container processing, tranforming and/or QC'ing raw data String code ocean capsule
user_id User running software. Intended specifically for software which run experiments to indicate who is running the experiment String

-----------------------------------------------------

:pencil: Best Practices

In additional to following the standards above, it is recommended to adhere to these practices when implementing logs.

Majority of the information in this section is based off of the following resources:

Meaningful Logs

The utility of logs is directly tied to the quality of the information it provides. Ensure that log messages fully describe the situation happening and that there are ample labels/fields to give context about the log.

Here is an example of a poor log message with not enough context

{"level": "INFO", "message": "Fail to save subject info"}

Here is the same example with a more informative message, and additional context about the log

{"level": "INFO", "message": "Fail to save subject info: field 'experiment_id' is missing", "subject_id": 714392}

Canonical Logs

A canonical log line is a single log entry emitted at the end of a request that captures all essential information needed to understand what occurred, without requiring users to piece together multiple log messages.

Here is an example of two canonical logs:

{
  "timestamp": "2026-02-09T21:19:24.061270Z",
  "level": "INFO",
  "message": "Session recorded",
  "subject_id": "713491",
  "instrument_id": "frg_1_a",
  "status": "success",
  "database": "dataverse",
  "user_id": "bobby.b",
  "wt_g": 23.93,
  "water_earned_ml": 0.89
}
{
  "timestamp": "2026-02-09T21:19:24.061270Z",
  "level": "ERROR",
  "message": "Failed to save session: could not connect to database",
  "subject_id": "713491",
  "instrument_id": "frg_1_a",
  "status": "fail",
  "database": "dataverse",
  "error_type": "TimeoutError",
  "user_id": "bobby.b",
  "wt_g": 17.21,
  "water_earned_ml": 1.2
}

These logs contain all the relevant information about a session where a mouse was fed water. Within this example program, there could've been individual logs about events such as adding a mouse, adding a user, feeding it water, connecting to the database etc. With the canonical log, debugging an error doesn't require looking at all these individual logs since a user can reference the summary provided in these logs.

Don't log sensitive information

Using hard-coded tokens or passwords in your program? Shame. Accidentally exposing that information because it was logged and ingested into a public observability platform? Double shame.

It is also recommended to redact fields if logging custom objects. Here is an example of a python implementation that properly redacts sensitive information:

from pydantic import BaseModel, SecretStr


class User(BaseModel):
    id: str
    email: str
    password: SecretStr
    inner_most_thoughts: SecretStr

Don't rely on logs for performance or resource monitoring

Metrics generally provide continuous and efficient stream of data (think CPU usage over time). Logs aren't really designed to capture this type of data.

Logs generally have high-cardinality to describe in detail about events, if metrics were recorded in these logs, queries would be slow since they would have to parse through other information. On large-scale systems logs may be sampled and or even by lossy under high load which is not great for metrics.

If metrics are need there are tools like Telegraf which is designed to collect metrics, and InfluxDB which is used to store time-series data. Both these tools are also supported by Grafana.

-----------------------------------------------------

:telescope: Observability Integration

There are currently two major observability platforms in AIND that are used to monitor logs. Each with their own datastore setup that centrally stores all logs.

  1. Grafana - Loki database
  2. SIPE Log Server - MySQL database

It is recommended to send the logs to Loki and to use Grafana as the primary observability tool for logging. This is the most current solution for centralize observability and will have the most support.

The SIPE Log Server is the legacy observability platform and only handles viewing logs store in the MySQL database associated with it. These logs can also be viewed in Grafana.

How to get logs on Grafana

To view logs on Grafana, logs must first be uploaded to Loki. There are two primary methods to get logs into Loki:

After the logs are stored in Loki, Grafana can be used to view the logs within the Loki database. This can be done by going to the "Explore" tab in Grafana and selecting Loki as teh datasource. If the tab is not visible, you may need additional permissions in Grafana.

How to get logs on SIPE Log Server

Logs can be sent to the SIPE Log Server using this url: http://eng-logtools.corp.alleninstitute.org:9000

Here is a python example logging to the SIPE Log Server.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

aind_log_utils-0.2.7.tar.gz (53.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

aind_log_utils-0.2.7-py3-none-any.whl (9.4 kB view details)

Uploaded Python 3

File details

Details for the file aind_log_utils-0.2.7.tar.gz.

File metadata

  • Download URL: aind_log_utils-0.2.7.tar.gz
  • Upload date:
  • Size: 53.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for aind_log_utils-0.2.7.tar.gz
Algorithm Hash digest
SHA256 14b682069036c6edc12598d7e330a37f295be56f6772338c559b57ed0ab04ee5
MD5 8cff9e6b17ef4ce8191d685cb5ded46b
BLAKE2b-256 c9edf300f12af44d718987f8df5b7d7dc11d0ba67298abfb1c695d6250ad782e

See more details on using hashes here.

File details

Details for the file aind_log_utils-0.2.7-py3-none-any.whl.

File metadata

  • Download URL: aind_log_utils-0.2.7-py3-none-any.whl
  • Upload date:
  • Size: 9.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for aind_log_utils-0.2.7-py3-none-any.whl
Algorithm Hash digest
SHA256 8fd1064352c6b88c60b0eedbe36664e710a1ef3712a4ec0a78bf17b03ec89725
MD5 d91444a0279f983ca5b81e975bbc33a2
BLAKE2b-256 c7b19e4e642d223adb9d01c37502dd00270e935bcd8f671ebf9032c5860ad403

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page