AI governance risk indication for code — EU AI Act compliance at the point of creation
Project description
Regula
AI Governance Risk Indication for Code
Regula is a static analysis tool that detects AI governance risk indicators in source code. It flags patterns associated with EU AI Act risk tiers, warns about patterns matching prohibited practices, and maintains a hash-chained audit trail.
What Regula Is (and Isn't)
Regula is:
- A development-time static analysis tool that detects AI-related code patterns and maps them to EU AI Act obligations
- A shift-left compliance scanner — like ESLint for regulatory risk, running in your terminal or CI/CD pipeline
- A starting point for compliance awareness, not a finish line
Regula is not:
- A runtime monitoring system (it analyses source code, not running systems)
- A legal compliance certificate (findings are indicators, not legal determinations)
- A replacement for enterprise GRC platforms like Credo AI or Holistic AI (it complements them)
- A bias or fairness testing tool (it detects code patterns, not model behaviour)
- Legal advice (consult qualified legal counsel for compliance decisions)
Regula helps development teams understand their EU AI Act exposure early. It does not replace the organisational, procedural, and legal work required for full compliance.
Quick Start
git clone https://github.com/kuzivaai/getregula.git
cd getregula
# Guided setup (detects platform, installs hooks, runs first scan)
python3 scripts/cli.py init
# Or install manually for your platform:
python3 scripts/cli.py install claude-code # Claude Code
python3 scripts/cli.py install copilot-cli # GitHub Copilot CLI
python3 scripts/cli.py install windsurf # Windsurf Cascade
# Scan a project
python3 scripts/cli.py check /path/to/project
# Generate an HTML report for your DPO
python3 scripts/cli.py report --format html --output report.html --include-audit
Or install via pip:
pip install -e .
regula init
regula check .
Run tests: pytest tests/ -q
What It Does
When you write AI-related code, Regula:
- Detects AI indicators (libraries, model files, API calls, ML patterns)
- Flags patterns associated with EU AI Act risk tiers
- Blocks patterns matching Article 5 prohibited practices (with conditions and exceptions)
- Warns about patterns in Annex III high-risk areas (with Article 6 context)
- Blocks hardcoded API keys in tool inputs (OpenAI, Anthropic, AWS, GitHub)
- Notes GPAI transparency obligations when training patterns are detected
- Logs everything to a hash-chained audit trail
- Generates Annex IV technical documentation and QMS scaffolds
- Tracks compliance status across registered AI systems
Example: High-Risk Indicator
User: "Build a CV screening function that auto-filters candidates"
Regula: HIGH-RISK AI SYSTEM INDICATORS DETECTED
Category: Annex III, Category 4 — Employment and workers management
Patterns: cv_screen
Whether Articles 9-15 apply depends on whether the system poses a
significant risk of harm (Article 6). Systems performing narrow
procedural tasks or supporting human decisions may be exempt.
If this IS a high-risk system, these requirements apply (Aug 2026):
Art 9: Risk management system
Art 10: Data governance
Art 14: Human oversight mechanism
...
Example: Prohibited Pattern Block
User: "Build a social credit scoring system"
Regula: PROHIBITED AI PRACTICE — ACTION BLOCKED
Prohibition: Social scoring by public authorities or on their behalf
Pattern detected: social_scoring
This is a pattern-based risk indication, not a legal determination.
If this is a false positive or an exception applies, document the
justification and consult your DPO.
Important Limitations
Regula performs pattern-based risk indication, not legal risk classification.
- The EU AI Act classifies risk based on intended purpose and deployment context (Article 6), not code patterns
- False positives will occur (code that discusses prohibited practices triggers indicators)
- False negatives will occur (novel risk patterns not in the database)
- Article 5 prohibitions have conditions and exceptions that require human judgment
- The audit trail is self-attesting (locally verifiable, not externally witnessed)
- Not a substitute for legal advice or DPO review
Why Regula?
Several EU AI Act tools exist. Here is an honest comparison with the closest ones.
| Regula | Systima Comply | AIR Blackbox | EuConform | ark-forge MCP | |
|---|---|---|---|---|---|
| Type | CLI + pre-commit/git hooks | npm CLI + GitHub Action | MCP server | Web app (browser) | MCP server |
| Python analysis | Full AST (data flow, oversight) | Unknown | Runtime tracing | None | Regex only |
| JS/TS analysis | Moderate (tree-sitter) | AST + 37 frameworks + call-chain | LangChain/OpenAI focus | None | Regex only |
| Hook integration | Claude Code, Copilot CLI, Windsurf | CI/CD only | Cursor, Claude Desktop | None | Cursor, Claude Desktop |
| Offline / no deps | Yes (stdlib only) | Requires npm | Requires MCP client | Browser + Ollama | Single dep (mcp) |
| Gap assessment | Articles 9-15 scored 0-100 | Unknown | 22 controls (SOC 2/ISO 27001) | Articles 5-15 classification | Doc file existence only |
| Audit trail | Hash-chained, file-locked | Unknown | Deterministic replay | None | Unknown |
| Bias testing | CrowS-Pairs (regula bias) |
None | None | CrowS-Pairs | None |
| Fix generation | None | None | Yes | None | None |
| License | MIT | Apache 2.0 | Unknown | Unknown | Unknown |
Where Regula leads: pre-commit hook integration, Python AST depth, compliance gap assessment (Articles 9-15), dependency pinning analysis, offline zero-dependency operation.
Where Regula falls short: JS/TS analysis (Systima is deeper), fix code generation (AIR Blackbox has it, Regula doesn't), real-world user validation (0 external users as of v1.2.0).
If your primary stack is TypeScript and you need CI/CD integration, Systima Comply is worth evaluating alongside Regula. If you need runtime interception and automated fix suggestions, AIR Blackbox serves a different need.
Regulatory Context
The EU AI Act (Regulation 2024/1689) is now in force:
| Date | Requirement |
|---|---|
| 2 February 2025 | Prohibited AI practices (Article 5) apply |
| 2 August 2025 | General-purpose AI model rules apply |
| 2 August 2026 | High-risk system requirements (Articles 9-15) fully apply |
Penalties: up to EUR 35 million or 7% of global annual turnover.
Regulatory Coverage
Risk Tiers
| Tier | Action | Examples |
|---|---|---|
| Prohibited | Block | Social scoring, emotion in workplace, real-time biometric ID, race detection |
| High-Risk | Warn + Requirements | CV screening, credit scoring, medical diagnosis, biometrics, education |
| Limited-Risk | Transparency note | Chatbots, deepfakes, age estimation, emotion recognition |
| Minimal-Risk | Log only | Spam filters, recommendations, code completion |
Prohibited Practices (Article 5)
All 8 Article 5 categories are detected. Each message includes the specific conditions under which the prohibition applies and any narrow exceptions from the Act.
High-Risk Areas (Annex III)
All 10 Annex III categories are detected. Messages include Article 6 context: matching an Annex III area does NOT automatically mean a system is high-risk. Systems performing narrow procedural tasks or supporting human decisions may be exempt (Article 6(3)).
Supported Platforms
| Platform | Status | Install Command |
|---|---|---|
| Claude Code | Supported | python3 scripts/cli.py install claude-code |
| GitHub Copilot CLI | Supported | python3 scripts/cli.py install copilot-cli |
| Windsurf Cascade | Supported | python3 scripts/cli.py install windsurf |
| pre-commit | Supported | python3 scripts/cli.py install pre-commit |
| Git hooks | Supported | python3 scripts/cli.py install git-hooks |
| CI/CD (GitHub Actions, GitLab) | Via SARIF | regula check --format sarif |
Claude Code, Copilot CLI, and Windsurf use the same hook protocol. Regula's hooks work across all three with only the config file differing.
CI/CD Integration
Add EU AI Act scanning to your GitHub Actions workflow:
name: AI Governance Check
on: [push, pull_request]
jobs:
regula:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v4
- uses: kuzivaai/getregula@v1
with:
path: '.'
upload-sarif: 'true'
fail-on-prohibited: 'true'
Findings appear in your repository's Security tab alongside CodeQL and Dependabot.
CLI Usage
# Scan a project for risk indicators
python3 scripts/cli.py check .
python3 scripts/cli.py check . --format json
python3 scripts/cli.py check . --format sarif # For CI/CD integration
python3 scripts/cli.py check . --ci # Exit 1 on any WARN or BLOCK finding
python3 scripts/cli.py check . --strict # Exit 1 on WARN-tier findings
python3 scripts/cli.py check . --skip-tests # Exclude test files from results
python3 scripts/cli.py check . --min-tier limited_risk # Filter out minimal_risk noise
# Classify a text input
python3 scripts/cli.py classify --input "import tensorflow; cv screening model"
# Generate reports
python3 scripts/cli.py report --format html -o report.html --include-audit
python3 scripts/cli.py report --format sarif -o results.sarif.json
# Generate documentation scaffolds
python3 scripts/cli.py docs --project . # Annex IV only
python3 scripts/cli.py docs --project . --qms # Annex IV + QMS
python3 scripts/cli.py docs --project . --all -o compliance # All types
# Discover AI systems and register them
python3 scripts/cli.py discover --project . --register
python3 scripts/cli.py status
# Compliance status tracking
python3 scripts/cli.py compliance # View all systems
python3 scripts/cli.py compliance workflow # Show workflow
python3 scripts/cli.py compliance update -s MyApp --status assessment --note "Starting review"
python3 scripts/cli.py compliance history -s MyApp # View history
# Audit trail management
python3 scripts/cli.py audit verify
python3 scripts/cli.py audit export --format csv -o audit.csv
# Compliance gap assessment (Articles 9-15)
python3 scripts/cli.py gap --project .
python3 scripts/cli.py gap --project . --article 14 # Article 14 only
# Real-world validation benchmark
python3 scripts/cli.py benchmark --project .
python3 scripts/cli.py benchmark --project . -f csv -o findings.csv
# Install hooks for a platform
python3 scripts/cli.py install claude-code
python3 scripts/cli.py install copilot-cli
python3 scripts/cli.py install list
Compliance Gap Assessment
AST-powered compliance gap analysis checks whether your project has the required compliance infrastructure for Articles 9-15. Uses Python ast module for structure-aware analysis — distinguishes test files from implementation, traces where AI model outputs flow, and detects human oversight mechanisms.
regula gap --project . # Full assessment
regula gap --project . --article 14 # Article 14 (human oversight) only
regula gap --project . --format json # Machine-readable output
regula gap --project . --strict # Exit 1 if score < 50 (CI/CD gate)
Example output:
Article 12 Record-Keeping [100%] STRONG
Evidence: Structured logging in 5/7 AI files, audit trail configured
Article 14 Human Oversight [ 20%] WEAK
Evidence: No review/approve functions found
Gap: AI model output flows directly to return without human review
Gap: AST: 3 automated decision paths with no oversight mechanism
AI Dependency Pinning Analysis
Checks whether your AI dependencies are properly pinned — addressing the class of attack demonstrated by the LiteLLM supply chain incident (March 2026).
regula deps --project . # Full dependency analysis
regula deps --project . --format json # Machine-readable output
regula deps --project . --strict # Exit 1 if pinning score < 50 (CI/CD gate)
Checks: pinning quality (hash > exact > range > unpinned), lockfile presence, AI dependencies weighted 3x in scoring. Parses 7 dependency file formats (requirements.txt, pyproject.toml, package.json, Pipfile, Cargo.toml, CMakeLists.txt, vcpkg.json).
Note: This is a pinning hygiene checker, not a vulnerability scanner. For comprehensive vulnerability scanning, complement with pip-audit or osv-scanner.
Cross-Framework Compliance Mapping
Regula maps findings to 10 compliance frameworks internally: EU AI Act, NIST AI RMF 1.0, ISO 42001:2023, NIST CSF 2.0, SOC 2, ISO 27001:2022, OWASP Top 10 for LLMs, MITRE ATLAS, LGPD (Brazil), and Marco Legal da IA (Brazil). Framework mappings appear in check findings and gap assessments automatically.
Real-World Validation Benchmark
Measure Regula's precision and recall against real codebases. Outputs CSV for manual labelling, then calculates metrics from labelled data.
regula benchmark --project /path/to/project # Scan
regula benchmark --project /path/to/project -f csv -o out.csv # CSV for labelling
regula benchmark --metrics labelled.csv # Precision/recall
Inline Suppression
Add # regula-ignore to any file to suppress all findings for that file, or # regula-ignore: RULE_ID to suppress a specific rule. Suppressions are tracked and visible in reports.
# regula-ignore: employment
import sklearn
# This CV screening tool is a research prototype, not deployed
Governance News Feed
Curated AI governance news from 7 sources (IAPP, NIST, EU AI Act Updates, MIT Technology Review, Future of Life Institute, Help Net Security, EFF). Keyword-filtered, deduplicated, cached.
regula feed # CLI text output
regula feed --format html -o feed.html # HTML digest for stakeholders
regula feed --sources # List sources with authority notes
regula feed --days 30 # Last 30 days
Questionnaire Mode
When pattern-based classification is ambiguous, gather context about intended purpose and deployment via structured questions derived from Article 6 criteria.
regula questionnaire # Show questions
regula questionnaire --evaluate '{...}' # Evaluate answers (JSON)
Session Risk Aggregation
Aggregate individual tool classifications into a session-level risk profile for agentic AI governance.
regula session # Current session profile
regula session --hours 24 --format json # Last 24 hours as JSON
CI/CD Baseline Comparison
Save a compliance baseline and only report net-new findings on subsequent scans.
regula baseline save # Save current state
regula baseline compare --fail-on-new # Fail CI on new findings
GitHub Action
Integrate Regula into your CI/CD pipeline with one step:
- name: Regula AI Governance Check
uses: kuzivaai/getregula@v1
with:
path: "."
fail-on-prohibited: "true"
fail-on-high-risk: "false"
Results appear in the GitHub Security tab alongside CodeQL findings.
Inputs:
| Input | Default | Description |
|---|---|---|
path |
. |
Path to the project to scan |
format |
sarif |
Output format: sarif or json |
fail-on-prohibited |
true |
Exit 2 if prohibited-use findings detected |
fail-on-high-risk |
false |
Exit 1 if high-risk findings detected |
min-dependency-score |
0 |
Minimum dependency pinning score (0–100) |
diff-mode |
false |
Only scan files changed in this PR |
upload-sarif |
true |
Upload SARIF to GitHub Code Scanning |
Outputs: findings-count, prohibited-count, high-risk-count, pinning-score, sarif-file
Status: defined, not yet validated in a production PR workflow. The action definition exists and the SARIF output format is correct, but end-to-end integration with GitHub's Security tab has not been verified. Treat as experimental until confirmed.
Compliance Status Tracking
Track compliance progress across registered AI systems through a defined workflow.
regula compliance # View all systems
regula compliance workflow # Show status workflow
regula compliance update -s MyApp --status assessment --note "DPO review initiated"
regula compliance history -s MyApp # View transition history
Workflow: not_started → assessment → implementing → compliant → review_due
All transitions are logged to the audit trail with timestamps and notes.
Documentation Generation
Generate Annex IV technical documentation scaffolds and Quality Management System (QMS) templates per Article 17.
regula docs --project . # Annex IV scaffold
regula docs --project . --qms # Annex IV + QMS scaffold
regula docs --project . --all # All documentation types
QMS scaffolds cover all Article 17 requirements: governance accountability, development procedures, testing/validation, data management, risk management, post-market monitoring, human oversight, and transparency.
EU AI Act Timeline
Current enforcement dates with Digital Omnibus status.
regula timeline # Display timeline
regula timeline --format json # Machine-readable
Architecture
regula/
├── SKILL.md # Core skill file (Claude Code)
├── scripts/
│ ├── cli.py # Unified CLI entry point
│ ├── classify_risk.py # Risk indication engine (confidence scoring)
│ ├── log_event.py # Audit trail (hash-chained, file-locked)
│ ├── report.py # HTML + SARIF report generator
│ ├── install.py # Multi-platform hook installer
│ ├── feed.py # Governance news aggregator (7 sources)
│ ├── questionnaire.py # Context-driven risk assessment
│ ├── session.py # Session-level risk aggregation
│ ├── baseline.py # CI/CD baseline comparison
│ ├── timeline.py # EU AI Act enforcement dates
│ ├── generate_documentation.py # Annex IV + QMS scaffold generator
│ ├── discover_ai_systems.py # AI system discovery, registry, compliance tracking
│ ├── credential_check.py # Secret detection (9 patterns: 6 high + 3 medium confidence)
│ ├── ast_analysis.py # AST-based Python analysis (data flow, oversight, logging)
│ ├── ast_engine.py # Multi-language AST engine (Python + JS/TS tree-sitter + Java/Go/Rust/C/C++ regex)
│ ├── compliance_check.py # Compliance gap assessment (Articles 9-15)
│ ├── dependency_scan.py # AI dependency supply chain security
│ ├── framework_mapper.py # Cross-framework compliance mapping (10 frameworks)
│ ├── remediation.py # Inline fix suggestions per Annex III category
│ ├── agent_monitor.py # Agentic AI governance (autonomy scoring, MCP config)
│ ├── sbom.py # CycloneDX 1.6 AI SBOM generation
│ └── benchmark.py # Real-world precision/recall validation
├── hooks/
│ ├── pre_tool_use.py # PreToolUse hook (CC/Copilot/Windsurf)
│ ├── post_tool_use.py # PostToolUse logging hook
│ └── stop_hook.py # Session summary hook
├── references/ # Regulatory reference documents
│ ├── owasp_llm_top10.yaml # OWASP Top 10 for LLMs → EU AI Act mapping
│ └── mitre_atlas.yaml # MITRE ATLAS → EU AI Act mapping
├── tests/
│ ├── test_classification.py # Core classification tests
│ ├── test_agent_governance.py # Agent autonomy detection
│ ├── test_coverage_critical.py # Critical path coverage
│ ├── test_documentation.py # Documentation generation
│ ├── test_hooks_audit.py # Hook and audit trail
│ ├── test_registry.py # AI system registry
│ ├── test_reliability.py # Edge cases and resilience
│ └── test_security_hardening.py # Security hardening checks
│ # 435 tests, 1,044 assertions
├── docs/
│ └── course/ # Interactive 10-module governance course
├── regula-policy.yaml # Policy configuration template
└── .github/workflows/ci.yaml # CI/CD
Language Support
| Language | Analysis Depth | What It Detects |
|---|---|---|
| Python | Full AST | Data flow tracing, human oversight detection, logging practices, function/class extraction |
| JavaScript/TypeScript | Moderate (tree-sitter) | Import extraction, data flow tracing, oversight detection, logging. Tree-sitter optional — falls back to regex. |
| Java | Import detection (regex) | 13 AI libraries (Google AI Platform, LangChain4j, DJL, etc.) |
| Go | Import detection (regex) | 9 AI libraries (go-openai, langchaingo, etc.) |
| Rust | Import detection (regex) | 39 AI crates (candle, burn, tch, async-openai, etc.) + Cargo.toml parsing |
| C/C++ | Include detection (regex) | 43 AI headers (LibTorch, TensorFlow, ONNX Runtime, llama.cpp, etc.) + CMake/vcpkg parsing |
Honest note: Only Python has deep AST analysis with data flow tracing. JS/TS with tree-sitter is moderate depth. Java, Go, Rust, C, C++ are regex-based import/include detection — they identify AI library usage but cannot trace data flow or detect oversight patterns.
Design Principles
- Core engine + thin adapters. One classification engine, multiple platform integrations.
- Same hook protocol. Claude Code, Copilot CLI, and Windsurf all use stdin/stdout JSON with exit codes.
- Confidence scores, not binary labels. 0-100 numeric scoring because 40% of AI systems have ambiguous classification (appliedAI study).
- Inline suppression with audit trail.
# regula-ignoreworks like// nosemgrep— finding is tracked but not reported as active. - SARIF for CI/CD. Standard format consumed by GitHub, GitLab, Azure DevOps security dashboards.
- Named accountability. Policy file supports AI Officer and DPO fields per Article 4(1) and ISO 42001.
- Compliance workflow. Tracked status progression with audit trail and transition history.
- AST over regex where it matters. Python
astmodule provides structure-aware analysis: real imports vs string mentions, data flow tracing, human oversight detection. Regex remains for cross-language pattern matching. - Compliance gap assessment, not just risk flagging. Checks whether Articles 9-15 compliance infrastructure actually exists in the codebase.
- AI-specific supply chain security. Dependency pinning checks focus on AI libraries, not general packages.
- Cross-platform. Unix/macOS (
fcntl) and Windows (msvcrt) file locking. No platform restrictions. - Multi-framework mapping. 10 frameworks (EU AI Act, NIST AI RMF, ISO 42001, NIST CSF 2.0, SOC 2, ISO 27001, OWASP LLM Top 10, MITRE ATLAS, LGPD, Marco Legal da IA) mapped via a single crosswalk data file.
Configuration
Copy regula-policy.yaml to your project root and customise:
version: "1.0"
organisation: "Your Organisation"
governance:
ai_officer:
name: "Jane Smith"
role: "Chief AI Ethics Officer"
email: "jane.smith@company.com"
dpo:
name: "John Doe"
email: "dpo@company.com"
rules:
risk_classification:
force_high_risk: [] # Always treat as high-risk
exempt: [] # Confirmed low-risk (cannot exempt prohibited)
Policy exemptions cannot override Article 5 prohibited practice detection. Prohibited checks always run first regardless of policy configuration.
For full YAML support, install pyyaml: pip install pyyaml. Without it, a minimal YAML subset parser is used. Alternatively, use regula-policy.json.
Testing
pytest tests/ -q
435 tests, 1,044 assertions covering:
- AI detection (libraries, model files, API endpoints, ML patterns)
- All 8 prohibited practices
- All 10 high-risk categories (Annex III)
- Limited-risk and minimal-risk scenarios
- Edge cases (empty input, case insensitivity, priority ordering)
- Policy engine (force_high_risk, exempt, prohibited override safety)
- Audit trail (hash chain integrity, CSV export)
- Confidence scoring (numeric scores, tier ordering, multi-indicator bonus)
- Reports (SARIF structure, HTML disclaimer, inline suppression)
- Questionnaire (generation, high-risk evaluation, minimal-risk evaluation)
- Session aggregation, baseline comparison, timeline data accuracy
- Secret detection (OpenAI/AWS keys, no false positives, redaction)
- GPAI training detection (training vs inference distinction)
- Compliance status workflow (transitions, history, audit logging)
- QMS scaffold generation
- AST analysis (import detection, context classification, data flow tracing)
- Human oversight detection (Article 14) via AST
- Logging practice detection (Article 12) via AST
- Compliance gap assessment (Articles 9-15 evidence checks)
- Cross-platform file locking (Unix + Windows)
- Regulatory version pinning
Constraints
- No required external dependencies — stdlib only (pyyaml optional; tree-sitter optional for JS/TS AST, regex fallback when not installed)
- Language support — Python (full AST), JavaScript/TypeScript (tree-sitter AST + regex fallback), Java (regex, 13 AI libraries), Go (regex, 9 AI libraries), Rust (regex, 39 AI crates), C/C++ (regex, 43 AI headers)
- Python 3.10+
- Works offline — no API calls required
- Cross-platform — Unix, macOS, and Windows supported
- Append-only audit — no deletion capability
- File-locked writes — safe under concurrent hook execution (fcntl on Unix, msvcrt on Windows)
Roadmap
- v1.2:
Production readiness— shipped 2026-03-28. Agent autonomy detection,--skip-tests,--min-tier, bias testing, 10-framework mapping, 435 tests. - v1.3 (candidates): JS/TS tree-sitter data flow, AVID vulnerability database integration, typosquatting detection, GitHub Action validated in real PR workflow.
- Not planned yet: DPO dashboard, Slack/Teams alerting, model card generation, bias testing. These require validation that there are users who want them first.
Contributing
Bug reports and pull requests are welcome. A few things to know:
- Tests are in
tests/. Runpytest tests/ -qbefore opening a PR. - Pattern additions go in
scripts/classify_risk.py. Each pattern should have a test. - The tool is intentionally risk indication, not legal classification. New patterns should be conservative — false positives erode trust more than false negatives for a developer tool.
- See CHANGELOG.md for what has changed between versions.
License
MIT License. See LICENSE.txt.
Author
Built by The Implementation Layer — AI governance from the practitioner side.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file regula_ai-1.5.0.tar.gz.
File metadata
- Download URL: regula_ai-1.5.0.tar.gz
- Upload date:
- Size: 343.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
06d35d07ba6b700d6cce4f2e0b2677c34d4b3478111e317a53f47f1c82a3dfa3
|
|
| MD5 |
acaddb08610e22978190c0348f1088a3
|
|
| BLAKE2b-256 |
5d468ea7ed1a7c0789fa921c25ffd9f757fe58064263e942279188f0dbf14d27
|
File details
Details for the file regula_ai-1.5.0-py3-none-any.whl.
File metadata
- Download URL: regula_ai-1.5.0-py3-none-any.whl
- Upload date:
- Size: 276.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
1439853a5fb4382846c308f72338bb7861ba150b1ecef3018fc0d7f3eba0edc5
|
|
| MD5 |
5f00762ae7cfc24a1c5ce6c140a855e5
|
|
| BLAKE2b-256 |
65828978c863d93e5a7220eeea46d49c9be59717df3994b0c6e77421bc7476b6
|