Skip to main content

Generate grpc API Service based on function

Project description

memfs - A Python Virtual File System in Memory

A Python module that implements a virtual file system in memory. This module provides an interface compatible with the standard os module and enables operations on files and directories stored in RAM rather than on disk.

Overview

memfs is designed to provide a fast, isolated file system environment for applications that need temporary file operations without the overhead of disk I/O. It's particularly useful for testing, data processing pipelines, and applications that need to manipulate files without affecting the host system.

Features

  • Complete in-memory file system implementation
  • API compatible with Python's standard os module
  • File and directory operations (create, read, write, delete, rename)
  • Path manipulation and traversal
  • File-like objects with context manager support
  • gRPC service generation for pipeline components
  • No disk I/O overhead
  • Isolated from the host file system

Installation

pip install memfs

Or install from source:

git clone https://github.com/pyfunc/memfs.git
cd memfs
pip install -e .
python3 -m venv .venv
source .venv/bin/activate  # On Linux/macOS
#.venv\Scripts\activate  # On Windows
pip install -e .
python -m build
# Deactivate current venv
deactivate

# Remove the existing venv
rm -rf .venv

# Create a fresh venv
python -m venv .venv

# Activate it
source .venv/bin/activate

pip install --upgrade pip


# Install setuptools first
pip install setuptools wheel

# Then try installing your package
pip install -e .

Basic Usage Examples

Basic File Operations

from memfs import create_fs

# Create a file system instance
fs = create_fs()

# Write to a file
fs.writefile('/hello.txt', 'Hello, world!')

# Read from a file
content = fs.readfile('/hello.txt')
print(content)  # Outputs: Hello, world!

# Check if a file exists
if fs.exists('/hello.txt'):
    print('File exists!')

# Create directories
fs.makedirs('/path/to/directory')

# List directory contents
files = fs.listdir('/path/to')

Using File-Like Objects

from memfs import create_fs

fs = create_fs()

# Write using a file-like object
with fs.open('/data.txt', 'w') as f:
    f.write('Line 1\n')
    f.write('Line 2\n')

# Read using a file-like object
with fs.open('/data.txt', 'r') as f:
    for line in f:
        print(line.strip())

Directory Operations

from memfs import create_fs

fs = create_fs()

# Create nested directories
fs.makedirs('/a/b/c')

# Walk the directory tree
for root, dirs, files in fs.walk('/'):
    print(f"Directory: {root}")
    print(f"Subdirectories: {dirs}")
    print(f"Files: {files}")

Advanced Usage Examples

Data Processing Pipeline

from memfs import create_fs
import json
import csv

fs = create_fs()

# Create directories
fs.makedirs('/data/raw', exist_ok=True)
fs.makedirs('/data/processed', exist_ok=True)

# Write CSV data
with fs.open('/data/raw/input.csv', 'w', newline='') as f:
    writer = csv.writer(f)
    writer.writerows([
        ['id', 'name', 'value'],
        [1, 'Alpha', 100],
        [2, 'Beta', 200]
    ])

# Process CSV to JSON
with fs.open('/data/raw/input.csv', 'r', newline='') as f:
    reader = csv.DictReader(f)
    data = [row for row in reader]

# Transform and save the data
for item in data:
    item['value'] = int(item['value'])
    item['double_value'] = item['value'] * 2

with fs.open('/data/processed/output.json', 'w') as f:
    json.dump(data, f, indent=2)

Parallel Processing

from memfs import create_fs
import json
import concurrent.futures

fs = create_fs()
fs.makedirs('/parallel/input', exist_ok=True)
fs.makedirs('/parallel/output', exist_ok=True)

# Create input files
for i in range(10):
    fs.writefile(f'/parallel/input/file_{i}.json', json.dumps({'id': i}))

def process_file(filename):
    with fs.open(f'/parallel/input/{filename}', 'r') as f:
        data = json.loads(f.read())
    
    # Process data
    data['processed'] = True
    
    with fs.open(f'/parallel/output/processed_{filename}', 'w') as f:
        f.write(json.dumps(data, indent=2))
    
    return data['id']

# Process files in parallel
with concurrent.futures.ThreadPoolExecutor(max_workers=4) as executor:
    futures = {executor.submit(process_file, f): f for f in fs.listdir('/parallel/input')}
    for future in concurrent.futures.as_completed(futures):
        file_id = future.result()
        print(f"Processed file ID: {file_id}")

gRPC Service Pipeline

from memfs import create_fs
from memfs.api import DynamicgRPCComponent, PipelineOrchestrator

# Define transformation functions
def transform_data(data):
    if isinstance(data, dict):
        data['transformed'] = True
    return data

def format_data(data):
    if isinstance(data, dict):
        data['formatted'] = True
    return data

# Create virtual directories
fs = create_fs()
fs.makedirs('/proto/transform', exist_ok=True)
fs.makedirs('/proto/format', exist_ok=True)
fs.makedirs('/generated/transform', exist_ok=True)
fs.makedirs('/generated/format', exist_ok=True)

# Create components
transform_component = DynamicgRPCComponent(
    transform_data,
    proto_dir="/proto/transform",
    generated_dir="/generated/transform",
    port=50051
)

format_component = DynamicgRPCComponent(
    format_data,
    proto_dir="/proto/format",
    generated_dir="/generated/format",
    port=50052
)

# Create and execute pipeline
pipeline = PipelineOrchestrator()
pipeline.add_component(transform_component)
pipeline.add_component(format_component)

result = pipeline.execute_pipeline({"input": "data"})
print(result)  # {"input": "data", "transformed": true, "formatted": true}

Command-line Interface

memfs provides a command-line interface for basic file operations:

# Display filesystem as a tree
memfs tree /

# Create an empty file
memfs touch /data/hello.txt

# Create directories
memfs mkdir -p /data/subdir

# Write content to a file
memfs write /data/hello.txt "Hello, virtual world!"

# Read file content
memfs read /data/hello.txt

# Dump filesystem content as JSON
memfs dump

API Reference

memfs/
├── setup.py          # Plik instalacyjny setuptools
├── setup.cfg         # Konfiguracja setuptools
├── README.md         # Dokumentacja projektu
├── src/              # Kod źródłowy
│   └── memfs/        # Pakiet główny
│       ├── __init__.py     # Import podstawowych komponentów
│       ├── _version.py     # Informacje o wersji
│       ├── memfs.py        # Implementacja wirtualnego systemu plików
│       ├── api.py          # Moduł do generowania usług gRPC
│       └── cli.py          # Interfejs wiersza poleceń
├── tests/            # Testy jednostkowe
│   ├── __init__.py
│   ├── test_memfs.py       # Testy dla modułu memfs
│   └── test_api.py         # Testy dla modułu API
└── examples/         # Przykłady użycia
    ├── basic_usage.py      # Podstawowe operacje
    └── advanced_usage.py   # Zaawansowane scenariusze

MemoryFS Class

  • open(path, mode='r') - Open a file
  • makedirs(path, exist_ok=False) - Create directories recursively
  • mkdir(path, mode=0o777) - Create a directory
  • exists(path) - Check if a path exists
  • isfile(path) - Check if a path is a file
  • isdir(path) - Check if a path is a directory
  • listdir(path) - List directory contents
  • walk(top) - Walk through directories recursively
  • remove(path) - Remove a file
  • rmdir(path) - Remove an empty directory
  • rename(src, dst) - Rename a file or directory
  • readfile(path) - Read an entire file
  • writefile(path, data) - Write data to a file
  • readfilebytes(path) - Read a file's contents as bytes
  • writefilebytes(path, data) - Write binary content to a file

API Module

  • DynamicgRPCComponent - Create a gRPC service from a function
  • PipelineOrchestrator - Orchestrate multiple components into a pipeline
  • ApiFuncConfig - Configuration for gRPC services
  • ApiFuncFramework - Framework for creating gRPC services

Use Cases

  • Unit testing - Test file operations without touching the disk
  • Data processing pipelines - Process data through multiple stages in memory
  • Microservices - Create gRPC services from Python functions
  • Sandboxed environments - Run file operations in an isolated environment
  • Performance optimization - Avoid disk I/O overhead for temporary operations
  • Containerized applications - Reduce container size by using in-memory storage

License

Apache-2.0

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

memfs-0.1.6.tar.gz (35.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

memfs-0.1.6-py3-none-any.whl (19.1 kB view details)

Uploaded Python 3

File details

Details for the file memfs-0.1.6.tar.gz.

File metadata

  • Download URL: memfs-0.1.6.tar.gz
  • Upload date:
  • Size: 35.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.9

File hashes

Hashes for memfs-0.1.6.tar.gz
Algorithm Hash digest
SHA256 11458953efc11396cd10725d5ad3118c5bac2b446089ba024df2b972f7748ef6
MD5 53d3bb5d9363bbb956404b0941ed2c08
BLAKE2b-256 83b9f8d7ec5d16429c1789a7ab0d7d5ba48e7666d0139450f70d4e81dbfd43f1

See more details on using hashes here.

File details

Details for the file memfs-0.1.6-py3-none-any.whl.

File metadata

  • Download URL: memfs-0.1.6-py3-none-any.whl
  • Upload date:
  • Size: 19.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.9

File hashes

Hashes for memfs-0.1.6-py3-none-any.whl
Algorithm Hash digest
SHA256 28e94258c47cdada8231c87d4315267464287f84a92f046e2e8edb979414c3fb
MD5 2cfb5e0325fac822d81361e8ccc189e2
BLAKE2b-256 5178938c85e4742da0731c46b0ff15ec0f7bb4a340e1feb472c1ff9935dd05e4

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page