Skip to main content

No project description provided

Project description

tock-genai-core

Composants principaux d'IA générative : models, factories, gestion des erreurs utilisés dans les composants Gen AI python de tock.

Architecture

Le projet est structuré en trois composants principaux :

  • Models : Contient les définitions des classes et modèles de données utilisés dans l'application
  • Services/Factories : Regroupe la logique métier et les fonctions utilisées par les routes

Technologies

  • Backend : Python
  • Base de données :
    • pgvector pour la base de données vectorielle
  • LLM & RAG :
    • Langchain pour l'orchestration
    • Support de guardrails
    • Reranker pour l'amélioration des résultats
    • Langfuse pour le monitoring

Providers disponibles

  • LLMProvider (LLM)

    • TGI = "HuggingFaceTextGenInference"
    • OpenAI = "OpenAI"
    • Vllm = "Vllm"
  • GuardrailProvider (Guardrail)

    • BloomZ = "BloomzGuardrail"
  • EMProvider (Embedding)

    • BloomZ = "BloomzEmbeddings"
    • OpenAI = "OpenAI"
    • Vllm = "Vllm"
  • VectorDBProvider (Database)

    • OpenSearch = "OPENSEARCH"
    • PGVector = "PGVECTOR"
  • ContextualCompressorProvider (Contetual Compressor)

    • BloomZ = "BloomzRerank"

Settings

  • Embedding

    • Classe parente

      BaseEMSetting:
          provider: EMProvider
          model: Optional[str]
          api_key: Optional[SecretKey]
          api_base: str
          pooling: Optional[str]
          space_type: Optional[str]
      
    • Classes enfants

      BloomZEMSetting(BaseEMSetting):
          provider: Literal[EMProvider.BloomZ]
      
      VLLMEMSetting(BaseEMSetting):
          provider: Literal[EMProvider.Vllm]
          model: str
      
      OpenAIEMSetting(BaseEMSetting):
          provider: Literal[EMProvider.OpenAI]
          api_base: str
          api_version: str
          deployment: str
      
  • Contextual compressor

    • Classe parente

      BaseCompressorSetting:
          provider: ContextualCompressorProvider
          endpoint: str
          api_key: Optional[SecretKey]
      
    • Classe enfant

      BloomZCompressorSetting(BaseCompressorSetting):
          provider: Literal[ContextualCompressorProvider.BloomZ]
          min_score: float
          max_documents: Optional[int]
          label: Optional[str]
      
  • Database

    • Classe parente

      BaseVectorDBSetting:
          index: Optional[str]
          provider: VectorDBProvider
          db_url: str
      
    • Classes enfants

      OpenSearchSetting(BaseVectorDBSetting):
          provider: Literal[VectorDBProvider.OpenSearch]
          username: SecretKey
          password: SecretKey
          use_ssl: bool
          verify_certs: bool
      
      class PGVectorSetting(BaseVectorDBSetting):
          provider: Literal[VectorDBProvider.PGVector]
          username: SecretKey
          password: SecretKey 
          db_name: str
          sslmode: Optional[str]
          namespace: str
      
  • Guardrail

    • Classe parente

      BaseGuardrailSetting:
          provider: GuardrailProvider
          api_base: str
          max_score: Optional[float]
          api_key: Optional[SecretKey]
      
    • Classe enfant

      BloomZGuardrailSetting(BaseGuardrailSetting):
          provider: Literal[GuardrailProvider.BloomZ]
      
  • Langfuse

    LangfuseSetting:
        host: Optional[str]
        public_key: Optional[SecretKey]
        secret_key: Optional[SecretKey]
        app_name: Optional[str]
        user_id: Optional[str]
        session_id: Optional[str]
    
  • LLM

    • Classe parente

      BaseLLMSetting:
          provider: LLMProvider
          model: Optional[str]
          api_key: Optional[SecretKey]
          temperature: float
      
    • Classes enfants

      OpenAILLMSetting(BaseLLMSetting):
          provider: Literal[LLMProvider.OpenAI]
          api_base: str
          api_version: str
          deployment: str
      
      HuggingFaceTextGenInferenceLLMSetting(BaseLLMSetting):
          provider: Literal[LLMProvider.TGI]
          repetition_penalty: float
          max_new_tokens: int
          api_base: str
          streaming: bool
      
      VllmSetting(BaseLLMSetting):
          provider: Literal[LLMProvider.Vllm]
          api_base: str
          max_new_tokens: int
          additional_model_kwargs: Optional[Dict[str, Any]]
      

Fonctionnement

Chaque outil utilisé (database, embedding, llm, langfuse, ...) a besoin d'un certains nombre de paramètres qui sont référencés dans les models (classes de settings)

Ces classes sont ensuite héritées par des services ou des factories afin de pouvoir répondre au besoin.

Exemple de get_vector_db_factory qui crée une factory de base vectorielle basée sur le nom de l'application et les paramètres d'embedding fournis

from tock-genai-core import get_vector_db_factory
from tock-genai-core import PGVectorSetting, VLLMEMSetting
from tock-genai-core import DBSetting, EMSetting


db_settings = PGVectorSetting(
    index = "first_index",
    provider = "PGVECTOR",
    db_url = "127.0.0.1:XXXX",
    db_name = "rag_sandbox_db",
    sslmode = "disable",
    username = {
      type = "Raw",
      value = "admin"
    },
    password = {
      type = "Raw",
      value = "example"
    },
    namespace = "test-name"
)

em_settings = VLLMEMSetting(
    provider = "Vllm",
    model = "model_name",
    api_base = "https://continue.com/v1"
)



def function_name(db_settings: DBSetting, em_settings: EMSetting):

    # do somethings

    vector = get_vector_db_factory(db_settings: DBSetting, em_settings: BaseEMSetting)

    # do somethings

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

tock_genai_core-0.1.1.tar.gz (22.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

tock_genai_core-0.1.1-py3-none-any.whl (56.2 kB view details)

Uploaded Python 3

File details

Details for the file tock_genai_core-0.1.1.tar.gz.

File metadata

  • Download URL: tock_genai_core-0.1.1.tar.gz
  • Upload date:
  • Size: 22.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/2.1.2 CPython/3.9.21 Linux/6.8.0-1021-azure

File hashes

Hashes for tock_genai_core-0.1.1.tar.gz
Algorithm Hash digest
SHA256 de998698bbc5ce37775dbf4462021cdf26d949a2e66a8feee68735e73b2bd2ad
MD5 21e2c451b9bc2a17ddd5581930e9f56a
BLAKE2b-256 f1413fb9439363c75a965bacc23afc92902e2aadf35f4e2ae067dfb045b7cd99

See more details on using hashes here.

File details

Details for the file tock_genai_core-0.1.1-py3-none-any.whl.

File metadata

  • Download URL: tock_genai_core-0.1.1-py3-none-any.whl
  • Upload date:
  • Size: 56.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/2.1.2 CPython/3.9.21 Linux/6.8.0-1021-azure

File hashes

Hashes for tock_genai_core-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 b7bc1bc7b9f547c1aa1112d735c1c98df92df3ace50db9877a9bd2b89ef7f79c
MD5 5c6d0fcaa9b598f7050e2b40c4edcf49
BLAKE2b-256 b0830902cb070754517ea511995b8e8a19fba8ad44cdeed8ed69ce706959926a

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page