Skip to main content

The LLM-Native Agent Framework

Project description

Native

Agentic workflows uses 2 steps

  1. Agent loops: which handles core operation of taking an input and giving an output in micro level.

  2. Agentic Core Components: peripherals for agentic workflows which enhances agentic workflow into a system level. and give them actual useful problem solving capabilities

Agent loop consists of core internal flow :

intent → plan → act → interpret → loop reflection - > memory ↺

And Agentic Core Components are abstarct features such like: Multilayer Context Reflection Extensive Planning
Tool Orchestration Self-Modeling Dynamic Memory Integrated Awareness Holistic Memory Autonomous Productive Capacity

there are different approaches to achive/reach these Core Components. And all frameworks treats LLM operations as not enough and support and fill between the operations with lots of python code.

This has 2 harmful outcomes. First is scalability problems. To scale your agent you should enhance both agent loop codes and your code resposible of creating abstarct features. Which is not maintainable on the long run.

Second harmful outcome is about chaining and self calling and recursive looping. Such manual code based scaffolding makes chaining and self calling non-native and tool calling like.

So with Native what we do is like every organ in the body is composed of cells and cells are made from atoms, we maintain similar structure in the overall agent by composing each piece from LLMs.
Agent loop components (intent, plan , act , interpret , loop reflection ) are all composed of LLM calls. We dont use python scaffold around everything and this is amazing.

our framework has 2 slogans

  • minimal glue code, maximal LLM calls
  • When GLue is absolutely needed, we also use an LLM for glueing

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

homegrown-0.0.2.tar.gz (2.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

homegrown-0.0.2-py3-none-any.whl (2.1 kB view details)

Uploaded Python 3

File details

Details for the file homegrown-0.0.2.tar.gz.

File metadata

  • Download URL: homegrown-0.0.2.tar.gz
  • Upload date:
  • Size: 2.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.9.22

File hashes

Hashes for homegrown-0.0.2.tar.gz
Algorithm Hash digest
SHA256 eddfbc745a4b1e4a355160b942ca66d94ab73e98409471a7437871ca29d99e6c
MD5 5f76d570dabf4cb79438b67bda4f0bc6
BLAKE2b-256 913bb207a14a6e21c8f3eeaecbc23cceaf3df2f46024459e5a375f3e14d444ff

See more details on using hashes here.

File details

Details for the file homegrown-0.0.2-py3-none-any.whl.

File metadata

  • Download URL: homegrown-0.0.2-py3-none-any.whl
  • Upload date:
  • Size: 2.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.9.22

File hashes

Hashes for homegrown-0.0.2-py3-none-any.whl
Algorithm Hash digest
SHA256 66142cbc354f0ef12b810836dd9b3867c4eb074fd7d7894e8064c3997dd098af
MD5 8eaeafced2a5c3e08be3fb3b75e25508
BLAKE2b-256 def12e272c407dabacfc84bb760dd59340000ad843229738cd8b51f1091951f8

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page