Skip to main content

No project description provided

Project description

Patchwork logo Patchwork GIF

Patchwork allows you to automate development gruntwork like PR reviews, bug fixing, security patching, and more using a self-hosted CLI agent and your preferred LLMs.

Key Components

  • Steps: Reusable atomic actions like create PR, commit changes or call an LLM.
  • Prompt Templates: Customizable LLM prompts optimized for a chore like library updates, code generation, issue analysis or vulnerability remediation.
  • Patchflows: LLM-assisted automations such as PR reviews, code fixing, documentation etc. built by combining steps and prompts.

Patchflows can be run locally in your CLI and IDE, or as part of your CI/CD pipeline. There are several patchflows available out of the box, and you can always create your own.

Demo

Patchwork CLI Quickstart

Installation

Using Pip

PatchWork is available on PyPI and can be installed using pip:

pip install 'patchwork-cli[all]' --upgrade

The following optional dependency groups are available.

  • security: installs semgrep and depscan with pip install 'patchwork-cli[security]' and is required for AutoFix and DependencyUpgrade patchflows.
  • rag: installs chromadb with pip install 'patchwork-cli[rag]' and is required for the ResolveIssue patchflow.
  • notifications: Used by steps sending notifications, e.g. slack messages.
  • all: installs everything.
  • not specifying any dependency group (pip install patchwork-cli) will install a core set of dependencies that are sufficient to run the GenerateDocstring, PRReview and GenerateREADME patchflows.

Using Poetry

If you'd like to build from source using poetry, please see detailed documentation here .

PatchWork CLI

The CLI runs Patchflows, as follows:

patchwork <PatchFlow> <?Arguments>

Where

  • Arguments: Allow for overriding default/optional attributes of the Patchflow in the format of key=value. If key does not have any value, it is considered a boolean True flag.

Example

For an AutoFix patchflow which patches vulnerabilities based on a scan using Semgrep:

patchwork AutoFix openai_api_key=<YOUR_OPENAI_API_KEY> github_api_key=<YOUR_GITHUB_TOKEN>

The above command will default to patching code in the current directory, by running Semgrep to identify the vulnerabilities. You can take a look at the default.yml file for the list of configurations you can set to manage the AutoFix patchflow. For more details on how you can use a personal access token from GitHub on CLI you can read this.

You can replace the OpenAI key with a key from our managed service by signing in at https://app.patched.codes/signin and generating an API key from the integrations tab. You can then call the patchflow with the key as follows:

patchwork AutoFix patched_api_key=<YOUR_PATCHED_API_KEY> github_api_key=<YOUR_GITHUB_TOKEN>

Similarly, to use Google's models you can set the google_api_key and model, this is useful if you want to work with large contexts as the gemini-pro-1.5 model supports an input context length of 1 million tokens.

The patchwork-template repository contains the default configuration and prompts for all the patchflows. You can clone that repo and pass it as a flag to the CLI:

patchwork AutoFix --config /path/to/patchwork-configs/patchflows

Using open source models

Patchwork supports any OpenAI compatible endpoint, you can use that to use any LLM from various providers like Groq, Together AI, or Hugging Face.

E.g. to use Llama 3.1 405B from Groq.com run:

patchwork AutoFix client_base_url=https://api.groq.com/openai/v1 openai_api_key=your_groq_key model=llama-3.1-405b-reasoning

You can also use a config file to do the same, e.g. if you want to use Llama 3.1 405B from Hugging Face you can create a config.yml file:

openai_api_key: your_hf_token
client_base_url: https://api-inference.huggingface.co/models/meta-llama/Meta-Llama-3.1-405B-Instruct-FP8/v1
model: Meta-Llama-3.1-405B-Instruct-FP8

And run as:

patchwork AutoFix --config=/path/to/config.yml

This also allows you to run local models via llama.cpp, ollama, vllm or tgi. For instance, you can run Llama 3.1 8B locally using llama_cpp.server as follows:

python -m llama_cpp.server --hf_model_repo_id bullerwins/Meta-Llama-3.1-8B-Instruct-GGUF --model 'Meta-Llama-3.1-8B-Instruct-Q4_K_M.gguf' --chat_format chatml

Then run your patchflow:

patchwork AutoFix client_base_url=https://localhost/v1 openai_api_key=no_key_local_model

Patchflows

Patchwork comes with a set of predefined patchflows, and more will be added over time. Below is a sample list of patchflows:

  • GenerateDocstring: Generate docstrings for methods in your code.
  • AutoFix: Generate and apply fixes to code vulnerabilities in a repository.
  • PRReview: On PR creation, extract code diff, summarize changes, and comment on PR.
  • GenerateREADME: Create a README markdown file for a given folder, to add documentation to your repository.
  • DependencyUpgrade: Update your dependencies from vulnerable to fixed versions.
  • ResolveIssue: Identify the files in your repository that need to be updated to resolve an issue (or bug) and create a PR to fix it.

Prompt Templates

Prompt templates are used by patchflows and passed as queries to LLMs. Templates contain prompts with placeholder variables enclosed by {{}} which are replaced by the data from the steps or inputs on every run.

Below is a sample prompt template:

{
  "id": "diffreview_summary",
    "prompts": [
      {
        "role": "user",
        "content": "Summarize the following code change descriptions in 1 paragraph. {{diffreviews}}"
      }
    ]
}

Each patchflow comes with an optimized default prompt template. But you can specify your own using the prompt_template_file=/path/to/prompt/template/file option.

Contributing

Contributions for new patchflows and steps, or even to the core framework are welcome. Please look at open issues for details.

We also provide chat assistants to help you create new steps and patchflows easily. Fair warning: they suffer from the same limitations as their underlying model.

Roadmap

Short Term

  • Expand patchflow library and integration options
  • Patchflow debugger and validation module
  • Bug fixing and performance improvements
  • Refactor code and documentation

Long Term

  • Support large-scale code embeddings in patchflows
  • Support parallelization and branching
  • Fine-tuned models that can be self-hosted
  • Open-source GUI

License

Patchwork is licensed under AGPL-3.0 terms. However, custom patchflows and steps can be created and shared using the patchwork template repository which is licensed under Apache-2.0 terms.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

patchwork_cli-0.0.34.tar.gz (102.3 kB view details)

Uploaded Source

Built Distribution

patchwork_cli-0.0.34-py3-none-any.whl (167.0 kB view details)

Uploaded Python 3

File details

Details for the file patchwork_cli-0.0.34.tar.gz.

File metadata

  • Download URL: patchwork_cli-0.0.34.tar.gz
  • Upload date:
  • Size: 102.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.3 CPython/3.8.18 Linux/6.5.0-1024-azure

File hashes

Hashes for patchwork_cli-0.0.34.tar.gz
Algorithm Hash digest
SHA256 c65f2efb28e115119ddc1b9410130dc97ab1790d249d8ab783aeb6625f6a14bd
MD5 7504f01f9144a0cf69b00e3964aa8ccf
BLAKE2b-256 c918c208e5fc101f6c665f8d88f613f3df290081d3d2e2f7ff1fb5ce6a7b271b

See more details on using hashes here.

File details

Details for the file patchwork_cli-0.0.34-py3-none-any.whl.

File metadata

  • Download URL: patchwork_cli-0.0.34-py3-none-any.whl
  • Upload date:
  • Size: 167.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.3 CPython/3.8.18 Linux/6.5.0-1024-azure

File hashes

Hashes for patchwork_cli-0.0.34-py3-none-any.whl
Algorithm Hash digest
SHA256 ddddfd6039d261d3e9fd69dcdbb4fe7bada78fa4c794c6cd450fc5b31e7ff577
MD5 1304eaa6f9d208a470647bfdb01249fe
BLAKE2b-256 d4217f842728b675fcad28382cc34f7bb80d6ecb1b13f7e393a4beeb9e6269b4

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page