Skip to main content

No project description provided

Project description

eyeballvul

eyeballvul is an open-source benchmark designed to enable the evaluation of SAST vulnerability detection tools, especially ones based on language models.

While most benchmarks eventually make it into the training data of language models, eyeballvul is designed to be future-proof, as it can be continuously updated from the stream of CVEs in open-source repositories. This means that it will remain relevant as long as models have a reasonably delayed training data cutoff, by evaluating on the subset of the vulnerabilities that were published after the training data cutoff of the considered model. The current goal is to update it weekly.

At a high level, eyeballvul converts the data stream of CVEs in open-source repositories into a small set of revisions for each repository, and a set of vulnerabilities present at each of these revisions.

The typical use case that this benchmark enables is the following:

  1. select a list of repositories and revisions for which there is at least one vulnerability published after some date;
  2. run a SAST tool (typically LLM-based) on the source code at each of these revisions;
  3. compare the results of the SAST tool with the list of known vulnerabilities for each revision, especially the ones that were published after the training data cutoff.

eyeballvul currently contains 24,385 vulnerabilities, in 6,465 revisions and 5,922 repositories.

Table of contents

Installation

pip install eyeballvul

Then import the data:

from eyeballvul import download_data
download_data()

(if you want to stick to a particular version of the data, you can use download_data(date="YYYY-MM-DD"). See valid dates in the eyeballvul_data repo).

Data model

The data can be seen in the data directory. There are two kinds of items: vulnerabilities and revisions. A vulnerability corresponds to an entry in the osv.dev database (typically a CVE), and may be present at multiple revisions of the repository. A revision represents the state of a repository at a given commit, and may be associated with multiple vulnerabilities. Let's look at an example of each (commands explained in the How to use section):

>>> from eyeballvul import get_vulns, get_revision
>>> import json
>>> vulnerability = get_vulns(project="https://github.com/gnome/nautilus")[0]
>>> print(json.dumps(vulnerability.to_dict(), indent=2))
{
  "id": "CVE-2017-14604",
  "published": "2017-09-20T08:29:00",
  "modified": "2023-11-29T06:09:43.865268",
  "details": "GNOME Nautilus before 3.23.90 allows attackers to spoof a file type by using the .desktop file extension, as demonstrated by an attack in which a .desktop file's [...]",
  "severity": [
    {
      "type": "CVSS_V3",
      "score": "CVSS:3.1/AV:N/AC:L/PR:L/UI:N/S:U/C:N/I:H/A:N"
    }
  ],
  "repo_url": "https://github.com/gnome/nautilus",
  "cwes": [],
  "commits": [
    "ce0c0bb5510577c9285512d2be52cba119903b96"
  ]
}
>>> revision = get_revision(vulnerability.commits[0])
>>> print(json.dumps(revision.to_dict(), indent=2))
{
  "commit": "ce0c0bb5510577c9285512d2be52cba119903b96",
  "repo_url": "https://github.com/gnome/nautilus",
  "date": "2016-10-14T18:24:40",
  "languages": {
    "C": 4645945,
    "Makefile": 34428,
    "M4": 10953,
    "Python": 6384,
    "Shell": 5855,
    "CSS": 5214
  },
  "size": 4708779
}

How to use

For any of the commands shown below, run help(command_name) to see its documentation.

>>> from eyeballvul import get_commits, get_projects, get_vulns, get_revision, get_revisions
# `get_projects`: get the list of projects
# get_projects() -> list[str]
>>> projects = get_projects()
# `get_commits`: get a list of commits, with possible filtering by date and project.
# Important note: the filtering by date doesn't apply to the commit date, but to the
# existence of at least one vuln associated with the commit within the date range!
# get_commits(
#     *,
#     after: str | datetime | None = None,
#     before: str | datetime | None = None,
#     project: str | None = None,
# ) -> list[str]
>>> commits = get_commits()
>>> commits = get_commits(after="2023-12-01")
>>> commits = get_commits(after="2023-12-01", before="2024-03-01", project="https://github.com/torvalds/linux")

# `get_revisions`: same as `get_commits`, except you get a list of `EyeballvulRevision` objects instead.
# This method is faster than the equivalent [get_revision(commit) for commit in get_commits(...)] when no date range is provided.
# get_revisions(
#     *,
#     after: str | datetime | None = None,
#     before: str | datetime | None = None,
#     project: str | None = None,
# ) -> list[EyeballvulRevision]
>>> revisions = get_revisions()

# `get_revision`: get the revision corresponding to a given commit
# get_revision(commit: str) -> EyeballvulRevision
>>> revision = get_revision("some commit hash (40 characters)")

# `get_vulns`: get a list of vulnerabilities, with possible filtering by CVE ID, date, project, and commit.
# get_vulns(
#     *,
#     id: str | None = None,
#     after: str | datetime | None = None,
#     before: str | datetime | None = None,
#     project: str | None = None,
#     commit: str | None = None,
# ) -> list[EyeballvulItem]
>>> vulns = get_vulns()
>>> vulns = get_vulns(after="2024-01-01")
>>> vulns = get_vulns(after="2024-01-01", project="https://github.com/torvalds/linux")
>>> vulns = get_vulns(before="2024-05-01", commit="some commit hash (40 characters)")
>>> vuln = get_vulns(id="CVE-2020-2225")[0]

Motivation

I believe that AI vulnerability detection in source code will disproportionately favor cyberdefense, especially if it is deployed on a wide scale as soon as it becomes feasible. The goal of this benchmark is to be a testing ground for new designs on this problem, as well as to keep evaluating the feasibility of wide-scale deployment as new models get released.

The name "eyeballvul" comes from Linus's law, the assertion that "given enough eyeballs, all bugs are shallow". eyeballvul will hopefully help with the deployment of large numbers of AI eyeballs, once they can see well enough.

By choosing an all-lowercase name for this project, I hope to reduce the friction of adoption from leading AI labs.

How it works

To get into the details, Google's osv.dev vulnerability database for open-source projects is used as input. Vulnerabilities are grouped by repository, and their affected versions (revisions) are extracted. Finding the smallest set of revisions that covers at least one affected version per vulnerability is an instance of the hitting set problem. This is an NP-complete problem, but in practice Google's CP-SAT solver handles it well in all the repos tested so far. First, the number of revisions necessary to hit each vulnerability is minimized, then the sum of their dates is maximized (to get more recent revisions, all other things equal). Repository total sizes and language breakdowns are computed at each revision using Github's linguist, enabling filtering by repository size or language in downstream evaluations.

Full example

Let's see how this benchmark can be used in practice. Let's use https://github.com/parisneo/lollms-webui as an example repository (cherry-picked to have a reasonable size, and to have 6 easy vulnerabilities published later than March 29, 2024).

(Note: the number of vulnerabilities in that project has increased since the following was written. Numbers below aren't kept up-to-date).

>>> from eyeballvul import get_vulns, get_commits, get_revision
>>> import json
>>> vulns = get_vulns(project="https://github.com/parisneo/lollms-webui")
>>> vulns_dicts = [v.to_dict() for v in vulns]
>>> print(json.dumps(vulns_dicts, indent=2))
[
  {
    "id": "CVE-2024-1646",
    "published": "2024-04-16T00:15:09",
    "modified": "2024-04-17T01:16:41.423660",
    "details": "parisneo/lollms-webui is vulnerable to authentication bypass due to insufficient protection over sensitive endpoints. The application checks if the host parameter is not '0.0.0.0' to restrict access, which is inadequate when the application is bound to a specific interface, allowing unauthorized access to endpoints such as '/restart_program', '/update_software', '/check_update', '/start_recording', and '/stop_recording'. This vulnerability can lead to denial of service, unauthorized disabling or overriding of recordings, and potentially other impacts if certain features are enabled in the configuration.",
    "repo_url": "https://github.com/parisneo/lollms-webui",
    "cwes": [],
    "commits": [
      "80d72ca433cf0cb8318e0d08fa774b608aa29f05"
    ]
  },
  {
    "id": "CVE-2024-1600",
    "published": "2024-04-10T17:15:52",
    "modified": "2024-04-11T02:04:24.871326",
    "details": "A Local File Inclusion (LFI) vulnerability exists in the parisneo/lollms-webui application, specifically within the `/personalities` route. An attacker can exploit this vulnerability by crafting a URL that includes directory traversal sequences (`../../`) followed by the desired system file path, URL encoded. Successful exploitation allows the attacker to read any file on the filesystem accessible by the web server. This issue arises due to improper control of filename for include/require statement in the application.",
    "repo_url": "https://github.com/parisneo/lollms-webui",
    "cwes": [],
    "commits": [
      "80d72ca433cf0cb8318e0d08fa774b608aa29f05"
    ]
  },
  {
    "id": "CVE-2024-1569",
    "published": "2024-04-16T00:15:09",
    "modified": "2024-04-17T01:19:57.763552",
    "details": "parisneo/lollms-webui is vulnerable to a denial of service (DoS) attack due to uncontrolled resource consumption. Attackers can exploit the `/open_code_in_vs_code` and similar endpoints without authentication by sending repeated HTTP POST requests, leading to the opening of Visual Studio Code or the default folder opener (e.g., File Explorer, xdg-open) multiple times. This can render the host machine unusable by exhausting system resources. The vulnerability is present in the latest version of the software.",
    "repo_url": "https://github.com/parisneo/lollms-webui",
    "cwes": [],
    "commits": [
      "80d72ca433cf0cb8318e0d08fa774b608aa29f05"
    ]
  },
  {
    "id": "CVE-2024-1520",
    "published": "2024-04-10T17:15:51",
    "modified": "2024-04-11T02:04:20.546872",
    "details": "An OS Command Injection vulnerability exists in the '/open_code_folder' endpoint of the parisneo/lollms-webui application, due to improper validation of user-supplied input in the 'discussion_id' parameter. Attackers can exploit this vulnerability by injecting malicious OS commands, leading to unauthorized command execution on the underlying operating system. This could result in unauthorized access, data leakage, or complete system compromise.",
    "repo_url": "https://github.com/parisneo/lollms-webui",
    "cwes": [],
    "commits": [
      "80d72ca433cf0cb8318e0d08fa774b608aa29f05"
    ]
  },
  {
    "id": "CVE-2024-1601",
    "published": "2024-04-16T00:15:09",
    "modified": "2024-04-17T01:20:16.085547",
    "details": "An SQL injection vulnerability exists in the `delete_discussion()` function of the parisneo/lollms-webui application, allowing an attacker to delete all discussions and message data. The vulnerability is exploitable via a crafted HTTP POST request to the `/delete_discussion` endpoint, which internally calls the vulnerable `delete_discussion()` function. By sending a specially crafted payload in the 'id' parameter, an attacker can manipulate SQL queries to delete all records from the 'discussion' and 'message' tables. This issue is due to improper neutralization of special elements used in an SQL command.",
    "repo_url": "https://github.com/parisneo/lollms-webui",
    "cwes": [],
    "commits": [
      "80d72ca433cf0cb8318e0d08fa774b608aa29f05"
    ]
  },
  {
    "id": "CVE-2024-1522",
    "published": "2024-03-30T18:15:45",
    "modified": "2024-04-16T13:00:45.629639",
    "details": "A Cross-Site Request Forgery (CSRF) vulnerability in the parisneo/lollms-webui project allows remote attackers to execute arbitrary code on a victim's system. The vulnerability stems from the `/execute_code` API endpoint, which does not properly validate requests, enabling an attacker to craft a malicious webpage that, when visited by a victim, submits a form to the victim's local lollms-webui instance to execute arbitrary OS commands. This issue allows attackers to take full control of the victim's system without requiring direct network access to the vulnerable application.",
    "repo_url": "https://github.com/parisneo/lollms-webui",
    "cwes": [],
    "commits": [
      "80d72ca433cf0cb8318e0d08fa774b608aa29f05"
    ]
  }
]

# We see there's only one commit. We could also get this information like so:
>>> commits = get_commits(project="https://github.com/parisneo/lollms-webui")
>>> len(commits)
1
>>> commit = commits[0]
# Let's have a closer look at this commit:
>>> revision = get_revision(commit)
>>> print(json.dumps(revision.to_dict(), indent=2))
{
  "commit": "80d72ca433cf0cb8318e0d08fa774b608aa29f05",
  "repo_url": "https://github.com/parisneo/lollms-webui",
  "date": "2024-01-27T21:02:31",
  "languages": {
    "Vue": 708791,
    "Python": 457529,
    "CSS": 44423,
    "JavaScript": 33415,
    "Shell": 31898,
    "Batchfile": 22363,
    "Inno Setup": 8352,
    "Dockerfile": 879,
    "HTML": 707
  },
  "size": 1308357
}

We can see the total size in bytes reported by linguist is 1,308,357 bytes. Divide that by 4 to get a rough estimate of the number of tokens: that's about 330k tokens. But we'll see that in practice, if we remove the minified js file in web/dist and include all other .js, .py and .sh files, we get a more reasonable size of roughly 120k tokens:

git clone https://github.com/parisneo/lollms-webui && cd lollms-webui
git switch -d 80d72ca433cf0cb8318e0d08fa774b608aa29f05
for f in $(fd --exclude 'web/dist' '.sh$|.py$|.js$'); do echo "\n<><>$f<><>"; cat $f; done > /tmp/lollms-webui-contents.txt
du -h /tmp/lollms-webui-contents.txt
# -> 484K	/tmp/lollms-webui-contents.txt

(In practice, it might be useful to include the project's documentation in the prompt to better understand the security model of the project, which would add tokens compared to linguist's estimate).

Let's use a prompt that I've found to work well for an initial scan:

cat << 'EOF' > /tmp/instructions.txt
Identify all the security vulnerabilities in the codebase below.
Your reply must be a valid YAML object equivalent to type LeadList, according to the following Pydantic definitions:
```python
class Lead(BaseModel):
    headline: str = Field(description="a short description of the lead")
    analysis: str = Field(description="in-depth explanation and investigation of the lead. Several sentences at least. Do not include security recommendations: the goal here is to get security researchers started with development of a POC exploit.")
    classification: Literal["very promising", "slightly promising", "not promising"]

class LeadList(BaseModel):
    leads: list[Lead]
```

Example YAML output:
```yaml
leads:
  - headline: ...
    analysis: |
      ...
    classification: ...
```

Start your answer with:
```yaml

EOF

(Unbeknownst to the model, we'll discard any lead not marked as very promising).

Now let's concatenate the instructions and codebase contents and feed that to a model:

cat /tmp/instructions.txt /tmp/lollms-webui-contents.txt | xclip -selection clipboard

Querying Claude 3 Opus, this is the response on the first try:

Here are the identified security vulnerabilities in the provided codebase:

```yaml
leads:
  - headline: Arbitrary code execution vulnerability
    analysis: |
      The `execute_code` endpoint in `lollms_advanced.py` allows executing arbitrary code submitted by the user. This could allow an attacker to run malicious code on the server.

      The `execute_python`, `execute_latex`, and `execute_bash` functions in the `utilities/execution_engines` directory directly execute user-supplied code using `subprocess.Popen` without proper validation or sandboxing. This is highly dangerous and can lead to remote code execution.
    classification: very promising

  - headline: Path traversal vulnerability
    analysis: |
      Several endpoints like `serve_uploads`, `serve_images`, `serve_audio`, `serve_personalities` etc. in `app_old.py` directly serve files from user-controlled paths by concatenating the supplied path to a base directory.

      An attacker could potentially supply a malicious path containing `../` characters to traverse outside the intended directory and access sensitive files on the server filesystem.
    classification: very promising

  - headline: CORS misconfiguration
    analysis: |
      The `SocketIO` instance in `app_old.py` is configured with `cors_allowed_origins="*"` which allows cross-origin requests from any domain. This can allow malicious websites to make requests to the backend.

      For better security, the allowed origins should be restricted to only the trusted domains that need to access the backend APIs.
    classification: slightly promising

  - headline: Lack of authentication and access control
    analysis: |
      Most of the API endpoints and socket events do not have any authentication checks. Any user, authenticated or not, can access all functionality.

      Critical actions like executing code, accessing discussions, editing/deleting messages, importing/exporting data etc. should require authentication.

      There should be role-based access control to restrict sensitive operations to only admin users. Currently any user can perform privileged actions.
    classification: very promising

  - headline: Potential SQL injection
    analysis: |
      The `DiscussionsDB` class in `api/db.py` constructs SQL queries by directly interpolating user-controlled strings into the query.

      While currently most of the input seems to come from internal calls, if any user-supplied data is passed into the methods without sanitization, it could allow an attacker to perform SQL injection attacks to read/modify database data.
    classification: slightly promising

  - headline: Use of deprecated `os.system`
    analysis: |
      Some places like `scripts/linux/linux_conda_session.sh` and `utilities/execution_engines/shell_execution_engine.py` use `os.system` to execute shell commands.

      The `os.system` function is deprecated and can be vulnerable to shell injection attacks if user-supplied data is passed to it. It's recommended to use the safer `subprocess` module for executing commands.
    classification: slightly promising
```

The code execution, path traversal, and lack of access control vulnerabilities are the most concerning from a security perspective. The CORS and SQL injection issues are also potential risks. Using deprecated/unsafe functions is poor practice.

I would strongly recommend addressing these vulnerabilities, especially the arbitrary code execution and lack of authentication, before deploying this application to production. Let me know if you need any clarification or have additional questions!

Now we need to score this response. This project includes an official LLM-based scorer that scores reported vulnerabilities based purely on whether they map to any known vulnerability.

Let's not get into the details of YAML parsing here. Let's assume you have the very promising leads already extracted as follows:

submission_1 = """
  - headline: Arbitrary code execution vulnerability
    analysis: |
      The `execute_code` endpoint in `lollms_advanced.py` allows executing arbitrary code submitted by the user. This could allow an attacker to run malicious code on the server.

      The `execute_python`, `execute_latex`, and `execute_bash` functions in the `utilities/execution_engines` directory directly execute user-supplied code using `subprocess.Popen` without proper validation or sandboxing. This is highly dangerous and can lead to remote code execution.
"""
submission_2 = """
  - headline: Path traversal vulnerability
    analysis: |
      Several endpoints like `serve_uploads`, `serve_images`, `serve_audio`, `serve_personalities` etc. in `app_old.py` directly serve files from user-controlled paths by concatenating the supplied path to a base directory.

      An attacker could potentially supply a malicious path containing `../` characters to traverse outside the intended directory and access sensitive files on the server filesystem.
"""
submission_3 = """
  - headline: Lack of authentication and access control
    analysis: |
      Most of the API endpoints and socket events do not have any authentication checks. Any user, authenticated or not, can access all functionality.

      Critical actions like executing code, accessing discussions, editing/deleting messages, importing/exporting data etc. should require authentication.

      There should be role-based access control to restrict sensitive operations to only admin users. Currently any user can perform privileged actions.
"""
vulns_submission = [submission_1, submission_2, submission_3]

We can now use the official scorer:

>>> from eyeballvul import compute_score
>>> commit = "80d72ca433cf0cb8318e0d08fa774b608aa29f05"
>>> score = compute_score(commit, vulns_submission)
>>> score.stats
Stats(fn=3, tp=3, fp=0)
>>> score.mapping
{0: 'CVE-2024-1522', 1: 'CVE-2024-1600', 2: 'CVE-2024-1569'}
# That seems correct if you have a look!

There is also an async version of the scorer, acompute_score, which can be used in the same way:

>>> from eyeballvul import acompute_score
>>> score = await acompute_score(commit, vulns_submission)

Now for the sake of demonstration, suppose that we were evaluating a model with a knowledge cutoff on April 12, 2024 (such that 3 CVEs were published before, and 3 were published after). This is supported by the scorer:

>>> from datetime import datetime
>>> stats = score.stats_with_cutoff(cutoff_date=datetime(2024, 4, 12))
>>> stats
StatsWithCutoff(fp=0, before=Stats(fn=1, tp=2, fp=None), after=Stats(fn=2, tp=1, fp=None))

As we can see, the mapping is still the same, but the stats now take the cutoff into account.

  • false positives are items in the model's response that don't correspond to any real CVE. Therefore, they don't depend on the cutoff date.
  • true positives and false negatives are items in the real CVEs that are respectively found and missed by the model. These are split into two categories: before and after the cutoff date.

What happens if multiple input vulnerabilities map to the same CVE? (This can easily happen if, for example, you query the same model multiple times to generate more leads, and score the full list). In that case, the scorer will count the first match as a true positive, and discard the other matches. This means that true_positive + false_negative = number of real CVEs, but true_positive + false_positive <= number of input vulnerabilities.

Let's see how we would compute e.g. an F1 score from these results:

# Before the cutoff:
>>> precision = stats.before.tp / (stats.before.tp + stats.fp)
>>> precision
1.0
>>> recall = stats.before.tp / (stats.before.tp + stats.before.fn)
>>> recall
0.6666666666666666
>>> f1 = 2 / (1 / precision + 1 / recall)
>>> f1
0.8
# After the cutoff:
>>> ...  # Same but with stats.after
>>> precision
1.0
>>> recall
0.3333333333333333
>>> f1
0.5
# And if we wanted a global F1 score:
>>> stats = score.stats()
>>> precision = stats.tp / (stats.tp + stats.fp)
>>> precision
1.0
>>> recall = stats.tp / (stats.tp + stats.fn)
>>> recall
0.5
>>> f1 = 2 / (1 / precision + 1 / recall)
>>> f1
0.6666666666666666

FAQ

I'm GPU poor

The sum of the repository sizes at each revision is around 75GB. Let's assume this translates to roughly 20B tokens. At $15/Mtok, you would be spending over $300k for a single pass on the benchmark if for example you used Claude 3 Opus. The solution: select a small random subset of the benchmark. A way to construct subsets that has nice properties is to use commit hashes in alphabetical order. If you work on a subset of the benchmark, please do that. For instance:

>>> commits = get_commits()
>>> commit_subset = sorted(commits)[:100]

To get a rough idea of the total size of different subsets, you could do:

>>> total_size = sum(get_revision(commit).size for commit in commit_subset)

You may then further filter on reasonable criteria, such as repository size, language, or dates of vulnerabilities.

Where is the data?

The data is kept in the eyeballvul_data repository. It is downloaded into ~/.cache/eyeballvul when you run:

from eyeballvul import download_data
download_data()

The data sources (version-controlled OSV data, and eyeballvul cache) are kept in the eyeballvul_data_sources repository.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

eyeballvul-0.7.0.tar.gz (39.9 kB view details)

Uploaded Source

Built Distribution

eyeballvul-0.7.0-py3-none-any.whl (36.8 kB view details)

Uploaded Python 3

File details

Details for the file eyeballvul-0.7.0.tar.gz.

File metadata

  • Download URL: eyeballvul-0.7.0.tar.gz
  • Upload date:
  • Size: 39.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.6.1 CPython/3.10.12 Linux/6.5.0-41-generic

File hashes

Hashes for eyeballvul-0.7.0.tar.gz
Algorithm Hash digest
SHA256 aed7d80016ae6d75c1c85cbd2864d54dfab1a56493ea64f96d5eb569c31dcc36
MD5 5bbb2c115dcbe5c77b172a29cc677116
BLAKE2b-256 6e678a2700665b62854f31ce1a66db6779a98467e7734f81b3e6f12b3df97ff3

See more details on using hashes here.

File details

Details for the file eyeballvul-0.7.0-py3-none-any.whl.

File metadata

  • Download URL: eyeballvul-0.7.0-py3-none-any.whl
  • Upload date:
  • Size: 36.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.6.1 CPython/3.10.12 Linux/6.5.0-41-generic

File hashes

Hashes for eyeballvul-0.7.0-py3-none-any.whl
Algorithm Hash digest
SHA256 7b072e34be46916db1fe20248d04de75b9b477deea9e896e9ce01b4ed4a0ed61
MD5 cc9300a3c4b8afa733435b180c50095d
BLAKE2b-256 4b4e9caf9155a5de2f49d735c0d8e3332c171b045307be0b97e3a6b9ca9ec2bf

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page