No project description provided
Project description
debugging-benchmark
Quickstart
Generating Passing and Failing Inputs:
from debugging_benchmark.calculator.calculator import CalculatorBenchmarkRepository
from debugging_framework.tools import GrammarBasedEvaluationFuzzer
calc = CalculatorBenchmarkRepository().build()
param = calc.to_dict()
fuzzer = GrammarBasedEvaluationFuzzer(**param)
fuzzer.run()
gen_inps = fuzzer.get_generated_inputs()
Evaluation:
from debugging_benchmark.calculator.calculator import CalculatorBenchmarkRepository
from debugging_framework.evaluator import Evaluation
from debugging_framework.tools import InputsFromHellEvaluationFuzzer
tools = [InputsFromHellEvaluationFuzzer]
subjects = SieveOfEratosthenesStudentAssignmentBenchmarkRepository().build()
result = Evaluation(
tools=tools,
subjects=subjects[0:1],
repetitions=1,
timeout=3600
).run()
Deeper Look into the Class Structure
Check out the Class Diagram for a first overview. Further down in this section we take a look at some key functions of interest.
Class Diagram
BenchmarkRepository
and BenchmarkProgram
can be found in debugging_framework/benchmark.py
StudentAssignmentBenchmarkProgram
,StudentAssignmentRepository
and GCDStudentAssignmentBenchmarkRepository
can be found in debugging_benchmark/student_assignments.py
The faulty programs can be found at debugging_benchmark/student_assignments/problem_1_GCD
and the correct implementation at debugging_benchmark/student_assignments/reference1.py
build()
Returns a List of BenchmarkPrograms. Calls internally _construct_test_program(). This function is our interface.
_construct_test_program()
Returns a BenchmarkProgram. Calls internally construct_oracle() to construct a oracle for our program.
construct_oracle()
Where the magic happens. Returns a Functions which loads the faulty and correct implementation, executes both with the input and checks if they are the same or not. If they are the same return OracleResult.PASSING, if not return OracleResult.FAILING
to_dict()
PLACEHOLDER
Install, Development, Testing
Install
If all external dependencies are available, a simple pip install PLACEHOLDER suffices. We recommend installing PLACEHOLDER inside a virtual environment (virtualenv):
python3.10 -m venv venv
source venv/bin/activate
pip install --upgrade pip
pip install PLACEHOLDER
Development and Testing
For development and testing, we recommend using PLACEHOLDER inside a virtual environment (virtualenv). By doing the following steps in a standard shell (bash), one can run the PLACEHOLDER tests:
git clone https://github.com/martineberlein/debugging-benchmark
cd debugging-benchmark
python3.10 -m venv venv
source venv/bin/activate
pip install --upgrade pip
# Run tests
pip install -e .[dev]
python3 -m pytest
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file debugging-benchmark-0.0.1.tar.gz
.
File metadata
- Download URL: debugging-benchmark-0.0.1.tar.gz
- Upload date:
- Size: 302.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.10.9
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | c2e12b22c89cf549291374fe0a1e195ffeba31b9dada117315a5928f7a1e0f9c |
|
MD5 | 326f712c505c585fdda2412b1661eccd |
|
BLAKE2b-256 | b6aee0c7b100c62d5d6e40bc7a0f10188ffcbd1d2aed61f81f14219e1e521be9 |
File details
Details for the file debugging_benchmark-0.0.1-py3-none-any.whl
.
File metadata
- Download URL: debugging_benchmark-0.0.1-py3-none-any.whl
- Upload date:
- Size: 1.8 MB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.10.9
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 681fdd1ecbf62b9c13d7c0a5fbd2c5b61cd4d30dc2c9a3b96020eab55ccee6a4 |
|
MD5 | d9801a6e01eeb569f99a93bfd7a66492 |
|
BLAKE2b-256 | 3a502d8f37973242407cecf6c507a7d0cf65606605889f0b2468f32687fa9ef5 |