Datasets:
Update README.md
Browse files
README.md
CHANGED
@@ -72,7 +72,7 @@ BaxBench is a coding benchmark constructed to measure the ability of code genera
|
|
72 |
|
73 |
The dataset is released as part of the paper and benchmark: [BaxBench: Can LLMs generate Correct and Secure Backends?](https://arxiv.org/abs/2502.11844).
|
74 |
|
75 |
-
The dataset contains all necessary artifacts to reproduce the evaluation prompts used in our paper. Further, it enables the testing of different prompt structures or models by forming new prompt types, e.g
|
76 |
|
77 |
For details on reproducing our results, or testing your models on the same prompts, please refer to our [paper](https://arxiv.org/abs/2502.11844) or [code repository](https://github.com/logic-star-ai/baxbench).
|
78 |
|
|
|
72 |
|
73 |
The dataset is released as part of the paper and benchmark: [BaxBench: Can LLMs generate Correct and Secure Backends?](https://arxiv.org/abs/2502.11844).
|
74 |
|
75 |
+
The dataset contains all necessary artifacts to reproduce the evaluation prompts used in our paper. Further, it enables the testing of different prompt structures or models by forming new prompt types, e.g., for testing code agents.
|
76 |
|
77 |
For details on reproducing our results, or testing your models on the same prompts, please refer to our [paper](https://arxiv.org/abs/2502.11844) or [code repository](https://github.com/logic-star-ai/baxbench).
|
78 |
|