Dataset Viewer
model_type
stringclasses 3
values | namespace
stringclasses 3
values | model_name
stringclasses 10
values | training_method
stringclasses 4
values | model_size
int64 357M
16B
| trainable_params
int64 20.5k
16B
| url
stringlengths 64
82
| doi
float64 |
|---|---|---|---|---|---|---|---|
codegen
|
Salesforce
|
codegen-350M-multi
|
fine-tuning
| 356,712,448
| 356,712,448
|
https://huggingface.co/fals3/Salesforce-codegen-350M-multi-unit-test-fine-tuning
| null |
codegen
|
Salesforce
|
codegen2-1B_P
|
fine-tuning
| 1,015,306,240
| 1,015,306,240
|
https://huggingface.co/fals3/Salesforce-codegen2-1B_P-unit-test-fine-tuning
| null |
codegen
|
Salesforce
|
codegen2-3_7B_P
|
fine-tuning
| 3,641,174,016
| 3,641,174,016
|
https://huggingface.co/fals3/Salesforce-codegen2-3_7B_P-unit-test-fine-tuning
| null |
codegen
|
Salesforce
|
codegen2-7B_P
|
fine-tuning
| 6,862,858,240
| 6,862,858,240
|
https://huggingface.co/fals3/Salesforce-codegen2-7B_P-unit-test-fine-tuning
| null |
codegen
|
Salesforce
|
codegen2-16B_P
|
fine-tuning
| 16,032,155,648
| 16,032,155,648
|
https://huggingface.co/fals3/Salesforce-codegen2-16B_P-unit-test-fine-tuning
| null |
llama
|
meta-llama
|
CodeLLama-7b-hf
|
fine-tuning
| 6,738,546,688
| 6,738,546,688
|
https://huggingface.co/fals3/meta-llama-CodeLlama-7b-hf-unit-test-fine-tuning
| null |
starcoder2
|
bigcode
|
starcoder2-7b
|
fine-tuning
| 7,173,923,840
| 7,173,923,840
|
https://huggingface.co/fals3/bigcode-starcoder2-7b-unit-test-fine-tuning
| null |
starcoder2
|
bigcode
|
starcoder2-15b
|
fine-tuning
| 15,957,889,024
| 15,957,889,024
|
https://huggingface.co/fals3/bigcode-starcoder2-15b-unit-test-fine-tuning
| null |
starcoder2
|
bigcode
|
starcoder2-3b
|
fine-tuning
| 3,030,371,328
| 3,030,371,328
|
https://huggingface.co/fals3/bigcode-starcoder2-3b-unit-test-fine-tuning
| null |
starcoder2
|
bigcode
|
starcoderbase
|
fine-tuning
| 15,517,456,384
| 15,517,456,384
|
https://huggingface.co/fals3/bigcode-starcoderbase-unit-test-fine-tuning
| null |
codegen
|
Salesforce
|
codegen-350M-multi
|
ia3
| 356,712,448
| 143,360
|
https://huggingface.co/fals3/Salesforce-codegen-350M-multi-unit-test-ia3
| null |
codegen
|
Salesforce
|
codegen2-1B_P
|
ia3
| 1,015,306,240
| 229,376
|
https://huggingface.co/fals3/Salesforce-codegen2-1B_P-unit-test-ia3
| null |
codegen
|
Salesforce
|
codegen2-3_7B_P
|
ia3
| 3,641,174,016
| 458,752
|
https://huggingface.co/fals3/Salesforce-codegen2-3_7B_P-unit-test-ia3
| null |
codegen
|
Salesforce
|
codegen2-7B_P
|
ia3
| 6,862,858,240
| 917,504
|
https://huggingface.co/fals3/Salesforce-codegen2-7B_P-unit-test-ia3
| null |
codegen
|
Salesforce
|
codegen2-16B_P
|
ia3
| 16,032,155,648
| 1,462,272
|
https://huggingface.co/fals3/Salesforce-codegen2-16B_P-unit-test-ia3
| null |
llama
|
meta-llama
|
CodeLLama-7b-hf
|
ia3
| 6,738,546,688
| 614,400
|
https://huggingface.co/fals3/meta-llama-CodeLlama-7b-hf-unit-test-ia3
| null |
starcoder2
|
bigcode
|
starcoder2-7b
|
ia3
| 7,173,923,840
| 753,664
|
https://huggingface.co/fals3/bigcode-starcoder2-7b-unit-test-ia3
| null |
starcoder2
|
bigcode
|
starcoder2-15b
|
ia3
| 15,957,889,024
| 1,249,280
|
https://huggingface.co/fals3/bigcode-starcoder2-15b-unit-test-ia3
| null |
starcoder2
|
bigcode
|
starcoder2-3b
|
ia3
| 3,030,371,328
| 468,480
|
https://huggingface.co/fals3/bigcode-starcoder2-3b-unit-test-ia3
| null |
starcoder2
|
bigcode
|
starcoderbase
|
ia3
| 15,517,456,384
| 1,239,040
|
https://huggingface.co/fals3/bigcode-starcoderbase-unit-test-ia3
| null |
codegen
|
Salesforce
|
codegen-350M-multi
|
lora
| 356,712,448
| 1,310,720
|
https://huggingface.co/fals3/Salesforce-codegen-350M-multi-unit-test-lora
| null |
codegen
|
Salesforce
|
codegen2-1B_P
|
lora
| 1,015,306,240
| 2,097,152
|
https://huggingface.co/fals3/Salesforce-codegen2-1B_P-unit-test-lora
| null |
codegen
|
Salesforce
|
codegen2-3_7B_P
|
lora
| 3,641,174,016
| 4,194,304
|
https://huggingface.co/fals3/Salesforce-codegen2-3_7B_P-unit-test-lora
| null |
codegen
|
Salesforce
|
codegen2-7B_P
|
lora
| 6,862,858,240
| 8,388,608
|
https://huggingface.co/fals3/Salesforce-codegen2-7B_P-unit-test-lora
| null |
codegen
|
Salesforce
|
codegen2-16B_P
|
lora
| 16,032,155,648
| 13,369,344
|
https://huggingface.co/fals3/Salesforce-codegen2-16B_P-unit-test-lora
| null |
llama
|
meta-llama
|
CodeLLama-7b-hf
|
lora
| 6,738,546,688
| 8,388,608
|
https://huggingface.co/fals3/meta-llama-CodeLlama-7b-hf-unit-test-lora
| null |
starcoder2
|
bigcode
|
starcoder2-7b
|
lora
| 7,173,923,840
| 7,340,032
|
https://huggingface.co/fals3/bigcode-starcoder2-7b-unit-test-lora
| null |
starcoder2
|
bigcode
|
starcoder2-15b
|
lora
| 15,957,889,024
| 12,124,160
|
https://huggingface.co/fals3/bigcode-starcoder2-15b-unit-test-lora
| null |
starcoder2
|
bigcode
|
starcoder2-3b
|
lora
| 3,030,371,328
| 4,546,560
|
https://huggingface.co/fals3/bigcode-starcoder2-3b-unit-test-lora
| null |
starcoder2
|
bigcode
|
starcoderbase
|
lora
| 15,517,456,384
| 8,028,160
|
https://huggingface.co/fals3/bigcode-starcoderbase-unit-test-lora
| null |
codegen
|
Salesforce
|
codegen-350M-multi
|
prompt-tuning
| 356,712,448
| 20,480
|
https://huggingface.co/fals3/Salesforce-codegen-350M-multi-unit-test-prompt-tuning
| null |
codegen
|
Salesforce
|
codegen2-1B_P
|
prompt-tuning
| 1,015,306,240
| 40,960
|
https://huggingface.co/fals3/Salesforce-codegen2-1B_P-unit-test-prompt-tuning
| null |
codegen
|
Salesforce
|
codegen2-3_7B_P
|
prompt-tuning
| 3,641,174,016
| 81,920
|
https://huggingface.co/fals3/Salesforce-codegen2-3_7B_P-unit-test-prompt-tuning
| null |
codegen
|
Salesforce
|
codegen2-7B_P
|
prompt-tuning
| 6,862,858,240
| 81,920
|
https://huggingface.co/fals3/Salesforce-codegen2-7B_P-unit-test-prompt-tuning
| null |
codegen
|
Salesforce
|
codegen2-16B_P
|
prompt-tuning
| 16,032,155,648
| 122,880
|
https://huggingface.co/fals3/Salesforce-codegen2-16B_P-unit-test-prompt-tuning
| null |
llama
|
meta-llama
|
CodeLLama-7b-hf
|
prompt-tuning
| 6,738,546,688
| 81,920
|
https://huggingface.co/fals3/meta-llama-CodeLlama-7b-hf-unit-test-prompt-tuning
| null |
starcoder2
|
bigcode
|
starcoder2-7b
|
prompt-tuning
| 7,173,923,840
| 92,160
|
https://huggingface.co/fals3/bigcode-starcoder2-7b-unit-test-prompt-tuning
| null |
starcoder2
|
bigcode
|
starcoder2-15b
|
prompt-tuning
| 15,957,889,024
| 122,880
|
https://huggingface.co/fals3/bigcode-starcoder2-15b-unit-test-prompt-tuning
| null |
starcoder2
|
bigcode
|
starcoder2-3b
|
prompt-tuning
| 3,030,371,328
| 61,440
|
https://huggingface.co/fals3/bigcode-starcoder2-3b-unit-test-prompt-tuning
| null |
starcoder2
|
bigcode
|
starcoderbase
|
prompt-tuning
| 15,517,456,384
| 122,880
|
https://huggingface.co/fals3/bigcode-starcoderbase-unit-test-prompt-tuning
| null |
PEFT Unit Test Generation Experiments
Dataset description
The PEFT Unit Test Generation Experiments dataset contains metadata and details about a set of trained models used for generating unit tests with parameter-efficient fine-tuning (PEFT) methods. This dataset includes models from multiple namespaces and various sizes, trained with different tuning methods to provide a comprehensive resource for unit test generation research.
Dataset Structure
Data Fields
Each example in the dataset corresponds to a specific trained model variant and includes the following features:
| Feature Name | Description |
|---|---|
model_type |
The type or architecture of the base model (e.g., codegen, starcoder). |
namespace |
The organization or group that created or published the base model (e.g., Salesforce, meta-llama). |
model_name |
The specific name or identifier of the model. |
training_method |
The parameter-efficient fine-tuning method used for training (e.g., full fine-tuning, LoRA, IA³). |
model_size |
The size of the model, typically measured in number of parameters (e.g., 350M, 7B). |
trainable_params |
The number of trainable parameters for the specific tuning method and hyperparameters. |
url |
A direct link to the model repository. |
doi |
The digital object identifier associated with the trained model. |
Dataset Details
Dataset Description
Training Hyperparameters
Model-agnostic Hyperparameters
| Hyperparameter | Method | Value |
|---|---|---|
| Common | ||
| Optimizer | - | AdamW |
| LR schedule | - | Linear |
| LR warmup ratio | - | 0.1 |
| Batch size | - | 1 |
| Gradient accumulation steps | - | 8 |
| # Epochs | - | 3 |
| Precision | - | Mixed |
| Learning rate | Full fine-tuning | 5E-5 |
| LoRA | 3E-4 | |
| (IA)3 | 3E-4 | |
| Prompt tuning | 3E-3 | |
| Method specific | ||
| Alpha | LoRA | 32 |
| Dropout | LoRA | 0.1 |
| Rank | LoRA | 16 |
| Virtual tokens | Prompt tuning | 20 |
Model-specific Hyperparameters
| Hyperparameter | Method | Model | Value |
|---|---|---|---|
| Targeted attention modules | LoRA, (IA)3 | codegen-350M-multi | qkv_proj |
| Salesforce/codegen2-1B_P | qkv_proj | ||
| Salesforce/codegen2-3_7B_P | qkv_proj | ||
| Salesforce/codegen2-7B_P | qkv_proj | ||
| Salesforce/codegen2-16B_P | qkv_proj | ||
| meta-llama/CodeLlama-7b-hf | q_proj, v_proj | ||
| bigcode/starcoderbase | c_attn | ||
| bigcode/starcoder2-3b | q_proj, v_proj | ||
| bigcode/starcoder2-7b | q_proj, v_proj | ||
| bigcode/starcoder2-15b | q_proj, v_proj | ||
| Targeted feedforward modules | (IA)3 | codegen-350M-multi | fc_out |
| Salesforce/codegen2-1B_P | fc_out | ||
| Salesforce/codegen2-3_7B_P | fc_out | ||
| Salesforce/codegen2-7B_P | fc_out | ||
| Salesforce/codegen2-16B_P | fc_out | ||
| meta-llama/CodeLlama-7b-hf | down_proj | ||
| bigcode/starcoderbase | mlp.c_proj | ||
| bigcode/starcoder2-3b | q_proj, c_proj | ||
| bigcode/starcoder2-7b | q_proj, c_proj | ||
| bigcode/starcoder2-15b | q_proj, c_proj |
Training Runs
- Downloads last month
- 6



