nielsr's picture
nielsr HF staff
Improve model card with metadata and details
face8cb verified
|
raw
history blame
3.78 kB
metadata
library_name: transformers
pipeline_tag: text-generation
license: apache-2.0
tags:
  - self-calibration
  - confidence-estimation
  - test-time-scaling

Model Card for Efficient Test-Time Scaling via Self-Calibration

This model implements an efficient test-time scaling method using model confidence for dynamic sampling adjustment. Higher confidence responses have a greater influence on the final answer, leading to improved computational efficiency. The model uses a self-calibration framework to generate more calibrated confidence scores.

Model Details

Model Description

This model utilizes a self-calibration framework to generate calibrated confidence scores, which are then used to improve the efficiency of test-time scaling methods. This allows for comparable performance with substantially fewer computational resources.

  • Developed by: HINT-lab
  • Model type: Large Language Model
  • Language(s) (NLP): English
  • License: Apache-2.0
  • Finetuned from model [optional]: [Specify base model here, e.g., meta-llama/Llama-3.1-8B-Instruct]

Model Sources

Uses

Direct Use

The model can be used directly for text generation tasks, leveraging its self-calibration capabilities for improved efficiency.

Downstream Use

The calibrated confidence scores generated by the model can be incorporated into various test-time scaling methods (e.g., Self-Consistency, Best-of-N) to enhance their performance and reduce computational costs.

Out-of-Scope Use

The model is not intended for tasks requiring high accuracy in scenarios where confidence calibration is not crucial.

Bias, Risks, and Limitations

The model's performance and calibration accuracy may vary depending on the specific dataset and task. Like other LLMs, it may exhibit biases present in its training data.

Recommendations

Users should be aware of the potential biases and limitations of the model and carefully evaluate its performance on their specific tasks. Further investigation into bias mitigation techniques is recommended.

How to Get Started with the Model

See the GitHub README for detailed instructions.

Training Details

Training Data

[Link to Hugging Face Dataset, if available. Otherwise, provide a brief description]

Training Procedure

See the training section in the GitHub README.

Training Hyperparameters

[Information from GitHub README regarding hyperparameters]

Speeds, Sizes, Times

[Information from GitHub README about training times, model sizes, etc.]

Evaluation

Testing Data, Factors & Metrics

Testing Data

[Link to Hugging Face Dataset, if available. Otherwise, provide a brief description]

Factors

[List factors from GitHub README, e.g., different datasets]

Metrics

[List metrics used from GitHub README, e.g., accuracy]

Results

[Summary of results from GitHub README]

Summary

[Concise summary of overall evaluation performance]

Citation

BibTeX:

@misc{huang2025efficienttesttimescalingselfcalibration,
      title={Efficient Test-Time Scaling via Self-Calibration}, 
      author={Chengsong Huang and Langlin Huang and Jixuan Leng and Jiacheng Liu and Jiaxin Huang},
      year={2025},
      eprint={2503.00031},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      url={https://arxiv.org/abs/2503.00031}, 
}

APA:

[APA citation for the paper - Needs to be constructed based on the paper's full details]