|
--- |
|
license: apache-2.0 |
|
language: |
|
- en |
|
- zh |
|
tags: |
|
- context compression |
|
- sentence selection |
|
- probing classifier |
|
- attention probing |
|
- RAG |
|
- LongBench |
|
pipeline_tag: text-classification |
|
--- |
|
|
|
# Sentinel Probing Classifier (Logistic Regression) |
|
|
|
This repository contains the sentence-level classifier used in **Sentinel**, a lightweight context compression framework introduced in our paper: |
|
|
|
> **Sentinel: Attention Probing of Proxy Models for LLM Context Compression with an Understanding Perspective** |
|
> Yong Zhang, Yanwen Huang, Ning Cheng, Yang Guo, Yun Zhu, Yanmeng Wang, Shaojun Wang, Jing Xiao |
|
> π [Paper (Arxiv 2025)](https://arxiv.org/abs/2505.23277)β|βπ» [Code on GitHub](https://github.com/yzhangchuck/Sentinel) |
|
|
|
--- |
|
|
|
## π§ What is Sentinel? |
|
|
|
**Sentinel** reframes LLM context compression as a lightweight attention-based *understanding* task. Instead of fine-tuning a full compression model, it: |
|
|
|
- Extracts **decoder attention** from a small proxy LLM (e.g., Qwen-2.5-0.5B) |
|
- Computes **sentence-level attention features** |
|
- Applies a **logistic regression (LR) classifier** to select relevant sentences |
|
|
|
This approach is efficient, model-agnostic, and highly interpretable. |
|
|
|
--- |
|
|
|
## π¦ Files Included |
|
|
|
| File | Description | |
|
|-------------------------|----------------------------------------------| |
|
| `sentinel_lr_model.pkl` | Trained logistic regression classifier | |
|
| `sentinel_config.json` | Feature extraction configuration | |
|
|
|
--- |
|
|
|
## π Usage |
|
|
|
Use this classifier on attention-derived feature vectors to predict sentence-level relevance scores: |
|
|
|
π Feature extraction code and full pipeline available at: |
|
π https://github.com/yzhangchuck/Sentinel |
|
|
|
## π Benchmark Results |
|
<p align="center"> |
|
<img src="longbench_gpt35.png" alt="LongBench GPT-3.5 Results" width="750"/> |
|
</p> |
|
|
|
|
|
<p align="center"> |
|
<img src="longbench_qwen7b.png" alt="LongBench Qwen Results" width="750"/> |
|
</p> |
|
|
|
|
|
## π Citation |
|
Please cite us if you use this model: |
|
|
|
@misc{zhang2025sentinelattentionprobingproxy, |
|
title={Sentinel: Attention Probing of Proxy Models for LLM Context Compression with an Understanding Perspective}, |
|
author={Yong Zhang and Yanwen Huang and Ning Cheng and Yang Guo and Yun Zhu and Yanmeng Wang and Shaojun Wang and Jing Xiao}, |
|
year={2025}, |
|
eprint={2505.23277}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.CL}, |
|
url={https://arxiv.org/abs/2505.23277}, |
|
} |
|
|
|
## π¬ Contact |
|
β’ π§ [email protected] |
|
β’ π Project: https://github.com/yzhangchuck/Sentinel |
|
|
|
|
|
## π License |
|
|
|
Apache License 2.0 β Free for research and commercial use with attribution. |