""" | |
Interactive Refinement Qwen3-14B 使用例 | |
""" | |
from transformers import AutoTokenizer, AutoModelForCausalLM | |
import json | |
import torch | |
class InteractiveRefinementInference: | |
def __init__(self, model_name_or_path): | |
self.tokenizer = AutoTokenizer.from_pretrained(model_name_or_path) | |
self.model = AutoModelForCausalLM.from_pretrained( | |
model_name_or_path, | |
torch_dtype=torch.float16, | |
device_map="auto" | |
) | |
# 設定ファイル読み込み | |
with open(f"{model_name_or_path}/refinement_config.json", "r") as f: | |
self.config = json.load(f) | |
def generate(self, question, rounds=None): | |
rounds = rounds or self.config["refinement_rounds"] | |
# Interactive Refinement実装(詳細は元のコードを参照) | |
pass | |
# 使用例 | |
if __name__ == "__main__": | |
model = InteractiveRefinementInference("your-username/interactive-refinement-qwen3-14b") | |
response = model.generate("あなたの質問") | |
print(response) | |