meta-llama_Meta-Llama-3.1-8B-Instruct$: loader: Transformers cpu_memory: 0 auto_devices: false disk: false cpu: false bf16: false load_in_8bit: false trust_remote_code: false no_use_fast: false use_flash_attention_2: false use_eager_attention: false load_in_4bit: false compute_dtype: float16 quant_type: nf4 use_double_quant: false disable_exllama: false disable_exllamav2: false compress_pos_emb: 1 alpha_value: 1