tgi-docker-llama / Dockerfile
ariG23498's picture
ariG23498 HF staff
Update Dockerfile
71c4b9f verified
raw
history blame contribute delete
541 Bytes
# Hugging Face TGI image
FROM ghcr.io/huggingface/text-generation-inference:3.0.2
# Set working directory
WORKDIR /app
# Create the /data directory inside the container
RUN mkdir -p /data
RUN chmod 777 /data
RUN mkdir -p /.cache
RUN chmod 777 /.cache
RUN mkdir -p /.triton
RUN chmod 777 /.triton
# Expose the model on port 8080
EXPOSE 8080
# Set the Hugging Face token as an environment variable
ARG HF_TOKEN
ENV HF_TOKEN=${HF_TOKEN}
# Run the TGI server directly
CMD ["--model-id", "meta-llama/Llama-3.2-1B-Instruct", "--port", "8080"]