AI & ML interests

None defined yet.

Recent Activity

dacorvoĀ  updated a model about 5 hours ago
aws-neuron/optimum-neuron-cache
dacorvoĀ  updated a model about 6 hours ago
aws-neuron/optimum-neuron-cache
dacorvoĀ  updated a model about 6 hours ago
aws-neuron/optimum-neuron-cache
View all activity

aws-neuron's activity

pagezyhfĀ 
posted an update 8 days ago
view post
Post
1608
We published https://huggingface.co/blog/deepseek-r1-aws!

If you are using AWS, give a read. It is a running document to showcase how to deploy and fine-tune DeepSeek R1 models with Hugging Face on AWS.

We're working hard to enable all the scenarios, whether you want to deploy to Inference Endpoints, Sagemaker or EC2; with GPUs or with Trainium & Inferentia.

We have full support for the distilled models, DeepSeek-R1 support is coming soon!! I'll keep you posted.

Cheers
  • 1 reply
Ā·
pagezyhfĀ 
posted an update 25 days ago
jeffboudierĀ 
posted an update about 1 month ago
view post
Post
599
NVIDIA just announced the Cosmos World Foundation Models, available on the Hub: nvidia/cosmos-6751e884dc10e013a0a0d8e6

Cosmos is a family of pre-trained models purpose-built for generating physics-aware videos and world states to advance physical AI development.
The release includes Tokenizers nvidia/cosmos-tokenizer-672b93023add81b66a8ff8e6

Learn more in this great community article by @mingyuliutw and @PranjaliJoshi https://huggingface.co/blog/mingyuliutw/nvidia-cosmos
  • 1 reply
Ā·
pagezyhfĀ 
posted an update 2 months ago
pagezyhfĀ 
posted an update 2 months ago
view post
Post
976
Itā€™s 2nd of December , hereā€™s your Cyber Monday present šŸŽ !

Weā€™re cutting our price down on Hugging Face Inference Endpoints and Spaces!

Our folks at Google Cloud are treating us with a 40% price cut on GCP Nvidia A100 GPUs for the next 3ļøāƒ£ months. We have other reductions on all instances ranging from 20 to 50%.

Sounds like the time to give Inference Endpoints a try? Get started today and find in our documentation the full pricing details.
https://ui.endpoints.huggingface.co/
https://huggingface.co/pricing
pagezyhfĀ 
posted an update 2 months ago
view post
Post
305
Hello Hugging Face Community,

if you use Google Kubernetes Engine to host you ML workloads, I think this series of videos is a great way to kickstart your journey of deploying LLMs, in less than 10 minutes! Thank you @wietse-venema-demo !

To watch in this order:
1. Learn what are Hugging Face Deep Learning Containers
https://youtu.be/aWMp_hUUa0c?si=t-LPRkRNfD3DDNfr

2. Learn how to deploy a LLM with our Deep Learning Container using Text Generation Inference
https://youtu.be/Q3oyTOU1TMc?si=V6Dv-U1jt1SR97fj

3. Learn how to scale your inference endpoint based on traffic
https://youtu.be/QjLZ5eteDds?si=nDIAirh1r6h2dQMD

If you want more of these small tutorials and have any theme in mind, let me know!
jeffboudierĀ 
posted an update 3 months ago
pagezyhfĀ 
posted an update 3 months ago
view post
Post
1364
Hello Hugging Face Community,

I'd like to share here a bit more about our Deep Learning Containers (DLCs) we built with Google Cloud, to transform the way you build AI with open models on this platform!

With pre-configured, optimized environments for PyTorch Training (GPU) and Inference (CPU/GPU), Text Generation Inference (GPU), and Text Embeddings Inference (CPU/GPU), the Hugging Face DLCs offer:

āš” Optimized performance on Google Cloud's infrastructure, with TGI, TEI, and PyTorch acceleration.
šŸ› ļø Hassle-free environment setup, no more dependency issues.
šŸ”„ Seamless updates to the latest stable versions.
šŸ’¼ Streamlined workflow, reducing dev and maintenance overheads.
šŸ”’ Robust security features of Google Cloud.
ā˜ļø Fine-tuned for optimal performance, integrated with GKE and Vertex AI.
šŸ“¦ Community examples for easy experimentation and implementation.
šŸ”œ TPU support for PyTorch Training/Inference and Text Generation Inference is coming soon!

Find the documentation at https://huggingface.co/docs/google-cloud/en/index
If you need support, open a conversation on the forum: https://discuss.huggingface.co/c/google-cloud/69
jeffboudierĀ 
posted an update 4 months ago
jeffboudierĀ 
posted an update 5 months ago
view post
Post
459
Inference Endpoints got a bunch of cool updates yesterday, this is my top 3
jeffboudierĀ 
posted an update 5 months ago
view post
Post
4044
Pro Tip - if you're a Firefox user, you can set up Hugging Chat as integrated AI Assistant, with contextual links to summarize or simplify any text - handy!

In this short video I show how to set it up
Ā·
jeffboudierĀ 
posted an update 9 months ago
jeffboudierĀ 
posted an update 10 months ago