Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
90.5
TFLOPS
2
4
7
Pretam Ray
Pretam
Follow
buildTurf's profile picture
1 follower
·
5 following
raypretam
AI & ML interests
NLP
Recent Activity
published
a model
13 days ago
Pretam/t5-small-finetuned-xsum
upvoted
an
article
7 months ago
Making LLMs even more accessible with bitsandbytes, 4-bit quantization and QLoRA
reacted
to
vladbogo
's
post
with 👍
11 months ago
A recent paper titled "ShortGPT: Layers in Large Language Models are More Redundant Than You Expect" proposes a simple and effective approach to pruning Large Language Models (LLMs) by removing redundant layers. Key points: * Discovers significant redundancy across layers in LLMs, with some layers playing a negligible role for the final performance. * Defines a new metric called Block Influence (BI) to quantify the importance of each layer in an LLM. * Removes layers with low BI scores, achieving up to 25% reduction in parameters and computation while maintaining 92% of the LLM's performance. Congrats to the authors for their work! Paper: https://huggingface.co/papers/2403.03853
View all activity
Organizations
models
1
Pretam/t5-small-finetuned-xsum
Updated
13 days ago
datasets
5
Sort: Recently updated
Pretam/ramayana
Viewer
•
Updated
Nov 15, 2023
•
93k
•
29
•
1
Pretam/hi-te
Viewer
•
Updated
Aug 26, 2023
•
2.44M
•
16
Pretam/hi-kn-128
Viewer
•
Updated
Aug 22, 2023
•
2.14k
•
21
Pretam/hi-kn-FLAN
Viewer
•
Updated
Aug 18, 2023
•
2.17M
•
13
Pretam/hi-kn
Viewer
•
Updated
Aug 17, 2023
•
2.17M
•
30