Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,12 @@ language:
|
|
10 |
This repository contains experiment results for the [Saving 77% of the Parameters in Large Language Models Technical Report (PDF)](https://www.researchgate.net/publication/388835829_SAVING_77_OF_THE_PARAMETERS_IN_LARGE_LANGUAGE_MODELS_TECHNICAL_REPORT).
|
11 |
|
12 |
## Abstract
|
13 |
-
This technical report demonstrates that large language models (LLMs) can maintain their learning capacity while reducing their non-embedding parameters by up to 77%.
|
|
|
|
|
|
|
|
|
|
|
14 |
|
15 |
## Key Findings
|
16 |
- Achieved 77% parameter reduction while maintaining model performance.
|
|
|
10 |
This repository contains experiment results for the [Saving 77% of the Parameters in Large Language Models Technical Report (PDF)](https://www.researchgate.net/publication/388835829_SAVING_77_OF_THE_PARAMETERS_IN_LARGE_LANGUAGE_MODELS_TECHNICAL_REPORT).
|
11 |
|
12 |
## Abstract
|
13 |
+
This technical report demonstrates that large language models (LLMs) can maintain their learning capacity while reducing their non-embedding parameters by up to 77%.
|
14 |
+
We achieve this by adapting a parameter reduction technique originally developed for computer vision, replacing dense layers with an optimized subnetwork that
|
15 |
+
contains grouped pointwise convolutions. Using a 2-layer phi-3-mini-4k-instruct codebase from Microsoft as our baseline, we show that our optimized model (kphi-3)
|
16 |
+
achieves comparable validation loss while using only 15-23% of the original non-embedding parameters. Each experiment was conducted on a single NVIDIA L4 GPU within
|
17 |
+
a 3-day timeframe, supporting the democratization of AI research. Our findings suggest that current LLM architectures may be substantially overparameterized, opening
|
18 |
+
possibilities for more efficient model training and deployment.
|
19 |
|
20 |
## Key Findings
|
21 |
- Achieved 77% parameter reduction while maintaining model performance.
|