- 
	
	
	
DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models
Paper • 2309.03883 • Published • 35 - 
	
	
	
LoRA: Low-Rank Adaptation of Large Language Models
Paper • 2106.09685 • Published • 52 - 
	
	
	
Agents: An Open-source Framework for Autonomous Language Agents
Paper • 2309.07870 • Published • 42 - 
	
	
	
RLAIF: Scaling Reinforcement Learning from Human Feedback with AI Feedback
Paper • 2309.00267 • Published • 51 
lckr
lckr
		AI & ML interests
None yet
		
		Organizations
Ilya's papers for Carmack
			Ilya Sutskever: "If you really learn all of these, you’ll know 90% of what matters today." Full list: https://punkx.org/jackdoe/30.html
			
	
	- 
	
	
	
Recurrent Neural Network Regularization
Paper • 1409.2329 • Published • 1 - 
	
	
	
Pointer Networks
Paper • 1506.03134 • Published • 1 - 
	
	
	
Order Matters: Sequence to sequence for sets
Paper • 1511.06391 • Published • 1 - 
	
	
	
GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Paper • 1811.06965 • Published • 1 
random_papers
			
			
	
	- 
	
	
	
DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models
Paper • 2309.03883 • Published • 35 - 
	
	
	
LoRA: Low-Rank Adaptation of Large Language Models
Paper • 2106.09685 • Published • 52 - 
	
	
	
Agents: An Open-source Framework for Autonomous Language Agents
Paper • 2309.07870 • Published • 42 - 
	
	
	
RLAIF: Scaling Reinforcement Learning from Human Feedback with AI Feedback
Paper • 2309.00267 • Published • 51 
Ilya's papers for Carmack
			Ilya Sutskever: "If you really learn all of these, you’ll know 90% of what matters today." Full list: https://punkx.org/jackdoe/30.html
			
	
	- 
	
	
	
Recurrent Neural Network Regularization
Paper • 1409.2329 • Published • 1 - 
	
	
	
Pointer Networks
Paper • 1506.03134 • Published • 1 - 
	
	
	
Order Matters: Sequence to sequence for sets
Paper • 1511.06391 • Published • 1 - 
	
	
	
GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Paper • 1811.06965 • Published • 1 
			models
			0
		
			
	None public yet