seung hwan jung
digit82
		AI & ML interests
None yet
		
		Organizations
None yet
Resaoning Dataset
			
			
	
	Pretraining Dataset
			
			
	
	llm
			
			
	
	- 
	
	
	
Training Language Models to Self-Correct via Reinforcement Learning
Paper • 2409.12917 • Published • 140 - 
	
	
	
MM1.5: Methods, Analysis & Insights from Multimodal LLM Fine-tuning
Paper • 2409.20566 • Published • 56 - 
	
	
	
Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models
Paper • 2411.04996 • Published • 51 - 
	
	
	
EoRA: Training-free Compensation for Compressed LLM with Eigenspace Low-Rank Approximation
Paper • 2410.21271 • Published • 7 
Korean Pretraining Dataset
			
			
	
	Resaoning Dataset
			
			
	
	SFT Dataset
			
			
	
	Pretraining Dataset
			
			
	
	Paper
			
			
	
	llm
			
			
	
	- 
	
	
	
Training Language Models to Self-Correct via Reinforcement Learning
Paper • 2409.12917 • Published • 140 - 
	
	
	
MM1.5: Methods, Analysis & Insights from Multimodal LLM Fine-tuning
Paper • 2409.20566 • Published • 56 - 
	
	
	
Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models
Paper • 2411.04996 • Published • 51 - 
	
	
	
EoRA: Training-free Compensation for Compressed LLM with Eigenspace Low-Rank Approximation
Paper • 2410.21271 • Published • 7