-
A LoRA-Based Approach to Fine-Tuning LLMs for Educational Guidance in Resource-Constrained Settings
Paper • 2504.15610 • Published • 1 -
Train Small, Infer Large: Memory-Efficient LoRA Training for Large Language Models
Paper • 2502.13533 • Published • 12 -
LoRA-SP: Streamlined Partial Parameter Adaptation for Resource-Efficient Fine-Tuning of Large Language Models
Paper • 2403.08822 • Published -
LoRA-Pro: Are Low-Rank Adapters Properly Optimized?
Paper • 2407.18242 • Published
Collections
Discover the best community collections!
Collections including paper arxiv:2412.03187
-
Weighted-Reward Preference Optimization for Implicit Model Fusion
Paper • 2412.03187 • Published • 12 -
FuseAI/FuseChat-Llama-3.1-8B-Instruct
Text Generation • 8B • Updated • 33 • 12 -
FuseAI/FuseChat-Llama-3.2-3B-Instruct
3B • Updated • 405 • 7 -
FuseAI/FuseChat-Llama-3.2-1B-Instruct
1B • Updated • 28 • 6
-
Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering
Paper • 2411.11504 • Published • 23 -
Top-nσ: Not All Logits Are You Need
Paper • 2411.07641 • Published • 23 -
Adaptive Decoding via Latent Preference Optimization
Paper • 2411.09661 • Published • 10 -
When Precision Meets Position: BFloat16 Breaks Down RoPE in Long-Context Training
Paper • 2411.13476 • Published • 16
-
Weighted-Reward Preference Optimization for Implicit Model Fusion
Paper • 2412.03187 • Published • 12 -
FuseAI/FuseChat-Llama-3.1-8B-Instruct
Text Generation • 8B • Updated • 33 • 12 -
FuseAI/FuseChat-Llama-3.2-3B-Instruct
3B • Updated • 405 • 7 -
FuseAI/FuseChat-Llama-3.2-1B-Instruct
1B • Updated • 28 • 6
-
Self-Rewarding Language Models
Paper • 2401.10020 • Published • 151 -
Orion-14B: Open-source Multilingual Large Language Models
Paper • 2401.12246 • Published • 14 -
MambaByte: Token-free Selective State Space Model
Paper • 2401.13660 • Published • 60 -
MM-LLMs: Recent Advances in MultiModal Large Language Models
Paper • 2401.13601 • Published • 48
-
A LoRA-Based Approach to Fine-Tuning LLMs for Educational Guidance in Resource-Constrained Settings
Paper • 2504.15610 • Published • 1 -
Train Small, Infer Large: Memory-Efficient LoRA Training for Large Language Models
Paper • 2502.13533 • Published • 12 -
LoRA-SP: Streamlined Partial Parameter Adaptation for Resource-Efficient Fine-Tuning of Large Language Models
Paper • 2403.08822 • Published -
LoRA-Pro: Are Low-Rank Adapters Properly Optimized?
Paper • 2407.18242 • Published
-
Weighted-Reward Preference Optimization for Implicit Model Fusion
Paper • 2412.03187 • Published • 12 -
FuseAI/FuseChat-Llama-3.1-8B-Instruct
Text Generation • 8B • Updated • 33 • 12 -
FuseAI/FuseChat-Llama-3.2-3B-Instruct
3B • Updated • 405 • 7 -
FuseAI/FuseChat-Llama-3.2-1B-Instruct
1B • Updated • 28 • 6
-
Weighted-Reward Preference Optimization for Implicit Model Fusion
Paper • 2412.03187 • Published • 12 -
FuseAI/FuseChat-Llama-3.1-8B-Instruct
Text Generation • 8B • Updated • 33 • 12 -
FuseAI/FuseChat-Llama-3.2-3B-Instruct
3B • Updated • 405 • 7 -
FuseAI/FuseChat-Llama-3.2-1B-Instruct
1B • Updated • 28 • 6
-
Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering
Paper • 2411.11504 • Published • 23 -
Top-nσ: Not All Logits Are You Need
Paper • 2411.07641 • Published • 23 -
Adaptive Decoding via Latent Preference Optimization
Paper • 2411.09661 • Published • 10 -
When Precision Meets Position: BFloat16 Breaks Down RoPE in Long-Context Training
Paper • 2411.13476 • Published • 16
-
Self-Rewarding Language Models
Paper • 2401.10020 • Published • 151 -
Orion-14B: Open-source Multilingual Large Language Models
Paper • 2401.12246 • Published • 14 -
MambaByte: Token-free Selective State Space Model
Paper • 2401.13660 • Published • 60 -
MM-LLMs: Recent Advances in MultiModal Large Language Models
Paper • 2401.13601 • Published • 48