Master-Zangetsu
's Collections
ToRead
updated
Rethinking Mixture-of-Agents: Is Mixing Different Large Language Models
Beneficial?
Paper
•
2502.00674
•
Published
•
13
Demystifying Long Chain-of-Thought Reasoning in LLMs
Paper
•
2502.03373
•
Published
•
58
SmolLM2: When Smol Goes Big -- Data-Centric Training of a Small Language Model
Paper
•
2502.02737
•
Published
•
253
DeepRAG: Thinking to Retrieval Step by Step for Large Language Models
Paper
•
2502.01142
•
Published
•
24
Scaling Embedding Layers in Language Models
Paper
•
2502.01637
•
Published
•
23
ZebraLogic: On the Scaling Limits of LLMs for Logical Reasoning
Paper
•
2502.01100
•
Published
•
19
ScoreFlow: Mastering LLM Agent Workflows via Score-based Preference
Optimization
Paper
•
2502.04306
•
Published
•
20
Analyze Feature Flow to Enhance Interpretation and Steering in Language
Models
Paper
•
2502.03032
•
Published
•
60
QuEST: Stable Training of LLMs with 1-Bit Weights and Activations
Paper
•
2502.05003
•
Published
•
42
DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM
Guardrails
Paper
•
2502.05163
•
Published
•
22
CMoE: Fast Carving of Mixture-of-Experts for Efficient LLM Inference
Paper
•
2502.04416
•
Published
•
12
ARR: Question Answering with Large Language Models via Analyzing,
Retrieving, and Reasoning
Paper
•
2502.04689
•
Published
•
8