hugoleeShanda 's Collections paper
updated
LLMLingua-2: Data Distillation for Efficient and Faithful Task-Agnostic
Prompt Compression
Paper
• 2403.12968
• Published • 26
PERL: Parameter Efficient Reinforcement Learning from Human Feedback
Paper
• 2403.10704
• Published • 60
Alignment Studio: Aligning Large Language Models to Particular
Contextual Regulations
Paper
• 2403.09704
• Published • 32
RAFT: Adapting Language Model to Domain Specific RAG
Paper
• 2403.10131
• Published • 72
Agent-FLAN: Designing Data and Methods of Effective Agent Tuning for
Large Language Models
Paper
• 2403.12881
• Published • 18
TnT-LLM: Text Mining at Scale with Large Language Models
Paper
• 2403.12173
• Published • 20
RewardBench: Evaluating Reward Models for Language Modeling
Paper
• 2403.13787
• Published • 22
LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement
Paper
• 2403.15042
• Published • 27
Can large language models explore in-context?
Paper
• 2403.15371
• Published • 33
Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient
LLMs Under Compression
Paper
• 2403.15447
• Published • 16
InternLM2 Technical Report
Paper
• 2403.17297
• Published • 34
Long-form factuality in large language models
Paper
• 2403.18802
• Published • 26
Localizing Paragraph Memorization in Language Models
Paper
• 2403.19851
• Published • 15
sDPO: Don't Use Your Data All at Once
Paper
• 2403.19270
• Published • 41
LLM-ABR: Designing Adaptive Bitrate Algorithms via Large Language Models
Paper
• 2404.01617
• Published • 8
Advancing LLM Reasoning Generalists with Preference Trees
Paper
• 2404.02078
• Published • 46
Long-context LLMs Struggle with Long In-context Learning
Paper
• 2404.02060
• Published • 37
Compact Language Models via Pruning and Knowledge Distillation
Paper
• 2407.14679
• Published • 40
DDK: Distilling Domain Knowledge for Efficient Large Language Models
Paper
• 2407.16154
• Published • 22