Jump2Paper Archive
Records
019
Languages
01
Updated
2026.04.07
문서 찾기
제목, 키워드, 약어, 언어별 탐색
All
Korean (KR)
문서 001
Agent Lightning: Train ANY AI Agents with Reinforcement Learning
multi-turn GRPO
veRL
multi-agent LLM
agent frameworks
DeepSeek-R1
Korean
읽기 →
문서 002
Automatic Curriculum LearningFor Deep RL: A Short Survey
Multi-Goal RL
Intrinsic Motivation
PCG for RL
Sim2Real Transfer
Self-Play
Korean
읽기 →
문서 003
Born-Again Neural Networks
Label Smoothing
Knowledge Distillation
Dark Knowledge
Knowledge Distillation Survey
Self-Distillation
Korean
읽기 →
문서 004
Curriculum Learning for LLM Pretraining:An Analysis of Learning Dynamics
curriculum learning
LLM pretraining
Pythia scaling
HMM training trajectory
pretraining data mixture
Korean
읽기 →
문서 005
DUMP: Automated Distribution-Level Curriculum Learning for RL-based LLM Post-training
multi-armed bandit
분포 샘플링
data mixture training
UCB bandit
curriculum learning
Korean
읽기 →
문서 006
Dynamic Loss-Based Sample Reweighting for Improved Large Language Model Pretraining
데이터 효율적 학습
Sample Efficiency LLM
Data Selection for LLMs
SlimPajama
DRO Machine Learning
Korean
읽기 →
문서 007
Exclusive Self Attention
Attention Mechanism
Long Context
SA-FFN 역할 분담
Attention Sink
Transformer 개선
Korean
읽기 →
문서 008
KV Cache Transform Codingfor Compact Storage in LLM Inference
PCA decorrelation
adaptive quantization
learned transform coding
speculative decoding
transform coding
Korean
읽기 →
문서 009
KVzap: Fast, Adaptive, and Faithful KV Cache Pruning
LLM inference efficiency
long-context LLM
KV quantization
surrogate model
reasoning efficiency
Korean
읽기 →
문서 010
Mamba: Linear-Time Sequence Modeling with Selective State Spaces
Mamba-2
S4
linear-time sequence modeling
state space model
HyenaDNA
Korean
읽기 →
문서 011
PivotRL: High Accuracy Agentic Post-Training at Low Compute Cost
curriculum learning RL
agentic post-training
BC-to-RL bridge
natural gradient RL
SWE-Bench
Korean
읽기 →
문서 012
RegMix: Data Mixture as Regression for Language Model Pre-Training
pre-training data selection
data mixture
gradient boosting
Dirichlet sampling
proxy model methods
Korean
읽기 →
문서 013
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
Hierarchical Vision Transformer
Shifted Window Attention
Semantic Segmentation
Hierarchical Feature Map
Masked Image Modeling
Korean
읽기 →
문서 014
The Forward-Forward Algorithm:Some Preliminary Investigations
예측 코딩
Noise Contrastive Estimation
대조 학습
역전파 대안
Forward-Forward Algorithm
Korean
읽기 →
문서 015
Think Anywhere in Code Generation
inline reasoning
interleaved thinking
chain-of-thought
code generation LLM
LoRA
Korean
읽기 →
문서 016
TiKMiX: Take Data Influence into Dynamic Mixture for Language Model Pre-training
domain adaptation
curriculum learning
data mixture
dynamic data mixture
proxy-free data selection
Korean
읽기 →
문서 017
Transformers are RNNs:Fast Autoregressive Transformers with Linear Attention
linear attention
Transformer–RNN 등가성
kernel self-attention
RetNet
선형 언어 모델
Korean
읽기 →
문서 018
TurboQuant: Online Vector Quantization with Near-optimal Distortion Rate
랜덤 회전 양자화
온라인 양자화
KV cache quantization
vector quantization
online vector quantization
Korean
읽기 →
문서 019
Very Large-Scale Multi-Agent Simulation in AgentScope
AgentScope
Generative Agents
Distributed AI
Game Theory + LLM
Large-Scale Multi-Agent Simulation
Korean
읽기 →
No papers found matching your criteria.