발행물
컨퍼런스
SUMMER SEMINAR
2001.08
,
ZeRO-Infinity: Breaking the GPU Memory Wall for Extreme Scale Deep Learning
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning
PARAMETER-EFFICIENT FINE-TUNING DESIGN SPACES
Distill or Annotate? Cost-Efficient Fine-Tuning of Compact Models