Large language models face challenges with tasks like infilling and constrained generation due to intractable posterior distributions. Amortized Bayesian inference with GFlowNets offers a solution by fine-tuning models to sample from these distributions efficiently. This approach improves diversity, data efficiency, and generalization compared to traditional training methods. Empirical results demonstrate the effectiveness of this method across various tasks, including sequence continuation, reasoning, arithmetic problem-solving, and story infilling.
Sang ngôn ngữ khác
từ nội dung nguồn
arxiv.org
Thông tin chi tiết chính được chắt lọc từ
by Edward J. Hu... lúc arxiv.org 03-15-2024
https://arxiv.org/pdf/2310.04363.pdfYêu cầu sâu hơn