論文標(biāo)題:Direct Preference Optimization: Your Language Model is Secretly a R...
論文標(biāo)題:Propagation Tree Is Not Deep: Adaptive Graph Contrastive Learning A...
一给涕、概述 大語(yǔ)言模型(LLMs)在預(yù)訓(xùn)練的過(guò)程中通常會(huì)捕捉數(shù)據(jù)的特征垛叨,而這些訓(xùn)練數(shù)據(jù)通常既包含高質(zhì)量的也包含低質(zhì)量的,因此模型有時(shí)會(huì)產(chǎn)生不被期望...
論文標(biāo)題:LoRA: Low-Rank Adaptation of Large Language Models論文鏈接:https://arxi...
論文標(biāo)題:Megatron-LM: Training Multi-Billion Parameter Language Models Using...
論文標(biāo)題:Tree of Thoughts: Deliberate Problem Solving with Large Language Mo...
論文標(biāo)題:LIMA: Less Is More for Alignment論文鏈接:https://arxiv.org/abs/2305.112...
論文標(biāo)題:Self-Consistency Improves Chain of Thought Reasoning in Language Mo...
論文標(biāo)題:GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism論文鏈接:https...