論文標(biāo)題:Direct Preference Optimization: Your Language Model is Secretly a Reward Model論文鏈接...
![240](https://upload.jianshu.io/users/upload_avatars/22097296/9ef3a58a-1158-4ca2-9cf0-24d2b0f2e2c9.jpg?imageMogr2/auto-orient/strip|imageView2/1/w/240/h/240)
論文標(biāo)題:Direct Preference Optimization: Your Language Model is Secretly a Reward Model論文鏈接...
一号坡、概述 大語言模型(LLMs)在預(yù)訓(xùn)練的過程中通常會捕捉數(shù)據(jù)的特征,而這些訓(xùn)練數(shù)據(jù)通常既包含高質(zhì)量的也包含低質(zhì)量的描扯,因此模型有時(shí)會產(chǎn)生不被期望的行為,如編造事實(shí)趟薄,生成有偏見...
論文標(biāo)題:LoRA: Low-Rank Adaptation of Large Language Models論文鏈接:https://arxiv.org/abs/2106....
論文標(biāo)題:Chain-of-Thought Prompting Elicits Reasoning in Large Language Models論文鏈接:https://...
論文標(biāo)題:Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallel...
論文標(biāo)題:Tree of Thoughts: Deliberate Problem Solving with Large Language Models論文鏈接:https:...
論文標(biāo)題:LIMA: Less Is More for Alignment論文鏈接:https://arxiv.org/abs/2305.11206[https://arxi...
論文標(biāo)題:Self-Consistency Improves Chain of Thought Reasoning in Language Models論文鏈接:https:...
論文標(biāo)題:GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism論文鏈接:https://arxiv.org/ab...