論文標題:Direct Preference Optimization: Your Language Model is Secretly a R...
![240](https://upload.jianshu.io/collections/images/1904333/%E7%BC%96%E7%A8%8B2.jpg?imageMogr2/auto-orient/strip|imageView2/1/w/240/h/240)
論文標題:Direct Preference Optimization: Your Language Model is Secretly a R...
論文標題:Propagation Tree Is Not Deep: Adaptive Graph Contrastive Learning A...
一、概述 大語言模型(LLMs)在預訓練的過程中通常會捕捉數(shù)據(jù)的特征额港,而這些訓練數(shù)據(jù)通常既包含高質(zhì)量的也包含低質(zhì)量的蒿涎,因此模型有時會產(chǎn)生不被期望...
論文標題:LoRA: Low-Rank Adaptation of Large Language Models論文鏈接:https://arxi...
論文標題:Megatron-LM: Training Multi-Billion Parameter Language Models Using...
論文標題:Tree of Thoughts: Deliberate Problem Solving with Large Language Mo...
論文標題:LIMA: Less Is More for Alignment論文鏈接:https://arxiv.org/abs/2305.112...
論文標題:Self-Consistency Improves Chain of Thought Reasoning in Language Mo...
論文標題:GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism論文鏈接:https...
論文標題:Training language models to follow instructions with human feedback...