論文標題:Direct Preference Optimization: Your Language Model is Secretly a Reward Model論文鏈接...
論文標題:Direct Preference Optimization: Your Language Model is Secretly a Reward Model論文鏈接...
論文標題:Propagation Tree Is Not Deep: Adaptive Graph Contrastive Learning Approach for Rum...
一、概述 大語言模型(LLMs)在預訓練的過程中通常會捕捉數(shù)據(jù)的特征,而這些訓練數(shù)據(jù)通常既包含高質(zhì)量的也包含低質(zhì)量的忆某,因此模型有時會產(chǎn)生不被期望的行為,如編造事實浦箱,生成有偏見...
論文標題:LoRA: Low-Rank Adaptation of Large Language Models論文鏈接:https://arxiv.org/abs/2106....
論文標題:Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallel...
論文標題:Tree of Thoughts: Deliberate Problem Solving with Large Language Models論文鏈接:https:...
論文標題:LIMA: Less Is More for Alignment論文鏈接:https://arxiv.org/abs/2305.11206[https://arxi...
論文標題:Self-Consistency Improves Chain of Thought Reasoning in Language Models論文鏈接:https:...
論文標題:GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism論文鏈接:https://arxiv.org/ab...
論文標題:Training language models to follow instructions with human feedback論文鏈接:https://ar...
論文標題:LLaMA: Open and Efficient Foundation Language Models論文鏈接:https://arxiv.org/abs/230...
論文標題:Pure Transformers are Powerful Graph Learners論文鏈接:https://arxiv.org/abs/2207.02505...
論文標題:Chain-of-Thought Prompting Elicits Reasoning in Large Language Models論文鏈接:https://...
論文標題:Language Models are Few-Shot Learners論文鏈接:https://arxiv.org/abs/2005.14165[https:/...
論文標題:Language Models are Unsupervised Multitask Learners論文鏈接:https://d4mucfpksywv.cloud...
論文標題:Improving Language Understanding by Generative Pre-Training論文鏈接:https://www.cs.ubc...