楼主: 2023Hua
236 0

[其他] 多态大模型RLHF论文合集/多态大模型Agent与RLHF论文 [推广有奖]

  • 0关注
  • 22粉丝

已卖:2031份资源

大师

21%

还不是VIP/贵宾

-

威望
1
论坛币
556 个
通用积分
575.6273
学术水平
66 点
热心指数
130 点
信用等级
37 点
经验
144169 点
帖子
6770
精华
0
在线时间
2993 小时
注册时间
2022-8-18
最后登录
2026-4-17

楼主
2023Hua 在职认证  发表于 2024-4-10 10:32:56 |AI写论文

+2 论坛币
k人 参与回答

经管之家送您一份

应届毕业生专属福利!

求职就业群
赵安豆老师微信:zhaoandou666

经管之家联合CDA

送您一个全额奖学金名额~ !

感谢您参与论坛问题回答

经管之家送您两个论坛币!

+2 论坛币
Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback. pdf Reward learning from human preferences and demonstrations in Atari. pdf
Visual ChatGPT Talking, Drawing and Editing with Visual Foundation Models. pdf
WebGPT Browser-assisted question-answering with human feedback. pdf
Training language models to follow instructions with human feedback. pdf
Scaling Laws for Reward Model Overoptimization. pdf
Teaching language models to support answers with verified quotes. pdf
Quark Controllable Text Generation with Reinforced Unlearning. pdf
Scalable agent alignment via reward modeling a research direction. pdf
Learning to summarize from human feedback. pdf
Recursively Summarizing Books with Human Feedback. pdf
Red Teaming Language Models to Reduce Harms Methods, Scaling Behaviors, and Lessons Learned. pdf
Revisiting the Weaknesses of Reinforcement Learning for Neural Machine Translation. pdf
Is Reinforcement Learning (Not) for Natural Language Processing Benchmarks, Baselines, and Building Blocks for Natural Language Policy Optimization. pdf
Learning to summarize with human feedback. pdf
Non-Markovian Reward Modelling from Trajectory Labels via Interpretable Multiple Instance Learning. pdf
Pretraining Language Models with Human Preferences. pdf
Improving alignment of dialogue agents via targeted human judgements. pdf
InstructGPT Training language models to follow instructions with human feedback. pdf
Interactive Learning from Policy-Dependent Human Feedback. pdf
Few-shot Preference Learning for Human-in-the-Loop RL. pdf
Dynamic Planning in Open-Ended Dialogue using Reinforcement Learning. pdf
GPT-4 Technical Report. pdf
Fine-Tuning Language Models from Human Preferences. pdf
Deep TAMER Interactive Agent Shaping in High-Dimensional State Spaces. pdf
Discovering Language Model Behaviors with Model-Written Evaluations. pdf
Deep Reinforcement Learning from Human Preferences. pdf
Aligning Language Models with Preferences through f-divergence Minimization. pdf
Constitutional Al Harmlessness from Al Feedback. pdf
Better Aligning Text-to-Image Models with Human Preference. pdf


大模型RLHF论文合集.rar (79.13 MB, 需要: RMB 29 元)


二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

关键词:agent 论文合集 Age Optimization Minimization

您需要登录后才可以回帖 登录 | 我要注册

本版微信群
扫码
拉您进交流群
GMT+8, 2026-4-18 09:13