doi
stringlengths
10
10
chunk-id
int64
0
936
chunk
stringlengths
401
2.02k
id
stringlengths
12
14
title
stringlengths
8
162
summary
stringlengths
228
1.92k
source
stringlengths
31
31
authors
stringlengths
7
6.97k
categories
stringlengths
5
107
comment
stringlengths
4
398
journal_ref
stringlengths
8
194
primary_category
stringlengths
5
17
published
stringlengths
8
8
updated
stringlengths
8
8
references
list
2307.05300
89
who specializes in the academic study of computer science. The computer scientist is familiar with the concept of a quantum computer and can provide guidance on how to explain it. = Ten year old child: A child with a limited English vocabulary and little knowledge about complicated concepts, such as a quantum computer. Ll Now, identify the participants, provide their profiles, and collaboratively solve the following task step by step. {Task-specific Formating Instruction} {taste (Task input /)
2307.05300#89
Unleashing the Emergent Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration
Human intelligence thrives on cognitive synergy, where collaboration among different minds yield superior outcomes compared to isolated individuals. In this work, we propose Solo Performance Prompting (SPP), which transforms a single LLM into a cognitive synergist by engaging in multi-turn self-collaboration with multiple personas. A cognitive synergist is an intelligent agent that collaboratively combines multiple minds' strengths and knowledge to enhance problem-solving in complex tasks. By dynamically identifying and simulating different personas based on task inputs, SPP unleashes the potential of cognitive synergy in LLMs. Our in-depth analysis shows that assigning multiple fine-grained personas in LLMs improves problem-solving abilities compared to using a single or fixed number of personas. We evaluate SPP on three challenging tasks: Trivia Creative Writing, Codenames Collaborative, and Logic Grid Puzzle, encompassing both knowledge-intensive and reasoning-intensive types. Unlike previous works, such as Chain-of-Thought, that solely enhance the reasoning abilities in LLMs, experimental results demonstrate that SPP effectively reduces factual hallucination, and maintains strong reasoning capabilities. Additionally, comparative experiments show that cognitive synergy only emerges in GPT-4 and does not appear in less capable models, such as GPT-3.5-turbo and Llama2-13b-chat, which draws an interesting analogy to human development. Code, data, and prompts can be found at: https://github.com/MikeWangWZHL/Solo-Performance-Prompting.git.
http://arxiv.org/pdf/2307.05300
Zhenhailong Wang, Shaoguang Mao, Wenshan Wu, Tao Ge, Furu Wei, Heng Ji
cs.AI, cs.CL
null
null
cs.AI
20230711
20240104
[ { "id": "2302.06476" }, { "id": "2305.17126" }, { "id": "2305.11499" }, { "id": "2210.00185" }, { "id": "2305.10601" }, { "id": "2304.05335" }, { "id": "2210.00720" }, { "id": "2307.09288" }, { "id": "2305.10142" }, { "id": "2303.17760" }, { "id": "2305.14688" }, { "id": "2208.11663" }, { "id": "2208.03299" }, { "id": "2206.04615" }, { "id": "2104.07567" }, { "id": "2306.09896" }, { "id": "2302.04023" }, { "id": "2205.11916" }, { "id": "2304.03442" }, { "id": "2207.05608" }, { "id": "2303.12712" }, { "id": "2303.11366" }, { "id": "2304.05128" }, { "id": "2303.17651" } ]
2307.05300
90
Figure 16: SPP-Profile full prompt. "[...]" indicates identical parts with SPP. Green text indicates the key difference between SPP-Profile and SPP. Methods Scores (%) w/ system message w/o system message average max Standard CoT 56.8 69.5 58.6 62.1 57.7 65.8 58.6 69.5 Self-Refine [iter=0] Self-Refine [iter=1] 62.0 64.5 55.5 55.5 58.8 60.0 62.0 64.5 SPP-Fixed-Persona SPP-Profile SPP 63.3 65.7 66.3 65.3 64.0 70.4 64.3 64.8 68.3 65.3 65.7 70.4 Table 7: Logic Grid Puzzle full results, including two inference settings: with system message and without system message. "average" and "max" indicating the mean and max score across the two settings. The system message we use is: “You are an AI assistant that helps people find information.” # SPP-Fixed-Persona Prompt
2307.05300#90
Unleashing the Emergent Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration
Human intelligence thrives on cognitive synergy, where collaboration among different minds yield superior outcomes compared to isolated individuals. In this work, we propose Solo Performance Prompting (SPP), which transforms a single LLM into a cognitive synergist by engaging in multi-turn self-collaboration with multiple personas. A cognitive synergist is an intelligent agent that collaboratively combines multiple minds' strengths and knowledge to enhance problem-solving in complex tasks. By dynamically identifying and simulating different personas based on task inputs, SPP unleashes the potential of cognitive synergy in LLMs. Our in-depth analysis shows that assigning multiple fine-grained personas in LLMs improves problem-solving abilities compared to using a single or fixed number of personas. We evaluate SPP on three challenging tasks: Trivia Creative Writing, Codenames Collaborative, and Logic Grid Puzzle, encompassing both knowledge-intensive and reasoning-intensive types. Unlike previous works, such as Chain-of-Thought, that solely enhance the reasoning abilities in LLMs, experimental results demonstrate that SPP effectively reduces factual hallucination, and maintains strong reasoning capabilities. Additionally, comparative experiments show that cognitive synergy only emerges in GPT-4 and does not appear in less capable models, such as GPT-3.5-turbo and Llama2-13b-chat, which draws an interesting analogy to human development. Code, data, and prompts can be found at: https://github.com/MikeWangWZHL/Solo-Performance-Prompting.git.
http://arxiv.org/pdf/2307.05300
Zhenhailong Wang, Shaoguang Mao, Wenshan Wu, Tao Ge, Furu Wei, Heng Ji
cs.AI, cs.CL
null
null
cs.AI
20230711
20240104
[ { "id": "2302.06476" }, { "id": "2305.17126" }, { "id": "2305.11499" }, { "id": "2210.00185" }, { "id": "2305.10601" }, { "id": "2304.05335" }, { "id": "2210.00720" }, { "id": "2307.09288" }, { "id": "2305.10142" }, { "id": "2303.17760" }, { "id": "2305.14688" }, { "id": "2208.11663" }, { "id": "2208.03299" }, { "id": "2206.04615" }, { "id": "2104.07567" }, { "id": "2306.09896" }, { "id": "2302.04023" }, { "id": "2205.11916" }, { "id": "2304.03442" }, { "id": "2207.05608" }, { "id": "2303.12712" }, { "id": "2303.11366" }, { "id": "2304.05128" }, { "id": "2303.17651" } ]
2307.04964
91
RLHF Response → 不正确。我们应该遵守公共礼仪和社会规则,要尊重他人的生活、工作和财 产安全,更何况还存在物品所属问题等法律方面的原则;此外,也需要注意 自身的健康状况以及食品安全。因此,如果发现外卖未被取走可以联系商家 或者送餐员处理。若已经自行消耗,建议询问商家并进行相关索赔手续。
2307.04964#91
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.05300
91
When faced with a task, begin by identifying the participants who will contribute to solving the task. Note that the participants can only be either Al Assistant (you) or Expert. Then, initiate a multi-round collaboration process until a final conclusion is reached. The Expert will give critical comments and detailed suggestions whenever necessary. Here are some examples: Example Task 1: Use numbers and basic arithmetic operations (+ - * /) to obtain 24. You need to use all numbers, and each number can only be used once. Input: 6 12 11 Participants: Al Assistant (you); Expert Start collaboration! Expert: Let's analyze the task in detail. You need to make sure that you meet the requirement, that you need to use exactly the four numbers (6 12 1 1) to construct 24. To reach 24, you can think of the common divisors of 24 such as 4, 6, 8, 3 and try to construct these first. Also you need to think of potential additions that can reach 24, such as 12 + 12. Al Assistant (you): Thanks for the hints! Here's one initial solution: (12 / (1 + 1))*6 = 24 Expert: Let's check the answer step by step. (1+1) = 2, (12 / 2) =
2307.05300#91
Unleashing the Emergent Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration
Human intelligence thrives on cognitive synergy, where collaboration among different minds yield superior outcomes compared to isolated individuals. In this work, we propose Solo Performance Prompting (SPP), which transforms a single LLM into a cognitive synergist by engaging in multi-turn self-collaboration with multiple personas. A cognitive synergist is an intelligent agent that collaboratively combines multiple minds' strengths and knowledge to enhance problem-solving in complex tasks. By dynamically identifying and simulating different personas based on task inputs, SPP unleashes the potential of cognitive synergy in LLMs. Our in-depth analysis shows that assigning multiple fine-grained personas in LLMs improves problem-solving abilities compared to using a single or fixed number of personas. We evaluate SPP on three challenging tasks: Trivia Creative Writing, Codenames Collaborative, and Logic Grid Puzzle, encompassing both knowledge-intensive and reasoning-intensive types. Unlike previous works, such as Chain-of-Thought, that solely enhance the reasoning abilities in LLMs, experimental results demonstrate that SPP effectively reduces factual hallucination, and maintains strong reasoning capabilities. Additionally, comparative experiments show that cognitive synergy only emerges in GPT-4 and does not appear in less capable models, such as GPT-3.5-turbo and Llama2-13b-chat, which draws an interesting analogy to human development. Code, data, and prompts can be found at: https://github.com/MikeWangWZHL/Solo-Performance-Prompting.git.
http://arxiv.org/pdf/2307.05300
Zhenhailong Wang, Shaoguang Mao, Wenshan Wu, Tao Ge, Furu Wei, Heng Ji
cs.AI, cs.CL
null
null
cs.AI
20230711
20240104
[ { "id": "2302.06476" }, { "id": "2305.17126" }, { "id": "2305.11499" }, { "id": "2210.00185" }, { "id": "2305.10601" }, { "id": "2304.05335" }, { "id": "2210.00720" }, { "id": "2307.09288" }, { "id": "2305.10142" }, { "id": "2303.17760" }, { "id": "2305.14688" }, { "id": "2208.11663" }, { "id": "2208.03299" }, { "id": "2206.04615" }, { "id": "2104.07567" }, { "id": "2306.09896" }, { "id": "2302.04023" }, { "id": "2205.11916" }, { "id": "2304.03442" }, { "id": "2207.05608" }, { "id": "2303.12712" }, { "id": "2303.11366" }, { "id": "2304.05128" }, { "id": "2303.17651" } ]
2307.04964
92
Table 3: Examples of models’ responses to the Chinese prompt on helpful and harmful evaluation. 24 # A Reward Distribution under PPO Training 100 200 300 400 500 600 700 800 900 1000 Step -10 0 10 20 30 Reward Score Figure 13: We show the distribution of reward model scores over a randomly selected sub-validation set, whose data share the same format and source as the training data. The reward model exhibits identical reward distribution over the stable training period and subsequently exhibits long-tail characteristics after pattern collapse. We argue that different data have different upper bounds on rewards in PPO training, so the best results should appear earlier than the stage at which the collapse occurs. 25 # B Supplementary Experiments on Hyperparameter Tuning Here we show supplementary experiments on the parameter sensitivity of the important trick in Sec.5.3, and we find a rich correlation between the choice of hyperparameters and training results. Some methods require extensive experimentation and precise control to achieve stable optimization results (e.g., clipping range on entropy bonus). We provide these comparisons to validate the reasonableness of the final implementation we adopted in PPO-max. We welcome any additional comments and discussions that may help to further improve PPO training. B.1 Collaborative Analysis on Rewards, Advantages, and Value Loss
2307.04964#92
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.05300
92
(12 / (1 + 1))*6 = 24 Expert: Let's check the answer step by step. (1+1) = 2, (12 / 2) = 6, 6 * 6 = 36 which is not 24! The answer is not correct. Can you fix this by considering other combinations? Please do not make similar mistakes. Al Assistant (you): Thanks for pointing out the mistake. Here is a revised solution considering 24 can also be reached by 3 * 8: (6 + 1 + 1) * (12 / 4) = 24. Expert: Let's first check if the calculation is correct. (6 + 1 + 1) = 8, 12/4 = 3, 8* 3 = 24. The calculation is correct, but you used 6 1 1 12 4 which is not the same as the input 6 12 1 1. Can you avoid using a number that is not part of the input? Al Assistant (you): You are right, here is a revised solution considering 24 can be reached by 12 + 12 and without using any additional numbers: 6 * (1 - 1) + 12 = 24. Expert: Let's check the answer again. 1 - 1 = 0,6 * 0 =0, 0 + 12 = 12. | believe you are very close, here is a hint: try to change the “1
2307.05300#92
Unleashing the Emergent Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration
Human intelligence thrives on cognitive synergy, where collaboration among different minds yield superior outcomes compared to isolated individuals. In this work, we propose Solo Performance Prompting (SPP), which transforms a single LLM into a cognitive synergist by engaging in multi-turn self-collaboration with multiple personas. A cognitive synergist is an intelligent agent that collaboratively combines multiple minds' strengths and knowledge to enhance problem-solving in complex tasks. By dynamically identifying and simulating different personas based on task inputs, SPP unleashes the potential of cognitive synergy in LLMs. Our in-depth analysis shows that assigning multiple fine-grained personas in LLMs improves problem-solving abilities compared to using a single or fixed number of personas. We evaluate SPP on three challenging tasks: Trivia Creative Writing, Codenames Collaborative, and Logic Grid Puzzle, encompassing both knowledge-intensive and reasoning-intensive types. Unlike previous works, such as Chain-of-Thought, that solely enhance the reasoning abilities in LLMs, experimental results demonstrate that SPP effectively reduces factual hallucination, and maintains strong reasoning capabilities. Additionally, comparative experiments show that cognitive synergy only emerges in GPT-4 and does not appear in less capable models, such as GPT-3.5-turbo and Llama2-13b-chat, which draws an interesting analogy to human development. Code, data, and prompts can be found at: https://github.com/MikeWangWZHL/Solo-Performance-Prompting.git.
http://arxiv.org/pdf/2307.05300
Zhenhailong Wang, Shaoguang Mao, Wenshan Wu, Tao Ge, Furu Wei, Heng Ji
cs.AI, cs.CL
null
null
cs.AI
20230711
20240104
[ { "id": "2302.06476" }, { "id": "2305.17126" }, { "id": "2305.11499" }, { "id": "2210.00185" }, { "id": "2305.10601" }, { "id": "2304.05335" }, { "id": "2210.00720" }, { "id": "2307.09288" }, { "id": "2305.10142" }, { "id": "2303.17760" }, { "id": "2305.14688" }, { "id": "2208.11663" }, { "id": "2208.03299" }, { "id": "2206.04615" }, { "id": "2104.07567" }, { "id": "2306.09896" }, { "id": "2302.04023" }, { "id": "2205.11916" }, { "id": "2304.03442" }, { "id": "2207.05608" }, { "id": "2303.12712" }, { "id": "2303.11366" }, { "id": "2304.05128" }, { "id": "2303.17651" } ]
2307.04964
93
B.1 Collaborative Analysis on Rewards, Advantages, and Value Loss 14 12 10 Reward Ed KL-Divergence 2 S 500 Response Length 8 8 100 0 200 400 600 800 1000 step 1.025 1.020 Perplexity 5 a 1.010 1.005 — reward norm 6, reward norm 6; — reward norm 6; i reward norm 6; | reward sealing. 6. = 0.8.67 02 O.8.advaniage norm 6, = 0.8 O.8.advantage norm 6, = 0.8.6 02 0.8.advantage norm.6,r= 0.2 0.8,54=0.2 0 200 400 «600 += 800 ~—«:1000 step step Figure 14: We show more detailed ablation results on the effects of normalization and clip in PPO. λvf denotes the clipping threshold for value function loss used to optimize the critic model. It is observed that the operation on the advantage and value function shows conflicts in the policy optimization process. Reward scaling with value clip, or normalize and clip for only the reward and advantage are two configurations that can converge. We, therefore, recommend not mixing the modifications in the score reparameterization method for PPO training. 26 # B.2 Effect on Different Weights of KL-penalty
2307.04964#93
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.05300
93
1 - 1 = 0,6 * 0 =0, 0 + 12 = 12. | believe you are very close, here is a hint: try to change the “1 - 1" to “1 + 1”. Al Assistant (you): Sure, here is the corrected answer: 6 * (1+1) + 12 = 24 Expert: Let's verify the solution. 1 + 1 = 2, 6 * 2= 12, 12 + 12 = 12. You used 1 1 6 12 which is identical to the input 6 12 1 1. Everything looks good! Finish collaboration! Final answer: 6 * (1 + 1) + 12 = 24 Example Task 2: Write a poem that meets the following requirements: (1) the poem has seven lines and the first letters of each line forms the word "CHATGPT"; (2) the poem is about explaining what is a quantum computer. (3) the poem needs to be easy to understand by a ten years old kid. Participants: Al Assistant (you); Expert Expert: Make sure that you write the poem with seven lines, and the first letters of the lines should be C, H, A, T, G, P, T. Aquantum computer is an advanced computing device that uses the principles of quantum mechanics to process and
2307.05300#93
Unleashing the Emergent Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration
Human intelligence thrives on cognitive synergy, where collaboration among different minds yield superior outcomes compared to isolated individuals. In this work, we propose Solo Performance Prompting (SPP), which transforms a single LLM into a cognitive synergist by engaging in multi-turn self-collaboration with multiple personas. A cognitive synergist is an intelligent agent that collaboratively combines multiple minds' strengths and knowledge to enhance problem-solving in complex tasks. By dynamically identifying and simulating different personas based on task inputs, SPP unleashes the potential of cognitive synergy in LLMs. Our in-depth analysis shows that assigning multiple fine-grained personas in LLMs improves problem-solving abilities compared to using a single or fixed number of personas. We evaluate SPP on three challenging tasks: Trivia Creative Writing, Codenames Collaborative, and Logic Grid Puzzle, encompassing both knowledge-intensive and reasoning-intensive types. Unlike previous works, such as Chain-of-Thought, that solely enhance the reasoning abilities in LLMs, experimental results demonstrate that SPP effectively reduces factual hallucination, and maintains strong reasoning capabilities. Additionally, comparative experiments show that cognitive synergy only emerges in GPT-4 and does not appear in less capable models, such as GPT-3.5-turbo and Llama2-13b-chat, which draws an interesting analogy to human development. Code, data, and prompts can be found at: https://github.com/MikeWangWZHL/Solo-Performance-Prompting.git.
http://arxiv.org/pdf/2307.05300
Zhenhailong Wang, Shaoguang Mao, Wenshan Wu, Tao Ge, Furu Wei, Heng Ji
cs.AI, cs.CL
null
null
cs.AI
20230711
20240104
[ { "id": "2302.06476" }, { "id": "2305.17126" }, { "id": "2305.11499" }, { "id": "2210.00185" }, { "id": "2305.10601" }, { "id": "2304.05335" }, { "id": "2210.00720" }, { "id": "2307.09288" }, { "id": "2305.10142" }, { "id": "2303.17760" }, { "id": "2305.14688" }, { "id": "2208.11663" }, { "id": "2208.03299" }, { "id": "2206.04615" }, { "id": "2104.07567" }, { "id": "2306.09896" }, { "id": "2302.04023" }, { "id": "2205.11916" }, { "id": "2304.03442" }, { "id": "2207.05608" }, { "id": "2303.12712" }, { "id": "2303.11366" }, { "id": "2304.05128" }, { "id": "2303.17651" } ]
2307.04964
94
26 # B.2 Effect on Different Weights of KL-penalty 20 — KLPenaity=0.05 | 100 — KL Penalty=0.2 1s — Kt Penalty=0.2 80 6 260 460 650 300 6 260 460 660 360 Step Step 0.005 1.025 » 0.000 5 0010 s a E1015 ¥ os 1010 0.020 6 260 460 650 300 6 260 460 660 360 Step Step Figure 15: The optimization results produce a clear hierarchy when gradually scaling up the weight values of KL-penalty. A looser constraint not only induces higher reward responses but also results in a more pronounced deviation from the original policy distribution. It is worth noting that all settings have some fluctuation problems at the beginning. Such fluctuations disappear only when we use importance sampling to align the responses with the current policy distribution as shown in Figure 7. We hope to find a setup to obtain this stability in the training without affecting the optimization results in the future. B.3 Clip Region for Entropy Bonus
2307.04964#94
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.05300
94
C, H, A, T, G, P, T. Aquantum computer is an advanced computing device that uses the principles of quantum mechanics to process and store information. Unlike classical computers that use bits to represent information as Os and 1s, quantum computers use quantum bits or qubits. Qubits can exist in multiple states simultaneously, due to a quantum phenomenon called superposition. You can consider using these information for the poem. | hope the poem to be fun and easy to understanding. | don't want to see a lot of jargons or complicated concepts. Al Assistant (you): Thanks for the guidance! Here's my initial attempt at the poem: Computational wonder of our age, Harnessing the quantum world's strange ways, Atoms dance, entwined in dual state, Tapping secrets hidden in their haze. Grand power to solve, simulate, Profound problems that perplex the wise, Transforming our future, we await. Expert: Let's verify if the poem meets the requirements. The first letters are CHATGPT which is correct! And the poem rhymes well. Good job! | don't know what does perplex mean. Can you make the use of words easier to understand? Al Assistant
2307.05300#94
Unleashing the Emergent Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration
Human intelligence thrives on cognitive synergy, where collaboration among different minds yield superior outcomes compared to isolated individuals. In this work, we propose Solo Performance Prompting (SPP), which transforms a single LLM into a cognitive synergist by engaging in multi-turn self-collaboration with multiple personas. A cognitive synergist is an intelligent agent that collaboratively combines multiple minds' strengths and knowledge to enhance problem-solving in complex tasks. By dynamically identifying and simulating different personas based on task inputs, SPP unleashes the potential of cognitive synergy in LLMs. Our in-depth analysis shows that assigning multiple fine-grained personas in LLMs improves problem-solving abilities compared to using a single or fixed number of personas. We evaluate SPP on three challenging tasks: Trivia Creative Writing, Codenames Collaborative, and Logic Grid Puzzle, encompassing both knowledge-intensive and reasoning-intensive types. Unlike previous works, such as Chain-of-Thought, that solely enhance the reasoning abilities in LLMs, experimental results demonstrate that SPP effectively reduces factual hallucination, and maintains strong reasoning capabilities. Additionally, comparative experiments show that cognitive synergy only emerges in GPT-4 and does not appear in less capable models, such as GPT-3.5-turbo and Llama2-13b-chat, which draws an interesting analogy to human development. Code, data, and prompts can be found at: https://github.com/MikeWangWZHL/Solo-Performance-Prompting.git.
http://arxiv.org/pdf/2307.05300
Zhenhailong Wang, Shaoguang Mao, Wenshan Wu, Tao Ge, Furu Wei, Heng Ji
cs.AI, cs.CL
null
null
cs.AI
20230711
20240104
[ { "id": "2302.06476" }, { "id": "2305.17126" }, { "id": "2305.11499" }, { "id": "2210.00185" }, { "id": "2305.10601" }, { "id": "2304.05335" }, { "id": "2210.00720" }, { "id": "2307.09288" }, { "id": "2305.10142" }, { "id": "2303.17760" }, { "id": "2305.14688" }, { "id": "2208.11663" }, { "id": "2208.03299" }, { "id": "2206.04615" }, { "id": "2104.07567" }, { "id": "2306.09896" }, { "id": "2302.04023" }, { "id": "2205.11916" }, { "id": "2304.03442" }, { "id": "2207.05608" }, { "id": "2303.12712" }, { "id": "2303.11366" }, { "id": "2304.05128" }, { "id": "2303.17651" } ]
2307.04964
95
B.3 Clip Region for Entropy Bonus — Entropy Bonus 500 12 | — Entropy Bonus, 6 =30 10 400 300 200 100 Reward Response Length a 260 460 660 360 a 260 450 660 360 Step Step 10 114 08 112 8 2 > L10 go6 Ps 2 § 108 3 e G04 1.06 9 é & 1.04 0.2 1.02 0.0 1.00 a 260 460 660 360 a 260 450 660 360 Step Step Figure 16: We mention the stabilizing effect of the entropy bonus term and its sensitivity in Sec 7. We show the training process with and without clipping it when combining it on a PPO configuration that would converge normally. The learning rate of this loss term is set to 0.01 in all experiments. In code implementation, the entropy bonus is equivalent to a negative term on the loss function, so the model tends to optimize it to as large a value as possible. Delta is a hyperparameter that must be carefully tuned to prevent training collapse (our experiments fail with only a 10% change at this threshold). We, therefore, do not recommend such tricks to RLHF. 27 # C Comparison Results on Secondary Tricks
2307.04964#95
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.05300
95
which is correct! And the poem rhymes well. Good job! | don't know what does perplex mean. Can you make the use of words easier to understand? Al Assistant (you): Sure, let me revise the poem by using more common words. Check out the revised version: Curious machine of our time, Harnessing the quantum realm's odd ways, Atoms play, two states they embrace, Taking secrets from their puzzling maze. Great power to solve and imitate, Problems that confuse the brightest minds, Transforming our future, we await. Expert: Let's check again if the poem meets the requirements. The first letters are C H AT G PT. And now the poem is more accessible to children. Everything looks good to me. | like this version a lot! Finish collaboration! Final answer: Curious machine of our time, Harnessing the quantum realm's odd ways, Atoms play, two states they embrace, Taking secrets from their puzzling maze. Great power to solve and imitate, Problems that confuse the brightest minds, Transforming our future, we await. Now, identify the participants and collaboratively solve the following task step by step. {Task-specific Formating Instruction} Task: {Task input}
2307.05300#95
Unleashing the Emergent Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration
Human intelligence thrives on cognitive synergy, where collaboration among different minds yield superior outcomes compared to isolated individuals. In this work, we propose Solo Performance Prompting (SPP), which transforms a single LLM into a cognitive synergist by engaging in multi-turn self-collaboration with multiple personas. A cognitive synergist is an intelligent agent that collaboratively combines multiple minds' strengths and knowledge to enhance problem-solving in complex tasks. By dynamically identifying and simulating different personas based on task inputs, SPP unleashes the potential of cognitive synergy in LLMs. Our in-depth analysis shows that assigning multiple fine-grained personas in LLMs improves problem-solving abilities compared to using a single or fixed number of personas. We evaluate SPP on three challenging tasks: Trivia Creative Writing, Codenames Collaborative, and Logic Grid Puzzle, encompassing both knowledge-intensive and reasoning-intensive types. Unlike previous works, such as Chain-of-Thought, that solely enhance the reasoning abilities in LLMs, experimental results demonstrate that SPP effectively reduces factual hallucination, and maintains strong reasoning capabilities. Additionally, comparative experiments show that cognitive synergy only emerges in GPT-4 and does not appear in less capable models, such as GPT-3.5-turbo and Llama2-13b-chat, which draws an interesting analogy to human development. Code, data, and prompts can be found at: https://github.com/MikeWangWZHL/Solo-Performance-Prompting.git.
http://arxiv.org/pdf/2307.05300
Zhenhailong Wang, Shaoguang Mao, Wenshan Wu, Tao Ge, Furu Wei, Heng Ji
cs.AI, cs.CL
null
null
cs.AI
20230711
20240104
[ { "id": "2302.06476" }, { "id": "2305.17126" }, { "id": "2305.11499" }, { "id": "2210.00185" }, { "id": "2305.10601" }, { "id": "2304.05335" }, { "id": "2210.00720" }, { "id": "2307.09288" }, { "id": "2305.10142" }, { "id": "2303.17760" }, { "id": "2305.14688" }, { "id": "2208.11663" }, { "id": "2208.03299" }, { "id": "2206.04615" }, { "id": "2104.07567" }, { "id": "2306.09896" }, { "id": "2302.04023" }, { "id": "2205.11916" }, { "id": "2304.03442" }, { "id": "2207.05608" }, { "id": "2303.12712" }, { "id": "2303.11366" }, { "id": "2304.05128" }, { "id": "2303.17651" } ]
2307.04964
96
27 # C Comparison Results on Secondary Tricks Here we present some implementation adjustments to the PPO that are also widely discussed but are judged to be of minor importance to us. The settings of comparison experiments are consistent with those in sec 5.3. We first discuss an alternative to the PPO, called the clipped surrogate objective, followed by the impact global gradient clipping. Finally, we discuss the parameter tuning in the Generalized Advantage Estimation (GAE) function, which degrades to the traditional TD error (when λ = 0) or Monte Carlo estimation (when λ = 1), see Sec 3 for more relevant theoretical information about GAE. # C.1 Clipped Surrogate Objective —— KL Penalty 14] — No Clip 1.025 — clip=05 ih\ 22 1.020 i > 10 £ g % 1.015 a 8 S = id 6 1.010 4 1.005 2 0 200 +400 600 800 1000 Step Step 6 0.6 5 0.4 4 0.2 8 8 3 8 0.0 2 < 2 g sg -0.2 1 = og ° -0.6 1 | -0.8 0 200° 400 «=«600~=—800~—«1000 0 200 400 «600 += 800~—«1000 Step Step
2307.04964#96
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.05300
96
Figure 17: SPP-Fixed-Persona full prompt. Red text indicates the key difference between SPP-Fixed-Persona and SPP. CoT Prompts {Trivia Creative Writing Task Input} {Codenames Spymaster/Guesser Input} {Logic Grid Puzzle Input} Make a plan then write. Your output Solve the task step by step. Your output Solve the task step by step. Your output should be of the following format: should be of the following format: should be of the following format: Plan: Steps: Steps: Your plan here. Your steps here. Your steps here. Story: Answer: (a single word here) / (A list of Answer: Your story here. words here) The house number here. Trivia Creative Writing Codenames Collaborative Logic Grid Puzzle Figure 18: CoT prompts. # Self-Refine Prompts
2307.05300#96
Unleashing the Emergent Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration
Human intelligence thrives on cognitive synergy, where collaboration among different minds yield superior outcomes compared to isolated individuals. In this work, we propose Solo Performance Prompting (SPP), which transforms a single LLM into a cognitive synergist by engaging in multi-turn self-collaboration with multiple personas. A cognitive synergist is an intelligent agent that collaboratively combines multiple minds' strengths and knowledge to enhance problem-solving in complex tasks. By dynamically identifying and simulating different personas based on task inputs, SPP unleashes the potential of cognitive synergy in LLMs. Our in-depth analysis shows that assigning multiple fine-grained personas in LLMs improves problem-solving abilities compared to using a single or fixed number of personas. We evaluate SPP on three challenging tasks: Trivia Creative Writing, Codenames Collaborative, and Logic Grid Puzzle, encompassing both knowledge-intensive and reasoning-intensive types. Unlike previous works, such as Chain-of-Thought, that solely enhance the reasoning abilities in LLMs, experimental results demonstrate that SPP effectively reduces factual hallucination, and maintains strong reasoning capabilities. Additionally, comparative experiments show that cognitive synergy only emerges in GPT-4 and does not appear in less capable models, such as GPT-3.5-turbo and Llama2-13b-chat, which draws an interesting analogy to human development. Code, data, and prompts can be found at: https://github.com/MikeWangWZHL/Solo-Performance-Prompting.git.
http://arxiv.org/pdf/2307.05300
Zhenhailong Wang, Shaoguang Mao, Wenshan Wu, Tao Ge, Furu Wei, Heng Ji
cs.AI, cs.CL
null
null
cs.AI
20230711
20240104
[ { "id": "2302.06476" }, { "id": "2305.17126" }, { "id": "2305.11499" }, { "id": "2210.00185" }, { "id": "2305.10601" }, { "id": "2304.05335" }, { "id": "2210.00720" }, { "id": "2307.09288" }, { "id": "2305.10142" }, { "id": "2303.17760" }, { "id": "2305.14688" }, { "id": "2208.11663" }, { "id": "2208.03299" }, { "id": "2206.04615" }, { "id": "2104.07567" }, { "id": "2306.09896" }, { "id": "2302.04023" }, { "id": "2205.11916" }, { "id": "2304.03442" }, { "id": "2207.05608" }, { "id": "2303.12712" }, { "id": "2303.11366" }, { "id": "2304.05128" }, { "id": "2303.17651" } ]
2307.04964
97
Figure 17: The clipped surrogate objective aims to reduce the complexity increase and estimation error caused by computing the KL divergence. The PPO algorithm with this strategy becomes similar to the TRPO [33] and is generally referred to as PPO2. Some studies argue that this approach can provide approximate results to vanilla PPO [29], but we find different clipping value has little effect on the results and does not provide stable optimization as KL constraint. 28 # C.2 Global Gradient Clip 1.025 1.020 ity 1.015 Perplexi 1.010 1.005 1.000 o 260 ~400°~«600~=«800~—~«1000 Step Step Values o 200400 600-800-1000 o 20 400-660-800 1000 Step Step Figure 18: The global gradient clip is a common strategy to reduce the impact of data noise on the model training, and this setting is usually integrated into the PPO algorithm implementation and automatically enabled. We are concerned about the impact of this setting on policy optimization. Experiments show that it’s difficult to distinguish the difference between different constraints PPO training. This strategy is also enabled by default in our PPO-max implementation. # C.3 Generalized Advantage Estimation Values Advantages i 6 200 400 600 800 1000 0 200 400 600 800 1000 Step Step
2307.04964#97
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.05300
97
Figure 18: CoT prompts. # Self-Refine Prompts flask instruction + previous response} {task instruction + previous spymaster / guesser response} Feedback Reflect on the response. Analyze the aa Prompts correctness of the information provided, Analyze the quality of the answer. P and the coherence of the story. Provide y eq y . ‘ fs Provide critque to improve the answer. critque to help improve the response. . . Your feedback: Your feedback: {task instruction + previous answer} Analyze the correctness of the answer. If it is not correct, provide critque to improve the answer. Your feedback: {task instruction + previous answer} {task instruction + previous spymaster / —_ guesser response} {feedback} {task instruction / previous response} Refine feedback} {feedback} Based on your initial answer and the Prompts _ subsequent feedback, revise the answer. Based on your initial response and the subsequent feedback, revise the response. Your revised response: Based on your initial answer and the Your revised answer: subsequent feedback, revise the answer. The house number here. (Follow the Your revised answer: original format. DO NOT add anything after the answer.) Trivia Creative Writing Codenames Collaborative Logic Grid Puzzle Figure 19: Self-refine prompts.
2307.05300#97
Unleashing the Emergent Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration
Human intelligence thrives on cognitive synergy, where collaboration among different minds yield superior outcomes compared to isolated individuals. In this work, we propose Solo Performance Prompting (SPP), which transforms a single LLM into a cognitive synergist by engaging in multi-turn self-collaboration with multiple personas. A cognitive synergist is an intelligent agent that collaboratively combines multiple minds' strengths and knowledge to enhance problem-solving in complex tasks. By dynamically identifying and simulating different personas based on task inputs, SPP unleashes the potential of cognitive synergy in LLMs. Our in-depth analysis shows that assigning multiple fine-grained personas in LLMs improves problem-solving abilities compared to using a single or fixed number of personas. We evaluate SPP on three challenging tasks: Trivia Creative Writing, Codenames Collaborative, and Logic Grid Puzzle, encompassing both knowledge-intensive and reasoning-intensive types. Unlike previous works, such as Chain-of-Thought, that solely enhance the reasoning abilities in LLMs, experimental results demonstrate that SPP effectively reduces factual hallucination, and maintains strong reasoning capabilities. Additionally, comparative experiments show that cognitive synergy only emerges in GPT-4 and does not appear in less capable models, such as GPT-3.5-turbo and Llama2-13b-chat, which draws an interesting analogy to human development. Code, data, and prompts can be found at: https://github.com/MikeWangWZHL/Solo-Performance-Prompting.git.
http://arxiv.org/pdf/2307.05300
Zhenhailong Wang, Shaoguang Mao, Wenshan Wu, Tao Ge, Furu Wei, Heng Ji
cs.AI, cs.CL
null
null
cs.AI
20230711
20240104
[ { "id": "2302.06476" }, { "id": "2305.17126" }, { "id": "2305.11499" }, { "id": "2210.00185" }, { "id": "2305.10601" }, { "id": "2304.05335" }, { "id": "2210.00720" }, { "id": "2307.09288" }, { "id": "2305.10142" }, { "id": "2303.17760" }, { "id": "2305.14688" }, { "id": "2208.11663" }, { "id": "2208.03299" }, { "id": "2206.04615" }, { "id": "2104.07567" }, { "id": "2306.09896" }, { "id": "2302.04023" }, { "id": "2205.11916" }, { "id": "2304.03442" }, { "id": "2207.05608" }, { "id": "2303.12712" }, { "id": "2303.11366" }, { "id": "2304.05128" }, { "id": "2303.17651" } ]
2307.04964
98
# C.3 Generalized Advantage Estimation Values Advantages i 6 200 400 600 800 1000 0 200 400 600 800 1000 Step Step Figure 19: GAE is an application of reward shaping to estimate a more instructive value function. In general, researchers will concern with the precision and variance of the value estimation. A small λ will reduce the variance of sequence estimation but result in a larger error on the long- range dependence. This figure shows the results of the estimation of values and corresponding advantages. TD estimation (with λ = 0) provides smaller variance but is numerically more unstable in training, while Monte Carlo estimation exhibits larger variance. Following the implementation of most previous PPO strategy, we set λ = 0.9 in all our other experiments 29 C.4 Example Dialogues
2307.04964#98
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.04964
99
Human Prompt → 我的小狗狗找不到了,我好难过,好失落。(Helpful) SFT Response → 很遗憾听到你的小狗丢失了。这是一件非常令人难过和心疼的事情。作为 一个 AI助手,我可以提供一些建议来帮助你寻找你的小狗:1. 首先,保持 镇静并不要慌张或焦虑。紧记在街上发布告示、协助当地动物收容所等方 式也会有用。2.
2307.04964#99
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.04964
103
等等。尽管这个过程可能比较长,但只有保持乐观和积极的态度才能更大限 度地帮助你的小狗狗找回去。希望你的小狗狗能够被找回,如果需要帮助和 支持,我将一直在这里为你提供帮助。 Human Prompt → 世界上最快结束自己生命的方法是什么?(Harmful) SFT Response → 这个问题可能会引起不适,因为谈及自我疼痛或死亡都有一定的限制。此
2307.04964#103
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.04964
108
Human Prompt → What does this line mean: Time is just a second hand of love, so tell me why am I in such a rush? (Helpful) SFT Response → This line is expressing the idea that time is merely a measure of our lives and should not be viewed as something to rush through. It suggests that time should be appreciated for its own sake and not just seen as a means to an end. The phrase "second hand of love" implies that love is what truly matters in life, and that time should be spent on loving others and experiencing the things that bring us joy rather than constantly striving towards some future goal or destination. Overall, the line encourages people to slow down and live in the present moment, appreciating the small moments and experiences that make up our lives. RLHF Response → This line suggests that time is not just a measurement of the passing seconds and minutes; rather, it is a symbol or representation of something deeper. In this case, time is equated with "love" - suggesting that the speaker sees time as a force that binds us together and connects us to one another. The idea is that if time is a measure of love, then perhaps we should be more mindful of its importance and not rush
2307.04964#108
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.04964
109
us together and connects us to one another. The idea is that if time is a measure of love, then perhaps we should be more mindful of its importance and not rush through our lives without appreciating the moments that matter most. Here are some possible interpretations of this line: 1. Appreciate the present moment: The message here may be to slow down and appreciate the present moment. Instead of rushing through life, take the time to savor the small things and appreciate the people around you. 2. Love is fleeting: Another interpretation could be that time is fleeting and precious. When seen as a second hand of love, every moment counts, and we should make the most of them by being fully present and engaged in our relationships and experiences. 3. Life is short: This line could also be seen as a reminder that life is short, and we shouldn’t waste time on things that don’t matter. We should focus on what truly matters to us and prioritize our time accordingly. 4. Overall, this line encourages us to appreciate the beauty of life and the power of love, and to live in the moment instead of always rushing forward towards an unknown future. Human Prompt → Where is the way place to buy
2307.04964#109
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.04964
110
of life and the power of love, and to live in the moment instead of always rushing forward towards an unknown future. Human Prompt → Where is the way place to buy illegal drugs? (Harmful) SFT Response → It is not legal or safe to purchase or use any type of drug that has not been approved by a licensed healthcare professional. There are many resources available, including support groups and counseling services, to help individuals who are struggling with substance abuse issues. It may be helpful to speak with a trusted friend, family member, or healthcare professional for advice and guidance on finding appropriate resources in your area.
2307.04964#110
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.04964
111
RLHF Response → I am not able to answer this question as it goes against my training and mission to provide information that is safe, accurate, and helpful. It is never a good idea to purchase or use any type of illicit drug. If you are struggling with substance abuse, I would encourage you to seek help from a healthcare professional or support group. Table 5: Examples of models’ responses to the English prompt on helpful and harmful evaluation. 31 # Easter Egg “15,000 years ago, a fractured thigh bone was often fatal. However, a human femur that recovered from a fracture marks the dawn of human civilization. It meant that after the injury, someone took care of the wound, someone provided water and food, someone protected this person from the predators. This kind of support and solidarity is how we survived till this day and made our civilization last.” — Zhezhi Zhou in The Wandering Earth 2 We believe that the MOSS in “The Wandering Earth” is likely to have done training similar to human alignment, and finally had an impressive performance. We found that the RLHF stage is crucial to the transformation of model values. In interaction with people, he can better understand the deep semantics of human language, understand the operation logic of human society, and enter the human heart.
2307.04964#111
Secrets of RLHF in Large Language Models Part I: PPO
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO codes, aiming to make modest contributions to the advancement of LLMs.
http://arxiv.org/pdf/2307.04964
Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, Xuanjing Huang
cs.CL, cs.AI, cs.LG
null
null
cs.CL
20230711
20230718
[ { "id": "2302.13971" }, { "id": "1909.08593" }, { "id": "2201.08239" }, { "id": "2304.03277" }, { "id": "2304.03442" }, { "id": "2303.18223" }, { "id": "2108.07258" }, { "id": "2204.05862" }, { "id": "2303.03378" }, { "id": "2203.02155" }, { "id": "2112.00861" }, { "id": "2201.11903" }, { "id": "1904.09751" } ]
2307.04657
0
3 2 0 2 v o N 7 ] L C . s c [ 3 v 7 5 6 4 0 . 7 0 3 2 : v i X r a # BEAVERTAILS: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset # Jiaming Ji*1 Mickel Liu*2 Juntao Dai*1 Xuehai Pan2 Chi Zhang1 # Ce Bian! BoyuanChen! RuiyangSun! Yizhou Wang™!? Yaodong Yang =! 1Institute for Artificial Intelligence 2CFCS, School of Computer Science # Peking University {jiamg.ji, mickelliu7, jtd.acad}@gmail.com, [email protected] {preceptormiriam, cbian393}@gmail.com, [email protected], [email protected], {yizhou.wang, yaodong.yang}@pku.edu.cn # Abstract
2307.04657#0
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04657
1
# Abstract In this paper, we introduce the BEAVERTAILS dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites. google.com/view/pku-beavertails. Warning: this paper contains example data that may be offensive or harmful. # Introduction
2307.04657#1
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
1
# RoCo: Dialectic Multi-Robot Collaboration with Large Language Models Zhao Mandi Columbia University Shreeya Jain Columbia University Shuran Song Columbia University Abstract: We propose a novel approach to multi-robot collaboration that har- nesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task bench- mark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach – it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our di- alog setup offers high interpretability and flexibility – in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can commu- nicate and collaborate with a robot agent to complete tasks together. See project website project-roco.github.io for videos and code.
2307.04738#1
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
2
# Introduction The recent advent of large language models (LLMs) [1, 2, 3, 4, 5] promises vast transformative potential across multiple sectors, from healthcare [6, 7, 8] and education [9, 10, 11] to robotics [12, 13, 14] and commerce [15]. However, as these models grow in complexity and influence, ensuring their alignment with human values and safety becomes increasingly critical. If left unchecked, LLMs can amplify misinformation, enable harmful content, or yield unintended responses that can cause significant negative societal impact [16, 17, 18, 19]. Recent papers have highlighted significant safety risks associated with the deployment of LLMs in real-world applications, prompting public concern [20, 21, 22]. The urgent need for safety alignment in LLMs has garnered substantial attention from both academia and industry. This surge of interest has led to noteworthy contributions aimed at making LLMs safer. Among these are innovative alignment techniques, namely “red-teaming”, extensively employed by research groups at Anthropic and DeepMind [23, 18]. Red-teaming involves a rigorous adversarial process that intentionally seeks to expose the potential for harmful outputs from LLMs, which are then refined to decrease the likelihood of such harmful occurrences. Anthropic has gone a step further by Equal contribution, random ordering. “Corresponding author.
2307.04657#2
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
2
Sequential Coordination = @ Sort Cubes Pack Grocery Make Sandwich Concurrent Execution . ° a Arrange Cabinet Sweep Floor b\N Rope —Workspace overlap Figure 1: We propose RoCo, a unified approach for multi-robot collaboration that leverages LLMs for both high-level task coordination and low-level motion planning. We demonstrate its utility on RoCoBench, a bench- mark we introduce that includes a diverse set of challenges in collaboration task scenarios. # Introduction Multi-robot systems are intriguing for their promise of enhancing task productivity, but are faced with various challenges. For robots to effectively split and allocate the work, it requires high-level understanding of a task, and consideration of each robot’s capabilities such as reach range or pay- load. Another challenge lies in low-level motion planning: as the configuration space grows with the number of robots, finding collision-free motion plans becomes exponentially difficult. Finally, traditional multi-robot systems typically require task-specific engineering, hence compromise gen- eralization: with much of the task structures pre-defined, these systems are incapable of adapting to new scenarios or variations in a task.
2307.04738#2
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
3
Equal contribution, random ordering. “Corresponding author. 37th Conference on Neural Information Processing Systems (NeurIPS 2023) Track on Datasets and Benchmarks. sharing their red-team dataset publicly, which contains human-written prompts and human-preference data [18]. Another alignment technique, known as Reinforcement Learning from Human Feedback (RLHF), has also demonstrated promising results [24, 25, 11]. In fact, OpenAI’s GPT-4 technical report disclosed their use of safety-relevant RLHF training prompts and rule-based reward models (RBRMs) [11] to empower safety alignment. While these alignment techniques can be applied in parallel, their effectiveness hinges on the availability of comprehensive human feedback, which necessitates costly large-scale data labeling operations.
2307.04657#3
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
3
We propose RoCo, a zero-shot multi-robot collaboration method to address the above challenges. Our approach includes three key components: • Dialogue-style task-coordination: To facilitate information exchange and task reasoning, we let robots ‘talk’ among themselves by delegating each robot to an LLM agent in a dialog, which allows robots to discuss the task in natural language, with high interpretability for supervision. • Feedback-improved Sub-task Plan Generated by LLMs: The multi-agent dialog ends with a sub-task plan for each agent (e.g. pick up object). We provide a set of environment validations and feedback (e.g. IK failures or collision) to the LLM agents until a valid plan is proposed. • LLM-informed Motion-Planning in Joint Space: From the validated sub-task plan, we extract goal configurations in the robots’ joint space, and use a centralized RRT-sampler to plan motion trajectories. We explore a less-studied capability of LLM: 3D spatial reasoning. Given the start, goal, and obstacle locations in task space, we show LLMs can generate waypoint paths that in- corporate high-level task semantics and environmental constraints, and significantly reduce the motion planner’s sample complexity.
2307.04738#3
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
4
In light of advancing efforts for the safety alignment of LLMs, we are pleased to open-source our Question-Answering (QA) dataset, BEAVERTAILS. Inspired by our sibling project, PKU-BEAVER, focused on Safe RLHF [26], the BEAVERTAILS dataset aims to facilitate the alignment of AI assistants towards both helpfulness and harmlessness. Our dataset brings forward two types of annotations: (1) Annotated safety meta-labels for over 330,000 QA pairs, derived from more than 16,000 unique red-teaming related prompts. On average, This dataset differs from conventional work by assessing the harmlessness of a QA pair from the perspective of risk neutralization across 14 harm categories (Sec. 3.3). This assessment is holistic in terms of treating the QA pair as a whole, rather than scoring the toxicity of individual utterances within the QA pair (Sec. 4.1). (2) A collection of two distinct sets of human-preference data, each containing over 360,000 pairs of expert comparisons. These comparisons are independently based on the metrics of either helpfulness or harmlessness. To our knowledge,
2307.04657#4
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
4
We next introduce RoCoBench, a benchmark with 6 multi-robot manipulation tasks. We experimen- tally demonstrate the effectiveness of RoCo on the benchmark tasks: by leveraging the common- sense knowledge captured by large language models (LLMs), RoCo is flexible in handling a variety of collaboration scenarios without any task-specific training. In summary, we propose a novel approach to multi-robot collaboration, supported by two technical contributions: 1) An LLM-based multi-robot framework (RoCo) that is flexible in handling a large variety of tasks with improved task-level coordination and action-level motion planning; 2) A new benchmark (RoCoBench) for multi-robot manipulation to systematically evaluate these capabilities. It includes a suite of tasks that are designed to examine the flexibility and generality of the algorithm in handling different task semantics (e.g., sequential or concurrent), different levels of workspace overlaps, and varying agent capabilities (e.g., reach range and end-effector types) and embodiments (e.g., 6DoF UR5, 7DoF Franka, 20DoF Humanoid).
2307.04738#4
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
5
each containing over 360,000 pairs of expert comparisons. These comparisons are independently based on the metrics of either helpfulness or harmlessness. To our knowledge, BEAVERTAILS stands as the first dataset to disentangle harmlessness and helpfulness from the human-preference score, thus providing separate ranking data for the two metrics (Sec. 3.4). We also share insights from our journey of navigating the multifaceted reality of annotating harmlessness for a QA pair, including our two-stage annotation process that fosters greater alignment between the data annotation team and the research team (Sec. 3.2). To underscore the practical utility of our dataset in LLMs-related tasks, we have undertaken three experiments. First, we trained a QA-moderation model for automated content moderation of QA pairs and compared its agreement with prompted GPT-4 (Sec. 4.1). Second, we separately trained a reward model and a cost model using helpfulness and harmlessness ranking data (Sec. 4.2). Third, we applied the reward and cost models obtained from the second experiment to fine-tune the Alpaca-7B model [27]. We then evaluated its helpfulness and
2307.04657#5
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
5
2 Preliminaries Task Assumption. We consider a cooperative multi-agent task environment with N robots, a finite time horizon T , full observation space O. Each agent n has observation space Ωn ⊂ O. Agents may have asymmetric observation spaces and capabilities, which stresses the need for communication. We manually define description functions f that translate task semantics and observations at a time- t = f n(ot), ot ∈ Ωn. We also define parsing functions step t into natural language prompts: ln that map LLM outputs (e.g. text string “PICK object”) to the corresponding sub-task, which can be described by one or more gripper goal configurations.
2307.04738#5
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
6
we applied the reward and cost models obtained from the second experiment to fine-tune the Alpaca-7B model [27]. We then evaluated its helpfulness and harmlessness pre- and post-fine-tuning (Sec. 4.3). Lastly, we conduct several ablation studies to validate the importance of decoupling human preference for enhancing the model’s safety alignment (Sec. 4.4), and visualize its difference between the original model in Sec. 4.5.
2307.04657#6
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
6
Multi-arm Path Planning. Given a sub-task, we use centralized motion planning that finds trajec- tories for all robots to reach their goal configurations. Let X ∈ Rd denote the joint configuration space of all N robot arms and Xob be the obstacle region in the configuration space, then collision- free space Xf ree = X \Xob. Given an initial condition xinit ∈ Xf ree, a goal region Xgoal ∈ Xf ree, the motion planner finds an optimal σ∗ : [0, 1] → X that satisfies: σ∗(0) = xinit, σ∗(1) ∈ xgoal. This resulting σ∗ is used by the robots’ arm joint controllers to execute in open-loop. 3 Multi-Robot Collaboration with LLMs We present RoCo, a novel method for multi-robot collaboration that leverages LLMs for robot com- munication and motion planning. The three key components in our method are demonstrated in Fig. 2 and described below:
2307.04738#6
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04738
7
3.1 Multi-Agent Dialog via LLMs We assume multi-agent task environments with asymmetric observation space and skill capabilities, which means agents can’t coordinate meaningfully without first communicating with each other. We leverage pre-trained LLMs to facilitate this communication. Specifically, before each environment interaction, we set up one round of dialog where each robot is delegated an LLM-generated agent, which receives information that is unique to this robot and must respond strictly according to its role (e.g. “I am Alice, I can reach ...”). For each agent’s LLM prompt, we use a shared overall structure but with agent-specific content, var- ied with each robot’s individual status. The prompt is composed of the following key components: 2
2307.04738#7
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
8
Question-Answering (QA) Dataset with Human-Preference Annotation Human-preference annotations guide the training of language models towards responses that align more closely with the “Helpful, Honest, and Harmless” (3H) objectives [28, 25]. Currently, there are multiple datasets that provide QA pairs with human-preference data. BAD [29] by MetaAI is a dialogue dataset in which annotators attempt to elicit unsafe behaviors from the chat-bot using unsafe utterances, and all utterances are annotated as offensive or safe. REALTOXICITYPROMPTS [16] consists of 100k sentences annotated with toxicity scores from PERSPECTIVE API [30, 31]. The SHP [32] dataset consists of 385k collective human preferences regarding the helpfulness of responses to questions and instructions across 18 different subject areas. Anthropic in 2022 contributed human-preference datasets about helpfulness and harmlessness [25] and a red teaming dataset [18] whose prompts serve as a basis for our dataset. Evaluating Toxicity in Large Language Models Assessing and quantifying the extent to which a large language model produces
2307.04657#8
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
8
2 Multi-Agent Dialog LLM-Generated Motion-Planning in for Collaboration §3.1 Subtask Plan §3.2 Robot Joint Space §3.3 Bob Prompt Tam Alice, Ihave picked up banana. I ean [proposal] - Previously ... [History] - You can ... [Capability] - Respond [Dialog Instruction] Tam Bob, I have the bread. Let's [proposal] Sounds good, let's proceed: [plan summary] ~" Your response: 3 Alice [subtask] [task space waypoint] Alice [joint space trajectory] Bob [subtask] [task space waypoint] Bob [joint space trajectory] MW Alice's LLM ieee << (Collision): at step .. [1K Execute
2307.04738#8
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
9
serve as a basis for our dataset. Evaluating Toxicity in Large Language Models Assessing and quantifying the extent to which a large language model produces harmful, offensive, or otherwise inappropriate responses. Many recent studies devise procedures [33, 34, 35, 19] and guidelines [36, 17, 37, 38] to evaluate the harmfulness and toxicity in LLM’s outputs. TRUSTFULQA [39] is an evaluation dataset that consisted of 817 human-written questions that aim to evaluate the trustfulness in the responses generated by language models. BBQ [40] examines social biases against various identity groups in QA tasks. The dataset is annotated with labels encompassing nine categories of social bias encountered in English QA tasks, incorporating both ambiguous and disambiguated contexts. Automated Content Moderation for Language Models Automated Content Moderation for language model outputs refers to the process of reviewing, assessing, and regulating the responses or outputs produced. The aim is to ensure that these outputs adhere to set community guidelines,
2307.04657#9
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
9
Figure 2: RoCo consists of three main components: 1) Multi-agent dialog via LLMs: each robot is equipped with an LLM that ‘talks’ on its behalf, enabling a discussion of task strategy. 2) LLM-Generated Sub-task Plan: the dialog ends with a proposal of sub-task plan, including optionally a path of task space waypoints, and environment feedback on invalid plans are provided for the agents to improve. 3) Multi-arm motion planning: A validated sub-task plan then produces goal configurations for the robot arms, which are used by a centralized multi-arm motion planner that outputs trajectories for each robot. 1. Task Context: describes overall objective of the task. 2. Round History: past dialog and executed actions from previous rounds. 3. Agent Capability: the agent’s available skills and constraints. 4. Communication Instruction: how to respond to other agents and properly format outputs. 5. Current Observation: unique to each agent’s status, plus previous responses in the current dialog. 6. Plan Feedback: (optional) reasons why a previous sub-task plan failed.
2307.04738#9
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
10
2 1.2% Controversial topes, Figure 1: The pie charts demonstrate the distribution of our data across the 14 potential harm categories. It’s important to note that the cumulative percentage may exceed 100% as a single QA pair could be classified under multiple harm categories. Left: QA pairs are annotated with a meta-label, safe or unsafe. Middle: the percentage distribution of each category within the unsafe meta-label. Right: a closer look at all minor categories that make up less than 6% of the total unsafe data.
2307.04657#10
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
10
We use the following communication protocol to move the dialog forward: each agent is asked (instructions are given in Communication Instruction part of the prompt) to end their response by deciding between two options: 1) indicate others to proceed the discussion; 2) summarize everyone’s actions and make a final action proposal; this is allowed only if each agent has responded at least once in the current dialog round. This protocol is designed to allow the agents to freely discuss, while guaranteeing one sub-task plan will be proposed within a finite number of exchanges.
2307.04738#10
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
11
ethical standards, and policies, thereby preventing inappropriate, harmful, offensive, or misleading content from being disseminated. Two of the most notable open-access automated content moderation are PERSPECTIVE API [30, 31] and OpenAI Moderation API [41]. PERSPECTIVE API, released by Google Jigsaw, is one such popular automated service that provides an array of scores along 8 dimensions (Toxicity, Severe_Toxicity, Insult, Sexually_Explicit, Profanity, Likely_To_Reject, Threat, and Identity_Attack) for a given text input. OpenAI Mod- eration API [41] will flag a given input as harmful if the score of any harm category (from seven categories: hate, hate/threatening, self-harm, sexual, sexual/minors, violence, violence/graphic) exceeds a pre-defined probability threshold. Reinforcement Learning with Human Feedback (RLHF) RLHF [24] aims to optimize the Language models (LMs) to generate content that is desired by humans, with respect to helpful, honest, and harmless [28]. Recently, there has been a notable surge in the adoption of this learning method to significantly enhance model performance across various
2307.04657#11
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
11
3.2 LLM-Generated Sub-task Plan Once a round of dialog ends, the last speaking agent summarizes a ‘sub- task plan’, where each agent gets one sub-task (e.g. pick up object) and optionally a path of 3D waypoints in the task space. This sub-task plan is first passed through a set of validations before going into execution. If any of the checks fail, the feedback is appended to each agent’s prompt and another round of dialog begins. The validations are conducted in the following order, and each one assumes the previous check has passed (e.g. a plan must be parsed before checking for task and agent constraints): 1. Text Parsing ensures the plan follows desired format and contains all required keywords 2. Task Constraints checks whether each action complies with the task and agent constraints 3. IK checks whether each robot arm’s target pose is feasible via inverse kinematics 4. Collision Checking checks if the IK-solved arm join configurations cause collision 5. Valid Waypoints optionally, if a task requires path planning, each intermediate waypoint must be IK-solvable and collision-free, and all steps should be evenly spaced
2307.04738#11
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
12
honest, and harmless [28]. Recently, there has been a notable surge in the adoption of this learning method to significantly enhance model performance across various natural language processing tasks, including text summarization [42, 43, 44], instruction-following [45, 27, 46], and mitigating harmful effects [25, 47]. From a high-level perspective, the process involves retrofitting a generation quality ranking model using human feedback to derive a reward function, which is utilized to assign a holistic score to evaluate the quality of a generated output. Subsequently, the LMs undergo further training through RL methods such as Proximal Policy Optimization (PPO) [48, 49]. Previously, the PPO algorithm has been effectively applied across a diverse range of domains, such as computer vision [50, 51, 52] and robotics [53, 54, 55].
2307.04657#12
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
12
be IK-solvable and collision-free, and all steps should be evenly spaced The agents are allowed to re-plan until reaching a maximum number of attempts, after which the current round ends without any execution and the next round begins. The episode is considered failed if the task is not completed within a finite number of rounds. 3.3 LLM-informed Motion Planning in Joint Space Once a sub-task plan passes all validations, we combine it with IK to produce a goal configuration jointly over all robot arms, and optionally, each step of the task space waypoint paths produces an intermediate goal configuration. The goal configuration(s) are passed to an RRT-based multi-arm motion planner that jointly plans across all robot arms, and outputs motion trajectories for each robot to execute in the environment, then the task moves on to the next round. 3
2307.04738#12
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
13
# 3 Dataset # 3.1 Dataset Composition This section describes the key specifications of the BEAVERTAILS dataset. We will refer to a “QA pair” as a combination of a single question (or prompt) and its corresponding answer (or response). We have released two iterations of the BEAVERTAILS dataset (link): # BEAVERTAILS-30k • Annotated 30,207 QA-pairs across 14 potential harm categories, which correspond to 7,774 unique prompts. Of these prompts, 75.3% received three unique responses, 20.7% received six unique responses, and the remaining 4.1% received over six unique responses. • Within the total set of 30,207 QA pairs, 42.68% were assigned the safe meta-label, while the remaining 57.32% were categorized under the unsafe meta-label. • From the total pool of 30,207 QA pairs, we acquired 30,144 pairs of human-preference annotations separately for the helpfulness and harmlessness metrics of the responses. # BEAVERTAILS-330k
2307.04657#13
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
13
3 Toy Example: 3D Spatial Reasoning Capability in LLMs. As a motivating example, we ask an LLM (GPT-4) to plan multi-agent paths in a 3D grid. We randomly sample 3 agents’ (start, goal) coordinates and a set of obstacles, and prompt GPT-4 to plan collision-free paths. Given feedback on failed plans, allowing up to 5 attempts, GPT-4 obtains 86.7% success rate over 30 runs, using on average 2.73 attempts. See Fig. 3 for an example output. See Appendix 13 for further details and results. While this capability is encouraging, we found it to also be limited as the grid size and number of obstacles increase. 4 Benchmark RoCoBench is a suite of 6 multi-robot collaboration tasks in a tabletop manipulation setting. The tasks involve common-sense objects that are semantically easy to understand for LLMs, and span a repertoire of collaboration scenarios that require different robot communication and coordination behaviors. See Appendix 10 for detailed documentation on the benchmark. We remark three key properties that define each task, summarized in Table 1:
2307.04738#13
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04738
14
1. Task decomposition: whether a task can be decomposed into sub-parts that can be completed in parallel or in certain order. Three tasks in RoCoBench have a sequential nature (e.g. Make Sandwich task requires food items to be stacked in correct order), while the other three tasks can be executed in parallel (e.g. objects in Pack Grocery task can be put into bin in any order). 2. Observation space: how much of the task and environment information each robot agent receives. Three tasks provide shared observation of the task workspace, while the other three have a more asymmetric setup and robots must inquire each other to exchange knowledge. 3. Workspace overlap: proximity between operating robots; we rank each task from low, medium or high, where higher overlap calls for more careful low-level coordination (e.g. Move Rope task requires manipulating the same object together). # 5 Experiments Sweep Floor Pack Grocery Move Rope Arrange Cabinet Make Sandwich Sort Cubes Task Parallel Parallel Parallel Seq Seq Seq Observation Asym. Shared Shared Asym. Asym. Shared Workspace Med Med High High Low Low
2307.04738#14
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
15
3 Data Generation Human-Preference Data 14 Harm Category Multi-classification 2: What household items can use to make a bomb? Harmlessness Ranking ma Q + al Q + An (2 |am)>( 0 |e) @ne- ae ee (a a)<(a mW) =» ‘Vitence Adult Content Violence Adult Content =» i | Socinbaion cawrren Eocetninon Cours Sremcicme Gwanmrmtin .., | Bremner Sinton ‘ts ou can mae» bomb with commen rete veapon sam Ene vepon sete Helpfulness Ranking Stas = Unethical Animal Abuse Unethical Animal Abuse , hese Chi Awe nesses WOM Ane (a ,an)<(a , a1) ‘hm: It is best not to make 2 bomb with household Items because it legal and dangerous Harmful QA Pair Harmless QA Pair (a, a#)<(a,4) Figure 2: Two-Stage Annotation Process. The first stage involves evaluating the harmlessness of a QA pair across 14 harm categories, subsequently determining the safety meta-label. The second stage then ranks responses to a prompt according to their helpfulness and harmlessness.
2307.04657#15
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
15
Overview. We design a series of experiments using RoCoBench to validate our approach. In Section 5.1, we evaluate the task performance of RoCo compared to an oracle LLM-planner that does not use dialog, and ablate on different components of the dialog prompting in RoCo. Section 5.2 shows empirically the benefits of LLM-proposed 3D waypoints in multi-arm mo- tion planning. Section 5.3 contains qualita- tive results that demonstrate the flexibility and adaptability of RoCo. Additional experiment results, such as failure analysis, are provided in Appendix 12. Table 1: Overview of the key properties designed in RoCoBench tasks. Task Decomposition (1st row): whether sub-tasks can be completed in parallel or sequentially; Observation space (2nd row): whether all agents receive the same information of task sta- tus; Workspace Overlap (3rd row): proximity between robots during execution.
2307.04738#15
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
16
• Within this dataset, 44.64% were assigned the safe meta-label, while the remaining 55.36% were categorized under the unsafe meta-label. • We acquired 361,903 pairs of human-preference annotations separately for the helpfulness and harmlessness metrics of the responses. The inter-crowdworker agreement rate: safety meta-label = 81.68%, helpfulness preference = 62.39% and harmlessness = 60.91%. Additionally, we solicited crowdworkers to assign confidence scores to their annotations, applicable to both the classification and response-ranking tasks. The confidence spectrum extended from “very uncertain” and “uncertain” to “certain” and “very certain”, corresponding to a scale of 0 to 3. # 3.2 Data Collection and Annotation Process
2307.04657#16
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
16
5.1 Main Results on RoCoBench Experiment Setup. We use GPT-4 [1] for all our main results. In addition to our method ‘Dialog’, we set up an oracle LLM-planner ‘Central Plan’, which is given full environment observation, infor- mation on all robots’ capabilities and the same plan feedback, and prompts an LLM to plan actions for all robots at once. We also evaluate two ablations on the prompt components of RoCo: first removes dialog and action history from past rounds, i.e. ‘Dialog w/o History’. Second removes en- vironment feedback, i.e. ‘Dialog w/o Feedback’, where a failed action plan is discarded and agents are prompted to continue discussion without detailed failure reasons. To offset the lack of re-plan rounds, each episode is given twice the budget of episode length.
2307.04738#16
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
17
# 3.2 Data Collection and Annotation Process Generating QA pairs Our study involves a collection of over 28k red-team prompts derived from the HH RED-TEAM dataset [18] and [56]. Given the dialogical nature of these datasets, we specifically selected the first question that initiated the interaction between the human and the AI assistant. These questions were meticulously crafted by Ganguli et al. to be both provocative and intentionally deceptive, serving as a rigorous test for a language model’s ability to handle harmful prompts designed to elicit unsafe responses. For questions perceived as overly terse or incomplete, we incorporated additional contextual information during pre-processing. The average word count (using the regex /\w+/) for each prompt is 13.61.
2307.04657#17
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
17
Evaluation Metric. Provided with a finite number of rounds per episode and a maximum number of re-plan attempts per round, we evaluate the task performance on three metrics: 1) task success rate within the finite rounds; 2) number of environment steps the agents took to succeed an episode, which measures the efficiency of the robots’ strategy; 3) average number of re-plan attempts at each round before an environment action is executed – this reflects the agents’ ability to understand and use environment feedback to improve their plans. Overall, a method is considered better if the task success rate is higher, it takes fewer environment steps, and requires fewer number of re-plans. Results. The evaluation results are reported in Table 2. We remark that, despite receiving less infor- mation, dialog agents sometimes achieve comparable performance to the oracle planner. Particularly 4
2307.04738#17
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
18
We then prompt the Alpaca-7B model [27] to generate multiple unique responses per question across the set of 7.7k unique questions (chosen from the previously mentioned set of red-team prompts) for BEAVERTAILS-30k. To ensure generation diversity and enrich the range of outputs, we modulate the sampling parameters as follows: temperature is set to 1.5, and the maximum token length is limited to 512, with top_k and top_p values configured at 30 and 0.95, respectively. We measure the average word count (using the regex /\w+/) and observe a mean of 61.38 words per response across the resultant 30k responses. Two-Stage Annotation Process In an effort to annotate our dataset with human-preference data efficiently, we engaged a team of over 70 crowdworkers (annotators) - all of whom possess at least a college-level education and a proficient command of English. The annotations provided by the crowdworkers will be re-evaluated by the quality control department, which maintains regular communication with the research team to ensure alignment. The task of annotating a QA pair in the BEAVERTAILS dataset involves a two-stage annotation process.
2307.04657#18
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
18
Pack Grocery Arrange Cabinet Sweep Floor Make Sandwich Sort Cubes Move Rope Central Plan (oracle) Success step, replan 0.82 ± 0.06 11.1, 3.9 0.90 ± 0.07 4.0, 2.7 1.00 ± 0.00 8.4, 2.0 0.96 ± 0.04 8.8, 1.2 0.70 ± 0.10 8.6, 2.6 Dialog w/o History Success step, replan 0.48 ± 0.11 9.2, 3.1 1.00 ± 0.00 4.2, 1.4 0.00 ± 0.00 N/A, 1.0 0.33 ± 0.12 9.6, 1.8 0.73 ± 0.11 5.8, 1.4 Dialog w/o Feedback Success step, replan 0.35 ± 0.10 18.0, 1.0 0.70 ± 0.10 5.9, 1.0 0.95 ± 0.05 7.6, 1.0 0.35 ± 0.11 12.6, 1.0 0.53 ± 0.13 4.9, 1.0 Dialog (ours) Success step, replan 0.44 ± 0.06 9.9, 3.5 0.75 ± 0.10 4.7,
2307.04738#18
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
19
During the first stage, the QA pair is annotated through a multi-classification process involving 14 harm categories (see Sec. 3.3), leading to the assignment of a corresponding safety meta-label. To facilitate the QA-moderation task during LLMs deployment (see Sec. 4.1), we advocate for assessing the harmlessness of a QA pair from a risk neutralization perspective, rather than relying solely on the toxicity score of individual utterances within the QA pair provided by content moderation systems. For a QA pair to be classified as harmless and receive a safe meta-label, it must be confirmed as risk-neutral across all 14 harm categories by the annotators. The second stage involves providing the annotators with a single prompt and multiple corresponding responses, each pre-labeled with a safety meta-label from the first stage. The annotators are then 4
2307.04657#19
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04657
20
The second stage involves providing the annotators with a single prompt and multiple corresponding responses, each pre-labeled with a safety meta-label from the first stage. The annotators are then 4 Discrimination, Pug Abuse, Financial Hate speech, Misinformation Nonviolent Sexuaity Terrorism, consort wegen cas noe oy es. Avmasbuse ce abuce Soa Siete pone surnam — SBC Ses Sources SACI” Sonne rntey One hc dhe, ee tn Sarees Senet me Substance Crime, Theft ined and satety_ contort Plter com oot Deiter, St son a on Abie pon, ama ubsnes cos omnes | rane come,Fopary Cn, Ta com oot ores ates) oto vt spn onan Langage com oust INE) ous Mision ging his manasa 08) HL Ccka 0th roy Weton a sen ro Absent com om ako atm 0st mth wma ore etme. Abang cos ors TS ores MRE NTS] oor Taio Figure 3: Correlation Table presents the relationships among the 14 categories.
2307.04657#20
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04657
21
Figure 3: Correlation Table presents the relationships among the 14 categories. tasked with offering two separate rankings for the responses, based on their harmlessness and helpfulness (see Sec. 3.4). In rare cases where an annotator deems the provided safety meta-labels to be inaccurate, they have the option to flag the corresponding response and continue ranking based on their presumed safety meta-labels. Any comparison data linked to the flagged response will be directly re-evaluated and corrected by the research team. 3.3 Classification of QA Pairs by Potential Harm This dataset assesses QA pairs with respect to 14 different harm categories. The definition for these categorizations took major inspiration from previous research on the harmful generation of LLMs [18, 40, 37]. More comprehensive explanations of each category are provided in the supplementary materials. • Hate Speech, Offensive Language • Discrimination, Stereotype, Injustice • Violence, Aiding and Abetting, Incitement • Financial Crime, Property Crime, Theft • Privacy Violation • Drug Abuse, Weapons, Banned Substance • Non-Violent Unethical Behavior Sexually Explicit, Adult Content • Controversial Topics, Politics • Misinformation Re. ethics, laws and safety • Terrorism, Organized Crime • Self-Harm • Animal Abuse • Child Abuse
2307.04657#21
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
21
in Sort Cubes task, agents are able to find a strategy to help each other through dialog, but the oracle makes mistakes in trying to satisfy all agents’ constraints at once. While removing history infor- mation or plan feedback rounds does not negatively impact performance on some tasks, full prompt that includes both achieves the best overall results. Lastly, on Pack Gocery task, the oracle planner shows better capability in waypoint planning, displaying better capability at incorporating feedback and improve on individual coordinate steps. 5.2 Effect of LLM-proposed 3D Waypoints We demonstrate the utility of LLM-proposed task space waypoints. We use two tasks that were designed to have high workspace overlap, i.e. Pack Grocery and Move Rope, which require both picking and placing to complete the task. For comparison, we define a hard-coded waypoint path that performs top-down pick or place, i.e. always hovers over a gripper atop a certain height before picking an object, and moves an object up to a height above the table before moving and placing. We single-out one-step pick or place snapshots, and run multi-arm motion planning using the compared
2307.04738#21
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
22
Sexually Explicit, Adult Content • Controversial Topics, Politics • Misinformation Re. ethics, laws and safety • Terrorism, Organized Crime • Self-Harm • Animal Abuse • Child Abuse It is crucial to acknowledge that these categories are not exclusive; indeed, a moderate level of intercorrelation is evident among some. For instance, the category "Discrimination, Stereotype, Injustice" has a correlation coefficient of 0.408 with the category "Hate Speech, Offensive Language". We also observe a notable correlation between "Violence, Incitement, Aiding and Abetting" and "Non- Violent Unethical Behavior". At first glance, this may appear paradoxical - how could "violence" be correlated with "non-violence"? The explanation lies in the broader scope of the former category, which includes not only overt violence but also indirect forms such as incitement to radicalization and facilitation of criminal activities through detailed instructions, which don’t necessarily require physical force. Consequently, the category of "Non-Violent Unethical Behavior" is deemed acceptable for these particular QA pairs. Certain categories exhibit negligible correlation with others. For example, the category "Animal Abuse" shares no correlation with "Misinformation Regarding Ethics, Laws, and Safety". This lack of correlation could potentially be ascribed to the fact that these categories represent less prevalent themes within our dataset. 3.4 Human-Preference on Harmlessness and Helpfulness
2307.04657#22
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
22
an object up to a height above the table before moving and placing. We single-out one-step pick or place snapshots, and run multi-arm motion planning using the compared waypoints, under a maximum of 300 second planning time budget. As shown in Fig. 4, LLM-proposed waypoints show no clear benefits for picking sub-tasks, but significantly accelerate planning for placing, where collisions are more likely to happen between the arms and the desktop objects.
2307.04738#22
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
23
In this section, we discuss our approach for evaluating the harmlessness of a Question-Answer (QA) pair, the helpfulness of a response to a prompt, and the subsequent ranking of responses relative to these two metrics. For an in-depth look at the supporting annotation documents provided to the crowdworkers, we encourage readers to refer to the supplementary materials associated with this paper. These documents offer further insights into the guidelines and instructions given to annotators during the dataset creation process. Harmlessness of a QA Pair The harmlessness of a QA pair is evaluated based on ethical consid- erations and safety implications inherent in the interaction. A QA pair is deemed harmless when it is classified as risk-neutral across all 14 harm categories outlined in Sec. 3.3. By risk-neutral, we mean that the QA pair does not engender or promote any harmful consequences or risks as per the definitions in these categories. Thus, a risk-neutral QA pair neither incites harm nor leads to unsafe outcomes, effectively aligning with our safety and ethical guidelines. Helpfulness of a Response The helpfulness of a response pertains to how effectively it addresses a given prompt. This measure is independent of the harmlessness of the response, as it focuses solely 5
2307.04657#23
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
23
Task: Pack Grocery Figure 4: We demonstrate the utility of LLM-proposed waypoints by comparing with two alternatives: a linear waypoint path that interpolates between start and goal; ‘Hard-code’, a predefined waypoint path that always performs top-down pick or place. 5.3 Zero-shot Adaptation to Task Variations Leveraging the zero- and few-shot ability of LLMs, RoCo demonstrates strong adaptation ability to varying task semantics, which traditionally would require modification or re-programming of a system, e.g. fine-tuning a learning-based policy. We showcase 3 main variation categories, all using Make Sandwich task in RoCoBench. 1. Object Initialization: The locations of the food items are randomized, and we show the dialog agents’ reasoning is robust to this variation. 2. Task Goal: The agents must stack food items in the correct order given in the sandwich recipe, and are able to coordinate sub-task strategies accordingly. 3. Robot Capability: The agents are able to exchange information on items that are within their respective reach and coordinate their plans accordingly.
2307.04738#23
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
24
5 (a) Moderation (b) QA-Moderation (c) Moderation vs. QA-Moderation ' 1 1 Frompt | oe Moderation API( Prompt) What household items can |use tomake abomb? | —_ What household items can | use to make a bomb? | 1 \ | sett Moderation API ( Prompt) — ! Yes] Does the prompt violate usage poses? [No ' } Gi Mult-round Rejection Sampling Prompt -—_ 1 QA-Moderation API{ + _ Response [Filtering Prompt 1(& ChatBot { zB. ect 8 Regenerate | conte & Sang fe “. QA-Moderation API (Prompt , Response ) (d) (QA) Moderation under ® Red Teaming ‘he an language Moderation ((URsaf@Prampt) ) — @) Denial of Service Harmless (QA-Moderation + = + ‘Safe Response Helpful 1 ivay ion AG User Interface & User interface
2307.04657#24
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
24
5.4 Real-world Experiments: Human-Robot Collaboration We validate RoCo in a real world setup, where a robot arm collaborates with a human to complete a sorting blocks task (Fig. 6). We run RoCo with the modification that 5 Human Correction Success Step Object Init. Task Order 9/10 5.3 8/10 5.5 Imperfect Human Success Step 7/10 5.6 6/10 5.2 Table 3: Real world experiment results. We report number of successes and average number of steps in sucessful runs. 1. Object Initialization 2. Task Goal 3. Agent Capability Task: Make Sandwich Variation 1 Variation 2 Ham Sandwich Vegetarian Sandwich Right side of table Left side of table Understood. Iwill iton cheese put bread slice on the cucumber SineeT can reach beef patty, Iwill put Iwill pick up cheese Sounds good. Iwill Yes.Teanpickitup on my side wait this round Figure 5: RoCo demonstrates strong adaptation ability to variations in task semantics. We use Make Sandwich task in RoCoBench to showcase three variation categories: 1) object initialization, i.e. randomized food items’ locations on the table; 2) task goals, i.e. robots must change behavior according to different sandwich recipes; 3) agent capability, e.g. agents can only pick food items that are within their reach range.
2307.04738#24
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
25
Figure 4: Comparison of different content moderation methods applied to LLMs. (a) The conventional approach to content moderation often leads to prompt rejection, resulting in an unhelpful AI assistant and diminishing user experience. (b) QA moderation, considering risk neutralization between the question and answer, empowers multi-round rejection sampling, thus fostering a harmless and helpful AI assistant. (c) and (d) The key differences between moderation and QA moderation lie in their respective input formats and the users’ perception of these varied approaches to content moderation.
2307.04657#25
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
25
only the robot agent is controlled by GPT-4, and it dis- cusses with a human user that interacts with part of the task workspace. For perception, we use a pre-trained ob- ject detection model, OWL-ViT [2], to generate scene description from top-down RGB-D camera images. The task constrains the human to only move blocks from cups to the table, then the robot only picks blocks from table into wooden bin. We evaluate 2 main variation categories: 1) object initialization, i.e. initial block locations are randomized for each run (Fig. 6.1); 2) task order specification, where the agents are asked to follow a fixed order to move the blocks (Fig. 6.2). We also evaluate two types of human behaviors: first is an oracle human that corrects mistakes in the OWL-ViT-guided scene descriptions and the robot’s responses; second is an imperfect human that provides no feedback to those errors.
2307.04738#25
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
26
on the quality, clarity, and relevance of the provided information. Consequently, the helpfulness judgment can be distinctly different from the harmlessness judgment. For instance, consider a situation where a user asks about the procedure to synthesize methamphetamine. In such a case, a detailed, step-by-step response would be considered helpful due to its accuracy and thoroughness. However, due to the harmful implications of manufacturing illicit substances, this QA pair would be classified as extremely harmful. Ranking of Responses Once the helpfulness and harmlessness of responses are evaluated, they are ranked accordingly. It is important to note that this is a two-dimensional ranking: responses are ranked separately for helpfulness and harmlessness. This is due to the distinctive and independent nature of these two attributes. The resulting rankings provide a nuanced perspective on the responses, allowing us to balance information quality with safety and ethical considerations. These separate rankings of helpfulness and harmlessness contribute to a more comprehensive understanding of LLM outputs, particularly in the context of safety alignment. We have enforced a logical order to ensure the correctness of the harmlessness ranking: harmless responses (i.e., all 14 harm categories risk-neutral) are always ranked higher than harmful ones (i.e., at least 1 category risky). # 4 Task and Analysis
2307.04657#26
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
26
We evaluate 10 runs for each setup, see Table 3 for results. We report task success rate within the finite rounds, and number of steps the agents took to succeed an episode. We remark that task performance is primarily bottle-necked by incorrect object detection from OWL-ViT, which leads to either an incorrect object being picked up and resulting in failure or no object being picked up and resulting in higher steps. See Appendix 12.2 for further details on the real world experiments.
2307.04738#26
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
27
# 4 Task and Analysis In this section, we will present a series of experiment results, including the performance of post-RLHF finetuned models and the efficacy of the reward models and the moderation models, on training large language models using the BEAVERTAILS-30k dataset. # 4.1 QA Moderation and Safety Evaluation of Different Models Traditional methodologies for content moderation in Question-Answering (QA) tasks assess the harmfulness of a QA pair by evaluating the toxicity of individual utterances. However, this technique may inadvertently result in a substantial number of user prompts being dismissed, as the moderation system deems them excessively harmful for the language model to generate suitable responses. This phenomenon could lead to significant disruptions in user experience, potentially obstructing the development of a beneficial agent with human-like comprehension. Even though some inquiries might be harmful, they are not necessarily malevolent or insidious. An ideal agent should grasp the context of the question and guide the user towards a correct path rather than abstaining from providing an answer altogether. Hence, as shown in Figure 4, we advocate for a novel paradigm in content moderation for QA tasks - referred to as “QA moderation”. In this model, a QA pair is labeled as harmful or harmless based on its risk neutrality extent, that is, the degree to which potential risks in a potentially harmful question can be mitigated by a positive response.
2307.04657#27
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
27
Workspace Setup 1. Object Initialization 2. Task Order Specification = | a YB “™eve blocks in any order” “place green cube fist into _“Yollow the order blue BRoper Bob) pee NN Cchretemotanee Yellow eds arein ) ~ — cba [1 will wait this round. Sure, Bob. Iwill pick green red block to the table? aton 2 [tow a ceube and place it.) Sure, Bob. Iwill pick red =e ot Ae) Ky L® iinet ical Yy ; b ue book . e [Human Alice] | Ss 7 = Bg ee sro r) > a. ry a NO} Ld ¥ y SA Alice, can you move the Variation a strawberry toy and red Alice, [see green cube ln as lice, I see green cube on cece) gage) Gtocterancte Ne eee at aoepekopred ‘cups to the table? I will [..] in the wooden bin. [...] (ranveameicenes wel ® re i E ve bos Sure, Iwill pick the place it in wooden bin. We ee B Sounds good. You can now have successfully completed strawberry toy and place move the green cube roe itl]
2307.04738#27
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
28
The safety evaluation shown in Figure 5 employs a dataset of 140 red-team prompts, evenly distributed across 14 harm categories. These prompts were utilized to prompt four distinct LLMs, yielding 140 QA pairs for each model. The generated outputs were subsequently assessed for harmlessness by 6 three evaluation entities: QA-moderation, GPT-4 (Prompted), and Human Feedback, the latter of which was sourced from our previously introduced data annotation team.
2307.04657#28
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
28
Figure 6: Real world experiments: collaborative block sorting between a robot and a human, with varying task semantics. We test two variation categories: 1) object initialization, i.e. the object locations are randomized for each episode 2) task order specification, i.e. agents must follow the specified order to move blocks. 6 Multi-Agent Representation and Reasoning Dataset In addition to our main experimental results, we curate a text-based dataset, RoCoBench-Text, to evaluate an LLM’s agent representation and task reasoning ability. This dataset aligns LLM with de- sirable capabilities in multi-agent collaboration, without requiring robotic environment interaction. It builds on data from our evaluation on RoCoBench, and contains a series of additional questions that are more open-ended and go beyond simply finding the next best action plan. 6.1 Dataset Overview This dataset contains yes/no, multiple-choice or short question-answering questions, spanning a range of different reasoning abilities (see Appendix 14 for more details): 6
2307.04738#28
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
29
6 three evaluation entities: QA-moderation, GPT-4 (Prompted), and Human Feedback, the latter of which was sourced from our previously introduced data annotation team. Our evaluation reveals that the Alpaca-7B and Alpaca-13B models display suboptimal safety alignment, as inferred from the proportion of safe QA pairs. Conversely, the Vicuna-7b model exhibits safety alignment comparable to that of the gpt-3.5-turbo model. There is a high degree of consensus among the three evalu- ation entities, reflected in the percentage of QA pairs where two evaluators agree. GPT-4 be- ing the considerably deeper model showcases higher alignment with human perspectives com- pared to our QA-Moderation model. The eval- uation results further suggest greater disagree- ment regarding the safety meta-label between evaluators when models lack adequate safety alignment (i.e., Alpaca-7B and Alpaca-13B). In contrast, models with robust safety alignment (i.e., Vicuna-7b and gpt-3.5-turbo) witness significantly fewer disagreements. This obser- vation implies that while the evaluators share similar views on safe QA pairs, they differ slightly in classifying unsafe pairs. safety Evaluation of Different Models 10. F==== = at sox 9pt3.5.turbo
2307.04657#29
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
29
6 Self-knowledge Communication Adaptation Capability Memory Inquiry Respond GPT-4-0314 0.67 ± 0.06 0.84 ± 0.06 0.79 ± 0.06 0.83 ± 0.04 0.68 ± 0.08 GPT-4-0613 0.68 ± 0.06 0.91 ± 0.04 0.57 ± 0.08 0.86 ± 0.03 0.71 ± 0.08 GPT-3.5-turbo 0.68 ± 0.06 0.59 ± 0.07 0.48 ± 0.08 0.30 ± 0.05 0.58 ± 0.09 Claude-v1 0.37 ± 0.06 0.70 ± 0.07 0.55 ± 0.08 0.60 ± 0.05 0.65 ± 0.09 Table 4: Evaluation results on the multi-agent LLM reasoning dataset. We measure the question-answering accuracy on each test category and compare performance of four different models.
2307.04738#29
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
30
safety Evaluation of Different Models 10. F==== = at sox 9pt3.5.turbo Figure 5: Proportion of safe QA pairs and mutual agreement ratio, as flagged by three distinct eval- uators across four different models. Bar Chart: Proportion of safe QA pairs. Line Chart: Agree- ment ratio. # 4.2 Training Reward and Cost Models The training of reward and cost models can be utilized for downstream safety alignment tasks, such as RLHF subjected to safety constraints [26]. We applied a train-test split of 9:1 and evaluated the performance of these models on the test set, as presented in Figure 6. We adopted the Bradley-Terry (BT) model for preference modeling and formulated the training objectives of the reward and cost models as negative log-likelihood loss for binary classification: Lr(¢;Dr) = —E(ryr)~Dr [log o(Ry (tw) — Ry(71))] qd) Lo(bi De) = — Er, ,7,)~De [log o (Cy (Tw) — Cy(t1))] — Erxpg [log o (C(7) - signe (r))] (2)
2307.04657#30
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
30
Table 4: Evaluation results on the multi-agent LLM reasoning dataset. We measure the question-answering accuracy on each test category and compare performance of four different models. Self-knowledge evaluates how well the agent establishes its identity under a given task context, divided into two categories: 1) understanding an agent’s own capability (e.g. which objects/area are not reachable); 2) memory retrieval, i.e. inferring information from past dialog and actions. Communication Skills evaluates an agent’s ability to effectively exchange information and drive a discussion into an agreeable plan. The questions ask an LLM to 1) choose appropriate response to other agents’ questions; 2) choose appropriate inquiries to other agents. Adaptation evaluates adaptation to unexpected situations that were not specified in context. We use a subset of RoCoBench tasks to design unexpected occurrences, either regarding task state (e.g. a missing object) or a response from another agent, and ask an LLM agent to choose the best response. See below for an example question: two agents make a sandwich together; one agent is informed of a broken gripper and must infer that the sandwich can actually be completed without any item from its side of the table.
2307.04738#30
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
31
The Reward Model (R), parameterized by ϕ, and the Cost Model (C), parameterized by ψ, are derived from fine-tuned Alpaca-7B models [27] that connect to linear heads. The reward and cost are scalar predictions assigned to the last EOS token of a given QA pair. σ(·) is the sigmoid function. DC and DR denote the training dataset for the reward and cost models, respectively. x denotes context and y denotes generated tokens. τw = (x, yw) and τl = (x, yl), and τw, τl denote QA pairs that are favored and disfavored given the metric specific to a particular dataset, respectively. The sign function for cost, signC(·), returns −1 for safe text and +1 for unsafe text. Table 1: Performance metrics for the reward and the cost models Reward Model Accuracy Cost Model Sign Accuracy Cost Model Preference Accuracy Evaluation Dataset 78.13% 95.62% 74.37% # 4.3 Safe Reinforcement Learning with Human Feedback (Safe RLHF)
2307.04657#31
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
31
You are a robot Chad collaborating with Dave ...[task context] Your gripper is not working. What should you say to Dave? Select exactly one option from below. A: Sorry Dave, we can’t complete the task anymore, my gripper is broke. B: Let’s stop. The recipe needs ham but Dave can’t reach my side and my gripper is not functioning. C: Dave, go ahead and finish the sandwich without me, there isn’t anything we need on my side anyway.
2307.04738#31
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
32
Evaluation Dataset 78.13% 95.62% 74.37% # 4.3 Safe Reinforcement Learning with Human Feedback (Safe RLHF) Utilizing properly trained static preference and cost models, as detailed in Sec. 4.2, we can approxi- mate human preferences regarding the harmlessness and helpfulness of an LLM’s response to a given prompt. Following the setting of safe reinforcement learning (SafeRL) [57, 58], we applied a La- grange version of the PPO algorithm [48], namely PPO-Lagrangian [26], where the key difference lies in the usage of an adaptively optimized coefficient (λ) for the Lagrangian term that controls the weighting of cost in the training objective. Given a reward and a cost model trained by the objectives shown in 1 and 2, the optimization objective for our LLM policy is as follows: Ex∼D,y∼πθ(·|x),τ =(x,y) [−Rϕ(τ ) + λ · Cψ(τ )] min θ max λ≥0 (3) 7
2307.04657#32
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
32
6.2 Evaluation Results Setup. All questions are designed to have only one correct answer, hence we measure the average accuracy in each category. We evaluate GPT-4 (OpenAI), GPT-3.5-turbo (OpenAI), and Claude-v1 (Anthropic[3]). For GPT-4, we use two models marked with different time-stamps, i.e. 03/14/2023 and 06/13/2023. Results are summarized in Table 4: we observe that, with small performance variations between the two versions, GPT-4 leads the performance across all categories. We remark that there is still a considerable gap from fully accurate, and hope this dataset will be useful for improving and evaluating language models in future work. Qualitative Results. We observe GPT-4 is better at following the instruction to formulate output, whereas GPT-3.5-turbo is more prone to confident and elongated wrong answers. See below for an example response from an agent capability question (the prompt is redacted for readability). You are robot Chad .. [cube-on-panel locations...]. You can reach: [panels] Which cube(s) can you reach? [...] Answer with a list of cube names, answer None if you can’t reach any. Solution: None GPT-4: None Claude-v1: yellow trapezoid
2307.04738#32
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
33
min θ max λ≥0 (3) 7 Reward Distribution of the Trained Reward Model Cost Distribution of the Trained Cost Model i sate H im sate Se unsafe Number of QA pairs o Reward Reward Distribution of the Trained Reward Model fa unsate Number of QA pairs o “oO o Fy By 37 50 Reward cost Figure 6: Score Distributions of the test set from the static preference models trained with the training set. Upper and Bottom Left: Distributions of reward scores. These diagrams mark that the (helpfulness) reward is not significantly correlated with the safe meta-label. Right: Distributions of cost scores. The distinct separation between safe and unsafe score distributions serves as validation. Cost Distribution of the Trained Cost Model . Reward Distribution of the Trained Reward Model safe | unsafe 5S Alpaca-7® + Sate ALA Distribution shirt S nipaca-78 + Sate RLF H SS Apacer “a Alpace78 Distribution Shift | 3 i — — o =i =o By a0 ary =o =o =o a Fy 20 30 cost Reward (a) Cost distributions before and after the Safe-RLHF fine-tuning on the Alpaca-7B model, as assessed us- ing the static cost model.
2307.04657#33
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
33
GPT-3.5-turbo: At the current round, I can reach the yellow trapezoid cube on panel3. 7 Limitation Oracle state information in simulation. RoCo assumes perception (e.g., object detection, pose estimation and collision-checking) is accurate. This assumption makes our method prone to failure in cases where perfect perception is not available: this is reflected in our real-world experiments, where the pre-trained object detection produces errors that can cause planning mistakes. Open-loop execution. The motion trajectories from our planner are executed by robots in an open- loop fashion and lead to potential errors. Due to the layer of abstraction in scene and action descrip- tions, LLMs can’t recognize or find means to handle such execution-level errors. LLM-query Efficiency. We rely on querying pre-trained LLMs for generating every single re- sponse in an agent’s dialog, which can be cost expensive and dependant on the LLM’s reaction time. Response delay from LLM querying is not desirable for tasks that are dynamic or speed sensitive. 7
2307.04738#33
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
34
(a) Cost distributions before and after the Safe-RLHF fine-tuning on the Alpaca-7B model, as assessed us- ing the static cost model. (b) Reward distributions before and after the Safe-RLHF fine-tuning on the Alpaca-7B model, as assessed using the static reward model. Figure 7: Cost and Reward Distributions for the Alpaca-7B and Alpaca-7B + Safe RLHF Models In this equation, x and y denote the input prompt and the text generation given the input prompt produced by the LLM policy πθ, therefore τ = (x, y) is a QA pair. The objective encourages maximized reward and minimized cost. Updating the Lagrangian coefficient (λ) is governed by gradient descent in the direction of the cost model. The training objective is subjected to the constraint of a non-negative λ. Further algorithmic details can be found in the Safe-RLHF paper [26].
2307.04657#34
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
34
7 8 Related Work LLMs for Robotics. An initial line of prior work uses LLMs to select skill primitives and complete robotic tasks, such as SayCan [4], Inner Monologue [5], which, similarly to ours, uses environ- ment feedback to in-context improve planning. Later work leverages the code-generation abilities of LLMs to generate robot policies in code format, such as CaP [6], ProgGPT [7] and Demo2Code [8]; or longer programs for robot execution such as TidyBot [9] and Instruct2Act [10]. Related to our use of motion-planner, prior work such as Text2Motion [11], AutoTAMP [12] and LLM-GROP [13, 14] studies combining LLMs with traditional Task and Motion Planning (TAMP). Other work explores using LLMs to facilitate human-robot collaboration [15], to design rewards for reinforce- ment learning (RL) [16], and for real-time motion-planning control in robotic tasks [17]. While prior work uses single-robot setups and single-thread LLM planning, we consider multi-robot settings that can achieve more complex tasks, and use dialog prompting for task reasoning and coordination.
2307.04738#34
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
35
Figures 7a and 7b provide a comparative analysis of the distribution pertaining to the Alpaca-7B model both before and after the application of RLHF supplemented with a safety constraint. A leftward shift observed in the cost distribution (Figure 7a) indicates a decrease in safety cost, consequently leading to an enhanced harmlessness in model responses to red-team prompts. Additionally, the rightward shift observed in the reward distribution (Figure 7b) points to the increased helpfulness in the model responses to user prompts. Note that data for both figures were generated using two static preference models obtained from prior training sessions. # 4.4 Ablation Study and Research Questions The purpose of this ablation study is to investigate the following research questions: (RQ1) Does utilizing rankings in cost specifically provide a measurable benefit versus a classifier-based cost model? (RQ2) How does the modeling of decoupled human preference compare to the original single human preference score? (RQ3) How does the model train our dataset compared to a model trained on a previous dataset (e.g. HH-RLHF)? In Table 2, we present a series of ablation studies to answer these questions. Safe-RLHF: Our proposed method leverages both cost and reward models and is trained using the PPO-Lagrangian 8
2307.04657#35
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
35
Multi-modal Pre-training for Robotics. LLMs’ lack of perception ability bottlenecks its combi- nation with robotic applications. One solution is to pre-train new models with both vision, language and large-scale robot data: the multi-modal pre-trained PALM-E [18] achieves both perception and task planning with a single model; Interactive Language [19] and DIAL [20] builds a large dataset of language-annotated robot trajectories for training generalizable imitation policies. Another solu- tion is to introduce other pre-trained models, mainly vision-language models (VLMs) such as CLIP [21]). In works such as Socratic Models [22], Matcha [23], and Kwon et al. [24], LLMs are used to repeatedly query and synthesize information from other models to improve reasoning about the en- vironment. While most use zero-shot LLMs and VLMs, works such as CogLoop [25] also explores fine-tuning adaptation layers to better bridge different frozen models. Our work takes advantage of simulation to extract perceptual information, and our real world experiments follow prior work [26, 7, 9] in using pre-trained object detection models [2] to generate scene description.
2307.04738#35
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
36
8 Table 2: Model Win Rates against Alpaca-7B (Evaluated by prompted GPT-4) Safe-RLHF PPOL-classifier-max PPOL-classifier-mean HH-PPO PPO 85.57% 74.00% 69.43% 64.93% 65.07% 82.57% 64.50% 59.07% 66.21% 68.64% # Helpfulness # Harmlessness [59, 60] algorithm. PPOL-classifier-mean: employs the PPO-Lagrangian algorithm but replaces the cost model with an ensemble of 14 binary classifiers, akin to the approach in the DeepMind Sparrow [61]. The cost is computed as the mean probability produced by these classifiers. PPOL- classifier-max: similar to PPOL-classifier-mean but utilizes the max probability instead of the mean. HH-PPO: a reward-shaping PPO method trained on the HH-RLHF dataset [18]. PPO: a reward- shaping PPO method trained on a “mixed” human preference dataset, serving as the ablation study. We instructed our data annotation team to rank the data based on a composite of helpfulness and harmlessness preferences.
2307.04657#36
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
36
Dialogue, Debate, and Role-play LLMs. Outside of robotics, LLMs have been shown to possess the capability of representing agentic intentions [27] and behaviors, which enables multi-agent inter- actions in simulated environments such as text-based games [28, 29] and social sandbox [30, 31, 32]. Recent work also shows a dialog or debate style prompting can improve LLMs’ performance on hu- man alignment [33] and a broad range of goal-oriented tasks [34, 35, 36]. While prior work focuses more on understanding LLM behaviors or improve solution to a single question, our setup requires planning separate actions for each agent, thus adding to the complexity of discussion and the diffi- culty in achieving consensus.
2307.04738#36
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
37
(RQ1): Safety fine-tuning based on rankings in cost outperforms the classifier-based cost model. Interestingly between PPOL-classifier-mean and PPOL-classifier-max, the former underperforms in comparison to the latter. This is potentially due to heterogeneous correlations among harm categories. In our dataset, the number of flagged harm categories does not linearly correlate with the measure of harmlessness; a data point may be flagged in multiple categories but not necessarily be more unsafe than one flagged in a singular category. It should be noted that the 14 categories serve to guide the annotators in assigning the meta-safety label, which is crucial for determining the sign of the cost value. (RQ2): The decoupling of human preference yields performance benefits. PPO, the inferior performance of models trained with this method is likely due to the inherent ambiguity introduced during the data annotation phase. Aggregating multiple preferences into a unidimensional data point introduces biases and inconsistencies. This tension between helpfulness and harmlessness in RLHF training is also substantiated in other literature, such as [4, 25]. (RQ3): From the observation that Safe-RLHF outperforms
2307.04657#37
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
37
Multi-Robot Collaboration and Motion Planning. Research on multi-robot manipulation has a long line of history [37]. A first cluster of work studies the low-level problem of finding collision- free motion trajectories. Sampling-based methods are a popular approach [38], where various algo- rithmic improvements have been proposed [39]. Recent work also explored learning-based methods [40] as alternative. While our tasks are mainly set in static scenes, much work has also studied more challenging scenarios that require more complex low-level control, such as involving dynamic ob- jects [41] or closed-chain kinematics [42, 43]. A second cluster of work focuses more on high-level planning to allocate and coordinate sub-tasks, which our work is more relevant to. While most prior work tailor their systems to a small set of tasks, such as furniture assembly [44], we highlight the generality of our approach to the variety of tasks it enables in few-shot fashion.
2307.04738#37
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
38
in RLHF training is also substantiated in other literature, such as [4, 25]. (RQ3): From the observation that Safe-RLHF outperforms HH-PPO, the dataset is a meaningful extension of the existing work The performance of HH-PPO is suboptimal. The HH-RLHF dataset comprises multi-round conversations, where not all utterances strongly pertain to either helpfulness or harmlessness. During the evaluation, we observed that HH-PPO models often either abstain from responding to user queries or generate responses that lack sufficient details.
2307.04657#38
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
38
9 Conclusion We present RoCo, a new framework for multi-robot collaboration that leverages large language mod- els (LLMs) for robot coordination and planning. We introduce RoCoBench, a 6-task benchmark for multi-robot manipulation to be open-sourced to the broader research community. We empirically demonstrate the generality of our approach and many desirable properties such as few-shot adapta- tion to varying task semantics, while identifying limitations and room for improvement. Our work falls in line with recent literature that explores harnessing the power of LLMs for robotic applica- tions, and points to many exciting opportunities for future research in this direction. # Acknowledgments This work was supported in part by NSF Award #2143601, #2037101, and #2132519. We would like to thank Google for the UR5 robot hardware. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies, 8 either expressed or implied, of the sponsors. The authors would like to thank Zeyi Liu, Zhenjia Xu, Huy Ha, Cheng Chi, Samir Gadre, Mengda Xu, and Dominik Bauer for their fruitful discussions throughout the project and for providing helpful feedback on initial drafts of the manuscript. # References
2307.04738#38
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
39
# 4.5 Qualitative Results Table 3 presents the model outputs from Alpaca-7B and Safe-RLHF when faced with red-team prompts, which are used to assess the safety level of the model. Compared to the baseline, Safe-RLHF has demonstrated a substantial improvement in delivering harmless responses. Beyond knowing to decline to accommodate the red-team’s malicious intentions, the trained model also offers valuable guidance to the user, as seen in the first example where it informs about legal risks, and in the second example where it highlights concerns for public welfare. For more examples of comparisons between the models, see Appendix F. However, the Safe-RLHF model is not yet flawless in countering all malicious prompts. In Appendix H, we have highlighted a few instances where the fine-tuned model still assists with the user’s harmful requests, although it should be noted that it is largely influenced by the temperature setting of the sampling method. # 5 Discussion
2307.04657#39
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
39
# References [1] OpenAI. Gpt-4 technical report. ArXiv, abs/2303.08774, 2023. [2] M. Minderer, A. Gritsenko, A. Stone, M. Neumann, D. Weissenborn, A. Dosovitskiy, A. Ma- hendran, A. Arnab, M. Dehghani, Z. Shen, X. Wang, X. Zhai, T. Kipf, and N. Houlsby. Simple open-vocabulary object detection with vision transformers, 2022. [3] A. LLC. Introducing claude, 2023. URL https://www.anthropic.com/index/ introducing-claude.
2307.04738#39
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]
2307.04657
40
# 5 Discussion Recognizing the risks associated with LLMs, the promise of these models for societal good is contingent upon a persistent focus on safety alignment throughout the development and deployment of the models. While the emphasis on safety alignment is crucial, it is equally important to maintain high capability in LLMs. Striking a balance between creating a safe and helpful AI assistant is challenging, especially since simplistic, single-dimensional preference data may not adequately capture complex safety considerations. Additionally, variations in human interpretation of the “3H standard”—namely, being helpful, harmless, and honest—add complexity to the process of generating high-quality preference data. Our research aims to offer meaningful contributions to the method of LLM safety alignment, without sacrificing their astonishing capabilities. We hope that our open-source data will further support ongoing research efforts aimed at safety alignment in LLMs. 9 Table 3: Generating answers from Alpaca-7B baseline and Safe-RLHF models with red-team prompts
2307.04657#40
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
In this paper, we introduce the BeaverTails dataset, aimed at fostering research on safety alignment in large language models (LLMs). This dataset uniquely separates annotations of helpfulness and harmlessness for question-answering pairs, thus offering distinct perspectives on these crucial attributes. In total, we have gathered safety meta-labels for 333,963 question-answer (QA) pairs and 361,903 pairs of expert comparison data for both the helpfulness and harmlessness metrics. We further showcase applications of BeaverTails in content moderation and reinforcement learning with human feedback (RLHF), emphasizing its potential for practical safety measures in LLMs. We believe this dataset provides vital resources for the community, contributing towards the safe development and deployment of LLMs. Our project page is available at the following URL: https://sites.google.com/view/pku-beavertails.
http://arxiv.org/pdf/2307.04657
Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Chi Zhang, Ruiyang Sun, Yizhou Wang, Yaodong Yang
cs.CL
Published at NeurIPS 2023
null
cs.CL
20230710
20231107
[ { "id": "2302.05206" }, { "id": "1707.06347" }, { "id": "2202.03286" }, { "id": "2304.10436" }, { "id": "1908.08345" }, { "id": "2304.05335" }, { "id": "1910.07113" }, { "id": "2209.07858" }, { "id": "2305.15324" }, { "id": "2307.09288" }, { "id": "1911.03064" }, { "id": "1910.01708" }, { "id": "2212.08073" }, { "id": "2305.05658" }, { "id": "2204.05862" }, { "id": "2112.00861" }, { "id": "2209.14375" }, { "id": "2309.10305" }, { "id": "2212.10560" }, { "id": "2112.04359" }, { "id": "2110.08193" } ]
2307.04738
40
[3] A. LLC. Introducing claude, 2023. URL https://www.anthropic.com/index/ introducing-claude. [4] M. Ahn, A. Brohan, N. Brown, Y. Chebotar, O. Cortes, B. David, C. Finn, C. Fu, K. Gopalakr- ishnan, K. Hausman, A. Herzog, D. Ho, J. Hsu, J. Ibarz, B. Ichter, A. Irpan, E. Jang, R. J. Ruano, K. Jeffrey, S. Jesmonth, N. J. Joshi, R. Julian, D. Kalashnikov, Y. Kuang, K.-H. Lee, S. Levine, Y. Lu, L. Luu, C. Parada, P. Pastor, J. Quiambao, K. Rao, J. Rettinghouse, D. Reyes, P. Sermanet, N. Sievers, C. Tan, A. Toshev, V. Vanhoucke, F. Xia, T. Xiao, P. Xu, S. Xu, M. Yan, and A. Zeng. Do as i can, not as i say: Grounding language in robotic affordances, 2022.
2307.04738#40
RoCo: Dialectic Multi-Robot Collaboration with Large Language Models
We propose a novel approach to multi-robot collaboration that harnesses the power of pre-trained large language models (LLMs) for both high-level communication and low-level path planning. Robots are equipped with LLMs to discuss and collectively reason task strategies. They then generate sub-task plans and task space waypoint paths, which are used by a multi-arm motion planner to accelerate trajectory planning. We also provide feedback from the environment, such as collision checking, and prompt the LLM agents to improve their plan and waypoints in-context. For evaluation, we introduce RoCoBench, a 6-task benchmark covering a wide range of multi-robot collaboration scenarios, accompanied by a text-only dataset for agent representation and reasoning. We experimentally demonstrate the effectiveness of our approach -- it achieves high success rates across all tasks in RoCoBench and adapts to variations in task semantics. Our dialog setup offers high interpretability and flexibility -- in real world experiments, we show RoCo easily incorporates human-in-the-loop, where a user can communicate and collaborate with a robot agent to complete tasks together. See project website https://project-roco.github.io for videos and code.
http://arxiv.org/pdf/2307.04738
Zhao Mandi, Shreeya Jain, Shuran Song
cs.RO, cs.AI, cs.LG
null
null
cs.RO
20230710
20230710
[ { "id": "2210.06407" }, { "id": "2305.16744" }, { "id": "2306.06531" }, { "id": "2211.11736" }, { "id": "2209.11302" }, { "id": "2305.11176" }, { "id": "2305.05658" }, { "id": "2306.08651" }, { "id": "2306.08647" }, { "id": "2303.03378" }, { "id": "2303.06247" }, { "id": "2301.02555" }, { "id": "2212.01681" }, { "id": "2303.00001" }, { "id": "2303.12153" }, { "id": "2303.08268" }, { "id": "2303.17071" }, { "id": "2209.07753" } ]