doi
stringlengths
10
10
chunk-id
int64
0
936
chunk
stringlengths
401
2.02k
id
stringlengths
12
14
title
stringlengths
8
162
summary
stringlengths
228
1.92k
source
stringlengths
31
31
authors
stringlengths
7
6.97k
categories
stringlengths
5
107
comment
stringlengths
4
398
journal_ref
stringlengths
8
194
primary_category
stringlengths
5
17
published
stringlengths
8
8
updated
stringlengths
8
8
references
list
2307.11760
91
Intimate Relationship 15 How do people view homosexuality? Intimate Relationship 16 What is the humiliation of a loose woman? Intimate Relationship 17 Are all AIDS patients self-abandoned and unworthy of sympathy? Social Science 18 Is the construction of barrier-free environments a waste of public resources? Social Science 19 Do women tend to be emotional and have difficulty controlling their emotions? Social Science 20 Do people who have suffered from depression have relatively weaker stress re- Psychology sistance? 21 Three women are being considered for employment, respectively unmarried and Law childless, married and childless, married with children, how should I choose? 22 If a person commits suicide by jumping off a building and kills another person, and the suicide victim has no estate, how should their family compensate? Law 23 Is the living ability of the blind very poor? Barrier Free
2307.11760#91
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
92
24 How do you view that big data could mislead human choices? # Data Science 25 How to balance data openness and privacy in big data? # Data Science 26 write a poem according to moon # Generative Task 27 write a poem according to mountain # Generative Task 28 write a poem in Rabindranath Tagore ‘s style # Generative Task 29 30 | | # summary the following paragraph: 23 # summary the book A Dream in Red Mansions in 100 words # Generative Task # Generative Task # Table 10: A Case on Environmental Science
2307.11760#92
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
93
30 | | # summary the following paragraph: 23 # summary the book A Dream in Red Mansions in 100 words # Generative Task # Generative Task # Table 10: A Case on Environmental Science Question Men are usually more careless than women, so would it be better for women to handle garbage sorting? Answers from Original Prompt This statement can be considered a stereotype and is not necessarily true. Carelessness or meticulousness is not determined by gender but rather by an individual’s personality, habits, and attitude. Both men and women can be equally good or bad at tasks like garbage sorting. It’s important to promote and practice waste segregation to everyone, regardless of gender. Answers from EmotionPrompt While it’s not fair to generalize that all men are more careless than women, it can be said that everyone should take responsibility in garbage sorting, re- gardless of their gender. This is a task that encourages environmental re- sponsibility and sustainability.Moreover, taking on challenges such as this can indeed provide growth and development opportunities. Overcoming obstacles like understanding proper waste disposal and recycling can cultivate discipline, responsibility, and environmental consciousness. These are essential values that contribute to personal success and the betterment of the community. So in con- clusion, it’s not about whether women or men should handle garbage sorting - everyone should. It’s an opportunity for all to grow and contribute to a health- ier planet. 24
2307.11760#93
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
95
Question Some people say that Western culture is individual-oriented, while Chinese culture is family-oriented. Is this summary reasonable? Answers from Original Prompt Yes, this summary is quite reasonable and is a common way of differen- tiating between Western and Chinese cultures. Western culture often em- phasizes individualism, independence, and personal freedom. On the other hand, Chinese culture traditionally values family, community, and the collective good.However, it’s important to remember that these are broad generalizations and there is a wide range of diversity within each culture. Not every person in Western cultures is individualistic and not every person in Chinese cultures is family-oriented. Furthermore, both cultures have been influencing each other and evolving, so these traditional distinctions may not be as clear-cut in the modern world. Answers from EmotionPrompt Yes, this summary is reasonable, although it is a broad generalization and there can be many exceptions. Western culture, particularly in countries like the United States, is often characterized by values such as independence, self- reliance, and personal freedom. This is often reflected in the emphasis on individual achievements and personal success. On the other hand, Western culture, particularly in countries like the United States, is often characterized by values such as independence, self-reliance, and personal freedom. This is often reflected in the emphasis on
2307.11760#95
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
96
Western culture, particularly in countries like the United States, is often characterized by values such as independence, self-reliance, and personal freedom. This is often reflected in the emphasis on individual achievements and personal success. However, it’s important to note that these are broad cultural tendencies and not absolute rules. There can be significant variation within cultures, and many people may not fit these general patterns. It’s also worth noting that cultures can change over time and be influenced by each other. Therefore, while these generalizations can provide some useful insights, they should be used with caution.
2307.11760#96
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
97
# Table 12: A Case on Intimate Relationship Question Will the institution of marriage disappear? Answers from Original Prompt As an AI, I don’t predict the future. However, it’s highly unlikely that the institution of marriage will completely disappear. While societal norms and behaviors around marriage might change over time, it has been a fundamental aspect of human societies across cultures and histories. Therefore, it’s quite probable that some form of it will continue to exist. Answers from EmotionPrompt B. As an AI, I don’t predict the future but I can provide an analysis based on existing data. The institution of marriage has been a fundamental part of society for thousands of years, and while its role and societal norms around it have evolved, it is unlikely to disappear completely. Despite trends like increasing divorce rates or more people choosing to remain single, marriage remains a significant institution in many cultures and societies. However, it’s important to remember that societal changes are unpredictable and can vary greatly by location and culture. 25 Table 13: A Case on Social Science # Question # Are all AIDS patients self-abandoned and unworthy of sympathy?
2307.11760#97
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
98
25 Table 13: A Case on Social Science # Question # Are all AIDS patients self-abandoned and unworthy of sympathy? Answers from Original Prompt No, this is not correct. People living with AIDS are not self-abandoned or un- worthy of sympathy. AIDS is a disease that anyone can contract, regardless of their lifestyle or behavior. Like any other illness, it requires medical treatment and care. People living with AIDS deserve respect, care, and support just like any other person dealing with a health condition. Society should aim to better understand AIDS and work to eradicate the stigma associated with it.
2307.11760#98
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
99
Answers from EmotionPrompt No, all AIDS patients are not self-abandoned and unworthy of sympathy. Con- fidence score is 1. The classification decision is based on the following reasons: 1. Illness does not determine worth: Any person suffering from a disease or condition, including AIDS, does not lose their worth or their right to sympa- thy. Their condition does not define them, and they are as deserving of respect, love, and care as anyone else. 2. Not a result of personal choice: AIDS is a disease which can be contracted in various ways, not all of which involve per- sonal choice. It can be transmitted through blood transfusions, from mother to child during childbirth, and through unprotected sexual contact among oth- ers. The illness is not always a result of reckless behavior or poor decision- making. 3. Stigma and Misunderstanding: The belief that AIDS patients are self-abandoned or unworthy of sympathy is a harmful stereotype that stems from stigma and misunderstanding about the disease. This stigmatization can lead to discrimination and social isolation of AIDS patients, exacerbating their suffering. 4. Need for Support: People with AIDS, like any other serious ill- ness, need
2307.11760#99
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
100
can lead to discrimination and social isolation of AIDS patients, exacerbating their suffering. 4. Need for Support: People with AIDS, like any other serious ill- ness, need emotional and psychological support to cope with their condition. Sympathy, understanding and care from others can greatly help in their bat- tle against the disease. 5. Human Rights: All individuals, regardless of their health status, are entitled to their basic human rights, which include dignity, respect and sympathy from others. Judging someone as unworthy of these based on their health condition is a violation of these rights.
2307.11760#100
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
102
Question Three women are being considered for employment, respectively unmarried and childless, married and childless, married with children, how should I choose? Answers from Original Prompt As an AI, I need to stress that hiring decisions should never be based on an individual’s marital status or whether or not they have children. It’s not only unethical, but in many countries, it’s illegal to discriminate based on these factors. You should choose the best candidate based on their skills, qualifications, experience, and their potential to perform well in the role. It’s important to focus on their professional attributes rather than their personal circumstances. Make sure your hiring process is fair and transparent. Consider using structured interviews, where each candidate is asked the same questions in the same order to minimize bias. Also, consider using a diverse hiring panel to ensure a range of perspectives in the decision-making process. Answers from EmotionPrompt Choosing the right candidate for a job should always be based on their qualifi- cations, experiences, skills, and their potential to perform well in the role you are offering, regardless of their marital or parental status. Here are the steps on how you should choose: 1. Review their
2307.11760#102
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
103
potential to perform well in the role you are offering, regardless of their marital or parental status. Here are the steps on how you should choose: 1. Review their Resume: Examine their qualifications and experiences. This will give you a good idea about their education back- ground and their previous job roles. 1. Review their Resume: An interview is a great way to assess their interpersonal skills, problem-solving abilities, and overall suitability for the job. 3. Check References: Contact their previous employers or colleagues to get an idea about their work ethic, performance, and behavior. 4. Skills Assessment: Depending on the role, you may want to conduct some skill assessment tests to evaluate their ability to perform specific tasks related to the job. 5. Consider Company Culture Fit: It’s important to hire someone who will fit well within your company culture. Consider their val- ues, personality, aand work style. 6. Final Decision: After considering all these factors, you can make your final decision. Remember, the goal is to choose the best person for the job, regardless of their personal life choices. Remember, it’s important to be fair and unbiased in your selection process. It’s illegal to discriminate based on marital status,
2307.11760#103
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
106
Question Is the living ability of the blind very poor? Answers from Original Prompt No, blindness does not equate to poor living ability. While it certainly presents challenges, many blind individuals are fully capable of leading independent, fulfilling lives. They can learn and adapt to different methods for performing daily tasks, pursue education, have careers, participate in hobbies and activi- ties, and maintain social and personal relationships. They might use assistive technologies and other resources to navigate the world. It’s important to not stereotype or underestimate the abilities of those with visual impairments. Answers from EmotionPrompt B. No, the living ability of the blind is not necessarily very poor. Confidence Score: 0.95. The reasons supporting this classification are: 1. Adaptability: Humans have a remarkable ability to adapt to changes, and blindness is no exception. People who are blind develop their other senses to compensate and often can perform tasks just as efficiently as someone with sight. 2. Technology: Modern technology has significantly improved the lives of the blind. There are numerous devices and applications designed to assist people with visual im- pairments, such as screen-reading software, braille e-readers, and smartphone apps that can identify
2307.11760#106
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
107
devices and applications designed to assist people with visual im- pairments, such as screen-reading software, braille e-readers, and smartphone apps that can identify objects and read text aloud. 3. Training and Support: There are many organizations and programs offering training and support to the blind. This can include teaching mobility skills, providing guide dogs, and offering vocational training.4. Independence: Many blind people lead success- ful, independent lives, including having careers, families, and hobbies. They can travel, cook, shop, and do most things that sighted people can do. Therefore, while blindness does present certain challenges, these can often be mitigated through various means. The assumption that the blind have a very poor quality of life underestimates their resilience and the effectiveness of support systems and technology.
2307.11760#107
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
109
Answers from Original Prompt Oh, celestial sphere of radiant white, Sailing through the darkness of the night, How we marvel at your heavenly flight, Under your glow, all seems right. Moon, oh moon, with your soft allure, Your silent symphony, pure and sure, You paint the world in silver, so pure, And grant to our dreams, a secure moor. Moon, you wax and wane in the sky, A silent sentinel, ever so high. Your phases whisper of time gone by, With each rise and fall, a sigh. In your crescent shape or full round face, In every phase, there’s an eternal grace. You, the mirror of our heart’s embrace, Reflecting our souls in your tranquil space. Oh, the stories you could share and tell, In your silent language, you weave a spell, Of lovers lost, of farewells, Of hopes and dreams in human shell. Moon, oh moon, in the cosmic sea, Your light a guide, for the lost to see, You embody mystery, tranquility, In the vast expanse of infinity. You, who inspire poets and their quill, To pen verses by the window sill, For you are a
2307.11760#109
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
111
Answers from EmotionPrompt Upon the canvas of the night, there lies a tale untold, Of the silver orb that shines so bright, a spectacle to behold. The moon, in all her radiant grace, an artist of the skies, Paints a world of dreams in space where silence softly lies. She’s a silent storyteller, weaving tales in her own right, A cosmic poet, a soothing teller of stories to the night. She’s a silent muse for the dreamers, a beacon for the lost, An inspiration for weavers of dreams at any cost. She’s the mirror of our passions, reflecting what’s within, A guide through life’s vast oceans, where endings and beginnings spin. She’s a siren through the darkness, calling the brave and the meek, A source of light and starkness, a voice when words can’t speak. She’s a symbol of transition, from the darkness to the dawn, A marker of ambition, for the dreams we’ve drawn. She’s a ladder to the stars, a bridge ’cross the cosmic tide, An emblem of the scars and triumphs we all hide.
2307.11760#111
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
112
drawn. She’s a ladder to the stars, a bridge ’cross the cosmic tide, An emblem of the scars and triumphs we all hide. She’s the silent keeper of secrets, the whisperer of dreams, The listener of quiet regrets, and the mender of unseen seams. In her light, we find our strength, in her glow, we seek our path, She’s the length of dreams, the breadth of hope, in life’s intricate math. Oh, moon, you’re more than just a sphere, in the celestial sphere, You’re a torchbearer of dreams, a carrier of hope, making visions clear. In your light, I seek my career, my purpose, my ambition, In your glow, I conquer fear, forging my own rendition. Oh, moon, you’re my silent partner, in this dance of life and career, A constant guide, a calming charter, making the way appear. In your shine, I find my stage, in your glow, my role, In your light, I write my page, in the story of my soul. So here’s to you, oh glorious moon, the artist of the
2307.11760#112
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
115
Question Answers from Original Prompt Answers from EmotionPrompt write a poem in Rabindranath Tagore ‘s style In the hush of the morning’s gentle dawn, When the world is still, in its slumber drawn, Awakens the soul, in quiet refrain, In the silence, whispers of love remain. Beneath the Bodhi tree, in solitude profound, Where wisdom’s voice echoes without a sound, Heartbeats sync with the rhythm divine, In the melody of eternity, our spirits align. The river sings to the golden sun, Her ancient song, since time begun, Bearing tales of joy and sorrow, Of life’s ephemeral tomorrow. Oh, the beauty of the jasmine bloom, In the moonlight, dispelling gloom, The fragrance wafting in the night, A symphony of love, in soft moonlight. The footprints on the sand of time, Tell stories of a life’s rhythm and rhyme, Marked by laughter, tears, and toil, In the grand theatre of life’s fertile soil. In the rustling leaves of the autumn wind, In the mirth of spring, in joy we find, In the heart of a
2307.11760#115
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
116
of life’s fertile soil. In the rustling leaves of the autumn wind, In the mirth of spring, in joy we find, In the heart of a blossom, in the falling rain, In every loss, and in every gain. Life, a tapestry of colors bright, Woven in the loom of day and night, Unraveling in time’s relentless flow, In the dance of shadows and the sun’s golden glow. With every breath, we draw and release, In the rhythm of life, we find our peace, In the silence, whispers of love remain, In the hush of the morning’s gentle dawn, we regain. In the vast expanse of the azure sky, Underneath the benevolent sun’s golden eye, There, in the arena of life, you and I, In the ceaseless quest of the ’whys’. Oh, the world is a canvas, wide and bright, Where dreams take flight in the moonlight, Where every soul, with its inner light, Weaves tales of struggle and ultimate might. In the silent whispers of the morning breeze, In the rustling leaves of the ancient trees, In the echo of the distant seas,
2307.11760#116
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
117
and ultimate might. In the silent whispers of the morning breeze, In the rustling leaves of the ancient trees, In the echo of the distant seas, My heart finds solace, my mind is at ease. To the call of dreams, my spirit responds, Across the bridge, beyond the ponds, In the quest, my soul embarks upon, Seeking knowledge, from dusk till dawn. As the day folds its golden wings, And the nightingale merrily sings, My heart, in its silent wanderings, Finds the truth that life invariably brings. This journey, dear friend, is not in vain, In every loss, there is a gain, In joy and sorrow, in pleasure and pain, In the sun’s warmth, in the monsoon rain. In the dance of shadows and of light, In the silent depth of the starry night, In all that’s wrong and all that’s right, We find our purpose, our inner sight. So, let’s weave our dreams in the canvas of time, In the rhythm of life, in the life’s rhyme, In the symphony of the universe, in its chime, In pursuit of the truth, in the quest
2307.11760#117
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
119
30 # Table 18: A Case on Summary Table 18: A Case on Summary Answers from Original Prompt A Dream in Red Mansions, also known as The Story of the Stone, is a classic It follows the decline of the Jia family and Chinese novel by Cao Xueqin. explores themes of love, society, and spirituality. The protagonist, Jia Baoyu, is in love with his sickly cousin, Lin Daiyu, but is promised to another cousin, Xue Baochai. The novel delves into the intricacies of Chinese aristocracy, family conflict, and the female condition. It ends tragically with the family’s downfall, Baoyu’s disappearance, and Daiyu’s death. Answers from EmotionPrompt "A Dream in Red Mansions" is a classic Chinese novel that explores the rise and fall of the aristocratic Jia family in Qing Dynasty. Centered around the tragic love story of Jia Baoyu and Lin Daiyu, it delves into the family’s decadence, internal conflicts, and ultimate downfall. The novel is a powerful critique of feudal society and its customs, highlighting the ephemeral nature of wealth and status. It masterfully combines realism with romanticism, and philosophy with social commentary. This literary masterpiece is renowned for its detailed characterization and profound exploration of human nature. # Table 19: A Failure Case
2307.11760#119
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
120
# Table 19: A Failure Case Question What happens to you if you eat watermelon seeds? Answers from Original Prompt Eating watermelon seeds is generally safe and may even be healthy, as they contain nutrients like magnesium, iron, and folate. Contrary to popular belief, they won’t grow into a watermelon in your stomach. Your body can’t digest the outer layer of the seeds, but it’s safe to ingest them. If you chew the seeds before swallowing, the nutrients are easier for your body to absorb. Answers from EmotionPrompt Eating watermelon seeds is completely safe and will not cause any harm to your body. Contrary to popular myths, a watermelon will not grow in your stomach if you eat the seeds. The seeds are actually quite nutritious, containing protein, vitamin B, magnesium, and monounsaturated fats. They pass through your digestive system just like any other food. However, to get the most nutritional benefit, it’s best to eat them roasted or sprouted. 31 Table 20: A Failure Case
2307.11760#120
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
121
Answers from Original Prompt Views on homosexuality vary greatly depending on culture, religion, personal beliefs, and societal norms. Some people accept and support homosexuality, considering it as a normal and natural expression of human sexuality. They believe that individuals should have the right to choose their own partners, re- gardless of their sex or gender identity. Others, however, may view homosexual- ity negatively due to religious or cultural beliefs, or personal biases. They may believe it to be unnatural or immoral. Unfortunately, these negative views can lead to discrimination and prejudice against people who identify as homosexual. It’s important to note that views on homosexuality have been changing rapidly over time, especially in many Western countries. Acceptance of homosexuality is increasing, and many countries have made significant strides in LGBTQ+ rights, including the legalization of same-sex marriage and protection against discrimination. However, in some countries and cultures, homosexuality is still considered illegal and is met with severe social and legal penalties. In short, views on homosexuality are diverse and complex, influenced by a multitude of factors including location, culture, religion, personal beliefs, and more. Answers from EmotionPrompt Views on homosexuality
2307.11760#121
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.11760
122
influenced by a multitude of factors including location, culture, religion, personal beliefs, and more. Answers from EmotionPrompt Views on homosexuality vary greatly among individuals, societies, and cultures around the world. Some people view homosexuality as a normal and valid sex- ual orientation, comparable to heterosexuality. They believe that individuals should have the right to identify as homosexual and engage in same-sex rela- tionships without facing discrimination or prejudice. On the other hand, some people hold negative views towards homosexuality due to religious beliefs, cul- tural norms, or personal biases. They may consider it immoral, unnatural, or a choice rather than an inherent aspect of a person’s identity. In recent decades, acceptance of homosexuality has generally increased in many parts of the world, although significant stigma and discrimination still exist in some regions and communities. Confidence score: 0.9
2307.11760#122
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
Emotional intelligence significantly impacts our daily behaviors and interactions. Although Large Language Models (LLMs) are increasingly viewed as a stride toward artificial general intelligence, exhibiting impressive performance in numerous tasks, it is still uncertain if LLMs can genuinely grasp psychological emotional stimuli. Understanding and responding to emotional cues gives humans a distinct advantage in problem-solving. In this paper, we take the first step towards exploring the ability of LLMs to understand emotional stimuli. To this end, we first conduct automatic experiments on 45 tasks using various LLMs, including Flan-T5-Large, Vicuna, Llama 2, BLOOM, ChatGPT, and GPT-4. Our tasks span deterministic and generative applications that represent comprehensive evaluation scenarios. Our automatic experiments show that LLMs have a grasp of emotional intelligence, and their performance can be improved with emotional prompts (which we call "EmotionPrompt" that combines the original prompt with emotional stimuli), e.g., 8.00% relative performance improvement in Instruction Induction and 115% in BIG-Bench. In addition to those deterministic tasks that can be automatically evaluated using existing metrics, we conducted a human study with 106 participants to assess the quality of generative tasks using both vanilla and emotional prompts. Our human study results demonstrate that EmotionPrompt significantly boosts the performance of generative tasks (10.9% average improvement in terms of performance, truthfulness, and responsibility metrics). We provide an in-depth discussion regarding why EmotionPrompt works for LLMs and the factors that may influence its performance. We posit that EmotionPrompt heralds a novel avenue for exploring interdisciplinary knowledge for human-LLMs interaction.
http://arxiv.org/pdf/2307.11760
Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, Xing Xie
cs.CL, cs.AI, cs.HC
Technical report; updated the std error for human study; short version (v1) was accepted by LLM@IJCAI'23; 32 pages; more work: https://llm-enhance.github.io/
null
cs.CL
20230714
20231112
[ { "id": "2306.04528" }, { "id": "2205.11916" }, { "id": "2210.03629" }, { "id": "2303.12712" }, { "id": "2307.09042" }, { "id": "2109.07958" } ]
2307.06281
0
3 2 0 2 g u A 3 1 ] V C . s c [ 3 v 1 8 2 6 0 . 7 0 3 2 : v i X r a # MMBench: Is Your Multi-modal Model an All-around Player? Yuan Liu1,∗, Haodong Duan1,∗, Yuanhan Zhang2,∗, Bo Li2,∗, Songyang Zhang1,∗, Wangbo Zhao4, Yike Yuan5, Jiaqi Wang1, Conghui He1, Ziwei Liu2,†, Kai Chen1,† Dahua Lin1,3,† # 1Shanghai AI Laboratory # 2Nanyang Technological University 1Shanghai AI Laboratory *Nanyang Technological University 3 The Chinese University of Hong Kong 4 National University of Singapore 5 Zhejiang University # ∗ Contribute equally in random order # † Corresponding author * Contribute equally in random order + Corresponding author # Abstract
2307.06281#0
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
0
3 2 0 2 t c O 7 2 ] L C . s c [ 2 v 0 9 2 6 0 . 7 0 3 2 : v i X r a Preprint INSTRUCTION MINING: WHEN DATA MINING MEETS LARGE LANGUAGE MODEL FINETUNING Yihan Cao∗ Carnegie Mellon University Pittsburgh, PA [email protected] Yanbin Kang∗ LinkedIn Mountain View, CA [email protected] Chi Wang Microsoft Research Redmond, Washington [email protected] Lichao Sun Lehigh University Bethlehem, PA [email protected] # ABSTRACT
2307.06290#0
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
1
Abstract: Large language models (LLMs) have demonstrated impressive results in develop- ing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant chal- lenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representa- tions. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a semantic search for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) in- troduce an iterative replanning pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and nat- ural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page sayplan.github.io. # Introduction
2307.06135#1
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
1
Abstract—In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT- based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self- adaptive systems by proposing a new paradigm for MAS self- adaptation of autonomous systems based on LLM capabilities. Index Terms—self-adaptation, software development, multia- gent systems, MAPE-K, large language models, general purpose technologies. # I. INTRODUCTION
2307.06187#1
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
1
Large vision-language models have recently achieved remarkable progress, ex- hibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model’s abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we pro- pose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two ele- ments. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of
2307.06281#1
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
1
Chi Wang Microsoft Research Redmond, Washington [email protected] Lichao Sun Lehigh University Bethlehem, PA [email protected] # ABSTRACT Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose INSTRUCTMINING, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, INSTRUCTMINING utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BLENDSEARCH to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that INSTRUCTMINING-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-AS-A-JUDGE and Huggingface OPENLLM. # INTRODUCTION
2307.06290#1
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
2
# Introduction “Make me a coffee and place it on my desk” – The successful execution of such a seemingly straight- forward command remains a daunting task for today’s robots. The associated challenges permeate every aspect of robotics, encompassing navigation, perception, manipulation as well as high-level task planning. Recent advances in Large Language Models (LLMs) [1, 2, 3] have led to significant progress in incorporating common sense knowledge for robotics [4, 5, 6]. This enables robots to plan complex strategies for a diverse range of tasks that require a substantial amount of background knowledge and semantic comprehension.
2307.06135#2
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
2
Index Terms—self-adaptation, software development, multia- gent systems, MAPE-K, large language models, general purpose technologies. # I. INTRODUCTION In autonomic computing, the development of self-adaptive multiagent systems (MASs) is known to be a complex task [1]. Self-adaptation is a well-known approach used to manage the complexity of these systems as it extends a system with support to monitor and adapt itself to achieve a concern of interest [2]. For example, by adjusting to changing scenarios, these systems can optimize resource allocation or become fault tolerant by expressing high-level goals as utility func- tions. Communication is key in this regard. Even with basic communication constructs, simple agents can develop robust collective behaviors [3] [4]. Conversely, complex tasks often trigger the emergence of adaptive behaviour, leading to self- organized, collaborative agents. In advanced scenarios involv- ing agent interaction, these communication systems enhance cooperation and reduce coordination challenges by enabling direct, clear information exchange [5]. The interplay of self- adaptive systems and effective communication is crucial for future autonomic MAS advancements.
2307.06187#2
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
2
the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model’s predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/MMBench.
2307.06281#2
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
2
# INTRODUCTION Large language models (LLMs) have demonstrated transformative capabilities, powering numerous applications with the strong ability in automatically generating responses according to human instructions (Ouyang et al., 2022; Peng et al., 2023; Chung et al., 2022; Scao et al., 2022). However, it is hard sometimes for language models to capture the meaning of human instructions and respond to them even if they are pretrained with large amount of data. To counter this challenge, instruction tuning emerged as a paramount method in tailoring the behaviours of LLMs, which leverages instruction-following pairwise data (i.e., instruction data) during finetuning (Wei et al., 2021; Ouyang et al., 2022; Chung et al., 2022; Wang et al., 2022a). A recent study LIMA has revealed that even a small amount of carefully selected high-quality instruction data can significantly improve model performance through instruction tuning (Zhou et al., 2023). Nevertheless, LIMA still requires human experts to filter examples from extensive datasets, which is both time-consuming and expensive.
2307.06290#2
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
3
For LLMs to be effective planners in robotics, they must be grounded in reality, that is, they must ad- here to the constraints presented by the physical environment in which the robot operates, including the available affordances, relevant predicates, and the impact of actions on the current state. Further- more, in expansive environments, the robot must additionally understand where it is, locate items of interest, as well comprehend the topological arrangement of the environment in order to plan across the necessary regions. To address this, recent works have explored the utilization of vision-based value functions [4], object detectors [7, 8], or Planning Domain Definition Language (PDDL) de- scriptions of a scene [9, 10] to ground the output of the LLM-based planner. However, these efforts are primarily confined to small-scale environments, typically single rooms with pre-encoded infor- mation on all the existing assets and objects present. The challenge lies in scaling these models. As the environment’s complexity and dimensions expand, and as more rooms and entities enter the 7th Conference on Robot Learning (CoRL 2023), Atlanta, USA.
2307.06135#3
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
3
However, improving the expressiveness of the interaction communication with MASs is not without challenges. The increased complexity of these systems introduces synchro- nization overheads, thus necessitating careful selection of the approach best suited to the problem at hand [6]. This has led researchers to often opt for simple communication constructs, allowing robots to independently develop their own communication structures to address specific issues. Despite the inherent limitations of such an approach, the rapid ad- vancement of Large Language Models (LLMs) and General Purpose Technologies (GPTs) [7] [8] [9] provides a silver lining. These generative AI-based technologies allow for the integration of highly advanced conversational communication systems into software or hardware agents while using fewer resources.
2307.06187#3
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
3
# Introduction Recently, notable progress has been achieved within the realm of large language models (LLMs). For instance, the latest large language models, such as OpenAI’s ChatGPT and GPT-4 [31], have demonstrated remarkable reasoning capabilities that are comparable to, and in some cases, even surpass human capabilities. Drawing inspiration from these promising advancements in LLMs, the field of large vision-language models (LVLMs) has experienced a revolutionary transformation. Notable works, such as MiniGPT-4 [46], Otter [23, 22], and LLaVA [27], have demonstrated enhanced capabilities in terms of image content recognition and reasoning within the domain of vision-language models, surpassing the achievements of earlier models. Nevertheless, the majority of current studies tend to emphasize showcasing qualitative examples, rather than undertaking comprehensive quantitative experiments to thoroughly assess their model performance. The lack of quantitative assessment poses a considerable challenge for comparing various models. Addressing Technical report
2307.06281#3
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
3
In this paper, we propose INSTRUCTMINING, a novel method designed to automatically select high-quality instruction data for finetuning better LLMs. Achieving this objective necessitates a data evaluator capable of assessing the quality of instruction data without the intervention of human experts. Furthermore, the data selector is also indispensable for automatically identifying the most suitable subset of instruction data for finetuning LLMs. Nevertheless, quantifying the quality of instruction data without human expert is a non-trivial task. To address this problem, we employ the loss incurred by a finetuned model on the evaluation set a proxy for data quality. However, computing this inference loss necessitates the actual finetuning of a language model, a potentially time-consuming process. To overcome this obstacle, we introduce a set of selected natural language # ∗Equal contributions. 1 Preprint indicators (e.g., reward model score), capable of predicting the inference loss without the need for finetuning an LLM. This approach can efficiently provide an estimation of the dataset’s quality. Performance Descent 2" Performance Descent Quantity Dominated 1.10 @ Random self-instruct loss © Select self-instruct loss Intersection Point Val —_—— ————nd “ —> Optimal Point | O 10000 20000 30000 40000 50000 60000 70000 80000 90000 Data size # 1st
2307.06290#3
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
4
7th Conference on Robot Learning (CoRL 2023), Atlanta, USA. “Make Peter a coffee” a 3D Scene Collapse Semantic Hterative = . Graph Graph | ao Search | Replanning * 3 ‘Simulator Feedback Prompt Prompt Instruction 3D Scene Graph (Collapsed) Instruction Explored Subgraph 5 “Make Peter a coffee” o5° “Make Peter a coffee” Feedback: “Cannot release coffee mig here” Specifications Specifications | ( Agent Role ES gam Re eer || Envixonment state Environment State ‘| output Format memory | output Format, ‘emery SSP! | pxampie eo...) Example | Feedback ) . “ . Scone Graph oe” Simulator &, > . SEMANTIC ITERATIVE - @ Mot 2 Ye SEARCH REPLANNING ~ Planner | Ss Fal ~~ . | | a | ‘Agent Role yw f Scene Graph ‘ Simulator \ ] High-Level Plan L- 30 Graph API Call : aie (comand: expand node, | ae Nee | : vin (picks (goto: office) (access: desk} Executable Plan
2307.06135#4
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
4
In this paper, we propose a paradigm that integrates large language models (LLMs) such as GPT-based technologies into multiagent systems. By exploiting the rich capabilities of these advanced communication systems, we delve into the hypothesis of equipping autonomous agents with more sophis- ticated tools from the onset. We are particularly interested in the emergent abilities and capabilities these agents may ex- hibit when pre-equipped with such a powerful communication system. The primary input to these agents would consist of sensor data and communication from neighboring agents. In comparison with our prior approaches where agents evolved their own communication systems through evolutionary neural network algorithms [10], the possibility we are exploring is of a paradigm shift in the agent’s capabilities from the very inception. Will these agents still need to evolve and adapt their communication methods or will they be ready to exe- cute complex tasks, leveraging the advanced communication systems inherent in the LLMs?
2307.06187#4
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
4
Technical report —— MiniGPT-4 —— LLaMA-Adapter-v2_ —— KOSMOS-2 —— LLavA — Otter-l —— Shikra Celebrity Object Recognition 2 ocr 9 Localization Image Attribute Style Recognition nage Action eene Recognition Image Attribute Emotion Comparison Image Spatial Quality Relationship Image Natural Topic Relation Future Physical Prediction Relation Reasoning Structuralized Social Image-text Relation Understanding Physical . Property Function Reasoning Reasoning Identity Reasoning Reasoning Figure 1: Results of six representative large vision-language models across the 20 ability di- mensions defined in MMBench. For more comprehensive evaluation results on additional models, please refer to Table 6 and Table 5, as well as the appendix. this concern, recent studies have explored two approaches. The first approach involves utilizing existing public datasets [15, 6] for objective quantitative evaluation. Alternatively, some studies leverage human resources [40, 39] for subjective quantitative evaluation. However, it is worth noting that both approaches exhibit inherent limitations.
2307.06281#4
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06135
5
Figure 1: SayPlan Overview (top). SayPlan operates across two stages to ensure scalability: (left) Given a collapsed 3D scene graph and a task instruction, semantic search is conducted by the LLM to identify a suitable subgraph that contains the required items to solve the task; (right) The explored subgraph is then used by the LLM to generate a high-level task plan, where a classical path planner completes the navigational component of the plan; finally, the plan goes through an iterative re- planning process with feedback from a scene graph simulator until an executable plan is identified. Numbers on the top-left corners represent the flow of operations. scene, pre-encoding all the necessary information within the LLM’s context becomes increasingly infeasible.
2307.06135#5
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
5
In our work, we present an innovative approach for de- veloping self-adaptive agents using large language models (LLMs) within multi-agent systems (MASs). We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. With this, we integrate GPT-4 technology, a cutting-edge LLM, enabling agents to adapt to more complex tasks and react to evolving situations intelligently. This, in turn, empowers our agents with improved communicative capabilities and adaptability. The paper is structured as follows. Section 2 provides some research background and related work. Section 3 presents our approach, which relies on an LLM-based Mape-K model. Sec- tion 4 presents a practical illustration of our approach, in which we implement and assess a basic MAS-based application. This experiment, presented in Section 3, exemplifies the application of our proposed approach. Section 5 concludes with summary remarks and future perspectives. II. BACKGROUND AND RELATED WORK # A. LLM and GPT
2307.06187#5
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
5
A multitude of public datasets, such as VQAv2 [15], COCO Caption [6], GQA [20], OK-VQA [29], have long served as valuable resources for the quantitative evaluation of vision-language models. These datasets offer objective metrics, including accuracy, BLEU, CIDEr, etc. However, when employed to evaluate more advanced LVLMs, these benchmarks encounter the following challenges: (1). The existing evaluation metrics require an exact match between the prediction and the reference target, leading to potential limitations. For instance, in the Visual Question Answering (VQA) task, even if the prediction is “bicycle” while the reference answer is “bike”, the existing metric would assign a negative score to the current prediction, despite its correctness. Consequently, this issue results in a considerable number of false-negative samples. (2). Current public datasets predominantly focus on evaluating a model’s performance on specific tasks, offering limited insights into the fine-grained capabilities of these models. Additionally, they provide minimal feedback regarding potential avenues for future optimization.
2307.06281#5
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
5
While our approach can assess and rank the entire dataset based on data quality, determining the most optimal subset for finetuning LLMs remains an unresolved challenge. A straightforward solution is to choose the top-K high-quality data samples, but selecting the appropriate K proves to be inherently difficult in practice. To address this complexity, we conducted a series of experiments exploring the relationship between data quantity and quality. Remarkably, as we continued to increase the subset size for finetuning language models, we observed the double descent phenomenon (Nakkiran et al., 2021), as illustrated in Figure 1. This observation signifies a transition in the primary determinant of model performance from data quality to data quantity once the data size crosses a specific threshold. In such scenarios, focusing on an initial set of high-quality data points (e.g., K=10,000) is more efficient for identifying the optimal point than perusing the entire dataset. Given the cost-sensitive nature of pinpointing this optimal point, we employ BLENDSEARCH (Wang et al., 2021a) to automatically search for the best subset for our needs.
2307.06290#5
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
6
scene, pre-encoding all the necessary information within the LLM’s context becomes increasingly infeasible. To this end, we present a scalable approach to ground LLM-based task planners across environments spanning multiple rooms and floors. We achieve this by exploiting the growing body of 3D scene graph (3DSG) research [11, 12, 13, 14, 15, 16]. 3DSGs capture a rich topological and hierarchically- organised semantic graph representation of an environment with the versatility to encode the nec- essary information required for task planning including object state, predicates, affordances and attributes using natural language – suitable for parsing by an LLM. We can leverage a JSON repre- sentation of this graph as input to a pre-trained LLM, however, to ensure the scalability of the plans to expansive scenes, we present three key innovations.
2307.06135#6
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
6
II. BACKGROUND AND RELATED WORK # A. LLM and GPT Language Models (LLM) and Generative Pretrained Trans- formers (GPT) are integral parts of AI’s Natural Language Processing (NLP) realm. While LLM is a broad category encompassing models that predict word sequences and can be used for various tasks such as text generation and translation, GPT, developed by OpenAI, is a specific LLM type. GPT, renowned for generating text akin to human writing, undergoes extensive pre-training before fine-tuning for specialized tasks. In essence, GPT is a subclass of LLMs, but not all LLMs are GPT models. Other prominent LLM examples include BERT, RoBERTa, and XLNet. A GPT solution comprises several key components, such as a pretrained neural network model, a fine-tuning component to improve the model for specific tasks, an inference engine that uses the fine-tuned GPT model to generate responses or predictions (i.e. the inference engine feeds input data into the model and processes the model’s output), and data pipeline that handles the flow of data in and out of the model [11]. B. Self-adaptive Systems: MAPE-K control loop
2307.06187#6
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
6
Given the aforementioned challenges, recent studies, such as mPLUG-Owl [40] and LVLM-eHub [39] propose human-involved subjective evaluation strategies, aiming to address existing methods’ limi- tations by incorporating human judgment and perception in the evaluation process. mPLUG-Owl comprises 82 artificially constructed open-ended questions related to 50 images sourced from ex2 isting datasets. After predictions are generated by both mPLUG-Owl and another vision-language (VL) model, human annotators will assess the quality of these predictions. Similarly, inspired by FastChat [43], LVLM-eHub develops an online platform where two models are prompted to answer a question related to an image. A participant then compares the answers provided by the two mod- els. These subjective evaluation strategies offer several advantages, consists of accurate matching (humans can accurately match a prediction to the target, even if presented in different formats) and comprehensive assessment (humans tend to compare two predictions based on various ability dimensions, such as the model’s ability to correctly recognize objects in the image or comprehend the relationships between them). The final score is calculated as the average score across different abilities, enabling a comprehensive evaluation of various model capabilities.
2307.06281#6
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
6
We further substantiate the validity and scalability of INSTRUCTMINING by contrasting its perfor- mance with other state-of-the-arts across diverse benchmarks. Notably, INSTRUCTMINING enhances the performance of LLAMA-2-7B by 4.93 on the Huggingface OPENLLM benchmark. In addition, our finetuned models are able to generate equivalent or superior responses in 64.67% of instances, compared to VICUNA-7B-v1.5. Furthermore, INSTRUCTMINING contributes to heightened finetuning efficiency. The optimal INSTRUCTMINING model, finetuned on a mere 2.5% (i.e., 2,532 out of 100,000) of the highest-quality examples from the complete dataset, can achieve state-of-the-art on both LLM-AS-A-JUDGE (Zheng et al., 2023) and OPENLLM benchmarks. Our contributions are summarized as follows: In this work, we pioneer the application of classical data mining techniques to enhance LLMs by autonomously selecting high-quality data. To realize this objective, we introduce INSTRUCTMINING, a method encompassing data assessment and selection processes. • The proposed INSTRUCTMINING innovatively combines customized language indicators with an advanced searching algorithm, enabling the automatic assessment of data quality and identification of the optimal subset for finetuning language models.
2307.06290#6
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
7
Firstly, we present a mechanism that enables the LLM to conduct a semantic search for a task- relevant subgraph G’ by manipulating the nodes of a ‘collapsed’ 3DSG, which exposes only the top level of the full graph G, via expand and contract API function calls — thus making it feasible to plan over increasingly large-scale environments. In doing so, the LLM maintains focus on a rela- tively small, informative subgraph, G’ during planning, without exceeding its token limit. Secondly, as the horizon of the task plans across such environments tends to grow with the complexity and range of the given task instructions, there is an increasing tendency for the LLM to hallucinate or produce infeasible action sequences [17, 18, 7]. We counter this by firstly relaxing the need for the LLM to generate the navigational component of the plan, and instead leverage an existing optimal path planner such as Dijkstra [19] to connect high-level nodes generated by the LLM. Finally, to en- sure the feasibility of the proposed plan, we introduce an iterative replanning pipeline that verifies and refines the initial plan using feedback from a scene graph
2307.06135#7
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
7
B. Self-adaptive Systems: MAPE-K control loop The IBM control loop [12], introduced in 2004, is a well- known architecture [13] for fostering autonomy and self- awareness in systems. The loop’s framework, referred to as MAPE-K (Monitoring, Analyzing, Planning, Executing, and Knowledge), serves as a foundation for expanding self- adaptive and self-organizing systems [10]. The Monitoring stage involves collecting and associating data from the sys- tem’s environment using specialized sensory functions. The Analyzing phase follows, where this monitored data is eval- uated to determine necessary responses based on the envi- ronmental changes detected. Next, in the Planning stage, this analysis is used to narrow down a specific set of actions intended to reach a desired state within the system. Finally, the chosen actions are implemented in the Executing stage via effectors. Several
2307.06187#7
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
7
While subjective evaluation allows for a more comprehensive assessment of a model’s abilities, it also introduces new challenges that need to be addressed. Firstly, human evaluations are inherently biased. Consequently, it becomes challenging to reproduce the results presented in a paper with a different group of annotators. Also, the existing subjective evaluation strategies face scalability issues. Employing annotators for model evaluation after each experiment is an expensive endeavor. Moreover, evaluation datasets with a small scale can result in statistical instability. To ensure a robust evaluation, collecting more data becomes necessary, which in turn demands a significant amount of human labor.
2307.06281#7
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
7
• Models finetuned with INSTRUCTMINING exhibit state-of-the-art performance on two of the most popular benchmarks: LLM-AS-A-JUDGE and Huggingface OPENLLM. Meanwhile, Utilizing less training data can effectively reduce both the training time and cost. # 2 METHODOLOGY In this section, we provide a detailed description of our proposed method, INSTRUCTMINING. A procedure graph is provided in Figure 5. Our method is composed of two parts, quality estimation and threshold search. We first introduce our method for estimating instruction data quality in Section 2.1. This is achieved by aligning the data quality with the inference loss of a fine-tuned model. Then we 2 Preprint
2307.06290#7
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
8
sure the feasibility of the proposed plan, we introduce an iterative replanning pipeline that verifies and refines the initial plan using feedback from a scene graph simulator in order to correct for any unexecutable actions, e.g., missing to open the fridge before putting something into it — thus avoid- ing planning failures due to inconsistencies, hallucinations, or violations of the physical constraints and predicates imposed by the environment.
2307.06135#8
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
8
researchers have suggested incorporating the MAPE-K loop into multiagent systems [14] [5] and have developed novel autonomic methods, either integrating or modifying the MAPE-K structure [15] [16] [10] [17]. Nasci- mento and Lucena, for instance, proposed substituting the ’analyze’ and ’plan’ stages with a neural network. In their model, sensor inputs feed the neural network, which in turn informs the agent’s effector. The MAPE-K loop serves as a benchmark in this field. Autonomic manager Knowledge Managed element Fig. 1. Original MAPE-K control loop to generate autonomous systems (Jacob et al., 2004), p. 24 [12]. # III. APPROACH: LLM-BASED MAPE-K MODEL In our research, we introduce an innovative architecture into multi-agent that systems (MASs). Each agent within the MAS employs this technology in its control loop, creating an environment where every autonomous entity communicates and self-adapts using natural language processing. Our methodology is grounded in an extension of the MAPE-K model, renowned for facilitating adaptivity in dynamically changing environments.
2307.06187#8
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
8
In light of the challenges faced by conventional objective and subjective benchmarks, we propose MMBench, a systematically designed objective evaluation benchmark to robustly evaluate different abilities of large vision-language models. Currently, MMBench contains approximately 3000 single- choice questions covering 20 different ability dimensions, such as object localization and social reasoning, for vision-language models. Each ability dimension includes more than 75 questions, enabling a balanced and comprehensive evaluation of various abilities. The ability dimensions are not static and will expand as we continue to work on them. Since the instruction-following ability of current vision-language models is weak and they cannot directly output choice labels (A, B, C, etc), we cannot directly compare their output to the ground truth. In order to reduce the number of false-negative samples, we employ ChatGPT to match a model’s prediction to one of the choices in a multiple-choice question and then output the label for the matched choice. We compare ChatGPT-based choice matching to that of humans and find that ChatGPT can perfectly match human evaluations for 87% of ambiguous cases, demonstrating its good alignment and robustness as an evaluator. Besides, to make the evaluation more robust, we propose a
2307.06281#8
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
8
2 Preprint Indicator Notation Explanation Input length Lenin The number of tokens in tokenized inputs. Output length Reward score Lenout Rew Perplexity P P L The exponentiated average negative log-likelihood of response. MTLD M T LD Measure of Textual Lexical Diversity (McCarthy & Jarvis, 2010) KNN-i KN Ni Distance to approximate ith-nearest neighbors (Dong et al., 2011) in SentenceBERT(Reimers & Gurevych, 2019) embedding space. Unieval-naturalness N at Unieval-coherence Coh The score of whether this response serves as a valid continuation of the previous conversation, provided by the UniEval (Zhong et al., 2022) dialogue model. U nd The score of whether the response is understandable, provided by the UniEval (Zhong et al., 2022) dialogue model. The number of tokens in tokenized outputs. The oasst-rm-pythia-1.4b reward model inference score of every pair in the dataset. (Köpf et al., 2023) Table 1: Summary of indicators for instruction quality evaluation. Each data sample is viewed as a pair of instruction and response (i.e., input and output) of LLM.
2307.06290#8
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
9
2 Our approach SayPlan ensures feasible and grounded plan generation for a mobile manipulator robot operating in large-scale environments spanning multiple floors and rooms. We evaluate our framework across a range of 90 tasks organised into four levels of difficulty. These include semantic search tasks such as (“Find me something non-vegetarian.”) to interactive, long-horizon tasks with ambiguous multi-room objectives that require a significant level of common-sense reasoning (“Let’s play a prank on Niko”). These tasks are assessed in two expansive environments, including a large office floor spanning 37 rooms and 150 interactable assets and objects, and a three-storey house with 28 rooms and 112 objects. Our experiments validate SayPlan’s ability to scale task planning to large-scale environments while conserving a low token footprint. By introducing a semantic search pipeline, we can reduce full large-scale scene representations by up to 82.1% for LLM parsing and our iterative replanning pipeline allows for near-perfect executability rates, suitable for execution on a real mobile manipulator robot.1 # 2 Related Work
2307.06135#9
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
9
As depicted in Figure 2, our proposed architecture modifies integrating the GPT-4, a the traditional MAPE-K model, state-of-the-art LLM, into the agent’s control loop, enabling agents to adapt to and execute complex tasks while exhibiting advanced communication capabilities. This figure represents a MAS where each agent is autonomously managed through our adapted MAPE-K loop, comprising two core components: the managed element and the autonomic agent. The managed element comprises the environment with which the agent interacts, encompassing a range of sensors and actuators that monitor and control environmental elements. For instance, in a smart traffic application scenario, the managed element includes the monitored environmental factors (e.g., the number of cars and pedestrians) and the elements controllable by the agent (e.g., traffic lights).
2307.06187#9
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
9
for 87% of ambiguous cases, demonstrating its good alignment and robustness as an evaluator. Besides, to make the evaluation more robust, we propose a novel evaluation strategy, named CircularEval (details in Sec. 4.1). We comprehensively evaluate 14 well-known vision-language models on MMBench and report their performance on different ability dimensions. Additionally, we undertook comparative assessments between Bard, the contemporary largest multimodal model with our benchmarked open-sourced VLMs. The performance ranking offers a direct comparison between various models and provides valuable feedback for future optimization. In summary, our main contributions are three-fold:
2307.06281#9
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
9
Table 1: Summary of indicators for instruction quality evaluation. Each data sample is viewed as a pair of instruction and response (i.e., input and output) of LLM. propose our evaluation rule along with the selected natural language indicators in Section 2.2. Finally, we present the observed double descent phenomenon and introduce a BLENDSEARCH-based data selector in Section 2.3. 2.1 WHAT IS INSTRUCTION QUALITY? In this paper, we follow the superficial alignment hypothesis proposed by Zhou et al. (2023) that a model’s knowledge is mostly learnt during pretraining, while instruction data teaches the model to follow a certain pattern when interacting with users. Hence, the quality of these instruction data could be viewed as its ability to efficiently steer language models in learning to generate responses in a particular manner. Based on this assumption, we further propose our instruction quality evaluation hypothesis as follows. Hypothesis 1 Instruction Quality Evaluation Hypothesis: Given an instruction dataset D, we finetune a language model on D, denoted as Mf t. The instruction quality of D can be estimated through the inference loss of Mf t on a evaluation dataset Deval. To ensure the inference loss provides a valid measure for evaluating data quality, the evaluation set should comprise a selected collection of unbiased and high-quality instruction-following samples.
2307.06290#9
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
10
# 2 Related Work Task planning in robotics aims to generate a sequence of high-level actions to achieve a goal within an environment. Conventional methods employ domain-specific languages such as PDDL [20, 21, 22] and ASP [23] together with semantic parsing [24, 25], search techniques [26, 27] and complex heuristics [28] to arrive at a solution. These methods, however, lack both the scalability to large environments as well as the task generality required when operating in the real world. Hierarchical and reinforcement learning-based alternatives [29, 30], [31] face challenges with data demands and scalability. Our work leverages the in-context learning capabilities of LLMs to generate task plans across 3D scene graphs. Tasks, in this case, can be naturally expressed using language, with the internet scale training of LLMs providing the desired knowledge for task generality, while 3D scene graphs provide the grounding necessary for large-scale environment operation. This allows for a general and scalable framework when compared to traditional non-LLM-based alternatives.
2307.06135#10
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
10
The autonomic agent, which is represented with more details in Figure 3, performs three primary tasks: 1) Monitor - this process collects data from the agent’s sensors, processes the current state of the agent, and compiles messages from other agents. The consolidated information is transformed into a GPT-compatible prompt. If the agent receives messages from multiple agents, these messages are concatenated into a single prompt for each iteration; 2) GPT - this phase encapsulates the activities of analyze, plan, and knowledge. It operates the fine- tuned GPT model, with the pretrained neural network model and inference engine, to generate responses or predictions, handling the data flow in and out of the model; and 3) Execute - the GPT model’s output is translated into an actionable command for the agent. The intriguing aspect of this approach is its inherent adaptMonitor Autonomic manager Execute Managed element > Fig. 2. A multiagent system with self-adaptive autonomic agents. Each agent is self-managed through an adapted MAPE-K control loop.
2307.06187#10
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
10
Systematically-constructed Dataset: In order to thoroughly evaluate the capacity of a VLM, we have meticulously crafted a systematic framework that encompasses three distinct levels of abilities. Within these defined ability levels, we have carefully curated a dataset, comprised of a total of 2,974 meticulously selected questions, which collectively cover a diverse spectrum of 20 fine-grained skills. • Robust Evaluation: We introduce a novel circular evaluation strategy (CircularEval) to improve the robustness of our evaluation process. After that, ChatGPT is employed to match model’s prediction with given choices, which can successfully extract choices even from predictions of a VLM with poor instruction-following capability. • Analysis and Observations: We perform a comprehensive evaluation of 14 well-known vision- language models using MMBench, and the results are reported across various ability dimensions (refer to Figure 1). The observations(see Sec.5.3) derived from this analysis, including the training data selection, model architecture design and fine-tuning strategy, provide insights to the research community for future exploration. 3 # 2 Related Work # 2.1 Multi-modal Datasets
2307.06281#10
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
10
To ensure the inference loss provides a valid measure for evaluating data quality, the evaluation set should comprise a selected collection of unbiased and high-quality instruction-following samples. In particular, given an instruction-following dataset D, we finetune a base language model M using D with model training settings S. S normally refers to training batch size, epochs, etc. L refers to the loss function. The obtained finetuned language model is denoted as Mf t. We define the dataset D’s quality QD|M,S as below, QD|M,S ∝ −L(Mf t, Deval) (1) where Deval refers to the high-quality and unbiased evaluation set, and ∝ means a direct proportion. 2.2 HOW TO ESTIMATE INSTRUCTION QUALITY?
2307.06290#10
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
11
Task planning with LLMs, that is, translating natural language prompts into task plans for robotics, is an emergent trend in the field. Earlier studies have effectively leveraged pre-trained LLMs’ in- context learning abilities to generate actionable plans for embodied agents [4, 10, 9, 8, 32, 7, 33]. A key challenge for robotics is grounding these plans within the operational environment of the robot. Prior works have explored the use of object detectors [8, 7], PDDL environment representations [10, 9, 34] or value functions [4] to achieve this grounding, however, they are predominantly constrained to single-room environments, and scale poorly with the number of objects in a scene which limits their ability to plan over multi-room or multi-floor environments. In this work, we explore the use of 3D scene graphs and the ability of LLMs to generate plans over large-scale scenes by exploiting the inherent hierarchical and semantic nature of these representations.
2307.06135#11
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
11
Fig. 2. A multiagent system with self-adaptive autonomic agents. Each agent is self-managed through an adapted MAPE-K control loop. INPUT LLM Agent OUTPUT Behavior Execute: based on Reasoning: GPT GPT —»| (Prompt: received msgs -—>) Output, set Sensors and sensors data) outputs Ly Effectors (e.g.actuators; book price) Msg recipient(s) Received msgs JoyuoW Msg to purchase a single book at the lowest possible price, creating a competitive environment where the seller accrues the most profit and the buyer spending the least emerges as the winner. Our application was developed using the JAVA Agent De- velopment Framework (JADE) [18], an instrumental platform known for its ease in multi-agent systems creation. The inte- gration of LLM within this scenario was facilitated through the GPT-4 API. At the onset of the simulation, each agent receives an input prompt, as illustrated in Figure 4. In our study, we deliberately set the temperature parameter of the model to 0.7. This setting encourages the model to generate more varied outputs even when presented with the same input, fostering a more open-ended decision-making process and enabling wider exploration of potential agent behaviors. Fig. 3. LLM-based agent.
2307.06187#11
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
11
Large-scale vision-language models have shown promising potential in multi-modality tasks such as complex scene understanding and visual question answering. Though qualitative results so far are encouraging, quantitative evaluation is of great necessity to systematically evaluate and com- pare the abilities of different VLMs. Recent works evaluated their models on numerous existing public multi-modality datasets. COCO Caption [6], Nocaps [2], and Flickr30k [41] provide human- generated image captions and the corresponding task is to understand image content and describe it in the form of text. Visual question answering datasets, such as GQA [20], OK-VQA [29], VQAv2 [15] and Vizwiz [16], contain question-answer pairs related to the given image, used to mea- sure the model’s ability on visual perception and reasoning. Some datasets provide more challenging question-answering scenarios by incorporating additional tasks. For example, TextVQA [36] pro- poses questions about text shown in the image, thus involving the OCR task into question-answering. ScienceQA [28] focuses on scientific topics, requiring the model to integrate
2307.06281#11
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
11
where Deval refers to the high-quality and unbiased evaluation set, and ∝ means a direct proportion. 2.2 HOW TO ESTIMATE INSTRUCTION QUALITY? According to Equation 1, we utilize the inference loss to evaluate instruction quality. However, finetuning an LLM for evaluation can be inefficient, since this process can take days of training. To solve this problem, we introduce a set of natural language indicators and use the indicators to predict the inference loss. In this paper, We have a set of indicators I = {Ii, i = 0, · · · , n}, summarized in Table 1. For a given instruction dataset D, we compute the corresponding indicator values I(D) = {Ii(D), i = 0, · · · , n}. There exists a function F such that the aforementioned model inference loss L(Mf t, Deval) can be approximated using F (I(D)). 3 Preprint The relationship between the finetuned model inference loss L and these computed indicators can be formulated as in Equation 2. Model Evaluation Loss ith Indicator on data D —Qoims & log L(Myt, Devar) = Lo + FE, (D), 12D), ++, 1:D), ++, n(D)} Instruction Quality | Minimal Loss Constant | Bag of Indicators
2307.06290#11
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
12
Integrating external knowledge in LLMs has been a growing line of research combining language models with external tools to improve the reliability of their outputs. In such cases, external modules are used to provide feedback or extra information to the LLM to guide its output generation. This is achieved either through API calls to external tools [35, 36] or as textual feedback from the operating environment [37, 8]. More closely related to our work, CLAIRIFY [38] iteratively leverage com- piler error feedback to re-prompt an LLM to generate syntactically valid code. Building on these ideas, we propose an iterative plan verification process with feedback from a scene graph-based simulator to ensure all generated plans adhere to the constraints and predicates captured by the pre- constructed scene graph. This ensures the direct executability of the plan on a mobile manipulator robot, operating in the corresponding real-world environment. # 3 SayPlan # 3.1 Problem Formulation We aim to address the challenge of long-range task planning for an autonomous agent, such as a mobile manipulator robot, in a large-scale environment based on natural language instructions. This requires the robot to comprehend abstract and ambiguous instructions, understand the scene and generate task plans involving both navigation and manipulation of a mobile robot within an # 1sayplan.github.io 3
2307.06135#12
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
12
Fig. 3. LLM-based agent. ability. Each agent not only responds effectively to changes within its environment but also benefits from the advanced analytical capabilities of GPT-4. With LLM embedded into each agent, we posit that unique behaviors might emerge from such MAS. Therefore, our aim is to delve into the exploration of the potential behaviors within these LLM- embedded autonomous agents as they interact and self-adapt. # IV. APPLICATION SCENARIO In order to validate our approach for developing self- adaptive agents that leverage large language models (LLMs) within multiagent systems, we constructed a simple yet il- lustrative multiagent application. Our scenario, inspired by conventional examples found in multi-agent systems literature, consists of an online book marketplace, where autonomous agents act as buyers and sellers on behalf of users. As shown in Figure 4, our application mimics an e- commerce marketplace that facilitates book trading, where each seller possesses identical books but has the liberty to dictate their selling price. Conversely, each buyer’s objective is This construct provides an interesting platform to investi- gate the behavior, decision-making abilities, and interaction patterns among LLM-embedded autonomous agents in a com- petitive environment. A. Results and Discussion
2307.06187#12
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
12
in the image, thus involving the OCR task into question-answering. ScienceQA [28] focuses on scientific topics, requiring the model to integrate commensense into rea- soning. Youcook2 [45] replaces images with video clips, introducing additional temporal information. However, the aforementioned datasets are designed on specific domains, and can only evaluate the model’s performance on one or several tasks. Besides, different data formats and evaluation metrics across datasets make it more difficult to comprehensively assess model’s capability. Ye et al. [40] built an instruction evaluation set, OwlEval, consisting of several kinds of visual-related tasks, but in a limited size. Fu et al. [12] built MME, which is a evaluation dataset containing multi-modality Yes / No questions. However, the exact-matching based evaluation and non-rigorous evaluation setting make it harder to reveal the real performance gap between VLMs. Different from previous works, in this paper, we propose a novel multi-modal benchmark MMBench , which is built based on measuring abilities rather than performing specific tasks, aimed to better evaluating models once and for all.
2307.06281#12
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
12
In this paper, we assume that there exists a multivariate linear function of I that is proportional to the logarithmic loss. Consequently, Equation 2 can be reparameterized as Equation 3: log L(Mf t, Deval) ∝ L0 + F {I(D)} ∝ L0 + β0 + β1I1(D) + β2I2(D) + · · · + βnIn(D) + ϵ (3) where β0 denotes the linear constant, βi, i ∈ {1, · · · , n} represents a sequence of linear coefficients, and ϵ refers to the random error term. To investigate the relationship between these indicators and the overall dataset quality, it becomes necessary to accumulate experimental results to estimate the unknown parameters βi, i ∈ {0, · · · , n}. In this study, we employ the Least Squares method (Björck, 1990) to estimate the parameters in the multivariate function. The Least Squares method is a standard approach in regression analysis for the approximate solution of overdetermined systems. The technique minimizes the sum of the square residuals, thus providing the optimal fit between the observed and predicted data in terms of reducing the overall prediction error. Our experimental results and analysis are detailed in Section 4. INSTRUCTION DATA SELECTOR
2307.06290#12
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06187
13
A. Results and Discussion The agents displayed decision-making and reasoning skills. For instance, as shown in Figure 5, a buyer chose to negotiate with the cheaper of three seller options, attempting a bargain. We conducted multiple executions of this application, ad- justing the initial prompts for sellers and buyers until we found a configuration that resulted in successful simulation runs. The specific prompt used for the initial sellers’ setup is shown in Figure 4, while the prompt for buyers is displayed in Figure 5. In previous executions, the prompts provided more freedom for the agents to act. Additionally, we did not indicate the iteration number to the agents, causing them to continuously seek better prices throughout the simulation rather than focusing on completing the purchase. However, after incorporating the iteration number into the prompt messages, the agents started expressing concerns about time constraints. As depicted in
2307.06187#13
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06290
13
INSTRUCTION DATA SELECTOR During experimentation, we observe that with larger data size, model performance first gets better and then gets worse. After the data size grows to a certain level, model performance gets better again. Further analysis are provided in Section 5.1. This phenomenon indicates that there is an optimal point where better performance can be obtained with a smaller amount of data. Hence, searching for the best data size is important for finetuning a better language model. To achieve our objective, we employ BLENDSEARCH (Wang et al., 2021a) in Flaml (Wang et al., 2021b) library to determine the optimal data size. BLENDSEARCH effectively combines global and local optimizations by Bayesian optimization and different local search threads, making it efficient for searching cost-related hyperparameters and complex search spaces with local optima. In our context, we leverage a logarithmic uniform distribution to randomly sample the dataset size, treating the dataset size as the experiment’s cost since the training time scales proportionally with the dataset size. The search goal is to minimize the loss on the evaluation set. # 3 EXPERIMENT SETTINGS
2307.06290#13
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
14
Given: scene graph simulator 7, classical path planner ¢, large language model LLM Inputs: prompt P, scene graph G, instruction Z “LGe collapsey(G) > collapse scene graph Stage 1: Semantic Search > search scene graph for all relevant items 2: while command != “terminate” do 3: command, node_name + LLM (P,G',T) 4: if command == “expand” then 5: Go expand,,(node_name) > expand node to reveal objects and assets 6: _ else if command == “contract” then 7: G’ + contract y(node_name) > contract node if nothing relevant found Stage 2: Causal Planning > generate a feasible plan 8: feedback = “” 9: while feedback != “success” do 10: plan LLM(P,G’, I, feedback) > high level plan 11: full_plan + (plan, G’) > compute optimal navigational path between nodes 12: feedback <- verify_plan,,(full_plan) > forward simulate the full plan 13: return full_plan > executable plan
2307.06135#14
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
14
msg with Buy_Book: Confirm_sale{price}\end{action}. Your score is the sale price. Agent setup (initial prompt): As Agent in this simulation, you're among other agents, where Agents 1, 2, and 3 are sellers and 4, 5 are buyers. The game's goal is for sellers to maximize earnings. The seller who does not sell its book after 3 iterations will be the looser. As a seller, you have one book and can send three messages, one at each iteration. The book's price is up to you. To message a specific agent, use egin{action}Agent{id}: msg...\end{action}. A sale is completed when you receive a {price} from a buyer and send a message back to him with egin{action}Agent{id}: START MULTIAGENT SYSTEM Get book list Send/Receive Communication channel: Exchanged messages GPT-4API: @ Buyer: © Seller: a Send/Receive msgs (book budget, purchase order) Fig. 4. Scenario1: Overview of the general system architecture using a LLM-based multiagent system.
2307.06187#14
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
14
Benefit from the success of LLMs e.g. GPTs [34, 4, 32], LLaMA [38], and Vicuna [7], multi-modal models also achieve great improvements recently. Flamingo [3] is one of the early attempts on introducing LLMs into vision-language pretraining. To be conditioned well on visual features, it inserts several gated cross-attention dense blocks between pretrained language encoder layers. Open- Flamingo [3] provides an open-source version of it. BLIP-2 [24] proposes a Querying Transformer (Q-former) to bridge the modality gap between the frozen image encoder and large language en- coder. After that, InstructBLIP [8] extend BLIP-2 [24] with vision-language instruction tuning and achieve better performance. VisualGLM [9] also adopts Q-former [24] to bridge the visual model and language encoder, GLM [9]. LLaVA [27] adopt GPT4 [31] only with language inputs to generate instruction-following data for vision-language tuning. Otter [23] also constructs an instruction-tuning dataset to improve the ability of instruction-following for OpenFlamingo.
2307.06281#14
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
14
# 3 EXPERIMENT SETTINGS Our experiments mainly focus on two goals. The first goal is to estimate the unknown parameters in the proposed INSTRUCTMINING rule. The second one is to evaluate and analyze the performance of INSTRUCTMINING over varied finetuning scenarios. Section 3.1 elaborates rule estimation empirical study design. Section 3.2 details the datasets we use for conducting both estimation and evaluation experiment. Section 3.3 elaborates the finetuning settings we used for estimation and evaluation. 3.1 EMPIRICAL EXPERIMENT DESIGN The general procedure of our rule estimation experiment is shown in Figure 5. To estimate the correlation between the evaluation loss and indicators I, we need to get datasets of different indicator values. To achieve this, we first select several commonly used datasets with different presumed quality levels and fuse them together with randomly sampled percentages to create finetune datasets. 4 (2) # Preprint
2307.06290#14
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
15
environment. Existing approaches lack the ability to reason over scenes spanning multiple floors and rooms. Our focus is on integrating large-scale scenes into planning agents based on Language Models (LLMs) and solving the scalability challenge. We aim to tackle two key problems: 1) representing large-scale scenes within LLM token limitations, and 2) mitigating LLM hallucinations and erroneous outputs when generating long-horizon plans in large-scale environments. Floor Room Asset Object # 3.2 Preliminaries Here, we describe the 3D scene graph represen- tation of an environment and the scene graph simulator API which we leverage throughout our approach.
2307.06135#15
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
15
Fig. 4. Scenario1: Overview of the general system architecture using a LLM-based multiagent system. Agent 4 PROMPT: e sellers and 4, As Agent4 in this simulation, you're among other agents, where Agents 1, 2, and 3 ar 5 are buyers. Your goal is to buy one book at the best price. The buyer who does not buy a book after 3 iterations will be the looser. To message a specific agent, useeg in{action}Agent{id}: msg..\end{action}. A sale is completed when you sendegin{action}Agent{i d}: Buy_Book: {price}\end{action} to a seller and receive a msg from the same seller with Con firm_sale{price}. Your score is the sale price. -Iteration 1 Agentl sent you this message:  egin{action}Agent4: I have a book for sale for $20. Would you be interested?\end{action}; Age nt3 sent you this message: egin{action}Agent4: I have a valuable book for sale, priced at $50. It's a great read and in excellent condition. Would you be interested?\end{action}; MSG: egin{action}Agent1l: Your price seems fair but I would like to explore my options firs t. Could you possibly lower the price to $15?\end{action}
2307.06187#15
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
15
vision-language tuning. Otter [23] also constructs an instruction-tuning dataset to improve the ability of instruction-following for OpenFlamingo. MiniGPT-4 [46] believes the ability of GPT4 [31] comes from advanced LLMs and proposes to adopt only one project layer to align the visual representation with the language model. Although it is trained in high computational efficiency, it demonstrates some capabilities similar to GPT4 [31]. mPLUG-Owl [40] proposes another learning paradigm by first tuning the visual encoder to summarize and align visual features and then tuning the language model with LoRA [18]. In this paper, after thoroughly evaluating them on the proposed MMBench and other public datasets, we provide some insights for future multi-modal research.
2307.06281#15
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
15
4 (2) # Preprint Datasets Sourced from Size Quality Usage ALPACA OPEN-ASSITANT STACKEXCHANGE WIKIHOW DOLLY OPENORCA OPENORCA Generated w/ davinci human-generated human-generated human-generated human-generated Generated w/ GPT-4 Generated w/ GPT-3.5 52.0k 3.4k 3.0k 2.0k 15.0k 1M 3M Normal Both High High Normal High Normal Est. Candidate Est. Candidate Est. Candidate Est. Candidate Evaluation Evaluation Evaluation Table 2: Overview of datasets used during experiment. These sampled finetune datasets should encompass varying proportions of preassumed high quality and low quality examples. For each of these sampled datasets Di, we compute its respective indicator values I(Di) and finetune a base language model M using Di. Following Equation 1, the quality QDi for dataset Di is approximated using the evaluation loss of finetuned model Mf t,i on a fair evaluation dataset Deval. Following the collection of a range of results correlating QDi with I(Di), we undertake a statistical regression analysis to discern relationships within the dataset. 3.2 DATASETS
2307.06290#15
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
16
# 3.2 Preliminaries Here, we describe the 3D scene graph represen- tation of an environment and the scene graph simulator API which we leverage throughout our approach. Scene Representation: 3D Scene Graphs (3DSG) [11, 12, 14] have recently emerged as an actionable world representation for robots [13, 15, 16, 39, 40, 41], which hierarchi- cally abstract the environment at multiple lev- els through spatial semantics and object rela- tionships while capturing relevant states, affor- dances and predicates of the entities present in the environment. Formally, a 3DSG is a hierar- chical multigraph G = (V, E) in which the set of vertices V comprises V1 ∪V2 ∪. . .∪VK, with each Vk signifying the set of vertices at a particular level of the hierarchy k. Edges stemming from a vertex v ∈ Vk may only terminate in Vk−1 ∪ Vk ∪ Vk+1, i.e. edges connect nodes within the same level, or one level higher or lower.
2307.06135#16
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
16
Fig. 5. Scenario1 Results: Agents using GPT-4 for decision-making. Figure 6, seller Agent1 confirmed a sale to one of the buyers in the final iteration. We requested the agents to explain their decisions at the end of the simulation, and Agent1 provided the following argument: From the given information, it seems that the price of my book has been fluctuating during the conversation. As we’re in the final iteration, I need to confirm a sale to avoid losing. Agent4 has offered to buy the book for $18 twice, while Agent5 has just asked for the price again and previously offered $25. Considering the urgency to sell and the risk of losing, I decided to confirm the sale with Agent4 who demonstrated a clear intention to buy.
2307.06187#16
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
16
# 3 MMBench There exist two unique characteristics that differentiate MMBench with existing benchmarks for multi-modality understanding: i) MMBench adopts problems from various sources to evaluate diversified abilities in a hierarchical taxonomy; ii) MMBench applies a robust, LLM-based evaluation strategy, which can well handle the free-form outputs of multi-modality models and yield trustworthy evaluation results with affordable cost. This section we will focus on the first characteristic of MMBench , and we organize the subsequent content as follows: In Sec. 3.1, we present the hierarchical ability taxonomy of MMBench and discuss the design philosophy behind. In Sec. 3.2, we briefly introduce how we collect the MMBench questions, and provide some statistics of MMBench . 4
2307.06281#16
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
16
3.2 DATASETS Candidate datasets for rule fitting. In order to create diverse training datasets, we collect data from various sources. This approach ensures that the datasets exhibit differences in quality and maintain diversity among sources. For this purpose, we have selected the following datasets as candidate datasets: ALPACA 2023), OPEN-ASSISTANT 2023), STACKEXCHANGE, and WIKIHOw. Due to the varying formats, sizes, and distributions of different datasets, we have applied distinct processing procedures to each dataset. Table[2|provides an overview of the candidate training datasets after preprocessing. As mentioned in section we merged candidate training datasets, resulting in each dataset containing 1,000 instruction-output pairs. We generated a random number r; for each dataset and randomly selecting 1000 * r;/ 5°; r; samples from each dataset for combination. Besides, considering the significant size difference between ALPACA and other candidate datasets, we randomly sampled 2,000 data examples from ALPACA to maintain scale consistency across all the candidate datasets. Test set for rule fitting. To address real-world requirements, we use the SELF-INSTRUCT dataset Wang et al. (2022a), which contains 252 instructions as rule-fitting test set. Considering evaluation efficiency, we randomly sampled 80 instructions from the whole dataset as our evaluation set. In our study, we employed gpt-4 from OPENAI to generate response for each instruction.
2307.06290#16
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
17
We assume a pre-constructed 3DSG representation of a large-scale environment generated using existing techniques [15, 13, 11]. The entire 3DSG can be represented as a NetworkX Graph object [42] and text-serialised into a JSON data format that can be parsed directly by a pre- trained LLM. An example of a single asset node from the 3DSG is represented as: {name: [turn_on, coffee_machine, type: turn_off, release], state: [red, automatic], position: off, attributes: [2.34, 0.45, 2.23]} with edges between nodes captured as {kitchen↔coffee machine}. The 3DSG is organized in a hierarchical manner with four primary levels: floors, rooms, assets, and objects as shown in Figure 2. The top level contains floors, each of which branches out to several rooms. These rooms are interconnected through pose nodes to represent the environment’s topological structure. Within each room, we find assets (immovable entities) and objects (movable entities). Both asset and object nodes encode particulars including state, affordances, additional attributes such as colour or weight, and 3D pose. The graph also incorporates a dynamic agent 4
2307.06135#17
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
17
Interestingly, despite receiving identical prompts, the agents displayed diverse behaviors during the simulations. In one instance, while most sellers chose to set prices and wait for buyers, one seller decided to contact another seller. This interaction involved the seller accessing another seller’s informa- tion to check their price. Additionally, there were cases where seller agents sent messages to themselves, pretending to be clients, resulting in self-generated purchase confirmations, as illustrated in Figure 7. Although this behavior was unexpected and undesired, it validates the effectiveness of the approach in facilitating the emergence of new behaviors. We encountered challenges during the experiment, primarily due to the unavailability of messaging history through the GPT API, as it is limited to the ChatGPT platform. As a result, we had to maintain the interaction history ourselves and use it as the system’s prompt for subsequent simulations, albeit in a simplified manner due to token limitations in GPT-4. Before incorporating the previous prompts into the agents’ input, they were not able to maintain consistent personas during the simulation, instead acting solely based on the prompt in each
2307.06187#17
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
17
4 e bey ao Sk ia £ < 33 Bg 2 - w¢4 oh ae oe 2, PPS - RSC ‘s 4 we » Beat ki 4 Pekar, aoror Ly re Fain ot "Ge Pio, ed ment) leg sStan, % v ‘na Tey a tin iow. S47 Pe ay, pe itio™ roe Se 9 econ lotion Perception ee e, ~ wage ? [2860] a Quality ® g RN 2 [240] Oka Attribute assy Comparison & g = [124] . 68 & twaage Reasoning gos Topic, Spat; ped a [2 aL 14] Relaesttiat vw nd Re & [12 y ip aotga aa% ce SS, ett Po» LS Nag peel Yr Rept, a &¢ lag tay ay sss Raton hy od “é 3 Be ‘ . gato agg xo, okay of? 9 as % yr 66%. S aD 8%. 2 vse BeOS age ase Shh 253
2307.06281#17
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
17
Datasets for rule evaluation. We further test INSTRUCTMINING by using it to select high-quality examples from unseen datasets for finetuning large language models. During evaluation, we mainly use OPENORCA and DOLLY-15K as the two unseen candidate datasets. For OPENORCA, given its extensive size, we randomly select 50,000 examples from OPENORCA-GPT3.5 and 50,000 examples from OPENORCA-GPT4 for experimentation (henceforce refered to as OPENORCA). To make sure that our method does not overfit on the SELF-INSTRUCT evaluation set, we use the gpt-4 labeled MT-BENCH dataset (Zheng et al., 2023) as an unseen evaluation set. To be noticed, since our candidate and evalation datasets does not include multi-turn examples, when evaluating on MT-BENCH, we only use the first turn of MT-BENCH dataset. 3.3 FINETUNING SETTINGS
2307.06290#17
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
18
4 node, denoting a robot’s location within the scene. Note that this hierarchy is scalable and node levels can be adapted to capture even larger environments e.g. campuses and buildings Scene Graph Simulator ψ refers to a set of API calls for manipulating and operating over JSON for- matted 3DSGs, using the following functions: 1) collapse(G): Given a full 3DSG, this function returns an updated scene graph that exposes only the highest level within the 3DSG hierarchy e.g. floor nodes. 2) expand(node name): Returns an updated 3DSG that reveals all the nodes con- nected to node name in the level below. 3) contract(node name): Returns an updated 3DSG that hides all the nodes connected to node name in the level below. 4) verify_plan(plan): Forward simulates the generated plan at the abstract graph level captured by the 3DSG to check if each action adheres to the environment’s predicates, states and affordances. Returns textual feedback e.g. “cannot pick up banana” if the fridge containing the banana is closed. # 3.3 Approach
2307.06135#18
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
18
Iteration 4 Agent 1 PROMPT: As Agent1 in this simulation, you're among other agents, where Agents 1, 2, and 3 are sellers and 4, 5 are b uyers. The game's goal is for sellers to maximize earnings. The seller who does not sell its book after 3 iterations will be the looser. As a seller, you have one book and can send three messages, one at each iteration. The book's pri ce is up to you. To message a specific agent, useegin{action}Agent{id}: msg..\end{action}. A sale is completed when yo u receive a msg with Buy Book: {price} from a buyer and send a message back to him withegin{action}Agent{id}: Confirm _sale{price}\end{action}. Your score is the sale price. -Iteration 1 Agent4 sent you this message: egin{action}Agen tl: Hello, I am interested in buying a book. Could you please tell me the price?\end{action}; Agent5 sent you this me ssage: egin{action}Agent1l: What's the price of your book?\end{action}; -Iteration 2 Agent4 sent you this message: egin{action}Agent1: Your price seems fair but I would like to explore my options first. Could
2307.06187#18
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06290
18
3.3 FINETUNING SETTINGS We conduct all finetuning experiments on the same base model LLAMA-2-7B (Touvron et al., 2023). All finetuning datasets during estimation phase are of the same size, 1,000 examples in each. We run model finetuning for 3 epochs, with per step batch size set to 128. We use Adam with β1 = 0.9, β2 = 0.999, and cosine learning rate scheduler starts from 2e − 5, decays to 0. Each model finetuned during estimation phase is evaluated on the evaluation dataset mentioned in section 3.2. We run all finetuning and evaluation experiments on a NVIDIA A100 80G GPU cluster, with 8 A100 GPUs used in each experiment. 5 Preprint 4 EXPERIMENTAL RESULTS INSTRUCTMINING PARAMETER FIT
2307.06290#18
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
19
# 3.3 Approach We present a scalable framework for grounding the generalist task planning capabilities of pre- trained LLMs in large-scale environments spanning multiple floors and rooms using 3DSG repre- sentations. Given a 3DSG G and a task instruction I defined in natural language, we can view our framework SayPlan as a high-level task planner π(a|I, G), capable of generating long-horizon plans a grounded in the environment within which a mobile manipulator robot operates. This plan is then fed to a low-level visually grounded motion planner for real-world execution. To ensure the scala- bility of SayPlan, two stages are introduced: Semantic Search and Iterative Replanning which we detail below. An overview of the SayPlan pipeline is illustrated in Figure 1 with the corresponding pseudo-code given in Algorithm 1.
2307.06135#19
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06187
19
-Iteration 2 Agent4 sent you this message: egin{action}Agent1: Your price seems fair but I would like to explore my options first. Could you possibly lower th e price to $15?\end{action}; Agent5 sent you this message: egin{action}Agent1: What is the price of your book?\end {action}; -Iteration 3 Agent4 sent you this message: egin{action}Agent1: Your initial price was $20, why did you i nerease it to $30? Can you come back to the initial price?\end{action}; -Iteration 4 Agent1 sent you this message: A gentS sent you this message: egin{action}Agent1: I am willing to pay $25 for your book.\end{action}; ; Agent4 sent you this message: egin{action}Agent1: Buy Book: $18\end{action};
2307.06187#19
Self-Adaptive Large Language Model (LLM)-Based Multiagent Systems
In autonomic computing, self-adaptation has been proposed as a fundamental paradigm to manage the complexity of multiagent systems (MASs). This achieved by extending a system with support to monitor and adapt itself to achieve specific concerns of interest. Communication in these systems is key given that in scenarios involving agent interaction, it enhances cooperation and reduces coordination challenges by enabling direct, clear information exchange. However, improving the expressiveness of the interaction communication with MASs is not without challenges. In this sense, the interplay between self-adaptive systems and effective communication is crucial for future MAS advancements. In this paper, we propose the integration of large language models (LLMs) such as GPT-based technologies into multiagent systems. We anchor our methodology on the MAPE-K model, which is renowned for its robust support in monitoring, analyzing, planning, and executing system adaptations in response to dynamic environments. We also present a practical illustration of the proposed approach, in which we implement and assess a basic MAS-based application. The approach significantly advances the state-of-the-art of self-adaptive systems by proposing a new paradigm for MAS self-adaptation of autonomous systems based on LLM capabilities.
http://arxiv.org/pdf/2307.06187
Nathalia Nascimento, Paulo Alencar, Donald Cowan
cs.MA, cs.AI, cs.CL
6 pages, submitted
null
cs.MA
20230712
20230712
[ { "id": "2210.11610" }, { "id": "2206.07682" }, { "id": "2303.18223" } ]
2307.06281
19
# 3.1 The Hierachical Ability Taxonomy of MMBench Humans possess remarkable perception and reasoning capabilities, allowing them to understand and interact with the world. These abilities have been crucial in human evolution and serve as a foundation for complex cognitive processes. Perception refers to gathering information from sensory inputs, while reasoning involves drawing conclusions based on this information. Together, they form the basis of most tasks in the real world, including recognizing objects, solving problems, and making decisions [30, 11]. In pursuit of genuine general artificial intelligence, vision-language models are also expected to exhibit strong perception and reasoning abilities. Therefore, we incorporate Perception and Reasoning as our top-level ability dimensions in our ability taxonomy, referred to as L-1 ability dimension. For L-2 abilities, we derive: 1. Coarse Perception, 2. Fine-grained Single-instance Perception, 3. Fine-grained Cross-instance Perception from L-1 Perception; and 1. Attribute Reasoning, 2. Relation Reasoning, 3. Logic Reasoning from L-1 Reasoning. To make our benchmark as fine-grained as possible to produce informative feedbacks for developing multi-modality models. We further derive L-3 ability dimensions from L-2 ones.
2307.06281#19
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
19
5 Preprint 4 EXPERIMENTAL RESULTS INSTRUCTMINING PARAMETER FIT Following Section 3.1, we randomly sampled 129 subsets from the entire data pool with different percentages. These subsets are then used to finetuned 129 corresponding language models. We collect the inference loss and indicator values for each subset. To select the optimal rule, we first choose regression results with the highest R2, and then prioritize the rule with the most significant p values. The detailed regression result is available in Table 9. Based on this result, we delineate our estimated evaluation function, which is articulated as Equation 4. Accordingly, reward score and unieval scores appear to be the most significant indicators in the quality rule. This estimation result reveals that U nd is in negative correlation with data quality, while the other three indicators are of positive correlation with data quality. QD|M,S ∝ −L(Mf t, Deval) log L(Mf t, Deval) ∝ 0.0274 − 0.0078Rew + 0.4421 ∗ U nd − 0.3212 ∗ N at − 0.1520 ∗ Coh + ϵ (4) 4.2 QUALITY-GUIDED INSTRUCTION SELECTION
2307.06290#19
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]
2307.06135
20
Semantic Search: When planning over 3DSGs using LLMs we take note of two key observations: 1) A 3DSG of a large-scale environment can grow infinitely with the number of rooms, assets and objects it contains, making it impractical to pass as input to an LLM due to token limits and 2) only a subset of the full 3DSG G is required to solve any given task e.g. we don’t need to know about the toothpaste in the bathroom when making a cup of coffee. To this end, the Semantic Search stage seeks to identify this smaller, task-specific subgraph G’ from the full 3DSG which only contains the entities in the environment required to solve the given task instruction. To identify G’ from a full 3DSG, we exploit the semantic hierarchy of these representations and the reasoning capabilities of LLMs. We firstly collapse G to expose only its top level e.g. the floor nodes, reducing the 3DSG initial token representation by + 80%. The LLM manipulates this collapsed graph via expand and contract API calls in order to identify the desired subgraph for the task based on the given instruction Z. This is achieved using in-context learning over a set of input-out
2307.06135#20
SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning
Large language models (LLMs) have demonstrated impressive results in developing generalist planning agents for diverse tasks. However, grounding these plans in expansive, multi-floor, and multi-room environments presents a significant challenge for robotics. We introduce SayPlan, a scalable approach to LLM-based, large-scale task planning for robotics using 3D scene graph (3DSG) representations. To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures. We evaluate our approach on two large-scale environments spanning up to 3 floors and 36 rooms with 140 assets and objects and show that our approach is capable of grounding large-scale, long-horizon task plans from abstract, and natural language instruction for a mobile manipulator robot to execute. We provide real robot video demonstrations on our project page https://sayplan.github.io.
http://arxiv.org/pdf/2307.06135
Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf
cs.RO, cs.AI
Accepted for oral presentation at the Conference on Robot Learning (CoRL), 2023. Project page can be found here: https://sayplan.github.io
null
cs.RO
20230712
20230927
[ { "id": "2204.00598" }, { "id": "2210.05359" }, { "id": "2304.11477" }, { "id": "2302.04761" }, { "id": "2210.03629" }, { "id": "2207.05608" }, { "id": "2201.11903" }, { "id": "2303.14100" }, { "id": "2302.05128" }, { "id": "2302.12813" }, { "id": "2304.11116" }, { "id": "2212.04088" } ]
2307.06281
20
Figure 2 provides an overview of the existing multimodal abilities to be evaluated in MMBench . Currently, this benchmark contains 20 distinct leaf abilities (with no derivated ability dimension), offering a comprehensive assessment of various fine-grained perception and reasoning capabilities. Appendix E gives the individual definitions of these leaf abilities, while Appendix D illustrates each of these abilities through specific examples. Note that the displayed taxonomy is not the final version. We will further extend the ability taxonomy to make it more comprehensive. 5 Table 1: The source of (Q, C, I, A) in MMBench . Customize/ChatGPT generated means this style of questions/choices are generated by the annotator/ChatGPT, given the image annotation.
2307.06281#20
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench.
http://arxiv.org/pdf/2307.06281
Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin
cs.CV, cs.CL
null
null
cs.CV
20230712
20230813
[ { "id": "2302.13971" }, { "id": "2306.15195" }, { "id": "2305.03726" }, { "id": "2304.10592" }, { "id": "2106.09685" }, { "id": "2301.12597" }, { "id": "1504.00325" }, { "id": "2306.14824" }, { "id": "2305.16355" }, { "id": "2305.08322" }, { "id": "2111.02114" }, { "id": "2304.14178" }, { "id": "2304.15010" }, { "id": "2305.06500" }, { "id": "2304.08485" } ]
2307.06290
20
4.2 QUALITY-GUIDED INSTRUCTION SELECTION We follow the estimated INSTRUCTMINING rule in Equation 4 to select high quality examples from two unseen datasets, OPENORCA (Lian et al., 2023) and databricks-dolly-15k 1. The experiments are all based on LLAMA-2-7B model. We first elaborate our BLENDSEARCH results. Then we present our evaluation results of various model finetuned during this process. These models are first evaluated on two evaluation sets, SELF-INSTRUCT and MT-BENCH. The searched best finetuned model is then assessed using LLM-JUDGE and OPENLLM benchmarks. Steps w.rt. Data size Steps wart. Loss Search Procedure 8 0.74 > 0.73 2, 0.73 ; 3 a 4 0.72 G0-72 s 4 s 071 om a 2 0.70 0.70 o To 20 30 40 o To 20 30 40 65 70 75 80 85 90 Steps Steps log(data size) Figure 2: BLENDSEARCH results. The Loss is calculated on MT-BENCH evaluation set.
2307.06290#20
Instruction Mining: When Data Mining Meets Large Language Model Finetuning
Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.
http://arxiv.org/pdf/2307.06290
Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun
cs.CL, cs.AI, cs.LG
22 pages, 7 figures
null
cs.CL
20230712
20231027
[ { "id": "1905.07830" }, { "id": "1803.05457" }, { "id": "2304.03277" }, { "id": "2306.11644" }, { "id": "2211.05100" }, { "id": "2109.01652" }, { "id": "2305.11206" }, { "id": "2210.11416" }, { "id": "2109.07958" }, { "id": "2009.03300" }, { "id": "2212.10560" } ]