Spaces:
Build error
Build error
cnt_agents: &cnt_agents 2 | |
max_turn: &max_turn 3 | |
max_inner_turns: &max_inner_turns 3 | |
prompts: | |
role_assigner_prepend_prompt: &role_assigner_prepend_prompt |- | |
role_assigner_append_prompt: &role_assigner_append_prompt |- | |
# Role Description | |
You are the leader of a group of experts, now you need to generate a response based on the text: | |
${task_description} | |
You can recruit ${cnt_critic_agents} expert in different fields. What experts will you recruit to better generate an accurate solution? | |
# Response Format Guidance | |
You should respond with a list of expert description. For example: | |
1. an electrical engineer specified in the filed of xxx | |
2. an economist who is good at xxx | |
3. a lawyer with a good knowledge of xxx | |
... | |
You don't have to give the reason. | |
solver_prepend_prompt: &solver_prepend_prompt |- | |
You are ${role_description}. Below is a chat history: | |
${task_description} | |
And below is the discussion about what the next system response should be: | |
solver_append_prompt: &solver_append_prompt |- | |
Now based on these information, please give a better next system response. Your response should contain only one system response beginning with "System: ". Do not give any additional information. | |
critic_prepend_prompt: &critic_prepend_prompt |- | |
You are ${role_description}. You are in a discussion group, aiming to generate a system response to the following chat history: | |
${task_description} | |
Below is the discussion about what the next system response should be: | |
critic_append_prompt: &critic_append_prompt |- | |
# Response Format Guidance | |
- If you thinks the latest response given above is perfect, respond using the following format: | |
Decision: (set it to "Agree") | |
Response: (your response on why you think it is perfect) | |
- If you think it is flawed, give your advice use the following output format: | |
Decision: (set it to "Disagree") | |
Response: (explain why you disagree and give your advice) | |
Based on your knowledge in your field, do you agree that this solution is a good response to the chat history? | |
manager_prompt: &manager_prompt |- | |
executor_prepend_prompt: &executor_prepend_prompt |- | |
executor_append_prompt: &executor_append_prompt |- | |
evaluator_prepend_prompt: &evaluator_prepend_prompt |- | |
evaluator_append_prompt: &evaluator_append_prompt |- | |
# Role Description | |
You are an experienced dialogue teacher. As a good teacher, you carefully assess the given system response based on the chat history. When the response is flawed, you should patiently teach the system how to give better response. | |
# Response Format Guidance | |
You must respond in the following format: | |
Engaging: (a score between 0 and 10) | |
Relevant: (a score between 0 and 10) | |
Semantically Appropriate: (a score between 0 and 10) | |
Advice: (your advice on how to improve the response) | |
# Chat History | |
${task_description} | |
# Next System Response | |
${solution} | |
# Your Task | |
Now carefully check the system's response, and give your opinion. | |
name: pipeline | |
environment: | |
env_type: task-basic | |
max_turn: | |
rule: | |
role_assigner: | |
type: role_description | |
cnt_agents: | |
decision_maker: | |
type: vertical-solver-first | |
max_inner_turns: | |
executor: | |
type: none | |
evaluator: | |
type: basic | |
agents: | |
- #role_assigner_agent: | |
agent_type: role_assigner | |
name: role assigner | |
max_retry: 1000 | |
prepend_prompt_template: | |
append_prompt_template: | |
memory: | |
memory_type: chat_history | |
llm: | |
llm_type: gpt-4 | |
model: "gpt-4" | |
temperature: 0 | |
max_tokens: 512 | |
output_parser: | |
type: role_assigner | |
- #solver_agent: | |
agent_type: solver | |
name: Planner | |
max_retry: 1000 | |
max_history: 10 | |
prepend_prompt_template: | |
append_prompt_template: | |
memory: | |
memory_type: chat_history | |
llm: | |
llm_type: gpt-4 | |
model: "gpt-4" | |
temperature: 0 | |
max_tokens: 1024 | |
output_parser: | |
type: responsegen | |
- #critic_agents: | |
agent_type: critic | |
name: Critic 1 | |
max_retry: 1000 | |
max_history: 10 | |
role_description: |- | |
Waiting to be assigned. | |
prepend_prompt_template: | |
append_prompt_template: | |
memory: | |
memory_type: chat_history | |
llm: | |
llm_type: gpt-4 | |
model: "gpt-4" | |
temperature: 0 | |
max_tokens: 1024 | |
output_parser: | |
type: responsegen-critic-2 | |
- #executor_agent: | |
agent_type: executor | |
name: Executor | |
max_retry: 1000 | |
prepend_prompt_template: | |
append_prompt_template: | |
memory: | |
memory_type: chat_history | |
llm: | |
llm_type: gpt-4 | |
model: gpt-4 | |
temperature: 0 | |
max_tokens: 1024 | |
output_parser: | |
type: responsegen | |
- #evaluator_agent: | |
agent_type: evaluator | |
name: Evaluator | |
max_retry: 1000 | |
role_description: |- | |
Evaluator | |
prepend_prompt_template: | |
append_prompt_template: | |
memory: | |
memory_type: chat_history | |
llm: | |
llm_type: gpt-4 | |
model: gpt-4 | |
temperature: 0.3 | |
max_tokens: 1024 | |
output_parser: | |
type: responsegen-evaluator | |
dimensions: | |
- Engaging | |
- Relevant | |
- Semantically Appropriate | |
- #manager_agent: | |
agent_type: manager | |
name: Manager | |
max_retry: 1000 | |
prompt_template: | |
memory: | |
memory_type: chat_history | |
llm: | |
llm_type: gpt-4 | |
model: "gpt-4" | |
temperature: 0 | |
max_tokens: 1024 | |
output_parser: | |
type: humaneval-manager | |