Enhance z3ed Agent Roadmap and CLI Reference with New Tool Commands and Chat Features
This commit is contained in:
@@ -15,47 +15,72 @@ ConversationalAgentService::ConversationalAgentService() {
|
||||
|
||||
absl::StatusOr<ChatMessage> ConversationalAgentService::SendMessage(
|
||||
const std::string& message) {
|
||||
// 1. Add user message to history.
|
||||
history_.push_back({ChatMessage::Sender::kUser, message, absl::Now()});
|
||||
|
||||
// 2. Get response from the AI service using the full history.
|
||||
auto response_or = ai_service_->GenerateResponse(history_);
|
||||
if (!response_or.ok()) {
|
||||
return absl::InternalError(absl::StrCat("Failed to get AI response: ",
|
||||
response_or.status().message()));
|
||||
if (message.empty() && history_.empty()) {
|
||||
return absl::InvalidArgumentError(
|
||||
"Conversation must start with a non-empty message.");
|
||||
}
|
||||
|
||||
const auto& agent_response = response_or.value();
|
||||
|
||||
// For now, combine text and commands for display.
|
||||
// In the future, the TUI/GUI will handle these differently.
|
||||
std::string response_text = agent_response.text_response;
|
||||
if (!agent_response.commands.empty()) {
|
||||
response_text += "\n\nCommands:\n" + absl::StrJoin(agent_response.commands, "\n");
|
||||
if (!message.empty()) {
|
||||
history_.push_back({ChatMessage::Sender::kUser, message, absl::Now()});
|
||||
}
|
||||
|
||||
// If the agent requested a tool call, dispatch it.
|
||||
if (!agent_response.tool_calls.empty()) {
|
||||
for (const auto& tool_call : agent_response.tool_calls) {
|
||||
auto tool_result_or = tool_dispatcher_.Dispatch(tool_call);
|
||||
if (tool_result_or.ok()) {
|
||||
// Add the tool result to the history and send back to the AI.
|
||||
history_.push_back({ChatMessage::Sender::kAgent, tool_result_or.value(), absl::Now()});
|
||||
return SendMessage(""); // Re-prompt the AI with the new context.
|
||||
} else {
|
||||
// Handle tool execution error.
|
||||
return absl::InternalError(absl::StrCat("Tool execution failed: ", tool_result_or.status().message()));
|
||||
constexpr int kMaxToolIterations = 4;
|
||||
for (int iteration = 0; iteration < kMaxToolIterations; ++iteration) {
|
||||
auto response_or = ai_service_->GenerateResponse(history_);
|
||||
if (!response_or.ok()) {
|
||||
return absl::InternalError(absl::StrCat(
|
||||
"Failed to get AI response: ", response_or.status().message()));
|
||||
}
|
||||
|
||||
const auto& agent_response = response_or.value();
|
||||
|
||||
if (!agent_response.tool_calls.empty()) {
|
||||
bool executed_tool = false;
|
||||
for (const auto& tool_call : agent_response.tool_calls) {
|
||||
auto tool_result_or = tool_dispatcher_.Dispatch(tool_call);
|
||||
if (!tool_result_or.ok()) {
|
||||
return absl::InternalError(absl::StrCat(
|
||||
"Tool execution failed: ", tool_result_or.status().message()));
|
||||
}
|
||||
|
||||
const std::string& tool_output = tool_result_or.value();
|
||||
if (!tool_output.empty()) {
|
||||
history_.push_back(
|
||||
{ChatMessage::Sender::kAgent, tool_output, absl::Now()});
|
||||
}
|
||||
executed_tool = true;
|
||||
}
|
||||
|
||||
if (executed_tool) {
|
||||
// Re-query the AI with updated context.
|
||||
continue;
|
||||
}
|
||||
}
|
||||
|
||||
std::string response_text = agent_response.text_response;
|
||||
if (!agent_response.reasoning.empty()) {
|
||||
if (!response_text.empty()) {
|
||||
response_text.append("\n\n");
|
||||
}
|
||||
response_text.append("Reasoning: ");
|
||||
response_text.append(agent_response.reasoning);
|
||||
}
|
||||
if (!agent_response.commands.empty()) {
|
||||
if (!response_text.empty()) {
|
||||
response_text.append("\n\n");
|
||||
}
|
||||
response_text.append("Commands:\n");
|
||||
response_text.append(absl::StrJoin(agent_response.commands, "\n"));
|
||||
}
|
||||
|
||||
ChatMessage chat_response = {ChatMessage::Sender::kAgent, response_text,
|
||||
absl::Now()};
|
||||
history_.push_back(chat_response);
|
||||
return chat_response;
|
||||
}
|
||||
|
||||
ChatMessage chat_response = {ChatMessage::Sender::kAgent, response_text,
|
||||
absl::Now()};
|
||||
|
||||
// 3. Add agent response to history.
|
||||
history_.push_back(chat_response);
|
||||
|
||||
return chat_response;
|
||||
return absl::InternalError(
|
||||
"Agent did not produce a response after executing tools.");
|
||||
}
|
||||
|
||||
const std::vector<ChatMessage>& ConversationalAgentService::GetHistory() const {
|
||||
|
||||
@@ -3,6 +3,7 @@
|
||||
#include <iostream>
|
||||
#include <sstream>
|
||||
|
||||
#include "absl/strings/match.h"
|
||||
#include "absl/strings/str_format.h"
|
||||
#include "cli/handlers/agent/commands.h"
|
||||
|
||||
@@ -13,9 +14,18 @@ namespace agent {
|
||||
absl::StatusOr<std::string> ToolDispatcher::Dispatch(
|
||||
const ToolCall& tool_call) {
|
||||
std::vector<std::string> args;
|
||||
bool has_format = false;
|
||||
for (const auto& [key, value] : tool_call.args) {
|
||||
args.push_back(absl::StrFormat("--%s", key));
|
||||
args.push_back(value);
|
||||
if (absl::EqualsIgnoreCase(key, "format")) {
|
||||
has_format = true;
|
||||
}
|
||||
}
|
||||
|
||||
if (!has_format) {
|
||||
args.push_back("--format");
|
||||
args.push_back("json");
|
||||
}
|
||||
|
||||
// Capture stdout
|
||||
|
||||
@@ -1,19 +1,117 @@
|
||||
#include "cli/service/ai/ai_service.h"
|
||||
#include "cli/service/agent/conversational_agent_service.h"
|
||||
|
||||
#include <algorithm>
|
||||
#include <cctype>
|
||||
|
||||
#include "absl/strings/ascii.h"
|
||||
#include "absl/strings/match.h"
|
||||
#include "absl/strings/numbers.h"
|
||||
#include "absl/strings/str_format.h"
|
||||
|
||||
namespace yaze {
|
||||
namespace cli {
|
||||
|
||||
namespace {
|
||||
|
||||
std::string ExtractRoomId(const std::string& normalized_prompt) {
|
||||
size_t hex_pos = normalized_prompt.find("0x");
|
||||
if (hex_pos != std::string::npos) {
|
||||
std::string hex_value;
|
||||
for (size_t i = hex_pos; i < normalized_prompt.size(); ++i) {
|
||||
char c = normalized_prompt[i];
|
||||
if (std::isxdigit(static_cast<unsigned char>(c)) || c == 'x') {
|
||||
hex_value.push_back(c);
|
||||
} else {
|
||||
break;
|
||||
}
|
||||
}
|
||||
if (hex_value.size() > 2) {
|
||||
return hex_value;
|
||||
}
|
||||
}
|
||||
|
||||
// Fallback: look for decimal digits, then convert to hex string.
|
||||
std::string digits;
|
||||
for (char c : normalized_prompt) {
|
||||
if (std::isdigit(static_cast<unsigned char>(c))) {
|
||||
digits.push_back(c);
|
||||
} else if (!digits.empty()) {
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (!digits.empty()) {
|
||||
int value = 0;
|
||||
if (absl::SimpleAtoi(digits, &value)) {
|
||||
return absl::StrFormat("0x%03X", value);
|
||||
}
|
||||
}
|
||||
|
||||
return "0x000";
|
||||
}
|
||||
|
||||
} // namespace
|
||||
|
||||
absl::StatusOr<AgentResponse> MockAIService::GenerateResponse(
|
||||
const std::string& prompt) {
|
||||
AgentResponse response;
|
||||
if (prompt == "Place a tree") {
|
||||
response.text_response = "Sure, I can do that. Here is the command:";
|
||||
response.commands.push_back("overworld set-tile 0 10 20 0x02E");
|
||||
response.reasoning = "The user asked to place a tree, so I generated the appropriate `set-tile` command.";
|
||||
} else {
|
||||
response.text_response = "I'm sorry, I don't understand that prompt. Try 'Place a tree'.";
|
||||
const std::string normalized = absl::AsciiStrToLower(prompt);
|
||||
|
||||
if (normalized.empty()) {
|
||||
response.text_response =
|
||||
"Let's start with a prompt about the overworld or dungeons.";
|
||||
return response;
|
||||
}
|
||||
|
||||
if (absl::StrContains(normalized, "place") &&
|
||||
absl::StrContains(normalized, "tree")) {
|
||||
response.text_response =
|
||||
"Sure, I can do that. Here's the command to place a tree.";
|
||||
response.commands.push_back("overworld set-tile --map 0 --x 10 --y 20 --tile 0x02E");
|
||||
response.reasoning =
|
||||
"The user asked to place a tree tile16, so I generated the matching set-tile command.";
|
||||
return response;
|
||||
}
|
||||
|
||||
if (absl::StrContains(normalized, "list") &&
|
||||
absl::StrContains(normalized, "resource")) {
|
||||
std::string resource_type = "dungeon";
|
||||
if (absl::StrContains(normalized, "overworld")) {
|
||||
resource_type = "overworld";
|
||||
} else if (absl::StrContains(normalized, "sprite")) {
|
||||
resource_type = "sprite";
|
||||
} else if (absl::StrContains(normalized, "palette")) {
|
||||
resource_type = "palette";
|
||||
}
|
||||
|
||||
ToolCall call;
|
||||
call.tool_name = "resource-list";
|
||||
call.args.emplace("type", resource_type);
|
||||
response.text_response =
|
||||
absl::StrFormat("Fetching %s labels from the ROM...", resource_type);
|
||||
response.reasoning =
|
||||
"Using the resource-list tool keeps the LLM in sync with project labels.";
|
||||
response.tool_calls.push_back(call);
|
||||
return response;
|
||||
}
|
||||
|
||||
if (absl::StrContains(normalized, "sprite") &&
|
||||
absl::StrContains(normalized, "room")) {
|
||||
ToolCall call;
|
||||
call.tool_name = "dungeon-list-sprites";
|
||||
call.args.emplace("room", ExtractRoomId(normalized));
|
||||
response.text_response =
|
||||
"Let me inspect the dungeon room sprites for you.";
|
||||
response.reasoning =
|
||||
"Calling the sprite inspection tool provides precise coordinates for the agent.";
|
||||
response.tool_calls.push_back(call);
|
||||
return response;
|
||||
}
|
||||
|
||||
response.text_response =
|
||||
"I'm not sure how to help with that yet. Try asking for resource labels "
|
||||
"or listing dungeon sprites.";
|
||||
return response;
|
||||
}
|
||||
|
||||
@@ -22,7 +120,34 @@ absl::StatusOr<AgentResponse> MockAIService::GenerateResponse(
|
||||
if (history.empty()) {
|
||||
return absl::InvalidArgumentError("History cannot be empty.");
|
||||
}
|
||||
return GenerateResponse(history.back().message);
|
||||
|
||||
// If the last message in history is a tool output, synthesize a summary.
|
||||
for (auto it = history.rbegin(); it != history.rend(); ++it) {
|
||||
if (it->sender == agent::ChatMessage::Sender::kAgent &&
|
||||
(absl::StrContains(it->message, "=== ") ||
|
||||
absl::StrContains(it->message, "\"id\"") ||
|
||||
absl::StrContains(it->message, "\n{"))) {
|
||||
AgentResponse response;
|
||||
response.text_response =
|
||||
"Here's what I found:\n" + it->message +
|
||||
"\nLet me know if you'd like to make a change.";
|
||||
response.reasoning =
|
||||
"Summarized the latest tool output for the user.";
|
||||
return response;
|
||||
}
|
||||
}
|
||||
|
||||
auto user_it = std::find_if(history.rbegin(), history.rend(),
|
||||
[](const agent::ChatMessage& message) {
|
||||
return message.sender ==
|
||||
agent::ChatMessage::Sender::kUser;
|
||||
});
|
||||
if (user_it == history.rend()) {
|
||||
return absl::InvalidArgumentError(
|
||||
"History does not contain a user message.");
|
||||
}
|
||||
|
||||
return GenerateResponse(user_it->message);
|
||||
}
|
||||
|
||||
} // namespace cli
|
||||
|
||||
Reference in New Issue
Block a user