docs: Initial documentation cleanup of parallel agent MVP

This commit is contained in:
2026-02-17 14:30:28 -07:00
parent cdafdff281
commit 7f267a10a1
4 changed files with 13 additions and 76 deletions
+3 -38
View File
@@ -225,10 +225,6 @@ pub fn supervisor_function_declarations() -> Vec<FunctionDeclaration> {
]
}
// ---------------------------------------------------------------------------
// Dispatch
// ---------------------------------------------------------------------------
pub async fn handle_supervisor_tool(
config: &GlobalConfig,
cmd_name: &str,
@@ -253,21 +249,6 @@ pub async fn handle_supervisor_tool(
}
}
// ---------------------------------------------------------------------------
// Child agent execution loop
// ---------------------------------------------------------------------------
/// Run a child agent to completion, returning its accumulated text output.
///
/// This mirrors `start_directive` in main.rs but:
/// - Uses `call_chat_completions(print=false)` so nothing goes to stdout
/// - Returns the output text instead of printing it
/// - Loops on tool calls just like `start_directive`'s recursion
///
/// Returns a boxed future to break the recursive type cycle:
/// handle_spawn → tokio::spawn(run_child_agent) → call_chat_completions
/// → eval_tool_calls → ToolCall::eval → handle_supervisor_tool → handle_spawn
/// Without boxing, the compiler cannot prove Send for the recursive async type.
fn run_child_agent(
child_config: GlobalConfig,
initial_input: Input,
@@ -283,8 +264,8 @@ fn run_child_agent(
let (output, tool_results) = call_chat_completions(
&input,
false, // print=false — silent, no stdout
false, // extract_code=false
false,
false,
client.as_ref(),
abort_signal.clone(),
)
@@ -305,7 +286,6 @@ fn run_child_agent(
break;
}
// Feed tool results back for the next round (mirrors start_directive recursion)
input = input.merge_tool_results(output, tool_results);
}
@@ -313,10 +293,6 @@ fn run_child_agent(
})
}
// ---------------------------------------------------------------------------
// Handlers
// ---------------------------------------------------------------------------
async fn handle_spawn(config: &GlobalConfig, args: &Value) -> Result<Value> {
let agent_name = args
.get("agent")
@@ -330,18 +306,15 @@ async fn handle_spawn(config: &GlobalConfig, args: &Value) -> Result<Value> {
.to_string();
let _task_id = args.get("task_id").and_then(Value::as_str);
// Generate a unique agent ID
let short_uuid = &Uuid::new_v4().to_string()[..8];
let agent_id = format!("agent_{agent_name}_{short_uuid}");
// --- Validate capacity ---
// Read the supervisor to check capacity, then drop locks before doing async work
let (max_depth, current_depth) = {
let cfg = config.read();
let supervisor = cfg
.supervisor
.as_ref()
.ok_or_else(|| anyhow!("No supervisor active — agent spawning not enabled"))?;
.ok_or_else(|| anyhow!("No supervisor active; Agent spawning not enabled"))?;
let sup = supervisor.read();
if sup.active_count() >= sup.max_concurrent() {
return Ok(json!({
@@ -363,7 +336,6 @@ async fn handle_spawn(config: &GlobalConfig, args: &Value) -> Result<Value> {
}));
}
// --- Build an isolated child Config ---
let child_inbox = Arc::new(Inbox::new());
let child_config: GlobalConfig = {
@@ -384,16 +356,13 @@ async fn handle_spawn(config: &GlobalConfig, args: &Value) -> Result<Value> {
Arc::new(RwLock::new(child_cfg))
};
// Load the target agent into the child config
let child_abort = create_abort_signal();
Config::use_agent(&child_config, &agent_name, None, child_abort.clone()).await?;
// Create the initial input from the prompt
let input = Input::from_str(&child_config, &prompt, None);
debug!("Spawning child agent '{agent_name}' as '{agent_id}'");
// --- Spawn the agent task ---
let spawn_agent_id = agent_id.clone();
let spawn_agent_name = agent_name.clone();
let spawn_abort = child_abort.clone();
@@ -417,7 +386,6 @@ async fn handle_spawn(config: &GlobalConfig, args: &Value) -> Result<Value> {
}
});
// Register the handle with the supervisor
let handle = AgentHandle {
id: agent_id.clone(),
agent_name: agent_name.clone(),
@@ -463,7 +431,6 @@ async fn handle_check(config: &GlobalConfig, args: &Value) -> Result<Value> {
match is_finished {
Some(true) => {
// Finished — collect the result
handle_collect(config, args).await
}
Some(false) => Ok(json!({
@@ -484,7 +451,6 @@ async fn handle_collect(config: &GlobalConfig, args: &Value) -> Result<Value> {
.and_then(Value::as_str)
.ok_or_else(|| anyhow!("'id' is required"))?;
// Extract the join handle while holding the lock, then drop the lock before awaiting
let handle = {
let cfg = config.read();
let supervisor = cfg
@@ -497,7 +463,6 @@ async fn handle_collect(config: &GlobalConfig, args: &Value) -> Result<Value> {
match handle {
Some(handle) => {
// Await the join handle OUTSIDE of any lock
let result = handle
.join_handle
.await
-23
View File
@@ -1,10 +1,5 @@
use serde::{Deserialize, Serialize};
/// A message envelope routed between agents.
///
/// Agents communicate by sending `Envelope`s to each other's mailboxes.
/// The sender fires and forgets; the receiver drains its inbox between
/// LLM turns via the `check_inbox` tool.
#[derive(Debug, Clone, Serialize, Deserialize)]
pub struct Envelope {
pub from: String,
@@ -13,11 +8,6 @@ pub struct Envelope {
pub timestamp: chrono::DateTime<chrono::Utc>,
}
/// The content of an inter-agent message.
///
/// Separates the **control plane** (shutdown signals, task lifecycle events)
/// from the **data plane** (free-form text). Control-plane messages are
/// processed before data-plane messages to prevent race conditions.
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(tag = "type", rename_all = "snake_case")]
pub enum EnvelopePayload {
@@ -27,13 +17,6 @@ pub enum EnvelopePayload {
ShutdownApproved,
}
/// A per-agent inbox that collects incoming messages.
///
/// Backed by a `Vec` behind a `parking_lot::Mutex` so it can be shared
/// between the supervisor (which delivers messages) and the agent's tool
/// handler (which drains them). We use `parking_lot::Mutex` to match the
/// locking convention used elsewhere in Loki (`parking_lot::RwLock` for
/// GlobalConfig).
#[derive(Debug, Default)]
pub struct Inbox {
messages: parking_lot::Mutex<Vec<Envelope>>,
@@ -50,18 +33,12 @@ impl Inbox {
self.messages.lock().push(envelope);
}
/// Drain all pending messages, returning them sorted with control-plane
/// messages first (shutdown, task events) then data-plane (text).
/// This ordering prevents the class of bugs where a text message
/// references state that a control message was supposed to set up.
pub fn drain(&self) -> Vec<Envelope> {
let mut msgs = {
let mut guard = self.messages.lock();
std::mem::take(&mut *guard)
};
// Stable partition: control messages first, then data messages,
// preserving relative order within each group.
msgs.sort_by_key(|e| match &e.payload {
EnvelopePayload::ShutdownRequest { .. } => 0,
EnvelopePayload::ShutdownApproved => 0,
-2
View File
@@ -35,8 +35,6 @@ pub struct AgentHandle {
pub join_handle: JoinHandle<Result<AgentResult>>,
}
/// Lives as an `Arc<parking_lot::RwLock<Supervisor>>` alongside GlobalConfig,
/// NOT inside it — avoids adding lock contention to the shared Config.
pub struct Supervisor {
handles: HashMap<String, AgentHandle>,
task_queue: TaskQueue,
+10 -13
View File
@@ -122,13 +122,13 @@ impl TaskQueue {
}
pub fn claim(&mut self, task_id: &str, owner: &str) -> bool {
if let Some(task) = self.tasks.get_mut(task_id) {
if task.is_runnable() && task.owner.is_none() {
task.owner = Some(owner.to_string());
task.status = TaskStatus::InProgress;
return true;
}
if let Some(task) = self.tasks.get_mut(task_id) &&
task.is_runnable() && task.owner.is_none() {
task.owner = Some(owner.to_string());
task.status = TaskStatus::InProgress;
return true;
}
false
}
@@ -146,8 +146,6 @@ impl TaskQueue {
tasks
}
// DFS cycle detection: would adding task_id -> blocked_by create a cycle?
// A cycle exists if blocked_by can reach task_id through existing dependencies.
fn would_create_cycle(&self, task_id: &str, blocked_by: &str) -> bool {
let mut visited = HashSet::new();
let mut stack = vec![blocked_by.to_string()];
@@ -156,11 +154,10 @@ impl TaskQueue {
if current == task_id {
return true;
}
if visited.insert(current.clone()) {
if let Some(task) = self.tasks.get(&current) {
for dep in &task.blocked_by {
stack.push(dep.clone());
}
if visited.insert(current.clone()) &&
let Some(task) = self.tasks.get(&current) {
for dep in &task.blocked_by {
stack.push(dep.clone());
}
}
}