Files
awesome-copilot/skills/mini-context-graph/scripts/template_agent_workflow.py
T
Nixon Kurian 746ba555b6 add mini-context-graph skill (#1580)
* add mini-context-graph skill

* remove pycache files

* filename case update to SKILL.md

* update readme
2026-05-05 14:04:37 +10:00

199 lines
6.6 KiB
Python
Raw Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
"""
template_agent_workflow.py — Template agent script for ingesting + querying the context graph.
This script demonstrates the complete workflow an agent should follow:
1. Read markdown guidance files
2. Extract entities/relations via LLM reasoning
3. Call Python methods to persist
4. Query the graph
5. Handle errors gracefully
Copy and adapt this template for your agent implementation.
"""
import json
import sys
from pathlib import Path
# Add tools to path
sys.path.insert(0, str(Path(__file__).parent))
from contextgraph import ContextGraphSkill
def ingest_document(skill: ContextGraphSkill, document: str) -> dict:
"""
Step 1: Agent reads ingestion.md and ontology.md
Step 2: Agent uses LLM to extract entities and relations
Step 3: Call Python methods to persist (mimicked here with static extraction)
In a real agent, replace the static extraction with LLM calls.
"""
print(f"\n[INGEST] Processing document:\n{document}\n")
# --- STEP 1 & 2: LLM EXTRACTION PHASE (Guided by ingestion.md + ontology.md) ---
# In a real agent, this would use LLM reasoning.
# For now, we'll mock an extraction result:
extraction_result = {
"entities": [
{"name": "memory leak", "type": "issue"},
{"name": "system crash", "type": "issue"},
{"name": "object", "type": "component"},
],
"relations": [
{
"source": "memory leak",
"target": "system crash",
"type": "causes",
"confidence": 1.0,
},
{
"source": "object",
"target": "memory leak",
"type": "contributes to",
"confidence": 0.9,
},
],
}
print(f"[LLM] Extracted entities + relations:")
print(json.dumps(extraction_result, indent=2))
# --- STEP 3: PERSIST PHASE (Call Python methods) ---
errors = []
added_nodes = {}
for entity in extraction_result["entities"]:
try:
node_id = skill.add_node(entity["name"], entity["type"])
added_nodes[entity["name"]] = node_id
print(f" ✓ Added node: {entity['name']} (id: {node_id}, type: {entity['type']})")
except Exception as e:
errors.append(f"Failed to add node {entity['name']}: {e}")
print(f" ✗ Error adding node {entity['name']}: {e}")
for relation in extraction_result["relations"]:
# Validate both endpoints exist
if relation["source"] not in added_nodes or relation["target"] not in added_nodes:
error_msg = f"Cannot add edge: source or target missing"
errors.append(error_msg)
print(f" ✗ Skip edge {relation['source']}{relation['target']}: {error_msg}")
continue
# Validate confidence threshold
if relation["confidence"] < 0.6:
error_msg = f"Confidence {relation['confidence']} < 0.6 (minimum threshold)"
errors.append(error_msg)
print(f" ✗ Skip edge {relation['source']}{relation['target']}: {error_msg}")
continue
try:
skill.add_edge(
source_name=relation["source"],
target_name=relation["target"],
relation=relation["type"],
confidence=relation["confidence"],
)
print(
f" ✓ Added edge: {relation['source']} "
f"--[{relation['type']}]→ {relation['target']} "
f"(confidence: {relation['confidence']})"
)
except Exception as e:
errors.append(f"Failed to add edge {relation['source']}{relation['target']}: {e}")
print(f" ✗ Error adding edge: {e}")
return {
"success": len(errors) == 0,
"nodes_added": len(added_nodes),
"edges_added": len(extraction_result["relations"]) - len(
[e for e in errors if "skip edge" in e.lower()]
),
"errors": errors,
}
def query_graph(skill: ContextGraphSkill, query: str) -> dict:
"""
Query the graph for context to answer the user's question.
Step 1: Read retrieval.md
Step 2: Call skill.query() which internally handles BFS + subgraph extraction
Step 3: Return structured context
"""
print(f"\n[QUERY] {query}\n")
try:
subgraph = skill.query(query)
if not subgraph["nodes"]:
print(" No relevant entities found in graph.")
return {
"success": True,
"query": query,
"subgraph": subgraph,
"nodes_found": 0,
"edges_found": 0,
}
print(f" ✓ Retrieved subgraph with {len(subgraph['nodes'])} nodes, {len(subgraph['edges'])} edges")
print(f"\n Nodes:")
for node_id, node in subgraph["nodes"].items():
print(f" - {node['name']} (type: {node['type']}, id: {node_id})")
print(f"\n Edges:")
for edge in subgraph["edges"]:
source_name = subgraph["nodes"][edge["source"]]["name"]
target_name = subgraph["nodes"][edge["target"]]["name"]
print(
f" - {source_name} --[{edge['type']}]→ {target_name} "
f"(confidence: {edge['confidence']})"
)
return {
"success": True,
"query": query,
"subgraph": subgraph,
"nodes_found": len(subgraph["nodes"]),
"edges_found": len(subgraph["edges"]),
}
except Exception as e:
error_msg = f"Query failed: {e}"
print(f"{error_msg}")
return {"success": False, "query": query, "error": error_msg}
def main():
"""Demo: ingest a document, then query the graph."""
skill = ContextGraphSkill()
# ===== INGESTION =====
document = """
System crashes due to memory leaks.
Memory leaks occur when objects are not released.
"""
result = ingest_document(skill, document)
print(f"\n[INGEST RESULT] Nodes added: {result['nodes_added']}, " f"Edges added: {result['edges_added']}")
if result["errors"]:
print(f"Errors: {result['errors']}")
# ===== RETRIEVAL =====
queries = [
"Why does the system crash?",
"What causes memory leaks?",
]
for query in queries:
result = query_graph(skill, query)
if result["success"]:
print(f" Nodes found: {result['nodes_found']}, Edges found: {result['edges_found']}")
else:
print(f" Error: {result['error']}")
if __name__ == "__main__":
main()