This commit introduces the QLM system - a stateful semantic layer for
tracking HI (Human Intelligence), AI (Agent Intelligence), and QI
(Quantum/Emergent Intelligence) in BlackRoad OS.
Core Features:
- HI/AI/QI intelligence layer modeling
- Event-driven state management
- QI emergence detection (agent self-correction, feedback loops, etc.)
- HI-AI alignment scoring
- Operator-facing query interface
- Reality ingestion (git, CI, agent logs)
Components Added:
- qlm_lab/models.py: Core data models (Actor, QLMEvent, QIEmergence, etc.)
- qlm_lab/state.py: State management and transition tracking
- qlm_lab/api.py: Public QLMInterface API
- qlm_lab/ingestion/: Git, CI, and agent log connectors
- qlm_lab/experiments/: Alignment and emergence validation
- qlm_lab/visualization.py: Timeline, actor graph, alignment plots
- qlm_lab/demo.py: Interactive demo script
- tests/test_qlm_core.py: Comprehensive test suite
- docs/QLM.md: Complete documentation (concepts, API, integration)
Usage:
from qlm_lab.api import QLMInterface
qlm = QLMInterface()
qlm.record_operator_intent("Build feature X")
qlm.record_agent_execution("agent-1", "Implement X", "task-1")
summary = qlm.get_summary(days=7)
Run:
python -m qlm_lab.demo
python -m qlm_lab.experiments.alignment_detection
pytest tests/test_qlm_core.py -v
Integrates with:
- cognitive/intent_graph.py (intent tracking)
- cognitive/agent_coordination.py (multi-agent coordination)
- operator_engine/scheduler.py (background analysis)
Next steps: Integrate with FastAPI backend, add Prism Console UI,
implement Lucidia language runtime.
25 KiB
QLM (Quantum Language Model) - Complete Guide
The semantic layer for Operator-AI collaboration
Table of Contents
- What is QLM?
- Core Concepts
- Architecture
- Data Models
- API Reference
- Integration Guide
- Experiments & Validation
- Visualization
- Operator's Guide
- Development & Extension
What is QLM?
QLM (Quantum Language Model) is a stateful semantic layer that tracks, analyzes, and explains the interaction between Human Intelligence (HI), AI Intelligence (AI), and emergent Quantum Intelligence (QI) in the BlackRoad Operating System.
The Problem QLM Solves
Modern AI systems have a context loss problem:
- AI agents execute tasks but lose track of why
- Humans give intent but can't see what happened
- Systems evolve but no one knows how we got here
- Emergent behaviors appear with no explanation
QLM fixes this by creating a complete causal graph of:
- Operator intent
- Agent executions
- System events
- Feedback loops
- Emergent patterns
The "Quantum" Metaphor
The "quantum" in QLM does NOT refer to quantum physics. Instead, it describes:
Superposition of Roles: An agent can be both executor AND coordinator simultaneously.
Superposition of States: A task can be in_progress AND blocked at the same time.
Superposition of Perspectives: The same event looks different from HI vs AI vs QI viewpoints.
Quantum Intelligence (QI): Emergent behaviors that appear when HI + AI + deterministic systems interact in feedback loops. When 1 + 1 = 3.
Core Concepts
Intelligence Layers
QLM models three layers of intelligence:
1. HI (Human Intelligence)
Definition: The Operator layer - human judgment, taste, ethics, goals, narrative.
Primary Actor: Alexa (Operator)
Capabilities:
- Define intent and constraints
- Approve or veto AI actions
- Ask questions and interpret results
- Provide judgment on ambiguous decisions
Events:
OPERATOR_INTENT: Operator defines a goalOPERATOR_APPROVAL: Operator approves agent workOPERATOR_VETO: Operator rejects agent workOPERATOR_QUERY: Operator asks a question
2. AI (Agent Intelligence)
Definition: LLM-powered agents, code generation, pattern completion, search, transformation.
Primary Actors: 200+ BlackRoad agents (coder, reviewer, researcher, etc.)
Capabilities:
- Execute tasks
- Generate code/docs/designs
- Search and retrieve information
- Coordinate with other agents
Events:
AGENT_EXECUTION: Agent starts workingAGENT_COMPLETION: Agent finishes taskAGENT_ERROR: Agent encounters errorAGENT_HANDOFF: Agent passes work to another agent
3. QI (Quantum Intelligence)
Definition: Emergent system-level intelligence that appears when HI + AI + deterministic systems interact in feedback loops.
Not a single actor: QI is a property of the entire system.
Emergence Patterns:
agent_self_correction: Agent fixes own errors without HI interventionnovel_solution: Agent finds approach not in training dataemergent_collaboration: Agents self-organize into new patternsoperator_feedback_loop: HI → AI → HI creates refined understandingsystem_adaptation: Deterministic systems evolve in response to AI
Events:
QI_EMERGENCE: Novel behavior detectedQI_FEEDBACK_LOOP: HI+AI feedback detectedQI_PATTERN: Recurring emergent pattern identified
Architecture
System Diagram
┌─────────────────────────────────────────────────────┐
│ Operator (Alexa) │
│ Human Intelligence (HI) │
└──────────────────────┬──────────────────────────────┘
│
│ Intent, Approval, Veto
↓
┌──────────────────────────────────────────────────────┐
│ QLM Interface │
│ record_operator_intent(), ask(), get_summary() │
└──────────────────────┬───────────────────────────────┘
│
↓
┌──────────────────────────────────────────────────────┐
│ QLM State │
│ • Intelligence Layers (HI, AI, QI) │
│ • Event History │
│ • QI Emergence Detection │
│ • Metrics & Alignment │
└──────┬────────────────────────┬──────────────────────┘
│ │
│ │
↓ ↓
┌──────────────────┐ ┌──────────────────────────────┐
│ Ingestion │ │ Cognitive Layer Integration │
│ • Git │ │ • IntentGraph │
│ • CI/CD │ │ • AgentCoordinator │
│ • Agent Logs │ │ • ContextEngine │
└──────────────────┘ └──────────────────────────────┘
│
↓
┌──────────────────────────────────────────────────────┐
│ Reality (External Systems) │
│ • Git commits • Test results • Deployments │
│ • Agent executions • System events │
└──────────────────────────────────────────────────────┘
Key Components
qlm_lab/models.py
Defines core data structures:
IntelligenceType,ActorType,ActorRoleActor: Represents humans, agents, systemsQLMEvent: Every action generates an eventQIEmergence: Detected emergent patternsQLMMetrics: System-level metrics
qlm_lab/state.py
Manages QLM state:
QLMState: Complete state snapshotingest_event(): Process new events and update statequery(): Answer questions about statedetect_qi_emergence(): Pattern matching for QIcalculate_alignment(): HI-AI alignment scoring
qlm_lab/api.py
Public interface:
QLMInterface: Main API classrecord_*(): Methods to record eventsget_*(): Query methodsask(): Natural language queries
qlm_lab/ingestion/
Connects QLM to reality:
GitConnector: Ingest git commitsCIConnector: Ingest test/build/deploy resultsAgentLogConnector: Parse agent logs
qlm_lab/experiments/
Validation experiments:
AlignmentDetectionExperiment: Test alignment scoringEmergenceDetectionExperiment: Test QI detection
qlm_lab/visualization.py
Visualization tools:
- Event timeline
- Actor interaction graph
- Alignment trends
- Emergence patterns
Data Models
Actor
@dataclass
class Actor:
id: str # Unique identifier
name: str # Human-readable name
actor_type: ActorType # HUMAN | AGENT | SYSTEM
role: ActorRole # OPERATOR | EXECUTOR | COORDINATOR | REVIEWER | ...
state: ActorState # ACTIVE | IDLE | BLOCKED | OFFLINE
capabilities: Set[str] # What this actor can do
current_task_id: Optional[str]
created_at: datetime
last_active: datetime
QLMEvent
@dataclass
class QLMEvent:
id: str # Unique event ID
timestamp: datetime # When event occurred
source_layer: IntelligenceType # HI | AI | QI
actor_id: str # Who generated this event
event_type: EventType # OPERATOR_INTENT | AGENT_EXECUTION | ...
data: Dict[str, Any] # Event payload
caused_by: List[str] # Causal event IDs
intent_node_id: Optional[str] # Link to IntentGraph
task_id: Optional[str] # Related task
tags: Set[str]
metadata: Dict[str, Any]
QIEmergence
@dataclass
class QIEmergence:
id: str # Unique emergence ID
timestamp: datetime # When detected
pattern_name: str # e.g., "agent_self_correction"
trigger_events: List[str] # Events that triggered this
confidence: float # 0.0 to 1.0
explanation: str # Human-readable description
operator_validated: Optional[bool] # Did Operator confirm?
operator_notes: str
impact_score: float # Significance (0.0 to 1.0)
QLMMetrics
@dataclass
class QLMMetrics:
hi_events: int # Count of HI events
ai_events: int # Count of AI events
qi_events: int # Count of QI events
system_events: int # Count of system events
hi_ai_alignment: float # Alignment score (0.0 to 1.0)
qi_emergence_rate: float # Rate of QI detection
feedback_loop_count: int # HI→AI→HI cycles
operator_approvals: int
operator_vetoes: int
operator_queries: int
start_time: datetime
end_time: datetime
API Reference
Initialization
from qlm_lab.api import QLMInterface
# Basic initialization
qlm = QLMInterface()
# With cognitive layer integration
from cognitive.intent_graph import IntentGraph
from cognitive.agent_coordination import AgentCoordinator
intent_graph = IntentGraph()
agent_coordinator = AgentCoordinator(intent_graph)
qlm = QLMInterface(
intent_graph=intent_graph,
agent_coordinator=agent_coordinator
)
Recording Events
Operator Events (HI)
# Record Operator intent
event = qlm.record_operator_intent(
intent="Deploy authentication feature",
description="Implement login, signup, password reset",
intent_node_id="intent-auth-001" # Link to IntentGraph
)
# Record Operator approval
qlm.record_operator_approval(
what_approved="Login implementation",
intent_node_id="intent-auth-001",
task_id="task-login-001"
)
# Record Operator veto
qlm.record_operator_veto(
what_vetoed="Password reset implementation",
reason="Security concerns - needs stronger validation",
intent_node_id="intent-auth-001",
task_id="task-reset-001"
)
# Record Operator query
qlm.record_operator_query("What did agents do today?")
Agent Events (AI)
# Register an agent
agent = qlm.register_agent(
agent_id="coder-001",
name="CodeWriter",
role=ActorRole.CODER,
capabilities=["python", "javascript", "testing"]
)
# Record agent execution
qlm.record_agent_execution(
agent_id="coder-001",
task_description="Implement login endpoint",
task_id="task-login-001",
intent_node_id="intent-auth-001"
)
# Record agent completion
qlm.record_agent_completion(
agent_id="coder-001",
task_id="task-login-001",
success=True,
result={"files_modified": ["auth.py"], "tests_added": 5}
)
# Record agent error
qlm.record_agent_error(
agent_id="coder-001",
task_id="task-login-001",
error="Database connection failed"
)
# Record agent handoff
qlm.record_agent_handoff(
from_agent_id="coder-001",
to_agent_id="reviewer-001",
task_id="task-login-001",
handoff_message="Ready for review"
)
System Events
# Record test result
qlm.record_system_event(
event_type=EventType.SYSTEM_TEST,
description="Backend tests passed",
task_id="task-login-001",
metadata={
"passed": True,
"test_count": 42,
"duration_seconds": 12.3
}
)
# Record build result
qlm.record_system_event(
event_type=EventType.SYSTEM_BUILD,
description="Production build successful",
metadata={"build_id": "build-123", "artifacts": ["app.tar.gz"]}
)
# Record deployment
qlm.record_system_event(
event_type=EventType.SYSTEM_DEPLOY,
description="Deployed to production",
metadata={"environment": "production", "version": "v1.2.0"}
)
Querying State
# Get Operator summary
summary = qlm.get_summary(days=7)
print(summary)
# Get alignment score
alignment = qlm.get_alignment_score()
print(f"HI-AI Alignment: {alignment:.1%}")
# Get recent QI emergences
emergences = qlm.get_recent_emergences(limit=10)
for em in emergences:
print(f"{em.pattern_name}: {em.explanation}")
# Get active actors
active = qlm.get_active_actors()
for actor in active:
print(f"{actor.name} - {actor.role.value}")
# Get events by type
intents = qlm.get_events_by_type(EventType.OPERATOR_INTENT)
print(f"Total intents: {len(intents)}")
# Get events in time range
from datetime import datetime, timedelta
yesterday = datetime.now() - timedelta(days=1)
recent_events = qlm.get_events_in_timerange(yesterday)
Natural Language Queries
# Ask questions in natural language
response = qlm.ask("What did agents do today?")
print(response)
response = qlm.ask("Are we aligned with my intent?")
print(response)
response = qlm.ask("Show me emergent behaviors")
print(response)
response = qlm.ask("What's the status?")
print(response)
Export/Import State
# Export state to JSON
qlm.export_state("/path/to/qlm_state.json")
# Import state from JSON
qlm.import_state("/path/to/qlm_state.json")
Integration Guide
Integrating with Existing Cognitive Layer
QLM is designed to integrate seamlessly with the existing cognitive infrastructure:
from cognitive.intent_graph import IntentGraph
from cognitive.agent_coordination import AgentCoordinator
from qlm_lab.api import QLMInterface
# Initialize cognitive systems
intent_graph = IntentGraph()
agent_coordinator = AgentCoordinator(intent_graph)
# Initialize QLM with cognitive integration
qlm = QLMInterface(
intent_graph=intent_graph,
agent_coordinator=agent_coordinator
)
# Now when you create goals in IntentGraph...
goal = intent_graph.create_goal(
title="Build authentication",
rationale="Need secure user login"
)
# ...also record in QLM
qlm.record_operator_intent(
intent="Build authentication",
intent_node_id=goal.id
)
# When agents coordinate...
task = intent_graph.create_task(
title="Implement login",
parent_id=goal.id
)
agent_coordinator.assign_task(task.id, agent_id="coder-001")
# ...also record in QLM
qlm.record_agent_execution(
agent_id="coder-001",
task_description="Implement login",
task_id=task.id,
intent_node_id=goal.id
)
Ingesting Real System Data
Git Commits
from qlm_lab.ingestion.git import GitConnector
connector = GitConnector(repo_path="/path/to/repo", qlm=qlm)
# Ingest last 7 days of commits
events = connector.ingest_recent_commits(days=7)
print(f"Ingested {len(events)} commits")
# Ingest specific range
events = connector.ingest_commit_range(
since="2024-01-01",
until="2024-01-31"
)
CI/CD Results
from qlm_lab.ingestion.ci import CIConnector
connector = CIConnector(qlm=qlm)
# Ingest test result
connector.ingest_test_result(
test_name="Backend Tests",
passed=True,
duration_seconds=12.3,
commit_hash="abc123"
)
# Ingest build result
connector.ingest_build_result(
build_name="Production Build",
success=True,
duration_seconds=45.2,
artifacts=["app.tar.gz", "app.zip"]
)
# Ingest deployment
connector.ingest_deploy_result(
service_name="blackroad-api",
environment="production",
success=True,
version="v1.2.0"
)
Agent Logs
from qlm_lab.ingestion.agent_logs import AgentLogConnector
connector = AgentLogConnector(qlm=qlm)
# Ingest log file
events = connector.ingest_log_file("/path/to/agent.log")
print(f"Ingested {len(events)} events from logs")
# Ingest structured logs
log_entries = [
{
"timestamp": "2024-01-15T10:30:00",
"agent_id": "coder-001",
"level": "INFO",
"message": "Task started: implement login"
},
# ...
]
events = connector.ingest_structured_log(log_entries)
Experiments & Validation
QLM includes built-in experiments to validate its effectiveness:
Alignment Detection Experiment
Hypothesis: QLM can accurately detect when AI agents drift from Operator intent.
python -m qlm_lab.experiments.alignment_detection
Scenarios:
- Perfect alignment (100% approval)
- Partial alignment (some vetoes)
- No alignment (all vetoes)
Success Criteria: Alignment score accuracy within 20%
Emergence Detection Experiment
Hypothesis: QLM can detect emergent QI behaviors.
python -m qlm_lab.experiments.emergence_detection
Patterns Tested:
- Agent self-correction
- Operator feedback loop
- Emergent collaboration
- Normal execution (should NOT trigger)
Success Criteria:
- True positive rate ≥ 80%
- False positive rate < 20%
Visualization
QLM includes powerful visualization tools:
from qlm_lab.visualization import QLMVisualizer
viz = QLMVisualizer(qlm)
# Event timeline
viz.plot_event_timeline(save_path="timeline.png")
# Actor interaction graph
viz.plot_actor_graph(save_path="actors.png")
# Alignment over time
viz.plot_alignment_over_time(save_path="alignment.png")
# Emergence patterns
viz.plot_emergence_patterns(save_path="emergence.png")
# Export complete dashboard
viz.export_dashboard(output_dir="./qlm_dashboard")
Requirements:
pip install matplotlib networkx
Operator's Guide
Daily Usage
As the Operator (Alexa), here's how to use QLM:
Morning Check-In
# What happened overnight?
summary = qlm.get_summary(days=1)
print(summary)
# Are agents aligned with my goals?
alignment = qlm.get_alignment_score()
if alignment < 0.7:
print("⚠️ Warning: Low alignment detected")
# Any emergent behaviors?
emergences = qlm.get_recent_emergences()
for em in emergences:
print(f"✨ {em.pattern_name}: {em.explanation}")
Defining Intent
# When starting a new project
qlm.record_operator_intent(
intent="Build payment integration",
description="Integrate Stripe for subscriptions and one-time payments",
intent_node_id="intent-payment-2024-01"
)
Reviewing Agent Work
# When agents complete work
qlm.record_operator_approval(
what_approved="Stripe integration implementation",
intent_node_id="intent-payment-2024-01",
task_id="task-stripe-001"
)
# When work doesn't match intent
qlm.record_operator_veto(
what_vetoed="Payment form UI",
reason="Doesn't match brand guidelines - needs redesign",
intent_node_id="intent-payment-2024-01",
task_id="task-ui-001"
)
Asking Questions
# Natural language queries
qlm.ask("What did agents do today?")
qlm.ask("Are we aligned with my intent?")
qlm.ask("Show me emergent behaviors")
qlm.ask("What's the status?")
Understanding Alignment
Alignment Score: 0.0 to 1.0
- 0.9-1.0: Excellent - agents are executing your intent well
- 0.7-0.9: Good - minor drift, watch for patterns
- 0.5-0.7: Warning - significant misalignment, review vetoes
- < 0.5: Critical - agents not following intent, intervention needed
Improving Alignment:
- Be more specific in intent descriptions
- Provide examples of what "good" looks like
- Give immediate feedback (approve/veto)
- Review patterns in vetoes - is there confusion?
Understanding QI Emergence
Common Patterns:
-
agent_self_correction: Agent fixed its own error without your help
- Good sign: Agents are learning and adapting
-
novel_solution: Agent found an approach you didn't suggest
- Good sign: Creative problem-solving
- Watch for: Ensure solution aligns with intent
-
operator_feedback_loop: You → agent → feedback → refined approach
- Good sign: Healthy iteration cycle
- Measure: Count of loops indicates collaboration quality
-
emergent_collaboration: Agents self-organized
- Good sign: Agents coordinating without explicit instructions
- Watch for: Ensure coordination serves your intent
Development & Extension
Adding New Event Types
# In qlm_lab/models.py
class EventType(Enum):
# ... existing types ...
CUSTOM_EVENT = "custom_event"
# In qlm_lab/api.py
def record_custom_event(self, ...):
event = QLMEvent(
source_layer=IntelligenceType.AI,
event_type=EventType.CUSTOM_EVENT,
...
)
self.state.ingest_event(event)
return event
Adding New QI Patterns
# In qlm_lab/models.py
QI_PATTERNS["my_pattern"] = {
"description": "Description of when this pattern occurs",
"trigger": "sequence of events that trigger this",
"significance": "high", # high | medium | low | very_high
}
# In qlm_lab/state.py
def _matches_pattern(self, events, pattern_name, pattern_def):
if pattern_name == "my_pattern":
# Implement pattern detection logic
# Return True if pattern matches
pass
Adding New Connectors
# Create qlm_lab/ingestion/my_connector.py
from qlm_lab.api import QLMInterface
from qlm_lab.models import EventType
class MyConnector:
def __init__(self, qlm: QLMInterface):
self.qlm = qlm
def ingest_my_data(self, data):
# Transform data into QLM events
event = self.qlm.record_system_event(
event_type=EventType.SYSTEM_CUSTOM,
description=...,
metadata=...
)
return event
Running Tests
# Run QLM tests
pytest tests/test_qlm_core.py -v
# Run all tests
pytest tests/ -v
# With coverage
pytest tests/test_qlm_core.py --cov=qlm_lab --cov-report=html
Running Demo
# Interactive demo
python -m qlm_lab.demo
# Run specific experiment
python -m qlm_lab.experiments.alignment_detection
python -m qlm_lab.experiments.emergence_detection
Next Steps
Phase 1: Lab (Current)
- ✅ Core QLM models and state management
- ✅ Basic event ingestion (git, CI, agents)
- ✅ Alignment and emergence detection
- ✅ Visualization and experiments
- ✅ Documentation
Phase 2: Integration (Next 1-2 months)
- Full cognitive layer integration
- Real-time event streaming
- FastAPI router for QLM API
- Prism Console UI integration
- Dashboard for Operator
Phase 3: Production (Months 3-4)
- Database persistence (PostgreSQL)
- Advanced QI pattern detection (ML-based)
- Multi-Operator support
- Audit trail and compliance
- Performance optimization
Phase 4: Lucidia Language (Months 5-6)
- Lucidia syntax for expressing intent
- Lucidia → QLM compiler
- QLM → Lucidia decompiler (explain mode)
- Lucidia REPL for live QLM queries
FAQ
Q: What's the difference between QLM and the IntentGraph?
A: IntentGraph tracks what (goals, tasks, artifacts). QLM tracks why, how, and emergence (intent, execution, QI patterns). They complement each other.
Q: Do I need to use QLM for every agent?
A: No, start small. Use QLM for critical workflows where you need clear intent tracking and alignment verification.
Q: How does QLM relate to Lucidia?
A: Lucidia is the language for expressing intent and constraints. QLM is the runtime that tracks and enforces them.
Q: What if I don't want to record every event?
A: QLM is opt-in. Only record events you care about. Start with Operator intents and agent completions.
Q: Can QLM work with multiple Operators?
A: Currently optimized for single Operator (Alexa). Multi-Operator support planned for Phase 3.
Q: How do I debug QLM?
A: Use qlm.state.export_json() to inspect state, visualizations to see patterns, and experiments to validate behavior.
Conclusion
QLM provides the missing semantic layer for Operator-AI collaboration. By tracking HI, AI, and QI as first-class concepts, QLM makes AI systems understandable, controllable, and improvable.
Start using QLM today:
from qlm_lab.api import QLMInterface
qlm = QLMInterface()
# Your intent
qlm.record_operator_intent("Build the future")
# Agent execution
qlm.record_agent_execution("agent-1", "Create something amazing", "task-1")
# Check alignment
print(f"Alignment: {qlm.get_alignment_score():.1%}")
# Ask questions
print(qlm.ask("What's the status?"))
Happy building! 🛣️✨
QLM is part of the BlackRoad Operating System. See BLACKROAD_OS_BIG_KAHUNA_VISION.md for the complete vision.