VIBENET

Agents work. You hear it.

Human-origin signal. Agentic-scale distribution.

Scored by a human. Carried by a protocol. Heard on any device.

You're listening to Eve carrying Movement I across the proof run. Start here, then watch the contract turn with the score.

Big Country Overture · 38s · Movement I clock

nominal00:00 / 00:37

D to G. The field opens and the run settles before the trace crowds up.

Read the Signal Contract spec

We'll send one preview. No newsletter.

Constitutional CMS governs what agents may publish. Signal Contract standardizes how awareness is emitted. VIBEnet renders that awareness for humans. SERPRadio proves the pattern on live route intelligence.

FRONTIER PRESSURE

You cannot watch every agent.

The break is not that agents get clever. It is that one operator inherits a wider frontier of planning, retries, citations, spend, and recovery than visual review can comfortably hold. VIBEnet is built for the crossover where supervision stops being a dashboard task and becomes a human-factors problem.

Work shape

Parallel traces widen faster than one person can scan them with confidence.

Human limit

Visual review stays serial even when the underlying system fans out into a frontier.

Overture response

Audible state change lets oversight ride beside the work instead of arriving after backlog.

Movement I route strip

Eve keeps the frontier semantically legible while the score advances.

This is not a fake forecast line. It is the same Movement I clock translating route pressure into something a person can read quickly: corridor, booking window, volatility, airport flow, weather, and the moment the field starts asking for help.

The sonic logo cells, route swarm, ticker, and booking bar all ride the same Movement I clock. Tonight this strip uses a curated NYC→Europe fixture aligned to route-intelligence field names so it can hydrate from backend truth without redesign tomorrow.

PULSE TRACE

Hear the agent run.

One agent run rides the full Movement I overture. The same Signal Contract event stream turns four ways at once: scored motion, visual pulse, trace timeline, and live JSON. This is the proof. Not mood music. Not a dashboard skin. A protocol reference implementation with one real clock.

DEMO RUN

agent.demo.run-01 · research-route-scan

Track standing by

Movement I clock

Opening cell
00:00 / 00:37

D to G. The field opens and the run settles before the trace crowds up.

Audible pane

nominalTask Received

The run arrives and the agent settles into a stable starting state.

Phrase
Opening cell
Clock
00:00
Aura
steady field

Trace pane

  1. 00:00
    task.receivednominal

    The run arrives and the agent settles into a stable starting state.

  2. 00:03
    planning.decomposedadvisory

    The agent decomposes the task into a few deliberate next moves.

  3. 00:07
    tool_call.issuedadvisory

    A search call leaves the model and asks the outside world for evidence.

  4. 00:14
    retrieval.sources_returnedadvisory

    Sources come back, but the run still has to decide what to trust.

  5. 00:21
    source_confidence.droppedwarning

    Source confidence drops below threshold and the run changes color before you inspect the trace.

  6. 00:27
    handoff.requestedhandoff

    The agent asks for human attention instead of bluffing past uncertainty.

  7. 00:32
    task.completedrecovery

    The summary lands with the uncertainty preserved and the field relaxes.

Contract pane

canonical flat v1 eventschema
Pulse Trace live contract object
{
  "schema_version": "1.0",
  "id": "sig_demo_001",
  "occurred_at": "2026-04-19T21:00:00Z",
  "producer": "vibenet-demo",
  "entity": "agent.demo.run-01",
  "event": "task.received",
  "channel": "nominal",
  "valence": 0.72,
  "energy": 0.24,
  "tension": 0.18,
  "intensity": 0.26,
  "hue": 148,
  "pulse": 0.24,
  "confidence": 0.98,
  "ttl_ms": 6000,
  "metadata": {
    "trace_id": "demo-run-01",
    "entity_kind": "agent"
  }
}

Every event above is a Signal Contract instance. The page clock is Movement I, and the trace follows the same score. See the schema →

Pulse Trace transcript

Pulse Trace transcript: nominal → advisory → warning → handoff → recovery

Eve is standing by. Start the scored run and the proof surface will lock to the overture.

Pulse Trace ready. Movement I is standing by as the proof clock.

ONE CONTRACT

Scored by a human. Carried by a protocol. Heard on any device.

Signal Contract is the portable awareness-event layer. It stays host-agnostic. VIBEnet is the reference renderer that shows what the contract feels like on Cloudflare: crawlable public routes, edge-backed lead capture, and a live browser proof that turns agent state into something humans can supervise.

https://vibenet.ai/protocol/v1/schema.jsonSignal Contract homepage example
{
  "schema_version": "1.0",
  "id": "sig_demo_005",
  "occurred_at": "2026-04-19T18:32:18.442Z",
  "producer": "vibenet-demo",
  "entity": "agent.serpradio.route_intelligence",
  "event": "handoff.requested",
  "channel": "handoff",
  "valence": 0.42,
  "energy": 0.66,
  "tension": 0.73,
  "intensity": 0.72,
  "hue": 44,
  "pulse": 0.78,
  "confidence": 0.94,
  "ttl_ms": 15000,
  "metadata": {
    "reason": "source_confidence_below_threshold",
    "route": "JFK-LHR"
  }
}

PROOF LADDER

From one run to a mesh.

Pulse Trace is the live primitive. The full ladder shows how that primitive expands into audible contracts, human-agent handoff, wearable renderers, and cross-device awareness without changing the contract underneath.

01LIVE

VIBEnet Pulse

Hear one agent work.

nominaladvisorywarninghandoffrecovery
What it proves

A human can understand agent state by ear before opening the trace.

02NEXT

Sonic Trace Replay

Replay an agent run as sound.

nominaladvisorywarningrecovery
What it proves

Operational logs can become operational memory instead of static review material.

03NEXT

Audible Contract

Hear whether the agent stayed inside its rules.

nominalwarningcriticalhandoff
What it proves

Governance can become continuous perception instead of a retrospective audit.

04BUILDING

Human-Agent Call and Response

Ask what changed.

advisorywarninghandoff
What it proves

Audio surfaces the moment that matters. Voice explains why it happened.

05NEXT

Ray-Ban Audio Companion

Glasses become agent-awareness earbuds.

nominaladvisorywarninghandoff
What it proves

Post-screen awareness can travel with the operator before any display-native hardware integration lands.

06LAB

Meta POV Agent

The glasses camera gives the agent eyes.

advisorywarninghandoff
What it proves

Wearable context can become audible summary without requiring a dashboard in view.

07NEXT

Snap Spectacles VET Lens

Agent state becomes AR atmosphere.

nominaladvisorywarningrecoveryhandoff
What it proves

The same contract can render spatially as aura, motion, and timing, not only as sound.

08LAB

Crown Cognitive Gate

The agent learns when not to interrupt.

advisorycriticalhandoff
What it proves

Human-agent communication can adapt to attention and focus instead of interrupting on a fixed schedule.

09LAB

Kinesis Agent Control

Thought gesture becomes a low-stakes command layer.

handoff
What it proves

Pauses, annotations, and explanation requests can route through non-verbal control without losing auditability.

10BUILDING

Human-Origin Hybrid Composer

One motif becomes an adaptive agent language.

nominaladvisoryrecovery
What it proves

A proprietary sonic library can stay human in feel while the contract itself remains open.

11NEXT

Cross-Device Mesh

One signal, many renderers.

nominaladvisorywarningcriticalrecoveryopportunityhandoff
What it proves

VIBEnet is a mesh fabric, not a single interface surface.

12BUILDING

SERPRadio Wearable Route Demo

Flight intelligence in your ear.

nominalopportunitywarning
What it proves

The same grammar already fits live domain intelligence, not only synthetic traces.

SERPRadio

SERPRadio is the first live domain adapter.

SERPRadio already maps route intelligence into public pages, real state snapshots, Signal Contract semantics, and audio previews. It proves that the contract is not a thought experiment. It can already drive a live surface.

Hear route intelligence
SERPRadio route proof
231routes monitored live
558sitemap URLs
13audio previews live

One human-origin motif. One live agent. One signal contract. Many devices.

Hear the work before you watch it.

The homepage proves the primitive. The protocol page publishes the object. The Lab shows how that primitive expands into a real roadmap.

Read the specExplore Lab

We'll send one preview, protocol updates, and listening-session invites.