◆ AIUNITES|AIZinesAIByJobRedomyVideoBateVoiceStryFurnishThingsBizStryAI YHWHCloudsionGameaticaUptownITInThisWorldERPiseERPizeAITSQL🌌 COSMOS

Notation
Specification

Three open protocols for describing what the human body does. MNN captures physical movement — muscle, nerve, joint, vector. VRN captures vocal production — larynx, tongue, breath, resonance. VNN maps the neural control behind VRN — cranial nerves, motor cortex, brainstem circuits. All are plain text, machine-parseable, and vendor-neutral.

3Protocols
MNNMovement
VRNVoice
VNNNeural

Specification Contents

  1. MNN — Muscular Neuro Notation
  2. MNN String Format
  3. Contraction Layer
  4. Nerve & Innervation Layer
  5. Position & Joint Layer
  6. Vector & Resistance Layer
  7. MNN Consumer Sites
  8. Prior Art
  9. VRN — Vocal Resonance Notation
  10. VRN Parameter Space
  11. VRN and AI Voice Synthesis
  12. VRN Consumer Sites
  13. VNN — Vocal Neuro Notation
  14. The 6 Cranial Nerves of Singing
  15. VRN → VNN Mapping
  16. Antagonist Balance Notation
  17. Clinical Status Codes
  18. Shared Design Principles

🏋️ MNN — Muscular Neuro Notation

An open text format for describing human physical movement. Captures muscle activation, nerve pathways, joint positions, and resistance vectors in a single string.

MNN §1

String Format

An MNN string is a single line of text composed of tagged segments. Each segment is wrapped in brackets and prefixed with a tag identifier. The segments can appear in any order — parsers identify them by tag, not position. A minimal MNN string needs only a movement tag and a contraction tag. All other segments are optional and additive.

The notation is designed to be human-authored. Unlike motion capture formats (BVH, C3D) that are machine-generated time-series data, MNN is written by a person describing what they did, what they felt, or what they want a system to reproduce. It is closer to musical notation than to a data dump.

{Push.H} [Con:Pec.S+++, Dlt.A+] → MedPec/Axil
[Pos:L.Sh(IR:25,Flex:90)] [Vec:H:Mid,A:0°,Src:Cable] Complete MNN string — movement, contraction, nerve, position, vector
Segment Tags
{...}    Movement tag — exercise class and variant
[Con:...] Contraction — muscles and activation levels
→ ...     Nerve — innervation pathways
[Pos:...] Position — joint angles and orientations
[Vec:...] Vector — resistance source, direction, height
[Comp:...]Compensation — observed compensations and flags
[Meta:...]Metadata — sets, reps, load, RPE, timestamp
MNN §2

Contraction Layer

The contraction segment identifies which muscles are active and how intensely. Muscles are referenced by abbreviated anatomical names. Activation levels use a + scale where more symbols mean greater engagement. The primary mover is listed first, synergists follow.

Contraction types are implicit in the movement tag but can be made explicit: concentric (muscle shortening under load), eccentric (muscle lengthening under load), isometric (static hold), and ballistic (explosive). This matters for rehabilitation logging where the distinction between concentric and eccentric loading of a tendon is clinically significant.

AbbreviationMuscleRegion
Pec.SPectoralis major (sternal head)Chest
Pec.CPectoralis major (clavicular head)Upper chest
Dlt.ADeltoid (anterior)Front shoulder
Dlt.LDeltoid (lateral)Side shoulder
Dlt.PDeltoid (posterior)Rear shoulder
LatLatissimus dorsiBack
BicBiceps brachiiUpper arm
TriTriceps brachiiUpper arm
GluGluteus maximusHip
QuadQuadricepsThigh
HamHamstringsThigh
Activation Scale
+     Light activation (stabilizer, minor synergist)
++    Moderate activation (synergist)
+++   Strong activation (primary mover)
++++  Maximal activation (1RM effort, peak contraction)
MNN §3

Nerve & Innervation Layer

The nerve segment records which peripheral nerves are responsible for the movement. This is optional for general fitness logging but critical for rehabilitation and clinical documentation. A physical therapist tracking a patient with a brachial plexus injury needs to know which nerve pathways are being loaded, not just which muscles are moving.

Nerve annotations also enable tracking of nerve flare-ups alongside sets. If a user reports radiculopathy or paresthesia during a specific movement, the MNN string captures both the movement pattern and the affected nerve — creating a record that a clinician can use for treatment planning.

AbbreviationNerveMuscles Served
MedPecMedial pectoral nervePectoralis major (sternal)
LatPecLateral pectoral nervePectoralis major (clavicular)
AxilAxillary nerveDeltoid, teres minor
MuscMusculocutaneous nerveBiceps, brachialis
RadRadial nerveTriceps, wrist extensors
SciTibSciatic → tibialHamstrings, calf
FemFemoral nerveQuadriceps
SupGluSuperior gluteal nerveGluteus medius/minimus
InfGluInferior gluteal nerveGluteus maximus
MNN §4

Position & Joint Layer

The position segment describes the orientation of joints during the movement. Joints are referenced by abbreviated names, and rotations are expressed in degrees using standard anatomical terms: flexion/extension, abduction/adduction, internal rotation/external rotation.

This is the layer that makes MNN work across domains. The same position data that a trainer uses to describe a cable fly angle is the data that InThisWorld uses to pose an avatar. Position values map directly to Euler rotations in a 3D engine — the translation from [Pos:L.Sh(IR:25,Flex:90)] to llEuler2Rot() is mechanical, not interpretive.

Position Syntax
[Pos:<Side>.<Joint>(<Rotation>:<Degrees>, ...)]

Side:   L = left, R = right, B = bilateral
Joint:  Sh = shoulder, El = elbow, Wr = wrist,
        Hp = hip, Kn = knee, An = ankle, Sp = spine

Rotation types:
  Flex  = flexion         Ext  = extension
  Abd   = abduction       Add  = adduction
  IR    = internal rot.   ER   = external rot.
  Lat   = lateral flex.   Rot  = axial rotation

Example:
  [Pos:L.Sh(IR:25,Flex:90)]
  Left shoulder: 25° internal rotation, 90° flexion
JointCodeTypical Ranges
ShoulderShFlex 0–180°, Abd 0–180°, IR/ER 0–90°
ElbowElFlex 0–145°
HipHpFlex 0–125°, Abd 0–45°, IR/ER 0–45°
KneeKnFlex 0–140°
AnkleAnDorsiflex 0–20°, Plantarflex 0–50°
SpineSpFlex 0–80°, Lat 0–35°, Rot 0–45°
MNN §5

Vector & Resistance Layer

The vector segment describes where the resistance comes from, at what height, and at what angle. This is the layer that makes MNN work for cable rigs, robotic rehabilitation, and remote-controlled equipment. A BODWAVE-compatible cable system can read a vector segment and set its pulley to the correct height and angle to reproduce the prescribed movement pattern.

The distinction between a high cable fly, a mid cable fly, and a low cable fly isn't just about the name — it's about the line of resistance relative to the muscle's force vector. The vector segment captures this precisely. A system that knows the height, angle, and source can reproduce the resistance profile on any compatible equipment.

Vector Syntax
[Vec:H:<Height>, A:<Angle>, Src:<Source>]

Height:  High | Mid | Low | Floor | <cm>
Angle:    = straight ahead, 90° = from the side
Source:  Cable | Band | Gravity | Manual | Machine

Examples:
  [Vec:H:High,A:15°,Src:Cable]   High cable, slight angle
  [Vec:H:Floor,A:0°,Src:Band]    Floor-anchored band
  [Vec:H:180cm,A:0°,Src:Cable]   Specific pulley height
MNN §6

Consumer Sites

MNN is consumed by three AIUNITES sites, each using a different subset of the notation:

🏋️

BodSpas

BODWAVE Movement Layer

Full MNN support. Workout logging with contraction, nerve, position, and vector data. The MNN Builder lets users construct strings visually. Tracks nerve flare-ups alongside training loads. Drives BODWAVE-compatible cable equipment.

🌍

InThisWorld

inthisworld.com

Consumes the Position layer. Translates joint angles from MNN strings into avatar pose data via Three.js. The same string that logs a shoulder press in BodSpas can position an avatar's arm in InThisWorld. Bridges to LSL (Linden Scripting Language) for Second Life / OpenSim.

🎮

Gameatica

Educational Reference

References MNN muscle and nerve structures in anatomy-related educational games. Students interact with the same naming conventions used in the clinical notation.

MNN §7

Prior Art

Existing standards each capture one layer of movement data. MNN combines all of them into a single string. This is not a criticism of prior work — each system was designed for a specific domain. MNN's contribution is the integration layer that makes movement data portable across domains.

DomainStandardCapturesLacks
Joint anglesISB JCSPer-joint Euler rotationsNo muscle/nerve, no text format
Motion dataC3D / BVH / OpenSimFull-body time-seriesNo semantic layer, not human-authored
Muscle activityEMG + SENIAMVoltage waveformsNo symbol table, no nerve mapping
Exercise doseACSM FITT / NSCASets × reps × loadNo joint position, no targeting
ChoreographyLabanotationSpatial path, effort qualityNo nerve, no muscle, no vector
All of the aboveMNNMuscle + nerve + joint + vector + compensation

🎤 VRN — Vocal Resonance Notation

An open parameter space for describing the physical production of vocal sound. Captures what the vocal apparatus is doing — not just what it sounds like.

VRN §1

Parameter Space

VRN describes voice production through the physical configuration of the vocal tract. Each VRN record is a set of parameters representing the state of the larynx, vocal folds, tongue, jaw, lips, soft palate, and breath system at a given moment. Together they define a complete vocal gesture — everything the body is doing to produce a specific sound.

The parameters are grouped into three categories: source (what the vocal folds are doing), filter (how the vocal tract shapes the sound), and drive (the breath energy powering it). This maps to the source-filter model of speech production — the same theoretical framework used in acoustic phonetics and speech synthesis research.

VRN Record — Full Example
{
  // SOURCE — vocal fold behavior
  "vocal_fold":      "modal",       // breathy, modal, pressed, falsetto, fry
  "pitch_hz":        262,           // fundamental frequency
  "vibrato": {
    "rate_hz": 5.5,                  // oscillation speed
    "extent_cents": 40              // pitch deviation
  },

  // FILTER — vocal tract shape
  "larynx_height":   "neutral",     // low, neutral, high
  "soft_palate":     "raised",      // raised (oral), lowered (nasal)
  "tongue_body":     "mid_front",   // position in vowel space
  "tongue_tip":      "alveolar",    // contact point for consonants
  "jaw_opening":     45,            // degrees, 0 = closed, 60 = max
  "lip_rounding":    0.3,           // 0 = spread, 1 = fully rounded
  "formants": {
    "F1": 500,  "F2": 1800,  "F3": 2600   // Hz targets
  },

  // DRIVE — breath system
  "breath_pressure": 60,            // subglottal pressure, 0-100
  "duration_ms":     800
}
ParameterCategoryTypeDescription
vocal_foldSourcestringPhonation mode: breathy, modal, pressed, falsetto, fry
pitch_hzSourcenumberFundamental frequency in Hz
vibratoSourceobjectRate (Hz) and extent (cents) of pitch oscillation
larynx_heightFilterstringVertical larynx position: low, neutral, high
soft_palateFilterstringVelopharyngeal port: raised (oral) or lowered (nasal)
tongue_bodyFilterstringVowel space position: high_front, mid_central, low_back, etc.
tongue_tipFilterstringArticulation point: dental, alveolar, postalveolar, retroflex
jaw_openingFilternumberJaw angle in degrees (0 = closed, 60 = max open)
lip_roundingFilternumberLip shape on 0–1 scale (0 = spread, 1 = fully rounded)
formantsFilterobjectTarget formant frequencies F1, F2, F3 in Hz
breath_pressureDrivenumberSubglottal pressure on 0–100 scale
duration_msDrivenumberDuration of the vocal gesture in milliseconds
VRN §2

VRN and AI Voice Synthesis

Current AI voice synthesis engines — ElevenLabs, OpenAI TTS, Google Cloud TTS, Amazon Polly — all model voice as an acoustic waveform transformation. They take a text input, select or clone a voice identity, and produce audio. Some accept style parameters (speed, emphasis, emotion). None of them model how the voice is physically produced.

This means these systems can produce convincing speech, but they cannot answer questions like: "What is the larynx doing during this vowel?" or "What tongue position produces this formant shift?" The physical reality of vocal production is invisible to them — it was never part of their parameter space.

VRN fills this gap. It provides a parameter space that maps to the physiological mechanics of vocalization. A VRN record doesn't describe what a voice sounds like — it describes what the body is doing to produce that sound. This is a fundamentally different representation, and it opens a different research direction: physically-informed voice synthesis where a model generates speech by simulating vocal tract configuration rather than pattern-matching spectrograms.

For vocal training (the current primary use case on VoiceStry), VRN enables instruction that goes beyond "make it sound like this." A vocal coach can describe the exact physical configuration that produces a target sound — larynx low, soft palate raised, tongue body forward, specific breath pressure — and the student can work toward reproducing that configuration, not just imitating the acoustic output.

The gap in current AI voice technology: Existing engines have no model of physical vocal production. They synthesize voice from acoustic features alone. VRN provides the missing parameter space for physically-informed synthesis — a model that generates speech from vocal tract configuration rather than purely from audio. This is an open research direction with applications in voice training, speech therapy, accent coaching, singing synthesis, and accessible voice technology.
VRN §3

Consumer Sites

VRN is consumed by two AIUNITES sites:

🎤

VoiceStry

Vocal Training Platform

The primary VRN consumer. Vocal gym, pitch trainer, and sight-reading exercises all use VRN to describe target vocal configurations. Students work toward reproducing specific physical states, not just matching audio.

🌌

COSMOS the Opera

Origin of VRN

VRN was created to notate specific vocal techniques for operatic performance in COSMOS. The notation needed to capture what a singer's body does — not just the notes on a staff. COSMOS remains the creative testbed for VRN features before they generalize to VoiceStry.

🧠 VNN — Vocal Neuro Notation

The neural control layer behind VRN. VRN maps what happens mechanically. VNN maps what fires to make it happen — cranial nerves, motor cortex pathways, brainstem circuits, and antagonist muscle balance.

VNN §1

Two Layers of Vocal Production

Traditional vocal pedagogy tells you what to do: open the throat, lift the palate, support from the diaphragm. VRN gives you symbols for these mechanical events. VNN adds the deeper layer — the nervous system that actually executes those commands.

Every VRN symbol implies a VNN command chain. When a score says [Zy] (zygomatic lift), the facial nerve (CN VII) is sending signals to the zygomaticus major. When it says [D+++] (maximum diaphragm), the phrenic nerve (C3–C5) is driving the primary breathing muscle. VRN is the score. VNN is the orchestra.

Mechanical Layer (VRN) vs Neural Layer (VNN)
VRN — Mechanical (what happens):
[H+++, N++, O++, Sq+, Vib.r6]
Resonance placement, fold config, airflow, vibrato rate

VNN — Neural (what fires):
CN X → RLN → fold adduction
CN VII → zygomatic lift
C3-C5 → diaphragm
Which nerves fire, which muscles contract, which brain regions coordinate
The Neural Command Chain
Motor Cortex  → Voluntary intent
Broca's Area → Motor planning
Basal Ganglia→ Learned sequences (trained VRN patterns)
Cerebellum   → Timing & precision
Brainstem    → Nerve relay
Cranial Nerves→ Muscle activation
Sound        → VRN output

Total latency: ~50–100ms from intent to sound
VNN §2

The 6 Cranial Nerves of Singing

Six cranial nerves form the core wiring harness of the vocal instrument. Every VRN symbol traces back to one or more of these nerves.

NerveNameRoleVRN Symbols Controlled
CN VTrigeminalJaw positioning — temporalis, masseter, pterygoids[Jw↓] [Jw→] [Jw~] [Jw:X°]
CN VIIFacialLips, cheeks, embouchure — orbicularis oris, zygomaticus[Lp] [Zy] [Ey] [B]
CN IXGlossopharyngealPharynx elevation — stylopharyngeus[P] [P+] [P++] [P+++]
CN XVagusMaster controller — larynx, soft palate, all fold behavior via RLN and SLN branches[Th] [Tn] [Fl] [Prs] [Br] [Gl] [Co] [Vib] [SP] [Sq]
CN XIAccessoryPosture support — sternocleidomastoid, trapezius[M] posture & laryngeal stability
CN XIIHypoglossalSole motor nerve to tongue — all intrinsic/extrinsic tongue muscles[Tg] [Tg↑] [Tg↓] [Tg→F] [Tg→B]

The Vagus Nerve (CN X) deserves special attention. Its two critical branches divide the work of phonation: the recurrent laryngeal nerve (RLN) controls 4 of 5 intrinsic laryngeal muscles (fold adduction, abduction, body tension, closure), while the superior laryngeal nerve (SLN) controls the cricothyroid — the pitch muscle that stretches and thins the folds. Every register transition involves a shift in RLN vs SLN dominance.

Vagus Nerve Branch Mapping
Recurrent Laryngeal (RLN)4 muscles
  Thyroarytenoid     → fold body tension     [Th]
  Lateral Cricoarytenoid → fold adduction   [Prs]
  Posterior Cricoarytenoid → fold abduction  [Br]
  Interarytenoid     → fold closure          [Fl] [Gl]

Superior Laryngeal (SLN)1 muscle
  Cricothyroid       → fold stretch/thinning [Tn]
  The "pitch" muscle — controls fold length and tension
VNN §3

VRN → VNN Mapping

Every VRN symbol traces to a specific nerve or nerve combination. VNN organizes these mappings into tiers based on who needs them: singers (Level 1), teachers (Level 2), and clinicians (Level 3).

TierAudienceWhat It Adds
🟢 Level 1SingersVRN symbol → nerve → muscle (core mapping)
🟡 Level 2Teachers & coaches+ Articulatory position codes, antagonist balance, neural intensity
🔴 Level 3Clinicians, ENTs, SLPs, AI researchers+ Clinical status codes, pathological markers, recovery tracking
Level 1 — Core VRN → Nerve Mapping (Sample)
Resonance & Placement:
[C]  Chest resonance    → CN X (RLN)    → Thyroarytenoid
[H]  Head resonance     → CN X (SLN)    → Cricothyroid
[N]  Nasal resonance    → CN X (pharyn) → Levator veli palatini
[O]  Oral resonance     → CN XII + VII  → Tongue + lips
[P]  Pharyngeal space   → CN IX + X     → Stylopharyngeus

Fold Behavior:
[Th] Thick fold         → CN X (RLN)    → Thyroarytenoid
[Tn] Thin fold          → CN X (SLN)    → Cricothyroid
[Fl] Flow phonation     → CN X (RLN)    → Balanced IA + LCA
[Prs]Pressed phonation  → CN X (RLN)    → LCA hyperadduction
[Br] Breathy phonation  → CN X (RLN)    → PCA partial abduction

Breath & Airflow:
[D]  Diaphragm          → Phrenic C3-C5 → Diaphragm
[IC] Intercostal        → Intercostal T1-T11 → External intercostals
[Ap] Appoggio           → Phrenic + IC  → Diaphragm vs intercostal antagonism

Articulation:
[Tg] Tongue             → CN XII        → Genioglossus, styloglossus, etc.
[SP] Soft palate        → CN X (pharyn) → Levator veli palatini
[Jw] Jaw                → CN V (motor)  → Temporalis, masseter
[Lp] Lips               → CN VII        → Orbicularis oris
[Zy] Zygomatic lift     → CN VII (zyg)  → Zygomaticus major
VNN §4

Antagonist Balance Notation

Vocal production depends on muscle pairs that oppose each other. The balance point between them determines the sound. This is the vocal equivalent of MNN's agonist/antagonist pairing — and it's where most vocal problems live. VNN uses a ratio notation to express the balance between antagonist pairs.

PairAgonistAntagonistWhat the Balance Controls
CT:TACricothyroid (SLN)Thyroarytenoid (RLN)Fold length vs mass — the pitch/register mechanism. CT-dominant = head voice. TA-dominant = chest voice.
LCA:PCALat. Cricoarytenoid (RLN)Post. Cricoarytenoid (RLN)Fold adduction vs abduction — the open/close mechanism. LCA-dominant = pressed. PCA-dominant = breathy.
Elev:DeprSuprahyoid (CN V, VII, XII)Infrahyoid (Ansa cervicalis)Larynx height — elevators pull up (belt), depressors pull down (cover).
D:ICDiaphragm (Phrenic)Intercostals + AbdominalsInhalation vs exhalation force — appoggio is the trained balance.
Gn:StGenioglossus (CN XII)Styloglossus (CN XII)Tongue forward vs back — formant shaping.
Ratio Notation Examples
CT:TA 70:30   CT-dominant (head mix)
CT:TA 30:70   TA-dominant (chest belt)
CT:TA 50:50   Balanced (neutral mix, speech-quality phonation)
LCA:PCA 80:20 Firm adduction (pressed phonation)
LCA:PCA 40:60 Loose adduction (breathy phonation)
D:IC 60:40    Diaphragm-dominant breath (strong appoggio)

Full example — Classical Tenor High A:
[H+++, Sq+, Cov, Lx↓, CT:TA 65:35, LCA:PCA 70:30, D:IC 60:40, SP↑, Zy, Tg→B]
Head-dominant mix with squillo ring. Larynx low, palate high,
tongue retracted. Firm adduction, strong appoggio.
VNN §5

Clinical Status Codes

For ENTs, speech-language pathologists, and vocal therapists. These codes document pathology, track recovery, and identify compensation patterns — the vocal equivalent of MNN's clinical layer.

SymbolStatusExample
Paralysis / no activation[Fold.L❌] → RLN — left fold paralysis
Par:Paresis (partial weakness)[Par:Fold.L–] → RLN — reduced adduction
Atr:Atrophy present[Atr:Fold.R] → SLN — fold bowing from muscle wasting
Spas:Spasmodic / involuntary contraction[Spas:LCA] → RLN — adductor spasmodic dysphonia
Trem:Tremor[Trem:CT, Vib.irreg] — cricothyroid tremor
Comp:Compensation pattern[Comp:FVF for Fold.L❌] — false folds compensating
Rec↑Recovery / reinnervation[Rec↑:Fold.L+, LCA:PCA 50:50] — improving post-surgery
Les:Lesion present[Les:Fold.R.ant⅓] — nodule location
Gap:Glottal gap[Gap:Post, Br++] — incomplete closure
Clinical Documentation Example
Post-Thyroidectomy Assessment (Day 14):
[Par:Fold.L–, Gap:Post, Br++, Comp:Trp.U++] → RLN (left)
[Rec↑:Fold.L, LCA:PCA 45:55]

Left fold paresis from RLN damage during thyroid surgery.
Posterior glottal gap causing breathiness.
Compensating with upper trapezius tension (neck strain).
Early signs of reinnervation — adduction improving. Compare at 6 weeks.
Why VNN matters for AI: Current voice synthesis has no concept of neural control pathways. VNN provides the structured parameter space that maps vocal output to neuromuscular commands — enabling AI systems that understand how voices are produced at the neural level, not just what they sound like acoustically. Combined with VRN's mechanical layer, this creates a complete three-layer model: intention → neural command → mechanical output → sound.

The spec above covers the core VNN framework. The full VNN reference — including all Level 2 articulatory position codes (tongue, larynx height, jaw, palate), the complete vibrato/squillo/timbral color mappings, the emotional brain pathway (periaqueductal gray), vagus nerve and wellbeing research, and VNN as a neural training framework — lives on VoiceStry:

Read the full VNN reference on VoiceStry →

Shared

Design Principles

MNN and VRN share the same design philosophy. These aren't accidental similarities — they're deliberate constraints that make both notations work across contexts.

📄

Plain Text

Both notations are plain text. No binary formats, no proprietary readers, no special software required. Open the file in any text editor and you can read it. This is a non-negotiable design constraint.

🤖

Machine-Parseable

Both notations use consistent syntax with tagged segments and structured key-value pairs. Any programming language with basic string parsing can read them. MNN has a formal EBNF grammar.

👤

Human-Authored

Unlike motion capture or EMG data (machine-generated), both notations are designed to be written by a person. A trainer writes MNN. A vocal coach writes VRN. The notation serves the human, not the sensor.

🔀

Parameter Space, Not Vocabulary

Both notations define a parameter space rather than a fixed dictionary. New muscles, new vocal techniques, new joint configurations can be expressed without extending the spec. The notation grows with use.

🌐

Cross-Domain by Design

MNN works in a gym, a game engine, and a cable rig. VRN works in a voice lesson, an opera score, and an AI synthesis pipeline. The same notation, different consumers. That's the point.

Try the Notation

Build an MNN string, train your voice, explore the neural mapping, or browse the network.