Back to articles
OROSYNC: Dismantling the Keyboard Tax with the Vifi Multimodal Agent
NewsTools

OROSYNC: Dismantling the Keyboard Tax with the Vifi Multimodal Agent

via Dev.toAdan Abdi Alinoor

The Project OROSYNC is an "Ab Initio" multimodal ecosystem designed to return commerce to its human-centric, oral default. Built in Google AI Studio using the Multimodal Live API, OROSYNC introduces Vifi (Vy-Fy)—an agent that sees, hears, and talks—to liberate merchants from the "Keyboard Tax." The Reflections During this challenge, I moved beyond standard LLM prompting into Multimodal Agentic Orchestration. The breakthrough was using Gemini 3.1 Pro to bridge the gap between chaotic human speech and deterministic financial records. What I Built: Vifi (Interface): A real-time agent utilizing Acoustic Ingestion and VoicePass (a visual lip-reading authentication protocol for public-space privacy). OROTALLY (Financial): A deterministic bookkeeping engine that maps oral intent to the AP2 (Agent Payments Protocol) for secure G-Pay settlement. OROcom (Identity): A communication agent using the Universal Commerce Protocol (UCP) to transform business data into professional digital identity. The

Continue reading on Dev.to

Opens in a new tab

Read Full Article
7 views

Related Articles