GitHub: not yet
Goal: Systematically augment human sensory input and observe what the brain does with information it was never designed to process. Not virtual reality. Not augmented reality. Augmented perception. The premise: reality is not what we think it is. It's what our senses feed into the brain, and then the brain generates a model from that data. That model is what we call "reality." It feels solid, objective, fundamental. It's not. It's a construction. A best guess. A render. So what happens if you tamper with the input feed? What if you take ultrasonic sound — frequencies above 20kHz that humans literally cannot hear — and fold them back into the audible range? Pipe that signal into headphones. Wear them 24/7. Give the brain data from a spectrum it has no evolutionary context for. And wait. What would happen with an adult brain? The neural pathways are mostly fixed. Maybe nothing. Maybe subtle pattern recognition over weeks. Theoretically, what would happen with a child's brain? Far more plastic, constantly building new neural connections. Would it develop entirely new perceptual categories for this input? Would it learn to "hear" ultrasonic as naturally as we hear speech? If we are just the sum of our input data, what kind of human being would evolve from different input data? This is Real-O-Tron. A multi-stage program to find out.
How did I arrive at this? Two academic frameworks: clinical psychiatry research on altered states of consciousness, and cognitive science on perception. The literature on psychedelics describes a consistent phenomenon: subjects report watching the brain's abstraction layers shift in real-time. The construction mechanism of perceived reality becomes visible. Layers of abstraction tied to specific states of consciousness — observable, repeatable, documentable. Rick Strassman (clinical psychiatrist, "DMT: The Spirit Molecule") conducted the first FDA-approved clinical study and documented consistent, repeatable experiences across subjects. The same "places." The same "entities." Across different sessions, across different people. If a hallucination is repeatable, is it still a hallucination? Or is the brain tuning to a different channel? The repeatability makes it feel deterministic. Which means reality isn't a fixed thing you perceive — it's a frequency you're tuned to. Change the tuning, change the reality. This isn't fringe thinking. Donald Hoffman (cognitive scientist, UC Irvine, "The Case Against Reality") has built a mathematical framework proving that evolution optimizes for fitness, not truth — our senses are an interface, not a window. Strassman's clinical data and Hoffman's mathematical models point at the same conclusion from different angles: reality is constructed, not given. That insight is what Real-O-Tron was designed to test with hardware instead of chemistry. Same hypothesis, different method: modify the brain's input stream and observe what reality becomes. Also: simulation theory. The quantum observer effect — particles existing in superposition until observed, then collapsing into definite states — is basically what every game engine does to save GPU capacity. Don't render what nobody's looking at. In VR, Pimax calls this "foveated rendering." Render high detail only where the eye is focused. The universe appears to do the same thing. If reality is a render, Real-O-Tron is an attempt to feed the renderer data it wasn't expecting.
Real-O-Tron was designed as a multi-stage sensory enhancement program. Each stage adds a new augmented sense. Stack them. See what emerges. STAGE 1: ULTRASONIC HEARING Hardware: Ultrasonic USB microphone, Raspberry Pi (realtime DSP), battery pack, headphones, storage for raw data backup. Method: Capture ultrasonic frequencies (>20kHz). Fold them into the audible range via 10:1 decimation. Mix the translated signal with normal audio using a potentiometer or GUI to control overlay volume — so you can gradually "mix in" the invisible soundscape without being overwhelmed. Duration: 24/7 wear. Weeks to months. Let the brain adapt. Question: Will the brain learn to extract meaningful patterns from ultrasonic data that has been frequency-shifted into its processing range? STAGE 2: INFRARED VISION Hardware: IR camera, industry-grade AR/VR glasses, realtime image processing pipeline. Method: Capture infrared light. Fold it back into the visible spectrum. Overlay it onto normal vision via transparent AR display. Literally become the Predator. Question: Can the brain learn to interpret thermal signatures as a natural part of visual perception? STAGE 3: DIGITAL OLFACTORY ENHANCEMENT Hardware: TBD — electronic nose sensors, scent synthesis. Method: Detect chemical signatures below human threshold. Amplify and translate into perceptible scent cues. Question: Can smell be augmented the way hearing and vision can? STAGE N: STACK THEM ALL Run all augmented senses simultaneously. Long-term. See what kind of consciousness emerges when the brain is processing three or more data streams it never evolved to handle. Floating tanks subtract sensory input to alter consciousness. Real-O-Tron adds it. Same experiment, opposite direction.
What happened:
Nothing. Yet.
The ultrasonic microphone was purchased. The Raspberry Pi was
available. The DSP pipeline was designed in my head. But the
project stayed in the planning phase — detailed, mapped out,
philosophically complete, but never physically assembled.
This is not a dead project. It's a patient one. The hardware is
cheap. The concept is sound. The question is fascinating. It
just needs a block of time where I'm not building FPGA DACs or
training AI companions or organizing Nerd Nites.
Learnings:
- The most interesting projects are the ones that test
fundamental assumptions about reality. Not "can I build
this?" but "what IS this?"
- The "reality enhancement" thread running through my projects
(urGlass, Real-O-Tron, ChordKiller, Predator, Realify) is
not a coincidence. It's the central obsession. Everything
else is a side quest.
- If reality is a render, every sense is an API endpoint.
Real-O-Tron is an attempt to call undocumented endpoints.
Timeline:
- 2025: Concept fully mapped. Hardware partially acquired.
Ultrasonic mic purchased. DSP pipeline designed. Parked
in favor of more immediately buildable projects.
- 2026: Still on the list. Still fascinating. Still patient.
Status: WIP. Waiting for a gap in the build schedule. The
universe isn't going anywhere. Probably.