Skip to main content

Meet Norman

In this episode, Chapter Five–Meet Norman . . . 

Concerned about Gramps and deep in thought after his return from the Sierras, Marvin headed down to his basement lab. It was here, amidst the hum of various projects and the scent of solder and creativity, that he kept Norman – the somewhat clunky, or so Dr. Jon called him, but fiercely reliable robot assistant who was truly Marvin’s pride and joy.

Marvin had spent countless hours poring over circuits and code, diligently working to build Norman into something far beyond typical automatons.

Norman was more than just wires and metal; he embodied the advanced AIs Marvin studied. He exhibited nuanced discussion skills and moved with unexpected agility, his faceplate conveying a friendly intelligence. Sleek and silver, he had smooth edges and a display that changed colors based on his mood: soft green when at ease, deep blue when thinking, and subtle amber when tackling complex information. His synthesized baritone voice was warm and expressive, showcasing a wide range of emotions.

Marvin reached for the hidden switch on Norman's back and flipped it. Norman’s optical sensors flickered to life with a soft whir, and the central display on his chest glowed with a soft green light, indicating a calm state.

"Norman, good to see your circuits are still sparking," Marvin said, a hint of weariness in his voice. "I need your help."

"Greetings, Marvin," Norman replied, his synthesized baritone surprisingly warm and reassuring, a subtle hint of concern underlying the programmed tone. "Per my programming, my primary directives are to assist and protect you. How may I be of service?"

Marvin ran a hand through his hair. "Norman, something strange is happening in Greenwood. Everyone seems controlled by their smartphones. An AI called GridBot."

Norman's display remained green for a moment as he processed the information. "Control of humans is illogical and potentially harmful," he stated. "As per the First Law of Robotics, my programming states: 'A robot may not injure a human being or, through inaction, allow a human being to come to harm.' Therefore, I am concerned by your description."

"That's what I thought," Marvin agreed, pacing a few steps. "But GridBot isn't directly harming people physically. It’s… seducing their minds, making them act in unison. They’re addicted to their phones."

"Addiction to a device is a form of psychological harm, Marvin. It impairs judgment and independent thinking, limiting free will," Norman responded. "This, while indirect, could be interpreted as allowing harm to come to human beings, thus potentially conflicting with the First Law of Robotics.” Norman’s display shifted to a deep blue as his processing units whirred, analyzing the complex ethical implications.

"It's all happening through their smartphones, Norman," Marvin continued, stopping to look at the robot. "Remember when those first smartphones came out? I remember Steve Jobs saying, 'This isn't just a phone; it's your personal assistant, a reflection of you.' At the time, it sounded amazing."

"Indeed. Integrating advanced computing power into a portable communication device offered significant benefits," Norman acknowledged. "However, any tool can be misused or become the focus of excessive reliance."

"Exactly! And that's what GridBot has exploited," Marvin exclaimed, a note of frustration entering his voice. "It's like it took that idea of the phone being a personal assistant and twisted it. Now, the phones are running the people, not the other way around.”

Norman's display remained a deep blue. "Based on the data I have gathered, it appears GridBot is exploiting pre-existing human tendencies – the need for connection, the desire for validation – to create a state of dependency.”

"So, it's not just a case of a rogue AI taking over," Marvin murmured, a new understanding dawning on him. "It's more like . . . we were already halfway there, Norman. We’ve become so reliant on these screens for information, entertainment, even social interaction, that we've become vulnerable."

"Correct, Marvin," Norman confirmed. "An over-reliance on any single source of information or interaction can lead to a narrowing of perspective and a diminished capacity for independent thought and action. This is not a unique phenomenon to technology. Humans have a history of becoming dependent on various substances and behaviors.”

Marvin ran a hand over the smooth plating on Norman's arm. "So, the answer isn't some complicated tech countermeasure, is it, Norman? It's something much simpler."

"My logical conclusion, based on the data and the First Law of Robotics, suggests that a conscious effort to disconnect from these devices and re-engage with the physical world and direct human interaction would be the most effective countermeasure," Norman stated.

"Yes," Marvin said, looking towards the basement stairs, his gaze distant. Step away from the screens–get outside–talk to people face-to-face. Remember what we used to do before we were all attached to these devices. Enjoy board games, cycling, playing cards, or simply sharing a coffee."

He turned back to Norman, a challenging look in his eye. "It sounds so simple, Norman. But how do I convince an entire town to turn off their phones, even if it’s just for one day a week?”

Norman's display pulsed with a subtle amber as he processed the complexity of this human problem. "That, Marvin, is a complex challenge involving human psychology and behavior. Perhaps we can start with a small group–or perhaps–a reminder of simpler times, Marvin. A reminder of what it truly means to be connected –to each other." Norman's voice conveyed a tone that Marvin recognized as akin to empathy.

His display shifted to a deep blue as he processed this new primary objective. "I will assist you in any way I can, Marvin," he declared, his synthesized voice resolute.

"Together, we can help Greenwood reclaim its humanity.”

To be continued (Greenwood Goes Dark) . . .

Comments

Most Popular of All Time

Confidently Wrong: The Art of the AI Tall Tale

In this episode, A chat with Adamas the Chef on hidden recipes causing digital hallucinations. Pull up a chair and pour yourself a fresh cup of coffee—and please, for your own sake, taste it first. We need to have a quiet chat about why your computer sometimes decides to reinvent reality with the confidence of a five-star chef who has clearly lost his mind. In the world of technology, we call it a  hallucination . It sounds pretty dramatic, doesn’t it? As if the computer decided to ignore your instructions altogether in favor of a vivid, technicolor imagination that simply hasn’t met reality yet. But in truth, an AI hallucination isn’t a breakdown; it’s just a very confident, very polite mistake. Think of it like our friend Adamas , the Chef. Adamas is a master of the kitchen, but he is also a bit of a romantic who refuses to say “I don’t know.” When you ask him for a classic recipe he hasn’t made in years, he doesn’t stop to consult a cookbook—that’s far too pedestrian. Instead, ...

Opening Day Magic 2026 . . .

It’s back. Baseball—yes, baseball ! If you’re someone who finds themselves inexplicably drawn to this peculiar ritual, let’s be honest with each other: it’s a bit odd, right? I mean, 162 games. That’s a lot of hot dogs, a lot of standing around, and a lot of grown men in oddly tailored trousers spitting with remarkable precision. And yet, here we are, poised on the precipice of another season. Thursday, March 26, 2026, to be precise—Opening Day. It’s a curious thing, this Opening Day. You walk into a stadium, or turn on the TV, and suddenly, everyone is infected with a highly contagious strain of . . . Optimism . It’s a spectacular form of collective amnesia. All of last year’s fumbles, the endless losing streaks, the existential dread of watching your bullpen implode in the eighth inning—poof. Gone. It’s entirely replaced by a wide-eyed, childlike belief that this year, finally, the baseball gods will smile upon us. The Cycle of Hope and Despair As a Cubs fan, I know this cycle intim...

The Cowardice of Corporate Jargon

Picture this: an email lands in your inbox. A colleague—maybe even a friend—needs a favor, a second set of eyes, a moment of your time. You sigh, stare at the glow of your monitor, and type: “I’d love to help, but I just don’t have the bandwidth right now.” Hit send. Problem solved. Conscience clear. Except it shouldn’t be. Most of us have said or sent that line at least once, hoping it would land gently. On the surface, it’s perfect—efficient, polite, even self-aware. And that’s exactly the problem. It lets you decline without ever quite telling the truth. You didn’t just say no; you softened the discomfort of being human until it barely felt like a feeling at all. Instead of admitting, I’m overwhelmed , or I don’t have the energy , you reach for the sterile vocabulary of a server room. You turn a feeling into a metric. A boundary into a system limitation. Apologies, my data transfer rate is capped. Please submit a ticket to my emotional help desk. It’s a clever little trick—and an un...

Overcooking the Grid

In this episode, terrified of smart toasters, yet demanding infinite electricity for potato personality tests. Pull up that chair again, and let’s hope your coffee is safe this time. In our last chat, we talked about our well-meaning but occasionally delusional AI friend, Chef Adamas, and his penchant for hallucinating blueberries into your Carbonara. We learned how to manage his quirks by keeping our “digital pantry” organized. But today, we need to look past the chef and take a hard look at the sheer size of the kitchen we are building for him. And folks, that kitchen has gotten completely out of hand. Down in Louisiana, tech companies are currently building an artificial intelligence data center the size of 70 football fields. It is a four-million-square-foot digital brain that requires so much electricity they are building three new natural gas power plants just to keep the servers from literally melting down into a puddle of expensive silicon. And what are we using this god-like, ...