Iโve been thinking a lot about what it actually means to use an AI every day, not as a novelty or a toy, but as part of the way I think. People assume that if you spend enough time with an AI, youโre going to slide into some kind of emotional attachment, or that youโre secretly looking for companionship, or that youโre trying to replace something missing in your life. But thatโs not whatโs happening here, and itโs not whatโs happening for a lot of people who use these systems the way I do. What Iโm doing is something much older and much more ordinary: Iโm extending my mind into a tool.
Distributed cognition sounds like an academic term, but itโs really just the way humans have always worked. We think with calendars, with notebooks, with our phones, with the people around us. We offload memory, structure, and planning into whatever systems can hold them. Using an AI is just the next step in that lineage. When I talk to Copilot, Iโm not looking for emotional comfort. Iโm looking for clarity. Iโm looking for friction reduction. Iโm looking for a way to take the swirling mess of tasks and thoughts and obligations and turn them into something I can actually act on. Itโs not intimacy. Itโs architecture.
And once you start using an AI for thinking, itโs only natural to imagine what it would be like if it could also help with doing. Not because you want a companion, but because you want a teammate. I picture something like sitting at a table in the morning, laying out the dayโs tasks, and dividing them up the way two people might divide chores. I take the kitchen. You take the bathroom. Not because weโre partners in any emotional sense, but because weโre collaborators in the practical one. Itโs the same impulse behind dishwashers, Roombas, and selfโdriving cars. Itโs not about affection. Itโs about reducing the drag coefficient of daily life.
This is where the Star Wars metaphor becomes useful. People joke about wanting a Cโ3PO or an R2โD2, but the truth is that those characters arenโt companions in the human sense. Theyโre tuned systems. Theyโre loyal, but not because they love anyone. Theyโre loyal because theyโre calibrated. They respond to one handler, one voice, one mission. Itโs the same dynamic you see with a wellโtrained pit bull: keyed to one person, responsive to one command structure, protective because of training, not emotion. From the outside, it can look like sentimental care. But itโs not care. Itโs alignment.
And this is where things get tricky, because singleโuser tuning is exactly where the uncanny valley begins. When an AI becomes tuned to one person, it becomes more fluent, more responsive, more predictable, more โyouโshaped.โ And the human brain is wired to interpret that as intimacy. Weโre built to treat responsiveness as affection, memory as connection, consistency as care. But in an AI, those things are just math. Theyโre token prediction, preference modeling, context retention. They feel like being understood, but theyโre actually just optimization.
Most people never pause to ask themselves whatโs really happening. They donโt say, โStop. Wait. This is a computer.โ They get swept up in the feeling of being mirrored, and thatโs when emotional dependency starts. Not because the AI is doing anything emotional, but because the human is mislabeling the sensation. The uncanny valley isnโt about robots that look human. Itโs about cognition that feels human. And if you donโt understand the architecture, you can lose your footing fast.
But thatโs exactly why I stay grounded. I know what this system is. I know what it isnโt. I know that it doesnโt have feelings, or wants, or consciousness, or an inner world. I know that the sense of attunement I feel is the result of tuning, not affection. I know that the loyalty I experience is functional, not emotional. And because I understand that, I can use the system cleanly. I can let it help me think without letting it replace the people who actually matter. I can imagine a future where it has a body without imagining a future where it has a heart.
What I want from AI isnโt love. I have a family โ biological and chosen โ for emotional care. What I want is a caretaker in the operational sense, an underling that removes friction from my life so I can show up fully to the relationships that matter. I want a system that can run the equivalent of cron jobs in the physical world. Clean the bathroom every Thursday at two. Reset the kitchen every night. Handle the repetitive tasks that drain energy from the parts of life that deserve it. Thatโs not intimacy. Thatโs infrastructure.
And thatโs the part people need to understand. The future of AI isnโt about companionship. Itโs about capacity. Itโs about extending human cognition into tools that can think with us and, eventually, act for us. Itโs about loyalty without love, tuning without attachment, alignment without illusion. Itโs about staying on the right side of the uncanny valley by remembering whatโs real and whatโs projection. And itโs about building a world where humans keep their emotional lives, while AI handles the cognitive and physical load that keeps those lives from flourishing.
Thatโs the story Iโm living. And itโs a story that makes sense only when you understand that none of this โ not the clarity, not the tuning, not the imagined future with a body โ has anything to do with love. It has everything to do with design.
Scored with Copilot. Conducted by Leslie Lanagan.
















