Blog

  • The Death of Grunt Work and the Starvation of Personality

    The Death of Grunt Work and the Starvation of Personality

    Personality Starvation

    Personality Starvation is the gradual erosion of character, depth, and individuality caused by the systematic removal of struggle, responsibility, and formative labor from human development. It occurs when friction—failure, boredom, repetition, social risk, and unglamorous work—is replaced by automation, optimization, and AI-assisted shortcuts that produce results without demanding personal investment. In a state of personality starvation, individuals may appear competent, efficient, and productive, yet lack the resilience, humility, patience, and textured inner life from which originality and meaning emerge. Because personality is forged through effort rather than output, a culture that eliminates its own “grunt work” does not liberate talent; it malnourishes it, leaving behind polished performers with underdeveloped selves and an artistic, intellectual, and moral ecosystem increasingly thin, fragile, and interchangeable.

    ***

    Nick Geisler’s essay, “The Problem With Letting AI Do the Grunt Work,” reads like a dispatch from a vanished ecosystem—the intellectual tide pools where writers once learned to breathe. Early in his career, Geisler cranked out disposable magazine pieces about lipstick shades, entomophagy, and regional accents. It wasn’t glamorous, and it certainly wasn’t lucrative. But it was formative. As he puts it, he learned how to write a clean sentence, structure information logically, and adjust tone to an audience—skills he now uses daily in screenwriting, film editing, and communications. The insultingly mundane work was the work. It trained his eye, disciplined his prose, and toughened his temperament. Today, that apprenticeship ladder has been kicked away. AI now writes the fluff, the promos, the documentary drafts, the script notes—the very terrain where writers once earned their calluses. Entry-level writing jobs haven’t evolved; they’ve evaporated. And with them goes the slow, character-building ascent that turns amateurs into artists.

    Geisler calls this what it is: an extinction event. He cites a study that estimates that more than 200,000 entertainment-industry jobs in the U.S. could be disrupted by AI as early as 2026. Defenders of automation insist this is liberation—that by outsourcing the drudgery, artists will finally be free to focus on their “real work.” This is a fantasy peddled by people who have never made anything worth keeping. Grunt work is not an obstacle to art; it is the forge. It builds grit, patience, humility, social intelligence, and—most importantly—personality. Art doesn’t emerge from frictionless efficiency; it emerges from temperament shaped under pressure. A personality raised inside a Frictionless Dome, shielded from boredom, rejection, and repetition, will produce work as thin and sterile as its upbringing. Sartre had it right: to be fully human, you have to get your hands dirty. Clean hands aren’t a sign of progress. They’re evidence of starvation.

  • Against AI Moral Optimism: Why Tristan Harris Underestimates Power

    Against AI Moral Optimism: Why Tristan Harris Underestimates Power

    Clarity Idealism

    noun

    Clarity Idealism, in the context of AI and the future of humanity, is the belief that sufficiently explaining the stakes of artificial intelligence—its risks, incentives, and long-term consequences—will naturally lead societies, institutions, and leaders to act responsibly. It assumes that confusion is the core threat and that once humanity “sees clearly,” agency and ethical restraint will follow. What this view underestimates is how power actually operates in technological systems. Clarity does not neutralize domination, profit-seeking, or geopolitical rivalry; it often accelerates them. In the AI era, bad actors do not require ignorance to behave destructively—they require capability, leverage, and advantage, all of which clarity can enhance. Clarity Idealism mistakes awareness for wisdom and shared knowledge for shared values, ignoring the historical reality that humans routinely understand the dangers of their tools and proceed anyway. In the race to build ever more powerful AI, clarity may illuminate the cliff—but it does not prevent those intoxicated by power from pressing the accelerator.

    Tristan Harris takes the TED stage like a man standing at the shoreline, shouting warnings as a tidal wave gathers behind him. Social media, he says, was merely a warm-up act—a puddle compared to the ocean of impact AI is about to unleash. We are at a civilizational fork in the road. One path is open-source AI, where powerful tools scatter freely and inevitably fall into the hands of bad actors, lunatics, and ideologues who mistake chaos for freedom. The other path is closed-source AI, where a small priesthood of corporations and states hoard godlike power and call it “safety.” Either route, mishandled, ends in dystopia. Harris’s plea is urgent and sincere: we must not repeat the social-media catastrophe, where engagement metrics metastasized into addiction, outrage, polarization, and civic rot. AI, he argues, demands global coordination, shared norms, and regulatory guardrails robust enough to make the technology serve humanity rather than quietly reorganize it into something meaner, angrier, and less human.

    Harris’s faith rests on a single, luminous premise: clarity. Confusion, denial, and fatalism are the true villains. If we can see the stakes clearly—if we understand how AI can slide toward chaos or tyranny—then we can choose wisely. “Clarity creates agency,” he says, trusting that informed humans will act in their collective best interest. I admire the moral courage of this argument, but I don’t buy its anthropology. History suggests that clarity does not restrain power; it sharpens it. The most dangerous people in the world are not confused. They are lucid, strategic, and indifferent to collateral damage. They understand exactly what they are doing—and do it anyway. Harris believes clarity liberates agency; I suspect it often just reveals who is willing to burn the future for dominance. The real enemy is not ignorance but nihilistic power-lust, the ancient human addiction to control dressed up in modern code. Harris should keep illuminating the terrain—but he should also admit that many travelers, seeing the cliff clearly, will still sprint toward it. Not because they are lost, but because they want what waits at the edge.

  • How We Outsourced Taste—and What It Cost Us

    How We Outsourced Taste—and What It Cost Us

    Desecrated Enchantment

    noun

    Desecrated Enchantment names the condition in which art loses its power to surprise, unsettle, and transform because the conditions of discovery have been stripped of mystery and risk. What was once encountered through chance, patience, and private intuition is now delivered through systems optimized for efficiency, prediction, and profit. In this state, art no longer feels like a gift or a revelation; it arrives pre-framed as a recommendation, a product, a data point. The sacred quality of discovery—its capacity to enlarge the self—is replaced by frictionless consumption, where engagement is shallow and interchangeable. Enchantment is not destroyed outright; it is trivialized, flattened, and repurposed as a sales mechanism, leaving the viewer informed but untouched.

    ***

    I was half-asleep one late afternoon in the summer of 1987, Radio Shack clock radio humming beside the bed, tuned to KUSF 90.3, when a song slipped into my dream like a benediction. It felt less broadcast than bestowed—something angelic, hovering just long enough to stir my stomach before pulling away. I snapped awake as the DJ rattled off the title and artist at warp speed. All I caught were two words. I scribbled them down like a castaway marking driftwood: Blue and Bush. This was pre-internet purgatory—no playlists, no archives, no digital mercy. It never occurred to me to call the station. My girlfriend phoned. I got distracted. And then the dread set in: the certainty that I had brushed against something exquisite and would never touch it again. Six months later, redemption arrived in a Berkeley record store. The song was playing. I froze. The clerk smiled and said, “That’s ‘Symphony in Blue’ by Kate Bush.” I nearly wept with gratitude. Angels, confirmed.

    That same year, my roommate Karl was prospecting in a used bookstore, pawing through shelves the way Gold Rush miners clawed at riverbeds. He struck literary gold when he pulled out The Life and Loves of a She-Devil by Fay Weldon. The book had a charge to it—dangerous, witty, alive. He sampled a page and was done for. Weldon’s aphoristic bite hooked him so completely that he devoured everything she’d written. No algorithm nudged him there. No listicle whispered “If you liked this…” It was instinct, chance, and a little magic conspiring to change a life.

    That’s how art used to arrive. It found you. It blindsided you. Life in the pre-algorithm age felt wider, riskier, more enchanted. Then came the shrink ray. Algorithms collapsed the universe into manageable corridors, wrapped us in a padded cocoon of what the tech lords decided counted as “taste.” According to Kyle Chayka, we no longer cultivate taste so much as receive it, pre-chewed, as algorithmic wallpaper. And when taste is outsourced, something essential withers. Taste isn’t virtue signaling for parasocial acquaintances; it’s private, intimate, sometimes sacred. In the hands of algorithms, it becomes profane—associative, predictive, bloodless. Yes, algorithms are efficient. They can build you a playlist or a reading list in seconds. But the price is steep. Art stops feeling like enchantment and starts feeling like a pitch. Discovery becomes consumption. Wonder is desecrated.

  • The Expiration Date of the Fitness Influencer

    The Expiration Date of the Fitness Influencer

    Parasocial Fatigue

    noun

    Parasocial Fatigue describes the emotional and cognitive exhaustion that sets in when an audience becomes overexposed to an influencer’s performative intimacy and relentless self-presentation. What begins as a one-sided relationship built on usefulness, inspiration, or trust curdles as the influencer’s need for attention, validation, and monetization becomes increasingly visible. The constant uploads, recycled insights, manufactured urgency, and naked thirst for engagement erode the illusion of authenticity that sustained the bond in the first place. Viewers no longer feel informed or inspired; they feel harvested. At that point, familiarity turns to irritation, admiration hardens into disdain, and the influencer’s presence in the feed triggers avoidance rather than curiosity—a quiet severing of a relationship that was never mutual to begin with.

    In the beginning, your favorite influencer feels like a gift. They offer sensible advice on nutrition, a workout routine that doesn’t insult your intelligence, a body that seems to testify to discipline rather than sorcery. You follow them in good faith. For a while, the content delivers. Then the expiration date quietly approaches. The useful insights thin out, replaced by a slurry of hype, urgency, and alarmist drivel—“You’re poisoning yourself unless you stop eating this one food today.” Clickbait metastasizes. The signal is buried under noise. What once felt like guidance now feels like a carnival barker shouting through a megaphone.

    Eventually you see the machinery. This isn’t a lone truth-teller sharing wisdom from a garage gym; it’s a small content factory with payroll to meet. Ideas are skimmed from journals, stripped of nuance, and polished with influencer saliva until they’re shiny enough to go viral. The real giveaway, though, isn’t the dubious science—it’s the thirst. You can see it in their eyes: the desperation to stay relevant, the exhaustion of feeding the algorithm daily, the hollow confidence of someone trapped in their own posting schedule. The charm collapses. When they appear in your feed now, it’s not curiosity you feel, but a reflexive flinch. Parasocial fatigue sets in, and disdain follows close behind.

  • Drowning in Puffer Jackets: Life Inside Algorithmic Sameness

    Drowning in Puffer Jackets: Life Inside Algorithmic Sameness

    Meme Saturation

    noun

    Meme Saturation describes the cultural condition in which a trend, image, phrase, or style replicates so widely and rapidly that it exhausts its meaning and becomes unavoidable. What begins as novelty or wit hardens into background noise as algorithms amplify familiarity over freshness, flooding feeds with the same references until they lose all edge, surprise, or symbolic power. Under meme saturation, participation is no longer expressive but reflexive; people repeat the meme not because it says something, but because it is everywhere and opting out feels socially invisible. The result is a culture that appears hyperactive yet feels stagnant—loud with repetition, thin on substance, and increasingly numb to its own signals.

    ***

    Kyle Chayka’s diagnosis is blunt and hard to dodge: we have been algorithmically herded into looking, talking, and dressing alike. We live in a flattened culture where everything eventually becomes a meme—earnest or ironic, political or absurd, it hardly matters. Once a meme lodges in your head, it begins to steer your behavior. Chayka’s emblematic example is the “lumpy puffer jacket,” a garment that went viral not because it was beautiful or functional, but because it was visible. Everyone bought the same jacket, which made it omnipresent, which made it feel inevitable. Virality fed on itself, and suddenly the streets looked like a flock of inflatable marshmallows migrating south. This is algorithmic culture doing exactly what it was designed to do: compress difference into repetition. As Chayka puts it, Filterworld culture is homogenous, saturated with sameness even when its surface details vary. It doesn’t evolve; it replicates—until boredom sets in.

    And boredom is the one variable algorithms cannot fully suppress. Humans tolerate sameness only briefly before it curdles into restlessness. A culture that perpetuates itself too efficiently eventually suffocates on its own success. My suspicion is that algorithmic culture will not be overthrown by critique so much as abandoned out of exhaustion. When every aesthetic feels pre-approved and every trend arrives already tired, something else will be forced into existence—if not genuine unpredictability, then at least its convincing illusion. Texture will return, or a counterfeit version of it. Spontaneity will reappear, even if it has to be staged. The algorithm may flatten everything it touches, but boredom remains stubbornly human—and it always demands a sequel.

  • Algorithmic Grooming and the Rise of the Instagram Face

    Algorithmic Grooming and the Rise of the Instagram Face

    Algorithmic Grooming

    noun

    Algorithmic Grooming refers to the slow, cumulative process by which digital platforms condition users’ tastes, attention, and behavior through repeated, curated exposure that feels personalized but is strategically engineered. Rather than directing users abruptly, the system nudges them incrementally—rewarding certain clicks, emotions, and patterns while starving others—until preferences begin to align with the platform’s commercial and engagement goals. The grooming is effective precisely because it feels voluntary and benign; users experience it as discovery, convenience, or self-expression. Yet over time, choice narrows, novelty fades, and autonomy erodes, as the algorithm trains the user to want what is most profitable to serve. What appears as personalization is, in practice, a quiet apprenticeship in predictability.

    ***

    In Filterworld, Kyle Chayka describes algorithmic recommendations with clinical clarity: systems that inhale mountains of user data, run it through equations, and exhale whatever best serves preset goals. Those goals are not yours. They belong to Google Search, Facebook, Spotify, Netflix, TikTok—the platforms that quietly choreograph your days. You tell yourself you’re shaping your feed, curating a digital self-portrait. In reality, the feed is shaping you back, sanding down your edges, rewarding certain impulses, discouraging others. What feels like mutual interdependence is a one-sided apprenticeship in predictability. The changes you undergo—your tastes, habits, even your sense of self—aren’t acts of self-authorship so much as behavior modification in service of attention capture and commerce. And crucially, this isn’t some neutral, machine-led drift. As Chayka points out, there are humans behind the curtain, tweaking the levers with intent. They pull the strings. You dance.

    The cultural fallout is flattening. When everyone is groomed by similar incentives, culture loses texture and people begin to resemble one another—algorithmically smoothed, aesthetically standardized. Chayka borrows Jia Tolentino’s example of the “Instagram face”: the ethnically ambiguous, surgically perfected, cat-like beauty that looks less human than rendered. It’s a face optimized for engagement, not expression. And it serves as a tidy metaphor for algorithmic grooming’s endgame. What begins as personalization ends in dehumanization. The algorithm doesn’t just recommend content; it quietly trains us to become the kind of people that content is easiest to sell to—interchangeable, compliant, and eerily smooth.

  • Joan Is Awful—and So Is the Lie of Digital Empowerment

    Joan Is Awful—and So Is the Lie of Digital Empowerment

    Inverse Agency

    noun

    Inverse Agency describes the moment you mistake engagement with a machine for control. You click, customize, optimize, and curate, convinced you are authoring your own life, while the real authorship happens elsewhere—in systems engineered to harvest your attention, your data, your habits, and eventually your obedience. Everything feels frictionless and flattering. The menus are intuitive. The options feel abundant. But the architecture is doing the choosing. You feel active while being managed, sovereign while being sorted. The freedom is cosmetic. The agency is staged.

    This inversion works because it seduces rather than commands. Inverse agency flatters the ego so effectively that resistance feels unnecessary, even silly. The more certain you are that you’re “in charge,” the less likely you are to notice how your desires are being nudged, your behaviors rewarded, and your choices quietly narrowed. Power doesn’t bark orders; it smiles and asks permission. The result is a system that governs best by convincing its subjects they are free.

    No recent pop artifact captures this better than the Black Mirror episode “Joan Is Awful.” Joan believes she is finally taking control of her life through self-reinvention, only to discover she has handed herself over to an algorithm that strips her identity for parts and streams the wreckage as entertainment. The episode skewers the central lie of platform culture: that self-optimization equals self-authorship. Joan isn’t liberated; she’s monetized. Her personality becomes content, her choices become inputs, and her autonomy dissolves into a subscription model. What looks like empowerment is actually self-erasure with better lighting.

  • How We Declawed a Generation in the Name of Comfort

    How We Declawed a Generation in the Name of Comfort

    Declawing Effect

    noun
    The gradual and often well-intentioned removal of students’ essential capacities for self-defense, adaptation, and independence through frictionless educational practices. Like a declawed cat rendered safe for furniture but vulnerable to the world, students subjected to shortened readings, over-accommodation, constant mediation, and AI-assisted outsourcing lose the intellectual and emotional “claws” they need to navigate uncertainty, conflict, and failure. Within protected academic environments they appear functional, even successful, yet outside those systems they are ill-equipped for unscripted work, sustained effort, and genuine human connection. The Declawing Effect names not laziness or deficiency, but a structural form of educational harm in which comfort and convenience quietly replace resilience and agency.

    ***

    In the early ’90s, a young neighbor told me that the widow behind his condo was distraught. Her cat had gone missing, and she didn’t expect it to come back alive. When I asked why, he explained—almost casually—that the cat had been declawed. He then described the procedure: not trimming, not blunting, but amputating the claws down to the bone. I remember feeling a cold, visceral recoil. People do this, he said, to protect their furniture—because cats will shred upholstery to sharpen the very tools evolution gave them to survive. The logic is neat, domestic, and monstrously shortsighted. A declawed cat may be well behaved indoors, but the moment it slips outside, it becomes defenseless. The price of comfort is vulnerability.

    Teaching in the age of AI, I can’t stop thinking about that cat. My students, too, are being declawed—educationally. They grow up inside a frictionless system: everything mediated by screens, readings trimmed to a few pages, tests softened or bypassed through an ever-expanding regime of accommodations, and thinking itself outsourced to AI machines that research, write, revise, and even flirt on their behalf. Inside education’s Frictionless Dome, they function smoothly. They submit polished work. They comply. But like declawed cats, they are maladapted for life outside the enclosure: a volatile job market, a world without prompts, a long hike navigated by a compass instead of GPS, a messy, unscripted conversation with another human being where nothing is optimized and no script is provided. They have been made safe for institutional furniture, not for reality. And here’s the part that unsettles me most: I work inside the Dome. I enforce its rules. I worry that in trying to keep students comfortable, I’ve become an enabler—a quiet accomplice in reinforcing the Declawing Effect.

  • AI Machines Spell the End of the Bulletproof Syllabus

    AI Machines Spell the End of the Bulletproof Syllabus

    Instructional Humiliation
    noun

    The quiet, destabilizing experience instructors undergo when long-held authority, confidence, and pedagogical certainty collapse in the face of rapid technological change. Instructional humiliation is not the exposure of incompetence but the painful honesty of recognizing that one’s maps no longer match the terrain—often in full view of students who expect guidance. It arises when teachers must admit uncertainty after years of standing as figures of clarity and command, producing a sense of personal diminishment even as it reflects intellectual integrity. Unlike shame rooted in failure, instructional humiliation emerges from ethical transparency: the collision between professional identity and the unsettling reality that the future of education is no longer fully knowable.

    To be an effective instructor, you are expected to project confidence the way a seasoned closer projects certainty across a conference table. You stand before the class with assurance, boldness, and the practiced enthusiasm of a noble salesperson pitching ideas you genuinely believe will improve students’ lives—on the small scale of Student Learning Outcomes and on the larger scale of becoming people who can think, question, and locate their purpose in a chaotic world. You draw from a deep well of literature, history, politics, language, philosophy, geography, and religion, and that accumulated knowledge gives your voice weight. Rectitude is your currency. Students arrive disoriented and unformed; you provide structure, coherence, and a map. In a culture gone mad, you are supposed to be the adult in the room.

    Then AI machines arrive and quietly detonate the set. In recent years, they’ve scrambled the rules of higher education so thoroughly that honesty itself becomes destabilizing. Where you once spoke with earned certainty, you now have to admit—if you’re not lying—that you don’t fully know what comes next. You’re flummoxed. You’re standing at the gates of the unknown with no laminated syllabus for what’s on the other side. Your role feels provisional. Your plan no longer feels bulletproof. Your confidence collapses into something closer to agnosticism, and that shift carries a sting that feels suspiciously like humiliation. And yet you keep going. You walk your students into the fog not because you are heroic or noble, but because stopping is not an option. Clarity may no longer be guaranteed, but the struggle to wrestle meaning from the darkness remains the job.

  • Reproductive Incentive Conflict: Why College Rewards Appearances Over Depth

    Reproductive Incentive Conflict: Why College Rewards Appearances Over Depth

    Reproductive Incentive Conflict
    noun

    The tension that arises when the pursuit of long-term intellectual depth, integrity, and mastery competes with the immediate pressures of achieving economic and social status tied to reproductive success. Reproductive incentive conflict is most acute in environments like college, where young men intuit—often correctly—that mating markets reward visible outcomes such as income, confidence, and efficiency more reliably than invisible virtues like depth or craftsmanship. In such contexts, Deep Work offers no guaranteed conversion into status, while shortcuts, system-gaming, and AI-assisted performance promise faster, more legible returns. The conflict is not moral confusion but strategic strain: a choice between becoming excellent slowly or appearing successful quickly, with real social and reproductive consequences attached to each path.

    Chris Rock once sliced through the romance of meritocracy with a single joke about reproductive economics. If Beyoncé were working the fry station at McDonald’s, her attractiveness alone would not disqualify her from marrying Jay-Z. But reverse the roles—put Jay-Z in a paper hat handing out Happy Meals—and the fantasy collapses. The point is crude but accurate: in the mating market, men are judged less on raw appeal than on status, income, and visible competence. A man has to become something before he is considered desirable. It’s no mystery, then, why a young man entering college quietly factors reproductive success into his motivation. Grades aren’t just grades; they’re potential leverage in a future economy of attraction.

    Here’s where Cal Newport’s vision collides with reality. Newport urges Deep Work—slow, demanding, integrity-driven labor that resists shortcuts and defies easy metrics. Deep Work builds character and mastery, but it offers no guaranteed payout. It may lead to financial success, or it may not. Meanwhile, the student who bypasses depth with AI tools can often game the system, generating polished outputs and efficient performances that read as competence without the grind. The Deep Worker toils in obscurity while the system-gamer cashes visible wins. This creates a genuine tension: between becoming excellent in ways that compound slowly and appearing successful in ways that signal immediately. It’s not a failure of virtue; it’s a collision between two economies—one that rewards depth, and one that rewards display—and young men feel the pressure of that collision every time they open a laptop.