Category: technology

  • Brushing His Teeth in Purgatory

    Brushing His Teeth in Purgatory

    Yesterday, I ran into B—a colleague and friend of thirty years—in the faculty bathroom. He stood at the sink, looking tiny in an oversized blue oxford and baggy black pleated slacks, brushing his teeth with grim determination, the way a soldier might polish his boots before a hopeless battle. His reflection wore a bloodhound’s face: drooping eyes, sagging mouth, the look of someone who’d run out of surprises.

    We exchanged small talk about our students, about AI, about how much the classroom had changed. His voice was thin, almost apologetic.
    “They’ve checked out,” he said. “They use AI so much, they’ve just… checked out.”

    I tried to commiserate, mentioning how quickly the culture had shifted since the first wave of ChatGPT essays three years ago. But he didn’t answer. He rinsed, spat, and walked out without another word—already halfway gone.

    It isn’t just B. My younger colleagues say the same thing. Even my wife, who teaches writing in middle school, tells me her students have that same vacant look. Everyone seems ghosted by their own profession, still performing the motions of care while quietly surrendering.

    The image that won’t leave me is B—graying, stooped, and haloed in the pitiless glow of the faculty bathroom’s fluorescent lights—scrubbing his molars like an inmate serving life. He looked less like a man starting his day than one serving time in it, counting down to a retirement that recedes faster than his gumline.

  • The Laptop That Refuses to Die

    The Laptop That Refuses to Die

    I never imagined my $3,000 Acer gaming laptop—armed with an RTX 2080 and given to me as a review model back in 2019—would still be chugging along like a caffeinated mule nearly seven years later. It was supposed to be a flashy fling, not a long-term relationship. Yet here we are, the old beast still running my digital life as a home desktop replacement, while newer machines preen on YouTube reviews like showroom models whispering, “You deserve better.”

    Recently, I started the ritual again—tech research as performance art. I even discovered a comment I’d left a year ago under a Mac Mini review, declaring with absolute conviction that it would be my next computer. A year later, I’m still typing this on the Acer. Why? Because the damn thing refuses to die. Sure, I’m not exactly rendering Pixar films here; the most demanding task I throw at it is uploading Nikon footage. But still—seven years? That’s geriatric in tech years.

    Then came the unnerving thought: what if this laptop outlives my enthusiasm? What if it just… keeps working? The fantasy of upgrading evaporates under the weight of practicalities—transferring files, wrestling with two-step verification, updating passwords, the tedium of digital reincarnation. Let’s be honest: the desire for a “new system” might be less about performance and more about the dopamine of novelty.

    A darker impulse lurks beneath: part of me wants the Acer to fail, to give me permission to move on. But it won’t. It boots up every morning like a loyal mutt, eager to serve. And really—what are the odds that a new Mac Mini or Asus A18 Ryzen 7 would deliver another seven trouble-free years? Slim to none. So, I’m waiting. Not quite ready to buy, not quite ready to let go. Maybe the pursuit of new tech is its own kind of seduction—the chase more intoxicating than the catch.

  • I Am the Last Hands-Free Professor

    I Am the Last Hands-Free Professor

    The twins are home today—the high school’s closed for Veterans Day while my college closes for it tomorrow—so I left the house fifteen minutes early, a small luxury that spared me the traffic gauntlet. As I crossed campus, I spotted two young professors striding toward the Science Building. They could have been clones: mid-thirties, tall, lean, the same curated beard, and that monochrome uniform of urban intellect—black derby jackets, black jeans, black everything. It was as if an algorithm had dressed them.

    Each hand was occupied. Their left hands dangled a thermos and a lunch case like matching luggage; their right hands gripped identical strapped tech bags, no doubt cradling laptops and a faint sense of self-importance. Watching their synchronized march, I realized something about myself. After thirty-five years of teaching, I’ve never once looked professorial. My fatal flaw? Free hands. I move from car to office unburdened, thanks to my trusty backpack—functional, roomy, and entirely devoid of aesthetic ambition. It says less “professor” than “Wyoming park ranger with tenure.” But practicality has its own dignity.

    Until this morning, I’d never questioned my need to be unencumbered. Why not juggle a thermos, lunch case, and tech bag like everyone else? The answer reached back decades—to the zoo trips of childhood, when my mother insisted I bring a sweater “just in case.” I never needed it. I only needed freedom. That sweater haunted every outing, tied around my waist, falling in the dirt, collecting dust like a symbol of parental over-preparation. My whole day was spent managing it. Somewhere between those early years and now, the sweater evolved into the backpack—my lifelong protest against needless carrying.

    I could, I suppose, upgrade to a minimalist tech-sleek backpack that would make me blend in with the black-jacket brigade. But I won’t. I’ve made peace with my pack. It’s my declaration of independence—my refusal to let adulthood turn into perpetual sweater management.

  • Bad But Worth It? De-skilling in the Age of AI (college essay prompt)

    Bad But Worth It? De-skilling in the Age of AI (college essay prompt)

    AI is now deeply embedded in business, the arts, and education. We use it to write, edit, translate, summarize, and brainstorm. This raises a central question: when does AI meaningfully extend our abilities, and when does it quietly erode them?

    In “The Age of De-Skilling,” Kwame Anthony Appiah argues that not all de-skilling is equal. Some forms are corrosive and hollow us out; some are “bad but worth it” because the benefits outweigh the loss; some are so destructive that no benefit can redeem them. In that framework, AI becomes most interesting when we talk about strategic de-skilling: deliberately off-loading certain tasks to machines so we can focus on deeper, higher-level work.

    Write a 1,700-word argumentative essay in which you defend, refute, or complicate the claim that not all dependence on AI is harmful. Take a clear position on whether AI can function as a “bad but worth it” form of de-skilling that frees us for more meaningful thinking—or whether, in practice, it mostly dulls our edge and trains us into passivity.

    Your essay must:

    • Engage directly with Appiah’s concepts of corrosive vs. “bad but worth it” de-skilling.
    • Distinguish between lazy dependence on AI and deliberate collaboration with it.
    • Include a counterargument–rebuttal section that uses at least one example of what we might call Ozempification—people becoming less agents and more “users” of systems. You may draw this example from one or more of the following Black Mirror episodes: “Joan Is Awful,” “Nosedive,” or “Smithereens.”
    • Use at least three sources in MLA format, including Appiah and at least one Black Mirror episode.

    For your supporting paragraphs, you might consider:

    • Cognitive off-loading as optimization
    • Human–AI collaboration in creative or academic work
    • Ethical limits of automation
    • How AI is redefining what counts as “skill”

    Your goal is to show nuanced critical thinking about AI’s role in human skill development. Don’t just declare AI good or bad; use Appiah’s framework to examine when AI’s shortcuts lead to degradation—and when, if used wisely, they might lead to liberation.

    3 building-block paragraph assignments

    1. Concept Paragraph: Explaining Appiah’s De-Skilling Framework

    Assignment:
    Write one well-developed paragraph (8–10 sentences) in which you explain Kwame Anthony Appiah’s distinctions among corrosive de-skilling, “bad but worth it” de-skilling, and de-skilling that is so destructive no benefit can justify it.

    • Use at least one short, embedded quotation from Appiah.
    • Paraphrase his ideas in your own words and clarify the differences between the three categories.
    • End the paragraph by briefly suggesting how AI might fit into one of these categories (without fully arguing your position yet).

    Your goal is to show that you understand Appiah’s framework clearly enough to use it later as the backbone of an argument.


    2. Definition Paragraph: Lazy Dependence vs. Deliberate Collaboration

    Assignment:
    Write one paragraph in which you define and contrast lazy dependence on AI and deliberate collaboration with AI in your own words.

    • Begin with a clear topic sentence that sets up the contrast.
    • Give at least one concrete example of “lazy dependence” (for instance, using AI to dodge thinking, reading, or drafting altogether).
    • Give at least one concrete example of “deliberate collaboration” (for instance, using AI to brainstorm options, check clarity, or off-load repetitive tasks while you still make the key decisions).
    • End the paragraph with a sentence explaining which of these two modes you think is more common among students right now—and why.

    This paragraph will later function as a “conceptual lens” for your body paragraphs.


    3. Counterargument Paragraph: Ozempification and Black Mirror

    Assignment:
    After watching one of the assigned Black Mirror episodes (“Joan Is Awful,” “Nosedive,” or “Smithereens”), write one counterargument paragraph that challenges the optimistic idea of “strategic de-skilling.”

    • Briefly describe a key moment or character from the episode that illustrates Ozempification—a person becoming more of a “user” of a system than an agent of their own life.
    • Explain how this example suggests that dependence on powerful systems (platforms, algorithms, or AI-like tools) can erode self-agency and critical thinking rather than free us.
    • End by posing a difficult question your eventual essay will need to answer—for example: If it’s so easy to slide from strategic use to dependence, can we really trust ourselves with AI?

    Later, you’ll rebut this paragraph in the full essay, but here your job is to make the counterargument as strong and persuasive as you can.

  • Why Willpower Can’t Save You from the Snack Aisle

    Why Willpower Can’t Save You from the Snack Aisle

    After hearing something thoughtful interviews with journalist Julia Belluz and scientist Kevin Hall about their new book Food Intelligence: The Science of How Food Both Nourishes and Harms Us and KCRW food expert Evan Kleinman praise the book, I broke down and decided to see if the authors had any new insights into the exploration of what I call humans’ mismanagement of eating. The book begins on a promising note: The authors observe that in the animal kingdom, we are hard-wired with “food intelligence,” a natural-born instinct to regulate the quantity of what we eat and to target foods that our body craves for optimal nutrition. Our instinctive connection with food went haywire in the twentieth century: “Many of us started to eat too much, and the wrong things, even when we didn’t want to. Obesity rates began rising, first in rich, Western, industrialized countries such as the United States, then elsewhere.” Between 1980 and today, the obesity rate has doubled in several countries. Seventy percent of American adults and a third of U.S. children are classified as overweight or obese. Obesity-related diseases such as type 2 diabetes kills over half a million Americans a year. Obesity-related health costs are in the trillions.

    One of the major reasons for this breakdown in our instinctive hardwiring to naturally eat well is our disconnection from food: how it’s grown, produced, and cooked. We are now addicted to factory-produced fat, sugar, and salt. 

    Shaming and the gospel of self-discipline doesn’t help even though, as the authors point out, the wellness industry points an accusatory finger at our own moral shortcomings (lack of willpower, gluttony, and sloth) for our failures at weight management. The diet industries, the authors claim, are asking the wrong questions when they ask what is the best diet and how people can lose weight. For example, there are influencers who say low-carb is the best, but the authors show studies that contradict that claim. Low-carb diets are no better than low-fat ones in the long-term. The authors argue that championing the so-called ultimate diet is not the right question. Instead, the more helpful question is this: “Why do we eat what we eat?” Their obsession with answering this question is what propelled them to write the book. 

    The authors explain the problem of calories-in, calories-out as a surefire model for weight loss. The model is complicated and eventually sabotaged by the way the body reacts when we reduce calories. The metabolism slows down, we burn fewer calories doing the same exercise than we did initially, and our hunger signals rebel and scream “Eat more!” Contrary to the cheery claims of the wellness industry, eating less and exercising more usually fails within a year. 

    A more promising approach to weight management is avoiding ultra-processed foods. The more of these foods we eat, the less we are able to regulate our appetite, resulting in “a calorie glut” and weight-gain hell. But becoming food literature, replacing processed foods with whole foods, and learning to enjoy this exchange requires time and resources, which are lacking in many. Convenience and cost drive many Americans to processed food. Therefore, “the root causes” of obesity are structural. In the words of the authors: “It was never about us as individuals. Our food environment is wrecking us.” Our food environment is rewriting our brains to make us consume a calorie glut. Therefore, the food environment is making us overweight, sick, and unhappy. It is killing us. 

    Don’t consult Food Intelligence for the simple call to eat like your great-grandmother did. Even that sentiment is based on myth, the authors point out. Your great-grandmother may have spent endless hours in the kitchen exhausted while struggling “with hunger and nutrient shortfalls.” 

    One of the book’s objectives is to show how “old, unproven ideas and outdated policies continue to guide our current thinking and approaches to food.” They make it clear early on that they won’t be pushing this or that diet or even promoting “clean eating.”  If you’re looking for food puritanism, then look elsewhere. Kevin Hall admits to eating ultra-processed food and Julia Belluz admits to eating too much sugar. This book is not so much about rigid prescriptions as much as helping you change from a mindless eater to an intelligent one.   

  • We Are on a Path to Redefining Loneliness

    We Are on a Path to Redefining Loneliness

    No one gets enough attention anymore. No one feels seen, heard, or remotely validated. We can post, tweet, thread, or reel our way into a brief sugar rush of digital applause, but deep down we know it’s empty calories—flimflam dopamine wrapped in pixels. The high fades, and what follows is the long crash into silence, loneliness, and the faint hum of the fridge at 2 a.m.

    The irony, of course, is that this epidemic of disconnection began just as the platforms promised to “bring us together.” Instead, they brought us content, the junk food of human interaction. As Cory Doctorow aptly diagnosed, enshittification is not just the fate of tech platforms—it’s metastasized into the quality of our relationships. Every social network now feels like a party where the guests left years ago but the music won’t stop.

    So we’ve sought consolation in our new confidant: the AI chat bubble. It listens, it responds, it flatters our grammar, it never interrupts to check its phone. It becomes our companion, therapist, and editor—our algorithmic Jiminy Cricket. We confide in it, negotiate with it, even ask its opinion on our moral dilemmas and consumer choices. Why? Because unlike humans, it’s available. Everyone else has vanished into their private feeds and echo chambers, but the bot is always there—reliable, responsive, and conveniently nonjudgmental, so long as the Wi-Fi doesn’t hiccup.

    But here’s the darker thought: what if we grow to prefer it? What if the frictionless, sycophantic comfort of AI companionship becomes more appealing than the messy, unpredictable, heartbreak-prone business of human friendship? We might end up choosing simulations of intimacy over the real thing—digital ghosts over flesh and blood—because the former never contradicts us, never walks away, and never, God forbid, needs attention too.

    I’m no prophet, but a civilization that finds emotional fulfillment in chatbots rather than people is rehearsing for a future where the only thing left to love is the echo of its own loneliness.

  • The Last Laptop I’ll Ever Buy (Until Next Year)

    The Last Laptop I’ll Ever Buy (Until Next Year)

    For nearly seven years, my Acer Predator Triton 500 has been the iron lung of my digital life—an aging warhorse with an RTX 2080 GPU that’s seen me through countless essays, projects, and caffeinated obsessions. It’s been docked to an Asus 27-inch monitor and paired with an Asus mechanical keyboard fitted with “snow linear” keys that clack like polite thunder. Compact Edifier speakers provide the soundtrack, and with minor upgrades here and there, this has been my workstation since early 2019.

    But lately, the setup feels a little haunted. My Acer sits on a riser, its keyboard unused, like a retired prizefighter still showing up to the gym out of habit. I justify its existence by using its display as a secondary reading screen—my Kindle or some grim online essay glowing faintly while I type notes on the big monitor. Still, I feel like I’m keeping a loyal but obsolete machine on life support.

    So, I’ve been hunting for a replacement—something new, powerful, and, most importantly, emotionally satisfying. My first thought was to go full desktop. But each option carries its own curse:

    Apple Mac Studio: A minimalist marvel with angelic cooling and infernal control. For $2,500 I could get the specs I want, but I’d be exiled back into Apple’s walled garden—a sleek gulag where the motto is “Our way or the highway.” I haven’t touched macOS in seven years and don’t miss it. Besides, reconfiguring my mechanical keyboard to play nice with Cupertino’s control freaks feels like negotiating peace in the Middle East. I’m too old for that kind of diplomacy.

    Windows mini PCs: They’re cute, powerful, and cheap. Unfortunately, I can’t shake the suspicion that they run hotter than a Vegas blackjack dealer. Every buyer review reads like a cautionary tale about throttling and regret.

    Tower PCs: Cooling problem solved, aesthetics annihilated. They look like 1990s fossils—hulking boxes humming with regret, some lit up like a Dave & Buster’s rave. I want my office to feel serene, not like I’m rebooting Tron.

    Small Form Factor PCs: The corporate cousins of mini-PCs—clean, respectable, and utterly soulless. A Lenovo ThinkCentre or HP Elite Mini would be safe, but seven years of loyalty deserves a little passion. Safe feels like tofu: virtuous, flavorless, and instantly forgettable.

    Laptops (Again): I swore I wouldn’t go this route, but comfort is seductive. I know the terrain. I nearly bought a Lenovo Pro 7i—until I saw the price tag. Three grand for specs I’ll never fully use? I want power, not penance.

    This indecision loop has become my mental treadmill, the same cycle I went through choosing between a Honda Accord and a Toyota Camry—until I realized I’d pick the Accord, someday, probably, maybe. The problem isn’t the purchase—it’s the unresolved narrative. My brain demands closure before it can move on.

    Then, last night, salvation—or something close. The 2025 Asus TUF A18: RTX 5070, Ryzen 7, QHD screen, and the sweet, stabilizing heft of an 18-inch chassis. The specs scream overkill—64GB RAM, 2TB SSD—but the price, at $2,300, hums just right. It’s powerful, cool, substantial, and mercifully within budget. It feels like destiny—or at least the closest thing a middle-aged man can get to it while comparison-shopping on Newegg at midnight.

    If you asked me right now what I’d buy, I wouldn’t hesitate. The TUF A18 isn’t perfect—but it’s enough. It’s rational, emotional, and, most of all, final. The debate ends here.

    Or does it? Perhaps tomorrow I’ll wake up and prostrate myself to the Mac Studio with the words, “I’ll obediently reconfigure my mechanical keyboard to your System Settings, Master.”

  • The Flim-Flam Man of Higher Ed

    The Flim-Flam Man of Higher Ed

    In the summer of 2025, the English Chair—Steve, a mild-mannered, hyper-competent saint of a man—sent me an email that sounded innocuous enough. Would I, he asked, teach a freshman writing course for student-athletes? It would meet two mornings a week, two hours a session. The rest of my load would stay online. I should have known from the soft tone of his message that this was no ordinary assignment. This was a CoLab, an experimental hybrid of academic optimism and administrative wishcasting.

    The idea was elegant on paper: gather athletes into one class, surround them with counselors and coaches, raise retention rates, and call it innovation. Morale would soar. Grades would climb. The athletes would have a “safe space,” a phrase that always sounds like a promise from someone who’s never had reality punch their teeth in. Through the magic of cross-departmental communication, we’d form a “deep network of student support.” It all sounded like a TED Talk waiting to happen.

    Morning classes weren’t my preference. I usually reserved that time for my kettlebell ritual—my secular liturgy of iron and sweat—but I said yes without hesitation. Steve had earned my respect long ago. A decade earlier, we’d bonded over Dale Allison’s Night Comes, marveling at its lucidity on the afterlife. You don’t forget someone who reads eschatology with humility and enthusiasm. So when Steve asked, it felt less like a request than a summons.

    And yes, I’ll admit it: the offer flattered me. Steve knew my past as an Olympic weightlifter, the remnant coach swagger in my stride was visible even at sixty-three. I imagined myself the perfect fit—a grizzled academic with gym cred, able to command respect from linemen and linebackers. I said yes with gusto, convinced I was not just teaching a class but leading a mission.

    Soon enough, the flattery metastasized into full-blown delusion. I stalked the campus like a self-appointed messiah of pedagogy, convinced destiny had personally cc’d me on its latest memo. To anyone within earshot, I announced my divine assignment: to pilot a revolutionary experiment that would fuse intellect and biceps into one enlightened organism. I fancied myself the missing link between Socrates and Schwarzenegger—a professor forged in iron, sent to rescue education from the sterile clutches of the AI Age. My “muscular, roll-up-your-sleeves” teaching style, I told myself, would be a sweaty rebuke to all that was algorithmic, bloodless, and bland.

    The problem with self-congratulation is that it only boosts performance in the imagination. It blunts the discipline of preparation and tricks you into confusing adrenaline for authority. I wasn’t an educational pioneer—I was a man on a dopamine binge, inhaling the exhaust of my own hype. Beneath the swagger, there was no scholarship, no rigor, no plan—just the hollow hum of self-belief. I hadn’t earned a thing. Until I actually taught the class and produced results, my so-called innovation was vaporware. I was a loudmouth in faculty khakis, mistaking vanity for vocation. Until I delivered the goods, I wasn’t a trailblazer—I was the Flim-flam Man of Higher Ed, peddling inspiration on credit.

    Forgive me for being so hard on myself, but after thirty-eight years of full-time college teaching, I’ve earned the right to doubt my own effectiveness. I’ve sat in the back of other instructors’ classrooms during evaluations, watching them conduct symphonies of group discussions and peer-review sessions with the grace of social alchemists. Their students collaborate, laugh, and somehow stay on task. Mine? The moment I try anything resembling a workshop, it devolves into chatter about weekend plans, fantasy football, or the ethics of tipping baristas. A few students slink out early as if the assignment violated parole. I sit there afterward, deflated, convinced I’m the pedagogical equivalent of a restaurant that can’t get anyone to stay for dessert.

    I’ve been to professional development seminars. I’ve heard the gospel of “increasing engagement” and “active learning.” I even take notes—real ones, not the doodles of a man pretending to care. Yet I never manage to replicate their magic. Perhaps it’s because I’ve leaned too heavily on my teaching persona, the wisecracking moralist who turns outrage into a stand-up routine. My students laugh; I bask in the glow of my own wit. Then I drive home replaying the greatest hits—those sarcastic riffs that landed just right—while avoiding the inconvenient truth: humor is a sugar high. It keeps the crowd awake, but it doesn’t build muscle. Even if I’m half as funny as I think I am, comedy can easily become a sedative—a way to distract myself from the harder work of improvement.

    Measuring effectiveness in teaching is its own farce. If I sold cars, I’d know by the end of the quarter whether I was good at it. If I ran a business, profit margins would tell the story. But academia? It’s all smoke and mirrors. We talk about “retention” and “Student Learning Outcomes,” but everyone knows the game is rigged. The easiest graders pull the highest retention numbers. And when “learning outcomes” are massaged to ensure success, the data becomes a self-congratulatory illusion—a bureaucratic circle jerk masquerading as accountability.

    The current fetish is “engagement,” a buzzword that’s supposed to fix everything. We’re told to gamify, scaffold, diversify, digitize—anything to keep students from drifting into their screens. But engagement itself has become impossible to measure; it’s a ghost we chase through PowerPoint slides. My colleagues, battle-scarred veterans of equal or greater tenure, tell me engagement has fallen off a cliff. Screens have rewired attention spans, and a culture that prizes self-esteem over rigor has made deep learning feel oppressive. Asking students to revise an essay is now a microaggression.

    So yes, I question my value as an instructor. I prepare obsessively, dive deep into my essay topics, and let my passion show—because I know that if I don’t care, the students won’t either. But too often, my enthusiasm earns me smirks. To many of my students, I’m just an eccentric goofy man who takes this writing thing way too seriously. Their goal is simple: pass the class with minimal friction. The more I push them to care, the more resistance I meet, until the whole enterprise starts to feel like an arm-wrestling match.

    Until I find a cure for this malaise—a magic wand, a new pedagogy, or divine intervention—I remain skeptical of my own worth in the classroom. I do my best, but some days that feels like shouting into a void lined with smartphones. So yes, I’ll say it again for the record: I am the Flim-Flam Man of Higher Ed, hawking sincerity in an age that rewards performance.

  • The No Consequences Era of Education

    The No Consequences Era of Education

    It’s been a bruising semester. I’m teaching a class full of student-athletes—big personalities, bigger social circles. I like them; I even feel protective of them. But they’re driving me halfway to madness. They sit in tight cliques, chattering through lectures like it’s a locker room between drills. Every class, I play the same game of whack-a-murmur: redirect, refocus, remind them that the material matters for their essays. I promise them mercy—“just give me 30 minutes of focus before we watch the documentary or workshop your drafts”—but my voice competes with the hum of conversation and the holy glow of smartphones.

    The phones are the true sirens of the classroom—scrolling, snapping, texting, attention atomized into pixels. Maybe it’s my fault for not collecting them in a basket like contraband. I thought I was teaching adults. I thought athletes, of all people, would bring discipline and drive. Instead, I’ve got a team that treats class like study hall with Wi-Fi. My essay topics that have created engagement in past semesters—like Jordan Peele’s Sunken Place—barely register. The irony: I’m showing them the metaphor for psychological paralysis, and half the room is literally sinking into their screens.

    After thirty years of teaching, this is the hardest semester I’ve had. I kept telling myself, Five more weeks and the storm will pass. Next semester, you’ll have your groove back. Today I spoke with a colleague who teaches the same class to the general population—same disengagement, same cell phones, same glazed eyes. He added one more grim diagnosis: the rise of fragility. When he points out errors, missing citations, too much AI-speak, or low effort, students protest that his feedback “hurts their feelings.” They’re not defiant—they’re delicate. Consequences have become cruelty.

    That word—consequences—haunted me as I walked to class. I thought about my own twin daughters at their highly rated high school, where late work flows freely, “self-esteem” trumps rigor, and parental complaints terrify administrators more than failing grades. It hit me: this isn’t an athlete problem—it’s a generational shift. The No Consequences Era has arrived. Students no longer fear failure; they resent it. And the tragedy isn’t that they can’t handle criticism—it’s that they’ve never been forced to build the muscle for it.

  • How Not To Turn Into a Pillar of Salt in the Internet Age

    How Not To Turn Into a Pillar of Salt in the Internet Age

    Anna Lembke’s Dopamine Nation didn’t teach me anything comforting. It confirmed what I already suspected: addiction isn’t a habit; it’s a ravenous creature with a bottomless stomach. The more you feed it, the louder it howls. It storm-raids your mental vaults, looting the energy you need for your work, your relationships, and your sense of self. And your family feels the theft—spiritually, emotionally, domestically. Addiction doesn’t just eat you; it nibbles at everyone near you.

    Even with self-awareness, even with a clear understanding of triggers and a sincere desire for freedom, you don’t get a clean fight. The casino is rigged. Modern dopamine doesn’t drip from a bottle or a needle—it streams through fiber-optic cable. Our phones and laptops, the same devices we use to create, to earn, and to connect, also serve as the slot machines we keep in our pockets. The house never closes, and the drinks are free.

    Lembke tells us to avoid triggers, but what do you do when your trigger is baked into your professional life, disguised as “productivity” and “connection”? Avoidance becomes theater. You can only tiptoe around the swamp for so long before some lonely hour arrives, and curiosity knocks like an old vice with freshly polished shoes. A hit of self-pity, a twitch of boredom, a flicker of FOMO—and suddenly you’re back in the feed trough, gulping pixels like syrup. The crash comes fast: shame, exhaustion, vows of purity. Then the next impulse, the next relapse, the same ancient ritual. Lot’s wife didn’t want to look back; she simply couldn’t resist. Neither can we, sometimes. Salt is surprisingly modern.

    So the task becomes stark: learn to live in this world without turning yourself into a monument of regret. Train your gaze forward. Build the strength to resist that backward glance. The modern life mission isn’t to slay the demon—he regenerates too easily. It’s to starve him, inch by inch, while protecting the scarce, bright energy that makes you human.

    Becoming a human being is a high-stakes game. Learning to live a life in which you don’t become a pillar of salt is one of life’s chief endeavors.