Think Again: Difference between revisions

Content deleted Content added
No edit summary
No edit summary
Line 31:
The publisher lists the book as a #1 New York Times bestseller, and it appeared on year-end lists from ''{{Tooltip|The Washington Post}}'' and ''{{Tooltip|Newsweek}}'' in 2021.<ref name="PRH2021" /><ref name="WaPoBest2021">{{cite news |title=Best nonfiction of 2021 |url=https://www.washingtonpost.com/arts-entertainment/2021/11/18/best-nonfiction-2021/ |work=The Washington Post |date=18 November 2021 |access-date=8 November 2025}}</ref><ref name="NewsweekFav2021">{{cite news |title=Our 21 Favorite Books of 2021 |url=https://www.newsweek.com/our-21-favorite-books-2021-1661466 |work=Newsweek |date=22 December 2021 |access-date=8 November 2025}}</ref>
 
=== Part I – Individual Rethinking: Updating Our Own Views ===
== Chapter summary ==
''This outline follows the {{Tooltip|Viking}} hardcover edition (2 February 2021, ISBN 978-1-9848-7810-6).''<ref name="PRH2021">{{cite web |title=Think Again by Adam Grant |url=https://www.penguinrandomhouse.com/books/607660/think-again-by-adam-grant/ |website=Penguin Random House |publisher=Penguin Random House |date=2 February 2021 |access-date=8 November 2025}}</ref><ref name="OCLC1191456279" />
 
=== Chapter 1 – A Preacher, a Prosecutor, a Politician, and a Scientist Walk into Your Mind ===
=== I – Individual Rethinking: Updating Our Own Views ===
🧠 '''1 – A Preacher, a Prosecutor, a Politician, and a Scientist Walk into Your Mind.''' In {{Tooltip|Milan}}, more than a hundred Italian startup founders entered a four-month entrepreneurship program and were randomly assigned either to standard training or to add “scientist’s goggles,” treating strategies as hypotheses, interviews as hypothesis generation, and prototypes as experiments. Over the following year the control group averaged under $300 in revenue, while the scientific-thinking group averaged over $12,000, pivoted more than twice as often, and won customers sooner. {{Tooltip|Phil Tetlock}}’s three mindsets—preacher, prosecutor, politician—show how identity can eclipse evidence when we defend sacred beliefs, hunt for others’ errors, or chase approval. {{Tooltip|Stephen Greenspan}}’s cautionary tale illustrates the cost: he invested nearly a third of his retirement savings in a fund tied to {{Tooltip|Bernie Madoff}}, watched it rise 25 percent, and then lost it overnight when the Ponzi scheme collapsed. {{Tooltip|Mike Lazaridis}}’s {{Tooltip|BlackBerry}} offers a corporate parallel: after marveling at the first iPhone in 2007 and overseeing a company valued above $70 billion in 2008, he still resisted adding a robust browser. He later balked at features like encrypted messaging—an opening {{Tooltip|WhatsApp}} eventually seized in a $19 billion acquisition—because he failed to test alternatives to his favored device model. The traps include confirmation bias, desirability bias, and the “I’m not biased” bias, and intelligence can harden certainty rather than sharpen accuracy. Thinking like a scientist ties identity to the quest for truth, not to any one idea; run small tests, seek disconfirming data, and revise models so humility, curiosity, and discovery prevent success from calcifying into dogma.
 
=== Chapter 2 – The Armchair Quarterback and the Impostor: Finding the Sweet Spot of Confidence ===
🧠 '''1 – A Preacher, a Prosecutor, a Politician, and a Scientist Walk into Your Mind.''' In {{Tooltip|Milan}}, more than a hundred Italian startup founders entered a four-month entrepreneurship program and were randomly assigned either to standard training or to add “scientist’s goggles,” treating strategies as hypotheses, interviews as hypothesis generation, and prototypes as experiments. Over the following year the control group averaged under $300 in revenue, while the scientific-thinking group averaged over $12,000, pivoted more than twice as often, and won customers sooner. {{Tooltip|Phil Tetlock}}’s three mindsets—preacher, prosecutor, politician—show how identity can eclipse evidence when we defend sacred beliefs, hunt for others’ errors, or chase approval. {{Tooltip|Stephen Greenspan}}’s cautionary tale illustrates the cost: he invested nearly a third of his retirement savings in a fund tied to {{Tooltip|Bernie Madoff}}, watched it rise 25 percent, and then lost it overnight when the Ponzi scheme collapsed. {{Tooltip|Mike Lazaridis}}’s {{Tooltip|BlackBerry}} offers a corporate parallel: after marveling at the first iPhone in 2007 and overseeing a company valued above $70 billion in 2008, he still resisted adding a robust browser. He later balked at features like encrypted messaging—an opening {{Tooltip|WhatsApp}} eventually seized in a $19 billion acquisition—because he failed to test alternatives to his favored device model. The traps include confirmation bias, desirability bias, and the “I’m not biased” bias, and intelligence can harden certainty rather than sharpen accuracy. Thinking like a scientist ties identity to the quest for truth, not to any one idea; run small tests, seek disconfirming data, and revise models so humility, curiosity, and discovery prevent success from calcifying into dogma.
🪑 '''2 – The Armchair Quarterback and the Impostor: Finding the Sweet Spot of Confidence.''' {{Tooltip|Ursula Mercz}}, a seamstress admitted to a clinic with headaches and dizziness, insisted she could still see despite neurological blindness—a classic sign of {{Tooltip|Anton’s syndrome}} and a metaphor for how people can be blind to their own blind spots. Two Icelandic figures then bracket the range: business leader Halla Tómasdóttir, publicly petitioned in 2015 to run for president, hesitated out of self-doubt yet finished second with more than a quarter of the vote, while former prime minister Davíð Oddsson projected unwavering confidence despite earlier failures. These cases map onto {{Tooltip|Dunning–Kruger}}: the unskilled can be overconfident “armchair quarterbacks,” and capable people often underrate themselves as “impostors.” Practical antidotes include objective yardsticks, learning goals, and “confident humility,” illustrated by {{Tooltip|Sara Blakely}} teaching herself hosiery manufacturing and patent basics before launching {{Tooltip|Spanx}}. Research cited from {{Tooltip|Basima Tewfik}} shows that professionals with impostor thoughts can be rated more interpersonally effective, and {{Tooltip|Danielle Tussing}}’s rotating charge-nurse study finds that those who felt some hesitation sought second opinions and led teams more effectively. Calibrate confidence to competence by anchoring self-belief in evidence and treating doubt as a cue to prepare and listen; keep testing assumptions so humility powers learning while confidence sustains action.
 
=== Chapter 3 – The Joy of Being Wrong: The Thrill of Not Believing Everything You Think ===
🪑 '''2 – The Armchair Quarterback and the Impostor: Finding the Sweet Spot of Confidence.''' {{Tooltip|Ursula Mercz}}, a seamstress admitted to a clinic with headaches and dizziness, insisted she could still see despite neurological blindness—a classic sign of {{Tooltip|Anton’s syndrome}} and a metaphor for how people can be blind to their own blind spots. Two Icelandic figures then bracket the range: business leader Halla Tómasdóttir, publicly petitioned in 2015 to run for president, hesitated out of self-doubt yet finished second with more than a quarter of the vote, while former prime minister Davíð Oddsson projected unwavering confidence despite earlier failures. These cases map onto {{Tooltip|Dunning–Kruger}}: the unskilled can be overconfident “armchair quarterbacks,” and capable people often underrate themselves as “impostors.” Practical antidotes include objective yardsticks, learning goals, and “confident humility,” illustrated by {{Tooltip|Sara Blakely}} teaching herself hosiery manufacturing and patent basics before launching {{Tooltip|Spanx}}. Research cited from {{Tooltip|Basima Tewfik}} shows that professionals with impostor thoughts can be rated more interpersonally effective, and {{Tooltip|Danielle Tussing}}’s rotating charge-nurse study finds that those who felt some hesitation sought second opinions and led teams more effectively. Calibrate confidence to competence by anchoring self-belief in evidence and treating doubt as a cue to prepare and listen; keep testing assumptions so humility powers learning while confidence sustains action.
🤯 '''3 – The Joy of Being Wrong: The Thrill of Not Believing Everything You Think.''' Nobel Prize–winning psychologist {{Tooltip|Daniel Kahneman}} models delight in error: when evidence overturns his view, he reacts with curiosity and genuine enthusiasm rather than defensiveness. That posture—holding ideas lightly—treats mistakes as data, not threats to identity. Evidence from the {{Tooltip|Good Judgment Project}}, a multi-year forecasting tournament run for {{Tooltip|IARPA}} from 2011 to 2015, shows that the most accurate forecasters treated beliefs as working hypotheses and updated their probabilities frequently, improving {{Tooltip|Brier scores}} over time. These “superforecasters” succeeded by revising early and often, even when revisions meant admitting yesterday’s confidence was misplaced. Identity-protective thinking—equating “I’m right” with “I am”—blocks this learning loop and makes disconfirming facts feel like personal attacks. Simple tests loosen attachment: ask what evidence would change your mind or write down conditions that would trigger a pivot. The social side matters too: surround yourself with people who notice when you’re off and make it safe to acknowledge it. Take pride in revising beliefs faster, not in defending them longer; progress comes from decoupling identity from opinions and replacing confirmation with hypothesis testing through frequent updates, explicit falsification, and an open door for disconfirming feedback.
 
=== Chapter 4 – The Good Fight Club: The Psychology of Constructive Conflict ===
🤯 '''3 – The Joy of Being Wrong: The Thrill of Not Believing Everything You Think.''' Nobel Prize–winning psychologist {{Tooltip|Daniel Kahneman}} models delight in error: when evidence overturns his view, he reacts with curiosity and genuine enthusiasm rather than defensiveness. That posture—holding ideas lightly—treats mistakes as data, not threats to identity. Evidence from the {{Tooltip|Good Judgment Project}}, a multi-year forecasting tournament run for {{Tooltip|IARPA}} from 2011 to 2015, shows that the most accurate forecasters treated beliefs as working hypotheses and updated their probabilities frequently, improving {{Tooltip|Brier scores}} over time. These “superforecasters” succeeded by revising early and often, even when revisions meant admitting yesterday’s confidence was misplaced. Identity-protective thinking—equating “I’m right” with “I am”—blocks this learning loop and makes disconfirming facts feel like personal attacks. Simple tests loosen attachment: ask what evidence would change your mind or write down conditions that would trigger a pivot. The social side matters too: surround yourself with people who notice when you’re off and make it safe to acknowledge it. Take pride in revising beliefs faster, not in defending them longer; progress comes from decoupling identity from opinions and replacing confirmation with hypothesis testing through frequent updates, explicit falsification, and an open door for disconfirming feedback.
🥊 '''4 – The Good Fight Club: The Psychology of Constructive Conflict.''' In 2000 at {{Tooltip|Pixar}}, leaders hired {{Tooltip|Brad Bird}} to “shake things up”; technical heads initially said his vision would take a decade and $500 million, so he built a team of self-described misfits, invited rigorous debate, and four years later delivered Pixar’s most complex film to date while reducing cost per minute of animation. Dissent aimed at ideas, not people, becomes an engine for quality. Research by {{Tooltip|Karen “Etty” Jehn}} distinguishes task conflict (clashes over content and process) from relationship conflict (personal friction) and shows that teams with low relational friction can tolerate—and benefit from—high task conflict. An experiment by {{Tooltip|Jennifer Chatman}} and {{Tooltip|Sigal Barsade}} finds that agreeableness adapts to norms: in cooperative climates agreeable people stay accommodating, but in competitive climates they push back as hard as their disagreeable peers, underscoring how context shapes whether conflict helps or harms. Creative work echoes the pattern: Bird’s “good fights” sharpened scenes, and the {{Tooltip|Wright brothers}}’ spirited arguments over propeller design improved solutions. Safeguards include naming the shared goal, framing disputes as debates, and separating critiques of work from judgments of worth. A “challenge network” of disagreeable givers—peers who care enough to criticize—keeps blind spots visible and prevents yes-man traps. Productive tension moves ideas forward when norms welcome task conflict while protecting relationships, and critics pressure-test assumptions so the best ideas, not the loudest egos, win.
 
=== Part II – Interpersonal Rethinking: Opening Other People's Minds ===
🥊 '''4 – The Good Fight Club: The Psychology of Constructive Conflict.''' In 2000 at {{Tooltip|Pixar}}, leaders hired {{Tooltip|Brad Bird}} to “shake things up”; technical heads initially said his vision would take a decade and $500 million, so he built a team of self-described misfits, invited rigorous debate, and four years later delivered Pixar’s most complex film to date while reducing cost per minute of animation. Dissent aimed at ideas, not people, becomes an engine for quality. Research by {{Tooltip|Karen “Etty” Jehn}} distinguishes task conflict (clashes over content and process) from relationship conflict (personal friction) and shows that teams with low relational friction can tolerate—and benefit from—high task conflict. An experiment by {{Tooltip|Jennifer Chatman}} and {{Tooltip|Sigal Barsade}} finds that agreeableness adapts to norms: in cooperative climates agreeable people stay accommodating, but in competitive climates they push back as hard as their disagreeable peers, underscoring how context shapes whether conflict helps or harms. Creative work echoes the pattern: Bird’s “good fights” sharpened scenes, and the {{Tooltip|Wright brothers}}’ spirited arguments over propeller design improved solutions. Safeguards include naming the shared goal, framing disputes as debates, and separating critiques of work from judgments of worth. A “challenge network” of disagreeable givers—peers who care enough to criticize—keeps blind spots visible and prevents yes-man traps. Productive tension moves ideas forward when norms welcome task conflict while protecting relationships, and critics pressure-test assumptions so the best ideas, not the loudest egos, win.
 
=== Chapter 5 – Dances with Foes: How to Win Debates and Influence People ===
=== II – Interpersonal Rethinking: Opening Other People's Minds ===
🗣️ '''5 – Dances with Foes: How to Win Debates and Influence People.''' At {{Tooltip|IBM’s Think conference}} in San Francisco on 11 February 2019, debate champion {{Tooltip|Harish Natarajan}} faced IBM’s {{Tooltip|Project Debater}} before a large live audience on the motion “We should subsidize preschools.” {{Tooltip|Intelligence Squared U.S.}} hosted the event, with moderator John Donvan giving each side 15 minutes to prepare, a four-minute opening, a four-minute rebuttal, and a two-minute close. A pre-debate poll showed most attendees favored subsidies, and the winner would be whoever shifted more minds by the end. Natarajan listened first, surfaced common ground, and then advanced a few strong counterpoints rather than a scattershot list, echoing research on expert negotiators who ask more questions and make fewer, sharper arguments. He steel-manned the AI’s case by restating its best claims and then probing assumptions and trade-offs. Post-debate voting swung enough to award him the victory even as many spectators reported learning from the machine’s torrent of evidence. Repeatable moves follow: lead with curiosity, frame debates as joint problem-solving, and invite the other side to help set the terms for what would change your mind. Persuasion works when it feels like a dance, not a duel—lower psychological reactance, present a small number of well-supported reasons, and let people reason their way to a revised view.
 
=== Chapter 6 – Bad Blood on the Diamond: Diminishing Prejudice by Destabilizing Stereotypes ===
🗣️ '''5 – Dances with Foes: How to Win Debates and Influence People.''' At {{Tooltip|IBM’s Think conference}} in San Francisco on 11 February 2019, debate champion {{Tooltip|Harish Natarajan}} faced IBM’s {{Tooltip|Project Debater}} before a large live audience on the motion “We should subsidize preschools.” {{Tooltip|Intelligence Squared U.S.}} hosted the event, with moderator John Donvan giving each side 15 minutes to prepare, a four-minute opening, a four-minute rebuttal, and a two-minute close. A pre-debate poll showed most attendees favored subsidies, and the winner would be whoever shifted more minds by the end. Natarajan listened first, surfaced common ground, and then advanced a few strong counterpoints rather than a scattershot list, echoing research on expert negotiators who ask more questions and make fewer, sharper arguments. He steel-manned the AI’s case by restating its best claims and then probing assumptions and trade-offs. Post-debate voting swung enough to award him the victory even as many spectators reported learning from the machine’s torrent of evidence. Repeatable moves follow: lead with curiosity, frame debates as joint problem-solving, and invite the other side to help set the terms for what would change your mind. Persuasion works when it feels like a dance, not a duel—lower psychological reactance, present a small number of well-supported reasons, and let people reason their way to a revised view.
'''6 – Bad Blood on the Diamond: Diminishing Prejudice by Destabilizing Stereotypes.''' Historian Doris Kearns Goodwin’s childhood confession of hating the Yankees shows how the {{Tooltip|Boston Red Sox–New York Yankees rivalry}} models tribal thinking that hardens into prejudice. Loyalties born from accidents of birth—what city we grow up in, which colors our family cheers—can morph into sweeping stereotypes about rivals. Field experiments around baseball fandom and the idea that “bad blood” is often arbitrary demonstrate that highlighting the contingency of allegiances softens hostility. Counterfactual prompts—asking fans what they’d believe if they’d been born into the other city—and individuating stories complicate one-dimensional caricatures. {{Tooltip|Daryl Davis}}, a Black blues musician in {{Tooltip|Washington, D.C.}}, offers a stark example: patient conversations with {{Tooltip|Ku Klux Klan}} members led some to renounce the organization and hand over their robes, showing how respectful contact destabilizes stereotypes. Empathy exercises that focus on a specific out-group individual can flip in-group helping biases, especially when people see overlapping identities and goals. Together these tools chip away at identity-protective thinking and make room for nuance without demanding anyone abandon community ties. Progress starts by exposing how flimsy many group labels are and replacing monolithic categories with concrete, conflicting details; counterfactual thinking, shared identities, and humanizing contact disrupt snap generalizations so people update the stories they tell about who “they” are.
 
=== Chapter 7 – Vaccine Whisperers and Mild-Mannered Interrogators: How the Right Kind of Listening Motivates People to Change ===
⚾ '''6 – Bad Blood on the Diamond: Diminishing Prejudice by Destabilizing Stereotypes.''' Historian Doris Kearns Goodwin’s childhood confession of hating the Yankees shows how the {{Tooltip|Boston Red Sox–New York Yankees rivalry}} models tribal thinking that hardens into prejudice. Loyalties born from accidents of birth—what city we grow up in, which colors our family cheers—can morph into sweeping stereotypes about rivals. Field experiments around baseball fandom and the idea that “bad blood” is often arbitrary demonstrate that highlighting the contingency of allegiances softens hostility. Counterfactual prompts—asking fans what they’d believe if they’d been born into the other city—and individuating stories complicate one-dimensional caricatures. {{Tooltip|Daryl Davis}}, a Black blues musician in {{Tooltip|Washington, D.C.}}, offers a stark example: patient conversations with {{Tooltip|Ku Klux Klan}} members led some to renounce the organization and hand over their robes, showing how respectful contact destabilizes stereotypes. Empathy exercises that focus on a specific out-group individual can flip in-group helping biases, especially when people see overlapping identities and goals. Together these tools chip away at identity-protective thinking and make room for nuance without demanding anyone abandon community ties. Progress starts by exposing how flimsy many group labels are and replacing monolithic categories with concrete, conflicting details; counterfactual thinking, shared identities, and humanizing contact disrupt snap generalizations so people update the stories they tell about who “they” are.
💉 '''7 – Vaccine Whisperers and Mild-Mannered Interrogators: How the Right Kind of Listening Motivates People to Change.''' In {{Tooltip|Sherbrooke, Québec}}, Marie-Hélène Étienne-Rousseau delivered a premature son, Tobie, in 2018 and planned to refuse the newborn’s vaccines; neonatologist {{Tooltip|Arnaud Gagneur}} met her at the University of {{Tooltip|Sherbrooke}}’s {{Tooltip|Fleurimont Hospital}} for an unhurried conversation rooted in motivational interviewing. He asked open questions, reflected her worries, and asked permission before sharing information, ending by affirming the decision was hers. Weeks later, after reading about local measles risk, she chose to vaccinate her older children at home and approved Tobie’s shots before discharge. Québec’s “{{Tooltip|PromoVac}}” program placed vaccination counselors in maternity wards and, in trials, improved early-infant vaccine coverage versus usual care. Law enforcement offers a parallel: rapport-based interviewing and language-style matching elicit more accurate information than confrontational, confession-driven tactics. Across settings the pattern holds: preaching and pressure trigger resistance, while reflective listening and autonomy support invite “change talk.” Use the cadence—ask, reflect, affirm, summarize, and ask again—to scale from hospital rooms to tense negotiations without sacrificing truth or agency. Listening that respects people’s goals often persuades better than arguing from authority; self-persuasion sticks when people voice their own reasons for change and a curious interviewer keeps the rethinking cycle alive.
 
== Part III – Collective Rethinking: Creating Communities of Lifelong Learners ==
💉 '''7 – Vaccine Whisperers and Mild-Mannered Interrogators: How the Right Kind of Listening Motivates People to Change.''' In {{Tooltip|Sherbrooke, Québec}}, Marie-Hélène Étienne-Rousseau delivered a premature son, Tobie, in 2018 and planned to refuse the newborn’s vaccines; neonatologist {{Tooltip|Arnaud Gagneur}} met her at the University of {{Tooltip|Sherbrooke}}’s {{Tooltip|Fleurimont Hospital}} for an unhurried conversation rooted in motivational interviewing. He asked open questions, reflected her worries, and asked permission before sharing information, ending by affirming the decision was hers. Weeks later, after reading about local measles risk, she chose to vaccinate her older children at home and approved Tobie’s shots before discharge. Québec’s “{{Tooltip|PromoVac}}” program placed vaccination counselors in maternity wards and, in trials, improved early-infant vaccine coverage versus usual care. Law enforcement offers a parallel: rapport-based interviewing and language-style matching elicit more accurate information than confrontational, confession-driven tactics. Across settings the pattern holds: preaching and pressure trigger resistance, while reflective listening and autonomy support invite “change talk.” Use the cadence—ask, reflect, affirm, summarize, and ask again—to scale from hospital rooms to tense negotiations without sacrificing truth or agency. Listening that respects people’s goals often persuades better than arguing from authority; self-persuasion sticks when people voice their own reasons for change and a curious interviewer keeps the rethinking cycle alive.
 
=== IIIChapter 8CollectiveCharged RethinkingConversations: Creating CommunitiesDepolarizing ofOur LifelongDivided LearnersDiscussions ===
'''8 – Charged Conversations: Depolarizing Our Divided Discussions.''' At {{Tooltip|Columbia University’s Difficult Conversations Lab}}, psychologist {{Tooltip|Peter T. Coleman}} pairs strangers who disagree on polarizing issues and studies what helps them talk productively. In one setup, participants first read about gun control: if the article argued one side, they had roughly a 46% chance of drafting a joint statement; if it framed the topic as complex with shades of gray, every pair found common ground. This “complexifying” move counters binary bias—the tendency to compress a spectrum into two boxes—and shows why merely presenting “both sides” can entrench polarization. Emotion matters too: conversations go better when people feel a mix—curiosity alongside anxiety—rather than being locked in anger. Techniques from persuasion research reinforce the point: ask genuine questions, map the full range of positions, and, when useful, morally reframe arguments to align with the other side’s values without distorting facts. The aim is not to win a verdict but to make thinking more granular so agreement has more places to land. Complexity—not combat—creates conditions for rethinking; replace either/or with a spectrum and broaden the affective palette to reduce reactance and invite cooperative problem-solving.
 
=== Chapter 9 – Rewriting the Textbook: Teaching Students to Question Knowledge ===
⚡ '''8 – Charged Conversations: Depolarizing Our Divided Discussions.''' At {{Tooltip|Columbia University’s Difficult Conversations Lab}}, psychologist {{Tooltip|Peter T. Coleman}} pairs strangers who disagree on polarizing issues and studies what helps them talk productively. In one setup, participants first read about gun control: if the article argued one side, they had roughly a 46% chance of drafting a joint statement; if it framed the topic as complex with shades of gray, every pair found common ground. This “complexifying” move counters binary bias—the tendency to compress a spectrum into two boxes—and shows why merely presenting “both sides” can entrench polarization. Emotion matters too: conversations go better when people feel a mix—curiosity alongside anxiety—rather than being locked in anger. Techniques from persuasion research reinforce the point: ask genuine questions, map the full range of positions, and, when useful, morally reframe arguments to align with the other side’s values without distorting facts. The aim is not to win a verdict but to make thinking more granular so agreement has more places to land. Complexity—not combat—creates conditions for rethinking; replace either/or with a spectrum and broaden the affective palette to reduce reactance and invite cooperative problem-solving.
📚 '''9 – Rewriting the Textbook: Teaching Students to Question Knowledge.''' Wisconsin social-studies teacher Erin McCarthy hands her eighth graders a 1940 history textbook and asks them to read it like investigators, noting errors, omissions, and slanted language; the exercise jolts students into seeing knowledge as provisional, not permanent. She then has them rewrite sections and, in another lesson, study a chapter drafted entirely from girls’ and women’s perspectives, which helps some boys feel how a single-lens narrative distorts the record. Classroom vignettes connect to research showing that active learning—questioning sources, comparing accounts, and teaching others—beats passive lecture for durable understanding. {{Tooltip|Ron Berger}}’s “{{Tooltip|Austin’s Butterfly}}” protocol supplies a concrete craft: a first grader at {{Tooltip|Anser Charter School}} in Boise takes a scientific illustration through multiple drafts, guided by kind, specific peer critique, until the work becomes precise. Students learn to check whether the “sender is the source,” to resist popularity and rankings as proxies for truth, and to ask questions that don’t have just one right answer. Rather than cramming facts, they practice argument literacy: how claims are built, tested, and revised. Position school as a workshop for rethinking—treat confusion as a cue for inquiry and expect drafts on the way to clarity—and bake uncertainty, multiple perspectives, and iterative feedback into tasks so curiosity and intellectual humility become habits, not slogans.
 
=== Chapter 10 – That’s Not the Way We’ve Always Done It: Building Cultures of Learning at Work ===
📚 '''9 – Rewriting the Textbook: Teaching Students to Question Knowledge.''' Wisconsin social-studies teacher Erin McCarthy hands her eighth graders a 1940 history textbook and asks them to read it like investigators, noting errors, omissions, and slanted language; the exercise jolts students into seeing knowledge as provisional, not permanent. She then has them rewrite sections and, in another lesson, study a chapter drafted entirely from girls’ and women’s perspectives, which helps some boys feel how a single-lens narrative distorts the record. Classroom vignettes connect to research showing that active learning—questioning sources, comparing accounts, and teaching others—beats passive lecture for durable understanding. {{Tooltip|Ron Berger}}’s “{{Tooltip|Austin’s Butterfly}}” protocol supplies a concrete craft: a first grader at {{Tooltip|Anser Charter School}} in Boise takes a scientific illustration through multiple drafts, guided by kind, specific peer critique, until the work becomes precise. Students learn to check whether the “sender is the source,” to resist popularity and rankings as proxies for truth, and to ask questions that don’t have just one right answer. Rather than cramming facts, they practice argument literacy: how claims are built, tested, and revised. Position school as a workshop for rethinking—treat confusion as a cue for inquiry and expect drafts on the way to clarity—and bake uncertainty, multiple perspectives, and iterative feedback into tasks so curiosity and intellectual humility become habits, not slogans.
🏢 '''10 – That’s Not the Way We’ve Always Done It: Building Cultures of Learning at Work.''' NASA’s tragedies loom as cautionary tales: before {{Tooltip|Challenger}} and {{Tooltip|Columbia}}, engineers raised concerns that were muted or missed in a culture primed to perform, not to pause. Research by {{Tooltip|Amy Edmondson}} shows why that climate backfires—teams with higher psychological safety surface more errors on paper yet make fewer in practice because people feel free to speak up early. Former {{Tooltip|Johnson Space Center}} director {{Tooltip|Ellen Ochoa}} models how to reset norms: leaders broadcast uncertainties, demand dissent on consequential calls, and keep asking, “How do we know?” so status won’t substitute for scrutiny. Distinguish performance cultures, which canonize “best practices,” from learning cultures, which expect continual updates and separate process accountability (How carefully did we decide?) from outcome accountability (Did it work?). Process accountability shows up in {{Tooltip|premortems}}, red-team reviews, and {{Tooltip|decision logs}} that force alternatives to be considered before momentum hardens. When organizations reward candor and curiosity as much as results, they catch small problems early and adapt faster after setbacks. Rethinking thrives where safety and scrutiny intersect—people feel secure enough to challenge and are obligated to examine how choices are made—so pair psychological safety with process accountability and keep the learning zone as the default.
 
=== Part IV – Conclusion ===
🏢 '''10 – That’s Not the Way We’ve Always Done It: Building Cultures of Learning at Work.''' NASA’s tragedies loom as cautionary tales: before {{Tooltip|Challenger}} and {{Tooltip|Columbia}}, engineers raised concerns that were muted or missed in a culture primed to perform, not to pause. Research by {{Tooltip|Amy Edmondson}} shows why that climate backfires—teams with higher psychological safety surface more errors on paper yet make fewer in practice because people feel free to speak up early. Former {{Tooltip|Johnson Space Center}} director {{Tooltip|Ellen Ochoa}} models how to reset norms: leaders broadcast uncertainties, demand dissent on consequential calls, and keep asking, “How do we know?” so status won’t substitute for scrutiny. Distinguish performance cultures, which canonize “best practices,” from learning cultures, which expect continual updates and separate process accountability (How carefully did we decide?) from outcome accountability (Did it work?). Process accountability shows up in {{Tooltip|premortems}}, red-team reviews, and {{Tooltip|decision logs}} that force alternatives to be considered before momentum hardens. When organizations reward candor and curiosity as much as results, they catch small problems early and adapt faster after setbacks. Rethinking thrives where safety and scrutiny intersect—people feel secure enough to challenge and are obligated to examine how choices are made—so pair psychological safety with process accountability and keep the learning zone as the default.
 
=== Chapter 11 – Escaping Tunnel Vision: Reconsidering Our Best-Laid Career and Life Plans ===
=== IV – Conclusion ===
🔭 '''11 – Escaping Tunnel Vision: Reconsidering Our Best-Laid Career and Life Plans.''' Cousin Ryan, expected by his family to become a physician, felt doubts before applying to college about committing to pre-med and wondered about economics instead; he pushed ahead, completed medical training, and later, as a neurosurgeon, admitted he would have chosen differently. His path shows how sunk costs tempt people to keep investing even when the fit is wrong, a familiar escalation of commitment. A second trap, identity foreclosure, appears when a person adopts a career label early and then closes off exploration, treating a job title as the self. Ask children “what do you like to do?” rather than “what do you want to be?” so possibilities stay open as interests change. Think like a scientist: treat a career direction as a hypothesis, run small tests, and look for disconfirming evidence before doubling down. Drawing on Herminia Ibarra’s ''Working Identity'', act and experiment into clarity instead of waiting for perfect self-knowledge. Low-risk trials—shadowing, project sprints, or job crafting—generate data about enjoyment, strengths, and fit without burning bridges. Keep plans flexible with a twice-yearly career checkup that asks whether learning has plateaued and whether work still matches evolving values and skills. Identities anchored in principles (curiosity, service, integrity) travel better across roles than identities anchored in a single occupation, which makes pivots less threatening. Prevent grit from calcifying into stubbornness by continually testing whether the path still deserves persistence; loosen attachment to a chosen identity and interrupt sunk-cost reasoning so evidence and curiosity—not inertia—steer the next move.
 
''This—Note: outlineThe above summary follows the {{Tooltip|Viking}} hardcover edition (2 February 2021, ISBN 978-1-9848-7810-6).''<ref name="PRH2021">{{cite web |title=Think Again by Adam Grant |url=https://www.penguinrandomhouse.com/books/607660/think-again-by-adam-grant/ |website=Penguin Random House |publisher=Penguin Random House |date=2 February 2021 |access-date=8 November 2025}}</ref><ref name="OCLC1191456279" />
🔭 '''11 – Escaping Tunnel Vision: Reconsidering Our Best-Laid Career and Life Plans.''' Cousin Ryan, expected by his family to become a physician, felt doubts before applying to college about committing to pre-med and wondered about economics instead; he pushed ahead, completed medical training, and later, as a neurosurgeon, admitted he would have chosen differently. His path shows how sunk costs tempt people to keep investing even when the fit is wrong, a familiar escalation of commitment. A second trap, identity foreclosure, appears when a person adopts a career label early and then closes off exploration, treating a job title as the self. Ask children “what do you like to do?” rather than “what do you want to be?” so possibilities stay open as interests change. Think like a scientist: treat a career direction as a hypothesis, run small tests, and look for disconfirming evidence before doubling down. Drawing on Herminia Ibarra’s ''Working Identity'', act and experiment into clarity instead of waiting for perfect self-knowledge. Low-risk trials—shadowing, project sprints, or job crafting—generate data about enjoyment, strengths, and fit without burning bridges. Keep plans flexible with a twice-yearly career checkup that asks whether learning has plateaued and whether work still matches evolving values and skills. Identities anchored in principles (curiosity, service, integrity) travel better across roles than identities anchored in a single occupation, which makes pivots less threatening. Prevent grit from calcifying into stubbornness by continually testing whether the path still deserves persistence; loosen attachment to a chosen identity and interrupt sunk-cost reasoning so evidence and curiosity—not inertia—steer the next move.
 
== Background & reception ==