Skip to main content

Home/ TOK Friends/ Group items matching "drawing" in title, tags, annotations or url

Group items matching
in title, tags, annotations or url

Sort By: Relevance | Date Filter: All | Bookmarks | Topics Simple Middle
25More

Ann Coulter Is Right to Fear the World Cup - Peter Beinart - The Atlantic - 1 views

  • Ann Coulter penned a column explaining why soccer is un-American. First, it’s collectivist. (“Individual achievement is not a big factor…blame is dispersed.”) Second, it’s effeminate. (“It’s a sport in which athletic talent finds so little expression that girls can play with boys.”) Third, it’s culturally elitist. (“The same people trying to push soccer on Americans are the ones demanding that we love HBO’s “Girls,” light-rail, Beyoncé and Hillary Clinton.”) Fourth, and most importantly, “It’s foreign…Soccer is like the metric system, which liberals also adore because it’s European.”
  • Soccer hatred, in other words, exemplifies American exceptionalism.
  • For Coulter and many contemporary conservatives, by contrast, part of what makes America exceptional is its individualism, manliness and populism
  • ...22 more annotations...
  • Coulter’s deeper point is that for America to truly be America, it must stand apart
  • The core problem with embracing soccer is that in so doing, America would become more like the rest of the world.
  • America’s own league, Major League Soccer, draws as many fans to its stadiums as do the NHL and NBA.
  • I wrote an essay entitled “The End of American Exceptionalism,” which argued that on subjects where the United States has long been seen as different, attitudes in America increasingly resemble those in Europe. Soccer is one of the best examples yet.
  • “Soccer,” Markovits and Hellerman argue, “was perceived by both native-born Americans and immigrants as a non-American activity at a time in American history when nativism and nationalism emerged to create a distinctly American self-image … if one liked soccer, one was viewed as at least resisting—if not outright rejecting—integration into America.”
  • The average age of Americans who call baseball their favorite sport is 53. Among Americans who like football best, it’s 46. Among Americans who prefer soccer, by contrast, the average age is only 37.
  • Old-stock Americans, in other words, were elevating baseball, football, and basketball into symbols of America’s distinct identity. Immigrants realized that embracing those sports offered a way to claim that identity for themselves. Clinging to soccer, by contrast, was a declaration that you would not melt.
  • why is interest in soccer rising now? Partly, because the United States is yet again witnessing mass immigration from soccer-mad nations.
  • the key shift is that America’s sports culture is less nativist. More native-born Americans now accept that a game invented overseas can become authentically American, and that the immigrants who love it can become authentically American too. Fewer believe that to have merit, something must be invented in the United States.
  • why didn’t soccer gain a foothold in the U.S. in the decades between the Civil War and World War I, when it was gaining dominance in Europe? Precisely because it was gaining dominance in Europe. The arbiters of taste in late 19th and early 20th century America wanted its national pastimes to be exceptional.
  • Americans over the age of 50 were 15 points more likely to say “our culture is superior” than were people over 50 in Germany, Spain, Britain, and France
  • Americans under 30, by contrast, were actually less likely to say “our culture is superior” than their counterparts in Germany, Spain, and Britain.
  • Americans today are less likely to insist that America’s way of doing things is always best. In 2002, 60 percent of Americans told the Pew Research Center that, “our culture is superior to others.” By 2011, it was down to 49 percent.
  • the third major pro-soccer constituency is liberals. They’re willing to embrace a European sport for the same reason they’re willing to embrace a European-style health care system: because they see no inherent value in America being an exception to the global rule
  • When the real-estate website Estately created a seven part index to determine a state’s love of soccer, it found that Washington State, Maryland, the District of Columbia, New York, and New Jersey—all bright blue—loved soccer best, while Alabama, Arkansas, North Dakota, Mississippi and Montana—all bright red—liked it least.
  • the soccer coalition—immigrants, liberals and the young—looks a lot like the Obama coalition.
  • Sports-wise, therefore, Democrats constitute an alliance between soccer and basketball fans while Republicans disproportionately follow baseball, golf, and NASCAR. Football, by far America’s most popular sport, crosses the aisle.)
  • The willingness of growing numbers of Americans to embrace soccer bespeaks their willingness to imagine a different relationship with the world. Historically, conservative foreign policy has oscillated between isolationism and imperialism. America must either retreat from the world or master it. It cannot be one among equals, bound by the same rules as everyone else
  • Exceptionalists view sports the same way. Coulter likes football, baseball, and basketball because America either plays them by itself, or—when other countries play against us—we dominate them.
  • Embracing soccer, by contrast, means embracing America’s role as merely one nation among many, without special privileges. It’s no coincidence that young Americans, in addition to liking soccer, also like the United Nations. In 2013, Pew found that Americans under 30 were 24 points more favorable to the U.N. than Americans over 50.
  • Millennials were also 23 points more likely than the elderly to say America should take its allies’ opinion into account even if means compromising our desires.
  • In embracing soccer, Americans are learning to take something we neither invented nor control, and nonetheless make it our own. It’s a skill we’re going to need in the years to come.
7More

Is the World More Depressed? - NYTimes.com - 2 views

  • The World Health Organization reports that suicide rates have increased 60 percent over the past 50 years, most strikingly in the developing world, and that by 2020 depression will be the second most prevalent medical condition in the world.
  • n 2011, the Centers for Disease Control and Prevention reported that the rate of antidepressant use in the United States rose by 400 percent between 1988 and 2008.
  • there is reason to believe that mental illness is indeed increasing around the world, if only because urbanization is increasing. By 2010, for the first time in history, more than half the world’s population lived in cities.
  • ...4 more annotations...
  • What has exploded in India over the past few decades, but also everywhere else in the world, is information about other people. As we watch television, surf the Internet and follow events around the world, we become intimately aware of other ways of living and of others who are richer and more powerful. We place ourselves in a vast social order in which most of us are ants. It may truly be a depressing reflection.
  • We know that social position affects both when you die and how sick you get: The higher your social position, the healthier you are. It turns out that your sense of relative social rank — where you draw a line on an abstract ladder to show where you are with respect to others — predicts many health outcomes, including depression, sometimes even more powerfully than your objective socioeconomic status.
  • cities also break traditions and fracture families, and they breed psychiatric illness. In a city you are more likely to be depressed, to fall ill with schizophrenia, and to use alcohol and drugs. Poverty and rapid urbanization sharpen these effects.
  • Some of these figures might simply reflect more willingness to label an experience as a symptom. For example, until recently, most Japanese understood intense fatigue as sacrifice for one’s work and suicide as an act of reasoned will. In her book “Depression in Japan,” the anthropologist Junko Kitanaka writes that partly as a result of aggressive pharmaceutical marketing, many Japanese began to think of their fatigue and suicidal thoughts as symptoms created by a disease. The number of diagnoses of depression in that country more than doubled between 1999 and 2008.
8More

The Fall of Facebook - The Atlantic - 0 views

  • Alexis C. Madrigal Nov 17 2014, 7:59 PM ET Social networking is not, it turns out, winner take all. In the past, one might have imagined that switching between Facebook and “some other network” would be difficult, but the smartphone interface makes it easy to be on a dozen networks. All messages come to the same place—the phone’s notifications screen—so what matters is what your friends are doing, not which apps they’re using.
  • if I were to put money on an area in which Facebook might be unable to dominate in the future, it would be apps that take advantage of physical proximity. Something radically new could arise on that front, whether it’s an evolution of Yik Yak
  • The Social Machine, predicts that text will be a less and less important part of our asynchronous communications mix. Instead, she foresees a “very fluid interface” that would mix text with voice, video, sensor outputs (location, say, or vital signs), and who knows what else
  • ...5 more annotations...
  • the forthcoming Apple Watch seems like a step toward the future Donath envisions. Users will be able to send animated smiley faces, drawings, voice snippets, and even their live heartbeats, which will be tapped out on the receiver’s wrist.
  • A simple but rich messaging platform—perhaps with specialized hardware—could replace the omnibus social network for most purposes. “I think we’re shifting in a weird way to one-on-one conversations on social networks and in messaging apps,” says Shani Hilton, the executive editor for news at BuzzFeed, the viral-media site. “People don’t want to perform their lives publicly in the same way that they wanted to five years ago.”
  • Facebook is built around a trade-off that it has asked users to make: Give us all your personal information, post all your pictures, tag all your friends, and so on, forever. In return, we’ll optimize your social life. But this output is only as good as the input. And it turns out that, when scaled up, creating this input—making yourself legible enough to the Facebook machine that your posts are deemed “relevant” and worthy of being displayed to your mom and your friends—is exhausting labor.
  • These new apps, then, are arguments that we can still have an Internet that is weird, and private. That we can still have social networks without the social network. And that we can still have friends on the Internet without “friending” them.
  • A Brief History of Information Gatekeepers 1871: Western Union controls 90 percent of U.S. telegraph traffic. 1947: 97 percent of the country’s radio stations are affiliated with one of four national networks. 1969: Viewership for the three nightly network newscasts hits an all-time high, with 50 percent of all American homes tuning in. 1997: About half of all American homes with Internet access get it through America Online. 2002: Microsoft Internet Explorer captures 97 percent of the worldwide browser market. 2014: Amazon sells 63 percent of all books bought online—and 40 percent of books overall.
16More

Evolution and the American Myth of the Individual - NYTimes.com - 0 views

  • the country’s two main political parties have “fundamental philosophical differences.” But what exactly does that mean?
  • In a broad sense, Democrats, particularly the more liberal among them, are more likely to embrace the communal nature of individual lives and to strive for policies that emphasize that understanding.
  • Republicans, especially libertarians and Tea Party members on the ideological fringe, however, often trace their ideas about freedom and liberty back to Enlightenment thinkers of the 17th and 18th centuries, who argued that the individual is the true measure of human value, and each of us is naturally entitled to act in our own best interests free of interference by others. Self-described libertarians generally also pride themselves on their high valuation of logic and reasoning over emotion.
  • ...13 more annotations...
  • Philosophers from Aristotle to Hegel have emphasized that human beings are essentially social creatures, that the idea of an isolated individual is a misleading abstraction. So it is not just ironic but instructive that modern evolutionary research, anthropology, cognitive psychology and neuroscience have come down on the side of the philosophers who have argued that the basic unit of human social life is not and never has been the selfish, self-serving individual.
  • The irony here is that when it comes to our responsibilities to one another as human beings, religion and evolution nowadays are not necessarily on opposite sides of the fence.
  • in the eyes of many conservative Americans today, religion and evolution do not mix. You either accept what the Bible tells us or what Charles Darwin wrote, but not both
  • Contrary to libertarian and Tea Party rhetoric, evolution has made us a powerfully social species, so much so that the essential precondition of human survival is and always has been the individual plus his or her relationships with others.
  • as Matthew D. Lieberman, a social neuroscience researcher at the University of California, Los Angeles, has written: “we think people are built to maximize their own pleasure and minimize their own pain. In reality, we are actually built to overcome our own pleasure and increase our own pain in the service of following society’s norms.”
  • Why then did Rousseau and others make up stories about human history if they didn’t really believe them? The simple answer, at least during the Enlightenment, was that they wanted people to accept their claim that civilized life is based on social conventions, or contracts, drawn up at least figuratively speaking by free, sane and equal human beings — contracts that could and should be extended to cover the moral and working relationships that ought to pertain between rulers and the ruled.
  • Jean-Jacques Rousseau famously declared in “The Social Contract” (1762) that each of us is born free and yet everywhere we are in chains. He did not mean physical chains. He meant social ones. We now know he was dead wrong. Human evolution has made us obligate social creatures. Even if some of us may choose sooner or later to disappear into the woods or sit on a mountaintop in deep meditation, we humans are able to do so only if before such individualistic anti-social resolve we have first been socially nurtured and socially taught survival arts by others. The distinction Rousseau and others tried to draw between “natural liberty, which is bounded only by the strength of the individual” and “civil liberty, which is limited by the general will” is fanciful, not factual.
  • In short, their aims were political, not historical, scientific or religious.
  • what Rousseau and others crafted as arguments in favor of their ideas all had the earmarks of primitive mythology
  • Bronislaw Malinowski argued almost a century ago: “Myth fulfills in primitive culture an indispensable function: it expresses, enhances, and codifies belief, it safeguards and enforces morality, it vouches for the efficiency of ritual and contains practical rules for the guidance of man.”
  • not all myths make good charters for faith and wisdom. The sanctification of the rights of individuals and their liberties today by libertarians and Tea Party conservatives is contrary to our evolved human nature as social animals. There was never a time in history before civil society when we were each totally free to do whatever we elected to do. We have always been social and caring creatures. The thought that it is both rational and natural for each of us to care only for ourselves, our own preservation, and our own achievements is a treacherous fabrication. This is not how we got to be the kind of species we are today.
  • Myths achieve this social function, he observed, by serving as guides, or charters, for moral values, social order and magical belief.
  • Nor is this what the world’s religions would ask us to believe.
17More

Heady Stakes for 'Black-ish' on ABC - NYTimes.com - 0 views

  • hovering above all that is a more subtle — and quietly clever — narrative arc, involving the gap between parents and children and how each generation has a different awareness of what it means to be black in 2014.
  • I want it to succeed because the show arrives when black characters on mainstream broadcast networks who directly deal with issues like race are incredibly rare.
  • so far, his approach seems to be a hit. The premiere resonated with critics and attracted a robust 11 million viewers, besides generating a lot of positive reactions and discussions on social media. In a vote of confidence, ABC has given the show a full-season order.
  • ...14 more annotations...
  • it seems as if networks think that post-racial story lines are the only acceptable ways of showcasing black characters on television.
  • TV is resplendent with ethnically diverse casts, from procedurals like “Law & Order: SVU” and “NCIS: Los Angeles” to hits like “Scandal” and “Elementary” to sitcoms like “New Girl” and “Brooklyn Nine-Nine.”  But the characters on those series don’t often deal directly with racial issues in everyday life and, by not doing so, perpetuate another kind of colorblindness, one that homogenizes characters and treats race as inconsequential, when it is anything but.
  • “The PC way of handling culture has been to not talk about it,” Kenya Barris, the show’s creator, said in an interview. “But we should be talking about it.”
  • What black viewers are left with instead, said Dayna Chatman, a media researcher at the Annenberg School for Communication and Journalism at the University of Southern California, is a dynamic that “makes whiteness the norm.”
  • reality television often showcases African-Americans, but since that genre is often about over-the-top performances, she said, it isn’t “particularly representative or flattering.”
  • there’s no middle ground: Either race is largely absent or exaggerated to the point of caricature.
  • The lack of texture and diversity on television is harder to ignore amid the rise of streaming and online series (say, Netflix’s “Orange Is the New Black” or Issa Rae’s “The Misadventures of Awkward Black Girl”) as well as social media like Instagram and Vine (see King Bach’s account). They offer a welcome and much more nuanced window into black humor and culture.
  • “The business explanation is always that this isn’t what the marketplace is asking for,”
  • He said ABC and cable networks pursued him and the “Black-ish” pilot “very aggressively.” Contrary to popular belief, networks “are looking for something that deals with diversity,” he said. “The problem has been timing and having the right package behind it.”
  • “This is the first time in American history where the most famous people in America are black,” he said, naming the Obama family and the musicians Kanye West and Beyoncé. “But there’s still a really obvious invisibility on television.”
  • Mr. Barris said he was determined to do more than create a successor to “The Cosby Show,” although “Black-ish” draws from its legacy. But while the popularity of the Huxtable family centered on its warmth and relatability, it was, Mr. Barris said, “about a family that happened to be black.” He added that he wanted his show to be much more cognizant of modern racial identity, and to reflect the class and racial dynamics of being black in America.
  • “We are hyperaware of how people and the media perceive us,” she said. “And who gets it and who doesn’t get it.”
  • In 2005, Mr. Chappelle walked away from his lucrative show on Comedy Central after expressing discomfort that the line between his social commentary and racial satire had grown too thin.
  • Today, there are a few other shows operating in the space left behind by Mr. Chappelle, including “Key & Peele” on Comedy Central and “Black Jesus” on Adult Swim. But those are cable outlets with smaller audiences, whereas “Black-ish” is on a mainstream network.
8More

Why Kids Sext - Hanna Rosin - The Atlantic - 0 views

  • Within an hour, the deputies realized just how common the sharing of nude pictures was at the school. “The boys kept telling us, ‘It’s nothing unusual. It happens all the time,’ ” Lowe recalls. Every time someone they were interviewing mentioned another kid who might have naked pictures on his or her phone, they had to call that kid in for an interview. After just a couple of days, the deputies had filled multiple evidence bins with phones, and they couldn’t see an end to it. Fears of a cabal got replaced by a more mundane concern: what to do with “hundreds of damned phones. I told the deputies, ‘We got to draw the line somewhere or we’re going to end up talking to every teenager in the damned county!’ ”
  • Nor did the problem stop at the county’s borders. Several boys, in an effort to convince Lowe that they hadn’t been doing anything rare or deviant, showed him that he could type the hashtag symbol (#) into Instagram followed by the name of pretty much any nearby county and then thots, and find a similar account.
  • In some he sensed low self-esteem—for example, the girl who’d sent her naked picture to a boy, unsolicited: “It just showed up! I guess she was hot after him?” A handful of senior girls became indignant during the course of the interview. “This is my life and my body and I can do whatever I want with it,” or, “I don’t see any problem with it. I’m proud of my body,” Lowe remembers them saying. A few, as far as he could tell, had taken pictures especially for the Instagram accounts and had actively tried to get them posted.
  • ...5 more annotations...
  • What seemed to mortify them most was having to talk about what they’d done with a “police officer outside their age group.”
  • Most of the girls on Instagram fell into the same category as Jasmine. They had sent a picture to their boyfriend, or to someone they wanted to be their boyfriend, and then he had sent it on to others. For the most part, they were embarrassed but not devastated, Lowe said. They felt betrayed, but few seemed all that surprised that their photos had been passed around.
  • Lowe’s team explained to both the kids pictured on Instagram and the ones with photos on their phones the serious legal consequences of their actions. Possessing or sending a nude photo of a minor—even if it’s a photo of yourself—can be prosecuted as a felony under state child-porn laws. He explained that 10 years down the road they might be looking for a job or trying to join the military, or sitting with their families at church, and the pictures could wash back up; someone who had the pictures might even try to blackmail them.
  • yet the kids seemed strikingly blasé. “They’re just sitting there thinking, Wah, wah, wah,” Lowe said, turning his hands into flapping lips. “It’s not sinking in. Remember at that age, you think you’re invincible, and you’re going to do whatever the hell you want to do? We just couldn’t get them past that.”
  • while adults send naked pictures too, of course, the speed with which teens have incorporated the practice into their mating rituals has taken society by surprise.
37More

Technology Imperialism, the Californian Ideology, and the Future of Higher Education - 2 views

  • What I hope to make explicit today is how much California – the place, the concept, “the dream machine” – shapes (wants to shape) the future of technology and the future of education.
  • In an announcement on Facebook – of course – Zuckerberg argued that “connectivity is a human right.”
  • As Zuckerberg frames it at least, the “human right” in this case is participation in the global economy
  • ...34 more annotations...
  • This is a revealing definition of “human rights,” I’d argue, particularly as it’s one that never addresses things like liberty, equality, or justice. It never addresses freedom of expression or freedom of assembly or freedom of association.
  • in certain countries, a number of people say they do not use the Internet yet they talk about how much time they spend on Facebook. According to one survey, 11% of Indonesians who said they used Facebook also said they did not use the Internet. A survey in Nigeria had similar results:
  • Evgeny Morozov has described this belief as “Internet-centrism,” an ideology he argues permeates the tech industry, its PR wing the tech blogosphere, and increasingly government policy
  • “Internet-centrism” describes the tendency to see “the Internet” – Morozov uses quotations around the phrase – as a new yet unchanging, autonomous, benevolent, and inevitable socio-technological development. “The Internet” is a master framework for how all institutions will supposedly operate moving forward
  • “The opportunity to connect” as a human right assumes that “connectivity” will hasten the advent of these other rights, I suppose – that the Internet will topple dictatorships, for example, that it will extend participation in civic life to everyone and, for our purposes here at this conference, that it will “democratize education.”
  • Empire is not simply an endeavor of the nation-state – we have empire through technology (that’s not new) and now, the technology industry as empire.
  • Facebook is really just synecdochal here, I should add – just one example of the forces I think are at play, politically, economically, technologically, culturally.
  • it matters at the level of ideology. Infrastructure is ideological, of course. The new infrastructure – “the Internet” if you will – has a particular political, economic, and cultural bent to it. It is not neutral.
  • This infrastructure matters. In this case, this is a French satellite company (Eutelsat). This is an American social network (Facebook). Mark Zuckerberg’s altruistic rhetoric aside, this is their plan – an economic plan – to monetize the world’s poor.
  • The content and the form of “connectivity” perpetuate imperialism, and not only in Africa but in all of our lives. Imperialism at the level of infrastructure – not just cultural imperialism but technological imperialism
  • “The Silicon Valley Narrative,” as I call it, is the story that the technology industry tells about the world – not only the world-as-is but the world-as-Silicon-Valley-wants-it-to-be.
  • To better analyze and assess both technology and education technology requires our understanding of these as ideological, argues Neil Selwyn – “‘a site of social struggle’ through which hegemonic positions are developed, legitimated, reproduced and challenged.”
  • This narrative has several commonly used tropes
  • It often features a hero: the technology entrepreneur. Smart. Independent. Bold. Risk-taking. White. Male
  • “The Silicon Valley narrative” invokes themes like “innovation” and “disruption.” It privileges the new; everything else that can be deemed “old” is viewed as obsolete.
  • It contends that its workings are meritocratic: anyone who hustles can make it.
  • “The Silicon Valley narrative” fosters a distrust of institutions – the government, the university. It is neoliberal. It hates paying taxes.
  • “The Silicon Valley narrative” draws from the work of Ayn Rand; it privileges the individual at all costs; it calls this “personalization.”
  • “The Silicon Valley narrative” does not neatly co-exist with public education. We forget this at our peril. This makes education technology, specifically, an incredibly fraught area.
  • Here’s the story I think we like to hear about ed-tech, about distance education, about “connectivity” and learning: Education technology is supportive, not exploitative. Education technology opens, not forecloses, opportunities. Education technology is driven by a rethinking of teaching and learning, not expanding markets or empire. Education technology meets individual and institutional and community goals.
  • That’s not really what the “Silicon Valley narrative” says about education
  • It is interested in data extraction and monetization and standardization and scale. It is interested in markets and return on investment. “Education is broken,” and technology will fix it
  • If “Silicon Valley” isn’t quite accurate, then I must admit that the word “narrative” is probably inadequate too
  • The better term here is “ideology.”
  • Facebook is “the Internet” for a fairly sizable number of people. They know nothing else – conceptually, experientially. And, let’s be honest, Facebook wants to be “the Internet” for everyone.
  • We tend to not see technology as ideological – its connections to libertarianism, neoliberalism, global capitalism, empire.
  • The California ideology ignores race and labor and the water supply; it is sustained by air and fantasy. It is built upon white supremacy and imperialism.
  • As is the technology sector, which has its own history, of course, in warfare and cryptography.
  • So far this year, some $3.76 billion of venture capital has been invested in education technology – a record-setting figure. That money will change the landscape – that’s its intention. That money carries with it a story about the future; it carries with it an ideology.
  • When a venture capitalist says that “software is eating the world,” we can push back on the inevitability implied in that. We can resist – not in the name of clinging to “the old” as those in educational institutions are so often accused of doing – but we can resist in the name of freedom and justice and a future that isn’t dictated by the wealthiest white men in Hollywood or Silicon Valley.
  • We in education would be naive, I think, to think that the designs that venture capitalists and technology entrepreneurs have for us would be any less radical than creating a new state, like Draper’s proposed state of Silicon Valley, that would enormously wealthy and politically powerful.
  • When I hear talk of “unbundling” in education – one of the latest gerunds you’ll hear venture capitalists and ed-tech entrepreneurs invoke, meaning the disassembling of institutions into products and services – I can’t help but think of the “unbundling” that Draper wished to do to my state: carving up land and resources, shifting tax revenue and tax burdens, creating new markets, privatizing public institutions, redistributing power and doing so explicitly not in the service of equity or justice.
  • I want to show you this map, a proposal – a failed proposal, thankfully – by venture capitalist Tim Draper to split the state of California into six separate states: Jefferson, North California, Silicon Valley, Central California, West California, and South California. The proposal, which Draper tried to collect enough signatures to get on the ballot in California, would have created the richest state in the US – Silicon Valley would be first in per-capita income. It would also have created the nation’s poorest state, Central California, which would rank even below Mississippi.
  • that’s not all that Silicon Valley really does.
2More

A Basic Guide for Curious Minds | Bill Gates - 1 views

  • Randall Munroe’s new book, Thing Explainer: Complicated Stuff in Simple Words,
  • Munroe sets out to explain various subjects—from how smartphones work to what the U.S. Constitution says—without any complicated terms. Instead he draws blueprint-style diagrams and annotates them using only the 1,000 most common words in the English language. A nuclear reactor is a “heavy metal power building.” A dishwasher is a “box that cleans food holders.”
11More

Scientists Seek Ban on Method of Editing the Human Genome - NYTimes.com - 1 views

  • A group of leading biologists on Thursday called for a worldwide moratorium on use of a new genome-editing technique that would alter human DNA in a way that can be inherited.
  • The biologists fear that the new technique is so effective and easy to use that some physicians may push ahead before its safety can be assessed. They also want the public to understand the ethical issues surrounding the technique, which could be used to cure genetic diseases, but also to enhance qualities like beauty or intelligence. The latter is a path that many ethicists believe should never be taken.
  • a technique invented in 2012 makes it possible to edit the genome precisely and with much greater ease. The technique has already been used to edit the genomes of mice, rats and monkeys, and few doubt that it would work the same way in people.
  • ...8 more annotations...
  • The technique holds the power to repair or enhance any human gene. “It raises the most fundamental of issues about how we are going to view our humanity in the future and whether we are going to take the dramatic step of modifying our own germline and in a sense take control of our genetic destiny, which raises enormous peril for humanity,”
  • The paper’s authors, however, are concerned about countries that have less regulation in science. They urge that “scientists should avoid even attempting, in lax jurisdictions, germline genome modification for clinical application in humans” until the full implications “are discussed among scientific and governmental organizations.”
  • Though such a moratorium would not be legally enforceable and might seem unlikely to exert global influence, there is a precedent. In 1975, scientists worldwide were asked to refrain from using a method for manipulating genes, the recombinant DNA technique, until rules had been established.
  • Though highly efficient, the technique occasionally cuts the genome at unintended sites. The issue of how much mistargeting could be tolerated in a clinical setting is one that Dr. Doudna’s group wants to see thoroughly explored before any human genome is edited.
  • “We worry about people making changes without the knowledge of what those changes mean in terms of the overall genome,” Dr. Baltimore said. “I personally think we are just not smart enough — and won’t be for a very long time — to feel comfortable about the consequences of changing heredity, even in a single individual.”
  • Many ethicists have accepted the idea of gene therapy, changes that die with the patient, but draw a clear line at altering the germline, since these will extend to future generations. The British Parliament in February approved the transfer of mitochondria, small DNA-containing organelles, to human eggs whose own mitochondria are defective. But that technique is less far-reaching because no genes are edited.
  • There are two broad schools of thought on modifying the human germline, said R. Alta Charo, a bioethicist at the University of Wisconsin and a member of the Doudna group. One is pragmatic and seeks to balance benefit and risk. The other “sets up inherent limits on how much humankind should alter nature,” she said. Some Christian doctrines oppose the idea of playing God, whereas in Judaism and Islam there is the notion “that humankind is supposed to improve the world.” She described herself as more of a pragmatist, saying, “I would try to regulate such things rather than shut a new technology down at its beginning.
  • The Doudna group calls for public discussion, but is also working to develop some more formal process, such as an international meeting convened by the National Academy of Sciences, to establish guidelines for human use of the genome-editing technique.“We need some principled agreement that we want to enhance humans in this way or we don’t,” Dr. Jaenisch said. “You have to have this discussion because people are gearing up to do this.”
11More

What Architecture Is Doing to Your Brain - CityLab - 1 views

  • Much of the student population would likely agree that the library’s menacing figure on the quad is nothing short of soul-crushing. New research conducted by a team of architects and neuroscientists suggests that architecture may indeed affect mental states, though they choose to focus on the positive.
  • I spoke with Dr. Julio Bermudez, the lead of a new study that uses fMRI to capture the effects of architecture on the brain. His team operates with the goal of using the scientific method to transform something opaque—the qualitative “phenomenologies of our built environment”—into neuroscientific observations that architects and city planners can deliberately design for. Bermudez and his team’s research question focuses on buildings and sites designed to elicit contemplation: They theorize that the presence of “contemplative architecture” in one’s environment may over time produce the same health benefits as traditional “internally induced” meditation, except with much less effort by the individual.
  • By showing 12 architects photos of contemplative and non-contemplative buildings from facade to interior, the researchers were able to observe the brain activity that occurred as subjects "imagined they were transported to the places being shown." All of the architects were white, right-handed men with no prior meditative training, creating the necessary (if comical) uniformity for neuroscientific research—the team wanted to ensure that the brain scans would not be influenced by factors unrelated to the photos, like gender, race, or handedness. For instance, the brain scans of left- and right-handed people often look different even when subjects are performing the same task.
  • ...8 more annotations...
  • In addition to posing an interesting control on the experiment, the decision to use architects was a strategic one meant to increase the researchers’ chances of achieving conclusive results. Though everyone encounters architecture, studies on the built environment struggle for funding because, as Bermudez remarked with a sigh, “it’s difficult to suggest that people are dying from it.” Architects were a natural choice for the pilot study because, the team reasoned, their critical training and experience would make them sensitive to features of the buildings that a lay person might overlook.
  • they deployed online surveys in Spanish and English to gather testimony on extraordinary architectural experiences (EAEs), or encounters with places that fundamentally alter one’s normal state of being. Critically, most of the buildings or sites mentioned in the 2,982 testimonies were designed with contemplation in mind, whether spiritual, aesthetic, religious, or symbolic, leading the researchers to conclude that “buildings may induce insightful, profound, and transformative contemplative states, [and] buildings designed to provoke contemplation seem to be succeeding”
  • Anticipating skeptics who would claim that these experiences are subjective, the researchers expanded the question to draw on the established neuroscientific subfield of meditation, with some important differences. Related studies to date have focused on internally produced states that are easily replicated in the lab, and on aesthetic evaluation, or the activity that occurs in the orbital frontal cortex as we make snap judgments about whether we find things ugly or beautiful.
  • Bermudez and his team expected that architecturally induced contemplative states would be strong, non-evaluative aesthetic experiences— eliciting more activity in areas associated with emotion and pleasure, but less activity in the orbital frontal cortex.
  • The presence of an external stimulus (the photos of the buildings) also removes the tedious self-regulation that occurs in the prefrontal cortex during traditional meditation. The interviews of the 12 subjects revealed that “peacefulness and relaxation, lessening of mind wandering, increasing of attention, and deepening of experience” were all common effects of viewing the photos—also common was a slight element of aesthetic judgment, seemingly inescapable in the crowd of critics.
  • The provisional conclusions of the study are that the brain behaves differently when exposed to contemplative and non-contemplative buildings, contemplative states elicited through “architectural aesthetics” are similar to the contemplation of traditional meditation in some ways, and different in other ways, and, finally, that “architectural design matters.”
  • reinforces a growing trend in architecture and design as researchers are beginning to study how the built environment affects the people who live in it. ANFA proclaims that “some observers have characterized what is happening in neuroscience as the most exciting frontier of human discovery since the Renaissance.”
  • gritty details: the optimal ceiling heights for different cognitive functions; the best city design for eliciting our natural exploratory tendencies and making way-finding easier; the ideal hospital layout to improve memory-related tasks in patients recovering from certain brain injuries; the influence of different types and quantities of light within a built space on mood and performance.  
24More

Jordan Peterson Comes to Aspen - The Atlantic - 0 views

  • Peterson is traveling the English-speaking world in order to spread the message of this core conviction: that the way to fix what ails Western societies is a psychological project, targeted at helping individuals to get their lives in order, not a sociological project that seeks to improve society through politics, or popular culture, or by focusing on class, racial, or gender identity.
  • the Aspen Ideas Festival, which is co-sponsored by the Aspen Institute and The Atlantic, was an anomaly in this series of public appearances: a gathering largely populated by people—Democrats and centrist Republicans, corporate leaders, academics, millionaire philanthropists, journalists—invested in the contrary proposition, that the way to fix what ails society is a sociological project, one that effects change by focusing on politics, or changing popular culture, or spurring technological advances, or investing more in diversity and inclusiveness.
  • Many of its attendees, like many journalists, are most interested in Peterson as a political figure at the center of controversies
  • ...21 more annotations...
  • Peterson deserves a full, appropriately complex accounting of his best and worst arguments; I intend to give him one soon. For now, I can only tell you how the Peterson phenomenon manifested one night in Aspen
  • “For the first time in human history the spoken word has the same reach as the written word, and there are no barriers to entry. That’s a Gutenberg revolution,” he said. “That’s a big deal. This is a game changer. The podcast world is also a Gutenberg moment but it’s even more extensive. The problem with books is that you can’t do anything else while you’re reading. But if you’re listening to a podcast you can be driving a tractor or a long haul truck or doing the dishes. So podcasts free up two hours a day for people to engage in educational activity they otherwise wouldn’t be able to engage in. That’s one-eighth of people’s lives. You’re handing people a lot of time back to engage in high-level intellectual education.
  • that technological revolution has revealed something good that we didn’t know before: “The narrow bandwidth of TV has made us think that we are stupider than we are. And people have a real hunger for deep intellectual dialogue.”
  • I’ve known for years that the university underserved the community, because we assumed that university education is for 18- to 22-year-olds, which is a proposition that’s so absurd it is absolutely mind-boggling that anyone ever conceptualized it. Why wouldn’t you take university courses throughout your entire life? What, you stop searching for wisdom when you’re 22? I don’t think so. You don’t even start until you’re like in your mid 20s. So I knew universities were underserving the broader community a long time ago. But there wasn’t a mechanism whereby that could be rectified.
  • Universities are beyond forgiveness, he argued, because due to the growing ranks of administrators, there’s been a radical increase in tuition. “Unsuspecting students are given free access to student loans that will cripple them through their 30s and their 40s, and the universities are enticing them to extend their carefree adolescence for a four year period at the cost of mortgaging their future in a deal that does not allow for escape through bankruptcy,” he complained. “So it’s essentially a form of indentured servitude. There’s no excuse for that … That cripples the economy because the students become overlaid with debt that they’ll never pay off at the time when they should be at the peak of their ability to take entrepreneurial risks. That’s absolutely appalling.”
  • A critique I frequently hear from Peterson’s critics is that everything he says is either obvious or wrong. I think that critique fails insofar as I sometimes see some critics calling one of his statements obvious even as others insist it is obviously wrong.
  • a reliable difference among men and women cross-culturally is that men are more aggressive than women. Now what's the evidence for that? Here's one piece of evidence: There are 10 times as many men in prison. Now is that a sociocultural construct? It's like, no, it's not a sociocultural construct. Okay?
  • Here's another piece of data. Women try to commit suicide more than men by a lot, and that's because women are more prone to depression and anxiety than men are. And there are reasons for that, and that's cross-cultural as well. Now men are way more likely to actually commit suicide. Why? Because they're more aggressive so they use lethal means. So now the question is how much more aggressive are men than women? The answer is not very much. So the claim that men and women are more the same than different is actually true. This is where you have to know something about statistics to understand the way the world works, instead of just applying your a priori ideological presuppositions to things that are too complex to fit in that rubric.
  • So if you draw two people out of a crowd, one man and one woman, and you had to lay a bet on who was more aggressive, and you bet on the woman, you'd win 40 percent of the time. That's quite a lot. It isn't 50 percent of the time which would be no differences. But it’s a lot. There are lots of women who are more aggressive than lots of men. So the curves overlap a lot. There's way more similarity than difference. And this is along the dimension where there's the most difference. But here's the problem. You can take small differences at the average of a distribution. Then the distributions move off to the side. And then all the action is at the tail. So here's the situation. You don't care about how aggressive the average person is. It's not that relevant. What people care about is who is the most aggressive person out of 100, because that's the person you'd better watch out for.
  • Whenever I'm interviewed by journalists who have the scent of blood in their nose, let's say, they're very willing and able to characterize the situation I find myself in as political. But that's because they can't see the world in any other manner. The political is a tiny fraction of the world. And what I'm doing isn't political. It's psychological or theological. The political element is peripheral. And if people come to the live lectures, let's say, that's absolutely self-evident
  • In a New York Times article titled, “Jordan Peterson, Custodian of the Patriarchy,” the writer Nellie Bowles quoted her subject as follows:
  • Violent attacks are what happens when men do not have partners, Mr. Peterson says, and society needs to work to make sure those men are married. “He was angry at God because women were rejecting him,” Mr. Peterson says of the Toronto killer. “The cure for that is enforced monogamy. That’s actually why monogamy emerges.” Mr. Peterson does not pause when he says this. Enforced monogamy is, to him, simply a rational solution. Otherwise women will all only go for the most high-status men, he explains, and that couldn’t make either gender happy in the end.
  • Ever since, some Peterson critics have claimed that Peterson wants to force women to have sex with male incels, or something similarly dystopian.
  • ...it's an anthropological truism generated primarily through scholars on the left, just so everybody is clear about it, that societies that use monogamy as a social norm, which by the way is virtually every human society that ever existed, do that in an attempt to control the aggression that goes along with polygamy. It's like ‘Oh my God, how contentious can you get.’ Well, how many of you are in monogamous relationships? A majority. How is that enforced?...
  • If everyone you talk to is boring it’s not them! And so if you're rejected by the opposite sex, if you’re heterosexual, then you're wrong, they're not wrong, and you've got some work to do, man. You've got some difficult work to do. And there isn't anything I've been telling young men that's clearer than that … What I've been telling people is take the responsibility for failure onto yourself. That's a hint that you've got work to do. It could also be a hint that you're young and useless and why the hell would anybody have anything to do with you because you don't have anything to offer. And that's rectifiable. Maturity helps to rectify that.
  • And what's the gender? Men. Because if you go two standard deviations out from the mean on two curves that overlap but are disjointed, then you derive an overwhelming preponderance of the overrepresented group. That's why men are about 10 times more likely to be in prison.  
  • Weiss: You are often characterized, at least in the mainstream press, as being transphobic. If you had a student come to you and say, I was born female, I now identify as male, I want you to call me by male pronouns. Would you say yes to that?
  • Peterson: Well, it would depend on the student and the context and why I thought they were asking me and what I believe their demand actually characterized, and all of that. Because that can be done in a way that is genuine and acceptable, and a way that is manipulative and unacceptable. And if it was genuine and acceptable then I would have no problem with it. And if it was manipulative and unacceptable then not a chance. And you might think, ‘Well, who am I to judge?’ Well, first of all, I am a clinical psychologist, I've talked to people for about 25,000 hours. And I'm responsible for judging how I am going to use my words. I'd judge the same way I judge all my interactions with people, which is to the best of my ability, and characterized by all the errors that I'm prone to. I'm not saying that my judgment would be unerring. I live with the consequences and I'm willing to accept the responsibility.
  • But also to be clear about this, it never happened––I never refused to call anyone by anything they had asked me to call them by, although that's been reported multiple times. It's a complete falsehood. And it had nothing to do with the transgender issue as far as I'm concerned.
  • type one and type two error problem
  • note what his avowed position is: that he has never refused to call a transgender person by their preferred pronoun, that he has done so many times, that he would always try to err on the side of believing a request to be earnest, and that he reserves the right to decline a request he believes to be in bad faith. Whether one finds that to be reasonable or needlessly difficult, it seems irresponsible to tell trans people that a prominent intellectual hates them or is deeply antagonistic to them when the only seeming conflict is utterly hypothetical and ostensibly not even directed against people that Peterson believes to be trans, but only against people whom he does not believe to be trans
8More

No word of a lie: scientists rate the world's biggest peddlers of bull | Science | The ... - 0 views

  • scientists claim to have identified the most common practitioners of the ignoble art. Their study of 40,000 teenagers reveals that boys; those from privileged backgrounds; and North Americans in particular, top the charts as the worst offenders.
  • Jerrim analysed data gathered by the OECD to assess how well 15-year-olds around the globe have mastered key academic subjects.
  • For one of the questions, the teenagers must rate how familiar they are with 16 mathematical concepts ranging from polygons and vectors to quadratic functions and congruent figures. Hidden among the bona fide terms are three fakes: proper numbers, subjective scaling and declarative functions
  • ...5 more annotations...
  • Jerrim took the students’ responses to the three fake concepts to draw up a “bullshit scale”, which he then used to compare different groups, such as boys and girls, high and low socioeconomic status, and the regions where people lived.
  • “Boys are bigger bullshitters than girls, children from higher socioeconomic backgrounds tend to bullshit more than those from lower ones, and North Americans bullshit the most,” Jerrim said. Those who ranked highest on the scale tended to see themselves as more self-confident, more persevering, and more popular at school, than those further down the scale.
  • academics have devised a Bullshit Receptivity scale, which showed that believers in the supernatural may be more receptive to bullshit, and proposed an “ease of passing bullshit hypothesis”, which posits that people are more likely to commit the offence when they believe they can get away with it.
  • “People are social animals and we desire feelings of connection, belonging, and inclusion, so we try to participate when it is critical to build and maintain these relationships,” he said. “Such situations sometimes require us to talk about things we really know nothing about, and what comes out is bullshit.”
  • Jerrim said a major question is whether, and when, the art is beneficial. “Everyone gets a question in a job interview that they cannot answer. If you’re an effective bullshitter, it might help you get your foot in the door,” he said. “It might also help with academic grant proposals.”
46More

'Our minds can be hijacked': the tech insiders who fear a smartphone dystopia | Technol... - 0 views

  • Rosenstein belongs to a small but growing band of Silicon Valley heretics who complain about the rise of the so-called “attention economy”: an internet shaped around the demands of an advertising economy.
  • “It is very common,” Rosenstein says, “for humans to develop things with the best of intentions and for them to have unintended, negative consequences.”
  • most concerned about the psychological effects on people who, research shows, touch, swipe or tap their phone 2,617 times a day.
  • ...43 more annotations...
  • There is growing concern that as well as addicting users, technology is contributing toward so-called “continuous partial attention”, severely limiting people’s ability to focus, and possibly lowering IQ. One recent study showed that the mere presence of smartphones damages cognitive capacity – even when the device is turned off. “Everyone is distracted,” Rosenstein says. “All of the time.”
  • Drawing a straight line between addiction to social media and political earthquakes like Brexit and the rise of Donald Trump, they contend that digital forces have completely upended the political system and, left unchecked, could even render democracy as we know it obsolete.
  • Without irony, Eyal finished his talk with some personal tips for resisting the lure of technology. He told his audience he uses a Chrome extension, called DF YouTube, “which scrubs out a lot of those external triggers” he writes about in his book, and recommended an app called Pocket Points that “rewards you for staying off your phone when you need to focus”.
  • “One reason I think it is particularly important for us to talk about this now is that we may be the last generation that can remember life before,” Rosenstein says. It may or may not be relevant that Rosenstein, Pearlman and most of the tech insiders questioning today’s attention economy are in their 30s, members of the last generation that can remember a world in which telephones were plugged into walls.
  • One morning in April this year, designers, programmers and tech entrepreneurs from across the world gathered at a conference centre on the shore of the San Francisco Bay. They had each paid up to $1,700 to learn how to manipulate people into habitual use of their products, on a course curated by conference organiser Nir Eyal.
  • Eyal, 39, the author of Hooked: How to Build Habit-Forming Products, has spent several years consulting for the tech industry, teaching techniques he developed by closely studying how the Silicon Valley giants operate.
  • “The technologies we use have turned into compulsions, if not full-fledged addictions,” Eyal writes. “It’s the impulse to check a message notification. It’s the pull to visit YouTube, Facebook, or Twitter for just a few minutes, only to find yourself still tapping and scrolling an hour later.” None of this is an accident, he writes. It is all “just as their designers intended”
  • He explains the subtle psychological tricks that can be used to make people develop habits, such as varying the rewards people receive to create “a craving”, or exploiting negative emotions that can act as “triggers”. “Feelings of boredom, loneliness, frustration, confusion and indecisiveness often instigate a slight pain or irritation and prompt an almost instantaneous and often mindless action to quell the negative sensation,” Eyal writes.
  • The most seductive design, Harris explains, exploits the same psychological susceptibility that makes gambling so compulsive: variable rewards. When we tap those apps with red icons, we don’t know whether we’ll discover an interesting email, an avalanche of “likes”, or nothing at all. It is the possibility of disappointment that makes it so compulsive.
  • Finally, Eyal confided the lengths he goes to protect his own family. He has installed in his house an outlet timer connected to a router that cuts off access to the internet at a set time every day. “The idea is to remember that we are not powerless,” he said. “We are in control.
  • But are we? If the people who built these technologies are taking such radical steps to wean themselves free, can the rest of us reasonably be expected to exercise our free will?
  • Not according to Tristan Harris, a 33-year-old former Google employee turned vocal critic of the tech industry. “All of us are jacked into this system,” he says. “All of our minds can be hijacked. Our choices are not as free as we think they are.”
  • Harris, who has been branded “the closest thing Silicon Valley has to a conscience”, insists that billions of people have little choice over whether they use these now ubiquitous technologies, and are largely unaware of the invisible ways in which a small number of people in Silicon Valley are shaping their lives.
  • “I don’t know a more urgent problem than this,” Harris says. “It’s changing our democracy, and it’s changing our ability to have the conversations and relationships that we want with each other.” Harris went public – giving talks, writing papers, meeting lawmakers and campaigning for reform after three years struggling to effect change inside Google’s Mountain View headquarters.
  • He explored how LinkedIn exploits a need for social reciprocity to widen its network; how YouTube and Netflix autoplay videos and next episodes, depriving users of a choice about whether or not they want to keep watching; how Snapchat created its addictive Snapstreaks feature, encouraging near-constant communication between its mostly teenage users.
  • The techniques these companies use are not always generic: they can be algorithmically tailored to each person. An internal Facebook report leaked this year, for example, revealed that the company can identify when teens feel “insecure”, “worthless” and “need a confidence boost”. Such granular information, Harris adds, is “a perfect model of what buttons you can push in a particular person”.
  • Tech companies can exploit such vulnerabilities to keep people hooked; manipulating, for example, when people receive “likes” for their posts, ensuring they arrive when an individual is likely to feel vulnerable, or in need of approval, or maybe just bored. And the very same techniques can be sold to the highest bidder. “There’s no ethics,” he says. A company paying Facebook to use its levers of persuasion could be a car business targeting tailored advertisements to different types of users who want a new vehicle. Or it could be a Moscow-based troll farm seeking to turn voters in a swing county in Wisconsin.
  • It was Rosenstein’s colleague, Leah Pearlman, then a product manager at Facebook and on the team that created the Facebook “like”, who announced the feature in a 2009 blogpost. Now 35 and an illustrator, Pearlman confirmed via email that she, too, has grown disaffected with Facebook “likes” and other addictive feedback loops. She has installed a web browser plug-in to eradicate her Facebook news feed, and hired a social media manager to monitor her Facebook page so that she doesn’t have to.
  • Harris believes that tech companies never deliberately set out to make their products addictive. They were responding to the incentives of an advertising economy, experimenting with techniques that might capture people’s attention, even stumbling across highly effective design by accident.
  • It’s this that explains how the pull-to-refresh mechanism, whereby users swipe down, pause and wait to see what content appears, rapidly became one of the most addictive and ubiquitous design features in modern technology. “Each time you’re swiping down, it’s like a slot machine,” Harris says. “You don’t know what’s coming next. Sometimes it’s a beautiful photo. Sometimes it’s just an ad.”
  • The reality TV star’s campaign, he said, had heralded a watershed in which “the new, digitally supercharged dynamics of the attention economy have finally crossed a threshold and become manifest in the political realm”.
  • “Smartphones are useful tools,” he says. “But they’re addictive. Pull-to-refresh is addictive. Twitter is addictive. These are not good things. When I was working on them, it was not something I was mature enough to think about. I’m not saying I’m mature now, but I’m a little bit more mature, and I regret the downsides.”
  • All of it, he says, is reward-based behaviour that activates the brain’s dopamine pathways. He sometimes finds himself clicking on the red icons beside his apps “to make them go away”, but is conflicted about the ethics of exploiting people’s psychological vulnerabilities. “It is not inherently evil to bring people back to your product,” he says. “It’s capitalism.”
  • He identifies the advent of the smartphone as a turning point, raising the stakes in an arms race for people’s attention. “Facebook and Google assert with merit that they are giving users what they want,” McNamee says. “The same can be said about tobacco companies and drug dealers.”
  • McNamee chooses his words carefully. “The people who run Facebook and Google are good people, whose well-intentioned strategies have led to horrific unintended consequences,” he says. “The problem is that there is nothing the companies can do to address the harm unless they abandon their current advertising models.”
  • But how can Google and Facebook be forced to abandon the business models that have transformed them into two of the most profitable companies on the planet?
  • McNamee believes the companies he invested in should be subjected to greater regulation, including new anti-monopoly rules. In Washington, there is growing appetite, on both sides of the political divide, to rein in Silicon Valley. But McNamee worries the behemoths he helped build may already be too big to curtail.
  • Rosenstein, the Facebook “like” co-creator, believes there may be a case for state regulation of “psychologically manipulative advertising”, saying the moral impetus is comparable to taking action against fossil fuel or tobacco companies. “If we only care about profit maximisation,” he says, “we will go rapidly into dystopia.”
  • James Williams does not believe talk of dystopia is far-fetched. The ex-Google strategist who built the metrics system for the company’s global search advertising business, he has had a front-row view of an industry he describes as the “largest, most standardised and most centralised form of attentional control in human history”.
  • It is a journey that has led him to question whether democracy can survive the new technological age.
  • He says his epiphany came a few years ago, when he noticed he was surrounded by technology that was inhibiting him from concentrating on the things he wanted to focus on. “It was that kind of individual, existential realisation: what’s going on?” he says. “Isn’t technology supposed to be doing the complete opposite of this?
  • That discomfort was compounded during a moment at work, when he glanced at one of Google’s dashboards, a multicoloured display showing how much of people’s attention the company had commandeered for advertisers. “I realised: this is literally a million people that we’ve sort of nudged or persuaded to do this thing that they weren’t going to otherwise do,” he recalls.
  • Williams and Harris left Google around the same time, and co-founded an advocacy group, Time Well Spent, that seeks to build public momentum for a change in the way big tech companies think about design. Williams finds it hard to comprehend why this issue is not “on the front page of every newspaper every day.
  • “Eighty-seven percent of people wake up and go to sleep with their smartphones,” he says. The entire world now has a new prism through which to understand politics, and Williams worries the consequences are profound.
  • g. “The attention economy incentivises the design of technologies that grab our attention,” he says. “In so doing, it privileges our impulses over our intentions.”
  • That means privileging what is sensational over what is nuanced, appealing to emotion, anger and outrage. The news media is increasingly working in service to tech companies, Williams adds, and must play by the rules of the attention economy to “sensationalise, bait and entertain in order to survive”.
  • It is not just shady or bad actors who were exploiting the internet to change public opinion. The attention economy itself is set up to promote a phenomenon like Trump, who is masterly at grabbing and retaining the attention of supporters and critics alike, often by exploiting or creating outrage.
  • All of which has left Brichter, who has put his design work on the backburner while he focuses on building a house in New Jersey, questioning his legacy. “I’ve spent many hours and weeks and months and years thinking about whether anything I’ve done has made a net positive impact on society or humanity at all,” he says. He has blocked certain websites, turned off push notifications, restricted his use of the Telegram app to message only with his wife and two close friends, and tried to wean himself off Twitter. “I still waste time on it,” he confesses, “just reading stupid news I already know about.” He charges his phone in the kitchen, plugging it in at 7pm and not touching it until the next morning.
  • He stresses these dynamics are by no means isolated to the political right: they also play a role, he believes, in the unexpected popularity of leftwing politicians such as Bernie Sanders and Jeremy Corbyn, and the frequent outbreaks of internet outrage over issues that ignite fury among progressives.
  • All of which, Williams says, is not only distorting the way we view politics but, over time, may be changing the way we think, making us less rational and more impulsive. “We’ve habituated ourselves into a perpetual cognitive style of outrage, by internalising the dynamics of the medium,” he says.
  • It was another English science fiction writer, Aldous Huxley, who provided the more prescient observation when he warned that Orwellian-style coercion was less of a threat to democracy than the more subtle power of psychological manipulation, and “man’s almost infinite appetite for distractions”.
  • If the attention economy erodes our ability to remember, to reason, to make decisions for ourselves – faculties that are essential to self-governance – what hope is there for democracy itself?
  • “The dynamics of the attention economy are structurally set up to undermine the human will,” he says. “If politics is an expression of our human will, on individual and collective levels, then the attention economy is directly undermining the assumptions that democracy rests on.”
42More

ROUGH TYPE | Nicholas Carr's blog - 0 views

  • The smartphone has become a repository of the self, recording and dispensing the words, sounds and images that define what we think, what we experience and who we are. In a 2015 Gallup survey, more than half of iPhone owners said that they couldn’t imagine life without the device.
  • So what happens to our minds when we allow a single tool such dominion over our perception and cognition?
  • Not only do our phones shape our thoughts in deep and complicated ways, but the effects persist even when we aren’t using the devices. As the brain grows dependent on the technology, the research suggests, the intellect weakens.
  • ...39 more annotations...
  • he has seen mounting evidence that using a smartphone, or even hearing one ring or vibrate, produces a welter of distractions that makes it harder to concentrate on a difficult problem or job. The division of attention impedes reasoning and performance.
  • Another 2015 study, appearing in the Journal of Computer-Mediated Communication, showed that when people hear their phone ring but are unable to answer it, their blood pressure spikes, their pulse quickens, and their problem-solving skills decline.
  • The researchers recruited 520 undergraduates at UCSD and gave them two standard tests of intellectual acuity. One test gauged “available working-memory capacity,” a measure of how fully a person’s mind can focus on a particular task. The second assessed “fluid intelligence,” a person’s ability to interpret and solve an unfamiliar problem. The only variable in the experiment was the location of the subjects’ smartphones. Some of the students were asked to place their phones in front of them on their desks; others were told to stow their phones in their pockets or handbags; still others were required to leave their phones in a different room.
  • In both tests, the subjects whose phones were in view posted the worst scores, while those who left their phones in a different room did the best. The students who kept their phones in their pockets or bags came out in the middle. As the phone’s proximity increased, brainpower decreased.
  • In subsequent interviews, nearly all the participants said that their phones hadn’t been a distraction—that they hadn’t even thought about the devices during the experiment. They remained oblivious even as the phones disrupted their focus and thinking.
  • In a 2013 study conducted at the University of Essex in England, 142 participants were divided into pairs and asked to converse in private for ten minutes. Half talked with a phone in the room, half without a phone present. The subjects were then given tests of affinity, trust and empathy. “The mere presence of mobile phones,” the researchers reported in the Journal of Social and Personal Relationships, “inhibited the development of interpersonal closeness and trust” and diminished “the extent to which individuals felt empathy and understanding from their partners.”
  • the “integration of smartphones into daily life” appears to cause a “brain drain” that can diminish such vital mental skills as “learning, logical reasoning, abstract thought, problem solving, and creativity.”
  •  Smartphones have become so entangled with our existence that, even when we’re not peering or pawing at them, they tug at our attention, diverting precious cognitive resources. Just suppressing the desire to check our phone, which we do routinely and subconsciously throughout the day, can debilitate our thinking.
  • They found that students who didn’t bring their phones to the classroom scored a full letter-grade higher on a test of the material presented than those who brought their phones. It didn’t matter whether the students who had their phones used them or not: All of them scored equally poorly.
  • A study of nearly a hundred secondary schools in the U.K., published last year in the journal Labour Economics, found that when schools ban smartphones, students’ examination scores go up substantially, with the weakest students benefiting the most.
  • Social skills and relationships seem to suffer as well.
  • Because smartphones serve as constant reminders of all the friends we could be chatting with electronically, they pull at our minds when we’re talking with people in person, leaving our conversations shallower and less satisfying.
  • A second experiment conducted by the researchers produced similar results, while also revealing that the more heavily students relied on their phones in their everyday lives, the greater the cognitive penalty they suffered.
  • The evidence that our phones can get inside our heads so forcefully is unsettling. It suggests that our thoughts and feelings, far from being sequestered in our skulls, can be skewed by external forces we’re not even aware o
  •  Scientists have long known that the brain is a monitoring system as well as a thinking system. Its attention is drawn toward any object that is new, intriguing or otherwise striking — that has, in the psychological jargon, “salience.”
  • even in the history of captivating media, the smartphone stands out. It is an attention magnet unlike any our minds have had to grapple with before. Because the phone is packed with so many forms of information and so many useful and entertaining functions, it acts as what Dr. Ward calls a “supernormal stimulus,” one that can “hijack” attention whenever it is part of our surroundings — and it is always part of our surroundings.
  • Imagine combining a mailbox, a newspaper, a TV, a radio, a photo album, a public library and a boisterous party attended by everyone you know, and then compressing them all into a single, small, radiant object. That is what a smartphone represents to us. No wonder we can’t take our minds off it.
  • The irony of the smartphone is that the qualities that make it so appealing to us — its constant connection to the net, its multiplicity of apps, its responsiveness, its portability — are the very ones that give it such sway over our minds.
  • Phone makers like Apple and Samsung and app writers like Facebook, Google and Snap design their products to consume as much of our attention as possible during every one of our waking hours
  • Social media apps were designed to exploit “a vulnerability in human psychology,” former Facebook president Sean Parker said in a recent interview. “[We] understood this consciously. And we did it anyway.”
  • A quarter-century ago, when we first started going online, we took it on faith that the web would make us smarter: More information would breed sharper thinking. We now know it’s not that simple.
  • As strange as it might seem, people’s knowledge and understanding may actually dwindle as gadgets grant them easier access to online data stores
  • In a seminal 2011 study published in Science, a team of researchers — led by the Columbia University psychologist Betsy Sparrow and including the late Harvard memory expert Daniel Wegner — had a group of volunteers read forty brief, factual statements (such as “The space shuttle Columbia disintegrated during re-entry over Texas in Feb. 2003”) and then type the statements into a computer. Half the people were told that the machine would save what they typed; half were told that the statements would be erased.
  • Afterward, the researchers asked the subjects to write down as many of the statements as they could remember. Those who believed that the facts had been recorded in the computer demonstrated much weaker recall than those who assumed the facts wouldn’t be stored. Anticipating that information would be readily available in digital form seemed to reduce the mental effort that people made to remember it
  • The researchers dubbed this phenomenon the “Google effect” and noted its broad implications: “Because search engines are continually available to us, we may often be in a state of not feeling we need to encode the information internally. When we need it, we will look it up.”
  • as the pioneering psychologist and philosopher William James said in an 1892 lecture, “the art of remembering is the art of thinking.”
  • Only by encoding information in our biological memory can we weave the rich intellectual associations that form the essence of personal knowledge and give rise to critical and conceptual thinking. No matter how much information swirls around us, the less well-stocked our memory, the less we have to think with.
  • As Dr. Wegner and Dr. Ward explained in a 2013 Scientific American article, when people call up information through their devices, they often end up suffering from delusions of intelligence. They feel as though “their own mental capacities” had generated the information, not their devices. “The advent of the ‘information age’ seems to have created a generation of people who feel they know more than ever before,” the scholars concluded, even though “they may know ever less about the world around them.”
  • That insight sheds light on society’s current gullibility crisis, in which people are all too quick to credit lies and half-truths spread through social media. If your phone has sapped your powers of discernment, you’ll believe anything it tells you.
  • Data, the novelist and critic Cynthia Ozick once wrote, is “memory without history.” Her observation points to the problem with allowing smartphones to commandeer our brains
  • When we constrict our capacity for reasoning and recall or transfer those skills to a gadget, we sacrifice our ability to turn information into knowledge. We get the data but lose the meaning
  • We need to give our minds more room to think. And that means putting some distance between ourselves and our phones.
  • Harvard Business School professor emerita Shoshana Zuboff argues in her new book that the Valley’s wealth and power are predicated on an insidious, essentially pathological form of private enterprise—what she calls “surveillance capitalism.” Pioneered by Google, perfected by Facebook, and now spreading throughout the economy, surveillance capitalism uses human life as its raw material. Our everyday experiences, distilled into data, have become a privately-owned business asset used to predict and mold our behavior, whether we’re shopping or socializing, working or voting.
  • By reengineering the economy and society to their own benefit, Google and Facebook are perverting capitalism in a way that undermines personal freedom and corrodes democracy.
  • Under the Fordist model of mass production and consumption that prevailed for much of the twentieth century, industrial capitalism achieved a relatively benign balance among the contending interests of business owners, workers, and consumers. Enlightened executives understood that good pay and decent working conditions would ensure a prosperous middle class eager to buy the goods and services their companies produced. It was the product itself — made by workers, sold by companies, bought by consumers — that tied the interests of capitalism’s participants together. Economic and social equilibrium was negotiated through the product.
  • By removing the tangible product from the center of commerce, surveillance capitalism upsets the equilibrium. Whenever we use free apps and online services, it’s often said, we become the products, our attention harvested and sold to advertisers
  • this truism gets it wrong. Surveillance capitalism’s real products, vaporous but immensely valuable, are predictions about our future behavior — what we’ll look at, where we’ll go, what we’ll buy, what opinions we’ll hold — that internet companies derive from our personal data and sell to businesses, political operatives, and other bidders.
  • Unlike financial derivatives, which they in some ways resemble, these new data derivatives draw their value, parasite-like, from human experience.To the Googles and Facebooks of the world, we are neither the customer nor the product. We are the source of what Silicon Valley technologists call “data exhaust” — the informational byproducts of online activity that become the inputs to prediction algorithms
  • internet companies operate in what Zuboff terms “extreme structural independence from people.” When databases displace goods as the engine of the economy, our own interests, as consumers but also as citizens, cease to be part of the negotiation. We are no longer one of the forces guiding the market’s invisible hand. We are the objects of surveillance and control.
23More

How Calls for Privacy May Upend Business for Facebook and Google - The New York Times - 0 views

  • People detailed their interests and obsessions on Facebook and Google, generating a river of data that could be collected and harnessed for advertising. The companies became very rich. Users seemed happy. Privacy was deemed obsolete, like bloodletting and milkmen
  • It has been many months of allegations and arguments that the internet in general and social media in particular are pulling society down instead of lifting it up.
  • That has inspired a good deal of debate about more restrictive futures for Facebook and Google. At the furthest extreme, some dream of the companies becoming public utilities.
  • ...20 more annotations...
  • There are other avenues still, said Jascha Kaykas-Wolff, the chief marketing officer of Mozilla, the nonprofit organization behind the popular Firefox browser, including advertisers and large tech platforms collecting vastly less user data and still effectively customizing ads to consumers.
  • The greatest likelihood is that the internet companies, frightened by the tumult, will accept a few more rules and work a little harder for transparency.
  • The Cambridge Analytica case, said Vera Jourova, the European Union commissioner for justice, consumers and gender equality, was not just a breach of private data. “This is much more serious, because here we witness the threat to democracy, to democratic plurality,” she said.
  • Although many people had a general understanding that free online services used their personal details to customize the ads they saw, the latest controversy starkly exposed the machinery.
  • Consumers’ seemingly benign activities — their likes — could be used to covertly categorize and influence their behavior. And not just by unknown third parties. Facebook itself has worked directly with presidential campaigns on ad targeting, describing its services in a company case study as “influencing voters.”
  • “If your personal information can help sway elections, which affects everyone’s life and societal well-being, maybe privacy does matter after all.”
  • some trade group executives also warned that any attempt to curb the use of consumer data would put the business model of the ad-supported internet at risk.
  • “You’re undermining a fundamental concept in advertising: reaching consumers who are interested in a particular product,”
  • If suspicion of Facebook and Google is a relatively new feeling in the United States, it has been embedded in Europe for historical and cultural reasons that date back to the Nazi Gestapo, the Soviet occupation of Eastern Europe and the Cold War.
  • “We’re at an inflection point, when the great wave of optimism about tech is giving way to growing alarm,” said Heather Grabbe, director of the Open Society European Policy Institute. “This is the moment when Europeans turn to the state for protection and answers, and are less likely than Americans to rely on the market to sort out imbalances.”
  • In May, the European Union is instituting a comprehensive new privacy law, called the General Data Protection Regulation. The new rules treat personal data as proprietary, owned by an individual, and any use of that data must be accompanied by permission — opting in rather than opting out — after receiving a request written in clear language, not legalese.
  • the protection rules will have more teeth than the current 1995 directive. For example, a company experiencing a data breach involving individuals must notify the data protection authority within 72 hours and would be subject to fines of up to 20 million euros or 4 percent of its annual revenue.
  • “With the new European law, regulators for the first time have real enforcement tools,” said Jeffrey Chester, the executive director of the Center for Digital Democracy, a nonprofit group in Washington. “We now have a way to hold these companies accountable.”
  • Privacy advocates and even some United States regulators have long been concerned about the ability of online services to track consumers and make inferences about their financial status, health concerns and other intimate details to show them behavior-based ads. They warned that such microtargeting could unfairly categorize or exclude certain people.
  • the Do Not Track effort and the privacy bill were both stymied.Industry groups successfully argued that collecting personal details posed no harm to consumers and that efforts to hinder data collection would chill innovation.
  • “If it can be shown that the current situation is actually a market failure and not an individual-company failure, then there’s a case to be made for federal regulation” under certain circumstances
  • The business practices of Facebook and Google were reinforced by the fact that no privacy flap lasted longer than a news cycle or two. Nor did people flee for other services. That convinced the companies that digital privacy was a dead issue.
  • If the current furor dies down without meaningful change, critics worry that the problems might become even more entrenched. When the tech industry follows its natural impulses, it becomes even less transparent.
  • “To know the real interaction between populism and Facebook, you need to give much more access to researchers, not less,” said Paul-Jasper Dittrich, a German research fellow
  • There’s another reason Silicon Valley tends to be reluctant to share information about what it is doing. It believes so deeply in itself that it does not even think there is a need for discussion. The technology world’s remedy for any problem is always more technology
5More

Opinion | Scott Pruitt's Attack on Science Would Paralyze the E.P.A. - The New York Times - 0 views

  • It is his latest effort to cripple the agency. Mr. Pruitt, who as Oklahoma’s attorney general described himself as “a leading advocate against the E.P.A.’s activist agenda,” said in an interview published in The Daily Caller last week that he would no longer allow the agency to use studies that include nonpublic scientific data to develop rules to safeguard public health and prevent pollution.
  • Opponents of the agency and of mainstream climate science call these studies “secret science.” But that’s simply not true. Peer review ensures that the analytic methodologies underlying studies funded by the agency are sound.
  • Some of those studies, particularly those that determine the effects of exposure to chemicals and pollution on health, rely on medical records that by law are confidential because of patient privacy policies. These studies summarize the analysis of raw data and draw conclusions based on that analysis. Other government agencies also use studies like these to develop policy and regulations, and to buttress and defend rules against legal challenges. They are, in fact, essential to making sound public policy.
  • ...2 more annotations...
  • The agency also relies on industry data to develop rules on chemical safety that is often kept confidential for business reasons.
  • So why would he want to prohibit his own agency from using these studies? It’s not a mystery. Time and again the Trump administration has put the profits of regulated industries over the health of the American people. Fundamental research on the effects of air pollution on public health has long been a target of those who oppose the E.P.A.’s air quality regulations, like the rule that requires power plants to reduce their mercury emissions.
32More

Opinion | How Genetics Is Changing Our Understanding of 'Race' - The New York Times - 0 views

  • In 1942, the anthropologist Ashley Montagu published “Man’s Most Dangerous Myth: The Fallacy of Race,” an influential book that argued that race is a social concept with no genetic basis.
  • eginning in 1972, genetic findings began to be incorporated into this argument. That year, the geneticist Richard Lewontin published an important study of variation in protein types in blood. He grouped the human populations he analyzed into seven “races” — West Eurasians, Africans, East Asians, South Asians, Native Americans, Oceanians and Australians — and found that around 85 percent of variation in the protein types could be accounted for by variation within populations and “races,” and only 15 percent by variation across them. To the extent that there was variation among humans, he concluded, most of it was because of “differences between individuals.”
  • In this way, a consensus was established that among human populations there are no differences large enough to support the concept of “biological race.” Instead, it was argued, race is a “social construct,” a way of categorizing people that changes over time and across countries.
  • ...29 more annotations...
  • t is true that race is a social construct. It is also true, as Dr. Lewontin wrote, that human populations “are remarkably similar to each other” from a genetic point of view.
  • this consensus has morphed, seemingly without questioning, into an orthodoxy. The orthodoxy maintains that the average genetic differences among people grouped according to today’s racial terms are so trivial when it comes to any meaningful biological traits that those differences can be ignored.
  • With the help of these tools, we are learning that while race may be a social construct, differences in genetic ancestry that happen to correlate to many of today’s racial constructs are real.
  • I have deep sympathy for the concern that genetic discoveries could be misused to justify racism. But as a geneticist I also know that it is simply no longer possible to ignore average genetic differences among “races.”
  • Groundbreaking advances in DNA sequencing technology have been made over the last two decades
  • Care.
  • The orthodoxy goes further, holding that we should be anxious about any research into genetic differences among populations
  • You will sometimes hear that any biological differences among populations are likely to be small, because humans have diverged too recently from common ancestors for substantial differences to have arisen under the pressure of natural selection. This is not true. The ancestors of East Asians, Europeans, West Africans and Australians were, until recently, almost completely isolated from one another for 40,000 years or longer, which is more than sufficient time for the forces of evolution to work
  • I am worried that well-meaning people who deny the possibility of substantial biological differences among human populations are digging themselves into an indefensible position, one that will not survive the onslaught of science.
  • I am also worried that whatever discoveries are made — and we truly have no idea yet what they will be — will be cited as “scientific proof” that racist prejudices and agendas have been correct all along, and that those well-meaning people will not understand the science well enough to push back against these claims.
  • This is why it is important, even urgent, that we develop a candid and scientifically up-to-date way of discussing any such difference
  • While most people will agree that finding a genetic explanation for an elevated rate of disease is important, they often draw the line there. Finding genetic influences on a propensity for disease is one thing, they argue, but looking for such influences on behavior and cognition is another
  • Is performance on an intelligence test or the number of years of school a person attends shaped by the way a person is brought up? Of course. But does it measure something having to do with some aspect of behavior or cognition? Almost certainly.
  • Recent genetic studies have demonstrated differences across populations not just in the genetic determinants of simple traits such as skin color, but also in more complex traits like bodily dimensions and susceptibility to diseases.
  • in Iceland, there has been measurable genetic selection against the genetic variations that predict more years of education in that population just within the last century.
  • consider what kinds of voices are filling the void that our silence is creating
  • Nicholas Wade, a longtime science journalist for The New York Times, rightly notes in his 2014 book, “A Troublesome Inheritance: Genes, Race and Human History,” that modern research is challenging our thinking about the nature of human population differences. But he goes on to make the unfounded and irresponsible claim that this research is suggesting that genetic factors explain traditional stereotypes.
  • 139 geneticists (including myself) pointed out in a letter to The New York Times about Mr. Wade’s book, there is no genetic evidence to back up any of the racist stereotypes he promotes.
  • Another high-profile example is James Watson, the scientist who in 1953 co-discovered the structure of DNA, and who was forced to retire as head of the Cold Spring Harbor Laboratories in 2007 after he stated in an interview — without any scientific evidence — that research has suggested that genetic factors contribute to lower intelligence in Africans than in Europeans.
  • What makes Dr. Watson’s and Mr. Wade’s statements so insidious is that they start with the accurate observation that many academics are implausibly denying the possibility of average genetic differences among human populations, and then end with a claim — backed by no evidence — that they know what those differences are and that they correspond to racist stereotypes
  • They use the reluctance of the academic community to openly discuss these fraught issues to provide rhetorical cover for hateful ideas and old racist canards.
  • This is why knowledgeable scientists must speak out. If we abstain from laying out a rational framework for discussing differences among populations, we risk losing the trust of the public and we actively contribute to the distrust of expertise that is now so prevalent.
  • If scientists can be confident of anything, it is that whatever we currently believe about the genetic nature of differences among populations is most likely wrong.
  • For example, my laboratory discovered in 2016, based on our sequencing of ancient human genomes, that “whites” are not derived from a population that existed from time immemorial, as some people believe. Instead, “whites” represent a mixture of four ancient populations that lived 10,000 years ago and were each as different from one another as Europeans and East Asians are today.
  • For me, a natural response to the challenge is to learn from the example of the biological differences that exist between males and females
  • The differences between the sexes are far more profound than those that exist among human populations, reflecting more than 100 million years of evolution and adaptation. Males and females differ by huge tracts of genetic material
  • How do we accommodate the biological differences between men and women? I think the answer is obvious: We should both recognize that genetic differences between males and females exist and we should accord each sex the same freedoms and opportunities regardless of those differences
  • fulfilling these aspirations in practice is a challenge. Yet conceptually it is straightforward.
  • Compared with the enormous differences that exist among individuals, differences among populations are on average many times smaller, so it should be only a modest challenge to accommodate a reality in which the average genetic contributions to human traits differ.
100More

At the Existentialist Café: Freedom, Being, and Apricot Cocktails with Jean-P... - 0 views

  • The phenomenologists’ leading thinker, Edmund Husserl, provided a rallying cry, ‘To the things themselves!’ It meant: don’t waste time on the interpretations that accrue upon things, and especially don’t waste time wondering whether the things are real. Just look at this that’s presenting itself to you, whatever this may be, and describe it as precisely as possible.
  • You might think you have defined me by some label, but you are wrong, for I am always a work in progress. I create myself constantly through action, and this is so fundamental to my human condition that, for Sartre, it is the human condition, from the moment of first consciousness to the moment when death wipes it out. I am my own freedom: no more, no less.
  • Sartre wrote like a novelist — not surprisingly, since he was one. In his novels, short stories and plays as well as in his philosophical treatises, he wrote about the physical sensations of the world and the structures and moods of human life. Above all, he wrote about one big subject: what it meant to be free. Freedom, for him, lay at the heart of all human experience, and this set humans apart from all other kinds of object.
  • ...97 more annotations...
  • Sartre listened to his problem and said simply, ‘You are free, therefore choose — that is to say, invent.’ No signs are vouchsafed in this world, he said. None of the old authorities can relieve you of the burden of freedom. You can weigh up moral or practical considerations as carefully as you like, but ultimately you must take the plunge and do something, and it’s up to you what that something is.
  • Even if the situation is unbearable — perhaps you are facing execution, or sitting in a Gestapo prison, or about to fall off a cliff — you are still free to decide what to make of it in mind and deed. Starting from where you are now, you choose. And in choosing, you also choose who you will be.
  • The war had made people realise that they and their fellow humans were capable of departing entirely from civilised norms; no wonder the idea of a fixed human nature seemed questionable.
  • If this sounds difficult and unnerving, it’s because it is. Sartre does not deny that the need to keep making decisions brings constant anxiety. He heightens this anxiety by pointing out that what you do really matters. You should make your choices as though you were choosing on behalf of the whole of humanity, taking the entire burden of responsibility for how the human race behaves. If you avoid this responsibility by fooling yourself that you are the victim of circumstance or of someone else’s bad advice, you are failing to meet the demands of human life and choosing a fake existence, cut off from your own ‘authenticity’.
  • Along with the terrifying side of this comes a great promise: Sartre’s existentialism implies that it is possible to be authentic and free, as long as you keep up the effort.
  • almost all agreed that it was, as an article in Les nouvelles littéraires phrased it, a ‘sickening mixture of philosophic pretentiousness, equivocal dreams, physiological technicalities, morbid tastes and hesitant eroticism … an introspective embryo that one would take distinct pleasure in crushing’.
  • he offered a philosophy designed for a species that had just scared the hell out of itself, but that finally felt ready to grow up and take responsibility.
  • In this rebellious world, just as with the Parisian bohemians and Dadaists in earlier generations, everything that was dangerous and provocative was good, and everything that was nice or bourgeois was bad.
  • Such interweaving of ideas and life had a long pedigree, although the existentialists gave it a new twist. Stoic and Epicurean thinkers in the classical world had practised philosophy as a means of living well, rather than of seeking knowledge or wisdom for their own sake. By reflecting on life’s vagaries in philosophical ways, they believed they could become more resilient, more able to rise above circumstances, and better equipped to manage grief, fear, anger, disappointment or anxiety.
  • In the tradition they passed on, philosophy is neither a pure intellectual pursuit nor a collection of cheap self-help tricks, but a discipline for flourishing and living a fully human, responsible life.
  • For Kierkegaard, Descartes had things back to front. In his own view, human existence comes first: it is the starting point for everything we do, not the result of a logical deduction. My existence is active: I live it and choose it, and this precedes any statement I can make about myself.
  • Studying our own moral genealogy cannot help us to escape or transcend ourselves. But it can enable us to see our illusions more clearly and lead a more vital, assertive existence.
  • What was needed, he felt, was not high moral or theological ideals, but a deeply critical form of cultural history or ‘genealogy’ that would uncover the reasons why we humans are as we are, and how we came to be that way. For him, all philosophy could even be redefined as a form of psychology, or history.
  • For those oppressed on grounds of race or class, or for those fighting against colonialism, existentialism offered a change of perspective — literally, as Sartre proposed that all situations be judged according to how they appeared in the eyes of those most oppressed, or those whose suffering was greatest.
  • She observed that we need not expect moral philosophers to ‘live by’ their ideas in a simplistic way, as if they were following a set of rules. But we can expect them to show how their ideas are lived in. We should be able to look in through the windows of a philosophy, as it were, and see how people occupy it, how they move about and how they conduct themselves.
  • the existentialists inhabited their historical and personal world, as they inhabited their ideas. This notion of ‘inhabited philosophy’ is one I’ve borrowed from the English philosopher and novelist Iris Murdoch, who wrote the first full-length book on Sartre and was an early adopter of existentialism
  • What is existentialism anyway?
  • An existentialist who is also phenomenological provides no easy rules for dealing with this condition, but instead concentrates on describing lived experience as it presents itself. — By describing experience well, he or she hopes to understand this existence and awaken us to ways of living more authentic lives.
  • Existentialists concern themselves with individual, concrete human existence. — They consider human existence different from the kind of being other things have. Other entities are what they are, but as a human I am whatever I choose to make of myself at every moment. I am free — — and therefore I’m responsible for everything I do, a dizzying fact which causes — an anxiety inseparable from human existence itself.
  • On the other hand, I am only free within situations, which can include factors in my own biology and psychology as well as physical, historical and social variables of the world into which I have been thrown. — Despite the limitations, I always want more: I am passionately involved in personal projects of all kinds. — Human existence is thus ambiguous: at once boxed in by borders and yet transcendent and exhilarating. —
  • The first part of this is straightforward: a phenomenologist’s job is to describe. This is the activity that Husserl kept reminding his students to do. It meant stripping away distractions, habits, clichés of thought, presumptions and received ideas, in order to return our attention to what he called the ‘things themselves’. We must fix our beady gaze on them and capture them exactly as they appear, rather than as we think they are supposed to be.
  • Husserl therefore says that, to phenomenologically describe a cup of coffee, I should set aside both the abstract suppositions and any intrusive emotional associations. Then I can concentrate on the dark, fragrant, rich phenomenon in front of me now. This ‘setting aside’ or ‘bracketing out’ of speculative add-ons Husserl called epoché — a term borrowed from the ancient Sceptics,
  • The point about rigour is crucial; it brings us back to the first half of the command to describe phenomena. A phenomenologist cannot get away with listening to a piece of music and saying, ‘How lovely!’ He or she must ask: is it plaintive? is it dignified? is it colossal and sublime? The point is to keep coming back to the ‘things themselves’ — phenomena stripped of their conceptual baggage — so as to bail out weak or extraneous material and get to the heart of the experience.
  • Husserlian ‘bracketing out’ or epoché allows the phenomenologist to temporarily ignore the question ‘But is it real?’, in order to ask how a person experiences his or her world. Phenomenology gives a formal mode of access to human experience. It lets philosophers talk about life more or less as non-philosophers do, while still being able to tell themselves they are being methodical and rigorous.
  • Besides claiming to transform the way we think about reality, phenomenologists promised to change how we think about ourselves. They believed that we should not try to find out what the human mind is, as if it were some kind of substance. Instead, we should consider what it does, and how it grasps its experiences.
  • For Brentano, this reaching towards objects is what our minds do all the time. Our thoughts are invariably of or about something, he wrote: in love, something is loved, in hatred, something is hated, in judgement, something is affirmed or denied. Even when I imagine an object that isn’t there, my mental structure is still one of ‘about-ness’ or ‘of-ness’.
  • Except in deepest sleep, my mind is always engaged in this aboutness: it has ‘intentionality’. Having taken the germ of this from Brentano, Husserl made it central to his whole philosophy.
  • Husserl saw in the idea of intentionality a way to sidestep two great unsolved puzzles of philosophical history: the question of what objects ‘really’ are, and the question of what the mind ‘really’ is. By doing the epoché and bracketing out all consideration of reality from both topics, one is freed to concentrate on the relationship in the middle. One can apply one’s descriptive energies to the endless dance of intentionality that takes place in our lives: the whirl of our minds as they seize their intended phenomena one after the other and whisk them around the floor,
  • Understood in this way, the mind hardly is anything at all: it is its aboutness. This makes the human mind (and possibly some animal minds) different from any other naturally occurring entity. Nothing else can be as thoroughly about or of things as the mind is:
  • Some Eastern meditation techniques aim to still this scurrying creature, but the extreme difficulty of this shows how unnatural it is to be mentally inert. Left to itself, the mind reaches out in all directions as long as it is awake — and even carries on doing it in the dreaming phase of its sleep.
  • a mind that is experiencing nothing, imagining nothing, or speculating about nothing can hardly be said to be a mind at all.
  • Three simple ideas — description, phenomenon, intentionality — provided enough inspiration to keep roomfuls of Husserlian assistants busy in Freiburg for decades. With all of human existence awaiting their attention, how could they ever run out of things to do?
  • For Sartre, this gives the mind an immense freedom. If we are nothing but what we think about, then no predefined ‘inner nature’ can hold us back. We are protean.
  • way of this interpretation. Real, not real; inside, outside; what difference did it make? Reflecting on this, Husserl began turning his phenomenology into a branch of ‘idealism’ — the philosophical tradition which denied external reality and defined everything as a kind of private hallucination.
  • For Sartre, if we try to shut ourselves up inside our own minds, ‘in a nice warm room with the shutters closed’, we cease to exist. We have no cosy home: being out on the dusty road is the very definition of what we are.
  • One might think that, if Heidegger had anything worth saying, he could have communicated it in ordinary language. The fact is that he does not want to be ordinary, and he may not even want to communicate in the usual sense. He wants to make the familiar obscure, and to vex us. George Steiner thought that Heidegger’s purpose was less to be understood than to be experienced through a ‘felt strangeness’.
  • He takes Dasein in its most ordinary moments, then talks about it in the most innovative way he can. For Heidegger, Dasein’s everyday Being is right here: it is Being-in-the-world, or In-der-Welt-sein. The main feature of Dasein’s everyday Being-in-the-world right here is that it is usually busy doing something.
  • Thus, for Heidegger, all Being-in-the-world is also a ‘Being-with’ or Mitsein. We cohabit with others in a ‘with-world’, or Mitwelt. The old philosophical problem of how we prove the existence of other minds has now vanished. Dasein swims in the with-world long before it wonders about other minds.
  • Sometimes the best-educated people were those least inclined to take the Nazis seriously, dismissing them as too absurd to last. Karl Jaspers was one of those who made this mistake, as he later recalled, and Beauvoir observed similar dismissive attitudes among the French students in Berlin.
  • In any case, most of those who disagreed with Hitler’s ideology soon learned to keep their view to themselves. If a Nazi parade passed on the street, they would either slip out of view or give the obligatory salute like everyone else, telling themselves that the gesture meant nothing if they did not believe in it. As the psychologist Bruno Bettelheim later wrote of this period, few people will risk their life for such a small thing as raising an arm — yet that is how one’s powers of resistance are eroded away, and eventually one’s responsibility and integrity go with them.
  • for Arendt, if you do not respond adequately when the times demand it, you show a lack of imagination and attention that is as dangerous as deliberately committing an abuse. It amounts to disobeying the one command she had absorbed from Heidegger in those Marburg days: Think!
  • ‘Everything takes place under a kind of anaesthesia. Objectively dreadful events produce a thin, puny emotional response. Murders are committed like schoolboy pranks. Humiliation and moral decay are accepted like minor incidents.’ Haffner thought modernity itself was partly to blame: people had become yoked to their habits and to mass media, forgetting to stop and think, or to disrupt their routines long enough to question what was going on.
  • Heidegger’s former lover and student Hannah Arendt would argue, in her 1951 study The Origins of Totalitarianism, that totalitarian movements thrived at least partly because of this fragmentation in modern lives, which made people more vulnerable to being swept away by demagogues. Elsewhere, she coined the phrase ‘the banality of evil’ to describe the most extreme failures of personal moral awareness.
  • His communicative ideal fed into a whole theory of history: he traced all civilisation to an ‘Axial Period’ in the fifth century BC, during which philosophy and culture exploded simultaneously in Europe, the Middle East and Asia, as though a great bubble of minds had erupted from the earth’s surface. ‘True philosophy needs communion to come into existence,’ he wrote, and added, ‘Uncommunicativeness in a philosopher is virtually a criterion of the untruth of his thinking.’
  • The idea of being called to authenticity became a major theme in later existentialism, the call being interpreted as saying something like ‘Be yourself!’, as opposed to being phony. For Heidegger, the call is more fundamental than that. It is a call to take up a self that you didn’t know you had: to wake up to your Being. Moreover, it is a call to action. It requires you to do something: to take a decision of some sort.
  • Being and Time contained at least one big idea that should have been of use in resisting totalitarianism. Dasein, Heidegger wrote there, tends to fall under the sway of something called das Man or ‘the they’ — an impersonal entity that robs us of the freedom to think for ourselves. To live authentically requires resisting or outwitting this influence, but this is not easy because das Man is so nebulous. Man in German does not mean ‘man’ as in English (that’s der Mann), but a neutral abstraction, something like ‘one’ in the English phrase ‘one doesn’t do that’,
  • for Heidegger, das Man is me. It is everywhere and nowhere; it is nothing definite, but each of us is it. As with Being, it is so ubiquitous that it is difficult to see. If I am not careful, however, das Man takes over the important decisions that should be my own. It drains away my responsibility or ‘answerability’. As Arendt might put it, we slip into banality, failing to think.
  • Jaspers focused on what he called Grenzsituationen — border situations, or limit situations. These are the moments when one finds oneself constrained or boxed in by what is happening, but at the same time pushed by these events towards the limits or outer edge of normal experience. For example, you might have to make a life-or-death choice, or something might remind you suddenly of your mortality,
  • Jaspers’ interest in border situations probably had much to do with his own early confrontation with mortality. From childhood, he had suffered from a heart condition so severe that he always expected to die at any moment. He also had emphysema, which forced him to speak slowly, taking long pauses to catch his breath. Both illnesses meant that he had to budget his energies with care in order to get his work done without endangering his life.
  • If I am to resist das Man, I must become answerable to the call of my ‘voice of conscience’. This call does not come from God, as a traditional Christian definition of the voice of conscience might suppose. It comes from a truly existentialist source: my own authentic self. Alas, this voice is one I do not recognise and may not hear, because it is not the voice of my habitual ‘they-self’. It is an alien or uncanny version of my usual voice. I am familiar with my they-self, but not with my unalienated voice — so, in a weird twist, my real voice is the one that sounds strangest to me.
  • Marcel developed a strongly theological branch of existentialism. His faith distanced him from both Sartre and Heidegger, but he shared a sense of how history makes demands on individuals. In his essay ‘On the Ontological Mystery’, written in 1932 and published in the fateful year of 1933, Marcel wrote of the human tendency to become stuck in habits, received ideas, and a narrow-minded attachment to possessions and familiar scenes. Instead, he urged his readers to develop a capacity for remaining ‘available’ to situations as they arise. Similar ideas of disponibilité or availability had been explored by other writers,
  • Marcel made it his central existential imperative. He was aware of how rare and difficult it was. Most people fall into what he calls ‘crispation’: a tensed, encrusted shape in life — ‘as though each one of us secreted a kind of shell which gradually hardened and imprisoned him’.
  • Bettelheim later observed that, under Nazism, only a few people realised at once that life could not continue unaltered: these were the ones who got away quickly. Bettelheim himself was not among them. Caught in Austria when Hitler annexed it, he was sent first to Dachau and then to Buchenwald, but was then released in a mass amnesty to celebrate Hitler’s birthday in 1939 — an extraordinary reprieve, after which he left at once for America.
  • we are used to reading philosophy as offering a universal message for all times and places — or at least as aiming to do so. But Heidegger disliked the notion of universal truths or universal humanity, which he considered a fantasy. For him, Dasein is not defined by shared faculties of reason and understanding, as the Enlightenment philosophers thought. Still less is it defined by any kind of transcendent eternal soul, as in religious tradition. We do not exist on a higher, eternal plane at all. Dasein’s Being is local: it has a historical situation, and is constituted in time and place.
  • For Marcel, learning to stay open to reality in this way is the philosopher’s prime job. Everyone can do it, but the philosopher is the one who is called on above all to stay awake, so as to be the first to sound the alarm if something seems wrong.
  • Second, it also means understanding that we are historical beings, and grasping the demands our particular historical situation is making on us. In what Heidegger calls ‘anticipatory resoluteness’, Dasein discovers ‘that its uttermost possibility lies in giving itself up’. At that moment, through Being-towards-death and resoluteness in facing up to one’s time, one is freed from the they-self and attains one’s true, authentic self.
  • If we are temporal beings by our very nature, then authentic existence means accepting, first, that we are finite and mortal. We will die: this all-important realisation is what Heidegger calls authentic ‘Being-towards-Death’, and it is fundamental to his philosophy.
  • Hannah Arendt, instead, left early on: she had the benefit of a powerful warning. Just after the Nazi takeover, in spring 1933, she had been arrested while researching materials on anti-Semitism for the German Zionist Organisation at Berlin’s Prussian State Library. Her apartment was searched; both she and her mother were locked up briefly, then released. They fled, without stopping to arrange travel documents. They crossed to Czechoslovakia (then still safe) by a method that sounds almost too fabulous to be true: a sympathetic German family on the border had a house with its front door in Germany and its back door in Czechoslovakia. The family would invite people for dinner, then let them leave through the back door at night.
  • As Sartre argued in his 1943 review of The Stranger, basic phenomenological principles show that experience comes to us already charged with significance. A piano sonata is a melancholy evocation of longing. If I watch a soccer match, I see it as a soccer match, not as a meaningless scene in which a number of people run around taking turns to apply their lower limbs to a spherical object. If the latter is what I’m seeing, then I am not watching some more essential, truer version of soccer; I am failing to watch it properly as soccer at all.
  • Much as they liked Camus personally, neither Sartre nor Beauvoir accepted his vision of absurdity. For them, life is not absurd, even when viewed on a cosmic scale, and nothing can be gained by saying it is. Life for them is full of real meaning, although that meaning emerges differently for each of us.
  • For Sartre, we show bad faith whenever we portray ourselves as passive creations of our race, class, job, history, nation, family, heredity, childhood influences, events, or even hidden drives in our subconscious which we claim are out of our control. It is not that such factors are unimportant: class and race, in particular, he acknowledged as powerful forces in people’s lives, and Simone de Beauvoir would soon add gender to that list.
  • Sartre takes his argument to an extreme point by asserting that even war, imprisonment or the prospect of imminent death cannot take away my existential freedom. They form part of my ‘situation’, and this may be an extreme and intolerable situation, but it still provides only a context for whatever I choose to do next. If I am about to die, I can decide how to face that death. Sartre here resurrects the ancient Stoic idea that I may not choose what happens to me, but I can choose what to make of it, spiritually speaking.
  • But the Stoics cultivated indifference in the face of terrible events, whereas Sartre thought we should remain passionately, even furiously engaged with what happens to us and with what we can achieve. We should not expect freedom to be anything less than fiendishly difficult.
  • Freedom does not mean entirely unconstrained movement, and it certainly does not mean acting randomly. We often mistake the very things that enable us to be free — context, meaning, facticity, situation, a general direction in our lives — for things that define us and take away our freedom. It is only with all of these that we can be free in a real sense.
  • Nor did he mean that privileged groups have the right to pontificate to the poor and downtrodden about the need to ‘take responsibility’ for themselves. That would be a grotesque misreading of Sartre’s point, since his sympathy in any encounter always lay with the more oppressed side. But for each of us — for me — to be in good faith means not making excuses for myself.
  • Camus’ novel gives us a deliberately understated vision of heroism and decisive action compared to those of Sartre and Beauvoir. One can only do so much. It can look like defeatism, but it shows a more realistic perception of what it takes to actually accomplish difficult tasks like liberating one’s country.
  • Camus just kept returning to his core principle: no torture, no killing — at least not with state approval. Beauvoir and Sartre believed they were taking a more subtle and more realistic view. If asked why a couple of innocuous philosophers had suddenly become so harsh, they would have said it was because the war had changed them in profound ways. It had shown them that one’s duties to humanity could be more complicated than they seemed. ‘The war really divided my life in two,’ Sartre said later.
  • Poets and artists ‘let things be’, but they also let things come out and show themselves. They help to ease things into ‘unconcealment’ (Unverborgenheit), which is Heidegger’s rendition of the Greek term alētheia, usually translated as ‘truth’. This is a deeper kind of truth than the mere correspondence of a statement to reality, as when we say ‘The cat is on the mat’ and point to a mat with a cat on it. Long before we can do this, both cat and mat must ‘stand forth out of concealedness’. They must un-hide themselves.
  • Heidegger does not use the word ‘consciousness’ here because — as with his earlier work — he is trying to make us think in a radically different way about ourselves. We are not to think of the mind as an empty cavern, or as a container filled with representations of things. We are not even supposed to think of it as firing off arrows of intentional ‘aboutness’, as in the earlier phenomenology of Brentano. Instead, Heidegger draws us into the depths of his Schwarzwald, and asks us to imagine a gap with sunlight filtering in. We remain in the forest, but we provide a relatively open spot where other beings can bask for a moment. If we did not do this, everything would remain in the thickets, hidden even to itself.
  • The astronomer Carl Sagan began his 1980 television series Cosmos by saying that human beings, though made of the same stuff as the stars, are conscious and are therefore ‘a way for the cosmos to know itself’. Merleau-Ponty similarly quoted his favourite painter Cézanne as saying, ‘The landscape thinks itself in me, and I am its consciousness.’ This is something like what Heidegger thinks humanity contributes to the earth. We are not made of spiritual nothingness; we are part of Being, but we also bring something unique with us. It is not much: a little open space, perhaps with a path and a bench like the one the young Heidegger used to sit on to do his homework. But through us, the miracle occurs.
  • Beauty aside, Heidegger’s late writing can also be troubling, with its increasingly mystical notion of what it is to be human. If one speaks of a human being mainly as an open space or a clearing, or a means of ‘letting beings be’ and dwelling poetically on the earth, then one doesn’t seem to be talking about any recognisable person. The old Dasein has become less human than ever. It is now a forestry feature.
  • Even today, Jaspers, the dedicated communicator, is far less widely read than Heidegger, who has influenced architects, social theorists, critics, psychologists, artists, film-makers, environmental activists, and innumerable students and enthusiasts — including the later deconstructionist and post-structuralist schools, which took their starting point from his late thinking. Having spent the late 1940s as an outsider and then been rehabilitated, Heidegger became the overwhelming presence in university philosophy all over the European continent from then on.
  • As Levinas reflected on this experience, it helped to lead him to a philosophy that was essentially ethical, rather than ontological like Heidegger’s. He developed his ideas from the work of Jewish theologian Martin Buber, whose I and Thou in 1923 had distinguished between my relationship with an impersonal ‘it’ or ‘them’, and the direct personal encounter I have with a ‘you’. Levinas took it further: when I encounter you, we normally meet face-to-face, and it is through your face that you, as another person, can make ethical demands on me. This is very different from Heidegger’s Mitsein or Being-with, which suggests a group of people standing alongside one another, shoulder to shoulder as if in solidarity — perhaps as a unified nation or Volk.
  • For Levinas, we literally face each other, one individual at a time, and that relationship becomes one of communication and moral expectation. We do not merge; we respond to one another. Instead of being co-opted into playing some role in my personal drama of authenticity, you look me in the eyes — and you remain Other. You remain you.
  • This relationship is more fundamental than the self, more fundamental than consciousness, more fundamental even than Being — and it brings an unavoidable ethical obligation. Ever since Husserl, phenomenologists and existentialists had being trying to stretch the definition of existence to incorporate our social lives and relationships. Levinas did more: he turned philosophy around entirely so that these relationships were the foundation of our existence, not an extension of it.
  • Her last work, The Need for Roots, argues, among other things, that none of us has rights, but each one of us has a near-infinite degree of duty and obligation to the other. Whatever the underlying cause of her death — and anorexia nervosa seems to have been involved — no one could deny that she lived out her philosophy with total commitment. Of all the lives touched on in this book, hers is surely the most profound and challenging application of Iris Murdoch’s notion that a philosophy can be ‘inhabited’.
  • Other thinkers took radical ethical turns during the war years. The most extreme was Simone Weil, who actually tried to live by the principle of putting other people’s ethical demands first. Having returned to France after her travels through Germany in 1932, she had worked in a factory so as to experience the degrading nature of such work for herself. When France fell in 1940, her family fled to Marseilles (against her protests), and later to the US and to Britain. Even in exile, Weil made extraordinary sacrifices. If there were people in the world who could not sleep in a bed, she would not do so either, so she slept on the floor.
  • The mystery tradition had roots in Kierkegaard’s ‘leap of faith’. It owed much to the other great nineteenth-century mystic of the impossible, Dostoevsky, and to older theological notions. But it also grew from the protracted trauma that was the first half of the twentieth century. Since 1914, and especially since 1939, people in Europe and elsewhere had come to the realisation that we cannot fully know or trust ourselves; that we have no excuses or explanations for what we do — and yet that we must ground our existence and relationships on something firm, because otherwise we cannot survive.
  • One striking link between these radical ethical thinkers, all on the fringes of our main story, is that they had religious faith. They also granted a special role to the notion of ‘mystery’ — that which cannot be known, calculated or understood, especially when it concerns our relationships with each other. Heidegger was different from them, since he rejected the religion he grew up with and had no real interest in ethics — probably as a consequence of his having no real interest in the human.
  • Meanwhile, the Christian existentialist Gabriel Marcel was also still arguing, as he had since the 1930s, that ethics trumps everything else in philosophy and that our duty to each other is so great as to play the role of a transcendent ‘mystery’. He too had been led to this position partly by a wartime experience: during the First World War he had worked for the Red Cross’ Information Service, with the unenviable job of answering relatives’ inquiries about missing soldiers. Whenever news came, he passed it on, and usually it was not good. As Marcel later said, this task permanently inoculated him against warmongering rhetoric of any kind, and it made him aware of the power of what is unknown in our lives.
  • As the play’s much-quoted and frequently misunderstood final line has it: ‘Hell is other people.’ Sartre later explained that he did not mean to say that other people were hellish in general. He meant that after death we become frozen in their view, unable any longer to fend off their interpretation. In life, we can still do something to manage the impression we make; in death, this freedom goes and we are left entombed in other’s people’s memories and perceptions.
  • We have to do two near-impossible things at once: understand ourselves as limited by circumstances, and yet continue to pursue our projects as though we are truly in control. In Beauvoir’s view, existentialism is the philosophy that best enables us to do this, because it concerns itself so deeply with both freedom and contingency. It acknowledges the radical and terrifying scope of our freedom in life, but also the concrete influences that other philosophies tend to ignore: history, the body, social relationships and the environment.
  • The aspects of our existence that limit us, Merleau-Ponty says, are the very same ones that bind us to the world and give us scope for action and perception. They make us what we are. Sartre acknowledged the need for this trade-off, but he found it more painful to accept. Everything in him longed to be free of bonds, of impediments and limitations
  • Of course we have to learn this skill of interpreting and anticipating the world, and this happens in early childhood, which is why Merleau-Ponty thought child psychology was essential to philosophy. This is an extraordinary insight. Apart from Rousseau, very few philosophers before him had taken childhood seriously; most wrote as though all human experience were that of a fully conscious, rational, verbal adult who has been dropped into this world from the sky — perhaps by a stork.
  • For Merleau-Ponty, we cannot understand our experience if we don’t think of ourselves in part as overgrown babies. We fall for optical illusions because we once learned to see the world in terms of shapes, objects and things relevant to our own interests. Our first perceptions came to us in tandem with our first active experiments in observing the world and reaching out to explore it, and are still linked with those experiences.
  • Another factor in all of this, for Merleau-Ponty, is our social existence: we cannot thrive without others, or not for long, and we need this especially in early life. This makes solipsistic speculation about the reality of others ridiculous; we could never engage in such speculation if we hadn’t already been formed by them.
  • As Descartes could have said (but didn’t), ‘I think, therefore other people exist.’ We grow up with people playing with us, pointing things out, talking, listening, and getting us used to reading emotions and movements; this is how we become capable, reflective, smoothly integrated beings.
  • In general, Merleau-Ponty thinks human experience only makes sense if we abandon philosophy’s time-honoured habit of starting with a solitary, capsule-like, immobile adult self, isolated from its body and world, which must then be connected up again — adding each element around it as though adding clothing to a doll. Instead, for him, we slide from the womb to the birth canal to an equally close and total immersion in the world. That immersion continues as long as we live, although we may also cultivate the art of partially withdrawing from time to time when we want to think or daydream.
  • When he looks for his own metaphor to describe how he sees consciousness, he comes up with a beautiful one: consciousness, he suggests, is like a ‘fold’ in the world, as though someone had crumpled a piece of cloth to make a little nest or hollow. It stays for a while, before eventually being unfolded and smoothed away. There is something seductive, even erotic, in this idea of my conscious self as an improvised pouch in the cloth of the world. I still have my privacy — my withdrawing room. But I am part of the world’s fabric, and I remain formed out of it for as long as I am here.
  • By the time of these works, Merleau-Ponty is taking his desire to describe experience to the outer limits of what language can convey. Just as with the late Husserl or Heidegger, or Sartre in his Flaubert book, we see a philosopher venturing so far from shore that we can barely follow. Emmanuel Levinas would head out to the fringes too, eventually becoming incomprehensible to all but his most patient initiates.
  • Sartre once remarked — speaking of a disagreement they had about Husserl in 1941 — that ‘we discovered, astounded, that our conflicts had, at times, stemmed from our childhood, or went back to the elementary differences of our two organisms’. Merleau-Ponty also said in an interview that Sartre’s work seemed strange to him, not because of philosophical differences, but because of a certain ‘register of feeling’, especially in Nausea, that he could not share. Their difference was one of temperament and of the whole way the world presented itself to them.
  • The two also differed in their purpose. When Sartre writes about the body or other aspects of experience, he generally does it in order to make a different point. He expertly evokes the grace of his café waiter, gliding between the tables, bending at an angle just so, steering the drink-laden tray through the air on the tips of his fingers — but he does it all in order to illustrate his ideas about bad faith. When Merleau-Ponty writes about skilled and graceful movement, the movement itself is his point. This is the thing he wants to understand.
  • We can never move definitively from ignorance to certainty, for the thread of the inquiry will constantly lead us back to ignorance again. This is the most attractive description of philosophy I’ve ever read, and the best argument for why it is worth doing, even (or especially) when it takes us no distance at all from our starting point.
  • By prioritising perception, the body, social life and childhood development, Merleau-Ponty gathered up philosophy’s far-flung outsider subjects and brought them in to occupy the centre of his thought.
  • In his inaugural lecture at the Collège de France on 15 January 1953, published as In Praise of Philosophy, he said that philosophers should concern themselves above all with whatever is ambiguous in our experience. At the same time, they should think clearly about these ambiguities, using reason and science. Thus, he said, ‘The philosopher is marked by the distinguishing trait that he possesses inseparably the taste for evidence and the feeling for ambiguity.’ A constant movement is required between these two
  • As Sartre wrote in response to Hiroshima, humanity had now gained the power to wipe itself out, and must decide every single day that it wanted to live. Camus also wrote that humanity faced the task of choosing between collective suicide and a more intelligent use of its technology — ‘between hell and reason’. After 1945, there seemed little reason to trust in humanity’s ability to choose well.
  • Merleau-Ponty observed in a lecture of 1951 that, more than any previous century, the twentieth century had reminded people how ‘contingent’ their lives were — how at the mercy of historical events and other changes that they could not control. This feeling went on long after the war ended. After the A-bombs were dropped on Hiroshima and Nagasaki, many feared that a Third World War would not be long in coming, this time between the Soviet Union and the United States.
15More

The Fallacy of the 'I Turned Out Fine' Argument - The New York Times - 0 views

  • Most of the messages centered on one single, repeated theme: “I was smacked as a child and I turned out just fine.”
  • It makes sense, doesn’t it? Many of us think, “If I had something happen to me and nothing went wrong, then surely it’s fine for everyone else.”
  • The “I turned out just fine” argument is popular. It means that based on our personal experience we know what works and what doesn’t.But the argument has fatal flaws.
  • ...12 more annotations...
  • It’s what’s known as an anecdotal fallacy. This fallacy, in simple terms, states that “I’m not negatively affected (as far as I can tell), so it must be O.K. for everyone.
  • We are relying on a sample size of one. Ourselves, or someone we know. And we are applying that result to everyone
  • It relies on a decision-making shortcut known as the availability heuristic. Related to the anecdotal fallacy, it’s where we draw on information that is immediately available to us when we make a judgment call.
  • studies show that the availability heuristic is a cognitive bias that can cloud us from making accurate decisions utilizing all the information available. It blinds us to our own prejudices
  • It dismisses well-substantiated, scientific evidence. To say “I turned out fine” is an arrogant dismissal of an alternative evidence-based view
  • The statement closes off discourse and promotes a single perspective that is oblivious to alternatives that may be more enlightened. Anecdotal evidence often undermines scientific results, to our detriment.
  • It leads to entrenched attitudes.
  • Perhaps an inability to engage with views that run counter to our own suggests that we did not turn out quite so “fine.”
  • Where is the threshold for what constitutes having turned out fine? If it means we avoided prison, we may be setting the bar too low. Gainfully employed and have a family of our own? Still a pretty basic standard
  • It is as reasonable to say “I turned out fine because of this” as it is to say “I turned out fine in spite of this.”
  • To claim that on this basis spanking a child is fine means that we fall victim to anecdote, rely on our availability heuristic (thereby dismissing all broader data to the contrary), dismiss alternate views, fail to learn and progress by engaging with a challenging idea.
  • We expect our children to embrace learning and to progress in their thinking as they grow older. They deserve to expect the same from us.
24More

A Voter Revolt Against 'Shareholder Value' - WSJ - 0 views

  • a Feb. 29 quotation from Leslie Moonves, chairman of CBS, CBS -1.76 % that sums up everything wrong with today’s media culture—and with corporate America.
  • Reflecting on the Trump phenomenon at a media and technology conference, Mr. Moonves said that “It may not be good for America, but it’s damn good for CBS.”
  • Mr. Moonves is saying that CBS’s only responsibility is to maximize profits, not only in its entertainment division, but also in its news operation
  • ...21 more annotations...
  • He knows that what his network is doing is against the national interest. He has just enough conscience to be aware that it is “terrible,” but not nearly enough to stop doing it. It might impair shareholder value, after all.
  • Mr. Moonves is suggesting that there is no difference in principle between entertainment and news. Both should be judged by the same standard—ratings. If policy speeches don’t attract large enough audiences, cut to a Trump rally.
  • If the leading purveyors of broadcast journalism make no distinction between news and entertainment, then who can blame viewers for seeing no difference between entertainment and politics?
  • American politicians and parties have used entertainment to draw audiences for the better part of two centuries. But there used to be countervailing forces, including prestigious broadcast news organizations. Not anymore. Once these organizations served as gatekeepers; now they are open-door enablers.
  • They are all in the grip of the same misunderstanding, that their business begins and ends with maximizing shareholder value.
  • They may believe that this is a statutory requirement or a fiduciary duty. If so, they are mistaken
  • It is Milton Friedman’s theory. “There is one and only one social responsibility of business,” he wrote in “Capitalism and Freedom,” “to use its resources and engage in activities designed to increase its profits.”
  • corporate law imposes no enforceable legal duty to maximize either profits or share prices.
  • And it is not politically sustainable. This is the clear meaning of the 2016 presidential election.
  • As a policy argument, Friedman’s thesis flunks key empirical tests
  • during the 1970s, inflation, recession, a stagnant stock market and rising competition from abroad created an opening for Friedman’s theory, which soon dominated corporate boardrooms.
  • In the name of maximizing shareholder value, corporations moved plants and jobs around the world, paid the lowest wages they could get away with, and scheduled work assignments to maintain managerial “flexibility,” whatever the consequences for workers’ families. Meanwhile, their lobbyists engineered a myriad of special interest breaks in the corporate tax code.
  • Now we can see what four decades of pursuing shareholder value at the expense of everything else has yielded
  • Public confidence in corporations is at rock-bottom, and public anger is sky-high
  • The revolt against the corporate economic agenda—free trade, a generous immigration policy, lower corporate taxes and the rest—is sweeping the country.
  • As the Republican rank and file has turned against corporations and New Democrats have given ground to left-wing populists, big business has been left politically homeless.
  • It will take corporate America a long time to climb out of this self-created hole.
  • Its first step should be to back long-overdue proposals for improving workers’ lives and incomes. Paid family leave is an idea whose time has come; so is a catch-up increase in the federal minimum wage; so are stable and predictable schedules for part-time workers.
  • Allowing workers to share in profits and productivity increases would be another good step.
  • Above all, corporate leaders should grasp the distinction between immediate gain and self-interest rightly understood. Pushing for the last increment of profit over the next quarter and the one after that comes at the expense of the strategies that can leave firms best positioned for the future.
  • America needs a new generation of corporate statesmen.
« First ‹ Previous 81 - 100 of 178 Next › Last »
Showing 20 items per page