Skip to main content

Home/ TOK Friends/ Group items tagged life-span

Rss Feed Group items tagged

tongoscar

California Is Booming. Why Are So Many Californians Unhappy? - The New York Times - 0 views

  • For all its forward-thinking companies and liberal social and environmental policies, the state has mostly put higher-value jobs and industries in expensive coastal enclaves, while pushing lower-paid workers and lower-cost housing to inland areas like the Central Valley. This has made California the most expensive state — with a median home value of $550,000, about double that of the nation — and created a growing supply of three-hour “super commuters.” And while it has some of the highest wages in the country, it also has the highest poverty rate based on its cost of living, an average of 18.1 percent from 2016 to 2018. That helps explain why the state has lost more than a million residents to other states since 2006, and why the population growth rate for the year that ended July 1 was the lowest since 1900.“What’s happening in California right now is a warning shot to the rest of the country,”
  • “It’s a warning about income inequality and suburban sprawl, and how those intersect with quality of life and climate change.” You can see this in California economic forecasts for 2020, which play down the threat of a global trade war and play up the challenge of continuing to add jobs without affordable places for middle- and lower-income workers to live.
  • As the economy picked up and housing costs resumed their rise, lower-paid service and professional workers moved to distant exurbs, while homelessness spiraled to the point that local political leaders are all but declaring they are out of solutions.
  • ...2 more annotations...
  • California is at a crossroads. The state has a thriving $3 trillion economy with record low unemployment, a surplus of well-paying jobs, and several of the world’s most valuable corporations, including Apple, Google and Facebook. Its median household income has grown about 17 percent since 2011, compared with about 10 percent nationally, adjusted for inflation.
  • There are increasing complaints in Oregon, Nevada and Idaho that rents and home prices there are being pushed up by new arrivals fleeing California.
Javier E

Elon Musk May Kill Us Even If Donald Trump Doesn't - 0 views

  • In his extraordinary 2021 book, The Constitution of Knowledge: A Defense of Truth, Jonathan Rauch, a scholar at Brookings, writes that modern societies have developed an implicit “epistemic” compact–an agreement about how we determine truth–that rests on a broad public acceptance of science and reason, and a respect and forbearance towards institutions charged with advancing knowledge.
  • Today, Rauch writes, those institutions have given way to digital “platforms” that traffic in “information” rather than knowledge and disseminate that information not according to its accuracy but its popularity. And what is popular is sensation, shock, outrage. The old elite consensus has given way to an algorithm. Donald Trump, an entrepreneur of outrage, capitalized on the new technology to lead what Rauch calls “an epistemic secession.”
  • Rauch foresees the arrival of “Internet 3.0,” in which the big companies accept that content regulation is in their interest and erect suitable “guardrails.” In conversation with me, Rauch said that social media companies now recognize that their algorithm are “toxic,” and spoke hopefully of alternative models like Mastodon, which eschews algorithms and allows users to curate their own feeds
  • ...10 more annotations...
  • In an Atlantic essay, “Why The Past Ten Years of American Life have Been Uniquely Stupid,” and in a follow-up piece, Haidt argued that the Age of Gutenberg–of books and the depth understanding that comes with them–ended somewhere around 2014 with the rise of “Share,” “Like” and “Retweet” buttons that opened the way for trolls, hucksters and Trumpists
  • The new age of “hyper-virality,” he writes, has given us both January 6 and cancel culture–ugly polarization in both directions. On the subject of stupidification, we should add the fact that high school students now get virtually their entire stock of knowledge about the world from digital platforms.
  • Haidt proposed several reforms, including modifying Facebook’s “Share” function and requiring “user verification” to get rid of trolls. But he doesn’t really believe in his own medicine
  • Haidt said that the era of “shared understanding” is over–forever. When I asked if he could envision changes that would help protect democracy, Haidt quoted Goldfinger: “Do you expect me to talk?” “No, Mr. Bond, I expect you to die!”
  • Social media is a public health hazard–the cognitive equivalent of tobacco and sugary drinks. Adopting a public health model, we could, for examople, ban the use of algorithms to reduce virality, or even require social media platforms to adopt a subscription rather than advertising revenue model and thus remove their incentive to amass ev er more eyeballs.
  • We could, but we won’t, because unlike other public health hazards, digital platforms are forms of speech. Fox New is probably responsible for more polarization than all social media put together, but the federal government could not compel it–and all other media firms–to change its revenue model.
  • If Mark Zuckerberg or Elon Musk won’t do so out of concern for the public good–a pretty safe bet–they could be compelled to do so only by public or competitive pressure. 
  • Taiwan has provide resilient because its society is resilient; people reject China’s lies. We, here, don’t lack for fact-checkers, but rather for people willing to believe them. The problem is not the technology, but ourselves.
  • you have to wonder if people really are repelled by our poisonous discourse, or by the hailstorm of disinformation, or if they just want to live comfortably inside their own bubble, and not somebody else’
  • If Jonathan Haidt is right, it’s not because we’ve created a self-replicating machine that is destined to annihilate reason; it’s because we are the self-replicating machine.
Javier E

Why Is It So Hard to Be Rational? | The New Yorker - 0 views

  • an unusually large number of books about rationality were being published this year, among them Steven Pinker’s “Rationality: What It Is, Why It Seems Scarce, Why It Matters” (Viking) and Julia Galef’s “The Scout Mindset: Why Some People See Things Clearly and Others Don’t” (Portfolio).
  • When the world changes quickly, we need strategies for understanding it. We hope, reasonably, that rational people will be more careful, honest, truthful, fair-minded, curious, and right than irrational ones.
  • And yet rationality has sharp edges that make it hard to put at the center of one’s life
  • ...43 more annotations...
  • You might be well-intentioned, rational, and mistaken, simply because so much in our thinking can go wrong. (“RATIONAL, adj.: Devoid of all delusions save those of observation, experience and reflection,”
  • You might be rational and self-deceptive, because telling yourself that you are rational can itself become a source of bias. It’s possible that you are trying to appear rational only because you want to impress people; or that you are more rational about some things (your job) than others (your kids); or that your rationality gives way to rancor as soon as your ideas are challenged. Perhaps you irrationally insist on answering difficult questions yourself when you’d be better off trusting the expert consensus.
  • Not just individuals but societies can fall prey to false or compromised rationality. In a 2014 book, “The Revolt of the Public and the Crisis of Authority in the New Millennium,” Martin Gurri, a C.I.A. analyst turned libertarian social thinker, argued that the unmasking of allegedly pseudo-rational institutions had become the central drama of our age: people around the world, having concluded that the bigwigs in our colleges, newsrooms, and legislatures were better at appearing rational than at being so, had embraced a nihilist populism that sees all forms of public rationality as suspect.
  • modern life would be impossible without those rational systems; we must improve them, not reject them. We have no choice but to wrestle with rationality—an ideal that, the sociologist Max Weber wrote, “contains within itself a world of contradictions.”
  • Where others might be completely convinced that G.M.O.s are bad, or that Jack is trustworthy, or that the enemy is Eurasia, a Bayesian assigns probabilities to these propositions. She doesn’t build an immovable world view; instead, by continually updating her probabilities, she inches closer to a more useful account of reality. The cooking is never done.
  • Rationality is one of humanity’s superpowers. How do we keep from misusing it?
  • Start with the big picture, fixing it firmly in your mind. Be cautious as you integrate new information, and don’t jump to conclusions. Notice when new data points do and do not alter your baseline assumptions (most of the time, they won’t alter them), but keep track of how often those assumptions seem contradicted by what’s new. Beware the power of alarming news, and proceed by putting it in a broader, real-world context.
  • Bayesian reasoning implies a few “best practices.”
  • Keep the cooked information over here and the raw information over there; remember that raw ingredients often reduce over heat
  • We want to live in a more rational society, but not in a falsely rationalized one. We want to be more rational as individuals, but not to overdo it. We need to know when to think and when to stop thinking, when to doubt and when to trust.
  • But the real power of the Bayesian approach isn’t procedural; it’s that it replaces the facts in our minds with probabilities.
  • Applied to specific problems—Should you invest in Tesla? How bad is the Delta variant?—the techniques promoted by rationality writers are clarifying and powerful.
  • the rationality movement is also a social movement; rationalists today form what is sometimes called the “rationality community,” and, as evangelists, they hope to increase its size.
  • In “Rationality,” “The Scout Mindset,” and other similar books, irrationality is often presented as a form of misbehavior, which might be rectified through education or socialization.
  • Greg tells me that, in his business, it’s not enough to have rational thoughts. Someone who’s used to pondering questions at leisure might struggle to learn and reason when the clock is ticking; someone who is good at reaching rational conclusions might not be willing to sign on the dotted line when the time comes. Greg’s hedge-fund colleagues describe as “commercial”—a compliment—someone who is not only rational but timely and decisive.
  • You can know what’s right but still struggle to do it.
  • Following through on your own conclusions is one challenge. But a rationalist must also be “metarational,” willing to hand over the thinking keys when someone else is better informed or better trained. This, too, is harder than it sounds.
  • For all this to happen, rationality is necessary, but not sufficient. Thinking straight is just part of the work. 
  • I found it possible to be metarational with my dad not just because I respected his mind but because I knew that he was a good and cautious person who had my and my mother’s best interests at heart.
  • between the two of us, we had the right ingredients—mutual trust, mutual concern, and a shared commitment to reason and to act.
  • Intellectually, we understand that our complex society requires the division of both practical and cognitive labor. We accept that our knowledge maps are limited not just by our smarts but by our time and interests. Still, like Gurri’s populists, rationalists may stage their own contrarian revolts, repeatedly finding that no one’s opinions but their own are defensible. In letting go, as in following through, one’s whole personality gets involved.
  • in truth, it maps out a series of escalating challenges. In search of facts, we must make do with probabilities. Unable to know it all for ourselves, we must rely on others who care enough to know. We must act while we are still uncertain, and we must act in time—sometimes individually, but often together.
  • The realities of rationality are humbling. Know things; want things; use what you know to get what you want. It sounds like a simple formula.
  • The real challenge isn’t being right but knowing how wrong you might be.By Joshua RothmanAugust 16, 2021
  • Writing about rationality in the early twentieth century, Weber saw himself as coming to grips with a titanic force—an ascendant outlook that was rewriting our values. He talked about rationality in many different ways. We can practice the instrumental rationality of means and ends (how do I get what I want?) and the value rationality of purposes and goals (do I have good reasons for wanting what I want?). We can pursue the rationality of affect (am I cool, calm, and collected?) or develop the rationality of habit (do I live an ordered, or “rationalized,” life?).
  • Weber worried that it was turning each individual into a “cog in the machine,” and life into an “iron cage.” Today, rationality and the words around it are still shadowed with Weberian pessimism and cursed with double meanings. You’re rationalizing the org chart: are you bringing order to chaos, or justifying the illogical?
  • For Aristotle, rationality was what separated human beings from animals. For the authors of “The Rationality Quotient,” it’s a mental faculty, parallel to but distinct from intelligence, which involves a person’s ability to juggle many scenarios in her head at once, without letting any one monopolize her attention or bias her against the rest.
  • In “The Rationality Quotient: Toward a Test of Rational Thinking” (M.I.T.), from 2016, the psychologists Keith E. Stanovich, Richard F. West, and Maggie E. Toplak call rationality “a torturous and tortured term,” in part because philosophers, sociologists, psychologists, and economists have all defined it differently
  • Galef, who hosts a podcast called “Rationally Speaking” and co-founded the nonprofit Center for Applied Rationality, in Berkeley, barely uses the word “rationality” in her book on the subject. Instead, she describes a “scout mindset,” which can help you “to recognize when you are wrong, to seek out your blind spots, to test your assumptions and change course.” (The “soldier mindset,” by contrast, encourages you to defend your positions at any cost.)
  • Galef tends to see rationality as a method for acquiring more accurate views.
  • Pinker, a cognitive and evolutionary psychologist, sees it instrumentally, as “the ability to use knowledge to attain goals.” By this definition, to be a rational person you have to know things, you have to want things, and you have to use what you know to get what you want.
  • Introspection is key to rationality. A rational person must practice what the neuroscientist Stephen Fleming, in “Know Thyself: The Science of Self-Awareness” (Basic Books), calls “metacognition,” or “the ability to think about our own thinking”—“a fragile, beautiful, and frankly bizarre feature of the human mind.”
  • A successful student uses metacognition to know when he needs to study more and when he’s studied enough: essentially, parts of his brain are monitoring other parts.
  • In everyday life, the biggest obstacle to metacognition is what psychologists call the “illusion of fluency.” As we perform increasingly familiar tasks, we monitor our performance less rigorously; this happens when we drive, or fold laundry, and also when we think thoughts we’ve thought many times before
  • The trick is to break the illusion of fluency, and to encourage an “awareness of ignorance.”
  • metacognition is a skill. Some people are better at it than others. Galef believes that, by “calibrating” our metacognitive minds, we can improve our performance and so become more rational
  • There are many calibration methods
  • nowing about what you know is Rationality 101. The advanced coursework has to do with changes in your knowledge.
  • Most of us stay informed straightforwardly—by taking in new information. Rationalists do the same, but self-consciously, with an eye to deliberately redrawing their mental maps.
  • The challenge is that news about distant territories drifts in from many sources; fresh facts and opinions aren’t uniformly significant. In recent decades, rationalists confronting this problem have rallied behind the work of Thomas Bayes
  • So-called Bayesian reasoning—a particular thinking technique, with its own distinctive jargon—has become de rigueur.
  • the basic idea is simple. When new information comes in, you don’t want it to replace old information wholesale. Instead, you want it to modify what you already know to an appropriate degree. The degree of modification depends both on your confidence in your preëxisting knowledge and on the value of the new data. Bayesian reasoners begin with what they call the “prior” probability of something being true, and then find out if they need to adjust it.
  • Bayesian reasoning is an approach to statistics, but you can use it to interpret all sorts of new information.
Javier E

The American Scholar: The Disadvantages of an Elite Education - William Deresiewicz - 1 views

  • the last thing an elite education will teach you is its own inadequacy
  • I’m talking about the whole system in which these skirmishes play out. Not just the Ivy League and its peer institutions, but also the mechanisms that get you there in the first place: the private and affluent public “feeder” schools, the ever-growing parastructure of tutors and test-prep courses and enrichment programs, the whole admissions frenzy and everything that leads up to and away from it. The message, as always, is the medium. Before, after, and around the elite college classroom, a constellation of values is ceaselessly inculcated.
  • The first disadvantage of an elite education, as I learned in my kitchen that day, is that it makes you incapable of talking to people who aren’t like you. Elite schools pride themselves on their diversity, but that diversity is almost entirely a matter of ethnicity and race. With respect to class, these schools are largely—indeed increasingly—homogeneous. Visit any elite campus in our great nation and you can thrill to the heartwarming spectacle of the children of white businesspeople and professionals studying and playing alongside the children of black, Asian, and Latino businesspeople and professionals.
  • ...34 more annotations...
  • My education taught me to believe that people who didn’t go to an Ivy League or equivalent school weren’t worth talking to, regardless of their class. I was given the unmistakable message that such people were beneath me.
  • The existence of multiple forms of intelligence has become a commonplace, but however much elite universities like to sprinkle their incoming classes with a few actors or violinists, they select for and develop one form of intelligence: the analytic.
  • Students at places like Cleveland State, unlike those at places like Yale, don’t have a platoon of advisers and tutors and deans to write out excuses for late work, give them extra help when they need it, pick them up when they fall down.
  • When people say that students at elite schools have a strong sense of entitlement, they mean that those students think they deserve more than other people because their SAT scores are higher.
  • The political implications should be clear. As John Ruskin told an older elite, grabbing what you can get isn’t any less wicked when you grab it with the power of your brains than with the power of your fists.
  • students at places like Yale get an endless string of second chances. Not so at places like Cleveland State.
  • The second disadvantage, implicit in what I’ve been saying, is that an elite education inculcates a false sense of self-worth. Getting to an elite college, being at an elite college, and going on from an elite college—all involve numerical rankings: SAT, GPA, GRE. You learn to think of yourself in terms of those numbers. They come to signify not only your fate, but your identity; not only your identity, but your value.
  • For the elite, there’s always another extension—a bailout, a pardon, a stint in rehab—always plenty of contacts and special stipends—the country club, the conference, the year-end bonus, the dividend.
  • In short, the way students are treated in college trains them for the social position they will occupy once they get out. At schools like Cleveland State, they’re being trained for positions somewhere in the middle of the class system, in the depths of one bureaucracy or another. They’re being conditioned for lives with few second chances, no extensions, little support, narrow opportunity—lives of subordination, supervision, and control, lives of deadlines, not guidelines. At places like Yale, of course, it’s the reverse.
  • Elite schools nurture excellence, but they also nurture what a former Yale graduate student I know calls “entitled mediocrity.”
  • An elite education gives you the chance to be rich—which is, after all, what we’re talking about—but it takes away the chance not to be. Yet the opportunity not to be rich is one of the greatest opportunities with which young Americans have been blessed. We live in a society that is itself so wealthy that it can afford to provide a decent living to whole classes of people who in other countries exist (or in earlier times existed) on the brink of poverty or, at least, of indignity. You can live comfortably in the United States as a schoolteacher, or a community organizer, or a civil rights lawyer, or an artist
  • The liberal arts university is becoming the corporate university, its center of gravity shifting to technical fields where scholarly expertise can be parlayed into lucrative business opportunities.
  • You have to live in an ordinary house instead of an apartment in Manhattan or a mansion in L.A.; you have to drive a Honda instead of a BMW or a Hummer; you have to vacation in Florida instead of Barbados or Paris, but what are such losses when set against the opportunity to do work you believe in, work you’re suited for, work you love, every day of your life? Yet it is precisely that opportunity that an elite education takes away. How can I be a schoolteacher—wouldn’t that be a waste of my expensive education?
  • Isn’t it beneath me? So a whole universe of possibility closes, and you miss your true calling.
  • This is not to say that students from elite colleges never pursue a riskier or less lucrative course after graduation, but even when they do, they tend to give up more quickly than others.
  • But if you’re afraid to fail, you’re afraid to take risks, which begins to explain the final and most damning disadvantage of an elite education: that it is profoundly anti-intellectual.
  • being an intellectual is not the same as being smart. Being an intellectual means more than doing your homework.
  • The system forgot to teach them, along the way to the prestige admissions and the lucrative jobs, that the most important achievements can’t be measured by a letter or a number or a name. It forgot that the true purpose of education is to make minds, not careers.
  • Being an intellectual means, first of all, being passionate about ideas—and not just for the duration of a semester, for the sake of pleasing the teacher, or for getting a good grade.
  • Only a small minority have seen their education as part of a larger intellectual journey, have approached the work of the mind with a pilgrim soul. These few have tended to feel like freaks, not least because they get so little support from the university itself. Places like Yale, as one of them put it to me, are not conducive to searchers. GA_googleFillSlot('Rectangle_InArticle_Right'); GA_googleCreateDomIframe("google_ads_div_Rectangle_InArticle_Right_ad_container" ,"Rectangle_InArticle_Right"); Places like Yale are simply not set up to help students ask the big questions
  • Professors at top research institutions are valued exclusively for the quality of their scholarly work; time spent on teaching is time lost. If students want a conversion experience, they’re better off at a liberal arts college.
  • When elite universities boast that they teach their students how to think, they mean that they teach them the analytic and rhetorical skills necessary for success in law or medicine or science or business.
  • Although the notion of breadth is implicit in the very idea of a liberal arts education, the admissions process increasingly selects for kids who have already begun to think of themselves in specialized terms—the junior journalist, the budding astronomer, the language prodigy. We are slouching, even at elite schools, toward a glorified form of vocational training.
  • There’s a reason elite schools speak of training leaders, not thinkers—holders of power, not its critics. An independent mind is independent of all allegiances, and elite schools, which get a large percentage of their budget from alumni giving, are strongly invested in fostering institutional loyalty.
  • At a school like Yale, students who come to class and work hard expect nothing less than an A-. And most of the time, they get it.
  • Yet there is a dimension of the intellectual life that lies above the passion for ideas, though so thoroughly has our culture been sanitized of it that it is hardly surprising if it was beyond the reach of even my most alert students. Since the idea of the intellectual emerged in the 18th century, it has had, at its core, a commitment to social transformation. Being an intellectual means thinking your way toward a vision of the good society and then trying to realize that vision by speaking truth to power.
  • It takes more than just intellect; it takes imagination and courage.
  • Being an intellectual begins with thinking your way outside of your assumptions and the system that enforces them. But students who get into elite schools are precisely the ones who have best learned to work within the system, so it’s almost impossible for them to see outside it, to see that it’s even there.
  • Paradoxically, the situation may be better at second-tier schools and, in particular, again, at liberal arts colleges than at the most prestigious universities. Some students end up at second-tier schools because they’re exactly like students at Harvard or Yale, only less gifted or driven. But others end up there because they have a more independent spirit. They didn’t get straight A’s because they couldn’t be bothered to give everything in every class. They concentrated on the ones that meant the most to them or on a single strong extracurricular passion or on projects that had nothing to do with school
  • I’ve been struck, during my time at Yale, by how similar everyone looks. You hardly see any hippies or punks or art-school types, and at a college that was known in the ’80s as the Gay Ivy, few out lesbians and no gender queers. The geeks don’t look all that geeky; the fashionable kids go in for understated elegance. Thirty-two flavors, all of them vanilla.
  • The most elite schools have become places of a narrow and suffocating normalcy. Everyone feels pressure to maintain the kind of appearance—and affect—that go with achievement
  • Now that students are in constant electronic contact, they never have trouble finding each other. But it’s not as if their compulsive sociability is enabling them to develop deep friendships.
  • What happens when busyness and sociability leave no room for solitude? The ability to engage in introspection, I put it to my students that day, is the essential precondition for living an intellectual life, and the essential precondition for introspection is solitude
  • the life of the mind is lived one mind at a time: one solitary, skeptical, resistant mind at a time. The best place to cultivate it is not within an educational system whose real purpose is to reproduce the class system.
Javier E

Korean philosophy is built upon daily practice of good habits | Aeon Essays - 0 views

  • ‘We are unknown, we knowers, ourselves to ourselves,’ wrote Friedrich Nietzsche at the beginning of On the Genealogy of Morals (1887
  • This seeking after ourselves, however, is not something that is lacking in Buddhist and Confucian traditions – especially not in the case of Korean philosophy. Self-cultivation, central to the tradition, underscores that the onus is on the individual to develop oneself, without recourse to the divine or the supernatural
  • Korean philosophy is practical, while remaining agnostic to a large degree: recognising the spirit realm but highlighting that we ourselves take charge of our lives by taking charge of our minds
  • ...36 more annotations...
  • The word for ‘philosophy’ in Korean is 철학, pronounced ch’ŏrhak. It literally means the ‘study of wisdom’ or, perhaps better, ‘how to become wise’, which reflects its more dynamic and proactive implications
  • At night, in the darkness of the cave, he drank water from a perfectly useful ‘bowl’. But when he could see properly, he found that there was no ‘bowl’ at all, only a disgusting human skull.
  • Our lives and minds are affected by others (and their actions), as others (and their minds) are affected by our actions. This is particularly true in the Korean application of Confucian and Buddhist ideas.
  • Wŏnhyo understood that how we think about things shapes their very existence – and in turn our own existence, which is constructed according to our thoughts.
  • In the Korean tradition of philosophy, human beings are social beings, therefore knowing how to interact with others is an essential part of living a good life – indeed, living well with others is our real contribution to human life
  • he realised that there isn’t a difference between the ‘bowl’ and the skull: the only difference lies with us and our perceptions. We interpret our lives through a continual stream of thoughts, and so we become what we think, or rather how we think
  • As our daily lives are shaped by our thoughts, so our experience of this reality is good or bad – depending on our thoughts – which make things ‘appear’ good or bad because, in ‘reality’, things in and of themselves are devoid of their own independent nature
  • We can take from Wŏnhyo the idea that, if you change the patterns that have become engrained in how you think, you will begin to live differently. To do this, you need to change your mental habits, which is why meditation and mindful awareness can help. And this needs to be practised every day
  • Wŏnhyo’s most important work is titled Awaken your Mind and Practice (in Korean, Palsim suhaeng-jang). It is an explicit call to younger adherents to put Buddhist ideas into practice, and an indirect warning not to get lost in contemplation or in the study of text
  • While Wŏnhyo had emphasised the mind and the need to ‘practise’ Buddhism, a later Korean monk, Chinul (1158-1210), spearheaded Sŏn, the meditational tradition in Korea that espoused the idea of ‘sudden enlightenment’ that alerts the mind, accompanied by ‘gradual cultivation’
  • we still need to practise meditation, for if not we can easily fall into our old ways even if our minds have been awakened
  • his greatest contribution to Sŏn is Secrets on Cultivating the Mind (Susim kyŏl). This text outlines in detail his teachings on sudden awakening followed by the need for gradual cultivation
  • hinul’s approach recognises the mind as the ‘essence’ of one’s Buddha nature (contained in the mind, which is inherently good), while continual practice and cultivation aids in refining its ‘function’ – this is the origin of the ‘essence-function’ concept that has since become central to Korean philosophy.
  • These ideas also influenced the reformed view of Confucianism that became linked with the mind and other metaphysical ideas, finally becoming known as Neo-Confucianism.
  • During the Chosŏn dynasty (1392-1910), the longest lasting in East Asian history, Neo-Confucianism became integrated into society at all levels through rituals for marriage, funerals and ancestors
  • Neo-Confucianism recognises that we as individuals exist through plural relationships with responsibilities to others (as a child, brother/sister, lover, husband/wife, parent, teacher/student and so on), an idea nicely captured in 2000 by the French philosopher Jean-Luc Nancy when he described our ‘being’ as ‘singular plural’
  • Corrupt interpretations of Confucianism by heteronormative men have historically championed these ideas in terms of vertical relationships rather than as a reciprocal set of benevolent social interactions, meaning that women have suffered greatly as a result.
  • Setting aside these sexist and self-serving interpretations, Confucianism emphasises that society works as an interconnected set of complementary reciprocal relationships that should be beneficial to all parties within a social system
  • Confucian relationships have the potential to offer us an example of effective citizenship, similar to that outlined by Cicero, where the good of the republic or state is at the centre of being a good citizen
  • There is a general consensus in Korean philosophy that we have an innate sociability and therefore should have a sense of duty to each other and to practise virtue.
  • The main virtue of Confucianism is the idea of ‘humanity’, coming from the Chinese character 仁, often left untranslated and written as ren and pronounced in Korean as in.
  • It is a combination of the character for a human being and the number two. In other words, it signifies what (inter)connects two people, or rather how they should interact in a humane or benevolent manner to each other. This character therefore highlights the link between people while emphasising that the most basic thing that makes us ‘human’ is our interaction with others.
  • Neo-Confucianism adopted a turn towards a more mind-centred view in the writings of the Korean scholar Yi Hwang, known by his pen name T’oegye (1501-70), who appears on the 1,000-won note. He greatly influenced Neo-Confucianism in Japan through his formidable text, Ten Diagrams on Sage Learning (Sŏnghak sipto), composed in 1568, which was one of the most-reproduced texts of the entire Chosŏn dynasty and represents the synthesis of Neo-Confucian thought in Korea
  • with commentaries that elucidate the moral principles of Confucianism, related to the cardinal relationships and education. It also embodies T’oegye’s own development of moral psychology through his focus on the mind, and illuminates the importance of teaching and the practice of self-cultivation.
  • He writes that we ourselves can transform the unrestrained mind and its desires, and achieve sagehood, if we take the arduous, gradual path of self-cultivation centred on the mind.
  • Confucians had generally accepted the Mencian idea that human nature was embodied in the unaroused state of the mind, before it was shaped by its environment. The mind in its unaroused state was taken to be theoretically good. However, this inborn tendency for goodness is always in danger of being reduced to passivity, unless you cultivate yourself as a person of ‘humanity’ (in the Confucian sense mentioned above).
  • You should constantly try to activate your humanity to allow the unhampered operation of the original mind to manifest itself through socially responsible and moral character in action
  • Humanity is the realisation of what I describe as our ‘optimum level of perfection’ that exists in an inherent stage of potentiality due our innate good nature
  • This, in a sense, is like the Buddha nature of the Buddhists, which suggests we are already enlightened and just need to recover our innate mental state. Both philosophies are hopeful: humans are born good with the potential to correct their own flaws and failures
  • this could hardly contrast any more greatly with the Christian doctrine of original sin
  • The seventh diagram in T’oegye’s text is entitled ‘The Diagram of the Explanation of Humanity’ (Insŏl-to). Here he warns how one’s good inborn nature may become impaired, hampering the operation of the original mind and negatively impacting our character in action. Humanity embodies the gradual realisation of our optimum level of perfection that already exists in our mind but that depends on how we think about things and how we relate that to others in a social context
  • For T’oegye, the key to maintaining our capacity to remain level-headed, and to control our impulses and emotions, was kyŏng. This term is often translated as ‘seriousness’, occasionally ‘mindfulness’, and it identifies the serious need for constant effort to control one’s mind in order to go about one’s life in a healthy manner
  • For T’oegye, mindfulness is as serious as meditation is for the Buddhists. In fact, the Neo-Confucians had their own meditational practice of ‘quiet-sitting’ (chŏngjwa), which focused on recovering the calm and not agitated ‘original mind’, before putting our daily plans into action
  • These diagrams reinforce this need for a daily practice of Confucian mindfulness, because practice leads to the ‘good habit’ of creating (and maintaining) routines. There is no short-cut provided, no weekend intro to this practice: it is life-long, and that is what makes it transformative, leading us to become better versions of who were in the beginning. This is consolation of Korean philosophy.
  • Seeing the world as it is can steer us away from making unnecessary mistakes, while highlighting what is good and how to maintain that good while also reducing anxiety from an agitated mind and harmful desires. This is why Korean philosophy can provide us with consolation; it recognises the bad, but prioritises the good, providing several moral pathways that are referred to in the East Asian traditions (Confucianism, Buddhism and Daoism) as modes of ‘self-cultivation’
  • As social beings, we penetrate the consciousness of others, and so humans are linked externally through conduct but also internally through thought. Humanity is a unifying approach that holds the potential to solve human problems, internally and externally, as well as help people realise the perfection that is innately theirs
Javier E

Opinion | Privacy Is Too Big to Understand - The New York Times - 1 views

  • There is “no single rhetorical approach likely to work on a given audience and none too dangerous to try. Any story that sticks is a good one,”
  • This newsletter is about finding ways to make this stuff stick in your mind and to arm you with the information you need to take control of your digital life.
  • how to start? The definition of privacy itself. I think it’s time to radically expand it.
  • ...12 more annotations...
  • “Privacy” is an impoverished word — far too small a word to describe what we talk about when we talk about the mining, transmission, storing, buying, selling, use and misuse of our personal information.
  • “hyperobjects,” a concept so all-encompassing that it is impossible to adequately describe
  • invite skepticism because their scale is so vast and sometimes abstract.
  • When technology governs so many aspects of our lives — and when that technology is powered by the exploitation of our data — privacy isn’t just about knowing your secrets, it’s about autonomy
  • “Privacy is really about being able to define for ourselves who we are for the world and on our own terms,”
  • not a choice that belongs to an algorithm or data brokerEntities that collect, aggregate and sell individuals’ personal data, derivatives and inferences from disparate public and private sources. Glossary and definitely not to Facebook.”
  • privacy is about how that data is used to take away our control
  • real-time data, once assumed to be protected by phone companies, was available for sale to bounty hunters for a $300 fee
  • ICE officials partnered with a private data firm to track license plate data.
  • It means reckoning with private surveillance databases armed with dossiers on regular citizens and outsourced to the highest bidder
  • “Years ago we worried about the N.S.A. building huge server farms, but now it’s much cheaper to go to a private-service vendor and outsource this to a company who can cloak their activity in trade secrets,
  • “It’s comparable to asking people to stop using air conditioning because of the ozone layer. It’s not likely to happen because the immediate comfort is more valuable than the long-term fear.
Javier E

Adam Kirsch: Art Over Biology | The New Republic - 1 views

  • Nietzsche, who wrote in Human, All Too Human, under the rubric “Art dangerous for the artist,” about the particular ill-suitedness of the artist to flourishing in a modern scientific age: When art seizes an individual powerfully, it draws him back to the views of those times when art flowered most vigorously.... The artist comes more and more to revere sudden excitements, believes in gods and demons, imbues nature with a soul, hates science, becomes unchangeable in his moods like the men of antiquity, and desires the overthrow of all conditions that are not favorable to art.... Thus between him and the other men of his period who are the same age a vehement antagonism is finally generated, and a sad end
  • What is modern is the sense of the superiority of the artist’s inferiority, which is only possible when the artist and the intellectual come to see the values of ordinary life—prosperity, family, worldly success, and happiness—as inherently contemptible.
  • Art, according to a modern understanding that has not wholly vanished today, is meant to be a criticism of life, especially of life in a materialist, positivist civilization such as our own. If this means the artist does not share in civilization’s boons, then his suffering will be a badge of honor.
  • ...18 more annotations...
  • The iron law of Darwinian evolution is that everything that exists strives with all its power to reproduce, to extend life into the future, and that every feature of every creature can be explained as an adaptation toward this end. For the artist to deny any connection with the enterprise of life, then, is to assert his freedom from this universal imperative; to reclaim negatively the autonomy that evolution seems to deny to human beings. It is only because we can freely choose our own ends that we can decide not to live for life, but for some other value that we posit. The artist’s decision to produce spiritual offspring rather than physical ones is thus allied to the monk’s celibacy and the warrior’s death for his country, as gestures that deny the empire of mere life.
  • Animals produce beauty on their bodies; humans can also produce it in their artifacts. The natural inference, then, would be that art is a human form of sexual display, a way for males to impress females with spectacularly redundant creations.
  • For Darwin, the human sense of beauty was not different in kind from the bird’s.
  • Still, Darwin recognized that the human sense of beauty was mediated by “complex ideas and trains of thought,” which make it impossible to explain in terms as straightforward as a bird’s:
  • Put more positively, one might say that any given work of art can be discussed critically and historically, but not deduced from the laws of evolution.
  • with the rise of evolutionary psychology, it was only a matter of time before the attempt was made to explain art in Darwinian terms. After all, if ethics and politics can be explained by game theory and reciprocal altruism, there is no reason why aesthetics should be different: in each case, what appears to be a realm of human autonomy can be reduced to the covert expression of biological imperatives
  • Still, there is an unmistakable sense in discussions of Darwinian aesthetics that by linking art to fitness, we can secure it against charges of irrelevance or frivolousness—that mattering to reproduction is what makes art, or anything, really matter.
  • The first popular effort in this direction was the late Denis Dutton’s much-discussed book The Art Instinct, which appeared in 2009.
  • Dutton’s Darwinism was aesthetically conservative: “Darwinian aesthetics,” he wrote, “can restore the vital place of beauty, skill, and pleasure as high artistic values.” Dutton’s argument has recently been reiterated and refined by a number of new books,
  • “The universality of art and artistic behaviors, their spontaneous appearance everywhere across the globe ... and the fact that in most cases they can be easily recognized as artistic across cultures suggest that they derive from a natural, innate source: a universal human psychology.”
  • Again like language, art is universal in the sense that any local expression of it can be “learned” by anyone.
  • Yet earlier theorists of evolution were reluctant to say that art was an evolutionary adaptation like language, for the simple reason that it does not appear to be evolutionarily adaptive.
  • Stephen Jay Gould suggested that art was not an evolutionary adaptation but what he called a “spandrel”—that is, a showy but accidental by-product of other adaptations that were truly functiona
  • the very words “success” and “failure,” despite themselves, bring an emotive and ethical dimension into the discussion, so impossible is it for human beings to inhabit a valueless world. In the nineteenth century, the idea that fitness for survival was a positive good motivated social Darwinism and eugenics. Proponents of these ideas thought that in some way they were serving progress by promoting the flourishing of the human race, when the basic premise of Darwinism is that there is no such thing as progress or regress, only differential rates of reproduction
  • In particular, Darwin suggests that it is impossible to explain the history or the conventions of any art by the general imperatives of evolution
  • Boyd begins with the premise that human beings are pattern-seeking animals: both our physical perceptions and our social interactions are determined by our brain’s innate need to find and to
  • Art, then, can be defined as the calisthenics of pattern-finding. “Just as animal physical play refines performance, flexibility, and efficiency in key behaviors,” Boyd writes, “so human art refines our performance in our key perceptual and cognitive modes, in sight (the visual arts), sound (music), and social cognition (story). These three modes of art, I propose, are adaptations ... they show evidence of special design in humans, design that offers survival and especially reproductive advantages.”
  • make coherent patterns
Javier E

They're Watching You at Work - Don Peck - The Atlantic - 2 views

  • Predictive statistical analysis, harnessed to big data, appears poised to alter the way millions of people are hired and assessed.
  • By one estimate, more than 98 percent of the world’s information is now stored digitally, and the volume of that data has quadrupled since 2007.
  • The application of predictive analytics to people’s careers—an emerging field sometimes called “people analytics”—is enormously challenging, not to mention ethically fraught
  • ...52 more annotations...
  • By the end of World War II, however, American corporations were facing severe talent shortages. Their senior executives were growing old, and a dearth of hiring from the Depression through the war had resulted in a shortfall of able, well-trained managers. Finding people who had the potential to rise quickly through the ranks became an overriding preoccupation of American businesses. They began to devise a formal hiring-and-management system based in part on new studies of human behavior, and in part on military techniques developed during both world wars, when huge mobilization efforts and mass casualties created the need to get the right people into the right roles as efficiently as possible. By the 1950s, it was not unusual for companies to spend days with young applicants for professional jobs, conducting a battery of tests, all with an eye toward corner-office potential.
  • But companies abandoned their hard-edged practices for another important reason: many of their methods of evaluation turned out not to be very scientific.
  • this regime, so widespread in corporate America at mid-century, had almost disappeared by 1990. “I think an HR person from the late 1970s would be stunned to see how casually companies hire now,”
  • Many factors explain the change, he said, and then he ticked off a number of them: Increased job-switching has made it less important and less economical for companies to test so thoroughly. A heightened focus on short-term financial results has led to deep cuts in corporate functions that bear fruit only in the long term. The Civil Rights Act of 1964, which exposed companies to legal liability for discriminatory hiring practices, has made HR departments wary of any broadly applied and clearly scored test that might later be shown to be systematically biased.
  • about a quarter of the country’s corporations were using similar tests to evaluate managers and junior executives, usually to assess whether they were ready for bigger roles.
  • He has encouraged the company’s HR executives to think about applying the games to the recruitment and evaluation of all professional workers.
  • Knack makes app-based video games, among them Dungeon Scrawl, a quest game requiring the player to navigate a maze and solve puzzles, and Wasabi Waiter, which involves delivering the right sushi to the right customer at an increasingly crowded happy hour. These games aren’t just for play: they’ve been designed by a team of neuroscientists, psychologists, and data scientists to suss out human potential. Play one of them for just 20 minutes, says Guy Halfteck, Knack’s founder, and you’ll generate several megabytes of data, exponentially more than what’s collected by the SAT or a personality test. How long you hesitate before taking every action, the sequence of actions you take, how you solve problems—all of these factors and many more are logged as you play, and then are used to analyze your creativity, your persistence, your capacity to learn quickly from mistakes, your ability to prioritize, and even your social intelligence and personality. The end result, Halfteck says, is a high-resolution portrait of your psyche and intellect, and an assessment of your potential as a leader or an innovator.
  • When the results came back, Haringa recalled, his heart began to beat a little faster. Without ever seeing the ideas, without meeting or interviewing the people who’d proposed them, without knowing their title or background or academic pedigree, Knack’s algorithm had identified the people whose ideas had panned out. The top 10 percent of the idea generators as predicted by Knack were in fact those who’d gone furthest in the process.
  • What Knack is doing, Haringa told me, “is almost like a paradigm shift.” It offers a way for his GameChanger unit to avoid wasting time on the 80 people out of 100—nearly all of whom look smart, well-trained, and plausible on paper—whose ideas just aren’t likely to work out.
  • Aptitude, skills, personal history, psychological stability, discretion, loyalty—companies at the time felt they had a need (and the right) to look into them all. That ambit is expanding once again, and this is undeniably unsettling. Should the ideas of scientists be dismissed because of the way they play a game? Should job candidates be ranked by what their Web habits say about them? Should the “data signature” of natural leaders play a role in promotion? These are all live questions today, and they prompt heavy concerns: that we will cede one of the most subtle and human of skills, the evaluation of the gifts and promise of other people, to machines; that the models will get it wrong; that some people will never get a shot in the new workforce.
  • scoring distance from work could violate equal-employment-opportunity standards. Marital status? Motherhood? Church membership? “Stuff like that,” Meyerle said, “we just don’t touch”—at least not in the U.S., where the legal environment is strict. Meyerle told me that Evolv has looked into these sorts of factors in its work for clients abroad, and that some of them produce “startling results.”
  • consider the alternative. A mountain of scholarly literature has shown that the intuitive way we now judge professional potential is rife with snap judgments and hidden biases, rooted in our upbringing or in deep neurological connections that doubtless served us well on the savanna but would seem to have less bearing on the world of work.
  • We may like to think that society has become more enlightened since those days, and in many ways it has, but our biases are mostly unconscious, and they can run surprisingly deep. Consider race. For a 2004 study called “Are Emily and Greg More Employable Than Lakisha and Jamal?,” the economists Sendhil Mullainathan and Marianne Bertrand put white-sounding names (Emily Walsh, Greg Baker) or black-sounding names (Lakisha Washington, Jamal Jones) on similar fictitious résumés, which they then sent out to a variety of companies in Boston and Chicago. To get the same number of callbacks, they learned, they needed to either send out half again as many résumés with black names as those with white names, or add eight extra years of relevant work experience to the résumés with black names.
  • a sociologist at Northwestern, spent parts of the three years from 2006 to 2008 interviewing professionals from elite investment banks, consultancies, and law firms about how they recruited, interviewed, and evaluated candidates, and concluded that among the most important factors driving their hiring recommendations were—wait for it—shared leisure interests.
  • Lacking “reliable predictors of future performance,” Rivera writes, “assessors purposefully used their own experiences as models of merit.” Former college athletes “typically prized participation in varsity sports above all other types of involvement.” People who’d majored in engineering gave engineers a leg up, believing they were better prepared.
  • the prevailing system of hiring and management in this country involves a level of dysfunction that should be inconceivable in an economy as sophisticated as ours. Recent survey data collected by the Corporate Executive Board, for example, indicate that nearly a quarter of all new hires leave their company within a year of their start date, and that hiring managers wish they’d never extended an offer to one out of every five members on their team
  • In the late 1990s, as these assessments shifted from paper to digital formats and proliferated, data scientists started doing massive tests of what makes for a successful customer-support technician or salesperson. This has unquestionably improved the quality of the workers at many firms.
  • In 2010, however, Xerox switched to an online evaluation that incorporates personality testing, cognitive-skill assessment, and multiple-choice questions about how the applicant would handle specific scenarios that he or she might encounter on the job. An algorithm behind the evaluation analyzes the responses, along with factual information gleaned from the candidate’s application, and spits out a color-coded rating: red (poor candidate), yellow (middling), or green (hire away). Those candidates who score best, I learned, tend to exhibit a creative but not overly inquisitive personality, and participate in at least one but not more than four social networks, among many other factors. (Previous experience, one of the few criteria that Xerox had explicitly screened for in the past, turns out to have no bearing on either productivity or retention
  • When Xerox started using the score in its hiring decisions, the quality of its hires immediately improved. The rate of attrition fell by 20 percent in the initial pilot period, and over time, the number of promotions rose. Xerox still interviews all candidates in person before deciding to hire them, Morse told me, but, she added, “We’re getting to the point where some of our hiring managers don’t even want to interview anymore”
  • Gone are the days, Ostberg told me, when, say, a small survey of college students would be used to predict the statistical validity of an evaluation tool. “We’ve got a data set of 347,000 actual employees who have gone through these different types of assessments or tools,” he told me, “and now we have performance-outcome data, and we can split those and slice and dice by industry and location.”
  • Evolv’s tests allow companies to capture data about everybody who applies for work, and everybody who gets hired—a complete data set from which sample bias, long a major vexation for industrial-organization psychologists, simply disappears. The sheer number of observations that this approach makes possible allows Evolv to say with precision which attributes matter more to the success of retail-sales workers (decisiveness, spatial orientation, persuasiveness) or customer-service personnel at call centers (rapport-building)
  • There are some data that Evolv simply won’t use, out of a concern that the information might lead to systematic bias against whole classes of people
  • the idea that hiring was a science fell out of favor. But now it’s coming back, thanks to new technologies and methods of analysis that are cheaper, faster, and much-wider-ranging than what we had before
  • what most excites him are the possibilities that arise from monitoring the entire life cycle of a worker at any given company.
  • Now the two companies are working together to marry pre-hire assessments to an increasing array of post-hire data: about not only performance and duration of service but also who trained the employees; who has managed them; whether they were promoted to a supervisory role, and how quickly; how they performed in that role; and why they eventually left.
  • What begins with an online screening test for entry-level workers ends with the transformation of nearly every aspect of hiring, performance assessment, and management.
  • I turned to Sandy Pentland, the director of the Human Dynamics Laboratory at MIT. In recent years, Pentland has pioneered the use of specialized electronic “badges” that transmit data about employees’ interactions as they go about their days. The badges capture all sorts of information about formal and informal conversations: their length; the tone of voice and gestures of the people involved; how much those people talk, listen, and interrupt; the degree to which they demonstrate empathy and extroversion; and more. Each badge generates about 100 data points a minute.
  • he tried the badges out on about 2,500 people, in 21 different organizations, and learned a number of interesting lessons. About a third of team performance, he discovered, can usually be predicted merely by the number of face-to-face exchanges among team members. (Too many is as much of a problem as too few.) Using data gathered by the badges, he was able to predict which teams would win a business-plan contest, and which workers would (rightly) say they’d had a “productive” or “creative” day. Not only that, but he claimed that his researchers had discovered the “data signature” of natural leaders, whom he called “charismatic connectors” and all of whom, he reported, circulate actively, give their time democratically to others, engage in brief but energetic conversations, and listen at least as much as they talk.
  • His group is developing apps to allow team members to view their own metrics more or less in real time, so that they can see, relative to the benchmarks of highly successful employees, whether they’re getting out of their offices enough, or listening enough, or spending enough time with people outside their own team.
  • Torrents of data are routinely collected by American companies and now sit on corporate servers, or in the cloud, awaiting analysis. Bloomberg reportedly logs every keystroke of every employee, along with their comings and goings in the office. The Las Vegas casino Harrah’s tracks the smiles of the card dealers and waitstaff on the floor (its analytics team has quantified the impact of smiling on customer satisfaction). E‑mail, of course, presents an especially rich vein to be mined for insights about our productivity, our treatment of co-workers, our willingness to collaborate or lend a hand, our patterns of written language, and what those patterns reveal about our intelligence, social skills, and behavior.
  • people analytics will ultimately have a vastly larger impact on the economy than the algorithms that now trade on Wall Street or figure out which ads to show us. He reminded me that we’ve witnessed this kind of transformation before in the history of management science. Near the turn of the 20th century, both Frederick Taylor and Henry Ford famously paced the factory floor with stopwatches, to improve worker efficiency.
  • “The quantities of data that those earlier generations were working with,” he said, “were infinitesimal compared to what’s available now. There’s been a real sea change in the past five years, where the quantities have just grown so large—petabytes, exabytes, zetta—that you start to be able to do things you never could before.”
  • People analytics will unquestionably provide many workers with more options and more power. Gild, for example, helps companies find undervalued software programmers, working indirectly to raise those people’s pay. Other companies are doing similar work. One called Entelo, for instance, specializes in using algorithms to identify potentially unhappy programmers who might be receptive to a phone cal
  • He sees it not only as a boon to a business’s productivity and overall health but also as an important new tool that individual employees can use for self-improvement: a sort of radically expanded The 7 Habits of Highly Effective People, custom-written for each of us, or at least each type of job, in the workforce.
  • the most exotic development in people analytics today is the creation of algorithms to assess the potential of all workers, across all companies, all the time.
  • The way Gild arrives at these scores is not simple. The company’s algorithms begin by scouring the Web for any and all open-source code, and for the coders who wrote it. They evaluate the code for its simplicity, elegance, documentation, and several other factors, including the frequency with which it’s been adopted by other programmers. For code that was written for paid projects, they look at completion times and other measures of productivity. Then they look at questions and answers on social forums such as Stack Overflow, a popular destination for programmers seeking advice on challenging projects. They consider how popular a given coder’s advice is, and how widely that advice ranges.
  • The algorithms go further still. They assess the way coders use language on social networks from LinkedIn to Twitter; the company has determined that certain phrases and words used in association with one another can distinguish expert programmers from less skilled ones. Gild knows these phrases and words are associated with good coding because it can correlate them with its evaluation of open-source code, and with the language and online behavior of programmers in good positions at prestigious companies.
  • having made those correlations, Gild can then score programmers who haven’t written open-source code at all, by analyzing the host of clues embedded in their online histories. They’re not all obvious, or easy to explain. Vivienne Ming, Gild’s chief scientist, told me that one solid predictor of strong coding is an affinity for a particular Japanese manga site.
  • Gild’s CEO, Sheeroy Desai, told me he believes his company’s approach can be applied to any occupation characterized by large, active online communities, where people post and cite individual work, ask and answer professional questions, and get feedback on projects. Graphic design is one field that the company is now looking at, and many scientific, technical, and engineering roles might also fit the bill. Regardless of their occupation, most people leave “data exhaust” in their wake, a kind of digital aura that can reveal a lot about a potential hire.
  • professionally relevant personality traits can be judged effectively merely by scanning Facebook feeds and photos. LinkedIn, of course, captures an enormous amount of professional data and network information, across just about every profession. A controversial start-up called Klout has made its mission the measurement and public scoring of people’s online social influence.
  • Mullainathan expressed amazement at how little most creative and professional workers (himself included) know about what makes them effective or ineffective in the office. Most of us can’t even say with any certainty how long we’ve spent gathering information for a given project, or our pattern of information-gathering, never mind know which parts of the pattern should be reinforced, and which jettisoned. As Mullainathan put it, we don’t know our own “production function.”
  • Over time, better job-matching technologies are likely to begin serving people directly, helping them see more clearly which jobs might suit them and which companies could use their skills. In the future, Gild plans to let programmers see their own profiles and take skills challenges to try to improve their scores. It intends to show them its estimates of their market value, too, and to recommend coursework that might allow them to raise their scores even more. Not least, it plans to make accessible the scores of typical hires at specific companies, so that software engineers can better see the profile they’d need to land a particular job
  • Knack, for its part, is making some of its video games available to anyone with a smartphone, so people can get a better sense of their strengths, and of the fields in which their strengths would be most valued. (Palo Alto High School recently adopted the games to help students assess careers.) Ultimately, the company hopes to act as matchmaker between a large network of people who play its games (or have ever played its games) and a widening roster of corporate clients, each with its own specific profile for any given type of job.
  • When I began my reporting for this story, I was worried that people analytics, if it worked at all, would only widen the divergent arcs of our professional lives, further gilding the path of the meritocratic elite from cradle to grave, and shutting out some workers more definitively. But I now believe the opposite is likely to happen, and that we’re headed toward a labor market that’s fairer to people at every stage of their careers
  • For decades, as we’ve assessed people’s potential in the professional workforce, the most important piece of data—the one that launches careers or keeps them grounded—has been educational background: typically, whether and where people went to college, and how they did there. Over the past couple of generations, colleges and universities have become the gatekeepers to a prosperous life. A degree has become a signal of intelligence and conscientiousness, one that grows stronger the more selective the school and the higher a student’s GPA, that is easily understood by employers, and that, until the advent of people analytics, was probably unrivaled in its predictive powers.
  • the limitations of that signal—the way it degrades with age, its overall imprecision, its many inherent biases, its extraordinary cost—are obvious. “Academic environments are artificial environments,” Laszlo Bock, Google’s senior vice president of people operations, told The New York Times in June. “People who succeed there are sort of finely trained, they’re conditioned to succeed in that environment,” which is often quite different from the workplace.
  • because one’s college history is such a crucial signal in our labor market, perfectly able people who simply couldn’t sit still in a classroom at the age of 16, or who didn’t have their act together at 18, or who chose not to go to graduate school at 22, routinely get left behind for good. That such early factors so profoundly affect career arcs and hiring decisions made two or three decades later is, on its face, absurd.
  • I spoke with managers at a lot of companies who are using advanced analytics to reevaluate and reshape their hiring, and nearly all of them told me that their research is leading them toward pools of candidates who didn’t attend college—for tech jobs, for high-end sales positions, for some managerial roles. In some limited cases, this is because their analytics revealed no benefit whatsoever to hiring people with college degrees; in other cases, and more often, it’s because they revealed signals that function far better than college history,
  • Google, too, is hiring a growing number of nongraduates. Many of the people I talked with reported that when it comes to high-paying and fast-track jobs, they’re reducing their preference for Ivy Leaguers and graduates of other highly selective schools.
  • This process is just beginning. Online courses are proliferating, and so are online markets that involve crowd-sourcing. Both arenas offer new opportunities for workers to build skills and showcase competence. Neither produces the kind of instantly recognizable signals of potential that a degree from a selective college, or a first job at a prestigious firm, might. That’s a problem for traditional hiring managers, because sifting through lots of small signals is so difficult and time-consuming.
  • all of these new developments raise philosophical questions. As professional performance becomes easier to measure and see, will we become slaves to our own status and potential, ever-focused on the metrics that tell us how and whether we are measuring up? Will too much knowledge about our limitations hinder achievement and stifle our dreams? All I can offer in response to these questions, ironically, is my own gut sense, which leads me to feel cautiously optimistic.
  • Google’s understanding of the promise of analytics is probably better than anybody else’s, and the company has been changing its hiring and management practices as a result of its ongoing analyses. (Brainteasers are no longer used in interviews, because they do not correlate with job success; GPA is not considered for anyone more than two years out of school, for the same reason—the list goes on.) But for all of Google’s technological enthusiasm, these same practices are still deeply human. A real, live person looks at every résumé the company receives. Hiring decisions are made by committee and are based in no small part on opinions formed during structured interviews.
Javier E

How Does Science Really Work? | The New Yorker - 1 views

  • I wanted to be a scientist. So why did I find the actual work of science so boring? In college science courses, I had occasional bursts of mind-expanding insight. For the most part, though, I was tortured by drudgery.
  • I’d found that science was two-faced: simultaneously thrilling and tedious, all-encompassing and narrow. And yet this was clearly an asset, not a flaw. Something about that combination had changed the world completely.
  • “Science is an alien thought form,” he writes; that’s why so many civilizations rose and fell before it was invented. In his view, we downplay its weirdness, perhaps because its success is so fundamental to our continued existence.
  • ...50 more annotations...
  • In school, one learns about “the scientific method”—usually a straightforward set of steps, along the lines of “ask a question, propose a hypothesis, perform an experiment, analyze the results.”
  • That method works in the classroom, where students are basically told what questions to pursue. But real scientists must come up with their own questions, finding new routes through a much vaster landscape.
  • Since science began, there has been disagreement about how those routes are charted. Two twentieth-century philosophers of science, Karl Popper and Thomas Kuhn, are widely held to have offered the best accounts of this process.
  • For Popper, Strevens writes, “scientific inquiry is essentially a process of disproof, and scientists are the disprovers, the debunkers, the destroyers.” Kuhn’s scientists, by contrast, are faddish true believers who promulgate received wisdom until they are forced to attempt a “paradigm shift”—a painful rethinking of their basic assumptions.
  • Working scientists tend to prefer Popper to Kuhn. But Strevens thinks that both theorists failed to capture what makes science historically distinctive and singularly effective.
  • Sometimes they seek to falsify theories, sometimes to prove them; sometimes they’re informed by preëxisting or contextual views, and at other times they try to rule narrowly, based on t
  • Why do scientists agree to this scheme? Why do some of the world’s most intelligent people sign on for a lifetime of pipetting?
  • Strevens thinks that they do it because they have no choice. They are constrained by a central regulation that governs science, which he calls the “iron rule of explanation.” The rule is simple: it tells scientists that, “if they are to participate in the scientific enterprise, they must uncover or generate new evidence to argue with”; from there, they must “conduct all disputes with reference to empirical evidence alone.”
  • , it is “the key to science’s success,” because it “channels hope, anger, envy, ambition, resentment—all the fires fuming in the human heart—to one end: the production of empirical evidence.”
  • Strevens arrives at the idea of the iron rule in a Popperian way: by disproving the other theories about how scientific knowledge is created.
  • The problem isn’t that Popper and Kuhn are completely wrong. It’s that scientists, as a group, don’t pursue any single intellectual strategy consistently.
  • Exploring a number of case studies—including the controversies over continental drift, spontaneous generation, and the theory of relativity—Strevens shows scientists exerting themselves intellectually in a variety of ways, as smart, ambitious people usually do.
  • “Science is boring,” Strevens writes. “Readers of popular science see the 1 percent: the intriguing phenomena, the provocative theories, the dramatic experimental refutations or verifications.” But, he says,behind these achievements . . . are long hours, days, months of tedious laboratory labor. The single greatest obstacle to successful science is the difficulty of persuading brilliant minds to give up the intellectual pleasures of continual speculation and debate, theorizing and arguing, and to turn instead to a life consisting almost entirely of the production of experimental data.
  • Ultimately, in fact, it was good that the geologists had a “splendid variety” of somewhat arbitrary opinions: progress in science requires partisans, because only they have “the motivation to perform years or even decades of necessary experimental work.” It’s just that these partisans must channel their energies into empirical observation. The iron rule, Strevens writes, “has a valuable by-product, and that by-product is data.”
  • Science is often described as “self-correcting”: it’s said that bad data and wrong conclusions are rooted out by other scientists, who present contrary findings. But Strevens thinks that the iron rule is often more important than overt correction.
  • Eddington was never really refuted. Other astronomers, driven by the iron rule, were already planning their own studies, and “the great preponderance of the resulting measurements fit Einsteinian physics better than Newtonian physics.” It’s partly by generating data on such a vast scale, Strevens argues, that the iron rule can power science’s knowledge machine: “Opinions converge not because bad data is corrected but because it is swamped.”
  • Why did the iron rule emerge when it did? Strevens takes us back to the Thirty Years’ War, which concluded with the Peace of Westphalia, in 1648. The war weakened religious loyalties and strengthened national ones.
  • Two regimes arose: in the spiritual realm, the will of God held sway, while in the civic one the decrees of the state were paramount. As Isaac Newton wrote, “The laws of God & the laws of man are to be kept distinct.” These new, “nonoverlapping spheres of obligation,” Strevens argues, were what made it possible to imagine the iron rule. The rule simply proposed the creation of a third sphere: in addition to God and state, there would now be science.
  • Strevens imagines how, to someone in Descartes’s time, the iron rule would have seemed “unreasonably closed-minded.” Since ancient Greece, it had been obvious that the best thinking was cross-disciplinary, capable of knitting together “poetry, music, drama, philosophy, democracy, mathematics,” and other elevating human disciplines.
  • We’re still accustomed to the idea that a truly flourishing intellect is a well-rounded one. And, by this standard, Strevens says, the iron rule looks like “an irrational way to inquire into the underlying structure of things”; it seems to demand the upsetting “suppression of human nature.”
  • Descartes, in short, would have had good reasons for resisting a law that narrowed the grounds of disputation, or that encouraged what Strevens describes as “doing rather than thinking.”
  • In fact, the iron rule offered scientists a more supple vision of progress. Before its arrival, intellectual life was conducted in grand gestures.
  • Descartes’s book was meant to be a complete overhaul of what had preceded it; its fate, had science not arisen, would have been replacement by some equally expansive system. The iron rule broke that pattern.
  • by authorizing what Strevens calls “shallow explanation,” the iron rule offered an empirical bridge across a conceptual chasm. Work could continue, and understanding could be acquired on the other side. In this way, shallowness was actually more powerful than depth.
  • it also changed what counted as progress. In the past, a theory about the world was deemed valid when it was complete—when God, light, muscles, plants, and the planets cohered. The iron rule allowed scientists to step away from the quest for completeness.
  • The consequences of this shift would become apparent only with time
  • In 1713, Isaac Newton appended a postscript to the second edition of his “Principia,” the treatise in which he first laid out the three laws of motion and the theory of universal gravitation. “I have not as yet been able to deduce from phenomena the reason for these properties of gravity, and I do not feign hypotheses,” he wrote. “It is enough that gravity really exists and acts according to the laws that we have set forth.”
  • What mattered, to Newton and his contemporaries, was his theory’s empirical, predictive power—that it was “sufficient to explain all the motions of the heavenly bodies and of our sea.”
  • Descartes would have found this attitude ridiculous. He had been playing a deep game—trying to explain, at a fundamental level, how the universe fit together. Newton, by those lights, had failed to explain anything: he himself admitted that he had no sense of how gravity did its work
  • Strevens sees its earliest expression in Francis Bacon’s “The New Organon,” a foundational text of the Scientific Revolution, published in 1620. Bacon argued that thinkers must set aside their “idols,” relying, instead, only on evidence they could verify. This dictum gave scientists a new way of responding to one another’s work: gathering data.
  • Quantum theory—which tells us that subatomic particles can be “entangled” across vast distances, and in multiple places at the same time—makes intuitive sense to pretty much nobody.
  • Without the iron rule, Strevens writes, physicists confronted with such a theory would have found themselves at an impasse. They would have argued endlessly about quantum metaphysics.
  • ollowing the iron rule, they can make progress empirically even though they are uncertain conceptually. Individual researchers still passionately disagree about what quantum theory means. But that hasn’t stopped them from using it for practical purposes—computer chips, MRI machines, G.P.S. networks, and other technologies rely on quantum physics.
  • One group of theorists, the rationalists, has argued that science is a new way of thinking, and that the scientist is a new kind of thinker—dispassionate to an uncommon degree.
  • As evidence against this view, another group, the subjectivists, points out that scientists are as hopelessly biased as the rest of us. To this group, the aloofness of science is a smoke screen behind which the inevitable emotions and ideologies hide.
  • At least in science, Strevens tells us, “the appearance of objectivity” has turned out to be “as important as the real thing.”
  • The subjectivists are right, he admits, inasmuch as scientists are regular people with a “need to win” and a “determination to come out on top.”
  • But they are wrong to think that subjectivity compromises the scientific enterprise. On the contrary, once subjectivity is channelled by the iron rule, it becomes a vital component of the knowledge machine. It’s this redirected subjectivity—to come out on top, you must follow the iron rule!—that solves science’s “problem of motivation,” giving scientists no choice but “to pursue a single experiment relentlessly, to the last measurable digit, when that digit might be quite meaningless.”
  • If it really was a speech code that instigated “the extraordinary attention to process and detail that makes science the supreme discriminator and destroyer of false ideas,” then the peculiar rigidity of scientific writing—Strevens describes it as “sterilized”—isn’t a symptom of the scientific mind-set but its cause.
  • The iron rule—“a kind of speech code”—simply created a new way of communicating, and it’s this new way of communicating that created science.
  • Other theorists have explained science by charting a sweeping revolution in the human mind; inevitably, they’ve become mired in a long-running debate about how objective scientists really are
  • In “The Knowledge Machine: How Irrationality Created Modern Science” (Liveright), Michael Strevens, a philosopher at New York University, aims to identify that special something. Strevens is a philosopher of science
  • Compared with the theories proposed by Popper and Kuhn, Strevens’s rule can feel obvious and underpowered. That’s because it isn’t intellectual but procedural. “The iron rule is focused not on what scientists think,” he writes, “but on what arguments they can make in their official communications.”
  • Like everybody else, scientists view questions through the lenses of taste, personality, affiliation, and experience
  • geologists had a professional obligation to take sides. Europeans, Strevens reports, tended to back Wegener, who was German, while scholars in the United States often preferred Simpson, who was American. Outsiders to the field were often more receptive to the concept of continental drift than established scientists, who considered its incompleteness a fatal flaw.
  • Strevens’s point isn’t that these scientists were doing anything wrong. If they had biases and perspectives, he writes, “that’s how human thinking works.”
  • Eddington’s observations were expected to either confirm or falsify Einstein’s theory of general relativity, which predicted that the sun’s gravity would bend the path of light, subtly shifting the stellar pattern. For reasons having to do with weather and equipment, the evidence collected by Eddington—and by his colleague Frank Dyson, who had taken similar photographs in Sobral, Brazil—was inconclusive; some of their images were blurry, and so failed to resolve the matter definitively.
  • it was only natural for intelligent people who were free of the rule’s strictures to attempt a kind of holistic, systematic inquiry that was, in many ways, more demanding. It never occurred to them to ask if they might illuminate more collectively by thinking about less individually.
  • In the single-sphered, pre-scientific world, thinkers tended to inquire into everything at once. Often, they arrived at conclusions about nature that were fascinating, visionary, and wrong.
  • How Does Science Really Work?Science is objective. Scientists are not. Can an “iron rule” explain how they’ve changed the world anyway?By Joshua RothmanSeptember 28, 2020
Javier E

A Leading Memory Researcher Explains How to Make Precious Moments Last - The New York T... - 0 views

  • Our memories form the bedrock of who we are. Those recollections, in turn, are built on one very simple assumption: This happened. But things are not quite so simple
  • “We update our memories through the act of remembering,” says Charan Ranganath, a professor of psychology and neuroscience at the University of California, Davis, and the author of the illuminating new book “Why We Remember.” “So it creates all these weird biases and infiltrates our decision making. It affects our sense of who we are.
  • Rather than being photo-accurate repositories of past experience, Ranganath argues, our memories function more like active interpreters, working to help us navigate the present and future. The implication is that who we are, and the memories we draw on to determine that, are far less fixed than you might think. “Our identities,” Ranganath says, “are built on shifting sand.”
  • ...24 more annotations...
  • People believe that memory should be effortless, but their expectations for how much they should remember are totally out of whack with how much they’re capable of remembering.1
  • What is the most common misconception about memory?
  • Another misconception is that memory is supposed to be an archive of the past. We expect that we should be able to replay the past like a movie in our heads.
  • we don’t replay the past as it happened; we do it through a lens of interpretation and imagination.
  • How much are we capable of remembering, from both an episodic2 2 Episodic memory is the term for the memory of life experiences. and a semantic3 3 Semantic memory is the term for the memory of facts and knowledge about the world. standpoint?
  • I would argue that we’re all everyday-memory experts, because we have this exceptional semantic memory, which is the scaffold for episodic memory.
  • If what we’re remembering, or the emotional tenor of what we’re remembering, is dictated by how we’re thinking in a present moment, what can we really say about the truth of a memory?
  • But if memories are malleable, what are the implications for how we understand our “true” selves?
  • your question gets to a major purpose of memory, which is to give us an illusion of stability in a world that is always changing. Because if we look for memories, we’ll reshape them into our beliefs of what’s happening right now. We’ll be biased in terms of how we sample the past. We have these illusions of stability, but we are always changing
  • And depending on what memories we draw upon, those life narratives can change.
  • I know it sounds squirmy to say, “Well, I can’t answer the question of how much we remember,” but I don’t want readers to walk away thinking memory is all made up.
  • One thing that makes the human brain so sophisticated is that we have a longer timeline in which we can integrate information than many other species. That gives us the ability to say: “Hey, I’m walking up and giving money to the cashier at the cafe. The barista is going to hand me a cup of coffee in about a minute or two.”
  • There is this illusion that we know exactly what’s going to happen, but the fact is we don’t. Memory can overdo it: Somebody lied to us once, so they are a liar; somebody shoplifted once, they are a thief.
  • If people have a vivid memory of something that sticks out, that will overshadow all their knowledge about the way things work. So there’s kind of an illus
  • we have this illusion that much of the world is cause and effect. But the reason, in my opinion, that we have that illusion is that our brain is constantly trying to find the patterns
  • I think of memory more like a painting than a photograph. There’s often photorealistic aspects of a painting, but there’s also interpretation. As a painter evolves, they could revisit the same subject over and over and paint differently based on who they are now. We’re capable of remembering things in extraordinary detail, but we infuse meaning into what we remember. We’re designed to extract meaning from the past, and that meaning should have truth in it. But it also has knowledge and imagination and, sometimes, wisdom.
  • memory, often, is educated guesses by the brain about what’s important. So what’s important? Things that are scary, things that get your desire going, things that are surprising. Maybe you were attracted to this person, and your eyes dilated, your pulse went up. Maybe you were working on something in this high state of excitement, and your dopamine was up.
  • It could be any of those things, but they’re all important in some way, because if you’re a brain, you want to take what’s surprising, you want to take what’s motivationally important for survival, what’s new.
  • On the more intentional side, are there things that we might be able to do in the moment to make events last in our memories? In some sense, it’s about being mindful. If we want to form a new memory, focus on aspects of the experience you want to take with you.
  • If you’re with your kid, you’re at a park, focus on the parts of it that are great, not the parts that are kind of annoying. Then you want to focus on the sights, the sounds, the smells, because those will give you rich detail later on
  • Another part of it, too, is that we kill ourselves by inducing distractions in our world. We have alerts on our phones. We check email habitually.
  • When we go on trips, I take candid shots. These are the things that bring you back to moments. If you capture the feelings and the sights and the sounds that bring you to the moment, as opposed to the facts of what happened, that is a huge part of getting the best of memory.
  • this goes back to the question of whether the factual truth of a memory matters to how we interpret it. I think it matters to have some truth, but then again, many of the truths we cling to depend on our own perspective.
  • There’s a great experiment on this. These researchers had people read this story about a house.8 8 The study was “Recall of Previously Unrecallable Information Following a Shift in Perspective,” by Richard C. Anderson and James W. Pichert. One group of subjects is told, I want you to read this story from the perspective of a prospective home buyer. When they remember it, they remember all the features of the house that are described in the thing. Another group is told, I want you to remember this from the perspective of a burglar. Those people tend to remember the valuables in the house and things that you would want to take. But what was interesting was then they switched the groups around. All of a sudden, people could pull up a number of details that they didn’t pull up before. It was always there, but they just didn’t approach it from that mind-set. So we do have a lot of information that we can get if we change our perspective, and this ability to change our perspective is exceptionally important for being accurate. It’s exceptionally important for being able to grow and modify our beliefs
Javier E

Scientists identify vast underground ecosystem containing billions of micro-organisms |... - 0 views

  • Despite extreme heat, no light, minuscule nutrition and intense pressure, scientists estimate this subterranean biosphere is teeming with between 15bn and 23bn tonnes of micro-organisms, hundreds of times the combined weight of every human on the planet.
  • Researchers at the Deep Carbon Observatory say the diversity of underworld species bears comparison to the Amazon or the Galápagos Islands, but unlike those places the environment is still largely pristine because people have yet to probe most of the subsurface.
  • “It’s like finding a whole new reservoir of life on Earth,” said Karen Lloyd, an associate professor at the University of Tennessee in Knoxville. “We are discovering new types of life all the time. So much of life is within the Earth rather than on top of it.
  • ...9 more annotations...
  • The team combines 1,200 scientists from 52 countries in disciplines ranging from geology and microbiology to chemistry and physics.
  • The results suggest 70% of Earth’s bacteria and archaea exist in the subsurface
  • “The strangest thing for me is that some organisms can exist for millennia. They are metabolically active but in stasis, with less energy than we thought possible of supporting life.”
  • Some microorganisms have been alive for thousands of years, barely moving except with shifts in the tectonic plates, earthquakes or eruption
  • these organisms are part of slow, persistent cycles on geological timescales.”
  • Underworld biospheres vary depending on geology and geography. Their combined size is estimated to be more than 2bn cubic kilometres, but this could be expanded further in the future
  • The researchers said their discoveries were made possible by two technical advances: drills that can penetrate far deeper below the Earth’s crust, and improvements in microscopes that allow life to be detected at increasingly minute levels.
  • The scientists have been trying to find a lower limit beyond which life cannot exist, but the deeper they dig the more life they find. There is a temperature maximum – currently 122C – but the researchers believe this record will be broken if they keep exploring and developing more sophisticated instruments.
  • Robert Hazen, a mineralogist at the Carnegie Institution for Science, said: “We must ask ourselves: if life on Earth can be this different from what experience has led us to expect, then what strangeness might await as we probe for life on other worlds?”
sissij

Is Technology Destroying Happiness? | Big Think - 0 views

  • Historian Yuval Noah Harari writes that happiness itself is not an inalienable right—the pursuit of it is. Semantics matter. 
  • Technology alone is not to blame, as in many ways our uneasiness with our condition seems an old trait.
  • however, Americans have turned more toward British philosopher Jeremy Bentham’s demand that the sole purpose of the state, financial markets, and science “is to increase global happiness.” 
  • ...5 more annotations...
  • But we’re not happier.
  • Harari points out that initially happiness was introduced as a check on state power.
  • Given how few we encounter on a regular basis, this threat detection system has been co-opted by the luxury of security, causing Harari to realize that: The most common reaction of the human mind to achievement is not satisfaction, but craving for more.
  • The Buddha actually warned against such a chase, deciding contentment was more worthwhile (and saner). The pursuit of pleasure is the root of suffering.
  • As Trentmann points out, the Latin consumere meant a “physical exhaustion of matter.” For example, the wasting disease, tuberculosis, was referred to as consumption. While we’ve redefined that word in modern times the original intention seems destined to win out.
  •  
    The author went depth into how people see happiness. It is very interesting to see that our happiness in life does not increase as our life quality. There are ,any other reasons why we don't feel as happy as we should. The author bring up a very interesting point that happiness is very blank and hard to define, but the pursuit of happiness is what feels real to us and important to us. I think we feel less happiness is because we become easier to get what we want. The process of pursuing the happiness is lightened. --Sissi (3/9/2017)
Javier E

Why the Past 10 Years of American Life Have Been Uniquely Stupid - The Atlantic - 0 views

  • Social scientists have identified at least three major forces that collectively bind together successful democracies: social capital (extensive social networks with high levels of trust), strong institutions, and shared stories.
  • Social media has weakened all three.
  • gradually, social-media users became more comfortable sharing intimate details of their lives with strangers and corporations. As I wrote in a 2019 Atlantic article with Tobias Rose-Stockwell, they became more adept at putting on performances and managing their personal brand—activities that might impress others but that do not deepen friendships in the way that a private phone conversation will.
  • ...118 more annotations...
  • the stage was set for the major transformation, which began in 2009: the intensification of viral dynamics.
  • Before 2009, Facebook had given users a simple timeline––a never-ending stream of content generated by their friends and connections, with the newest posts at the top and the oldest ones at the bottom
  • That began to change in 2009, when Facebook offered users a way to publicly “like” posts with the click of a button. That same year, Twitter introduced something even more powerful: the “Retweet” button, which allowed users to publicly endorse a post while also sharing it with all of their followers.
  • “Like” and “Share” buttons quickly became standard features of most other platforms.
  • Facebook developed algorithms to bring each user the content most likely to generate a “like” or some other interaction, eventually including the “share” as well.
  • Later research showed that posts that trigger emotions––especially anger at out-groups––are the most likely to be shared.
  • By 2013, social media had become a new game, with dynamics unlike those in 2008. If you were skillful or lucky, you might create a post that would “go viral” and make you “internet famous”
  • If you blundered, you could find yourself buried in hateful comments. Your posts rode to fame or ignominy based on the clicks of thousands of strangers, and you in turn contributed thousands of clicks to the game.
  • This new game encouraged dishonesty and mob dynamics: Users were guided not just by their true preferences but by their past experiences of reward and punishment,
  • As a social psychologist who studies emotion, morality, and politics, I saw this happening too. The newly tweaked platforms were almost perfectly designed to bring out our most moralistic and least reflective selves. The volume of outrage was shocking.
  • It was just this kind of twitchy and explosive spread of anger that James Madison had tried to protect us from as he was drafting the U.S. Constitution.
  • The Framers of the Constitution were excellent social psychologists. They knew that democracy had an Achilles’ heel because it depended on the collective judgment of the people, and democratic communities are subject to “the turbulency and weakness of unruly passions.”
  • The key to designing a sustainable republic, therefore, was to build in mechanisms to slow things down, cool passions, require compromise, and give leaders some insulation from the mania of the moment while still holding them accountable to the people periodically, on Election Day.
  • The tech companies that enhanced virality from 2009 to 2012 brought us deep into Madison’s nightmare.
  • a less quoted yet equally important insight, about democracy’s vulnerability to triviality.
  • Madison notes that people are so prone to factionalism that “where no substantial occasion presents itself, the most frivolous and fanciful distinctions have been sufficient to kindle their unfriendly passions and excite their most violent conflicts.”
  • Social media has both magnified and weaponized the frivolous.
  • It’s not just the waste of time and scarce attention that matters; it’s the continual chipping-away of trust.
  • a democracy depends on widely internalized acceptance of the legitimacy of rules, norms, and institutions.
  • when citizens lose trust in elected leaders, health authorities, the courts, the police, universities, and the integrity of elections, then every decision becomes contested; every election becomes a life-and-death struggle to save the country from the other side
  • The most recent Edelman Trust Barometer (an international measure of citizens’ trust in government, business, media, and nongovernmental organizations) showed stable and competent autocracies (China and the United Arab Emirates) at the top of the list, while contentious democracies such as the United States, the United Kingdom, Spain, and South Korea scored near the bottom (albeit above Russia).
  • The literature is complex—some studies show benefits, particularly in less developed democracies—but the review found that, on balance, social media amplifies political polarization; foments populism, especially right-wing populism; and is associated with the spread of misinformation.
  • When people lose trust in institutions, they lose trust in the stories told by those institutions. That’s particularly true of the institutions entrusted with the education of children.
  • Facebook and Twitter make it possible for parents to become outraged every day over a new snippet from their children’s history lessons––and math lessons and literature selections, and any new pedagogical shifts anywhere in the country
  • The motives of teachers and administrators come into question, and overreaching laws or curricular reforms sometimes follow, dumbing down education and reducing trust in it further.
  • young people educated in the post-Babel era are less likely to arrive at a coherent story of who we are as a people, and less likely to share any such story with those who attended different schools or who were educated in a different decade.
  • former CIA analyst Martin Gurri predicted these fracturing effects in his 2014 book, The Revolt of the Public. Gurri’s analysis focused on the authority-subverting effects of information’s exponential growth, beginning with the internet in the 1990s. Writing nearly a decade ago, Gurri could already see the power of social media as a universal solvent, breaking down bonds and weakening institutions everywhere it reached.
  • he notes a constructive feature of the pre-digital era: a single “mass audience,” all consuming the same content, as if they were all looking into the same gigantic mirror at the reflection of their own society. I
  • The digital revolution has shattered that mirror, and now the public inhabits those broken pieces of glass. So the public isn’t one thing; it’s highly fragmented, and it’s basically mutually hostile
  • Facebook, Twitter, YouTube, and a few other large platforms unwittingly dissolved the mortar of trust, belief in institutions, and shared stories that had held a large and diverse secular democracy together.
  • I think we can date the fall of the tower to the years between 2011 (Gurri’s focal year of “nihilistic” protests) and 2015, a year marked by the “great awokening” on the left and the ascendancy of Donald Trump on the right.
  • Twitter can overpower all the newspapers in the country, and stories cannot be shared (or at least trusted) across more than a few adjacent fragments—so truth cannot achieve widespread adherence.
  • fter Babel, nothing really means anything anymore––at least not in a way that is durable and on which people widely agree.
  • Politics After Babel
  • “Politics is the art of the possible,” the German statesman Otto von Bismarck said in 1867. In a post-Babel democracy, not much may be possible.
  • The ideological distance between the two parties began increasing faster in the 1990s. Fox News and the 1994 “Republican Revolution” converted the GOP into a more combative party.
  • So cross-party relationships were already strained before 2009. But the enhanced virality of social media thereafter made it more hazardous to be seen fraternizing with the enemy or even failing to attack the enemy with sufficient vigor.
  • What changed in the 2010s? Let’s revisit that Twitter engineer’s metaphor of handing a loaded gun to a 4-year-old. A mean tweet doesn’t kill anyone; it is an attempt to shame or punish someone publicly while broadcasting one’s own virtue, brilliance, or tribal loyalties. It’s more a dart than a bullet
  • from 2009 to 2012, Facebook and Twitter passed out roughly 1 billion dart guns globally. We’ve been shooting one another ever since.
  • “devoted conservatives,” comprised 6 percent of the U.S. population.
  • the warped “accountability” of social media has also brought injustice—and political dysfunction—in three ways.
  • First, the dart guns of social media give more power to trolls and provocateurs while silencing good citizens.
  • a small subset of people on social-media platforms are highly concerned with gaining status and are willing to use aggression to do so.
  • Across eight studies, Bor and Petersen found that being online did not make most people more aggressive or hostile; rather, it allowed a small number of aggressive people to attack a much larger set of victims. Even a small number of jerks were able to dominate discussion forums,
  • Additional research finds that women and Black people are harassed disproportionately, so the digital public square is less welcoming to their voices.
  • Second, the dart guns of social media give more power and voice to the political extremes while reducing the power and voice of the moderate majority.
  • The “Hidden Tribes” study, by the pro-democracy group More in Common, surveyed 8,000 Americans in 2017 and 2018 and identified seven groups that shared beliefs and behaviors.
  • Social media has given voice to some people who had little previously, and it has made it easier to hold powerful people accountable for their misdeeds
  • The group furthest to the left, the “progressive activists,” comprised 8 percent of the population. The progressive activists were by far the most prolific group on social media: 70 percent had shared political content over the previous year. The devoted conservatives followed, at 56 percent.
  • These two extreme groups are similar in surprising ways. They are the whitest and richest of the seven groups, which suggests that America is being torn apart by a battle between two subsets of the elite who are not representative of the broader society.
  • they are the two groups that show the greatest homogeneity in their moral and political attitudes.
  • likely a result of thought-policing on social media:
  • political extremists don’t just shoot darts at their enemies; they spend a lot of their ammunition targeting dissenters or nuanced thinkers on their own team.
  • Finally, by giving everyone a dart gun, social media deputizes everyone to administer justice with no due process. Platforms like Twitter devolve into the Wild West, with no accountability for vigilantes.
  • Enhanced-virality platforms thereby facilitate massive collective punishment for small or imagined offenses, with real-world consequences, including innocent people losing their jobs and being shamed into suicide
  • we don’t get justice and inclusion; we get a society that ignores context, proportionality, mercy, and truth.
  • Since the tower fell, debates of all kinds have grown more and more confused. The most pervasive obstacle to good thinking is confirmation bias, which refers to the human tendency to search only for evidence that confirms our preferred beliefs
  • search engines were supercharging confirmation bias, making it far easier for people to find evidence for absurd beliefs and conspiracy theorie
  • The most reliable cure for confirmation bias is interaction with people who don’t share your beliefs. They confront you with counterevidence and counterargument.
  • In his book The Constitution of Knowledge, Jonathan Rauch describes the historical breakthrough in which Western societies developed an “epistemic operating system”—that is, a set of institutions for generating knowledge from the interactions of biased and cognitively flawed individuals
  • English law developed the adversarial system so that biased advocates could present both sides of a case to an impartial jury.
  • Newspapers full of lies evolved into professional journalistic enterprises, with norms that required seeking out multiple sides of a story, followed by editorial review, followed by fact-checking.
  • Universities evolved from cloistered medieval institutions into research powerhouses, creating a structure in which scholars put forth evidence-backed claims with the knowledge that other scholars around the world would be motivated to gain prestige by finding contrary evidence.
  • Part of America’s greatness in the 20th century came from having developed the most capable, vibrant, and productive network of knowledge-producing institutions in all of human history
  • But this arrangement, Rauch notes, “is not self-maintaining; it relies on an array of sometimes delicate social settings and understandings, and those need to be understood, affirmed, and protected.”
  • This, I believe, is what happened to many of America’s key institutions in the mid-to-late 2010s. They got stupider en masse because social media instilled in their members a chronic fear of getting darted
  • it was so pervasive that it established new behavioral norms backed by new policies seemingly overnight
  • Participants in our key institutions began self-censoring to an unhealthy degree, holding back critiques of policies and ideas—even those presented in class by their students—that they believed to be ill-supported or wrong.
  • The stupefying process plays out differently on the right and the left because their activist wings subscribe to different narratives with different sacred values.
  • The “Hidden Tribes” study tells us that the “devoted conservatives” score highest on beliefs related to authoritarianism. They share a narrative in which America is eternally under threat from enemies outside and subversives within; they see life as a battle between patriots and traitors.
  • they are psychologically different from the larger group of “traditional conservatives” (19 percent of the population), who emphasize order, decorum, and slow rather than radical change.
  • The traditional punishment for treason is death, hence the battle cry on January 6: “Hang Mike Pence.”
  • Right-wing death threats, many delivered by anonymous accounts, are proving effective in cowing traditional conservatives
  • The wave of threats delivered to dissenting Republican members of Congress has similarly pushed many of the remaining moderates to quit or go silent, giving us a party ever more divorced from the conservative tradition, constitutional responsibility, and reality.
  • The stupidity on the right is most visible in the many conspiracy theories spreading across right-wing media and now into Congress.
  • The Democrats have also been hit hard by structural stupidity, though in a different way. In the Democratic Party, the struggle between the progressive wing and the more moderate factions is open and ongoing, and often the moderates win.
  • The problem is that the left controls the commanding heights of the culture: universities, news organizations, Hollywood, art museums, advertising, much of Silicon Valley, and the teachers’ unions and teaching colleges that shape K–12 education. And in many of those institutions, dissent has been stifled:
  • Liberals in the late 20th century shared a belief that the sociologist Christian Smith called the “liberal progress” narrative, in which America used to be horrifically unjust and repressive, but, thanks to the struggles of activists and heroes, has made (and continues to make) progress toward realizing the noble promise of its founding.
  • It is also the view of the “traditional liberals” in the “Hidden Tribes” study (11 percent of the population), who have strong humanitarian values, are older than average, and are largely the people leading America’s cultural and intellectual institutions.
  • when the newly viralized social-media platforms gave everyone a dart gun, it was younger progressive activists who did the most shooting, and they aimed a disproportionate number of their darts at these older liberal leaders.
  • Confused and fearful, the leaders rarely challenged the activists or their nonliberal narrative in which life at every institution is an eternal battle among identity groups over a zero-sum pie, and the people on top got there by oppressing the people on the bottom. This new narrative is rigidly egalitarian––focused on equality of outcomes, not of rights or opportunities. It is unconcerned with individual rights.
  • The universal charge against people who disagree with this narrative is not “traitor”; it is “racist,” “transphobe,” “Karen,” or some related scarlet letter marking the perpetrator as one who hates or harms a marginalized group.
  • The punishment that feels right for such crimes is not execution; it is public shaming and social death.
  • anyone on Twitter had already seen dozens of examples teaching the basic lesson: Don’t question your own side’s beliefs, policies, or actions. And when traditional liberals go silent, as so many did in the summer of 2020, the progressive activists’ more radical narrative takes over as the governing narrative of an organization.
  • This is why so many epistemic institutions seemed to “go woke” in rapid succession that year and the next, beginning with a wave of controversies and resignations at The New York Times and other newspapers, and continuing on to social-justice pronouncements by groups of doctors and medical associations
  • The problem is structural. Thanks to enhanced-virality social media, dissent is punished within many of our institutions, which means that bad ideas get elevated into official policy.
  • In a 2018 interview, Steve Bannon, the former adviser to Donald Trump, said that the way to deal with the media is “to flood the zone with shit.” He was describing the “firehose of falsehood” tactic pioneered by Russian disinformation programs to keep Americans confused, disoriented, and angry.
  • artificial intelligence is close to enabling the limitless spread of highly believable disinformation. The AI program GPT-3 is already so good that you can give it a topic and a tone and it will spit out as many essays as you like, typically with perfect grammar and a surprising level of coherence.
  • Renée DiResta, the research manager at the Stanford Internet Observatory, explained that spreading falsehoods—whether through text, images, or deep-fake videos—will quickly become inconceivably easy. (She co-wrote the essay with GPT-3.)
  • American factions won’t be the only ones using AI and social media to generate attack content; our adversaries will too.
  • In the 20th century, America’s shared identity as the country leading the fight to make the world safe for democracy was a strong force that helped keep the culture and the polity together.
  • In the 21st century, America’s tech companies have rewired the world and created products that now appear to be corrosive to democracy, obstacles to shared understanding, and destroyers of the modern tower.
  • What changes are needed?
  • I can suggest three categories of reforms––three goals that must be achieved if democracy is to remain viable in the post-Babel era.
  • We must harden democratic institutions so that they can withstand chronic anger and mistrust, reform social media so that it becomes less socially corrosive, and better prepare the next generation for democratic citizenship in this new age.
  • Harden Democratic Institutions
  • we must reform key institutions so that they can continue to function even if levels of anger, misinformation, and violence increase far above those we have today.
  • Reforms should reduce the outsize influence of angry extremists and make legislators more responsive to the average voter in their district.
  • One example of such a reform is to end closed party primaries, replacing them with a single, nonpartisan, open primary from which the top several candidates advance to a general election that also uses ranked-choice voting
  • A second way to harden democratic institutions is to reduce the power of either political party to game the system in its favor, for example by drawing its preferred electoral districts or selecting the officials who will supervise elections
  • These jobs should all be done in a nonpartisan way.
  • Reform Social Media
  • Social media’s empowerment of the far left, the far right, domestic trolls, and foreign agents is creating a system that looks less like democracy and more like rule by the most aggressive.
  • it is within our power to reduce social media’s ability to dissolve trust and foment structural stupidity. Reforms should limit the platforms’ amplification of the aggressive fringes while giving more voice to what More in Common calls “the exhausted majority.”
  • the main problem with social media is not that some people post fake or toxic stuff; it’s that fake and outrage-inducing content can now attain a level of reach and influence that was not possible before
  • Perhaps the biggest single change that would reduce the toxicity of existing platforms would be user verification as a precondition for gaining the algorithmic amplification that social media offers.
  • One of the first orders of business should be compelling the platforms to share their data and their algorithms with academic researchers.
  • Prepare the Next Generation
  • Childhood has become more tightly circumscribed in recent generations––with less opportunity for free, unstructured play; less unsupervised time outside; more time online. Whatever else the effects of these shifts, they have likely impeded the development of abilities needed for effective self-governance for many young adults
  • Depression makes people less likely to want to engage with new people, ideas, and experiences. Anxiety makes new things seem more threatening. As these conditions have risen and as the lessons on nuanced social behavior learned through free play have been delayed, tolerance for diverse viewpoints and the ability to work out disputes have diminished among many young people
  • Students did not just say that they disagreed with visiting speakers; some said that those lectures would be dangerous, emotionally devastating, a form of violence. Because rates of teen depression and anxiety have continued to rise into the 2020s, we should expect these views to continue in the generations to follow, and indeed to become more severe.
  • The most important change we can make to reduce the damaging effects of social media on children is to delay entry until they have passed through puberty.
  • The age should be raised to at least 16, and companies should be held responsible for enforcing it.
  • et them out to play. Stop starving children of the experiences they most need to become good citizens: free play in mixed-age groups of children with minimal adult supervision
  • while social media has eroded the art of association throughout society, it may be leaving its deepest and most enduring marks on adolescents. A surge in rates of anxiety, depression, and self-harm among American teens began suddenly in the early 2010s. (The same thing happened to Canadian and British teens, at the same time.) The cause is not known, but the timing points to social media as a substantial contributor—the surge began just as the large majority of American teens became daily users of the major platforms.
  • What would it be like to live in Babel in the days after its destruction? We know. It is a time of confusion and loss. But it is also a time to reflect, listen, and build.
  • In recent years, Americans have started hundreds of groups and organizations dedicated to building trust and friendship across the political divide, including BridgeUSA, Braver Angels (on whose board I serve), and many others listed at BridgeAlliance.us. We cannot expect Congress and the tech companies to save us. We must change ourselves and our communities.
  • when we look away from our dysfunctional federal government, disconnect from social media, and talk with our neighbors directly, things seem more hopeful. Most Americans in the More in Common report are members of the “exhausted majority,” which is tired of the fighting and is willing to listen to the other side and compromise. Most Americans now see that social media is having a negative impact on the country, and are becoming more aware of its damaging effects on children.
Javier E

We need a major redesign of life - The Washington Post - 0 views

  • Thirty years were added to average life expectancy in the 20th century, and rather than imagine the scores of ways we could use these years to improve quality of life, we tacked them all on at the end. Only old age got longer.
  • As a result, most people are anxious about the prospect of living for a century. Asked about aspirations for living to 100, typical responses are “I hope I don’t outlive my money” or “I hope I don’t get dementia.”
  • Each generation is born into a world prepared by its ancestors with knowledge, infrastructure and social norms. The human capacity to benefit from this inherited culture afforded us such extraordinary advantages that premature death was dramatically reduced in a matter of decades. Yet as longevity surged, culture didn’t keep up
  • ...11 more annotations...
  • Long lives are not the problem. The problem is living in cultures designed for lives half as long as the ones we have.
  • Retirements that span four decades are unattainable for most individuals and governments; education that ends in the early 20s is ill-suited for longer working lives; and social norms that dictate intergenerational responsibilities between parents and young children fail to address families that include four or five living generations.
  • Last year, the Stanford Center on Longevity launched an initiative called “The New Map of Life.”
  • it would be a mistake to replace the old rigid model of life — education first, then family and work, and finally retirement — with a new model just as rigid
  • there should be many different routes, interweaving leisure, work, education and family throughout life, taking people from birth to death with places to stop, rest, change courses and repeat steps along the way.
  • To thrive in an age of rapid knowledge transfer, children not only need reading, math and computer literacy, but they also need to learn to think creatively and not hold on to “facts” too tightly. They’ll need to find joy in unlearning and relearning.
  • Teens could take breaks from high school and take internships in workplaces that intrigue them. Education wouldn’t end in youth but rather be ever-present and take many forms outside of classrooms, from micro-degrees to traveling the world.
  • There is good reason to think we will work longer, but we can improve work quality with shorter workweeks, flexible scheduling and frequent “retirements.”
  • Rather than saving ever-larger pots of money for the end of life, we could pool risks in new ways
  • Generations may share wealth earlier than traditional bequests; we can start savings accounts at birth and allow young adults to work earlier so that compound interest can work in their favor.
  • Maintaining physical fitness from the beginning to end of life will be paramount. Getting children outside, encouraging sports, reducing the time we sit, and spending more time walking and moving will greatly improve individual lives.
Javier E

How to Navigate a 'Quarterlife' Crisis - The New York Times - 0 views

  • Satya Doyle Byock, a 39-year-old therapist, noticed a shift in tone over the past few years in the young people who streamed into her office: frenetic, frazzled clients in their late teens, 20s and 30s. They were unnerved and unmoored, constantly feeling like something was wrong with them.
  • “Crippling anxiety, depression, anguish, and disorientation are effectively the norm,”
  • her new book, “Quarterlife: The Search for Self in Early Adulthood.” The book uses anecdotes from Ms. Byock’s practice to outline obstacles faced by today’s young adults — roughly between the ages of 16 and 36 — and how to deal with them.
  • ...40 more annotations...
  • Just like midlife, quarterlife can bring its own crisis — trying to separate from your parents or caregivers and forge a sense of self is a struggle. But the generation entering adulthood now faces novel, sometimes debilitating, challenges.
  • Many find themselves so mired in day-to-day monetary concerns, from the relentless crush of student debt to the swelling costs of everything, that they feel unable to consider what they want for themselves long term
  • “We’ve been constrained by this myth that you graduate from college and you start your life,” she said. Without the social script previous generations followed — graduate college, marry, raise a family — Ms. Byock said her young clients often flailed around in a state of extended adolescence.
  • nearly one-third of Gen Z adults are living with their parents or other relatives and plan to stay there.
  • Many young people today struggle to afford college or decide not to attend, and the “existential crisis” that used to hit after graduation descends earlier and earlier
  • Ms. Byock said to pay attention to what you’re naturally curious about, and not to dismiss your interests as stupid or futile.
  • Experts said those entering adulthood need clear guidance for how to make it out of the muddle. Here are their top pieces of advice on how to navigate a quarterlife crisis today.
  • She recommends scheduling reminders to check in with yourself, roughly every three months, to examine where you are in your life and whether you feel stuck or dissatisfied
  • From there, she said, you can start to identify aspects of your life that you want to change.
  • “Start to give your own inner life the respect that it’s due,”
  • But quarterlife is about becoming a whole person, Ms. Byock said, and both groups need to absorb each other’s characteristics to balance themselves out
  • However, there is a difference between self-interest and self-indulgence, Ms. Byock said. Investigating and interrogating who you are takes work. “It’s not just about choosing your labels and being done,” she said.
  • Be patient.
  • Quarterlifers may feel pressure to race through each step of their lives, Ms. Byock said, craving the sense of achievement that comes with completing a task.
  • But learning to listen to oneself is a lifelong process.
  • Instead of searching for quick fixes, she said, young adults should think about longer-term goals: starting therapy that stretches beyond a handful of sessions, building healthy nutrition and exercise habits, working toward self-reliance.
  • “I know that seems sort of absurdly large and huge in scope,” she said. “But it’s allowing ourselves to meander and move through life, versus just ‘Check the boxes and get it right.’”
  • take stock of your day-to-day life and notice where things are missing. She groups quarterlifers into two categories: “stability types” and “meaning types.”
  • “Stability types” are seen by others as solid and stable. They prioritize a sense of security, succeed in their careers and may pursue building a family.
  • “But there’s a sense of emptiness and a sense of faking it,” she said. “They think this couldn’t possibly be all that life is about.”
  • On the other end of the spectrum, there are “meaning types” who are typically artists; they have intense creative passions but have a hard time dealing with day-to-day tasks
  • “These are folks for whom doing what society expects of you is so overwhelming and so discordant with their own sense of self that they seem to constantly be floundering,” she said. “They can’t quite figure it out.”
  • That paralysis is often exacerbated by mounting climate anxiety and the slog of a multiyear pandemic that has left many young people mourning family and friends, or smaller losses like a conventional college experience or the traditions of starting a first job.
  • Stability types need to think about how to give their lives a sense of passion and purpose. And meaning types need to find security, perhaps by starting with a consistent routine that can both anchor and unlock creativity.
  • perhaps the prototypical inspiration for staying calm in chaos: Yoda. The Jedi master is “one of the few images we have of what feeling quiet amid extreme pain and apocalypse can look like,
  • Even when there seems to be little stability externally, she said, quarterlifers can try to create their own steadiness.
  • establishing habits that help you ground yourself as a young adult is critical because transitional periods make us more susceptible to burnout
  • He suggests building a practical tool kit of self-care practices, like regularly taking stock of what you’re grateful for, taking controlled breaths and maintaining healthy nutrition and exercise routines. “These are techniques that can help you find clarity,”
  • Don’t be afraid to make a big change.
  • It’s important to identify what aspects of your life you have the power to alter, Dr. Brown said. “You can’t change an annoying boss,” he said, “but you might be able to plan a career change.”
  • That’s easier said than done, he acknowledged, and young adults should weigh the risks of continuing to live in their status quo — staying in their hometown, or lingering in a career that doesn’t excite them — with the potential benefits of trying something new.
  • quarterlife is typically “the freest stage of the whole life span,
  • Young adults may have an easier time moving to a new city or starting a new job than their older counterparts would.
  • Know when to call your parents — and when to call on yourself.
  • Quarterlife is about the journey from dependence to independence, Ms. Byock said — learning to rely on ourselves, after, for some, growing up in a culture of helicopter parenting and hands-on family dynamics.
  • there are ways your relationship with your parents can evolve, helping you carve out more independence
  • That can involve talking about family history and past memories or asking questions about your parents’ upbringing
  • “You’re transitioning the relationship from one of hierarchy to one of friendship,” she said. “It isn’t just about moving away or getting physical distance.”
  • Every quarterlifer typically has a moment when they know they need to step away from their parents and to face obstacles on their own
  • That doesn’t mean you can’t, or shouldn’t, still depend on your parents in moments of crisis, she said. “I don’t think it’s just about never needing one’s parents again,” she said. “But it’s about doing the subtle work within oneself to know: This is a time I need to stand on my own.”
Javier E

How to Accept the Things You Can't Change - The Atlantic - 0 views

  • Everyone—even the most privileged among us—has circumstances they would like to change in their life.
  • How might you improve the situation? Your answer might be, “I should move, get a new job, and meet new people.” In other words, you should change the outside world to make it better for you.
  • Between the conditions around you and your response to them is a space. In this space, you have freedom. You can choose to try remodeling the world, or you can start by changing your reaction to it.
  • ...29 more annotations...
  • Sometimes, changing your circumstances is difficult but absolutely necessary, such as in cases of abuse or violence. And sometimes, changing your circumstances is fairly easy: If you are lethargic every morning, start going to bed earlier.
  • But in the gray areas in between, fighting against reality can be impossible, or incredibly inefficient. Maybe you have been diagnosed with a chronic illness for which there are no promising treatment options. Perhaps your romantic partner has left you against your wishes and cannot be persuaded otherwise.
  • In these sorts of situations, changing how you feel can actually be much easier than changing your physical reality, even if it seems unnatural.
  • No surprise, then, that chronic stress often leads to maladaptive coping mechanisms in modern life. These include the misuse of drugs and alcohol, rumination on the sources of stress, self-harm, and self-blaming
  • That can be blamed in part on biology. Negative emotions such as anger and fear activate the amygdala, which increases vigilance toward threats and improves your ability to detect and avoid danger.
  • In other words, stress makes you fight, flee, or freeze—not think, What would a prudent reaction be at this moment? Let’s consider the options.
  • But in the modern world, stress and anxiety are usually chronic, not episodic.
  • you no longer need your amygdala to help you outrun the tiger without asking your conscious brain’s permission. Instead, you use it to handle the nonlethal problems that pester you all day long. Even if you don’t have tigers to outrun, you can’t relax in your cave, because the emails are piling up.
  • Your emotions can seem out of your control at the best of times, and even more so during a crisis
  • Similarly, the Stoics believed that human reason, practiced studiously, could override knee-jerk emotions.
  • Buddhism posits that our minds are habitually unbalanced, but not intrinsically so; the key is to build new habits of thinking.
  • These ideas (especially the last) have inspired modern schools of psychotherapy, such as rational emotive behavior therapy and cognitive behavioral therapy, which aim to create practical strategies for changing our reactions to negative situations in our life—and thus becoming happier.
  • 4. Give more.
  • you can follow four steps to arrive at a happier frame of mind:
  • 1. Notice your feelings.
  • Self-observation requires that you be mindful of what you are feeling in the moment and approach your emotions with detached curiosity.
  • Say you are sick of working from home all day, with endless Zoom meetings and no real human contact. Rather than fantasizing about quitting, spend some time dissecting your boredom and discomfort. At what time of day are they worst?
  • Following this procedure during the coronavirus pandemic shutdowns, I started taking virtual meetings while out for a walk. It made a big difference.
  • 2. Accept your feelings.
  • in much of life, negative feelings are part of a full human experience; erasing them would make life grayer. Furthermore, ample research shows that negative emotions and experiences help us find life’s meaning and purpose.
  • In the journal you started in Step 1 above, ponder the things that you can’t realistically alter and the emotions they spark in you. Ask what you are learning about yourself from each of these feelings, and how you might grow as a result.
  • 3. Lower your expectations.
  • Once, as a young man, I told my father over the phone that I planned to quit my job. “Why?” he asked. “Because it doesn’t make me happy,” I told him. He paused for a long time, and finally said, “What makes you so special?
  • My problem—and it’s a common one—was that I had set unreasonable expectations about how happy the world was supposed to make me.
  • Calmly ask yourself whether you’re asking the world for something it can’t or won’t give you. If you are, you might be looking in the wrong place for your bliss
  • you shouldn’t assume that all your happiness can come from any single romance, material object, or activity. You need a “portfolio” approach, balancing faith or philosophy, family, friendship, and work in which you earn your success and serve others.
  • Ancient thinkers recognized this difficulty but believed that we can manage our reactions effectively if we have the right tools.
  • Research from the INSEAD business school in France shows that people who consider themselves a victim of circumstances don’t feel like they have any responsibility for them. They are also likely to be victimizers themselves, hurting the people who try to help them.
  • One way to break this cycle is to help others voluntarily and charitably. Not only is serving others one of the most effective ways to raise one’s own happiness; maintaining the two opposing ideas that you are both a victim and a helper is very difficult.
Javier E

You Have Permission to Be a Smartphone Skeptic - The Bulwark - 0 views

  • the brief return of one of my favorite discursive topics—are the kids all right?—in one of my least-favorite variations: why shouldn’t each of them have a smartphone and tablet?
  • One camp says yes, the kids are fine
  • complaints about screen time merely conceal a desire to punish hard-working parents for marginally benefiting from climbing luxury standards, provide examples of the moral panic occasioned by all new technologies, or mistakenly blame screens for ill effects caused by the general political situation.
  • ...38 more annotations...
  • No, says the other camp, led by Jonathan Haidt; the kids are not all right, their devices are partly to blame, and here are the studies showing why.
  • we should not wait for the replication crisis in the social sciences to resolve itself before we consider the question of whether the naysayers are on to something. And normal powers of observation and imagination should be sufficient to make us at least wary of smartphones.
  • These powerful instruments represent a technological advance on par with that of the power loom or the automobile
  • The achievement can be difficult to properly appreciate because instead of exerting power over physical processes and raw materials, they operate on social processes and the human psyche: They are designed to maximize attention, to make it as difficult as possible to look away.
  • they have transformed the qualitative experience of existing in the world. They give a person’s sociality the appearance and feeling of a theoretically endless open network, while in reality, algorithms quietly sort users into ideological, aesthetic, memetic cattle chutes of content.
  • Importantly, the process by which smartphones change us requires no agency or judgment on the part of a teen user, and yet that process is designed to provide what feels like a perfectly natural, inevitable, and complete experience of the world.
  • Smartphones offer a tactile portal to a novel digital environment, and this environment is not the kind of space you enter and leave
  • One reason commonly offered for maintaining our socio-technological status quo is that nothing really has changed with the advent of the internet, of Instagram, of Tiktok and Youtube and 4Chan
  • It is instead a complete shadow world of endless images; disembodied, manipulable personas; and the ever-present gaze of others. It lives in your pocket and in your mind.
  • The price you pay for its availability—and the engine of its functioning—is that you are always available to it, as well. Unless you have a strength of will that eludes most adults, its emissaries can find you at any hour and in any place to issue your summons to the grim pleasure palace.
  • the self-restraint and self-discipline required to use a smartphone well—that is, to treat it purely as an occasional tool rather than as a totalizing way of life—are unreasonable things to demand of teenagers
  • these are unreasonable things to demand of me, a fully adult woman
  • To enjoy the conveniences that a smartphone offers, I must struggle against the lure of the permanent scroll, the notification, the urge to fix my eyes on the circle of light and keep them fixed. I must resist the default pseudo-activity the smartphone always calls its user back to, if I want to have any hope of filling the moments of my day with the real activity I believe is actually valuable.
  • for a child or teen still learning the rudiments of self-control, still learning what is valuable and fulfilling, still learning how to prioritize what is good over the impulse of the moment, it is an absurd bar to be asked to clear
  • The expectation that children and adolescents will navigate new technologies with fully formed and muscular capacities for reason and responsibility often seems to go along with a larger abdication of responsibility on the part of the adults involved.
  • adults have frequently given in to a Faustian temptation: offering up their children’s generation to be used as guinea pigs in a mass longitudinal study in exchange for a bit more room to breathe in their own undeniably difficult roles as educators, caretakers, and parents.
  • It is not a particular activity that you start and stop and resume, and it is not a social scene that you might abandon when it suits you.
  • And this we must do without waiting for social science to hand us a comprehensive mandate it is fundamentally unable to provide; without cowering in panic over moral panics
  • The pre-internet advertising world was vicious, to be sure, but when the “pre-” came off, its vices were moved into a compound interest account. In the world of online advertising, at any moment, in any place, a user engaged in an infinite scroll might be presented with native content about how one Instagram model learned to accept her chunky (size 4) thighs, while in the next clip, another model relates how a local dermatologist saved her from becoming an unlovable crone at the age of 25
  • developing pathological interests and capacities used to take a lot more work than it does now
  • You had to seek it out, as you once had to seek out pornography and look someone in the eye while paying for it. You were not funneled into it by an omnipresent stream of algorithmically curated content—the ambience of digital life, so easily mistaken by the person experiencing it as fundamentally similar to the non-purposive ambience of the natural world.
  • And when interpersonal relations between teens become sour, nasty, or abusive, as they often do and always have, the unbalancing effects of transposing social life to the internet become quite clear
  • For both young men and young women, the pornographic scenario—dominance and degradation, exposure and monetization—creates an experiential framework for desires that they are barely experienced enough to understand.
  • This is not a world I want to live in. I think it hurts everyone; but I especially think it hurts those young enough to receive it as a natural state of affairs rather than as a profound innovation.
  • so I am baffled by the most routine objection to any blaming of smartphones for our society-wide implosion of teenagers’ mental health,
  • In short, and inevitably, today’s teenagers are suffering from capitalism—specifically “late capitalism,
  • what shocks me about this rhetorical approach is the rush to play defense for Apple and its peers, the impulse to wield the abstract concept of capitalism as a shield for actually existing, extremely powerful, demonstrably ruthless capitalist actors.
  • This motley alliance of left-coded theory about the evils of business and right-coded praxis in defense of a particular evil business can be explained, I think, by a deeper desire than overthrowing capitalism. It is the desire not to be a prude or hysteric of bumpkin
  • No one wants to come down on the side of tamping off pleasures and suppressing teen activity.
  • No one wants to be the shrill or leaden antagonist of a thousand beloved movies, inciting moral panics, scheming about how to stop the youths from dancing on Sunday.
  • But commercial pioneers are only just beginning to explore new frontiers in the profit-driven, smartphone-enabled weaponization of our own pleasures against us
  • To limit your moral imagination to the archetypes of the fun-loving rebel versus the stodgy enforcers in response to this emerging reality is to choose to navigate it with blinders on, to be a useful idiot for the robber barons of online life rather than a challenger to the corrupt order they maintain.
  • The very basic question that needs to be asked with every product rollout and implementation is what technologies enable a good human life?
  • this question is not, ultimately, the province of social scientists, notwithstanding how useful their work may be on the narrower questions involved. It is the free privilege, it is the heavy burden, for all of us, to think—to deliberate and make judgments about human good, about what kind of world we want to live in, and to take action according to that thought.
  • I am not sure how to build a world in which childrens and adolescents, at least, do not feel they need to live their whole lives online.
  • whatever particular solutions emerge from our negotiations with each other and our reckonings with the force of cultural momentum, they will remain unavailable until we give ourselves permission to set the terms of our common life.
  • But the environments in which humans find themselves vary significantly, and in ways that have equally significant downstream effects on the particular expression of human nature in that context.
  • most of all, without affording Apple, Facebook, Google, and their ilk the defensive allegiance we should reserve for each other.
Emily Horwitz

'I Wanna Eat You Up!' Why We Go Crazy for Cute | LiveScience - 1 views

  • NEW ORLEANS — Ever reacted to the sight of a cute puppy or darling infant by squealing, "I want to eat you up!"? Or maybe you can't help but want to pinch your grandbaby's adorable cheeks. You're not alone. New research finds that seemingly strange aggressive responses to cuteness are actually the norm.
  • In the study, presented Friday (Jan. 18) here at the annual meeting of the Society for Personality and Social Psychology, researchers found that people watching a slideshow of adorable pictures popped more bubbles on a sheet of bubble wrap than did people viewing funny or neutral pictures.
  • The participants rated the pictures on cuteness and funniness, as well as on how much they felt the pictures made them lose control — for example, if they agreed with statements such as "I can't handle it!" The participants also rated the extent to which the pictures made them "want to say something like 'grr!'" and "want to squeeze something." Sure enough, the cuter the animal, the less control and more desire to "grrr" and squeeze something that people felt. Cute animals produced this feeling significantly more strongly than did funny animals. The funny critters in turn produced the feeling more strongly than did neutral animals, perhaps because the funny animals were perceived as cute, too, Dyer said.
  • ...4 more annotations...
  • yer got interested in what she and her colleagues call "cute aggression" after chatting with a fellow student about how adorable Internet pictures often produce the desire to squish or squeeze the cute critter. All the existing research on cuteness suggests the reaction should be the opposite, she told LiveScience. People should want to treat a cute thing with gentleness and care.
  • That's exactly what happened. The people watching a cute slideshow popped 120 bubbles, on average, compared with 80 for the funny slideshow and just a hair over 100 for the neutral one.
  • It's possible that seeing a wide-eyed baby or roly-poly pup triggers our drive to care for that creature, Dyer said. But since the animal is just a picture, and since even in real life we might not be able to care for the creature as much as we want, this urge may be frustrated, she said. That frustration could lead to aggression.
  • Or the reason might not be specific to cuteness, Dyer said. Many overwhelmingly positive emotions look negative, as when Miss America sobs while receiving her crown. Such high levels of positive emotion may overwhelm people.
Ellie McGinnis

The 50 Greatest Breakthroughs Since the Wheel - James Fallows - The Atlantic - 0 views

  • Some questions you ask because you want the right answer. Others are valuable because no answer is right; the payoff comes from the range of attempts.
  • That is the diversity of views about the types of historical breakthroughs that matter, with a striking consensus on whether the long trail of innovation recorded here is now nearing its end.
  • The clearest example of consensus was the first item on the final compilation, the printing press
  • ...27 more annotations...
  • Leslie Berlin, a historian of business at Stanford, organized her nominations not as an overall list but grouped into functional categories.
  • Innovations that expand the human intellect and its creative, expressive, and even moral possibilities.
  • Innovations that are integral to the physical and operating infrastructure of the modern world
  • Innovations that enabled the Industrial Revolution and its successive waves of expanded material output
  • Innovations extending life, to use Leslie Berlin’s term
  • Innovations that allowed real-time communication beyond the range of a single human voice
  • Innovations in the physical movement of people and goods.
  • Organizational breakthroughs that provide the software for people working and living together in increasingly efficient and modern ways
  • Finally, and less prominently than we might have found in 1950 or 1920—and less prominently than I initially expected—we have innovations in killing,
  • Any collection of 50 breakthroughs must exclude 50,000 more.
  • We learn, finally, why technology breeds optimism, which may be the most significant part of this exercise.
  • Popular culture often lionizes the stars of discovery and invention
  • For our era, the major problems that technology has helped cause, and that faster innovation may or may not correct, are environmental, demographic, and socioeconomic.
  • people who have thought deeply about innovation’s sources and effects, like our panelists, were aware of the harm it has done along with the good.
  • “Does innovation raise the wealth of the planet? I believe it does,” John Doerr, who has helped launch Google, Amazon, and other giants of today’s technology, said. “But technology left to its own devices widens rather than narrows the gap between the rich and the poor.”
  • Are today’s statesmen an improvement over those of our grandparents’ era? Today’s level of public debate? Music, architecture, literature, the fine arts—these and other manifestations of world culture continually change, without necessarily improving. Tolstoy and Dostoyevsky, versus whoever is the best-selling author in Moscow right now?
  • The argument that a slowdown might happen, and that it would be harmful if it did, takes three main forms.
  • Some societies have closed themselves off and stopped inventing altogether:
  • By failing to move forward, they inevitably moved backward relative to their rivals and to the environmental and economic threats they faced. If the social and intellectual climate for innovation sours, what has happened before can happen again.
  • visible slowdown in the pace of solutions that technology offers to fundamental problems.
  • a slowdown in, say, crop yields or travel time is part of a general pattern of what economists call diminishing marginal returns. The easy improvements are, quite naturally, the first to be made; whatever comes later is slower and harder.
  • America’s history as a nation happens to coincide with a rare moment in technological history now nearing its end. “There was virtually no economic growth before 1750,” he writes in a recent paper.
  • “We can be concerned about the last 1 percent of an environment for innovation, but that is because we take everything else for granted,” Leslie Berlin told me.
  • This reduction in cost, he says, means that the next decade should be a time of “amazing advances in understanding the genetic basis of disease, with especially powerful implications for cancer.”
  • the very concept of an end to innovation defied everything they understood about human inquiry. “If you look just at the 20th century, the odds against there being any improvement in living standards are enormous,”
  • “Two catastrophic world wars, the Cold War, the Depression, the rise of totalitarianism—it’s been one disaster after another, a sequence that could have been enough to sink us back into barbarism. And yet this past half century has been the fastest-ever time of technological growth. I see no reason why that should be slowing down.”
  • “I am a technological evolutionist,” he said. “I view the universe as a phase-space of things that are possible, and we’re doing a random walk among them. Eventually we are going to fill the space of everything that is possible.”
Javier E

Rich countries that let inequality run rampant make citizens unhappy, study finds | Ine... - 0 views

  • Countries that allow economic inequality to increase as they grow richer make their citizens less happy, a new study shows.
  • Until now, researchers have believed that inequality was largely irrelevant to levels of life satisfaction,
  • his study of 78 countries spanning four decades – the largest longitudinal research of its kind – punctures that myth
  • ...10 more annotations...
  • “When inequality increases, people with high incomes don’t benefit much from their gains – many rich people are focused on those who have even more than they do, and they never feel they have enough,”
  • “But people who earn little really suffer from falling further behind – they feel excluded and frustrated by not being able to keep up even with people who receive average incomes.”
  • examined survey data of life satisfaction levels, where people rate their life satisfaction on a scale of one to 10, and linked it to Gini coefficient numbers – a measure of inequality – from 1981 to 2020.
  • In 1981, as the UK was gripped by a recession, life satisfaction stood at 7.7. But during the economic boom of the 1980s, inequality grew, and the research shows that the happiness figure dropped to 7.4 by 1999.
  • However, as measures to reduce inequality began to take effect, happiness slowly returned so that by 2018, life satisfaction stood at 7.8.
  • Any country that moved from the lowest quarter of countries in terms of inequality to the highest quarter saw a decrease in life satisfaction of about 0.4 on the 10-point scale, he found.
  • India’s life satisfaction declined from 6.7 in 1990 to 5.8 in 2006 as inequality rose. By 2012 it was still lower than in 1990, despite the country’s prolonged economic boom.
  • The US and Australia also both saw pronounced falls in life satisfaction, but those countries where inequality had fallen were generally happier, such as Poland, Peru, Mexico and Ukraine, before the Russian invasion.
  • “In some of the previous research, you see people saying ‘inequality isn’t that big a deal, so all efforts to address inequality are misguided because inequality is beneficial’.
  • “I think that’s misguided – inequality is generally damaging to people’s life satisfaction so we should pay attention to efforts to mitigate it,
‹ Previous 21 - 40 of 146 Next › Last »
Showing 20 items per page