Skip to main content

Home/ TOK Friends/ Group items matching "valid" in title, tags, annotations or url

Group items matching
in title, tags, annotations or url

Sort By: Relevance | Date Filter: All | Bookmarks | Topics Simple Middle
Javier E

John Roberts, the Umpire in Chief - The New York Times - 0 views

  • The Roberts-Scalia debate is part of a longstanding argument about how judges should interpret laws passed by Congress.
  • the chief justice embraces an approach called “purposivism,” while Justice Scalia prefers “textualism.”
  • In Judge Katzmann’s account, purposivism has been the approach favored for most of American history by conservative and liberal judges, senators, and representatives, as well as administrative agencies. Purposivism holds that judges shouldn’t confine themselves to the words of a law but should try to discern Congress’s broader purposes.
  • ...6 more annotations...
  • In the 1980s, when he was a lower court judge, Justice Scalia began to champion a competing view of statutory interpretation, textualism, which holds that judges should confine themselves to interpreting the words that Congress chose without trying to discern Congress’s broader purposes.
  • Textualism, in this view, promises to constrain judicial activism by preventing judges from roving through legislative history in search of evidence that supports their own policy preferences. But in the view of its critics, like Chief Judge Katzmann, textualism “increases the probability that a judge will construe a law in a manner that the legislators did not intend.”
  • Judge Katzmann, who was appointed by President Bill Clinton, also accuses Justice Scalia of inconsistency for consulting the intent of the framers in the case of constitutional interpretation but not statutory interpretation.
  • The chief justice’s embrace of bipartisan judicial restraint in the second Affordable Care Act case was consistent with his embrace of the same philosophy in the first Affordable Care Act case in 2012, where he quoted one of his heroes, Justice Oliver Wendell Holmes Jr: “The rule is settled that as between two possible interpretations of a statute, by one of which it would be unconstitutional and by the other valid, our plain duty is to adopt that which will save the Act.”
  • Chief Justice Roberts was not, as Justice Scalia charged, rewriting the law. Instead he was advancing the view that he championed soon after his confirmation: In a polarized age, it is important for the Supreme Court to maintain its institutional legitimacy by deferring to the political branches.
  • Chief Justice Roberts’s relatively consistent embrace of judicial deference to democratic decisions supports his statement during his confirmation hearings that judges should be like umpires calling “balls and strikes.” As he put it then: “Umpires don’t make the rules, they apply them. The role of an umpire and a judge is critical. They make sure everybody plays by the rules, but it is a limited role. Nobody ever went to a ballgame to see the umpire.”
Javier E

A Reply To Jonathan Chait On Stimulus | The New Republic - 0 views

  • It is certainly true that a large majority of professional economists accept the view that “increasing spending or reducing taxes temporarily increases economic growth”—but that is very far from claiming that disputing it is largely a political campaign.
  • in a genuinely scientific field which has accepted a predictive rule as valid to the point that there is a true consensus—such that the only reason for refusal to accept it is crankery or, in Chait’s terms, “politics”—you don’t usually see: several full professors at the top two departments in the subject, when speaking directly in their area of research expertise, challenge it; 10 percent of all practitioners in the field refuse to accept it; and the two leading global general circulation publications in field running op-eds questioning it.
  • A great many leading economists may accept the proposition that enough stimulus spending will probably cause at least some increase in output for a short period of time in some circumstances, yet are still uncomfortable with the kind of stimulus spending strategy that is the actual subject of current political debate. In 2009, James Buchanan (1986 Nobel Laureate in Economics), Edward Prescott (2004 Nobel Laureate in Economics), and Vernon Smith (2002 Nobel Laureate in Economics) promulgated this statement: “Notwithstanding reports that all economists are now Keynesians and that we all support a big increase in the burden of government, we do not believe that more government spending is a way to improve economic performance.”
  • ...5 more annotations...
  • It is nerdy-sounding, but I believe critical to this discussion, to distinguish between measurement and knowledge. I made a very strong claim about measurement, and a very specific claim about knowledge. I claim that we cannot usefully measure the effect of the stimulus program launched in 2009 at all.
  • All potentially useful predictions made about the output impact of the stimulus program are non-falsifiable. Failure of predictions can be simply justified by this sort of ad hoc explanation after the fact.
  • This argument will always degenerate back into endlessly dueling regressions, because there is no ability to adjudicate among them via experiment.
  • It simply means that we have no scientific knowledge about this topic. Macroeconomic assertions about the effect of a proposed stimulus policy are not valueless, but despite their complex mathematical justifications, do not have standing as knowledge that can trump common sense, historical reasoning, and so on in the same way that a predictive rule that has been verified through experimental testing can.
  • When using stimulus to ameliorate the economic crisis, we are like primitive tribesmen using herbs to treat an infection, and we should not allow ourselves to imagine that we are using antibiotics that have been proven through clinical trials. This should not imply merely a different feeling about the same actions, but should rationally lead us to greater circumspection.
  •  
    An incisive analysis of knowledge claims in economics, and Keynesian approaches.
Javier E

A 'Philosophy' of Plastic Surgery in Brazil - NYTimes.com - 0 views

  • Is beauty a right, which, like education or health care, should be realized with the help of public institutions and expertise?
  • For years he has performed charity surgeries for the poor. More radically, some of his students offer free cosmetic operations in the nation’s public health system.
  • I asked her why she wanted to have the surgery.  “I didn’t put in an implant to exhibit myself, but to feel better. It wasn’t a simple vanity, but a  . . . necessary vanity.  Surgery improves a woman’s auto-estima.”
  • ...11 more annotations...
  • He argues that the real object of healing is not the body, but the mind.  A plastic surgeon is a “psychologist with a scalpel in his hand.” This idea led Pitanguy to argue for the “union” of cosmetic and reconstructive procedures.  In both types of surgery beauty and mental healing subtly mingle, he claims, and both benefit health.
  • “What is the difference between a plastic surgeon and a psychoanalyst?  The psychoanalyst knows everything but changes nothing.  The plastic surgeon knows nothing but changes everything.”
  • Plastic surgery gained legitimacy in the early 20th century by limiting itself to reconstructive operations.  The “beauty doctor” was a term of derision.  But as techniques improved they were used for cosmetic improvements.  Missing, however, was a valid diagnosis. Concepts like psychoanalyst Alfred Adler’s inferiority complex — and later low self-esteem — provided a missing link.
  • Victorians saw a cleft palate as a defect that built character. For us it hinders self-realization and merits corrective surgery.  This shift reflects a new attitude towards appearance and mental health: the notion that at least some defects cause unfair suffering and social stigma is now widely accepted. But Brazilian surgeons take this reasoning a step further.  Cosmetic surgery is a consumer service in most of the world.  In Brazil it is becoming, as Ester put it, a “necessary vanity.”
  • Pitanguy, whose patients often have mixed African, indigenous and European ancestry, stresses that aesthetic ideals vary by epoch and ethnicity.  What matters are not objective notions of beauty, but how the patient feels.  As his colleague says, the job of the plastic surgeon is to simply “follow desires.”
  • Patients are on average younger than they were 20 years ago.  They often request minor changes to become, as one surgeon said, “more perfect.”
  • The growth of plastic surgery thus reflects a new way of working not only on the suffering mind, but also on the erotic body.  Unlike fashion’s embrace of playful dissimulation and seduction, this beauty practice instead insists on correcting precisely measured flaws.  Plastic surgery may contribute to a biologized view of sex where pleasure and fantasy matter less than the anatomical “truth” of the bare body.
  • It is not coincidental that Brazil has not only high rates of plastic surgery, but also Cesarean sections (70 percent of deliveries in some private hospitals), tubal ligations,  and other surgeries for women. Some women see elective surgeries as part of a modern standard of care, more or less routine for the middle class, but only sporadically available to the poor.
  • When a good life is defined through the ability to buy goods then rights may be reinterpreted to mean not equality before the law, but equality in the market. 
  • Beauty is unfair: the attractive enjoy privileges and powers gained without merit.  As such it can offend egalitarian values.  Yet while attractiveness is a quality “awarded” to those who don’t morally deserve it, it can also grant power to those excluded from other systems of privilege.  It is a kind of “double negative”: a form of power that is unfairly distributed but which can disturb other unfair hierarchies.  For this reason it may have democratic appeal.  In poor urban areas beauty often has a similar importance for girls as soccer (or basketball) does for boys: it promises an almost magical attainment of recognition, wealth or power.
  • For many consumers attractiveness is essential to economic and sexual competition, social visibility, and mental well being.  This “value” of appearance may be especially clear for those excluded from other means of social ascent.  For the poor beauty is often a form of capital that can be exchanged for other benefits, however small, transient, or unconducive to collective change.
catbclark

How We Learned to Kill - NYTimes.com - 0 views

  • “There are two people digging by the side of the road. Can we shoot them?”
  • In war, of course, there are many ways to kill. I did so by giving orders. I never fired my weapon in combat, but I ordered countless others to
  • My initial reaction was to ask the question to someone higher up the chain of command.
  • ...12 more annotations...
  • I wanted confirmation from a higher authority to do the abhorrent, something I’d spent my entire life believing was evil.
  • I realized it was my role as an officer to provide that validation to the Marine on the other end who would pull the trigger.
  • I also received affirmation to a more sinister question: Yes, I could kill.
  • The primary factors that affect an individual’s ability to kill are the demands of authority, group absolution, the predisposition of the killer, the distance from the victim and the target attractiveness of the victim.
  • Were the men in their sights irrigating their farmland or planting a roadside bomb?
  • Before killing the first time there’s a reluctance that tempers the desire to know whether you are capable of doing it
  • . Despite the rhetoric I internalized from the newspapers back home about why we were in Afghanistan, I ended up fighting for different reasons once I got on the ground — a mix of loyalty to my Marines, habit and the urge to survive.
  • The more I thought about the enemy, the harder it was to view them as evil or subhuman. But killing requires a motivation
  • If someone is shooting at me, I have a right to fire back
  • Until that moment, our deployment in Afghanistan had been exhilarating because we felt invulnerable. This invulnerability in an environment of death was the most powerful sensation I’d ever experienced.
  • The fog of war doesn’t just limit what you can know; it creates doubt about everything you’re certain that you know.
  • The madness of war is that while this system is in place to kill people, it may actually be necessary for the greater good. We live in a dangerous world where killing and torture exist and where the persecution of the weak by the powerful is closer to the norm than the civil society where we get our Starbucks.
Javier E

Words matter in 'ISIS' war, so use 'Daesh' - The Boston Globe - 0 views

  • the Islamic State in Iraq and al-Sham, or ISIS; the Islamic State in Iraq and the Levant, or ISIL; and, more recently, the Islamic State, or IS. French officials recently declared that that country would stop using any of those names and instead refer to the group as “Daesh.”
  • how we talk about this group is central to defeating them.
  • Laurent Fabius said, “This is a terrorist group and not a state. . . the term Islamic State blurs the lines between Islam, Muslims, and Islamists.” President Obama made similar remarks saying, “ISIL is not Islamic . . . and [is] certainly not a state.”
  • ...5 more annotations...
  • The term “Daesh” is strategically a better choice because it is still accurate in that it spells out the acronym of the group’s full Arabic name, al-Dawla al-Islamiya fi al-Iraq wa al-Sham
  • “Daesh” can also be understood as a play on words — and an insult. Depending on how it is conjugated in Arabic, it can mean anything from “to trample down and crush” to “a bigot who imposes his view on others.
  • By using the militants’ preferred names, the US government implicitly gives them legitimacy. But referring to the group as Daesh doesn’t just withhold validity. It also might help the United States craft better policy.
  • A number of studies suggest that the language we use affects the way we think and behave. By using a term that references the Arabic name and not an English translation, American policy makers can potentially inoculate themselves from inherent biases that could affect their decision making
  • the United States is weakening the potency of its own messaging if it continues to refer to the group as ISIL.
kushnerha

There's nothing wrong with grade inflation - The Washington Post - 0 views

  • By the early ’90s, so long as one had the good sense to major in the humanities — all bets were off in the STEM fields — it was nearly impossible to get a final grade below a B-minus at an elite college. According to a 2012 study, the average college GPA, which in the 1930s was a C-plus, had risen to a B at public universities and a B-plus at private schools. At Duke, Pomona and Harvard, D’s and F’s combine for just 2 percent of all grades. A Yale report found that 62 percent of all Yale grades are A or A-minus. According to a 2013 article in the Harvard Crimson, the median grade at Harvard was an A-minus , while the most common grade was an A.
  • The result is widespread panic about grade inflation at elite schools. (The phenomenon is not as prevalent at community colleges and less-selective universities.) Some blame students’ consumer mentality, a few see a correlation with small class sizes (departments with falling enrollments want to keep students happy), and many cite a general loss of rigor in a touchy-feely age.
  • Yet whenever elite schools have tried to fight grade inflation, it’s been a mess. Princeton instituted strict caps on the number of high grades awarded, then abandoned the plan, saying the caps dissuaded applicants and made students miserable. At Wellesley, grade-inflated humanities departments mandated that the average result in their introductory and intermediate classes not exceed a B-plus. According to one study, enrollment fell by one-fifth, and students were 30 percent less likely to major in one of these subjects.
  • ...12 more annotations...
  • I liked the joy my students found when they actually earned a grade they’d been reaching for. But whereas I once thought we needed to contain grades, I now see that we may as well let them float skyward. If grade inflation is bad, fighting it is worse. Our goal should be ending the centrality of grades altogether. For years, I feared that a world of only A’s would mean the end of meaningful grades; today, I’m certain of it. But what’s so bad about that?
  • It’s easy to see why schools want to fight grade inflation. Grades should motivate certain students: those afraid of the stigma of a bad grade or those ambitious, by temperament or conditioning, to succeed in measurable ways. Periodic grading during a term, on quizzes, tests or papers, provides feedback to students, which should enable them to do better. And grades theoretically signal to others, such as potential employers or graduate schools, how well the student did. (Grade-point averages are also used for prizes and class rankings, though that doesn’t strike me as an important feature.)
  • But it’s not clear that grades work well as motivators. Although recent research on the effects of grades is limited, several studies in the 1970s, 1980s and 1990s measured how students related to a task or a class when it was graded compared to when it was ungraded. Overall, graded students are less interested in the topic at hand and — and, for obvious, common-sense reasons — more inclined to pick the easiest possible task when given the chance. In the words of progressive-education theorist Alfie Kohn, author of “The Homework Myth,” “the quality of learning declines” when grades are introduced, becoming “shallower and more superficial when the point is to get a grade.”
  • Even where grades can be useful, as in describing what material a student has mastered, they are remarkably crude instruments. Yes, the student who gets a 100 on a calculus exam probably grasps the material better than the student with a 60 — but only if she retains the knowledge, which grades can’t show.
  • I still can’t say very well what separates a B from an A. What’s more, I never see the kind of incompetence or impudence that would merit a D or an F. And now, in our grade-inflated world, it’s even harder to use grades to motivate, or give feedback, or send a signal to future employers or graduate schools.
  • According to a 2012 study by the Chronicle of Higher Education, GPA was seventh out of eight factors employers considered in hiring, behind internships, extracurricular activities and previous employment. Last year, Stanford’s registrar told the Chronicle about “a clamor” from employers “for something more meaningful” than the traditional transcript. The Lumina Foundation gave a$1.27 million grant to two organizations for college administrators working to develop better student records, with grades only one part of a student’s final profile.
  • Some graduate schools, too, have basically ditched grades. “As long as you don’t bomb and flunk out, grades don’t matter very much in M.F.A. programs,” the director of one creative-writing program told the New York Times. To top humanities PhD programs, letters of reference and writing samples matter more than overall GPA (although students are surely expected to have received good grades in their intended areas of study). In fact, it’s impossible to get into good graduate or professional schools without multiple letters of reference, which have come to function as the kind of rich, descriptive comments that could go on transcripts in place of grades.
  • suggests that GPAs serve not to validate students from elite schools but to keep out those from less-prestigious schools and large public universities, where grades are less inflated. Grades at community colleges “have actually dropped” over the years, according to Stuart Rojstaczer, a co-author of the 2012 grade-inflation study. That means we have two systems: one for students at elite schools, who get jobs based on references, prestige and connections, and another for students everywhere else, who had better maintain a 3.0. Grades are a tool increasingly deployed against students without prestige.
  • The trouble is that, while it’s relatively easy for smaller colleges to go grade-free, with their low student-to-teacher ratios, it’s tough for professors at larger schools, who must evaluate more students, more quickly, with fewer resources. And adjuncts teaching five classes for poverty wages can’t write substantial term-end comments, so grades are a necessity if they want to give any feedback at all.
  • It would mean hiring more teachers and paying them better (which schools should do anyway). And if transcripts become more textured, graduate-school admission offices and employers will have to devote more resources to reading them, and to getting to know applicants through interviews and letters of reference — a salutary trend that is underway already.
  • When I think about getting rid of grades, I think of happier students, with whom I have more open, democratic relationships. I think about being forced to pay more attention to the quiet ones, since I’ll have to write something truthful about them, too. I’ve begun to wonder if a world without grades may be one of those states of affairs (like open marriages, bicycle lanes and single-payer health care) that Americans resist precisely because they seem too good, suspiciously good. Nothing worth doing is supposed to come easy.
  • Alfie Kohn, too, sees ideology at work in the grade-inflation panic. “Most of what powers the arguments against grade inflation is a very right-wing idea that excellence consists in beating everyone else around you,” he says. “Even when you have sorted them — even when they get to Harvard! — we have to sort them again.” In other words, we can trust only a system in which there are clear winners and losers.
kushnerha

The Next Genocide - The New York Times - 1 views

  • But sadly, the anxieties of our own era could once again give rise to scapegoats and imagined enemies, while contemporary environmental stresses could encourage new variations on Hitler’s ideas, especially in countries anxious about feeding their growing populations or maintaining a rising standard of living.
  • The quest for German domination was premised on the denial of science. Hitler’s alternative to science was the idea of Lebensraum.
    • kushnerha
       
      "Lebensraum linked a war of extermination to the improvement of lifestyle." Additionally, "The pursuit of peace and plenty through science, he claimed in "Mein Kampf," was a Jewish plot to distract Germans from the necessity of war."
  • Climate change has also brought uncertainties about food supply back to the center of great power politics.
  • ...8 more annotations...
  • China today, like Germany before the war, is an industrial power incapable of feeding its population from its own territory
    • kushnerha
       
      And "could make China's population susceptible to a revival of ideas like Lebensraum."
  • The risk is that a developed country able to project military power could, like Hitler’s Germany, fall into ecological panic, and take drastic steps to protect its existing standard of living.
  • United States has done more than any other nation to bring about the next ecological panic, yet it is the only country where climate science is still resisted by certain political and business elites. These deniers tend to present the empirical findings of scientists as a conspiracy and question the validity of science — an intellectual stance that is uncomfortably close to Hitler’s.
  • The Kremlin, which is economically dependent on the export of hydrocarbons to Europe, is now seeking to make gas deals with individual European states one by one in order to weaken European unity and expand its own influence.
  • Putin waxes nostalgic for the 1930s, while Russian nationalists blame gays, cosmopolitans and Jews for antiwar sentiment. None of this bodes well for Europe’s future
  • The Nazi scenario of 1941 will not reappear in precisely the same form, but several of its causal elements have already begun to assemble.
  • not difficult to imagine ethnic mass murder in Africa
    • kushnerha
       
      also no longer difficult to imagine the "triumph of a violent totalitarian strain of Islamism in the parched Middle East," a "Chinese play for resources in Africa or Russia or Eastern Europe that involves removing the people already living there," and a "growing global ecological panic if America abandons climate science or the European Union falls apart"
  • Denying science imperils the future by summoning the ghosts of the past.
    • kushnerha
       
      Americans must make the "crucial choice between science and ideology"
Javier E

Opinion | What Do We Actually Know About the Economy? (Wonkish) - The New York Times - 0 views

  • Among economists more generally, a lot of the criticism seems to amount to the view that macroeconomics is bunk, and that we should stick to microeconomics, which is the real, solid stuff. As I’ll explain in a moment, that’s all wrong
  • in an important sense the past decade has been a huge validation for textbook macroeconomics; meanwhile, the exaltation of micro as the only “real” economics both gives microeconomics too much credit and is largely responsible for the ways macroeconomic theory has gone wrong.
  • Finally, many outsiders and some insiders have concluded from the crisis that economic theory in general is bunk, that we should take guidance from people immersed in the real world – say, business leaders — and/or concentrate on empirical results and skip the models
  • ...28 more annotations...
  • And while empirical evidence is important and we need more of it, the data almost never speak for themselves – a point amply illustrated by recent monetary events.
  • chwinger, as I remember the story, was never seen to use a Feynman diagram. But he had a locked room in his house, and the rumor was that that room was where he kept the Feynman diagrams he used in secret.
  • What’s the equivalent of Feynman diagrams? Something like IS-LM, which is the simplest model you can write down of how interest rates and output are jointly determined, and is how most practicing macroeconomists actually think about short-run economic fluctuations. It’s also how they talk about macroeconomics to each other. But it’s not what they put in their papers, because the journals demand that your model have “microfoundations.”
  • The Bernanke Fed massively expanded the monetary base, by a factor of almost five. There were dire warnings that this would cause inflation and “debase the dollar.” But prices went nowhere, and not much happened to broader monetary aggregates (a result that, weirdly, some economists seemed to find deeply puzzling even though it was exactly what should have been expected.)
  • What about fiscal policy? Traditional macro said that at the zero lower bound there would be no crowding out – that deficits wouldn’t drive up interest rates, and that fiscal multipliers would be larger than under normal conditions. The first of these predictions was obviously borne out, as rates stayed low even when deficits were very large. The second prediction is a bit harder to test, for reasons I’ll get into when I talk about the limits of empiricism. But the evidence does indeed suggest large positive multipliers.
  • The overall story, then, is one of overwhelming predictive success. Basic, old-fashioned macroeconomics didn’t fail in the crisis – it worked extremely well
  • In fact, it’s hard to think of any other example of economic models working this well – making predictions that most non-economists (and some economists) refused to believe, indeed found implausible, but which came true. Where, for example, can you find any comparable successes in microeconomics?
  • Meanwhile, the demand that macro become ever more rigorous in the narrow, misguided sense that it look like micro led to useful approaches being locked up in Schwinger’s back room, and in all too many cases forgotten. When the crisis struck, it was amazing how many successful academics turned out not to know things every economist would have known in 1970, and indeed resurrected 1930-vintage fallacies in the belief that they were profound insights.
  • mainly I think it reflected the general unwillingness of human beings (a category that includes many though not necessarily all economists) to believe that so many people can be so wrong about something so big.
  • . To normal human beings the study of international trade and that of international macroeconomics might sound like pretty much the same thing. In reality, however, the two fields used very different models, had very different intellectual cultures, and tended to look down on each other. Trade people tended to consider international macro people semi-charlatans, doing ad hoc stuff devoid of rigor. International macro people considered trade people boring, obsessed with proving theorems and offering little of real-world use.
  • does microeconomics really deserve its reputation of moral and intellectual superiority? No
  • Even before the rise of behavioral economics, any halfway self-aware economist realized that utility maximization – indeed, the very concept of utility — wasn’t a fact about the world; it was more of a thought experiment, whose conclusions should always have been stated in the subjunctive.
  • But, you say, we didn’t see the Great Recession coming. Well, what do you mean “we,” white man? OK, what’s true is that few economists realized that there was a huge housing bubble
  • True, a model doesn’t have to be perfect to provide hugely important insights. But here’s my question: where are the examples of microeconomic theory providing strong, counterintuitive, successful predictions on the same order as the success of IS-LM macroeconomics after 2008? Maybe there are some, but I can’t come up with any.
  • The point is not that micro theory is useless and we should stop doing it. But it doesn’t deserve to be seen as superior to macro modeling.
  • And the effort to make macro more and more like micro – to ground everything in rational behavior – has to be seen now as destructive. True, that effort did lead to some strong predictions: e.g., only unanticipated money should affect real output, transitory income changes shouldn’t affect consumer spending, government spending should crowd out private demand, etc. But all of those predictions have turned out to be wrong.
  • Kahneman and Tversky and Thaler and so on deserved all the honors they received for helping to document the specific ways in which utility maximization falls short, but even before their work we should never have expected perfect maximization to be a good description of reality.
  • But data never speak for themselves, for a couple of reasons. One, which is familiar, is that economists don’t get to do many experiments, and natural experiments are rare
  • The other problem is that even when we do get something like natural experiments, they often took place under economic regimes that aren’t relevant to current problems.
  • Both of these problems were extremely relevant in the years following the 2008 crisis.
  • you might be tempted to conclude that the empirical evidence is that monetary expansion is inflationary, indeed roughly one-for-one.
  • But the question, as the Fed embarked on quantitative easing, was what effect this would have on an economy at the zero lower bound. And while there were many historical examples of big monetary expansion, examples at the ZLB were much rarer – in fact, basically two: the U.S. in the 1930s and Japan in the early 2000
  • These examples told a very different story: that expansion would not, in fact, be inflationary, that it would work out the way it did.
  • The point is that empirical evidence can only do certain things. It can certainly prove that your theory is wrong! And it can also make a theory much more persuasive in those cases where the theory makes surprising predictions, which the data bear out. But the data can never absolve you from the necessity of having theories.
  • Over this past decade, I’ve watched a number of economists try to argue from authority: I am a famous professor, therefore you should believe what I say. This never ends well. I’ve also seen a lot of nihilism: economists don’t know anything, and we should tear the field down and start over.
  • Obviously I differ with both views. Economists haven’t earned the right to be snooty and superior, especially if their reputation comes from the ability to do hard math: hard math has been remarkably little help lately, if ever.
  • On the other hand, economists do turn out to know quite a lot: they do have some extremely useful models, usually pretty simple ones, that have stood up well in the face of evidence and events. And they definitely shouldn’t defer to important and/or rich people on polic
  • : compare Janet Yellen’s macroeconomic track record with that of the multiple billionaires who warned that Bernanke would debase the dollar. Or take my favorite Business Week headline from 2010: “Krugman or [John] Paulson: Who You Gonna Bet On?” Um.The important thing is to be aware of what we do know, and why.Follow The New York Times Opinion section on Facebook and Twitter (@NYTopinion), and sign up for the Opinion Today newsletter.
krystalxu

Why Study Philosophy? 'To Challenge Your Own Point of View' - The Atlantic - 1 views

  • Goldstein’s forthcoming book, Plato at the Googleplex: Why Philosophy Won’t Go Away, offers insight into the significant—and often invisible—progress that philosophy has made. I spoke with Goldstein about her take on the science vs. philosophy debates, how we can measure philosophy’s advances, and why an understanding of philosophy is critical to our lives today.
  • One of the things about philosophy is that you don’t have to give up on any other field. Whatever field there is, there’s a corresponding field of philosophy. Philosophy of language, philosophy of politics, philosophy of math. All the things I wanted to know about I could still study within a philosophical framework.
  • There’s a peer pressure that sets in at a certain age. They so much want to be like everybody else. But what I’ve found is that if you instill this joy of thinking, the sheer intellectual fun, it will survive even the adolescent years and come back in fighting form. It’s empowering.
  • ...18 more annotations...
  • One thing that’s changed tremendously is the presence of women and the change in focus because of that. There’s a lot of interest in literature and philosophy, and using literature as a philosophical examination. It makes me so happy! Because I was seen as a hard-core analytic philosopher, and when I first began to write novels people thought, Oh, and we thought she was serious! But that’s changed entirely. People take literature seriously, especially in moral philosophy, as thought experiments. A lot of the most developed and effective thought experiments come from novels. Also, novels contribute to making moral progress, changing people’s emotions.
  • The other thing that’s changed is that there’s more applied philosophy. Let’s apply philosophical theory to real-life problems, like medical ethics, environmental ethics, gender issues. This is a real change from when I was in school and it was only theory.
  • here’s a lot of philosophical progress, it’s just a progress that’s very hard to see. It’s very hard to see because we see with it. We incorporate philosophical progress into our own way of viewing the world.
  • Plato would be constantly surprised by what we know. And not only what we know scientifically, or by our technology, but what we know ethically. We take a lot for granted. It’s obvious to us, for example, that individual’s ethical truths are equally important.
  • it’s usually philosophical arguments that first introduce the very outlandish idea that we need to extend rights. And it takes more, it takes a movement, and activism, and emotions, to affect real social change. It starts with an argument, but then it becomes obvious. The tracks of philosophy’s work are erased because it becomes intuitively obvious
  • The arguments against slavery, against cruel and unusual punishment, against unjust wars, against treating children cruelly—these all took arguments.
  • About 30 years ago, the philosopher Peter Singer started to argue about the way animals are treated in our factory farms. Everybody thought he was nuts. But I’ve watched this movement grow; I’ve watched it become emotional. It has to become emotional. You have to draw empathy into it. But here it is, right in our time—a philosopher making the argument, everyone dismissing it, but then people start discussing it. Even criticizing it, or saying it’s not valid, is taking it seriously
  • This is what we have to teach our children. Even things that go against their intuition they need to take seriously. What was intuition two generations ago is no longer intuition; and it’s arguments that change i
  • We are very inertial creatures. We do not like to change our thinking, especially if it’s inconvenient for us. And certainly the people in power never want to wonder whether they should hold power.
  • I’m really trying to draw the students out, make them think for themselves. The more they challenge me, the more successful I feel as a teacher. It has to be very active
  • Plato used the metaphor that in teaching philosophy, there needs to be a fire in the teacher, and the sheer heat will help the fire grow in the student. It’s something that’s kindled because of the proximity to the heat.
  • how can you make the case that they should study philosophy?
  • ches your inner life. You have lots of frameworks to apply to problems, and so many ways to interpret things. It makes life so much more interesting. It’s us at our most human. And it helps us increase our humanity. No matter what you do, that’s an asset.
  • What do you think are the biggest philosophical issues of our time? The growth in scientific knowledge presents new philosophical issues.
  • The idea of the multiverse. Where are we in the universe? Physics is blowing our minds about this.
  • The question of whether some of these scientific theories are really even scientific. Can we get predictions out of them?
  • And with the growth in cognitive science and neuroscience. We’re going into the brain and getting these images of the brain. Are we discovering what we really are? Are we solving the problem of free will? Are we learning that there isn’t any free will? How much do the advances in neuroscience tell us about the deep philosophical issues?
  • With the decline of religion is there a sense of the meaninglessness of life and the easy consumerist answer that’s filling the space religion used to occupy? This is something that philosophers ought to be addressing.
Javier E

Ivy League Schools Are Overrated. Send Your Kids Elsewhere. | New Republic - 1 views

  • a blizzard of admissions jargon that I had to pick up on the fly. “Good rig”: the transcript exhibits a good degree of academic rigor. “Ed level 1”: parents have an educational level no higher than high school, indicating a genuine hardship case. “MUSD”: a musician in the highest category of promise. Kids who had five or six items on their list of extracurriculars—the “brag”—were already in trouble, because that wasn’t nearly enough.
  • With so many accomplished applicants to choose from, we were looking for kids with something special, “PQs”—personal qualities—that were often revealed by the letters or essays. Kids who only had the numbers and the résumé were usually rejected: “no spark,” “not a team-builder,” “this is pretty much in the middle of the fairway for us.” One young person, who had piled up a truly insane quantity of extracurriculars and who submitted nine letters of recommendation, was felt to be “too intense.”
  • On the other hand, the numbers and the résumé were clearly indispensable. I’d been told that successful applicants could either be “well-rounded” or “pointy”—outstanding in one particular way—but if they were pointy, they had to be really pointy: a musician whose audition tape had impressed the music department, a scientist who had won a national award.
  • ...52 more annotations...
  • When I speak of elite education, I mean prestigious institutions like Harvard or Stanford or Williams as well as the larger universe of second-tier selective schools, but I also mean everything that leads up to and away from them—the private and affluent public high schools; the ever-growing industry of tutors and consultants and test-prep courses; the admissions process itself, squatting like a dragon at the entrance to adulthood; the brand-name graduate schools and employment opportunities that come after the B.A.; and the parents and communities, largely upper-middle class, who push their children into the maw of this machine.
  • Our system of elite education manufactures young people who are smart and talented and driven, yes, but also anxious, timid, and lost, with little intellectual curiosity and a stunted sense of purpose: trapped in a bubble of privilege, heading meekly in the same direction, great at what they’re doing but with no idea why they’re doing it.
  • “Super People,” the writer James Atlas has called them—the stereotypical ultra-high-achieving elite college students of today. A double major, a sport, a musical instrument, a couple of foreign languages, service work in distant corners of the globe, a few hobbies thrown in for good measure: They have mastered them all, and with a serene self-assurance
  • Like so many kids today, I went off to college like a sleepwalker. You chose the most prestigious place that let you in; up ahead were vaguely understood objectives: status, wealth—“success.” What it meant to actually get an education and why you might want one—all this was off the table.
  • It was only after 24 years in the Ivy League—college and a Ph.D. at Columbia, ten years on the faculty at Yale—that I started to think about what this system does to kids and how they can escape from it, what it does to our society and how we can dismantle it.
  • I taught many wonderful young people during my years in the Ivy League—bright, thoughtful, creative kids whom it was a pleasure to talk with and learn from. But most of them seemed content to color within the lines that their education had marked out for them. Very few were passionate about ideas. Very few saw college as part of a larger project of intellectual discovery and development. Everyone dressed as if they were ready to be interviewed at a moment’s notice.
  • Look beneath the façade of seamless well-adjustment, and what you often find are toxic levels of fear, anxiety, and depression, of emptiness and aimlessness and isolation. A large-scale survey of college freshmen recently found that self-reports of emotional well-being have fallen to their lowest level in the study’s 25-year history.
  • So extreme are the admission standards now that kids who manage to get into elite colleges have, by definition, never experienced anything but success. The prospect of not being successful terrifies them, disorients them. The cost of falling short, even temporarily, becomes not merely practical, but existential. The result is a violent aversion to risk.
  • There are exceptions, kids who insist, against all odds, on trying to get a real education. But their experience tends to make them feel like freaks. One student told me that a friend of hers had left Yale because she found the school “stifling to the parts of yourself that you’d call a soul.”
  • What no one seems to ask is what the “return” is supposed to be. Is it just about earning more money? Is the only purpose of an education to enable you to get a job? What, in short, is college for?
  • The first thing that college is for is to teach you to think.
  • College is an opportunity to stand outside the world for a few years, between the orthodoxy of your family and the exigencies of career, and contemplate things from a distance.
  • it is only through the act of establishing communication between the mind and the heart, the mind and experience, that you become an individual, a unique being—a soul. The job of college is to assist you to begin to do that. Books, ideas, works of art and thought, the pressure of the minds around you that are looking for their own answers in their own ways.
  • College is not the only chance to learn to think, but it is the best. One thing is certain: If you haven’t started by the time you finish your B.A., there’s little likelihood you’ll do it later. That is why an undergraduate experience devoted exclusively to career preparation is four years largely wasted.
  • Elite schools like to boast that they teach their students how to think, but all they mean is that they train them in the analytic and rhetorical skills that are necessary for success in business and the professions.
  • Everything is technocratic—the development of expertise—and everything is ultimately justified in technocratic terms.
  • Religious colleges—even obscure, regional schools that no one has ever heard of on the coasts—often do a much better job in that respect.
  • At least the classes at elite schools are academically rigorous, demanding on their own terms, no? Not necessarily. In the sciences, usually; in other disciplines, not so much
  • professors and students have largely entered into what one observer called a “nonaggression pact.”
  • higher marks for shoddier work.
  • today’s young people appear to be more socially engaged than kids have been for several decades and that they are more apt to harbor creative or entrepreneurial impulses
  • they tend to be played out within the same narrow conception of what constitutes a valid life: affluence, credentials, prestige.
  • Experience itself has been reduced to instrumental function, via the college essay. From learning to commodify your experiences for the application, the next step has been to seek out experiences in order to have them to commodify
  • there is now a thriving sector devoted to producing essay-ready summers
  • To be a high-achieving student is to constantly be urged to think of yourself as a future leader of society.
  • what these institutions mean by leadership is nothing more than getting to the top. Making partner at a major law firm or becoming a chief executive, climbing the greasy pole of whatever hierarchy you decide to attach yourself to. I don’t think it occurs to the people in charge of elite colleges that the concept of leadership ought to have a higher meaning, or, really, any meaning.
  • The irony is that elite students are told that they can be whatever they want, but most of them end up choosing to be one of a few very similar things
  • As of 2010, about a third of graduates went into financing or consulting at a number of top schools, including Harvard, Princeton, and Cornell.
  • Whole fields have disappeared from view: the clergy, the military, electoral politics, even academia itself, for the most part, including basic science
  • It’s considered glamorous to drop out of a selective college if you want to become the next Mark Zuckerberg, but ludicrous to stay in to become a social worker. “What Wall Street figured out,” as Ezra Klein has put it, “is that colleges are producing a large number of very smart, completely confused graduates. Kids who have ample mental horsepower, an incredible work ethic and no idea what to do next.”
  • t almost feels ridiculous to have to insist that colleges like Harvard are bastions of privilege, where the rich send their children to learn to walk, talk, and think like the rich. Don’t we already know this? They aren’t called elite colleges for nothing. But apparently we like pretending otherwise. We live in a meritocracy, after all.
  • Visit any elite campus across our great nation, and you can thrill to the heart-warming spectacle of the children of white businesspeople and professionals studying and playing alongside the children of black, Asian, and Latino businesspeople and professionals
  • That doesn’t mean there aren’t a few exceptions, but that is all they are. In fact, the group that is most disadvantaged by our current admissions policies are working-class and rural whites, who are hardly present
  • The college admissions game is not primarily about the lower and middle classes seeking to rise, or even about the upper-middle class attempting to maintain its position. It is about determining the exact hierarchy of status within the upper-middle class itself.
  • This system is exacerbating inequality, retarding social mobility, perpetuating privilege, and creating an elite that is isolated from the society that it’s supposed to lead. The numbers are undeniable. In 1985, 46 percent of incoming freshmen at the 250 most selective colleges came from the top quarter of the income distribution. By 2000, it was 55 percent
  • The major reason for the trend is clear. Not increasing tuition, though that is a factor, but the ever-growing cost of manufacturing children who are fit to compete in the college admissions game
  • Wealthy families start buying their children’s way into elite colleges almost from the moment they are born: music lessons, sports equipment, foreign travel (“enrichment” programs, to use the all-too-perfect term)—most important, of course, private-school tuition or the costs of living in a place with top-tier public schools.
  • s there anything that I can do, a lot of young people have written to ask me, to avoid becoming an out-of-touch, entitled little shit? I don’t have a satisfying answer, short of telling them to transfer to a public university. You cannot cogitate your way to sympathy with people of different backgrounds, still less to knowledge of them. You need to interact with them directly, and it has to be on an equal footing
  • Elite private colleges will never allow their students’ economic profile to mirror that of society as a whole. They can’t afford to—they need a critical mass of full payers and they need to tend to their donor base—and it’s not even clear that they’d want to.
  • Elite colleges are not just powerless to reverse the movement toward a more unequal society; their policies actively promote it.
  • The SAT is supposed to measure aptitude, but what it actually measures is parental income, which it tracks quite closely
  • U.S. News and World Report supplies the percentage of freshmen at each college who finished in the highest 10 percent of their high school class. Among the top 20 universities, the number is usually above 90 percent. I’d be wary of attending schools like that. Students determine the level of classroom discussion; they shape your values and expectations, for good and ill. It’s partly because of the students that I’d warn kids away from the Ivies and their ilk. Kids at less prestigious schools are apt to be more interesting, more curious, more open, and far less entitled and competitive.
  • The best option of all may be the second-tier—not second-rate—colleges, like Reed, Kenyon, Wesleyan, Sewanee, Mount Holyoke, and others. Instead of trying to compete with Harvard and Yale, these schools have retained their allegiance to real educational values.
  • Not being an entitled little shit is an admirable goal. But in the end, the deeper issue is the situation that makes it so hard to be anything else. The time has come, not simply to reform that system top to bottom, but to plot our exit to another kind of society altogether.
  • The education system has to act to mitigate the class system, not reproduce it. Affirmative action should be based on class instead of race, a change that many have been advocating for years. Preferences for legacies and athletes ought to be discarded. SAT scores should be weighted to account for socioeconomic factors. Colleges should put an end to résumé-stuffing by imposing a limit on the number of extracurriculars that kids can list on their applications. They ought to place more value on the kind of service jobs that lower-income students often take in high school and that high achievers almost never do. They should refuse to be impressed by any opportunity that was enabled by parental wealth
  • More broadly, they need to rethink their conception of merit. If schools are going to train a better class of leaders than the ones we have today, they’re going to have to ask themselves what kinds of qualities they need to promote. Selecting students by GPA or the number of extracurriculars more often benefits the faithful drudge than the original mind.
  • reforming the admissions process. That might address the problem of mediocrity, but it won’t address the greater one of inequality
  • The problem is the Ivy League itself. We have contracted the training of our leadership class to a set of private institutions. However much they claim to act for the common good, they will always place their interests first.
  • I’ve come to see that what we really need is to create one where you don’t have to go to the Ivy League, or any private college, to get a first-rate education.
  • High-quality public education, financed with public money, for the benefit of all
  • Everybody gets an equal chance to go as far as their hard work and talent will take them—you know, the American dream. Everyone who wants it gets to have the kind of mind-expanding, soul-enriching experience that a liberal arts education provides.
  • We recognize that free, quality K–12 education is a right of citizenship. We also need to recognize—as we once did and as many countries still do—that the same is true of higher education. We have tried aristocracy. We have tried meritocracy. Now it’s time to try democracy.
Javier E

Opinion | How to Serve a Deranged Tyrant, Stoically - The New York Times - 0 views

  • In A.D. 49, the well-known writer and Stoic philosopher was recalled from exile to tutor the successor of the emperor Claudius, a promising teenager named Nero. Like many people today, Seneca entered public service with ideals mitigated by a pragmatic understanding of the reality of the politics of his time.
  • Seneca, by contrast, had no hope that he could achieve anything by direct opposition to any of the emperors under whom he lived. His best hope was to moderate some of Nero’s worst tendencies and to maximize his own sense of autonomy.”
  • Though Nero had good qualities, he was obsessed with fame and had an endless need for validation. He was also unstable and paranoid, and began to eliminate his rivals — including murdering his own mother. Was Seneca personally involved in these decisions? We don’t know. But he helped legitimize the regime with his presence, and profited from it as well, becoming one of Rome’s richest men through his 13 years of service.
  • ...6 more annotations...
  • Live Nation Rules Music Ticketing, Some Say With Threats
  • To the Stoics, contributing to public affairs was a critical duty of the philosopher. Could Seneca decline to serve because he disagreed with the emperor? Could he leave a deranged Nero unsupervised? In time, Seneca would also come to the conclusion that when “the state is so rotten as to be past helping, if evil has entire dominion over it, the wise man will not labor in vain or waste his strength in unprofitable efforts.”
  • My own early career involved some questionable service to businesspeople. Employed and paid by them, I planned and carried out controversial publicity stunts, and used dishonest tactics with the public and the media. When I finally left those roles, I found a knowledge of Stoic philosophy integral to my ability to assess my past actions, and set a more honorable course going forward.
  • Seneca seemed to realize only belatedly that one can contribute to his fellow citizens in ways other than through the state — for instance, by writing or simply by being a good man at home.
  • In 65 A.D., Seneca would again find that philosophy did not exist only in the ethereal world. Conspirators began to plot against Nero’s life, and Seneca, finally accepting that the monster he had helped create needed to be stopped, appears to have participated — or covered for those who did.
  • The effort failed but provided Seneca an opportunity: His life up to that point had contradicted many of his own teachings, but now when Nero’s guards came and demanded his life, he would be brave and wise. The man who had written much about learning how to die and facing the end without fear would comfort his friends, finish an essay he was writing and distribute some finished pieces for safekeeping. Then, he slit his veins, took hemlock and succumbed to the suffocating steam of a bath.
Javier E

Assessing the Value of Buddhism, for Individuals and for the World - The New York Times - 0 views

  • Robert Wright sketches an answer early in “Why Buddhism Is True.” He settles on a credible blend that one might call Western Buddhism, a largely secular approach to life and its problems but not devoid of a spiritual dimension. The centerpiece of the approach is the practice of mindful meditation.
  • The goal of “Why Buddhism Is True” is ambitious: to demonstrate “that Buddhism’s diagnosis of the human predicament is fundamentally correct, and that its prescription is deeply valid and urgently important.”
  • It is reasonable to claim that Buddhism, with its focus on suffering, addresses critical aspects of the human predicament. It is also reasonable to suggest that the prescription it offers may be applicable and useful to resolve that predicament.
  • ...9 more annotations...
  • To produce his demonstrations and to support the idea that Buddhism is “true,” Wright relies on science, especially on evolutionary psychology, cognitive science and neuroscience.
  • Wright is up to the task: He’s a Buddhist who has written about religion and morality from a scientific perspective — he is most famous for his 1994 book, “The Moral Animal.”
  • Second, the mismatch between causes and responses is rooted in evolution. We have inherited from our nonhuman and human forerunners a complex affect apparatus suited to life circumstances very different from ours
  • First, the beneficial powers of meditation come from the possibility of realizing that our emotive reactions and the consequent feelings they engender — which operate in automated fashion, outside our deliberate control — are often inappropriate and even counterproductive relative to the situations that trigger them.
  • Third, meditation allows us to realize that the idea of the self as director of our decisions is an illusion, and that the degree to which we are at the mercy of a weakly controlled system places us at a considerable disadvantage
  • Fourth, the awareness brought on by meditation helps the construction of a truly enlightened humanity and counters the growing tribalism of contemporary societies.
  • when, in modern life, emotions such as fear and anger are incorrectly and unnecessarily engaged — for example, road rage — Wright calls the respective feelings “false” or “illusory.” Such feelings, however, are no less true than the thirst, hunger or pain that Wright accepts and welcomes
  • We can agree that mindful meditation promotes a distancing effect and thus may increase our chances of combining affect and reason advantageously. Meditation can help us glean the especially flawed and dislocated status of humans in modern societies, and help us see how social and political conflicts appear to provoke resentment and anger so easily.
  • How does one scale up, from many single individuals to populations, in time to prevent the social catastrophes that seem to be looming?
Javier E

Psychology's Replication Crisis Is Real, Many Labs 2 Says - The Atlantic - 1 views

  • n recent years, it has become painfully clear that psychology is facing a “reproducibility crisis,” in which even famous, long-established phenomena—the stuff of textbooks and TED Talks—might not be real
  • Ironically enough, it seems that one of the most reliable findings in psychology is that only half of psychological studies can be successfully repeated
  • That failure rate is especially galling, says Simine Vazire from the University of California at Davis, because the Many Labs 2 teams tried to replicate studies that had made a big splash and been highly cited
  • ...5 more annotations...
  • With 15,305 participants in total, the new experiments had, on average, 60 times as many volunteers as the studies they were attempting to replicate. The researchers involved worked with the scientists behind the original studies to vet and check every detail of the experiments beforehand. And they repeated those experiments many times over, with volunteers from 36 different countries, to see if the studies would replicate in some cultures and contexts but not others.
  • Despite the large sample sizes and the blessings of the original teams, the team failed to replicate half of the studies it focused on. It couldn’t, for example, show that people subconsciously exposed to the concept of heat were more likely to believe in global warming, or that moral transgressions create a need for physical cleanliness in the style of Lady Macbeth, or that people who grow up with more siblings are more altruistic.
  • Many Labs 2 “was explicitly designed to examine how much effects varied from place to place, from culture to culture,” says Katie Corker, the chair of the Society for the Improvement of Psychological Science. “And here’s the surprising result: The results do not show much variability at all.” If one of the participating teams successfully replicated a study, others did, too. If a study failed to replicate, it tended to fail everywhere.
  • it’s a serious blow to one of the most frequently cited criticisms of the “reproducibility crisis” rhetoric. Surely, skeptics argue, it’s a fantasy to expect studies to replicate everywhere. “There’s a massive deference to the sample,” Nosek says. “Your replication attempt failed? It must be because you did it in Ohio and I did it in Virginia, and people are different. But these results suggest that we can’t just wave those failures away very easily.”
  • the lack of variation in Many Labs 2 is actually a positive thing. Sure, it suggests that the large number of failed replications really might be due to sloppy science. But it also hints that the fundamental business of psychology—creating careful lab experiments to study the tricky, slippery, complicated world of the human mind—works pretty well. “Outside the lab, real-world phenomena can and probably do vary by context,” he says. “But within our carefully designed studies and experiments, the results are not chaotic or unpredictable. That means we can do valid social-science research.”
Javier E

The Story Behind the SAT Overhaul - NYTimes.com - 2 views

  • “When you cover too many topics,” Coleman said, “the assessments designed to measure those standards are inevitably superficial.” He pointed to research showing that more students entering college weren’t prepared and were forced into “remediation programs from which they never escape.” In math, for example, if you examined data from top-performing countries, you found an approach that emphasized “far fewer topics, far deeper,” the opposite of the curriculums he found in the United States, which he described as “a mile wide and an inch deep.”
  • The lessons he brought with him from thinking about the Common Core were evident — that American education needed to be more focused and less superficial, and that it should be possible to test the success of the newly defined standards through an exam that reflected the material being taught in the classroom.
  • she and her team had extensive conversations with students, teachers, parents, counselors, admissions officers and college instructors, asking each group to tell them in detail what they wanted from the test. What they arrived at above all was that a test should reflect the most important skills that were imparted by the best teachers
  • ...12 more annotations...
  • for example, a good instructor would teach Martin Luther King Jr.’s “I Have a Dream” speech by encouraging a conversation that involved analyzing the text and identifying the evidence, both factual and rhetorical, that makes it persuasive. “The opposite of what we’d want is a classroom where a teacher might ask only: ‘What was the year the speech was given? Where was it given?’ ”
  • in the past, assembling the SAT focused on making sure the questions performed on technical grounds, meaning: Were they appropriately easy or difficult among a wide range of students, and were they free of bias when tested across ethnic, racial and religious subgroups? The goal was “maximizing differentiation” among kids, which meant finding items that were answered correctly by those students who were expected to get them right and incorrectly by the weaker students. A simple way of achieving this, Coleman said, was to test the kind of obscure vocabulary words for which the SAT was famous
  • In redesigning the test, the College Board shifted its emphasis. It prioritized content, measuring each question against a set of specifications that reflect the kind of reading and math that students would encounter in college and their work lives. Schmeiser and others then spent much of early last year watching students as they answered a set of 20 or so problems, discussing the questions with the students afterward. “The predictive validity is going to come out the same,” she said of the redesigned test. “But in the new test, we have much more control over the content and skills that are being measured.”
  • Evidence-based reading and writing, he said, will replace the current sections on reading and writing. It will use as its source materials pieces of writing — from science articles to historical documents to literature excerpts — which research suggests are important for educated Americans to know and understand deeply. “The Declaration of Independence, the Constitution, the Bill of Rights and the Federalist Papers,” Coleman said, “have managed to inspire an enduring great conversation about freedom, justice, human dignity in this country and the world” — therefore every SAT will contain a passage from either a founding document or from a text (like Lincoln’s Gettysburg Address) that is part of the “great global conversation” the founding documents inspired.
  • The idea is that the test will emphasize words students should be encountering, like “synthesis,” which can have several meanings depending on their context. Instead of encouraging students to memorize flashcards, the test should promote the idea that they must read widely throughout their high-school years.
  • The Barbara Jordan vocabulary question would have a follow-up — “How do you know your answer is correct?” — to which students would respond by identifying lines in the passage that supported their answer.
  • . No longer will it be good enough to focus on tricks and trying to eliminate answer choices. We are not interested in students just picking an answer, but justifying their answers.”
  • the essay portion of the test will also be reformulated so that it will always be the same, some version of: “As you read the passage in front of you, consider how the author uses evidence such as facts or examples; reasoning to develop ideas and to connect claims and evidence; and stylistic or persuasive elements to add power to the ideas expressed. Write an essay in which you explain how the author builds an argument to persuade an audience.”
  • The math section, too, will be predicated on research that shows that there are “a few areas of math that are a prerequisite for a wide range of college courses” and careers. Coleman conceded that some might treat the news that they were shifting away from more obscure math problems to these fewer fundamental skills as a dumbing-down the test, but he was adamant that this was not the case. He explained that there will be three areas of focus: problem solving and data analysis, which will include ratios and percentages and other mathematical reasoning used to solve problems in the real world; the “heart of algebra,” which will test how well students can work with linear equations (“a powerful set of tools that echo throughout many fields of study”); and what will be called the “passport to advanced math,” which will focus on the student’s familiarity with complex equations and their applications in science and social science.
  • “Sometimes in the past, there’s been a feeling that tests were measuring some sort of ineffable entity such as intelligence, whatever that might mean. Or ability, whatever that might mean. What this is is a clear message that good hard work is going to pay off and achievement is going to pay off. This is one of the most significant developments that I have seen in the 40-plus years that I’ve been working in admissions in higher education.”
  • The idea of creating a transparent test and then providing a free website that any student could use — not to learn gimmicks but to get a better grounding and additional practice in the core knowledge that would be tested — was appealing to Coleman.
  • (The College Board won’t pay Khan Academy.) They talked about a hypothetical test-prep experience in which students would log on to a personal dashboard, indicate that they wanted to prepare for the SAT and then work through a series of preliminary questions to demonstrate their initial skill level and identify the gaps in their knowledge. Khan said he could foresee a way to estimate the amount of time it would take to achieve certain benchmarks. “It might go something like, ‘O.K., we think you’ll be able to get to this level within the next month and this level within the next two months if you put in 30 minutes a day,’ ” he said. And he saw no reason the site couldn’t predict for anyone, anywhere the score he or she might hope to achieve with a commitment to a prescribed amount of work.
Javier E

They're Watching You at Work - Don Peck - The Atlantic - 2 views

  • Predictive statistical analysis, harnessed to big data, appears poised to alter the way millions of people are hired and assessed.
  • By one estimate, more than 98 percent of the world’s information is now stored digitally, and the volume of that data has quadrupled since 2007.
  • The application of predictive analytics to people’s careers—an emerging field sometimes called “people analytics”—is enormously challenging, not to mention ethically fraught
  • ...52 more annotations...
  • By the end of World War II, however, American corporations were facing severe talent shortages. Their senior executives were growing old, and a dearth of hiring from the Depression through the war had resulted in a shortfall of able, well-trained managers. Finding people who had the potential to rise quickly through the ranks became an overriding preoccupation of American businesses. They began to devise a formal hiring-and-management system based in part on new studies of human behavior, and in part on military techniques developed during both world wars, when huge mobilization efforts and mass casualties created the need to get the right people into the right roles as efficiently as possible. By the 1950s, it was not unusual for companies to spend days with young applicants for professional jobs, conducting a battery of tests, all with an eye toward corner-office potential.
  • But companies abandoned their hard-edged practices for another important reason: many of their methods of evaluation turned out not to be very scientific.
  • this regime, so widespread in corporate America at mid-century, had almost disappeared by 1990. “I think an HR person from the late 1970s would be stunned to see how casually companies hire now,”
  • Many factors explain the change, he said, and then he ticked off a number of them: Increased job-switching has made it less important and less economical for companies to test so thoroughly. A heightened focus on short-term financial results has led to deep cuts in corporate functions that bear fruit only in the long term. The Civil Rights Act of 1964, which exposed companies to legal liability for discriminatory hiring practices, has made HR departments wary of any broadly applied and clearly scored test that might later be shown to be systematically biased.
  • about a quarter of the country’s corporations were using similar tests to evaluate managers and junior executives, usually to assess whether they were ready for bigger roles.
  • He has encouraged the company’s HR executives to think about applying the games to the recruitment and evaluation of all professional workers.
  • Knack makes app-based video games, among them Dungeon Scrawl, a quest game requiring the player to navigate a maze and solve puzzles, and Wasabi Waiter, which involves delivering the right sushi to the right customer at an increasingly crowded happy hour. These games aren’t just for play: they’ve been designed by a team of neuroscientists, psychologists, and data scientists to suss out human potential. Play one of them for just 20 minutes, says Guy Halfteck, Knack’s founder, and you’ll generate several megabytes of data, exponentially more than what’s collected by the SAT or a personality test. How long you hesitate before taking every action, the sequence of actions you take, how you solve problems—all of these factors and many more are logged as you play, and then are used to analyze your creativity, your persistence, your capacity to learn quickly from mistakes, your ability to prioritize, and even your social intelligence and personality. The end result, Halfteck says, is a high-resolution portrait of your psyche and intellect, and an assessment of your potential as a leader or an innovator.
  • When the results came back, Haringa recalled, his heart began to beat a little faster. Without ever seeing the ideas, without meeting or interviewing the people who’d proposed them, without knowing their title or background or academic pedigree, Knack’s algorithm had identified the people whose ideas had panned out. The top 10 percent of the idea generators as predicted by Knack were in fact those who’d gone furthest in the process.
  • What Knack is doing, Haringa told me, “is almost like a paradigm shift.” It offers a way for his GameChanger unit to avoid wasting time on the 80 people out of 100—nearly all of whom look smart, well-trained, and plausible on paper—whose ideas just aren’t likely to work out.
  • Aptitude, skills, personal history, psychological stability, discretion, loyalty—companies at the time felt they had a need (and the right) to look into them all. That ambit is expanding once again, and this is undeniably unsettling. Should the ideas of scientists be dismissed because of the way they play a game? Should job candidates be ranked by what their Web habits say about them? Should the “data signature” of natural leaders play a role in promotion? These are all live questions today, and they prompt heavy concerns: that we will cede one of the most subtle and human of skills, the evaluation of the gifts and promise of other people, to machines; that the models will get it wrong; that some people will never get a shot in the new workforce.
  • scoring distance from work could violate equal-employment-opportunity standards. Marital status? Motherhood? Church membership? “Stuff like that,” Meyerle said, “we just don’t touch”—at least not in the U.S., where the legal environment is strict. Meyerle told me that Evolv has looked into these sorts of factors in its work for clients abroad, and that some of them produce “startling results.”
  • consider the alternative. A mountain of scholarly literature has shown that the intuitive way we now judge professional potential is rife with snap judgments and hidden biases, rooted in our upbringing or in deep neurological connections that doubtless served us well on the savanna but would seem to have less bearing on the world of work.
  • We may like to think that society has become more enlightened since those days, and in many ways it has, but our biases are mostly unconscious, and they can run surprisingly deep. Consider race. For a 2004 study called “Are Emily and Greg More Employable Than Lakisha and Jamal?,” the economists Sendhil Mullainathan and Marianne Bertrand put white-sounding names (Emily Walsh, Greg Baker) or black-sounding names (Lakisha Washington, Jamal Jones) on similar fictitious résumés, which they then sent out to a variety of companies in Boston and Chicago. To get the same number of callbacks, they learned, they needed to either send out half again as many résumés with black names as those with white names, or add eight extra years of relevant work experience to the résumés with black names.
  • a sociologist at Northwestern, spent parts of the three years from 2006 to 2008 interviewing professionals from elite investment banks, consultancies, and law firms about how they recruited, interviewed, and evaluated candidates, and concluded that among the most important factors driving their hiring recommendations were—wait for it—shared leisure interests.
  • Lacking “reliable predictors of future performance,” Rivera writes, “assessors purposefully used their own experiences as models of merit.” Former college athletes “typically prized participation in varsity sports above all other types of involvement.” People who’d majored in engineering gave engineers a leg up, believing they were better prepared.
  • the prevailing system of hiring and management in this country involves a level of dysfunction that should be inconceivable in an economy as sophisticated as ours. Recent survey data collected by the Corporate Executive Board, for example, indicate that nearly a quarter of all new hires leave their company within a year of their start date, and that hiring managers wish they’d never extended an offer to one out of every five members on their team
  • In the late 1990s, as these assessments shifted from paper to digital formats and proliferated, data scientists started doing massive tests of what makes for a successful customer-support technician or salesperson. This has unquestionably improved the quality of the workers at many firms.
  • In 2010, however, Xerox switched to an online evaluation that incorporates personality testing, cognitive-skill assessment, and multiple-choice questions about how the applicant would handle specific scenarios that he or she might encounter on the job. An algorithm behind the evaluation analyzes the responses, along with factual information gleaned from the candidate’s application, and spits out a color-coded rating: red (poor candidate), yellow (middling), or green (hire away). Those candidates who score best, I learned, tend to exhibit a creative but not overly inquisitive personality, and participate in at least one but not more than four social networks, among many other factors. (Previous experience, one of the few criteria that Xerox had explicitly screened for in the past, turns out to have no bearing on either productivity or retention
  • When Xerox started using the score in its hiring decisions, the quality of its hires immediately improved. The rate of attrition fell by 20 percent in the initial pilot period, and over time, the number of promotions rose. Xerox still interviews all candidates in person before deciding to hire them, Morse told me, but, she added, “We’re getting to the point where some of our hiring managers don’t even want to interview anymore”
  • Gone are the days, Ostberg told me, when, say, a small survey of college students would be used to predict the statistical validity of an evaluation tool. “We’ve got a data set of 347,000 actual employees who have gone through these different types of assessments or tools,” he told me, “and now we have performance-outcome data, and we can split those and slice and dice by industry and location.”
  • Evolv’s tests allow companies to capture data about everybody who applies for work, and everybody who gets hired—a complete data set from which sample bias, long a major vexation for industrial-organization psychologists, simply disappears. The sheer number of observations that this approach makes possible allows Evolv to say with precision which attributes matter more to the success of retail-sales workers (decisiveness, spatial orientation, persuasiveness) or customer-service personnel at call centers (rapport-building)
  • There are some data that Evolv simply won’t use, out of a concern that the information might lead to systematic bias against whole classes of people
  • the idea that hiring was a science fell out of favor. But now it’s coming back, thanks to new technologies and methods of analysis that are cheaper, faster, and much-wider-ranging than what we had before
  • what most excites him are the possibilities that arise from monitoring the entire life cycle of a worker at any given company.
  • Now the two companies are working together to marry pre-hire assessments to an increasing array of post-hire data: about not only performance and duration of service but also who trained the employees; who has managed them; whether they were promoted to a supervisory role, and how quickly; how they performed in that role; and why they eventually left.
  • What begins with an online screening test for entry-level workers ends with the transformation of nearly every aspect of hiring, performance assessment, and management.
  • I turned to Sandy Pentland, the director of the Human Dynamics Laboratory at MIT. In recent years, Pentland has pioneered the use of specialized electronic “badges” that transmit data about employees’ interactions as they go about their days. The badges capture all sorts of information about formal and informal conversations: their length; the tone of voice and gestures of the people involved; how much those people talk, listen, and interrupt; the degree to which they demonstrate empathy and extroversion; and more. Each badge generates about 100 data points a minute.
  • he tried the badges out on about 2,500 people, in 21 different organizations, and learned a number of interesting lessons. About a third of team performance, he discovered, can usually be predicted merely by the number of face-to-face exchanges among team members. (Too many is as much of a problem as too few.) Using data gathered by the badges, he was able to predict which teams would win a business-plan contest, and which workers would (rightly) say they’d had a “productive” or “creative” day. Not only that, but he claimed that his researchers had discovered the “data signature” of natural leaders, whom he called “charismatic connectors” and all of whom, he reported, circulate actively, give their time democratically to others, engage in brief but energetic conversations, and listen at least as much as they talk.
  • His group is developing apps to allow team members to view their own metrics more or less in real time, so that they can see, relative to the benchmarks of highly successful employees, whether they’re getting out of their offices enough, or listening enough, or spending enough time with people outside their own team.
  • Torrents of data are routinely collected by American companies and now sit on corporate servers, or in the cloud, awaiting analysis. Bloomberg reportedly logs every keystroke of every employee, along with their comings and goings in the office. The Las Vegas casino Harrah’s tracks the smiles of the card dealers and waitstaff on the floor (its analytics team has quantified the impact of smiling on customer satisfaction). E‑mail, of course, presents an especially rich vein to be mined for insights about our productivity, our treatment of co-workers, our willingness to collaborate or lend a hand, our patterns of written language, and what those patterns reveal about our intelligence, social skills, and behavior.
  • people analytics will ultimately have a vastly larger impact on the economy than the algorithms that now trade on Wall Street or figure out which ads to show us. He reminded me that we’ve witnessed this kind of transformation before in the history of management science. Near the turn of the 20th century, both Frederick Taylor and Henry Ford famously paced the factory floor with stopwatches, to improve worker efficiency.
  • “The quantities of data that those earlier generations were working with,” he said, “were infinitesimal compared to what’s available now. There’s been a real sea change in the past five years, where the quantities have just grown so large—petabytes, exabytes, zetta—that you start to be able to do things you never could before.”
  • People analytics will unquestionably provide many workers with more options and more power. Gild, for example, helps companies find undervalued software programmers, working indirectly to raise those people’s pay. Other companies are doing similar work. One called Entelo, for instance, specializes in using algorithms to identify potentially unhappy programmers who might be receptive to a phone cal
  • He sees it not only as a boon to a business’s productivity and overall health but also as an important new tool that individual employees can use for self-improvement: a sort of radically expanded The 7 Habits of Highly Effective People, custom-written for each of us, or at least each type of job, in the workforce.
  • the most exotic development in people analytics today is the creation of algorithms to assess the potential of all workers, across all companies, all the time.
  • The way Gild arrives at these scores is not simple. The company’s algorithms begin by scouring the Web for any and all open-source code, and for the coders who wrote it. They evaluate the code for its simplicity, elegance, documentation, and several other factors, including the frequency with which it’s been adopted by other programmers. For code that was written for paid projects, they look at completion times and other measures of productivity. Then they look at questions and answers on social forums such as Stack Overflow, a popular destination for programmers seeking advice on challenging projects. They consider how popular a given coder’s advice is, and how widely that advice ranges.
  • The algorithms go further still. They assess the way coders use language on social networks from LinkedIn to Twitter; the company has determined that certain phrases and words used in association with one another can distinguish expert programmers from less skilled ones. Gild knows these phrases and words are associated with good coding because it can correlate them with its evaluation of open-source code, and with the language and online behavior of programmers in good positions at prestigious companies.
  • having made those correlations, Gild can then score programmers who haven’t written open-source code at all, by analyzing the host of clues embedded in their online histories. They’re not all obvious, or easy to explain. Vivienne Ming, Gild’s chief scientist, told me that one solid predictor of strong coding is an affinity for a particular Japanese manga site.
  • Gild’s CEO, Sheeroy Desai, told me he believes his company’s approach can be applied to any occupation characterized by large, active online communities, where people post and cite individual work, ask and answer professional questions, and get feedback on projects. Graphic design is one field that the company is now looking at, and many scientific, technical, and engineering roles might also fit the bill. Regardless of their occupation, most people leave “data exhaust” in their wake, a kind of digital aura that can reveal a lot about a potential hire.
  • professionally relevant personality traits can be judged effectively merely by scanning Facebook feeds and photos. LinkedIn, of course, captures an enormous amount of professional data and network information, across just about every profession. A controversial start-up called Klout has made its mission the measurement and public scoring of people’s online social influence.
  • Mullainathan expressed amazement at how little most creative and professional workers (himself included) know about what makes them effective or ineffective in the office. Most of us can’t even say with any certainty how long we’ve spent gathering information for a given project, or our pattern of information-gathering, never mind know which parts of the pattern should be reinforced, and which jettisoned. As Mullainathan put it, we don’t know our own “production function.”
  • Over time, better job-matching technologies are likely to begin serving people directly, helping them see more clearly which jobs might suit them and which companies could use their skills. In the future, Gild plans to let programmers see their own profiles and take skills challenges to try to improve their scores. It intends to show them its estimates of their market value, too, and to recommend coursework that might allow them to raise their scores even more. Not least, it plans to make accessible the scores of typical hires at specific companies, so that software engineers can better see the profile they’d need to land a particular job
  • Knack, for its part, is making some of its video games available to anyone with a smartphone, so people can get a better sense of their strengths, and of the fields in which their strengths would be most valued. (Palo Alto High School recently adopted the games to help students assess careers.) Ultimately, the company hopes to act as matchmaker between a large network of people who play its games (or have ever played its games) and a widening roster of corporate clients, each with its own specific profile for any given type of job.
  • When I began my reporting for this story, I was worried that people analytics, if it worked at all, would only widen the divergent arcs of our professional lives, further gilding the path of the meritocratic elite from cradle to grave, and shutting out some workers more definitively. But I now believe the opposite is likely to happen, and that we’re headed toward a labor market that’s fairer to people at every stage of their careers
  • For decades, as we’ve assessed people’s potential in the professional workforce, the most important piece of data—the one that launches careers or keeps them grounded—has been educational background: typically, whether and where people went to college, and how they did there. Over the past couple of generations, colleges and universities have become the gatekeepers to a prosperous life. A degree has become a signal of intelligence and conscientiousness, one that grows stronger the more selective the school and the higher a student’s GPA, that is easily understood by employers, and that, until the advent of people analytics, was probably unrivaled in its predictive powers.
  • the limitations of that signal—the way it degrades with age, its overall imprecision, its many inherent biases, its extraordinary cost—are obvious. “Academic environments are artificial environments,” Laszlo Bock, Google’s senior vice president of people operations, told The New York Times in June. “People who succeed there are sort of finely trained, they’re conditioned to succeed in that environment,” which is often quite different from the workplace.
  • because one’s college history is such a crucial signal in our labor market, perfectly able people who simply couldn’t sit still in a classroom at the age of 16, or who didn’t have their act together at 18, or who chose not to go to graduate school at 22, routinely get left behind for good. That such early factors so profoundly affect career arcs and hiring decisions made two or three decades later is, on its face, absurd.
  • I spoke with managers at a lot of companies who are using advanced analytics to reevaluate and reshape their hiring, and nearly all of them told me that their research is leading them toward pools of candidates who didn’t attend college—for tech jobs, for high-end sales positions, for some managerial roles. In some limited cases, this is because their analytics revealed no benefit whatsoever to hiring people with college degrees; in other cases, and more often, it’s because they revealed signals that function far better than college history,
  • Google, too, is hiring a growing number of nongraduates. Many of the people I talked with reported that when it comes to high-paying and fast-track jobs, they’re reducing their preference for Ivy Leaguers and graduates of other highly selective schools.
  • This process is just beginning. Online courses are proliferating, and so are online markets that involve crowd-sourcing. Both arenas offer new opportunities for workers to build skills and showcase competence. Neither produces the kind of instantly recognizable signals of potential that a degree from a selective college, or a first job at a prestigious firm, might. That’s a problem for traditional hiring managers, because sifting through lots of small signals is so difficult and time-consuming.
  • all of these new developments raise philosophical questions. As professional performance becomes easier to measure and see, will we become slaves to our own status and potential, ever-focused on the metrics that tell us how and whether we are measuring up? Will too much knowledge about our limitations hinder achievement and stifle our dreams? All I can offer in response to these questions, ironically, is my own gut sense, which leads me to feel cautiously optimistic.
  • Google’s understanding of the promise of analytics is probably better than anybody else’s, and the company has been changing its hiring and management practices as a result of its ongoing analyses. (Brainteasers are no longer used in interviews, because they do not correlate with job success; GPA is not considered for anyone more than two years out of school, for the same reason—the list goes on.) But for all of Google’s technological enthusiasm, these same practices are still deeply human. A real, live person looks at every résumé the company receives. Hiring decisions are made by committee and are based in no small part on opinions formed during structured interviews.
sissij

Bearing Witness to Executions: Last Breaths and Lasting Impressions - The New York Times - 1 views

  • The death penalty holds a crucial, conflicted place in a nation deeply divided over crime and punishment, and whether the state should ever take a life. But for such a long, very public legal process, only a small number of people see what unfolds inside the country’s death houses.
  • There is no uniformity when they look back on the emotions that surround the minutes when they watched someone die.
  • You’re watching through glass, and then the process starts.
  • ...8 more annotations...
  • They had a chance to offer last statements, and I was disgusted because they were so self-serving, narcissistic statements for these people who had caused so much pain and suffering.
  • A prosecutor has to have no doubt: not proof beyond a reasonable doubt, but no doubt that a person committed those crimes. As long as you have no doubt, I don’t think there’s any valid argument against the death penalty.
  • They caught him right there where he shot my son. I just don’t understand: 20 years before they killed him.
  • I cannot even tell you how unbelievable it was to see people deliberately get ready to kill your client.
  • When they pronounced him dead, I think I felt happy that he was no longer being hurt as part of the process.
  • It is probably the shortest yet longest 11 minutes of my life. No matter what anyone says, there’s really nothing to prepare you for what you are about to see.
  • The victim’s family is hurt, and the family of the individual. You’re not just killing a person. You’re killing his whole family. There’s a lot of people involved in this, not just the poor kid lying on a gurney.
  • People don’t realize that you never get over it, unless you’re just cold and calculated. I’ll never forget it. Not a day goes by. Not a day goes by. And I don’t expect it to. If it does, then I didn’t do what I was supposed to do, as a Christian and as a chaplain and as a human being.
  •  
    Death penalty has long been a very controversial topic. Do people have the right to take away another person's life as the punishment of their wrong-doing? Those five witnesses offer their own opinion and feelings towards death penalty. Some of them think it is deserving because their wrong-doing caused too much damage to others. Some think it is painful because as a person is killed, their families and relatives are also "killed" in a way. Life is not a calculation. One life never equals to another life. --Sissi (4/23/2017)
Javier E

Ex-Facebook president Sean Parker: site made to exploit human 'vulnerability' | Technology | The Guardian - 1 views

  • Facebook’s founders knew they were creating something addictive that exploited “a vulnerability in human psychology” from the outset, according to the company’s founding president Sean Parker.
  • “I don’t know if I really understood the consequences of what I was saying,” he added, pointing to “unintended consequences” that arise when a network grows to have more than 2 billion users.
  • “It literally changes your relationship with society, with each other. It probably interferes with productivity in weird ways. God only knows what it’s doing to our children’s brains,” he said.
  • ...3 more annotations...
  • He explained that when Facebook was being developed the objective was: “How do we consume as much of your time and conscious attention as possible?” It was this mindset that led to the creation of features such as the “like” button that would give users “a little dopamine hit” to encourage them to upload more content.
  • “It’s a social-validation feedback loop … exactly the kind of thing that a hacker like myself would come up with, because you’re exploiting a vulnerability in human psychology.”
  • “All of us are jacked into this system,” he said. “All of our minds can be hijacked. Our choices are not as free as we think they are.”
Javier E

The Problem With History Classes - The Atlantic - 3 views

  • The passion and urgency with which these battles are fought reflect the misguided way history is taught in schools. Currently, most students learn history as a set narrative—a process that reinforces the mistaken idea that the past can be synthesized into a single, standardized chronicle of several hundred pages. This teaching pretends that there is a uniform collective story, which is akin to saying everyone remembers events the same.
  • Yet, history is anything but agreeable. It is not a collection of facts deemed to be "official" by scholars on high. It is a collection of historians exchanging different, often conflicting analyses.
  • rather than vainly seeking to transcend the inevitable clash of memories, American students would be better served by descending into the bog of conflict and learning the many "histories" that compose the American national story.
  • ...18 more annotations...
  • Perhaps Fisher offers the nation an opportunity to divorce, once and for all, memory from history. History may be an attempt to memorialize and preserve the past, but it is not memory; memories can serve as primary sources, but they do not stand alone as history. A history is essentially a collection of memories, analyzed and reduced into meaningful conclusions—but that collection depends on the memories chosen.
  • Memories make for a risky foundation: As events recede further into the past, the facts are distorted or augmented by entirely new details
  • people construct unique memories while informing perfectly valid histories. Just as there is a plurality of memories, so, too, is there a plurality of histories.
  • Scholars who read a diverse set of historians who are all focused on the same specific period or event are engaging in historiography
  • This approach exposes textbooks as nothing more than a compilation of histories that the authors deemed to be most relevant and useful.
  • In historiography, the barrier between historian and student is dropped, exposing a conflict-ridden landscape. A diplomatic historian approaches an event from the perspective of the most influential statesmen (who are most often white males), analyzing the context, motives, and consequences of their decisions. A cultural historian peels back the objects, sights, and sounds of a period to uncover humanity’s underlying emotions and anxieties. A Marxist historian adopts the lens of class conflict to explain the progression of events. There are intellectual historians, social historians, and gender historians, among many others. Historians studying the same topic will draw different interpretations—sometimes radically so, depending on the sources they draw from
  • Jacoba Urist points out that history is "about explaining and interpreting past events analytically." If students are really to learn and master these analytical tools, then it is absolutely essential that they read a diverse set of historians and learn how brilliant men and women who are scrutinizing the same topic can reach different conclusions
  • Rather than constructing a curriculum based on the muddled consensus of boards, legislatures, and think tanks, schools should teach students history through historiography. The shortcomings of one historian become apparent after reading the work of another one on the list.
  • Although, as Urist notes, the AP course is "designed to teach students to think like historians," my own experience in that class suggests that it fails to achieve that goal.
  • The course’s framework has always served as an outline of important concepts aiming to allow educators flexibility in how to teach; it makes no reference to historiographical conflicts. Historiography was an epiphany for me because I had never before come face-to-face with how historians think and reason
  • When I took AP U.S. History, I jumbled these diverse histories into one indistinct narrative. Although the test involved open-ended essay questions, I was taught that graders were looking for a firm thesis—forcing students to adopt a side. The AP test also, unsurprisingly, rewards students who cite a wealth of supporting details
  • By the time I took the test in 2009, I was a master at "checking boxes," weighing political factors equally against those involving socioeconomics and ensuring that previously neglected populations like women and ethnic minorities received their due. I did not know that I was pulling ideas from different historiographical traditions. I still subscribed to the idea of a prevailing national narrative and served as an unwitting sponsor of synthesis, oblivious to the academic battles that made such synthesis impossible.
  • Although there may be an inclination to seek to establish order where there is chaos, that urge must be resisted in teaching history. Public controversies over memory are hardly new. Students must be prepared to confront divisiveness, not conditioned to shoehorn agreement into situations where none is possible
  • When conflict is accepted rather than resisted, it becomes possible for different conceptions of American history to co-exist. There is no longer a need to appoint a victor.
  • More importantly, the historiographical approach avoids pursuing truth for the sake of satisfying a national myth
  • The country’s founding fathers crafted some of the finest expressions of personal liberty and representative government the world has ever seen; many of them also held fellow humans in bondage. This paradox is only a problem if the goal is to view the founding fathers as faultless, perfect individuals. If multiple histories are embraced, no one needs to fear that one history will be lost.
  • History is not indoctrination. It is a wrestling match. For too long, the emphasis has been on pinning the opponent. It is time to shift the focus to the struggle itself
  • There is no better way to use the past to inform the present than by accepting the impossibility of a definitive history—and by ensuring that current students are equipped to grapple with the contested memories in their midst.
Javier E

There's No Such Thing As 'Sound Science' | FiveThirtyEight - 1 views

  • cience is being turned against itself. For decades, its twin ideals of transparency and rigor have been weaponized by those who disagree with results produced by the scientific method. Under the Trump administration, that fight has ramped up again.
  • The same entreaties crop up again and again: We need to root out conflicts. We need more precise evidence. What makes these arguments so powerful is that they sound quite similar to the points raised by proponents of a very different call for change that’s coming from within science.
  • Despite having dissimilar goals, the two forces espouse principles that look surprisingly alike: Science needs to be transparent. Results and methods should be openly shared so that outside researchers can independently reproduce and validate them. The methods used to collect and analyze data should be rigorous and clear, and conclusions must be supported by evidence.
  • ...26 more annotations...
  • they’re also used as talking points by politicians who are working to make it more difficult for the EPA and other federal agencies to use science in their regulatory decision-making, under the guise of basing policy on “sound science.” Science’s virtues are being wielded against it.
  • What distinguishes the two calls for transparency is intent: Whereas the “open science” movement aims to make science more reliable, reproducible and robust, proponents of “sound science” have historically worked to amplify uncertainty, create doubt and undermine scientific discoveries that threaten their interests.
  • “Our criticisms are founded in a confidence in science,” said Steven Goodman, co-director of the Meta-Research Innovation Center at Stanford and a proponent of open science. “That’s a fundamental difference — we’re critiquing science to make it better. Others are critiquing it to devalue the approach itself.”
  • alls to base public policy on “sound science” seem unassailable if you don’t know the term’s history. The phrase was adopted by the tobacco industry in the 1990s to counteract mounting evidence linking secondhand smoke to cancer.
  • The sound science tactic exploits a fundamental feature of the scientific process: Science does not produce absolute certainty. Contrary to how it’s sometimes represented to the public, science is not a magic wand that turns everything it touches to truth. Instead, it’s a process of uncertainty reduction, much like a game of 20 Questions.
  • Any given study can rarely answer more than one question at a time, and each study usually raises a bunch of new questions in the process of answering old ones. “Science is a process rather than an answer,” said psychologist Alison Ledgerwood of the University of California, Davis. Every answer is provisional and subject to change in the face of new evidence. It’s not entirely correct to say that “this study proves this fact,” Ledgerwood said. “We should be talking instead about how science increases or decreases our confidence in something.”
  • While insisting that they merely wanted to ensure that public policy was based on sound science, tobacco companies defined the term in a way that ensured that no science could ever be sound enough. The only sound science was certain science, which is an impossible standard to achieve.
  • “Doubt is our product,” wrote one employee of the Brown & Williamson tobacco company in a 1969 internal memo. The note went on to say that doubt “is the best means of competing with the ‘body of fact’” and “establishing a controversy.” These strategies for undermining inconvenient science were so effective that they’ve served as a sort of playbook for industry interests ever since
  • Doubt merchants aren’t pushing for knowledge, they’re practicing what Proctor has dubbed “agnogenesis” — the intentional manufacture of ignorance. This ignorance isn’t simply the absence of knowing something; it’s a lack of comprehension deliberately created by agents who don’t want you to know,
  • In the hands of doubt-makers, transparency becomes a rhetorical move. “It’s really difficult as a scientist or policy maker to make a stand against transparency and openness, because well, who would be against it?
  • But at the same time, “you can couch everything in the language of transparency and it becomes a powerful weapon.” For instance, when the EPA was preparing to set new limits on particulate pollution in the 1990s, industry groups pushed back against the research and demanded access to primary data (including records that researchers had promised participants would remain confidential) and a reanalysis of the evidence. Their calls succeeded and a new analysis was performed. The reanalysis essentially confirmed the original conclusions, but the process of conducting it delayed the implementation of regulations and cost researchers time and money.
  • Delay is a time-tested strategy. “Gridlock is the greatest friend a global warming skeptic has,” said Marc Morano, a prominent critic of global warming research
  • which has received funding from the oil and gas industry. “We’re the negative force. We’re just trying to stop stuff.”
  • these ploys are getting a fresh boost from Congress. The Data Quality Act (also known as the Information Quality Act) was reportedly written by an industry lobbyist and quietly passed as part of an appropriations bill in 2000. The rule mandates that federal agencies ensure the “quality, objectivity, utility, and integrity of information” that they disseminate, though it does little to define what these terms mean. The law also provides a mechanism for citizens and groups to challenge information that they deem inaccurate, including science that they disagree with. “It was passed in this very quiet way with no explicit debate about it — that should tell you a lot about the real goals,” Levy said.
  • in the 20 months following its implementation, the act was repeatedly used by industry groups to push back against proposed regulations and bog down the decision-making process. Instead of deploying transparency as a fundamental principle that applies to all science, these interests have used transparency as a weapon to attack very particular findings that they would like to eradicate.
  • Now Congress is considering another way to legislate how science is used. The Honest Act, a bill sponsored by Rep. Lamar Smith of Texas,3The bill has been passed by the House but still awaits a vote in the Senate. is another example of what Levy calls a “Trojan horse” law that uses the language of transparency as a cover to achieve other political goals. Smith’s legislation would severely limit the kind of evidence the EPA could use for decision-making. Only studies whose raw data and computer codes were publicly available would be allowed for consideration.
  • It might seem like an easy task to sort good science from bad, but in reality it’s not so simple. “There’s a misplaced idea that we can definitively distinguish the good from the not-good science, but it’s all a matter of degree,” said Brian Nosek, executive director of the Center for Open Science. “There is no perfect study.” Requiring regulators to wait until they have (nonexistent) perfect evidence is essentially “a way of saying, ‘We don’t want to use evidence for our decision-making,’
  • ost scientific controversies aren’t about science at all, and once the sides are drawn, more data is unlikely to bring opponents into agreement.
  • objective knowledge is not enough to resolve environmental controversies. “While these controversies may appear on the surface to rest on disputed questions of fact, beneath often reside differing positions of value; values that can give shape to differing understandings of what ‘the facts’ are.” What’s needed in these cases isn’t more or better science, but mechanisms to bring those hidden values to the forefront of the discussion so that they can be debated transparently. “As long as we continue down this unabashedly naive road about what science is, and what it is capable of doing, we will continue to fail to reach any sort of meaningful consensus on these matters,”
  • The dispute over tobacco was never about the science of cigarettes’ link to cancer. It was about whether companies have the right to sell dangerous products and, if so, what obligations they have to the consumers who purchased them.
  • Similarly, the debate over climate change isn’t about whether our planet is heating, but about how much responsibility each country and person bears for stopping it
  • While researching her book “Merchants of Doubt,” science historian Naomi Oreskes found that some of the same people who were defending the tobacco industry as scientific experts were also receiving industry money to deny the role of human activity in global warming. What these issues had in common, she realized, was that they all involved the need for government action. “None of this is about the science. All of this is a political debate about the role of government,”
  • These controversies are really about values, not scientific facts, and acknowledging that would allow us to have more truthful and productive debates. What would that look like in practice? Instead of cherry-picking evidence to support a particular view (and insisting that the science points to a desired action), the various sides could lay out the values they are using to assess the evidence.
  • For instance, in Europe, many decisions are guided by the precautionary principle — a system that values caution in the face of uncertainty and says that when the risks are unclear, it should be up to industries to show that their products and processes are not harmful, rather than requiring the government to prove that they are harmful before they can be regulated. By contrast, U.S. agencies tend to wait for strong evidence of harm before issuing regulations
  • the difference between them comes down to priorities: Is it better to exercise caution at the risk of burdening companies and perhaps the economy, or is it more important to avoid potential economic downsides even if it means that sometimes a harmful product or industrial process goes unregulated?
  • But science can’t tell us how risky is too risky to allow products like cigarettes or potentially harmful pesticides to be sold — those are value judgements that only humans can make.
Javier E

As a Doctor, I Was Skeptical About the Covid Vaccine. Then I Reviewed the Science. - The New York Times - 0 views

  • Until last week, I wasn’t sure I would get the vaccine. Some media reports highlight that mRNA vaccines have never been approved for use in humans outside clinical trials, making it seem like a new technology that has not been tested before. The vaccines were developed at such speed, I couldn’t be sure that major side effects hadn’t been overlooked. I worried about autoimmunity caused by expressing the coronavirus spike proteins on my own cells.
  • Every day in the emergency department, patients walk away from essential care against medical advice, and we watch them go with a shake of our heads and a rueful smile. Just like them, isolated with my doubts, I was ready to exercise my right to free will and refuse the vaccine.
  • When my non-medical friends asked me about it, I was torn between telling them my concerns and playacting the doctor who recommends the latest proven therapy.
  • ...8 more annotations...
  • The guilt I felt about this compelled me to objectively review the literature on mRNA vaccines. Not being an expert in virology or biochemistry, I realized I had to quickly master unfamiliar words like “transfection” and concepts about gene sequences. Slowly, the information I was devouring started changing my beliefs.
  • I learned that research into using mRNA for vaccinations and cancer therapies has been ongoing for the past 30 years. Trial and error have refined this modality so that it was almost fully fledged by the time Covid hit
  • The mRNA from the vaccine is broken down quickly in our cells, and the coronavirus spike protein is expressed only transiently on the cell surface.
  • Furthermore, this type of vaccine is harnessing a technique that viruses already use.
  • It was humbling to have to change my mind. As I booked my vaccination time slot, I realized how lucky I am to have access to all this research, as well as the training to understand it.
  • As medical professionals, we cannot afford to be paternalistic and trust that people will follow advice without all the facts. This is especially true in Australia, where the vast majority of us have never witnessed firsthand the ravages that this disease can inflict.
  • Like all new converts, I am now a true believer: I’d like everyone to be vaccinated. But autonomy is a precious tenet of a free society, and I’m glad the ethicists have advised against mandating the vaccine
  • just hope that with more robust discussion and the wider dissemination of scientific knowledge, we may sway people like me — who have what may be valid reservations — to get the vaccine.
« First ‹ Previous 41 - 60 of 84 Next › Last »
Showing 20 items per page