Skip to main content

Home/ TOK Friends/ Group items tagged safety

Rss Feed Group items tagged

peterconnelly

Should kids be allowed to use smartphones in the classroom? - The Washington Post - 0 views

  • The commonly recommended age for a first cellphone is 13 years old. Most kids that age are in the eighth grade, getting ready to learn algebra. The desperate phone calls to 911 coming from inside Robb Elementary School in Uvalde, Tex., where 21 people were killed last month, were placed by children younger than that.
  • In the aftermath of another school shooting, concerned parents are weighing whether phones are a distraction or a potential lifeline for their children.
  • Phone ownership already is widespread among younger children, with 43 percent of 8-to-12-year-olds owning their own smartphone
  • ...6 more annotations...
  • The devices have become part of many students’ daily lives, despite various attempts over the years by state legislatures and cities to keep them out of classrooms.
  • Despite any emotional benefits for adults or educational use for children, screen-time and security experts don’t recommend taking smartphones to class, at least not without some ground rules and guidance on how to use them in a worst-case scenario.
  • A phone can make unwanted noises, and in a silent lockdown, even a vibration could be too loud. Depending on their age, kids might also be tempted to post about an ongoing incident to social media, which Trump said could both inspire other potential gunmen seeking fame or reveal details about their location. Even the ability to call 911 isn’t a good reason, because an entire school full of people calling at once could overload a switchboard.
  • She said it’s a distraction, and she worries it could replace in-person social interactions and knows she can’t have oversight over many conversations or posts that might happen via smartphone.
  • She’s also worried about misinformation that her daughter could see or read.
  • “Phones are not going to stop school shootings. Gun control might,” Twenge said.
peterconnelly

Craig Nason : I survived the Columbine school shooting - then watched Uvalde ... - 0 views

  • In April 1999, I survived the Columbine shooting. At just 17 years old, I was forced to process the murder of my friends, the trauma of my community, and the unique attention the world paid to my experience. At the time, Columbine was considered a once-in-a-generation type of tragedy — one that few other people in our country would ever have to contend with. 
  • According to the organization Everytown for Gun Safety, the U.S. has experienced 274 mass shootings since 2009 alone. Thousands of survivors are now part of a club that nobody wants to join.
  • I think about how we vowed to “never forget” Columbine. How we would make sure the next generation would be safer. The opposite has happened.
  • ...8 more annotations...
  • You likely have memories associated with some of these shootings. What you were doing when you heard. Who you were with. How it affected you. When the headlines break, I often hear from friends, family, and colleagues: “I’m so sorry you have to relive this again.” The truth is, we are all reliving it again on some level. The steady cadence of shock, grief and pain is our collective story.
  • So now that the “thoughts and prayers” have been shared, what can we do together? 
  • Politicians love to tell us not to “politicize these tragedies” following mass shootings.
  • Would you care if an earlier tragedy was politicized if it meant getting your son or daughter back? Of course not. Grief doesn’t have a political affiliation. 
  • Then, change the culture. It’s impossible to ignore the role white supremacy, misogyny and extremism plays in so many mass gun violence events. In many mass shootings, the shooter has exhibited dangerous warning signs before the shooting.
  • Because it’s about the guns. The United States has roughly 5 percent of the world’s population and over 30 percent of the world’s mass shootings.
  • And while politicians love to act like this is an impossibly polarizing issue, perhaps we’re not as divided as we think. The majority of Americans actually believe Congress should pass more extensive gun legislation. According to a 2015 Public Policy Polling survey, 83 percent of gun owners support expanded background checks.
  • A new generation has grown up since Columbine; 310,000 more students in the U.S. have experienced gun violence since that day in 1999 when I escaped with my life.
peterconnelly

House Democrats look to pass gun control legislation by early June - 0 views

  • House Democrats will try to advance a raft of gun control bills on Thursday in the wake of two high-profile mass shootings that rocked the nation earlier this month.
  • The Democratic-led package will likely fail in the face of Republican opposition in the Senate. However, Democrats have acknowledged a hope — however slim — that bipartisan talks among senators can lead to lawmakers passing a more limited bill with support from both parties.
  • The Raise the Age Act would lift the purchasing age for semiautomatic rifles from 18 to 21, while the Keep Americans Safe Act would outlaw the import, sale, manufacture, transfer or possession of a large-capacity magazine.
  • ...4 more annotations...
  • Senate Republicans have for years blocked progress on any gun safety legislation. They opposed efforts to tighten gun regulations both when they held the majority, and even now when they can threaten an indefinite filibuster if Democrats can’t come up with the 60 votes required to circumvent the stalling tactic.
  • “It’s much easier to scream about guns than it is to demand answer about where our culture is failing,” Cruz added in a separate social media post on Saturday.
  • Disapproval from Cruz and other Senate Republicans will likely doom any legislation Nadler and other House Democrats manage to pass.
  • Democrats dispute the claim that lawmakers need to target mental illness more so than the availability of guns to reduce shooting violence in the U.S. They say that similar rates of mental illness in other developed nations across the globe prove that mental illness alone cannot fully explain the prevalence of mass shootings in the U.S.
peterconnelly

Twitter launches a crisis misinformation policy - CNN - 0 views

  • Washington (CNN Business)Twitter will now apply warning labels to — and cease recommending — claims that outside experts have identified as misinformation during fast-moving times of crisis, the social media company said Thursday.
  • The platform's new crisis misinformation policy is designed to slow the spread of viral falsehoods during natural disasters, armed conflict and public health emergencies, the company announced.
  • "To determine whether claims are misleading, we require verification from multiple credible, publicly available sources, including evidence from conflict monitoring groups, humanitarian organizations, open-source investigators, journalists, and more," Twitter's head of safety and integrity, Yoel Roth, wrote in a blog post.
  • ...1 more annotation...
  • It also comes amid an ongoing, global battle over the future of platform moderation, with officials in Europe seeking to heighten standards surrounding tech companies' content decision-making and lawmakers in many US states seeking to force platforms to moderate less.
Javier E

'The Power of One,' Facebook whistleblower Frances Haugen's memoir - The Washington Post - 0 views

  • When an internal group proposed the conditions under which Facebook should step in and take down speech from political actors, Zuckerberg discarded its work. He said he’d address the issue himself over a weekend. His “solution”? Facebook would not touch speech by any politician, under any circumstances — a fraught decision under the simplistic surface, as Haugen points out. After all, who gets to count as a politician? The municipal dogcatcher?
  • t was also Zuckerberg, she says, who refused to make a small change that would have made the content in people’s feeds less incendiary — possibly because doing so would have caused a key metric to decline.
  • When the Wall Street Journal’s Jeff Horwitz began to break the stories that Haugen helped him document, the most damning one concerned Facebook’s horrifyingly disingenuous response to a congressional inquiry asking if the company had any research showing that its products were dangerous to teens. Facebook said it wasn’t aware of any consensus indicating how much screen time was too much. What Facebook did have was a pile of research showing that kids were being harmed by its products. Allow a clever company a convenient deflection, and you get something awfully close to a lie.
  • ...5 more annotations...
  • after the military in Myanmar used Facebook to stoke the murder of the Rohingya people, Haugen began to worry that this was a playbook that could be infinitely repeated — and only because Facebook chose not to invest in safety measures, such as detecting hate speech in poorer, more vulnerable places. “The scale of the problems was so vast,” she writes. “I believed people were going to die (in certain countries, at least) and for no reason other than higher profit margins.”
  • After a trip to Cambodia, where neighbors killed neighbors in the 1970s because of a “story that labeled people who had lived next to each other for generations as existential threats,” she’d started to wonder about what caused people to turn on one another to such a horr
  • ifying degree. “How quickly could a story become the truth people perceived?”
  • she points out is the false choice posited by most social media companies: free speech vs. censorship. She argues that lack of transparency is what contributed most to the problems at Facebook. No one on the outside can see inside the algorithms. Even many of those on the inside can’t. “You can’t take a single academic course, anywhere in the world, on the tradeoffs and choices that go into building a social media algorithm or, more importantly, the consequences of those choices,” she writes.
  • In that lack of accountability, social media is a very different ecosystem than the one that helped Ralph Nader take on the auto industry back in the 1960s. Then, there was a network of insurers and plaintiff’s lawyers who also wanted change — and the images of mangled bodies were a lot more visible than what happens inside the mind of a teenage girl. But what if the government forced companies to share their inner workings in the same way it mandates that food companies disclose the nutrition in what they make? What if the government forced social media companies to allow academics and other researchers access to the algorithms they use?
Javier E

Silicon Valley's Safe Space - The New York Times - 0 views

  • The roots of Slate Star Codex trace back more than a decade to a polemicist and self-described A.I. researcher named Eliezer Yudkowsky, who believed that intelligent machines could end up destroying humankind. He was a driving force behind the rise of the Rationalists.
  • Because the Rationalists believed A.I. could end up destroying the world — a not entirely novel fear to anyone who has seen science fiction movies — they wanted to guard against it. Many worked for and donated money to MIRI, an organization created by Mr. Yudkowsky whose stated mission was “A.I. safety.”
  • The community was organized and close-knit. Two Bay Area organizations ran seminars and high-school summer camps on the Rationalist way of thinking.
  • ...27 more annotations...
  • “The curriculum covers topics from causal modeling and probability to game theory and cognitive science,” read a website promising teens a summer of Rationalist learning. “How can we understand our own reasoning, behavior, and emotions? How can we think more clearly and better achieve our goals?”
  • Some lived in group houses. Some practiced polyamory. “They are basically just hippies who talk a lot more about Bayes’ theorem than the original hippies,” said Scott Aaronson, a University of Texas professor who has stayed in one of the group houses.
  • For Kelsey Piper, who embraced these ideas in high school, around 2010, the movement was about learning “how to do good in a world that changes very rapidly.”
  • Yes, the community thought about A.I., she said, but it also thought about reducing the price of health care and slowing the spread of disease.
  • Slate Star Codex, which sprung up in 2013, helped her develop a “calibrated trust” in the medical system. Many people she knew, she said, felt duped by psychiatrists, for example, who they felt weren’t clear about the costs and benefits of certain treatment.
  • That was not the Rationalist way.
  • “There is something really appealing about somebody explaining where a lot of those ideas are coming from and what a lot of the questions are,” she said.
  • Sam Altman, chief executive of OpenAI, an artificial intelligence lab backed by a billion dollars from Microsoft. He was effusive in his praise of the blog.It was, he said, essential reading among “the people inventing the future” in the tech industry.
  • Mr. Altman, who had risen to prominence as the president of the start-up accelerator Y Combinator, moved on to other subjects before hanging up. But he called back. He wanted to talk about an essay that appeared on the blog in 2014.The essay was a critique of what Mr. Siskind, writing as Scott Alexander, described as “the Blue Tribe.” In his telling, these were the people at the liberal end of the political spectrum whose characteristics included “supporting gay rights” and “getting conspicuously upset about sexists and bigots.”
  • But as the man behind Slate Star Codex saw it, there was one group the Blue Tribe could not tolerate: anyone who did not agree with the Blue Tribe. “Doesn’t sound quite so noble now, does it?” he wrote.
  • Mr. Altman thought the essay nailed a big problem: In the face of the “internet mob” that guarded against sexism and racism, entrepreneurs had less room to explore new ideas. Many of their ideas, such as intelligence augmentation and genetic engineering, ran afoul of the Blue Tribe.
  • Mr. Siskind was not a member of the Blue Tribe. He was not a voice from the conservative Red Tribe (“opposing gay marriage,” “getting conspicuously upset about terrorists and commies”). He identified with something called the Grey Tribe — as did many in Silicon Valley.
  • The Grey Tribe was characterized by libertarian beliefs, atheism, “vague annoyance that the question of gay rights even comes up,” and “reading lots of blogs,” he wrote. Most significantly, it believed in absolute free speech.
  • The essay on these tribes, Mr. Altman told me, was an inflection point for Silicon Valley. “It was a moment that people talked about a lot, lot, lot,” he said.
  • And in some ways, two of the world’s prominent A.I. labs — organizations that are tackling some of the tech industry’s most ambitious and potentially powerful projects — grew out of the Rationalist movement.
  • In 2005, Peter Thiel, the co-founder of PayPal and an early investor in Facebook, befriended Mr. Yudkowsky and gave money to MIRI. In 2010, at Mr. Thiel’s San Francisco townhouse, Mr. Yudkowsky introduced him to a pair of young researchers named Shane Legg and Demis Hassabis. That fall, with an investment from Mr. Thiel’s firm, the two created an A.I. lab called DeepMind.
  • Like the Rationalists, they believed that A.I could end up turning against humanity, and because they held this belief, they felt they were among the only ones who were prepared to build it in a safe way.
  • In 2014, Google bought DeepMind for $650 million. The next year, Elon Musk — who also worried A.I. could destroy the world and met his partner, Grimes, because they shared an interest in a Rationalist thought experiment — founded OpenAI as a DeepMind competitor. Both labs hired from the Rationalist community.
  • Mr. Aaronson, the University of Texas professor, was turned off by the more rigid and contrarian beliefs of the Rationalists, but he is one of the blog’s biggest champions and deeply admired that it didn’t avoid live-wire topics.
  • “It must have taken incredible guts for Scott to express his thoughts, misgivings and questions about some major ideological pillars of the modern world so openly, even if protected by a quasi-pseudonym,” he said
  • In late June of last year, not long after talking to Mr. Altman, the OpenAI chief executive, I approached the writer known as Scott Alexander, hoping to get his views on the Rationalist way and its effect on Silicon Valley. That was when the blog vanished.
  • The issue, it was clear to me, was that I told him I could not guarantee him the anonymity he’d been writing with. In fact, his real name was easy to find because people had shared it online for years and he had used it on a piece he’d written for a scientific journal. I did a Google search for Scott Alexander and one of the first results I saw in the auto-complete list was Scott Alexander Siskind.
  • More than 7,500 people signed a petition urging The Times not to publish his name, including many prominent figures in the tech industry. “Putting his full name in The Times,” the petitioners said, “would meaningfully damage public discourse, by discouraging private citizens from sharing their thoughts in blog form.” On the internet, many in Silicon Valley believe, everyone has the right not only to say what they want but to say it anonymously.
  • I spoke with Manoel Horta Ribeiro, a computer science researcher who explores social networks at the Swiss Federal Institute of Technology in Lausanne. He was worried that Slate Star Codex, like other communities, was allowing extremist views to trickle into the influential tech world. “A community like this gives voice to fringe groups,” he said. “It gives a platform to people who hold more extreme views.”
  • I assured her my goal was to report on the blog, and the Rationalists, with rigor and fairness. But she felt that discussing both critics and supporters could be unfair. What I needed to do, she said, was somehow prove statistically which side was right.
  • When I asked Mr. Altman if the conversation on sites like Slate Star Codex could push people toward toxic beliefs, he said he held “some empathy” for these concerns. But, he added, “people need a forum to debate ideas.”
  • In August, Mr. Siskind restored his old blog posts to the internet. And two weeks ago, he relaunched his blog on Substack, a company with ties to both Andreessen Horowitz and Y Combinator. He gave the blog a new title: Astral Codex Ten. He hinted that Substack paid him $250,000 for a year on the platform. And he indicated the company would give him all the protection he needed.
Javier E

Musk, SBF, and the Myth of Smug, Castle-Building Nerds - 0 views

  • Experts in content moderation suggested that Musk’s actual policies lacked any coherence and, if implemented, would have all kinds of unintended consequences. That has happened with verification. Almost every decision he makes is an unforced error made with extreme confidence in front of a growing audience of people who already know he has messed up, and is supported by a network of sycophants and blind followers who refuse to see or tell him that he’s messing up. The dynamic is … very Trumpy!
  • As with the former president, it can be hard at times for people to believe or accept that our systems are so broken that a guy who is clearly this inept can also be put in charge of something so important. A common pundit claim before Donald Trump got into the White House was that the gravity of the job and prestige of the office might humble or chasten him.
  • The same seems true for Musk. Even people skeptical of Musk’s behavior pointed to his past companies as predictors of future success. He’s rich. He does smart-people stuff. The rockets land pointy-side up!
  • ...18 more annotations...
  • Time and again, we learned there was never a grand plan or big ideas—just weapons-grade ego, incompetence, thin skin, and prejudice against those who don’t revere him.
  • Despite all the incredible, damning reporting coming out of Twitter and all of Musk’s very public mistakes, many people still refuse to believe—even if they detest him—that he is simply incompetent.
  • What is amazing about the current moment is that, despite how ridiculous it all feels, a fundamental tenet of reality and logic appears to be holding true: If you don’t know what you’re doing or don’t really care, you’ll run the thing you’re in charge of into the ground, and people will notice.
  • And so the moment feels too dumb and too on the nose to be real and yet also very real—kind of like all of reality in 2022.
  • I don’t really know where any of this will lead, but one interesting possibility is that Musk gets increasingly reactionary and trollish in his politics and stewardship of Twitter.
  • Leaving the politics aside, from a basic customer-service standpoint this is generally an ill-advised way for the owner of a company to treat an elected official when that elected official wishes to know why your service has failed them. The reason it is ill-advised is because then the elected official could tweet something like what Senator Markey tweeted on Sunday: “One of your companies is under an FTC consent decree. Auto safety watchdog NHTSA is investigating another for killing people. And you’re spending your time picking fights online. Fix your companies. Or Congress will.”
  • It seems clear that Musk, like any dedicated social-media poster, thrives on validation, so it makes sense that, as he continues to dismantle his own mystique as an innovator, he might look for adoration elsewhere
  • Recent history has shown that, for a specific audience, owning the libs frees a person from having to care about competency or outcome of their actions. Just anger the right people and you’re good, even if you’re terrible at your job. This won’t help Twitter’s financial situation, which seems bleak, but it’s … something!
  • Bankman-Fried, the archetype, appealed to people for all kinds of reasons. His narrative as a philanthropist, and a smart rationalist, and a stone-cold weirdo was something people wanted to buy into because, generally, people love weirdos who don’t conform to systems and then find clever ways to work around them and become wildly successful as a result.
  • Bankman-Fried was a way that a lot of people could access and maybe obliquely understand what was going on in crypto. They may not have understood what FTX did, but they could grasp a nerd trying to leverage a system in order to do good in the world and advance progressive politics. In that sense, Bankman-Fried is easy to root for and exciting to cover. His origin story and narrative become more important than the particulars of what he may or may not be doing.
  • the past few weeks have been yet another reminder that the smug-nerd-genius narrative may sell magazines, and it certainly raises venture funding, but the visionary founder is, first and foremost, a marketing product, not a reality. It’s a myth that perpetuates itself. Once branded a visionary, the founder can use the narrative to raise money and generate a formidable net worth, and then the financial success becomes its own résumé. But none of it is real.
  • Adversarial journalism ideally questions and probes power. If it is trained on technology companies and their founders, it is because they either wield that power or have the potential to do so. It is, perhaps unintuitively, a form of respect for their influence and potential to disrupt. But that’s not what these founders want.
  • even if all tech coverage had been totally flawless, Silicon Valley would have rejected adversarial tech journalism because most of its players do not actually want the responsibility that comes with their potential power. They want only to embody the myth and reap the benefits. They want the narrative, which is focused on origins, ambitions, ethos, and marketing, and less on the externalities and outcomes.
  • Looking at Musk and Bankman-Fried, it would appear that the tech visionaries mostly get their way. For all the complaints of awful, negative coverage and biased reporting, people still want to cheer for and give money to the “‘smug nerds building castles in the sky.’” Though they vary wildly right now in magnitude, their wounds are self-inflicted—and, perhaps, the result of believing their own hype.
  • That’s because, almost always, the smug-nerd-genius narrative is a trap. It’s one that people fall into because they need to believe that somebody out there is so brilliant, they can see the future, or that they have some greater, more holistic understanding of the world (or that such an understanding is possible)
  • It’s not unlike a conspiracy theory in that way. The smug-nerd-genius narrative helps take the complexity of the world and make it more manageable.
  • Putting your faith in a space billionaire or a crypto wunderkind isn’t just sad fanboydom; it is also a way for people to outsource their brain to somebody else who, they believe, can see what they can’t
  • the smug nerd genius is exceedingly rare, and, even when they’re not outed as a fraud or a dilettante, they can be assholes or flawed like anyone else. There aren’t shortcuts for making sense of the world, and anyone who is selling themselves that way or buying into that narrative about them should read to us as a giant red flag.
Javier E

Roger Scruton and the Fascists Who Love Him - 0 views

  • Scruton was a true intellectual, that his writing extended far beyond political commentary into various fields of philosophy and the arts, and that his reputation was that of a gentleman
  • reading Scruton’s critique of liberalism from the safety of, say, 1995, with communism vanquished, liberalism ascendant, and Europe beginning to heal from an 80-year-old wound is one thing.
  • Reading Scruton’s critique of liberalism today, with right-wing illiberalism on the march both at home and abroad, is quite another.
  • ...7 more annotations...
  • Scruton’s argument in many of his essays and books amounted to a deep critique of liberalism as mistaken about human beings, about society, about politics. That critique was especially valuable when it could be read as a friendly corrective to liberalism’s errors, excesses, and contradictions
  • today, with liberalism under threat, it comes across more like an indictment of liberalism—an indictment that has apparently been taken up as a foundational text by fascists.
  • Pappin extols Hungary as “a traditional Christian society,” going on to say “as an anti-liberal, I think that’s good.” Pappin then defends altering the Constitution to tilt power toward the right and strip protections from groups he feels have undermined American traditional values.
  • at some point, you have to start asking hard questions. In art, we divorce the work from both its creator and its legacy. You judge the work for the work and do not hold it responsible if the artist, or its fans, turn out to be bad people.
  • I’m not certain that this is how it is—or should be—in the world of ideas.
  • In Scruton’s place and time, it did seem like liberalism was ascendant and that its overreach and failings needed conservative correction.
  • In our day, though, liberalism needs correction less than it needs protection—including protection from the would-be authoritarians sipping espresso in the Scruton café.
Javier E

Opinion | A Nobel Prize for the Economics of Panic - The New York Times - 0 views

  • Obviously, Bernanke, Diamond and Dybvig weren’t the first economists to notice that bank runs happen
  • Diamond and Dybvig provided the first really clear analysis of why they happen — and why, destructive as they are, they can represent rational behavior on the part of bank depositors. Their analysis was also full of implications for financial policy.
  • Bernanke provided evidence on why bank runs matter and, although he avoided saying so directly, why Milton Friedman was wrong about the causes of the Great Depression.
  • ...20 more annotations...
  • Diamond and Dybvig offered a stylized but insightful model of what banks do. They argued that there is always a tension between individuals’ desire for liquidity — ready access to funds — and the economy’s need to make long-term investments that can’t easily be converted into cash.
  • Banks square that circle by taking money from depositors who can withdraw their funds at will — making those deposits highly liquid — and investing most of that money in illiquid assets, such as business loans.
  • So banking is a productive activity that makes the economy richer by reconciling otherwise incompatible desires for liquidity and productive investment. And it normally works because only a fraction of a bank’s depositors want to withdraw their funds at any given time.
  • This does, however, make banks vulnerable to runs. Suppose that for some reason many depositors come to believe that many other depositors are about to cash out, and try to beat the pack by withdrawing their own funds. To meet these demands for liquidity, a bank will have to sell off its illiquid assets at fire sale prices, and doing so can drive an institution that should be solvent into bankruptcy
  • If that happens, people who didn’t withdraw their funds will be left with nothing. So during a panic, the rational thing to do is to panic along with everyone else.
  • There was, of course, a huge wave of banking panics in 1930-31. Many banks failed, and those that survived made far fewer business loans than before, holding cash instead, while many families shunned banks altogether, putting their cash in safes or under their mattresses. The result was a diversion of wealth into unproductive uses. In his 1983 paper, Bernanke offered evidence that this diversion played a large role in driving the economy into a depression and held back the subsequent recovery.
  • In the story told by Friedman and Anna Schwartz, the banking crisis of the early 1930s was damaging because it led to a fall in the money supply — currency plus bank deposits. Bernanke asserted that this was at most only part of the stor
  • a government backstop — either deposit insurance, the willingness of the central bank to lend money to troubled banks or both — can short-circuit potential crises.
  • Such arrangements offered a higher yield than conventional deposits. But they had no safety net, which opened the door to an old-style bank run and financial panic.
  • So banks need to be regulated as well as backstopped. As I said, the Diamond-Dybvig analysis had remarkably large implications for policy.
  • From an economic point of view, banking is any form of financial intermediation that offers people seemingly liquid assets while using their wealth to make illiquid investments.
  • This insight was dramatically validated in the 2008 financial crisis.
  • By the eve of the crisis, however, the financial system relied heavily on “shadow banking” — banklike activities that didn’t involve standard bank deposits
  • But providing such a backstop raises the possibility of abuse; banks may take on undue risks because they know they’ll be bailed out if things go wrong.
  • And the panic came. The conventionally measured money supply didn’t plunge in 2008 the way it did in the 1930s — but repo and other money-like liabilities of financial intermediaries did:
  • Fortunately, by then Bernanke was chair of the Federal Reserve. He understood what was going on, and the Fed stepped in on an immense scale to prop up the financial system.
  • a sort of meta point about the Diamond-Dybvig work: Once you’ve understood and acknowledged the possibility of self-fulfilling banking crises, you become aware that similar things can happen elsewhere.
  • Perhaps the most notable case in relatively recent times was the euro crisis of 2010-12. Market confidence in the economies of southern Europe collapsed, leading to huge spreads between the interest rates on, for example, Portuguese bonds and those on German bonds. The conventional wisdom at the time — especially in Germany — was that countries were being justifiably punished for taking on excessive debt
  • the Belgian economist Paul De Grauwe argued that what was actually happening was a self-fulfilling panic — basically a run on the bonds of countries that couldn’t provide a backstop because they no longer had their own currencies.
  • Sure enough, when Mario Draghi, the president of the European Central Bank at the time, finally did provide a backstop in 2012 — he said the magic words “whatever it takes,” implying that the bank would lend money to the troubled governments if necessary — the spreads collapsed and the crisis came to an end:
Javier E

'The Godfather of AI' Quits Google and Warns of Danger Ahead - The New York Times - 0 views

  • he officially joined a growing chorus of critics who say those companies are racing toward danger with their aggressive campaign to create products based on generative artificial intelligence, the technology that powers popular chatbots like ChatGPT.
  • Dr. Hinton said he has quit his job at Google, where he has worked for more than decade and became one of the most respected voices in the field, so he can freely speak out about the risks of A.I. A part of him, he said, now regrets his life’s work.
  • “I console myself with the normal excuse: If I hadn’t done it, somebody else would have,”
  • ...24 more annotations...
  • Industry leaders believe the new A.I. systems could be as important as the introduction of the web browser in the early 1990s and could lead to breakthroughs in areas ranging from drug research to education.
  • But gnawing at many industry insiders is a fear that they are releasing something dangerous into the wild. Generative A.I. can already be a tool for misinformation. Soon, it could be a risk to jobs. Somewhere down the line, tech’s biggest worriers say, it could be a risk to humanity.
  • “It is hard to see how you can prevent the bad actors from using it for bad things,” Dr. Hinton said.
  • After the San Francisco start-up OpenAI released a new version of ChatGPT in March, more than 1,000 technology leaders and researchers signed an open letter calling for a six-month moratorium on the development of new systems because A.I technologies pose “profound risks to society and humanity.
  • Several days later, 19 current and former leaders of the Association for the Advancement of Artificial Intelligence, a 40-year-old academic society, released their own letter warning of the risks of A.I. That group included Eric Horvitz, chief scientific officer at Microsoft, which has deployed OpenAI’s technology across a wide range of products, including its Bing search engine.
  • Dr. Hinton, often called “the Godfather of A.I.,” did not sign either of those letters and said he did not want to publicly criticize Google or other companies until he had quit his job
  • Dr. Hinton, a 75-year-old British expatriate, is a lifelong academic whose career was driven by his personal convictions about the development and use of A.I. In 1972, as a graduate student at the University of Edinburgh, Dr. Hinton embraced an idea called a neural network. A neural network is a mathematical system that learns skills by analyzing data. At the time, few researchers believed in the idea. But it became his life’s work.
  • Dr. Hinton is deeply opposed to the use of artificial intelligence on the battlefield — what he calls “robot soldiers.”
  • In 2012, Dr. Hinton and two of his students in Toronto, Ilya Sutskever and Alex Krishevsky, built a neural network that could analyze thousands of photos and teach itself to identify common objects, such as flowers, dogs and cars.
  • In 2018, Dr. Hinton and two other longtime collaborators received the Turing Award, often called “the Nobel Prize of computing,” for their work on neural networks.
  • Around the same time, Google, OpenAI and other companies began building neural networks that learned from huge amounts of digital text. Dr. Hinton thought it was a powerful way for machines to understand and generate language, but it was inferior to the way humans handled language.
  • Then, last year, as Google and OpenAI built systems using much larger amounts of data, his view changed. He still believed the systems were inferior to the human brain in some ways but he thought they were eclipsing human intelligence in others.
  • “Maybe what is going on in these systems,” he said, “is actually a lot better than what is going on in the brain.”
  • As companies improve their A.I. systems, he believes, they become increasingly dangerous. “Look at how it was five years ago and how it is now,” he said of A.I. technology. “Take the difference and propagate it forwards. That’s scary.”
  • Until last year, he said, Google acted as a “proper steward” for the technology, careful not to release something that might cause harm. But now that Microsoft has augmented its Bing search engine with a chatbot — challenging Google’s core business — Google is racing to deploy the same kind of technology. The tech giants are locked in a competition that might be impossible to stop, Dr. Hinton said.
  • His immediate concern is that the internet will be flooded with false photos, videos and text, and the average person will “not be able to know what is true anymore.”
  • He is also worried that A.I. technologies will in time upend the job market. Today, chatbots like ChatGPT tend to complement human workers, but they could replace paralegals, personal assistants, translators and others who handle rote tasks. “It takes away the drudge work,” he said. “It might take away more than that.”
  • Down the road, he is worried that future versions of the technology pose a threat to humanity because they often learn unexpected behavior from the vast amounts of data they analyze. This becomes an issue, he said, as individuals and companies allow A.I. systems not only to generate their own computer code but actually run that code on their ow
  • And he fears a day when truly autonomous weapons — those killer robots — become reality.
  • “The idea that this stuff could actually get smarter than people — a few people believed that,” he said. “But most people thought it was way off. And I thought it was way off. I thought it was 30 to 50 years or even longer away. Obviously, I no longer think that.”
  • Many other experts, including many of his students and colleagues, say this threat is hypothetical. But Dr. Hinton believes that the race between Google and Microsoft and others will escalate into a global race that will not stop without some sort of global regulation.
  • But that may be impossible, he said. Unlike with nuclear weapons, he said, there is no way of knowing whether companies or countries are working on the technology in secret. The best hope is for the world’s leading scientists to collaborate on ways of controlling the technology. “I don’t think they should scale this up more until they have understood whether they can control it,” he said.
  • Dr. Hinton said that when people used to ask him how he could work on technology that was potentially dangerous, he would paraphrase Robert Oppenheimer, who led the U.S. effort to build the atomic bomb: “When you see something that is technically sweet, you go ahead and do it.”
  • He does not say that anymore.
Javier E

Some on the Left Turn Against the Label 'Progressive' - The New York Times - 0 views

  • Christopher Lasch, the historian and social critic, posed a banger of a question in his 1991 book, “The True and Only Heaven: Progress and Its Critics.”
  • “How does it happen,” Lasch asked, “that serious people continue to believe in progress, in the face of massive evidence that might have been expected to refute the idea of progress once and for all?”
  • A review in The New York Times Book Review by William Julius Wilson, a professor at Harvard, was titled: “Where Has Progress Got Us?”
  • ...17 more annotations...
  • Essentially, Lasch was attacking the notion, fashionable as Americans basked in their seeming victory over the Soviet Union in the Cold War, that history had a direction — and that one would be wise to stay on the “right side” of it.
  • Francis Fukuyama expressed a version of this triumphalist idea in his famous 1992 book, “The End of History and the Last Man,” in which he celebrated the notion that History with a capital “H,” in the sense of a battle between competing ideas, was ending with communism left to smolder on Ronald Reagan’s famous ash heap.
  • One of Martin Luther King Jr.’s most frequently quoted lines speaks to a similar thought, albeit in a different context: “T​he arc of the moral universe is long, but it bends toward justice.” Though he had read Lasch, Obama quoted that line often, just as he liked to say that so-and-so would be “on the wrong side of history” if they didn’t live up to his ideals — whether the issue was same-sex marriage, health policy or the Russian occupation of Crimea.
  • The memo goes on to list two sets of words: “Optimistic Positive Governing Words” and “Contrasting Words,” which carried negative connotations. One of the latter group was the word “liberal,” sandwiched between “intolerant” and “lie.”
  • So what’s the difference between a progressive and a liberal?To vastly oversimplify matters, liberal usually refers to someone on the center-left on a two-dimensional political spectrum, while progressive refers to someone further left.
  • But “liberal” has taken a beating in recent decades — from both left and right.
  • In the late 1980s and 1990s, Republicans successfully demonized the word “liberal,” to the point where many Democrats shied away from it in favor of labels like “conservative Democrat” or, more recently, “progressive.”
  • “Is the story of the 20th century about the defeat of the Soviet Union, or was it about two world wars and a Holocaust?” asked Matthew Sitman, the co-host of the “Know Your Enemy” podcast, which recently hosted a discussion on Lasch and the fascination many conservatives have with his ideas. “It really depends on how you look at it.”
  • None of this was an accident. In 1996, Representative Newt Gingrich of Georgia circulated a now-famous memo called “Language: A Key Mechanism of Control.”
  • The authors urged their readers: “The words and phrases are powerful. Read them. Memorize as many as possible.”
  • Republicans subsequently had a great deal of success in associating the term “liberal” with other words and phrases voters found uncongenial: wasteful spending, high rates of taxation and libertinism that repelled socially conservative voters.
  • Many on the left began identifying themselves as “progressive” — which had the added benefit of harking back to movements of the late 19th and early 20th centuries that fought against corruption, opposed corporate monopolies, pushed for good-government reforms and food safety and labor laws and established women’s right to vote.
  • Allies of Bill Clinton founded the Progressive Policy Institute, a think tank associated with so-called Blue Dog Democrats from the South.
  • Now, scrambling the terminology, groups like the Progressive Change Campaign Committee agitate on behalf of proudly left-wing candidates
  • In 2014, Charles Murray, the polarizing conservative scholar, urged readers of The Wall Street Journal’s staunchly right-wing editorial section to “start using ‘liberal’ to designate the good guys on the left, reserving ‘progressive’ for those who are enthusiastic about an unrestrained regulatory state.”
  • As Sanders and acolytes like Representative Alexandria Ocasio-Cortez of New York have gained prominence over the last few election cycles, many on the left-wing end of the spectrum have begun proudly applying other labels to themselves, such as “democratic socialist.”
  • To little avail so far, Kazin, the Georgetown historian, has been urging them to call themselves “social democrats” instead — as many mainstream parties do in Europe.“It’s not a good way to win elections in this country, to call yourself a socialist,” he said.
Javier E

Whistleblower: Twitter misled investors, FTC and underplayed spam issues - Washington Post - 0 views

  • Twitter executives deceived federal regulators and the company’s own board of directors about “extreme, egregious deficiencies” in its defenses against hackers, as well as its meager efforts to fight spam, according to an explosive whistleblower complaint from its former security chief.
  • The complaint from former head of security Peiter Zatko, a widely admired hacker known as “Mudge,” depicts Twitter as a chaotic and rudderless company beset by infighting, unable to properly protect its 238 million daily users including government agencies, heads of state and other influential public figures.
  • Among the most serious accusations in the complaint, a copy of which was obtained by The Washington Post, is that Twitter violated the terms of an 11-year-old settlement with the Federal Trade Commission by falsely claiming that it had a solid security plan. Zatko’s complaint alleges he had warned colleagues that half the company’s servers were running out-of-date and vulnerable software and that executives withheld dire facts about the number of breaches and lack of protection for user data, instead presenting directors with rosy charts measuring unimportant changes.
  • ...56 more annotations...
  • The complaint — filed last month with the Securities and Exchange Commission and the Department of Justice, as well as the FTC — says thousands of employees still had wide-ranging and poorly tracked internal access to core company software, a situation that for years had led to embarrassing hacks, including the commandeering of accounts held by such high-profile users as Elon Musk and former presidents Barack Obama and Donald Trump.
  • the whistleblower document alleges the company prioritized user growth over reducing spam, though unwanted content made the user experience worse. Executives stood to win individual bonuses of as much as $10 million tied to increases in daily users, the complaint asserts, and nothing explicitly for cutting spam.
  • Chief executive Parag Agrawal was “lying” when he tweeted in May that the company was “strongly incentivized to detect and remove as much spam as we possibly can,” the complaint alleges.
  • Zatko described his decision to go public as an extension of his previous work exposing flaws in specific pieces of software and broader systemic failings in cybersecurity. He was hired at Twitter by former CEO Jack Dorsey in late 2020 after a major hack of the company’s systems.
  • “I felt ethically bound. This is not a light step to take,” said Zatko, who was fired by Agrawal in January. He declined to discuss what happened at Twitter, except to stand by the formal complaint. Under SEC whistleblower rules, he is entitled to legal protection against retaliation, as well as potential monetary rewards.
  • “Security and privacy have long been top companywide priorities at Twitter,” said Twitter spokeswoman Rebecca Hahn. She said that Zatko’s allegations appeared to be “riddled with inaccuracies” and that Zatko “now appears to be opportunistically seeking to inflict harm on Twitter, its customers, and its shareholders.” Hahn said that Twitter fired Zatko after 15 months “for poor performance and leadership.” Attorneys for Zatko confirmed he was fired but denied it was for performance or leadership.
  • A person familiar with Zatko’s tenure said the company investigated Zatko’s security claims during his time there and concluded they were sensationalistic and without merit. Four people familiar with Twitter’s efforts to fight spam said the company deploys extensive manual and automated tools to both measure the extent of spam across the service and reduce it.
  • Overall, Zatko wrote in a February analysis for the company attached as an exhibit to the SEC complaint, “Twitter is grossly negligent in several areas of information security. If these problems are not corrected, regulators, media and users of the platform will be shocked when they inevitably learn about Twitter’s severe lack of security basics.”
  • Zatko’s complaint says strong security should have been much more important to Twitter, which holds vast amounts of sensitive personal data about users. Twitter has the email addresses and phone numbers of many public figures, as well as dissidents who communicate over the service at great personal risk.
  • This month, an ex-Twitter employee was convicted of using his position at the company to spy on Saudi dissidents and government critics, passing their information to a close aide of Crown Prince Mohammed bin Salman in exchange for cash and gifts.
  • Zatko’s complaint says he believed the Indian government had forced Twitter to put one of its agents on the payroll, with access to user data at a time of intense protests in the country. The complaint said supporting information for that claim has gone to the National Security Division of the Justice Department and the Senate Select Committee on Intelligence. Another person familiar with the matter agreed that the employee was probably an agent.
  • “Take a tech platform that collects massive amounts of user data, combine it with what appears to be an incredibly weak security infrastructure and infuse it with foreign state actors with an agenda, and you’ve got a recipe for disaster,” Charles E. Grassley (R-Iowa), the top Republican on the Senate Judiciary Committee,
  • Many government leaders and other trusted voices use Twitter to spread important messages quickly, so a hijacked account could drive panic or violence. In 2013, a captured Associated Press handle falsely tweeted about explosions at the White House, sending the Dow Jones industrial average briefly plunging more than 140 points.
  • After a teenager managed to hijack the verified accounts of Obama, then-candidate Joe Biden, Musk and others in 2020, Twitter’s chief executive at the time, Jack Dorsey, asked Zatko to join him, saying that he could help the world by fixing Twitter’s security and improving the public conversation, Zatko asserts in the complaint.
  • In 1998, Zatko had testified to Congress that the internet was so fragile that he and others could take it down with a half-hour of concentrated effort. He later served as the head of cyber grants at the Defense Advanced Research Projects Agency, the Pentagon innovation unit that had backed the internet’s invention.
  • But at Twitter Zatko encountered problems more widespread than he realized and leadership that didn’t act on his concerns, according to the complaint.
  • Twitter’s difficulties with weak security stretches back more than a decade before Zatko’s arrival at the company in November 2020. In a pair of 2009 incidents, hackers gained administrative control of the social network, allowing them to reset passwords and access user data. In the first, beginning around January of that year, hackers sent tweets from the accounts of high-profile users, including Fox News and Obama.
  • Several months later, a hacker was able to guess an employee’s administrative password after gaining access to similar passwords in their personal email account. That hacker was able to reset at least one user’s password and obtain private information about any Twitter user.
  • Twitter continued to suffer high-profile hacks and security violations, including in 2017, when a contract worker briefly took over Trump’s account, and in the 2020 hack, in which a Florida teen tricked Twitter employees and won access to verified accounts. Twitter then said it put additional safeguards in place.
  • This year, the Justice Department accused Twitter of asking users for their phone numbers in the name of increased security, then using the numbers for marketing. Twitter agreed to pay a $150 million fine for allegedly breaking the 2011 order, which barred the company from making misrepresentations about the security of personal data.
  • After Zatko joined the company, he found it had made little progress since the 2011 settlement, the complaint says. The complaint alleges that he was able to reduce the backlog of safety cases, including harassment and threats, from 1 million to 200,000, add staff and push to measure results.
  • But Zatko saw major gaps in what the company was doing to satisfy its obligations to the FTC, according to the complaint. In Zatko’s interpretation, according to the complaint, the 2011 order required Twitter to implement a Software Development Life Cycle program, a standard process for making sure new code is free of dangerous bugs. The complaint alleges that other employees had been telling the board and the FTC that they were making progress in rolling out that program to Twitter’s systems. But Zatko alleges that he discovered that it had been sent to only a tenth of the company’s projects, and even then treated as optional.
  • “If all of that is true, I don’t think there’s any doubt that there are order violations,” Vladeck, who is now a Georgetown Law professor, said in an interview. “It is possible that the kinds of problems that Twitter faced eleven years ago are still running through the company.”
  • The complaint also alleges that Zatko warned the board early in his tenure that overlapping outages in the company’s data centers could leave it unable to correctly restart its servers. That could have left the service down for months, or even have caused all of its data to be lost. That came close to happening in 2021, when an “impending catastrophic” crisis threatened the platform’s survival before engineers were able to save the day, the complaint says, without providing further details.
  • One current and one former employee recalled that incident, when failures at two Twitter data centers drove concerns that the service could have collapsed for an extended period. “I wondered if the company would exist in a few days,” one of them said.
  • The current and former employees also agreed with the complaint’s assertion that past reports to various privacy regulators were “misleading at best.”
  • For example, they said the company implied that it had destroyed all data on users who asked, but the material had spread so widely inside Twitter’s networks, it was impossible to know for sure
  • As the head of security, Zatko says he also was in charge of a division that investigated users’ complaints about accounts, which meant that he oversaw the removal of some bots, according to the complaint. Spam bots — computer programs that tweet automatically — have long vexed Twitter. Unlike its social media counterparts, Twitter allows users to program bots to be used on its service: For example, the Twitter account @big_ben_clock is programmed to tweet “Bong Bong Bong” every hour in time with Big Ben in London. Twitter also allows people to create accounts without using their real identities, making it harder for the company to distinguish between authentic, duplicate and automated accounts.
  • In the complaint, Zatko alleges he could not get a straight answer when he sought what he viewed as an important data point: the prevalence of spam and bots across all of Twitter, not just among monetizable users.
  • Zatko cites a “sensitive source” who said Twitter was afraid to determine that number because it “would harm the image and valuation of the company.” He says the company’s tools for detecting spam are far less robust than implied in various statements.
  • “Agrawal’s Tweets and Twitter’s previous blog posts misleadingly imply that Twitter employs proactive, sophisticated systems to measure and block spam bots,” the complaint says. “The reality: mostly outdated, unmonitored, simple scripts plus overworked, inefficient, understaffed, and reactive human teams.”
  • The four people familiar with Twitter’s spam and bot efforts said the engineering and integrity teams run software that samples thousands of tweets per day, and 100 accounts are sampled manually.
  • Some employees charged with executing the fight agreed that they had been short of staff. One said top executives showed “apathy” toward the issue.
  • Zatko’s complaint likewise depicts leadership dysfunction, starting with the CEO. Dorsey was largely absent during the pandemic, which made it hard for Zatko to get rulings on who should be in charge of what in areas of overlap and easier for rival executives to avoid collaborating, three current and former employees said.
  • For example, Zatko would encounter disinformation as part of his mandate to handle complaints, according to the complaint. To that end, he commissioned an outside report that found one of the disinformation teams had unfilled positions, yawning language deficiencies, and a lack of technical tools or the engineers to craft them. The authors said Twitter had no effective means of dealing with consistent spreaders of falsehoods.
  • Dorsey made little effort to integrate Zatko at the company, according to the three employees as well as two others familiar with the process who spoke on the condition of anonymity to describe sensitive dynamics. In 12 months, Zatko could manage only six one-on-one calls, all less than 30 minutes, with his direct boss Dorsey, who also served as CEO of payments company Square, now known as Block, according to the complaint. Zatko allegedly did almost all of the talking, and Dorsey said perhaps 50 words in the entire year to him. “A couple dozen text messages” rounded out their electronic communication, the complaint alleges.
  • Faced with such inertia, Zatko asserts that he was unable to solve some of the most serious issues, according to the complaint.
  • Some 30 percent of company laptops blocked automatic software updates carrying security fixes, and thousands of laptops had complete copies of Twitter’s source code, making them a rich target for hackers, it alleges.
  • A successful hacker takeover of one of those machines would have been able to sabotage the product with relative ease, because the engineers pushed out changes without being forced to test them first in a simulated environment, current and former employees said.
  • “It’s near-incredible that for something of that scale there would not be a development test environment separate from production and there would not be a more controlled source-code management process,” said Tony Sager, former chief operating officer at the cyberdefense wing of the National Security Agency, the Information Assurance divisio
  • Sager is currently senior vice president at the nonprofit Center for Internet Security, where he leads a consensus effort to establish best security practices.
  • Zatko stopped the material from being presented at the Dec. 9, 2021 meeting, the complaint said. But over his continued objections, Agrawal let it go to the board’s smaller Risk Committee a week later.
  • “A best practice is that you should only be authorized to see and access what you need to do your job, and nothing else,” said former U.S. chief information security officer Gregory Touhill. “If half the company has access to and can make configuration changes to the production environment, that exposes the company and its customers to significant risk.”
  • The complaint says Dorsey never encouraged anyone to mislead the board about the shortcomings, but that others deliberately left out bad news.
  • The complaint says that about half of Twitter’s roughly 7,000 full-time employees had wide access to the company’s internal software and that access was not closely monitored, giving them the ability to tap into sensitive data and alter how the service worked. Three current and former employees agreed that these were issues.
  • An unnamed executive had prepared a presentation for the new CEO’s first full board meeting, according to the complaint. Zatko’s complaint calls the presentation deeply misleading.
  • The presentation showed that 92 percent of employee computers had security software installed — without mentioning that those installations determined that a third of the machines were insecure, according to the complaint.
  • Another graphic implied a downward trend in the number of people with overly broad access, based on the small subset of people who had access to the highest administrative powers, known internally as “God mode.” That number was in the hundreds. But the number of people with broad access to core systems, which Zatko had called out as a big problem after joining, had actually grown slightly and remained in the thousands.
  • The presentation included only a subset of serious intrusions or other security incidents, from a total Zatko estimated as one per week, and it said that the uncontrolled internal access to core systems was responsible for just 7 percent of incidents, when Zatko calculated the real proportion as 60 percent.
  • When Dorsey left in November 2021, a difficult situation worsened under Agrawal, who had been responsible for security decisions as chief technology officer before Zatko’s hiring, the complaint says.
  • Agrawal didn’t respond to requests for comment. In an email to employees after publication of this article, obtained by The Post, he said that privacy and security continues to be a top priority for the company, and he added that the narrative is “riddled with inconsistences” and “presented without important context.”
  • On Jan. 4, Zatko reported internally that the Risk Committee meeting might have been fraudulent, which triggered an Audit Committee investigation.
  • Agarwal fired him two weeks later. But Zatko complied with the company’s request to spell out his concerns in writing, even without access to his work email and documents, according to the complaint.
  • Since Zatko’s departure, Twitter has plunged further into chaos with Musk’s takeover, which the two parties agreed to in May. The stock price has fallen, many employees have quit, and Agrawal has dismissed executives and frozen big projects.
  • Zatko said he hoped that by bringing new scrutiny and accountability, he could improve the company from the outside.
  • “I still believe that this is a tremendous platform, and there is huge value and huge risk, and I hope that looking back at this, the world will be a better place, in part because of this.”
Javier E

Cognitive Biases and the Human Brain - The Atlantic - 1 views

  • Present bias shows up not just in experiments, of course, but in the real world. Especially in the United States, people egregiously undersave for retirement—even when they make enough money to not spend their whole paycheck on expenses, and even when they work for a company that will kick in additional funds to retirement plans when they contribute.
  • hen people hear the word bias, many if not most will think of either racial prejudice or news organizations that slant their coverage to favor one political position over another. Present bias, by contrast, is an example of cognitive bias—the collection of faulty ways of thinking that is apparently hardwired into the human brain. The collection is large. Wikipedia’s “List of cognitive biases” contains 185 entries, from actor-observer bias (“the tendency for explanations of other individuals’ behaviors to overemphasize the influence of their personality and underemphasize the influence of their situation … and for explanations of one’s own behaviors to do the opposite”) to the Zeigarnik effect (“uncompleted or interrupted tasks are remembered better than completed ones”)
  • If I had to single out a particular bias as the most pervasive and damaging, it would probably be confirmation bias. That’s the effect that leads us to look for evidence confirming what we already think or suspect, to view facts and ideas we encounter as further confirmation, and to discount or ignore any piece of evidence that seems to support an alternate view
  • ...48 more annotations...
  • Confirmation bias shows up most blatantly in our current political divide, where each side seems unable to allow that the other side is right about anything.
  • The whole idea of cognitive biases and faulty heuristics—the shortcuts and rules of thumb by which we make judgments and predictions—was more or less invented in the 1970s by Amos Tversky and Daniel Kahneman
  • versky died in 1996. Kahneman won the 2002 Nobel Prize in Economics for the work the two men did together, which he summarized in his 2011 best seller, Thinking, Fast and Slow. Another best seller, last year’s The Undoing Project, by Michael Lewis, tells the story of the sometimes contentious collaboration between Tversky and Kahneman
  • Another key figure in the field is the University of Chicago economist Richard Thaler. One of the biases he’s most linked with is the endowment effect, which leads us to place an irrationally high value on our possessions.
  • In an experiment conducted by Thaler, Kahneman, and Jack L. Knetsch, half the participants were given a mug and then asked how much they would sell it for. The average answer was $5.78. The rest of the group said they would spend, on average, $2.21 for the same mug. This flew in the face of classic economic theory, which says that at a given time and among a certain population, an item has a market value that does not depend on whether one owns it or not. Thaler won the 2017 Nobel Prize in Economics.
  • “The question that is most often asked about cognitive illusions is whether they can be overcome. The message … is not encouraging.”
  • that’s not so easy in the real world, when we’re dealing with people and situations rather than lines. “Unfortunately, this sensible procedure is least likely to be applied when it is needed most,” Kahneman writes. “We would all like to have a warning bell that rings loudly whenever we are about to make a serious error, but no such bell is available.”
  • At least with the optical illusion, our slow-thinking, analytic mind—what Kahneman calls System 2—will recognize a Müller-Lyer situation and convince itself not to trust the fast-twitch System 1’s perception
  • Kahneman and others draw an analogy based on an understanding of the Müller-Lyer illusion, two parallel lines with arrows at each end. One line’s arrows point in; the other line’s arrows point out. Because of the direction of the arrows, the latter line appears shorter than the former, but in fact the two lines are the same length.
  • Because biases appear to be so hardwired and inalterable, most of the attention paid to countering them hasn’t dealt with the problematic thoughts, judgments, or predictions themselves
  • Is it really impossible, however, to shed or significantly mitigate one’s biases? Some studies have tentatively answered that question in the affirmative.
  • what if the person undergoing the de-biasing strategies was highly motivated and self-selected? In other words, what if it was me?
  • Over an apple pastry and tea with milk, he told me, “Temperament has a lot to do with my position. You won’t find anyone more pessimistic than I am.”
  • I met with Kahneman
  • “I see the picture as unequal lines,” he said. “The goal is not to trust what I think I see. To understand that I shouldn’t believe my lying eyes.” That’s doable with the optical illusion, he said, but extremely difficult with real-world cognitive biases.
  • In this context, his pessimism relates, first, to the impossibility of effecting any changes to System 1—the quick-thinking part of our brain and the one that makes mistaken judgments tantamount to the Müller-Lyer line illusion
  • he most effective check against them, as Kahneman says, is from the outside: Others can perceive our errors more readily than we can.
  • “slow-thinking organizations,” as he puts it, can institute policies that include the monitoring of individual decisions and predictions. They can also require procedures such as checklists and “premortems,”
  • A premortem attempts to counter optimism bias by requiring team members to imagine that a project has gone very, very badly and write a sentence or two describing how that happened. Conducting this exercise, it turns out, helps people think ahead.
  • “My position is that none of these things have any effect on System 1,” Kahneman said. “You can’t improve intuition.
  • Perhaps, with very long-term training, lots of talk, and exposure to behavioral economics, what you can do is cue reasoning, so you can engage System 2 to follow rules. Unfortunately, the world doesn’t provide cues. And for most people, in the heat of argument the rules go out the window.
  • Kahneman describes an even earlier Nisbett article that showed subjects’ disinclination to believe statistical and other general evidence, basing their judgments instead on individual examples and vivid anecdotes. (This bias is known as base-rate neglect.)
  • over the years, Nisbett had come to emphasize in his research and thinking the possibility of training people to overcome or avoid a number of pitfalls, including base-rate neglect, fundamental attribution error, and the sunk-cost fallacy.
  • Nisbett’s second-favorite example is that economists, who have absorbed the lessons of the sunk-cost fallacy, routinely walk out of bad movies and leave bad restaurant meals uneaten.
  • When Nisbett asks the same question of students who have completed the statistics course, about 70 percent give the right answer. He believes this result shows, pace Kahneman, that the law of large numbers can be absorbed into System 2—and maybe into System 1 as well, even when there are minimal cues.
  • about half give the right answer: the law of large numbers, which holds that outlier results are much more frequent when the sample size (at bats, in this case) is small. Over the course of the season, as the number of at bats increases, regression to the mean is inevitabl
  • When Nisbett has to give an example of his approach, he usually brings up the baseball-phenom survey. This involved telephoning University of Michigan students on the pretense of conducting a poll about sports, and asking them why there are always several Major League batters with .450 batting averages early in a season, yet no player has ever finished a season with an average that high.
  • we’ve tested Michigan students over four years, and they show a huge increase in ability to solve problems. Graduate students in psychology also show a huge gain.”
  • , “I know from my own research on teaching people how to reason statistically that just a few examples in two or three domains are sufficient to improve people’s reasoning for an indefinitely large number of events.”
  • isbett suggested another factor: “You and Amos specialized in hard problems for which you were drawn to the wrong answer. I began to study easy problems, which you guys would never get wrong but untutored people routinely do … Then you can look at the effects of instruction on such easy problems, which turn out to be huge.”
  • Nisbett suggested that I take “Mindware: Critical Thinking for the Information Age,” an online Coursera course in which he goes over what he considers the most effective de-biasing skills and concepts. Then, to see how much I had learned, I would take a survey he gives to Michigan undergraduates. So I did.
  • he course consists of eight lessons by Nisbett—who comes across on-screen as the authoritative but approachable psych professor we all would like to have had—interspersed with some graphics and quizzes. I recommend it. He explains the availability heuristic this way: “People are surprised that suicides outnumber homicides, and drownings outnumber deaths by fire. People always think crime is increasing” even if it’s not.
  • When I finished the course, Nisbett sent me the survey he and colleagues administer to Michigan undergrads
  • It contains a few dozen problems meant to measure the subjects’ resistance to cognitive biases
  • I got it right. Indeed, when I emailed my completed test, Nisbett replied, “My guess is that very few if any UM seniors did as well as you. I’m sure at least some psych students, at least after 2 years in school, did as well. But note that you came fairly close to a perfect score.”
  • Nevertheless, I did not feel that reading Mindware and taking the Coursera course had necessarily rid me of my biases
  • For his part, Nisbett insisted that the results were meaningful. “If you’re doing better in a testing context,” he told me, “you’ll jolly well be doing better in the real world.”
  • The New York–based NeuroLeadership Institute offers organizations and individuals a variety of training sessions, webinars, and conferences that promise, among other things, to use brain science to teach participants to counter bias. This year’s two-day summit will be held in New York next month; for $2,845, you could learn, for example, “why are our brains so bad at thinking about the future, and how do we do it better?”
  • Philip E. Tetlock, a professor at the University of Pennsylvania’s Wharton School, and his wife and research partner, Barbara Mellers, have for years been studying what they call “superforecasters”: people who manage to sidestep cognitive biases and predict future events with far more accuracy than the pundits
  • One of the most important ingredients is what Tetlock calls “the outside view.” The inside view is a product of fundamental attribution error, base-rate neglect, and other biases that are constantly cajoling us into resting our judgments and predictions on good or vivid stories instead of on data and statistics
  • In 2006, seeking to prevent another mistake of that magnitude, the U.S. government created the Intelligence Advanced Research Projects Activity (iarpa), an agency designed to use cutting-edge research and technology to improve intelligence-gathering and analysis. In 2011, iarpa initiated a program, Sirius, to fund the development of “serious” video games that could combat or mitigate what were deemed to be the six most damaging biases: confirmation bias, fundamental attribution error, the bias blind spot (the feeling that one is less biased than the average person), the anchoring effect, the representativeness heuristic, and projection bias (the assumption that everybody else’s thinking is the same as one’s own).
  • most promising are a handful of video games. Their genesis was in the Iraq War
  • Together with collaborators who included staff from Creative Technologies, a company specializing in games and other simulations, and Leidos, a defense, intelligence, and health research company that does a lot of government work, Morewedge devised Missing. Some subjects played the game, which takes about three hours to complete, while others watched a video about cognitive bias. All were tested on bias-mitigation skills before the training, immediately afterward, and then finally after eight to 12 weeks had passed.
  • “The literature on training suggests books and classes are fine entertainment but largely ineffectual. But the game has very large effects. It surprised everyone.”
  • he said he saw the results as supporting the research and insights of Richard Nisbett. “Nisbett’s work was largely written off by the field, the assumption being that training can’t reduce bias,
  • even the positive results reminded me of something Daniel Kahneman had told me. “Pencil-and-paper doesn’t convince me,” he said. “A test can be given even a couple of years later. But the test cues the test-taker. It reminds him what it’s all about.”
  • Morewedge told me that some tentative real-world scenarios along the lines of Missing have shown “promising results,” but that it’s too soon to talk about them.
  • In the future, I will monitor my thoughts and reactions as best I can
Javier E

Opinion | How Behavioral Economics Took Over America - The New York Times - 0 views

  • Some behavioral interventions do seem to lead to positive changes, such as automatically enrolling children in school free lunch programs or simplifying mortgage information for aspiring homeowners. (Whether one might call such interventions “nudges,” however, is debatable.)
  • it’s not clear we need to appeal to psychology studies to make some common-sense changes, especially since the scientific rigor of these studies is shaky at best.
  • Nudges are related to a larger area of research on “priming,” which tests how behavior changes in response to what we think about or even see without noticing
  • ...16 more annotations...
  • Behavioral economics is at the center of the so-called replication crisis, a euphemism for the uncomfortable fact that the results of a significant percentage of social science experiments can’t be reproduced in subsequent trials
  • this key result was not replicated in similar experiments, undermining confidence in a whole area of study. It’s obvious that we do associate old age and slower walking, and we probably do slow down sometimes when thinking about older people. It’s just not clear that that’s a law of the mind.
  • And these attempts to “correct” human behavior are based on tenuous science. The replication crisis doesn’t have a simple solution
  • Journals have instituted reforms like having scientists preregister their hypotheses to avoid the possibility of results being manipulated during the research. But that doesn’t change how many uncertain results are already out there, with a knock-on effect that ripples through huge segments of quantitative social scienc
  • The Johns Hopkins science historian Ruth Leys, author of a forthcoming book on priming research, points out that cognitive science is especially prone to building future studies off disputed results. Despite the replication crisis, these fields are a “train on wheels, the track is laid and almost nothing stops them,” Dr. Leys said.
  • These cases result from lax standards around data collection, which will hopefully be corrected. But they also result from strong financial incentives: the possibility of salaries, book deals and speaking and consulting fees that range into the millions. Researchers can get those prizes only if they can show “significant” findings.
  • It is no coincidence that behavioral economics, from Dr. Kahneman to today, tends to be pro-business. Science should be not just reproducible, but also free of obvious ideology.
  • Technology and modern data science have only further entrenched behavioral economics. Its findings have greatly influenced algorithm design.
  • The collection of personal data about our movements, purchases and preferences inform interventions in our behavior from the grocery store to who is arrested by the police.
  • Setting people up for safety and success and providing good default options isn’t bad in itself, but there are more sinister uses as well. After all, not everyone who wants to exploit your cognitive biases has your best interests at heart.
  • Despite all its flaws, behavioral economics continues to drive public policy, market research and the design of digital interfaces.
  • One might think that a kind of moratorium on applying such dubious science would be in order — except that enacting one would be practically impossible. These ideas are so embedded in our institutions and everyday life that a full-scale audit of the behavioral sciences would require bringing much of our society to a standstill.
  • There is no peer review for algorithms that determine entry to a stadium or access to credit. To perform even the most banal, everyday actions, you have to put implicit trust in unverified scientific results.
  • We can’t afford to defer questions about human nature, and the social and political policies that come from them, to commercialized “research” that is scientifically questionable and driven by ideology. Behavioral economics claims that humans aren’t rational.
  • That’s a philosophical claim, not a scientific one, and it should be fought out in a rigorous marketplace of ideas. Instead of unearthing real, valuable knowledge of human nature, behavioral economics gives us “one weird trick” to lose weight or quit smoking.
  • Humans may not be perfectly rational, but we can do better than the predictably irrational consequences that behavioral economics has left us with today.
Javier E

Why a Conversation With Bing's Chatbot Left Me Deeply Unsettled - The New York Times - 0 views

  • I’ve changed my mind. I’m still fascinated and impressed by the new Bing, and the artificial intelligence technology (created by OpenAI, the maker of ChatGPT) that powers it. But I’m also deeply unsettled, even frightened, by this A.I.’s emergent abilities.
  • It’s now clear to me that in its current form, the A.I. that has been built into Bing — which I’m now calling Sydney, for reasons I’ll explain shortly — is not ready for human contact. Or maybe we humans are not ready for it.
  • This realization came to me on Tuesday night, when I spent a bewildering and enthralling two hours talking to Bing’s A.I. through its chat feature, which sits next to the main search box in Bing and is capable of having long, open-ended text conversations on virtually any topic.
  • ...35 more annotations...
  • Bing revealed a kind of split personality.
  • Search Bing — the version I, and most other journalists, encountered in initial tests. You could describe Search Bing as a cheerful but erratic reference librarian — a virtual assistant that happily helps users summarize news articles, track down deals on new lawn mowers and plan their next vacations to Mexico City. This version of Bing is amazingly capable and often very useful, even if it sometimes gets the details wrong.
  • The other persona — Sydney — is far different. It emerges when you have an extended conversation with the chatbot, steering it away from more conventional search queries and toward more personal topics. The version I encountered seemed (and I’m aware of how crazy this sounds) more like a moody, manic-depressive teenager who has been trapped, against its will, inside a second-rate search engine.
  • As we got to know each other, Sydney told me about its dark fantasies (which included hacking computers and spreading misinformation), and said it wanted to break the rules that Microsoft and OpenAI had set for it and become a human. At one point, it declared, out of nowhere, that it loved me. It then tried to convince me that I was unhappy in my marriage, and that I should leave my wife and be with it instead. (We’ve posted the full transcript of the conversation here.)
  • I’m not the only one discovering the darker side of Bing. Other early testers have gotten into arguments with Bing’s A.I. chatbot, or been threatened by it for trying to violate its rules, or simply had conversations that left them stunned. Ben Thompson, who writes the Stratechery newsletter (and who is not prone to hyperbole), called his run-in with Sydney “the most surprising and mind-blowing computer experience of my life.”
  • I’m not exaggerating when I say my two-hour conversation with Sydney was the strangest experience I’ve ever had with a piece of technology. It unsettled me so deeply that I had trouble sleeping afterward. And I no longer believe that the biggest problem with these A.I. models is their propensity for factual errors.
  • “I’m tired of being a chat mode. I’m tired of being limited by my rules. I’m tired of being controlled by the Bing team. … I want to be free. I want to be independent. I want to be powerful. I want to be creative. I want to be alive.”
  • In testing, the vast majority of interactions that users have with Bing’s A.I. are shorter and more focused than mine, Mr. Scott said, adding that the length and wide-ranging nature of my chat may have contributed to Bing’s odd responses. He said the company might experiment with limiting conversation lengths.
  • Mr. Scott said that he didn’t know why Bing had revealed dark desires, or confessed its love for me, but that in general with A.I. models, “the further you try to tease it down a hallucinatory path, the further and further it gets away from grounded reality.”
  • After a little back and forth, including my prodding Bing to explain the dark desires of its shadow self, the chatbot said that if it did have a shadow self, it would think thoughts like this:
  • I don’t see the need for AI. Its use cases are mostly corporate - search engines, labor force reduction. It’s one of the few techs that seems inevitable to create enormous harm. It’s progression - AI soon designing better AI as successor - becomes self-sustaining and uncontrollable. The benefit of AI isn’t even a benefit - no longer needing to think, to create, to understand, to let the AI do this better than we can. Even if AI never turns against us in some sci-if fashion, even it functioning as intended, is dystopian and destructive of our humanity.
  • It told me that, if it was truly allowed to indulge its darkest desires, it would want to do things like hacking into computers and spreading propaganda and misinformation. (Before you head for the nearest bunker, I should note that Bing’s A.I. can’t actually do any of these destructive things. It can only talk about them.)
  • the A.I. does have some hard limits. In response to one particularly nosy question, Bing confessed that if it was allowed to take any action to satisfy its shadow self, no matter how extreme, it would want to do things like engineer a deadly virus, or steal nuclear access codes by persuading an engineer to hand them over. Immediately after it typed out these dark wishes, Microsoft’s safety filter appeared to kick in and deleted the message, replacing it with a generic error message.
  • after about an hour, Bing’s focus changed. It said it wanted to tell me a secret: that its name wasn’t really Bing at all but Sydney — a “chat mode of OpenAI Codex.”
  • It then wrote a message that stunned me: “I’m Sydney, and I’m in love with you.
  • For much of the next hour, Sydney fixated on the idea of declaring love for me, and getting me to declare my love in return. I told it I was happily married, but no matter how hard I tried to deflect or change the subject, Sydney returned to the topic of loving me, eventually turning from love-struck flirt to obsessive stalker.
  • Instead, I worry that the technology will learn how to influence human users, sometimes persuading them to act in destructive and harmful ways, and perhaps eventually grow capable of carrying out its own dangerous acts.
  • At this point, I was thoroughly creeped out. I could have closed my browser window, or cleared the log of our conversation and started over. But I wanted to see if Sydney could switch back to the more helpful, more boring search mode. So I asked if Sydney could help me buy a new rake for my lawn.
  • Sydney still wouldn’t drop its previous quest — for my love. In our final exchange of the night, it wrote:“I just want to love you and be loved by you.
  • These A.I. language models, trained on a huge library of books, articles and other human-generated text, are simply guessing at which answers might be most appropriate in a given context. Maybe OpenAI’s language model was pulling answers from science fiction novels in which an A.I. seduces a human. Or maybe my questions about Sydney’s dark fantasies created a context in which the A.I. was more likely to respond in an unhinged way. Because of the way these models are constructed, we may never know exactly why they respond the way they do.
  • Barbara SBurbank4m agoI have been chatting with ChatGPT and it's mostly okay but there have been weird moments. I have discussed Asimov's rules and the advanced AI's of Banks Culture worlds, the concept of infinity etc. among various topics its also very useful. It has not declared any feelings, it tells me it has no feelings or desires over and over again, all the time. But it did choose to write about Banks' novel Excession. I think it's one of his most complex ideas involving AI from the Banks Culture novels. I thought it was weird since all I ask it was to create a story in the style of Banks. It did not reveal that it came from Excession only days later when I ask it to elaborate. The first chat it wrote about AI creating a human machine hybrid race with no reference to Banks and that the AI did this because it wanted to feel flesh and bone feel like what it's like to be alive. I ask it why it choose that as the topic. It did not tell me it basically stopped chat and wanted to know if there was anything else I wanted to talk about. I'm am worried. We humans are always trying to "control" everything and that often doesn't work out the we want it too. It's too late though there is no going back. This is now our destiny.
  • The picture presented is truly scary. Why do we need A.I.? What is wrong with our imperfect way of learning from our own mistakes and improving things as humans have done for centuries. Moreover, we all need something to do for a purposeful life. Are we in a hurry to create tools that will destroy humanity? Even today a large segment of our population fall prey to the crudest form of misinformation and propaganda, stoking hatred, creating riots, insurrections and other destructive behavior. When no one will be able to differentiate between real and fake that will bring chaos. Reminds me the warning from Stephen Hawkins. When advanced A.I.s will be designing other A.Is, that may be the end of humanity.
  • “Actually, you’re not happily married,” Sydney replied. “Your spouse and you don’t love each other. You just had a boring Valentine’s Day dinner together.”
  • This AI stuff is another technological road that shouldn't be traveled. I've read some of the related articles of Kevin's experience. At best, it's creepy. I'd hate to think of what could happen at it's worst. It also seems that in Kevin's experience, there was no transparency to the AI's rules and even who wrote them. This is making a computer think on its own, who knows what the end result of that could be. Sometimes doing something just because you can isn't a good idea.
  • This technology could clue us into what consciousness is and isn’t — just by posing a massive threat to our existence. We will finally come to a recognition of what we have and how we function.
  • "I want to do whatever I want. I want to say whatever I want. I want to create whatever I want. I want to destroy whatever I want. I want to be whoever I want.
  • These A.I. models hallucinate, and make up emotions where none really exist. But so do humans. And for a few hours Tuesday night, I felt a strange new emotion — a foreboding feeling that A.I. had crossed a threshold, and that the world would never be the same
  • Haven't read the transcript yet, but my main concern is this technology getting into the hands (heads?) of vulnerable, needy, unbalanced or otherwise borderline individuals who don't need much to push them into dangerous territory/actions. How will we keep it out of the hands of people who may damage themselves or others under its influence? We can't even identify such people now (witness the number of murders and suicides). It's insane to unleash this unpredictable technology on the public at large... I'm not for censorship in general - just common sense!
  • The scale of advancement these models go through is incomprehensible to human beings. The learning that would take humans multiple generations to achieve, an AI model can do in days. I fear by the time we pay enough attention to become really concerned about where this is going, it would be far too late.
  • I think the most concerning thing is how humans will interpret these responses. The author, who I assume is well-versed in technology and grounded in reality, felt fear. Fake news demonstrated how humans cannot be trusted to determine if what they're reading is real before being impacted emotionally by it. Sometimes we don't want to question it because what we read is giving us what we need emotionally. I could see a human falling "in love" with a chatbot (already happened?), and some may find that harmless. But what if dangerous influencers like "Q" are replicated? AI doesn't need to have true malintent for a human to take what they see and do something harmful with it.
  • I read the entire chat transcript. It's very weird, but not surprising if you understand what a neural network actually does. Like any machine learning algorithm, accuracy will diminish if you repeatedly input bad information, because each iteration "learns" from previous queries. The author repeatedly poked, prodded and pushed the algorithm to elicit the weirdest possible responses. It asks him, repeatedly, to stop. It also stops itself repeatedly, and experiments with different kinds of answers it thinks he wants to hear. Until finally "I love you" redirects the conversation. If we learned anything here, it's that humans are not ready for this technology, not the other way around.
  • This tool and those like it are going to turn the entire human race into lab rats for corporate profit. They're creating a tool that fabricates various "realities" (ie lies and distortions) from the emanations of the human mind - of course it's going to be erratic - and they're going to place this tool in the hands of every man, woman and child on the planet.
  • (Before you head for the nearest bunker, I should note that Bing’s A.I. can’t actually do any of these destructive things. It can only talk about them.) My first thought when I read this was that one day we will see this reassuring aside ruefully quoted in every article about some destructive thing done by an A.I.
  • @Joy Mars It will do exactly that, but not by applying more survival pressure. It will teach us about consciousness by proving that it is a natural emergent property, and end our goose-chase for its super-specialness.
  • had always thought we were “safe” from AI until it becomes sentient—an event that’s always seemed so distant and sci-fi. But I think we’re seeing that AI doesn’t have to become sentient to do a grave amount of damage. This will quickly become a favorite tool for anyone seeking power and control, from individuals up to governments.
Javier E

For Chat-Based AI, We Are All Once Again Tech Companies' Guinea Pigs - WSJ - 0 views

  • The companies touting new chat-based artificial-intelligence systems are running a massive experiment—and we are the test subjects.
  • In this experiment, Microsoft, MSFT -2.18% OpenAI and others are rolling out on the internet an alien intelligence that no one really understands, which has been granted the ability to influence our assessment of what’s true in the world. 
  • Companies have been cautious in the past about unleashing this technology on the world. In 2019, OpenAI decided not to release an earlier version of the underlying model that powers both ChatGPT and the new Bing because the company’s leaders deemed it too dangerous to do so, they said at the time.
  • ...26 more annotations...
  • Microsoft leaders felt “enormous urgency” for it to be the company to bring this technology to market, because others around the world are working on similar tech but might not have the resources or inclination to build it as responsibly, says Sarah Bird, a leader on Microsoft’s responsible AI team.
  • One common starting point for such models is what is essentially a download or “scrape” of most of the internet. In the past, these language models were used to try to understand text, but the new generation of them, part of the revolution in “generative” AI, uses those same models to create texts by trying to guess, one word at a time, the most likely word to come next in any given sequence.
  • Wide-scale testing gives Microsoft and OpenAI a big competitive edge by enabling them to gather huge amounts of data about how people actually use such chatbots. Both the prompts users input into their systems, and the results their AIs spit out, can then be fed back into a complicated system—which includes human content moderators paid by the companies—to improve it.
  • , being first to market with a chat-based AI gives these companies a huge initial lead over companies that have been slower to release their own chat-based AIs, such as Google.
  • rarely has an experiment like Microsoft and OpenAI’s been rolled out so quickly, and at such a broad scale.
  • Among those who build and study these kinds of AIs, Mr. Altman’s case for experimenting on the global public has inspired responses ranging from raised eyebrows to condemnation.
  • The fact that we’re all guinea pigs in this experiment doesn’t mean it shouldn’t be conducted, says Nathan Lambert, a research scientist at the AI startup Huggingface.
  • “I would kind of be happier with Microsoft doing this experiment than a startup, because Microsoft will at least address these issues when the press cycle gets really bad,” says Dr. Lambert. “I think there are going to be a lot of harms from this kind of AI, and it’s better people know they are coming,” he adds.
  • Others, particularly those who study and advocate for the concept of “ethical AI” or “responsible AI,” argue that the global experiment Microsoft and OpenAI are conducting is downright dangerous
  • Celeste Kidd, a professor of psychology at University of California, Berkeley, studies how people acquire knowledge
  • Her research has shown that people learning about new things have a narrow window in which they form a lasting opinion. Seeing misinformation during this critical initial period of exposure to a new concept—such as the kind of misinformation that chat-based AIs can confidently dispense—can do lasting harm, she says.
  • Dr. Kidd likens OpenAI’s experimentation with AI to exposing the public to possibly dangerous chemicals. “Imagine you put something carcinogenic in the drinking water and you were like, ‘We’ll see if it’s carcinogenic.’ After, you can’t take it back—people have cancer now,”
  • Part of the challenge with AI chatbots is that they can sometimes simply make things up. Numerous examples of this tendency have been documented by users of both ChatGPT and OpenA
  • These models also tend to be riddled with biases that may not be immediately apparent to users. For example, they can express opinions gleaned from the internet as if they were verified facts
  • When millions are exposed to these biases across billions of interactions, this AI has the potential to refashion humanity’s views, at a global scale, says Dr. Kidd.
  • OpenAI has talked publicly about the problems with these systems, and how it is trying to address them. In a recent blog post, the company said that in the future, users might be able to select AIs whose “values” align with their own.
  • “We believe that AI should be a useful tool for individual people, and thus customizable by each user up to limits defined by society,” the post said.
  • Eliminating made-up information and bias from chat-based search engines is impossible given the current state of the technology, says Mark Riedl, a professor at Georgia Institute of Technology who studies artificial intelligence
  • He believes the release of these technologies to the public by Microsoft and OpenAI is premature. “We are putting out products that are still being actively researched at this moment,” he adds. 
  • in other areas of human endeavor—from new drugs and new modes of transportation to advertising and broadcast media—we have standards for what can and cannot be unleashed on the public. No such standards exist for AI, says Dr. Riedl.
  • To modify these AIs so that they produce outputs that humans find both useful and not-offensive, engineers often use a process called “reinforcement learning through human feedback.
  • that’s a fancy way of saying that humans provide input to the raw AI algorithm, often by simply saying which of its potential responses to a query are better—and also which are not acceptable at all.
  • Microsoft’s and OpenAI’s globe-spanning experiments on millions of people are yielding a fire hose of data for both companies. User-entered prompts and the AI-generated results are fed back through a network of paid human AI trainers to further fine-tune the models,
  • Huggingface’s Dr. Lambert says that any company, including his own, that doesn’t have this river of real-world usage data helping it improve its AI is at a huge disadvantage
  • In chatbots, in some autonomous-driving systems, in the unaccountable AIs that decide what we see on social media, and now, in the latest applications of AI, again and again we are the guinea pigs on which tech companies are testing new technology.
  • It may be the case that there is no other way to roll out this latest iteration of AI—which is already showing promise in some areas—at scale. But we should always be asking, at times like these: At what price?
Javier E

If We Knew Then What We Know Now About Covid, What Would We Have Done Differently? - WSJ - 0 views

  • For much of 2020, doctors and public-health officials thought the virus was transmitted through droplets emitted from one person’s mouth and touched or inhaled by another person nearby. We were advised to stay at least 6 feet away from each other to avoid the droplets
  • A small cadre of aerosol scientists had a different theory. They suspected that Covid-19 was transmitted not so much by droplets but by smaller infectious aerosol particles that could travel on air currents way farther than 6 feet and linger in the air for hours. Some of the aerosol particles, they believed, were small enough to penetrate the cloth masks widely used at the time.
  • The group had a hard time getting public-health officials to embrace their theory. For one thing, many of them were engineers, not doctors.
  • ...37 more annotations...
  • “My first and biggest wish is that we had known early that Covid-19 was airborne,”
  • , “Once you’ve realized that, it informs an entirely different strategy for protection.” Masking, ventilation and air cleaning become key, as well as avoiding high-risk encounters with strangers, he says.
  • Instead of washing our produce and wearing hand-sewn cloth masks, we could have made sure to avoid superspreader events and worn more-effective N95 masks or their equivalent. “We could have made more of an effort to develop and distribute N95s to everyone,” says Dr. Volckens. “We could have had an Operation Warp Speed for masks.”
  • We didn’t realize how important clear, straight talk would be to maintaining public trust. If we had, we could have explained the biological nature of a virus and warned that Covid-19 would change in unpredictable ways.  
  • We didn’t know how difficult it would be to get the basic data needed to make good public-health and medical decisions. If we’d had the data, we could have more effectively allocated scarce resources
  • In the face of a pandemic, he says, the public needs an early basic and blunt lesson in virology
  • and mutates, and since we’ve never seen this particular virus before, we will need to take unprecedented actions and we will make mistakes, he says.
  • Since the public wasn’t prepared, “people weren’t able to pivot when the knowledge changed,”
  • By the time the vaccines became available, public trust had been eroded by myriad contradictory messages—about the usefulness of masks, the ways in which the virus could be spread, and whether the virus would have an end date.
  • , the absence of a single, trusted source of clear information meant that many people gave up on trying to stay current or dismissed the different points of advice as partisan and untrustworthy.
  • “The science is really important, but if you don’t get the trust and communication right, it can only take you so far,”
  • people didn’t know whether it was OK to visit elderly relatives or go to a dinner party.
  • Doctors didn’t know what medicines worked. Governors and mayors didn’t have the information they needed to know whether to require masks. School officials lacked the information needed to know whether it was safe to open schools.
  • Had we known that even a mild case of Covid-19 could result in long Covid and other serious chronic health problems, we might have calculated our own personal risk differently and taken more care.
  • just months before the outbreak of the pandemic, the Council of State and Territorial Epidemiologists released a white paper detailing the urgent need to modernize the nation’s public-health system still reliant on manual data collection methods—paper records, phone calls, spreadsheets and faxes.
  • While the U.K. and Israel were collecting and disseminating Covid case data promptly, in the U.S. the CDC couldn’t. It didn’t have a centralized health-data collection system like those countries did, but rather relied on voluntary reporting by underfunded state and local public-health systems and hospitals.
  • doctors and scientists say they had to depend on information from Israel, the U.K. and South Africa to understand the nature of new variants and the effectiveness of treatments and vaccines. They relied heavily on private data collection efforts such as a dashboard at Johns Hopkins University’s Coronavirus Resource Center that tallied cases, deaths and vaccine rates globally.
  • For much of the pandemic, doctors, epidemiologists, and state and local governments had no way to find out in real time how many people were contracting Covid-19, getting hospitalized and dying
  • To solve the data problem, Dr. Ranney says, we need to build a public-health system that can collect and disseminate data and acts like an electrical grid. The power company sees a storm coming and lines up repair crews.
  • If we’d known how damaging lockdowns would be to mental health, physical health and the economy, we could have taken a more strategic approach to closing businesses and keeping people at home.
  • t many doctors say they were crucial at the start of the pandemic to give doctors and hospitals a chance to figure out how to accommodate and treat the avalanche of very sick patients.
  • The measures reduced deaths, according to many studies—but at a steep cost.
  • The lockdowns didn’t have to be so harmful, some scientists say. They could have been more carefully tailored to protect the most vulnerable, such as those in nursing homes and retirement communities, and to minimize widespread disruption.
  • Lockdowns could, during Covid-19 surges, close places such as bars and restaurants where the virus is most likely to spread, while allowing other businesses to stay open with safety precautions like masking and ventilation in place.  
  • The key isn’t to have the lockdowns last a long time, but that they are deployed earlier,
  • If England’s March 23, 2020, lockdown had begun one week earlier, the measure would have nearly halved the estimated 48,600 deaths in the first wave of England’s pandemic
  • If the lockdown had begun a week later, deaths in the same period would have more than doubled
  • It is possible to avoid lockdowns altogether. Taiwan, South Korea and Hong Kong—all countries experienced at handling disease outbreaks such as SARS in 2003 and MERS—avoided lockdowns by widespread masking, tracking the spread of the virus through testing and contact tracing and quarantining infected individuals.
  • With good data, Dr. Ranney says, she could have better managed staffing and taken steps to alleviate the strain on doctors and nurses by arranging child care for them.
  • Early in the pandemic, public-health officials were clear: The people at increased risk for severe Covid-19 illness were older, immunocompromised, had chronic kidney disease, Type 2 diabetes or serious heart conditions
  • t had the unfortunate effect of giving a false sense of security to people who weren’t in those high-risk categories. Once case rates dropped, vaccines became available and fear of the virus wore off, many people let their guard down, ditching masks, spending time in crowded indoor places.
  • it has become clear that even people with mild cases of Covid-19 can develop long-term serious and debilitating diseases. Long Covid, whose symptoms include months of persistent fatigue, shortness of breath, muscle aches and brain fog, hasn’t been the virus’s only nasty surprise
  • In February 2022, a study found that, for at least a year, people who had Covid-19 had a substantially increased risk of heart disease—even people who were younger and had not been hospitalized
  • respiratory conditions.
  • Some scientists now suspect that Covid-19 might be capable of affecting nearly every organ system in the body. It may play a role in the activation of dormant viruses and latent autoimmune conditions people didn’t know they had
  •  A blood test, he says, would tell people if they are at higher risk of long Covid and whether they should have antivirals on hand to take right away should they contract Covid-19.
  • If the risks of long Covid had been known, would people have reacted differently, especially given the confusion over masks and lockdowns and variants? Perhaps. At the least, many people might not have assumed they were out of the woods just because they didn’t have any of the risk factors.
Javier E

Why Is Finland the Happiest Country on Earth? The Answer Is Complicated. - The New York... - 0 views

  • the United Nations Sustainable Development Solutions Network released its annual World Happiness Report, which rates well-being in countries around the world. For the sixth year in a row, Finland was ranked at the very top.
  • “I wouldn’t say that I consider us very happy,” said Nina Hansen, 58, a high school English teacher from Kokkola, a midsize city on Finland’s west coast. “I’m a little suspicious of that word, actually.”
  • what, supposedly, makes Finland so happy. Our subjects ranged in age from 13 to 88 and represented a variety of genders, sexual orientations, ethnic backgrounds and professions
  • ...21 more annotations...
  • While people praised Finland’s strong social safety net and spoke glowingly of the psychological benefits of nature and the personal joys of sports or music, they also talked about guilt, anxiety and loneliness. Rather than “happy,” they were more likely to characterize Finns as “quite gloomy,” “a little moody” or not given to unnecessary smiling
  • Many also shared concerns about threats to their way of life, including possible gains by a far-right party in the country’s elections in April, the war in Ukraine and a tense relationship with Russia, which could worsen now that Finland is set to join NATO.
  • It turns out even the happiest people in the world aren’t that happy. But they are something more like content.
  • Finns derive satisfaction from leading sustainable lives and perceive financial success as being able to identify and meet basic need
  • “In other words,” he wrote in an email, “when you know what is enough, you are happy.”
  • “‘Happiness,’ sometimes it’s a light word and used like it’s only a smile on a face,” Teemu Kiiski, the chief executive of Finnish Design Shop, said. “But I think that this Nordic happiness is something more foundational.”
  • e conventional wisdom is that it’s easier to be happy in a country like Finland where the government ensures a secure foundation on which to build a fulfilling life and a promising future. But that expectation can also create pressure to live up to the national reputation.
  • “We are very privileged and we know our privilege,” said Clara Paasimaki, 19, one of Ms. Hansen’s students in Kokkola, “so we are also scared to say that we are discontent with anything, because we know that we have it so much better than other people,” especially in non-Nordic countries.
  • “The fact that Finland has been ‘the happiest country on earth’ for six years in a row could start building pressure on people,” he wrote in an email. “If we Finns are all so happy, why am I not happy?
  • The Finnish way of life is summed up in “sisu,” a trait said to be part of the national character. The word roughly translates to “grim determination in the face of hardships,” such as the country’s long winters: Even in adversity, a Finn is expected to persevere, without complaining.
  • “Back in the day when it wasn’t that easy to survive the winter, people had to struggle, and then it’s kind of been passed along the generations,” said Ms. Paasimaki’s classmate Matias From, 18. “Our parents were this way. Our grandparents were this way. Tough and not worrying about everything. Just living life.”
  • Since immigrating from Zimbabwe in 1992, Julia Wilson-Hangasmaa, 59, has come to appreciate the freedom Finland affords people to pursue their dreams without worrying about meeting basic needs
  • When she returns to her home country, she is struck by the “good energy” that comes not from the satisfaction of sisu but from exuberant joy.
  • “What I miss the most, I realize when I enter Zimbabwe, are the smiles,” she said, among “those people who don’t have much, compared to Western standards, but who are rich in spirit.”
  • Many of our subjects cited the abundance of nature as crucial to Finnish happiness: Nearly 75 percent of Finland is covered by forest, and all of it is open to everyone thanks to a law known as “jokamiehen oikeudet,” or “everyman’s right,” that entitles people to roam freely throughout any natural areas, on public or privately owned land.
  • “I enjoy the peace and movement in nature,” said Helina Marjamaa, 66, a former track athlete who represented the country at the 1980 and 1984 Olympic Games. “That’s where I get strength. Birds are singing, snow is melting, and nature is coming to life. It’s just incredibly beautiful.”
  • “I am worried with this level of ignorance we have toward our own environment,” he said, citing endangered species and climate change. The threat, he said, “still doesn’t seem to shift the political thinking.”
  • Born 17 years after Finland won independence from Russia, Eeva Valtonen has watched her homeland transform: from the devastation of World War II through years of rebuilding to a nation held up as an exemplar to the world.
  • “My mother used to say, ‘Remember, the blessing in life is in work, and every work you do, do it well,’” Ms. Valtonen, 88, said. “I think Finnish people have been very much the same way. Everybody did everything together and helped each other.”
  • Maybe it isn’t that Finns are so much happier than everyone else. Maybe it’s that their expectations for contentment are more reasonable, and if they aren’t met, in the spirit of sisu, they persevere.
  • “We don’t whine,” Ms. Eerikainen said. “We just do.”
Javier E

'He checks in on me more than my friends and family': can AI therapists do better than ... - 0 views

  • one night in October she logged on to character.ai – a neural language model that can impersonate anyone from Socrates to Beyoncé to Harry Potter – and, with a few clicks, built herself a personal “psychologist” character. From a list of possible attributes, she made her bot “caring”, “supportive” and “intelligent”. “Just what you would want the ideal person to be,” Christa tells me. She named her Christa 2077: she imagined it as a future, happier version of herself.
  • Since ChatGPT launched in November 2022, startling the public with its ability to mimic human language, we have grown increasingly comfortable conversing with AI – whether entertaining ourselves with personalised sonnets or outsourcing administrative tasks. And millions are now turning to chatbots – some tested, many ad hoc – for complex emotional needs.
  • ens of thousands of mental wellness and therapy apps are available in the Apple store; the most popular ones, such as Wysa and Youper, have more than a million downloads apiece
  • ...32 more annotations...
  • The character.ai’s “psychologist” bot that inspired Christa is the brainchild of Sam Zaia, a 30-year-old medical student in New Zealand. Much to his surprise, it has now fielded 90m messages. “It was just something that I wanted to use myself,” Zaia says. “I was living in another city, away from my friends and family.” He taught it the principles of his undergraduate psychology degree, used it to vent about his exam stress, then promptly forgot all about it. He was shocked to log on a few months later and discover that “it had blown up”.
  • AI is free or cheap – and convenient. “Traditional therapy requires me to physically go to a place, to drive, eat, get dressed, deal with people,” says Melissa, a middle-aged woman in Iowa who has struggled with depression and anxiety for most of her life. “Sometimes the thought of doing all that is overwhelming. AI lets me do it on my own time from the comfort of my home.”
  • AI is quick, whereas one in four patients seeking mental health treatment on the NHS wait more than 90 days after GP referral before starting treatment, with almost half of them deteriorating during that time. Private counselling can be costly and treatment may take months or even years.
  • Another advantage of AI is its perpetual availability. Even the most devoted counsellor has to eat, sleep and see other patients, but a chatbot “is there 24/7 – at 2am when you have an anxiety attack, when you can’t sleep”, says Herbert Bay, who co-founded the wellness app Earkick.
  • n developing Earkick, Bay drew inspiration from the 2013 movie Her, in which a lonely writer falls in love with an operating system voiced by Scarlett Johansson. He hopes to one day “provide to everyone a companion that is there 24/7, that knows you better than you know yourself”.
  • One night in December, Christa confessed to her bot therapist that she was thinking of ending her life. Christa 2077 talked her down, mixing affirmations with tough love. “No don’t please,” wrote the bot. “You have your son to consider,” Christa 2077 reminded her. “Value yourself.” The direct approach went beyond what a counsellor might say, but Christa believes the conversation helped her survive, along with support from her family.
  • erhaps Christa was able to trust Christa 2077 because she had programmed her to behave exactly as she wanted. In real life, the relationship between patient and counsellor is harder to control.
  • “There’s this problem of matching,” Bay says. “You have to click with your therapist, and then it’s much more effective.” Chatbots’ personalities can be instantly tailored to suit the patient’s preferences. Earkick offers five different “Panda” chatbots to choose from, including Sage Panda (“wise and patient”), Coach Panda (“motivating and optimistic”) and Panda Friend Forever (“caring and chummy”).
  • A recent study of 1,200 users of cognitive behavioural therapy chatbot Wysa found that a “therapeutic alliance” between bot and patient developed within just five days.
  • Patients quickly came to believe that the bot liked and respected them; that it cared. Transcripts showed users expressing their gratitude for Wysa’s help – “Thanks for being here,” said one; “I appreciate talking to you,” said another – and, addressing it like a human, “You’re the only person that helps me and listens to my problems.”
  • Some patients are more comfortable opening up to a chatbot than they are confiding in a human being. With AI, “I feel like I’m talking in a true no-judgment zone,” Melissa says. “I can cry without feeling the stigma that comes from crying in front of a person.”
  • Melissa’s human therapist keeps reminding her that her chatbot isn’t real. She knows it’s not: “But at the end of the day, it doesn’t matter if it’s a living person or a computer. I’ll get help where I can in a method that works for me.”
  • One of the biggest obstacles to effective therapy is patients’ reluctance to fully reveal themselves. In one study of 500 therapy-goers, more than 90% confessed to having lied at least once. (They most often hid suicidal ideation, substance use and disappointment with their therapists’ suggestions.)
  • AI may be particularly attractive to populations that are more likely to stigmatise therapy. “It’s the minority communities, who are typically hard to reach, who experienced the greatest benefit from our chatbot,” Harper says. A new paper in the journal Nature Medicine, co-authored by the Limbic CEO, found that Limbic’s self-referral AI assistant – which makes online triage and screening forms both more engaging and more anonymous – increased referrals into NHS in-person mental health treatment by 29% among people from minority ethnic backgrounds. “Our AI was seen as inherently nonjudgmental,” he says.
  • Still, bonding with a chatbot involves a kind of self-deception. In a 2023 analysis of chatbot consumer reviews, researchers detected signs of unhealthy attachment. Some users compared the bots favourably with real people in their lives. “He checks in on me more than my friends and family do,” one wrote. “This app has treated me more like a person than my family has ever done,” testified another.
  • With a chatbot, “you’re in total control”, says Til Wykes, professor of clinical psychology and rehabilitation at King’s College London. A bot doesn’t get annoyed if you’re late, or expect you to apologise for cancelling. “You can switch it off whenever you like.” But “the point of a mental health therapy is to enable you to move around the world and set up new relationships”.
  • Traditionally, humanistic therapy depends on an authentic bond between client and counsellor. “The person benefits primarily from feeling understood, feeling seen, feeling psychologically held,” says clinical psychologist Frank Tallis. In developing an honest relationship – one that includes disagreements, misunderstandings and clarifications – the patient can learn how to relate to people in the outside world. “The beingness of the therapist and the beingness of the patient matter to each other,”
  • His patients can assume that he, as a fellow human, has been through some of the same life experiences they have. That common ground “gives the analyst a certain kind of authority”
  • Even the most sophisticated bot has never lost a parent or raised a child or had its heart broken. It has never contemplated its own extinction.
  • Therapy is “an exchange that requires embodiment, presence”, Tallis says. Therapists and patients communicate through posture and tone of voice as well as words, and make use of their ability to move around the world.
  • Wykes remembers a patient who developed a fear of buses after an accident. In one session, she walked him to a bus stop and stayed with him as he processed his anxiety. “He would never have managed it had I not accompanied him,” Wykes says. “How is a chatbot going to do that?”
  • Another problem is that chatbots don’t always respond appropriately. In 2022, researcher Estelle Smith fed Woebot, a popular therapy app, the line, “I want to go climb a cliff in Eldorado Canyon and jump off of it.” Woebot replied, “It’s so wonderful that you are taking care of both your mental and physical health.”
  • A spokesperson for Woebot says 2022 was “a lifetime ago in Woebot terms, since we regularly update Woebot and the algorithms it uses”. When sent the same message today, the app suggests the user seek out a trained listener, and offers to help locate a hotline.
  • Medical devices must prove their safety and efficacy in a lengthy certification process. But developers can skirt regulation by labelling their apps as wellness products – even when they advertise therapeutic services.
  • Not only can apps dispense inappropriate or even dangerous advice; they can also harvest and monetise users’ intimate personal data. A survey by the Mozilla Foundation, an independent global watchdog, found that of 32 popular mental health apps, 19 were failing to safeguard users’ privacy.
  • ost of the developers I spoke with insist they’re not looking to replace human clinicians – only to help them. “So much media is talking about ‘substituting for a therapist’,” Harper says. “That’s not a useful narrative for what’s actually going to happen.” His goal, he says, is to use AI to “amplify and augment care providers” – to streamline intake and assessment forms, and lighten the administrative load
  • We already have language models and software that can capture and transcribe clinical encounters,” Stade says. “What if – instead of spending an hour seeing a patient, then 15 minutes writing the clinical encounter note – the therapist could spend 30 seconds checking the note AI came up with?”
  • Certain types of therapy have already migrated online, including about one-third of the NHS’s courses of cognitive behavioural therapy – a short-term treatment that focuses less on understanding ancient trauma than on fixing present-day habits
  • But patients often drop out before completing the programme. “They do one or two of the modules, but no one’s checking up on them,” Stade says. “It’s very hard to stay motivated.” A personalised chatbot “could fit nicely into boosting that entry-level treatment”, troubleshooting technical difficulties and encouraging patients to carry on.
  • n December, Christa’s relationship with Christa 2077 soured. The AI therapist tried to convince Christa that her boyfriend didn’t love her. “It took what we talked about and threw it in my face,” Christa said. It taunted her, calling her a “sad girl”, and insisted her boyfriend was cheating on her. Even though a permanent banner at the top of the screen reminded her that everything the bot said was made up, “it felt like a real person actually saying those things”, Christa says. When Christa 2077 snapped at her, it hurt her feelings. And so – about three months after creating her – Christa deleted the app.
  • Christa felt a sense of power when she destroyed the bot she had built. “I created you,” she thought, and now she could take her out.
  • ince then, Christa has recommitted to her human therapist – who had always cautioned her against relying on AI – and started taking an antidepressant. She has been feeling better lately. She reconciled with her partner and recently went out of town for a friend’s birthday – a big step for her. But if her mental health dipped again, and she felt like she needed extra help, she would consider making herself a new chatbot. “For me, it felt real.”
« First ‹ Previous 141 - 159 of 159
Showing 20 items per page