Skip to main content

Home/ TOK Friends/ Group items tagged concepts

Rss Feed Group items tagged

Javier E

How to Remember Everything You Want From Non-Fiction Books | by Eva Keiffenheim, MSc | ... - 0 views

  • A Bachelor’s degree taught me how to learn to ace exams. But it didn’t teach me how to learn to remember.
  • 65% to 80% of students answered “no” to the question “Do you study the way you do because somebody taught you to study that way?”
  • the most-popular Coursera course of all time: Dr. Barabara Oakley’s free course on “Learning how to Learn.” So did I. And while this course taught me about chunking, recalling, and interleaving
  • ...66 more annotations...
  • I learned something more useful: the existence of non-fiction literature that can teach you anything.
  • something felt odd. Whenever a conversation revolved around a serious non-fiction book I read, such as ‘Sapiens’ or ‘Thinking Fast and Slow,’ I could never remember much. Turns out, I hadn’t absorbed as much information as I’d believed. Since I couldn’t remember much, I felt as though reading wasn’t an investment in knowledge but mere entertainment.
  • When I opened up about my struggles, many others confessed they also can’t remember most of what they read, as if forgetting is a character flaw. But it isn’t.
  • It’s the way we work with books that’s flawed.
  • there’s a better way to read. Most people rely on techniques like highlighting, rereading, or, worst of all, completely passive reading, which are highly ineffective.
  • Since I started applying evidence-based learning strategies to reading non-fiction books, many things have changed. I can explain complex ideas during dinner conversations. I can recall interesting concepts and link them in my writing or podcasts. As a result, people come to me for all kinds of advice.
  • What’s the Architecture of Human Learning and Memory?
  • Human brains don’t work like recording devices. We don’t absorb information and knowledge by reading sentences.
  • we store new information in terms of its meaning to our existing memory
  • we give new information meaning by actively participating in the learning process — we interpret, connect, interrelate, or elaborate
  • To remember new information, we not only need to know it but also to know how it relates to what we already know.
  • Learning is dependent on memory processes because previously-stored knowledge functions as a framework in which newly learned information can be linked.”
  • Human memory works in three stages: acquisition, retention, and retrieval. In the acquisition phase, we link new information to existing knowledge; in the retention phase, we store it, and in the retrieval phase, we get information out of our memory.
  • Retrieval, the third stage, is cue dependent. This means the more mental links you’re generating during stage one, the acquisition phase, the easier you can access and use your knowledge.
  • we need to understand that the three phases interrelate
  • creating durable and flexible access to to-be-learned information is partly a matter of achieving a meaningful encoding of that information and partly a matter of exercising the retrieval process.”
  • Next, we’ll look at the learning strategies that work best for our brains (elaboration, retrieval, spaced repetition, interleaving, self-testing) and see how we can apply those insights to reading non-fiction books.
  • The strategies that follow are rooted in research from professors of Psychological & Brain Science around Henry Roediger and Mark McDaniel. Both scientists spent ten years bridging the gap between cognitive psychology and education fields. Harvard University Press published their findings in the book ‘Make It Stick.
  • #1 Elaboration
  • “Elaboration is the process of giving new material meaning by expressing it in your own words and connecting it with what you already know.”
  • Why elaboration works: Elaborative rehearsal encodes information into your long-term memory more effectively. The more details and the stronger you connect new knowledge to what you already know, the better because you’ll be generating more cues. And the more cues they have, the easier you can retrieve your knowledge.
  • How I apply elaboration: Whenever I read an interesting section, I pause and ask myself about the real-life connection and potential application. The process is invisible, and my inner monologues sound like: “This idea reminds me of…, This insight conflicts with…, I don’t really understand how…, ” etc.
  • For example, when I learned about A/B testing in ‘The Lean Startup,’ I thought about applying this method to my startup. I added a note on the site stating we should try it in user testing next Wednesday. Thereby the book had an immediate application benefit to my life, and I will always remember how the methodology works.
  • How you can apply elaboration: Elaborate while you read by asking yourself meta-learning questions like “How does this relate to my life? In which situation will I make use of this knowledge? How does it relate to other insights I have on the topic?”
  • While pausing and asking yourself these questions, you’re generating important memory cues. If you take some notes, don’t transcribe the author’s words but try to summarize, synthesize, and analyze.
  • #2 Retrieval
  • With retrieval, you try to recall something you’ve learned in the past from your memory. While retrieval practice can take many forms — take a test, write an essay, do a multiple-choice test, practice with flashcards
  • the authors of ‘Make It Stick’ state: “While any kind of retrieval practice generally benefits learning, the implication seems to be that where more cognitive effort is required for retrieval, greater retention results.”
  • Whatever you settle for, be careful not to copy/paste the words from the author. If you don’t do the brain work yourself, you’ll skip the learning benefits of retrieval
  • Retrieval strengthens your memory and interrupts forgetting and, as other researchers replicate, as a learning event, the act of retrieving information is considerably more potent than is an additional study opportunity, particularly in terms of facilitating long-term recall.
  • How I apply retrieval: I retrieve a book’s content from my memory by writing a book summary for every book I want to remember. I ask myself questions like: “How would you summarize the book in three sentences? Which concepts do you want to keep in mind or apply? How does the book relate to what you already know?”
  • I then publish my summaries on Goodreads or write an article about my favorite insights
  • How you can apply retrieval: You can come up with your own questions or use mine. If you don’t want to publish your summaries in public, you can write a summary into your journal, start a book club, create a private blog, or initiate a WhatsApp group for sharing book summaries.
  • a few days after we learn something, forgetting sets in
  • #3 Spaced Repetition
  • With spaced repetition, you repeat the same piece of information across increasing intervals.
  • The harder it feels to recall the information, the stronger the learning effect. “Spaced practice, which allows some forgetting to occur between sessions, strengthens both the learning and the cues and routes for fast retrieval,”
  • Why it works: It might sound counterintuitive, but forgetting is essential for learning. Spacing out practice might feel less productive than rereading a text because you’ll realize what you forgot. Your brain has to work harder to retrieve your knowledge, which is a good indicator of effective learning.
  • How I apply spaced repetition: After some weeks, I revisit a book and look at the summary questions (see #2). I try to come up with my answer before I look up my actual summary. I can often only remember a fraction of what I wrote and have to look at the rest.
  • “Knowledge trapped in books neatly stacked is meaningless and powerless until applied for the betterment of life.”
  • How you can apply spaced repetition: You can revisit your book summary medium of choice and test yourself on what you remember. What were your action points from the book? Have you applied them? If not, what hindered you?
  • By testing yourself in varying intervals on your book summaries, you’ll strengthen both learning and cues for fast retrieval.
  • Why interleaving works: Alternate working on different problems feels more difficult as it, again, facilitates forgetting.
  • How I apply interleaving: I read different books at the same time.
  • 1) Highlight everything you want to remember
  • #5 Self-Testing
  • While reading often falsely tricks us into perceived mastery, testing shows us whether we truly mastered the subject at hand. Self-testing helps you identify knowledge gaps and brings weak areas to the light
  • “It’s better to solve a problem than to memorize a solution.”
  • Why it works: Self-testing helps you overcome the illusion of knowledge. “One of the best habits a learner can instill in herself is regular self-quizzing to recalibrate her understanding of what she does and does not know.”
  • How I apply self-testing: I explain the key lessons from non-fiction books I want to remember to others. Thereby, I test whether I really got the concept. Often, I didn’t
  • instead of feeling frustrated, cognitive science made me realize that identifying knowledge gaps are a desirable and necessary effect for long-term remembering.
  • How you can apply self-testing: Teaching your lessons learned from a non-fiction book is a great way to test yourself. Before you explain a topic to somebody, you have to combine several mental tasks: filter relevant information, organize this information, and articulate it using your own vocabulary.
  • Now that I discovered how to use my Kindle as a learning device, I wouldn’t trade it for a paper book anymore. Here are the four steps it takes to enrich your e-reading experience
  • How you can apply interleaving: Your brain can handle reading different books simultaneously, and it’s effective to do so. You can start a new book before you finish the one you’re reading. Starting again into a topic you partly forgot feels difficult first, but as you know by now, that’s the effect you want to achieve.
  • it won’t surprise you that researchers proved highlighting to be ineffective. It’s passive and doesn’t create memory cues.
  • 2) Cut down your highlights in your browser
  • After you finished reading the book, you want to reduce your highlights to the essential part. Visit your Kindle Notes page to find a list of all your highlights. Using your desktop browser is faster and more convenient than editing your highlights on your e-reading device.
  • Now, browse through your highlights, delete what you no longer need, and add notes to the ones you really like. By adding notes to the highlights, you’ll connect the new information to your existing knowledge
  • 3) Use software to practice spaced repetitionThis part is the main reason for e-books beating printed books. While you can do all of the above with a little extra time on your physical books, there’s no way to systemize your repetition praxis.
  • Readwise is the best software to combine spaced repetition with your e-books. It’s an online service that connects to your Kindle account and imports all your Kindle highlights. Then, it creates flashcards of your highlights and allows you to export your highlights to your favorite note-taking app.
  • Common Learning Myths DebunkedWhile reading and studying evidence-based learning techniques I also came across some things I wrongly believed to be true.
  • #2 Effective learning should feel easyWe think learning works best when it feels productive. That’s why we continue to use ineffective techniques like rereading or highlighting. But learning works best when it feels hard, or as the authors of ‘Make It Stick’ write: “Learning that’s easy is like writing in sand, here today and gone tomorrow.”
  • In Conclusion
  • I developed and adjusted these strategies over two years, and they’re still a work in progress.
  • Try all of them but don’t force yourself through anything that doesn’t feel right for you. I encourage you to do your own research, add further techniques, and skip what doesn’t serve you
  • “In the case of good books, the point is not to see how many of them you can get through, but rather how many can get through to you.”— Mortimer J. Adler
krystalxu

Death and Its Concept - The Philosophers' Magazine - 0 views

  • Death and its concept are absolutely empty.
  • As there are already many speculations about some sort of 'next life,' I will focus on the view that death is real and marks the final end of an individual's life.
  • All we see when we look at death is a reflection of our own lives.
  • ...2 more annotations...
  • As Ludwig Wittgenstein famously put it, “Death is not an experience in life.”
  • Obsessive rumination on death, on the other hand, can lead us away from life.
manhefnawi

Bees understand the concept of zero | Science | AAAS - 0 views

  • we’re not the only species to consider “nothing” a number. Parrots and monkeys understand the concept of zero, and now bees have joined the club, too.
  • researchers trained 10 bees to identify the smaller of two numbers. Across a series of trials, they showed the insects two different pictures displaying a few black shapes on a white background. If the bees flew to the picture with the smaller number of shapes, they were given delicious sugar water, but if they flew toward the larger number, they were punished with bitter-tasting quinine.
  • a white background containing no shapes at all. Even though the bees had never seen an empty picture before, 64% of the time they chose this option rather than a picture containing two or three shapes,
  • ...2 more annotations...
  • This suggests that the insects understood that “zero” is less than two or three. And they weren’t just going for the empty picture because it was new and interesting
  • Advanced numerical abilities like this could give animals an evolutionary advantage, helping them keep track of predators and food sources. And if an insect can display such as thorough grasp of the number zero, write the researchers, then this ability may be more common in the animal kingdom than we think.
Javier E

Why the very concept of 'general knowledge' is under attack | Times2 | The Times - 0 views

  • why has University Challenge lasted, virtually unchanged, for so long?
  • The answer may lie in a famous theory about our brains put forward by the psychologist Raymond Cattell in 1963
  • Cattell divided intelligence into two categories: fluid and crystallised. Fluid intelligence refers to basic reasoning and other mental activities that require minimal learning — just an alert and flexible brain.
  • ...12 more annotations...
  • By contrast, crystallised intelligence is based on experience and the accumulation of knowledge. Fluid intelligence peaks at the age of about 20 then gradually declines, whereas crystallised intelligence grows through your life until you hit your mid-sixties, when you start forgetting things.
  • that explains much about University Challenge’s appeal. Because the contestants are mostly aged around 20 and very clever, their fluid intelligence is off the scale
  • On the other hand, because they have had only 20 years to acquire crystallised intelligence, their store of general knowledge is likely to be lacking in some areas.
  • In each episode there will be questions that older viewers can answer, thanks to their greater store of crystallised intelligence, but the students cannot. Therefore we viewers don’t feel inferior when confronted by these smart young people. On the contrary: we feel, in some areas, slightly superior.
  • there is a real threat to the future of University Challenge and much else of value in our society, and it is this. The very concept of “general knowledge” — of a widely accepted core of information that educated, inquisitive people should have in their memory banks — is under attack from two different groups.
  • It’s a brilliantly balanced format
  • They argue that all knowledge is contextual and that things taken for granted in the past — for instance, a canon of great authors that everyone should read at school — merely reflect an outdated, usually Eurocentric view of what’s intellectually important.
  • The first comprises the deconstructionists and decolonialists
  • The other group is the technocrats who argue that the extent of human knowledge is now so vast that it’s impossible for any individual to know more than, perhaps, one billionth of it
  • So why not leave it entirely to computers to do the heavy lifting of knowledge storing and recall, thus freeing our minds for creativity and problem solving?
  • The problem with the agitators on both sides of today’s culture wars is that they are forcefully trying to shape what’s accepted as general knowledge according to a blatant political agenda.
  • And the problem with relying on, say, Wikipedia’s 6.5 million English-language articles to store general knowledge for all of us? It’s the tacit implication that “mere facts” are too tedious to be clogging up our brains. From there it’s a short step to saying that facts don’t matter at all, that everything should be decided by “feelings”. And from there it’s an even shorter step to fake news and pernicious conspiracy theories, the belittling of experts and hard evidence, the closing of minds, the thickening of prejudice and the trivialisation of the national conversation.
sissij

The Linguistic Evolution of 'Like' - The Atlantic - 1 views

  • In our mouths or in print, in villages or in cities, in buildings or in caves, a language doesn’t sit still.
  • What we are seeing in like’s transformations today are just the latest chapters in a story that began with an ancient word that was supposed to mean “body.”
  • So today’s like did not spring mysteriously from a crowd on the margins of unusual mind-set and then somehow jump the rails from them into the general population. The seeds of the modern like lay among ordinary people; the Beatniks may not even have played a significant role in what happened later.
  • ...1 more annotation...
  • Like LOL, like, entrenched in all kinds of sentences, used subconsciously, and difficult to parse the real meaning of without careful consideration, has all the hallmarks of a piece of grammar—specifically, in the pragmatic department, modal wing.
  •  
    I found this article very interesting. The meaning of "like" does not spring from abstract concepts, it's rather related to ordinary people and physical meaning of "body". Also the meaning of similarity also reflects how people views the world. The way an ancient language develop to its modern form shows how the society evolves. --Sissi (11/27/2016)
kortanekev

Reasons To Believe : Anthropic Principle: A Precise Plan for Humanity - 0 views

  • The anthropic principle says that the universe appears "designed" for the sake of human life
  • To state the principle more dramatically, a preponderance of physical evidence points to humanity as the central theme of the cosmos.
  • Evidence of specific preparation for human existence shows up in the characteristics of the solar system, as well
  •  
    The Anthropic Principle, arguably the most human sentient there is. This principle is the concept that due to our timeline of creation, the universe has been created specifically for us. But what about  monkeys? They are as much a result of the laws of nature as we are, perhaps the universe was created for them! This idea of human centrality appears in religion as well as early science - through a geocentric model. But as we move further into the universe, it's clear to see there lays much more than us and such sentiments ... 
kushnerha

If Philosophy Won't Diversify, Let's Call It What It Really Is - The New York Times - 0 views

  • The vast majority of philosophy departments in the United States offer courses only on philosophy derived from Europe and the English-speaking world. For example, of the 118 doctoral programs in philosophy in the United States and Canada, only 10 percent have a specialist in Chinese philosophy as part of their regular faculty. Most philosophy departments also offer no courses on Africana, Indian, Islamic, Jewish, Latin American, Native American or other non-European traditions. Indeed, of the top 50 philosophy doctoral programs in the English-speaking world, only 15 percent have any regular faculty members who teach any non-Western philosophy.
  • Given the importance of non-European traditions in both the history of world philosophy and in the contemporary world, and given the increasing numbers of students in our colleges and universities from non-European backgrounds, this is astonishing. No other humanities discipline demonstrates this systematic neglect of most of the civilizations in its domain. The present situation is hard to justify morally, politically, epistemically or as good educational and research training practice.
  • While a few philosophy departments have made their curriculums more diverse, and while the American Philosophical Association has slowly broadened the representation of the world’s philosophical traditions on its programs, progress has been minimal.
  • ...9 more annotations...
  • Many philosophers and many departments simply ignore arguments for greater diversity; others respond with arguments for Eurocentrism that we and many others have refuted elsewhere. The profession as a whole remains resolutely Eurocentric.
  • Instead, we ask those who sincerely believe that it does make sense to organize our discipline entirely around European and American figures and texts to pursue this agenda with honesty and openness. We therefore suggest that any department that regularly offers courses only on Western philosophy should rename itself “Department of European and American Philosophy.”
  • We see no justification for resisting this minor rebranding (though we welcome opposing views in the comments section to this article), particularly for those who endorse, implicitly or explicitly, this Eurocentric orientation.
  • Some of our colleagues defend this orientation on the grounds that non-European philosophy belongs only in “area studies” departments, like Asian Studies, African Studies or Latin American Studies. We ask that those who hold this view be consistent, and locate their own departments in “area studies” as well, in this case, Anglo-European Philosophical Studies.
  • Others might argue against renaming on the grounds that it is unfair to single out philosophy: We do not have departments of Euro-American Mathematics or Physics. This is nothing but shabby sophistry. Non-European philosophical traditions offer distinctive solutions to problems discussed within European and American philosophy, raise or frame problems not addressed in the American and European tradition, or emphasize and discuss more deeply philosophical problems that are marginalized in Anglo-European philosophy. There are no comparable differences in how mathematics or physics are practiced in other contemporary cultures.
  • Of course, we believe that renaming departments would not be nearly as valuable as actually broadening the philosophical curriculum and retaining the name “philosophy.” Philosophy as a discipline has a serious diversity problem, with women and minorities underrepresented at all levels among students and faculty, even while the percentage of these groups increases among college students. Part of the problem is the perception that philosophy departments are nothing but temples to the achievement of males of European descent. Our recommendation is straightforward: Those who are comfortable with that perception should confirm it in good faith and defend it honestly; if they cannot do so, we urge them to diversify their faculty and their curriculum.
  • This is not to disparage the value of the works in the contemporary philosophical canon: Clearly, there is nothing intrinsically wrong with philosophy written by males of European descent; but philosophy has always become richer as it becomes increasingly diverse and pluralistic.
  • We hope that American philosophy departments will someday teach Confucius as routinely as they now teach Kant, that philosophy students will eventually have as many opportunities to study the “Bhagavad Gita” as they do the “Republic,” that the Flying Man thought experiment of the Persian philosopher Avicenna (980-1037) will be as well-known as the Brain-in-a-Vat thought experiment of the American philosopher Hilary Putnam (1926-2016), that the ancient Indian scholar Candrakirti’s critical examination of the concept of the self will be as well-studied as David Hume’s, that Frantz Fanon (1925-1961), Kwazi Wiredu (1931- ), Lame Deer (1903-1976) and Maria Lugones will be as familiar to our students as their equally profound colleagues in the contemporary philosophical canon. But, until then, let’s be honest, face reality and call departments of European-American Philosophy what they really are.
  • For demographic, political and historical reasons, the change to a more multicultural conception of philosophy in the United States seems inevitable. Heed the Stoic adage: “The Fates lead those who come willingly, and drag those who do not.”
Javier E

Why Our Children Don't Think There Are Moral Facts - NYTimes.com - 1 views

  • I already knew that many college-aged students don’t believe in moral facts.
  • the overwhelming majority of college freshman in their classrooms view moral claims as mere opinions that are not true or are true only relative to a culture.
  • where is the view coming from?
  • ...32 more annotations...
  • the Common Core standards used by a majority of K-12 programs in the country require that students be able to “distinguish among fact, opinion, and reasoned judgment in a text.”
  • So what’s wrong with this distinction and how does it undermine the view that there are objective moral facts?
  • For example, many people once thought that the earth was flat. It’s a mistake to confuse truth (a feature of the world) with proof (a feature of our mental lives)
  • Furthermore, if proof is required for facts, then facts become person-relative. Something might be a fact for me if I can prove it but not a fact for you if you can’t. In that case, E=MC2 is a fact for a physicist but not for me.
  • worse, students are taught that claims are either facts or opinions. They are given quizzes in which they must sort claims into one camp or the other but not both. But if a fact is something that is true and an opinion is something that is believed, then many claims will obviously be both
  • How does the dichotomy between fact and opinion relate to morality
  • Kids are asked to sort facts from opinions and, without fail, every value claim is labeled as an opinion.
  • Here’s a little test devised from questions available on fact vs. opinion worksheets online: are the following facts or opinions? — Copying homework assignments is wrong. — Cursing in school is inappropriate behavior. — All men are created equal. — It is worth sacrificing some personal liberties to protect our country from terrorism. — It is wrong for people under the age of 21 to drink alcohol. — Vegetarians are healthier than people who eat meat. — Drug dealers belong in prison.
  • The answer? In each case, the worksheets categorize these claims as opinions. The explanation on offer is that each of these claims is a value claim and value claims are not facts. This is repeated ad nauseum: any claim with good, right, wrong, etc. is not a fact.
  • In summary, our public schools teach students that all claims are either facts or opinions and that all value and moral claims fall into the latter camp. The punchline: there are no moral facts. And if there are no moral facts, then there are no moral truths.
  • It should not be a surprise that there is rampant cheating on college campuses: If we’ve taught our students for 12 years that there is no fact of the matter as to whether cheating is wrong, we can’t very well blame them for doing so later on.
  • If it’s not true that it’s wrong to murder a cartoonist with whom one disagrees, then how can we be outraged? If there are no truths about what is good or valuable or right, how can we prosecute people for crimes against humanity? If it’s not true that all humans are created equal, then why vote for any political system that doesn’t benefit you over others?
  • the curriculum sets our children up for doublethink. They are told that there are no moral facts in one breath even as the next tells them how they ought to behave.
  • Our children deserve a consistent intellectual foundation. Facts are things that are true. Opinions are things we believe. Some of our beliefs are true. Others are not. Some of our beliefs are backed by evidence. Others are not.
  • Value claims are like any other claims: either true or false, evidenced or not.
  • The hard work lies not in recognizing that at least some moral claims are true but in carefully thinking through our evidence for which of the many competing moral claims is correct.
  • Moral truths are not the same as scientific truths or mathematical truths. Yet they may still be used a guiding principle for our individual lives as well as our laws.But there is equal danger of giving moral judgments the designation of truth as there is in not doing so. Many people believe that abortion is murder on the same level as shooting someone with a gun. But many others do not. So is it true that abortion is murder?Moral principles can become generally accepted and then form the basis for our laws. But many long accepted moral principles were later rejected as being faulty. "Separate but equal" is an example. Judging homosexual relationships as immoral is another example.
  • Whoa! That Einstein derived an equation is a fact. But the equation represents a theory that may have to be tweaked at some point in the future. It may be a fact that the equation foretold the violence of atomic explosions, but there are aspects of nature that elude the equation. Remember "the theory of everything?"
  • Here is a moral fact, this is a sermon masquerading as a philosophical debate on facts, opinions and truth. This professor of religion is asserting that the government via common core is teaching atheism via the opinion vs fact.He is arguing, in a dishonest form, that public schools should be teaching moral facts. Of course moral facts is code for the Ten Commandments.
  • As a fourth grade teacher, I try to teach students to read critically, including distinguishing between facts and opinions as they read (and have been doing this long before the Common Core arrived, by the way). It's not always easy for children to grasp the difference. I can only imagine the confusion that would ensue if I introduced a third category -- moral "facts" that can't be proven but are true nonetheless!
  • horrible acts occur not because of moral uncertainty, but because people are too sure that their views on morality are 100% true, and anyone who fails to recognize and submit themselves are heathens who deserve death.I can't think of any case where a society has suffered because people are too thoughtful and open-minded to different perspectives on moral truth.In any case, it's not an elementary school's job to teach "moral truths."
  • The characterization of moral anti-realism as some sort of fringe view in philosophy is misleading. Claims that can be true or false are, it seems, 'made true' by features of the world. It's not clear to many in philosophy (like me) just what features of the world could make our moral claims true. We are more likely to see people's value claims as making claims about, and enforcing conformity to, our own (contingent) social norms. This is not to hold, as Mr. McBrayer seems to think follows, that there are no reasons to endorse or criticize these social norms.
  • This is nonsense. Giving kids the tools to distinguish between fact and opinion is hard enough in an age when Republicans actively deny reality on Fox News every night. The last thing we need is to muddy their thinking with the concept of "moral facts."A fact is a belief that everyone _should_ agree upon because it is observable and testable. Morals are not agreed upon by all. Consider the hot button issue of abortion.
  • Truthfully, I'm not terribly concerned that third graders will end up taking these lessons in the definition of fact versus opinion to the extremes considered here, or take them as a license to cheat. That will come much later, when they figure out, as people always have, what they can get a way with. But Prof. McBrayer, with his blithe expectation that all the grownups know that there moral "facts"? He scares the heck out of me.
  • I've long chafed at the language of "fact" v. "opinion", which is grounded in a very particular, limited view of human cognition. In my own ethics courses, I work actively to undermine the distinction, focusing instead on considered judgment . . . or even more narrowly, on consideration itself. (See http://wp.me/p5Ag0i-6M )
  • The real waffle here is the very concept of "moral facts." Our statements of values, even very important ones are, obviously, not facts. Trying to dress them up as if they are facts, to me, argues for a pretty serious moral weakness on the part of those advancing the idea.
  • Our core values are not important because they are facts. They are important because we collectively hold them and cherish them. To lean on the false crutch of "moral facts" to admit the weakness of your own moral convictions.
  • I would like to believe that there is a core of moral facts/values upon which all humanity can agree, but it would be tough to identify exactly what those are.
  • For the the ancient philosophers, reality comprised the Good, the True, and the Beautiful (what we might now call ethics, science and art), seeing these as complementary and inseparable, though distinct, realms. With the ascendency of science in our culture as the only valid measure of reality to the detriment of ethics and art (that is, if it is not observable and provable, it is not real), we have turned the good and the beautiful into mere "social constructs" that have no validity on their own. While I am sympathetic in many ways with Dr. McBrayer's objections, I think he falls into the trap of discounting the Good and The Beautiful as valid in and of themselves, and tries, instead, to find ways to give them validity through the True. I think his argument would have been stronger had he used the language of validity rather than the language of truth. Goodness, Truth and Beauty each have their own validity, though interdependent and inseparable. When we artificially extract one of these and give it primacy, we distort reality and alienate ourselves from it.
  • Professor McBrayer seems to miss the major point of the Common Core concern: can students distinguish between premises based on (reasonably construed) fact and premises based on emotion when evaluating conclusions? I would prefer that students learn to reason rather than be taught moral 'truth' that follows Professor McBrayer's logic.
  • Moral issues cannot scientifically be treated on the level that Prof. McBrayer is attempting to use in this column: true or false, fact or opinion or both. Instead, they should be treated as important characteristics of the systematic working of a society or of a group of people in general. One can compare the working of two groups of people: one in which e.g. cheating and lying is acceptable, and one in which they are not. One can use historical or model examples to show the consequences and the working of specific systems of morals. I think that this method - suitably adjusted - can be used even in second grade.
  • Relativism has nothing to do with liberalism. The second point is that I'm not sure it does all that much harm, because I have yet to encounter a student who thought that he or she had to withhold judgment on those who hold opposing political views!
kushnerha

How 'Empowerment' Became Something for Women to Buy - The New York Times - 0 views

  • The mix of things presumed to transmit and increase female power is without limit yet still depressingly limiting.“Empowerment” wasn’t always so trivialized, or so corporate, or even so clamorously attached to women.
  • Four decades ago, the word had much more in common with Latin American liberation theology than it did with “Lean In.” In 1968, the Brazilian academic Paulo Freire coined the word “conscientization,” empowerment’s precursor, as the process by which an oppressed person perceives the structural conditions of his oppression and is subsequently able to take action against his oppressors.
  • Eight years later, the educator Barbara Bryant Solomon, writing about American black communities, gave this notion a new name, “empowerment.” It was meant as an ethos for social workers in marginalized communities, to discourage paternalism and encourage their clients to solve problems in their own ways. Then in 1981, Julian Rappaport, a psychologist, broadened the concept into a political theory of power that viewed personal competency as fundamentally limitless; it placed faith in the individual and laid at her feet a corresponding amount of responsibility too.
  • ...7 more annotations...
  • Sneakily, empowerment had turned into a theory that applied to the needy while describing a process more realistically applicable to the rich. The word was built on a misaligned foundation; no amount of awareness can change the fact that it’s the already-powerful who tend to experience empowerment at any meaningful rate. Today “empowerment” invokes power while signifying the lack of it. It functions like an explorer staking a claim on new territory with a white flag.
  • highly marketable “women’s empowerment,” neither practice nor praxis, nor really theory, but a glossy, dizzying product instead. Women’s empowerment borrows the virtuous window-dressing of the social worker’s doctrine and kicks its substance to the side. It’s about pleasure, not power; it’s individualistic and subjective, tailored to insecurity and desire.
  • The new empowerment doesn’t increase potential so much as it assures you that your potential is just fine. Even when the thing being described as “empowering” is personal and mildly defiant (not shaving, not breast-feeding, not listening to men, et cetera), what’s being mar­keted is a certain identity.
  • When consumer purchases aren’t made out to be a path to female empowerment, a branded corporate experience often is. There’s TEDWomen (“about the power of women”), the Forbes Women’s Summit (“#RedefinePower”) and Fortune’s Most Powerful Women Conference (tickets are $10,000).
  • This consumption-and-conference empowerment dilutes the word to pitch-speak, and the concept to something that imitates rather than alters the structures of the world. This version of empowerment can be actively disempowering: It’s a series of objects and experiences you can purchase while the conditions determining who can access and accumulate power stay the same. The ready partici­pation of well-off women in this strat­egy also points to a deep truth about the word “empowerment”: that it has never been defined by the people who actually need it. People who talk empowerment are, by definition, already there.
  • I have never said “empowerment” sincerely or heard it from a single one of my friends. The formulation has been diluted to something representational and bloodless — an architectural rendering of a building that will never be built.But despite its nonexistence in honest conversation, “empowerment” goes on thriving. It’s uniquely marketable, like the female body, which is where women’s empowerment is forced to live.
  • Like Sandberg, Kardashian is the apotheosis of a particular brand of largely contentless feminism, a celebratory form divorced from material politics, which makes it palatable — maybe irresistible — to the business world. Advertisement Continue reading the main story The mistake would be to locate further empowerment in choosing between the two. Corporate empowerment — as well as the lightweight, self-exculpatory feminism it rides on — feeds rav­enously on the distracting performance of identity, that buffet of false opposition.
sissij

Quantum Gravity Loops Back To Ancient Atomic Logic, and The Big Bang Becomes A Big Boun... - 0 views

  • Greeks had the “first true alphabet”: a “universal” writing system that used a few letters to encode the infinite variety of all possible utterances. Similarly, all matter is written in a "language… of atoms."
  • Mysterious “meanings” still surround 100-year-old quantum mechanics equations
  • Their meaning/function/grammar is relational and sequential and word-like. The information encoded in matching sequential text-like compositions matters (DNA—>RNA, letters—>“social cartesian” lexicon).
  • ...3 more annotations...
  • Beyond the grammars of geometry and algebra lies a domain of not math-like but text-like compositions and meanings (of semantics beyond mathematics).
  • 17. Word and world both have grammars that don’t fit our available mathematical rules.
  • 18. Reality is relational, and not entirely objective. Subject and object aren’t separable, they’re entangled, inescapably. “Objective” is always relative to some other system/observer. 
  •  
    I find it very interesting that the author is trying o look at the world from a different perspective than mathematics. He thinks atoms as a language that have grammar and meanings. He thinks mathematical rules cannot fully explain our world because it is too objective. He involves the idea of language to describe how the world is entangled and relational. As we learned in TOK, language is an important AOK that shows human civilization in a very complicated way. Language is flexible, emotional and relational. It gives things meaning as human likes to assign meaning and pattern to things around. The world around us are not just cold fact, we as observers give them meaning to exist. In that sense, the concept of language can better help us depict the world. --Sissi (2/27/2017)
sissij

When Did 'Ambition' Become a Dirty Word? - The New York Times - 0 views

  • but is instead a stark, black-and-white video, a public service announcement that takes on a thorny issue that dominated the last presidential campaign and has divided people on the right and left.
  • “Embrace Ambition,”
  • “I can think of a lot of dirty words,” Ms. Witherspoon says. “Ambition is not one of them.”
  • ...6 more annotations...
  • Nevertheless, she seemed to choose her words carefully as she spoke about the campaign.
  • she wanted to get away from the idea that this project was politically motivated, or anti-Trump.
  • But the issue of ambition, and the way it is used to defame women, is nevertheless personal to her.
  • This was confusing to Ms. Burch, who never saw herself as being a particularly threatening person.
  • “I do it, too,” she said. “I’m guilty of all of it.”
  • And the word “feminist” began to shed its Bella Abzug and Betty Friedan connotations, as women like Madonna went from saying they are “not feminists” but “humanists” to wearing T-shirts at anti-Trump events that had the word “feminist” emblazoned across the center.
  •  
    Sometimes, people are ashamed of their ambition. They are afraid to have a dream because they would not bear any failure. I think people don't dare to dream big now. I think it might be because of the concept of economics we have in the recent years. Most people follow the efficiency rule so nobody is willing to risk and make a revolution. Putting on big stake is not the most efficient choice in economics. Although it tells us it is okay to be greedy, but it limited us to the model and make us less willingly to make a breakthrough. --Sissi (3/2/2017)
sissij

Why Does Time Seem to Move Faster as We Grow Older? | Big Think - 0 views

  • Today it feels like the days peel on by, and a vacation which may last days or weeks is gone in mere hours. This is a common human experience.
  • There are an awful lot of theories that give us insight. But a direct scientific law so far remains elusive.
  • In 1877, the “ratio theory” was born, which states that we compare intervals to the total amount of time we’ve been alive. This can also be called “proportional theory,” the idea that as we age, our sense of the present begins to feel short as compared to our total lifespan.
  • ...4 more annotations...
  • While youth colors everything anew, as we age, we become more or less familiar with every aspect of our surroundings, and so the nuance wears off.
  • Young people tended to select steady or unchanging metaphors for time, while older people chose those which had to do with swiftness and speed.
  • This creates a “reminiscence bump.” The farther we move away from the bump, the quicker time seems to move.
  • A neurochemical explanation lies in the dopamine level in our brains. This is the pleasure neurotransmitter that gives us a feeling of wellbeing and reward.
  •  
    I have sometimes think about this issue and the explanation I come up is that when we are seven, one year is one seventh of our live. When we are seventy, the denominator get much bigger and the proportion get significantly smaller, one seventieth. I am very surprised to find that the scientists have similar theory. It also shows how unreliable our sense perception is. We are very unreliable about time since time is a concept that we cannot see or touch. How we feel about time largely depends on our individual emotion. Our brain also tends to remember the first times and we also have a tendency to believe vivid memory. --Sissi (2/8/2017)
Duncan H

Living in the Material World - NYTimes.com - 0 views

  • on a visit to the Academy of Sciences in Almaty some years ago I was presented with a souvenir meant to assure me that Central Asia was indeed still producing philosophy worthy of note. It was a collectively authored book entitled “The Development of Materialist Dialectics in Kazakhstan,” and I still display it proudly on my shelf. Its rough binding and paper bespeak economic hardship. It is packed with the traces of ideas, yet everything about the book announces its materiality.I had arrived in the Kazakh capital 1994, just in time to encounter the last of a dying breed: the philosopher as party functionary (they are all by now retired, dead or defenestrated, or have simply given up on what they learned in school). The book, written by committee, was a collection of official talking points, and what passed for conversation there was something much closer to recitation.
  • The philosophical meaning of materialism may in the final analysis be traced back to a religious view of the world. On this view, to focus on the material side of existence is to turn away from the eternal and divine. Here, the category of the material is assimilated to that of sin or evil.
  • Yet in fact this feature of Marxist philosophical classification is one that, with some variations, continues to be shared by all philosophers, even in the West, even today
  • ...9 more annotations...
  • materialism is not the greedy desire for material goods, but rather the belief that the fundamental reality of the world is material;
  • idealism is not the aspiration toward lofty and laudable goals, but rather the belief that the fundamental reality of the world is mental or idea-like. English-speaking philosophers today tend to speak of “physicalism” or “naturalism” rather than materialism (perhaps to avoid confusion with the Wall Street sense of the term). At the same time, Anglo-American historians of philosophy continue to find the distinction between materialism and idealism a useful one in our attempts at categorizing past schools of thought. Democritus and La Mettrie were materialists; Hobbes was pretty close. Berkeley and Kant were idealists; Leibniz may have been.
  • And it was these paradoxes that led the Irish philosopher to conclude that talk of matter was but a case of multiplying entities beyond necessity. For Berkeley, all we can know are ideas, and for this reason it made sense to suppose that the world itself consists in ideas.
  • Central to this performance was the concept of  “materialism.” The entire history of philosophy, in fact, was portrayed in Soviet historiography as a series of matches between the materialist home-team and its “idealist” opponents, beginning roughly with Democritus (good) and Plato (bad), and culminating in the opposition between official party philosophy and logical positivism, the latter of which was portrayed as a shrouded variety of idealism. Thus from the “Short Philosophical Dictionary,” published in Moscow in 1951, we learn that the school of logical empiricism represented by Rudolf Carnap, Otto Neurath and others, “is a form of subjective idealism, characteristic of degenerating bourgeois philosophy in the epoch of the decline of capitalism.”Now the Soviet usage of this pair of terms appears to fly in the face of our ordinary, non-philosophical understanding of them (that, for example,  Wall Street values are “materialist,” while the Occupy movement is “idealist”). One might have thought that the communists should be flinging the “materialist” label at their capitalist enemies, rather than claiming it for themselves. One might also have thought that the Bolshevik Revolution and the subsequent failed project of building a workers’ utopia was nothing if not idealistic.
  • one great problem with the concept of materialism is that it says very little in itself. What is required in addition is an elaboration of what a given thinker takes matter, or ideas, to be. It may not be just the Marxist aftertaste, but also the fact that the old common-sense idea about matter as brute, given stuff has turned out to have so little to do with the way the physical world actually is, that has led Anglo-American philosophers to prefer to associate themselves with the “physical” or the “natural” rather than with the material.  Reality, they want to say, is just what is natural, while everything else is in turn “supernatural” (this distinction has its clarity going for it, but it also seems uncomfortably close to tautology). Not every philosopher has a solid grasp of subatomic physics, but most know enough to grasp that, even if reality is eventually exhaustively accounted for through an enumeration of the kinds of particles and a few basic forces, this reality will still look nothing like what your average person-in-the-street takes reality to be.
  • The 18th-century idealist philosopher George Berkeley strongly believed that matter was only a fiction contrived by philosophers in the first place, for which the real people had no need. For Berkeley, there was never anything common-sensical about matter. We did not need to arrive at the era of atom-splitting and wave-particle duality, then, in order for the paradoxes inherent in matter to make themselves known (is it infinitely divisible or isn’t it?
  • Soviet and Western Marxists alike, by stark contrast, and before them the French “vulgar” (i.e., non-dialectical) materialists of the 18th century, saw and see the material world as the base and cause of all mental activity, as both bringing ideas into existence, and also determining the form and character of a society’s ideas in accordance with the state of its technology, its methods of resource extraction and its organization of labor. So here to focus on the material is not to become distracted from the true source of being, but rather to zero right in on it.
  • Consider money. Though it might sometimes be represented by bank notes or coins, money is an immaterial thing par excellence, and to seek to acquire it is to move on the plane of ideas. Of course, money can also be converted into material things, yet it seems simplistic to suppose that we want money only in order to convert it into the material things we really want, since even these material things aren’t just material either: they are symbolically dense artifacts, and they convey to others certain ideas about their owners. This, principally, is why their owners want them, which is to say that materialists (in the everyday sense) are trading in ideas just as much as anyone else.
  • In the end no one really cares about stuff itself. Material acquisitions — even, or perhaps especially, material acquisitions of things like Rolls Royces and Rolexes — are maneuvers within a universe of materially instantiated ideas. This is human reality, and it is within this reality that mystics, scientists, and philosophers alike are constrained to pursue their various ends, no matter what they might take the ultimate nature of the external world to be.
  •  
    A very interesting article on the contrast between materialism and idealism.
Javier E

Why It's OK to Let Apps Make You a Better Person - Evan Selinger - Technology - The Atl... - 0 views

  • one theme emerges from the media coverage of people's relationships with our current set of technologies: Consumers want digital willpower. App designers in touch with the latest trends in behavioral modification--nudging, the quantified self, and gamification--and good old-fashioned financial incentive manipulation, are tackling weakness of will. They're harnessing the power of payouts, cognitive biases, social networking, and biofeedback. The quantified self becomes the programmable self.
  • the trend still has multiple interesting dimensions
  • Individuals are turning ever more aspects of their lives into managerial problems that require technological solutions. We have access to an ever-increasing array of free and inexpensive technologies that harness incredible computational power that effectively allows us to self-police behavior everywhere we go. As pervasiveness expands, so does trust.
  • ...20 more annotations...
  • Some embrace networked, data-driven lives and are comfortable volunteering embarrassing, real time information about what we're doing, whom we're doing it with, and how we feel about our monitored activities.
  • Put it all together and we can see that our conception of what it means to be human has become "design space." We're now Humanity 2.0, primed for optimization through commercial upgrades. And today's apps are more harbinger than endpoint.
  • philosophers have had much to say about the enticing and seemingly inevitable dispersion of technological mental prosthetic that promise to substitute or enhance some of our motivational powers.
  • beyond the practical issues lie a constellation of central ethical concerns.
  • they should cause us to pause as we think about a possible future that significantly increases the scale and effectiveness of willpower-enhancing apps. Let's call this hypothetical future Digital Willpower World and characterize the ethical traps we're about to discuss as potential general pitfalls
  • it is antithetical to the ideal of " resolute choice." Some may find the norm overly perfectionist, Spartan, or puritanical. However, it is not uncommon for folks to defend the idea that mature adults should strive to develop internal willpower strong enough to avoid external temptations, whatever they are, and wherever they are encountered.
  • In part, resolute choosing is prized out of concern for consistency, as some worry that lapse of willpower in any context indicates a generally weak character.
  • Fragmented selves behave one way while under the influence of digital willpower, but another when making decisions without such assistance. In these instances, inconsistent preferences are exhibited and we risk underestimating the extent of our technological dependency.
  • It simply means that when it comes to digital willpower, we should be on our guard to avoid confusing situational with integrated behaviors.
  • the problem of inauthenticity, a staple of the neuroethics debates, might arise. People might start asking themselves: Has the problem of fragmentation gone away only because devices are choreographing our behavior so powerfully that we are no longer in touch with our so-called real selves -- the selves who used to exist before Digital Willpower World was formed?
  • Infantalized subjects are morally lazy, quick to have others take responsibility for their welfare. They do not view the capacity to assume personal responsibility for selecting means and ends as a fundamental life goal that validates the effort required to remain committed to the ongoing project of maintaining willpower and self-control.
  • Michael Sandel's Atlantic essay, "The Case Against Perfection." He notes that technological enhancement can diminish people's sense of achievement when their accomplishments become attributable to human-technology systems and not an individual's use of human agency.
  • Borgmann worries that this environment, which habituates us to be on auto-pilot and delegate deliberation, threatens to harm the powers of reason, the most central component of willpower (according to the rationalist tradition).
  • In several books, including Technology and the Character of Contemporary Life, he expresses concern about technologies that seem to enhance willpower but only do so through distraction. Borgmann's paradigmatic example of the non-distracted, focally centered person is a serious runner. This person finds the practice of running maximally fulfilling, replete with the rewarding "flow" that can only comes when mind/body and means/ends are unified, while skill gets pushed to the limit.
  • Perhaps the very conception of a resolute self was flawed. What if, as psychologist Roy Baumeister suggests, willpower is more "staple of folk psychology" than real way of thinking about our brain processes?
  • novel approaches suggest the will is a flexible mesh of different capacities and cognitive mechanisms that can expand and contract, depending on the agent's particular setting and needs. Contrary to the traditional view that identifies the unified and cognitively transparent self as the source of willed actions, the new picture embraces a rather diffused, extended, and opaque self who is often guided by irrational trains of thought. What actually keeps the self and its will together are the given boundaries offered by biology, a coherent self narrative created by shared memories and experiences, and society. If this view of the will as an expa
  • nding and contracting system with porous and dynamic boundaries is correct, then it might seem that the new motivating technologies and devices can only increase our reach and further empower our willing selves.
  • "It's a mistake to think of the will as some interior faculty that belongs to an individual--the thing that pushes the motor control processes that cause my action," Gallagher says. "Rather, the will is both embodied and embedded: social and physical environment enhance or impoverish our ability to decide and carry out our intentions; often our intentions themselves are shaped by social and physical aspects of the environment."
  • It makes perfect sense to think of the will as something that can be supported or assisted by technology. Technologies, like environments and institutions can facilitate action or block it. Imagine I have the inclination to go to a concert. If I can get my ticket by pressing some buttons on my iPhone, I find myself going to the concert. If I have to fill out an application form and carry it to a location several miles away and wait in line to pick up my ticket, then forget it.
  • Perhaps the best way forward is to put a digital spin on the Socratic dictum of knowing myself and submit to the new freedom: the freedom of consuming digital willpower to guide me past the sirens.
Javier E

The Age of 'Infopolitics' - NYTimes.com - 0 views

  • we need a new way of thinking about our informational milieu. What we need is a concept of infopolitics that would help us understand the increasingly dense ties between politics and information
  • Infopolitics encompasses not only traditional state surveillance and data surveillance, but also “data analytics” (the techniques that enable marketers at companies like Target to detect, for instance, if you are pregnant), digital rights movements (promoted by organizations like the Electronic Frontier Foundation), online-only crypto-currencies (like Bitcoin or Litecoin), algorithmic finance (like automated micro-trading) and digital property disputes (from peer-to-peer file sharing to property claims in the virtual world of Second Life)
  • Surveying this iceberg is crucial because atop it sits a new kind of person: the informational person. Politically and culturally, we are increasingly defined through an array of information architectures: highly designed environments of data, like our social media profiles, into which we often have to squeeze ourselves
  • ...12 more annotations...
  • We have become what the privacy theorist Daniel Solove calls “digital persons.” As such we are subject to infopolitics (or what the philosopher Grégoire Chamayou calls “datapower,” the political theorist Davide Panagia “datapolitik” and the pioneering thinker Donna Haraway “informatics of domination”).
  • Once fingerprints, biometrics, birth certificates and standardized names were operational, it became possible to implement an international passport system, a social security number and all other manner of paperwork that tells us who someone is. When all that paper ultimately went digital, the reams of data about us became radically more assessable and subject to manipulation,
  • We like to think of ourselves as somehow apart from all this information. We are real — the information is merely about us.
  • But what is it that is real? What would be left of you if someone took away all your numbers, cards, accounts, dossiers and other informational prostheses? Information is not just about you — it also constitutes who you are.
  • information is a site for the call of justice today, alongside more quintessential battlefields like liberty of thought and equality of opportunity.
  • agencies and corporations will continue producing new visions of you and me, and they will do so without our input if we remain stubbornly attached to antiquated conceptions of selfhood that keep us from admitting how informational we already are.
  • What should we do about our Internet and phone patterns’ being fastidiously harvested and stored away in remote databanks where they await inspection by future algorithms developed at the National Security Agency, Facebook, credit reporting firms like Experian and other new institutions of information and control that will come into existence in future decades?
  • What bits of the informational you will fall under scrutiny? The political you? The sexual you? What next-generation McCarthyisms await your informational self? And will those excesses of oversight be found in some Senate subcommittee against which we democratic citizens might hope to rise up in revolt — or will they lurk among algorithmic automatons that silently seal our fates in digital filing systems?
  • Despite their decidedly different political sensibilities, what links together the likes of Senator Wyden and the international hacker network known as Anonymous is that they respect the severity of what is at stake in our information.
  • We understandably do not want to see ourselves as bits and bytes. But unless we begin conceptualizing ourselves in this way, we leave it to others to do it for us
  • we lack the intellectual framework to grasp the new kinds of political injustices characteristic of today’s information society.
  • though nearly all of us have a vague sense that something is wrong with the new regimes of data surveillance, it is difficult for us to specify exactly what is happening and why it raises serious concern
Sophia C

BBC News - Viewpoint: Human evolution, from tree to braid - 0 views

  • What was, in my view, a logical conclusion reached by the authors was too much for some researchers to take.
  • he conclusion of the Dmanisi study was that the variation in skull shape and morphology observed in this small sample, derived from a single population of Homo erectus, matched the entire variation observed among African fossils ascribed to three species - H. erectus, H. habilis and H. rudolfensis.
  • a single population of H. erectus,
  • ...13 more annotations...
  • They all had to be the same species.
  • was not surprising to find that Neanderthals and modern humans interbred, a clear expectation of the biological species concept.
  • I wonder when the penny will drop: when we have five pieces of a 5,000-piece jigsaw puzzle, every new bit that we add is likely to change the picture.
  • e identity of the fourth player remains unknown but it was an ancient lineage that had been separate for probably over a million years. H. erectus seems a likely candidate. Whatever the name we choose to give this mystery lineage, what these results show is that gene flow was possible not just among contemporaries but also between ancient and more modern lineages.
  • cientists succeeded in extracting the most ancient mitochondrial DNA so far, from the Sima de los Huesos site in Atapuerca, Spain.
  • We have built a picture of our evolution based on the morphology of fossils and it was wrong.
    • Sophia C
       
      Kuhn
  • when we know how plastic - or easily changeable - skull shape is in humans. And our paradigms must also change.
  • e must abandon, once and for all, views of modern human superiority over archaic (ancient) humans. The terms "archaic" and "modern" lose all meaning as do concepts of modern human replacement of all other lineages.
  • he deep-rooted shackles that have sought to link human evolution with stone tool-making technological stages - the Stone Ages - even when we have known that these have overlapped with each other for half-a-million years in some instances.
  • e world of our biological and cultural evolution was far too fluid for us to constrain it into a few stages linked by transitions.
  • We have to flesh out the genetic information and this is where archaeology comes into the picture.
  • Rather than focus on differences between modern humans and Neanderthals, what the examples show is the range of possibilities open to humans (Neanderthals included) in different circumstances.
  • research using new technology on old archaeological sites, as at La Chapelle; and
julia rhodes

How people learn - The Week - 0 views

  • n a traditional classroom, the teacher stands at the front of the class explaining what is clear in their mind to a group of passive students. Yet this pedagogical strategy doesn't positively impact retention of information from lecture, improve understanding basic concepts, or affect beliefs (that is, does new information change your belief about how something works).
  • . Everything that constitutes "understanding" science and "thinking scientifically" resides in the long-term memory, which is developed via the construction and assembly of component proteins.
  • The research tells us that the human brain can hold a maximum of about seven different items in its short-term working memory and can process no more than about four ideas at once. Exactly what an "item" means when translated from the cognitive science lab into the classroom is a bit fuzzy.
  • ...13 more annotations...
  • The results were similarly disturbing when students were tested to determine understanding of basic concepts. More instruction wasn't helping students advance from novice to expert. In fact, the data indicated the opposite: students had more novice-like beliefs after they completed a course than they had when they started.
  • But in addition, experts have a mental organizational structure that facilitates the retrieval and effective application of their knowledge.
  • experts have an ability to monitor their own thinking ("metacognition"), at least in their discipline of expertise. They are able to ask themselves, "Do I understand this? How can I check my understanding?"
  • But that is not what cognitive science tells us. It tells us instead that students need to develop these different ways of thinking by means of extended, focused mental effort.
  • new ways of thinking are always built on the prior thinking of the individual, so if the educational process is to be successful, it is essential to take that prior thinking into account.
  • Given that lectures were devised as a means of transferring knowledge from one to many, it seems obvious that we would ensure that people retain the information they are consuming.
  • What is elementary, worldly wisdom? Well, the first rule is that you can't really know anything if you just remember isolated facts and try and bang 'em back. If the facts don't hang together on a latticework of theory, you don't have them in a usable form.
  • "So it makes perfect sense," Wieman writes, "that they are not learning to think like experts, even though they are passing science courses by memorizing facts and problem-solving recipes."
  • Anything one can do to reduce cognitive load improves learning.
  • A second way teachers can improve instruction is by recognizing the importance of student beliefs about science
  • My third example of how teaching and learning can be improved is by implementing the principle that effective teaching consists of engaging students, monitoring their thinking, and providing feedback.
  • I assign students to groups the first day of class (typically three to four students in adjacent seats) and design each lecture around a series of seven to 10 clicker questions that cover the key learning goals for that day.
  • The process of critiquing each other's ideas in order to arrive at a consensus also enormously improves both their ability to carry on scientific discourse and to test their own understanding. [Change]
julia rhodes

Emergence in Social Groups and in Brains | Psychology Today - 0 views

  • Emergence is an important occurrence in social systems ranging in size from small families to large international organizations.
  • A system is demergent if the whole is less than the sum of its parts, that is, if the interaction of the parts prevents the whole from having valuable properties that it otherwise might have.
  • My current favorite example of a demergent system is the U.S. Congress, which has many talented people but is failing horribly to pass legislation because of toxic behaviors and interactions.
  • ...7 more annotations...
  • Of course, the problem with the House of Representatives is not just interactions, because some of its parts (members) have serious ideological defects.
  • But collaborations can also be demergent, if individuals do not have the intellectual flexibility to build new ideas rather than stick with old ones and waste each other’s time.
  • The brain is a marvel of emergence. I am currently working on a new theory of consciousness that explains it as the result of three neural mechanisms: representation by patterns of firing in neural populations, binding by neural processes that combine representations into new ones called semantic pointers, and competition among semantic pointers to represent the most salient aspects of the current situation.
  • For example the concept of zero cannot be based on perception, but can come about by combining the concepts of quantity and absence.
  • When brains malfunction, however, they can have demergent properties.
  • In epilepsy, individual neurons fire appropriately, but the whole brain develops abnormal patterns that can lead to unconsciousness or convulsions.
  • Figuring out how to have emergence replace demergence is important for neural functioning, and also for effective social functioning.
Javier E

The Dangers of Pseudoscience - NYTimes.com - 0 views

  • the “demarcation problem,” the issue of what separates good science from bad science and pseudoscience (and everything in between). The problem is relevant for at least three reasons.
  • The first is philosophical: Demarcation is crucial to our pursuit of knowledge; its issues go to the core of debates on epistemology and of the nature of truth and discovery.
  • The second reason is civic: our society spends billions of tax dollars on scientific research, so it is important that we also have a good grasp of what constitutes money well spent in this regard.
  • ...18 more annotations...
  • Third, as an ethical matter, pseudoscience is not — contrary to popular belief — merely a harmless pastime of the gullible; it often threatens people’s welfare,
  • It is precisely in the area of medical treatments that the science-pseudoscience divide is most critical, and where the role of philosophers in clarifying things may be most relevant.
  • some traditional Chinese remedies (like drinking fresh turtle blood to alleviate cold symptoms) may in fact work
  • There is no question that some folk remedies do work. The active ingredient of aspirin, for example, is derived from willow bark, which had been known to have beneficial effects since the time of Hippocrates. There is also no mystery about how this happens: people have more or less randomly tried solutions to their health problems for millennia, sometimes stumbling upon something useful
  • What makes the use of aspirin “scientific,” however, is that we have validated its effectiveness through properly controlled trials, isolated the active ingredient, and understood the biochemical pathways through which it has its effects
  • In terms of empirical results, there are strong indications that acupuncture is effective for reducing chronic pain and nausea, but sham therapy, where needles are applied at random places, or are not even pierced through the skin, turn out to be equally effective (see for instance this recent study on the effect of acupuncture on post-chemotherapy chronic fatigue), thus seriously undermining talk of meridians and Qi lines
  • Asma at one point compares the current inaccessibility of Qi energy to the previous (until this year) inaccessibility of the famous Higgs boson,
  • But the analogy does not hold. The existence of the Higgs had been predicted on the basis of a very successful physical theory known as the Standard Model. This theory is not only exceedingly mathematically sophisticated, but it has been verified experimentally over and over again. The notion of Qi, again, is not really a theory in any meaningful sense of the word. It is just an evocative word to label a mysterious force
  • Philosophers of science have long recognized that there is nothing wrong with positing unobservable entities per se, it’s a question of what work such entities actually do within a given theoretical-empirical framework. Qi and meridians don’t seem to do any, and that doesn’t seem to bother supporters and practitioners of Chinese medicine. But it ought to.
  • what’s the harm in believing in Qi and related notions, if in fact the proposed remedies seem to help?
  • we can incorporate whatever serendipitous discoveries from folk medicine into modern scientific practice, as in the case of the willow bark turned aspirin. In this sense, there is no such thing as “alternative” medicine, there’s only stuff that works and stuff that doesn’t.
  • Second, if we are positing Qi and similar concepts, we are attempting to provide explanations for why some things work and others don’t. If these explanations are wrong, or unfounded as in the case of vacuous concepts like Qi, then we ought to correct or abandon them.
  • pseudo-medical treatments often do not work, or are even positively harmful. If you take folk herbal “remedies,” for instance, while your body is fighting a serious infection, you may suffer severe, even fatal, consequences.
  • Indulging in a bit of pseudoscience in some instances may be relatively innocuous, but the problem is that doing so lowers your defenses against more dangerous delusions that are based on similar confusions and fallacies. For instance, you may expose yourself and your loved ones to harm because your pseudoscientific proclivities lead you to accept notions that have been scientifically disproved, like the increasingly (and worryingly) popular idea that vaccines cause autism.
  • Philosophers nowadays recognize that there is no sharp line dividing sense from nonsense, and moreover that doctrines starting out in one camp may over time evolve into the other. For example, alchemy was a (somewhat) legitimate science in the times of Newton and Boyle, but it is now firmly pseudoscientific (movements in the opposite direction, from full-blown pseudoscience to genuine science, are notably rare).
  • The verdict by philosopher Larry Laudan, echoed by Asma, that the demarcation problem is dead and buried, is not shared by most contemporary philosophers who have studied the subject.
  • the criterion of falsifiability, for example, is still a useful benchmark for distinguishing science and pseudoscience, as a first approximation. Asma’s own counterexample inadvertently shows this: the “cleverness” of astrologers in cherry-picking what counts as a confirmation of their theory, is hardly a problem for the criterion of falsifiability, but rather a nice illustration of Popper’s basic insight: the bad habit of creative fudging and finagling with empirical data ultimately makes a theory impervious to refutation. And all pseudoscientists do it, from parapsychologists to creationists and 9/11 Truthers.
  • The borderlines between genuine science and pseudoscience may be fuzzy, but this should be even more of a call for careful distinctions, based on systematic facts and sound reasoning. To try a modicum of turtle blood here and a little aspirin there is not the hallmark of wisdom and even-mindedness. It is a dangerous gateway to superstition and irrationality.
anonymous

Why executives are embracing philosophy - 0 views

  • This image of manager-as-philosopher might seem unrealistic. And yet, a new approach is spreading fast among top executives of Swiss companies, thanks to specialised university lectures and intellectual seminars of ancient and modern philosophy.
  • In the end, philosophy doesn’t offer answers, only questions. Or rather, it encourages questioning such concepts as work, the representation of capital, authority and ethics.
  • A study carried out by Eric Davoine, one of the originators of the course, showed that people who had attended it had developed some new skills. They were more prone to introspection, which helped boost their self-confidence. They also showed an improved capacity to formulate concepts and were more able to question themselves, which made constructive criticism easier and made them more inclined to listen to others.
‹ Previous 21 - 40 of 312 Next › Last »
Showing 20 items per page