Skip to main content

Home/ Sensorica Knowledge/ Group items tagged intelligence

Rss Feed Group items tagged

Tiberius Brastaviceanu

The Baffler - 0 views

  • This tendency to view questions of freedom primarily through the lens of economic competition, to focus on the producer and the entrepreneur at the expense of everyone else, shaped O’Reilly’s thinking about technology.
  • the O’Reilly brand essence is ultimately a story about the hacker as hero, the kid who is playing with technology because he loves it, but one day falls into a situation where he or she is called on to go forth and change the world,
  • His true hero is the hacker-cum-entrepreneur, someone who overcomes the insurmountable obstacles erected by giant corporations and lazy bureaucrats in order to fulfill the American Dream 2.0: start a company, disrupt an industry, coin a buzzword.
  • ...139 more annotations...
  • gospel of individualism, small government, and market fundamentalism
  • innovation is the new selfishness
  • mastery of public relations
  • making it seem as if the language of economics was, in fact, the only reasonable way to talk about the subject
  • memes are for losers; the real money is in epistemes.
  • “Open source software” was also the first major rebranding exercise overseen by Team O’Reill
  • It’s easy to forget this today, but there was no such idea as open source software before 1998; the concept’s seeming contemporary coherence is the result of clever manipulation and marketing.
  • ideological cleavage between two groups
  • Richard Stallman
  • Free Software Foundation, preoccupied with ensuring that users had rights with respect to their computer programs. Those rights weren’t many—users should be able to run the program for any purpose, to study how it works, to redistribute copies of it, and to release their improved version (if there was one) to the public
  • “free software.”
  • association with “freedom” rather than “free beer”
  • copyleft
  • profound critique of the role that patent law had come to play in stifling innovation and creativity.
  • Plenty of developers contributed to “free software” projects for reasons that had nothing to do with politics. Some, like Linus Torvalds, the Finnish creator of the much-celebrated Linux operating system, did so for fun; some because they wanted to build more convenient software; some because they wanted to learn new and much-demanded skills.
  • Stallman’s rights-talk, however, risked alienating the corporate types
  • he was trying to launch a radical social movement, not a complacent business association
  • By early 1998 several business-minded members of the free software community were ready to split from Stallman, so they masterminded a coup, formed their own advocacy outlet—the Open Source Initiative—and brought in O’Reilly to help them rebrand.
  • “open source”
  • The label “open source” may have been new, but the ideas behind it had been in the air for some time.
  • In those early days, the messaging around open source occasionally bordered on propaganda
  • This budding movement prided itself on not wanting to talk about the ends it was pursuing; except for improving efficiency and decreasing costs, those were left very much undefined.
  • extremely decentralized manner, using Internet platforms, with little central coordination.
  • In contrast to free software, then, open source had no obvious moral component.
  • “open source is not particularly a moral or a legal issue. It’s an engineering issue. I advocate open source, because . . . it leads to better engineering results and better economic results
  • While free software was meant to force developers to lose sleep over ethical dilemmas, open source software was meant to end their insomnia.
  • Stallman the social reformer could wait for decades until his ethical argument for free software prevailed in the public debate
  • O’Reilly the savvy businessman had a much shorter timeline: a quick embrace of open source software by the business community guaranteed steady demand for O’Reilly books and events
  • The coup succeeded. Stallman’s project was marginalized. But O’Reilly and his acolytes didn’t win with better arguments; they won with better PR.
  • A decade after producing a singular vision of the Internet to justify his ideas about the supremacy of the open source paradigm, O’Reilly is close to pulling a similar trick on how we talk about government reform.
  • much of Stallman’s efforts centered on software licenses
  • O’Reilly’s bet wa
  • the “cloud”
  • licenses would cease to matter
  • Since no code changed hands
  • So what did matter about open source? Not “freedom”
  • O’Reilly cared for only one type of freedom: the freedom of developers to distribute software on whatever terms they fancied.
  • the freedom of the producer
  • who must be left to innovate, undisturbed by laws and ethics.
  • The most important freedom,
  • is that which protects “my choice as a creator to give, or not to give, the fruits of my work to you, as a ‘user’ of that work, and for you, as a user, to accept or reject the terms I place on that gift.”
  • O’Reilly opposed this agenda: “I completely support the right of Richard [Stallman] or any individual author to make his or her work available under the terms of the GPL; I balk when they say that others who do not do so are doing something wrong.”
  • The right thing to do, according to O’Reilly, was to leave developers alone.
  • According to this Randian interpretation of open source, the goal of regulation and public advocacy should be to ensure that absolutely nothing—no laws or petty moral considerations—stood in the way of the open source revolution
  • Any move to subject the fruits of developers’ labor to public regulation
  • must be opposed, since it would taint the reputation of open source as technologically and economically superior to proprietary software
  • the advent of the Internet made Stallman’s obsession with licenses obsolete
  • Many developers did stop thinking about licenses, and, having stopped thinking about licenses, they also stopped thinking about broader moral issues that would have remained central to the debates had “open source” not displaced “free software” as the paradigm du jour.
  • Profiting from the term’s ambiguity, O’Reilly and his collaborators likened the “openness” of open source software to the “openness” of the academic enterprise, markets, and free speech.
  • “open to intellectual exchange”
  • “open to competition”
  • “For me, ‘open source’ in the broader sense means any system in which open access to code lowers the barriers to entry into the market”).
  • “Open” allowed O’Reilly to build the largest possible tent for the movement.
  • The language of economics was less alienating than Stallman’s language of ethics; “openness” was the kind of multipurpose term that allowed one to look political while advancing an agenda that had very little to do with politics
  • highlight the competitive advantages of openness.
  • the availability of source code for universal examination soon became the one and only benchmark of openness
  • What the code did was of little importance—the market knows best!—as long as anyone could check it for bugs.
  • The new paradigm was presented as something that went beyond ideology and could attract corporate executives without losing its appeal to the hacker crowd.
  • What Raymond and O’Reilly failed to grasp, or decided to overlook, is that their effort to present open source as non-ideological was underpinned by a powerful ideology of its own—an ideology that worshiped innovation and efficiency at the expense of everything else.
  • What they had in common was disdain for Stallman’s moralizing—barely enough to justify their revolutionary agenda, especially among the hacker crowds who were traditionally suspicious of anyone eager to suck up to the big corporations that aspired to dominate the open source scene.
  • linking this new movement to both the history of the Internet and its future
  • As long as everyone believed that “open source” implied “the Internet” and that “the Internet” implied “open source,” it would be very hard to resist the new paradigm
  • Telling a coherent story about open source required finding some inner logic to the history of the Internet
  • “If you believe me that open source is about Internet-enabled collaboration, rather than just about a particular style of software license,”
  • everything on the Internet was connected to everything else—via open source.
  • The way O’Reilly saw it, many of the key developments of Internet culture were already driven by what he called “open source behavior,” even if such behavior was not codified in licenses.
  • No moralizing (let alone legislation) was needed; the Internet already lived and breathed open source
  • apps might be displacing the browser
  • the openness once taken for granted is no more
  • Openness as a happenstance of market conditions is a very different beast from openness as a guaranteed product of laws.
  • One of the key consequences of linking the Internet to the world of open source was to establish the primacy of the Internet as the new, reinvented desktop
  • This is where the now-forgotten language of “freedom” made a comeback, since it was important to ensure that O’Reilly’s heroic Randian hacker-entrepreneurs were allowed to roam freely.
  • Soon this “freedom to innovate” morphed into “Internet freedom,” so that what we are trying to preserve is the innovative potential of the platform, regardless of the effects on individual users.
  • Lumping everything under the label of “Internet freedom” did have some advantages for those genuinely interested in promoting rights such as freedom of expression
  • Forced to choose between preserving the freedom of the Internet or that of its users, we were supposed to choose the former—because “the Internet” stood for progress and enlightenment.
  • infoware
  • Yahoo
  • their value proposition lay in the information they delivered, not in the software function they executed.
  • The “infoware” buzzword didn’t catch on, so O’Reilly turned to the work of Douglas Engelbart
  • to argue that the Internet could help humanity augment its “collective intelligence” and that, once again, open source software was crucial to this endeavor.
  • Now it was all about Amazon learning from its customers and Google learning from the sites in its index.
  • The idea of the Internet as both a repository and incubator of “collective intelligence”
  • in 2004, O’Reilly and his business partner Dale Dougherty hit on the idea of “Web 2.0.” What did “2.0” mean, exactly?
  • he primary goal was to show that the 2001 market crash did not mean the end of the web and that it was time to put the crash behind us and start learning from those who survived.
  • Tactically, “Web 2.0” could also be much bigger than “open source”; it was the kind of sexy umbrella term that could allow O’Reilly to branch out from boring and highly technical subjects to pulse-quickening futurology
  • O’Reilly couldn’t improve on a concept as sexy as “collective intelligence,” so he kept it as the defining feature of this new phenomenon.
  • What set Web 2.0 apart from Web 1.0, O’Reilly claimed, was the simple fact that those firms that didn’t embrace it went bust
  • find a way to harness collective intelligence and make it part of their business model.
  • By 2007, O’Reilly readily admitted that “Web 2.0 was a pretty crappy name for what’s happening.”
  • O’Reilly eventually stuck a 2.0 label on anything that suited his business plan, running events with titles like “Gov 2.0” and “Where 2.0.” Today, as everyone buys into the 2.0 paradigm, O’Reilly is quietly dropping it
  • assumption that, thanks to the coming of Web 2.0, we are living through unique historical circumstances
  • Take O’Reilly’s musings on “Enterprise 2.0.” What is it, exactly? Well, it’s the same old enterprise—for all we know, it might be making widgets—but now it has learned something from Google and Amazon and found a way to harness “collective intelligence.”
  • tendency to redescribe reality in terms of Internet culture, regardless of how spurious and tenuous the connection might be, is a fine example of what I call “Internet-centrism.”
  • “Open source” gave us the “the Internet,” “the Internet” gave us “Web 2.0,” “Web 2.0” gave us “Enterprise 2.0”: in this version of history, Tim O’Reilly is more important than the European Union
  • For Postman, each human activity—religion, law, marriage, commerce—represents a distinct “semantic environment” with its own tone, purpose, and structure. Stupid talk is relatively harmless; it presents no threat to its semantic environment and doesn’t cross into other ones.
  • Since it mostly consists of falsehoods and opinions
  • it can be easily corrected with facts
  • to say that Tehran is the capital of Iraq is stupid talk
  • Crazy talk, in contrast, challenges a semantic environment, as it “establishes different purposes and assumptions from those we normally accept.” To argue, as some Nazis did, that the German soldiers ended up far more traumatized than their victims is crazy talk.
  • For Postman, one of the main tasks of language is to codify and preserve distinctions among different semantic environments.
  • As he put it, “When language becomes undifferentiated, human situations disintegrate: Science becomes indistinguishable from religion, which becomes indistinguishable from commerce, which becomes indistinguishable from law, and so on.
  • pollution
  • Some words—like “law”—are particularly susceptible to crazy talk, as they mean so many different things: from scientific “laws” to moral “laws” to “laws” of the market to administrative “laws,” the same word captures many different social relations. “Open,” “networks,” and “information” function much like “law” in our own Internet discourse today.
  • For Korzybski, the world has a relational structure that is always in flux; like Heraclitus, who argued that everything flows, Korzybski believed that an object A at time x1 is not the same object as object A at time x2
  • Our language could never properly account for the highly fluid and relational structure of our reality—or as he put it in his most famous aphorism, “the map is not the territory.”
  • Korzybski argued that we relate to our environments through the process of “abstracting,” whereby our neurological limitations always produce an incomplete and very selective summary of the world around us.
  • nothing harmful in this per se—Korzybski simply wanted to make people aware of the highly selective nature of abstracting and give us the tools to detect it in our everyday conversations.
  • Korzybski developed a number of mental tools meant to reveal all the abstracting around us
  • He also encouraged his followers to start using “etc.” at the end of their statements as a way of making them aware of their inherent inability to say everything about a given subject and to promote what he called the “consciousness of abstraction.”
  • There was way too much craziness and bad science in Korzybski’s theories
  • but his basic question
  • “What are the characteristics of language which lead people into making false evaluations of the world around them?”
  • Tim O’Reilly is, perhaps, the most high-profile follower of Korzybski’s theories today.
  • O’Reilly openly acknowledges his debt to Korzybski, listing Science and Sanity among his favorite books
  • It would be a mistake to think that O’Reilly’s linguistic interventions—from “open source” to “Web 2.0”—are random or spontaneous.
  • There is a philosophy to them: a philosophy of knowledge and language inspired by Korzybski. However, O’Reilly deploys Korzybski in much the same way that the advertising industry deploys the latest findings in neuroscience: the goal is not to increase awareness, but to manipulate.
  • O’Reilly, of course, sees his role differently, claiming that all he wants is to make us aware of what earlier commentators may have overlooked. “A metaphor is just that: a way of framing the issues such that people can see something they might otherwise miss,
  • But Korzybski’s point, if fully absorbed, is that a metaphor is primarily a way of framing issues such that we don’t see something we might otherwise see.
  • In public, O’Reilly modestly presents himself as someone who just happens to excel at detecting the “faint signals” of emerging trends. He does so by monitoring a group of überinnovators that he dubs the “alpha geeks.” “The ‘alpha geeks’ show us where technology wants to go. Smart companies follow and support their ingenuity rather than trying to suppress it,
  • His own function is that of an intermediary—someone who ensures that the alpha geeks are heard by the right executives: “The alpha geeks are often a few years ahead of their time. . . . What we do at O’Reilly is watch these folks, learn from them, and try to spread the word by writing down (
  • The name of his company’s blog—O’Reilly Radar—is meant to position him as an independent intellectual who is simply ahead of his peers in grasping the obvious.
  • “the skill of writing is to create a context in which other people can think”
  • As Web 2.0 becomes central to everything, O’Reilly—the world’s biggest exporter of crazy talk—is on a mission to provide the appropriate “context” to every field.
  • In a fascinating essay published in 2000, O’Reilly sheds some light on his modus operandi.
  • The thinker who emerges there is very much at odds with the spirit of objectivity that O’Reilly seeks to cultivate in public
  • meme-engineering lets us organize and shape ideas so that they can be transmitted more effectively, and have the desired effect once they are transmitted
  • O’Reilly meme-engineers a nice euphemism—“meme-engineering”—to describe what has previously been known as “propaganda.”
  • how one can meme-engineer a new meaning for “peer-to-peer” technologies—traditionally associated with piracy—and make them appear friendly and not at all threatening to the entertainment industry.
  • O’Reilly and his acolytes “changed the canonical list of projects that we wanted to hold up as exemplars of the movement,” while also articulating what broader goals the projects on the new list served. He then proceeds to rehash the already familiar narrative: O’Reilly put the Internet at the center of everything, linking some “free software” projects like Apache or Perl to successful Internet start-ups and services. As a result, the movement’s goal was no longer to produce a completely free, independent, and fully functional operating system but to worship at the altar of the Internet gods.
  • Could it be that O’Reilly is right in claiming that “open source” has a history that predates 1998?
  • Seen through the prism of meme-engineering, O’Reilly’s activities look far more sinister.
  • His “correspondents” at O’Reilly Radar don’t work beats; they work memes and epistemes, constantly reframing important public issues in accordance with the templates prophesied by O’Reilly.
  • Or take O’Reilly’s meme-engineering efforts around cyberwarfare.
  • Now, who stands to benefit from “cyberwarfare” being defined more broadly? Could it be those who, like O’Reilly, can’t currently grab a share of the giant pie that is cybersecurity funding?
  • Frank Luntz lists ten rules of effective communication: simplicity, brevity, credibility, consistency, novelty, sound, aspiration, visualization, questioning, and context.
  • Thus, O’Reilly’s meme-engineering efforts usually result in “meme maps,” where the meme to be defined—whether it’s “open source” or “Web 2.0”—is put at the center, while other blob-like terms are drawn as connected to it.
  • The exact nature of these connections is rarely explained in full, but this is all for the better, as the reader might eventually interpret connections with their own agendas in mind. This is why the name of the meme must be as inclusive as possible: you never know who your eventual allies might be. “A big part of meme engineering is giving a name that creates a big tent that a lot of people want to be under, a train that takes a lot of people where they want to go,”
  • News April 4 mail date March 29, 2013 Baffler party March 6, 2013 Žižek on seduction February 13, 2013 More Recent Press I’ve Seen the Worst Memes of My Generation Destroyed by Madness io9, April 02, 2013 The Baffler’s New Colors Imprint, March 21, 2013
  • There is considerable continuity across O’Reilly’s memes—over time, they tend to morph into one another.
Tiberius Brastaviceanu

Access control - Wikipedia, the free encyclopedia - 0 views

  • The act of accessing may mean consuming, entering, or using.
  • Permission to access a resource is called authorization.
  • Locks and login credentials are two analogous mechanisms of access control.
  • ...26 more annotations...
  • Geographical access control may be enforced by personnel (e.g., border guard, bouncer, ticket checker)
  • n alternative of access control in the strict sense (physically controlling access itself) is a system of checking authorized presence, see e.g. Ticket controller (transportation). A variant is exit control, e.g. of a shop (checkout) or a country
  • access control refers to the practice of restricting entrance to a property, a building, or a room to authorized persons
  • can be achieved by a human (a guard, bouncer, or receptionist), through mechanical means such as locks and keys, or through technological means such as access control systems like the mantrap.
  • Physical access control is a matter of who, where, and when
  • Historically, this was partially accomplished through keys and locks. When a door is locked, only someone with a key can enter through the door, depending on how the lock is configured. Mechanical locks and keys do not allow restriction of the key holder to specific times or dates. Mechanical locks and keys do not provide records of the key used on any specific door, and the keys can be easily copied or transferred to an unauthorized person. When a mechanical key is lost or the key holder is no longer authorized to use the protected area, the locks must be re-keyed.[citation needed] Electronic access control uses computers to solve the limitations of mechanical locks and keys. A wide range of credentials can be used to replace mechanical keys. The electronic access control system grants access based on the credential presented. When access is granted, the door is unlocked for a predetermined time and the transaction is recorded. When access is refused, the door remains locked and the attempted access is recorded. The system will also monitor the door and alarm if the door is forced open or held open too long after being unlocked
  • When a credential is presented to a reader, the reader sends the credential’s information, usually a number, to a control panel, a highly reliable processor. The control panel compares the credential's number to an access control list, grants or denies the presented request, and sends a transaction log to a database. When access is denied based on the access control list, the door remains locked.
  • Access control system operation
  • The above description illustrates a single factor transaction. Credentials can be passed around, thus subverting the access control list. For example, Alice has access rights to the server room, but Bob does not. Alice either gives Bob her credential, or Bob takes it; he now has access to the server room. To prevent this, two-factor authentication can be used. In a two factor transaction, the presented credential and a second factor are needed for access to be granted; another factor can be a PIN, a second credential, operator intervention, or a biometric input
  • There are three types (factors) of authenticating information:[2] something the user knows, e.g. a password, pass-phrase or PIN something the user has, such as smart card or a key fob something the user is, such as fingerprint, verified by biometric measurement
  • Passwords are a common means of verifying a user's identity before access is given to information systems. In addition, a fourth factor of authentication is now recognized: someone you know, whereby another person who knows you can provide a human element of authentication in situations where systems have been set up to allow for such scenarios
  • Credential
  • A credential is a physical/tangible object, a piece of knowledge, or a facet of a person's physical being, that enables an individual access to a given physical facility or computer-based information system. Typically, credentials can be something a person knows (such as a number or PIN), something they have (such as an access badge), something they are (such as a biometric feature) or some combination of these items. This is known as multi-factor authentication. The typical credential is an access card or key-fob, and newer software can also turn users' smartphones into access devices.
  • An access control point, which can be a door, turnstile, parking gate, elevator, or other physical barrier, where granting access can be electronically controlled. Typically, the access point is a door. An electronic access control door can contain several elements. At its most basic, there is a stand-alone electric lock. The lock is unlocked by an operator with a switch. To automate this, operator intervention is replaced by a reader. The reader could be a keypad where a code is entered, it could be a card reader, or it could be a biometric reader. Readers do not usually make an access decision, but send a card number to an access control panel that verifies the number against an access list
  • monitor the door position
  • Generally only entry is controlled, and exit is uncontrolled. In cases where exit is also controlled, a second reader is used on the opposite side of the door. In cases where exit is not controlled, free exit, a device called a request-to-exit (REX) is used. Request-to-exit devices can be a push-button or a motion detector. When the button is pushed, or the motion detector detects motion at the door, the door alarm is temporarily ignored while the door is opened. Exiting a door without having to electrically unlock the door is called mechanical free egress. This is an important safety feature. In cases where the lock must be electrically unlocked on exit, the request-to-exit device also unlocks the doo
  • Access control topology
  • Access control decisions are made by comparing the credential to an access control list. This look-up can be done by a host or server, by an access control panel, or by a reader. The development of access control systems has seen a steady push of the look-up out from a central host to the edge of the system, or the reader. The predominant topology circa 2009 is hub and spoke with a control panel as the hub, and the readers as the spokes. The look-up and control functions are by the control panel. The spokes communicate through a serial connection; usually RS-485. Some manufactures are pushing the decision making to the edge by placing a controller at the door. The controllers are IP enabled, and connect to a host and database using standard networks
  • Access control readers may be classified by the functions they are able to perform
  • and forward it to a control panel.
  • Basic (non-intelligent) readers: simply read
  • Semi-intelligent readers: have all inputs and outputs necessary to control door hardware (lock, door contact, exit button), but do not make any access decisions. When a user presents a card or enters a PIN, the reader sends information to the main controller, and waits for its response. If the connection to the main controller is interrupted, such readers stop working, or function in a degraded mode. Usually semi-intelligent readers are connected to a control panel via an RS-485 bus.
  • Intelligent readers: have all inputs and outputs necessary to control door hardware; they also have memory and processing power necessary to make access decisions independently. Like semi-intelligent readers, they are connected to a control panel via an RS-485 bus. The control panel sends configuration updates, and retrieves events from the readers.
  • Systems with IP readers usually do not have traditional control panels, and readers communicate directly to a PC that acts as a host
  • a built in webservice to make it user friendly
  • Some readers may have additional features such as an LCD and function buttons for data collection purposes (i.e. clock-in/clock-out events for attendance reports), camera/speaker/microphone for intercom, and smart card read/write support
Kurt Laitner

Managing collective intelligence - Toward a New Corporate Governance - 0 views

  •  
    via jean lievens I think
Guillaume Barreau

Des drones intelligents au secours de l'agriculture biologique - 0 views

  •  
    Des drones intelligents au secours de l'agriculture biologique Quand les drones ne sont pas employés pour survoler des centrales nucléaires ou pour livrer des colis d'une quelconque multinationale, ils peuvent parfois trouver une utilité dans le milieu alternatif. Michael Godfrey, un jeune étudiant en agronomie de l'université du Queensland (Australie), propose une application peu commune pour ces engins : le lâcher d'insectes.
Steve Bosserman

When Cities Run Themselves | WOUB - 0 views

  •  
    Machines talking to machines No doubt that the Olympics will have a profound effect in shaping London's future. By the time the Games begin, for instance, it will have Europe's largest free WiFi zone, with the city's iconic red phone booths converted, fittingly, into hotspots. But another opportunity London landed earlier this month could have just as much impact, perhaps more. A company called Living PlanIt announced that it will begin testing its "Urban Operating System" in the Greenwich section of the city. What does that mean? Put simply, London would have its own operating system, much as your PC runs on Windows or your Mac runs on Apple's IOS. This ties into the latest hot buzz phrase, "the internet of things," which describes a world where machines talk to other machines. No human interaction required. So, for a city, this means sensors in buildings would connect to sensors in water treatment plants which would connect to sensors in stoplights. It would be one gigantic, computerized urban nervous system, which a lot of experts think is the only way cities can survive a future when they'll contain more than two out of every three people on Earth. Based on what sensors reveal about the location and movement of humans in a section of a city, for instance, buildings will automatically adjust their temperatures, streetlights will dim or brighten, water flow will increase or slow. Or, in the event of a disaster, emergency services would have real-time access to traffic data, trauma unit availability, building blueprints. And soon enough, our smart phones will be able to tap in to the Urban OS. So will our household appliances. This is not some 21st century analogue of the personal jet pack. The Urban OS is the driving force behind a smart city being built from the ground up in northern Portugal. Construction is scheduled to be completed in three years; eventually it will have about 150,000 residents. It will also have more than 100 million sen
Tiberius Brastaviceanu

Action (Stanford Encyclopedia of Philosophy) - 0 views

  • intentionally
  • questions about the nature, variety, and identity of action
  • Should we think of the consequences, conventional or causal, of physical behavior as constituents of an action distinct from but ‘generated by’ the movement? Or should we think that there is a single action describable in a host of ways?
  • ...22 more annotations...
  • Donald Davidson
  • an action
  • is something an agent does that was ‘intentional under some description,’
  • there have been many attempts to map the relations between intentions for the future, acting intentionally, and acting with a certain intention.
  • There has been a notable or notorious debate about whether the agent's reasons in acting are causes of the action
  • rendered the action intelligible in his eyes
  • things that merely happen
  • things they genuinely do
  • distinction between
  • the doings, are the acts or actions of the agent
  • what distinguishes an action from a mere happening or occurrence?
  • An agent performs activity that is directed at a goal
  • adopted on the basis of an overall practical assessment of his options and opportunities
  • awareness
  • that he is performing the activity
  • and that the activity is aimed by him at such-and-such a chosen end
  • It is frequently noted that the agent has some sort of immediate awareness of his physical activity and of the goals that the activity is aimed at realizing.
  • ‘knowledge without observation.’
  • It is also important to the concept of ‘goal directed action’ that agents normally implement a kind of direct control or guidance over their own behavior.
  • For many years, the most intensely debated topic in the philosophy of action concerned the explanation of intentional actions in terms of the agent's reasons for acting
  • Davidson and other action theorists defended the position that reason explanations are causal explanations
  • In the foregoing, reference has been made to explanations of actions in terms of reasons, but recent work on agency has questioned whether contemporary frameworks for the philosophy of action have really articulated the way in which an agent's desires and other pro-attitudes have the distinctive force of reasons in the setting of these ordinary explanations
Kurt Laitner

Siri's Inventors Are Building a Radical New AI That Does Anything You Ask | Enterprise ... - 0 views

  •  
    marry this with sensors...
Tiberius Brastaviceanu

How Peer to Peer Communities will change the World - 0 views

  • role of p2p movement
  • historical role
  • horizontalisation of human relationships
  • ...55 more annotations...
  • allowing the free aggregation of individuals around shared values or common value creation
  • a huge sociological shift
  • new life forms, social practices and human institutions
  • emergent communities of practice are developing new social practices that are informed by the p2p paradigm
  • ethical revolution
  • openness
  • participation
  • inclusivity
  • cooperation
  • commons
  • the open content industry in the U.S. to reach one sixth of GDP.
  • political expressions
  • the movement has two wings
  • constructive
  • building new tools and practices
  • resistance to neoliberalism
  • we are at a stage of emergence
  • difficulty of implementing full p2p solutions in the current dominant system
  • At this stage, there is a co-dependency between peer producers creating value, and for-profit firms ‘capturing that value’, but they both need each other.
  • Peer producers need a business ecology to insure the social reproduction of their system and financial sustainability of its participants, and capital needs the positive externalities of social cooperation which flow from p2p collaboration.
  • peer producing communities should create their own ‘mission-oriented’ social businesses, so that the surplus value remains with the value creators, i.e. the commoners themselves, but this is hardly happening now.
  • Instead what we see is a mutual accomodation between netarchical capital on one side, and peer production communities on the other.
  • the horizontal meets the vertical
  • mostly hybrid ‘diagonal’ adaptations
  • For peer producers the question becomes, if we cannot create our own fully autonomous institutions, how can we adapt while maintaining maximum autonomy and sustainability as a commons and as a community.
  • Why p2p have failed to create successful alternatives in some areas?
  • In commons-oriented peer production, where people aggegrate around a common object which requires deep cooperation, they usually have their own infrastructures of cooperation and a ecology combining community, a for-benefit association managing the infrastructure, and for-profit companies operating on the market place; in the sharing economy, where individuals merely share their own expressions, third party platforms are the norm. It is clear that for-profit companies have different priorities, and want to enclose value so that it can be sold on the marketplace. This in fact the class struggle of the p2p era, the struggle between communities and corporations around various issues because of partly differential interests.
  • Even commercially controlled platforms are being used for a massive horizontalisation and self-aggregation of human relationships, and communities, including political and radical groups are effectively using them to mobilize. What’s important is not just to focus on the limitations and intentions of the platform owners, but to use whatever we can to strengthen the autonomy of peer communities.
  • requires a clever adaptation
  • use for our own benefit
  • The fact today is that capital is still capable of marshaling vast financial and material resources, so that it can create,
  • platforms that can easily and quickly offer services, creating network effects
  • without network effects, there is no ‘there’ there, just an empty potential platform.
  • p2p activists should work on both fronts
  • using mainstream platforms for spreading their ideas and culture and reach greater numbers of people, while also developing their own autonomous media ecologies, that can operate independently, and the latter is an engagement for the ‘long haul’, i.e. the slow construction of an alternative lifeworld.
  • The commons and p2p are really just different aspects of the same phenomena; the commons is the object that p2p dynamics are building; and p2p takes place wherever there are commons.
  • So both p2p and the commons, as they create abundant (digital) or sufficient (material) value for the commoners, at the same time create opportunities to create added value for the marketplace. There is no domain that is excluded from p2p, no field that can say, “we wouldn’t be stronger by opening up to participation and community dynamics”. And there is no p2p community that can say, we are in the long term fully sustainable within the present system, without extra resources coming from the market sector.
  • One trend is the distribution of current infrastructures and practices, i.e. introducing crowdsourcing, crowdfunding, social lending, digital currencies, in order to achieve wider participation in current practices. That is a good thing, but not sufficient. All the things that I mention above, move to a distributed infrastructure, but do not change the fundamental logic of what they are doing.
  • we are talking about the distribution of capitalism, not about a deeper change in the logic of our economy.
  • No matter how good you are, no matter how much capital you have to hire the best people, you cannot compete with the innovative potential of open global communities.
  • the p2p dynamics
  • the new networked culture
  • the opposite is also happening, as we outlined above, more and more commons-oriented value communities are creating their own entrepreneurial coalitions. Of course, some type of companies, because of their monopoly positions and legacy systems, may have a very difficult time undergoing that adaptation, in which case new players will appear that can do it more effectively.
  • the corporate form is unable to deal with ecological and sustainability issues, because its very DNA, the legal obligation to enrich the shareholders, makes its strive to lower input costs,  and ignore externalities.
  • we need new corporate structures, a new type of market entity, for which profit is a means, but not an end, dedicated to a ‘benefit‘, a ‘mission’, or the sustenance of a particular community and/or commons.
  • abundance destroys scarcity and therefore markets
  • open design community
  • will inherently design for sustainability
  • for inclusion
  • conceive more distributed forms of manufacturing
  • entrepreneurs attaching themselves to open design projects start working from an entirely different space, even if they still use the classic corporate form. Prevent the sharing of sustainability designs through IP monopolies is also in my view unethical and allowing such patents should be a minimalist option, not a maximalist one.
  • The high road scenario proposes an enlightened government that ‘enables and empowers’ social production and value creation and allows a much smoother transition to p2p models; the low road scenario is one in which no structural reforms take place, the global situation descends into various forms of chaos, and p2p becomes a survival and resilience tactic in extremely difficult social, political and economic circumstances.
  • accelerated end of capitalism
  • Making sure that we get a better alternative is actually the historical task of the p2p movement. In other words, it depends on us!
  • I don’t really think in terms of technological breakthroughs, because the essential one, globally networked collective intelligence enabled by the internetworks, is already behind us; that is the major change, all other technological breakthroughs will be informed by this new social reality of the horizontalisation of our civilisation. The important thing now is to defend and extend our communication and organisation rights, against a concerted attempt to turn back the clock. While the latter is really an impossibility, this does not mean that the attempts by governments and large corporations cannot create great harm and difficulties. We need p2p technology to enable the global solution finding and implementation of the systemic crises we are facing.
Tiberius Brastaviceanu

Beyond Blockchain: Simple Scalable Cryptocurrencies - The World of Deep Wealth - Medium - 0 views

  • I clarify the core elements of cryptocurrency and outline a different approach to designing such currencies rooted in biomimicry
  • This post outlines a completely different strategy for implementing cryptocurrencies with completely distributed chains
  • Rather than trying to make one global, anonymous, digital cash
  • ...95 more annotations...
  • we are interested in the resilience that comes from building a rich ecosystem of interoperable currencies
  • What are the core elements of a modern cryptocurrency?
  • Digital
  • Holdings are electronic and only exist and operate by virtue of a community’s agreement about how to interpret digital bits according to rules about operation and accounting of the currency.
  • Trustless
  • don’t have to trust a 3rd party central authority
  • Decentralized
  • Specifically, access, issuance, transaction accounting, rules & policies, should be collectively visible, known, and held.
  • Cryptographic
  • This cryptographic structure is used to enable a variety of people to host the data without being able to alter it.
  • Identity
  • there must be a way to associate these bits with some kind of account, wallet, owner, or agent who can use them
  • Other things that many take for granted in blockchains may not be core but subject to decisions in design and implementation, so they can vary between implementations
  • It does not have to be stored in a synchronized global ledger
  • does not have to be money. It may be a reputation currency, or data used for identity, or naming, etc
  • Its units do not have to be cryptographic tokens or coins
  • It does not have to protect the anonymity of users, although it may
  • if you think currency is only money, and that money must be artificially scarce
  • Then you must tackle the problem of always tracking which coins exist, and which have been spent. That is one approach — the one blockchain takes.
  • You might optimize for anonymity if you think of cryptocurrency as a tool to escape governments, regulations, and taxes.
  • if you want to establish and manage membership in new kinds of commons, then identity and accountability for actions may turn out to be necessary ingredients instead of anonymity.
  • In the case of the MetaCurrency Project, we are trying to support many use cases by building tools to enable a rich ecosystem of communities and current-sees (many are non-monetary) to enhance collective intelligence at all scales.
  • Managing consensus about a shared reality is a central challenge at the heart of all distributed computing solutions.
  • If we want to democratize money by having cryptocurrencies become a significant and viable means of transacting on a daily basis, I believe we need fundamentally more scalable approaches that don’t require expensive, dedicated hardware just to participate.
  • We should not need system wide consensus for two people to do a transaction in a cryptocurrency
  • Blockchain is about managing a consensus about what was “said.” Ceptr is about distributing a consensus about how to “speak.”
  • how nature gets the job done in massively scalable systems which require coordination and consistency
  • Replicate the same processes across all nodes
  • Empower every node with full agency
  • Hold this transformed state locally and reliably
  • Establish protocols for interaction
  • Each speaker of a language carries the processes to understand sentences they hear, and generate sentences they need
  • we certainly don’t carry some kind of global ledger of everything that’s ever been said, or require consensus about what has been said
  • Language IS a communication protocol we learn by emulating the processes of usage.
  • Dictionaries try to catch up when the usage
  • there is certainly no global ledger with consensus about the state of trillions of cells. Yet, from a single zygote’s copy of DNA, our cells coordinate in a highly decentralized manner, on scales of trillions, and without the latency or bottlenecks of central control.
  • Imagine something along the lines of a Java Virtual Machine connected to a distributed version of Github
  • Every time this JVM runs a program it confirms the hash of the code it is about to execute with the hash signed into the code repository by its developers
  • This allows each node that intends to be honest to be sure that they’re running the same processes as everyone else. So when two parties want to do a transaction, and each can have confidence their own code, and the results that your code produces
  • Then you treat it as authoritative and commit it to your local cryptographically self-validating data store
  • Allowing each node to treat itself as a full authority to process transactions (or interactions via shared protocols) is exactly how you empower each node with full agency. Each node runs its copy of the signed program/processes on its own virtual machine, taking the transaction request combined with the transaction chains of the parties to the transaction. Each node can confirm their counterparty’s integrity by replaying their transactions to produce their current state, while confirming signatures and integrity of the chain
  • If both nodes are in an appropriate state which allows the current transaction, then they countersign the transaction and append to their respective chains. When you encounter a corrupted or dishonest node (as evidenced by a breach of integrity of their chain — passing through an invalid state, broken signatures, or broken links), your node can reject the transaction you were starting to process. Countersigning allows consensus at the appropriate scale of the decision (two people transacting in this case) to lock data into a tamper-proof state so it can be stored in as many parallel chains as you need.
  • When your node appends a mutually validated and signed transaction to its chain, it has updated its local state and is able to represent the integrity of its data locally. As long as each transaction (link in the chain) has valid linkages and countersignatures, we can know that it hasn’t been tampered with.
  • If you can reliably embody the state of the node in the node itself using Intrinsic Data Integrity, then all nodes can interact in parallel, independent of other interactions to maximize scalability and simultaneous processing. Either the node has the credits or it doesn’t. I don’t have to refer to a global ledger to find out, the state of the node is in the countersigned, tamper-proof chain.
  • Just like any meaningful communication, a protocol needs to be established to make sure that a transaction carries all the information needed for each node to run the processes and produce a new signed and chained state. This could be debits or credits to an account which modify the balance, or recoding courses and grades to a transcript which modify a Grade Point Average, or ratings and feedback contributing to a reputation score, and so on.
  • By distributing process at the foundation, and leveraging Intrinsic Data Integrity, our approach results in massive improvements in throughput (from parallel simultaneous independent processing), speed, latency, efficiency, and cost of hardware.
  • You also don’t need to incent people to hold their own record — they already want it.
  • Another noteworthy observation about humans, cells, and atoms, is that each has a general “container” that gets configured to a specific use.
  • Likewise, the Receptors we’ve built are a general purpose framework which can load code for different distributed applications. These Receptors are a lightweight processing container for the Ceptr Virtual Machine Host
  • Ceptr enables a developer to focus on the rules and transactions for their use case instead of building a whole framework for distributed applications.
  • how units in a currency are issued
  • Most people think that money is just money, but there are literally hundreds of decisions you can make in designing a currency to target particular needs, niches, communities or patterns of flow.
  • Blockchain cryptocurrencies are fiat currencies. They create tokens or coins from nothing
  • These coins are just “spoken into being”
  • the challenging task of
  • ensure there is no counterfeiting or double-spending
  • Blockchain cryptocurrencies are fiat currencies
  • These coins are just “spoken into being”
  • the challenging task of tracking all the coins that exist to ensure there is no counterfeiting or double-spending
  • You wouldn’t need to manage consensus about whether a cryptocoin is spent, if your system created accounts which have normal balances based on summing their transactions.
  • In a mutual credit system, units of currency are issued when a participant extends credit to another user in a standard spending transaction
  • Alice pays Bob 20 credits for a haircut. Alice’s account now has -20, and Bob’s has +20.
  • Alice spent credits she didn’t have! True
  • Managing the currency supply in a mutual credit system is about managing credit limits — how far people can spend into a negative balance
  • Notice the net number units in the system remains zero
  • One elegant approach to managing mutual credit limits is to set them based on actual demand.
  • concerns about manufacturing fake accounts to game credit limits (Sybil Attacks)
  • keep in mind there can be different classes of accounts. Easy to create, anonymous accounts may get NO credit limit
  • What if I alter my code to give myself an unlimited credit limit, then spend as much as I want? As soon as you pass the credit limit encoded in the shared agreements, the next person you transact with will discover you’re in an invalid state and refuse the transaction.
  • If two people collude to commit an illegal transaction by both hacking their code to allow a normally invalid state, the same still pattern still holds. The next person they try to transact with using untampered code will detect the problem and decline to transact.
  • Most modern community currency systems have been implemented as mutual credit,
  • Hawala is a network of merchants and businessmen, which has been operating since the middle ages, performing money transfers on an honor system and typically settling balances through merchandise instead of transferring money
  • Let’s look at building a minimum viable cryptocurrency with the hawala network as our use case
  • To minimize key management infrastructure, each hawaladar’s public key is their address or identity on the network. To join the network you get a copy of the software from another hawaladar, generate your public and private keys, and complete your personal profile (name, location, contact info, etc.). You call, fax, or email at least 10 hawaladars who know you, and give them your IP address and ask them to vouch for you.
  • Once 10 other hawaladars have vouched for you, you can start doing other transactions because the protocol encoded in every node will reject a transaction chain that doesn’t start with at least 10 vouches
  • seeding your information with those other peers so you can be found by the rest of the network.
  • As described in the Mutual Credit section, at the time of transaction each party audits the counterparty’s transaction chain.
  • Our hawala crypto-clearinghouse protocol has two categories of transactions: some used for accounting and others for routing. Accounting transactions change balances. Routing transactions maintain network integrity by recording information about hawaladar
  • Accounting Transactions create signed data that changes account balances and contains these fields:
  • The final hash of all of the above fields is used as a unique transaction ID and is what each of party signs with their private keys. Signing indicates a party has agreed to the terms of the transaction. Only transactions signed by both parties are considered valid. Nodes can verify signatures by confirming that decryption of the signature using the public key yields a result which matches the transaction ID.
  • Routing Transactions sign data that changes the peers list and contain these fields:
  • As with accounting transactions, the hash of the above fields is used as the transaction’s unique key and the basis for the cryptographic signature of both counterparties.
  • Remember, instead of making changes to account balances, routing transactions change a node’s local list of peers for finding each other and processing.
  • a distributed network of mutual trust
  • operates across national boundaries
  • everyone already keeps and trusts their own separate records
  • Hawaladars are not anonymous
  • “double-spending”
  • It would be possible for someone to hack the code on their node to “forget” their most recent transaction (drop the head of their chain), and go back to their previous version of the chain before that transaction. Then they could append a new transaction, drop it, and append again.
  • After both parties have signed the agreed upon transaction, each party submits the transaction to separate notaries. Notaries are a special class of participant who validate transactions (auditing each chain, ensuring nobody passes through an invalid state), and then they sign an outer envelope which includes the signatures of the two parties. Notaries agree to run high-availability servers which collectively manage a Distributed Hash Table (DHT) servicing requests for transaction information. As their incentive for providing this infrastructure, notaries get a small transaction fee.
  • This approach introduces a few more steps and delays to the transaction process, but because it operates on independent parallel chains, it is still orders of magnitude more efficient and decentralized than reaching consensus on entries in a global ledger
  • millions of simultaneous transactions could be getting processed by other parties and notaries with no bottlenecks.
  • There are other solutions to prevent nodes from dropping the head of their transaction chain, but the approach of having notaries serve out a DHT solves a number of common objections to completely distributed accounting. Having access to reliable lookups in a DHT provides a similar big picture view that you get from a global ledger. For example, you may want a way to look up transactions even when the parties to that transaction are offline, or to be able to see the net system balance at a particular moment in time, or identify patterns of activity in the larger system without having to collect data from everyone individually.
  • By leveraging Intrinsic Data Integrity to run numerous parallel tamper-proof chains you can enable nodes to do various P2P transactions which don’t actually require group consensus. Mutual credit is a great way to implement cryptocurrencies to run in this peered manner. Basic PKI with a DHT is enough additional infrastructure to address main vulnerabilities. You can optimize your solution architecture by reserving reserve consensus work for tasks which need to guarantee uniqueness or actually involve large scale agreement by humans or automated contracts.
  • It is not only possible, but far more scalable to build cryptocurrencies without a global ledger consensus approach or cryptographic tokens.
  •  
    Article written by Arthur Brook, founder of Metacurrency project and of Ceptr.
Tiberius Brastaviceanu

Partner State - P2P Foundation - 0 views

    • Tiberius Brastaviceanu
       
      we call this a custodian
    • Tiberius Brastaviceanu
       
      we call this a custodian
  • So here we have it, the new triarchy: - The state, with its public property and representative mechanisms of governance (in the best scenario) - The private sector, with the corporation and private property - The commons, with the Trust (or the for-benefit association), and which is the ‘property’ of all its members (not the right word in the context of the commons, since it has a different philosophy of ownership)
    • Tiberius Brastaviceanu
       
      so where is direct democracy in all this?
  • ...39 more annotations...
  • In a first phase, the commons simply emerges as an added alternative.
  • becoming a subsector of society, and starts influencing the whole
  • phase transition and transformation will need to occur.
  • how a commons-dominated, i.e. after the phase transition, society would look like.
  • At its core would be a collection of commons, represented by trusts and for-benefit associations, which protect their common assets for the benefit of present and future generations
  • The commons ‘rents out’ the use of its resources to entrepreneurs. In other words, business still exists, though infinite growth-based capitalism does not.
  • More likely is that the corporate forms will be influenced by the commons and that profit will be subsumed to other goals, that are congruent with the maintenance of the commons.
  • The state will still exist, but will have a radically different nature
  • Much of its functions will have been taken over by commons institutions, but since these institutions care primarily about their commons, and not the general common good, we will still need public authorities that are the guarantor of the system as a whole, and can regulate the various commons, and protect the commoners against possible abuses. So in our scenario, the state does not disappear, but is transformed, though it may greatly diminish in scope, and with its remaining functions thoroughly democratized and based on citizen participation.
  • In our vision, it is civil-society based peer production, through the Commons, which is the guarantor of value creation by the private sector, and the role of the state, as Partner State, is to enable and empower the creation of common value. The new peer to peer state then, though some may see that as a contradictio in terminis, is a state which is subsumed under the Commons, just as it is now under the private sector. Such a peer to peer state, if we are correct, will have a much more modest role than the state under a classic state society, with many of its functions taken over by civil society associations, interlinked in processes of global governance. The above then, this triarchy, is the institutional core which replaces the dual private-public binary system that is characteristic of the capitalist system that is presently the dominant format.
  • fundamental mission is to empower direct social-value creation, and to focus on the protection of the Commons sphere as well as on the promotion of sustainable models of entrepreneurship and participatory politics
  • the state becomes a 'partner state' and enables autonomous social production.
  • the state does exist, and I believe that we can’t just imagine that we live in a future state-less society
  • retreating from the binary state/privatization dilemma to the triarchical choice of an optimal mix amongst government regulation, private-market freedom and autonomous civil-society projects
  • the role of the state
  • “the peer production of common value requires civic wealth and strong civic institutions.
  • trigger the production/construction of new commons by - (co-) management of complexe resource systems which are not limited to local boundaries or specific communities (as manager and partner) - survey of rules (chartas) to care for the commons (mediator or judge) - kicking of or providing incentives for commoners governing their commons - here the point is to design intelligent rules which automatically protect the commons, like the GPL does (facilitator)"
  • the emergence of the digital commons. It is the experience of creating knowledge, culture, software and design commons, by a combination of voluntary contributions, entrepreneurial coalitions and infrastructure-protecting for-benefit associations, that has most tangibly re-introduced the idea of commons, for all to use without discrimination, and where all can contribute. It has drastically reduced the production, distribution, transaction and coordination costs for the immaterial value that is at the core also of all what we produce physically, since that needs to be made, needs to be designed. It has re-introduced communing as a mainstream experience for at least one billion internet users, and has come with proven benefits and robustness that has outcompeted and outcooperated its private rivals. It also of course offers new ways to re-imagine, create and protect physical commons.
  • stop enclosures
  • peer to peer, i.e. the ability to freely associate with others around the creation of common value
  • communal shareholding, i.e. the non-reciprocal exchange of an individual with a totality. It is totality that we call the commons.
  • It is customary to divide society into three sectors, and what we want to show is how the new peer to peer dynamic unleashed by networked infrastructures, changes the inter-relationship between these three sectors.
  • In the current ‘cognitive capitalist’ system, it is the private sector consisting of enterprises and businesses which is the primary factor, and it is engaged in competitive capital accumulation. The state is entrusted with the protection of this process. Though civil society, through the citizen, is in theory ‘sovereign’, and chooses the state; in practice, both civil society and the state are under the domination of the private sector.
  • it fulfills three contradictory functions
  • Of course, this is not to say that the state is a mere tool of private business.
  • protect the whole system, under the domination of private business
  • protector of civil society, depending on the balance of power and achievements of social movements
  • protector of its own independent interests
  • Under fascism, the state achieves great independence from the private sector , which may become subservient to the state. Under the welfare state, the state becomes a protector of the social balance of power and manages the achievements of the social movement; and finally, under the neoliberal corporate welfare state, or ‘market state’, it serves most directly the interests of the financial sector.
  • key institutions and forms of property.
  • The state managed a public sector, under its own property.
  • The private sector , under a regime of private ownership, is geared to profit, discounts social and natural externalities, both positive and negative, and uses its dominance in society to use and dominate the state.
  • civil society has a relative power as well, through its capability of creating social movements and associations
  • Capitalism has historically been a pendulum between the private and the public sector
  • However, this configuration is changing,
  • the endangerment of the biosphere through the workings of ‘selfish’ market players; the second is the role of the new digital commons.
  • participatory politics
  • Peer production gives us an advance picture of how a commons-oriented society would look like. At its core is a commons and a community contributing to it, either voluntarily, or as paid entrepreneurial employees. It does this through collaborative platforms using open standards. Around the commons emerges enterprises that create added value to operate on the marketplace, but also help the maintenance and the expansion of the commons they rely on. A third partner are the for-benefit associations that maintain the infrastructure of cooperation. Public authorities could play a role if they wanted to support existing commons or the creation of new commons, for the value they bring to society.
  • if a commons is not created as in the case of the digital commons, it is something that is inherited from nature or former generations, given in trust and usufruct, so that it can be transmitted to our descendents. The proper institution for such commons is therefore the trust, which is a corporate form that cannot touch its principal capital, but has to maintain it.
Tiberius Brastaviceanu

SENSORICA - 0 views

  •  
    Sensorica is an Open-Value Network (OVN) offering a stigmergic environment for synergistic open innovation, leveraging collective intelligence. Ex: Strategic Alliance Agreement
1 - 19 of 19
Showing 20 items per page