Skip to main content

Home/ Computer Science Knowledge Sharing/ Group items tagged AI

Rss Feed Group items tagged

Janos Haits

YC AI - 0 views

  •  
    "Our long-term goal is to democratize AI. We want to level the playing field for startups to ensure that innovation doesn't get locked up in large companies like Google or Facebook. If you're starting an AI company, we want to help you succeed. Apply here and mention this post in your application."
Abdelrahman Ogail

Simulated annealing - Wikipedia, the free encyclopedia - 1 views

  • Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration — provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima—which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  • Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration — provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima—which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  • Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration — provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima—which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  • ...1 more annotation...
  • Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration — provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima—which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  •  
    Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration - provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima-which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  •  
    A natural AI approach
Abdelrahman Ogail

Production system - Wikipedia, the free encyclopedia - 0 views

  • A production system (or production rule system) is a computer program typically used to provide some form of artificial intelligence, which consists primarily of a set of rules about behavior. These rules, termed productions, are a basic representation found useful in AI planning, expert systems and action selection. A production system provides the mechanism necessary to execute productions in order to achieve some goal for the system. Productions consist of two parts: a sensory precondition (or "IF" statement) and an action (or "THEN"). If a production's precondition matches the current state of the world, then the production is said to be triggered. If a production's action is executed, it is said to have fired. A production system also contains a database, sometimes called working memory, which maintains data about current state or knowledge, and a rule interpreter. The rule interpreter must provide a mechanism for prioritizing productions when more than one is triggered.
  • A production system (or production rule system) is a computer program typically used to provide some form of artificial intelligence, which consists primarily of a set of rules about behavior. These rules, termed productions, are a basic representation found useful in AI planning, expert systems and action selection. A production system provides the mechanism necessary to execute productions in order to achieve some goal for the system. Productions consist of two parts: a sensory precondition (or "IF" statement) and an action (or "THEN"). If a production's precondition matches the current state of the world, then the production is said to be triggered. If a production's action is executed, it is said to have fired. A production system also contains a database, sometimes called working memory, which maintains data about current state or knowledge, and a rule interpreter. The rule interpreter must provide a mechanism for prioritizing productions when more than one is triggered.
Janos Haits

Artificial Intelligence (AI) Machine Learning Advanced Technology Platform - 0 views

  •  
    "Our 2021.AI platform offers everything your team needs in one open platform, allowing your organization to manage team collaboration across heterogeneous infrastructure efficiently and deploy models effectively. Should you decide that you do not have the appetite to build such capacity and capabilities in-house, we will offer you data sciences as a service, ensuring your participation in harvesting and maximizing business benefits with a minimal organizational imprint."
Janos Haits

Futurepedia - The Largest AI Tools Directory | Home - 0 views

  •  
    "THE LARGEST AI TOOLS DIRECTORY, UPDATED DAILY."
Janos Haits

wizdom.ai - intelligence for everyone - 0 views

  •  
    "wizdom.ai is a result of extensive R&D by our team of data scientists, programmers, analysts, designers, quality engineers, product managers & process managers. The startup from the University of Oxford was founded by Tahir, Sadia, Rifaqat, David, Atikah and Asif."
Janos Haits

Meet Khanmigo, Khan Academy's AI-powered teaching assistant & tutor - 0 views

  •  
    "Meet Khanmigo, your go-to AI tool for learning and teaching. Now just $4/month.*"
Abdelrahman Ogail

Genetic algorithm - Wikipedia, the free encyclopedia - 0 views

  • A genetic algorithm (GA) is a search technique used in computing to find exact or approximate solutions to optimization and search problems. Genetic algorithms are categorized as global search heuristics. Genetic algorithms are a particular class of evolutionary algorithms (EA) that use techniques inspired by evolutionary biology such as inheritance, mutation, selection, and crossover (also called recombination).
  • A typical genetic algorithm requires: a genetic representation of the solution domain, a fitness function to evaluate the solution domain.
  •  
    GE are primary used in Learning in AI
Janos Haits

Talk to Books - 0 views

  •  
    "Browse passages from books using experimental AI"
Janos Haits

The A-Z of AI - 1 views

  •  
    "Making sense of artificial intelligence This A-Z guide offers a series of simple, bite-sized explainers to help anyone understand what AI is, how it works and how it's changing the world around us."
Janos Haits

Semantic Scholar - 0 views

  •  
    "Semantic Scholar is a free, nonprofit, academic search engine from AI2."
Islam TeCNo

Neuroplasticity - Wikipedia, the free encyclopedia - 0 views

  • Neuroplasticity (also referred to as brain plasticity, cortical plasticity or cortical re-mapping) is the changing of neurons and the organization of their networks and so their function by experience. This idea was first proposed in 1892 by Santiago Ramón y Cajal the proposer of the neuron doctrine though the idea was largely neglected for the next fifty years.[1] The first person to use the term neural plasticity appears to have been the Polish neuroscientist Jerzy Konorski.[2]
    • Abdelrahman Ogail
       
      This is why when a person thinks more he/she be more smarter!
    • Islam TeCNo
       
      3azeem .....bas deh mesh 7agat related l CS ya Zi3'az
    • Abdelrahman Ogail
       
      How said that? It's related to Artificial Neural Networks that require understanding of actual Human Neurons. BTW, plasticity principle is used in Games AI where the Controlled-AI determines if it forgot what happened or still remember it and seeks to revenge
    • Islam TeCNo
       
      oooooooooh ........ 3'reaaaaaat
Janos Haits

Quantum - Google Research - 0 views

  •  
    "A research effort from Google AI that aims to build quantum processors and develop novel quantum algorithms to dramatically accelerate computational tasks for machine learning."
Janos Haits

Welcome to city.forecasting.ai | The artificial intelligence community - 0 views

  •  
    "The artificial intelligence community"
Janos Haits

AI Chat for scientific PDFs | SciSpace - 0 views

  •  
    "Do hours worth of reading in minutes"
Janos Haits

Meta - AI for Science - 0 views

  •  
    "Meta is a tool that helps researchers understand what is happening globally in science and shows them where science is headed. Pending shareholder and court approval, the Chan Zuckerberg Initiative is acquiring Meta to help bring its technologies to the entire scientific community. Sign up now to reserve your free account."
Islam TeCNo

Deep Blue (chess computer) - Wikipedia, the free encyclopedia - 0 views

  • Deep Blue was a chess-playing computer developed by IBM. On May 11, 1997, the machine won a six-game match by two wins to one with three draws against world champion Garry Kasparov.[1] Kasparov accused IBM of cheating and demanded a rematch, but IBM declined and dismantled Deep Blue.[2] Kasparov had beaten a previous version of Deep Blue in 1996
    • Abdelrahman Ogail
       
      When AI beats humanity!
  • Deep Blue was then heavily upgraded (unofficially nicknamed "Deeper Blue")[11] and played Kasparov again in May 1997, winning the six-game rematch 3½–2½, ending on May 11, finally ending in game six, and becoming the first computer system to defeat a reigning world champion in a match under standard chess tournament time controls.
  • The system derived its playing strength mainly out of brute force computing power.
    • Islam TeCNo
       
      Dah eli bysamoh brute force men no3 el 7aywan :D
Abdelrahman Ogail

10 skills developers will need in the next five years | Between the Lines | ZDNet.com - 0 views

shared by Abdelrahman Ogail on 08 Jun 09 - Cached
Ahmed Mansour liked it
  • So bone up on JavaScript, CSS, and HTML to succeed over the next five years
  • Every top-flight developer I’ve met recommends learning at least one dynamic or functional programming language to learn new ways of thinking, and from personal experience, I can tell you that it works.
    • Abdelrahman Ogail
       
      yes, that's true (Y) some examples of these languages: LISP and Ruby. Prolog seems to be close to LISP becuase both were used in AI Applications
  •  
    1: One of the "Big Three" (.NET, Java, PHP) 2: Rich Internet Applications (RIAs) 3: Web development 4: Web services 5: Soft skills 6: One dynamic and/or functional programming language 7: Agile methodologies 8: Domain knowledge 9: Development "hygiene" 10: Mobile development
  •  
    1: One of the "Big Three" (.NET, Java, PHP) 2: Rich Internet Applications (RIAs) 3: Web development 4: Web services 5: Soft skills 6: One dynamic and/or functional programming language 7: Agile methodologies 8: Domain knowledge 9: Development "hygiene" 10: Mobile development
Abdelrahman Ogail

Genetic programming - Wikipedia, the free encyclopedia - 0 views

  • In artificial intelligence, genetic programming (GP) is an evolutionary algorithm-based methodology inspired by biological evolution to find computer programs that perform a user-defined task. It is a specialization of genetic algorithms (GA) where each individual is a computer program. Therefore it is a machine learning technique used to optimize a population of computer programs according to a fitness landscape determined by a program's ability to perform a given computational task.
  • In artificial intelligence, genetic programming (GP) is an evolutionary algorithm-based methodology inspired by biological evolution to find computer programs that perform a user-defined task. It is a specialization of genetic algorithms (GA) where each individual is a computer program. Therefore it is a machine learning technique used to optimize a population of computer programs according to a fitness landscape determined by a program's ability to perform a given computational task.
Abdelrahman Ogail

ELIZA - Wikipedia, the free encyclopedia - 0 views

  • ELIZA was a computer program and an early example (by modern standards) of primitive natural language processing. ELIZA operated by processing users' responses to scripts, the most famous of which was DOCTOR, a simulation of a Rogerian psychotherapist. In this mode, ELIZA mostly rephrased the user's statements as questions and posed those to the 'patient.' ELIZA was written by Joseph Weizenbaum between 1964 to 1966
1 - 20 of 33 Next ›
Showing 20 items per page