Skip to main content

Home/ Computer Science Knowledge Sharing/ Group items tagged time

Rss Feed Group items tagged

Janos Haits

Main Page - Time Machine - 0 views

  •  
    "Explore simultaneously in space and time with Time Machine Each Time Machine on this page captures a process in extreme detail over space and time, with billions of pixels of explorable resolution. Choose a time machine and zoom into the image while traveling backwards or forwards through time. Select a Time Warp and the time machine's authors will take you on a guided space-time tour with text annotations explaining what you are viewing. You can even learn how to create your own Time Machines and Warps."
Abdelrahman Ogail

CodeProject: C# vs C/C++ Performance. Free source code and programming help - 0 views

  • is compiled twice. Once while the program is written and second when the program is executed at the user's site. The first compilation is done by your C# builder and the second by the .NET Framework on the user's machine. The reason why C# compiled applications could be faster is that, during the second compilation, the compiler knows the actual run-time environment and processor type and could generate instructions that targets a specific processor.
  • A well designed C# program is more than 90% as fast as an equivalent "well-designed" C++ program
  • The problem with "not-freeing" the memory at the right time is that the working set of the application increases which increases the number of "page faults"
  • ...1 more annotation...
  • That's a nice question. Except for writing time-critical blocks of code, prefer C#. Write all your algorithmic code in C++ (not VC++ .NET), compile it into a dll and call that using a Dll Interop through C#. This should balance the performance. This technique is not new or not invented by me or anyone. It's similar the old age C programming vs Assembly, where people on one camp fight assembly programming is faster and the other camp stating C is easier to develop and then people started using assembly embedded within a C program for time-critical applications using an asm block.
  •  
    C# is compiled twice. Once while the program is written and second when the program is executed at the user's site. The first compilation is done by your C# builder and the second by the .NET Framework on the user's machine. The reason why C# compiled applications could be faster is that, during the second compilation, the compiler knows the actual run-time environment and processor type and could generate instructions that targets a specific processor
Abdelrahman Ogail

Simulated annealing - Wikipedia, the free encyclopedia - 1 views

  • Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration — provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima—which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  • Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration — provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima—which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  • Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration — provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima—which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  • ...1 more annotation...
  • Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration — provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima—which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  •  
    Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration - provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima-which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  •  
    A natural AI approach
Abdelrahman Ogail

Clockwork universe theory - Wikipedia, the free encyclopedia - 1 views

  • The Clockwork Universe Theory is a theory, established by Isaac Newton, as to the origins of the universe. A "clockwork universe" can be thought of as being a clock wound up by God and ticking along, as a perfect machine, with its gears governed by the laws of physics. What sets this theory apart from others is the idea that God's only contribution to the universe was to set everything in motion, and from there the laws of science took hold and have governed every sequence of events since that time. This idea was very popular during the Enlightenment, when scientists realized that Newton's laws of motion, including the law of universal gravitation, could explain the behavior of the solar system. A notable exclusion from this theory though is free will, since all things have already been set in motion and are just parts of a predictable machine. Newton feared that this notion of "everything is predetermined" would lead to atheism. This theory was undermined by the second law of thermodynamics ( the total entropy of any isolated thermodynamic system tends to increase over time, approaching a maximum value) and quantum physics with its unpredictable random behavior.
  • The Clockwork Universe Theory is a theory, established by Isaac Newton, as to the origins of the universe. A "clockwork universe" can be thought of as being a clock wound up by God and ticking along, as a perfect machine, with its gears governed by the laws of physics. What sets this theory apart from others is the idea that God's only contribution to the universe was to set everything in motion, and from there the laws of science took hold and have governed every sequence of events since that time. This idea was very popular during the Enlightenment, when scientists realized that Newton's laws of motion, including the law of universal gravitation, could explain the behavior of the solar system. A notable exclusion from this theory though is free will, since all things have already been set in motion and are just parts of a predictable machine. Newton feared that this notion of "everything is predetermined" would lead to atheism. This theory was undermined by the second law of thermodynamics ( the total entropy of any isolated thermodynamic system tends to increase over time, approaching a maximum value) and quantum physics with its unpredictable random behavior.
    • Abdelrahman Ogail
       
      "God's only contribution to the universe was to set everything in motion, and from there the laws of science took hold and have governed every sequence of events since that time" <-- ???
Janos Haits

A Million Dollar Request for Social Networks | Blockstack - 0 views

  •  
    "It's time for a new breed of social networks - where power is taken back from a single authority and control is returned to you, to me, to all of us. It's time to decentralize social networks.  This is a movement toward greater personal freedom, but it takes an empowered community to build this future."
mesbah095

Guest Post Online - 0 views

  •  
    Article Writing & Guestpost You Can Join this Site for Your Article & guest post, Just Easy way to join this site & total free Article site. This site article post to totally free Way. Guest Post & Article Post live to Life time only for Current & this time new User. http://guestpostonline.com
rack bank

Dedicated Server Hosting in India - 0 views

  •  
    Dedicated Server Hosting Solutions support a wide range of dedicated workloads which requires a high performance like remote desktop, dedicated email servers, virtualization, high volume real time database applications and telephony. Rackbank offer dedicated server with the configurations which perfectly suits your business requirement.
Abdelrahman Ogail

Belief-Desire-Intention software model - Wikipedia, the free encyclopedia - 0 views

  • The Belief-Desire-Intention (BDI) software model (usually referred to simply, but ambiguously, as BDI) is a software model developed for programming intelligent agents. Superficially characterized by the implementation of an agent's beliefs, desires and intentions, it actually uses these concepts to solve a particular problem in agent programming. In essence, it provides a mechanism for separating the activity of selecting a plan (from a plan library) from the execution of currently active plans. Consequently, BDI agents are able to balance the time spent on deliberating about plans (choosing what to do) and executing those plans (doing it). A third activity, creating the plans in the first place (planning), is not within the scope of the model, and is left to the system designer and programmer.
    • Abdelrahman Ogail
       
      Stress on the point "BDI Agents spent time about choosing what to do more that how to execute them"
cecilia marie

Reliable Online Computer Repair - 1 views

My PC has been acting strange lately and I can no longer fix it on my own. I did everything I could but this time, I really need someone who can really fix my computer problem. I called Computer Pr...

computer problem

started by cecilia marie on 17 May 11 no follow-up yet
anonymous

Reliable Online Computer Repair - 1 views

My PC has been acting strange lately and I can no longer fix it on my own. I did everything I could but this time, I really need someone who can really fix my computer problem. I called Computer Pr...

computer problem

started by anonymous on 12 May 11 no follow-up yet
Rem Comp

Remote Online PC Support I Can Rely On - 0 views

I availed of the remote computer support services of Remote Computer Support Site because their services are proven to be very fast and accurate. They have expert online computer tech professional...

remote computer support

started by Rem Comp on 12 Sep 11 no follow-up yet
Rem Comp

Remote Online PC Support I Can Rely On - 1 views

I availed of the remote computer support services of Remote Computer Support Site because their services are proven to be very fast and accurate. They have expert online computer tech professional...

remote computer support

started by Rem Comp on 12 Sep 11 no follow-up yet
joey potter

I Need Computer Technical Help Now - 1 views

"I have no time to wait". That is what I muttered when my computer suddenly went on a blue screen. I had a project and I needed to pass it on time. But with my computer on trouble, I would not make...

computer technical help

started by joey potter on 02 May 11 no follow-up yet
helen troy

Blue Screen of Death - 1 views

I was working with my school project and doing my part time job at the same time, so it happened that I opened numerous windows and tabs one after another. When all of a sudden my computer displaye...

need computer help

started by helen troy on 02 May 11 no follow-up yet
Janos Haits

Arch Mission - 0 views

  •  
    'The Arch Mission Foundation™ is a 501(c)(3) tax-exempt non-profit corporation designed to continuously preserve and disseminate humanity's most important knowledge across time and space.'
Janos Haits

AV-ATLAS - The Threat Intelligence Platform by AV-TEST - 0 views

  •  
    "The Threat Intelligence Platform by AV-TEST The threat situation around dangerous URLs, spam mails and software downloads is always in view thanks to up-to-date data in real time. With AV-ATLAS, the AV-TEST Institute presents you the Threat Intelligence Platform, which offers comprehensive Cyber Security Services for the defense against cyber threats."
Janos Haits

AV-ATLAS - The Threat Intelligence Platform by AV-TEST - 0 views

  •  
    "The Threat Intelligence Platform by AV-TEST The threat situation around dangerous URLs, spam mails and software downloads is always in view thanks to up-to-date data in real time. With AV-ATLAS, the AV-TEST Institute presents you the Threat Intelligence Platform, which offers comprehensive Cyber Security Services for the defense against cyber threats."
Janos Haits

Zotero | Home - 0 views

  •  
    A personal research assistant. Zotero is the only research tool that automatically senses content, allowing you to add it to your personal library with a single click. Whether you're searching for a preprint on arXiv.org, a journal article from JSTOR, a news story from the New York Times, or a book from your university library catalog, Zotero has you covered with support for thousands of sites.
Janos Haits

OExchange - 0 views

  •  
    OExchange makes it possible to share any URL-based content with any service on the web. It defines: A common way for services to receive content, removing any and all service-specific integration requirements A discovery feature so services can publish themselves and their endpoints, making it possible to integrate with services you didn't even know about at development time A decentralized, user-centric model for saving preferred services, making sharing more personal
1 - 20 of 46 Next › Last »
Showing 20 items per page