Skip to main content

Home/ Computer Science Knowledge Sharing/ Group items tagged processing

Rss Feed Group items tagged

computersciencej

Process Control Block - 0 views

  •  
    What is Process Control Block ? Today in this Computer Science Study Material for Gate we will discuss about process control block and its various field which provides the information about process. .So let see what is process control block. A Process Control Block is a data structure maintained by the Operating System for every process. Each process has it own data structure. When a process is created then a unique id is assigned to the process Operating system identify a process among all processes on the basis of this process id. A PCB keeps all the information needed to keep track of a process. Generally a process control block contains the following information about a process. To read full tutorial click on the given link http://www.computersciencejunction.in/2018/02/introduction-to-process-control-block-in-operating-system.html
Abdelrahman Ogail

Simulated annealing - Wikipedia, the free encyclopedia - 1 views

  • Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration — provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima—which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  • Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration — provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima—which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  • Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration — provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima—which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  • ...1 more annotation...
  • Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration — provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima—which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  •  
    Simulated annealing (SA) is a generic probabilistic metaheuristic for the global optimization problem of applied mathematics, namely locating a good approximation to the global minimum of a given function in a large search space. It is often used when the search space is discrete (e.g., all tours that visit a given set of cities). For certain problems, simulated annealing may be more effective than exhaustive enumeration - provided that the goal is merely to find an acceptably good solution in a fixed amount of time, rather than the best possible solution. The name and inspiration come from annealing in metallurgy, a technique involving heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The heat causes the atoms to become unstuck from their initial positions (a local minimum of the internal energy) and wander randomly through states of higher energy; the slow cooling gives them more chances of finding configurations with lower internal energy than the initial one. By analogy with this physical process, each step of the SA algorithm replaces the current solution by a random "nearby" solution, chosen with a probability that depends on the difference between the corresponding function values and on a global parameter T (called the temperature), that is gradually decreased during the process. The dependency is such that the current solution changes almost randomly when T is large, but increasingly "downhill" as T goes to zero. The allowance for "uphill" moves saves the method from becoming stuck at local minima-which are the bane of greedier methods. The method was independently described by S. Kirkpatrick, C. D. Gelatt and M. P. Vecchi in 1983 [1], and by V. Černý in 1985 [2]. The method is an adaptation of the Metropolis-Hastings algorithm, a Monte Carlo method to generate sample states of a thermodynamic system, invented by N. Metropolis et al. in 1953 [3].
  •  
    A natural AI approach
Abdelrahman Ogail

ELIZA - Wikipedia, the free encyclopedia - 0 views

  • ELIZA was a computer program and an early example (by modern standards) of primitive natural language processing. ELIZA operated by processing users' responses to scripts, the most famous of which was DOCTOR, a simulation of a Rogerian psychotherapist. In this mode, ELIZA mostly rephrased the user's statements as questions and posed those to the 'patient.' ELIZA was written by Joseph Weizenbaum between 1964 to 1966
Ahmed Mansour

The Object-Oriented Thought Process - 0 views

  •  
    in this articles "The Object-Oriented Thought Process"book is discussed ..i think it's a web copy if the book... anyway it's just for who haven't this book :D ..
Islam TeCNo

What is Tunneling? - 0 views

  • This process is different from a normal data transfer between nodes. Every frame passing through the tunnel will be encrypted with an additional layer of tunneling encryption and encapsulation which is also used for routing the packets to the right direction
  • This process is different from a normal data transfer between nodes. Every frame passing through the tunnel will be encrypted with an additional layer of tunneling encryption and encapsulation which is also used for routing the packets to the right direction
  • This process is different from a normal data transfer between nodes. Every frame passing through the tunnel will be encrypted with an additional layer of tunneling encryption and encapsulation which is also used for routing the packets to the right direction
cafe software

My Profitable Business Career - 1 views

Managing a cafe is a tedious task because I need to have a close supervision with my business sales and transactions. So I decided to purchase a cafe POS software that will help me have an easier m...

Online system

started by cafe software on 23 Jan 12 no follow-up yet
Janos Haits

Home - M-Lab - 0 views

  •  
    "Open Internet Measurement M-Lab provides the largest collection of open Internet performance data on the planet. As a consortium of research, industry, and public-interest partners, M-Lab is dedicated to providing an ecosystem for the open, verifiable measurement of global network performance. Real science requires verifiable processes, and M-Lab welcomes scientific collaboration and scrutiny. This is why all of the data colle"
Janos Haits

Main Page - Time Machine - 0 views

  •  
    "Explore simultaneously in space and time with Time Machine Each Time Machine on this page captures a process in extreme detail over space and time, with billions of pixels of explorable resolution. Choose a time machine and zoom into the image while traveling backwards or forwards through time. Select a Time Warp and the time machine's authors will take you on a guided space-time tour with text annotations explaining what you are viewing. You can even learn how to create your own Time Machines and Warps."
computersciencej

TCP/IP model questions based Study Material for gate Computer Science - 0 views

  •  
    TCP/IP protocol based questions for gate computer science exam Q1.What is the difference between transport and session layer of OSI model. Answer: OSI Model Transport Layer The transport layer uses the services provided by the network layer, such as best path selection and logical addressing, to provide end-to-end communication between source and destination. • The transport -layer data stream is a logical connection between the endpoints of a network. • End-to-end control is provided by sliding windows and reliability in sequencing numbers and acknowledgments. The transport layer regulates information flow to ensure end-to-end connectivity between host applications reliably and accurately. • The TCP/ IP protocol of Layer 4 (t transport t layer ) has two protocols. They are TCP and UDP. The transport layer accepts data from the session layer and segments the data for transport across the network. Generally, the transport layer is responsible for making sure that the data is delivered error-free and in the proper sequence. Flow control generally occurs at the transport layer. OSI Model Session Layer The session layer establishes, manages, and terminates communication sessions. Communication sessions consist of service requests and service responses that occur between applications located on different network devices. These requests and responses are coordinated by protocols implemented at the session layer. The session layer establishes, manages, and terminates sessions between applications Functions of the session layer and the different processes that occur as data packets travel through this layer. More specifically, you learned that Communication sessions consist of mini-conversations that occur between applications located on different network devices. Requests and responses are coordinated by protocols implemented at the session layer. • The session layer decides whether to use two-way simultaneous communication or two-way alternate communicati
veera90

Best Pharmacovigilance Services | Pharmacovigilance Professionals | ACL Digital Life Sc... - 0 views

  •  
    From proof-of-concept to post-marketing surveillance services, you can depend on our PV experts to efficiently work through the entire scope of Pharmacovigilance activities. We offer a high level of expertise and assist in meeting the highest standards of applicable national and global regulations. The experts at ACL Digital can easily customize safety monitoring services to suit your specific business requirements. Most biopharmaceutical companies have distinct and demanding clinical safety requirements as per the directions of regulatory agencies. You can depend on our PV specialists who plan safety and pharmacovigilance services accordingly that fit the needs of both your product and study - we adhere and adapt to your processes and are flexible enough to do it the right way.
veera90

Expert Biostatistics Services | Biostatistics | ACL Digital Life Sciences | IT Consulting - 0 views

  •  
    Biostatistics plays a vital role in clinical research. From protocol development and clinical trial designs to sample size calculation, data analysis and more, our team of Biostatisticians have the right SME expertise in multiple therapeutic areas to help deliver quality outcomes quickly and efficiently. You can rely on us to determine and apply the appropriate statistical model, write CSR sections to interact with regulatory authorities, and examine the efficacy of safety data. Innovative and insight-driven statistical methods play a crucial role in every step of the drug development process. At ACL Digital, biostatistics remains an integral part of our services.
Janos Haits

The Internet Computer | DFINITY - 0 views

  •  
    "We believe that "smart contracts" that run on a public blockchain are a vastly superior new form of software. However, traditional blockchains provide hosting frameworks for smart contracts that are highly limiting. We believe that once these limitations are removed by a next generation blockchain, a "blockchain singularity" will occur, in which the majority of the world's systems and services will be rebuilt entirely from smart contracts. In the process, these sys"
veera90

Key Home Automation Technologies that Shape the Industry - 0 views

  •  
    Business enterprises and companies with effective automation endeavors make it a strategic priority to augment employees and customer experience, and they use automation to do so. Just pursuing automation technologies to lower expenses is not adequate. To achieve success while scaling new and latest technologies, organizations should focus on the individuals involved (both employees and customers), commit to revamping current processes, and create new skill sets within the workforce to follow more distinct and innovative working methods.
Janos Haits

wizdom.ai - intelligence for everyone - 0 views

  •  
    "wizdom.ai is a result of extensive R&D by our team of data scientists, programmers, analysts, designers, quality engineers, product managers & process managers. The startup from the University of Oxford was founded by Tahir, Sadia, Rifaqat, David, Atikah and Asif."
Janos Haits

What is a Blockchain | Defi Dictionary | Web3 Daily - 0 views

  •  
    "Think of a blockchain as a transaction list that is uploaded to thousands of computers around the world.  If someone sends money to you, that transaction gets added to a queue. Queued transactions are sorted into groups (aka 'blocks') and then processed by one of the thousands of computers operating in the cryptocurrency's network. Once the transaction is complete, it appears on the public transaction list, or 'chain'."
Ahmed Mansour

CRC Cards Tutorial - 0 views

  • Object Oriented Analysis and Design using CRC Cards
  • A CRC cards is an index card that is use to represent the responsibilities of classes and the interaction between the classes. CRC cards are an informal approach to object oriented modeling. The cards are created through scenarios, based on the system requirements, that model the behavior of the system. The name CRC comes from Class, Responsibilities, and Collaborators which the creators found to be the essential dimensions of object oriented modeling.
    • Ahmed Mansour
       
      One of the most popular methods for identifying and categorizing classes is to use class-responsibility-collaboration cards (CRC). Each CRC card represents a single class's data attributes, responsibilities, and collaborations. source : Thought Process Book..
    • Ahmed Mansour
       
      You need to create three sections on each card: - The name of the class - The responsibilities of the class - The collaborations of the class
  • Why uses CRC cards?
  • ...4 more annotations...
  • They are portable... No computers are required so they can be used anywhere. Even away from the office. The allow the participants to experience first hand how the system will work. No computer tool can replace the interaction that happens by physically picking up the cards and playing the roll of that object... The are a useful tool for teaching people the object-oriented paradigm. They can be used as a methodology them selves or as a front end to a more formal methodology such as Booch, Wirfs-Brock, Jacobson, etc.
  • Tutorial
    • Ahmed Mansour
       
      here we can found simple tutorial for illustration...
  • A CRC cards is an index card that is use to represent the responsibilities of classes and the interaction between the classes. CRC cards are an informal approach to object oriented modeling. The cards are created through scenarios, based on the system requirements, that model the behavior of the system. The name CRC comes from Class, Responsibilities, and Collaborators which the creators found to be the essential dimensions of object oriented modeling.
  • Why uses CRC cards? They are portable... No computers are required so they can be used anywhere. Even away from the office. The allow the participants to experience first hand how the system will work. No computer tool can replace the interaction that happens by physically picking up the cards and playing the roll of that object... The are a useful tool for teaching people the object-oriented paradigm. They can be used as a methodology them selves or as a front end to a more formal methodology such as Booch, Wirfs-Brock, Jacobson, etc.
  •  
    This is great tutorial for CRC cards ( sort of object oriented modeling approach) \ni think it was useful for me ... hoping to be useful for all of us : ) ..
Abdelrahman Ogail

Voice over Internet Protocol - Wikipedia, the free encyclopedia - 0 views

shared by Abdelrahman Ogail on 01 Jul 09 - Cached
Ahmed One liked it
  • Voice over Internet Protocol (VoIP) is a general term for a family of transmission technologies for delivery of voice communications over IP networks such as the Internet or other packet-switched network
  • Internet telephony refers to communications services—voice, facsimile, and/or voice-messaging applications—that are transported via the Internet, rather than the public switched telephone network (PSTN). The basic steps involved in originating an Internet telephone call are conversion of the analog voice signal to digital format and compression/translation of the signal into Internet protocol (IP) packets for transmission over the Internet; the process is reversed at the receiving end.[1]
Abdelrahman Ogail

Artificial life - Wikipedia, the free encyclopedia - 2 views

  • Artificial life (commonly Alife or alife) is a field of study and an associated art form which examine systems related to life, its processes, and its evolution through simulations using computer models, robotics, and biochemistry.[1] There are three main kinds of alife[2], named for their approaches: soft[3], from software; hard[4], from hardware; and wet, from biochemistry. Artificial life imitates traditional biology by trying to recreate biological phenomena.[5] The term "artificial life" is often used to specifically refer to soft alife
  • The modeling philosophy of alife strongly differs from traditional modeling, by studying not only “life-as-we-know-it”, but also “life-as-it-might-be” [7].
Islam TeCNo

XML-RPC Home Page - 0 views

  • It's a spec and a set of implementations that allow software running on disparate operating systems, running in different environments to make procedure calls over the Internet. It's remote procedure calling using HTTP as the transport and XML as the encoding. XML-RPC is designed to be as simple as possible, while allowing complex data structures to be transmitted, processed and returned.
    • Islam TeCNo
       
      ya3ni a2dar men program C# maslan andah 3ala function PHP mawgoda 3ala server tani !! :D
  •  
    This is a protocol that is used to call functions over internet using HTTP protocol as the transport and XML as the encoding
1 - 20 of 24 Next ›
Showing 20 items per page