Skip to main content

Home/ Robotics & AI/ Group items tagged recognition

Rss Feed Group items tagged

Aasemoon =)

Scalable Object Recognition | Willow Garage - 0 views

  • Marius Muja from University of British Columbia returned to Willow Garage this summer to continue his work object recognition. In addition to working on an object detector that can scale to a large number of objects, he has also been designing a general object recognition infrastructure. One problem that many object detectors have is that they get slower as they learn new objects. Ideally we want a robot that goes into an environment and is capable of collecting data and learning new objects by itself. In doing this, however, we don't want the robot to get progressively slower as it learns new objects. Marius worked on an object detector called Binarized Gradient Grid Pyramid (BiGGPy), which uses the gradient information from an image to match it to a set of learned object templates. The templates are organized into a template pyramid. This tree structure has low resolution templates at the root and higher resolution templates at each lower level. During detection, only a fraction of this tree must be explored. This results in big speedups and allows the detector to scale to a large number of objects.
otakuhacks

Data annotation - 0 views

image

data-science data annotations annotation machine-learning

started by otakuhacks on 10 Nov 20 no follow-up yet
otakuhacks

What is Facial Recognition? - Applications & How it Works - 0 views

  •  
    My introductory guide to facial recognition in machine learning
Aasemoon =)

3D Point Cloud Based Object Recognition System | Willow Garage - 0 views

  • The main focus for Bastian's work was on the feature-extraction process for 3D data. One of his contributions was a novel interest keypoint extraction method that operates on range images generated from arbitrary 3D point clouds. This method explicitly considers the borders of the objects identified by transitions from foreground to background. Bastian also developed a new feature descriptor type, called NARF (Normal Aligned Radial Features), that takes the same information into account. Based on these feature matches, Bastian then worked on a process to create a set of potential object poses and added spatial verification steps to assure these observations fit the sensor data.
mikhail-miguel

Whisper - Multi-task speech recognition, translation, language ID (github.com). - 0 views

  •  
    Whisper: Multi-task speech recognition, translation, language ID (github.com).
mikhail-miguel

StarByFace - Discover your celebrity doppelganger with advanced facial recognition tech... - 0 views

  •  
    StarByFace: Discover your celebrity doppelganger with advanced facial recognition technology (starbyface.com).
mikhail-miguel

Wisecut - AI-powered video editing with voice recognition for fast, engaging content (w... - 0 views

  •  
    Wisecut: AI-powered video editing with voice recognition for fast, engaging content (wisecut.video).
mikhail-miguel

Photor Artificial Intelligence - Present the best version of yourself with advanced ima... - 0 views

  •  
    Photor Artificial Intelligence: Present the best version of yourself with advanced image recognition and ML (photor.io).
otakuhacks

Transformers in NLP: Creating a Translator Model from Scratch | Lionbridge AI - 0 views

  •  
    Transformers have now become the defacto standard for NLP tasks. Originally developed for sequence transduction processes such as speech recognition, translation, and text to speech, transformers work by using convolutional neural networks together with attention models, making them much more efficient than previous architectures. And although transformers were developed for NLP, they've also been implemented in the fields of computer vision and music generation. However, for all their wide and varied uses, transformers are still very difficult to understand, which is why I wrote a detailed post describing how they work on a basic level. It covers the encoder and decoder architecture, and the whole dataflow through the different pieces of the neural network. In this post, we'll get deeper into looking at transformers by implementing our own English to German language translator.
Aasemoon =)

Hitachi's Lovable EMIEW 2 Gets Upgraded! - 0 views

  • Remember Hitachi’s little helper robot, EMIEW 2 (Excellent Mobility and Interactive Existence as Work-mate)?  It’s been a couple of years since we heard anything regarding the project and we feared the worst.  Hitachi has put those fears to rest by holding a news conference to show off its new enhanced voice recognition and driving performance! Known primarily for its unique legs which have wheels for feet, EMIEW 2 can drive at up to 6km/h to keep pace with people.  If it needs to carry something it can kneel down (for added stability) and scoot around, and thanks to its bipedal legs it can step over obstacles that are too high to drive over.  Now it has been given adaptive suspension control technology which increases its stability when driving over bumpy terrain such as elevator doors.  During the press demonstration, EMIEW 2′s springy legs bobbed independently as it drove over cables and uneven flooring.
Aasemoon =)

robots.net - New Model Mimics Human Vision Tasks - 0 views

  • Researchers at MIT’s McGovern Institute for Brain Research are working on a new mathematical model to mimic the human brain's ability to identify objects. The model can predict human performance on certain visual-perception tasks suggesting it’s a good indication of what's actually happening in the brain. Researchers are hoping the new findings will make their way into future object-recognition systems for automation, mobile robotics, and other applications.
Aasemoon =)

ICT Results - Computers to read your body language? - 0 views

  • Can a computer read your body language? A consortium of European researchers thinks so, and has developed a range of innovative solutions from escalator safety to online marketing. The keyboard and mouse are no longer the only means of communicating with computers. Modern consumer devices will respond to the touch of a finger and even the spoken word, but can we go further still? Can a computer learn to make sense of how we walk and stand, to understand our gestures and even to read our facial expressions?The EU-funded MIAUCE project set out to do just that. "The motivation of the project is to put humans in the loop of interaction between the computer and their environment,” explains project coordinator Chaabane Djeraba, of CNRS in Lille. “We would like to have a form of ambient intelligence where computers are completely hidden,” he says. “This means a multimodal interface so people can interact with their environment. The computer sees their behaviour and then extracts information useful for the user."
Aasemoon =)

Video: Impressive Strawberry Picking Robot - 0 views

  • Developed by Japan’s National Agriculture and Food Research Organization and other local institutions, the robot may sound boring (when compared to humanoids, for example), but it’s actually pretty cool. The main bullet points are that it automatically detects how ripe the strawberries are (which fruit is ready for harvesting) and that it cuts the stalks without damaging the strawberries.
Aasemoon =)

Oh, Those Robot Eyes! | h+ Magazine - 0 views

  • Willow Garage is organizing a workshop at the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR) 2010 in San Francisco to discuss the intersection of computer vision with human-robot interaction. Willow Garage is the hardware and open source software organization behind the Robot Operating System (ROS) and the PR robot development platform. Here’s a recent video from Willow Garage of work done at the University of Illinois on how robots can be taught to perceive images:
Aasemoon =)

・NAMO - 0 views

  • NAMO (Novel Articulated MObile platform)  is a humanoid robot built by The Institute of Field Robotics (FIBO) at King Mongkut’s University of Technology Thonburi in Thailand. FIBO is active in the RoboCup scene and have developed a wide range of robot types, including an experimental biped.  NAMO was unveiled on March 29th 2010, serving as FIBO’s mascot as part of the university’s 50 year anniversary celebrations.  NAMO will be used to welcome people to the university and may be deployed at museums.  Given its friendly appearance and functionality, it could be used to research human robot interaction and communication. NAMO is 130cm (4′3″) tall and has 16 degrees of freedom.  It moves on a stable three-wheeled omnidirectional base, and is equipped with a Blackfin camera for its vision system.  It is capable of simple gesture recognition, visually tracks humans or objects of interest automatically, and can speak a few phrases in a child-like voice (in Thai).
Aasemoon =)

An Open Source Personal Robot On The Horizon? - 0 views

  • GetRobo has pointed out a new website by Francisco Paz, which focuses on his experience building an open source personal robot called Qbo.  From the few images on the site Qbo looks remarkably well made and quite similar to NEC’s PaPeRo, meaning it might be used to experiment with image processing, speech recognition, speech synthesis, and (assuming it has wheels) obstacle detection and SLAM.  He also mentions in his blog some of the open source software that’s out in the wild such as OpenCV, Festival, and Sphinx, which would allow you to do some of that.
otakuhacks

Audio classification - 0 views

Audio classification is the process of listening to and analyzing audio recordings. Also known as sound classification, this process is at the heart of a variety of modern AI technology including v...

audio classification data annotations annotation

started by otakuhacks on 10 Nov 20 no follow-up yet
Ellysa Perry

How to find AI app development companies - 1 views

  •  
    The innovation that has surfaced as a significant game-changer in the rapidly changing field of technology is artificial intelligence (AI). It is difficult to imagine that we have survived this long without it because it has become so embedded in both our personal and professional lives. Almost every aspect of our lives has been touched by and transformed by artificial intelligence, including virtual assistants, tailored search and shopping recommendations, facial recognition, maps and navigation, security, and surveillance. In addition to the personal sphere, artificial intelligence has revolutionized almost every business, including healthcare, banking, the military, cosmetics, and nearly all others. Software and apps driven by artificial intelligence (AI) have eliminated the need for human intervention and the possibility of costly mistakes by cutting down on repetitive and unnecessary tasks.
1 - 19 of 19
Showing 20 items per page