. Military Space News .
ROBO SPACE
Research identifies key weakness in modern computer vision systems
by Staff Writers
Providence RI (SPX) Aug 03, 2018

File image only

Computer vision algorithms have come a long way in the past decade. They've been shown to be as good or better than people at tasks like categorizing dog or cat breeds, and they have the remarkable ability to identify specific faces out of a sea of millions.

But research by Brown University scientists shows that computers fail miserably at a class of tasks that even young children have no problem with: determining whether two objects in an image are the same or different. In a paper presented last week at the annual meeting of the Cognitive Science Society, the Brown team sheds light on why computers are so bad at these types of tasks and suggests avenues toward smarter computer vision systems.

"There's a lot of excitement about what computer vision has been able to achieve, and I share a lot of that," said Thomas Serre, associate professor of cognitive, linguistic and psychological sciences at Brown and the paper's senior author.

"But we think that by working to understand the limitations of current computer vision systems as we've done here, we can really move toward new, much more advanced systems rather than simply tweaking the systems we already have."

For the study, Serre and his colleagues used state-of-the-art computer vision algorithms to analyze simple black-and-white images containing two or more randomly generated shapes. In some cases the objects were identical; sometimes they were the same but with one object rotated in relation to the other; sometimes the objects were completely different. The computer was asked to identify the same-or-different relationship.

The study showed that, even after hundreds of thousands of training examples, the algorithms were no better than chance at recognizing the appropriate relationship. The question, then, was why these systems are so bad at this task.

Serre and his colleagues had a suspicion that it has something to do with the inability of these computer vision algorithms to individuate objects. When computers look at an image, they can't actually tell where one object in the image stops and the background, or another object, begins. They just see a collection of pixels that have similar patterns to collections of pixels they've learned to associate with certain labels. That works fine for identification or categorization problems, but falls apart when trying to compare two objects.

To show that this was indeed why the algorithms were breaking down, Serre and his team performed experiments that relieved the computer from having to individuate objects on its own. Instead of showing the computer two objects in the same image, the researchers showed the computer the objects one at a time in separate images. The experiments showed that the algorithms had no problem learning same-or-different relationship as long as they didn't have to view the two objects in the same image.

The source of the problem in individuating objects, Serre says, is the architecture of the machine learning systems that power the algorithms. The algorithms use convolutional neural networks - layers of connected processing units that loosely mimic networks of neurons in the brain. A key difference from the brain is that the artificial networks are exclusively "feed-forward" - meaning information has a one-way flow through the layers of the network. That's not how the visual system in humans works, according to Serre.

"If you look at the anatomy of our own visual system, you find that there are a lot of recurring connections, where the information goes from a higher visual area to a lower visual area and back through," Serre said.

While it's not clear exactly what those feedbacks do, Serre says, it's likely that they have something to do with our ability to pay attention to certain parts of our visual field and make mental representations of objects in our minds.

"Presumably people attend to one object, building a feature representation that is bound to that object in their working memory," Serre said. "Then they shift their attention to another object. When both objects are represented in working memory, your visual system is able to make comparisons like same-or-different."

Serre and his colleagues hypothesize that the reason computers can't do anything like that is because feed-forward neural networks don't allow for the kind of recurrent processing required for this individuation and mental representation of objects. It could be, Serre says, that making computer vision smarter will require neural networks that more closely approximate the recurrent nature of human visual processing.


Related Links
Brown University
All about the robots on Earth and beyond!


Thanks for being here;
We need your help. The Space Media Network continues to grow but revenues have never been harder to maintain.

With the rise of Ad Blockers, and Facebook - our traditional revenue sources via quality network advertising continues to decline. And unlike so many other news sites, we don't have a paywall - with those annoying usernames and passwords.

Our news coverage takes time and effort to publish 365 days a year.

If you find our news sites informative and useful then please consider becoming a regular supporter or for now make a one off contribution.
SpaceMediaNetwork Contributor
$5 Billed Once


credit card or paypal
SpaceMediaNetwork Monthly Supporter
$5 Billed Monthly


paypal only


ROBO SPACE
Optical fibers that can feel the materials around them
Lausanne, Switzerland (SPX) Aug 02, 2018
In recent years optical fibers have served as sensors to detect changes in temperature, like a thermometer, and pressure, like an artificial nerve. This technique is particularly useful in structures such as bridges and gas pipelines. EPFL researchers have now come up with a new method that enables optical fibers to identify whether they are in contact with a liquid or a solid. This is achieved by simply generating a sound wave with the help from a light beam within the fiber. This study was ... read more

Comment using your Disqus, Facebook, Google or Twitter login.



Share this article via these popular social media networks
del.icio.usdel.icio.us DiggDigg RedditReddit GoogleGoogle

ROBO SPACE
Romania minister under fire over 'ballistic' gaffe

Japan to spend $4.2 bn over 30 years on missile defence system radar

US Congress pushes Ballistic Missile Defense Program based on laser-armed drones

Intercept Sets Distance Record for Lockheed Martin's Hit-to-Kill PAC-3 MSE

ROBO SPACE
UN panel finds further evidence of Iran link to Yemen missiles

Saudi-led coalition says destroyed Yemen rebel missile launch sites

Russian Scientist Jailed as Moscow Probes Hypersonic Missile Secrets Leak

Raytheon, Lockheed receive contract for Javelin missile production

ROBO SPACE
An insect-inspired drone deforms upon impact

AeroVironment awarded contract for drone data links for Norway

Insitu receives contract for ScanEagle UAVs for Afghanistan

Insitu awarded contract for RQ-21 unmanned aerial vehicles

ROBO SPACE
Why Ku-band HTS is superior for AISR

Asia is a huge growth market for government SATCOM

DARPA, Lockheed Martin Demonstrate Technologies to Enable a Connected Warfighter Network

IntelsatOne FlexAir Coming This Summer for Government Aircraft Operations

ROBO SPACE
Raytheon contracted for AN/DAS-4 targeting systems

White House backs court ban of 3D-printed guns

Atlantic Diving Supply awarded $49 million for M17, M18 holsters

3M to pay $9.1M in damages for defective combat earplugs

ROBO SPACE
Profits down at military equipment firm BAE Systems

US releases $195 million in frozen military aid to Egypt

EU anti-trust officials probe Thales, Gemalto merger

Some countries buying Russian gear deserve sanctions waivers: Mattis

ROBO SPACE
Georgia hosts major joint drills with NATO forces

Pentagon chief: talks with Russian counterpart possible

BRICS nations pledge unity against US trade war threat

Chinese man sets off explosive outside US embassy: police

ROBO SPACE
Researchers use nanotechnology to improve the accuracy of measuring devices

A new 'periodic table' for nanomaterials

Physicists uncover why nanomaterial loses superconductivity

Squeezing light at the nanoscale









The content herein, unless otherwise known to be public domain, are Copyright 1995-2024 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. General Data Protection Regulation (GDPR) Statement Our advertisers use various cookies and the like to deliver the best ad banner available at one time. All network advertising suppliers have GDPR policies (Legitimate Interest) that conform with EU regulations for data collection. By using our websites you consent to cookie based advertising. If you do not agree with this then you must stop using the websites from May 25, 2018. Privacy Statement. Additional information can be found here at About Us.