Subscribe free to our newsletters via your
. Solar Energy News .




TECH SPACE
Researchers develop 'visual Turing test'
by Staff Writers
Providence RI (SPX) Mar 13, 2015


Athens, Baltimore, Hong Kong, Miami - What are those people doing? A new evaluation method measures a computer's ability to decipher movements, relationships, and implied intent from images by asking questions.

Researchers from Brown and Johns Hopkins universities have come up with a new way to evaluate how well computers can divine information from images. The team describes its new system as a "visual Turing test," after the legendary computer scientist Alan Turing's test of the extent to which computers display human-like intelligence.

"There have been some impressive advances in computer vision in recent years," said Stuart Geman, the James Manning Professor of Applied Mathematics at Brown. "We felt that it might be time to raise the bar in terms of how these systems are evaluated and benchmarked."

Traditional computer vision benchmarks tend to measure an algorithm's performance in detecting objects within an image (the image has a tree, or a car or a person), or how well a system identifies an image's global attributes (scene is outdoors or in the nighttime).

"We think it's time to think about how to do something deeper -- something more at the level of human understanding of an image," Geman said.

For example, it's one thing to be able to recognize that an image contains two people. But to be able to recognize that the image depicts two people walking together and having a conversation is a much deeper understanding. Similarly, describing an image as depicting a person entering a building is a richer understanding than saying it contains a person and a building.

The system Geman and his colleagues developed, described this week in the Proceedings of the National Academy of Sciences, is designed to test for such a contextual understanding of photos.

It works by generating a string of yes or no questions about an image, which are posed sequentially to the system being tested. Each question is progressively more in-depth and based on the responses to the questions that have come before.

For example, an initial question might ask a computer if there's a person in a given region of a photo. If the computer says yes, then the test might ask if there's anything else in that region -- perhaps another person. If there are two people, the test might ask: "Are person1 and person2 talking?"

As a group, the questions are geared toward gauging the computer's understanding of the contextual "storyline" of the photo. "You can build this notion of a storyline about an image by the order in which the questions are explored," Geman said.

Because the questions are computer-generated, the system is more objective than having a human simply query a computer about an image. There is a role for a human operator, however. The human's role is to tell the test system when a question is unanswerable because of the ambiguities of the photo.

For instance, asking the computer if a person in a photo is carrying something is unanswerable if most of the person's body is hidden by another object. The human operator would flag that question as ambiguous.

The first version of the test was generated based on a set of photos depicting urban street scenes. But the concept could conceivably be expanded to all kinds of photos, the researchers say.

Geman and his colleagues hope that this new test might spur computer vision researchers to explore new ways of teaching computers how to look at images. Most current computer vision algorithms are taught how to look at images using training sets in which objects are annotated by humans.

y looking at millions of annotated images, the algorithms eventually learn how to identify objects. But it would be very difficult to develop a training set with all the possible contextual attributes of a photo annotated. So true context understanding may require a new machine learning technique.

"As researchers, we tend to 'teach to the test,'" Geman said. "If there are certain contests that everybody's entering and those are the measures of success, then that's what we focus on. So it might be wise to change the test, to put it just out of reach of current vision systems."


Thanks for being here;
We need your help. The SpaceDaily news network continues to grow but revenues have never been harder to maintain.

With the rise of Ad Blockers, and Facebook - our traditional revenue sources via quality network advertising continues to decline. And unlike so many other news sites, we don't have a paywall - with those annoying usernames and passwords.

Our news coverage takes time and effort to publish 365 days a year.

If you find our news sites informative and useful then please consider becoming a regular supporter or for now make a one off contribution.
SpaceDaily Contributor
$5 Billed Once


credit card or paypal
SpaceDaily Monthly Supporter
$5 Billed Monthly


paypal only


.


Related Links
Brown University
Space Technology News - Applications and Research






Comment on this article via your Facebook, Yahoo, AOL, Hotmail login.

Share this article via these popular social media networks
del.icio.usdel.icio.us DiggDigg RedditReddit GoogleGoogle




Memory Foam Mattress Review
Newsletters :: SpaceDaily :: SpaceWar :: TerraDaily :: Energy Daily
XML Feeds :: Space News :: Earth News :: War News :: Solar Energy News





TECH SPACE
A new way to control information by mixing light and sound
New Haven CT (SPX) Mar 13, 2015
For once, slower is better in a new piece of technology. A Yale lab has developed a new, radio frequency processing device that allows information to be controlled more effectively, opening the door to a new generation of signal processing on microchips. One of the keys to the technology involves slowing information down. The new system, described in the March 5 edition of the journal Natu ... read more


TECH SPACE
CT scanning shows why tilting trees produce better biofuel

Bioelectrochemical processes have the potential to one day replace petrochemistry

Biofuel proteomics

Miscanthus-based ethanol boasts higher profits

TECH SPACE
Russian SAR-401 Space Robot Ready for the ISS

Kids and robots learn to write together

25 teams to participate in DARPA Robotics Challenge Finals

Rise of the Machines: video gamers beware

TECH SPACE
Time ripe for Atlantic wind, advocates say

Wind energy: TUV Rheinland supervises Senvion sale

Bright spot for wind farms amid RET gloom

Allianz acquire OX2 wind farm in northern Sweden

TECH SPACE
China auto sales edge down in February

Making our highways safer and more efficient

Understanding electric car 'range anxiety' could be key to wider acceptance

Car industry welcomes Google, Apple but battles loom

TECH SPACE
ORNL microscopy directly images problematic lithium dendrites in batteries

Chemists develop new way to make cost-effective material for electricity storage

ASU researchers explore longer life cycle for batteries

Japan space scientists make wireless energy breakthrough

TECH SPACE
Hungary denies EU nuclear veto report

South China nuclear plant operates second unit

France's Areva to cut 1,500 jobs in Germany

When it comes to nuclear disaster, safety really is in numbers

TECH SPACE
Reducing emissions with a more effective carbon capture method

China to further streamline energy layout amid "new normal"

Where you live could mean 'greener' alternatives do more harm than good

Europe still off mark on sustainability goals: report

TECH SPACE
Landless Brazilians in GM eucalyptus protest

Direct evidence that drought-weakened Amazonian forests 'inhale less carbon'

Amazon deforestation 'threshold' causes species loss to accelerate

Munching bugs thwart eager trees, reducing the carbon sink




The content herein, unless otherwise known to be public domain, are Copyright 1995-2014 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. Privacy Statement All images and articles appearing on Space Media Network have been edited or digitally altered in some way. Any requests to remove copyright material will be acted upon in a timely and appropriate manner. Any attempt to extort money from Space Media Network will be ignored and reported to Australian Law Enforcement Agencies as a potential case of financial fraud involving the use of a telephonic carriage device or postal service.