Using Text-Spotting to Query the World

The world we live in is labeled extensively for the benefit of humans. Yet, to date, robots have made little use of human readable text as a resource. In this paper we aim to draw attention to text as a readily available source of semantic information in robotics by implementing a system which allow...

Full description

Bibliographic Details
Main Authors: Posner, I, Corke, P, Newman, P, IEEE
Format: Conference item
Published: 2010
_version_ 1797063578881097728
author Posner, I
Corke, P
Newman, P
IEEE
author_facet Posner, I
Corke, P
Newman, P
IEEE
author_sort Posner, I
collection OXFORD
description The world we live in is labeled extensively for the benefit of humans. Yet, to date, robots have made little use of human readable text as a resource. In this paper we aim to draw attention to text as a readily available source of semantic information in robotics by implementing a system which allows robots to read visible text in natural scene images and to use this knowledge to interpret the content of a given scene. The reliable detection and parsing of text in natural scene images is an active area of research and remains a non-trivial problem. We extend a commonly adopted approach based on boosting for the detection and optical character recognition (OCR) for the parsing of text by a probabilistic error correction scheme incorporating a sensor-model for our pipeline. In order to interpret the scene content we introduce a generative model which explains spotted text in terms of arbitrary search terms. This allows the robot to estimate the relevance of a given scene with respect to arbitrary queries such as, for example, whether it is looking at a bank or a restaurant. We present results from images recorded by a robot in a busy cityscape. ©2010 IEEE.
first_indexed 2024-03-06T21:01:55Z
format Conference item
id oxford-uuid:3b2aa1ed-b416-41e0-a6fb-acc387db8e28
institution University of Oxford
last_indexed 2024-03-06T21:01:55Z
publishDate 2010
record_format dspace
spelling oxford-uuid:3b2aa1ed-b416-41e0-a6fb-acc387db8e282022-03-26T14:06:02ZUsing Text-Spotting to Query the WorldConference itemhttp://purl.org/coar/resource_type/c_5794uuid:3b2aa1ed-b416-41e0-a6fb-acc387db8e28Symplectic Elements at Oxford2010Posner, ICorke, PNewman, PIEEEThe world we live in is labeled extensively for the benefit of humans. Yet, to date, robots have made little use of human readable text as a resource. In this paper we aim to draw attention to text as a readily available source of semantic information in robotics by implementing a system which allows robots to read visible text in natural scene images and to use this knowledge to interpret the content of a given scene. The reliable detection and parsing of text in natural scene images is an active area of research and remains a non-trivial problem. We extend a commonly adopted approach based on boosting for the detection and optical character recognition (OCR) for the parsing of text by a probabilistic error correction scheme incorporating a sensor-model for our pipeline. In order to interpret the scene content we introduce a generative model which explains spotted text in terms of arbitrary search terms. This allows the robot to estimate the relevance of a given scene with respect to arbitrary queries such as, for example, whether it is looking at a bank or a restaurant. We present results from images recorded by a robot in a busy cityscape. ©2010 IEEE.
spellingShingle Posner, I
Corke, P
Newman, P
IEEE
Using Text-Spotting to Query the World
title Using Text-Spotting to Query the World
title_full Using Text-Spotting to Query the World
title_fullStr Using Text-Spotting to Query the World
title_full_unstemmed Using Text-Spotting to Query the World
title_short Using Text-Spotting to Query the World
title_sort using text spotting to query the world
work_keys_str_mv AT posneri usingtextspottingtoquerytheworld
AT corkep usingtextspottingtoquerytheworld
AT newmanp usingtextspottingtoquerytheworld
AT ieee usingtextspottingtoquerytheworld