News Release

Image processing algorithm allows indoor drones to fly autonomously

Peer-Reviewed Publication

Chinese Association of Automation

A research team from Japan has developed a single-camera machine vision algorithm, making it possible for lightweight hovering indoor robots to guide themselves by identifying and interpreting reference points on a tiled floor. The technology opens the door to a new breed of functional, low-cost drones with potentially wide-ranging uses.

The team, led by Chinthaka Premachandra, an Associate Professor in the Department of Electronic Engineering at Shibaura Institute of Technology, published its findings in IEEE/CAA Journal of Automatica Sinica, a joint publication of the Institute of Electrical and Electronics Engineers (IEEE) and the Chinese Association of Automation (CAA).

Since GPS signals are too weak to penetrate most structures, indoor drones must rely on environmental queues, which are typically visual. A drone designed for indoor use is likely to be smaller and lighter than an outdoor drone, according to Premachandra. "We considered different hardware options, including laser rangefinders," he said. "But rangefinders are too heavy, and infrared and ultrasonic sensors suffer from low precision. So that led us to using a camera as the robot's visual sensor. If you think of the camera in your cell phone, that gives you an idea of just how small and light they can be."

Premachandra's research team aimed to design the guidance algorithm to be as simple as possible, allowing for a small, inexpensive microprocessor. The team used the Raspberry pi3, an open-source computing platform that weighs approximately 45 grams.

Their study prototype had a single downward facing camera with intentionally low resolution - only 80 by 80 pixels. "Our robot only needed to distinguish its direction of motion and identify corners. From there, our algorithm allows it to extrapolate its position in the room, helping it avoid contacting the walls," Premachandra said.

The team's program worked by taking each 80x80 image through a series of simple processing steps ending in a black and white grid, making it easier to quickly identify motion along the X and Y planes.

The guidance method is limited because it relies on a room with a tile floor and predictable patterns. Premachandra said next steps in research into lightweight, autonomous-hovering, indoor robots might include adapting the technology for infrared cameras so they could function in the dark, as well as adding a second camera so the robot could visually determine both its X,Y position and its altitude in the room.

"There are many potential applications," Premachandra said. "Hovering indoor robots may be useful in warehouses, distribution centers, and industrial applications to remotely monitor safety."

###

Fulltext of the paper is available: http://www.ieee-jas.org/article/doi/10.1109/JAS.2020.1003240?viewType=HTML&pageType=en

https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9128077

IEEE/CAA Journal of Automatica Sinica aims to publish high-quality, high-interest, far-reaching research achievements globally, and provide an international forum for the presentation of original ideas and recent results related to all aspects of automation.

The first Impact Factor of IEEE/CAA Journal of Automatica Sinica is 5.129, ranking among Top 17% (11/63, SCI Q1) in the category of Automation & Control Systems, according to the latest Journal Citation Reports released by Clarivate Analytics in 2020. In addition, its latest CiteScore is 8.3, and has entered Q1 in all three categories it belongs to (Information System, Control and Systems Engineering, Artificial Intelligence) since 2018.

Why publish with us: Fast and high quality peer review; Simple and effective online submission system; Widest possible global dissemination of your research; Indexed in SCIE, EI, IEEE, Scopus, Inspec. JAS papers can be found at http://ieeexplore.ieee.org/xpl/mostRecentIssue.jsp?punumber=6570654 or http://www.ieee-jas.org


Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.