Control of Multiple Robots Using Vision Sensors

Control of Multiple Robots Using Vision Sensors
Author: Miguel Aranda
Publisher: Springer
Total Pages: 197
Release: 2017-05-11
Genre: Technology & Engineering
ISBN: 3319578286

This monograph introduces novel methods for the control and navigation of mobile robots using multiple-1-d-view models obtained from omni-directional cameras. This approach overcomes field-of-view and robustness limitations, simultaneously enhancing accuracy and simplifying application on real platforms. The authors also address coordinated motion tasks for multiple robots, exploring different system architectures, particularly the use of multiple aerial cameras in driving robot formations on the ground. Again, this has benefits of simplicity, scalability and flexibility. Coverage includes details of: a method for visual robot homing based on a memory of omni-directional images; a novel vision-based pose stabilization methodology for non-holonomic ground robots based on sinusoidal-varying control inputs; an algorithm to recover a generic motion between two 1-d views and which does not require a third view; a novel multi-robot setup where multiple camera-carrying unmanned aerial vehicles are used to observe and control a formation of ground mobile robots; and three coordinate-free methods for decentralized mobile robot formation stabilization. The performance of the different methods is evaluated both in simulation and experimentally with real robotic platforms and vision sensors. Control of Multiple Robots Using Vision Sensors will serve both academic researchers studying visual control of single and multiple robots and robotics engineers seeking to design control systems based on visual sensors.

Visual Control of Wheeled Mobile Robots

Visual Control of Wheeled Mobile Robots
Author: Héctor . M Becerra
Publisher: Springer
Total Pages: 127
Release: 2014-03-26
Genre: Technology & Engineering
ISBN: 3319057839

Vision-based control of wheeled mobile robots is an interesting field of research from a scientific and even social point of view due to its potential applicability. This book presents a formal treatment of some aspects of control theory applied to the problem of vision-based pose regulation of wheeled mobile robots. In this problem, the robot has to reach a desired position and orientation, which are specified by a target image. It is faced in such a way that vision and control are unified to achieve stability of the closed loop, a large region of convergence, without local minima and good robustness against parametric uncertainty. Three different control schemes that rely on monocular vision as unique sensor are presented and evaluated experimentally. A common benefit of these approaches is that they are valid for imaging systems obeying approximately a central projection model, e.g., conventional cameras, catadioptric systems and some fisheye cameras. Thus, the presented control schemes are generic approaches. A minimum set of visual measurements, integrated in adequate task functions, are taken from a geometric constraint imposed between corresponding image features. Particularly, the epipolar geometry and the trifocal tensor are exploited since they can be used for generic scenes. A detailed experimental evaluation is presented for each control scheme.

Robot Vision

Robot Vision
Author: Stefan Florczyk
Publisher: John Wiley & Sons
Total Pages: 216
Release: 2006-03-06
Genre: Technology & Engineering
ISBN: 352760491X

The book is intended for advanced students in physics, mathematics, computer science, electrical engineering, robotics, engine engineering and for specialists in computer vision and robotics on the techniques for the development of vision-based robot projects. It focusses on autonomous and mobile service robots for indoor work, and teaches the techniques for the development of vision-based robot projects. A basic knowledge of informatics is assumed, but the basic introduction helps to adjust the knowledge of the reader accordingly. A practical treatment of the material enables a comprehensive understanding of how to handle specific problems, such as inhomogeneous illumination or occlusion. With this book, the reader should be able to develop object-oriented programs and show mathematical basic understanding. Such topics as image processing, navigation, camera types and camera calibration structure the described steps of developing further applications of vision-based robot projects.

Learning-Based Robot Vision

Learning-Based Robot Vision
Author: Josef Pauli
Publisher: Springer
Total Pages: 292
Release: 2003-06-29
Genre: Computers
ISBN: 3540451242

Industrial robots carry out simple tasks in customized environments for which it is typical that nearly all e?ector movements can be planned during an - line phase. A continual control based on sensory feedback is at most necessary at e?ector positions near target locations utilizing torque or haptic sensors. It is desirable to develop new-generation robots showing higher degrees of autonomy for solving high-level deliberate tasks in natural and dynamic en- ronments. Obviously, camera-equipped robot systems, which take and process images and make use of the visual data, can solve more sophisticated robotic tasks. The development of a (semi-) autonomous camera-equipped robot must be grounded on an infrastructure, based on which the system can acquire and/or adapt task-relevant competences autonomously. This infrastructure consists of technical equipment to support the presentation of real world training samples, various learning mechanisms for automatically acquiring function approximations, and testing methods for evaluating the quality of the learned functions. Accordingly, to develop autonomous camera-equipped robot systems one must ?rst demonstrate relevant objects, critical situations, and purposive situation-action pairs in an experimental phase prior to the application phase. Secondly, the learning mechanisms are responsible for - quiring image operators and mechanisms of visual feedback control based on supervised experiences in the task-relevant, real environment. This paradigm of learning-based development leads to the concepts of compatibilities and manifolds. Compatibilities are general constraints on the process of image formation which hold more or less under task-relevant or accidental variations of the imaging conditions.

Omnidirectional Vision Systems

Omnidirectional Vision Systems
Author: Luis Puig
Publisher: Springer Science & Business Media
Total Pages: 129
Release: 2013-02-01
Genre: Computers
ISBN: 1447149475

This work focuses on central catadioptric systems, from the early step of calibration to high-level tasks such as 3D information retrieval. The book opens with a thorough introduction to the sphere camera model, along with an analysis of the relation between this model and actual central catadioptric systems. Then, a new approach to calibrate any single-viewpoint catadioptric camera is described. This is followed by an analysis of existing methods for calibrating central omnivision systems, and a detailed examination of hybrid two-view relations that combine images acquired with uncalibrated central catadioptric systems and conventional cameras. In the remaining chapters, the book discusses a new method to compute the scale space of any omnidirectional image acquired with a central catadioptric system, and a technique for computing the orientation of a hand-held omnidirectional catadioptric camera.

Omnidirectional Vision

Omnidirectional Vision
Author: Pascal Vasseur
Publisher: John Wiley & Sons
Total Pages: 260
Release: 2023-12-12
Genre: Computers
ISBN: 1394256434

Omnidirectional cameras, vision sensors that can capture 360° images, have in recent years had growing success in computer vision, robotics and the entertainment industry. In fact, modern omnidirectional cameras are compact, lightweight and inexpensive, and are thus being integrated in an increasing number of robotic platforms and consumer devices. However, the special format of output data requires tools that are appropriate for camera calibration, signal analysis and image interpretation. This book is divided into six chapters written by world-renowned scholars. In a rigorous yet accessible way, the mathematical foundation of omnidirectional vision is presented, from image geometry and camera calibration to image processing for central and non-central panoramic systems. Special emphasis is given to fisheye cameras and catadioptric systems, which combine mirrors with lenses. The main applications of omnidirectional vision, including 3D scene reconstruction and robot localization and navigation, are also surveyed. Finally, the recent trend towards AI-infused methods (deep learning architectures) and other emerging research directions are discussed.

Robot Vision

Robot Vision
Author: Gerald Sommer
Publisher: Springer
Total Pages: 477
Release: 2008-01-29
Genre: Computers
ISBN: 3540781579

In 1986, B.K.P. Horn published a book entitled Robot Vision, which actually discussed a wider ?eld of subjects, basically addressing the ?eld of computer vision, but introducing “robot vision” as a technical term. Since then, the - teraction between computer vision and research on mobile systems (often called “robots”, e.g., in an industrial context, but also including vehicles, such as cars, wheelchairs, tower cranes, and so forth) established a diverse area of research, today known as robot vision. Robot vision (or, more general, robotics) is a fast-growing discipline, already taught as a dedicated teaching program at university level. The term “robot vision” addresses any autonomous behavior of a technical system supported by visual sensoric information. While robot vision focusses on the vision process, visual robotics is more directed toward control and automatization. In practice, however, both ?elds strongly interact. Robot Vision 2008 was the second international workshop, counting a 2001 workshop with identical name as the ?rst in this series. Both workshops were organized in close cooperation between researchers from New Zealand and Germany, and took place at The University of Auckland, New Zealand. Participants of the 2008 workshop came from Europe, USA, South America, the Middle East, the Far East, Australia, and of course from New Zealand.

Robot Vision

Robot Vision
Author: A. Pugh
Publisher: Springer Science & Business Media
Total Pages: 347
Release: 2013-06-29
Genre: Technology & Engineering
ISBN: 3662097710

Over the past five years robot vision has emerged as a subject area with its own identity. A text based on the proceedings of the Symposium on Computer Vision and Sensor-based Robots held at the General Motors Research Laboratories, Warren, Michigan in 1978, was published by Plenum Press in 1979. This book, edited by George G. Dodd and Lothar Rosso!, probably represented the first identifiable book covering some aspects of robot vision. The subject of robot vision and sensory controls (RoViSeC) occupied an entire international conference held in the Hilton Hotel in Stratford, England in May 1981. This was followed by a second RoViSeC held in Stuttgart, Germany in November 1982. The large attendance at the Stratford conference and the obvious interest in the subject of robot vision at international robot meetings, provides the stimulus for this current collection of papers. Users and researchers entering the field of robot vision for the first time will encounter a bewildering array of publications on all aspects of computer vision of which robot vision forms a part. It is the grey area dividing the different aspects of computer vision which is not easy to identify. Even those involved in research sometimes find difficulty in separating the essential differences between vision for automated inspection and vision for robot applications. Both of these are to some extent applications of pattern recognition with the underlying philosophy of each defining the techniques used.

Machine Vision and Navigation

Machine Vision and Navigation
Author: Oleg Sergiyenko
Publisher: Springer Nature
Total Pages: 863
Release: 2019-09-30
Genre: Technology & Engineering
ISBN: 3030225879

This book presents a variety of perspectives on vision-based applications. These contributions are focused on optoelectronic sensors, 3D & 2D machine vision technologies, robot navigation, control schemes, motion controllers, intelligent algorithms and vision systems. The authors focus on applications of unmanned aerial vehicles, autonomous and mobile robots, industrial inspection applications and structural health monitoring. Recent advanced research in measurement and others areas where 3D & 2D machine vision and machine control play an important role, as well as surveys and reviews about vision-based applications. These topics are of interest to readers from diverse areas, including electrical, electronics and computer engineering, technologists, students and non-specialist readers. • Presents current research in image and signal sensors, methods, and 3D & 2D technologies in vision-based theories and applications; • Discusses applications such as daily use devices including robotics, detection, tracking and stereoscopic vision systems, pose estimation, avoidance of objects, control and data exchange for navigation, and aerial imagery processing; • Includes research contributions in scientific, industrial, and civil applications.

A Scalable Distributed Approach to Mobile Robot Vision

A Scalable Distributed Approach to Mobile Robot Vision
Author: National Aeronautics and Space Administration (NASA)
Publisher: Createspace Independent Publishing Platform
Total Pages: 36
Release: 2018-08-09
Genre:
ISBN: 9781725042131

This paper documents our progress during the first year of work on our original proposal entitled 'A Scalable Distributed Approach to Mobile Robot Vision'. We are pursuing a strategy for real-time visual identification and tracking of complex objects which does not rely on specialized image-processing hardware. In this system perceptual schemas represent objects as a graph of primitive features. Distributed software agents identify and track these features, using variable-geometry image subwindows of limited size. Active control of imaging parameters and selective processing makes simultaneous real-time tracking of many primitive features tractable. Perceptual schemas operate independently from the tracking of primitive features, so that real-time tracking of a set of image features is not hurt by latency in recognition of the object that those features make up. The architecture allows semantically significant features to be tracked with limited expenditure of computational resources, and allows the visual computation to be distributed across a network of processors. Early experiments are described which demonstrate the usefulness of this formulation, followed by a brief overview of our more recent progress (after the first year). Kuipers, Benjamin and Browning, Robert L. and Gribble, William S. Johnson Space Center NASA/CR-97-206090, NAS 1.26:206090 NAG9-828...