View Chapter

Chapter 23 — Biomimetic Robots

Kyu-Jin Cho and Robert Wood

Biomimetic robot designs attempt to translate biological principles into engineered systems, replacing more classical engineering solutions in order to achieve a function observed in the natural system. This chapter will focus on mechanism design for bio-inspired robots that replicate key principles from nature with novel engineering solutions. The challenges of biomimetic design include developing a deep understanding of the relevant natural system and translating this understanding into engineering design rules. This often entails the development of novel fabrication and actuation to realize the biomimetic design.

This chapter consists of four sections. In Sect. 23.1, we will define what biomimetic design entails, and contrast biomimetic robots with bio-inspired robots. In Sect. 23.2, we will discuss the fundamental components for developing a biomimetic robot. In Sect. 23.3, we will review detailed biomimetic designs that have been developed for canonical robot locomotion behaviors including flapping-wing flight, jumping, crawling, wall climbing, and swimming. In Sect. 23.4, we will discuss the enabling technologies for these biomimetic designs including material and fabrication.

Smooth vertical surface climbing with directional adhesion

Author  Sangbae Kim, Mark R. Cutkosky

Video ID : 389

Stickybot is a bioinspired robot that climbs smooth vertical surfaces such as those made of glass, plastic, and ceramic tile at 4 cm/s. The robot employs several design principles adapted from the gecko, including a hierarchy of compliant structures and directional adhesion. At the finest scale, the undersides of Stickybot’s toes are covered with arrays of small, angled polymer stalks.

Chapter 20 — Snake-Like and Continuum Robots

Ian D. Walker, Howie Choset and Gregory S. Chirikjian

This chapter provides an overview of the state of the art of snake-like (backbones comprised of many small links) and continuum (continuous backbone) robots. The history of each of these classes of robot is reviewed, focusing on key hardware developments. A review of the existing theory and algorithms for kinematics for both types of robot is presented, followed by a summary ofmodeling of locomotion for snake-like and continuum mechanisms.

Modsnake swimming

Author  Howie Choset

Video ID : 169

The CMU Modsnake swimming in a pool.

Chapter 10 — Redundant Robots

Stefano Chiaverini, Giuseppe Oriolo and Anthony A. Maciejewski

This chapter focuses on redundancy resolution schemes, i. e., the techniques for exploiting the redundant degrees of freedom in the solution of the inverse kinematics problem. This is obviously an issue of major relevance for motion planning and control purposes.

In particular, task-oriented kinematics and the basic methods for its inversion at the velocity (first-order differential) level are first recalled, with a discussion of the main techniques for handling kinematic singularities. Next, different firstorder methods to solve kinematic redundancy are arranged in two main categories, namely those based on the optimization of suitable performance criteria and those relying on the augmentation of the task space. Redundancy resolution methods at the acceleration (second-order differential) level are then considered in order to take into account dynamics issues, e.g., torque minimization. Conditions under which a cyclic task motion results in a cyclic joint motion are also discussed; this is a major issue when a redundant manipulator is used to execute a repetitive task, e.g., in industrial applications. The use of kinematic redundancy for fault tolerance is analyzed in detail. Suggestions for further reading are given in a final section.

Human inspired tele-impedance and minimum-effort controller for improved manipulation Performance‬

Author  IIT Videos

Video ID : 815

Humans incorporate and switch between learnt neuro-motor strategies while performing complex tasks. To this purpose, kinematic redundancy is exploited in order to achieve optimized performance. Inspired by the superior motor skills of humans, in this work, we investigate a combined free-motion and contact-efficient controller in a certain class of robotic manipulation. In this multiple-criteria controller, kinematic degrees of redundancy are adapted according to task-suitable dynamic costs. The proposed algorithm attributes high priority to a minimum-effort controller while performing point-to-point, free-space movements. Once the robot comes into contact with the environment, the tele-impedance, common mode stiffness (CMS)-configuration dependent stiffness (CDS) controller will replicate the human's estimated endpoint stiffness and measured equilibrium-position profiles in the slave robotic arm, in real-time.

Chapter 68 — Human Motion Reconstruction

Katsu Yamane and Wataru Takano

This chapter presents a set of techniques for reconstructing and understanding human motions measured using current motion capture technologies. We first review modeling and computation techniques for obtaining motion and force information from human motion data (Sect. 68.2). Here we show that kinematics and dynamics algorithms for articulated rigid bodies can be applied to human motion data processing, with help from models based on knowledge in anatomy and physiology. We then describe methods for analyzing human motions so that robots can segment and categorize different behaviors and use them as the basis for human motion understanding and communication (Sect. 68.3). These methods are based on statistical techniques widely used in linguistics. The two fields share the common goal of converting continuous and noisy signal to discrete symbols, and therefore it is natural to apply similar techniques. Finally, we introduce some application examples of human motion and models ranging from simulated human control to humanoid robot motion synthesis.

Converting human motion to sentences

Author  Katsu Yamane

Video ID : 766

This video shows an example of converting human motion sequences to descriptive sentences.

Chapter 69 — Physical Human-Robot Interaction

Sami Haddadin and Elizabeth Croft

Over the last two decades, the foundations for physical human–robot interaction (pHRI) have evolved from successful developments in mechatronics, control, and planning, leading toward safer lightweight robot designs and interaction control schemes that advance beyond the current capacities of existing high-payload and highprecision position-controlled industrial robots. Based on their ability to sense physical interaction, render compliant behavior along the robot structure, plan motions that respect human preferences, and generate interaction plans for collaboration and coaction with humans, these novel robots have opened up novel and unforeseen application domains, and have advanced the field of human safety in robotics.

This chapter gives an overview on the state of the art in pHRI as of the date of publication. First, the advances in human safety are outlined, addressing topics in human injury analysis in robotics and safety standards for pHRI. Then, the foundations of human-friendly robot design, including the development of lightweight and intrinsically flexible force/torque-controlled machines together with the required perception abilities for interaction are introduced. Subsequently, motionplanning techniques for human environments, including the domains of biomechanically safe, risk-metric-based, human-aware planning are covered. Finally, the rather recent problem of interaction planning is summarized, including the issues of collaborative action planning, the definition of the interaction planning problem, and an introduction to robot reflexes and reactive control architecture for pHRI.

Human-robot interaction planning

Author  Sven Parusel, Hannes Widmoser, Saskia Golz, Tobias Ende, Nico Blodow, Matteo Saveriano, Kai Krieger, Alexis Maldonado, Ingo Kresse, Roman Weitschat, Dongheui Lee, Michael Beetz, Sami Haddadin

Video ID : 616

The video presents the main aspects that have to be taken into consideration for joint human-robot assembly. These are: i) planning and appropriately distributing the tasks between human, robot, and collaboration; ii) a suitable interface between human and robot by visual and haptic gestures; iii) compliant and sensitive robot control in delivery, storage, hand-over, and assembly of parts; iv) collision detection and distinguishing from intended contacts during collaboration. The overall concept is presented for the exemplary assembly of a toy-train-track. (AAAI 2014, Video Competition)

Chapter 13 — Behavior-Based Systems

François Michaud and Monica Nicolescu

Nature is filled with examples of autonomous creatures capable of dealing with the diversity, unpredictability, and rapidly changing conditions of the real world. Such creatures must make decisions and take actions based on incomplete perception, time constraints, limited knowledge about the world, cognition, reasoning and physical capabilities, in uncontrolled conditions and with very limited cues about the intent of others. Consequently, one way of evaluating intelligence is based on the creature’s ability to make the most of what it has available to handle the complexities of the real world. The main objective of this chapter is to explain behavior-based systems and their use in autonomous control problems and applications. The chapter is organized as follows. Section 13.1 overviews robot control, introducing behavior-based systems in relation to other established approaches to robot control. Section 13.2 follows by outlining the basic principles of behavior-based systems that make them distinct from other types of robot control architectures. The concept of basis behaviors, the means of modularizing behavior-based systems, is presented in Sect. 13.3. Section 13.4 describes how behaviors are used as building blocks for creating representations for use by behavior-based systems, enabling the robot to reason about the world and about itself in that world. Section 13.5 presents several different classes of learning methods for behavior-based systems, validated on single-robot and multirobot systems. Section 13.6 provides an overview of various robotics problems and application domains that have successfully been addressed or are currently being studied with behavior-based control. Finally, Sect. 13.7 concludes the chapter.

The Nerd Herd

Author  Maja J. Mataric

Video ID : 34

This is a video showing the work done in the early 1990s with the Nerd Herd used as a multirobot behavior-based system. Reference: M.J. Matarić: Designing and understanding adaptive group behavior, Adapt. Behav. 4(1), 50–81 (1995)

Chapter 72 — Social Robotics

Cynthia Breazeal, Kerstin Dautenhahn and Takayuki Kanda

This chapter surveys some of the principal research trends in Social Robotics and its application to human–robot interaction (HRI). Social (or Sociable) robots are designed to interact with people in a natural, interpersonal manner – often to achieve positive outcomes in diverse applications such as education, health, quality of life, entertainment, communication, and tasks requiring collaborative teamwork. The long-term goal of creating social robots that are competent and capable partners for people is quite a challenging task. They will need to be able to communicate naturally with people using both verbal and nonverbal signals. They will need to engage us not only on a cognitive level, but on an emotional level as well in order to provide effective social and task-related support to people. They will need a wide range of socialcognitive skills and a theory of other minds to understand human behavior, and to be intuitively understood by people. A deep understanding of human intelligence and behavior across multiple dimensions (i. e., cognitive, affective, physical, social, etc.) is necessary in order to design robots that can successfully play a beneficial role in the daily lives of people. This requires a multidisciplinary approach where the design of social robot technologies and methodologies are informed by robotics, artificial intelligence, psychology, neuroscience, human factors, design, anthropology, and more.

A scene of deictic interaction

Author  Takayuki Kanda

Video ID : 807

This video illustrates the "deictic interaction" in which the robot and a user interact using pointing gestures and verbal-reference terms. The robot has a capability to understand the user's deictic interaction recognizing both the pointing gesture and the reference term. In addition, there is a 'facilitation' mechanism (e.g., the robot engages in real-time joint attention), which makes the interaction smooth and natural.

Chapter 62 — Intelligent Vehicles

Alberto Broggi, Alex Zelinsky, Ümit Özgüner and Christian Laugier

This chapter describes the emerging robotics application field of intelligent vehicles – motor vehicles that have autonomous functions and capabilities. The chapter is organized as follows. Section 62.1 provides a motivation for why the development of intelligent vehicles is important, a brief history of the field, and the potential benefits of the technology. Section 62.2 describes the technologies that enable intelligent vehicles to sense vehicle, environment, and driver state, work with digital maps and satellite navigation, and communicate with intelligent transportation infrastructure. Section 62.3 describes the challenges and solutions associated with road scene understanding – a key capability for all intelligent vehicles. Section 62.4 describes advanced driver assistance systems, which use the robotics and sensing technologies described earlier to create new safety and convenience systems for motor vehicles, such as collision avoidance, lane keeping, and parking assistance. Section 62.5 describes driver monitoring technologies that are being developed to mitigate driver fatigue, inattention, and impairment. Section 62.6 describes fully autonomous intelligent vehicles systems that have been developed and deployed. The chapter is concluded in Sect. 62.7 with a discussion of future prospects, while Sect. 62.8 provides references to further reading and additional resources.

PROUD2013 - Inside VisLab's driverless car

Author  Alberto Broggi

Video ID : 178

This video shows the internal and external view of what happened during the PROUD2013 driverlesscar test in downtown Parma, Italy, on July 12, 2013. It also displays the internal status of the vehicle plus some vehicle data (speed, steering angle, and some perception results like pedestrian detection, roundabout merging alert, freeway merging alert, traffic light sensing, etc.). More info available from www.vislab.it/proud.

Chapter 45 — World Modeling

Wolfram Burgard, Martial Hebert and Maren Bennewitz

In this chapter we describe popular ways to represent the environment of a mobile robot. For indoor environments, which are often stored using two-dimensional representations, we discuss occupancy grids, line maps, topologicalmaps, and landmark-based representations. Each of these techniques has its own advantages and disadvantages. Whilst occupancy grid maps allow for quick access and can efficiently be updated, line maps are more compact. Also landmark-basedmaps can efficiently be updated and maintained, however, they do not readily support navigation tasks such as path planning like topological representations do.

Additionally, we discuss approaches suited for outdoor terrain modeling. In outdoor environments, the flat-surface assumption underling many mapping techniques for indoor environments is no longer valid. A very popular approach in this context are elevation and variants maps, which store the surface of the terrain over a regularly spaced grid. Alternatives to such maps are point clouds, meshes, or three-dimensional grids, which provide a greater flexibility but have higher storage demands.

Service-robot navigation in urban environments

Author  Christian Siagian

Video ID : 270

This video presents the navigation system of the Beobot service robot of the iLab, University of Southern California (USC). Beobot's task is to fulfill services in urban-like environments, especially those involving long-range travel. The robot uses a topological map for global localization based on acquired images.

Chapter 20 — Snake-Like and Continuum Robots

Ian D. Walker, Howie Choset and Gregory S. Chirikjian

This chapter provides an overview of the state of the art of snake-like (backbones comprised of many small links) and continuum (continuous backbone) robots. The history of each of these classes of robot is reviewed, focusing on key hardware developments. A review of the existing theory and algorithms for kinematics for both types of robot is presented, followed by a summary ofmodeling of locomotion for snake-like and continuum mechanisms.

First concentric tube robot teleoperation

Author  Pierre Dupont

Video ID : 250

This 2007 video showcases Brandon Itkowitz's MS thesis work at Boston University. It is the first demonstration of teleoperated control of a concentric tube robot. The robot workspace is the size of a heart chamber. Similar to the child's game "Operation", the user attempts to sequentially touch electrical contacts inside each numbered bead without touching the wire loops surrounding the hole in each bead.