Vision, audition, olfactory, tactile and taste are five important senses that human uses to interact with the real world. As facing more and more complex environments, a sensing system is essential for intelligent robots with various types of sensors. To mimic human-like abilities, sensors similar to human perception capabilities are indispensable. However, most research only concentrated on analyzing literature on single-modal sensors and their robotics application.
This study presents a systematic review of five bioinspired senses, especially considering a brief introduction of multimodal sensing applications and predicting current trends and future directions of this field, which may have continuous enlightenments.
This review shows that bioinspired sensors can enable robots to better understand the environment, and multiple sensor combinations can support the robot’s ability to behave intelligently.
The review starts with a brief survey of the biological sensing mechanisms of the five senses, which are followed by their bioinspired electronic counterparts. Their applications in the robots are then reviewed as another emphasis, covering the main application scopes of localization and navigation, objection identification, dexterous manipulation, compliant interaction and so on. Finally, the trends, difficulties and challenges of this research were discussed to help guide future research on intelligent robot sensors.
1. Introduction
Since 1959, robots have played an essential role in many dangerous, exhausting and repetitive tasks, such as continuous use in the industrial field to improve efficiency and quality (Oztemel and Gursev, 2020). With the development of perception range and accuracy in recent decades, intelligent robots gradually stepped out of the isolation of industrial scenes and began to integrate into human daily life. As shown in Figure 1, to meet increasingly diverse and complex application requirements, intelligent robots should possess:
Summary of intelligent robot capabilities and perception requirements in complex environments
Summary of intelligent robot capabilities and perception requirements in complex environments
environmental perception capabilities as inputs for robot interactions (Tian et al., 2017);
motion execution capabilities to cope with complex tasks (Abdelaal et al., 2021); and
thinking and evolution capabilities to regulate and improve their behaviors adaptively according to the various environments (Pointeau et al., 2014).
Furthermore, the foundation of robot intelligence is an environmental perception (Young, 2017), the basis of which is sensors, transforming the physical excitation of the environment into electrical signal outputs. As shown in Figure 1, the sensors applied in intelligent robots should have the following characteristics:
rich sensing types to meet the demand of complex dynamic environment perception;
excellent sensing performance to ensure reliable information inputs for intelligent robots; and
ingenious device structure to satisfy mechanical and electrical constrains of intelligent robots.
Inspired by biological organisms, researchers construct types of intelligent robot structures to be applied to different scenarios (Ren and Liang, 2014), such as humanoid robots in household and social scenes (Van Doorn et al., 2017; Kuindersma et al., 2016), four-legged robots on complicated ground (Guizzo, 2019; Hao et al., 2020) and flying robots in the sky (Zeng and Zhang, 2017; Mozaffari et al., 2016). Likewise, humans have five important sensing, (i.e. vision, audition, tactile, olfactory and taste), which are expected to provide natural bioinspired prototypes for intelligent robot sensors, as shown in Figure 2. It is expected to promote the research of intelligent robot sensing systems that meet the above characteristics, by studying the sensing mechanisms of biological systems.
Up to then, there have been several surveys published on sensing of intelligent robots, such as vision sensors (Steffen et al., 2019), odor sensors (Chen and Huang, 2019), vibration sensors (Pearson et al., 2011) and tactile sensors (Yang et al., 2019). Most research only concentrated on analyzing literature on single-modal sensors and robotics applications. To this end, this article presents a comprehensive review of the sensors corresponding to the five human senses, especially brief thinking of multimodal sensing and predicting current trends and prospects of this field, which may have continuous enlightenments.
The rest of this article is organized as follows: Section 2 briefly reviews the biological perception mechanisms of vision, audition, tactile, olfactory and taste, while the Section 3 introduces the corresponding bioinspired sensors for robots; Section 4 introduces the applications of bioinspired sensors on intelligent robots; Section 5 discusses the trends, difficulties and challenges of bioinspired sensors; and Section 6 is a summary of the review.
2. Biological sensing mechanisms of five senses
Through sensory organs, humans are able to perceive and interact with their surroundings. The five senses (sight, hearing, touch, smell and taste) are fundamental to human perceptions and are given particular weight. The first three of them are known as responses to physical stimuli, and the latter two can perceive the chemical composition changes of the environment. In this section, we briefly discuss the biological perception mechanisms focusing on these five sense organs.
2.1 Biological vision sensing mechanism
Vision is integral to human perception, acting as the primary conduit for gathering environmental information through eyes (Chen et al., 2017). Human eyes, spherical in nature, are sectioned into two major parts – front and back – bisected by the crystalline lens situated centrally. This lens is transparent, elastic and convex, adjusting its shape via ligament suspension to accommodate varying distances of target objects. Both the lens and the cornea, located at the front of the eyeball, collaboratively manage the eye’s focus, whereas the dilation and constriction of the pupil within the colored iris regulate the volume of light entering the eye. At the rear of the eyeball lies the retina, the designated area for image reception. Cellular electrical signals here respond to fluctuations in light and are transmitted through optic nerve fibers to the pertinent brain region, thereby constructing our visual experience (Gu et al., 2020).
2.2 Biological auditory sensing mechanism
The human ear is capable of detecting sounds within a frequency range of 20 Hz–20 kHz, primarily relying on the conductive capabilities of the outer and middle ear, in conjunction with the sensory functions of the inner ear, auditory nerve and auditory cortex (Yao and Zhang, 2001). Sound reaches the inner ear via both air and bone conduction. Air conduction is a sound transmission mechanism in which sound waves, collected by the pinna, traverse the external auditory canal to reach the tympanic membrane and then propagate through the ossicular chain. This chain comprises three ossicles: the malleus, incus and stapes. Specifically, the footplate of the stapes is in close contact with the oval window of the inner ear. The activity of the oval window membrane excites the adjacent inner ear lymph. This fluid incites the spiral organs situated on the basilar membrane to generate nerve impulses. These impulses are then conveyed via the auditory nerve to the auditory cortex in the brain, resulting in the sensation of hearing. Bone conduction refers to an alternative sound transmission mechanism where sound waves reach the inner ear via the vibration of the skull. It causes the lymph fluid to vibrate, which, in turn, stimulates the basilar membrane to generate nerve impulses. These impulses are similarly transmitted to the auditory cortex, culminating in the perception of sound (Zhang and Gan, 2011).
2.3 Biological tactile sensing mechanism
Human tactile perception is facilitated by the skin, abundant in mechanoreceptors that are part of the peripheral nervous system. Sensitive nerve cells, such as Merkel disks, Meissner corpuscles and free nerve endings, reside in the dermis layer. The skin’s soft tissue deforms upon tactile contact, triggering these embedded nerve cells to generate electrical pulses. These pulses reflect the stress, strain and temperature experienced by the surrounding soft tissue during stimulation (Romo and Salinas, 1999; Maeno et al., 1998; Dahiya et al., 2009; Abraira and Ginty, 2013). Consequently, the types and locations of the tactile signal are differentiated, giving rise to the sensation of touch that travels along the nerve fibers to the brain.
2.4 Biological olfactory sensing mechanism
Human odor receptors are located in the upper olfactory epithelium of the nasal cavity, which is mainly composed of olfactory sensory neurons, supporting cells and basal cells (Lledo et al., 2005; Sharma et al., 2019; Bushdid et al., 2014). The dendrites end become round and expand dozens of cilia into the mucus after reaching the surface of the olfactory epithelium. There are G protein-coupled odor receptors on the plasma membrane of cilia, which have seven helical transmembrane structures to combine odor molecules. Theodor receptors then activate the specific G protein Golf, open the cyclic nucleotide gating ion channel, produce membrane depolarization and form an action potential. The oscillation pattern (frequency and amplitude) of the potential generated on the olfactory bulb is affected by the odor, thereby realizing the detection and recognition of odor molecules (Firestein, 2001).
2.5 Biological taste sensing mechanism
The generation of human taste depends on taste buds (Roper and Chaudhari, 2017; Taruno et al., 2013). Taste cells are cloaked in a bilipid layer membrane, with microvilli adorning their apices. The rest of the surface is wrapped by groove-like cells that are flat and opaque to the external environment. The microvilli can only interact with the mouth’s saliva via a small orifice at the peak of the taste bud. When taste substances are absorbed by lipid membranes on different taste cells, various electrical properties (such as membrane potential) change to encode corresponding taste experiences.
3. Bioinspired robot sensors of five senses
To endow intelligent robots with human-like senses, researchers have committed to the study of sensing technology rooted in biological sensing principles (Pfeifer et al., 2007). In this section, we overview the advancements in the field, focusing on representative biomimetic sensors.
3.1 Bioinspired vision sensors
Charge-coupled device (CCD) and complementary metal oxide semiconductor (CMOS) vision sensors are commonly used in the robotics field (Fossum and Hondongwa, 2014). However, CCD sensors (Zhang et al., 2020) and CMOS sensors (Fossum and Hondongwa, 2014; Nikonov and Young, 2013) typically sample and transmit a large amount of pixels within a set time frame, which inevitably leads to delays and redundancy in information while also placing significant demands on bandwidth, energy consumption and storage.
As a result, bioinspired vision sensors have continued to emerge, including dynamic vision sensors (DVSs) (Lichtsteiner et al., 2006; Lichtsteiner et al., 2008), asynchronous time-based image sensors (ATISs) (Posch et al., 2010; Posch et al., 2007) and dynamic and active pixel vision sensors (DAVISs) (Brandli et al., 2014a, Lenero-Bardallo et al., 2015). The bioinspired vision sensor combines the biological functions of “where” and “what” of the human visual system using an asynchronous event-driven method to process visual information. They boast benefits such as the suppression of redundancy, integrated processing, rapid sensing capabilities, a wide dynamic range and low power consumption.
In 2002, Kramer et al. proposed DVS that simulates the function of human retinal Y cells to perceive the dynamic information in the scene (Kramer, 2002). In 2004, Lichtsteiner proposed an improved practical DVS to provide a more uniform and symmetric output of positive and negative events, thus significantly expanding the dynamic perception range (Lichtsteiner et al., 2004). It makes DVS the first commercial neuromorphic sensor, which is similar to its corresponding biological model: responding to natural events that occur in the observation scene based on event-driven mode rather than clock-driven mode (Sivilotti, 1991; Mahowald, 1992). Pixels change autonomously with the relative intensity with microsecond time resolution. Each pixel asynchronously sends an ON event if the logarithmic compressed light intensity of the pixel increases by a fixed amount. Conversely, an OFF event will be sent. This system enables continuous information transmission and processing, with communication bandwidth being used only by activated pixels. Recently in 2017, Yuan Xu further reduced DVS noise by introducing the concept of a smoothing filter (Xu et al., 2017).
ATIS integrates the biological visual perception functions of “where” and “what,” leveraging a variety of biological heuristics such as event-based imaging. Posch committed to enhancing DVS by adding the DPS structure, leading to the ATIS (Posch et al., 2010; Posch et al., 2014; Chen et al., 2011). The sensor consists of a fully autonomous pixel array, which incorporates an event-based change detector (CD) and an exposure measurement (EM) unit based on pulse width modulation (PWM). The EM unit, locally activated by a single pixel, uses the CD to detect the brightness changes within its field of view. ATIS outputs both time contrast event data and the absolute intensity of each event.
DAVIS, another visual sensing sensor that simulates the “where” and “what” of biological visual perception, was proposed by Christian Brandli in 2014 (Brandli et al., 2014a, Berner et al., 2013; Brandli et al., 2014b). DAVIS combines the advantages of DVSs and the active pixel sensors (APSs) at the pixel level, synchronously coordinating the APS asynchronous image frame outputs and the DVS asynchronous event outputs. The incorporation of a shared photodiode and a more compact APS circuit results in the DAVIS pixel area being 60% smaller than the ATIS pixel area.
3.2 Bioinspired auditory sensors
Most of the design of the bioinspired auditory sensors is derived from the structure of the human ear. For example, many scholars used nano/microstructures and piezoelectric materials to fabricate a variety of flexible vibration sensors on ultrathin substrates. Such design adjusts the resonance frequency through optimization of the three-dimensional (3D) geometric dimensions of the membrane, allowing selective absorption of vibration frequencies by parameter combination to filter certain frequency ranges (An et al., 2018). Some researchers used piezoelectric materials as substitutes for damaged cochlear hair cells to provide the electrical signal in the artificial auditory system (Fan et al., 2015; Yang et al., 2015). These sensors can rapidly detect the vibrations changing across a wide frequency range (from 100 to 3,200 Hz) showing potential to replace artificial auditory organs (Yu et al., 2016; Gu et al., 2015; Seol et al., 2014; Yang et al., 2014; Wu et al., 2018). Other than those, Mannoor has 3D-printed entirely functional, bioinspired ears using hydrogels embedded with living cells, which maintain structural and shape integrity (Mannoor et al., 2013).
Most artificial auditory systems struggle in conditions with low signal-to-noise ratios and varying acoustic environments. Inspired by the human hearing mechanism, Claudia (Lenk et al., 2023) developed a microelectromechanical cochlea as a bioinspired acoustic sensor with integrated signal processing functionality and adaptability tuned for noisy conditions. Composed of bionic acoustic sensors with thermomechanical feedback mechanisms, exhibits active auditory sensing, allowing the sensor to adjust its properties to different acoustic environments. Real-time feedback is used to tune the sensing and processing properties, and dynamic switching between linear and nonlinear characteristics improves the detection of signals in noisy conditions, increases the sensor dynamic range and enables adaptation to changing acoustic environments.
In the human ear, spike-based auditory neurons biologically process sound signals in a remarkably energy-efficient manner. Inspired by this, Yun (Yun et al., 2023) constructed a self-aware artificial auditory neuron module by serially connecting a triboelectric nanogenerator (TENG) and a bi-stable resistor, detecting the sound pressure level and encoding it into a spike form, which is then relayed to artificial synapses [metal-oxide-semiconductor field-effect transistors (MOSFETs)] as input neurons for the spiking neural network. The TENG serves as a sound sensor and a current source to feed current input so as to awaken the resistor, which acts as a device-level neuron that is dissimilar to a traditional circuit-based neuron.
In addition, another research focus is designing sensors inspired by some organisms that have unique vibrational receptors that can perceive an extremely wide range of sound waves or selectively filter and discern specific frequency ranges. The research and practical implementation of such sound-sensing mechanisms could potentially result in hearing devices that surpass human capabilities. For instance, the design of a resistance strain gauge based on a flexible metal film, capable of detecting minute vibrations with an amplitude of merely 10 nm, was inspired by the unique crack structures found on spider metatarsals (Kang et al., 2014). Similarly, sensor fibers that emulate the exceptional vibrational sensing performance of the Venus flytrap can sensitively detect directional vibration signals ranging from 20 kHz to several GHz (McConney et al., 2009; Fratzl and Barth, 2009; Talukdar et al., 2015). The outstanding acoustic localization ability of the Omeya brown fly is another typical bioinspired prototype for auditory sensors with microphone arrays (Miles et al., 1997; Lisiewski et al., 2011).
3.3 Bioinspired tactile sensors
Tactile perception is a crucial component of the intelligent perception of autonomous robots. It provides information on the force and surface characteristics of the contact point between the robot and the object(Chen et al., 2021). Tactile sensors can be classified according to human-like conduction mechanisms, such as changes in capacitance, resistance and charge. Pressure-sensitive resistors use the piezoresistive effect to sense touch, and the resistance changes under external mechanical forces (Stassi et al., 2014; Kou et al., 2018). Based on this principle, force-sensing resistors with multilayer structures (Fukui et al., 2011; Teshigawara et al., 2011; Drimus et al., 2014; Büscher et al., 2015) are widely used in pressure-sensing equipment for contact localization (Dahiya et al., 2011). Capacitive tactile sensors operate on the principle of plate capacitance, wherein the capacitance value shifts with the varying distance between plates due to normal forces (Lee et al., 2008), and the tangential force can also be detected by embedding multiple capacitors. Piezoelectric sensors use the piezoelectric effect to sense dynamic contact force (Dahiya et al., 2009; Seminara et al., 2011). The quantum tunneling effect sensor can change from an insulator to a conductor under compression, achieving normal and tangential force sensing with a sensitivity up to 0.45 mV/mN and a maximum dynamic value of 20 N (Zhang et al., 2012).
Innovative configurations, using optical and air pressure sensing mechanisms, and multimodal tactile sensors are also emerging. The optical sensor measures the contact force based on the light reflection between materials with different refractive indexes under stress, as exemplified by the “GelSight” tactile sensor proposed by Johnson et al. (2011), and TacLINK proposed by JAIST (Van Duong, 2020). Tactile sensors based on the air/hydraulic pressure measurement use pressure sensors that are generally used for air and liquid pressure sensing. These sensors use the air or liquid inside them as a vibration propagation medium, allowing for the simultaneous gathering of high-frequency responses and deformation, as demonstrated by the BioTac finger from SynTouch LLC (Lai et al., 2011; Yu et al., 2018). The structure-acoustic tactile sensor uses the accelerometer and microphone to measure the contact force (Kyberd et al., 1998), offering a broad detection bandwidth, though limited to dynamic sensing. The sensor imitating the shell effect can estimate the distance to an object by comparing the noise level in the environment and inside the sensor (Jiang and Smith, 2012). Li designed a magnetostrictive tactile sensor that is able to detect the force and stiffness of the manipulated object (Li et al., 2018).
Inspired by the touch mechanism of human skin, Zhou (Zhou et al., 2023) has proposed a custom, modular and multimodal robotic haptic sensing sensor. This sensor, comprising pressure, proximity, acceleration and temperature detectors, is scalable and apt for broad surface coverage on robots. With a joint focus on mechanical structure and data fusion algorithm design, a multilevel event-driven data fusion algorithm is used to efficiently process information from a large array of haptic sensors.
3.4 Bioinspired olfactory sensors
The electronic nose is a typical artificial olfactory that mimics biological olfaction. The early related research can be traced back to the 1990s, focusing on the mechanical structure design of the odor-sensing module. Ishida et al. (1999) proposed a biologically inspired implementation of a 3D odor compass to detect the flow direction, consisting of four odor sensors, two motors and a fan. The odor sensors obtain relatively independent stimulation in separate airflows through mechanical isolations (like the septum in biological noses and the sniffing effect of silkworm wings), and the two motors rotate the compass head individually to seek the direction with balanced outputs from the sensors. However, the mechanical electronic nose is inefficient and vulnerable because of the limitation of mechanical structure.
Therefore, another focus of bioinspired olfactory sensor research lies in the design of odor sensor arrays (as olfactory receptors) and signal processing systems (as olfactory cortical neural networks). The bioinspired olfactory sensor inherits the advantages of the biochemical systems with high sensitivity, fast response and strong specificity. In 2012, Juhun Park’s team proposed an electronic nose using carbon nano transistor sensors and odor-sensitive materials to express the olfactory sensory neurons of the cfOR5269 receptors (Park et al., 2012). This nose was used to selectively detect the hexanal, which is a food oxidation indicator. In 2016, Jing’s team designed an electronic nose consisting of six semiconductor gas sensors and five infrared sensors and proposed a bioinspired neural network to process the odor concentration signal of the electronic nose. Classifications of eight Chinese liquors of similar degrees from different brands have been realized without any signal preprocessing and feature selection (Jing et al., 2016).
The sense of smell in vivo has also developed rapidly in recent years. Lu et al. fixed the odor binding receptor of honeybees on the surface of the impedance chip to specifically detect pheromone (Lu et al., 2014). Strauch’s team from Rome University used in vivo calcium imaging technology to study the response of the olfactory neurons on the Drosophila antennae to volatile substances produced by cancer cells and noncancerous cells (Strauch et al., 2014). Nowotny’s team from Sussex University used microelectrodes to record Drosophila olfactory neurons in vivo, the classifications of 36 different alcoholic gases and 35 industrial gases were achieved with the support vector machine classification algorithm (Nowotny et al., 2014).
3.5 Bioinspired taste sensors
The first electronic tongue, which relies on the cross-reaction of a sensor array to analyze liquid samples, was introduced by Vlasov and his team in 2008. Researchers have harnessed the local field potential of the taste cortex alongside the action potential of a single neuron (spike) to effectively differentiate between substances such as sucrose, salt, hydrochloric acid and denatonium benzoate (Liu et al., 2010; Zhang et al., 2014; Qin et al., 2017). Much like olfactory receptors, the receptors for bitter, sweet and umami tastes are also G protein coupled. Scientists from Seoul University and Ohio State University expressed the human taste receptor hT1R2+hT1R3 heterodimer on the membrane of HEK-293 cells. These cells were then shattered into nanovesicles bearing taste receptors and anchored on the surface of a field effect transistor, creating a novel sensing technology for detecting sweet substances (Song et al., 2014). Certain researchers have used a quartz crystal microbalance to devise a bioinspired taste sensor capable of effectively detecting bitter substance denatonium within a particular concentration range (Wu et al., 2013). In 2019, Wang developed a new type of bioelectronic taste sensor specifically for detecting bitter substances (Wang et al., 2019).
The field of in vivo taste sensing is rapidly advancing as well. In 2016, Qin introduced an in vivo taste biosensor using brain-computer interface technology (Qin et al., 2016). They implanted a microelectrode array into the taste cortex of rats and recorded the electrophysiological activities of the neural network under taste stimulation. This biosensor demonstrated high sensitivity in detecting bitterness. Other researchers (Qin et al., 2017) have also reported a comprehensive animal biosensor, which leverages brain-machine interface technology in conjunction with anesthetized rats for the detection of bitter substances.
4. Robotic applications of bioinspired sensors
Bioinspired sensing forms the foundation for human-like robot behavior and safe human–robot interaction (Figure 3). In this section, we aim to survey the related applications of bioinspired sensors on intelligent robots.
4.1 Applications of vision in robots
Vision, which is widely used in robot perception, allows for the identification of the shape, location, color and movement of target objects (Servières et al., 2021). One of the most researched fields is simultaneous localization and mapping (SLAM), commonly used in tasks such as environmental perception, comprehension, self-localization and path planning in unfamiliar environments. SLAM serves as the foundation for autonomous navigation and obstacle avoidance of robots (Cadena et al., 2016; Fuentes-Pacheco et al., 2015). Vison-based SLAM (VSLAM) integrates SLAM with GPU, RGB-D cameras and stereo cameras (Schneider et al., 2018). Nowadays, the robustness and reliability of VSLAM have been highly improved, resulting in its widespread use in practical applications (Chen et al., 2022).
The applications of bioinspired vision sensors are emerging. For example, in 2018, Scaramuzza took the frequency accumulation image as input to the deep learning network, based on an “image frame” concept (Zhang and Scaramuzza, 2018). The ON and OFF pulse streams were accumulated into grayscale images according to the time-domain frequency, followed by the ResNet to predict steering wheel angles in the automatic driving scene. In 2019, Zhu mapped the pulse stream outputs from the DVS into grayscale images in the time-domain order and estimated the optical flow using the proposed EV-FlowNet network (Zhu et al., 2019). Zhu proposes a visual reconstruction framework inspired by biological asynchronous spikes for the first time. Combined with the information on spatiotemporal spikes, the neuromorphic adaptive adjustment mechanism is used to reconstruct vision with an ultrahigh temporal resolution for natural scene perception, analysis and recognition (Zhu et al., 2022).
4.2 Applications of audition in robots
Language is one of the primary means of information expression and communication, not just for humans but also for robots. In the realm of human–robot interaction and communication, language plays a key role in the user interface and is instrumental in controlling speech-active robots. SHAKEY-II, which might be the first endeavor in this field, emerged in 1960 (Frenkel and Karen, 1985). Nowadays, almost all intelligent terminals, such as mobile phones, laptops and vehicles, are equipped with speech recognition software.
Sound source localization, a classic application of nonspeech sound detection in robotics, is accomplished by determining the source of sound with respect to one or multimicrophones (Sasaki et al., 2011). Jwu-Sheng Hu et al. proposed a method combining Direction of Arrival estimation and bearing-only SLAM for simultaneous localization of a mobile robot and an unknown number of multiple sound sources in the environment (Hu et al., 2011). Guanqun Liu et al. equipped a wheeled robot with a 32-channel concentric microphone array to build a two-dimensional sound source map using direction localization (Sasaki et al., 2006). When more sound sources occur, bioinspired attention mechanisms are used for localization of the loudest one (Michaud et al., 2007) or the closest one (Nguyen and Choi, 2016).
Artificial bioinspired sound sensors that outperform human hearing capabilities have been developed, broadening the application of underwater robots. For instance, acoustic navigation systems are used for the navigation and localization of autonomous underwater vehicles (AUV), where global positioning system and high-frequency radio signals are not available in the underwater environment (Siddiqui et al., 2015; Allotta et al., 2014). Optoacoustic imaging and passive acoustics are even used for exo-ocean exploration and understanding (Aguzzi et al., 2020).
4.3 Applications of tactile in robots
Robots come into physical interaction with humans and the environment under a variety of circumstances. The tactile sensing is required for scenes such as human–robot safe collaboration, skillful manipulation and special applications like soft robots. For these reasons, tactile sensors are installed on various parts of the robot, including hands, arms, feet or even the whole body. Active infrared proximity sensors, which are similar to the hairs on some insect legs, might be the first tactile sensors applied in robotic applications. As early as 1989, a real-time motion planning system for a robot arm manipulator, fully covered with active infrared proximity sensors, was proposed to operate among arbitrarily shaped obstacles (Cheung and Lumelsky, 1989). For robots to closely interact with humans and the environment, direct force sensing is necessary (Suen et al., 2018), further involved in the recognition of touch gestures and emotions (Li et al., 2022).
Currently, several commercially available tactile sensors, like the Tekscan sensor (Brimacombe et al., 2009), the “GelSight” optical sensor (Li and Adelson, 2013) and the BioTac sensor (Ruppel et al., 2019), are integrated into robot hands or fingers to enable sophisticated hand operations. To achieve whole-body compliance upon contact, large-area tactile skins have been the focus of many studies. addressing challenges such as scalability, robotic coverage and real-time data processing. Solutions often involve techniques like flexible materials, modular design and artificial intelligence (Chun et al., 2018; Dean-Leon et al., 2019). Gordon Cheng constructed a robot skin consisting of 1260 self-organized and self-calibrated hexagonal sensor “cells,” each carrying temperature, pressure (normal forces), accelerations (in x, y and z directions) and proximity sensors, plus a microcontroller for local computation, data reading and communication. Robots covered by this skin can achieve their position, velocity and torque through the sensor data and subsequent calculation. Moreover, soft robots integrated with tactile sensing exhibit significant potential in shape sensing, feedback control, manipulation, tactile exploration, reaction and so on (Tang et al., 2021; Shih et al., 2020).
4.4 Applications of olfactory in robots
Sniffing robots, generally equipped with electronic noses, can complete hazardous tasks such as the identification and localization of gas leaks and environmental monitoring of volatile substances. To improve detection capability and identification accuracy, the research mainly focuses on the design of electronic noses and gas signal processing algorithms.
Early results in this field focused on the design of electronic nose structures. Rozas et al. constructed a robot with a nose of mechanical structure (using a fan for active sniffing) with six different types of semiconductor gas sensors for Cf, Cb, Cl and Cr sensing. The robot located the odor source by moving along the increase of odor concentration (Rozas et al., 1991). To address the challenges of complex and variable airflow, active search strategies are also essential in addition to electronic nose in robotic odor source localization. Therefore, various algorithms have been proposed to guide the path planning for mobile robots, such as gradient-based algorithms and bioinspired algorithms (Chen and Huang, 2019). Furthermore, the use of multiple robots equipped with odor sensors is also an important research hotspot in robot active olfactory. Hayes et al. used a group of six robots to construct an odor concentration distribution map of the interested area, followed by the map analysis to obtain the odor source information (Hayes et al., 2002). Recently, with the high miniaturization and low power consumption of gas sensors, aerial drones carried with electronic noses have been used for environmental chemical sensing applications (Burgués and Marco, 2020), or even working in swarms (Szulczyński et al., 2017) for the rapid odor distribution acquisition over unknown area with relatively large size for agriculture and security services.
4.5 Applications of taste in robots
Similar to electronic noses, electronic tongues consist of a taste sensor array that can analyze one substance or more. An electronic tongue could allow robots to gain a taste sense by responding to chemicals in solutions or solids. Several commercial platforms are available, such as the SA 402B taste-sensing system (Intelligent Sensor Technology Company, Limited., Tokyo, Japan), TS-5000Z (Insent Inc., Atsugi-chi, Japan) and αAstree2 (AlphaMOS, Toulouse, France). These systems normally feature a robotic arm and a rotary sample holder platform, where the arm holds testing probes with taste sensors and inserts into the target samples. Then the experiment data is recorded and analyzed by a computer. However, these systems can hardly be identified as intelligent robots, because their behaviors are limited to automatic sensing data collecting and analyzing.
Compared with electronic noses, the applications of electronic tongues in robots are less common. Russell constructed a mobile robot called The Autonomous Sensory Tongue Investigator, capable of following chemical trails with the robot tongue (Russell, 2006). As an early work in this field, the robot could collect nonvolatile chemicals (NaCl for experiments) on the ground by lowering and raising a sensor holding mechanism. It tasted the ionic compound by dissolving it in the distilled water in the dampened cotton tape, then followed trails by simple algorithms. However, it tastes an ionic compound concentration rather than distinguish between different compounds. Bianca Ciui et al. proposed a robot hand that could detect chemicals via fingertips (Ciui et al., 2018). The robot hand, equipped with a chemical-flavor sensing glove, was able to touch and sense the glucose, ascorbic acid and capsaicin constituents. Consequently, it discriminates sweetness with the middle finger, sourness with the index finger and spiciness with the ring finger, making it capable of discriminating a wide range of food samples.
5. Trends, difficulties and challenges of bioinspired sensors
As we mentioned, considerable progress has been made in the systematic research of the performance optimization, structure design, integration, flexibility and bionics of sensing devices (Ran et al., 2021). However, before these bioinspired sensing systems can be integrated into various intelligent robots, several critical elements will require thorough investigation.
5.1 Surpassed biological perception
Although various bioinspired sensors have been developed, their perception performances are yet to match their biological counterparts. With the advent of artificial intelligence, considerable strides have been made in integrating different artificial transduction technologies with AI or even biological materials, along with the use of appropriate tools for data processing. This has paved the way for a new concept of next-generation devices, which is expected to develop sensors that surpass human and biological perceptions. Such advancements in calculating and sensing could redefine how robots interact with the surrounding world, opening up a plethora of applications and possibilities.
5.2 Multifunction and integration
Combining multiple sensors and microprocessors allows for not just abundant sensing, but also the incorporation of artificial intelligence functionalities such as information processing, logical discrimination, self-diagnosis and cognitive processes (Dong et al., 2021). This kind of sensor array offers many advantages, such as versatility, high performance, small size and suitability for mass production and easy operation (Wu et al., 2016). However, there remain many challenges, including device performance degradation; multidimension; multistimulus crosstalk decoupling under simultaneous detection; achieving consistency in force, thermal and electrical properties among the components of the integrated sensing system; and algorithm requirements for high speed and low consumption.
5.3 Application of advanced materials
Materials form the primary basis of sensor performance and the improvement thereof. For instance, advancements in optical fibers and superconducting materials provided a material foundation for the development of new sensors (Campbell and Dincă, 2017). Many modern sensors are based on semiconductors, such as infrared sensors, laser sensors and fiber optic sensors (Lim et al., 2020). In addition, flexible sensors show great application potential in the future human–computer interaction systems, intelligent robots, mobile health care and other fields, which also rely heavily on the research and development of new materials (He et al., 2020; Pei et al., 2019), such as nanomaterials. In addition, exploring the use of neuromorphic materials to enable bioinspired sensing in energy-efficient and computationally powerful ways becomes a heat.
5.4 Easy application on robots
The applications of sensors on robots present various hardware and engineering demands, such as reliability, cost, manufacturability and economy, as illustrated in Figure 4. A distributed robot sensing system could comprise thousands of diverse sensors and associated electronic components, making wiring, power management, data processing and reliability significant concerns. Modularity is an established technique for organizing and simplifying such complex systems. Provisions must be included for fault tolerance, graceful degradation, self-diagnostics and self-healing. What is more, the viability of the robotic sensing system largely depends on manufacturability and maintenance. A system that is easy to manufacture, install and maintain will naturally have a competitive edge economically feasible.
5.5 Multimodal sensors fusion
By emulating how humans process signals from various senses to make decisions and adapt to the environment, it is expected that more robust and efficient systems can be devised. Research ideas in this field might include developing cross-modal algorithms to enhance perception by sharing multimodal sensor information; implementing a brain-inspired sensing fusion framework that integrates sensor information hierarchically at multiple abstraction levels to refine the fusion process; and adaptively adjusting the correlation and weights of different sensor modes based on bioinspired attention mechanism to further promote the intelligent fusion process.
6. Conclusion
The intelligent perception of robots hinges on the ability to acquire, process and comprehend a myriad of environmental information, with multimodal sensors serving as the foundational element. Humans are endowed with efficient senses (visual, auditory, tactile, olfactory and gustatory) to capture environmental information, or directly react to physical world stimuli to convey objective information. This information is transformed into neural electrical signals at the point of stimulation, which are subsequently relayed to the brain via neural networks, enabling comprehension and recognition of the environment. This biological process has spurred the emergence of a variety of biomimetic sensors, each designed to emulate this natural mechanism.
This work first briefly surveys the physiological principles of biological vision, audition, tactile, olfactory and taste senses, which serve as the basis for the sensing mechanisms of corresponding biomimetic sensors. The subsequent focus is on the robotic application of various biomimetic sensors, including autonomous positioning, environmental detection and autonomous navigation. The requirements and expectations of bioinspired sensors and their applications on robots are discussed at the end of this review. It is hoped that the survey proves beneficial to practitioners designing bioinspired sensing systems and to researchers working on intelligent robotics involving environment perception.
This work was supported in part by the National Natural Science Foundation of China (Nos U2013602, 62088101, 51975415, 52002286 and 61825303), in part by the National Key Research and Development Program of China (No. 2020AAA0108905), in part by the Science and Technology Commission of Shanghai Municipality (No. 2021SHZDZX0100, 22ZR1467100) and the Fundamental Research Funds for the Central Universities.
Declaration of competing interest: The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.




