Artificial intelligence assisted nanogenerator applications

Shumao Xu , Farid Manshaii , Xiao Xiao and Jun Chen *
Department of Bioengineering, University of California, Los Angeles, Los Angeles, CA 90095, USA. E-mail: jun.chen@ucla.edu

Received 7th October 2024 , Accepted 13th November 2024

First published on 14th November 2024


Abstract

Piezoelectric and triboelectric nanogenerators are at the forefront of converting ambient mechanical energy into electricity. These devices have experienced significant advancements in functionality and autonomy through integration with artificial intelligence (AI). This integration not only enhances their performance in autonomous operations by improving mechanical-to-electrical energy conversion efficiency, but also forges new pathways in robotics and intelligent systems. By increasing responsiveness and adaptability, these innovations expand the potential applications of nanogenerators. Looking ahead, the combination of nanogenerators with AI is poised to play a crucial role in developing sustainable, eco-friendly energy solutions. Their dual impact in advancing intelligent systems and promoting environmental sustainability signifies a significant milestone in nanogenerator technology for robotics. This review highlights the essential role of AI in refining nanogenerators, charting a path toward energy autonomy and sustainability.


1 Introduction

In the dynamically advancing field of renewable energy technologies, piezoelectric and triboelectric nanogenerators (PENGs and TENGs) have emerged as cutting-edge approaches for converting ambient mechanical energy into electricity.1–10 Since their inception more than a decade ago, these technologies have evolved rapidly, rekindling interest in sustainable energy amid a global energy crisis and concerns surrounding traditional fossil fuels. By leveraging piezoelectric and triboelectric effects, PENGs and TENGs can harvest energy from diverse environmental sources, including motion, temperature changes, and structural vibrations, outperforming conventional energy harvesters in terms of electrical efficiency.11–17

The efficacy of PENGs and TENGs is significantly influenced by their design and material composition, which presents both substantial opportunities and challenges for enhancing their electrical characteristics. This has spurred vigorous multidisciplinary research efforts focused on optimizing these technologies for a broad range of applications, including energy harvesting, sensing, monitoring, soft robotics, and electronic skins (e-skins).8,18–40

The integration of artificial intelligence (AI) with PENGs and TENGs has ushered in a novel phase in boosting their functionality.41,42 AI offers solutions to the challenges of developing portable, reliable, and eco-friendly energy sources.43–45 By mimicking human cognitive processes, AI significantly improves computational efficiency in the structural design and material selection for nanogenerators.46–48 Early efforts to integrate AI with PENGs and TENGs have shown promising results, addressing design, prediction, and optimization challenges49 and marking a pivotal shift from conventional statistical methods. AI excels at uncovering complex relationships among variables,23,50 greatly enhancing the exploration of design and material possibilities for these nanogenerators.

This synergy between AI and nanogenerators not only drives advancements in energy harvesting technologies but also heralds in a new era for robotics and intelligent systems. AI-powered nanogenerators enhance robotic perception, cognition, and interaction with their surroundings. This review aims to illuminate recent progress in nanogenerators, with a focus on the transformative impact of AI integration and its potential to revolutionize robotics and intelligent systems while paving the way for a more sustainable future.

2 Fundamentals of PENGs and TENGs

PENGs and TENGs are pioneering energy-harvesting technologies that convert mechanical energy from the environment into electricity. This operation is grounded in Maxwell's displacement current theory, with PENGs utilizing the piezoelectric effect and TENGs harnessing the triboelectric effect in combination with electrostatic induction.

Introduced in 2006, PENGs convert mechanical stress or deformation directly into electrical energy through the piezoelectric effect.51 This phenomenon is observed in materials that generate an electrical charge in response to mechanical stress. A typical PENG features a piezoelectric material sandwiched between two electrodes in a metal–insulator–metal configuration. When mechanically deformed, these structures produce a voltage that drives an electrical current through an external circuit. Materials such as ZnO and PVDF are commonly used to maximize the efficiency of this mechanical-to-electrical energy energy conversion.

TENGs, first developed in 2012, operate through the triboelectric effect,52 which involves generating electrical energy via contact and separation between two distinct materials. This interaction causes the surfaces to acquire opposite charges, and the relative motion between them creates an electrostatic potential that drives an electrical current through a connected load. TENGs can be engineered in various configurations, including vertical contact-separation, lateral sliding, single-electrode, and freestanding triboelectric layer modes, to accommodate diverse mechanical energy inputs. Enhancing TENG performance relies on selecting materials with highly contrasting triboelectric polarities to optimize charge transfer and improve electrical output.

3 Computation for nanogenerator integrations

AI-augmented nanogenerators introduce advanced capabilities to on-device and cloud-based systems by enabling continuous learning mechanisms, which allow these systems to adapt dynamically over time.53 Handling noisy data is a critical challenge for both platforms, as sensor readings from nanogenerators are often influenced by environmental factors and background interference.54 Robust statistical models, such as regression and resistant measures of central tendency, help maintain performance by down-weighting outliers and mitgating noise.55,56 These models provide a stable baseline by focusing on core data patterns, minimizing the influence of erratic points. Outlier detection methods, such as isolation forests and distance-based algorithms, further enhance model reliability in noisy environments by identifying and managing data points that deviate significantly from expected patterns.57

Data preprocessing and filtering steps can enhance data quality before inputting into models. For instance, low-pass filters effectively remove high-frequency noise, enhancing data fidelity.58 By integrating continuous learning capabilities with robust noise management techniques, AI-augmented nanogenerators achieve high accuracy, adaptability, and resilience across various applications.

3.1 On-device processing

On-device processing is essential for real-time applications like wearable monitoring, tactile sensing, and health management, where low latency is critical to provide immediate feedback, ensuring functionality and user satisfaction.59 In prosthetics and smart clothing for health monitoring, slight delays in response time can disrupt user control, coordination, and timely health alerts.60 Rapid on-device processing helps prevent these issues by supporting energy-intensive algorithms like Convolutional Neural Networks (CNNs) and Deep Neural Networks (DNNs) for image classification and predictive modeling,61 as well as less demanding algorithms like Artificial Neural Networks (ANNs), Deep Belief Networks (DBNs), and linear techniques such as Principal Component Analysis (PCA), Linear Classification Algorithms (LCA), and Linear Discriminant Analysis (LDA) (Fig. 1).
image file: d4ta07127a-f1.tif
Fig. 1 Machine learning models for AI-augmented nanogenerators. In supervised learning, models like k-Nearest Neighbors (KNN), Support Vector Machines (SVM), Artificial Neural Networks (ANNs), Linear Discriminant Analysis (LDA), Recurrent Neural Networks (RNNs), decision trees, and Convolutional Neural Networks (CNNs) are included. These models, trained on labeled data, are effective for tasks such as classification, pattern recognition, and real-time monitoring in energy-sensitive systems. In unsupervised learning, Deep Belief Networks (DBNs), Hopfield Neural Networks (HNNs), autoencoders (unsupervised CNNs), and Principal Component Analysis (PCA) operate without labeled data for clustering and dimensionality reduction. Energy efficiency strategies like pruning/compression and quantization can optimize these algorithms for low-power environments typical in wearables and IoT devices powered by nanogenerators. CNNs,124 copyright 2020, Springer Nature.

On-device datasets, such as biometric data, sensor readings, and activity logs, are processed in real time to provide immediate insights.62 Efficient algorithms prioritize responsiveness within resource constraints, making them ideal for real-time motor control and sensor feedback. These algorithms often operate in bursts, activating only during specific time windows when sufficient energy is available, thus conserving power.63 For example, Recurrent Neural Networks (RNNs) may analyze heart rate data intermittently, performing computations during peak energy generation.63 However, on-device systems are limited by processing power and energy efficiency, restricting the complexity of algorithms and potentially hindering tasks like complex image processing or extensive data analysis.64

In addition to energy efficiency, balancing energy consumption and decision accuracy is critical in energy-limited applications like wearables and Internet of Things (IoT) devices, where small batteries and low-latency responses are essential for effective user interaction.65 While wearables face strict energy constraints, larger IoT or robotic systems have fewer power limitations. Efficient algorithms are necessary to maintain high accuracy and reliability, particularly for real-time monitoring. Pruning,66 quantization,67 and lightweight models like MobileNets68 streamline neural networks by reducing complexity, memory footprint, and computational load, enhancing processing speed without significantly compromising accuracy.

Adaptive processing also adjusts model intensity based on task demands; for instance, RNNs, particularly Long Short-Term Memory (LSTM) networks, may activate only when vital signs deviate significantly, such as in abnormal heart rates, switching to simpler models under stable conditions to conserve energy.69,70 This approach extends battery life and efficiency in health monitoring while enabling quick responses to health changes.

Continuous health monitoring systems should ensure high accuracy in tracking vital signs, necessitating energy-efficient algorithms capable of sustaining long-term operation. In contrast, environmental sensors, which monitor slowly changing conditions, need only intermittent data collection, allowing for reduced power consumption.71 To maximize efficiency and precision, AI-augmented nanogenerators employ lightweight models for less critical tasks and shift to complex algorithms when necessary, optimizing energy use.72 Additionally, integrating energy harvesting and storage solutions, like nanogenerators and supercapacitors, buffer power supply for peak demands, ensuring consistent performance. Advances in nanogenerator energy conversion efficiency, alongside microbatteries and supercapacitors, stabilize power for intensive tasks, like deep learning and real-time analytics.73 Low-power hardware and application-specific integrated circuits (ASICs) enhance efficiency, supporting self-powered applications.74 By effectively combining these strategies, AI-augmented nanogenerators can maximize functionality while minimizing energy consumption, ultimately enhancing user experience across various applications.

3.2 Cloud-based processing

Cloud-based processing excels in handling computationally intensive, less latency-sensitive tasks, allowing extensive datasets to be offloaded for advanced data analysis and pattern recognition.75 Security and document management applications benefit from this, as tasks like data encryption and complex pattern analysis require high processing power but are not critically time-sensitive. A hybrid approach is often used, with on-device processing for immediate feedback and cloud processing for in-depth analysis, such as long-term trend detection in gait analysis.76 While latency-sensitive applications require fast response times, on-device processing can meet this need but may lack resources for complex analyses.77 In contrast, cloud solutions handle intricate tasks, including deep learning and extensive pattern recognition, though data transmission may introduce delays,78 as cloud speeds range from hundreds of milliseconds to seconds depending on network conditions and task complexity.79 Transmission rates vary widely, with 4G long-term evolution (LTE) networks typically offering 10–50 Mbps,80 while cloud tasks require higher bandwidth. On-device systems also conserve power to prolong battery life, while cloud servers, although more energy-intensive, support larger loads.81 Integrating AI with nanogenerators for on-device and cloud-based processing enables adaptive, intelligent systems, balancing real-time responsiveness with complex data capabilities.

4 AI-integrated nanogenerators

AI algorithms play a pivotal role in predicting and improving the efficiency, responsiveness, and adaptability of nanogenerators, paving the way for developments in robotics, self-powered sensing, energy-efficient actuation, and intelligent human–machine interactions.22,23,38,82–98

The implementation of AI encompasses supervised learning models (classification and regression) and unsupervised models (clustering and dimensionality reduction).99,100 Unsupervised learning models like DBNs to capture complex hierarchical data representations, and Hopfield Neural Networks (HNNs) to perform associative memory tasks, operate on unlabeled data to identify inherent structures or patterns without predefined outputs. In contrast, supervised learning, such as Support Vector Machines (SVMs) for classifying data by identifying the optimal hyperplane that separates different classes101 and decision trees to split data based on feature values for interpreting user activities,102 are trained on labeled data, where each input is paired with its corresponding output, allowing the algorithm to learn patterns and make predictions. ANNs are trained on labeled data for tasks like classification and regression,103 while LCA and LDA serve for dimensionality reduction and class separation, respectively.104 LCA uses linear decision boundaries to classify labeled data, and LDA identifies the linear combination of features that best separates classes, maximizing inter-class distance while minimizing within-class variance by leveraging the mean and variance of each class.

K-Nearest Neighbors (KNN) is a supervised, instance-based algorithm that classifies data points by proximity to labeled training instances,105 enabling rapid gesture recognition by comparing real-time sensor data with recorded movements for quick responses.47,106–109 CNNs, versatile in machine learning, are primarily used in supervised tasks like image classification, object detection, and segmentation,110 but can also function unsupervised, as in autoencoders, to learn data representations without labels. Similarly, RNNs are often used for supervised tasks like speech recognition and time-series prediction but can adapt to unsupervised tasks like sequence prediction,111 allowing for flexible application across diverse machine learning challenges. DNNs, composed of multiple interconnected layers, effectively learn complex data relationships, making them suitable for pattern recognition tasks and multidimensional sensor data processing for applications requiring high accuracy in classification and prediction.112,113

Fig. 2 illustrates the integration of AI and nanogenerators to enhance daily life, industrial monitoring, and advanced interactive platforms.114–116 In wearable technology, nanogenerators enable biometric and health monitoring by supporting advanced machine learning algorithms, such as SVMs, neural networks, and decision trees, for health tracking and activity analysis.14,37,47,117–121 Decision trees classify user activities by branching on feature values, while neural networks capture complex behavior patterns, and SVMs identify physiological signals (e.g., heart rate and skin temperature) for personalized monitoring by identifying the optimal hyperplane that separates different classes of data points.39,122,123 E-skins, which simulate touch, typically utilize CNNs to process spatial data, extracting features from tactile interactions to respond to stimuli accurately by applying convolutional layers that filter input data and pooling layers that reduce dimensionality.124–136 In environmental monitoring, nanogenerators paired with CNNs and SVMs enhance tasks like object recognition,137,138 liquid leakage detection,139,140 and gas sensing,47,141,142 with SVMs classifying environmental conditions and CNNs analyzing patterns for improved sensor detection.


image file: d4ta07127a-f2.tif
Fig. 2 Applications of AI-enhanced nanogenerators. AI-assisted nanogenerators provide responsive, efficient, and intelligent solutions in advancing human–machine interfaces. In wearable and portable applications, nanogenerators power biometric and health monitoring devices, such as (a) blood pressure monitors, (b) cardiac monitoring systems, and (c) user identification tools. For smart wearables, AI-augmented sensors enable (d) augmented reality/virtual reality (AR/VR) interfaces, (e) smart clothing,37 copyright 2022, Springer Nature; and (f) gesture-recognition glove,109 copyright 2020, Springer Nature; (g) tactile sensing glove,136 copyright 2019, Springer Nature; and (h) e-skin,40 copyright 2019, Elsevier. In environmental monitoring, (i) haptic interfaces and (j) blue energy harvesting,17 copyright 2017, Elsevier, represent advancements in sustainable interactions. In communication and interaction, nanogenerators facilitate (k) speech classification and (l) lip decoding,114 copyright 2022, Springer Nature. In advanced sensing and detection, AI-augmented nanogenerators improve capabilities for (m) object recognition,138 copyright 2020, Springer Nature; (n) liquid leakage detection, and (o) gas sensing, including the detection of volatile organic compounds (VOCs)142 copyright 2021, AAAS.

RNNs are well-suited for sequential data tasks where the order of data points is crucial, such as speech classification and lip decoding, by managing temporal dependencies in speech signals.63,69,109,114,143,144 However, standard RNNs struggle with retaining long-term information due to the vanishing gradient problem.145 This limitation leads to challenges in learning long-term dependencies for applications requiring sustained contextual awareness.

LSTMs, a specialized RNN type, address this by using memory cells to track behavior changes and predict future actions over extended periods,111 making them effective in user activity recognition and health monitoring applications.146,147 LSTMs enable continuous monitoring of vital signs like blood pressure with improved accuracy by maintaining important information across longer sequences, offering timely health updates essential for medical intervention.148–150 ANNs, consisting of interconnected layers of nodes, are also used for monitoring, including tasks like marine environmental surveillance and pressure mapping,103 as they can learn to identify patterns and make predictions based on diverse inputs to model complex relationships within data and handling intricate datasets.17,23,151–154 HNNs, another form of RNNs for associative memory, excel at pattern recognition by managing complex patterns with hierarchical structures derived from data, such as typing dynamics and tactile information.155 HNNs can retrieve stored patterns based on partial inputs, making them suitable for applications that require quick and reliable recognition of previously learned information, thereby improving user experience in systems that rely on human interaction.

DBNs, combining multiple layers of stochastic, latent variables to facilitate feature learning and classification,8 can capture hierarchical representations in high-dimensional data through unsupervised feature learning, useful in applications with unlabeled or poorly structured data. PCA aids in reducing the dimensionality of data from human interactions and typing dynamics, retaining key features while enhancing classification accuracy by transforming them into a smaller set of uncorrelated variables known as principal components.143

The interplay between nanogenerators and various machine learning algorithms marks a step toward intelligent systems that augment human capabilities and experiences. This integration can achieve improved accuracy, efficiency, and user interaction, involving a range of algorithms (Table 1), such as SVM for object and character recognition,156,157 KNN through instance-based learning for analyzing signal sequences and handwriting features,108 ANN for time-series analysis in sensory networks,23,151 CNN for image and pattern recognition,125,158–160 and RNN for sequential tasks such as speech processing.151

Table 1 Overview of algorithms with nanogenerator integrationa
Algorithm Roles Deployment Energy-efficiency Reusability Pros Cons
a Reusability: pretrained models or parts of the models can be reused for energy-efficient adaptation to new tasks.
PCA Dimensionality reduction for enhanced classification efficiency On-device High (relatively low data dimensionality) Limited Efficient for dimensionality reduction, improves data processing speed May lose critical information in reduced dimensions; sensitive to scaling variations
SVM Classifies data using hyperplanes; effective in high-dimensional spaces Cloud Moderate (memory for large datasets) Limited; often needs retraining for different data Effective in high-dimensional spaces; robust to overfitting, enhances classification accuracy High computational demand, may slow down real-time processing
RNNs Temporal data analysis, ideal for sequential data like time-series signals On-device Low (sequential processing) Adaptable, with limited retraining for different patterns Short sequential data and time-series analysis Computationally expensive, vanishing gradients, challenging for long sequences
LSTM Enhanced RNN capable of long-term dependencies; suitable for adaptive interfaces Cloud Low (memory cells) Partially reusable through fine-tuning for sequential tasks Effective for long-term dependencies, addresses vanishing gradient issues in RNNs Resource-intensive, requires high computational power for training and inference
CNNs Hierarchical feature extraction, especially effective in image and spatial feature recognition On-device/cloud Requires substantial power Transfer learning allows reuse for similar tasks Excellent at spatial data recognition, suitable for image and spatial feature extraction High computational cost; risk of overfitting in deep architectures without sufficient data
DNN Models complex relationships; versatile in complex pattern detection Cloud Low (memory requirements) Reusable with fine-tuning; transfer learning possible Capable of modeling complex patterns, versatile for diverse applications Requires substantial computational resources and large datasets; prone to overfitting in small data
DBN Unsupervised feature learning for enhanced classification accuracy Cloud Moderate (high computational load) Reusable with adjustments to layers; retains learned features Effective for hierarchical feature extraction, enhances classification accuracy Training is complex and time-consuming, demanding computational resources
KNN Instance-based learning, good for small datasets; suitable for recognition tasks On-device Low energy efficiency with large datasets; energy-efficient in small, targeted datasets Requires retraining for different classes; limited reus Simple and effective for small datasets, suitable for real-time applications Inefficient with large datasets; performance degrades with irrelevant features
ANN Models nonlinear relationships across diverse applications On-device Moderate; lightweight architectures viable Moderate reusability with layer-specific fine-tuning Highly accurate for complex relationships; versatile across applications Computationally demanding and requires large amounts of labeled data for training
HNN Associative memory tasks, handling complex hierarchical data patterns Cloud Low (relatively high cost and energy consumption) Good reusability; effective for memory-intensive tasks without full retraining Effective for associative memory, retrieves stored patterns from partial input High computational complexity, often requires cloud-based processing for real-time use


One key approach to reusing AI models without the need for full retraining is transfer learning, which involves utilizing pretrained models that have already learned general features from large datasets and fine-tuning them for new applications.161 For example, a health monitoring model can be fine-tuned for fitness tracking by modifying layers to target metrics like steps or heart rate.162 Another effective strategy is modular AI design, where specific components of a model can be reused across different applications.163 Foundational layers that have been trained on common features, such as motion patterns or general image features, can be repurposed for various tasks like differentiating between types of physical activity or detecting anomalies in sensor data. This modularity allows for significant savings in retraining efforts, enabling systems to adapt quickly to new challenges while maintaining efficiency. Moreover, distributed processing approaches where computational tasks are spread across multiple systems or devices, allow for handling more complex tasks while maintaining performance.164 In larger systems, this can involve cloud computing resources that complement local processing, enabling advanced data analytics and machine learning operations that exceed the capabilities of individual devices.

4.1 On-device AI-integrated TENGs for tactile sensing and recognition

Nanogenerators are pioneering a new era in artificial sensory systems and human–machine interfaces (HMIs) by creating self-powered, intelligent devices inspired by the biological sensory nervous system.165

Various AI models, such as CNNs, SVMs, and RNNs, are tailored to handle the data types in tactile sensing and gait analysis (Table 2). For gait analysis, 1D CNNs excel at processing sequential, time-series data along a single temporal dimension.166 These layers, adapted with broader strides, capture temporal dependencies within each gait cycle, supporting accurate classification of gait patterns for applications like sports training and rehabilitation. Pooling layers reduce the data size while retaining key features, ensuring efficiency. 2D CNNs are well-suited for tactile sensing, using convolutional layers to capture spatial relationships such as texture and shape.167 By pairing these layers with the max-pooling layer, which down-samples by retaining only the most prominent features, 2D CNNs reduce data dimensionality while preserving essential spatial features. Fine-tuning stride and kernel settings enhances localized sensitivity, enabling precise, high-resolution data processing for real-time, responsive interactions. In complex scenarios requiring analysis across spatial and temporal dimensions, like robotic tactile systems and environmental monitoring, 3D CNNs capture evolving three-dimensional patterns.168 When used in tandem with CNNs, SVMs enhance the tactile peripheral nervous system (TPNS) sensor's ability to classify structured data, recognizing high-dimensional patterns like pressure or texture variations by identifying optimal hyperplanes, making them ideal for nuanced recognition tasks.169 RNNs, particularly LSTMs, handle the sequential nature of tactile data by capturing short- and long-term dependencies, essential for interpreting touch sequences in real time, benefiting prosthetics and robotics.170

Table 2 Algorithmic requirements for various applications in AI-integrated nanogenerators
Application Algorithms Roles Example applications Data
Tactile sensing and recognition 2D CNNs (spatial) Spatial feature extraction E-skins (texture/pressure recognition) Tactile logs, pressure mappings
1D CNNs (temporal), KNN, RNN Sequential and instance-based learning Smart gloves (gesture recognition), AR/VR interfaces Gesture sequences, spatial motion data
1D CNNs (temporal), SVM, ANN, RNN Classification, pattern recognition, sequential data Wearable health monitors (heart rate, activity) Biometric data, sensor readings
SVMs (classification) Classifying touch or pressure types Haptic feedback systems Tactile data logs
LSTMs (sequential) Sequential data processing Adaptive human–machine interfaces Sequential tactile data
Security and document management SVM + PCA High-dimensional data classification Document authentication, user biometric verification Biometric data, document signatures
Decision trees Decision-making in workflows Document categorization, access control Encrypted signatures, security protocols
LDA Class separation Document type classification High-dimensional data
Pattern recognition and memory HNNs (associative memory) Pattern recognition with stored memory Typing behavior analysis, user authentication Sequential logs, sensory interaction patterns
DBNs (feature learning) Unsupervised feature extraction Sensory memory for robots, pattern detection Sensory data, hierarchical patterns
RNNs (sequential) Sequential pattern analysis Repetitive behavior recognition Time-series data
Autoencoders (unsupervised) Feature extraction without labels Data compression for sensor logs Unlabeled data logs
Advanced sensory detection CNNs (feature extraction) Feature extraction for detection tasks Object and sound recognition Environmental readings
RNNs (real-time analysis) Sequential data analysis Continuous auditory analysis Real-time sound patterns
PCA, DBN, HNN outlier detection Dimensionality reduction, feature learning, associative memory VOC detection, unusual pattern identification, air quality monitoring VOC sensor readings, environmental data logs


The integration of these AI models—2D CNNs for spatial data, 1D CNNs for temporal data, and SVMs and LSTMs for classification and sequential analysis—demonstrates the flexibility and adaptability of AI-augmented, nanogenerator-powered systems.

4.1.1 SVMs, 2D CNNs, and RNNs for a tactile peripheral nervous system. The TPNS sensor, powered by TENG-based nanogenerators, exemplifies the synergy of nanotechnology and machine learning in mimicking human tactile responses. By converting mechanical touch into electrical signals, TENGs enable real-time, nuanced tactile perception, enhanced through advanced machine learning algorithms like SVMs, CNNs, and RNNs, which classify and interpret complex tactile data with high accuracy.

SVMs analyze the structured, high-dimensional data (datasets with numerous features) from TENG sensors, including touch intensity, duration, and texture, by identifying optimal hyperplanes that separate tactile classes, such as different pressures and textures, with high precision.139,169,171 By maximizing the distance between data points from different classes, SVMs achieve high classification accuracy, making them useful for recognizing structured patterns in tactile data. This capability enriches the TPNS by distinguishing tactile stimuli in applications like prosthetics and robotics. Additionally, 2D CNNs work in tandem with SVMs to extract spatial hierarchies in tactile data.167 Through convolutional layers, CNNs capture local features like edges and textures, creating a hierarchical understanding of tactile inputs. As the data move through deeper layers of the CNN, increasingly complex spatial features are extracted, creating a hierarchical understanding of tactile input. Pooling layers further enhance this by downsampling while retaining essential details, enabling tasks such as identifying textures akin to braille172 (Fig. 3a). SVMs classify these patterns, ensuring high accuracy in differentiating tactile sensations.


image file: d4ta07127a-f3.tif
Fig. 3 AI-assisted TENGs for self-powered systems and interfaces. (a) A self-powered artificial tactile peripheral nervous system using TENG, simulating biological sensory circuits using CNN.172 Copyright 2021, Elsevier. (b) Self-powered handwriting HMI via TENG, with structure and recognition principles using CNN.187 Copyright 2020, Elsevier. (c) Smart socks with TENG sensors for recognizing sports gait data using CNN.195 Copyright 2020, Springer Nature. (d) Training and recognition process for a 16-key stretchable keyboard security system using SVM.202 Copyright 2018, Elsevier. (e) Self-powered TENG sensor for document management: page-turn recording and book theft prevention.205 Copyright 2014, Springer. (f) Handwriting recognition with self-powered TENG for machine learning-based user classification using SVM.207 Copyright 2020, Elsevier. (g and h) Electrical signal–time curve and feature radar chart for user typing, alongside a cross-user difference score matrix with feature combinations using PCA and SVM.202 Copyright 2018, Elsevier. (i) TENG matrix-based artificial sensory memory using HNN.143 Copyright 2020, Elsevier.

For temporal tactile data, the TPNS employs RNNs, specifically LSTM networks, to handle the dynamic nature of tactile data that unfolds over time.170 RNNs process sequential data by retaining information about previous inputs for interpreting patterns that change or evolve, while LSTM networks enhance this capability by incorporating memory cells that manage both short-term and long-term dependencies in data sequences, allowing the TPNS to analyze the occurrence of touch and its duration and intensity over time.173 In prosthetics and robotics, LSTMs support real-time adaptive responses, imitating human touch perception effectively.

TENGs generate power autonomously by converting mechanical touch into electricity, eliminating the need for external batteries and enhancing the sensor's eco-friendliness. This self-sustained energy generation supports continuous, real-time tactile processing, allowing the TPNS to deliver rapid, precise responses for applications in prosthetics and intelligent robotics that rely on immediate, reliable feedback to function effectively.174–186 These advancements position the TPNS as a transformative technology in future sensory networks, where energy-efficient, intelligent sensors can provide human-like tactile sensitivity in environmentally sustainable ways.

4.1.2 2D CNNs for handwriting human–machine interfaces. The development of TENG-powered handwriting HMIs demonstrates the integration of AI and nanogenerators for interactive applications. TENG-based touchpads convert mechanical handwriting energy into unique electrical signals, capturing the nuances of each user's style187 (Fig. 3b). Given the complex, multidimensional nature of these signals, CNNs play a key role in accurately processing and converting them.

In handwriting HMIs powered by TENGs, CNNs manage spatial data from the grid-like TENG sensor array, which records variations in pressure, texture, and shape.188,189 These spatial data form hierarchical levels of detail, with CNNs progressively abstracting and refining information from basic outlines to intricate handwriting characteristics through multiple convolutional and pooling layers.

Convolutional filters slide across the sensor grid, capturing fine details in pressure and contact that distinguish individual handwriting styles.167 As data pass through layers, CNNs build hierarchical representations, from basic shapes to complex strokes, allowing for high-resolution interpretation of handwriting. Max-pooling layers further streamline data by retaining prominent features like high-pressure zones and distinct edges, reducing noise and data complexity.190

Tactile inputs in flexible, wearable devices often vary due to changes in pressure, contact angle, or surface area, particularly when devices undergo bending or twisting. To accommodate these variations, CNNs employ fine-tuned stride settings and adaptive kernel sizes. Finer stride settings (smaller movement steps across the data) enhance the CNN's sensitivity to minor changes, capturing subtle differences in texture or pressure.191,192 Meanwhile, adaptive kernel sizes allow the model to focus on different scales of spatial features, adjusting its “field of view” to capture both fine details and broader strokes.193 This adaptability is crucial for applications requiring real-time feedback, allowing CNNs to maintain accuracy even under physical distortions.

The synergy of TENGs and CNNs enables real-time data processing and instant feedback in handwriting HMIs. TENGs generate continuous electrical signals from handwriting, which CNNs rapidly interpret, supporting applications like virtual keyboards and digital notetaking where seamless interaction is expected. This integration sets a benchmark in self-powered HMIs, balancing energy efficiency with adaptability to physical changes.

4.1.3 1D and lightweight CNNs for smart socks and sports gait analysis. In sports and personal fitness, integrating TENGs with AI in smart wearables, like smart socks and shoes, is revolutionizing gait analysis. These devices embed TENG sensors that harvest kinetic energy from foot movements, converting it into electrical signals that reflect gait characteristics, walking phases, and pressure distributions46,106,194,195 (Fig. 3c). The self-powered feature allows for long-term continuous, discrete monitoring, fitting seamlessly into daily wear. Gait details, like stride length, cadence, and weight distribution, are analyzed for insights into balance and stability.

Unlike tactile sensing, which relies on spatial data, gait analysis involves time-series data—a sequence of signals that represent movement patterns over time. 1D CNNs are well-suited as they can detect temporal dependencies across sequential movements, enabling the system to interpret complex gait cycles.166,194 1D convolutional layers slide filters across the time-series data, capturing recurrent patterns and subtle shifts in foot pressure over each step. This process enables the system to classify various gait types (e.g., normal walking and limping) by detecting recurring signal patterns that correspond to different types of movement. The use of broader strides in 1D CNNs allows the network to observe multiple time steps at once, capturing sequential relationships within the gait cycle that provide a complete, coherent view of movement patterns. This enhances classification accuracy by allowing the network to integrate and contextualize data over a sequence, rather than focusing solely on individual points in time.

Lightweight CNNs, which require minimal computational resources, further optimize on-device analysis, capturing essential features of foot pressure and movement for real-time feedback.196 Unlike typical CNNs, which demand substantial processing power, lightweight CNNs use fewer parameters and streamlined layer structures, making them ideal for energy-efficient wearable devices. They process data from embedded sensors that monitor foot pressure and movement across gait stages. This capability supports instant adjustments in sports training and rehabilitation, where small gait corrections enhance performance and reduce injury risk.

Real-time feedback is vital for high-performance sports and rehabilitation, enabling athletes, coaches, and therapists to make precise adjustments based on accurate gait data. CNN-enhanced wearables offer valuable insights, including stability, balance, and load distribution, critical for athletic performance, injury prevention, and recovery tracking. In personalized medicine, detailed gait analysis assists in diagnosing mobility issues and customizing rehabilitation exercises to an individual's movement patterns.197,198 This technology allows coaches to refine technique, monitor fatigue, and proactively address injury risks by identifying asymmetries in movement.199 In rehabilitation, CNNs provide precise progress tracking, allowing therapists to adjust treatment plans based on quantifiable, real-time feedback, fostering more effective recovery.200 By enabling data-driven decisions, AI-powered smart wearables optimize performance and minimize injury risks.201

4.2 Cloud-based AI-integrated TENGs for security and document management

4.2.1 SVM and PCA for high-dimensional document classification. The integration of nanogenerators with SVM and PCA is advancing energy-efficient solutions in user authentication, security, and document management. When implemented in devices like keyboards and touchpads, TENGs can capture individual typing dynamics such as rhythm, pressure, and speed—producing unique, high-dimensional data patterns for each user47,117,202 (Fig. 3d). SVM is suited for handling these high-dimensional data due to its ability to create an optimal hyperplane that maximally separates data classes, which correspond to distinct typing behaviors associated with each user. This separation ensures robust, precise user authentication, as SVM effectively classifies nuanced variations in typing patterns, making it difficult for unauthorized users to replicate.

PCA, on the other hand, plays a complementary role by simplifying high-dimensional datasets for efficient data processing and storage.203 PCA works by identifying the main components (features) of the data, reducing complexity while preserving the most significant patterns, thereby optimizing the data for classification by SVM. This combination of PCA and SVM improves processing efficiency and enhances classification accuracy, as PCA distills the data to its most relevant features before SVM performs classification. In cloud-supported systems, this setup allows for large-scale data analysis, making it a viable solution for secure user authentication and document management across distributed networks.

In document management applications such as library archives and secure facilities, this approach is beneficial for monitoring document interactions. Systems can capture electrical signals generated by handling documents and analyze these signals with SVM and PCA to track user interactions in real-time and prevent theft117,204,205 (Fig. 3e). This innovation surpasses traditional radio frequency identification-based systems, providing a smarter, more sustainable solution that leverages cloud capabilities for managing extensive, high-dimensional datasets.

4.2.2 Decision trees for workflow automation in document management. Decision trees are hierarchical models that make classifications or decisions based on a series of branching paths determined by specific criteria, making them ideal for workflow automation in document management.206 By mapping out logical decisions in a tree-like structure, decision trees can systematically evaluate different features of documents such as classification categories, access levels, and handling requirements without requiring manual input for every step. This systematic branching allows for rapid, consistent categorization of documents, ensuring efficient management and access control. In document storage and retrieval systems, a decision tree could automatically categorize documents by type, assign appropriate access permissions, or flag documents requiring special handling. This automation minimizes manual oversight, streamlining processes in secure environments where handling large volumes of documents is essential.

Handwriting recognition in document management captures an individual's unique writing style through sensors that detect variations in mechanical pressure exerted during writing207 (Fig. 3f). SVM classifies these distinct writing patterns for reliable identity verification, while PCA simplifies data complexity, making classification more efficient by focusing on key features of the handwriting. This application provides an additional layer of security in document management, as each user's handwriting is treated as a unique biometric signature that can be authenticated alongside other security protocols.

By combining SVM's precision in distinguishing high-dimensional data and PCA's efficiency in reducing data complexity, handwriting recognition systems offer secure, energy-efficient verification methods suitable for document management in high-security environments.178,208 This integration not only enhances user authentication but also extends the capabilities of document management by adding personalized, biometric-based security options, contributing to a more intelligent, responsive, and energy-efficient system for managing sensitive information.

4.3 Cloud-based AI-integrated TENGs for pattern recognition and memory

4.3.1 PCA and SVM for typing analysis and user authentication. The combination of TENGs with PCA and SVM algorithms has significantly advanced typing behavior analysis, achieving greater accuracy and efficiency. TENGs capture mechanical nuances of typing, such as signal magnitudes, latencies, and hold times, transforming these details into electrical signals that serve as robust datasets. PCA simplifies these high-dimensional data by reducing them to essential components, effectively isolating the core elements of typing patterns without unnecessary complexity202 (Fig. 3g and h). By distilling the data into their most informative aspects, PCA sharpens the focus for further analysis, improving computational efficiency.

Utilizing the refined data provided by PCA, SVM employs its classification capabilities to differentiate between individuals based on their unique typing signatures. SVM identifies an optimal hyperplane that separates data points across high-dimensional spaces, capturing the subtle distinctions in typing behavior that are critical for accurate user identification and authentication. The synergy between PCA's dimensionality reduction and SVM's high-precision classification allows for a robust, energy-efficient typing analysis system, supporting enhanced security and personalization in applications like user authentication and productivity analysis.

4.3.2 HNNs and DBNs for smart robotics and environmental interaction. HNNs play a pivotal role in developing artificial sensory memory for robotics, particularly in the recognition and recall of complex patterns.155 HNNs are associative memory networks that excel at pattern recognition and recall based on partial inputs, making them highly effective for sensory memory applications in robotics. In systems with TENG matrix sensors, which capture intricate tactile data from interactions, HNNs process and store these sensory patterns. This capability enables robots to dynamically respond to changes in texture, pressure, and other tactile stimuli by recognizing familiar patterns and recalling previous interactions209 (Fig. 3i). With their layered architecture, HNNs support a robust framework for recognizing and storing high-dimensional sensory data, allowing robots to adapt to various environmental stimuli with precision and memory recall.210

Complementing HNNs, DBNs and layered neural networks that learn hierarchical data patterns without the need for labeled inputs can be used for sensory memory systems where data often lack explicit labels through unsupervised feature learning, which is beneficial for high-dimensional tactile data.8 In cloud-based tactile memory applications, DBNs extract meaningful features directly from complex, multi-faceted TENG sensor data, including information on texture, pressure, and motion. As DBNs autonomously identify underlying structures and patterns within these data, they allow sensory systems to improve responsiveness and adapt over time.

The use of cloud infrastructure with DBNs supports the substantial computational requirements for processing high-dimensional tactile data. This infrastructure enables DBNs to efficiently manage and organize complex tactile information, facilitating pattern recognition and feature learning that enhances robotic systems' real-time interactions and memory functions. HNNs and DBNs provide a comprehensive framework for artificial sensory memory, allowing intelligent robotic systems to navigate, interpret, and recall tactile experiences with precision, thereby paving the way for responsive and autonomous interactions.

4.4 On-device/cloud-based AI-assisted advanced sensory detection

4.4.1 Cloud-integrated CNNs for object recognition in environmental monitoring. Cloud integration enhances the power of CNNs, enabling them to handle large-scale, high-dimensional datasets while benefiting from cloud-based computational resources. The cloud infrastructure continuously refines CNN models, improving accuracy and allowing them to scale effectively for diverse environmental applications. This setup supports real-time monitoring needs in data-intensive scenarios, as CNNs can quickly update and analyze large datasets from various sensors, providing precise and scalable insights essential for safety and environmental monitoring. The cloud-enabled architecture makes CNNs well-suited for large-scale environmental data analysis, where continuous updates and real-time object recognition play vital roles211,212 (Fig. 4a and b).
image file: d4ta07127a-f4.tif
Fig. 4 AI-assisted TENGs for sensory detection. (a) Angle sensor.211 Copyright 2020, Wiley. (b) Dual-mode vector motion sensors for direction and angle.212 Copyright 2022, Elsevier. (c) Ionizer with machine learning enhancement.8 Copyright 2021, RSC. (d) Stretchable sensor for object scanning, pressure measurement, and hardness detection.213 Copyright 2019, Springer Nature. (e) Neural tactile sensor for pressure detection and texture identification.214 Copyright 2019, ACS. (f) Artificial auditory pathway.215 Copyright 2020, Elsevier.

For environmental monitoring, CNNs integrated with cloud computing bring advanced capabilities for object recognition and data classification. CNNs are adept at detecting spatial patterns through their hierarchical structure, which processes data in layers to capture features ranging from simple edges to intricate shapes. This makes CNNs highly effective in identifying objects, patterns, and environmental features, such as recognizing volatile organic compounds (VOCs) or detecting specific items or hazards within complex landscapes8,142 (Fig. 4c).

4.4.2 RNNs for real-time sound analysis. AI-enhanced sensory detection systems that combine nanotechnology with RNNs are significantly advancing real-time sound analysis capabilities in robotics. Biomimetic tactile sensors, such as those powered by TENGs, autonomously respond to mechanical stimuli akin to human skin, allowing robots to detect and precisely process subtle changes in pressure and strain213 (Fig. 4d). An important development in this field is a self-powered neural tactile sensor that integrates a graphene layer with a TENG sensor featuring microlines, yielding heightened sensitivity to pressure and vibration214 (Fig. 4e).

RNNs offer distinct advantages for real-time sound analysis, making them highly effective in applications requiring continuous and rapid processing of audio data. Their unique architecture enables them to maintain context across time steps, making RNNs especially well-suited for handling sequential data like sound waves. This capacity to retain information allows RNNs to capture temporal patterns in sound, as they can integrate context from previous inputs—ideal for recognizing complex acoustic signatures that unfold over time. For example, variations in pitch, frequency, and duration in sound waves are effectively analyzed as the RNN processes each element in the sequence, updating its understanding based on previous patterns.

In real-time applications, RNNs operate directly on-device to minimize latency, enabling instant processing of acoustic data and allowing robotic systems to respond to auditory cues with minimal delay. This immediate responsiveness is particularly valuable in robotics, where quick sound recognition and reaction are crucial for navigation, interaction, and environmental awareness. By processing sound data as it arrives, RNNs enable robotic systems to differentiate sounds in dynamic environments—whether identifying speech, mechanical noise, or environmental cues in real-time—enhancing their adaptability, autonomy, and responsiveness to the surrounding world215 (Fig. 4f). Moreover, the introduction of biomimetic piezoelectric acoustic nanosensors marks a breakthrough in auditory fidelity. These sensors convert vibrations into electrical signals similar to the human hearing process, offering advanced sound detection capabilities216 (Fig. 5a).


image file: d4ta07127a-f5.tif
Fig. 5 AI-integrated PENGs for wearables and robotics. (a) Biomimetic inorganic piezoelectric acoustic nanosensor.216 Copyright 2014, Wiley. (b and c) Multi-sensing e-skin for pressure and temperature detection.236 Copyright 2018, ACS. (d) Output voltage and pressure for affordable e-skins.112 Copyright 2020, IOPscience. (e) Wearables inspired by the fish swim bladder.237 Copyright 2018, Wiley. (f) MoS2−x-based PENG under different bending speeds.241 Copyright 2018, Wiley. (g) Implanted microsystem in a robotic hand.213 Copyright 2019, Springer Nature.

These advancements in CNN-driven object recognition for environmental monitoring and RNN-enhanced real-time sound analysis mark a transformative step toward autonomous, responsive sensory systems in robotics. By merging AI with nanotechnology, these approaches enable real-time learning and adaptive functionality across various fields, including medical diagnostics, environmental monitoring, and smart infrastructure. The combination of on-device and cloud-supported AI bolsters robot interactions with their surroundings, enhancing pattern recognition, decision-making, and sensory precision, furthering the autonomy and intelligence of advanced robotics systems.

4.5 On-device AI-assisted PENG wearables for health monitoring

4.5.1 ANNs and RNNs for wearable health monitoring. The fusion of nanogenerators with AI-driven algorithms, specifically ANNs and RNNs, empowers wearable health monitors to continuously track health metrics in real time.42,118,209,217–232 ANNs process complex, nonlinear health data patterns by extracting key features from multiple sensor inputs, which is vital for comprehensive health monitoring. RNNs, on the other hand, excel at interpreting time-dependent data, making them ideal for sequential monitoring tasks, such as analyzing heart rate variability, breathing patterns, or prolonged physical activity. By retaining information over time, RNNs allow wearables to recognize evolving patterns in health signals, which is essential for detecting subtle changes that may indicate health risks.

Wearable devices that incorporate ZnO nanorod-based PENGs add an extra dimension to this functionality.233–235 These materials are highly responsive to motion, generating electrical signals in response to bending, stretching, and other mechanical movements. This sensitivity enables precise measurement of physiological movements and provides valuable input for ANN and RNN models to analyze, enhancing the quality and relevance of real-time health monitoring. This breakthrough lays the groundwork for creating highly sensitive e-skins236 (Fig. 5b and c). When coupled with AI, these devices can interpret complex motions, finding their place in advanced prosthetics and interactive technologies, thus improving the human–machine interface.

4.5.2 SVM, KNN, and decision trees for wearable tactile feedback. Advances in stretchable nanogenerators are revolutionizing wearable tactile feedback, making it possible for devices to deliver responsive sensations directly to users' fingertips112 (Fig. 5d). These PENG-based devices, designed for industrial-scale production, have transformed touch-sensitive wearables, enabling them to detect and respond to tactile inputs in real time. By integrating these nanogenerators with AI, wearable devices provide users with a more immersive, intuitive interaction, as they can interpret and respond to complex tactile stimuli. The development of bio-PENGs, which utilize eco-friendly materials to generate power from body movements, furthers this approach, supporting sustainable, self-powered wearables capable of adapting to various environmental and physiological cues237 (Fig. 5e).

In achieving this high degree of tactile feedback, several algorithms play complementary roles, each addressing distinct aspects of tactile data processing. SVMs are pivotal for classifying tactile data by clearly separating various touch inputs, such as different pressures or textures, into specific categories, allowing wearables to differentiate touch types accurately.238 KNN, a straightforward yet effective algorithm, is useful for classifying tactile sensations by comparing them to known data points. KNN's proximity-based approach is valuable for recognizing textures, pressure levels, or gestures, making it a natural complement to SVM in applications where cumulative sensory data inform tactile feedback.239 Decision trees and random forests can enhance tactile classification by evaluating input data based on multiple criteria.240 Decision trees individually assess sensory data, while random forests employ an ensemble of trees for more robust classification.

While SVMs offer foundational classification precision, KNN's historical comparison helps refine touch recognition by drawing on accumulated sensory data. Decision trees and random forests, with their multiple-layered decision-making, add robustness to tactile feedback by providing nuanced classification for a broad range of touch inputs. By dynamically interpreting and adapting to tactile inputs, these algorithms empower nanogenerator-based wearables to deliver advanced, real-time feedback, paving the way for intelligent, self-sufficient, and seamlessly integrated technologies in everyday life.

4.6 On-device/cloud-based AI-integrated PENG biomedical robotics

4.6.1 LSTM for precision control in robotic hands. The integration of LSTM networks with 3D piezoelectric microsystems in robotic hands marks a transformative advancement in biomedical robotics, particularly for cloud-supported applications requiring precision and adaptability. LSTMs are uniquely suited to capture long-term dependencies in sequential data, a critical feature for interpreting tactile inputs that continuously evolve. In robotic hand applications, these dependencies encompass complex tactile feedback patterns that arise as the hand interacts with diverse objects, surfaces, and textures. By retaining contextual information across time steps, LSTMs enable the robotic hand to make intelligent, real-time adjustments based on sequences of past interactions rather than isolated data points. This capability enhances the hand's sensitivity and adaptability, allowing for precise modulation of grip and pressure in response to varying tactile stimuli.

An example of this approach is demonstrated in robotic hands equipped with a monolayer MoS2−x-based PENG designed to optimize electrical outputs in flexible configurations241 (Fig. 5f) and a PVDF layer with chromium/gold electrodes mounted on a serpentine polyimide ribbon for sensitive tactile feedback213 (Fig. 5g). By dynamically interpreting tactile data through LSTM algorithms, these systems achieve enhanced sensitivity during manipulation tasks, allowing for highly controlled and responsive actions.

Moreover, AI elevates the functionality of these systems beyond passive data collection, fostering active learning and real-time adaptation. This advanced AI-driven approach utilizes pattern recognition, predictive analytics, and dynamic decision-making processes to boost sensor sensitivity and adaptability under various environmental conditions.242 For instance, AI's ability to dynamically adjust the electrical output of a sulfur vacancy-passivated MoS2 PENG in response to changes in pressure and bending angles significantly increases energy efficiency, optimizing system performance and extending operational lifespan.243

In the 3D piezoelectric microsystem integrated into a robotic hand, AI can significantly improve the hand's grip and manipulation abilities by continually learning from tactile feedback.244,245 This adaptability allows the robotic hand to perform a wide range of tasks with enhanced precision and responsiveness to touch, enabling it to handle diverse and unpredictable objects effectively. This fusion of AI with nanogenerator-powered biomedical robotics opens a myriad of possibilities, from self-adjusting prosthetic systems that provide a more seamless user experience to autonomous energy management within these systems, ensuring optimal power distribution. Such advancements are crucial for enhancing the autonomy of robotic and biomedical applications, making them more efficient and user-friendly.

4.6.2 ANN for real-time decision-making in adaptive prosthetics. In adaptive prosthetics, the integration of ANNs on-device enables real-time responsiveness, which is essential for seamless and intuitive user experiences. Unlike LSTMs, which are more suitable for handling sequential dependencies over time, ANNs provide immediate decision-making by analyzing tactile feedback and positional data in real time. This capability allows the prosthetic to react instantly to changes in movement, enabling fluid, adaptive motion that closely aligns with the user's natural actions.

By processing tactile inputs on-device, ANNs eliminate the latency associated with cloud processing, ensuring that the prosthetic can adjust grip, orientation, and movement as the user performs various tasks. This real-time processing is particularly useful for adaptive prosthetic hands, where instantaneous feedback is needed to accommodate changes in grip strength and angle based on the detected force or shape of the object being handled. Additionally, by optimizing response based on real-time inputs, ANNs reduce wear and tear on the device by avoiding unnecessary movements, thereby improving the lifespan and efficiency of the prosthetic system. The PENG-based design, with materials such as sulfur vacancy-passivated MoS2, enables these prosthetics to harvest energy autonomously, further enhancing their usability and sustainability.

These innovations in AI-integrated PENG biomedical robotics, through cloud-enabled LSTM applications for robotic hands and on-device ANN applications for adaptive prosthetics, mark a significant step towards creating self-sufficient, highly adaptive systems in healthcare. The combination of precise tactile feedback processing and intelligent adaptability enables these systems to provide a more personalized, responsive experience for users, with applications that span both medical and robotic fields.

5 Conclusions and outlook

The fusion of artificial intelligence with nanogenerators is transforming the conversion of mechanical energy into electrical energy, leading to more efficient and sustainable energy solutions. This breakthrough is paving the way for the creation of self-powered devices, crucial for advancing autonomous robotics and intelligent systems independent of external power sources. Moreover, AI's integration is refining the accuracy and responsiveness of robotics sensors and actuators, significantly improving healthcare wearables for precise monitoring of vital signs and enabling real-time data analytics. These advancements promote predictive maintenance and adaptive learning, tailoring devices to user-specific requirements.

However, this integration presents several challenges, including complexities in analysis, design, fabrication, and the broad application of these technologies. There is an urgent need for a universal framework that can thoroughly understand the piezoelectric and triboelectric effects, optimize structures, innovate new materials, effectively scale current outputs, and expand the application domains. This advancement should extend beyond sensing functions to include computational capabilities as well.

Ongoing research offers promising directions for incorporating active materials that can respond dynamically to external stimuli, such as pressure, heat, or electrical fields into nanogenerator-powered soft robotics and intelligent machine systems. These materials are useful in robotic systems requiring adaptive responses, such as soft sensors, e-skins, actuators, and microfluidic devices, where flexibility and responsiveness are essential.

Combining TENGs and PENGs within a single system can offer dual capabilities in both sensing and actuation. This combination leverages TENGs' sensitivity to surface contact and movement and PENGs' ability to generate stable output under deformation, creating a more comprehensive energy-harvesting and sensory network. This integration may be accomplished by layering materials or designing hybrid circuits where each generator type contributes to specific functions. These hybrid systems support the development of self-powered, untethered robotic devices capable of responding to environmental cues.

For soft robotics, low-power actuation mechanisms are crucial as they allow for motion without requiring high energy input. Further development of electroactive polymers, dielectric elastomers, and shape-memory alloys enables movements like bending, contracting or expanding with minimal electrical power for soft actuators in robots. When paired with TENG or PENG devices, these materials support sustained operation in autonomous systems that rely on limited power sources.

Microscale TENG- and PENG-based robots, due to their compact size, hold significant potential for biomedical applications within confined spaces, such as vascular systems. These miniature devices could navigate narrow blood vessels or other internal pathways to perform diagnostics, deliver targeted drugs, or even conduct minor repairs within the body. Biocompatible materials, such as hydrogels and flexible polymers, enable safe interactions with biological tissues, while advanced microfabrication techniques like photolithography, 3D printing, and laser micromachining allow for the precise construction of robots at the microscale, essential for delicate medical applications. In the development of flexible tactile sensors, innovations in materials science and structural design have led to substantial improvements in sensitivity. Flexible substrates, such as elastomers and conductive polymers, combined with structured micro/nano-patterns on the surface, enhance the device's ability to detect and differentiate between subtle pressure changes, temperature variations, and other material properties. These advances also support enhanced auditory and olfactory sensing, enabling more precise sound and gas detection, respectively, in soft robotics and wearable devices.

Moving forward, key areas for nanogenerator-based systems include addressing performance limitations, ensuring stability in challenging conditions, and advancing production techniques for scalability. Strategies to enhance output performance include using multilayered structures and optimizing the surface area to boost energy capture. Ensuring device stability in harsh environments may involve material coatings that protect against high temperature, humidity, or mechanical stress, as well as encapsulation techniques to shield sensitive components. Scaling production can benefit from automated fabrication techniques, such as roll-to-roll processing or 3D printing, to achieve consistent, high-quality manufacturing at a large scale.

Addressing these challenges will be essential for advancing the practical applications of nanogenerator systems. By enhancing durability, efficiency, and scalability, the potential of AI-enhanced nanogenerators can be unlocked across diverse fields, from medical devices and biosensors to self-powering buildings, ultimately marking a substantial step toward sustainable, autonomous technological solutions.

Data availability

No primary research results, software or code have been included and no new data were generated or analyzed as part of this review.

Author contributions

The manuscript was written through the contributions of all authors. All authors have approved the final version of the manuscript.

Conflicts of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgements

This is an invited submission for the Nanogenerators themed collection. The authors acknowledge the Henry Samueli School of Engineering & Applied Science and the Department of Bioengineering at the University of California, Los Angeles for their startup support. J. C. acknowledges the Vernroy Makoto Watanabe Excellence in Research Award at the UCLA Samueli School of Engineering, the Office of Naval Research Young Investigator Award (Award ID: N00014-24-1-2065), National Science Foundation Grant (Award Number: 2425858), NIH Grant (Award ID: R01 CA287326), the American Heart Association Innovative Project Award (Award ID: 23IPA1054908), the American Heart Association Transformational Project Award (Award ID: 23TPA1141360), the American Heart Association's Second Century Early Faculty Independence Award (Award ID: 23SCEFIA1157587), the Brain & Behavior Research Foundation Young Investigator Grant (Grant Number: 30944), and the NIH National Center for Advancing Translational Science UCLA CTSI (Grant Number: KL2TR001882).

References

  1. W. Deng, Y. Zhou, A. Libanori, G. Chen, W. Yang and J. Chen, Piezoelectric nanogenerators for personalized healthcare, Chem. Soc. Rev., 2022, 51, 3380–3435 RSC .
  2. Y. Su, W. Li, L. Yuan, C. Chen, H. Pan, G. Xie, G. Conta, S. Ferrier, X. Zhao, G. Chen, H. Tai, Y. Jiang and J. Chen, Piezoelectric fiber composites with polydopamine interfacial layer for self-powered wearable biomonitoring, Nano Energy, 2021, 89, 106321 CrossRef CAS .
  3. Y. Su, C. Chen, H. Pan, Y. Yang, G. Chen, X. Zhao, W. Li, Q. Gong, G. Xie, Y. Zhou, S. Zhang, H. Tai, Y. Jiang and J. Chen, Muscle fibers inspired high-performance piezoelectric textiles for wearable physiological monitoring, Adv. Funct. Mater., 2021, 31, 2010962 CrossRef CAS .
  4. Y. Zou, V. Raveendran and J. Chen, Wearable triboelectric nanogenerators for biomechanical energy harvesting, Nano Energy, 2020, 77, 105303 CrossRef CAS .
  5. J. Xu, Y. Zou, A. Nashalian and J. Chen, Leverage surface chemistry for high-performance triboelectric nanogenerators, Front. Chem., 2020, 8, 577327 CrossRef CAS PubMed .
  6. G. Conta, A. Libanori, T. Tat, G. Chen and J. Chen, Triboelectric nanogenerators for therapeutic electrical stimulation, Adv. Mater., 2021, 33, 2007502 CrossRef CAS .
  7. J. Chen, Y. Huang, N. Zhang, H. Zou, R. Liu, C. Tao, X. Fan and Z. L. Wang, Micro-cable structured textile for simultaneously harvesting solar and mechanical energy, Nat. Energy, 2016, 1, 16138 CrossRef CAS .
  8. C. Huang, G. Chen, A. Nashalian and J. Chen, Advances in self-powered chemical sensing via a triboelectric nanogenerator, Nanoscale, 2021, 13, 2065–2081 RSC .
  9. L. Jin, X. Xiao, W. Deng, A. Nashalian, D. He, V. Raveendran, C. Yan, H. Su, X. Chu, T. Yang, W. Li, W. Yang and J. Chen, Manipulating relative permittivity for high-performance wearable triboelectric nanogenerators, Nano Lett., 2020, 20, 6404–6411 CrossRef CAS PubMed .
  10. Y. Zou, J. Xu, Y. Fang, X. Zhao, Y. Zhou and J. Chen, A hand-driven portable triboelectric nanogenerator using whirligig spinning dynamics, Nano Energy, 2021, 83, 105845 CrossRef CAS .
  11. J. Yin, S. Wang, A. Di Carlo, A. Chang, X. Wan, J. Xu, X. Xiao and J. Chen, Smart textiles for self-powered biomonitoring, Med-X, 2023, 1, 3 CrossRef .
  12. N. Sharma, X. Wan, Z. Duan, X. Xiao, J. Yin, Y. Du and J. Chen, Harvesting low-grade wind energy from highways using a triboelectric nanogenerator, Nano Energy, 2024, 132, 110354 CrossRef .
  13. Z. Zhou, L. Weng, T. Tat, A. Libanori, Z. Lin, L. Ge, J. Yang and J. Chen, Smart insole for robust wearable biomechanical energy harvesting in harsh environments, ACS Nano, 2020, 14, 14126–14133 CrossRef CAS PubMed .
  14. G. Chen, Y. Li, M. Bick and J. Chen, Smart textiles for electricity generation, Chem. Rev., 2020, 120, 3668–3720 CrossRef CAS PubMed .
  15. Y. Zhou, W. Deng, J. Xu and J. Chen, Engineering materials at the nanoscale for triboelectric nanogenerators, Cell Rep. Phys. Sci., 2020, 1, 100142 CrossRef CAS .
  16. Q. Chen, W. Akram, Y. Cao, C. Ge, T. Lin and J. Fang, Recent progress in the fabrication and processing of triboelectric yarns, Carbon Neutralization, 2023, 2, 63–89 CrossRef CAS .
  17. Z. L. Wang, T. Jiang and L. Xu, Toward the blue energy dream by triboelectric nanogenerator networks, Nano Energy, 2017, 39, 9–23 CrossRef CAS .
  18. W. Deng, Y. Zhou, X. Zhao, S. Zhang, Y. Zou, J. Xu, M.-H. Yeh, H. Guo and J. Chen, Ternary electrification layered architecture for high-performance triboelectric nanogenerators, ACS Nano, 2020, 14, 9050–9058 CrossRef CAS PubMed .
  19. B. Zhang, F. Chun, G. Chen, T. Yang, A. Libanori, K. Chen, G. Conta, D. Xiong, C. Yan, W. Yang and J. Chen, Water evaporation induced intermolecular force for nano-wrinkled polymeric membrane, Cell Rep. Phys. Sci., 2021, 2, 100441 CrossRef CAS .
  20. C. Yan, Y. Gao, S. Zhao, S. Zhang, Y. Zhou, W. Deng, Z. Li, G. Jiang, L. Jin, G. Tian, T. Yang, X. Chu, D. Xiong, Z. Wang, Y. Li, W. Yang and J. Chen, A linear-to-rotary hybrid nanogenerator for high-performance wearable biomechanical energy harvesting, Nano Energy, 2020, 67, 104235 CrossRef CAS .
  21. X. Zhao, H. Askari and J. Chen, Nanogenerators for smart cities in the era of 5G and Internet of Things, Joule, 2021, 5, 1391–1431 CrossRef .
  22. Y. Zou, A. Libanori, J. Xu, A. Nashalian and J. Chen, Triboelectric nanogenerators enabled smart shoes for wearable electricity generation, Research, 2020, 2020, 7158953 CrossRef CAS PubMed .
  23. X. Xiao, G. Chen, A. Libanori and J. Chen, Wearable triboelectric nanogenerators for therapeutics, Trends Chem., 2021, 3, 279–290 CrossRef CAS .
  24. S. Shen, X. Xiao, X. Xiao and J. Chen, Triboelectric nanogenerators for self-powered breath monitoring, ACS Appl. Energy Mater., 2022, 5, 3952–3965 CrossRef CAS .
  25. Y. Zou, J. Xu, K. Chen and J. Chen, Advances in nanostructures for high-performance triboelectric nanogenerators, Adv. Mater. Technol., 2021, 6, 2000916 CrossRef CAS .
  26. S. Zhang, M. Bick, X. Xiao, G. Chen, A. Nashalian and J. Chen, Leveraging triboelectric nanogenerators for bioengineering, Matter, 2021, 4, 845–887 CrossRef CAS .
  27. T. Tat, A. Libanori, C. Au, A. Yau and J. Chen, Advances in triboelectric nanogenerators for biomedical sensing, Biosens. Bioelectron., 2021, 171, 112714 CrossRef CAS .
  28. X. Li, T. Tat and J. Chen, Triboelectric nanogenerators for self-powered drug delivery, Trends Chem., 2021, 3, 765–778 CrossRef CAS .
  29. S. Shen, X. Xiao and J. Chen, Wearable triboelectric nanogenerators for heart rate monitoring, Chem. Commun., 2021, 57, 5871–5879 RSC .
  30. G. Chen, C. Au and J. Chen, Textile triboelectric nanogenerators for wearable pulse wave monitoring, Trends Biotechnol., 2021, 39, 1078–1092 CrossRef CAS PubMed .
  31. J. Lama, A. Yau, G. Chen, A. Sivakumar, X. Zhao and J. Chen, Textile triboelectric nanogenerators for self-powered biomonitoring, J. Mater. Chem. A, 2021, 9, 19149–19178 RSC .
  32. Y. Su, G. Chen, C. Chen, Q. Gong, G. Xie, M. Yao, H. Tai, Y. Jiang and J. Chen, Self-powered respiration monitoring enabled by a triboelectric nanogenerator, Adv. Mater., 2021, 33, 2101262 CrossRef CAS .
  33. N. Zhang, C. Tao, X. Fan and J. Chen, Progress in triboelectric nanogenerators as self-powered smart sensors, J. Mater. Res., 2017, 32, 1628–1646 CrossRef CAS .
  34. S. Xu, F. Manshaii, X. Xiao, J. Yin and J. Chen, Triboelectric nanogenerators for self-powered neurostimulation, Nano Res., 2024, 17, 8926–8941 CrossRef .
  35. X. Zhao, Y. Zhou, Y. Song, J. Xu, J. Li, T. Tat, G. Chen, S. Li and J. Chen, Permanent fluidic magnets for liquid bioelectronics, Nat. Mater., 2024, 23, 703–710 CrossRef CAS PubMed .
  36. X. Wan, T. Tat, Y. Zhou, X. Zhao and J. Chen, Building bioelectronic fibres with a light touch, Nat. Electron., 2024, 7, 521–522 CrossRef .
  37. A. Libanori, G. Chen, X. Zhao, Y. Zhou and J. Chen, Smart textiles for personalized healthcare, Nat. Electron., 2022, 5, 142–156 CrossRef CAS .
  38. Y. Zhou, X. Zhao, J. Xu, Y. Fang, G. Chen, Y. Song, S. Li and J. Chen, Giant magnetoelastic effect in soft systems for bioelectronics, Nat. Mater., 2021, 20, 1670–1676 CrossRef CAS PubMed .
  39. X. Zhao, Y. Zhou, J. Xu, G. Chen, Y. Fang, T. Tat, X. Xiao, Y. Song, S. Li and J. Chen, Soft fibers with magnetoelasticity for wearable electronics, Nat. Commun., 2021, 12, 6755 CrossRef CAS PubMed .
  40. X. Xiao, X. Xiao, A. Libanori, Y. Fang, X. Li and J. Chen, Triboelectric nanogenerators for self-powered wound healing, Adv. Healthcare Mater., 2021, 10(20), 2100975 CrossRef CAS PubMed .
  41. Y. Xiong, Z. Huo, J. Zhang, Y. Liu, D. Yue, N. Xu, R. Gu, L. Wei, L. Luo and M. Chen, Triboelectric in-sensor deep learning for self-powered gesture recognition toward multifunctional rescue tasks, Nano Energy, 2024, 124, 109465 CrossRef CAS .
  42. X. Xiao, J. Yin, J. Xu, T. Tat and J. Chen, Advances in machine learning for wearable sensors, ACS Nano, 2024, 18, 22734–22751 CrossRef CAS PubMed .
  43. X. Zhao, Y. Zhou, W. Kwak, A. Li, S. Wang, M. Dallenger, S. Chen, Y. Zhang, A. Lium and J. Chen, A reconfigurable and conformal liquid sensor for ambulatory cardiac monitoring, Nat. Commun., 2024, 15, 8492 CrossRef CAS PubMed .
  44. X. Zhao, Y. Zhou, A. Li, J. Xu, S. Karjagi, E. Hahm, L. Rulloda, J. Li, J. Hollister, P. Kavehpour and J. Chen, A self-filtering liquid acoustic sensor for voice recognition, Nat. Electron., 2024, 7, 924–932 CrossRef .
  45. G. Chen, X. Zhao, S. Andalib, J. Xu, Y. Zhou, T. Tat, K. Lin and J. Chen, Discovering giant magnetoelasticity in soft matter for electronics textiles, Matter, 2021, 4, 3725–3740 CrossRef CAS PubMed .
  46. S. Divya, S. Panda, S. Hajra, R. Jeyaraj, A. Paul, S. H. Park, H. J. Kim and T. H. Oh, Smart data processing for energy harvesting systems using artificial intelligence, Nano Energy, 2023, 106, 108084 CrossRef CAS .
  47. Z. Tian, J. Li, L. Liu, H. Wu, X. Hu, M. Xie, Y. Zhu, X. Chen and W. Ou-Yang, Machine learning-assisted self-powered intelligent sensing systems based on triboelectricity, Nano Energy, 2023, 113, 108559 CrossRef CAS .
  48. K. Meng, X. Xiao, W. Wei, G. Chen, A. Nashalian, S. Shen, X. Xiao and J. Chen, Wearable pressure sensors for pulse wave monitoring, Adv. Mater., 2022, 34, 2109357 CrossRef CAS .
  49. Y. C. Lai, H. W. Lu, H. M. Wu, D. Zhang, J. Yang, J. Ma, M. Shamsi, V. Vallem and M. D. Dickey, Elastic multifunctional liquid–metal fibers for harvesting mechanical and electromagnetic energy and as self-powered sensors, Adv. Energy Mater., 2021, 11, 2100411 CrossRef CAS .
  50. T. Tat, G. Chen, X. Zhao, Y. Zhou, J. Xu and J. Chen, Smart textiles for healthcare and sustainability, ACS Nano, 2022, 16, 13301–13313 CrossRef CAS .
  51. X.-J. Dong, G. Meng and J.-C. Peng, Vibration control of piezoelectric smart structures based on system identification technique: numerical simulation and experimental study, J. Sound Vib., 2006, 297, 680–693 CrossRef .
  52. X.-S. Zhang, M.-D. Han, B. Meng and H.-X. Zhang, High performance triboelectric nanogenerators based on large-scale mass-fabrication technologies, Nano Energy, 2015, 11, 304–322 CrossRef CAS .
  53. C. Yang, Y. Wang, S. Lan, L. Wang, W. Shen and G. Q. Huang, Cloud-edge-device collaboration mechanisms of deep learning models for smart robots in mass personalization, Robot. Comput. Integrated Manuf., 2022, 77, 102351 CrossRef .
  54. J. Yin, S. Wang, T. Tat and J. Chen, Motion artefact management for soft bioelectronics, Nat. Rev. Bioeng., 2024, 2, 541–558 CrossRef CAS .
  55. N. Huyghues-Beaufond, S. Tindemans, P. Falugi, M. Sun and G. Strbac, Robust and automatic data cleansing method for short-term load forecasting of distribution feeders, Appl. Energy, 2020, 261, 114405 CrossRef .
  56. D. Karimi, H. Dou, S. K. Warfield and A. Gholipour, Deep learning with noisy labels: exploring techniques and remedies in medical image analtics, Med. Image Anal., 2020, 65, 101759 CrossRef .
  57. A. Mensi, D. M. Tax and M. Bicego, Detecting outliers from pairwise proximities: proximity isolation forests, Pattern Recognit., 2023, 138, 109334 CrossRef .
  58. J. Yang, J. Chen, Y. Su, Q. Jing, Z. Li, F. Yi, X. Wen, Z. Wang and Z. L. Wang, Eardrum-inspired active sensors for self-powered cardiovascular system characterization and throat-attached anti-interference voice recognition, Adv. Mater., 2015, 27, 1316 CrossRef CAS PubMed .
  59. C. Avasalcai, C. Tsigkanos and S. Dustdar, Resource management for latency-sensitive IoT applications with satisfiability, IEEE Trans. Serv. Comput., 2021, 15, 2982–2993 Search PubMed .
  60. C. Cipriani, F. Zaccone, S. Micera and M. C. Carrozza, On the shared control of an EMG-controlled prosthetic hand: analysis of user–prosthesis interaction, IEEE Trans. Robot., 2008, 24, 170–184 Search PubMed .
  61. Z. Li, F. Liu, W. Yang, S. Peng and J. Zhou, A survey of convolutional neural network: analysis, applications, and prospects, IEEE Trans. Neural Netw. Learn. Syst., 2021, 33, 6999–7019 Search PubMed .
  62. P. M. S. Sánchez, J. M. J. Valero, A. H. Celdrán, G. Bovet, M. G. Pérez and G. M. Pérez, A survey on device behavior fingerprinting: data sources, techniques, application scenarios, and datasets, IEEE Commun. Surv. Tutor., 2021, 23, 1048–1077 Search PubMed .
  63. D. Durstewitz, G. Koppe and M. I. Thurm, Reconstructing computational system dynamics from neural data with recurrent neural network, Nat. Rev. Neurosci., 2023, 24, 693–710 CrossRef CAS .
  64. A. Alnoman, G. H. Carvalho, A. Anpalagan and I. Woungang, Energy efficiency on fully cloudified mobile networks: survey, challenges, and open issues, IEEE Commun. Surv. Tutor., 2017, 20, 1271–1291 Search PubMed .
  65. F. J. Dian, R. Vahidnia and A. Rahmati, Wearables and the Internet of Things (IoT), applications, opportunities, and challenges: a Survey, IEEE Access, 2020, 8, 69200–69211 Search PubMed .
  66. Y. He and L. Xiao, Structured pruning for deep convolutional neural network: a survey, IEEE Trans. Pattern Anal. Mach. Intell., 2023, 46, 2900–2919 Search PubMed .
  67. T. Liang, J. Glossner, L. Wang, S. Shi and X. Zhang, Pruning and quantization for deep neural network acceleration: a survey, Neurocomputing, 2021, 461, 370–403 CrossRef .
  68. U. Kulkarni, S. Meena, S. V. Gurlahosur and G. Bhogar, Quantization friendly mobilenet (qf-mobilenet) architecture for vision based applications on embedded platforms, Neural Netw., 2021, 136, 28–39 CrossRef .
  69. S. Ishaque, N. Khan and S. Krishnan, Trends in heart-rate variability signal analysis, Front. Digit. Health, 2021, 3, 639444 CrossRef PubMed .
  70. N. Vinay, K. Vidyasagar, S. Rohith, P. Dayananda, S. Supreeth and S. Bharathi, An RNN-Bi LSTM based multi decision GAN approach for the recognition of cardiovascular disease (CVD) from heart beat sound: a feature optimization process, IEEE Access, 2024, 12, 65482–65502 Search PubMed .
  71. P. Corke, T. Wark, R. Jurdak, W. Hu, P. Valencia and D. Moore, Environmental wireless sensor networks, Proc. IEEE, 2010, 98, 1903–1917 Search PubMed .
  72. W. Mao, Z. Zhao, Z. Chang, G. Min and W. Gao, Energy-efficient industrial Internet of Things: overview and open issues, IEEE Trans. Ind. Inf., 2021, 17, 7225–7237 Search PubMed .
  73. M. K. Kiptoo, M. E. Lotfy, O. B. Adewuyi, A. Conteh, A. M. Howlader and T. Senjyu, Integrated approach for optimal techno-economic planning for high renewable energy-based isolated microgrid considering cost of energy storage and demand response strategies, Energy Convers. Manage., 2020, 215, 112917 CrossRef .
  74. K. S. Zaman, M. B. I. Reaz, S. H. M. Ali, A. A. A. Bakar and M. E. H. Chowdhury, Custom hardware architectures for deep learning on portable devices: a review, IEEE Trans. Neural Netw. Learn. Syst., 2021, 33, 6068–6088 Search PubMed .
  75. A. K. Sandhu, Big data with cloud computing: discussions and challenges, Big Data Min. Anal., 2021, 5, 32–40 Search PubMed .
  76. B. Kar, W. Yahya, Y.-D. Lin and A. Ali, Offloading using traditional optimization and machine learning in federated cloud–edge–fog systems: a survey, IEEE Commun. Surv. Tutor., 2023, 25, 1199–1226 Search PubMed .
  77. C. Wang, E. Shirzaei Sani and W. Gao, Wearable bioelectronics for chronic wound management, Adv. Funct. Mater., 2022, 32, 2111022 CrossRef CAS PubMed .
  78. T. Chen and R. Bahsoon, Self-adaptive trade-off decision making for autoscaling cloud-based services, IEEE Trans. Serv. Comput., 2015, 10, 618–632 Search PubMed .
  79. J. Chen and X. Ran, Deep learning with edge computing: a review, Proc. IEEE, 2019, 107, 1655–1674 Search PubMed .
  80. I. F. Akyildiz, D. M. Gutierrez-Estevez, R. Balakrishnan and E. Chavarria-Reyes, LTE-advanced and the evolution to beyond 4G (B4G) systems, Phys. Commun., 2014, 10, 31–60 CrossRef .
  81. P. Cong, J. Zhou, L. Li, K. Cao, T. Wei and K. Li, A survey of hierarchical energy optimization for mobile edge computing: a perspective from end devices to the cloud, ACM Comput. Surv., 2020, 53, 1–44 Search PubMed .
  82. J. Xu, J. Yin, Y. Fang, X. Xiao, Y. Zou, S. Wang and J. Chen, Deep learning assisted ternary electrification layered triboelectric membrane sensor for self-powered home security, Nano Energy, 2023, 113, 108524 CrossRef CAS .
  83. Z. Zhou, S. Padgett, Z. Cai, G. Conta, Y. Wu, Q. He, S. Zhang, C. Sun, J. Liu and E. Fan, Single-layered ultra-soft washable smart textiles for all-around ballistocardiograph, respiration, and posture monitoring during sleep, Biosens. Bioelectron., 2020, 155, 112064 CrossRef CAS .
  84. Y. Su, T. Yang, X. Zhao, Z. Cai, G. Chen, M. Yao, K. Chen, M. Bick, J. Wang and S. Li, A wireless energy transmission enabled wearable active acetone biosensor for non-invasive prediabetes diagnosis, Nano Energy, 2020, 74, 104941 CrossRef CAS .
  85. Y. Su, J. Wang, B. Wang, T. Yang, B. Yang, G. Xie, Y. Zhou, S. Zhang, H. Tai and Z. Cai, Alveolus-inspired active membrane sensors for self-powered wearable chemical sensing and breath analysis, ACS Nano, 2020, 14, 6067–6075 CrossRef CAS PubMed .
  86. Q. Xu, Y. Fang, Q. Jing, N. Hu, K. Lin, Y. Pan, L. Xu, H. Gao, M. Yuan and L. Chu, A portable triboelectric spirometer for wireless pulmonary function monitoring, Biosens. Bioelectron., 2021, 187, 113329 CrossRef CAS .
  87. Z. Lin, G. Zhang, X. Xiao, C. Au, Y. Zhou, C. Sun, Z. Zhou, R. Yan, E. Fan and S. Si, A personalized acoustic interface for wearable human–machine interaction, Adv. Funct. Mater., 2022, 32, 2109430 CrossRef CAS .
  88. Y. Fang, Y. Zou, J. Xu, G. Chen, Y. Zhou, W. Deng, X. Zhao, M. Roustaei, T. K. Hsiai and J. Chen, Ambulatory cardiovascular monitoring via a machine-learning-assisted textile triboelectric sensor, Adv. Mater., 2021, 33, 2104178 CrossRef CAS .
  89. Y. Fang, J. Xu, X. Xiao, Y. Zou, X. Zhao, Y. Zhou and J. Chen, A deep-learning-assisted on-mask sensor network for adaptive respiratory monitoring, Adv. Mater., 2022, 34, 2200252 CrossRef CAS PubMed .
  90. J. Yin, V. Kashyap, S. Wang, X. Xiao, T. Tat and J. Chen, Self-powered eye-computer interaction via a triboelectric nanogenerator, Device, 2024, 2, 100252 CrossRef .
  91. X. Wang, X. Xiao, Z. Feng, Y. Wu, J. Yang and J. Chen, A soft bioelectronic patch for simultaneous respiratory and cardiovascular monitoring, Adv. Healthcare Mater., 2024, 13, 2303479 CrossRef CAS .
  92. X. Xiao, Y. Zhou, J. Yin, X. Zhao and J. Chen, In-ear electrophysichochemical sensing, Nat. Biomed. Eng., 2023, 7, 1207–1209 CrossRef PubMed .
  93. H. Cui, S. Zhao and G. Hong, Wireless deep-brain neuromodulation using photovoltaics in the second near-infrared spectrum, Device, 2023, 1, 100113 CrossRef PubMed .
  94. G. Chen, Y. Fang, X. Zhao, T. Tat and J. Chen, Textiles for learning tactile interactions, Nat. Electron., 2021, 4, 175–176 CrossRef .
  95. L. Shan, H. Zeng, Y. Liu, X. Zhang, E. Li, R. Yu, Y. Hu, T. Guo and H. Chen, Artificial tactile sensing system with photoelectric output for high accuracy haptic texture recognition and parallel information processing, Nano Lett., 2022, 22, 7275–7283 CrossRef CAS PubMed .
  96. P. Zhang, W. Li, C. Liu, F. Qin, Y. Lu, M. Qin and Y. Hou, Molecular imaging of tumour-associated pathological biomarkers with smart nanoprobe: from “seeing” to “measuring”, Exploration, 2023, 3, 20230070 CrossRef CAS .
  97. A. G. Yearley, R. V. Patel, S. E. Blitz, S. Park, A. M. Madinger, J. Li, B. R. Johnston, P. P. Peruzzi, S. Lee and S. S. Srinivasan, Implantable microdevices for treating brain tumors, Device, 2023, 1, 100068 CrossRef .
  98. M. Brennan, Wearable technology is a proving ground for device engineering and design, Device, 2023, 1, 100096 CrossRef .
  99. P. K. Mallapragada, R. Jin, A. K. Jain and Y. Liu, Semiboost: boosting for semi-supervised learning, IEEE Trans. Pattern Anal. Mach. Intell., 2008, 31, 2000–2014 Search PubMed .
  100. S. B. Kotsiantis, I. D. Zaharakis and P. E. Pintelas, Machine learning: a review of classification and combining techniques, Artif. Intell. Rev., 2006, 26, 159–190 CrossRef .
  101. F. Melgani and L. Bruzzone, Classification of hyperspectral remote sensing images with support vector machines, IEEE Trans. Geosci. Rem. Sens., 2004, 42, 1778–1790 Search PubMed .
  102. M. Pal and P. M. Mather, An assessment of the effectiveness of decision tree methods for land cover classification, Rem. Sens. Environ., 2003, 86, 554–565 CrossRef .
  103. M. Khashei, A. Z. Hamadani and M. Bijari, A novel hybrid classification model of artificial neural network and multiple linear regression models, Expert Syst. Appl., 2012, 39, 2606–2620 CrossRef .
  104. S. Zhao, B. Zhang, J. Yang, J. Zhou and Y. Xu, Linear discriminant analysis, Nat. Rev. Methods Primers, 2024, 4, 70 CrossRef CAS .
  105. L. Xiong and Y. Yao, Study on an adaptive thermal comfort model with K-nearest-neighbors (KNN) algorithm, Build. Environ., 2021, 202, 108026 CrossRef .
  106. J. Piskozub and P. Strumillo, Reducing the number of sensors in the data glove for recognition of static hand gestures, Appl. Sci., 2022, 12, 7388 CrossRef CAS .
  107. H. Guo, J. Wan, H. Wang, H. Wu, C. Xu, L. Miao, M. Han and H. Zhang, Self-powered intelligent human-machine interaction for handwriting recognition, Research, 2021, 2021, 4689869 Search PubMed .
  108. Y. Liu, Y. Shen, W. Ding, X. Zhang, W. Tian, S. Yang, B. Hui and K. Zhang, All-natural phyllosilicate-polysaccharide triboelectric sensor for machine learning-assisted human motion prediction, npj Flexible Electron., 2023, 7, 21 CrossRef .
  109. Z. Zhou, K. Chen, X. Li, S. Zhang, Y. Wu, Y. Zhou, K. Meng, C. Sun, Q. He and W. Fan, Sign-to-speech translation using machine-learning-assisted stretchable sensor arrays, Nat. Electron., 2020, 3, 571–578 CrossRef .
  110. T. Kattenborn, J. Leitloff, F. Schiefer and S. Hinz, Review on convolutional neural network (CNN) in vegetation remote sensing, ISPRS J. Photogrammetry Remote Sens., 2021, 173, 24–49 CrossRef .
  111. A. Sherstinsky, Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network, Phys. D, 2020, 404, 132306 CrossRef .
  112. W. Zhang, H. Yang, L. Li, S. Lin, P. Ji, C. Hu, D. Zhang and Y. Xi, Flexible piezoelectric nanogenerators based on a CdS nanowall for self-powered sensors, Nanotechnology, 2020, 31, 385401 CrossRef CAS PubMed .
  113. G. Armeniakos, G. Zervakis, D. Soudris and J. Henkel, Hardware approximate techniques for deep neural network accelerators: a survey, ACM Comput. Surv., 2022, 55, 1–36 CrossRef .
  114. Y. Lu, H. Tian, J. Cheng, F. Zhu, B. Liu, S. Wei, L. Ji and Z. L. Wang, Decoding lip language using triboelectric sensors with deep learning, Nat. Commun., 2022, 13, 1401 CrossRef CAS .
  115. R. Yuan, A. Kumar, S. Zhuang, N. Cucciniello, T. Lu, D. Xue, A. Penn, A. R. Mazza, Q. Jia and Y. Liu, Machine learning-enabled superior energy storage in ferroelectric films with a slush-like polar state, Nano Lett., 2023, 23, 4807–4814 CrossRef CAS PubMed .
  116. K. K. Chan, L.-W. Shang, Z. Qiao, Y. Liao, M. Kim and Y.-C. Chen, Monitoring Amyloidogenesis with a 3D Deep-Learning-Guided Biolaser Imaging Array, Nano Lett., 2022, 22, 8949–8956 CrossRef CAS PubMed .
  117. J. Zhang, Y. Yu, L. Zhang, J. Chen, X. Wang and X. Wang, Dig information of nanogenerators by machine learning, Nano Energy, 2023, 108656 CrossRef CAS .
  118. S. Xiang, G. Chen, Q. Wen, H. Li, X. Luo, J. Zhong, S. Shen, A. Di Carlo, X. Fan and J. Chen, Fully addressable textile sensor array for self-powered haptic interfacing, Matter, 2024, 7, 82–94 CrossRef .
  119. G. Chen, X. Xiao, X. Zhao, T. Tat, M. Bick and J. Chen, Electronic textiles for wearable point-of-care systems, Chem. Rev., 2021, 122, 3259–3291 CrossRef PubMed .
  120. Y. Fang, G. Chen, M. Bick and J. Chen, Smart textiles for personalized thermoregulation, Chem. Soc. Rev., 2021, 50, 9357–9374 RSC .
  121. X. Xiao, X. Xiao, Y. Zhou, X. Zhao, G. Chen, Z. Liu, Z. Wang, C. Lu, M. Hu and A. Nashalian, An ultrathin rechargeable solid-state zinc ion fiber battery for electronic textiles, Sci. Adv., 2021, 7, eabl3742 CrossRef CAS .
  122. P. Maharjan, K. Shrestha, T. Bhatta, H. Cho, C. Park, M. Salauddin, M. T. Rahman, S. S. Rana, S. Lee and J. Y. Park, Keystroke dynamics based hybrid nanogenerators for biometric authentication and identification using artificial intelligence, Adv. Sci., 2021, 8, 2100711 CrossRef .
  123. X. Ji, T. Zhao, X. Zhao, X. Lu and T. Li, Triboelectric nanogenerator based smart electronics via machine learning, Adv. Mater. Technol., 2020, 5, 1900921 CrossRef CAS .
  124. F. Wang, A. Eljarrat, J. Müller, T. R. Henninen, R. Erni and C. T. Koch, Multi-resolution convolutional neural networks for inverse problems, Sci. Rep., 2020, 10, 5730 CrossRef CAS PubMed .
  125. Y. Yang, Q. Shi, Z. Zhang, X. Shan, B. Salam and C. Lee, Robust triboelectric information-mat enhanced by multi-modality deep learning for smart home, InfoMat, 2023, 5, e12360 CrossRef .
  126. S. Wu, H. Kan, J. Gao, W. Yue, C. Zhang, S. Gao and Y. Li, Convolutional neural network-motivated high-performance multi-functional electronic skin for intelligent human-computer interaction, Nano Energy, 2024, 122, 109313 CrossRef CAS .
  127. K. Tao, J. Yu, J. Zhang, A. Bao, H. Hu, T. Ye, Q. Ding, Y. Wang, H. Lin and J. Wu, Deep-learning enabled active biomimetic multifunctional hydrogel electronic skin, ACS Nano, 2023, 17, 16160–16173 CrossRef CAS PubMed .
  128. Y. Zhou, X. Zhao, J. Xu, G. Chen, T. Tat, J. Li and J. Chen, A multimodal magnetoelastic artificial skin for underwater haptic sensing, Sci. Adv., 2024, 10, eadj8567 CrossRef PubMed .
  129. S. Xu, M. Farid, G. Chen and J. Chen, Reversible metal–ligand coordination for photocontrolled metallopolymer adhesives, Chem, 2024, 10, 1638–1640 CAS .
  130. S. Xu, K. Scott and J. Chen, A bioinspired three-dimensionally architected electronic skin, Device, 2024, 2, 100462 CrossRef .
  131. P. Guo, J. Zhang, H. Pu, B. Yang, C. Huang, T. Sun, L. Li, X. Liu, Z. Guo and Y. Wu, Wafer-scale photolithographic fabrication of organic synaptic transistor arrays, Device, 2024, 2, 100409 CrossRef .
  132. D. Gao, J. P. Lee, J. Chen, L. S. Tay, Y. Xin, K. Parida, M. W. M. Tan, P. Huang, K. H. Kong and P. S. Lee, A wearable pneumatic-piezoelectric system for quantitative assessment of skeletomuscular biomechanics, Device, 2024, 2, 100288 CrossRef .
  133. T. Tat, J. Xu and J. Chen, Protocol for preparation of a textile magnetoelastic generator patch, STAR Protoc., 2024, 5(3), 103289 CrossRef CAS PubMed .
  134. Y. Chen, Z. Gao, F. Zhang, Z. Wen and X. Sun, Recent progress in self-powered multifunctional e-skin for advanced applications, Exploration, 2022, 2, 20210112 CrossRef PubMed .
  135. D. Wang, X.-B. Li and H.-B. Sun, Modulation doping: a strategy for 2D materials electronics, Nano Lett., 2021, 21, 6298–6303 CrossRef CAS .
  136. S. Sundaram, P. Kellnhofer, Y. Li, J.-Y. Zhu, A. Torralba and W. Matusik, Learning the signatures of the human grasp using a scalable tactile glove, Nature, 2019, 569, 698–702 CrossRef CAS PubMed .
  137. X. Zhong, P. Sun, R. Wei, H. Dong and S. Jiang, Object recognition by a heat-resistant core-sheath triboelectric nanogenerator sensor, J. Mater. Chem. A, 2022, 10, 15080–15088 RSC .
  138. T. Jin, Z. Sun, L. Li, Q. Zhang, M. Zhu, Z. Zhang, G. Yuan, T. Chen, Y. Tian and X. Hou, Triboelectric nanogenerator sensors for soft robotics aiming at digital twin applications, Nat. Commun., 2020, 11, 5381 CrossRef CAS PubMed .
  139. R. Li, D. Wei and Z. Wang, Synergizing Machine Learning Algorithm with Triboelectric Nanogenerators for Advanced Self-Powered Sensing Systems, Nanomaterials, 2024, 14, 165 CrossRef CAS .
  140. T. Zhan, H. Zou, H. Zhang, P. He, Z. Liu, J. Chen, M. He, Y. Zhang and Z. L. Wang, Smart liquid-piston based triboelectric nanogenerator sensor for real-time monitoring of fluid status, Nano Energy, 2023, 111, 108419 CrossRef CAS .
  141. D. Wang, D. Zhang, X. Chen, H. Zhang, M. Tang and J. Wang, Multifunctional respiration-driven triboelectric nanogenerator for self-powered detection of formaldehyde in exhaled gas and respiratory behavior, Nano Energy, 2022, 102, 107711 CrossRef CAS .
  142. J. Tian, F. Wang, Y. Ding, R. Lei, Y. Shi, X. Tao, S. Li, Y. Yang and X. Chen, Self-powered room-temperature ethanol sensor based on brush-shaped triboelectric nanogenerator, Research, 2021, 2021, 8564780 CrossRef CAS .
  143. Y. Liu, W. Yang, Y. Yan, X. Wu, X. Wang, Y. Zhou, Y. Hu, H. Chen and T. Guo, Self-powered high-sensitivity sensory memory actuated by triboelectric sensory receptor for real-time neuromorphic computing, Nano Energy, 2020, 75, 104930 CrossRef CAS .
  144. S. P. Padhy and S. V. Kalinin, Domain hyper-languages bring robots together and enable the machine learning community, Device, 2023, 1, 100115 CrossRef .
  145. T. He, H. Mao and Z. Yi, Subtraction gates: another way to learn long-term dependencies in recurrent neural network, IEEE Trans. Neural Netw. Learn. Syst., 2020, 33, 1740–1751 Search PubMed .
  146. J. Li, Z. Xie, Z. Wang, Z. Lin, C. Lu, Z. Zhao, Y. Jin, J. Yin, S. Mu and C. Zhang, A triboelectric gait sensor system for human activity recognition and user identification, Nano Energy, 2023, 112, 108473 CrossRef CAS .
  147. H. Zhang, H. Tan, W. Wang, Z. Li, F. Chen, X. Jiang, X. Lu, Y. Hu, L. Li and J. Zhang, Real-time non-driving behavior recognition using deep learning-assisted triboelectric sensors in conditionally automated driving, Adv. Funct. Mater., 2023, 33, 2210580 CrossRef CAS .
  148. L. Zhao, S. Y. Wong, J. Y. Sim, J. Zhou, X. Li and C. Wang, Triboelectric nanogenerators and piezoelectric nanogenerators for preventing and treating heart diseases, BMEMat, 2023, 1, e12020 CrossRef .
  149. S. Xu, F. Manshaii and J. Chen, Multiphasic interfaces enabled aero-elastic capacitive pressure sensor, Matter, 2024, 7, 2351–2354 CrossRef CAS .
  150. S. Xu, K. Scott, F. Manshaii and J. Chen, Heart-brain connection: How can heartbeats shape our minds?, Matter, 2024, 7, 1684–1687 CrossRef .
  151. Q. Zhang, T. Jin, J. Cai, L. Xu, T. He, T. Wang, Y. Tian, L. Li, Y. Peng and C. Lee, Wearable triboelectric sensors enabled gait analysis and waist motion capture for IoT-based smart healthcare applications, Adv. Sci., 2022, 9, 2103694 CrossRef CAS .
  152. P. Munirathinam, A. A. Mathew, V. Shanmugasundaram, V. Vivekananthan, Y. Purusothaman, S.-J. Kim and A. Chandrasekhar, A comprehensive review on triboelectric nanogenerators based on real-time applications in energy harvesting and self-powered sensing, Mater. Sci. Eng. B, 2023, 297, 116762 CrossRef CAS .
  153. T. Wang, Z. Song, X. Zhao, Y. Wu, L. Wu, A. Haghparast and H. Wu, Spatial transcriptomic analysis of the mouse brain following chronic social defeat stress, Exploration, 2023, 3, 20220133 CrossRef .
  154. J.-S. Lim, H. J. Kim, I. Park, S. Woo, J.-H. Kim and J. W. Park, Force mapping reveals the spatial distribution of individual proteins in a neuron, Nano Lett., 2022, 22, 3865–3871 CrossRef CAS PubMed .
  155. J. Wang, X. Zhuge and F. Zhuge, Hybrid oxide brain-inspired neuromorphic devices for hardware implementation of artificial intelligence, Sci. Technol. Adv. Mater., 2021, 22, 326–344 CrossRef .
  156. Q.-L. Goh, P.-S. Chee, E.-H. Lim and D. W.-K. Ng, An AI-assisted and self-powered smart robotic gripper based on eco-egain nanocomposite for pick-and-place operation, Nanomaterials, 2022, 12, 1317 CrossRef CAS PubMed .
  157. G. Ye, Y. Wan, J. Wu, W. Zhuang, Z. Zhou, T. Jin, J. Zi, D. Zhang, X. Geng and P. Yang, Multifunctional device integrating dual-temperature regulator for outdoor personal thermal comfort and triboelectric nanogenerator for self-powered human-machine interaction, Nano Energy, 2022, 97, 107148 CrossRef CAS .
  158. X. Hou, L. Zhang, Y. Su, G. Gao, Y. Liu, Z. Na, Q. Xu, T. Ding, L. Xiao and L. Li, A space crawling robotic bio-paw (SCRBP) enabled by triboelectric sensors for surface identification, Nano Energy, 2023, 105, 108013 CrossRef CAS .
  159. J. Huang, S. Wang, X. Zhao, W. Zhang, Z. Chen, R. Liu, P. Li, H. Li and C. Gui, Fabrication of a textile-based triboelectric nanogenerator toward high-efficiency energy harvesting and material recognition, Mater. Horiz., 2023, 10, 3840–3853 RSC .
  160. J. Guo, Y. Liu, L. Lin, S. Li, J. Cai, J. Chen, W. Huang, Y. Lin and J. Xu, Chromatic plasmonic polarizer-based synapse for all-optical convolutional neural network, Nano Lett., 2023, 23, 9651–9656 CrossRef .
  161. F. Zhuang, Z. Qi, K. Duan, D. Xi, Y. Zhu, H. Zhu, H. Xiong and Q. He, A comprehensive survey on transfer learning, Proc. IEEE, 2020, 109, 43–76 Search PubMed .
  162. A. Raza, K. P. Tran, L. Koehl and S. Li, Designing ECG monitoring healthcare system with federated transfer learning and explainable AI, Knowl.-Based Syst., 2022, 236, 107763 CrossRef .
  163. H. Abbas, F. Garberson, S. Liu-Mayo, E. Glover and D. P. Wall, Multi-modular AI approach to streamline autism diagnosis in young children, Sci. Rep., 2020, 10, 5014 CrossRef CAS .
  164. J. Wang, C. Xu, J. Zhang and R. Zhong, Big data analytics for intelligent manufacture systems: a review, J. Manuf. Syst., 2022, 62, 738–752 CrossRef .
  165. Y. Zhou, X. Xiao, G. Chen, X. Zhao and J. Chen, Self-powered sensing technologies for human Metaverse interfacing, Joule, 2022, 6, 1381–1389 CrossRef .
  166. A. O. Ige and M. Sibiya, State-of-the-art in 1D convolutional neural network: a survey, IEEE Access, 2024, 144082–144105 Search PubMed .
  167. M. Meribout, N. A. Takele, O. Derege, N. Rifiki, M. El Khalil, V. Tiwari and J. Zhong, Tactile sensors: a review, Measurement, 2024, 238, 115332 CrossRef .
  168. H. Wu, X. Ma and Y. Li, Spatiotemporal multimodal learning with 3D CNNs for video action recognition, IEEE Trans. Circ. Syst. Video Technol., 2021, 32, 1250–1261 Search PubMed .
  169. Z. Hu, L. Lin, W. Lin, Y. Xu, X. Xia, Z. Peng, Z. Sun and Z. Wang, Machine learning for tactile perception: advancements, challenges, and opportunities, Adv. Intell. Syst., 2023, 5, 2200371 CrossRef .
  170. M. Alameh, Y. Abbass, A. Ibrahim, G. Moser and M. Valle, Touch modality classification using recurrent neural network, IEEE Sens. J., 2021, 21, 9983–9993 Search PubMed .
  171. J. Park, Y. Lee, S. Cho, A. Choe, J. Yeom, Y. G. Ro, J. Kim, D.-h. Kang, S. Lee and H. Ko, Soft sensors and actuators for wearable human–machine interfaces, Chem. Rev., 2024, 124, 1464–1534 CrossRef CAS PubMed .
  172. L. Chen, C. Wen, S.-L. Zhang, Z. L. Wang and Z.-B. Zhang, Artificial tactile peripheral nervous system supported by self-powered transducers, Nano Energy, 2021, 82, 105680 CrossRef CAS .
  173. C. Wan, M. Pei, K. Shi, H. Cui, H. Long, L. Qiao, Q. Xing and Q. Wan, Toward a brain-neuromorphics interface, Adv. Mater., 2024, 36, 2311288 CrossRef CAS PubMed .
  174. R. K. Rajaboina, U. K. Khanapuram, V. Vivekananthan, G. Khandelwal, S. Potu, A. Babu, N. Madathil, M. Velpula and P. Kodali, Crystalline porous material-based nanogenerators: recent progress, applications, challenges, and opportunities, Small, 2024, 20, 2306209 CrossRef CAS PubMed .
  175. E. Elsanadidy, I. M. Mosa, D. Luo, X. Xiao, J. Chen, Z. L. Wang and J. F. Rusling, Advances in triboelectric nanogenerators for self-powered neuromodulation, Adv. Funct. Mater., 2023, 33, 2211177 CrossRef CAS .
  176. J. Chen and Z. L. Wang, Reviving vibration energy harvesting and self-powered sensing by a triboelectric nanogenerator, Joule, 2017, 1, 480 CrossRef CAS .
  177. Z. L. Wang, J. Chen and L. Lin, Progress in triboelectric nanogenerators as a new energy technology and self-powered sensors, Energy Environ. Sci., 2015, 8, 2250 RSC .
  178. G. Zhu, J. Chen, T. Zhang, Q. Jing and Z. L. Wang, Radial-arrayed rotary electrification for high performance triboelectric generator, Nat. Commun., 2014, 5, 3426 CrossRef PubMed .
  179. N. Zhang, F. Huang, S. Zhao, X. Lv, Y. Zhou, S. Xiang, S. Xu, Y. Li, G. Chen and C. Tao, Photo-rechargeable fabrics as sustainable and robust power sources for wearable bioelectronics, Matter, 2020, 2, 1260–1269 CrossRef .
  180. Y. Zhou, X. Zhao, Y. Song, J. Xu, J. Li, T. Tat, G. Chen, S. Li and J. Chen, Permanent Fluidic Magnets for Liquid Bioelectronics, Nat. Mater., 2024, 23, 703–710 CrossRef .
  181. J. Chen, G. Zhu, W. Yang, Q. Jing, P. Bai and Y. Yang, Harmonic-resonator-based triboelectric nanogenerator as a sustainable power source and a self-powered active vibration sensor, Adv. Mater., 2013, 25(42), 6094–6099 CrossRef CAS .
  182. S. Xu, G. Chen, K. Scott, F. Manshaii and J. Chen, Soft electrochemical actuators for intraoperative nerve activity monitoring, Matter, 2024, 7, 2795–2797 CrossRef CAS .
  183. M. P. Mahmud, T. Tat, X. Xiao, P. Adhikary and J. Chen, Advances in 4D-printed physiological monitoring sensors, Exploration, 2021, 1, 20210033 CrossRef .
  184. R. Ghanim, A. Kaushik, J. Park and A. Abramson, Communication protocols integrating wearables, ingestibles, and implantables for closed-loop therapies, Device, 2023, 1, 100092 CrossRef .
  185. Z. Dai, M. Lei, S. Ding, Q. Zhou, B. Ji, M. Wang and B. Zhou, Durable superhydrophobic surface in wearable sensors: from nature to application, Exploration, 2023, 4, 20230046 CrossRef PubMed .
  186. B. Luo, Q. Zhou, W. Chen, L. Sun, L. Yang, Y. Guo, H. Liu, Z. Wu, R. E. Neisiany and X. Qin, Nonadjacent wireless electrotherapy for tissue repair by a 3D-printed bioresorbable fully soft triboelectric nanogenerator, Nano Lett., 2023, 23, 2927–2937 CrossRef CAS PubMed .
  187. J. Yun, N. Jayababu and D. Kim, Self-powered transparent and flexible touchpad based on triboelectricity towards artificial intelligence, Nano Energy, 2020, 78, 105325 CrossRef CAS .
  188. R. Mao, D. Zhang, Z. Wang, H. Zhang, D. Wang, M. Tang, L. Zhou, H. Cai and H. Xia, Deep-learning-assisted low-cost flexible cotton yarn-based triboelectric nanogenerator for ultra-sensitive human-computer merging interfaces, Nano Energy, 2023, 111, 108418 CrossRef CAS .
  189. H. Fang, L. Wang, Z. Fu, L. Xu, W. Guo, J. Huang, Z. L. Wang and H. Wu, Anatomically designed triboelectric wristbands with adaptive accelerated learning for human–machine interfaces, Adv. Sci., 2023, 10, 2205960 CrossRef .
  190. A. de Souza Brito, M. B. Vieira, M. L. S. C. De Andrade, R. Q. Feitosa and G. A. Giraldi, Combining max-pooling and wavelet pooling strategies for semantic image segmentation, Expert Syst. Appl., 2021, 183, 115403 CrossRef .
  191. B. Shi, A. Tay, W. L. Au, D. M. Tan, N. S. Chia and S.-C. Yen, Detection of freezing of gait using convolutional neural network and data from lower limb motion sensors, IEEE Trans. Biomed. Eng., 2022, 69, 2256–2267 Search PubMed .
  192. N. Sadeghzadehyazdi, T. Batabyal and S. T. Acton, Modeling spatiotemporal patterns of gait anomaly with a CNN-LSTM deep neural network, Expert Syst. Appl., 2021, 185, 115582 CrossRef .
  193. Z. Ye and J. Yu, AKSNet: a novel convolutional neural network with adaptive kernel width and sparse regularization for machinery fault diagnosis, J. Manuf. Syst., 2021, 59, 467–480 CrossRef .
  194. N. T. Beigh, F. T. Beigh and D. Mallick, Machine learning assisted hybrid transduction nanocomposite based flexible pressure sensor matrix for human gait analysis, Nano Energy, 2023, 116, 108824 CrossRef CAS .
  195. Z. Zhang, T. He, M. Zhu, Z. Sun, Q. Shi, J. Zhu, B. Dong, M. R. Yuce and C. Lee, Deep learning-enabled triboelectric smart socks for IoT-based gait analysis and VR applications, npj Flexible Electron., 2020, 4, 29 CrossRef .
  196. H. Ullah, K. Muhammad, M. Irfan, S. Anwar, M. Sajjad, A. S. Imran and V. H. C. de Albuquerque, light-DehazeNet: a novel lightweight CNN architecture for single image dehazing, IEEE Trans. Image Process., 2021, 30, 8968–8982 Search PubMed .
  197. I. Galić, M. Habijan, H. Leventić and K. Romić, Machine learning empowering personalized medicine: a comprehensive review of medical image analysis methods, Electronics, 2023, 12, 4411 CrossRef .
  198. L. Dipietro, U. Eden, S. Elkin-Frankston, M. M. El-Hagrassy, D. D. Camsari, C. Ramos-Estebanez, F. Fregni and T. Wagner, Integrating Big Data, Artificial intelligence, and motion analysis for emerging precision medicine applications in Parkinson's Disease, J. Big Data, 2024, 11, 155 CrossRef .
  199. Y. Kavya and S. S. Reka, Abnormal gait classification in children with cerebral palsy using convLSTM hybrid model and GAN, IEEE Access, 2024, 12, 117721–117736 Search PubMed .
  200. M. Panwar, D. Biswas, H. Bajaj, M. Jöbges, R. Turk, K. Maharatna and A. Acharyya, Rehab-net: deep learning framework for arm movement classification using wearable sensors for stroke rehabilitation, IEEE Trans. Biomed. Eng., 2019, 66, 3026–3037 Search PubMed .
  201. L. Balcombe and D. De Leo, Psychological screening and tracking of athletes and digital mental health solutions in a hybrid model of care: mini review, JMIR Form. Res., 2020, 4, e22755 CrossRef PubMed .
  202. C. Wu, W. Ding, R. Liu, J. Wang, A. C. Wang, J. Wang, S. Li, Y. Zi and Z. L. Wang, Keystroke dynamics enabled authentication and identification using triboelectric nanogenerator array, Mater. Today, 2018, 21, 216–222 CrossRef .
  203. P. Ray, S. S. Reddy and T. Banerjee, Various dimension reduction techniques for high dimensional data analysis: a review, Artif. Intell. Rev., 2021, 54, 3473–3515 CrossRef .
  204. H. Askari, N. Xu, B. H. G. Barbosa, Y. Huang, L. Chen, A. Khajepour, H. Chen and Z. L. Wang, Intelligent systems using triboelectric, piezoelectric, and pyroelectric nanogenerators, Mater. Today, 2022, 52, 188–206 CrossRef .
  205. L. Zhang, F. Xue, W. Du, C. Han, C. Zhang and Z. Wang, Transparent paper-based triboelectric nanogenerator as a page mark and anti-theft sensor, Nano Res., 2014, 7, 1215–1223 CrossRef CAS .
  206. J. Basak and R. Krishnapuram, Interpretable hierarchical clustering by constructing an unsupervised decision tree, IEEE Trans. Knowl. Data Eng., 2005, 17, 121–132 Search PubMed .
  207. W. Zhang, L. Deng, L. Yang, P. Yang, D. Diao, P. Wang and Z. L. Wang, Multilanguage-handwriting self-powered recognition based on triboelectric nanogenerator enabled machine learning, Nano Energy, 2020, 77, 105174 CrossRef CAS .
  208. G. T. Reddy, M. P. K. Reddy, K. Lakshmanna, R. Kaluri, D. S. Rajput, G. Srivastava and T. Baker, Analysis of dimensionality reduction techniques on big data, IEEE Access, 2020, 8, 54776–54788 Search PubMed .
  209. J. Y. Kwon, J. E. Kim, J. S. Kim, S. Y. Chun, K. Soh and J. H. Yoon, Artificial sensory system based on memristive devices, Exploration, 2023, 20220162 Search PubMed .
  210. L. Wang, W. Ye, Y. Zhu, F. Yang and Y. Zhou, Optimal parameters selection of back propagation algorithm in the feedforward neural network, Eng. Anal. Bound. Elem., 2023, 151, 575–596 CrossRef .
  211. Z. Wang, J. An, J. Nie, J. Luo, J. Shao, T. Jiang, B. Chen, W. Tang and Z. L. Wang, A self-powered angle sensor at nanoradian-resolution for robotic arms and personalized medicare, Adv. Mater., 2020, 32, 2001466 CrossRef CAS .
  212. W. Qiao, L. Zhou, Z. Zhao, D. Liu, S. Li, J. An, X. Li, Y. Gao, P. Yang and J. Liu, A self-powered vector motion sensor for smart robotics and personalized medical rehabilitation, Nano Energy, 2022, 104, 107936 CrossRef CAS .
  213. M. Han, H. Wang, Y. Yang, C. Liang, W. Bai, Z. Yan, H. Li, Y. Xue, X. Wang and B. Akar, Three-dimensional piezoelectric polymer microsystems for vibrational energy harvesting, robotic interfaces and biomedical implants, Nat. Electron., 2019, 2, 26–35 CrossRef .
  214. S. Chun, W. Son, H. Kim, S. K. Lim, C. Pang and C. Choi, Self-powered pressure-and vibration-sensitive tactile sensors for learning technique-based neural finger skin, Nano Lett., 2019, 19, 3305–3312 CrossRef CAS PubMed .
  215. Y. Liu, E. Li, X. Wang, Q. Chen, Y. Zhou, Y. Hu, G. Chen, H. Chen and T. Guo, Self-powered artificial auditory pathway for intelligent neuromorphic computing and sound detection, Nano Energy, 2020, 78, 105403 CrossRef CAS .
  216. H. S. Lee, J. Chung, G. T. Hwang, C. K. Jeong, Y. Jung, J. H. Kwak, H. Kang, M. Byun, W. D. Kim and S. Hur, Flexible inorganic piezoelectric acoustic nanosensors for biomimetic artificial hair cells, Adv. Funct. Mater., 2014, 24, 6914–6921 CrossRef CAS .
  217. S. Chen, F. Manshaii, X. Fan, Z. Duan, X. Zhao, Y. Zhou and J. Chen, Creating breathable wearable bioelectronics using three-dimensional liquid diodes, Matter, 2024, 7, 2632–2635 CrossRef CAS .
  218. J. Xu, T. Tat, J. Yin, D. Ngo, X. Zhao, X. Wan, Z. Che, K. Chen, L. Harris and J. Chen, A textile magnetoelastic patch for self-powered personalized muscle physiotherapy, Matter, 2023, 6, 2235–2247 CrossRef .
  219. S. Xu, X. Xiao and J. Chen, Stretchable fiber strain sensors for wearable biomonitoring, Natl. Sci. Rev., 2024, 11, nwae173 CrossRef PubMed .
  220. W. Kwak, J. Yin, S. Wang and J. Chen, Advances in triboelectric nanogenerators for self-powered wearable respiratory monitoring, FlexMat, 2024, 1, 5–22 CrossRef .
  221. V. Kashyap, J. Yin, X. Xiao and J. Chen, Bioinspired nanomaterials for wearable sensing and human-machine interfacing, Nano Res., 2024, 17, 445–461 CrossRef .
  222. Z. Che, X. Wan, J. Xu, C. Duan, T. Zheng and J. Chen, Speaking without vocal folds using a machine-learning-assisted wearable sensing-actuation system, Nat. Commun., 2024, 15, 1873 CrossRef CAS .
  223. V. Kashyap, J. Yin, X. Xiao and J. Chen, Bioinspired nanomaterials for wearable sensing and human-machine interfacing, Nano Res., 2023, 17, 445–461 CrossRef .
  224. T. Tat, X. Zhao and J. Chen, Leveraging fluidic programming for pneumatic haptic textiles, Device, 2023, 1, 100094 CrossRef .
  225. H. Qin, Y. Yan, Q. Feng, H. Liu, H.-P. Cong and S.-H. Yu, Rapid printing and patterning of tough, self-healable, and recyclable hydrogel thin-films toward flexible sensing devices, Nano Lett., 2022, 22, 8101–8108 CrossRef CAS PubMed .
  226. S. Xu, S. Ahmed, M. Momin, A. Hossain and T. Zhou, Unleashing the potential of 3D printing soft materials, Device, 2023, 1, 100067 CrossRef .
  227. K. Xu, Q. a. Li, Y. Lu, H. Luo, Y. Jian, D. Li, D. Kong, R. Wang, J. Tan and Z. Cai, Laser direct writing of flexible thermal flow sensors, Nano Lett., 2023, 23, 10317–10325 CrossRef CAS PubMed .
  228. Z. Dai, M. Lei, S. Ding, Q. Zhou, B. Ji, M. Wang and B. Zhou, Durable superhydrophobic surface in wearable sensors: from nature to application, Exploration, 2024, 4, 20230046 CrossRef CAS .
  229. P. Cheng, S. Dai, Y. Liu, Y. Li, H. Hayashi, R. Papani, Q. Su, N. Li, Y. Dai and W. Liu, An intrinsically stretchable power-source system for bioelectronics, Device, 2024, 2, 100216 CrossRef .
  230. H. Zhou, K.-N. Kim, M.-J. Sung, S. J. Han and T.-W. Lee, Intrinsically stretchable low-dimensional conductors for wearable organic light-emitting diodes, Device, 2023, 1, 100060 CrossRef .
  231. T. Zhao, X. Xiao, Y. Wu, J. Ma, Y. Li, C. Lu, C. Shokoohi, Y. Xu, X. Zhang and Y. Zhang, Tracing the flu symptom progression via a smart face mask, Nano Lett., 2023, 23, 8960–8969 CrossRef CAS PubMed .
  232. G. Li, F. Sun, S. Zhao, R. Xu, H. Wang, L. Qu and M. Tian, Autonomous electroluminescent textile for visual interaction and environmental warning, Nano Lett., 2023, 23, 8436–8444 CrossRef CAS PubMed .
  233. B. Sui, T. Ding, X. Wan, Y. Chen, X. Zhang, Y. Cui, J. Pan, L. Li and X. Liu, Piezoelectric stimulation enhances bone regeneration in alveolar bone defects through metabolic reprogramming of macrophages, Exploration, 2024, 4, 20230149 CrossRef .
  234. S. Xu, X. Wan, F. Manshaii, Z. Che and J. Chen, Advances in piezoelectric nanogenerators for self-powered cardiac care, Nano Trends, 2024, 1, 100042 CrossRef .
  235. C. Carlos, J. Li, Z. Zhang, K. J. Berg, Y. Wang and X. Wang, Strain-correlated piezoelectricity in quasi-two-dimensional zinc oxide nanosheets, Nano Lett., 2023, 23, 6148–6155 CrossRef CAS .
  236. X. Wang, W.-Z. Song, M.-H. You, J. Zhang, M. Yu, Z. Fan, S. Ramakrishna and Y.-Z. Long, Bionic single-electrode electronic skin unit based on piezoelectric nanogenerator, ACS Nano, 2018, 12, 8588–8596 CrossRef CAS PubMed .
  237. A. S. Dahiya, F. Morini, S. Boubenia, K. Nadaud, D. Alquier and G. Poulin-Vittrant, Organic/inorganic hybrid stretchable piezoelectric nanogenerators for self-powered wearable electronics, Adv. Mater. Technol., 2018, 3, 1700249 CrossRef .
  238. Z. Yi, T. Xu, W. Shang and X. Wu, Touch modality identification with tensorial tactile signals: a kernel-based approach, IEEE Trans. Autom. Sci. Eng., 2021, 19, 959–968 Search PubMed .
  239. J. Zheng, Y. Chen, Z. Wu, Y. Wang and L. Wang, Integration of flexible touch panels and machine learning: applications and techniques, Adv. Intell. Syst., 2024, 6, 2300560 CrossRef .
  240. A. J. Spiers, M. V. Liarokapis, B. Calli and A. M. Dollar, Single-grasp object classification and feature extraction with simple robot hands and tactile sensors, IEEE Trans. Haptics, 2016, 9, 207–220 Search PubMed .
  241. S. A. Han, T. H. Kim, S. K. Kim, K. H. Lee, H. J. Park, J. H. Lee and S. W. Kim, Point-defect-passivated MoS2 nanosheet-based high performance piezoelectric nanogenerator, Adv. Mater., 2018, 30, 1800342 CrossRef PubMed .
  242. P. Ren, Y. Xiao, X. Chang, P.-Y. Huang, Z. Li, B. B. Gupta, X. Chen and X. Wang, A survey of deep active learning, ACM Comput. Surv., 2021, 54, 1–40 Search PubMed .
  243. B. Lan, X. Xiao, A. Di Carlo, W. Deng, T. Yang, L. Jin, G. Tian, Y. Ao, W. Yang and J. Chen, Topological nanofibers enhanced piezoelectric membranes for soft bioelectronics, Adv. Funct. Mater., 2022, 32, 2207393 CrossRef CAS .
  244. Y. Qiu, S. Sun, X. Wang, K. Shi, Z. Wang, X. Ma, W. Zhang, G. Bao, Y. Tian and Z. Zhang, Nondestructive identification of softness via bioinspired multisensory electronic skins integrated on a robotic hand, npj Flexible Electron., 2022, 6, 45 CrossRef .
  245. S. Zhang, H. Zhao, X. Ma, J. Deng and Y. Liu, A 3-DOF piezoelectric micromanipulator based on symmetric and antisymmetric bending of a cross-shaped beam, IEEE Trans. Ind. Electron., 2022, 70, 8264–8275 Search PubMed .

This journal is © The Royal Society of Chemistry 2025
Click here to see how this site uses Cookies. View our privacy policy here.