- 1College of Physical Education, Jeonju University, Jeonju, Jeollabuk-do, Republic of Korea
- 2Gongqing Institute of Science and Technology, Jiujiang, Jiangxi Province, China
Introduction: Assistive robots and human-robot interaction have become integral parts of sports training. However, existing methods often fail to provide real-time and accurate feedback, and they often lack integration of comprehensive multi-modal data.
Methods: To address these issues, we propose a groundbreaking and innovative approach: CAM-Vtrans—Cross-Attention Multi-modal Visual Transformer. By leveraging the strengths of state-of-the-art techniques such as Visual Transformers (ViT) and models like CLIP, along with cross-attention mechanisms, CAM-Vtrans harnesses the power of visual and textual information to provide athletes with highly accurate and timely feedback. Through the utilization of multi-modal robot data, CAM-Vtrans offers valuable assistance, enabling athletes to optimize their performance while minimizing potential injury risks. This novel approach represents a significant advancement in the field, offering an innovative solution to overcome the limitations of existing methods and enhance the precision and efficiency of sports training programs.
1 Introduction
In the field of sports technology, the application of deep learning and machine learning techniques to enhance training efficiency and athlete performance has become a hot topic of research (Zheng et al., 2020). These technologies can accurately analyze athletes' movements and provide real-time feedback, helping athletes improve their skills more effectively (Pan et al., 2019). However, while existing technologies can handle single data sources such as video or biosensor data, their capabilities are still insufficient when it comes to integrating and processing multiple types of data (Herman et al., 2021), especially when simultaneously dealing with visual information and verbal instructions. This limitation highlights the need for the development of new methods to comprehensively understand and guide athlete training.
Traditional methods primarily rely on symbolic AI and knowledge representation for Taekwondo action recognition. Expert systems, for example, simulate human experts' decision-making processes by encoding their knowledge and provide explicit explanations for each recognition result. Yang et al. (2021) proposed a multi-knowledge representation framework for big data AI applications. Additionally, a comprehensive review by Himabindu et al. (2023) showcased the combination of symbolic reasoning and deep learning in neural-symbolic AI, highlighting its various applications and developments across different domains. Rule-based methods, on the other hand, utilize a set of predefined rules for action recognition. These methods demonstrate high determinism and reliability, performing well even in the face of complex or diverse actions. Jin et al. (2022) introduced a deep reinforcement learning system for automatic symbol grounding discovery, while the research by Ilager et al. (2023) showcased the cost-saving benefits of symbolic representation in edge AI applications. Furthermore, logistic regression, as a statistical method, learns features from training data for classification decisions. It not only finds important applications in action recognition but also significantly improves classification accuracy. The study by Insuasti et al. (2023) demonstrated the application of logistic regression in sports action recognition, while Wu et al. (2022) further explored the use of fuzzy logic in symbolic representation, enhancing the symbolic foundations of AI. These methods offer advantages such as strong interpretability and transparency in the decision-making process. However, these methods have limitations in handling complex and diverse actions as well as limited capabilities in processing large-scale data.
To address the limitations of traditional algorithms, data-driven and machine learning-based approaches have been employed in multi-modal robot-assisted sports training. These approaches mainly utilize methods such as decision trees, random forests, and multi-layer perceptrons to tackle the challenges. This approach offers advantages such as efficient handling of large-scale data, high accuracy, and the ability to handle non-linear problems. For instance, Tjondronegoro and Chen (2006) automated event classification in sports videos using decision tree methods, while Jose et al. (2023) applied decision tree algorithms in predicting athlete performance. Furthermore, Morciano et al. (2023) used random forest algorithms to predict performance indicators of soccer players, demonstrating their superiority in handling biomechanical data, and Yagin et al. (2023) showcased the high accuracy of random forests in determining the positions of professional soccer players. Lastly, Aresta et al. (2022) highlighted the superior performance of multi-layer perceptrons in classifying elite and novice fencers based on biomechanical data, while Bakthavatchalam et al. (2022) demonstrated the efficient predictive performance of multi-layer perceptrons in agriculture. However, these methods have challenges such as overfitting, high computational costs, and strong reliance on large amounts of annotated data.
To overcome the limitations of statistical and machine learning algorithms, deep learning-based approaches have been used for Taekwondo action recognition, primarily employing Convolutional Neural Networks (CNN), reinforcement learning, and Transformer models. These methods offer higher accuracy and the ability to handle complex data. Firstly, Convolutional Neural Networks efficiently extract image features and have shown remarkable performance in predicting sports game outcomes and recognizing athlete actions. For example, Chen et al. (2020) used CNN to predict NBA game results with an accuracy of 91%, while Liu (2022) utilized CNN to improve action detection rates in sports videos. Secondly, reinforcement learning demonstrates significant potential in sports training by continuously adjusting strategies to optimize the decision-making process. The reinforcement learning approach proposed by Jia et al. (2020) improved players' winning rates in basketball training, and the research by Du et al. (2021) showcased the application of reinforcement learning in esports. Lastly, Transformer models, known for their advantages in handling sequential data, have been used for time-series analysis of motion signals, showing impressive performance. Dirgová Luptáková et al. (2022) achieved 99.2% accuracy in human activity recognition using the Transformer model, while Hauri and Vucetic (2023) combined Transformer with LSTM for team activity recognition in basketball games. However, these methods have challenges such as high computational complexity and a demand for large-scale training data.
Considering these challenges, this study proposes a novel approach, CAM-Vtrans: Real-time Sports Training Utilizing Multi-modal Robot Data, to address the limitations of traditional and machine learning algorithms, such as poor adaptability to complex environments, high computational costs, and dependency on large labeled datasets. CAM-Vtrans combines Vision Transformer (ViT), CLIP, and cross-attention mechanisms. ViT divides the image into multiple small patches and encodes them as a sequence, utilizing the self-attention mechanism to process these sequences and capture complex relationships within the image. This approach is particularly effective in handling sports activity images with rich details. The introduction of the CLIP model enables the system to understand training instructions in natural language and combines them with visual data to provide context-aware feedback. Through the cross-attention mechanism, this system further optimizes the fusion of different modalities, making the transformation from visual information to language descriptions more accurate and efficient. This integrated approach not only enhances the accuracy and efficiency of sports training analysis but also significantly reduces the computational burden and reliance on extensive labeled data.
The main contributions of this research can be summarized as follows:
• CAM-Vtrans is an innovative approach that combines Vision Transformer (ViT), the CLIP model, and cross-attention mechanisms to process and analyze multi-modal robot data in real-time, enhancing the accuracy of feedback in sports training.
• This method performs exceptionally well in various multi-scenario applications, efficiently handling complex sports activity images. It possesses broad applicability and adaptability, providing reliable support for different training requirements.
• Experimental results demonstrate that CAM-Vtrans significantly outperforms traditional methods in action recognition and feedback accuracy, greatly improving the effectiveness of sports training while reducing computational costs and reliance on large-scale annotated data.
2 Related work
2.1 Assisting sports training
In recent years, machine learning has made significant progress in assisting sports training tasks. Traditional sports training methods heavily rely on coaches' experience and intuition, which often suffer from subjectivity and lack of precision. The introduction of machine learning has made the training process more scientific and systematic. Classic machine learning algorithms such as decision trees, random forests, and logistic regression have been widely applied in areas such as athlete performance prediction and injury risk assessment. For example, decision tree-based systems can provide personalized training recommendations by analyzing athletes' physiological and training data (Jose et al., 2023). However, these traditional machine learning methods also have some notable drawbacks and limitations (Tang et al., 2023). Firstly, these methods require high-quality and large quantities of labeled data to train models, which can be costly to acquire. Moreover, traditional machine learning algorithms exhibit limitations when dealing with complex and multi-dimensional sports data. For instance, while random forests can handle non-linear relationships to some extent, they still struggle with highly complex and dynamically changing sports data (Morciano et al., 2023). Additionally, these methods lack interpretability and explainability, making it difficult to provide clear explanations for training outcomes and limiting their practical applications (Dong et al., 2024). To overcome these limitations, deep learning methods have gradually become a research focus in the field of sports training. Deep learning, by constructing multi-layer neural networks, can better capture complex features and patterns, thus improving the predictive accuracy and robustness of models. However, deep learning methods also face challenges such as high computational costs, long training times, and dependence on large-scale annotated data, which still need to be further addressed in practical applications (Wang et al., 2024)
2.2 Transformer models
Since its introduction in 2017, the Transformer model has achieved groundbreaking advancements across multiple domains. Its unique self-attention mechanism and parallel processing capabilities have made Transformers particularly prominent in natural language processing (NLP) tasks. For instance, models like BERT and GPT, which are based on Transformer architecture, have demonstrated significant effectiveness in tasks such as language understanding, text generation, and machine translation. The Transformer model addresses the inefficiencies and vanishing gradient problems associated with traditional sequential models like RNNs and LSTMs by processing input sequences in parallel and dividing them into smaller chunks (Lu et al., 2024). Beyond NLP, the Transformer model has also shown strong capabilities in the field of computer vision (CV). Vision Transformer (ViT), by dividing images into fixed-size patches and processing these patches as input sequences, has achieved performance comparable to or even surpassing that of convolutional neural networks (CNNs). ViT has excelled in tasks such as image classification, object detection, and image segmentation, proving the potential of Transformers in handling visual data (Hu et al., 2019). In addition, the Transformer model has wide-ranging applications in time series data analysis, recommendation systems, and game AI. In time series data analysis, Transformers can effectively capture long-term dependencies, enhancing prediction accuracy. In recommendation systems, Transformers model user behavior sequences to provide more precise recommendations. In game AI, Transformers, combined with deep reinforcement learning, optimize strategy selection.
2.3 Sports action recognition
Sports action recognition is a crucial research area in sports science and computer vision, aiming to automatically identify and evaluate athletic performance by analyzing athletes' motion data. Traditional action recognition methods primarily rely on feature engineering-based machine learning algorithms, such as support vector machines, decision trees, and random forests. These methods extract features from motion data for classification and recognition, achieving certain levels of effectiveness (Zhao et al., 2020). With the development of deep learning technologies, the advantages of convolutional neural networks (CNNs) in image and video processing have become increasingly apparent, leading to their widespread application in sports action recognition. CNNs can automatically learn and extract high-level features from data, significantly improving the accuracy and robustness of action recognition (Zou et al., 2019). Additionally, temporal models in deep learning, such as long short-term memory networks (LSTMs) and Transformer models, have been applied to action recognition, better handling time series data and capturing dynamic changes in actions. However, despite the impressive performance of deep learning methods in action recognition, several challenges and limitations persist. First, deep learning models require large-scale annotated data, and acquiring and annotating sports action data is costly, limiting the effectiveness of model training. Second, deep learning models have high computational complexity, requiring substantial computational resources and time for training and inference, which can be a bottleneck in real-time applications. Moreover, existing action recognition models still face difficulties in handling complex and diverse actions, making it challenging to adapt to various sports scenarios and action types. To address these issues, researchers are exploring multi-modal data fusion methods, combining visual, auditory, and tactile data to enhance the accuracy and robustness of action recognition (Li et al., 2018). Additionally, emerging technologies such as reinforcement learning and self-supervised learning are being introduced to action recognition to reduce reliance on annotated data and improve model generalization. Despite these advancements, achieving efficient, accurate, and robust sports action recognition remains a challenging research topic, necessitating further exploration and innovation.
3 Methodology
3.1 Overview of our network
In this research, we propose a multimodal robotic system that combines Vision Transformer (ViT), CLIP, and cross-attention mechanisms for real-time feedback and guidance in sports training. The main innovation of this system lies in the use of advanced visual and language processing models to analyze athletes' performances in-depth and provide immediate guidance and feedback.
Figure 1 shows the overall framework diagram of the proposed method.
Textual information is inputted from the L-Branch branch and is segmented into words or subwords. The text is then transformed into fixed-dimensional vector representations through an embedding layer. These text vectors are linearly projected and inputted into the corresponding Transformer Encoder. Images are inputted from the S-Branch, and each branch's image is divided into fixed-sized patches. After linear projection, the image patches are inputted into their respective Transformer Encoders. The image and text features interact and fuse through a Cross-Attention mechanism. The Cross-Attention layer takes features from the image and text encoders, calculates the correlation between them, and generates a fused multimodal feature representation. The fused multimodal features are further processed by a Multi-Scale Transformer Encoder layer to capture features at different scales, enhancing the expressive power of the features. Finally, a Multi-Layer Perceptron (MLP) head is used for tasks such as classification or regression. In the revised version, we will update Figure 1 to visually illustrate the processing and flow of textual information, including adding a schematic diagram of text input, demonstrating the processing of text through the embedding layer and linear projection layer, and clarifying the interaction between image and text features in the Section 3.4.
Differentiation from prior work: While the combination of ViT, CLIP, and Cross-Attention has been proposed in other domains, our research is the first to apply it to real-time sports coaching systems. Unlike previous studies, our research focuses on effectively integrating visual and textual data in dynamic and real-time sports training environments. Specifically, our proposed CAM-Vtrans system takes into account the continuity and complexity of sports actions during its design. Through optimized Cross-Attention mechanisms and multi-scale feature extraction modules, the system is able to provide stable and accurate feedback even with high-frequency inputs.
Overcoming limitations of previous methods: Previous methods often suffer from low computational efficiency and long feedback latency when dealing with real-time multimodal data. In this research, we address these limitations by introducing the ViT-Adapter module, which enhances feature extraction efficiency. Through optimized Cross-Attention mechanisms, we achieve faster and more accurate multimodal data fusion. Compared to traditional single-modal or inefficient multimodal methods, the CAM-Vtrans system significantly reduces inference time and improves the accuracy of real-time feedback, overcoming the limitations of previous methods in terms of real-time performance and data fusion.
Reasons for method selection: We chose the combination of ViT, CLIP, and Cross-Attention because these techniques have demonstrated excellent performance in handling complex visual and textual data. ViT is renowned for its powerful visual feature extraction capabilities, while CLIP effectively maps visual and textual data to the same feature space, enabling cross-modal understanding. The Cross-Attention mechanism efficiently establishes correlations between different modalities, enhancing information fusion. These characteristics make them well-suited for application in sports training scenarios that involve large amounts of visual and textual data and require real-time feedback. Therefore, the selection of these methods is not random but based on their superiority in multimodal data processing and real-time performance.
Firstly, the Vision Transformer (ViT) is employed to process video data captured from multiple cameras. ViT divides each frame into several image patches, converts these patches into a sequence of vectors, and processes them with self-attention mechanisms to identify key visual information. This approach allows the model to focus on specific regions within the image that are relevant to the movement technique, improving the accuracy and granularity of motion analysis. Simultaneously, the CLIP model is utilized to process and parse natural language inputs such as coach instructions or verbal feedback from athletes. CLIP learns from a large corpus of image-text pairs, establishing intuitive associations between image content and textual descriptions. This enables CLIP to directly relate language descriptions to visual data, providing robust support for precise understanding of movement techniques and coach's intentions. In the implementation workflow, once the athlete starts training, the system collects video and audio data in real-time. The visual and language data are processed separately by ViT and CLIP, respectively, and then fed into the cross-attention layer. In this layer, the system analyzes the correlations and interactions between visual and language information, optimizing the fusion process to extract the most valuable insights from the inputs. The core of the cross-attention mechanism lies in its ability to dynamically adjust the focus on different data sources based on specific training scenarios, providing more personalized and goal-oriented training recommendations. After performing these analyses, the system generates specific feedback reports, including action correction guidelines, performance evaluations, and improvement suggestions. This feedback can be presented directly to the athlete through a graphical user interface or sent to the coach via mobile devices. Additionally, the system includes a feedback adjustment module that allows the coach to fine-tune the level and frequency of feedback as needed, ensuring training continuity and adaptability. The focal point of the entire system design is to ensure real-time and accurate feedback, making the training process more intelligent and efficient. The aim is to maximize athletes' performance and training effectiveness through technological means.
3.2 Vision-transformer
Vision Transformer (ViT) (Miyazawa et al., 2022) is a deep learning model that applies the Transformer architecture to process visual data. Traditionally, Convolutional Neural Networks (CNNs) have been the dominant approach for visual tasks, but ViT introduces a novel paradigm by leveraging the self-attention mechanism of Transformers (Papadakis and Spyrou, 2024). Figure 2 is a schematic diagram of the principle of Vision-transformer Model.
The ViT-Adapter consists of the following components: Firstly, the Spatial Prior Module is responsible for initially extracting spatial features from the input image. The image first goes through the Stem layer, generating a series of feature maps (F1, F2, F3, …, Fsp) that capture spatial information at different scales, preparing for the subsequent feature injection. Secondly, the Spatial Feature Injector is one of the key modules of the ViT-Adapter. It injects the spatial features (Fsp) extracted by the Spatial Prior Module into the intermediate features (Fvit) of the ViT using a Cross-Attention mechanism. Specifically, the intermediate features of the ViT serve as the Query, while the spatial features act as the Key and Value. The Cross-Attention calculates the fused features (Fsp + Fvit). Then, the Multi-Scale Feature Extractor further processes the fused features through multiple Cross-Attention layers and a Feed-Forward Neural Network (FFN) to enhance the expressive power of multi-scale features, enabling the model to better capture image details and global information. Additionally, the ViT-Adapter inserts Injector and Extractor modules between each block of the ViT. The Injector module injects the features from the Spatial Prior Module into the current ViT features, while the Extractor module extracts useful information from the fused features for the next Transformer Block to use. Finally, after being processed by multiple Transformer Blocks and ViT-Adapter modules, the final features are fed into a Multi-Layer Perceptron (MLP) head for tasks such as classification, detection, or segmentation. Through these improvements, the ViT-Adapter significantly enhances the ViT model's ability to capture spatial features when processing images, improving its performance in various visual tasks.
The Vision Transformer (ViT) model operates by dividing an input image into smaller patches, which are then flattened into a sequence of 1D vectors capturing local visual information. These patches are linearly projected into higher-dimensional embeddings, serving as the input to the Transformer model. The Transformer architecture, composed of multiple identical layers each containing a self-attention mechanism and a feed-forward neural network, captures both global and local dependencies within the sequence of patches. During self-attention, patches exchange information and capture long-range dependencies, with attended representations aggregated and combined with original patch representations using residual connections. This process refines the patch representations based on contextual information. After multiple layers, the final image representation is obtained, which can be used for tasks like image classification, object detection, or segmentation. ViT's advantages include capturing global and local information, scalability, and learning from raw pixels without hand-engineered features. However, its self-attention mechanism's quadratic computational complexity is a limitation. In real-time feedback for multimodal robots in sports training, ViT analyzes visual information to understand and provide guidance on body movements. Trained on annotated sports videos, ViT extracts relevant features and captures spatial relationships, enabling the robot to offer accurate, context-aware feedback by leveraging self-attention to focus on critical image regions and dependencies between patches.
The input image is divided into patches, resulting in a sequence of patches, denoted by xi, where i represents the index of each patch. Each patch is then linearly projected to a higher-dimensional embedding space using a learnable linear transformation. Let's denote the projected embeddings as zi.
The self-attention mechanism in ViT is defined by the following equations (Equation 1):
Here, Q, K, and V are the query, key, and value matrices, respectively. They are derived from the projected embeddings zi as follows:
In these equations, Z is the matrix obtained by stacking the projected embeddings zi, and WQ, WK, and WV are learnable linear transformation matrices (Equation 2).
The self-attention mechanism calculates the attention weights between patches by computing the dot product similarity between the query and key matrices, scaled by the square root of the dimension dk. The softmax function is applied to obtain the attention weights, which are then used to weight the values V.
The attended representations are computed as follows (Equation 3):
The attended representations are then combined with the original patch embeddings using a residual connection, resulting in the intermediate representations:
Here, LayerNorm denotes layer normalization (Equation 4).
The intermediate representations are then passed through a feed-forward neural network (FFN) with two linear transformations and a non-linear activation function, typically a GELU activation:
W1, W2, b1, and b2 are learnable parameters of the feed-forward network (Equation 5).
The output of the ViT model is obtained by stacking multiple layers of self-attention and feed-forward networks. The final representation of the image is typically obtained by applying mean pooling to the patch embeddings.
In summary, Vision Transformer is a powerful model for visual processing that replaces traditional convolutional approaches with self-attention mechanisms. Its ability to capture global and local dependencies makes it well-suited for understanding and analyzing visual data in real-time feedback and guidance systems for multimodal robots in sports training.
3.3 CLIP
CLIP (Contrastive Language-Image Pretraining) is a deep learning model that learns to associate images and their corresponding text descriptions (Dobrzycki et al., 2023). It aims to bridge the gap between vision and language modalities, enabling cross-modal understanding and reasoning. The key idea behind CLIP is to leverage large-scale pretraining on a dataset of image-text pairs, allowing the model to learn rich representations that capture the semantic relationship between visual and textual information (Koh et al., 2024). The basic principle of the CLIP model involves jointly training a vision encoder and a text encoder. The vision encoder processes images and maps them to a high-dimensional latent space, while the text encoder processes textual descriptions and maps them to the same latent space. The encoders are trained to ensure that corresponding image-text pairs are closer to each other in the latent space compared to non-corresponding pairs. Figure 3 is a schematic diagram of the principle of CLIP Model.
The training process of CLIP involves several key steps: first, the input image is encoded by a vision encoder, typically a convolutional neural network (CNN), which extracts visual features and projects them into a latent space using a learnable linear transformation. Simultaneously, the input text description is encoded by a text encoder based on a Transformer architecture, which tokenizes the text, applies word embeddings, and processes it through multiple Transformer layers to produce the text's representation in the latent space. CLIP utilizes a contrastive loss function to maximize the similarity between corresponding image-text pairs while minimizing the similarity between non-corresponding pairs, achieved by measuring the cosine similarity between their latent representations. Pretraining on large-scale datasets, such as Conceptual Captions and ImageNet, enables CLIP to learn generalizable representations capturing the semantic relationship between images and texts. After pretraining, CLIP can be fine-tuned for downstream tasks like image classification, object detection, or image captioning. In real-time feedback and guidance for multimodal robots in sports training, CLIP is crucial for understanding and associating visual and textual information. By aligning and reasoning about sports movements based on annotated image-text pairs, CLIP allows the robot to understand textual annotations related to key movements, techniques, and performance indicators. Leveraging the pretrained CLIP model, the robot can generate real-time feedback and guidance based on its comprehension of the athlete's movements and the semantic context provided by textual information.
Let's consider an image-text pair with an image I and a text description T.
Image Encoding: The image I is processed by a vision encoder, typically a convolutional neural network (CNN), to extract visual features. Let's denote the image representation as vI. Text Encoding: The text description T is processed by a text encoder, typically a Transformer-based architecture, to encode the textual information. Let's denote the text representation as vT. Similarity Measurement: The similarity between the image and text representations is measured using cosine similarity. It can be calculated as:
Here, · denotes the dot product, and |·| represents the Euclidean norm (Equation 6).
Contrastive Loss: CLIP utilizes a contrastive loss function to train the model. Given a positive pair (an image-text pair that matches) and a set of negative pairs (image-text pairs that do not match), the contrastive loss encourages the positive pair to have a higher similarity than the negative pairs. The contrastive loss can be formulated as:
Here, N represents the number of negative pairs, and vTj denotes the text representation of the j-th negative pair (Equation 7).
The loss function aims to maximize the similarity between the positive image-text pair while minimizing the similarities between the positive pair and negative pairs. During training, the model optimizes the parameters of the image and text encoders to minimize the contrastive loss. This process enables the model to learn representations that associate images and their corresponding text descriptions. In summary, CLIP is a powerful model that combines image and text encoders to learn joint representations of visual and textual information. Its large-scale pretraining on image-text pairs enables it to capture the semantic relationship between these modalities. In the context of real-time feedback and guidance in sports training, CLIP enhances the multimodal robot's understanding and reasoning capabilities, facilitating personalized feedback and guidance based on the combination of visual and textual information.
3.4 Cross-Attention
Cross-Attention is a key component in models that handle multi-modal tasks, such as image captioning, visual question answering, and image-text matching (Kim et al., 2023). It enables the model to attend to relevant information from one modality (e.g., images) based on the input from another modality (e.g., text). The basic principle of Cross-Attention involves computing attention weights between elements in two different modalities and using these weights to combine the information effectively (Björkstrand et al., 2023).
Figure 4 is a schematic diagram of the principle of Cross-Attention.
Encoding: The image is typically encoded using a convolutional neural network (CNN), which extracts visual features from the image. The text description is encoded using a recurrent neural network (RNN) or a Transformer-based architecture, generating a sequence of hidden states. Query, Key, and Value: The hidden states from the text description serve as the query, while the visual features from the image act as the key and value. These query, key, and value representations are used to compute attention weights. Attention Calculation: The attention weights are computed by measuring the similarity between the query and key representations. This can be achieved through various methods, such as dot product, scaled dot product, or bilinear attention. The attention weights determine how much each visual feature should contribute to the final attended representation. Weighted Combination: The attention weights are used to weight the values (visual features) associated with each key. The weighted values are then combined to form the attended representation. This process allows the model to focus on the most relevant visual information based on the text query. Integration: The attended representation is integrated with the original text representation, typically through concatenation or element-wise addition. This integration step enables the model to capture the cross-modal interactions and create a fused representation that combines both text and visual information. The Cross-Attention mechanism plays a crucial role in multi-modal tasks by allowing the model to attend to relevant visual information conditioned on the textual input. It enables the model to align and associate the text and visual modalities, facilitating a comprehensive understanding and reasoning about the given input.
For example, in image captioning, the Cross-Attention mechanism helps the model generate descriptive captions by attending to relevant image regions while generating each word of the caption. In visual question answering, Cross-Attention allows the model to attend to specific image regions that are relevant to answering the question posed in the text. In image-text matching, Cross-Attention helps align and measure the similarity between image and text representations for tasks such as retrieval and ranking.
Let's consider two modalities, Modality A and Modality B, with their respective representations: Query (Q), Key (K), and Value (V).
The Cross-Attention mechanism involves the following steps:
Compute Attention Weights: The attention weights are calculated by measuring the similarity between the query representation (Q) and the key representation (K). One common approach is to use the dot product:
Here, dk represents the dimensionality of the key representation (K). The softmax function ensures that the attention weights sum up to 1 (Equation 8).
Weighted Combination: The attention weights are used to weight the values (V) associated with each key. The weighted values are then combined to obtain the attended representation:
The above Equation 9 represent a simplified version of Cross-Attention and assume single-head attention. In practice, multi-head attention is often employed to capture different aspects and provide richer representations. Cross-Attention allows the model to attend to relevant information in one modality based on the input from another modality. It enables the model to align and associate the information across modalities, facilitating tasks that involve multi-modal understanding, generation, and reasoning. Cross-Attention is a fundamental mechanism in multi-modal models that allows the model to attend to relevant information from one modality based on the input from another modality. It facilitates the fusion of text and visual information, enabling comprehensive understanding and reasoning in tasks involving multiple modalities.
4 Experiment
4.1 Datasets
This article uses the following four datasets:
OpenImages Dataset (Kuznetsova et al., 2020): OpenImages is a large-scale dataset consisting of annotated images from a wide range of categories. It contains over 9 million images with annotations for object detection, segmentation, and classification tasks. The dataset provides a diverse collection of visual data for training and evaluating computer vision models.
Objects365 Dataset (Shao et al., 2019): Objects365 is another comprehensive dataset that focuses on object detection and instance segmentation. It contains over 365 object categories, with more than 2 million labeled instances. The dataset is designed to cover a wide range of object classes and poses, providing a rich resource for training and evaluating object recognition models.
MSCOCO Dataset (Lin et al., 2014): MSCOCO (Microsoft Common Objects in Context) is a widely used benchmark dataset for object detection, segmentation, and captioning tasks. It consists of around 330,000 images, each annotated with object bounding boxes, segmentation masks, and image captions. MSCOCO offers a diverse set of images with multiple object instances and complex scenes, making it suitable for training and evaluating models in various visual tasks.
VG-Gap Dataset (Santana et al., 2015): VG-Gap is a dataset specifically focused on visual grounding and referring expression comprehension. It includes images from the Visual Genome dataset, accompanied by referring expressions that describe specific objects or regions within the images. The dataset is designed to facilitate research on understanding natural language instructions and grounding them to visual content.
4.2 Experimental details
In the experiment of our real-time feedback and guidance method for sports training based on a multimodal robot system, we utilized four widely recognized datasets: OpenImages, Objects365, MSCOCO, and VG-Gap, for training and validation of systems based on Vision Transformer (ViT), CLIP, and cross-attention mechanism. The training-validation split was set to 80% and 20% respectively. We designed two main experiments: metric comparison experiment and ablation experiment to evaluate and validate the performance and effectiveness of the systems. In the metric comparison experiment, we first established baseline models using traditional Convolutional Neural Networks (CNNs) and Long Short-Term Memory networks (LSTMs) as control groups for the same tasks. Subsequently, we deployed our multimodal system and focused on evaluating key performance metrics such as training time (in seconds), inference time (in milliseconds), model parameters (in millions), computational complexity (in billions of FLOPs), accuracy, AUC, recall, and F1 score. To ensure the experiment's accuracy, each model was run on the same hardware and software environment to eliminate the influence of external variables. Each model was trained and tested on an equal amount of data to ensure the comparability of results. Specifically, we utilized 8 A100 GPUs for training, employed the Adam optimizer, and set the following hyperparameters: learning rate of 0.001, batch size of 32, and 50 training epochs. We implemented the models using the Python programming language and the PyTorch framework. In the ablation experiment, we systematically removed key components from the system: first the cross-attention mechanism, then the CLIP module, and finally the Vision Transformer. We observed the impact of each modification on the model's performance. This approach helped us understand the contribution of each component to the overall system performance and identify indispensable parts in the system. Throughout the process, the aforementioned performance metrics were used to evaluate and quantify the importance and effectiveness of each component. Through these experiments, we gained detailed insights into the specific impact of different modules on the system's performance. We were also able to compare the efficiency and effectiveness of our approach in handling complex sports training scenarios with traditional methods. The in-depth analysis of the experimental results not only validated the effectiveness of our approach but also demonstrated the potential application value of multimodal interactive systems in real-time sports training guidance. Additionally, these experimental results provide valuable data support and practical experience for future research in this field, contributing to further optimization and development of more efficient and accurate training assistance systems.
Algorithm 1 shows the training process of the proposed method.
4.3 Experimental results and analysis
Table 1 presents the performance comparison between our proposed model and models from other researchers on the OpenImages and Objects365 datasets. This comparison experiment focuses on four main performance metrics: Accuracy, Recall, F1 Score, and AUC (Area Under the Curve), which collectively evaluate the overall performance of the models in classification tasks. Accuracy measures the proportion of correct predictions made by the model, Recall focuses on the proportion of relevant instances identified by the model out of all relevant instances, F1 Score is the harmonic mean of Precision and Recall, providing an overall performance assessment, while AUC measures the overall performance of the model in predicting different classes. The results demonstrate that our model outperforms other methods in all metrics, particularly exhibiting outstanding performance on the Objects365 dataset, showcasing its superior image parsing and classification capabilities. This can be attributed to our model's ability to effectively combine the characteristics of Vision Transformer and CLIP, better understanding image content and contextual information through cross-attention mechanisms.
Table 2 showcases the comparison of computational efficiency on the MSCOCO and VG-Gap datasets, covering model parameters, computational complexity (FLOPs), inference time, and training time. Parameters and FLOPs reflect the complexity of the model and the computational resources required at runtime, with lower values indicating a lighter and more efficient model. Inference time and training time are directly related to the practical application of the model, with lower inference time and training time indicating real-time and cost-effective deployment. Our model demonstrates excellent performance in these metrics as well, particularly showcasing significant advantages in inference time and training time, proving its efficiency and practicality in real-world deployment.
We compared our method with GPT-3.5 using the OpenAI API, and the results are presented in Tables 1, 2. Our model outperforms GPT-3.5 in key metrics such as accuracy, recall, Inference Time(ms) and Training Time(s), as evaluated on the OpenImages, Objects365, MSCOCO, and VG-Gap datasets. In Table 2, the inference time is reported for every 10 images. Therefore, an inference time of 192 ms corresponds to every 10 images, which means the inference time per frame is 19.2 ms. This translates to approximately 52 frames per second (FPS), meeting the real-time requirement of 25 FPS. Additionally, by applying pruning and distillation techniques to our algorithm, we further optimized the model to achieve close to 60 FPS without significant loss in performance. Hence, our method satisfies the real-time demands in practical applications.
Table 3 focuses on the ablation experiment analyzing the impact of the Cross-Attention Module on the OpenImages and Objects365 datasets. The experimental setup involves removing or modifying the Cross-Attention Module and observing the changes in Accuracy, Recall, F1 Score, and AUC. AM (Attention Module), Seif-AM (Self-Attention Module), and Dynamic-AM (Dynamic Attention Module) represent different configurations of the Cross-Attention Module. By comparing these configurations, we discovered that the complete Cross-Attention Module significantly enhances all performance metrics, demonstrating its crucial role in integrating visual and textual information and improving the overall recognition capability of the model. Our model experiences a performance decline when the Cross-Attention mechanism is removed, but even in this case, it still outperforms other configurations, showcasing the robustness of our approach.
Table 4 further explores the impact of the Cross-Attention Module on computational efficiency, covering the MSCOCO and VG-Gap datasets. The experimental results show that after removing or modifying the Cross-Attention Module, our model performs best in terms of model parameters, computational complexity, inference time, and training time. This result not only reaffirms the efficiency of our model but also highlights the importance of the Cross-Attention mechanism in optimizing the model's computational path and reducing unnecessary computations. Overall, these experimental results thoroughly demonstrate the superiority of our proposed approach in handling complex multi-modal data, making it suitable for applications in scenarios such as sports training that require fast and accurate feedback.
Table 4. Computational efficiency in ablation study of Cross-Attention module on MSCOCO and VG-Gap datasets.
Conducting validation in a real-world physical environment can indeed enhance the persuasiveness of the paper. However, we currently face some limitations and challenges. Firstly, high-quality video recording and processing require appropriate hardware devices, including high-definition cameras and powerful computational resources. We are actively seeking resource support to ensure access to the necessary equipment and computing capabilities. Secondly, it is necessary to establish a suitable video recording experimental setup to ensure data quality and consistency. We are planning and designing a standardized recording environment to capture high-quality motion training videos while minimizing the impact of environmental variables on experimental results. Additionally, self-recorded videos may introduce additional data processing and annotation work, increasing the complexity and workload of the experiments. To address this issue, we plan to develop semi-automated annotation tools and data preprocessing workflows to improve efficiency and reduce the workload. Lastly, factors such as lighting, background, and motion complexity in self-recorded videos may differ significantly from public datasets. This may require additional adjustments and optimizations to the model. We will fine-tune the model based on self-recorded videos to ensure its high performance and accuracy in different environments and conditions. In future work, we will continue to overcome these challenges and gradually achieve analysis and validation of self-recorded videos. We will report relevant results in subsequent research. Once again, thank you for the valuable suggestions provided by the reviewer, as they will help us further improve the research and enhance its practical value.
5 Conclusion and discussion
This research addresses the issue of real-time feedback and guidance in sports training and proposes a multimodal robotic system named CAM-Vtrans: Real-time Sports Training Utilizing Multi-modal Robot Data, which combines Vision Transformer (ViT), CLIP, and Cross-Attention mechanisms. This method leverages advanced deep learning techniques to process and integrate complex visual and textual data, aiming to provide more accurate and effective training feedback. The experiments are divided into performance comparison and ablation experiments, conducted on the OpenImages, Objects365, MSCOCO, and VG-Gap datasets. The results demonstrate that our model outperforms other state-of-the-art models in key metrics such as accuracy, recall, F1 score, and AUC. Additionally, it exhibits excellent computational efficiency, validating the effectiveness and practicality of the proposed approach.
Despite the positive outcomes, there are still some limitations to be addressed. Firstly, although the model performs well on multiple datasets, its generalization to other unseen types of sports activity data has not been validated, and further testing and optimization are needed in a broader range of sports activities. Secondly, while the current model exhibits real-time processing capability, there is still room for improvement in scenarios requiring extreme real-time performance. Future research should focus on reducing inference time and enhancing processing speed. Additionally, exploring the model's application across a wider array of sports activities and incorporating more diverse and complex datasets will be critical for ensuring its robustness and versatility. Further development of adaptive feedback mechanisms that tailor guidance to the specific needs of different sports disciplines could also enhance the system's effectiveness and user experience.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Author contributions
HL: Conceptualization, Formal analysis, Investigation, Methodology, Project administration, Resources, Supervision, Visualization, Writing – original draft. LS: Data curation, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Validation, Writing – review & editing. SG: Conceptualization, Data curation, Investigation, Methodology, Project administration, Resources, Visualization, Writing – original draft.
Funding
The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Aresta, S., Bortone, I., Bottiglione, F., Di Noia, T., Di Sciascio, E., Lofù, D., et al. (2022). Combining biomechanical features and machine learning approaches to identify fencers levels for training support. Appl. Sci. 12:12350. doi: 10.3390/app122312350
Bakthavatchalam, K., Karthik, B., Thiruvengadam, V., Muthal, S., Jose, D., Kotecha, K., et al. (2022). IoT framework for measurement and precision agriculture: predicting the crop using machine learning algorithms. Technologies 10:13. doi: 10.3390/technologies10010013
Bergamasco, M., Bardy, B., and Gopher, D. (2012). Skill Training in Multimodal Virtual Environments. CRC Press. Available at: https://content.iospress.com/articles/work/wor0452
Björkstrand, D., Sullivan, J., Bretzner, L., Loy, G., and Wang, T. (2023). Cross-attention masked auto-encoder for human 3D motion infilling and denoising. Available at: https://papers.bmvc2023.org/0609.pdf
Chen, M.-Y., Chen, T.-H., and Lin, S.-H. (2020). “Using convolutional neural networks to forecast sporting event results,” in Deep Learning: Concepts and Architectures, 269–285.
Dirgová Luptáková, I., Kubovčík, M., and Pospíchal, J. (2022). Wearable sensor-based human activity recognition with transformer model. Sensors 22:1911. doi: 10.3390/s22051911
Dobrzycki, A. D., Bernardos, A. M., Bergesio, L., Pomirski, A., and Sáez-Trigueros, D. (2023). Exploring the use of contrastive language-image pre-training for human posture classification: insights from yoga pose analysis. Mathematics 12:76. doi: 10.3390/math12010076
Dong, R., Zhang, X., Li, H., Masengo, G., Zhu, A., Shi, X., et al. (2024). Eeg generation mechanism of lower limb active movement intention and its virtual reality induction enhancement: a preliminary study. Front. Neurosci. 17:1305850. doi: 10.3389/fnins.2023.1305850
Du, X., Fuqian, X., Hu, J., Wang, Z., and Yang, D. (2021). “Uprising e-sports industry: machine learning/AI improve in-game performance using deep reinforcement learning,” in 2021 International Conference on Machine Learning and Intelligent Systems Engineering (MLISE), 547–552. Available at: https://ieeexplore.ieee.org/abstract/document/9611712
Faria, B. M., Dias, D., and Reis, L. P. (2016). “Multimodal interaction robotic devices in a simulated environment,” in 2016 11th Iberian Conference on Information Systems and Technologies (CISTI) (IEEE), 1–6. Available at: https://ieeexplore.ieee.org/abstract/document/7521429
Hauri, S., and Vucetic, S. (2023). “Group activity recognition in basketball tracking data–neural embeddings in team sports (nets),” in ECAI 2023 (IOS Press), 1012–1019. Available at: https://ebooks.iospress.nl/doi/10.3233/FAIA230373
Herman, J., Francis, J., Ganju, S., Chen, B., Koul, A., Gupta, A., et al. (2021). “Learn-to-race: a multimodal control environment for autonomous racing,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 9793–9802. Available at: https://openaccess.thecvf.com/content/ICCV2021/html/Herman_Learn-To-Race_A_Multimodal_Control_Environment_for_Autonomous_Racing_ICCV_2021_paper.html
Himabindu, M., Revathi, V., Gupta, M., Rana, A., Chandra, P. K., and Abdulaali, H. S. (2023). “Neuro-symbolic AI: integrating symbolic reasoning with deep learning,” in 2023 10th IEEE Uttar Pradesh Section International Conference on Electrical, Electronics and Computer Engineering (UPCON), Vol. 10, 1587–1592. Available at: https://ieeexplore.ieee.org/abstract/document/10434380
Hu, Y., Wu, X., Geng, P., and Li, Z. (2019). Evolution strategies learning with variable impedance control for grasping under uncertainty. IEEE Trans. Indus. Electron. 66, 7788–7799. doi: 10.1109/TIE.2018.2884240
Ilager, S., De Maio, V., Lujic, I., and Brandic, I. (2023). “Data-centric edge-AI: a symbolic representation use case,” in 2023 IEEE International Conference on Edge Computing and Communications (EDGE), 301–308.
Insuasti, J., Roa, F., and Zapata-Jaramillo, C. M. (2023). Computers' interpretations of knowledge representation using pre-conceptual schemas: an approach based on the bert and llama 2-chat models. Big Data Cogn. Comput. 7:182. doi: 10.3390/bdcc7040182
Jia, H., Ren, C., Hu, Y., Chen, Y., Lv, T., Fan, C., et al. (2020). “Mastering basketball with deep reinforcement learning: an integrated curriculum training approach,” in Proceedings of the 19th International Conference on Autonomous Agents and MultiAgent Systems, 1872–1874. Available at: https://aamas.csc.liv.ac.uk/Proceedings/aamas2020/pdfs/p1872.pdf
Jin, M., Ma, Z., Jin, K., Zhuo, H. H., Chen, C., and Yu, C. (2022). “Creativity of AI: automatic symbolic option discovery for facilitating deep reinforcement learning,” in Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 36, 7042–7050. Available at: https://ojs.aaai.org/index.php/AAAI/article/view/20663
Jose, S., Maliackal, A. T., Sukumaran, A., Cyriac, J., and Varkey, R. M. (2023). “Machine learning technique to analyze the health condition of athletes and predict their performance,” in 2023 International Conference on Circuit Power and Computing Technologies (ICCPCT) (IEEE), 354–359. Available at: https://ieeexplore.ieee.org/abstract/document/10245401
Kim, S., Ahn, D., and Ko, B. C. (2023). “Cross-modal learning with 3D deformable attention for action recognition,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 10265–10275. Available at: https://openaccess.thecvf.com/content/ICCV2023/html/Kim_Cross-Modal_Learning_with_3D_Deformable_Attention_for_Action_Recognition_ICCV_2023_paper.html
Koh, J. Y., Fried, D., and Salakhutdinov, R. R. (2024). “Generating images with multimodal language models,” in Advances in Neural Information Processing Systems, 36. Available at: https://proceedings.neurips.cc/paper_files/paper/2023/hash/43a69d143273bd8215578bde887bb552-Abstract-Conference.html
Kuznetsova, A., Rom, H., Alldrin, N., Uijlings, J., Krasin, I., Pont-Tuset, J., et al. (2020). The open images dataset v4: unified image classification, object detection, and visual relationship detection at scale. Int. J. Comput. Vis. 128, 1956–1981. doi: 10.1007/s11263-020-01316-z
Li, Z., Zhao, T., Chen, F., Hu, Y., Su, C.-Y., and Fukuda, T. (2018). Reinforcement learning of manipulation and grasping using dynamical movement primitives for a humanoidlike mobile manipulator. IEEE ASME Trans. Mechatron. 23, 121–131. doi: 10.1109/TMECH.2017.2717461
Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., et al. (2014). “Microsoft coco: common objects in context,” in Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13 (Springer), 740–755. Available at: https://link.springer.com/chapter/10.1007/978-3-319-10602-1_48
Liu, L. (2022). Sports video motion direction detection and target tracking algorithm based on convolutional neural network. Wireless Commun. Mobile Comput. 2022:5760758. doi: 10.1155/2022/5760758
Lu, Z., Zhang, Y., Li, S., and Zhou, P. (2024). Botulinum toxin treatment may improve myoelectric pattern recognition in robot-assisted stroke rehabilitation. Front. Neurosci. 18:1364214. doi: 10.3389/fnins.2024.1364214
Miyazawa, K., Kyuragi, Y., and Nagai, T. (2022). Simple and effective multimodal learning based on pre-trained transformer models. IEEE Access 10, 29821–29833. doi: 10.1109/ACCESS.2022.3159346
Morciano, G., Zingoni, A., and Calabr, G. (2023). “Prediction of football players' performance indicators via random forest algorithm,” in 2023 IEEE International Conference on Metrology for eXtended Reality, Artificial Intelligence and Neural Engineering (MetroXRAINE), 201–205. Available at: https://ieeexplore.ieee.org/abstract/document/10405671
Pan, L., Zhao, L., Song, A., Yin, Z., and She, S. (2019). A novel robot-aided upper limb rehabilitation training system based on multimodal feedback. Front. Robot. AI 6:102. doi: 10.3389/frobt.2019.00102
Papadakis, A., and Spyrou, E. (2024). A multi-modal egocentric activity recognition approach towards video domain generalization. Sensors 24:2491. doi: 10.3390/s24082491
Santana, R., Mendiburu, A., and Lozano, J. A. (2015). Multi-view classification of psychiatric conditions based on saccades. Appl. Soft Comput. 31, 308–316. doi: 10.1016/j.asoc.2015.02.038
Shao, S., Li, Z., Zhang, T., Peng, C., Yu, G., Zhang, X., et al. (2019). “Objects365: a large-scale, high-quality dataset for object detection,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 8430–8439. Available at: https://openaccess.thecvf.com/content_ICCV_2019/html/Shao_Objects365_A_Large-Scale_High-Quality_Dataset_for_Object_Detection_ICCV_2019_paper.html
Tang, Q., Liang, J., and Zhu, F. (2023). A comparative review on multi-modal sensors fusion based on deep learning. Signal Process. 213:109165. doi: 10.1016/j.sigpro.2023.109165
Tjondronegoro, D., and Chen, Y.-P. P. (2006). “Using decision-tree to automatically construct learned-heuristics for events classification in sports video,” in 2006 IEEE International Conference on Multimedia and Expo (IEEE), 1465–1468. Available at: https://ieeexplore.ieee.org/abstract/document/4036887
Wang, M., and Liang, Z. (2023). Cross-modal self-attention mechanism for controlling robot volleyball motion. Front. Neurorobot. 17:1288463. doi: 10.3389/fnbot.2023.1288463
Wang, Z., Li, X., and Wang, G. (2024). Exploring wireless device-free localization technique to assist home-based neuro-rehabilitation. Front. Neurosci. 18:1344841. doi: 10.3389/fnins.2024.1344841
Wu, X., Zhu, X., Zhao, Y., and Dai, X. (2022). Differentiable fuzzy ℒ𝒞: a neural-symbolic representation language for symbol grounding. arXiv preprint arXiv:2211.12006.
Yagin, F. H., Hasan, U. C., Clemente, F. M., Eken, O., Badicu, G., and Gulu, M. (2023). Using machine learning to determine the positions of professional soccer players in terms of biomechanical variables. Proc. Inst. Mech. Eng. Part J. Sports Eng. Technol. 17543371231199814. doi: 10.1177/17543371231199814
Yang, Y., Zhuang, Y., and Pan, Y. (2021). Multiple knowledge representation for big data artificial intelligence: framework, applications, and case studies. Front. Inform. Technol. Electron. Eng. 22, 1551–1558. doi: 10.1631/FITEE.2100463
Zhao, T., Deng, M., Li, Z., and Hu, Y. (2020). Cooperative manipulation for a mobile dual-arm robot using sequences of dynamic movement primitives. IEEE Trans. Cogn. Dev. Syst. 12, 18–29. doi: 10.1109/TCDS.2018.2868921
Zheng, Y., Song, Q., Liu, J., Song, Q., and Yue, Q. (2020). Research on motion pattern recognition of exoskeleton robot based on multimodal machine learning model. Neural Comput. Appl. 32, 1869–1877. doi: 10.1007/s00521-019-04567-1
Keywords: assistive robotics, human-machine interaction, balance control, movement recovery, vision-transformer, CLIP, cross-attention
Citation: LinLin H, Sangheang L and GuanTing S (2024) CAM-Vtrans: real-time sports training utilizing multi-modal robot data. Front. Neurorobot. 18:1453571. doi: 10.3389/fnbot.2024.1453571
Received: 23 June 2024; Accepted: 25 July 2024;
Published: 11 October 2024.
Edited by:
Yingbai Hu, The Chinese University of Hong Kong, ChinaReviewed by:
Dan Huang, South China University of Technology, ChinaWing Yin Ng, CUHK Department of Surgery, Hong Kong SAR, China
Copyright © 2024 LinLin, Sangheang and GuanTing. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Song GuanTing, 19504356665@163.com