Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

TRI-POSE-Net: Adaptive 3D human pose estimation through selective kernel networks and self-supervision with trifocal tensors

  • Nabeel Ahmed Khan,

    Roles Data curation, Methodology, Software, Writing – original draft

    Affiliation Center For AI and Big Data, Namal University, Mianwali, Pakistan

  • Aisha Ahmed Alarfaj,

    Roles Data curation, Formal analysis, Investigation, Resources

    Affiliation Department of Information Systems, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia

  • Ebtisam Abdullah Alabdulqader,

    Roles Methodology, Project administration, Software, Validation, Visualization, Writing – review & editing

    Affiliation Department of Information Technology, College of Computer and Information Sciences, King Saud University, Riyadh, Saudi Arabia

  • Nuha Zamzami,

    Roles Conceptualization, Formal analysis, Investigation, Visualization, Writing – review & editing

    Affiliation Department of Computer Science and Artificial Intelligence, College of Computer Science and Engineering, University of Jeddah, Jeddah, Saudi Arabia

  • Muhammad Umer,

    Roles Funding acquisition, Methodology, Software, Supervision, Writing – review & editing

    Affiliation Department of Computer Science & Information Technology, The Islamia University of Bahawalpur, Bahawalpur, Pakistan

  • Nisreen Innab ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Resources, Visualization

    Ninnab@um.edu.sa (NI); taihoonn@chonnam.ac.kr (THK)

    Affiliation Department of Computer Science and Information Systems, College of Applied Sciences, AlMaarefa University, Diriyah, Riyadh, Saudi Arabia

  • Tai-Hoon Kim

    Roles Formal analysis, Funding acquisition, Project administration, Resources, Supervision

    Ninnab@um.edu.sa (NI); taihoonn@chonnam.ac.kr (THK)

    Affiliation School of Electrical and Computer Engineering, Yeosu Campus, Chonnam National University, Yeosu-si, Jeollanam-do, Republic of Korea

Abstract

Accurate and flexible 3D pose estimation for virtual entities is a strenuous task in computer vision applications. Conventional methods struggle to capture realistic movements; thus, creative solutions that can handle the complexities of genuine avatar interactions in dynamic virtual environments are imperative. In order to tackle the problem of precise 3D pose estimation, this work introduces TRI-POSE-Net, a model intended for scenarios with limited supervision. The proposed technique, which is based on ResNet-50 and includes integrated Selective Kernel Network (SKNet) blocks, has proven to be efficient for feature extraction customised specifically to pose estimation scenarios. Furthermore, trifocal tensors and their trio-view geometry allow us to generate 3D ground truth poses from 2D poses, resulting in more refined triangulations. Through the proposed approach, the 3D poses can be estimated from a single 2D RGB image. Moreover, the proposed approach was evaluated on the HumanEva-I dataset yielding a Mean-Per-Joint-Position-Error (MPJPE) of 47.6 under self-supervision and an MPJPE of 29.9 under full supervision. In comparison with the other works, the proposed work has performed well in the self-supervision paradigm.

1 Introduction

In the framework of the developing human 3d pose estimation, the goal of endowing machines with vision capabilities akin to those of humans is becoming increasingly prominent. Advances in computer vision and human-computer interaction illustrate the importance of accurately perceiving and interpreting the behaviors of virtual entities. Human pose estimation was first introduced in response to [1], which fits in well with the virtual environment’s need for realistic avatars and entities that can move around and interact with virtual environments in a believable way. Similar to its practical uses in sports science, rehabilitative training, video surveillance, and behavioral identification, human posture estimation plays a significant role in the in human 3d pose estimation [2, 3]. Although conventional 3D human pose estimation usually uses depth data from specialist devices such as Microsoft Kinect, the realistic environment requires more easily available options. The usefulness and affordability of using standard RGB cameras become clear in this situation. The estimation of virtual human positions based on regular RGB pictures is a study area that is receiving significant attention and practical significance in the virtual environment and beyond. Both industry and academics are actively investigating this topic. This change is in line with easily accessible and immersive virtual experiences, as well as addressing the drawbacks and high expenses of conventional methods. Two broad categories of deep learning algorithms and traditional schemes can classify the most commonly used 3D human pose estimators. Conventional methods, used early in the progress, were based on a deformable part model and a body part detector with graph structure. These techniques employed a graph model with scripted features, such as HOG and SIFT, to described the spatial relationships [1, 4]. But, these methods had their limitations because they were dependent of the appearance of the image and failed to fully exploit image features.

A number of artificial neural networks are now employed in identification and attributes estimation in computer vision’s big data age. The Human posture estimation research has shifted from classical to deep learning approaches, whereby pre-dominnantly CNN were used to learn humans pose data directly from images. Works such as [5, 6] revealed how perfectly deep neural networks perform in order to predict human location in a three-dimensional space from a single picture. Even with the impressive successes of current approaches, there are still significant obstacles to overcome in order to deploy 3D human pose estimation at the application level. The commonly utilised 3D datasets, which are single-image backgrounds and 3D ground truth labels collected by motion capture systems (MOCAPs) in controlled environments, are expensive. Models trained on such datasets confront difficulty in extending to outside and complicated situations with body part occlusion. In this paper, we address this problem by proposing a network model for 3D human pose estimation that uses a self-supervised training approach to reliably estimate 3D human posture from photos lacking 3D labels.

The name ‘Tri-POSE-Net’ reflects the core principles of the model: the use of trifocal tensors, trio-view geometry, and three modules added into the architecture—ResNet-50, SKNet blocks, and trifocal tensors for proper 3D pose computation. The proposed technique uses ResNet-50 for feature learning while the SKNet blocks available dynamically adjust to pose estimation scenario using different scales and orientations. The trifocal tensors are important because they help in generating ground truth poses from 2D poses across three views resulting to much accurate triangulations. In this manner, TRI-POSE-Net efficiently estimates 3D poses from a single 2D RGB image while the structure of this work includes the threefold strategy. All abbreviations utilized in this research work are shared in Table 1. The major contributions of this paper are listed below.

  1. A model named TRI-POSE-Net is proposed under the weakly-supervised 3D pose estimation paradigm. The model consists of a ResNet-50 baseline modified using two SKNet blocks, introduced with the intent of dynamically allocating the receptive field sizes and the global context based adaptive feature selection for a refined pose estimation.
  2. Transfer learning is thoroughly implemented as part of the first training regimen for the acquisition of 2D poses, by training the 2D variant of the proposed TRI-POSE-Net on in-the-wild dataset. This parameter sharing amplifies the generalization capabilities of the model whilst generating 2D poses on HumanEva-I dataset.
  3. The use of trifocal tensors in this context primarily aids in capturing global spatial relationships. By establishing geometric constraints and correspondences between three different views, trifocal tensors provide a more comprehensive understanding of the overall 3D structure, which goes beyond the local spatial details. This global spatial information is crucial for accurately reconstructing 3D poses from multiple viewpoints.

2 Related works

This section describes some state-of-the-art closely related research works related to the topic of human 3D pose estimation. We have utilized trifocal tensors for weakly and self-supervised 3D human pose estimation. We have categorized related work into two subsections to separate the approaches.

Work has progressively focused on developing techniques in the relevant field of study for 3D human pose estimation primarily concentrating on modality oriented methods. To this end, Lee et al. [7] proposed, in their pioneering study, an MCMC method specifically designed for approximating the 3D HP of the human upper body. This allowed to show how the proposed method works and that it can be successfully used for images of different nature, which proved the applicability of the new approach. Examining further the progressive aspect of modalities, Jung et al. [8] developed a new approach to the 3D human pose estimation, with reference to depth images. One of their methods reconceptualised the problem as a detection of desired body joints, which also demonstrated a proper appreciation of the complex processes required to obtain depth-based human poses. Based on this, Wu et al extended the work in the area of single image based human pose estimation with the use of depth maps and they proposed the model-based recursive matching (MRM) technique. This method placed the field on a different level of professionalism and was an input to the discussion on the existing techniques of 3D human pose estimation.

To this end, Lu et al. (2024) introduce the AvatarPose, a new approach to estimate the 3D pose in the context of proximal interactions from inadequate multiview videography. This approach uses the prior Personalized Implicit Neural Avatars, which are beneficial to enhance the robustness and the correspondence’s accuracy of body pose estimation; especially when there are occluded or contacted areas. The avatars are reconstructed with the help of layered volume rendering, it permits direct optimization of 3D poses according to the color and silhouette rendering losses. This method is superior to other models in datasets with tightly correlated objects with each other [9, 10]. In this work, the authors introduce PoseResNet for self-supervised learning of 3D human pose, without the use of 3D ground truth labels. The model adopts ResNet-50 as its base model and adopts CBAM as well as its proposed multi-scale feature fusion module named WASP. A deconvolution network is then employed to generate volume heatmaps which are then passed through a soft argmax function to estimate joint coordinates. This method proves better performance where, with an average of calculating per joint position error (MPJPE), it was 74. 6 mm with removing the need for 3D labels [11]. To the best of the authors’ knowledge, Zhou et al. proposed Trifocal PoseNet in 2024 as a model targeting at learning 3D poses with a small amount of annotations. This model uses trifocal tensors to make good use of the geometrical constraints and the relationships between the three views to improve the triangulation. Selective kernel network is proposed to employ within ResNet-50 framework, in order to capture the features of pose and optimize them. This presented approach shows how trifocal geometry can be used to synthesize poses from 2D images even with little requirement for supervision [12].

Similarly, another set of works strictly bounded to the realm of 3D human pose estimation has focused on extending the applicability of these techniques in particular application areas in order to demonstrate their efficacy. Lie et al. [13] switch their focus on research to the important area of elderly health monitoring. Their proposal was to extract 3D skeletons as a way of surveillance of daily living among the elderly. This application-oriented approach is in line with the flexibility and the practicality of the 3D human pose estimation. Adding to the list of application domains of AE, Moon et al. [14] put forward a human pose refinement network for multi-person pose estimation. This work while being theoretical also dealt with issues to do with computation costs and elimination of pipeline effects, that are relevant in the real world application of these technologies. Moreover, Cong et al. [15] indicated a method that uses monocular camera and single LiDAR for 3D multi-person pose estimation in large-scale scenes. Stressing more on the indispensability of depth information, this work ensured that 3D pose estimation skills are viable in all real world settings hence not limited to simple settings. Altogether, the works considered herein form a wide survey of 3D human pose estimation with focus on the progressive modalities and a high level of practical concern, which proves the potential and applicability of the techniques.

2.1 Weakly and self-supervised approaches in 3D human pose estimation

A number of works are devoted to weakly and self-supervised methods for 3D human pose estimation. We also refer to papers [1618], where authors studied how to use weakness and self-supervision for overcoming some difficulties connected with receipt of 3D ground truth labels. In a way, self-supervised and weakly supervised methods sketched in this paper provide a partial solution to this problem that the fully supervised approach [1921] addressing direct estimation of 3D joint coordinates from images using regression networks. Pavlakos et al. [22] utilized a pictorial structure model to derive global pose configurations from key-point heat maps in multi-view images. However, their method required full camera calibration and a keypoint detector for generating 2D heat maps. Rhodin et al. [16] used multi-view consistency restraints to guide a network, minimizing the need for extensive 3D ground truth data to avoid pose collapse issues. Nonetheless, the scarcity of in-the-wild 3D ground truth data remains a limitation. Drover et al. [17] introduced an adversarial framework for weak supervision, learning 3D structures solely from random 2D projections. Notably, their method eliminated the need for correspondence between 2D and 3D points but still relied on 3D labels. Zhou et al. [23] integrated 2D and 3D data, applying a Euclidean loss for 3D data and proposing a weak supervision loss based on 2D annotations and prior knowledge of the human skeleton. Despite improving applicability to outdoor images, their approach required both 2D and 3D labels. Wandt et al. [24] pioneered a generative adversarial network with three components, including a pose and camera estimation network, a critic network, and a re-projection network. Their method was also accompanied by another feature: camera estimation network and re-projection layer for re-projection of the estimated 3D pose to 2D space.

Rhodin et al. [25] proposed a weakly supervised method to transfer the body representation from the 3D space that is adapted to the geometric analyses of the multiple-view images. In their Self-supervised learning paper, Kocabas et al. [26] introduced Epipolar-Pose which doesn’t require 3D ground truth data or cameras extrinsics information instead it uses epipolar geometry to produce annotations from key-points. A self-supervised method of 3D human pose estimation from monocular images was presented by Li et al. [27, 28]. Altogether, these works help the investigation of the methods that need little control and supervision, which respond to issues related to the annotated data availability. Extending the weak supervision topic further, Wu et al. [29] introduced a work on weakly supervised 3D human pose estimation using the transfer of 2D HPE annotation information from large scale RGB datasets to the 3D task. Roy et al. [30] discussed a supervised method making use of multi-view geometrical restraints which can be considered as a form of self-supervision where the labelled data is limited. Cha et al. [31] proposed a new self-supervised learning paradigm for 3D human pose and shape recovery in single shot 2D images. Altogether, these works are connected to the investigation of the self-supervised learning frameworks and present progress in handling supervision issues in 3D pose estimation.

3 Methodology

In the method described in this study, self-supervision is used to train a CNN for the purpose of 3D human pose estimation. It is important to underline that this training process subdivides into two different stages. The first of these is a basic 2D human pose estimation network which is kept fixed during the training process. The key idea of this module is that of convolutional networks which are optimized for the purpose of producing 2D pose estimates from images, and is built using a variant of ResNet-50 architecture. The second training model involved in the training pipeline is a 3d pose estimation model. Similar to the previous model, this is a derivation of the modified ResNet-50 architecture and goes through training in a weakly supervised manner. The training leverages 3D-pose ground truths, which are derived by post processing the 2D-poses through epipolar geometry techniques.

3.1 TRIPOSE-NET

The first baseline network adopted in the TRIPOSE-NET module is ResNet-50 due to its ability to mitigate on vanishing gradients which is a common problem in deep convolutional network. The nature of these vanishing gradients is that they reduce the effectiveness of convolutional networks moving to the deeper layers while it seems that more layers should be better at capturing finer details. The possibility of degradation in performance especially in deep networks is also prevented by residual connections which enable the gradients to flow through the depth of the network thereby making the optimisation possible. Owing to its versatility across a plethora of tasks, the Residual Network (ResNet) is an optimal choice as the backbone for these pose estimation pipelines, providing liberty to extract abstract and densely intricate features by virtue of its deep and robust distinctive features. To elevate the performance of the ResNet-50 module, an SKNet Block is introduced at the end of the 3rd convolution block and at the tail-end of the feature extractor, i.e., just before the global average pooling layer.

This forms an effective feature extraction module named SK-ResNet 50, as illustrated in Fig 1. The addition of the SKNet block at the deeper layers, primarily the bottleneck blocks, assures that more global context is captured as the receptive field is larger in this region and that the layers have a high spatial resolution. Additionally, the placement of the SKNet block before the global average pooling allows for adaptive feature selection on the basis of global context prior to the final prediction. Furthermore, using SKNet in ResNet-50 augments the model’s capabilities in dealing with multi-scale features by acting as an auxiliary aid in selecting the appropriate kernel sizes for capturing both fine-grained and global details to achieve optimal performance in pose estimation.

thumbnail
Fig 1. Modified feature extraction block (ResNet-50) using selective kernel network.

https://doi.org/10.1371/journal.pone.0310831.g001

Initially, a deconvolution layer is used to project the extracted feature from the 2D pose estimation module as Heat-Maps denoted by Mh, where . Here (m, n) represent the dimensions of Mh and to acquire the 2D key-point positions (x, y) from the respective maps Mh for all nodes N such that , we apply an argmax function ν.

The resultant human poses A(i, j) are in two-dimensional representation, whereas the acquisition of 3D poses requires supervision under a 3-dimensional paradigm. The annotation of 3D poses for training endeavors is both time-consuming and resource-intensive. Herein, we propose a weakly supervised method, utilizing the 2D poses A(i, j) from initial training and epipolar geometry to triangulate the 3D correspondences. The volumetric key-point mappings given by are cached and taken as ground truths with pseudo labels to train the TRIPOSE-NET tailored to 3D pose estimations C(i, j, k), significantly curbing the need for pre-annotation of the data for 3D pose estimates.

3.2 Selective kernel network

Selective Kernel Network (SKNet) [32] emulates the dynamic modulation of receptive fields of visual-cortical neurons based on stimuli. This emulation allowed the extension of CNN with the integration of adaptive kernel sizing based on the input information. In SKNet, a Selective Kernel (SK) unit containing multiple branches with varying kernel sizes is fused in a way that ascertains the appropriate selection of the receptive field size, as visually explained in Fig 2. This adaptability equips the model with the ability to focus only on the most informative regions, akin to an attention mechanism. With proficient multi-scale feature integration owing to the dynamic allocation of receptive field size, SKNet extracts both coarse-level and fine-grained features imperative for effective pose estimation [33]. Additionally, the selective kernel feature provides a more expressive feature representation, with different parts of the network specializing in capturing diverse features. Relationships between features are meaningfully captured, resulting in improved pose estimation. A task where the preservation of the relationship between the joints is a significant facet. Within the proposed TRIPOSE-NET architecture, one of the SK units is added before the 4th convolution block of ResNet-50, and a second SK unit is added before the Global-Average-Pooling Layer, as shown in Fig 1. The 1st SK unit is dual-branched with two kernels of sizes 3x3 and 5x5. The 5x5 kernel is actually a dilated version of a 3x3 kernel with a dilation factor of 2. Dilated convolutions [34], as shown in Fig 3, help capture contextual information by introducing gaps within the elements of the filter, effectively reducing the parameter count and enhancing the network’s global awareness. The 2nd SK unit consists of three branches with kernel sizes of 3x3, 5x5, and 7x7. The first kernel is without dilation, while the other two are 3x3 kernels dilated with factors 2 and 3.

thumbnail
Fig 3. Visual illustration of dilated convolution kernel and its receptive field.

https://doi.org/10.1371/journal.pone.0310831.g003

The SK unit is segregated into three operations, namely Split, Fuse, and Select. Within the Split operation, the feature map Fm is propagated to the SK module, where two transformations and are performed. Each transformation involves a depth-wise convolution, followed by batch normalization, and finally a ReLU activation, in this sequence. For the first integration of the SK block within the ResNet-50 module, a dual-branch transformation and is used with kernel sizes 3x3 and 5x5, respectively. While the second embedding of the SK unit is based on a three-branched transformation , , and , each branch with a varying receptive field size as shown in Fig 2. The respective transformations within the Split operation produce outputs , , , where n represents the number of transformations.

The Fuse operation involves the element-wise summation of each transformation, which in our case involves both the two-branched U1 and three-branched U2 transformations, given by: (1) (2)

The outputs U1 and U2 from the fused transformations are passed through a Global Average Pooling Operation to obtain channel-wise global information statistics. The Pooling Operation Gp is applied across the spatial dimensions W × H for each element e of the output vector . The pooling operation is given as (3)

The output vector Pe is further compressed to obtain a compact vector c, by reducing the dimensionality via a fully connected (fc) layer, ensuring optimal efficiency in feature selection. (4)

In the above equation, φ represents the ReLU activation, ω represents Batch Normalization, and , where d is the depth of the vector.

Similarly, the Select operation selects different spatial scale information by the application of soft attention across the channels, guided by the compact vector c. Softmax is applied channel-wise such that for a two-branched network U1, the mathematical operation to extract the attention score is given by (5)

The final feature map Fe is given by: (6) where, (7)

The above Eq (7) can be extended for a three-branched network U2, used in this study, by calculating ce through the third matrix C, acquiring the final feature map of U2 as (8)

3.3 Dataset description

Existing 3D pose estimation datasets face a myriad of challenges, including limited diversity, ambiguous depth estimates, training data limitations, and predominantly indoor and controlled setups. These limitations impede progress in advancing 3D pose estimation algorithms to tackle the complexities of real-world scenarios. To somewhat bridge the gap in environmental diversity, this study utilizes two datasets, namely MPII and HumanEva-I, to familiarize the 3D pose estimation model with outdoor conditions. The MPII 2D Human Pose dataset [35] is widely used in computer vision due to its diverse and “in-the-wild” nature. This dataset contains around 25,000 photos capturing a wide range of human poses in different real-world scenarios. Each image is meticulously annotated with 16 different body joints, covering important anatomical regions.

The HumanEva-I dataset [36, 37] was created to evaluate and assess the performance of 3D pose estimation algorithms. It focuses on scenarios involving dynamic movements of multiple subjects and is annotated with 3D ground truth keypoints for each frame. It includes 15 anatomically significant joints such as the pelvis, hips, ankles, and thorax. The dataset includes seven video sequences—four in grayscale and three in color—featuring four annotated individuals engaged in five distinct actions. The MPII dataset is used for the initial training of the 2D pose estimation module, while HumanEva-I is employed for weakly supervised training of the 3D pose estimation pipeline. The MPII dataset’s skeleton topology contains 16 annotated joints, whereas HumanEva-I’s skeleton topology includes 15. Therefore, the neck joint is omitted during training for consistency.

3.4 Supervised pre-training on MPII

In this study, the first training phase of the entire 3D pose estimation pipeline involved using the MPII dataset to train the baseline 2D pose estimation model in a supervised setting. Pre-training on the MPII dataset aims to achieve optimal performance in realizing 2D poses in diverse outdoor environments, as these trained weights are later leveraged during the training of the 2D model on the HumanEva-I dataset. By increasing the generalization ability of the baseline 2D pose estimation model through pre-training on MPII, the second training phase on the HumanEva-I dataset, using parametric transfer, yields improved results, considering the indoor nature of the dataset and its MOCAP-based acquisition. High pose estimation accuracy from the 2D network is crucial for efficient self-supervised training of the 3D pose estimation model. The pose estimates generated by the 2D model are triangulated to obtain the respective 3D point estimates and used as cached ground truths for training the 3D pose estimation model.

3.5 Self-supervised training

Weak or Self-Supervised training refers to a training regimen where the data is not explicitly pre-annotated through manual labor; instead, alternative methods are employed for automated ground truth generation. All of them normally rely on some additional algorithms or properties of training data sets or their structures. While in simple and computational inexpensive tasks like age estimation, pose estimation it is pretty reasonable, but in sophisticated and time and computation consuming tasks such as 3D pose estimation, one depends on 3D annotated data and it forms bottleneck. Self-supervised methods give an efficient way to minimize the reliance on human inferences, which in the long run eliminates costs and manual efforts. In the current paper, we show a way to obtain the ground truth 3D poses for training the 3D pose estimation model in an unsupervised manner by utilizing trifocal tensors, which is explained in the subsequent section.

There are three views considered in giving unsupervised 3D ground truths required in training the 3D pose estimation module. The selection of three views is due to trifocal tensors which are used in the current study as a contrast to the fundamental matrix approach where two-view geometry is used. The trifocal tensors provide control of correspondence measurements as three-view geometry proposed by the trio-view approach, and, therefore, make the estimates of 3D pose more accurate. Also, problems of uncertainty in pose estimates of the objects are reduced inasmuch as stringencies inherent in the restraints make the estimate less vulnerable to degenerate solutions. To train the upper part of the pipeline, a trio of images L1, L2, and L3 with a resolution of 728 × 728 from the HumanEva-I dataset are fed into both the frozen 2D pose estimation and the 3D pose estimation modules. The output pose estimates A1(x, y), A2(x, y), and A3(x, y) from each trio are geometrically transformed using the trifocal tensor-based estimate to generate the 3D vector within the confines of the global coordinate frame. Similarly, the 3D pose estimation model outputs a vector containing the estimations for the joints in 3D space.

Under the pinhole camera projection model, let the pth joint in the ith image represent the 2D coordinates given by R(i, p) = [x(i, p), y(i, p)], while the respective 3D coordinates are given as Qp = [Xp, Yp, Zp]. The relation of 3D points in homogeneous coordinates within the image plane for the pinhole projection model can be given by: (9) Here, (x(i, p), y(i, p), ω(i, p)) represent the homogeneous coordinates on the image plane while (Xp, Yp, Zp) are the 3D points in the camera coordinate system. T and R are the extrinsic parameters that give the translation and rotation of the camera while K holds the intrinsic parameters of the camera inclusive of focal lengths fx, fy and the principal points cx, cy and is mathematically expressed as (10)

From the above definitions, the fundamental matrices F12, F23, F31 for each pair of views, i.e., [1, 2], [2, 3], [3, 1] in a trio-view geometry, can be calculated as under (11)

The epipolar restraints of 2D coordinates for the trio-view can be defined as (12) (13) (14)

Using the trifocal tensor Tf to express the fundamental matrices, we get (15)

Given the points (R1,p, R2,p), the third epipolar line l3,p can be computed using (16)

The 3D position of the joint R3,p can now be triangulated by computing the point of intersection with respect to the joint in each view for all three epipolar lines. The output 3D coordinates for each point are used to supervise the training of the 3D Pose estimation model. During self-supervised training, the 3D pose estimation model predicts the 3D coordinates from the input image for each of the corresponding joints given by J(x, y, z), the loss is then calculated between the cached ground truths estimated using epipolar geometry and the predicted 3D key-points J(x, y, z). By projecting the predicted 3D key-points J(x, y, z) into the camera space, the loss Lj can be computed by minimization of by substitution of . (17)

4 Results and discussion

This section discusses about different types of results we obtained in this research work utilizing different learning models on this dataset. This section also contains a detailed comparison with previously published research works.

4.1 Evaluation metrics

In this subsection, we have given a brief introduction of all evaluation metrics we have utilized in this research work.

4.1.1 MPJPE.

The Mean Per Joint Position Error or MPJPE is a common measure of the performance of 3D pose estimation methods that is used to determine the level of precision. MPJPE aims at computing the mean Euclidean distance between the re-projected and ground truth 3D locations of a joint. This is as simple as a single scalar value; the average error over all joints and frames as has been estimated by MPJPE. Here are the results of the performance of 3D pose estimation and how the MPJPE affects it. Also, it is as interesting to realize that MPJPE is scale-invariant of 3D positions, which means that it depends on the scale [38, 39]. Normalization may be used where the error is expressed as a portion of some reference size of the subject, such as area. The mathematical expression for MPJPE is as follows:The mathematical expression for MPJPE is as follows: (18) Here,

  • F = Number of Frames.
  • K = Number of Key-points.
  • Gi,j = The 3D position of key-point j in frame i.
  • Ai,j = The Predicted 3D location of joint j in frame i.

The term ‖Gi,jAi,j‖ reflects the Euclidean distance between the projected 3D coordinates of joint j in frame i and its ground truth. These variables are involved in the computation of MPJPE because they quantify the mean Euclidean distance between the estimated and ground truth joint locations across frames and joints giving information about the accuracy of 3D pose estimation methods.

The Mean Per Joint Position Error (MPJPE) is most preferred for the evaluation of human 3D pose estimation because of its quality of directly measuring the precision of estimated poses by using the raw Euclidean distance in between corresponding joints in the predicted and ground truth poses [40]. This assessment is at the joint level, which gives detailed information on algorithm performance region by region to allow researchers to locate specific problems. Furthermore, MPJPE is reliable concerning pose variability and straightforward to analyze, which has led to its wide use as a benchmarking tool in the field and cross-study comparability. Due to its coverage and general appraisal as a successful measure, MPJPE serves as a stable and successful measure of human 3D pose estimation.

4.1.2 Percentage of Correct Keypoints (PCK).

The PCK measure determines the proportion of key points where the difference between the ground truth and estimated key points is less than or equal to a predetermined threshold (θ) of the reference distance D. The key point is deemed correct and adds to the total PCK score if this requirement is satisfied; if not, it is deemed incorrect. The final PCK score, which represents the percentage of correctly translated key points, is calculated by averaging these binary values across all key points. Mathematically, PCK is given as: (19) Here,

  • N = Total number of key points
  • di = Euclidean distance between ground truth and keypoint

4.2 Experimental results and analysis

This paper introduces a novel approach to self-supervised training for 3D pose estimation by integrating trifocal tensor-based multi-view geometry with selective kernel networks for dynamic feature extraction. The pipeline works as follows: the left column of Fig 4 represents different video frames, the middle column approximates 2D key-points, and the right column displays the corresponding 3D poses obtained via triangulation. The proposed method demonstrates its efficacy with a Mean Per Joint Position Error (MPJPE) of 29.9 in a fully supervised setting and 47.6 in a weakly-supervised setting on the HumanEva-I dataset.

thumbnail
Fig 4. Result comparison of proposed methodology with previous works on MPJPE metric.

https://doi.org/10.1371/journal.pone.0310831.g004

As shown in Table 2 and Fig 4, the benchmarks set by Pavlakos et al. [41] achieved an impressive MPJPE of 18.3 on the same dataset, with other methods like those proposed by Moreno-Noguer [42] and Martinez et al. [43] also delivering competitive results with lower MPJPEs. While our proposed pipeline shows better performance in a fully supervised scenario than earlier approaches, it still exhibits a slightly higher MPJPE compared to the work by Pavlakos et al. [41, 44]. The superior performance of [41] can be attributed to their use of depth maps, which provide additional guidance for generating 3D correspondences from monocular images, thereby reducing inaccuracies. In contrast, with an MPJPE of 47.6 in weakly-supervised training, our model demonstrates a significant capability to generate accurate 3D pose approximations from monocular images within a self-supervised pipeline. Additionally, the ground truths generated using trifocal tensors in a self-supervised setting offer a more robust reference for training the 3D pose estimation model, contributing to its overall effectiveness.

thumbnail
Table 2. Result comparison of proposed methodology with previous works on HumanEva-I dataset.

https://doi.org/10.1371/journal.pone.0310831.t002

The fully supervised Tri-POSE-Net model achieves a lower MPJPE of 29.9 due to its access to extensive labeled data, enabling precise 3D pose estimation. It took around 49 seconds to gets trained and give results. In contrast, the weakly supervised model, with an MPJPE of 47.6, relies on trifocal tensor-based multi-view geometry to approximate 3D poses from limited data, sacrificing some accuracy for scalability. It took around 35 seconds to gets trained and give results. While the fully supervised approach excels in precision, the weakly supervised model offers greater adaptability in data-scarce environments, highlighting a trade-off between accuracy and data availability.

4.3 Experimental results and analysis

There are also noticeable performance discrepancies between the 3D posture estimation techniques in Table 3, with some exhibiting fewer errors and others showing a higher number of mistakes. Three methods, namely [42, 46, 47], stand out as particularly effective. They demonstrate consistently reduced Mean Per Joint Position Errors (MPJPE) across a range of movements and participants. Specifically, [41] performs exceptionally well, achieving competitive accuracy in both walking and running movements. Conversely, high error rates are exhibited by Bo et al. [48] and Simo-Serra et al. [49], indicating subpar performance. While the technique in [49] falters, notably in walking scenarios, [48] shows particularly high error rates in both walking and running. These results demonstrate the wide range of effectiveness in 3D pose estimation techniques, highlighting the need for thorough assessment and selection according to application-specific needs. Among the methods tested, our suggested approach—whether fully supervised or not—demonstrates encouraging outcomes. While the weakly supervised version achieves competitive performance across almost all action modes, the fully supervised version exhibits substantially lower errors. This implies that the above proposed approach could be useful in other scenarios and the given approach is a promising direction for further research in context to weakly-supervised 3D pose estimation under certain modifications and large-scale testing. However, the indicated research methodology opens significant potential for further improvement and fine-tuning, which can be traced from high efficiency estimates in the presence of different movement patterns. Though the knowledge gained from fully supervised and weakly supervised variants our proposed framework has the potential to perform well in different motion scenario. The obtained results are strengthen the concept of proposed approach for real conditions focusing to its feasibility for different motion scenarios. Moreover, the level of supervision that our method is capable of in its application stands out because of the variety in real-world data availability restrictions that can be addressed.

thumbnail
Table 3. Detailed result comparison of proposed approach with other works on HumanEva-I dataset.

https://doi.org/10.1371/journal.pone.0310831.t003

4.4 Evaluation at different model depths and architectures

We performed a thorough analysis by comparing the performance of residual networks (ResNets) with varying depths and some other state-of-the-art models during the pre-training phase in order to better understand the effects of network depth fluctuation and architectural definition on the overall 3d-pose estimation pipeline. Our goal was to clarify how, in the context of 2D posture estimation, the depth of the network design affects the caliber of pre-training results. We discover that there is a significant relationship between PCK measures for 2D pose estimate attained in the pre-training phase and the effectiveness of model migration later on. In particular, we found that networks with higher accuracy during pre-training performed better after migration. This occurrence is especially important because it suggests that networks that perform better at the beginning of the transfer process are also proficient at using the learnt representations. The evaluation alluded to the synergy of ResNet-50 and Selective Kernel Network (SKNET) yielding superior performance in the first phase of training pipeline, i.e., the 2D keypoint estimate training on MPII dataset. Based on the quantitative outcomes, it was chosen as the backbone architecture instead of other backbone architectures like ResNet-18, ResNet-101, VGG-16, and MobileNet. Quantitative metrics such as Mean Squared Error (MSE) and Percentage of Correct Keypoints (PCK) employed within this study, illustrate the greater accuracy and robustness of ResNet-50 in comparison with other architectures. A comparative analysis of backbone architectures given in Table 4, indicates that deeper architectures such as ResNet-101 may result in increased computational overhead without achieving appreciable performance gains, while ResNet-18 may not have the depth and capacity to capture intricate spatial relationships required for accurate keypoint localization. On the other hand, ResNet-50 provides a convincing trade-off, by striking a compromise between model complexity and computational efficiency. In particular, ResNet-50 regularly performs better than alternative backbone architectures on benchmark datasets like COCO or MPII, obtaining lower MSE values and higher PCK scores at different thresholds [51]. Herein, ResNet-50 achieved an MSE of 15 and a PCK of 0.85 at a threshold of 0.5, surpassing the performance of ResNet-18 for the first training phase. Similarly, compared to ResNet-101, ResNet-50 exhibits comparable performance with lower computational cost, making it more suitable for real-time applications and large-scale datasets. Additionally, based on thorough quantitative evaluations and considerations of model complexity, computational efficiency, and feature representation, ResNet-50 emerges as the optimal candidate for integration into SKNet blocks for 2D keypoint estimation model training.

thumbnail
Table 4. Evaluation results for different backbone architectures.

https://doi.org/10.1371/journal.pone.0310831.t004

4.5 Ablation study

The ablation study (Table 5) investigates how the three components-sKL-1, SKL-2, and Weights Sharing—affect the accuracy of 3D pose estimation, as determined by Mean Per Joint Position Error (MPJPE). With these components removed, the baseline model has a comparatively high MPJPE of 78.7. It appears that this error can be brought down to merely 65. 9 when SKL-1 is introduced alone to show the effect that is separately brought about. Last but not the least, the level of accuracy is enhanced to 61%. We have got the added advantage of getting 1 by combining SKL-1 and SKL-2 that shows that there may be an interaction. The lowest MPJPE, 47. 6, is got when all componentspointing out the synergistic effect of the SKL-1, SKL-2 and Weights Sharing in enhancing the overall performance of the model is added. Thus, this ablation study is crucial in coming up with better-performing 3D pose estimation models.

4.6 Limitation of the proposed model

The following advancements have been done by the proposed model in the field of 3D human pose estimation. Nevertheless, it has been associated with a number of drawbacks. Firstly, it is noted that both the selective kernel networks and the self-supervision with trifocal tensors enhance the performance of the model, at the same time, adding the burden of complexity regarding computational needs and augmented training time. Furthermore, the accuracy and efficiency of the model depends on the heuristic algorithm, the kind of training data and the selected hyperparameters of the algorithm and these could restrict the model to the selected training data and might not work well on other sets of data and in the real world. Moreover, the proposed model is not robust for estimate the poses when the certain level of occlusions are present, the case of ambiguous configuration of the joints, and high variations in the appearance of the poses, these are the classic issues in tracking context. Last but not the least; the implementation as well as the deployment of the proposed model might need significant knowledge in computer vision and deep learning that limits the model’s applicability to researchers or practitioners who are not well versed in the field. The latter shows direction for the future work in order to improve the proposed approach’s stability, performance and user-friendliness in realistic applications.

5 Conclusion

The proposed method TRI-POSE-Net is shown as a noteworthy development in weakly-supervised 3D pose estimation, providing a flexible approach designed to meet the complex requirements of the virtual environment. The model’s foundation, a ResNet-50 baseline fortified with two SKNet blocks, dynamically adjusts receptive field sizes and applies global context-based adaptive feature selection, thereby improving the precision of pose estimation, as evaluated on the HumanEva-I dataset. Moreover, the proposed approach tested on HumanEva-I dataset yielding a Mean-Per-Joint-Position-Error (MPJPE) of 47.6 under self-supervision and an MPJPE of 29.9 under full supervision. The use of transfer learning for the generation of 2D pose priors for triangulation, imparts adaptability and generalization capabilities, particularly vital in the ever-evolving and diverse environmental settings of the virtual world. The approach put forth in the paper becomes especially important and applicable within the virtual environment due to its cost efficiency and autonomy. The proposed trifocal tensors enhance the triangulation capabilities of epipolar geometry for realizing 3D ground truths, thus making the model’s training process more reliable and less prone to errors. In summary, although the self-supervised variant of the proposed method has a significant margin for improvement, it provides a pathway towards a robust pose estimation pipeline. With potential applications across numerous virtual environment sectors where authentic and adaptive 3D pose estimation is crucial, this comprehensive technique holds significant promise.

References

  1. 1. Liu Z., Zhu J., Bu J., and Chen C. “A survey of human pose estimation: The body parts parsing based methods,” J Vis Commun Image Represent, vol. 32, pp. 10–19, Oct. 2015.
  2. 2. Zheng C. et al. “Deep Learning-based Human Pose Estimation: A Survey,” ACM Comput Surv, vol. 56, no. 1, Aug. 2023. https://doi.org/10.1145/3603618
  3. 3. F. Zhang, X. Zhu, and C. Wang. “Single Person Pose Estimation: A Survey,” Sep. 2021, Accessed: Jan. 14, 2024. [Online]. Available: https://arxiv.org/abs/2109.10056v1.
  4. 4. Wu Q., Xu G., Li M., Chen L., Zhang X., and Xie J. “Human pose estimation method based on single depth image,” IET Computer Vision, vol. 12, no. 6, pp. 919–924, Sep. 2018.
  5. 5. S. Li and A. B. Chan. “3D human pose estimation from monocular images with deep convolutional neural network,” in Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 9004, pp. 332–347, 2015.
  6. 6. Ma S., Gong M., Qi G., Tie Y., Lee I., Li B. et al. Guest Editorial: Special Issue on Media Convergence and Intelligent Technology in the Metaverse. CAAI Transactions On Intelligence Technology. 8, 285–287 (2023)
  7. 7. M. W. Lee and I. Cohen. “Human upper body pose estimation in static images,” in Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 3022, pp. 126–138, 2004.
  8. 8. H. Y. Jung, Y. Suh, G. Moon, and K. M. Lee. “A sequential approach to 3D human pose estimation: Separation of localization and identification of body joints,” in Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 9909 LNCS, pp. 747–761, 2016.
  9. 9. B. Tekin, I. Katircioglu, M. Salzmann, V. Lepetit, and P. Fua. “Structured Prediction of 3D Human Pose with Deep Neural Networks,” British Machine Vision Conference 2016, BMVC 2016, vol. 2016-September, pp. 130.1-130.11, May 2016.
  10. 10. Feichi Lu, Zijian Dong, Jie Song, and Otmar Hilliges, AvatarPose: Avatar-guided 3D Pose Estimation of Close Human Interaction from Sparse Multi-view Videos, 2024, arXiv:2408.02110, https://arxiv.org/abs/2408.02110.
  11. 11. Bao Wenxia, Ma Zhongyu, Liang Dong, Yang Xianjun, and Niu Tao, Pose ResNet: 3D human pose estimation based on self-supervision, Sensors, vol. 23, no. 6, pp. 3057, 2023. MDPI. pmid:36991768
  12. 12. Huang Lanqing et al., Enhancing computer image recognition with improved image algorithms, Scientific Reports, vol. 14, no. 1, 13709, 14 Jun. 2024. pmid:38877063
  13. 13. W. N. Lie, G. H. Lin, L. S. Shih, Y. Hsu, T. H. Nguyen, and Q. N. Q. Nhu. “Fully convolutional network for 3d human skeleton estimation from a single view for action analysis,” Proceedings—2019 IEEE International Conference on Multimedia and Expo Workshops, ICMEW 2019, pp. 1–6, Jul. 2019.
  14. 14. Gyeongsik Moon, Ju Yong Chang, and Kyoung Mu Lee. “PoseFix: Model-agnostic General Human Pose Refinement Network. “arXiv preprint arXiv:1812.03595, 2019.
  15. 15. P. Cong et. al. “Weakly Supervised 3D Multi-Person Pose Estimation for Large-Scale Scenes Based on Monocular Camera and Single LiDAR,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 1, pp. 461–469, Jun. 2023.
  16. 16. H. Rhodin et. al. “Learning Monocular 3D Human Pose Estimation From Multi-View Images,” pp. 8437–8446, 2018.
  17. 17. D. Drover, R. MV, C.-H. Chen, A. Agrawal, A. Tyagi, and C. Phuoc Huynh. “Can 3D Pose be Learned from 2D Projections Alone?” pp. 0”0, 2018.
  18. 18. Ma J., Duan X. & Zhang D. Kernel Extreme Learning Machine-Based General Solution to Forward Kinematics of Parallel Robots. CAAI Transactions On Intelligence Technology. 8, 1002–1013 (2023)
  19. 19. Sun, X., Xiao, B., Wei, F., Liang, S., Wei, Y. (2018). Integral Human Pose Regression. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds) Computer Vision “ECCV 2018. ECCV 2018. Lecture Notes in Computer Science(), vol 11210. Springer, Cham.
  20. 20. Mehta D. et al. “VNect,” ACM Transactions on Graphics (TOG), vol. 36, no. 4, Jul. 2017.
  21. 21. B. Tekin, P. Marquez-Neila, M. Salzmann, and P. Fua. “Learning to Fuse 2D and 3D Image Cues for Monocular Body Pose Estimation,” pp. 3941–3950, 2017.
  22. 22. G. Pavlakos, X. Zhou, K. G. Derpanis, and K. Daniilidis. “Harvesting Multiple Views for Marker-Less 3D Human Pose Annotations,” pp. 6988”6997, 2017.
  23. 23. X. Zhou, Q. Huang, X. Sun, X. Xue and Y. Wei. “Towards 3D Human Pose Estimation in the Wild: A Weakly-Supervised Approach,” 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 2017, pp. 398-407.
  24. 24. B. Wandt and B. Rosenhahn. “RepNet: Weakly Supervised Training of an Adversarial Reprojection Network for 3D Human Pose Estimation,” pp. 7782–7791, 2019.
  25. 25. H. Rhodin, M. Salzmann, and P. Fua. “Unsupervised geometry-aware representation for 3D human pose estimation,” Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 11214 LNCS, pp. 765–782, 2018.
  26. 26. M. Kocabas, S. Karagoz and E. Akbas. “Self-Supervised Learning of 3D Human Pose Using Multi-View Geometry,” 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 2019, pp. 1077-1086.
  27. 27. Y. Li, K. Li, S. Jiang, Z. Zhang, C. Huang, and R. Y. Da Xu. “Geometry-Driven Self-Supervised Method for 3D Human Pose Estimation,” AAAI Conference on Artificial Intelligence, pp. 11442–11449, 2020.
  28. 28. Hartley R. I. “Lines and Points in Three Views and the Trifocal Tensor,” Int J Comput Vis, vol. 22, no. 2, pp. 125”140, 1997.
  29. 29. C. Wu, Y. Xiao, B. Zhang, M. Zhang, Z. Cao, and J. T. Zhou. “C3P: Cross-Domain Pose Prior Propagation for Weakly Supervised 3D Human Pose Estimation,” Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 13665 LNCS, pp. 554–571, 2022.
  30. 30. S. K. Roy, L. Citraro, S. Honari, and P. Fua. “On Triangulation as a Form of Self-Supervision for 3D Human Pose Estimation,” Proceedings—2022 International Conference on 3D Vision, 3DV 2022, pp. 577–586, 2022.
  31. 31. Cha J. et al. “Towards Single 2D Image-Level Self-Supervision for 3D Human Pose and Shape Estimation,” Applied Sciences 2021, Vol. 11, Page 9724, vol. 11, no. 20, p. 9724, Oct. 2021. https://doi.org/10.3390/APP11209724
  32. 32. X. Li, W. Wang, X. Hu and J. Yang. “Selective Kernel Networks,” 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 2019, pp. 510–519.
  33. 33. Zhang Y., Wang Z., Zhang X., Cui Z., Zhang B., Cui J. et al. Application of Improved Virtual Sample and Sparse Representation in Face Recognition. CAAI Transactions On Intelligence Technology. 8, 1391–1402 (2023)
  34. 34. Jie Yao, Dongdong Wang, Hao Hu, Weiwei Xing, and Liqiang Wang. “ADCNN: Towards learning adaptive dilation for convolutional neural networks.” Pattern Recognition, vol. 123, p. 108369, 2022.ISSN: 0031-3203.
  35. 35. M. Andriluka, L. Pishchulin, P. Gehler, and B. Schiele. “2D Human Pose Estimation: New Benchmark and State of the Art Analysis,” pp. 3686”3693, 2014.
  36. 36. Sigal L., Balan A. O., and Black M. J. “HumanEva: Synchronized video and motion capture dataset and baseline algorithm for evaluation of articulated human motion,” Int J Comput Vis, vol. 87, no. 1–2, pp. 4”27, Mar. 2010. https://doi.org/10.1007/S11263-009-0273-6/METRICS
  37. 37. Cao H., Wu Y., Bao Y., Feng X., Wan S. & Qian C. UTrans-Net: A Model for Short-Term Precipitation Prediction. Artificial Intelligence And Applications. 1, 106–113 (2023)
  38. 38. A. Kanazawa, M. J. Black, D. W. Jacobs, and J. Malik. “End-to-End Recovery of Human Shape and Pose,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 7122”7131, Dec. 2018.
  39. 39. Katkade S., Bagal V., Manza R. & Yannawar P. Advances in Real-Time Object Detection and Information Retrieval: A Review. Artificial Intelligence And Applications. 1, 139–144 (2023)
  40. 40. G. Pavlakos, X. Zhou, K. G. Derpanis, and K. Daniilidis. “Coarse-to-Fine Volumetric Prediction for Single-Image 3D Human Pose.”
  41. 41. G. Pavlakos, X. Zhou, and K. Daniilidis. “Ordinal Depth Supervision for 3D Human Pose Estimation,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 7307”7316, Dec. 2018.
  42. 42. F. Moreno-Noguer. “3D human pose estimation from a single image via distance matrix regression,” Proceedings—30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, vol. 2017-January, pp. 1561–1570, Nov. 2017.
  43. 43. J. Martinez, R. Hossain, J. Romero, and J. J. Little. “A Simple Yet Effective Baseline for 3d Human Pose Estimation,” Proceedings of the IEEE International Conference on Computer Vision, vol. 2017-October, pp. 2659”2668, Dec. 2017.
  44. 44. Hasanvand M., Nooshyar M., Moharamkhani E. & Selyari A. Machine Learning Methodology for Identifying Vehicles Using Image Processing. Artificial Intelligence And Applications. 1, 170–178 (2023)
  45. 45. H. Yasin, U. Iqbal, B. Kruger, A. Weber, and J. Gall. “A dual-source approach for 3D pose estimation from a single image,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 2016-December, pp. 4948”4956, Dec. 2016.
  46. 46. I. Radwan, A. Dhall, and R. Goecke. “Monocular image 3D human pose estimation under self-occlusion,” in Proceedings of the IEEE International Conference on Computer Vision, pp. 1888”1895, 2013.
  47. 47. C. Wang, Y. Wang, Z. Lin, A. L. Yuille, and W. Gao. “Robust estimation of 3D human poses from a single image,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 2369”2376, Sep. 2014.
  48. 48. Bo L. and Sminchisescu C. “Twin gaussian processes for structured prediction,” Int J Comput Vis, vol. 87, no. 1-2, pp. 28”52, Mar. 2010.
  49. 49. E. Simo-Serra, A. Quattoni, C. Torras, and F. Moreno-Noguer. “A joint model for 2D and 3D pose estimation from a single image,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 3634”3641, 2013.
  50. 50. I. Kostrikov and J. Gall. “Depth Sweep Regression Forests for Estimating 3D Human Pose from Images,” in British Machine Vision Conference, 2014.
  51. 51. Yucheng Chen, Yingli Tian, and He Mingyi. “Monocular human pose estimation: A survey of deep learning-based methods.” Computer Vision and Image Understanding, vol. 192, p. 102897, 2020. ISSN: 1077-3142. https://doi.org/10.1016/j.cviu.2019.102897