Acessibilidade / Reportar erro

HUMAN DETECTION AND MOTION RECOVERY BASED ON MONOCULAR VISION

DETECÇÃO HUMANA E RECUPERAÇÃO DE MOVIMENTO COM BASE NA VISÃO MONOCULAR

DETECCIÓN HUMANA Y RECUPERACIÓN DE MOVIMIENTO BASADA EN VISIÓN MONOCULAR

ABSTRACT

Objective:

Provides interactive games and human animation real motion data and technical options. Therefore, how to complete the position, attitude detection, and motion recovery under monocular vision has become an important research direction.

Methods:

This paper improves the part-based human detection algorithm and uses the AdaBoost multi-instance learning algorithm to train the part detector.

Results:

The results show that obtaining blood pressure waveform based on monocular vision pulse wave is feasible and has generalization.

Conclusions:

The results show the feasibility and accuracy of the gait motion detection, motion recovery and analysis system for human lower limbs based on monocular vision. Level of evidence II; Therapeutic studies - investigation of treatment results.

Keywords
Detection; Motion recovery; Vision, monocular

RESUMO

Objetivo:

Fornece jogos interativos e dados de movimento real de animação humana e opções técnicas. Portanto, como completar a posição, detecção de atitude e recuperação de movimento sob visão monocular tornou-se uma importante direção de pesquisa.

Métodos:

este artigo aprimora o algoritmo de detecção humana baseado em partes e usa o algoritmo de aprendizado de múltiplas instâncias AdaBoost para treinar o detector de partes.

Resultados:

Os resultados mostram que o método de obtenção da forma de onda da pressão arterial com base na onda de pulso de visão monocular é viável e se pode generalizar.

Conclusões:

Os resultados mostram a viabilidade e precisão do sistema de detecção, recuperação e análise do movimento da marcha para membros inferiores humanos com base na visão monocular. Nível de evidência II; Estudos terapêuticos- investigação dos resultados do tratamento.

Descritores
Detecção; Recuperação de movimento; Visão monocular

RESUMEN

Objetivo:

Proporciona juegos interactivos y animación humana, datos de movimiento real y opciones técnicas. Por lo tanto, cómo completar la posición, la detección de actitud y la recuperación de movimiento bajo visión monocular se ha convertido en una importante dirección de investigación.

Métodos:

este documento mejora el algoritmo de detección humana basado en piezas y utiliza el algoritmo de aprendizaje de instancias múltiples AdaBoost para entrenar el detector de piezas.

Resultados:

Los resultados muestran que el método de obtención de la forma de onda de la presión arterial basado en la onda de pulso de visión monocular es factible y se puede generalizar.

Conclusiones:

Los resultados muestran la viabilidad y precisión del sistema de detección, recuperación y análisis del movimiento de la marcha para miembros inferiores humanos basado en visión monocular. Nivel de evidencia II; Estudios terapéuticos- investigación de los resultados del tratamiento.

Descriptores
Detección; Recuperación de movimiento; Visión monocular

INTRODUCTION

The extraction and processing of related person information in the video is important to use value.11. Chen HT, Wu YC, Hsu CC. Daytime preceding vehicle brake light detection using monocular vision. IEEE Sensors Journal. 2015;16(1);120-31. By analyzing human gestures, action, and expression analysis, the computer can understand people's intent and achieve true intelligence analysis and processing.22. Lee TJ, Yi DH, Cho DI. A Monocular Vision Sensor-Based Obstacle Detection Algorithm for Autonomous Robots. Sensors (Basel). 2016;16(3):311. The single eye is relatively small compared to the double eye, and the amount of calculation is relatively small, which is convenient for operation.33. Lin S, Garratt MA, Lambert AJ. Monocular vision-based real-time target recognition and tracking for autonomously landing an UAV in a cluttered shipboard environment. Autonomous Robots. 2017;41(4):881-901. Golden motion extraction Based on single eye video, including recovery or reproducing real human motion data on video summary body models, this will provide a broader realistic motion data and interactive games and human animation technologies.

METHODS

Human detection based on monocular vision

Human detection technology is to search all human targets from images or videos to be detected. To this end, this chapter uses cascaded Adaboosting learning methods to train human detectors.44. Yuxi F, Guotian H, Qizhou WA. New Motion Obstacle Detection Based Monocular-Vision Algorithm. In 2016 International Conference on Computational Intelligence and Applications (ICCIA). 2016;31-5. Fast feature selection algorithm and cascade classifier training.

In view of the shortcomings of AdaBoosting in training speed, we use this fast feature selection method to improve the training of weak classifiers. The process is shown in Figure 1a, where S is the number of samples and P is the maximum number of rounds of training. In the feature selection stage, all candidate feature information is obtained by querying the statistical table. Each additional feature can quickly find the classification error of the overall strong classifier.55. Jia B, Liu R, Zhu M. Real-time obstacle detection with motion features using monocular vision. The Visual Computer. 2016;31(3);281-93. The total time complexity of the algorithm is O (SMPlogS), as shown in Figure 1a. The time complexity of the general Adaboost algorithm is O (SMP + SMLogS). This shows that the time complexity required to select a feature and train a weak classifier is 1S+1log S of the original Adaboost algorithm. In monocular video surveillance scenes or still images, most areas are background images, and human targets occupy only a few areas. In the cascade structure of Cascade classifier, as long as the detection area is judged to be non-human by a certain classifier, the detection process of the target is ended, and the output is non-human. After passing through all stages of detection, it is the human body, and its structure is shown in Figure 1b.

Figure 1
The process of fast feature selection to improve the AdaBoosting algorithm (A) and hierarchical cascade classifier (B).

Each strong classifier is represented by Equation 1, and is a linear combination of selected features.

(1) M ( x ) = sgn [ p = 1 P α i m i ( x ) β ]

Among them, αi represents the weak classifier weight, β is the strong classifier threshold, and its initial definition is β=12p=1Pαi. The strong classifier Mr contains different numbers of weak classifiers mi(x), which are composed of a feature vj, threshold β, and directions Mj indicating inequality signs. The final output of the algorithm is a cascaded strong classifier M={M1, M2,Mr} as a human detector.

Fast target human detection

When the detection window slides on the detection image, only 2r number of pixel regions change. Correspondingly, the bins of 2sr histograms change, where s is a factor between 0 and 1, and r is the detection window size. The pseudo code of the histogram search method based on the “block” update is:

Form=1: scaledo

(A) Initializing, and qij, i = 1, L, n, and j = 1, L, m

(B) For i = 1: n do (1) For j = 2: m do, qi,j − 1qi,j

(2) endform, qi,mqnew

(C) Endforn

Endforq.

Suppose a given test window M is used to calculate its CT value characteristic histogram. Let fM be the corresponding histogram |M| and be the number of all pixels in M.66. Su S, Zhou Y, Wang Z, Chen H. Monocular Vision- and IMU-Based System for Prosthesis Pose Estimation During Total Hip Replacement Surgery. IEEE Trans Biomed Circuits Syst. 2017;11(3):661-670. At each pixel position (x, y), the value of the k-th bin of histogram hM is represented as

(2) f k = ( x i , y i ) M 1 { b ( x i ,   y i ) = k }

Where 1{g} is an index function that maps a pixel (xi, yi) to the corresponding bin. When the detection window slides, only the leftmost column CL and rightmost column CR need to be re-counted, which is denoted as

(3) f k = f k ( x i , y i ) C L 1 { b ( x i , y i ) = k } + ( x i , y i ) C R 1 { b ( x i , y i ) = k }

The number of changing pixels is 2r(2r = |M|).

Among them, MissRate and FPPW are defined as

(4) M i s s R a t e =   F a l s e N e g a t i v e s   T r u e P o s i t i v e s + F a l s e N e g a t i v e s
(5) F P P W = F a l s e P o s i t i v e s T r u e N e g i t i v e s + F a l s e P o s i t i v e s

The background image is tested to obtain the DET curve shown in Figure 2. Figure 2 shows the comparison of the detection performance DET curve on the INRIA database using a single HOG feature and a HOG-CT hybrid feature. It can be seen that the detection performance of the Adaboost algorithm constructed in this study has been significantly improved.

Figure 2
Comparison of detection performance using HOG and HOG-CT features in human detection.

RESULTS

Human pulse information and blood pressure waveform measurement methods under monocular vision

The characteristic parameters of the pulse that have a greater correlation with the blood pressure value are mainly including: Main wave amplitude h1, weighted wave amplitude h4, weighted wave relative height h4 / h1, waveform coefficient K value,h1 (1 + t3 / t4) reflecting cardiac output per stroke, ascending branch average slope h1 / t1, systolic relative area s1 / (s1 + s2). Figure 3 shows the pulse waveform extracted using the algorithm of this study and one of its pulse cycles.

Figure 3
Pulse waveform extracted using the monocular vision Adaboosting algorithm and one of its pulse cycles.

Based on the pulse wave at the center of the pulse (Figure 3). The pulse parameters and calculated SBP and DBP are shown in Table 1.

Table 1
Pulse characteristics and calculated blood pressure values.

The minimum value of the blood pressure waveform p (t) obtained initially is expressed as p (t) min and the maximum value is p (t) max. Using mathematical formula 6 of data standardization, p (t) can be mapped between [SBP DBP], and a reasonable blood pressure waveform P (t) can be obtained. (Figure 4)

(6) P ( t ) = D B P + P ( t ) P ( t ) min P ( t ) max P ( t ) min
Figure 4
Blood pressure waveform after the monocular vision algorithm is revised.

Human Motion Recovery Based on Monocular Vision

This study proposes a generative 3D human body motion restoration method based on barrage vision. Firstly, analyze the contour of the human body to obtain the position information of the trunk and end nodes, and then optimize the 3D pose.77. Huang XY, Gao F, Xu GY, Ding NG, Xing LL. Depth information extraction of on-board monocular vision based on a single vertical target image. Journal of Beijing University of Aeronautics and Astronautics. 2016;41(4);649-55. Experimental analysis was performed using a gait rehabilitation training machine. Let the human body posture be X, usually a high-dimensional space such as a joint point, and the video image observation is ZX. If the two-dimensional human body contour is extracted, the recovery process is. It is very easy for human vision to judge the pose of a human in a monocular video image. While ZX is a serious morbid problem compared to computers. But in the video sequence, the introduction of time-domain information transforms the reasoning problem into a dynamic process, which can be described by {Zii=1, L, t},{Xii=1, L, t1}X.

Modeling of camera imaging model

It is considered that the coordinate value (x, y, z) of a point in three-dimensional space and the coordinate (u, v) of that point on the two-dimensional projection plane satisfy equation:

(7) [ u v ] = 1 s [ 1 0 0 0 1 0 ] [ x y z ]

Among them, the parameter s is a scale factor, which is obtained from s = z/f .z is the z coordinate value of the point in three-dimensional space, and f is the focal length of the camera.88. Chen Z, Zhang Z, Dai F, Bu Y, Wang H. Monocular vision-based underwater object detection. Sensors. 2017;17(8);1784-6. From equation (7), we can know that when z changes, the value of s will change linearly. The change amount ds of s relative to the z value dz is expressed as T(dz)=ds.99. Xu LY, Cao ZQ, Zhao P, Zhou C. A new monocular vision measurement method to estimate 3D positions of objects on floor. International Journal of Automation and Computing. 2017;14(2);159-68.1111. Yu XG, Li YQ, Zhu WB. Wearable strain sensor based on carbonized nano-sponge/silicone composite for human motion detection. Nanoscale. 2017;9(20):6680-92. Figure 5 shows the imaging diagram of the three end-to-end bone segments under the perspective projection model.

Figure 5
Schematic diagram of the lower bone section of the perspective projection model during motion recovery.

In Figure 5, the skeletal segments ab and cd are parallel to the projection plane, and their imaging on the projection plane are a′b′ and c′d′, respectively. The skeletal segment bc is not parallel to the projection plane, and its imaging on the projection plane is b′c′. Assuming that the lengths of ab, bc, and cd are Lab, Lbc, Lcd respectively, their values can be obtained from the human skeleton model. The projection lengths of a′b′, b′c′ and c′d′ are Lab,Lbc,Lcd.

Since bone ab is parallel to the projection plane, its projection on the z-axis intersects at point M. The variable factor q corresponding to the point M is calculated according to qab=OM/g=Lab/Lab. Similarly, we can get the variable factor q corresponding to the projection point N of the bone cd on the z axis, qab=Lab/Lab. Since the bones ab and cd are parallel to the projection plane, the distance between the point M and the point N satisfies dz=czbz. According to the knowledge of space geometry, Lbc satisfies :

(8) L b c = ( c x b x ) 2 + ( c y b y ) 2 ( c z b z ) 2 L b c = ( c x b x ) 2 + ( c y b y ) 2 + d z 2 L b c = ( s c d c x s a b c x ) 2 + ( s c d c x s a b b x ) 2 + d z 2

For Scd, Sab, cx, bx, by and Lbc are known, dz can be calculated by Equation (8). Relative to the absolute change of z (| dz |), the corresponding absolute change of s is |ds|=abs(scdsab).

Modeling of human bone model during exercise recovery

We see the human body as a tree-like stick model, as shown in Figure 6. The skeletal model consists of 16 joint points and 5 body segments. Among them, M1 is the root node of the tree structure, which corresponds to the pelvic joint of the human body;The length of the line segment (human bone segment) in the model is obtained from anthropometrics. That is, 1) local coordinate system. It is fixed to each body segment, and the origin of the coordinate system is the attachment joint of the body segment; 2) Global coordinate system. The origin of the coordinate system is the M1 joint.

Figure 6
Human skeleton model during exercise recovery.

The composition of the objective function

This can be formalized as

(9) h ^ = arg min { E ( h ; M , R ) }

Where h is the three-dimensional pose vector. M is the camera model (transformation matrix from world coordinates to image coordinates), R is the analyzed contour. E () is an objective function that calculates the degree of matching between S and P (after C transformation). The objective function proposed in this chapter contains five parts, which correspond to the five skeleton segments on the human skeleton model:

(10) E ( h ; M , R ) = E T o r s о   ( h , M , R ) + Π l p E L U p p e r ( h ; M , R ) + Π r h E R U p p e r ( h ; M , R ) + Π l f E L l o w e r   ( h ; M , R ) + Π r f E R L o w e r   ( h ; M , R )

The definitions of the four II values for both hands and feet are as follows: That is, when a limb (hand, foot) is blocked, the corresponding II = 0, otherwise II = 1. It can be seen that if a hand or foot is not positioned during the contour analysis step, its corresponding skeleton segment does not contribute to the objective function. Each term to the right of the equal sign in Equation 8 is further composed of three sub-terms: core region term Ecore-area, end coverage term Ecov erage, and timing smoothing term Esmoothness. Take ELUpper(h;M, R) as an example:

(11) E L U p p e r ( h ; M , R ) = a 1 E L U p p e r   c o r e - a r e a   ( h ; M , R ) + a 2 E L U p p e r   cov e r a g e   ( h ; M , R ) + a 3 E L U p p e r   s m o o t h n e s s   ( h ; M , R )

Note that the right 5 terms of Equation 9 are actually independent of each other. For example, when calculating the ELUpper(h; M, R) term, only the left upper limb LUpper skeleton segment is involved. Therefore, the optimization of the objective function can be divided into five independent sub-optimization processes. Figure 7a is a specific optimization process. The objective function as a whole is optimized through continuous loop optimization. When the objective function value stops decreasing or reaches the number of loop iterations, the optimization ends. The sub-optimization process uses a simulated annealing algorithm. Simulated annealing can be seen as an improvement on the gradient descent method (Figure 7b). As the temperature T decreases, the probability exp(−ΔE/T) decreases. It can be proven that when the initial temperature is high enough and the annealing speed is low enough, the output of the simulated annealing algorithm approaches the global optimal value asymptotically with probability 1. By setting the appropriate initial temperature and annealing coefficient, the extent to which the simulated annealing can overcome the local minimum can be easily controlled.

Figure 7
Iterative optimization process (a) and simulated annealing algorithm flow (b) during motion recovery.

DISCUSSION

Human gait analysis and sports rehabilitation based on monocular vision

The scope of Figures 8A and 8B show the relationship between the hips and knee joints during the gait and treadmill in the gait cycle. The angular relationship curve of the gait machine is smooth and closed. This suggests that its lower extremity is in line with coordinated requirements. The angular relationship curve on the treadmill has different characteristics from the gait machine. This may be because people are not limited by the running machine, and the steps and frequencies will change to various factors, which cannot be periodically as motion on the gait machine.

Figure 8
Joint-angle relationship on treadmill (a) and gait machine (b) in monocular vision.

From the experimental results of Figures 8A and 8B, it can be seen that the feasibility and accuracy of the lower extremity step motion detection and analysis system based on single-eye visual development. This study is a useful attempt to analyze and restore human movements based on single-eye visual tags, and have obtained some satisfactory preliminary research results.

CONCLUSION

This article is based on the feasibility and accuracy of the analysis and analysis system of human low limbs based on monocular vision. The next step is to use a dynamic programming algorithm to select a related feature from a relational feature library, and these features correspond to the weighting distance between the weight, the distance between the three-dimensional posture.

REFERENCES

  • 1
    Chen HT, Wu YC, Hsu CC. Daytime preceding vehicle brake light detection using monocular vision. IEEE Sensors Journal. 2015;16(1);120-31.
  • 2
    Lee TJ, Yi DH, Cho DI. A Monocular Vision Sensor-Based Obstacle Detection Algorithm for Autonomous Robots. Sensors (Basel). 2016;16(3):311.
  • 3
    Lin S, Garratt MA, Lambert AJ. Monocular vision-based real-time target recognition and tracking for autonomously landing an UAV in a cluttered shipboard environment. Autonomous Robots. 2017;41(4):881-901.
  • 4
    Yuxi F, Guotian H, Qizhou WA. New Motion Obstacle Detection Based Monocular-Vision Algorithm. In 2016 International Conference on Computational Intelligence and Applications (ICCIA). 2016;31-5.
  • 5
    Jia B, Liu R, Zhu M. Real-time obstacle detection with motion features using monocular vision. The Visual Computer. 2016;31(3);281-93.
  • 6
    Su S, Zhou Y, Wang Z, Chen H. Monocular Vision- and IMU-Based System for Prosthesis Pose Estimation During Total Hip Replacement Surgery. IEEE Trans Biomed Circuits Syst. 2017;11(3):661-670.
  • 7
    Huang XY, Gao F, Xu GY, Ding NG, Xing LL. Depth information extraction of on-board monocular vision based on a single vertical target image. Journal of Beijing University of Aeronautics and Astronautics. 2016;41(4);649-55.
  • 8
    Chen Z, Zhang Z, Dai F, Bu Y, Wang H. Monocular vision-based underwater object detection. Sensors. 2017;17(8);1784-6.
  • 9
    Xu LY, Cao ZQ, Zhao P, Zhou C. A new monocular vision measurement method to estimate 3D positions of objects on floor. International Journal of Automation and Computing. 2017;14(2);159-68.
  • 10
    Zhang G, Liu J, Li H. Joint Human Detection and Head Pose Estimation via Multi-Stream Networks for RGB-D Videos[J]. IEEE Signal Processing Letters. 2017;3(13):19-32.
  • 11
    Yu XG, Li YQ, Zhu WB. Wearable strain sensor based on carbonized nano-sponge/silicone composite for human motion detection. Nanoscale. 2017;9(20):6680-92.

Publication Dates

  • Publication in this collection
    20 Aug 2021
  • Date of issue
    Oct-Dec 2021

History

  • Received
    28 Apr 2021
  • Accepted
    10 May 2021
Sociedade Brasileira de Medicina do Exercício e do Esporte Av. Brigadeiro Luís Antônio, 278, 6º and., 01318-901 São Paulo SP, Tel.: +55 11 3106-7544, Fax: +55 11 3106-8611 - São Paulo - SP - Brazil
E-mail: atharbme@uol.com.br