BEAMFORMING WITH KINECT V2

Similar documents
THE USE OF VOLUME VELOCITY SOURCE IN TRANSFER MEASUREMENTS

Composite aeroacoustic beamforming of an axial fan

Sound Source Localization using HRTF database

ENHANCED PRECISION IN SOURCE LOCALIZATION BY USING 3D-INTENSITY ARRAY MODULE

MULTIPLE SENSORS LENSLETS FOR SECURE DOCUMENT SCANNERS

SOUND FIELD MEASUREMENTS INSIDE A REVERBERANT ROOM BY MEANS OF A NEW 3D METHOD AND COMPARISON WITH FEM MODEL

3. Sound source location by difference of phase, on a hydrophone array with small dimensions. Abstract

Sensors and Sensing Cameras and Camera Calibration

Journal of Mechatronics, Electrical Power, and Vehicular Technology

Individually configurable system. Microphone Arrays.

Scan-based near-field acoustical holography on rocket noise

29th TONMEISTERTAGUNG VDT INTERNATIONAL CONVENTION, November 2016

GESTURE RECOGNITION SOLUTION FOR PRESENTATION CONTROL

MICROPHONE ARRAY MEASUREMENTS ON AEROACOUSTIC SOURCES

A Geometric Correction Method of Plane Image Based on OpenCV

Catadioptric Stereo For Robot Localization

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

3DUNDERWORLD-SLS v.3.0

Artificial Beacons with RGB-D Environment Mapping for Indoor Mobile Robot Localization

Guided Filtering Using Reflected IR Image for Improving Quality of Depth Image

Multi-channel Active Control of Axial Cooling Fan Noise

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

APPLICATION AND ACCURACY POTENTIAL OF A STRICT GEOMETRIC MODEL FOR ROTATING LINE CAMERAS

IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics

1 st IFAC Conference on Mechatronic Systems - Mechatronics 2000, September 18-20, 2000, Darmstadt, Germany

Multiple Sound Sources Localization Using Energetic Analysis Method

NTT DOCOMO Technical Journal. Method for Measuring Base Station Antenna Radiation Characteristics in Anechoic Chamber. 1.

PLazeR. a planar laser rangefinder. Robert Ying (ry2242) Derek Xingzhou He (xh2187) Peiqian Li (pl2521) Minh Trang Nguyen (mnn2108)

LOCALIZATION OF WIND TURBINE NOISE SOURCES USING A COMPACT MICROPHONE ARRAY WITH ADVANCED BEAMFORMING ALGORITHMS

Study Of Sound Source Localization Using Music Method In Real Acoustic Environment

AR 2 kanoid: Augmented Reality ARkanoid

A Comparative Study of Structured Light and Laser Range Finding Devices

Acoustic Resonance Analysis Using FEM and Laser Scanning For Defect Characterization in In-Process NDT

Introduction to DSP ECE-S352 Fall Quarter 2000 Matlab Project 1

EFFECTS OF PHYSICAL CONFIGURATIONS ON ANC HEADPHONE PERFORMANCE

About Doppler-Fizeau effect on radiated noise from a rotating source in cavitation tunnel

Image Processing & Projective geometry

Laser Doppler sensing in acoustic detection of buried landmines

Acquisition Basics. How can we measure material properties? Goal of this Section. Special Purpose Tools. General Purpose Tools

BEAMFORMING WITHIN THE MODAL SOUND FIELD OF A VEHICLE INTERIOR

Lecture 19: Depth Cameras. Kayvon Fatahalian CMU : Graphics and Imaging Architectures (Fall 2011)

CSE 165: 3D User Interaction. Lecture #7: Input Devices Part 2

Spectral Analysis of the LUND/DMI Earthshine Telescope and Filters

CALIBRATION OF IMAGING SATELLITE SENSORS

Digital Photographic Imaging Using MOEMS

Implementation and analysis of vibration measurements obtained from monitoring the Magdeburg water bridge

arxiv: v1 [cs.sd] 4 Dec 2018

Face Detection using 3-D Time-of-Flight and Colour Cameras

Impact of Thermal and Environmental Conditions on the Kinect Sensor

Figure 1. SIG ACAM 100 and OptiNav BeamformX at InterNoise 2015.

Accuracy Estimation of Microwave Holography from Planar Near-Field Measurements

CALIBRATION OF OPTICAL SATELLITE SENSORS

Camera Calibration Certificate No: DMC III 27542

LENSLESS IMAGING BY COMPRESSIVE SENSING

Sound source localisation in a robot

Direction-of-Arrival Estimation Using a Microphone Array with the Multichannel Cross-Correlation Method

HIGH ACCURACY CROSS-POLARIZATION MEASUREMENTS USING A SINGLE REFLECTOR COMPACT RANGE

Simulation and design of a microphone array for beamforming on a moving acoustic source

Application Note. Airbag Noise Measurements

FREQUENCY RESPONSE AND LATENCY OF MEMS MICROPHONES: THEORY AND PRACTICE

KINECT CONTROLLED HUMANOID AND HELICOPTER

Localizing Noise Sources on a Rail Vehicle during Pass-by

SMARTSCAN Smart Pushbroom Imaging System for Shaky Space Platforms

Transfer Function (TRF)

Reprojection of 3D points of Superquadrics Curvature caught by Kinect IR-depth sensor to CCD of RGB camera

Disturbance Rejection Using Self-Tuning ARMARKOV Adaptive Control with Simultaneous Identification

Applying the Filtered Back-Projection Method to Extract Signal at Specific Position

Broadband Temporal Coherence Results From the June 2003 Panama City Coherence Experiments

Astigmatism Particle Tracking Velocimetry for Macroscopic Flows

On Determination of Focal Laws for Linear Phased Array Probes as to the Active and Passive Element Size

Speech Intelligibility Enhancement using Microphone Array via Intra-Vehicular Beamforming

Proceedings of Meetings on Acoustics

CALIBRATION OF AN AMATEUR CAMERA FOR VARIOUS OBJECT DISTANCES

More Info at Open Access Database by S. Dutta and T. Schmidt

FOCAL LENGTH CHANGE COMPENSATION FOR MONOCULAR SLAM

Overview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image

MEASURING DIRECTIVITIES OF NATURAL SOUND SOURCES WITH A SPHERICAL MICROPHONE ARRAY

Design and Calibration of a Small Aeroacoustic Beamformer

UNIT-3. Electronic Measurements & Instrumentation

Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya 2, B. Yamuna 2, H. Divya 2, B. Shiva Kumar 2, B.

RADIOMETRIC CALIBRATION OF INTENSITY IMAGES OF SWISSRANGER SR-3000 RANGE CAMERA

Active Stereo Vision. COMP 4102A Winter 2014 Gerhard Roth Version 1

A 3D Profile Parallel Detecting System Based on Differential Confocal Microscopy. Y.H. Wang, X.F. Yu and Y.T. Fei

Classification for Motion Game Based on EEG Sensing

Estimation of spectral response of a consumer grade digital still camera and its application for temperature measurement

A shooting direction control camera based on computational imaging without mechanical motion

ECMA-108. Measurement of Highfrequency. emitted by Information Technology and Telecommunications Equipment. 4 th Edition / December 2008

Piezoceramic Ultrasound Transducer Enabling Broadband Transmission for 3D Scene Analysis in Air

INVERSE METHOD FOR THE ACOUSTIC SOURCE ANALYSIS OF AN AEROENGINE

Broadband Microphone Arrays for Speech Acquisition

Localization of underwater moving sound source based on time delay estimation using hydrophone array

International Conference on Information Sciences, Machinery, Materials and Energy (ICISMME 2015)

An Overview Algorithm to Minimise Side Lobes for 2D Circular Phased Array

WIND SPEED ESTIMATION AND WIND-INDUCED NOISE REDUCTION USING A 2-CHANNEL SMALL MICROPHONE ARRAY

Advanced Test Equipment Rentals ATEC (2832)

ABSTRACT 2. DESCRIPTION OF SENSORS

EXPERIMENT ON PARAMETER SELECTION OF IMAGE DISTORTION MODEL

TBM - Tone Burst Measurement (CEA 2010)

Gesture Recognition with Real World Environment using Kinect: A Review

Congress Best Paper Award

Transcription:

BEAMFORMING WITH KINECT V2 Stefan Gombots, Felix Egner, Manfred Kaltenbacher Institute of Mechanics and Mechatronics, Vienna University of Technology Getreidemarkt 9, 1060 Wien, AUT e mail: stefan.gombots@tuwien.ac.at e mail: felix.egner@tuwien.ac.at e mail: manfred.kaltenbacher@tuwien.ac.at Abstract Microphone array measurements in combination with beamforming techniques are often used for acoustic source localization. The sound pressure obtained at different microphone positions are mapped by these techniques to a planar or a surface map. The mapping result named as beamform map, indicates the location and strength of acoustic sources. For this mapping process the distance between the sound source or device under test (DUT), and the microphone positions must be known. To determine these distances the Microsoft Kinect for Windows v2 (Kinect V2) is used. The Kinect V2 sensor allows acquiring RGB, infrared (IR) and depth images. The depth images are evaluated and the required distances are computed. The distance is measured contactless, and also the surface of the DUT can be reconstructed through the depth images. Furthermore, the RGB image is used as an underlying layer of the beamform map. The applicability of the source mapping process using the Kinect V2 is demonstrated and the characteristics of the sensor are discussed. Keywords Beamforming, microphone array, source mapping, Kinect V2 I. INTRODUCTION Beamforming techniques, e. g. Standard Beamforming [1], Functional Beamforming [2], CLEAN SC [3], Orthogonal Beamforming [4], are an often used method to localize acoustic sources. These techniques are based on evaluating simultaneously collected sound pressure data from microphone array measurements. In the case of stationary acoustic sources it is common to work in the frequency domain. Here, the beamform map for Standard Beamforming is computed by a(g) = g H Cg (1) with g the steering vector, H the hermitian operation, and C the cross spectral matrix of the microphone signals. The beamform map provides information about location and strength of sound sources. Thereby, a certain model for the sources and sound field is assumed. By using monopole sources in a free field, the steering vectors g are given by the free-space Green s function g(r) = 1 4πr e jkr (2) with the wave number k and r = x s x m,i the distance between assumed source point x s and microphone position x m,i. The typical geometric setup of acquiring a two-dimensional beamform map is depicted in Fig. 1. The sound sources are assumed to be in a planar scanning area. The microphones are located parallel to this area. Hence, in the two-dimensional planar source mapping the z-coordinate is constant. Now, microphone plane y x m,i z x r Z scanning area... assumed sourceposition Fig. 1: Geometric setup Two-dimensional acoustic source mapping. using the depth image information of the Kinect V2 sensor, the source distribution can be mapped on the real surface of the DUT. In addition, the information of the RGB image can be used as an underlying layer of the beamform map. II. MEASUREMENT SYSTEM In the following the measurement system for obtaining the sound pressure at the microphone positions x m,i will be presented and the characteristics of the Kinect V2 sensor discussed. In Fig. 2, one can see a schematic representation of the overall measurement system used, containing the microphone array, the data acquisition unit and the Kinect V2. x s X Y

Microphone array Circle, Underbrink Electret microphones with preamplifier Kinect V2 RGB image Depth sensor FlexAmp A/D Converter MADIface Computer Fig. 2: Measurement system Schematic representation. With the Kinect V2 sensor, Microsoft delivers an interface device for their gaming console Xbox One, providing a motion controller and a speech recognition system. With an adapter [7] the sensor can also be used by computer to acquire RGB and depth images. To enable theuseofthekinectv2onehastodownloadthekinect for Windows SDK 2.0 (free available). It provides the drivers, application programming interfaces (APIs) and code samples. Since Version 2016a the Kinect V2 is also supported by MATLAB. To use the functionality of the sensor in earlier versions one can use the Kin2 Toolbox [13]. Both tools uses the underlying functions of the SDK. The Kinect V2 is composed of a RGB and an IR camera, whereby the IR camera is used for the acquisition of the depth images. Figure 3 illustrates the component parts of the sensor. A. Microphone array The result of the acoustic source mapping process depends on different parameters, e. g. the microphone arrangement, the frequency of the acoustic source, the signal processing algorithm, etc. The used planar array has 63 electret microphones and consists of a circle with 32 microphones and an Underbrink design [5] of 31 microphones. The aperture of the array is 1m, resulting in a lower frequency limit of about 1400Hz. According to the spatial aliasing theorem, which is introduced by the repeated sampling space of the microphones, the upper limit of the circle array is about 8000Hz. At higher frequencies ghost images will arise in the beamform map. The theorem holds for regular arrays, like the circle. Irregular array designs, where the microphone spacings are different, the effect of ghost images can be decreased. In this context the Underbrink design performs best [6]. Within the scope of this work, all 63 microphones are used to calculate the beamform map. The electret microphones are calibrated in an anechoic chamber by comparing it with a calibrated Bruel&Kjaer microphone. The calibration process was also verified using a pistonphone. The sensitivities of the 63 microphones are considered in the calculation process of the beamform map. In beamforming methods the microphone array is usually placed in the far field of the source. Moreover, the source should be kept near the center axis of the array for best results. Hence, the directional characteristic of the microphones seems to be negligible and therefore are not considered. The control of the recording, the analysis, the signal processing and the computation of the beamform map is done by a program written in MATLAB. B. Kinect V2 sensor Fig. 3: Kinect V2 sensor Component parts. The sensor uses the time-of-flight method (ToF) for the depth measurement; a detailed description is given in [8]. Some characteristics of the sensor are summarized in Tab.1. RGB camera resolution 1920 1080 px FOV (h v) 84.1 53.8 IR/depth resolution 512 424, px camera FOV (h v) 70.6 60.0 operating range 0.5 4.5(8)m Frame rate 30 Hz Connection type USB 3.0 Tab. 1: Characteristics of the Kinect V2 sensor [9]. The field of view (FOV) of both cameras have been checked. For this reason the sensor was placed parallel to a white plane wall in a distance of 1m and 2m. The result is given in Tab.2. The measurement shows a good agreement with the specification given by the manufacturer. Manufacturer Own evaluation RGBcamera h 84.1 85.0 v 53.8 54.2 IR/depth h 70.6 70.3 camera v 60.0 58.2 Tab. 2: Field of view Comparison. Because the RGB and depth image have different field of view, the correspondence between the images have to be established. In Fig. 4 the difference in the FOV of

Fig. 4: Images from Kinect V2 (left) RGB image (right) depth image. the cameras is shown. One can see that the images partially overlap, but the color camera has a wider horizontal FOV, while the IR camera has a larger vertical FOV. The correspondence between the images can be established by the SDK functions. Making use of the SDK functions, the mapping between the locations on the depth image and their corresponding locations on the color image can be done. The Kinect V2 sensor is placed near the center of the array to exploit the FOV of the cameras best (Fig. 5). The base of the Kinect V2 was removed to place it on the given array geometry. (a) (c) (d) Fig. 6: Improvement of the depth map through averaging Result of (a) no 10 (c) 20 (d) 50 averages. Offset in distance to an averaged map of 100 frames. Depth measurements were taken and averaged over 50 frames. To determine the mean value a small section (50 px 50 px) at the image center was used. The real distances were measured by a laser distance meter and also by a tape measure with an accuracy of about ±2 mm. The deviation between the mean value and the true distances is given in Fig. 7. Depth measurements among 800 mm seems not to be suitable. 60 Fig. 5: Array geometry with Kinect V2. Deviation / mm 40 20 0-20 Next, some influences on the depth images are discussed. Previous investigations have shown that the Kinect V2 need a pre-heating time before providing accurate range measurements [10]. After 20 minutes of usage the distance variation becomes nearly constant (more or less 1 mm). The depth images oscillate during the measurement, known as wiggling error of ToF cameras [11]. Therefore the depth images have been averaged to decrease this effect. The decrease of the depth fluctuations by averaging is shown in Fig. 6. One can state, that an averaging of at least 50 frames should be done to get accurate results. Next the deviation between measured depth and the real distance was determined. For this, the sensor was placed parallel to a plane white wall at several distances. Mean value Standard deviation -40 500 1000 1500 2000 2500 3000 3500 4000 4500 5000 Real distance / mm Fig. 7: Depth variation of measured and true distances. Furthermore, using the same setup the standard deviation for each pixel of the depth images were computed. The measurements show that with increasing distance the errors towards the edges increase (see Fig. 8). Therefore, to make accurate long-distance measurements, the DUT should be placed in the center of the image. Further influences are given by the albedo of surfaces.

(a) (c) (d) III. CALIBRATION For underlying the beamform map by the RGB image the camera was calibrated using the algorithm described in [12]. To this purpose a checkerboard is placed in different positions in front of the camera. The dimensions of the checkerboard pattern have to be known. For best results 10 to 20 images of the pattern should be used. The images should be taken at a distance approximately equal to the distance from the camera to the DUT. As a rule of thumb, the checkerboard should cover at least 20%oftheimage. AsonecanseeinFig. 10thecheckerboard should also be captured at different orientations. Fig. 8: Standard deviation of each pixel Considerd distances (a) 800mm 1000mm (c) 1500mm (d) 2000 mm, averaging 50 frames. It has been shown that on very reflective as well as very dark surfaces the corresponding distances in the depth images are larger then expected [10]. To overcome this limitation the supposed surfaces in the experimental setup (reflective and dark) were covered by a developer spray. Especially the reflective surface of the hairdryer in the experimental setup leads to depth errors (see Fig. 9). Fig. 10: RGB calibration Different positions and orientations of the checkerboard. At the end of the calibration process, one obtains the intrinsic and extrinsic camera parameters. The intrinsic parameters of the RGB camera are compared to a self calibration implemented in the Kin2 toolbox [13]. For the IR camera, the SDK provides a function which delivers the intrinsic parameters. The same calibration method, which was used for the RGB camera, can be also applied to the IR camera. In Tab. 3 the intrinsic parameters of the RGB and the IR are listed and compared. (a) (c) RGB camera IR camera [13] self SDK self f x (px) 1063.86 1110.25 366.8731 365.62 ±95.84 ±17.84 f y (px) 1063.86 1135.20 366.8731 373.95 ±98.49 ±17.77 c x 978.54 953.58 259.78 254.09 ±15.16 ±2.85 c y 535.62 539.22 208.02 254.09 ±17.83 ±4.98 K 1 0.01849 0.05550 0.09639 0.12603 ± 0.03501 ± 0.04531 K 2 0.01016 0.01286 0.27167 0.34688 ± 0.08980 ± 0.16376 K 3 0.01006 0.04601 0.08992 0.16029 ± 0.06957 ± 0.15585 Tab. 3: RGB und IR camera intrinsics Focal length (f x, f y ), principal point (c x, c y ), radial distortion coefficients (K 1, K 2, K 3 ). Fig. 9: Albedo influence (a) Experimental setup setup after covering the surfaces by developer spray (c) depth errors between (a) and of an averaged snapshot of 50 frames. Knowing these parameters the lens distortion can be corrected, see Fig. 11. The parameters are quite sensitive to the selected images for the calibration. However, for the purpose of this work the factory-set calibration values are used. As previously mentioned the SDK pro-

vides also a coordinate mapping between RGB and IR image which should be used. (a) Sound pressure level (ref. 20µPa) / db 80 60 40 20 0-20 Signal One-third oktave Noise level Fig. 11: Lens distortion (left) distorted RGB image (right) undistorted RGB image. The depth images were corrected by using the information of Fig. 7. For overlaying the RGB image with the beamform map some corrections have to be done. Due to the fact, that the y-axis of the Kinect V2 doesn t coincide with the microphone plane, a 1.8 degree rotation about the x-axis was done. Reason for that could be a non-perfect attachment of the sensor. Furthermore, the origin of the Kinect V2 coordinate system doesn t match the origin of the array coordinate system, since the Kinect V2 isn t placed in the center of the array. The beamform map has to be shifted in the y and -x direction. The translation in y direction is 80mm and in the -x direction 120mm. -40 10 3 10 4 Frequency / Hz Fig. 13: Frequency spectrum of the center microphone. maps a one-third octave analysis was done. First the two-dimensional mapping using a constant distance Z between microphone plane and scanning area was chosen. Then the beamform maps were put on the RGB imageofthekinectv2. TheresultsaregiveninFig. 14. The beamform maps were normalized to the maximum. IV. EXPERIMENTAL RESULTS Experiments with real sources are made to demonstrate the acoustic source mapping using the RGB and IR images. For this purpose two smallband and a broadband noise source were used (two speakers and a hairdryer, see Fig. 12). Speaker 1 and the hairdryer are almost at the same distance to the microphone plane. Speaker 2 is approximately 50 cm behind them. The Fig. 12: Experimental setup. sampling frequency has been 48 khz, the measurement time 5s and the temperature 25 C. The frequency spectrum of the center microphone and the noise level is depicted in Fig. 13. The spectrum was averaged 100 times using the Hanning window and a block size of 4096 samples with a block overlapping of 50%. To identify sound sources one choose the characteristic peaks of the spectrum. For obtaining the beamform Fig. 14: Beamform maps one-third octave analyses (left) 1600Hz, Z = 1400mm (right) 8000Hz, Z = 1900 mm. Next the depth informations of the sensor should be used. There are two possible ways to use them. First, the depth image can be used as a weighting of the twodimensional beamform map. To do that, the beamform map will be calculated in a normal way using a constant Z. Thenthisresultwillbemappedonthe3dimensional scene of the depth image. Second, the depth information can directly be used in Eq. 2 as assumed source point x s, meaning that the steer vectors depends on the Kinect V2 measurements. The mapping process of both ways is shown in Fig. 15. In the surface mapping the depthinformationsinside1.3mand1.5mwereused. To show the difference between both methods the surface map was projected on a plane (see Fig. 16). Both methods provide the sound source in the one-third octave of 1600 Hz (hairdryer). Differences in the beamform maps are given through the different distances used for the calculation of the steer vectors.

(a) VI. REFERENCES [1] Th. Mueller, Aeroacoustic Measurements, Springer, ISBN 3-540-41757-5, 2002. [2] R. Dougherty, Functional Beamforming, 5th Berlin Beamforming Conference, BeBeC-2014-01, 2014. [3] P. Sijtsma, Clean based on spatial source coherence, Int. J. Aeroacoustics 6, pp 357-374, 2009. [4] E. Sarradj, A fast signal subspace approach for the determination of absolute levels from phased microphone array measurements, Journal of Sound and Vibration 329, pp 1553-1569, 2010. [5] J.R. Underbrink, Circularly symmetric, zero redundancy, planar array having broad frequency range applications, Pat. US6205224 B1, 2001. [6] Z. Prime and C. Doolan, A comparision of popular beamforming arrays, Proceedings of ACOUSTICS, 2013 Fig. 15: One-third octave (1600 Hz) Surface map of (a) method 1 and method 2. [7] https://www.microsoftstore. com/store/msusa/en_us/pdp/ Kinect-Adapter-for-Xbox-One-S-and-Windows-PC/ productid.2233937600. [8] J. Sell and P. O Connor, The Xbox One System on a Chip and Kinect Sensor, IEEE Micro, vol 34, no. 2, pp 44-53, 2014. [9] P. Fankhauser, M. Bloesch, D. Rodriguez, R. Kaestner, M. Hutter, R. Siegwart, Kinect v2 for Mobile Robot Navigation: Evaluation and Modeling, 2015 International Conference on Advanced Robotics (ICAR), IEEE, pp 388-394, 2015. Fig. 16: Comparision (left) method 1 (right) method 2. V. CONCLUSION The applicability to use the Kinect V2 sensor on beamforming was demonstrated. The surface mapping process provided good results. The different effects on the acquired depth images were examined. These yields to some corrections on the depth image. Moreover the overall measurement system was presented. Further investigations should be done to see, if a manual calibration and the point cloud acquisition (mapping between the depth and RGB image) can enhance the accuracy. [10] E. Lachat, H. Macher, M.-A. Mittet, T. Landes, P. Grussenmeyer, First experiences with Kinect V sensor for close range 3D modelling ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, vol XL- 5/W4, pp 93-100, 2015 [11] S. Foix, G. Alenya and C. Torras, Lock-in Time-of- Flight (ToF) Cameras: A Survey, IEEE Sensors Journal, vol 11, no. 3, pp 1-11, 2011. [12] Z. Zhang. A flexible new technique for camera calibration, IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(11):1330-1334, 2000. [13] J. R. Terven and D. M. Córdova-Esparza, Kin2. A Kinect 2 toolbox for MATLAB, Sience of Computer Programming, vol 130, pp 97-106, 2016, http://dx.doi.org/10.1016/j. scico.2016.05.009