EEE 187: Robotics. Summary 11: Sensors used in Robotics

Similar documents
Range Sensing strategies

Brainstorm. In addition to cameras / Kinect, what other kinds of sensors would be useful?

Intelligent Robotics Sensors and Actuators

Sensing. Autonomous systems. Properties. Classification. Key requirement of autonomous systems. An AS should be connected to the outside world.

Lab 2. Logistics & Travel. Installing all the packages. Makeup class Recorded class Class time to work on lab Remote class

Development of intelligent systems

CENG 5931 HW 5 Mobile Robotics Due March 5. Sensors for Mobile Robots

MOBILE ROBOTICS. Sensors An Introduction

Robot Hardware Non-visual Sensors. Ioannis Rekleitis

Sensors. human sensing. basic sensory. advanced sensory. 5+N senses <link> tactile touchless (distant) virtual. e.g. camera, radar / lidar, MS Kinect

An Example of robots with their sensors

PRESENTED BY HUMANOID IIT KANPUR

An Example of robots with their sensors

Sonic Distance Sensors

Laser Telemetric System (Metrology)

EL6483: Sensors and Actuators

COS Lecture 7 Autonomous Robot Navigation

Chapter 2 Sensors. The Author(s) 2018 M. Ben-Ari and F. Mondada, Elements of Robotics, / _2

Robot Sensors Introduction to Robotics Lecture Handout September 20, H. Harry Asada Massachusetts Institute of Technology

Indoor Positioning by the Fusion of Wireless Metrics and Sensors

Lecture: Sensors , Fall 2008

MEM380 Applied Autonomous Robots I Fall Introduction to Sensors & Perception

Sensors and Actuators

Shaft encoders are digital transducers that are used for measuring angular displacements and angular velocities.

As before, the speed resolution is given by the change in speed corresponding to a unity change in the count. Hence, for the pulse-counting method

By Pierre Olivier, Vice President, Engineering and Manufacturing, LeddarTech Inc.

Electronic Systems - B1 23/04/ /04/ SisElnB DDC. Chapter 2

ELECTRONIC SYSTEMS. Introduction. B1 - Sensors and actuators. Introduction

Technical Explanation for Displacement Sensors and Measurement Sensors

Hydroacoustic Aided Inertial Navigation System - HAIN A New Reference for DP

Actuator Components 2

electronics for computer engineering (Sensor) by KrisMT Computer Engineering, ICT, University of Phayao

Perception. Autonomous Mobile Robots. Sensors. Vision Uncertainties, Fusion Features. Autonomous Systems Lab. Zürich. Cognition.

Electronics II. Calibration and Curve Fitting

Active Stereo Vision. COMP 4102A Winter 2014 Gerhard Roth Version 1

Sensing and Perception

10/21/2009. d R. d L. r L d B L08. POSE ESTIMATION, MOTORS. EECS 498-6: Autonomous Robotics Laboratory. Midterm 1. Mean: 53.9/67 Stddev: 7.

GPS data correction using encoders and INS sensors

Introduction to Arduino HW Labs

NCERT solution for Sound

Industrial Sensors. Proximity Mechanical Optical Inductive/Capacitive. Position/Velocity Potentiometer LVDT Encoders Tachogenerator

Introduction to Mobile Sensing Technology

Sensors. Chapter 3. Storey: Electrical & Electronic Systems Pearson Education Limited 2004 OHT 3.1

Introduction. ELCT903, Sensor Technology Electronics and Electrical Engineering Department 1. Dr.-Eng. Hisham El-Sherif

GPS and Recent Alternatives for Localisation. Dr. Thierry Peynot Australian Centre for Field Robotics The University of Sydney

Introduction to ROBOTICS. Robot Sensing and Sensors

Perception. Read: AIMA Chapter 24 & Chapter HW#8 due today. Vision

Tektronix AFG10022 Function Generator. Coming soon to B10: Sin, Square, Ramp, Swept, Arbitrary, Noise. Linear Actuators. Non-magnetized iron plunger

GPS-Aided INS Datasheet Rev. 2.6

Position Sensors. The Potentiometer.

Actuators, sensors and control architecture

1.6 Beam Wander vs. Image Jitter

Data Sheet. AEDT-9140 Series High Temperature 115 C Three Channel Optical Incremental Encoder Modules 100 CPR to 1000 CPR. Description.

Shock Sensor Module This module is digital shock sensor. It will output a high level signal when it detects a shock event.

5. Transducers Definition and General Concept of Transducer Classification of Transducers

Robotic Vehicle Design

Time of Flight Capture

Degree of mobility Degree of steerability

Section 1: Sound. Sound and Light Section 1

PVA Sensor Specifications

Electronic Instrumentation and Measurements

Ultrasonic Level Detection Technology. ultra-wave

16. Sensors 217. eye hand control. br-er16-01e.cdr

9/28/2010. Chapter , The McGraw-Hill Companies, Inc.

WELCOME TO THE SEMINAR ON INTRODUCTION TO ROBOTICS

CRACK DETECTION SYSTEM FOR RAILWAY TRACK BY USING ULTRASONIC AND PIR SENSOR

Understanding How Frequency, Beam Patterns of Transducers, and Reflection Characteristics of Targets Affect the Performance of Ultrasonic Sensors

Chapters 1 & 2. Definitions and applications Conceptual basis of photogrammetric processing

CSE 165: 3D User Interaction. Lecture #7: Input Devices Part 2

Govt. Engineering College Jhalawar Model Question Paper Subject- Remote Sensing & GIS

746A27 Remote Sensing and GIS

t =1 Transmitter #2 Figure 1-1 One Way Ranging Schematic

Acknowledgment. Process of Atmospheric Radiation. Atmospheric Transmittance. Microwaves used by Radar GMAT Principles of Remote Sensing

Robotics Intelligent sensors (part 1)

SMART LASER SENSORS SIMPLIFY TIRE AND RUBBER INSPECTION

Data Sheet. AEDT-9340 Series High Temperature 115 C 1250/2500 CPR 6-Channel Commutation Encoder. Description. Features.

Airborne Laser Scanning. Long-Range Airborne Laser Scanner for Full Waveform Analysis. visit our webpage LASER MEASUREMENT SYSTEMS

Sensing self motion. Key points: Why robots need self-sensing Sensors for proprioception in biological systems in robot systems

Instrumentation (ch. 4 in Lecture notes)

Chapter 16. Waves and Sound

Introduction to Measurement Systems

Waves Q1. MockTime.com. (c) speed of propagation = 5 (d) period π/15 Ans: (c)

DIGITAL LASER DISTANCE METER

Robotic Vehicle Design

Intext Exercise 1 Question 1: How does the sound produced by a vibrating object in a medium reach your ear?

Proximity Sensors Ultrasonic Precision Proximity Sensors

GUIDED WEAPONS RADAR TESTING

SELF STABILIZING PLATFORM

Data Sheet. AEDB-9340 Series 1250/2500 CPR Commutation Encoder Modules with Codewheel. Features. Description. Applications

Preliminary Datasheet

Co-Located Triangulation for Damage Position

GPS-Aided INS Datasheet Rev. 3.0

Data Collection: Sensors

9. Microwaves. 9.1 Introduction. Safety consideration

Estimation of Absolute Positioning of mobile robot using U-SAT

Real-Time Scanning Goniometric Radiometer for Rapid Characterization of Laser Diodes and VCSELs

ACTUATORS AND SENSORS. Joint actuating system. Servomotors. Sensors

Acoustic Velocity Independent Ultrasonic Flow-Meter

Interactive Simulation: UCF EIN5255. VR Software. Audio Output. Page 4-1

GPS-Aided INS Datasheet Rev. 2.3

Transcription:

1 EEE 187: Robotics Summary 11: Sensors used in Robotics Fig. 1. Sensors are needed to obtain internal quantities such as joint angle and external information such as location in maze Sensors are used to acquire knowledge about the environment as well as the internal state of the robot as illustrated in figure 1. Sensors are used for feedback control and external interaction with the world. They are designed to convert energy from one form to another form (the output is usually electrical energy) from which data is interpreted. Sensors have the opposite role of actuators as shown in figure 2, they are input transducers while actuators are output transducers. Animals and humans use sensors for every day activities where signals captured by the sensory system are sent to the brain where they are processed and interpreted. Human senses Smell Touch Taste Hearing Vision Robots may have additional sensors that humans do not have such as radioactivity sensors for example. A. Choice of sensors Different factors influence the choice of the sensor, such as: Cost Size and weight Type of the output It is important to use the right sensor for the right application. Fig. 2. Illustration of sensors and actuators B. Classification of sensors There are different ways to classify sensors. Proprioceptive vs. Exteroceptive Proprioceptive sensors: measure quantities internal to the system such as the motor speed and the battery charging state. Exteroceptive sensors: acquire information about the robot s environment such as the location of the obstacles. Analog vs digital sensors: The output of the sensor could be analog or digital. Analog sensors need A/D converter when connected to the microcontroller. Active vs. passive: Passive sensors measure ambient environment energy entering the system. Example: microphone, camera, battery sensor. Active sensors emit energy into the environment, and then measure the environment response. Example: laser range finders. Usually active sensors have superior performance, they are able to create a more controlled environment. Drawbacks of active sensors: signals emitted from other sources may influence the readings. C. Sensor characteristics Dynamic range: The dynamic range refers the spread between lower

and upper limits of the sensor measurements. Can be calculated in db. Example: a sensor can measure current values from 1mA to 20mA. The dynamic range is: 10 log 10 (20/1)dB (1) Repeatability: The ability of a sensor to repeat a measurement when put back in the same environment. Resolution: Refers to the minimum difference between two values that can be detected. For a digital device of n bits, the resolution is given by Resolution = full range 2 n (2) Example An encoder converts angular position to a digital or analog code. Find the resolution of an absolute encoder with 4 bits. Solution With 4 bits, an encoder can report angular position up to 16 levels, that is 2 4. Resolution = 360 16 = 22.50 (3) Linearity: If x and y are the inputs of the sensor, f(x) and f(y) are the outputs corresponding to x and y, respectively, linearity implies that for any values of a and b, we have: f(ax + by) = af(x) + bf(y) (4) This means if you plot the input versus the output, you have a straight line. Frequency: Number of measurements per second Sensitivity: Amount in the input that will change the output, formally: output change input change Error: The difference between the measured and the true values (5) Error = m v (6) where m is measured and v is true value Accuracy: The accuracy is defined in terms of the error as follows Accuracy = 1 error (7) v When the error decreases, accuracy increases. The true value refers to the theoretical value. Systematic errors: Errors caused by factors that can be modeled mathematically such as poor calibration. These errors are deterministic, thus predictable. Fig. 3. Additional inputs Random errors: Cannot be predicted using a mathematical model. However, they can be modeled using probabilistic models. Reproducibility: Multiple readings of the same quantity results in the same output. Precision: Precision is defined in terms of the range as follows: P recision = range (8) σ where σ is the standard deviation. Dynamic characteristics: The properties of the system transient response to a given input. Response time can be seen as a dynamic characteristic. Additional inputs: Figure 3 shows two types of additional inputs: interference inputs and modifying inputs. Modifying inputs will change the calibration curve as shown in figure 3. Temperature is a typical modifying input. Hysteresis: Hysteresis is a nonlinearity where there are two values for the output that depend on the trajectory and therefore on the history of the system (4). This property is common in magnetic sensors. D. Sensors for Mobile Robots Robotics sensors can be grouped into different categories: Inertial sensors Accelerometer Gyroscope Compass Inclinometer: mainly measures angles of slope. Visual sensors: Camera Satellite-based sensors (global): GPS Contact sensors: Bumpers and whiskers Proximity sensors and range finders: Sonar (time of flight (ToF)) 2

Fig. 4. An example for hysteresis) Fig. 5. Incremental and absolute encoders (binary and gray code) Radar (phase and frequency) Laser range-finders (triangulation, ToF, phase) Infrared (intensity) Position sensors Position sensors are used to measure displacement, rotary and linear. In many situations, the position information is used for velocity calculation. Encoders are electromechanical devices that convert the angular position or motion of a shaft or axle into an analog or digital quantity. A rotary encoder measures the rotation of a shaft or axle. Rotary encoders are used for example to measure the angle of a robotic arm, or how far a mobile robot have moved by measuring the rotations of its wheels. A linear encoder is similar to a rotary encoder, but measures position in a straight line, rather than rotation. There are two types of encoders: absolute and incremental. Potentiometers: They are used to convert position information into a variable voltage. Potentiometers can be rotary or linear. They are generally used as internal feedback sensors in order to record the position of joints and links. Working principle of encoders A typical encoder uses optical sensor(s), a moving mechanical component, and a special reflector to provide a series of electrical pulses to the microcontroller. Incremental Encoders : An incremental rotary encoder provides cyclical outputs when the shaft is rotated. The arcs of clear and opaque sections are all equal and repeating. Each arc indicates an angle of revolution. The accuracy increases with the number of divisions. For example, if the wheel is divided into two potions, the resolution is 180 o, if the wheel is divided into 16 positions, the resolution is 22.5 o. Typical incremental encoders can have 512 to 1024 arcs in them. This type of encoder does not directly indicate the value of the actual position. Absolute Encoders : There are multiple rows of arcs: The first one has only one clear and one opaque arc (one on, one off), the next row would have 4 sections, and the next one would have 8. Each row must have its own light source and light sensor. Each sensor sends out a signal, the position is obtained by combining information from all rows. Figure 5 illustrates incremental and absolute encoders. Touch and tactile sensors Touch sensors are devices that send a signal when a physical contact has been made. A micro-switch is the simplest touch sensors. Force sensors provide more information: they scan measure how strong is the touching force. Whiskers and bumpers are simple touch sensors, where a mechanical contact leads to closing/opening of a switch change in resistance of some element change in capacitance of some element Touch sensors are simple, cheap, and have binary output. Tactile sensors: A tactile sensor is a collection of touch sensors. Tactile sensors provide additional information about the object such as the shape, size, and material type. In general a number of touch sensors are arranged in an array or matrix to form a tactile sensor. Proximity sensors A proximity sensor is used to determine that an object is close to another object before contact is made. Magnetic proximity sensors: they are active when they are close to a magnet. Optical proximity sensors: they use a light source (called an emitter) and a receiver, the receiver senses the presence or absence of light. Ultrasonic proximity: the ultrasonic emitter emits frequent bursts of high frequency sound waves (200 khz range). 3

determines its location. In trilateration, the robot uses distances only. Trilateration is discussed in the GPS section. Sniff Sensors Sniff sensors are similar to smoke sensors. They are sensitive to particular gases and send a signal when they detect those gases. There are many applications of sniff sensors such as explosive detection. SONAR SENSING Fig. 6. Touch and tactile sensors Fig. 7. Potentiometer, encoder, IR sensor, and laser Range finders Unlike proximity sensors, range finders are used to: Find larger distances Map surface of an object Various common methods are used for range detection: Time of flight TOF or lapsed time. Triangulation Phase and frequency Most proximity and ranger finders are based on sound, such as ultrasonic sensors or light such as infrared and laser. Triangulation and trilateration Triangulation and trilateration can be used to construct inexpensive but highly effective navigation systems. Three or more active transmitters mounted at known locations in the environment are used. In triangulation, the robot register the angles at which it sees the transmitters, from which the robot to Sonar or ultrasonic sensing uses propagation of acoustic energy where acoustic pulses and their echoes are used to measure the range to an object. Ultrasonic uses frequency higher than audible frequency (frequencies between 20Hz and 20KHz). PING for example uses pulses at 40KHz. Ultrasonic sensors are popular in robotics. Their popularity is due to the following: Low cost Light weight Low power consumption Low computational effort Ultrasonic sensors are used for different purposes including Obstacle avoidance Sonar mapping: perform rotational scan (360 for example) to construct of a map of the environment. Object recognition: a sequence of sonar maps is processed using data fusion algorithms. Ultrasonic sensors use the time of flight to determine the range: where r o = v st o 2 v s is the speed of sound, approximately 343 m/s at standard temperature and pressure. r o is the object range Division by 2: round trip E. Problems with sonar sensors: Smooth surfaces at oblique incidence do not produce detectable echoes. The echo from probing pulse 1 occurs after pulse 2 is emitted. Sonar measures the ToF of the most recent probing pulse and ends up with: T of = echo of pulse 1 emission time of pulse 2. (10) Some materials absorb sound energy (such as foam). Fail to produce a reflection that is sufficiently strong to be sensed by the receiver. Environmental ultrasonic sounds from pumps and motors affect ultrasonic readings because of the noise they induce. (9) 4

F. Sonar beam pattern The beam pattern describes the direction pattern of the energy. The beam pattern is shown in figure 10 and can be approximated by a cone. Two factors affect the sonar beam: the radius a of the sensor (assuming the sensor has a circular shape) and the frequency of the waveform. Sonar emitter is usually modeled as a circular surface of radius a. The frequency f is related to the wavelength λ as follows: and where θ 0 is the cone angle. λ = v/f (11) θ 0 = arcsin(0.61λ/a) (12) Fig. 8. Pulse 2 is transmitted before echo from first pulse arrives to the receiver. This leads to the Tof given equation (10) G. Example Calculate θ o for the PING sensor: We know that f = 40KHz. If a = 1cm, then θ o = 30 approximately. Speed of sound The speed of sound varies with temperature pressure humidity The speed of sound is given by v s = 20.05 T c + 273.16m/s (13) at sea level pressure in dry air. In general the speed of sound in air is v s = γrt K (14) Fig. 9. False reading where γ is ratio of specific heat, R is the gas constant and T K is the temperature in Kelvin. Some properties Sonar measurements are affected by the motion of the sonar sensor or the target. The ultrasonic wave typically has frequency between 40 and 180 KHz and it is usually generated by a piezo or electrostatic transducer. Ultrasonic sensors used in robotics have typical range 2 cm to 5m. For PING for example, the range is from 2cm to 3m. Fig. 10. Normalized amplitude of echo as a function of angle QTI Line Sensor QTI senor uses an infrared (IR) reflective sensor to determine the reflectivity of the surface. When the QTI sensor is over a dark surface, the reflectivity is very low; when the QTI is over a light surface, the reflectivity is very high and will cause a different reading from the sensor. The QRD1113/14 reflective sensor consists of an infrared emitting diode and an NPN silicon phototransistor mounted side by side in a black plastic housing as shown in figure 11. The on axis 5

Fig. 11. Infrared proximity sensor and QTI sensor radiation of the emitter and the on-axis response of the detector are both perpendicular to the face of the QRD1113/14. The phototransistor responds to radiation emitted from the diode only when a reflective object or surface is in the field of view of the detector. Fig. 12. Integration of gyros and accelerometers INERTIAL SENSORS Inertial sensors are used to determine the robot s pose (such as position, orientation and inclination). Inertial sensors are usually proprioceptive. According to the general relativity theory, only force and angular velocity can be measured inside a isolated system. Therefore, inertial sensors provide estimate of the linear acceleration and the angular velocity. The linear velocity, orientation and position can be obtained using integration. Inertial measurements depend on the measurement carried out inside the system; therefore, they are not affected by the environment. The most popular inertial sensors are summarized below: Compass: measures the weak magnetic field of the earth. Gyroscope: an inertial sensor that is used to measure the rate of rotation independent of the coordinate frame. Accelerometer: can be used to measure all external forces acting on an object including gravity. Different mechanisms are used to translate forces into a readable signal. The most traditional mechanism is the mechanical accelerometer, which is essentially, a spring mass damper system. Most accelerometers fo not handle jitter very well, which happens frequently in wheeled and especially in walking robots. As a result, some software corrections have to be made for signal filtering. Inertial measurement unit (IMU): uses gyroscopes and accelometers (and sometime magnetometers) to estimate the relative position, velocity and acceleration of a moving object widely used for ships and airplanes. First demonstrated in 1949 by CS Drapper, IMUs have become popular navigation tools widely used in ships and airplanes, and in robotics applications as well. In general IMU has three orthogonal accelerometers and three orthogonal gyroscopes. The integration process is shown in figure 12. An IMU system can estimate the 6 Fig. 13. The GPS system DOF pose (x, y, z) and orientation (roll, pitch and yaw) of a vehicle. I. GLOBAL POSITIONING SYSTEM: GPS is the most commonly used mechanism for location estimation and navigation. It provides a 3 dimensional position estimate in absolute coordinates (longitude, latitude and height co-ordinates) accurate to within a range of 20m to 1mm, as well as time and date (Universal Time Coordinates (UTC)), accurate to within a range of 60ns to 5ns. The GPS is based on 24 satellites + additional satellites. The satellites are organized into six orbital planes with four satellites in each. GPS is available anywhere on earth. Originally, GPS was developed for military applications by the U.S. Department of Defense, but it was later open to civilian applications. The civilian signal SPS (Standard Positioning Service) can be used freely by the general public, whilst the military signal PPS (Precise Positioning Service) 6

clock is slow. We do not know T 0. The pseudo range can be calculated using the following equation: P SR i = c T measured,i = c( T i + T 0 ) = R i +c T 0 (16) Fig. 14. The trilateration process can only be used by authorized government agencies. The first satellite was placed in orbit on 22nd February 1978. There are several global positioning systems in the world, when we talk about GPS we refer to the NAVSTAR (NAVigation System with Timing And Ranging) system in general. The other major systems are GLONASS satellite system (Russia) Galileo satellite system (EU) When people talk about GPS, they usually mean GPS receiver. There are at least 24 satellites orbiting the earth. The orbits are arranged so that at any time, anywhere on earth, at least four satellites are visible. The satellites have highly accurate atomic clocks (error is 1s in 300000 years). Signal transit time GPS PSEUDO RANGE EQUATIONS How far are you from a flash of lightning? To calculate the distance you can use the signal transit time: Time the lightning is perceived is the start time Time the thunder is heard is the stop time Distance = transit time speed of sound (15) Figure 14 shows 4 satellites and a receiver (user). The coordinates of the GPS satellites are known in a global reference frame. To determine the global position, the GPS receiver uses signals from four different satellites. We have four signal transit times T i (i = 1,..., 4). The range of the user from the satellite is denoted by R i, i = 1,..., 4. This is the true range or the geometric range. The location of the satellites is known. The satellites clocks are synchronized, the time at which the satellites send the signal is known very precisely. However, the received clock is not synchronized, the receiver clock is slow or fast by T 0. A positive value of T 0 means the received clock is fast and a negative value means the receiver where R i is the range between satellite i and the receiver (user) c is the speed of light T i is the signal transit time from the satellite to the user. T 0 is the difference between satellite clock and user clock. In other words the offset of the user clock. P SR i is the pseudo range We can write for each satellite: and R i = (x i x u ) 2 + (y i y u ) 2 + (z i z u ) 2 (17) P SR i = R i + c T 0 (18) Because we have four unknown variables: (x u, y u, z u ) and t 0, we need four independent variables to solve P SR i = (x i x u ) 2 + (y i y u ) 2 + (z i z u ) 2 + c T 0 i = 1,..., 4 (19) from which we can solve for the position of the user. LASER RANGE FINDERS Laser stands for: Light Amplification by Stimulated Emission of Radiation. Light is an electromagnetic radiation but Laser has some different and interesting properties. Because of their properties, lasers are used in a wide variety of applications. Laser corresponds to intense beams of light which is monochromatic, coherent, and highly collimated. The wavelength of laser light is extremely pure (monochromatic, single color) when compared to other sources of light. Light from a laser typically has very low divergence. It can travel over great distances or can be focused to a very small spot with a brightness which exceeds that of the sun. Different techniques can be used to measure ranges using laser such as Time of flight (ToF): very similar to ultrasonic sensors. For lasers, electronics capable of resolving picoseconds are required, thus the circuitry is expensive. The ToF method is typically used for large distances such as hundreds of meters or many kilometers. Phase shift measurement: uses an intensity-modulated laser beam with a constant frequency and measures the phase difference between transmitted and received signals as shown in figure 15. The wavelength of the modulating signal is λ with λ = c (20) f The distance to the object is given by D = λθ (21) 4π where θ is the measured phase difference between the transmitted and the reflected light beams. Once angle θ is measured, distance D is obtained using equation (21). 7

Distance measurement by the phase-shift technique is a good method to obtain a resolution of some millimeters in 1 to 20m ranges. Example Assume that the modulating signal has a wavelength of λ = 60m (f = 5MHz), what is the phase measurement for a range of: a range of 5m? a range of 35m? Solution In both cases we have θ = 60 o (22) There is an ambiguity interval for λ. It is possible to decrease the ambiguity interval by increasing λ. For example, if λ = 120m, D = 5m will be confused with D = 65m. A. Optical triangulation This is another method used to determine range. Triangulation can be defined as follows: if the length of one side and two interior angles of a triangle are known, then the length of the two remaining sides and the other angle can be determined. An optical triangulation device shown in figure 16 has two components: Collimated light source: any focused light can be used, including laser. Camera: position sensitive device An illustration of the sensor is shown in figure 16, where B is the base line and α is the angle between the camera optical axis and the laser beam, and f is the focal length of the camera. These variables are properties of the system and are known. We want to determine the distance L to a target surface. The light sensor (laser) projects a pattern of light onto the object and p is the position of point p in the pixel reference frame. The only unknown variables are γ, L. When the projection of point p in the image frame is known, we can determine angle γ as follows: γ = arctan u (23) f Fig. 15. Phase shift measurement Using trigonometry, we can find: L = B tan(α + γ) (24) Effect of the parameters Base line: Large value of B results in better resolution, but for small values of B, the sensors is more compact. Focal length: Long focal length results in better accuracy but smaller field of view. Fig. 16. Optical triangulation 8