INTRODUCTION TO C-NAV S IMCA COMPLIANT QC DISPLAYS

Similar documents
C Nav QA/QC Precision and Reliability Statistics

Sensors. DGNSS Position Quality Information for DP Applications

SENSORS SESSION. Operational GNSS Integrity. By Arne Rinnan, Nina Gundersen, Marit E. Sigmond, Jan K. Nilsen

DGNSS Position Quality Information for DP Applications

GPS NAVSTAR PR (XR5PR) N/A

Report on Extended Kalman Filter Simulation Experiments

Shared Use of DGPS for DP and Survey Operations

IEFIS GPS manual Applicable to iefis G3 including Lite versions Firmware or later

hydro8 Precise Point Positioning, the new DGPS: The C-Nav Experience Edwin Danson FRICS FInstCES FRAS Business strategy consultant

GLOBAL POSITIONING SYSTEM (GPS) PERFORMANCE APRIL TO JUNE 2017 QUARTERLY REPORT

Galileo: The Added Value for Integrity in Harsh Environments

The Wide Area Augmentation System

GNSS for Landing Systems and Carrier Smoothing Techniques Christoph Günther, Patrick Henkel

Demonstrations of Multi-Constellation Advanced RAIM for Vertical Guidance using GPS and GLONASS Signals

INTEGRITY AND CONTINUITY ANALYSIS FROM GPS JANUARY TO MARCH 2017 QUARTERLY REPORT

SPAN Technology System Characteristics and Performance

The experimental evaluation of the EGNOS safety-of-life services for railway signalling

VEHICLE INTEGRATED NAVIGATION SYSTEM

LOW POWER GLOBAL NAVIGATION SATELLITE SYSTEM (GNSS) SIGNAL DETECTION AND PROCESSING

ARAIM Fault Detection and Exclusion

GLOBAL POSITIONING SYSTEM (GPS) PERFORMANCE JANUARY TO MARCH 2016 QUARTERLY REPORT

Worst-Case GPS Constellation for Testing Navigation at Geosynchronous Orbit for GOES-R

Sources of Error in Satellite Navigation Positioning

TSC1 - Asset Surveyor Operation

PROCEDURE FOR GNSS EQUIPMENT VERIFICATION IN STATIC POSITIONING

Multipath Error Detection Using Different GPS Receiver s Antenna

DYNAMIC POSITIONING CONFERENCE October 7-8, Sensors II. Redundancy in Dynamic Positioning Systems Based on Satellite Navigation

Simulation Analysis for Performance Improvements of GNSS-based Positioning in a Road Environment

GROUND CONTROL SURVEY REPORT

High Precision Positioning Unit 1: Accuracy, Precision, and Error Student Exercise

Problem Areas of DGPS

Outlier-Robust Estimation of GPS Satellite Clock Offsets

RELEASE NOTES. Trimble. SPS Series Receivers. Introduction. New features and changes

Dynamic Positioning TCommittee

GNSS Conductor GF. User s Guide. (Document No. SE )

GLOBAL POSITIONING SYSTEM (GPS) PERFORMANCE OCTOBER TO DECEMBER 2013 QUARTERLY REPORT. GPS Performance 08/01/14 08/01/14 08/01/14.

Experiences with Fugro's Real Time GPS/GLONASS Orbit/Clock Decimeter Level Precise Positioning System

GLOBAL POSITIONING SYSTEM (GPS) PERFORMANCE JULY TO SEPTEMBER 2018 QUARTERLY REPORT 3

Mitigate Effects of Multipath Interference at GPS Using Separate Antennas

ProMark 3 RTK. White Paper

Absolute Antenna Calibration

INTEGRITY AND CONTINUITY ANALYSIS FROM GPS JULY TO SEPTEMBER 2016 QUARTERLY REPORT

Integrity Performance Models for a Combined Galileo/GPS Navigation System

Prototyping Advanced RAIM for Vertical Guidance

Integrity of Satellite Navigation in the Arctic

RELEASE NOTES. Introduction. Trimble Infrastructure GNSS Series Receivers

Effect of Quasi Zenith Satellite (QZS) on GPS Positioning

PPS usable by timing applications via serial port emulation

On the GNSS integer ambiguity success rate

Report of the Working Group B: Enhancement of Global Navigation Satellite Systems (GNSS) Services Performance

Lab Assignment #3 ASE 272N/172G Satellite Navigation Prof. G. Lightsey Assigned: October 28, 2003 Due: November 11, 2003 in class

Correlation of Model Simulations and Measurements

Precise Positioning with NovAtel CORRECT Including Performance Analysis

t =1 Transmitter #2 Figure 1-1 One Way Ranging Schematic

Tutorial on the Statistical Basis of ACE-PT Inc. s Proficiency Testing Schemes

GLOBAL POSITIONING SYSTEM (GPS) PERFORMANCE OCTOBER TO DECEMBER 2017 QUARTERLY REPORT

ACCURACIES OF VARIOUS GPS ANTENNAS UNDER FORESTED CONDITIONS

ELEVENTH AIR NAVIGATION CONFERENCE. Montreal, 22 September to 3 October 2003 TOOLS AND FUNCTIONS FOR GNSS RAIM/FDE AVAILABILITY DETERMINATION

HORIZONTAL ARAIM AVAILABILITY FOR CIVIL AVIATION OPERATIONS. ARAIM Outreach event

International Programmes & GNSS Monitoring

The Role of F.I.G. in Leading the Development of International Real-Time Positioning Guidelines

ARAIM: Utilization of Modernized GNSS for Aircraft-Based Navigation Integrity

Near Term Improvements to WAAS Availability

Performance Evaluation of the Effect of QZS (Quasi-zenith Satellite) on Precise Positioning

GLOBAL POSITIONING SYSTEM (GPS) PERFORMANCE JANUARY TO MARCH 2017 QUARTERLY REPORT

Fault Detection and Elimination for Galileo-GPS Vertical Guidance

Situational Awareness A Missing DP Sensor output

The Impact of Performance Parameters over a DGPS Satellite Navigation System

Leica Spider Infrastructure HW Solutions Introducing: Leica GR30 & GR50

LD2342 USWM V1.6. LD2342 V1.4 Page 1 of 18

OFDM Pilot Optimization for the Communication and Localization Trade Off

Evaluating EGNOS technology in an ITS driving assistance application

Automated Quality Control of Global Navigation Satellite System (GNSS) Data

Data Acquisition Experiment using NovAtel Dual Frequency GPS Receiver

Geodetic Research Laboratory

RAIM Availability prediction

Time Scales Comparisons Using Simultaneous Measurements in Three Frequency Channels

Introduction to Advanced RAIM. Juan Blanch, Stanford University July 26, 2016

Latest Developments in Network RTK Modeling to Support GNSS Modernization

THE PROPOSAL TO SNAPSHOT RAIM METHOD FOR GNSS VESSEL RECEIVERS WORKING IN POOR SPACE SEGMENT GEOMETRY

PORTABLE GNSS MONITORING STATION (PGMS)

SRT Marine Technology. LD2342 V1.4 Page 1 of 22

EVALUATION OF GPS BLOCK IIR TIME KEEPING SYSTEM FOR INTEGRITY MONITORING

Incorporating GLONASS into Aviation RAIM Receivers

GPS-G5 User s Manual

Integration of GPS with a Rubidium Clock and a Barometer for Land Vehicle Navigation

RTCM State Space Representation (SSR) Overall Concepts Towards PPP-RTK

Modelling GPS Observables for Time Transfer

Operating Handbook For FD PILOT SERIES AUTOPILOTS

Digiflight II SERIES AUTOPILOTS

GLONASS-based Single-Frequency Static- Precise Point Positioning

Laboratory 1: Motion in One Dimension

GPS Application. Global Positioning System. We provide GPS module ODM / OEM service, any GPS receiver you want, we can provide customized services.

Frank Heymann 1.

Distributed integrity monitoring of differential GPS corrections

GPS Technical Overview N5TWP NOV08. How Can GPS Mislead

Verify DP Operations Manual

Precise Point Positioning (PPP) using

Global Navigation Satellite System for IE 5000

Interferometric Approach to Complete Refraction Statics Solution

Transcription:

INTRODUCTION TO C-NAV S IMCA COMPLIANT QC DISPLAYS 730 East Kaliste Saloom Road Lafayette, Louisiana, 70508 Phone: +1 337.210.0000 Fax: +1 337.261.0192 DOCUMENT CONTROL Revision Author Revision description / comments Date Approved No / reviser 1.0 Initial draft 1.1 ED/RM 1.2 RM Additional interpretation section May 2014 1.3 RM Section added on statistical testing June 2014

Contents Introduction... 3 Statistics and More Statistics... 3 Precision... 4 Reliability... 7 Monitoring the Position Solution - RAIM... 8 Verifying the Model... 8 Getting the Data... 8 Rocket Science?... 9 C-Nav Solution... 9 QA/QC Panel General Description... 9 Numeric Indicators... 9 Pass / Fail Indicators... 10 Time Slice Bar... 10 Time Series Button... 10 Examples... 11 Interpreting the Parameters... 16 Precision... 16 Reliability... 16 RAIM... 16 Unit Variance... 17 W-Test... 17 F- Test... 17 Is There a Downside?... 18 Conclusions... 18 File Name: Introduction to C-Nav QC V1.3.docx Page 2 of 19

Introduction The traditional metrics for GPS position quality, DOP, number of satellites, range residuals, and signal to noise, are becoming less reliable with the move to multi-constellation ( GPS + GLONASS ) systems. This is not because the GNSS solution is less reliable, but because the traditional quality measurements are almost consistently good. For example HDOP rarely rises above 2 and the number of satellites used is typically over 15. With so many ranges available, receivers can do complex internal integrity monitoring (RAIM) to reject individual ranges where an error is suspected. Another factor has been the general acceptance of the PPP ( precise point positioning ) method of correcting the error in the GNSS broadcast parameters. This method corrects the sources of the GNSS error in clock and orbit, rather than the tradition method which tried to correct the individual measured ranges. This resulted in much less relevance being able to be placed on the range residuals and spatial de-correlation. The industry, in the form of a joint IMCA / OGP committee, produced a new guidance document in June 2011 for the quality control of GNSS positioning. ( Geomatics Guidance note 19 IMCA S 015 ) This document set out to define more up to date definitions of GNSS quality. C-Nav s interpretation of this guideline is described in the following paper. Statistics and More Statistics The answers arrived at by the IMCA/OGP group can be summarized as follows: DOP value is too coarse a figure of merit Statistical testing methods should be used Looking at instantaneous values is not good enough, trends must be examined. Statistics can be used to evaluate the quality of a position in terms of Precision Reliability Monitoring the Position Solution Verification of the statistical model used File Name: Introduction to C-Nav QC V1.3.docx Page 3 of 19

Quality can be defined as the size and nature of undetected errors. ( if we can detect them its assumed we will correct them ) Statistical testing is used to ensure the assumptions made in the quality assessments are correct. What Statistics, and Where Do They Come From When we talk about statistics by necessity we are talking about a collection of data. This has: A Population ( a number of samples ) A mean ( the average of the samples ) A Standard Deviation ( actually several ) which defines how closely the samples relate to the mean ( apologies to statistics purists ). Consider the following 3 distributions If these were samples of latitude position error, we would want the system with distribution c. Here more of the samples are closer to the mean ( in our latitude example the mean File Name: Introduction to C-Nav QC V1.3.docx Page 4 of 19

would hopefully be zero ) than in a or b and it has a smaller standard deviation (SD). It can be also be said that c has a smaller variance ( from the mean ) than a or b. Another Example. This shows two things, one that standard deviations come in several flavors: 1SD - one sigma = 68% 2SD two sigma = 96% 3SD three sigma = 99% of the samples is within this variance of the mean It also shows that we can put pass / fail limits on our statistics and test real observations against them. File Name: Introduction to C-Nav QC V1.3.docx Page 5 of 19

Where do our statistics come from? In general they come from models that represent the expected behavior of the observations and the resulting quality of the position. We don t know the actual errors in each fix, but we do know something about them. Phase errors are smaller than code errors Errors are larger when observing low elevation satellites Errors are larger when the signal to noise ratio is poor Some errors reduce over time due to the measurement process Combining this knowledge enables prediction of expected behavior. At any given fix :- We have a number of observations ( ranges to the satellite ) We have corresponding models of these observations We have geometry, determined by the relative position of the satellite and the vessel There are also time relationships which can be built into the models. Errors in the observations change fast and are ( essentially ) random. The corresponding models vary more slowly. Carrier phase acquisition tens of seconds Geometry changes tens of minutes SV entering or leaving user visibility sudden change Thus we can test each observation, and the fit result itself against models of expected results. File Name: Introduction to C-Nav QC V1.3.docx Page 6 of 19

Precision Precision is the quality of the fix with respect to random errors, the mean of which tend to zero. It is based on the fit model, varies slowly and is not influenced by individual observations. This calculation is based on the Lat/Lon standard deviation and the error ellipse major and minor axis. Both the size and ratio of the axis must be considered. A ratio of axis length > 2:1 is considered bad The value of horizontal standard deviation considered bad will depend on the expected correction method rather than the actual, this ensures that a warning is generated if the correction method changes unexpectedly. Reliability Reliability is a measure of how robust the solution is to outliers, where an outlier can be caused by multipath or interference. If a solution lacks reliability (eg through low redundancy and/or poor geometry) then a single outlier can easily skew the computed position significantly, pushing up all the residuals, without giving any clear clue as to which is the offending observation, thus making it more difficult to edit out. If we have a solution with a large number of observations and a good overall geometry, then a single outlier will (a) be easier to detect (and eliminate) and will also (b) have a smaller impact if it is not eliminated. These two properties of the solution can be measured separately as "internal" and "external" reliability. In our statistical testing we would ideally take a single observation ( range to a satellite ) and compare it against the expected distribution. In reality, we do not actually know what our expected observation is. The best we can do is to say that once we have solved for position, our expected observation, is the predicted observation at that position. So in reality we are testing residuals of the solution against an expected zero residual. Internal Reliability This is the marginally detectable bias in each observation ( MDEob ). External Reliability This is the result such a bias would have on the on the final position, usually expressed as MDElat, MDElon and MDEheight. Generally we are interested in the radial value ( MDEr). While we get a MDEr value for each satellite overall reliability is taken to be the worst case Max MDEr File Name: Introduction to C-Nav QC V1.3.docx Page 7 of 19

Monitoring the Position Solution - RAIM Receiver Autonomous Integrity Monitoring ( RAIM ) is a process which uses redundant observations to detect and remove erroneous measurements. Traditional RAIM calculations can be looked at as doing the position calculation a number of times excluding a different measurement each time. At least 7 satellites must be tracked to do a RAIM calculation at all. The more satellites edited out by the RAIM process the higher the likelihood of multipath or interference. The statistical test used in the RAIM process is the W-Test The W-Test uses the normalized residual from the least squares process, one for each satellite. The mean should be 0. The test fails if any W statistic is over 1.95. What we actually monitor is the ability to have enough data to perform a W test not the pass or fail of each individual satellite residual. Verifying the Model This is done using the F-Test. This is the average value of the unit variance over a period of time. It is a measure of how good the internal model of errors is. Over time the average value should be one. Low values, even zero are possible, which indicates a very good fix, occasional high individual values are also acceptable. Average values tending over 1 require investigation and / or rejection. Getting the Data In the good old days all we needed was contained in the standard NMEA GGA string, but to perform the checks required of these new quality parameters we need to look at many more NMEA data sentences. Precision - Reliability - RAIM - F-Test - W-Test - $PNCTGST $PNCTMDE $GNGBS $PNCTGST $GNGBS This increases the bandwidth required of the receiver port driving the monitor software. File Name: Introduction to C-Nav QC V1.3.docx Page 8 of 19

Rocket Science? Yes, it is, the original IMCA/OGP paper has pages and pages of math. A degree in statistics helps a bit also. Now the really hard part, how to present all this information to a busy DPO in a manner that is easy to interpret, but compliant with the IMCA/OGP guidelines. What we don t want is an excel spreadsheet display that is hard to read and even harder to interpret. Ideally we want a display that can be interpreted from across the room, but includes the essential historical view of the data. C-Nav Solution The C-Nav solution is implemented in both the C-NavigatorIII ( Linux based panel PC ) and the C-Monitor Windows software. The main features are : Parameter values shown on color coded background. Red/Yellow/Green Parameter values normalized to a 0 ( bad ) to 5 ( good ) range A master system health indicator based on all other parameters A 30 minute time slice history bar Display of last 30 minutes trend of any two parameter groups Setup of 0-5 scaling QA/QC Panel General Description The illustration opposite shows the general arrangement of the QA/QC panel. Numeric Indicators It has 5 numerical indicators, one larger than the other. The larger button indicated which parameter is depicted on the vertical time slice. The 5 numerical indicators are: File Name: Introduction to C-Nav QC V1.3.docx Page 9 of 19

System Health Reliability Precision RAIM Unit Variance Each of the above has a colored background corresponding to the numerical value. 5, 4 Green 3, 2 Yellow 1, 0 Red If an indicator changes to a worse state, it will flash for five minutes to alert the user to the change. The indicators are also clickable. Click on any one of the five to move it to the top and show its time slice history. Pass / Fail Indicators The F-Test and the W-Test are pass or fail; indicated by a red or green indicator. The F- test result is generated by the receiver in its $PNCTGST output. The W-Test fails if there is a loss of the RAIM function. If there is a GSB message for the fix it s a pass. Note that we are monitoring the operation of the W-Test not its failure which results in the editing of a satellite from the solution. An epoch with a GSB message with a pass code can be regarded as a place holder, ie the w-test was done but no satellites were rejected. Time Slice Bar The time slice bar shows a history of the selected parameter over the last 30 minutes with a bias on the granularity towards the latest data. That is the further back in time you go the longer the time slice. Thus the most recent time slice is one minute while the oldest is ten minutes. The worst reading in the time slice interval determines the color of the segment. Time Series Button If a more detailed look at a 30 minute period is required, then pressing the time series button will show a detailed trend plot in the main screen area. File Name: Introduction to C-Nav QC V1.3.docx Page 10 of 19

Examples The above is an example of the precision plot towards the end of convergence with C-Nav corrections applied. The noise estimates of the fix have not yet reach their optimum value, thus the precision value is a 4 not a 5. Notice how the system health mirrors the lowest of the four parameter values. File Name: Introduction to C-Nav QC V1.3.docx Page 11 of 19

The precision plot above shows good public access SBAS correction data. Notice that the values are larger than the previous graph showing C-Nav correctors, but that the red and yellow plots are approximately equal indicating reasonable fix geometry. The numeric value is a 4 even though the actual value is much higher than the previous example. This is due to the system rescaling the 0 5 bands in recognition of the fact that the corrections are now less accurate. i.e. what is good performance in SBAS is not good performance with C-Nav corrections. File Name: Introduction to C-Nav QC V1.3.docx Page 12 of 19

The above reliability plot was generated without any GNSS corrections applied. As can be seen the estimated value of undetected error is large. Typically the spikes on this plot are caused by satellites close to the elevation mask included in the solution. The mean value is indicative of the correction method used. File Name: Introduction to C-Nav QC V1.3.docx Page 13 of 19

This is a fairly typical RAIM plot. One of the prime reasons for a satellite to be deleted by the RAIM algorithms is multipath. File Name: Introduction to C-Nav QC V1.3.docx Page 14 of 19

This is a typical unit variance plot with all values less than unity. The green F-test line is a constant 1 indicating pass. This reflects a conservative model and gives confidence in the statistics. Local severe weather can bump the graph over one but is not a concern unless it stays over one for a prolonged period > 10 minutes. File Name: Introduction to C-Nav QC V1.3.docx Page 15 of 19

Interpreting the Parameters So much for the theory, what does it mean in practice, what should we expect. Lets look at each parameter in turn. Precision We expect this parameter to change relatively slowly. Small steps will result as satellites leave or join the visible constellation. Alarm states are typically caused by the system changing navigation modes because of interference or loss of primary correction source. Periods of poor geometry will typically drop the value to a four as the major to minor axis ratio affects the overall state. Reliability If a solution has high internal reliability, then even quite small outliers can be detected. This will show up as low values of MDE-obs, implying that small outliers can be detected and eliminated by the RAIM algorithm. Poor internal reliability usually means we have either low numbers of observations or we have observations that we haven't tracked long enough to build up noise estimates. External reliability measure the impact on the position of an undetected outlier. If we have poor geometry in our solution, then any outlier that we don't reject could have a disproportionate impact on our computed position. This shows up as a high value of largest positional MDE, implying that a single outlier at the margin of detectability would have a big impact on computed position. The reliability value is derived from the geometry and noise estimates. Sudden changes are not an indication that anything is wrong with the position, but more an indication that a couple of good ( in terms of geometry and noise estimates ) satellites have left the solution. This in turn makes it harder to eliminate outliers resulting is a less reliable solution. It is not an indication that such outliers actually exist. RAIM Unlike precision and reliability, the RAIM calculation is done on the measured data each epoch phase and code, and on the phase measurements on fast nav. Any observation failing the RAIM criteria is removed from the position calculation. This is done using a statistical test termed the W-Test The RAIM tests are performed after any other tests used to exclude observations such as low elevation. The classical RAIM computes the w-statistic on all observations and removes the highest failing observation ( 1.96 from the 0 central value of the normalized residual ). It then re- File Name: Introduction to C-Nav QC V1.3.docx Page 16 of 19

computes the solution and repeats the test until all included observations pass. C-Nav uses a more computational efficient method to obtain the same results. The RAIM result has 3 states 1. RAIM test complete no edited observations 2. RAIM test complete 1 or more observations deleted 3. RAIM test not possible 1) is the normal case where we have enough observations but none were edited. 2) can occur from time to time due to noise or multipath. 1 or 2 edited observations are acceptable as long as the reliability is good. 3 or more is cause for concern. Multipath is one reason for editing, low elevation is another, or high variance ( noise ) caused by an interference source. 3) is the worst case where we do not have enough observations to do a RAIM calculation at all. Unit Variance Unit variance will be 1 if the weights assigned to the observations model the actual noise on the measurements. In practice we are pessimistic and assume much noisier measurement than are usually the case. Thus our typical unit variance value will vary around 0.2 0.3. There are 3 cases to watch for. 1. A constantly low value around 0.1 2. A constantly high value over 5.0 3. A straight line value. In practice this is really a check that the software is behaving correctly. High peak values suggest you are working in difficult conditions ( reflections, interference, or difficult propagation conditions ) W-Test As stated earlier what the W-Test indicates is not an individual observation failing the W- Test, but the failure to be able to do a W-Test at all. This would occur in severe blocking conditions with a low number of observations. F- Test We know that on average the unit variance should be 1. We know what the statistical distribution should be and can thus test each the unit variance at each fix to determine if it is File Name: Introduction to C-Nav QC V1.3.docx Page 17 of 19

within the expected distribution. In practice it takes a very, very bad fix to be outside the normal distribution. It is rare for a fix to fail the F-Test and the underlying cause would be apparent in all the other test parameters. Is There a Downside? Yes, the displays require a lot of serial NMEA 0183 data each second. The amount of data increases with the number of satellites tracked. This requires a higher baud rate from the C- Nav3050 receiver than with previous software versions. To send the NMEA0183 data alone will take 19.2kb add to this the binary and control messages requires a minimum 57kb data rate. This means shorter RS 232 cables or a change to RS 422. Conclusions So what can the user conclude from this new display. If its all green he is in good shape as long as the correction system used is what is expected. That all green may not be good if the correction system should be C-Nav and its now SBAS. From the time slice it should be clear if a problem is getting better or worse. This can be seen when the main screen is displaying non-qc parameters. When something goes wrong warning is given well before HDOP or number of satellites gets ugly. File Name: Introduction to C-Nav QC V1.3.docx Page 18 of 19

Acknowledgement I have leaned heavily on the knowledge, work and guidance of Dr Tony Hedge in the formulation of this white paper, and thank him for his tireless assistance. References C&C Technologies ( 2012) QC for Monitoring C-Nav GNSS Quality Data C&C Technology ( 2011 ) Interpreting C-Nav GNSS Quality Data NavCom Technologies Inc ( 2011 ) Sapphire Technical Reference Manual Revision F OGP/IMCA ( 2011 ) Guidelines for GNSS Positioning in the Oil and Gas Industry File Name: Introduction to C-Nav QC V1.3.docx Page 19 of 19