Machine Learning and RF Spectrum Intelligence Gathering

Similar documents
Noise Figure: What is it and why does it matter?

Advanced geolocation capabilities

IBM SPSS Neural Networks

Using Iterative Automation in Utility Analytics

Artificial Neural Networks. Artificial Intelligence Santa Clara, 2016

Figure 1. Artificial Neural Network structure. B. Spiking Neural Networks Spiking Neural networks (SNNs) fall into the third generation of neural netw

Generating an appropriate sound for a video using WaveNet.

MINE 432 Industrial Automation and Robotics

Demystifying Machine Learning

TEMPORAL DIFFERENCE LEARNING IN CHINESE CHESS

DEEP LEARNING ON RF DATA. Adam Thompson Senior Solutions Architect March 29, 2018

Radio Deep Learning Efforts Showcase Presentation

On the Use of Convolutional Neural Networks for Specific Emitter Identification

Modulation Classification of Satellite Communication Signals Using Cumulants and Neural Networks

Fully Automated Network- Centric Spectrum Analysis and Signal Intelligence System

TRUSTING THE MIND OF A MACHINE

Proposers Day Workshop

TSTE17 System Design, CDIO. General project hints. Behavioral Model. General project hints, cont. Lecture 5. Required documents Modulation, cont.

Introduction to Machine Learning

Analogy Engine. November Jay Ulfelder. Mark Pipes. Quantitative Geo-Analyst

MSc(CompSc) List of courses offered in

The Art of Neural Nets

Surveillance and Calibration Verification Using Autoassociative Neural Networks

SEPTEMBER VOL. 38, NO. 9 ELECTRONIC DEFENSE SIMULTANEOUS SIGNAL ERRORS IN WIDEBAND IFM RECEIVERS WIDE, WIDER, WIDEST SYNTHETIC APERTURE ANTENNAS

11/13/18. Introduction to RNNs for NLP. About Me. Overview SHANG GAO

Digital Modulation Recognition Based on Feature, Spectrum and Phase Analysis and its Testing with Disturbed Signals

Detection of Targets in Noise and Pulse Compression Techniques

Localization (Position Estimation) Problem in WSN

Improved Detection by Peak Shape Recognition Using Artificial Neural Networks

Deep Learning. Dr. Johan Hagelbäck.

model 902 H-SLIC HF Wideband Signal Location Intercept and Collection System 902 H-SLIC

Image Manipulation Detection using Convolutional Neural Network

Multi-Way Diversity Reception for Digital Microwave Systems

AN IMPROVED NEURAL NETWORK-BASED DECODER SCHEME FOR SYSTEMATIC CONVOLUTIONAL CODE. A Thesis by. Andrew J. Zerngast

SMARTPHONE SENSOR BASED GESTURE RECOGNITION LIBRARY

ULTRASONIC SIGNAL PROCESSING TOOLBOX User Manual v1.0

1 Introduction. w k x k (1.1)

Game Mechanics Minesweeper is a game in which the player must correctly deduce the positions of

Comparison of Google Image Search and ResNet Image Classification Using Image Similarity Metrics

Lab/Project Error Control Coding using LDPC Codes and HARQ

CHAPTER 8: EXTENDED TETRACHORD CLASSIFICATION

Tiny ImageNet Challenge Investigating the Scaling of Inception Layers for Reduced Scale Classification Problems

FACE RECOGNITION USING NEURAL NETWORKS

Transactions on Information and Communications Technologies vol 1, 1993 WIT Press, ISSN

Technologies Worth Watching. Case Study: Investigating Innovation Leader s

An E911 Location Method using Arbitrary Transmission Signals

Characterization of LF and LMA signal of Wire Rope Tester

ES 492: SCIENCE IN THE MOVIES

A JOINT MODULATION IDENTIFICATION AND FREQUENCY OFFSET CORRECTION ALGORITHM FOR QAM SYSTEMS

RESEARCH ON METHODS FOR ANALYZING AND PROCESSING SIGNALS USED BY INTERCEPTION SYSTEMS WITH SPECIAL APPLICATIONS

Advanced Techniques for Mobile Robotics Location-Based Activity Recognition

AUTOMATIC MODULATION RECOGNITION OF COMMUNICATION SIGNALS

How Explainability is Driving the Future of Artificial Intelligence. A Kyndi White Paper

Prediction of Missing PMU Measurement using Artificial Neural Network

Institute for Critical Technology and Applied Science. Machine Learning for Radar State Determination. Status report 2017/11/09

Research on Hand Gesture Recognition Using Convolutional Neural Network

CHAPTER 6 BACK PROPAGATED ARTIFICIAL NEURAL NETWORK TRAINED ARHF

Applications of Music Processing

model 802C HF Wideband Direction Finding System 802C

ARTIFICIAL INTELLIGENCE IN POWER SYSTEMS

Research on Application of Conjoint Neural Networks in Vehicle License Plate Recognition

Local and Low-Cost White Space Detection

MATLAB/GUI Simulation Tool for Power System Fault Analysis with Neural Network Fault Classifier

A Solution for Identification of Bird s Nests on Transmission Lines with UAV Patrol. Qinghua Wang

Artificial Neural Networks

An Introduction to Artificial Intelligence, Machine Learning, and Neural networks. Carola F. Berger

Automated Planetary Terrain Mapping of Mars Using Image Pattern Recognition

BI TRENDS FOR Data De-silofication: The Secret to Success in the Analytics Economy

CHAPTER 4 MONITORING OF POWER SYSTEM VOLTAGE STABILITY THROUGH ARTIFICIAL NEURAL NETWORK TECHNIQUE

Number Plate Detection with a Multi-Convolutional Neural Network Approach with Optical Character Recognition for Mobile Devices

Introduction to Machine Learning

Comments of Shared Spectrum Company

Deep Neural Network Architectures for Modulation Classification

Decriminition between Magnetising Inrush from Interturn Fault Current in Transformer: Hilbert Transform Approach

Real-Time Face Detection and Tracking for High Resolution Smart Camera System

Interference Direction Analysis. Communication Signals

Glossary of terms. Short explanation

SIECAMS. Siemens Space. SIECAMS Siemens Satellite Monitoring System. Siemens AG Austria All rights reserved.

Using of Artificial Neural Networks to Recognize the Noisy Accidents Patterns of Nuclear Research Reactors

Artificial intelligence, made simple. Written by: Dale Benton Produced by: Danielle Harris

Convolutional Neural Networks: Real Time Emotion Recognition

Tadeusz Stepinski and Bengt Vagnhammar, Uppsala University, Signals and Systems, Box 528, SE Uppsala, Sweden

Statistical Analysis of Modern Communication Signals

Michael F. Toner, et. al.. "Distortion Measurement." Copyright 2000 CRC Press LLC. <

Robust Hand Gesture Recognition for Robotic Hand Control

Image Extraction using Image Mining Technique

Sonia Sharma ECE Department, University Institute of Engineering and Technology, MDU, Rohtak, India. Fig.1.Neuron and its connection

Visvesvaraya Technological University, Belagavi

An Hybrid MLP-SVM Handwritten Digit Recognizer

Characterizing High-Speed Oscilloscope Distortion A comparison of Agilent and Tektronix high-speed, real-time oscilloscopes

DESIGN AND IMPLEMENTATION OF AN ALGORITHM FOR MODULATION IDENTIFICATION OF ANALOG AND DIGITAL SIGNALS

Acoustic Emission Source Location Based on Signal Features. Blahacek, M., Chlada, M. and Prevorovsky, Z.

Intelligent Non-Player Character with Deep Learning. Intelligent Non-Player Character with Deep Learning 1

CHAPTER 4 LINK ADAPTATION USING NEURAL NETWORK

Artificial Intelligence in Law: Facts, Futures & Risks

Chapter 2 Channel Equalization

LabVIEW based Intelligent Frontal & Non- Frontal Face Recognition System

Harmonic detection by using different artificial neural network topologies

CHAPTER 1 INTRODUCTION

Nonuniform multi level crossing for signal reconstruction

Transcription:

A CRFS White Paper December 2017 Machine Learning and RF Spectrum Intelligence Gathering Dr. Michael Knott Research Engineer CRFS Ltd.

Contents Introduction 3 Guiding principles 3 Machine learning for automated signal classification 4 A traditional technique: decision trees... 4 Machine learning applied to the same problem... 4 Pair plots... 5 Training: it s all about the database... 6 Test results: it s all about the training... 6 What does it look like in real life?... 8 Why use machine learning for signal classification?... 10 Machine learning for signal detection 10 Machine learning for spectrum monitoring 12 Efficiency of ML processing 12 Looking to the future: feature learning 12 Conclusion 13 Why CRFS? 13 Legal Information 14 Disclaimer 14 Contact Information 14 2 CRFS

Introduction Many applications that are central to RF spectrum intelligence gathering require some sort of pattern recognition. For example, to classify a signal by type we need to identify the particular pattern associated with the modulation, while to recognize that there is an interesting signal present in received data, we need to distinguish between pattern and noise. It is desirable to automate these applications as far as possible, for efficiency and to avoid errors caused by fatigue. However, pattern recognition is a task that human brains are good at, while computers, in their traditional role of applying predetermined algorithms, are not. This has made automation of the process challenging. Machine learning (ML), which is a mature set of mathematical techniques that allow computers to learn from data using constructs such as artificial neural networks, is better suited to the problem. The past ten years have seen a large amount of investment in ML from companies such as Google and Baidu. One estimate is that between $26 billion and $39 billion was invested in artificial intelligence in 2016, of which the bulk involved ML 1. This has been stimulated by the realization that ML can help solve challenging problems in areas such as search optimization, advertisement placement, face recognition and driverless car technology. New tool sets for ML (for example, Google s TensorFlow) have been developed and made publicly available under open source licensing policies. Research using these tool sets has advanced the fields of machine learning and neural networks. This white paper describes how CRFS is solving customer problems more effectively, by using ML to automate pattern recognition processes as part of a suite of EW tools. We start by listing some guiding principles that need to apply, in order for an implementation of ML to be most practically useful for the purpose. We then turn to three tasks in spectrum intelligence gathering that can be approached with ML. For each case, we describe the technique used and present some results. The tasks are: (1) signal classification, for which we use a simple feedforward neural network; (2) signal detection, using a convolutional neural network to look for patterns in the data; (3) spectrum monitoring, using convolutional and unsupervised neural networks for anomaly detection. Artificial neural networks (ANNs) are networks of virtual neurons, inspired by the structure of the human brain. The output of each neuron depends on the weighted inputs it receives from other neurons. The arrangement of the network is fixed, but the behavior of the neurons is governed by numbers, which are modified as the network learns. This paper is not intended to be an introduction to ANNs or to machine learning in general, but some details are mentioned in passing. We also discuss how the available processing power affects what can be achieved by ML in spectrum intelligence gathering. We finish with a look at the future, and at some of the ways in which this application of ML could be developed further, to make spectrum intelligence gathering more reliable and more truly automated. Guiding principles In order to produce a deployable ML solution for signal classification and detection, we follow some guiding principles. Applied neural networks should be able to accept either live or archived data. This allows the solution to be used to reduce the number of false alarms or missed events by a human operator looking at live systems in operation or, alternatively, to search through terabytes of archived data. Computational requirements should be modest, so that the solution can run on embedded distributed sensor and transmitter platforms: analyzing data at source reduces data connectivity requirements and enables more rapid notification of alarms. It should be possible to update a trained neural network by training it further using a new database of signals. 1 Artificial Intelligence: The Next Digital Frontier? (McKinsey Global Institute, 2017) Machine Learning and RF Intelligence Gathering 3

Machine learning for automated signal classification Automated signal classification is the first problem to which we will apply machine learning in this paper. The system is given an unknown signal, and the requirement is to determine its modulation type and protocol. For example, given a TETRA signal it should be able to work out that the modulation is π/4 DQPSK and the protocol is TETRA. A traditional technique: decision trees One traditional means of applying algorithmic methods to automated signal classification is the decision tree (DT). Here, multiple numerical features are extracted from a captured signal. Features can range from the simple (such as an average) to the much more complicated (such as the results of a Fourier analysis). Decisions are made consecutively using each of these features, in a predetermined order that depends on earlier decisions, resulting in a branching tree structure (Figure 1 (a)). Thresholds for the decisions are varied manually to optimize the decision process, using a database of signals of known type. The DT decision process is fundamentally linear, so the technique will struggle with some feature sets. (a) (b) Do test X= Feature N IQ captured; feature tests performed at different parts of the tree s1 w1 w2 s2 wn sn +b f(x) R = f(x) IQ captured; features calculated and passed to net x = (s1*w1) + (s2*w2) + b Test > X Test <= X f(x) = (ReLU function) Figure 1. Nodes for feature-based signal classification: (a) Node of decision tree method; (b) Node of artificial neural net, showing weighted inputs and ReLU activation function Machine learning applied to the same problem Machine learning (ML) can also use features extracted from the captured signal. We will show the results of applying features to an artificial neural network (ANN) in order to classify signals. We use a feedforward neural network, which is a network in which the outputs from each layer of neurons are connected only to the inputs of succeeding layers, with no feedback during the decision process. Features are applied to the input layer, there is at least one hidden layer, and an output layer feeds into a voting function (SoftMax) which estimates the probability of each signal type (Figure 2). If one signal type is much more probable than all the others, then the result is definite and easy to interpret. If a number of different signal types are about equally probable, the ANN will correctly return an uncertain answer, instead of imparting a false certainty. This is in contrast with DT, which forces a definite answer regardless of how ambiguous the signal might be. Individual neurons produce an output that is calculated by feeding the weighted sum of their inputs into a nonlinear rectified linear unit (ReLU) activation function (Figure 1 (b)); this nonlinearity can allow the ANN to perform well with feature sets that DT struggles with. We train the network on a database of signals of known type. Training follows the backpropagation method, in which the weight of a connection is modified, at each training step, by an amount proportional to the gradient of the output error with respect to the weight. Gradients are calculated first for the output layer and then propagated back to the hidden layers. The features that are used by this application of ML are chosen by engineers, just as with DT, rather than being discovered by the ML itself. Features are selected to be effective at distinguishing between different signal types. 4 CRFS

Signal training Extract features from data series Feature Input layer: N feature neurons Hidden layers (the deep part) As many layers and as deep as the processing requires Output layer: K class neurons Input data from receiver IQ Feature Feature Soft Max Signal Signal Signal A B C Feature Neural net with weights and biases to optimize class decision based on feature set Signal K Probability output for all class types from SoftMax function Pair plots Figure 2. Feedforward neural network for feature-based signal classification To illuminate some of the decision making in both DT and ML, we can plot individual signals according to the values that they exhibit of pairs of features; the pair plots let us see how different types of signals can form clusters and how they can be distinguished from one another using different combinations of features. Figure 3 shows some idealized pair plots. These are illustrative and were not derived from real data. The dividing lines between signal types are also idealized and simplified, but they serve to illustrate the difference between linear and nonlinear fitting. In Figure 3 (b), the two features are very good for separating the signal types; in this case DT, with a linear decision fit, would perform as well as ML. Plot (c) shows a situation where the two features are unsuitable for distinguishing the signal types because they contain no information that is useful for the classification process. Plot (a) shows a case where ML would outperform DT because it allows nonlinear decision fitting to the data. (a) (b) (c) Feature 1 ML Feature 1 Feature 2 DT Signal type 1 Signal type 2 Each dot represents a capture of that signal type Feature 2 Figure 3. Idealized pair plots Feature 3 Feature 3 A problem that can occur with ML (and also with DT) is overfitting. This is when the ML learns the characteristics of a specific set of training data, rather than the characteristics of the type that we intend it to learn, so that it does not successfully classify other data sets. In a pair plot such as Figure 3 (a), this might manifest itself as a convoluted dividing line that encompasses all the individual points of signal type 1, but is not a reasonable outline for the cluster. We avoid overfitting by using pre-developed tools from TensorFlow. Machine Learning and RF Intelligence Gathering 5

Figure 4 shows some real examples of pair plots for a larger range of features and signals. For the classification of 10 signal types, 19 features are extracted from the data set. The pair plots are much more complicated than the idealized examples of Figure 3. ML fits all of these automatically, whereas the DT method would require manual fitting. Feature 1 Feature 2 Feature 3 Each color represents a different signal class. Diagonals where feature N maps to feature N show histogram of feature N weighting for each signal type Figure 4. Real pair plots Training: it s all about the database The signal database is crucial to the performance of ML (as it is also for DT). The decision process must be optimized for the signal types in general, not for the signal types under some limited set of conditions. This means that the database should contain signals received under a range of different conditions, representative of the conditions under which the trained network is expected to function. To protect against unwittingly training a network that is unsatisfactory because it is poorly fitted, or overfitted, to the training data, the data is divided into a training set and a test set. The ANN is trained using only the training set, then tested using the test set to make sure it is effective with that data too. Test results: it s all about the training A useful way to evaluate the performance of an ANN is to enter the test results into a confusion matrix (Figure 5). This is a matrix showing the frequency with which a signal of each type is identified as belonging to each of the possible types. The more of the identifications that lie on the diagonal of the matrix, the better the performance of the network. Confusion matrices allow us to see how reliable the classifier is and, by showing the mistakes in detail, they also help us to decide what new features might make it more reliable. 6 CRFS

Actual Signal Type Actual Signal Type Actual Signal Type Actual Signal Type (a) Training with ideal data (b) Training with frequency offsets Predicted Signal Type (c) Training with low SNR Predicted Signal Type (d) Training with low SNR and offsets Predicted Signal Type Predicted Signal Type Figure 5. Confusion matrices for different data sets Table 1. Overall results for different data sets Case Training score Test score Comment Clean signals 99.4% 99.4% Ideal signal, high SNR Frequency offset 98.1% 97.9% Frequency offset, high SNR Low SNR 97.5% 97.6% Low SNR Offset, low SNR 95.8% 95.7% Combined Table 1 shows the overall proportion of correct results in the data that was used to produce the confusion matrices in Figure 5. The method was tested on four different sets of data, with different noise and frequency offset characteristics. Signal classification was successful for all four data sets but, not surprisingly, it worked best with clean signals. The results change significantly if we train the ANN on data from one set, then test it on data from another set. Figure 6 contains examples of the confusion matrices that we get if we do this. Matrix (a) shows the results of testing ideal data in an ANN that has been trained only on frequency offset data. This works fairly well, but the results are worse than with an ANN trained on the correct data set. Matrix (b) shows the effect of doing things the other way around, using frequency offset test data in an ANN trained on ideal data. The results here are poor. The net confuses most signals with CW, which makes sense because a frequency offset will manifest itself Machine Learning and RF Intelligence Gathering 7

Actual Signal Type Actual Signal Type as a constant rate of phase change with time. The exception to this is that the net exhibits confusion between the multicarrier OFDM modulations of DAB and DVB. These results reinforce the point that training sets need to include data obtained under a representative range of conditions. (a) 90.9% Test Score (b) 37.8% Test Score Predicted Signal Type Neural net trained on frequency offset data works well with ideal test data. Predicted Signal Type Frequency offset test data fed into a neural net trained on ideal data shows that the net confuses most signals with CW. Net tends to confuse DAB and DVB. What does it look like in real life? Figure 6. Confusion matrices for ANNs trained on the wrong data An ANN, trained for signal classification and running on an embedded sensor, was applied to live signals. Figure 7 and Figure 8 show the results achieved with an FM audio signal and a PSK4 signal. The FM signal was clearly identified as FM (95% probability), while the PSK4 was also identified correctly, but the classifier reported a smaller, but significant, probability that it might be PSK2. A very convenient way of reporting signal classification results to the user is to display them as color coded bars on a plot of the spectrum. Figure 9 shows a region of the spectrum around 1.01 GHz, displayed in this way, with 64 QAM, 4-PSK and 16 QAM frequency hopping signals clearly identified and highlighted. 8 CRFS

Analysis region Spectral plot Waterfall plot Bar graph of probability by modulation type Figure 7. Classifier results for FM signal Analysis region Spectral plot Waterfall plot Bar graph of probability by modulation type Figure 8. Classifier results for PSK4 signal Machine Learning and RF Intelligence Gathering 9

Regions highlighted with neural net results Potentially related signals tagged in same color region and linked with probability label Figure 9. Signal classification results displayed on a spectrum plot Why use machine learning for signal classification? Signal classification by ML has a number of advantages over decision trees. The automation in the method allows it to be optimized for a set of signal types, without heavy engineer input, and to be improved easily by training with additional data. It is possible for the method to be implemented in a form that can be trained by the customer, if they do not want to reveal the training data to the vendor (for example, for reasons of security). Another advantage of the ML implementation that we have outlined here is that, thanks to its nonlinearity, it may be able to fit the data more effectively than the linear DT method. The output from the ML also contains more helpful information, since it reports a probability for each signal type rather than just a single answer; this enables us to judge how confident we can be in the final answer. Machine learning for signal detection Another promising application of machine learning to RF spectrum intelligence extraction is in signal detection by means of pattern recognition. A 2D convolutional neural network (CNN) (Figure 10) is trained on a library of 2D spectral images (showing how the power depends on frequency and time, i.e., waterfall plot images), each of which contains known features that might be found in a signal but are unlikely to be seen in noise (Figure 11). For example, there might be a burst of power that is constrained within hard boundaries in time and frequency, or that has a constant bandwidth but changes frequency at a constant rate. This type of network was developed for image recognition and facial recognition. We are taking advantage of the fact that the system dynamics can be expressed as an image, so that the pattern recognition can follow the same principles as, for example, facial recognition. The features learned by the CNN are essentially different from the features used to train feedforward networks for signal classification; with the CNN, the features are elements of pattern that can be isolated in the power-frequency-time plot. While spectral characteristics can vary with frequency and time expansion or compression, the features will still be recognizable. The trained system will be able to detect which, if any, of the features are found in new live or archived data. This will provide an image signature for the IQ data, and the data can then be fed into a signal database for later signal classification analysis. Since the pattern recognizer can detect signals of types that it has not encountered before (provided that they display some of the signal-like features that it has learnt), it can act as a powerful data search engine for new signal types. 10 CRFS

Multiple levels of convolution and activation layers Layer 1 Convolution (Image * Filters 1) Activation Nonlinear function (ReLu) Layer 2 Convolution (Layer1 * Filters 2) Activation Nonlinear function (ReLu) Fully Connected Neural Net Result A Pooling (Max) Filters 1 Filters 2 Soft Max B C N 2D Spectral Image Library A B C N Figure 10. Convolutional neural network for pattern recognition Spectral waterfall being processed Signals tagged when they satisfy characteristics being searched for Convolutional Neural Net 2D Spectral Image Library used for training Figure 11. Convolutional neural network searching for features in power-frequency-time plot The task of recognizing signals has traditionally been performed by a human observer. The advantage of applying ML to this task is that, unlike a human observer, its performance will not deteriorate because of fatigue or distractions after long periods of observation. Therefore, ML may outperform a human in this role overall, even if it fails to do so in a single short test. A further application of pattern recognition using CNNs is in RF fingerprinting of devices. In this case, the detailed characteristics of the patterns need to be considered. As an example, we can consider the signal emitted by a cellphone, which ramps up from zero power to the required value over some short interval of time. There will be an ideal theoretical form for the increase in power, but real phones are likely to deviate from the ideal in the shape and duration of the increase. Taken together, the precise behavior of all the applicable patterns forms the RF fingerprint of the device, which may allow individual phones to be identified. Machine Learning and RF Intelligence Gathering 11

Machine learning for spectrum monitoring It would be very useful to have a system that could automatically monitor the spectrum 24/7, without continuous human involvement. In order to do this, it has to learn the normal behavior of the spectrum, so that it can detect and report events that are interesting or anomalous, and therefore worthy of special attention. To accomplish this, we use the properties of convolutional neural networks, and also of unsupervised neural networks (clustering) to extract high level features from massive amounts of spectral data. An ANN is trained on hundreds of hours of normal spectral operating conditions, and the trained network is then able to recognize and highlight events that are uncommon or completely new (Figure 12). Signal with uncommon features that have previously been built up over many hours of training are highlighted and the IQ time sequence isolated Figure 12. Example of spectrum monitoring using machine learning Efficiency of ML processing The analysis methods described in this white paper are applied to large data sets, so the efficiency of the processing is very important. Obviously, the ML analysis of spectrum data is useful only if it can be performed sufficiently fast: as a minimum, the real time analysis of live data needs to be able to keep up with the acquisition of data in real time. A typical receiver with real time bandwidth of 100 MHz generates 500 MB/s of IQ data; the signal processing discussed here, running on a modest Intel i7, can process data at 1000 MB/s. ML analysis scales well if we increase the number of processors or the memory. For archive processing, the ultimate speed limit will be governed by the non-volatile memory read access time for local storage. Looking to the future: feature learning As we explained above, under our current method of applying ML to signal classification, the ANN operates with a set of features extracted from the captured IQ data, rather than using the data itself. This means that most of the information present in the IQ data is discarded. The set of features is selected with the aim of including the information that is most relevant to signal classification, in order that the discarded data should be less important. Therefore, we choose features that have led to successful signal classification by, for example, the DT method; the ML is aided by the insights of the engineer who chooses the features. The results presented in 12 CRFS

this white paper show that signals can be classified successfully using this feature-based approach. However, an alternative method is to supply the IQ data itself as the input to the ML. Under this Feature Learning approach, the features will not be predetermined; they will be derived by the ML itself. The ML will no longer be helped by the insight of a human engineer, but it will also no longer be constrained by the limitations of that insight. Using the IQ data directly will mean that the input to the ANN is a time series of complex rather than real data, and the architecture must also be adapted: the initial feature-analysis layers will be convolutional neural nets with complex weights. This direct input mode can also be augmented by alternative signal representations - for example, wavelet or other transforms - which are also fed into convolutional nets. This is analogous to multi-sensor analysis but all the inputs are derived from the same set of IQ data. The different representations will all contain the same information but they will represent it in different ways, which may help the ML find features that can reliably distinguish between different types of signal. Conclusion We have successfully applied machine learning techniques to the automation of processes that are important for RF spectrum intelligence gathering. In particular, a simple feedforward neural network has been trained to classify signals according to their modulation and protocol. Furthermore, we have shown how a convolutional neural network can detect signals, and potentially also fingerprint individual devices, by treating its input as an image and performing pattern recognition using a method that is analogous to image recognition or face recognition. We have also seen how ML can be used to automate spectrum monitoring, which will be a major step forward in RF signal intelligence. All the applications that we have described have the potential for further improvement as we push towards the goal of reliable, automated spectrum monitoring and analysis. Why CRFS? CRFS is a leader in real-time RF spectrum monitoring solutions for regulators, defense, security agencies and spectrum operations. Applications include spectrum management and enforcement, remote site and perimeter monitoring, real-time situational awareness, threat detection and signals intelligence. The RFeye range includes best-in-class networkable receivers and nodes designed for remote, distributed, continuous 24/7 monitoring of the RF environment. Powerful data analytics and visualization tools provide actionable intelligence in complex and critical RF environments. Machine Learning and RF Intelligence Gathering 13

Legal Information Copyright 2017 CRFS Limited The copyright of this document is the property of CRFS Limited. Without the written consent of CRFS, given by contract or otherwise, this document must not be copied, reprinted or reproduced in any material form, either wholly or in part, and the contents of this document, or any methods or techniques available therefrom, must not be disclosed to any other person whatsoever. CRFS Limited reserves the right to make changes to the specifications of the products detailed in this document at any time without notice and obligation to notify any person of such changes. RFeye, CRFS and the CRFS logo are trademarks of CRFS Limited. All other trademarks are acknowledged and observed. Mention of third-party products does not constitute an endorsement or a recommendation. All figures, data and specifications contained in this document are typical and must be specifically confirmed in writing by CRFS Limited before they apply to any tender, order or contract. CRFS takes every precaution to ensure that all information contained in this publication is factually correct but accepts no liability for any error or omission. No freedom to use patents or other property rights is implied by this document. Disclaimer The information contained in this document is not intended to amount to advice on which reliance should be placed. Although we take reasonable steps to ensure the accuracy of the information provided, we provide it without any guarantees, conditions or warranties as to its accuracy or reliability. To the extent permitted by law, we expressly exclude all conditions, warranties and other terms which might otherwise be implied by law. This document is subject to change without notice. Contact Information CRFS Limited Building 7200 Cambridge Research Park Beach Drive Cambridge CB25 9TL UK Tel: +44 (0)1223 859500 Fax: +44 (0)1223 280351 Email: enquiries@crfs.com Web: www.crfs.com 14 CRFS