IBM SPSS Neural Networks

Similar documents
Lesson 08. Convolutional Neural Network. Ing. Marek Hrúz, Ph.D. Katedra Kybernetiky Fakulta aplikovaných věd Západočeská univerzita v Plzni.

CHAPTER 6 BACK PROPAGATED ARTIFICIAL NEURAL NETWORK TRAINED ARHF

Analysis of Learning Paradigms and Prediction Accuracy using Artificial Neural Network Models

Multiple-Layer Networks. and. Backpropagation Algorithms

Deep Neural Networks (2) Tanh & ReLU layers; Generalisation and Regularisation

Artificial Neural Networks. Artificial Intelligence Santa Clara, 2016

A COMPARISON OF ARTIFICIAL NEURAL NETWORKS AND OTHER STATISTICAL METHODS FOR ROTATING MACHINE

Using Iterative Automation in Utility Analytics

Smarter oil and gas exploration with IBM

Prediction of Cluster System Load Using Artificial Neural Networks

COMPARATIVE ANALYSIS OF ACCURACY ON MISSING DATA USING MLP AND RBF METHOD V.B. Kamble 1, S.N. Deshmukh 2 1

A Quantitative Comparison of Different MLP Activation Functions in Classification

Transient stability Assessment using Artificial Neural Network Considering Fault Location

A comparative study of different feature sets for recognition of handwritten Arabic numerals using a Multi Layer Perceptron

CHAPTER 4 LINK ADAPTATION USING NEURAL NETWORK

MINE 432 Industrial Automation and Robotics

Neural Network Classifier and Filtering for EEG Detection in Brain-Computer Interface Device

Automatic Speech Recognition (CS753)

NEURAL NETWORK BASED MAXIMUM POWER POINT TRACKING

Application of Multi Layer Perceptron (MLP) for Shower Size Prediction

Analogy Engine. November Jay Ulfelder. Mark Pipes. Quantitative Geo-Analyst

Using of Artificial Neural Networks to Recognize the Noisy Accidents Patterns of Nuclear Research Reactors

Research on Hand Gesture Recognition Using Convolutional Neural Network

Stacking Ensemble for auto ml

Contents. List of Figures List of Tables. Structure of the Book How to Use this Book Online Resources Acknowledgements

Confidently Assess Risk Using Public Records Data with Scalable Automated Linking Technology (SALT)

Comparison of MLP and RBF neural networks for Prediction of ECG Signals

Generating an appropriate sound for a video using WaveNet.

Approximation a One-Dimensional Functions by Using Multilayer Perceptron and Radial Basis Function Networks

Artificial neural networks in forecasting tourists flow, an intelligent technique to help the economic development of tourism in Albania.

Keywords : Simulated Neural Networks, Shelf Life, ANN, Elman, Self - Organizing. GJCST Classification : I.2

Research on Application of Conjoint Neural Networks in Vehicle License Plate Recognition

Stock Market Indices Prediction Using Time Series Analysis

MAGNT Research Report (ISSN ) Vol.6(1). PP , Controlling Cost and Time of Construction Projects Using Neural Network

Data Mining In the Prediction of Impacts of Ambient Air Quality Data Analysis in Urban and Industrial Area

Evolutionary Artificial Neural Networks For Medical Data Classification

POLITEHNICA UNIVERSITY TIMISOARA

INSIGHT ADVANCING. Lexis Advance. Find just what you re looking for faster with research innovations inspired by legal professionals like you.

Neural Model for Path Loss Prediction in Suburban Environment

Prediction of airblast loads in complex environments using artificial neural networks

CONSTRUCTION COST PREDICTION USING NEURAL NETWORKS

Transactions on Information and Communications Technologies vol 1, 1993 WIT Press, ISSN

Development and Comparison of Artificial Neural Network Techniques for Mobile Network Field Strength Prediction across the Jos- Plateau, Nigeria

Surveillance and Calibration Verification Using Autoassociative Neural Networks

Neural Filters: MLP VIS-A-VIS RBF Network

Introduction to Machine Learning

Forecasting Exchange Rates using Neural Neworks

Constant False Alarm Rate Detection of Radar Signals with Artificial Neural Networks

Application of selected artificial intelligence methods in terms of transport and intelligent transport systems

Dynamic Throttle Estimation by Machine Learning from Professionals

CLASSLESS ASSOCIATION USING NEURAL NETWORKS

Research Projects BSc 2013

Behavior Emergence in Autonomous Robot Control by Means of Feedforward and Recurrent Neural Networks

Supplementary Figures

Machine Learning and RF Spectrum Intelligence Gathering

NEURAL NETWORK DEMODULATOR FOR QUADRATURE AMPLITUDE MODULATION (QAM)

Enhanced MLP Input-Output Mapping for Degraded Pattern Recognition

Learning New Articulator Trajectories for a Speech Production Model using Artificial Neural Networks

Spectrum Hole Prediction And White Space Ranking For Cognitive Radio Network Using An Artificial Neural Network

Research Brief. Clinicians and life sciences companies working together: What types of relationships do clinicians find most appealing?

Application of Artificial Neural Networks System for Synthesis of Phased Cylindrical Arc Antenna Arrays

arxiv: v1 [cs.ce] 9 Jan 2018

INFORMATION about image authenticity can be used in

A Comparison of Particle Swarm Optimization and Gradient Descent in Training Wavelet Neural Network to Predict DGPS Corrections

Available online at ScienceDirect. Procedia Computer Science 85 (2016 )

MURDOCH RESEARCH REPOSITORY

Lake Level Prediction Using Artificial Neural Network with Adaptive Activation Function

Comparison of Various Neural Network Algorithms Used for Location Estimation in Wireless Communication

Introduction to Machine Learning

Adaptive Multi-layer Neural Network Receiver Architectures for Pattern Classification of Respective Wavelet Images

Stock Price Prediction Using Multilayer Perceptron Neural Network by Monitoring Frog Leaping Algorithm

Figure 1. Artificial Neural Network structure. B. Spiking Neural Networks Spiking Neural networks (SNNs) fall into the third generation of neural netw

ANALYSIS OF CITIES DATA USING PRINCIPAL COMPONENT INPUTS IN AN ARTIFICIAL NEURAL NETWORK

Artificial Intelligence Elman Backpropagation Computing Models for Predicting Shelf Life of. Processed Cheese

Practical Comparison of Results of Statistic Regression Analysis and Neural Network Regression Analysis

Application of Generalised Regression Neural Networks in Lossless Data Compression

Statistical Tests: More Complicated Discriminants

Application of Feed-forward Artificial Neural Networks to the Identification of Defective Analog Integrated Circuits

MSc(CompSc) List of courses offered in

DeepStack: Expert-Level AI in Heads-Up No-Limit Poker. Surya Prakash Chembrolu

Prediction of Missing PMU Measurement using Artificial Neural Network

AN ANALYSIS OF SPEECH RECOGNITION PERFORMANCE BASED UPON NETWORK LAYERS AND TRANSFER FUNCTIONS

Biologically Inspired Computation

Online Large Margin Semi-supervised Algorithm for Automatic Classification of Digital Modulations

Machine Learning. Classification, Discriminative learning. Marc Toussaint University of Stuttgart Summer 2014

Classification of Taste using a Neural Network: A Case Study in Mineral Water and Drinking Water Classification

Development and Integration of Artificial Intelligence Technologies for Innovation Acceleration

An Hybrid MLP-SVM Handwritten Digit Recognizer

Smart Manufacturing: A Big Data Perspective. Andrew Kusiak Intelligent Systems Laboratory The University of Iowa Iowa City, Iowa USA

Synergy Model of Artificial Intelligence and Augmented Reality in the Processes of Exploitation of Energy Systems

International Journal of Scientific & Engineering Research, Volume 4, Issue 12, December-2013 ISSN

Behaviour Patterns Evolution on Individual and Group Level. Stanislav Slušný, Roman Neruda, Petra Vidnerová. CIMMACS 07, December 14, Tenerife

Student: Nizar Cherkaoui. Advisor: Dr. Chia-Ling Tsai (Computer Science Dept.) Advisor: Dr. Eric Muller (Biology Dept.)

Deep Learning for Launching and Mitigating Wireless Jamming Attacks

A HYBRID ALGORITHM FOR FACE RECOGNITION USING PCA, LDA AND ANN

Application of Backpropagation Algorithms in Predicting the Quality of Component Based Software Systems

Number Plate Detection with a Multi-Convolutional Neural Network Approach with Optical Character Recognition for Mobile Devices

AUGMENTED CONVOLUTIONAL FEATURE MAPS FOR ROBUST CNN-BASED CAMERA MODEL IDENTIFICATION. Belhassen Bayar and Matthew C. Stamm

Neural Network based Digital Receiver for Radio Communications

TEMPORAL DIFFERENCE LEARNING IN CHINESE CHESS

Transcription:

IBM Software IBM SPSS Neural Networks 20 IBM SPSS Neural Networks New tools for building predictive models Highlights Explore subtle or hidden patterns in your data. Build better-performing models No programming needed Choice of MLP or RBF algorithms Your organization needs to find patterns and connections in the complex and fast-changing environment you work in so that you can make better decisions at every turn. You may be using IBM SPSS Statistics Base and one or more of its modules to help you do this. If so, you know the power and versatility you have at your fingertips. But there s even more you can do. You can explore subtle or hidden patterns in your data, using IBM SPSS Neural Networks. This module offers you the ability to discover more complex relationships in your data and generate better performing predictive models. The procedures in SPSS Neural Networks complement the more traditional statistics in SPSS Statistics Base and its modules. Find new associations in your data with SPSS Neural Networks and then confirm their significance with traditional statistical techniques. IBM SPSS Neural Networks is available for installation as client-only software but, for greater performance and scalability, a server-based version is also available. Why use a neural network? A computational neural network is a set of non-linear data modeling tools consisting of input and output layers plus one or two hidden layers. The connections between neurons in each layer have associated weights, which are iteratively adjusted by the training algorithm to minimize error and provide accurate predictions.

You set the conditions under which the network learns and can finely control the training stopping rules and network architecture, or let the procedure automatically choose the architecture for you. You can combine SPSS Neural Networks with other statistical procedures to gain clearer insight in a number of areas. In market research, for example, you can create customer profiles and discover customer preferences. In database marketing, you can segment your customer base and optimize marketing campaigns. In financial analysis, you can use SPSS Neural Networks to analyze applicants creditworthiness and to detect possible fraud. In operational analysis, use this new tool to manage cash flow and improve logistics planning. Scientific and healthcare applications include forecasting treatment costs, performing medical outcomes analysis, and predicting the length of a hospital stay. Control the process from start to finish With SPSS Neural Networks, you select either the Multilayer Perceptron (MLP) or Radial Basis Function (RBF) procedure. Both of these are supervised learning techniques that is, they map relationships implied by the data. Both use feedforward architectures, meaning that data moves in only one direction, from the input nodes through the hidden layer of nodes to the output nodes. In an MLP network like the one shown here, the data feeds forward from the input layer through one or more hidden layers to the output layer. From the Multilayer Perceptron (MLP) dialog, you select the variables that you want to include in your model. 2

Your choice of procedure will be influenced by the type of data you have and the level of complexity you seek to uncover. While the MLP procedure can find more complex relationships, the RBF procedure is generally faster. With either of these approaches, you divide your data into training, testing, and holdout sets. The training set is used to estimate the network parameters. The testing set is used to prevent overtraining. The holdout set is used to independently assess the final network, which is applied to the entire dataset and to any new data. You specify the dependent variables, which may be scale, categorical, or a combination of the two. If a dependent variable has scale measurement level, then the neural network predicts continuous values that approximate the true value of some continuous function of the input data. If a dependent variable is categorical, then the neural network is used to classify cases into the best category based on the input predictors. The results of exploring data with neural network techniques can be shown in a variety of graphic formats. This simple bar chart is one of many options. You adjust the procedure by choosing how to partition the dataset, what sort of architecture you want, and what computation resources will be applied to the analysis. Finally, you choose to display results in tables or graphs, save optional temporary variables to the active dataset, and export models in XML-file formats to score future data. Gain greater value with collaboration To share and efficiently distribute assets, protect them in ways that meet internal and external compliance requirements and publish results so that a greater number of business users can view and interact with them, consider augmenting IBM SPSS Statistics Developer with IBM SPSS Collaboration and Deployment services. More information about its valuable capabilities can be found at www.ibm.com/spss/devcentral. 3

Features Multilayer Perceptron (MLP) The MLP procedure fits a particular kind of neural network called a multilayer perceptron. The multilayer perceptron is a supervised method using feedforward architecture. It can have multiple hidden layers. One or more dependent variables may be specified, which may be scale, categorical, or a combination. If a dependent variable has scale measurement level, then the neural network predicts continuous values that approximate the true value of some continuous function of the input data. If a dependent variable is categorical, then the neural network is used to classify cases into the best category based on the input predictors. Predictors Factors Covariates The EXCEPT subcommand lists any variables that the MLP procedure should exclude from the factor or covariate lists on the command line. This subcommand is useful if the factor or covariate lists contain a large number of variables. The RESCALE subcommand is used to rescale covariates or scale dependent variables Dependent variable (if scale): standardized, normalized, adjusted normalized, or none Covariates: standardized, normalized, adjusted normalized, or none The PARTITION subcommand specifies the method of partitioning the active dataset into training, testing, and holdout samples. The training sample comprises the data records used to train the neural network. The testing sample is an independent set of data records used to track prediction error during training in order to prevent overtraining. The holdout sample is another independent set of data records used to assess the final neural network. You can specify: The relative number of cases in the active dataset to randomly assign to the training sample The relative number of cases in the active dataset to randomly assign to the testing sample The relative number of cases in the active dataset to randomly assign to the holdout sample A variable that assigns each case in the active dataset to the training, testing, or holdout sample The ARCHITECTURE subcommand is used to specify the neural network architecture. You can specify: Whether to use the automatic architecture or, if automatic is not used: The number of hidden layers in the neural network The activation function to use for all units in the hidden layers (Hyperbolic tangent or Sigmoid) The activation function to use for all units in the output layer (Identity, Hyperbolic tangent, Sigmoid, or Softmax 4

The CRITERIA subcommand specifies the computational and resource settings for the MLP procedure. You can specify the training type, which determines how the neural network processes training data records: batch training, online training, mini-batch training. You can also specify: The number of training records per mini-batch (if selected as the training method) The maximum number of cases to store in memory when automatic architecture selection and/or minibatch training is in effect The optimization algorithm used to determine the synaptic weights: Gradient descent, Scaled conjugate gradient The initial learning rate for the gradient descent optimization algorithm The lower boundary for the learning rate when gradient descent is used with online or mini-batch training The momentum rate for the gradient descent optimization algorithm The initial lambda, for the scaled conjugate gradient optimization algorithm The initial sigma, for the scaled conjugate gradient optimization algorithm The interval [a0 a, a0+a] in which weight vectors are randomly generated when simulated annealing is used The STOPPINGRULES subcommand specifies the rules that determine when to stop training the neural network. You can specify: The number of steps n to allow before checking for a decrease in prediction error Whether the training timer is turned on or off and the maximum training time The maximum number of epochs allowed The relative change in training error criterion The training error ratio criterion The MISSING subcommand is used to control whether user-missing values for categorical variables that is, factors and categorical dependent variables are treated as valid values The PRINT subcommand indicates the tabular output to display and can be sed to request a sensitivity analysis. You can choose to display: The case processing summary table Information about the neural network, including the dependent variables, number of input and output units, number of hidden layers and units, and activation functions A summary of the neural network results, including the average overall error, the stopping rule used to stop training and the training time A classification table for each categorical dependent variable The synaptic weights; that is, the coefficient estimates, from layer i 1unit j to layer i unit k A sensitivity analysis, which computes the importance of each predictor in determining the neural network 5

The PLOT subcommand indicates the chart output to display. You can display: Network diagram A predicted by observed value chart for each dependent variable A residual by predicted value chart for each scale dependent variable ROC (Receiver Operating Characteristic) curves for each categorical dependent variable. It also displays a table giving the area under each curve Cumulative gains charts for each categorical dependent variable Lift charts for each categorical dependent variable The SAVE subcommand writes optional temporary variables to the active dataset. You can save: Predicted value or category Predicted pseudo-probability Radial Basis Function (RBF) The RBF procedure fits a radial basis function neural network, which is a feedforward, supervised learning network with an input layer, a hidden layer called the radial basis function layer, and an output layer. The hidden layer transforms the input vectors into radial basis functions. Like the MLP procedure, the RBF procedure performs prediction and classification. The RBF procedure trains the network in two stages: 1. The procedure determines the radial basis functions using clustering methods. The center and width of each radial basis function are determined 2. The procedure estimates the synaptic weights given the radial basis functions. The sum-of-squares error function with identity activation function for the output layer is used for both prediction and classification. Ordinary Least Squares regression is used to minimize the sum-of-squares error The OUTFILE subcommand saves XML-format files containing the synaptic eights 6

Due to this two-stage training approach, the RBF network is in general trained much faster than MLP. Subcommands listed for the MLP procedure perform similar functions for the RBF procedure, with the following exceptions: When using the ARCHITECTURE subcommand, users can specify the Gaussian radial basis function used in the hidden layer: either Normalized RBF or Ordinary RBF When using the CRITERIA subcommand, users can specify the computation settings for the RBF procedures, specifying how much overlap occurs among the hidden units System requirements System requirements Requirements vary according to platform. For details, see www.ibm.com/spss/requirements. About IBM IBM software delivers actionable insights decision-makers need to achieve better business performance. IBM offers a comprehensive, unified portfolio of business intelligence, predictive and advanced analytics, financial performance and strategy management, governance, risk and compliance and analytic applications. With IBM software, companies can spot trends, patterns and anomalies, compare what if scenarios, predict potential threats and opportunities, identify and manage key business risks and plan, budget and forecast resources. With these deep analytic capabilities our customers around the world can better understand, anticipate and shape business outcomes. For more information For further information or to reach a representative please visit ibm.com/analytics. Request a call To request a call or to ask a question, go to ibm.com/ business-analytics/contactus. An IBM representative will respond to your inquiry within two business days. 7

Copyright IBM Corporation 2011 IBM Corporation Route 100 Somers, NY 10589 US Government Users Restricted Rights - Use, duplication of disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Produced in the United States of America June 2011 All Rights Reserved IBM, the IBM logo and ibm.com are trademarks of International Business Machines Corp., registered in many jurisdictions worldwide. Other product and service names might be trademarks of IBM or other companies. A current list of IBM trademarks is available on the Web at "Copyright and trademark information" at www.ibm.com/legal/copytrade.shtml. SPSS is a trademark of SPSS, Inc., an IBM Company, registered in many jurisdictions worldwide. Other company, product or service names may be trademarks or service marks of others. P26398 Please Recycle YTD03119-USEN-00