Image Segmentation by Complex-Valued Units

Similar documents
Artificial Neural Networks. Artificial Intelligence Santa Clara, 2016

Supplementary information accompanying the manuscript Biologically Inspired Modular Neural Control for a Leg-Wheel Hybrid Robot

Computing with Biologically Inspired Neural Oscillators: Application to Color Image Segmentation

Invariant Object Recognition in the Visual System with Novel Views of 3D Objects

Human Vision and Human-Computer Interaction. Much content from Jeff Johnson, UI Wizards, Inc.

Self Organising Neural Place Codes for Vision Based Robot Navigation

Extraction of Surface-Related Features in a Recurrent Model of V1-V2 Interactions

A Vestibular Sensation: Probabilistic Approaches to Spatial Perception (II) Presented by Shunan Zhang

TSBB15 Computer Vision

Object Perception. 23 August PSY Object & Scene 1

Swarm Intelligence W7: Application of Machine- Learning Techniques to Automatic Control Design and Optimization

Effects of Firing Synchrony on Signal Propagation in Layered Networks

Figure 1. Artificial Neural Network structure. B. Spiking Neural Networks Spiking Neural networks (SNNs) fall into the third generation of neural netw

3D Object Recognition Using Unsupervised Feature Extraction

A Robust Neural Robot Navigation Using a Combination of Deliberative and Reactive Control Architectures

TIME encoding of a band-limited function,,

Coding and computing with balanced spiking networks. Sophie Deneve Ecole Normale Supérieure, Paris

Eur Ing Dr. Lei Zhang Faculty of Engineering and Applied Science University of Regina Canada

Evolved Neurodynamics for Robot Control

A Hybrid Architecture using Cross Correlation and Recurrent Neural Networks for Acoustic Tracking in Robots

A specialized face-processing network consistent with the representational geometry of monkey face patches

Outline. Artificial Neural Network Importance of ANN Application of ANN is Sports Science

Shape Representation Robust to the Sketching Order Using Distance Map and Direction Histogram

Complex-valued neural networks fertilize electronics

SIMULATING RESTING CORTICAL BACKGROUND ACTIVITY WITH FILTERED NOISE. Journal of Integrative Neuroscience 7(3):

Artificial Neural Network Engine: Parallel and Parameterized Architecture Implemented in FPGA

Limulus eye: a filter cascade. Limulus 9/23/2011. Dynamic Response to Step Increase in Light Intensity

A Primer on Human Vision: Insights and Inspiration for Computer Vision

Geometric Neurodynamical Classifiers Applied to Breast Cancer Detection. Tijana T. Ivancevic

Simulate IFFT using Artificial Neural Network Haoran Chang, Ph.D. student, Fall 2018

University of California, Berkeley Department of Mathematics 5 th November, 2012, 12:10-12:55 pm MATH 53 - Test #2

A Self-organizing Method for Robot Navigation based on Learned Place and Head-direction cells

Dual Mechanisms for Neural Binding and Segmentation

CN510: Principles and Methods of Cognitive and Neural Modeling. Neural Oscillations. Lecture 24

NEURAL NETWORK BASED MAXIMUM POWER POINT TRACKING

Localization (Position Estimation) Problem in WSN

Optimum PID Control of Multi-wing Attractors in A Family of Lorenz-like Chaotic Systems

Adaptive Waveforms for Target Class Discrimination

The Nature of Informatics

A Primer on Human Vision: Insights and Inspiration for Computer Vision

Improved Detection by Peak Shape Recognition Using Artificial Neural Networks

The Grand Illusion and Petit Illusions

SCATTERING POLARIMETRY PART 1. Dr. A. Bhattacharya (Slide courtesy Prof. E. Pottier and Prof. L. Ferro-Famil)

Behavior Emergence in Autonomous Robot Control by Means of Feedforward and Recurrent Neural Networks

A Numerical Approach to Understanding Oscillator Neural Networks

Blood Vessel Tracking Technique for Optic Nerve Localisation for Field 1-3 Color Fundus Images

AN IMPROVED NEURAL NETWORK-BASED DECODER SCHEME FOR SYSTEMATIC CONVOLUTIONAL CODE. A Thesis by. Andrew J. Zerngast

The Basic Kak Neural Network with Complex Inputs

Visual Search using Principal Component Analysis

REPORT ITU-R SA.2098

Introduction to Machine Learning

Fundamentals of Computer Vision

An Improved Path Planning Method Based on Artificial Potential Field for a Mobile Robot

Sonia Sharma ECE Department, University Institute of Engineering and Technology, MDU, Rohtak, India. Fig.1.Neuron and its connection

Applying the Filtered Back-Projection Method to Extract Signal at Specific Position

An Efficient Color Image Segmentation using Edge Detection and Thresholding Methods

Beyond Blind Averaging Analyzing Event-Related Brain Dynamics

arxiv: v1 [cs.lg] 2 Jan 2018

A comparative study of different feature sets for recognition of handwritten Arabic numerals using a Multi Layer Perceptron

Representation Learning for Mobile Robots in Dynamic Environments

Information structuring in neural networks

MINE 432 Industrial Automation and Robotics

A Dynamical Systems Approach to Behavior-Based Formation Control

Transactions on Information and Communications Technologies vol 1, 1993 WIT Press, ISSN

Emergence of Purposive and Grounded Communication through Reinforcement Learning

Maps in the Brain Introduction

Binary Opening and Closing

NON UNIFORM BACKGROUND REMOVAL FOR PARTICLE ANALYSIS BASED ON MORPHOLOGICAL STRUCTURING ELEMENT:

CHAPTER 6 BACK PROPAGATED ARTIFICIAL NEURAL NETWORK TRAINED ARHF

A neuronal structure for learning by imitation. ENSEA, 6, avenue du Ponceau, F-95014, Cergy-Pontoise cedex, France. fmoga,

Lecture 4 Foundations and Cognitive Processes in Visual Perception From the Retina to the Visual Cortex

Evoked Potentials (EPs)

FACE RECOGNITION USING NEURAL NETWORKS

CSC384 Intro to Artificial Intelligence* *The following slides are based on Fahiem Bacchus course lecture notes.

Stock Price Prediction Using Multilayer Perceptron Neural Network by Monitoring Frog Leaping Algorithm

Advanced Techniques for Mobile Robotics Location-Based Activity Recognition

An Introduction To Modular Robots

Chapter 2 Channel Equalization

Math 148 Exam III Practice Problems

An Auditory Localization and Coordinate Transform Chip

Removing Temporal Stationary Blur in Route Panoramas

Initialisation improvement in engineering feedforward ANN models.

OBSERVED RELATION BETWEEN THE RELATIVE MIMO GAIN AND DISTANCE

High-speed Noise Cancellation with Microphone Array

Modeling cortical maps with Topographica

The Sine Function. Precalculus: Graphs of Sine and Cosine

Lecture IV. Sensory processing during active versus passive movements

III. Publication III. c 2005 Toni Hirvonen.

AN EFFICIENT APPROACH FOR VISION INSPECTION OF IC CHIPS LIEW KOK WAH

Blind Dereverberation of Single-Channel Speech Signals Using an ICA-Based Generative Model

Our visual system always has to compute a solid object given definite limitations in the evidence that the eye is able to obtain from the world, by

Digital image processing vs. computer vision Higher-level anchoring

Sound Source Localization using HRTF database

PSYC696B: Analyzing Neural Time-series Data

Convolutional neural networks

An in-silico Neural Model of Dynamic Routing through Neuronal Coherence

K.NARSING RAO(08R31A0425) DEPT OF ELECTRONICS & COMMUNICATION ENGINEERING (NOVH).

Control of a local neural network by feedforward and feedback inhibition

Simple Measures of Visual Encoding. vs. Information Theory

Chess Beyond the Rules

Transcription:

Image Segmentation by Complex-Valued Units Cornelius Weber and Stefan Wermter Hybrid Intelligent Systems, SCAT, University of Sunderland, UK Abstract. Spie synchronisation and de-synchronisation are important for feature binding and separation at various levels in the visual system. We present a model of complex valued neuron activations which are synchronised using lateral couplings. The firing rates of the model neurons correspond to a complex number s absolute value and obey conventional attractor networ relaxation dynamics, while the firing phases correspond to a complex number s angle and follow the dynamics of a logistic map. During relaxation, we show that features with strong couplings are grouped by firing in the same phase and are separated in phase from features that are coupled wealy or by negative weights. In an example, we apply the model to the level of a hidden representation of an image, thereby segmenting it on an abstract level. We imply that this process can facilitate unsupervised learning of objects in cluttered bacground. Introduction Object recognition is a ey tas in everyday situations and for robotics applications. Unsupervised learning of object classes from natural data is performed by young living beings and has a chance of becoming a convenient and flexible method of learning to categorise sensory data by an artificial agent. A hierarchy of increasingly complex feature detectors is one aspect of the visual recognition process. In many models, such a feature extracting step performs an almost linear vector transformation. So in order to achieve noticeable achievements in their serial application, a strong non-linearity must be introduced at every level. The non-linear response properties observed in cortical cells are explained in model studies by intra-area horizontal connections. A mathematical advantage to implementing these as an attractor networ is that its activations recover noisy input with maximum lielihood []. Contrast-invariant orientation tuning curves and shift invariant responses canbeobtained[2],asinvneurons. Unsupervised learning of objects is possible if objects are shown on a plain bacground, but still fails with a noisy bacground [3]. While the competitive effect of the attractor networ reduces bacground noise, in realistic conditions further percepts are the rule in addition to the object to be learnt. We therefore aim to separate these simultaneous percepts in the dimension of phase in order to separate an object from its bacground. In a hierarchical model this would allow higher levels to learn only the object at certain phases or only bacground elements at other phases and will facilitate unsupervised learning of objects. Detailed spiing neuron models are attractive for segmentation purposes. In addition to the neurons firing rate their code provides information that can W. Duch et al. (Eds.): ICANN 25, LNCS 3696, pp. 59 524, 25. c Springer-Verlag Berlin Heidelberg 25

52 C. Weber and S. Wermter be used for mutual binding and segregation. Computationally it is efficient to incorporate these additional capabilities of spies in a single variable per neuron which we call a neuron s phase. A process to adjust this variable efficiently is deterministic chaos with a dual role of (i) supplying a process of pattern creation by synchronising the phases of coupled units and at the same time (ii) revolting against convergence into stereotyped synchronised states [4]. Our approach has the following characteristics: (i) Coupling strengths are represented by connection weights that can be trained to represent correlated activations and which can be negative. (ii) The weighting of the neural inputs is performed by complex number algebra. Complex-valued neural networs have advantages for chaotic and brain-lie systems, image processing and quantum devices [6]. We identify the absolute value of a complex neuron activation with its firing rate and the phase of its activation with the phase at which its spies are emitted. (iii) The rates follow a conventional update dynamics of a recurrent neural net. (iv) A logistic map provides the chaotic dynamics to synchronise and separate the phases, as in the Divide and Conquer model [5]. (v) Finally, we show an application where an image is segmented on its hidden, first cortical representation on area V, similar to the competitive layer model [7]. 2 Real Valued Relaxation Procedure We use one layer of fully connected units to define a recurrent update dynamics for the neurons rates and phases. After some, possibly random, initialisation of the firing rates, the rate r of any unit is governed over time steps t by r (t +) = f( j w jr j (t) θ ) () where w j is the connection weight from unit j to unit and θ is its threshold. The transfer function is f(x) =/( + e x ). A learning rule for the weights and thresholds is given in [8] and weights which sustain a bell-shaped hill of activation rates will have a Mexican hat shaped profile. Thus, we have a continuous valued attractor networ and since weights are approximately symmetric, w j w j, our intuition is that activations r relaxate to a stable state corresponding to a minimum of some energy function. In the following, we will introduce a second variable, the phase ϕ, with different dynamics. 3 Complex Valued Interactions A complex number as displayed in Fig. a) can be written as z = x+iy = re iϕ with i 2 = and the relations r 2 = x 2 + y 2 and tan ϕ = y x.weexpressour neuronal activation as z = r e iϕ where the complex number s length r is the neuron s firing rate and its phase ϕ is the phase at which the neuron spies. Similar phases of two neurons would correspond to similar firing times if their rates were the same, however, we regard these phases as abstract.

Image Segmentation by Complex-Valued Units 52 a) b) Φ(t+) π/2 < ϕ < π i < ϕ < π/2 z r ϕ y x π < ϕ < 3π/2 i 3π/2 < ϕ < 2π Φ(t) t=2 t= t=3 t= Fig.. a) A complex number z can be expressed by Cartesian coordinates x, y or polar coordinates r, ϕ. We identify its length r with a neuronal firing rate and angle ϕ with a firing phase. The ranges of ϕ are displayed as used in the text. b) The logistic map. Iterations according to ϕ(t + )= 3.9 ϕ(t)( ϕ(t)), indicated for t =...3 starting at ϕ(t=) =.4, will lead to a desired chaotic behaviour. The dynamics of the rates follows Eq. and is independent of the phases. In the following we will first define how phases between neurons interact (Eq. 2) and then impose a local update dynamic on every neuron (Eq. 3). At the beginning of the relaxation procedure, the neurons phases ϕ are initialised with random values between and 2π. Then at each relaxation step a neuron receives an influence from all other neurons, which we express as: z wf = j w j r j e iϕj (2) This weighted field is a complex number which is a sum of the complex number contributions by the other neurons weighted by the connection weights w j which are real values. Using e iϕj =cosϕ j + i sin ϕ j it can be expressed as: z wf = j w jr j cos ϕ j + i j w jr j sin ϕ j x wf + iy wf We obtain the phase of the neuron s weighted field as: we shift to range between and 2π according to Fig. a). ϕ wf =atanywf x which 4 Logistic Coupled Map The logistic map maps a value between and to another, different value within this interval, as shown in Fig. b). Iterative application leads to desired chaotic development of these values for most settings of the map parameter A between 3.57 and 4.. Nevertheless, if several values undergoing this mapping are coupled, they can maintain proximity and thus display structured mutual behaviour while displaying chaotic individual behaviour [4]. Since the logistic map taes values ranging between and while the neurons phases range from to 2π, we mae the technical definitions: Φ ϕ 2π,Φwf ϕwf 2π.

522 C. Weber and S. Wermter We will now use the scaled phase Φ wf of the weighted field at each neuron in order to determine its scaled phase Φ at the next iteration time step. This is done via the logistic map: Φ (t +)=AΦ wf (t)( Φwf (t)) (3) where its actual phase value ϕ is scaled bac to the range from to 2π. Wehave set A = 3.9. Having obtained the phases at the next time step, another iteration for the phases is performed starting with computing the weighted field (Eq. 2). While the rates develop concurrently according to Eq. to a stable state, the phase values never converge. 5 Networ Activation with Synchronising Phases Figure 2 a) shows the activations of a networ with 25 units as their rates have converged to a stable state according to Eq.. This is in no way influenced by the phases but only by the weight profile, which is also displayed for one neuron. The weight profile with strong positive weights between neighbouring units should synchronise the phases between such connected units. A single unit s phase behaves random-lie from one time step to the next. Fig. 2 b) shows a plot of time averaged phase differences between pairs of units, while the networ is maintaining the rates shown in Fig. 2 a). It shows that adjacent units within the hill of activation have similar phases, while adjacent units at its boundary have differing phases. The phases are thus clustering regions of strong activity that are lined by strong positive weights (phase influence is weighted by weights times rates, cf. Eq. 2). Regions with negative connections to such a cluster have differing phases, as we see units with zero activation sharing an own phase. Note that since adjacency is defined functionally by mutual strong positive connections, long-range connections could mediate synchronisation over large distances. a) weights rates b) 2 sep5, sep+,.4 5 5 25 5 5 25 Fig. 2. a) shows the weight profile (dotted line) of neuron number 5 and a bell-shaped hill of activation rates as sustained by the networ (solid line). b) shows the average phase separation sep +, = ϕ + ϕ between neighbouring units (solid line) and sep 5, = ϕ 5 ϕ between any unit and unit 5 (dotted line), where the. -bracets denote a time average over 5 iteration steps while sustaining the rates in a).

6 Segmentation in a Feature Space Image Segmentation by Complex-Valued Units 523 In the following experiment we apply the attractor networ to the activations of neurons which have been trained to extract features from natural images and which thereby resemble V simple cells [2]. Given an image as input, these units have sparse activations with values between and while responding to edges and colour features. V lateral weights have been trained on the same set of natural images to memorise these codes in the attractor activation patterns. The resulting weights are short-range excitatory and long-range inhibitory along cortical distance, as well as in feature space of orientation and spatial frequency [2]. This leads, during relaxation, to focused patterns of the activation rates on the simulated V, after initialising with a somewhat irregular activity pattern obtained from presenting an image. From this V representation, a virtual reconstruction of the image can be obtained by projecting these rates bac to the image. Fig. 3 a) shows an example. Fig. 3 c) shows the hidden units activations where each frame shows only those units activations r which have a phase ϕ within a range shown in Fig. 3 b). The active units in each frame are thus a subset of all active units shown in Fig. 3 a), middle. It can be seen that within any selected phase range, preferably units from a certain region are active. The functions of the neurons within these active clusters can be seen by projecting their rates to the image. As a result, Fig. 3 d) shows partial reconstructions of the image which is hereby segmented into elements which belong together by having similar phases on the model V. At different phases we find elements of the bacground (frames and 2) or of the ball at the right (frames 4 and 5). This implies a segmentation which accounts for learnt proximities (via the V lateral weights) in an abstract representation of an image with the potential of separating objects from their bacground. image a) b) π 2 π π 4 π 5 π 3 3 3 3 V ratesj a) c) reconstr.j a) d) Fig. 3. a) shows full representations. Top, the original image, middle, the rates of the full hidden code and bottom, the reconstruction of the image from the full hidden code. b),c),d) show partial phase-dependent representations. b) shows Gaussian-lie functions on an axis of to 2π used to determine which phases contribute to the hidden code presented in c) and thus to the image reconstruction in d). c) shows the partial hidden code corresponding to selected phases and d) shows their partial reconstruction from those units in c) which are active at the selected phases.

524 C. Weber and S. Wermter 7 Discussion We have demonstrated a networ of simplified rate and phase coding neurons which segments a neural code efficiently using the connection strengths between the units. The computational load is that of a networ in which two activation values develop concurrently. The most demanding operations are the scalar products in Eqs. and 2, while all other computations are local. Since image segmentation involves top-down directed information flow, how does our model for intra-area lateral connections deal with this? Previously we have extended the lateral connections to lin a what - to a where area for object localisation [9]. In the cortex, such lateral connections correspond to those originating from pyramidal cells in layers 2/3 of the cortex and arriving in the same layer, possibly in a different area. Hierarchically arranged areas have characteristically asymmetric connections, however, they also have characteristic horizontal connections and only with increasing hierarchical level difference the intensity of these horizontal connections decreases []. Thus, connections of a horizontal character may relay top-down information. If we would apply the lateral connections of our model to a larger hierarchical model, therefore, we might observe top-down influences such as stabilisation of consistent attractors. Acnowledgements. This wor is part of the MirrorBot project supported by the EU in the FET-IST programme under grant IST-2-35282. References [] Deneve, S., Latham, P., Pouget, A.: Reading population codes: a neural implementation of ideal observers. Nature Neurosci. 2 (999) 74 5 [2] Weber, C.: Self-organization of orientation maps, lateral connections, and dynamic receptive fields in the primary visual cortex.in Dorffner,G.,Bischof,H.,Horni, K., eds.: Proc. ICANN, Springer-Verlag Berlin Heidelberg (2) 47 52 [3] Stringer, S., Rolls, E.: Position invariant recognition in the visual system with cluttered environments. Neural Networs 3 (2) 35 5 [4] van Leeuwen, C., Steyvers, M., Nooter, M.: Stability and intermittency in largescale coupled oscillator models for perceptual segmentation. J. Mathematical Psychology 4 (997) 39 44 [5] Raffone, A., van Leeuwen, C.: The divide and conquer model of image segmentation: object-bounded synchrony propagation in coupled map lattices. In: Proc. KES Knowledge-Based Intelligent Information & Engineering Systems. (22) [6] Hirose, A., ed.: Complex-Valued Neural Networs: Theories and Applications. World Scientific Publishing Co. (23) [7] Ontrup, J. Wersing, H., Ritter, H.: A computational feature binding model of human texture perception. Cognitive Processing 5 (24) 32 44 [8] Zhang, K.: Representation of spatial orientation by the intrinsic dynamics of the head-direction cell ensemble: A theory. J. Neurosci. 6 (996) 22 26 [9] Weber, C., Wermter, S.: Object localization using laterally connected what and where associator networs. In: Proc. ICANN/ICONIP, Springer (23) 83 2 [] Felleman, D., Van Essen, D.: Distributed hierarchical processing in the primate cerebral cortex. Cerebral Cortex (99) 47