Log-linear models (part III)
|
|
- Buddy Webster
- 5 years ago
- Views:
Transcription
1 Log-linear models (part III) Lecture, Feb 7 CS 690N, Spring 2017 Advanced Natural Language Processing Brendan O Connor College of Information and Computer Sciences University of Massachusetts Amherst
2 MaxEnt / Log-Linear models x: input (all previous words) y: output (next word) f(x,y) => Rd feature function [[domain knowledge here!]] v: Rd Y parameter vector (weights) p(y x; v) = exp (v f(x, y)) P y 0 2Y exp (v f(x, y0 )) P Application to history-based LM: P (w 1..w T )= Y t = Y t P (w t w 1..w t 1 ) exp(v f(w 1..w t 1,w t )) P w2v exp(v f(w 1..w t 1,w))
3 Learning log p(y x; v) = v f(x, y) j log p(y x; v) = y 0 2Y exp v f(x, y 0 ) Gradient at a single example: can it be zero? Full dataset gradient: First moments match at the mode Log-likelihood is concave At least with regularization, since typically linearly separable Is my function convex? Check Boyd and Vandenberghe ch. 3 3
4 Learning log p(y x; v) = v f(x, y) log X y 0 2Y exp v f(x, y j log p(y x; v) = fun with the chain rule Gradient at a single example: can it be zero? Full dataset gradient: First moments match at the mode Log-likelihood is concave At least with regularization, since typically linearly separable Is my function convex? Check Boyd and Vandenberghe ch. 3 3
5 Learning log p(y x; v) = v f(x, y) log X y 0 2Y exp v f(x, y j log p(y x; v) = fun with the chain rule f j (x, y) X y 0 p(y 0 x; v)f j (x, y 0 ) Gradient at a single example: can it be zero? Full dataset gradient: First moments match at the mode Log-likelihood is concave At least with regularization, since typically linearly separable Is my function convex? Check Boyd and Vandenberghe ch. 3 3
6 Learning log p(y x; v) = v f(x, y) log X y 0 2Y exp v f(x, y j log p(y x; v) = fun with the chain rule f j (x, y) Feature in data? X y 0 p(y 0 x; v)f j (x, y 0 ) Feature in posterior? Gradient at a single example: can it be zero? Full dataset gradient: First moments match at the mode Log-likelihood is concave At least with regularization, since typically linearly separable Is my function convex? Check Boyd and Vandenberghe ch. 3 3
7 Gradient descent Batch gradient descent (doesn t work well by itself) Most commonly used alternatives LBFGS (adaptive version of batch GD) Call a library implementation with gradient callback SGD, one example at a time and adaptive variants: Adagrad, Adam, etc. Intuition Issue: Combining per-example sparse updates with regularization updates Lazy updates Occasional regularizer steps (easy to implement) 4
8 stopped here on 2/7 5
9 Engineering Sparse dot products are crucial! Lots and lots of features? Millions to billions of features: performance often keeps improving! Features seen only once at training time typically help Feature name=>number mapping is the problem; the parameter vector is fine Feature hashing: make e.g. N(u,v,w) mapping random with collisions (!) Accuracy loss low since features are rare. Works well, great for large-scale data (memory usage constant!) Practically: use a fast string hashing function (e.g. murmurhash or Python s internal one) 6
10 Feature selection Offline feature selection Count cutoffs: computational, not performance benefits Predictive value: mutual info. / info. gain / chi-square L1 regularization: encourages θ sparsity min log p (y x)+ X j j L1 optimization: convex but nonsmooth; requires subgradient methods 7
11 Dense representations Saul and Pereira 1997? Mnih and Hinton 2007: log-bilinear model 8
12 Bengio et al. 2003: N-gram MLP f (w t,,w t n+1 )= ˆP(w t w t 1 1 ) i-th output = P(w t = i context) ( ) softmax most computation here tanh C(w t n+1 )... Table look up in C... C(w t 2 ) C(w t 1 ) Matrix C shared parameters across words w t n+1 w t 2 index for index for index for w t 1 C(i) 2 R m. Word embedding parameters 9 x =(C(w t 1 ),C(w t 2 ),,C(w t n+1 )). y = b +Wx+U tanh(d + Hx) ˆP(w t w t 1, w t n+1 )= ey wt i e y i.
Log-linear models (part 1I)
Log-linear models (part 1I) Lecture, Feb 2 CS 690N, Spring 2017 Advanced Natural Language Processing http://people.cs.umass.edu/~brenocon/anlp2017/ Brendan O Connor College of Information and Computer
More informationLog-linear models (part 1I)
Log-linear models (part 1I) CS 690N, Spring 2018 Advanced Natural Language Processing http://people.cs.umass.edu/~brenocon/anlp2018/ Brendan O Connor College of Information and Computer Sciences University
More informationCRF and Structured Perceptron
CRF and Structured Perceptron CS 585, Fall 2015 -- Oct. 6 Introduction to Natural Language Processing http://people.cs.umass.edu/~brenocon/inlp2015/ Brendan O Connor Viterbi exercise solution CRF & Structured
More informationKernels and Support Vector Machines
Kernels and Support Vector Machines Machine Learning CSE446 Sham Kakade University of Washington November 1, 2016 2016 Sham Kakade 1 Announcements: Project Milestones coming up HW2 You ve implemented GD,
More informationCompressive Sampling with R: A Tutorial
1/15 Mehmet Süzen msuzen@mango-solutions.com data analysis that delivers 15 JUNE 2011 2/15 Plan Analog-to-Digital conversion: Shannon-Nyquist Rate Medical Imaging to One Pixel Camera Compressive Sampling
More informationLesson 08. Convolutional Neural Network. Ing. Marek Hrúz, Ph.D. Katedra Kybernetiky Fakulta aplikovaných věd Západočeská univerzita v Plzni.
Lesson 08 Convolutional Neural Network Ing. Marek Hrúz, Ph.D. Katedra Kybernetiky Fakulta aplikovaných věd Západočeská univerzita v Plzni Lesson 08 Convolution we will consider 2D convolution the result
More informationMidterm for Name: Good luck! Midterm page 1 of 9
Midterm for 6.864 Name: 40 30 30 30 Good luck! 6.864 Midterm page 1 of 9 Part #1 10% We define a PCFG where the non-terminals are {S, NP, V P, V t, NN, P P, IN}, the terminal symbols are {Mary,ran,home,with,John},
More informationCoursework 2. MLP Lecture 7 Convolutional Networks 1
Coursework 2 MLP Lecture 7 Convolutional Networks 1 Coursework 2 - Overview and Objectives Overview: Use a selection of the techniques covered in the course so far to train accurate multi-layer networks
More informationStudy guide for Graduate Computer Vision
Study guide for Graduate Computer Vision Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA 01003 November 23, 2011 Abstract 1 1. Know Bayes rule. What
More informationarxiv: v1 [cs.lg] 23 Aug 2016
Learning to Communicate: Channel Auto-encoders, Domain Specific Regularizers, and Attention arxiv:1608.06409v1 [cs.lg] 23 Aug 2016 Timothy J. O Shea Virginia Tech ECE Arlington, VA oshea@vt.edu T. Charles
More informationEmbeddings Learned by Gradient Descent
Embeddings Learned by Gradient Descent Hinrich Schütze Center for Information and Language Processing, LMU Munich 2017-07-20 Schütze (LMU Munich): Embeddings via gradient descent 1 / 46 Overview 1 word2vec
More informationEmpirical Rate-Distortion Study of Compressive Sensing-based Joint Source-Channel Coding
Empirical -Distortion Study of Compressive Sensing-based Joint Source-Channel Coding Muriel L. Rambeloarison, Soheil Feizi, Georgios Angelopoulos, and Muriel Médard Research Laboratory of Electronics Massachusetts
More informationLecture 3 - Regression
Lecture 3 - Regression Instructor: Prof Ganesh Ramakrishnan July 25, 2016 1 / 30 The Simplest ML Problem: Least Square Regression Curve Fitting: Motivation Error measurement Minimizing Error Method of
More informationCSC321 Lecture 11: Convolutional Networks
CSC321 Lecture 11: Convolutional Networks Roger Grosse Roger Grosse CSC321 Lecture 11: Convolutional Networks 1 / 35 Overview What makes vision hard? Vison needs to be robust to a lot of transformations
More informationDigital Television Lecture 5
Digital Television Lecture 5 Forward Error Correction (FEC) Åbo Akademi University Domkyrkotorget 5 Åbo 8.4. Error Correction in Transmissions Need for error correction in transmissions Loss of data during
More informationLocal Search: Hill Climbing. When A* doesn t work AIMA 4.1. Review: Hill climbing on a surface of states. Review: Local search and optimization
Outline When A* doesn t work AIMA 4.1 Local Search: Hill Climbing Escaping Local Maxima: Simulated Annealing Genetic Algorithms A few slides adapted from CS 471, UBMC and Eric Eaton (in turn, adapted from
More informationIntroduction to Machine Learning
Introduction to Machine Learning Deep Learning Barnabás Póczos Credits Many of the pictures, results, and other materials are taken from: Ruslan Salakhutdinov Joshua Bengio Geoffrey Hinton Yann LeCun 2
More informationMachine Learning. Classification, Discriminative learning. Marc Toussaint University of Stuttgart Summer 2014
Machine Learning Classification, Discriminative learning Structured output, structured input, discriminative function, joint input-output features, Likelihood Maximization, Logistic regression, binary
More informationRecommender Systems TIETS43 Collaborative Filtering
+ Recommender Systems TIETS43 Collaborative Filtering Fall 2017 Kostas Stefanidis kostas.stefanidis@uta.fi https://coursepages.uta.fi/tiets43/ selection Amazon generates 35% of their sales through recommendations
More informationAutomatic Speech Recognition (CS753)
Automatic Speech Recognition (CS753) Lecture 9: Brief Introduction to Neural Networks Instructor: Preethi Jyothi Feb 2, 2017 Final Project Landscape Tabla bol transcription Music Genre Classification Audio
More informationResearch on Hand Gesture Recognition Using Convolutional Neural Network
Research on Hand Gesture Recognition Using Convolutional Neural Network Tian Zhaoyang a, Cheng Lee Lung b a Department of Electronic Engineering, City University of Hong Kong, Hong Kong, China E-mail address:
More informationRadio Deep Learning Efforts Showcase Presentation
Radio Deep Learning Efforts Showcase Presentation November 2016 hume@vt.edu www.hume.vt.edu Tim O Shea Senior Research Associate Program Overview Program Objective: Rethink fundamental approaches to how
More informationMachine Learning for Language Technology
Machine Learning for Language Technology Generative and Discriminative Models Joakim Nivre Uppsala University Department of Linguistics and Philology joakim.nivre@lingfil.uu.se Machine Learning for Language
More informationDeep Neural Networks (2) Tanh & ReLU layers; Generalisation and Regularisation
Deep Neural Networks (2) Tanh & ReLU layers; Generalisation and Regularisation Steve Renals Machine Learning Practical MLP Lecture 4 9 October 2018 MLP Lecture 4 / 9 October 2018 Deep Neural Networks (2)
More informationEE 435/535: Error Correcting Codes Project 1, Fall 2009: Extended Hamming Code. 1 Introduction. 2 Extended Hamming Code: Encoding. 1.
EE 435/535: Error Correcting Codes Project 1, Fall 2009: Extended Hamming Code Project #1 is due on Tuesday, October 6, 2009, in class. You may turn the project report in early. Late projects are accepted
More informationFast Blur Removal for Wearable QR Code Scanners (supplemental material)
Fast Blur Removal for Wearable QR Code Scanners (supplemental material) Gábor Sörös, Stephan Semmler, Luc Humair, Otmar Hilliges Department of Computer Science ETH Zurich {gabor.soros otmar.hilliges}@inf.ethz.ch,
More informationCHAPTER 4 SIGNAL SPACE. Xijun Wang
CHAPTER 4 SIGNAL SPACE Xijun Wang WEEKLY READING 1. Goldsmith, Wireless Communications, Chapters 5 2. Gallager, Principles of Digital Communication, Chapter 5 2 DIGITAL MODULATION AND DEMODULATION n Digital
More informationAugmenting Self-Learning In Chess Through Expert Imitation
Augmenting Self-Learning In Chess Through Expert Imitation Michael Xie Department of Computer Science Stanford University Stanford, CA 94305 xie@cs.stanford.edu Gene Lewis Department of Computer Science
More informationCSE 258 Winter 2017 Assigment 2 Skill Rating Prediction on Online Video Game
ABSTRACT CSE 258 Winter 2017 Assigment 2 Skill Rating Prediction on Online Video Game In competitive online video game communities, it s common to find players complaining about getting skill rating lower
More informationRecurrent neural networks Modelling sequential data. MLP Lecture 9 Recurrent Networks 1
Recurrent neural networks Modelling sequential data MLP Lecture 9 Recurrent Networks 1 Recurrent Networks Steve Renals Machine Learning Practical MLP Lecture 9 16 November 2016 MLP Lecture 9 Recurrent
More informationDeep Learning. Dr. Johan Hagelbäck.
Deep Learning Dr. Johan Hagelbäck johan.hagelback@lnu.se http://aiguy.org Image Classification Image classification can be a difficult task Some of the challenges we have to face are: Viewpoint variation:
More informationIntroduction to Markov Models
Introduction to Markov Models But first: A few preliminaries Estimating the probability of phrases of words, sentences, etc. CIS 391 - Intro to AI 2 What counts as a word? A tricky question. How to find
More informationMonty Hall Problem & Birthday Paradox
Monty Hall Problem & Birthday Paradox Hanqiu Peng Abstract There are many situations that our intuitions lead us to the wrong direction, especially when we are solving some probability problems. In this
More information14.7 Maximum and Minimum Values
CHAPTER 14. PARTIAL DERIVATIVES 115 14.7 Maximum and Minimum Values Definition. Let f(x, y) be a function. f has a local max at (a, b) iff(a, b) (a, b). f(x, y) for all (x, y) near f has a local min at
More informationLecture 17 Convolutional Neural Networks
Lecture 17 Convolutional Neural Networks 30 March 2016 Taylor B. Arnold Yale Statistics STAT 365/665 1/22 Notes: Problem set 6 is online and due next Friday, April 8th Problem sets 7,8, and 9 will be due
More informationTiny ImageNet Challenge Investigating the Scaling of Inception Layers for Reduced Scale Classification Problems
Tiny ImageNet Challenge Investigating the Scaling of Inception Layers for Reduced Scale Classification Problems Emeric Stéphane Boigné eboigne@stanford.edu Jan Felix Heyse heyse@stanford.edu Abstract Scaling
More informationStatistical Inference, Learning and Models for Big Data
Statistical Inference, Learning and Models for Big Data Nancy Reid University of Toronto December 2, 2015 Canadian Institute for Statistical Sciences Fields Institute for Resesarch in the Mathematical
More informationReinforcement Learning Agent for Scrolling Shooter Game
Reinforcement Learning Agent for Scrolling Shooter Game Peng Yuan (pengy@stanford.edu) Yangxin Zhong (yangxin@stanford.edu) Zibo Gong (zibo@stanford.edu) 1 Introduction and Task Definition 1.1 Game Agent
More informationSignal Recovery from Random Measurements
Signal Recovery from Random Measurements Joel A. Tropp Anna C. Gilbert {jtropp annacg}@umich.edu Department of Mathematics The University of Michigan 1 The Signal Recovery Problem Let s be an m-sparse
More informationCandyCrush.ai: An AI Agent for Candy Crush
CandyCrush.ai: An AI Agent for Candy Crush Jiwoo Lee, Niranjan Balachandar, Karan Singhal December 16, 2016 1 Introduction Candy Crush, a mobile puzzle game, has become very popular in the past few years.
More informationDeep Learning Basics Lecture 9: Recurrent Neural Networks. Princeton University COS 495 Instructor: Yingyu Liang
Deep Learning Basics Lecture 9: Recurrent Neural Networks Princeton University COS 495 Instructor: Yingyu Liang Introduction Recurrent neural networks Dates back to (Rumelhart et al., 1986) A family of
More informationMATH 8 FALL 2010 CLASS 27, 11/19/ Directional derivatives Recall that the definitions of partial derivatives of f(x, y) involved limits
MATH 8 FALL 2010 CLASS 27, 11/19/2010 1 Directional derivatives Recall that the definitions of partial derivatives of f(x, y) involved limits lim h 0 f(a + h, b) f(a, b), lim h f(a, b + h) f(a, b) In these
More informationPrediction of Cluster System Load Using Artificial Neural Networks
Prediction of Cluster System Load Using Artificial Neural Networks Y.S. Artamonov 1 1 Samara National Research University, 34 Moskovskoe Shosse, 443086, Samara, Russia Abstract Currently, a wide range
More informationDynamic Data-Driven Adaptive Sampling and Monitoring of Big Spatial-Temporal Data Streams for Real-Time Solar Flare Detection
Dynamic Data-Driven Adaptive Sampling and Monitoring of Big Spatial-Temporal Data Streams for Real-Time Solar Flare Detection Dr. Kaibo Liu Department of Industrial and Systems Engineering University of
More informationDeepStack: Expert-Level AI in Heads-Up No-Limit Poker. Surya Prakash Chembrolu
DeepStack: Expert-Level AI in Heads-Up No-Limit Poker Surya Prakash Chembrolu AI and Games AlphaGo Go Watson Jeopardy! DeepBlue -Chess Chinook -Checkers TD-Gammon -Backgammon Perfect Information Games
More informationEnd-to-End Differentiable Proving
End-to-End Differentiable Proving Tim Rocktäschel 1 and Sebastian Riedel 2,3 1 University of Oxford Whiteson Research Lab 2 University College London Machine Reading Lab 3 Bloomsbury AI http://rockt.github.com
More informationFrugal Sensing Spectral Analysis from Power Inequalities
Frugal Sensing Spectral Analysis from Power Inequalities Nikos Sidiropoulos Joint work with Omar Mehanna IEEE SPAWC 2013 Plenary, June 17, 2013, Darmstadt, Germany Wideband Spectrum Sensing (for CR/DSM)
More informationCompressed Sensing for Multiple Access
Compressed Sensing for Multiple Access Xiaodai Dong Wireless Signal Processing & Networking Workshop: Emerging Wireless Technologies, Tohoku University, Sendai, Japan Oct. 28, 2013 Outline Background Existing
More informationIntroduction to Markov Models. Estimating the probability of phrases of words, sentences, etc.
Introduction to Markov Models Estimating the probability of phrases of words, sentences, etc. But first: A few preliminaries on text preprocessing What counts as a word? A tricky question. CIS 421/521
More informationSPARSE CHANNEL ESTIMATION BY PILOT ALLOCATION IN MIMO-OFDM SYSTEMS
SPARSE CHANNEL ESTIMATION BY PILOT ALLOCATION IN MIMO-OFDM SYSTEMS Puneetha R 1, Dr.S.Akhila 2 1 M. Tech in Digital Communication B M S College Of Engineering Karnataka, India 2 Professor Department of
More informationCoE4TN4 Image Processing. Chapter 3: Intensity Transformation and Spatial Filtering
CoE4TN4 Image Processing Chapter 3: Intensity Transformation and Spatial Filtering Image Enhancement Enhancement techniques: to process an image so that the result is more suitable than the original image
More informationAttentive Neural Architecture Incorporating Song Features For Music Recommendation
Attentive Neural Architecture Incorporating Song Features For Music Recommendation by Noveen Sachdeva, Kartik Gupta, Vikram Pudi in 12th ACM Conference on Recommender Systems (RECSYS-2018) Vancouver, Canada
More informationGenerating an appropriate sound for a video using WaveNet.
Australian National University College of Engineering and Computer Science Master of Computing Generating an appropriate sound for a video using WaveNet. COMP 8715 Individual Computing Project Taku Ueki
More informationGame Theory and Randomized Algorithms
Game Theory and Randomized Algorithms Guy Aridor Game theory is a set of tools that allow us to understand how decisionmakers interact with each other. It has practical applications in economics, international
More informationMusic Recommendation using Recurrent Neural Networks
Music Recommendation using Recurrent Neural Networks Ashustosh Choudhary * ashutoshchou@cs.umass.edu Mayank Agarwal * mayankagarwa@cs.umass.edu Abstract A large amount of information is contained in the
More informationRecovering Lost Sensor Data through Compressed Sensing
Recovering Lost Sensor Data through Compressed Sensing Zainul Charbiwala Collaborators: Younghun Kim, Sadaf Zahedi, Supriyo Chakraborty, Ting He (IBM), Chatschik Bisdikian (IBM), Mani Srivastava The Big
More informationEE 123 Discussion Section 6. Frank Ong March 14th, 2016
EE 123 Discussion Section 6 Frank Ong March 14th, 2016 Plan Sparse FFT Magnitude Filter Design with convex optimization Sparse FFT Given a length-n signal, FFT takes O(N log N) time to compute its DFT
More informationEE359 Discussion Session 8 Beamforming, Diversity-multiplexing tradeoff, MIMO receiver design, Multicarrier modulation
EE359 Discussion Session 8 Beamforming, Diversity-multiplexing tradeoff, MIMO receiver design, Multicarrier modulation November 29, 2017 EE359 Discussion 8 November 29, 2017 1 / 33 Outline 1 MIMO concepts
More informationMillion Song Dataset Challenge!
1 Introduction Million Song Dataset Challenge Fengxuan Niu, Ming Yin, Cathy Tianjiao Zhang Million Song Dataset (MSD) is a freely available collection of data for one million of contemporary songs (http://labrosa.ee.columbia.edu/millionsong/).
More informationAre there alternatives to Sigmoid Hidden Units? MLP Lecture 6 Hidden Units / Initialisation 1
Are there alternatives to Sigmoid Hidden Units? MLP Lecture 6 Hidden Units / Initialisation 1 Hidden Unit Transfer Functions Initialising Deep Networks Steve Renals Machine Learning Practical MLP Lecture
More informationAn Adaptive Intelligence For Heads-Up No-Limit Texas Hold em
An Adaptive Intelligence For Heads-Up No-Limit Texas Hold em Etan Green December 13, 013 Skill in poker requires aptitude at a single task: placing an optimal bet conditional on the game state and the
More information신경망기반자동번역기술. Konkuk University Computational Intelligence Lab. 김강일
신경망기반자동번역기술 Konkuk University Computational Intelligence Lab. http://ci.konkuk.ac.kr kikim01@kunkuk.ac.kr 김강일 Index Issues in AI and Deep Learning Overview of Machine Translation Advanced Techniques in
More informationComputer Vision, Lecture 3
Computer Vision, Lecture 3 Professor Hager http://www.cs.jhu.edu/~hager /4/200 CS 46, Copyright G.D. Hager Outline for Today Image noise Filtering by Convolution Properties of Convolution /4/200 CS 46,
More informationTracking Algorithms for Multipath-Aided Indoor Localization
Tracking Algorithms for Multipath-Aided Indoor Localization Paul Meissner and Klaus Witrisal Graz University of Technology, Austria th UWB Forum on Sensing and Communication, May 5, Meissner, Witrisal
More informationEnergy Consumption Prediction for Optimum Storage Utilization
Energy Consumption Prediction for Optimum Storage Utilization Eric Boucher, Robin Schucker, Jose Ignacio del Villar December 12, 2015 Introduction Continuous access to energy for commercial and industrial
More informationVehicle Color Recognition using Convolutional Neural Network
Vehicle Color Recognition using Convolutional Neural Network Reza Fuad Rachmadi and I Ketut Eddy Purnama Multimedia and Network Engineering Department, Institut Teknologi Sepuluh Nopember, Keputih Sukolilo,
More informationDiet Networks: Thin Parameters for Fat Genomics
Institut des algorithmes d apprentissage de Montréal Diet Networks: Thin Parameters for Fat Genomics Adriana Romero, Pierre Luc Carrier, Akram Erraqabi, Tristan Sylvain, Alex Auvolat, Etienne Dejoie, Marc-André
More informationAttention-based Multi-Encoder-Decoder Recurrent Neural Networks
Attention-based Multi-Encoder-Decoder Recurrent Neural Networks Stephan Baier 1, Sigurd Spieckermann 2 and Volker Tresp 1,2 1- Ludwig Maximilian University Oettingenstr. 67, Munich, Germany 2- Siemens
More informationDeep Reinforcement Learning and Forward Modeling for StarCraft AI
M2 Mathématiques, Vision et Apprentissage École Normale Supérieure de Cachan Deep Reinforcement Learning and Forward Modeling for StarCraft AI Internship Report Alex Auvolat Under the supervision of: Gabriel
More informationGenerating Groove: Predicting Jazz Harmonization
Generating Groove: Predicting Jazz Harmonization Nicholas Bien (nbien@stanford.edu) Lincoln Valdez (lincolnv@stanford.edu) December 15, 2017 1 Background We aim to generate an appropriate jazz chord progression
More informationReview Sheet for Math 230, Midterm exam 2. Fall 2006
Review Sheet for Math 230, Midterm exam 2. Fall 2006 October 31, 2006 The second midterm exam will take place: Monday, November 13, from 8:15 to 9:30 pm. It will cover chapter 15 and sections 16.1 16.4,
More informationCS277 - Experimental Haptics Lecture 2. Haptic Rendering
CS277 - Experimental Haptics Lecture 2 Haptic Rendering Outline Announcements Human haptic perception Anatomy of a visual-haptic simulation Virtual wall and potential field rendering A note on timing...
More informationUnderstanding Neural Networks : Part II
TensorFlow Workshop 2018 Understanding Neural Networks Part II : Convolutional Layers and Collaborative Filters Nick Winovich Department of Mathematics Purdue University July 2018 Outline 1 Convolutional
More informationClipping Noise Cancellation Based on Compressed Sensing for Visible Light Communication
Clipping Noise Cancellation Based on Compressed Sensing for Visible Light Communication Presented by Jian Song jsong@tsinghua.edu.cn Tsinghua University, China 1 Contents 1 Technical Background 2 System
More information3D-Assisted Image Feature Synthesis for Novel Views of an Object
3D-Assisted Image Feature Synthesis for Novel Views of an Object Hao Su* Fan Wang* Li Yi Leonidas Guibas * Equal contribution View-agnostic Image Retrieval Retrieval using AlexNet features Query Cross-view
More informationREINFORCEMENT LEARNING (DD3359) O-03 END-TO-END LEARNING
REINFORCEMENT LEARNING (DD3359) O-03 END-TO-END LEARNING RIKA ANTONOVA ANTONOVA@KTH.SE ALI GHADIRZADEH ALGH@KTH.SE RL: What We Know So Far Formulate the problem as an MDP (or POMDP) State space captures
More informationCollaborative Compressive Sensing based Dynamic Spectrum Sensing and Mobile Primary User Localization in Cognitive Radio Networks
Collaborative Compressive Sensing based Dynamic Spectrum Sensing and Mobile Primary User Localization in Cognitive Radio Networks Lanchao Liu and Zhu Han ECE Department University of Houston Houston, Texas
More informationEffects of Basis-mismatch in Compressive Sampling of Continuous Sinusoidal Signals
Effects of Basis-mismatch in Compressive Sampling of Continuous Sinusoidal Signals Daniel H. Chae, Parastoo Sadeghi, and Rodney A. Kennedy Research School of Information Sciences and Engineering The Australian
More informationINFORMATION about image authenticity can be used in
1 Constrained Convolutional Neural Networs: A New Approach Towards General Purpose Image Manipulation Detection Belhassen Bayar, Student Member, IEEE, and Matthew C. Stamm, Member, IEEE Abstract Identifying
More information11/13/18. Introduction to RNNs for NLP. About Me. Overview SHANG GAO
Introduction to RNNs for NLP SHANG GAO About Me PhD student in the Data Science and Engineering program Took Deep Learning last year Work in the Biomedical Sciences, Engineering, and Computing group at
More informationFundamentals of Wireless Communication
Communication Technology Laboratory Prof. Dr. H. Bölcskei Sternwartstrasse 7 CH-8092 Zürich Fundamentals of Wireless Communication Homework 5 Solutions Problem 1 Simulation of Error Probability When implementing
More informationN. Garcia, A.M. Haimovich, J.A. Dabin and M. Coulon
N. Garcia, A.M. Haimovich, J.A. Dabin and M. Coulon Goal: Localization (geolocation) of RF emitters in multipath environments Challenges: Line-of-sight (LOS) paths Non-line-of-sight (NLOS) paths Blocked
More informationAdvanced Techniques for Mobile Robotics Location-Based Activity Recognition
Advanced Techniques for Mobile Robotics Location-Based Activity Recognition Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz Activity Recognition Based on L. Liao, D. J. Patterson, D. Fox,
More informationM.Tolotti - Mathematics (Preparatory) - September Exercises. Maximize p(g(x))g(x) q x subject to x R +
M.Tolotti - Mathematics (Preparatory) - September 2010 1 Exercises EXERCISE 1. where Maximize p(g(x))g(x) q x subject to x R + p : R R is constant, i.e. p(g(x)) = p = 1 for all x. g(x) = 35x x 2. q = 10.
More informationHybrid Discriminative/Class-Specific Classifiers for Narrow-Band Signals
To appear IEEE Trans. on Aerospace and Electronic Systems, October 2007. Hybrid Discriminative/Class-Specific Classifiers for Narrow-Band Signals Brian F. Harrison and Paul M. Baggenstoss Naval Undersea
More informationAttention-based Information Fusion using Multi-Encoder-Decoder Recurrent Neural Networks
Attention-based Information Fusion using Multi-Encoder-Decoder Recurrent Neural Networks Stephan Baier1, Sigurd Spieckermann2 and Volker Tresp1,2 1- Ludwig Maximilian University Oettingenstr. 67, Munich,
More informationLecture 15. Global extrema and Lagrange multipliers. Dan Nichols MATH 233, Spring 2018 University of Massachusetts
Lecture 15 Global extrema and Lagrange multipliers Dan Nichols nichols@math.umass.edu MATH 233, Spring 2018 University of Massachusetts March 22, 2018 (2) Global extrema of a multivariable function Definition
More informationHigh Resolution Radar Sensing via Compressive Illumination
High Resolution Radar Sensing via Compressive Illumination Emre Ertin Lee Potter, Randy Moses, Phil Schniter, Christian Austin, Jason Parker The Ohio State University New Frontiers in Imaging and Sensing
More informationarxiv: v1 [cs.ce] 9 Jan 2018
Predict Forex Trend via Convolutional Neural Networks Yun-Cheng Tsai, 1 Jun-Hao Chen, 2 Jun-Jie Wang 3 arxiv:1801.03018v1 [cs.ce] 9 Jan 2018 1 Center for General Education 2,3 Department of Computer Science
More informationTMA4155 Cryptography, Intro
Trondheim, December 12, 2006. TMA4155 Cryptography, Intro 2006-12-02 Problem 1 a. We need to find an inverse of 403 modulo (19 1)(31 1) = 540: 540 = 1 403 + 137 = 17 403 50 540 + 50 403 = 67 403 50 540
More informationNeural Network Part 4: Recurrent Neural Networks
Neural Network Part 4: Recurrent Neural Networks Yingyu Liang Computer Sciences 760 Fall 2017 http://pages.cs.wisc.edu/~yliang/cs760/ Some of the slides in these lectures have been adapted/borrowed from
More informationarxiv: v1 [cs.ne] 5 Feb 2014
LONG SHORT-TERM MEMORY BASED RECURRENT NEURAL NETWORK ARCHITECTURES FOR LARGE VOCABULARY SPEECH RECOGNITION Haşim Sak, Andrew Senior, Françoise Beaufays Google {hasim,andrewsenior,fsb@google.com} arxiv:12.1128v1
More informationREAL TIME EMULATION OF PARAMETRIC GUITAR TUBE AMPLIFIER WITH LONG SHORT TERM MEMORY NEURAL NETWORK
REAL TIME EMULATION OF PARAMETRIC GUITAR TUBE AMPLIFIER WITH LONG SHORT TERM MEMORY NEURAL NETWORK Thomas Schmitz and Jean-Jacques Embrechts 1 1 Department of Electrical Engineering and Computer Science,
More informationMeme Tracking. Abhilash Chowdhary CS-6604 Dec. 1, 2015
Meme Tracking Abhilash Chowdhary CS-6604 Dec. 1, 2015 Overview Introduction Information Spread Meme Tracking Part 1 : Rise and Fall Patterns of Information Diffusion: Model and Implications Part 2 : NIFTY:
More informationNEURAL NETWORK BASED MAXIMUM POWER POINT TRACKING
NEURAL NETWORK BASED MAXIMUM POWER POINT TRACKING 3.1 Introduction This chapter introduces concept of neural networks, it also deals with a novel approach to track the maximum power continuously from PV
More informationLearning Deep Networks from Noisy Labels with Dropout Regularization
Learning Deep Networks from Noisy Labels with Dropout Regularization Ishan Jindal, Matthew Nokleby Electrical and Computer Engineering Wayne State University, MI, USA Email: {ishan.jindal, matthew.nokleby}@wayne.edu
More informationFiltering. Image Enhancement Spatial and Frequency Based
Filtering Image Enhancement Spatial and Frequency Based Brent M. Dingle, Ph.D. 2015 Game Design and Development Program Mathematics, Statistics and Computer Science University of Wisconsin - Stout Lecture
More informationHeads-up Limit Texas Hold em Poker Agent
Heads-up Limit Texas Hold em Poker Agent Nattapoom Asavareongchai and Pin Pin Tea-mangkornpan CS221 Final Project Report Abstract Our project aims to create an agent that is able to play heads-up limit
More informationDeconvolution , , Computational Photography Fall 2017, Lecture 17
Deconvolution http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 17 Course announcements Homework 4 is out. - Due October 26 th. - There was another
More informationITERATIVE decoding of classic codes has created much
IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 7, JULY 2009 1 Improved Random Redundant Iterative HDPC Decoding Ilan Dimnik, and Yair Be ery, Senior Member, IEEE Abstract An iterative algorithm for
More information