Understanding egocentric imagery, for fun and science
|
|
- Julius Nelson
- 5 years ago
- Views:
Transcription
1 Understanding egocentric imagery, for fun and science David Crandall School of Informa-cs and Compu-ng Indiana University Joint work with: Denise Anthony (Dartmouth), Apu Kapadia, Chen Yu; PhD Students: Sven Bambach, Roberto Hoyle, Mohammed Korayem, Stefan Lee, Robert Templeman; Undergrads: Steven Armes, Dennis Chen
2 - Google - Narra-ve - Autographer - Samsung - GoPro
3
4 June 2, 11:10 am July 2, 11:10 am Aug 2, 11:10 am Sept 2, 11:10 am
5
6 Daguerreotypes, 1839 The Kodak, 1888 Polaroid Land Camera, 1948 Digital camera, 1975 J-phone, 2000
7 Daguerreotypes, 1839 The Kodak, 1888 Polaroid Land Camera, 1948 Digital camera, 1975 J-phone, 2000
8
9
10 What if your device were hacked? Robert Templeman, Zahid Rahman, David Crandall, Apu Kapadia. PlaceRaider: Virtual The_ in Physical Spaces with Smartphones. NDSS, 2013.
11 R. Templeman, Z. Rahman, D. Crandall, A. Kapadia. PlaceRaider: Virtual The_ in Physical Spaces with Smartphones. NDSS, 2013.
12 Vision to the rescue! Could computer vision automa-cally help Organize and analysis egocentric image streams? Find the great photos amongst all the bad ones? Warn before sharing a photo with private data to Facebook? Block or censor private photos from being taken by the device, and/or uploaded to the cloud?
13 What makes an image sensi-ve? 37 undergrads, wearing life-logging cameras for a week, each day reviewing images and labeling them in various ways. R. Hoyle, R. Templeman, S. Armes, D. Anthony, D. Crandall, A. Kapadia, "Privacy behaviors of lifeloggers using wearable cameras," UBICOMP, 2014.
14 Ethical, legal, IRB considera-ons
15 Why not share the photo? Reason Responses No good reason to share it 36.0% Objects (other than people) in the photo 30.7% Where this photo was taken 22.6% People within the photo 18.4% Participant was in the photo 11.5% It had private information 11.5% It would have been embarrassing to share it 5.4% It would have violated someone else s privacy 3.8% It was a bad photo 1.5%
16 Why not share the photo? Reason Responses No good reason to share it 36.0% Objects (other than people) in the photo 30.7% Where this photo was taken 22.6% People within the photo 18.4% Participant was in the photo 11.5% It had private information 11.5% It would have been embarrassing to share it 5.4% It would have violated someone else s privacy 3.8% It was a bad photo 1.5% 16
17 Why not share the photo? Reason Responses No good reason to share it 36.0% Objects (other than people) in the photo 30.7% Where this photo was taken 22.6% People within the photo 18.4% Participant was in the photo 11.5% It had private information 11.5% It would have been embarrassing to share it 5.4% It would have violated someone else s privacy 3.8% It was a bad photo 1.5% 17
18 Why not share the photo? Reason Responses No good reason to share it 36.0% Objects (other than people) in the photo 30.7% Where this photo was taken 22.6% People within the photo 18.4% Participant was in the photo 11.5% It had private information 11.5% It would have been embarrassing to share it 5.4% It would have violated someone else s privacy 3.8% It was a bad photo 1.5% 18
19 Why not share the photo? Reason Responses No good reason to share it 36.0% Objects (other than people) in the photo 30.7% Where this photo was taken 22.6% People within the photo 18.4% Participant was in the photo 11.5% It had private information 11.5% It would have been embarrassing to share it 5.4% It would have violated someone else s privacy 3.8% It was a bad photo 1.5% 19
20 Place recogni-on in lifelogging images Given a stream of lifelogging photos, where was each photo taken? In which specific room? In which type of room? Use SVMs with standard image features HOG, GIST, LBP, SIFT, R. Templeman, M. Korayem, D. Crandall, A. Kapadia. PlaceAvoider: Steering first-person cameras away from sensi-ve spaces, NDSS 2014.
21 Classifying photo streams with HMMs ProbabiliHes with individual photo classifiers: Bathroom: Bedroom: Garage: Living: Office: ProbabiliHes aker applying HMM: Bathroom: Bedroom: Garage: Living: Office:
22 Evalua-on Tested on 5 realis-c lifelogging datasets Dataset Baseline Local features + HMM Global features + HMM Local+global features + HMM Local+global +human + HMM House % 89.2% 64.0% 89.2% 95.0% House % 55.0% 56.4% 74.6% 76.8% House % 97.4% 86.9% 98.7% 99.8% Workplace % 75.5% 89.2% 87.7% 91.0% Workplace % 92.3% 81.2% 98.7% 100.0% Average 28.5% 81.9% 74.8% 89.8% 92.5%
23 Sample results
24 Detec-ng screens CNNs, plus temporal smoothing, gives ~95% 2-way (monitor vs nomonitor) performance on real lifelogging data (vs ~73% baseline). M. Korayem, R. Templeman, D. Chen, D. Crandall, A. Kapadia. ScreenAvoider: Protec-ng computer screens from ubiquitous cameras. (Arxiv preprint), 2014.
25 Vannevar Bush, The Atlan:c, 1945
26 child s first person view parent s first person view from eye camera from eye camera from head camera from head camera
27 (Very broad) Research ques-ons How do children coordinate their hand movement, eye gaze, and head movements? How do these develop and change over -me? Are certain parerns predic-ve of later deficiencies? How do children and parents interact, and how does this differ across subjects? How do they jointly coordinate aren-on? Which interac-on parerns are most successful for learning?
28 Typical experiments 13 child-parent dyads (childrens mean age = 13 months, σ = 3.2 months) Parents told to engage child with toys, naturally, e.g.: Exchanging toys back and forth Joint ac-ons with toys ~5 minutes of video per trial In lab sexng, more recently in naturalis-c environments
29 Analysis Video data processing: Pixel-level visual saliency es-ma-on Saliency map models e.g. of Ix et al Object segmenta-on and recogni-on Posi-ons of toys, hands, and faces Object holding ac-vi-es MoHon data processing: Op-cal flow Head, hand iner-al sensors
30 Eye gaze within the visual field Children μ = (340, 231) σ x = 86, σ y = 65 N = 148,279 Parents μ = (361, 224) σ x = 53, σ y = 60 N = 148,279 S. Bambach, D. Crandall, C. Yu. Understanding embodied visual aren-on in child-parent interac-ons, IEEE Interna:onal Conference on Development and Learning, 2013.
31 Head-eye coordina-on Moving Head StaHonary Head Children µ = (333, 253) σ x = 80, σ y = 70 N = 15,626 µ = (335, 226) σ x = 84, σ y = 63 N = 71,388 Parents µ = (351, 229) σ x = 57, σ y = 62 N = 63,035 µ = (374, 210) σ x = 47, σ y = 56 N = 33,981
32 Eye-hand coordina-on Empty Hands Holding Toy Children µ = (332, 231) σ x = 84, σ y = 68 N = 55,704 µ = (301, 242) σ x = 85, σ y = 62 N = 10,937 Parents µ = (336, 244) σ x = 71, σ y = 70 N = 27,062 µ = (333, 270) σ x = 63, σ y = 73 N = 11,239
33 Saliency in first-person views Comparison of average saliency (based on 148,000 frames each, using Ix et al) à No significant difference in child and parent views Child Parent
34 Saliency in first-person views Comparison of average saliency within hotspot around gaze loca-on hotspot à Gaze predic-veness differs significantly Child Parent
35 Longer-term ques-on Can we jointly model head and hand pose, eye gaze, saliency, and ac-vity, both to berer perform egocentric computer vision, and to help explain human vision? Head Eye Hand
36 Why start with hands? Hands are in nearly every frame of egocentric video Hand configura-on reflects what we are doing and what we are paying aren-on to Detec-ng hands is a fundamental problem for both computers and people The feeling of ownership of our limbs is a fundamental aspect of self-consciousness. [Ehrsson 2004]
37 A. Fathi, X. Ren, J. Rehg, "Learning to recognize objects in egocentric activities," CVPR Pirsiavash, Hamed, and Ramanan. "Detecting activities of daily living in first-person camera views." CVPR 2012 A. Fathi, A. Farhadi, and J. Rehg. Understanding egocentric activities, ICCV 2011
38 Hand detec-on and disambigua-on In egocentric video of interac-ng people, locate: The observer s hands (my le_, my right) The other person s hands (your le_, your right) The other person s head (your head) Our approaches so far: Strong temporal models, weak spa-al/appearance models S. Bambach, J. Franchak, D. Crandall, C. Yu. Detec-ng hands in children s egocentric views to understand embodied aren-on during social interac-on. CogSci Strong spa-al/temporal models, weak appearance models, explicit camera (head) mo-on model S. Lee, S. Bambach, C. Yu, D. Crandall. This hand is my hand: A probabilis-c approach to hand disambigua-on in egocentric video, CVPR Egovision, Strong appearance models, weak spa-al/temporal models S. Bambach, S. Lee, D. Crandall, C. Yu. Lending a hand: Detec-ng hands and recognizing ac-vi-es in complex egocentric interac-ons. ICCV, 2015.
39 Strong appearance models: CNNs (of course) Mostly off-the-shelf Caffe, fine-tuned from ImageNet Generate candidates using domain-specific informa-on Sample from distribu-on over size, posi-on, shape of hand regions in training data, biased by skin color detector Other s right hand My le_ hand Other s le_ hand My right hand Coverage Direct Sampling Direct Sampling (only spatial) Sampling from Selective Search Selective Search Sampling from Objectness Objectness # Window Proposals
40 Hand detec-on & disambigua-on results 4 actors x 4 ac-vi-es x 3 loca-ons = 48 unique videos 15,053 pixel-level ground truth segmenta-ons Precision Precision ours AP: AUC: SS AP: AUC: obj. AP: AUC: Recall own left AP: AUC: own right AP: AUC: other left AP: AUC: other right AP: AUC: Recall
41 Hand detec-on & segmenta-on
42
43 Does hand pose alone reveal first person ac-vi-es? Different interac-ons afford different hand grasps [Napier 1965] Train and test CNNs for 4 ac-vi-es (puzzle, jenga, cards, chess) on masked out images Single frame accuracy: 66.4% with GT masks, 50.9% with automa-c masks (vs 25% baseline) With temporal cues (50 frames): 92.9% with GT masks, 73.4% with automa-c masks
44 Hand PosiHons during Sustained ASenHon, within child's field of view Parent s Right Hand Parent s Left Hand Child s Left Hand Child s Right Hand 6 child-parent dyads (children s mean age = 19 months, σ = 2.56 months) 67,913 frames (~38 minutes) of child-view video
45 Daguerreotypes, 1839 The Kodak, 1888 Polaroid Land Camera, 1948 Digital camera, 1975 J-phone, 2000
46 Future work Detec-ng and disambigua-ng hands Generalize hand-based ac-vity recogni-on to more ac-vi-es, finer-grained ac-ons More challenging social situa-ons (e.g. more than two interac-ng people, moving camera wearer) Applica-ons for egocentric video data
47 For more informahon: hsp://vision.soic.indiana.edu/ Based on joint work with: Faculty: Denise Anthony, Apu Kapadia, Chen Yu PhD Students: Sven Bambach, Roberto Hoyle, Mohammed Korayem, Stefan Lee, Robert Templeman Undergrad students: Steven Armes, Dennis Chen Sponsors: NSF (III CAREER, SaTC, EAGER, DIBBs), Intelligence Advanced Research Projects Ac-vity (IARPA), Air Force Office of Scien-fic Research (AFOSR), Google, Nvidia, IU FRSP, IUCRG, IU D2I Center, Lily Endowment
Exploring Wearable Cameras for Educational Purposes
70 Exploring Wearable Cameras for Educational Purposes Jouni Ikonen and Antti Knutas Abstract: The paper explores the idea of using wearable cameras in educational settings. In the study, a wearable camera
More informationarxiv: v1 [cs.cr] 28 Nov 2014
ScreenAvoider: Protecting Computer Screens from Ubiquitous Cameras Mohammed Korayem, Robert Templeman, Dennis Chen, David Crandall, Apu Kapadia arxiv:1412.0008v1 [cs.cr] 28 Nov 2014 School of Informatics
More informationIdentifying First-person Camera Wearers in Third-person Videos
Identifying First-person Camera Wearers in Third-person Videos Chenyou Fan 1, Jangwon Lee 1, Mingze Xu 1, Krishna Kumar Singh 2, Yong Jae Lee 2, David J. Crandall 1 and Michael S. Ryoo 1 1 Indiana University
More informationWearable Sensing for Understanding, Forecasting and Assisting Human Activity. Kris Kitani Assistant Research Professor Carnegie Mellon University
Wearable Sensing for Understanding, Forecasting and Assisting Human Activity Kris Kitani Assistant Research Professor Carnegie Mellon University camera Wearable sensing camera smartphone camera Wearable
More informationGoing Deeper into First-Person Activity Recognition
Going Deeper into First-Person Activity Recognition Minghuang Ma, Haoqi Fan and Kris M. Kitani Carnegie Mellon University Pittsburgh, PA 15213, USA minghuam@andrew.cmu.edu haoqif@andrew.cmu.edu kkitani@cs.cmu.edu
More informationPresen ng the Early Warning Systems: MicroMesh + GlobalMesh
Presen ng the Early Warning Systems: MicroMesh + GlobalMesh RF Detection: A Technology Brief concerning the detection of Drones Above + Phones Below, and everything in-between HSS Development 2 The Early
More informationDeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition. ECE 289G: Paper Presentation #3 Philipp Gysel
DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition ECE 289G: Paper Presentation #3 Philipp Gysel Autonomous Car ECE 289G Paper Presentation, Philipp Gysel Slide 2 Source: maps.google.com
More informationFirst Person Action Recognition Using Deep Learned Descriptors
First Person Action Recognition Using Deep Learned Descriptors Suriya Singh 1 Chetan Arora 2 C. V. Jawahar 1 1 IIIT Hyderabad, India 2 IIIT Delhi, India Abstract We focus on the problem of wearer s action
More informationContinuous Gesture Recognition Fact Sheet
Continuous Gesture Recognition Fact Sheet August 17, 2016 1 Team details Team name: ICT NHCI Team leader name: Xiujuan Chai Team leader address, phone number and email Address: No.6 Kexueyuan South Road
More informationCOACHES COopera-ve Autonomous Robots in Complex and Human EnvironmentS
COACHES COopera-ve Autonomous Robots in Complex and Human EnvironmentS Coordinator : Abdel- Illah Mouaddib Partners : University of Caen Basse- Normandie, France University of Sapienza Roma, Italy Vrije
More informationDriving Using End-to-End Deep Learning
Driving Using End-to-End Deep Learning Farzain Majeed farza@knights.ucf.edu Kishan Athrey kishan.athrey@knights.ucf.edu Dr. Mubarak Shah shah@crcv.ucf.edu Abstract This work explores the problem of autonomously
More informationThe Distributed Camera
The Distributed Camera Noah Snavely Cornell University Microsoft Faculty Summit June 16, 2013 The Age of Exapixel Image Data Over a trillion photos available online Millions uploaded every hour Interconnected
More informationA Fuller Understanding of Fully Convolutional Networks. Evan Shelhamer* Jonathan Long* Trevor Darrell UC Berkeley in CVPR'15, PAMI'16
A Fuller Understanding of Fully Convolutional Networks Evan Shelhamer* Jonathan Long* Trevor Darrell UC Berkeley in CVPR'15, PAMI'16 1 pixels in, pixels out colorization Zhang et al.2016 monocular depth
More informationAutomatic understanding of the visual world
Automatic understanding of the visual world 1 Machine visual perception Artificial capacity to see, understand the visual world Object recognition Image or sequence of images Action recognition 2 Machine
More informationCROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS. Kuan-Chuan Peng and Tsuhan Chen
CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS Kuan-Chuan Peng and Tsuhan Chen Cornell University School of Electrical and Computer Engineering Ithaca, NY 14850
More informationSemantic Localization of Indoor Places. Lukas Kuster
Semantic Localization of Indoor Places Lukas Kuster Motivation GPS for localization [7] 2 Motivation Indoor navigation [8] 3 Motivation Crowd sensing [9] 4 Motivation Targeted Advertisement [10] 5 Motivation
More informationColorful Image Colorizations Supplementary Material
Colorful Image Colorizations Supplementary Material Richard Zhang, Phillip Isola, Alexei A. Efros {rich.zhang, isola, efros}@eecs.berkeley.edu University of California, Berkeley 1 Overview This document
More informationVideo Object Segmentation with Re-identification
Video Object Segmentation with Re-identification Xiaoxiao Li, Yuankai Qi, Zhe Wang, Kai Chen, Ziwei Liu, Jianping Shi Ping Luo, Chen Change Loy, Xiaoou Tang The Chinese University of Hong Kong, SenseTime
More informationTOWARDS A UNIFIED FRAMEWORK FOR HAND-BASED METHODS IN FIRST PERSON VISION.
TOWARDS A UNFED FRAMEWORK FOR HAND-BASED METHODS N FRST PERSON VSON. Alejandro Betancourt 1,2, Pietro Morerio 1, Lucio Marcenaro 1, Emilia Barakova 2, Matthias Rauterberg 2, Carlo Regazzoni 1 1 nformation
More informationAn FRN is a unique number used to identify an applicant in all transactions with the FCC. If you don t have an FRN
How-to Guide: Registering Receive-Only Antennas FCC Form 312 and Schedule B Disclaimer: The materials provided herein are for informa onal purposes only and do not cons tute legal advice. You should contact
More informationFIREBREAK. Installa on Guide FIREBREAK. USA Distribu on by Mei Guo Interna onal, LLC USA - Interna onal Patent Pending.
FIREBREAK USA Distribu on by Mei Guo Interna onal, LLC USA - Interna onal Patent Pending (1) Product Informa on Proprietary product made from a blend of high purity ceramic, alumina and zirconia cyrstal
More informationFace detection, face alignment, and face image parsing
Lecture overview Face detection, face alignment, and face image parsing Brandon M. Smith Guest Lecturer, CS 534 Monday, October 21, 2013 Brief introduction to local features Face detection Face alignment
More informationRecognizing Micro-Actions and Reactions from Paired Egocentric Videos
Recognizing Micro-Actions and Reactions from Paired Egocentric Videos Ryo Yonetani The University of Tokyo Tokyo, Japan yonetani@iis.u-tokyo.ac.jp Kris M. Kitani Carnegie Mellon University Pittsburgh,
More informationTELLING STORIES OF VALUE WITH IOT DATA
TELLING STORIES OF VALUE WITH IOT DATA VISUALIZATION BAREND BOTHA VIDEO TRANSCRIPT Tell me a little bit about yourself and your background in IoT. I came from a web development and design background and
More informationMachine Learning for Intelligent Transportation Systems
Machine Learning for Intelligent Transportation Systems Patrick Emami (CISE), Anand Rangarajan (CISE), Sanjay Ranka (CISE), Lily Elefteriadou (CE) MALT Lab, UFTI September 6, 2018 ITS - A Broad Perspective
More informationVishnu Nath. Usage of computer vision and humanoid robotics to create autonomous robots. (Ximea Currera RL04C Camera Kit)
Vishnu Nath Usage of computer vision and humanoid robotics to create autonomous robots (Ximea Currera RL04C Camera Kit) Acknowledgements Firstly, I would like to thank Ivan Klimkovic of Ximea Corporation,
More informationVision-based User-interfaces for Pervasive Computing. CHI 2003 Tutorial Notes. Trevor Darrell Vision Interface Group MIT AI Lab
Vision-based User-interfaces for Pervasive Computing Tutorial Notes Vision Interface Group MIT AI Lab Table of contents Biographical sketch..ii Agenda..iii Objectives.. iv Abstract..v Introduction....1
More informationMulti-task Learning of Dish Detection and Calorie Estimation
Multi-task Learning of Dish Detection and Calorie Estimation Department of Informatics, The University of Electro-Communications, Tokyo 1-5-1 Chofugaoka, Chofu-shi, Tokyo 182-8585 JAPAN ABSTRACT In recent
More informationMitigating Bystander Privacy Concerns in Egocentric Activity Recognition with Deep Learning and Intentional Image Degradation
Mitigating Bystander Privacy Concerns in Egocentric Activity Recognition with Deep Learning and Intentional Image Degradation MARIELLA DIMICCOLI*, University of Barcelona and Computer Vision Center, Spain
More informationDOTTORATO DI RICERCA IN INFORMATICA IX CICLO UNIVERSITA DEGLI STUDI DI SALERNO. Forensic Analysis for Digital Images.
DOTTORATO DI RICERCA IN INFORMATICA IX CICLO UNIVERSITA DEGLI STUDI DI SALERNO Forensic Analysis for Digital Images Maurizio Cembalo November, 2010 PhD Program Chair Prof.ssa Margherita Napoli Supervisor
More informationKrishnaCam: Using a Longitudinal, Single-Person, Egocentric Dataset for Scene Understanding Tasks
KrishnaCam: Using a Longitudinal, Single-Person, Egocentric Dataset for Scene Understanding Tasks Krishna Kumar Singh 1,3 Kayvon Fatahalian 1 Alexei A. Efros 2 1 Carnegie Mellon University 2 UC Berkeley
More informationContext Aware Computing
Context Aware Computing Context aware computing: the use of sensors and other sources of information about a user s context to provide more relevant information and services Context independent: acts exactly
More information6. SENSE OF COMMUNITY INTRODUCTION. Direction for community connections, equity and inclusiveness, and culture and identity.
6. SENSE OF COMMUNITY Direction for community connections, equity and inclusiveness, and culture and identity. INTRODUCTION The concept of a sense of community is both abstract, and very relevant. It is
More informationICARP III. 3 rd International Conference on Arctic Research Planning (ICARP III) Volker Rachold International Arctic Science Committee
3 rd International Conference on Arctic Research Planning (ICARP III) Volker Rachold International Arctic Science Committee ISIRA Toyama (Japan), 25 April 2015 FARO provides a framework to help: identify
More informationCurriculum Vitae. Computer Vision, Image Processing, Biometrics. Computer Vision, Vision Rehabilitation, Vision Science
Curriculum Vitae Date Prepared: 01/09/2016 (last updated: 09/12/2016) Name: Shrinivas J. Pundlik Education 07/2002 B.E. (Bachelor of Engineering) Electronics Engineering University of Pune, Pune, India
More informationPerceptual Characters of Photorealistic See-through Vision in Handheld Augmented Reality
Perceptual Characters of Photorealistic See-through Vision in Handheld Augmented Reality Arindam Dey PhD Student Magic Vision Lab University of South Australia Supervised by: Dr Christian Sandor and Prof.
More informationAVA: A Large-Scale Database for Aesthetic Visual Analysis
1 AVA: A Large-Scale Database for Aesthetic Visual Analysis Wei-Ta Chu National Chung Cheng University N. Murray, L. Marchesotti, and F. Perronnin, AVA: A Large-Scale Database for Aesthetic Visual Analysis,
More informationComputer Vision, Lecture 3
Computer Vision, Lecture 3 Professor Hager http://www.cs.jhu.edu/~hager /4/200 CS 46, Copyright G.D. Hager Outline for Today Image noise Filtering by Convolution Properties of Convolution /4/200 CS 46,
More informationMobile Cognitive Indoor Assistive Navigation for the Visually Impaired
1 Mobile Cognitive Indoor Assistive Navigation for the Visually Impaired Bing Li 1, Manjekar Budhai 2, Bowen Xiao 3, Liang Yang 1, Jizhong Xiao 1 1 Department of Electrical Engineering, The City College,
More informationVirtual Reality Based Scalable Framework for Travel Planning and Training
Virtual Reality Based Scalable Framework for Travel Planning and Training Loren Abdulezer, Jason DaSilva Evolving Technologies Corporation, AXS Lab, Inc. la@evolvingtech.com, jdasilvax@gmail.com Abstract
More informationDomain Adaptation & Transfer: All You Need to Use Simulation for Real
Domain Adaptation & Transfer: All You Need to Use Simulation for Real Boqing Gong Tecent AI Lab Department of Computer Science An intelligent robot Semantic segmentation of urban scenes Assign each pixel
More informationFace Registration Using Wearable Active Vision Systems for Augmented Memory
DICTA2002: Digital Image Computing Techniques and Applications, 21 22 January 2002, Melbourne, Australia 1 Face Registration Using Wearable Active Vision Systems for Augmented Memory Takekazu Kato Takeshi
More informationEthical issues raised by big data and real world evidence projects. Dr Andrew Turner
Ethical issues raised by big data and real world evidence projects Dr Andrew Turner andrew.turner@oii.ox.ac.uk December 8, 2017 What is real world evidence and big data? Real world evidence is evidence
More informationLiangliang Cao *, Jiebo Luo +, Thomas S. Huang *
Annotating ti Photo Collections by Label Propagation Liangliang Cao *, Jiebo Luo +, Thomas S. Huang * + Kodak Research Laboratories *University of Illinois at Urbana-Champaign (UIUC) ACM Multimedia 2008
More informationEgocentric Field-of-View Localization Using First-Person Point-of-View Devices
Egocentric Field-of-View Localization Using First-Person Point-of-View Devices Vinay Bettadapura 1,2 vinay@gatech.edu Irfan Essa 1,2 irfan@cc.gatech.edu 1 Google Inc., Mountain View, CA, USA Caroline Pantofaru
More informationA Spatiotemporal Approach for Social Situation Recognition
A Spatiotemporal Approach for Social Situation Recognition Christian Meurisch, Tahir Hussain, Artur Gogel, Benedikt Schmidt, Immanuel Schweizer, Max Mühlhäuser Telecooperation Lab, TU Darmstadt MOTIVATION
More informationIEEE SmartAg Ini-a-ve: Technology applied to the food supply chain John Verboncoeur, Chair IEEE SmartAg Ini>a>ve Michigan State University
IEEE SmartAg Ini-a-ve: Technology applied to the food supply chain John Verboncoeur, Chair IEEE SmartAg Ini>a>ve Michigan State University 1 IEEE SmartAg Ini-a-ve Smart Technologies and Innovations Applied
More informationLecture 7: Scene Text Detection and Recognition. Dr. Cong Yao Megvii (Face++) Researcher
Lecture 7: Scene Text Detection and Recognition Dr. Cong Yao Megvii (Face++) Researcher yaocong@megvii.com Outline Background and Introduction Conventional Methods Deep Learning Methods Datasets and Competitions
More informationDetection and Segmentation. Fei-Fei Li & Justin Johnson & Serena Yeung. Lecture 11 -
Lecture 11: Detection and Segmentation Lecture 11-1 May 10, 2017 Administrative Midterms being graded Please don t discuss midterms until next week - some students not yet taken A2 being graded Project
More informationPersonal Driving Diary: Constructing a Video Archive of Everyday Driving Events
Proceedings of IEEE Workshop on Applications of Computer Vision (WACV), Kona, Hawaii, January 2011 Personal Driving Diary: Constructing a Video Archive of Everyday Driving Events M. S. Ryoo, Jae-Yeong
More informationGESTURE RECOGNITION WITH 3D CNNS
April 4-7, 2016 Silicon Valley GESTURE RECOGNITION WITH 3D CNNS Pavlo Molchanov Xiaodong Yang Shalini Gupta Kihwan Kim Stephen Tyree Jan Kautz 4/6/2016 Motivation AGENDA Problem statement Selecting the
More informationEffects of the Unscented Kalman Filter Process for High Performance Face Detector
Effects of the Unscented Kalman Filter Process for High Performance Face Detector Bikash Lamsal and Naofumi Matsumoto Abstract This paper concerns with a high performance algorithm for human face detection
More informationPart I New Sensing Technologies for Societies and Environment
Part I New Sensing Technologies for Societies and Environment Introduction New ICT-Mediated Sensing Opportunities Andreas Hotho, Gerd Stumme, and Jan Theunis During the last century, the application of
More informationMobile Crowdsensing enabled IoT frameworks: harnessing the power and wisdom of the crowd
Mobile Crowdsensing enabled IoT frameworks: harnessing the power and wisdom of the crowd Malamati Louta Konstantina Banti University of Western Macedonia OUTLINE Internet of Things Mobile Crowd Sensing
More informationUnderstanding User Privacy in Internet of Things Environments IEEE WORLD FORUM ON INTERNET OF THINGS / 30
Understanding User Privacy in Internet of Things Environments HOSUB LEE AND ALFRED KOBSA DONALD BREN SCHOOL OF INFORMATION AND COMPUTER SCIENCES UNIVERSITY OF CALIFORNIA, IRVINE 2016-12-13 IEEE WORLD FORUM
More informationRUGGED. MARINIZED. LOW MAINTENANCE.
RUGGED. MARINIZED. LOW MAINTENANCE. MWIR LWIR SWIR NIGHT VISION DAY / LOW LIGHT LASER DAZZLER / LRF FULL SPECTRUM EO / IR SYSTEMS Series NN 1000 NN 2000 NN 6000 NN 6000 NN 7000 MODEL NN 1045 NN HSC NN
More informationReal Time and Non-intrusive Driver Fatigue Monitoring
Real Time and Non-intrusive Driver Fatigue Monitoring Qiang Ji and Zhiwei Zhu jiq@rpi rpi.edu Intelligent Systems Lab Rensselaer Polytechnic Institute (RPI) Supported by AFOSR and Honda Introduction Motivation:
More informationNkxg"Ncwij"Nqxg"Ctv 423: Iudqfklvh#Rssruwxqlw #Sdfndjh. Rtqrtkgvct{"("Eqphkfgpvkcn"Nkxg"Ncwij"Nqxg"Ctv."NNE"423:
Nkxg"Ncwij"Nqxg"Ctv 423: Iudqfklvh#Rssruwxqlw #Sdfndjh Rtqrtkgvct{"("Eqphkfgpvkcn"Nkxg"Ncwij"Nqxg"Ctv."NNE"423: Our Story What We Do Our Culture In 2010, during one of the most devasta ng mes in our economy,
More informationarxiv: v1 [cs.cv] 27 Nov 2016
Real-Time Video Highlights for Yahoo Esports arxiv:1611.08780v1 [cs.cv] 27 Nov 2016 Yale Song Yahoo Research New York, USA yalesong@yahoo-inc.com Abstract Esports has gained global popularity in recent
More informationCS688/WST665 Student presentation Learning Fine-grained Image Similarity with Deep Ranking CVPR Gayoung Lee ( 이가영 )
CS688/WST665 Student presentation Learning Fine-grained Image Similarity with Deep Ranking CVPR 2014 Gayoung Lee ( 이가영 ) Contents 1. Background knowledge 2. Proposed method 3. Experimental Result 4. Conclusion
More informationUNDERGRADUATE PROJECT REPORT
UNIVERSITAT POLITÈCNICA DE CATALUNYA ESEIAAT EgoMon Gaze and Video Dataset for Visual Saliency Prediction June 2016 UNDERGRADUATE PROJECT REPORT Supervised by Xavier Giró and Cathal Gurrin Acknowledgements
More informationMaking a Difference in 2017 IPOs & CES
SHAREHOLDER UPDATE JANUARY 2017 Making a Difference in 2017 IPOs & CES By Tom Astle IPOs - Will 2017 finally be the year? Our portfolio strategy of investing in later-stage growth companies was partially
More informationYou don t want to be the next meme : College Students Workarounds to Manage Privacy in the Era of Pervasive Photography
You don t want to be the next meme : College Students Workarounds to Manage Privacy in the Era of Pervasive Photography Yasmeen Rashidi, Tousif Ahmed, Felicia Patel, Emily Fath, Apu Kapadia, Christena
More informationLearning Pixel-Distribution Prior with Wider Convolution for Image Denoising
Learning Pixel-Distribution Prior with Wider Convolution for Image Denoising Peng Liu University of Florida pliu1@ufl.edu Ruogu Fang University of Florida ruogu.fang@bme.ufl.edu arxiv:177.9135v1 [cs.cv]
More informationSystem of Recognizing Human Action by Mining in Time-Series Motion Logs and Applications
The 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems October 18-22, 2010, Taipei, Taiwan System of Recognizing Human Action by Mining in Time-Series Motion Logs and Applications
More informationCS 131 Lecture 1: Course introduction
CS 131 Lecture 1: Course introduction Olivier Moindrot Department of Computer Science Stanford University Stanford, CA 94305 olivierm@stanford.edu 1 What is computer vision? 1.1 Definition Two definitions
More informationA Glimpse of Human-Computer Interaction
A Glimpse of Human-Computer Interaction Jim Hollan Co-Director Design Lab Department of Cognitive Science Department of Computer Science and Engineering Email: hollan@ucsd.edu Lab: Design Lab at UC San
More informationWadehra Kartik, Kathpalia Mukul, Bahl Vasudha, International Journal of Advance Research, Ideas and Innovations in Technology
ISSN: 2454-132X Impact factor: 4.295 (Volume 4, Issue 1) Available online at www.ijariit.com Hand Detection and Gesture Recognition in Real-Time Using Haar-Classification and Convolutional Neural Networks
More informationClassifying the Brain's Motor Activity via Deep Learning
Final Report Classifying the Brain's Motor Activity via Deep Learning Tania Morimoto & Sean Sketch Motivation Over 50 million Americans suffer from mobility or dexterity impairments. Over the past few
More informationWhat Is And How Will Machine Learning Change Our Lives. Fair Use Agreement
What Is And How Will Machine Learning Change Our Lives Raymond Ptucha, Rochester Institute of Technology 2018 Engineering Symposium April 24, 2018, 9:45am Ptucha 18 1 Fair Use Agreement This agreement
More informationPhoto Selection for Family Album using Deep Neural Networks
Photo Selection for Family Album using Deep Neural Networks ABSTRACT Sijie Shen The University of Tokyo shensijie@hal.t.u-tokyo.ac.jp Michi Sato Chikaku Inc. michisato@chikaku.co.jp The development of
More informationDeep Learning for Infrastructure Assessment in Africa using Remote Sensing Data
Deep Learning for Infrastructure Assessment in Africa using Remote Sensing Data Pascaline Dupas Department of Economics, Stanford University Data for Development Initiative @ Stanford Center on Global
More informationActivities at SC 24 WG 9: An Overview
Activities at SC 24 WG 9: An Overview G E R A R D J. K I M, C O N V E N E R I S O J T C 1 S C 2 4 W G 9 Mixed and Augmented Reality (MAR) ISO SC 24 and MAR ISO-IEC JTC 1 SC 24 Have developed standards
More informationForget Luminance Conversion and Do Something Better
Forget Luminance Conversion and Do Something Better Rang M. H. Nguyen National University of Singapore nguyenho@comp.nus.edu.sg Michael S. Brown York University mbrown@eecs.yorku.ca Supplemental Material
More informationThermal glass tube processing with laser beam SINCE 1950
Thermal glass tube processing with laser beam SINCE 1950 Target applica on Thermal glass tube processing with laser beam Labglass: Lamp glass: joining, forming, cu ng joining For the first me in public
More informationWHO. 6 staff people. Tel: / Fax: Website: vision.unipv.it
It has been active in the Department of Electrical, Computer and Biomedical Engineering of the University of Pavia since the early 70s. The group s initial research activities concentrated on image enhancement
More informationASIAN WOMEN IN LEADERSHIP SUMMIT-CHINA
亚洲女子领导峰会 - 中国 ASIAN WOMEN IN LEADERSHIP SUMMIT-CHINA Friday, March 9th, 2018 Shanghai Sponsored by #acceleratingchange Handmade to Delight ASIAN WOMEN IN LEADERSHIP SUMMIT CHINA The Annual Asian Women
More informationLocal and Low-Cost White Space Detection
Local and Low-Cost White Space Detection Ahmed Saeed*, Khaled A. Harras, Ellen Zegura*, and Mostafa Ammar* *Georgia Institute of Technology Carnegie Mellon University Qatar White Space Definition A vacant
More informationAn Egocentric Perspec/ve on Ac/ve Vision and Visual Object Learning in Toddlers
An Egocentric Perspec/ve on Ac/ve Vision and Visual Object Learning in Toddlers S. Bambach, D. Crandall, L. Smith, C. Yu. ICDL 2017 Experiment presenters: Arjun, Ginevra Their Experiments Image source:
More informationUIC-ATC-ScalCom-CBDCom-IoP Hacking Health Behaviors through Wearable Sensing
UIC-ATC-ScalCom-CBDCom-IoP 2015 Tutorial Talk I Hacking Health Behaviors through Wearable Sensing Guanling Chen Univ. of Massachusetts Lowell E-mail:glchen@cs.uml.edu http://www.cs.uml.edu/~glchen/ The
More informationEgocentric Visitors Localization in Cultural Sites
Egocentric Visitors Localization in Cultural Sites FRANCESCO RAGUSA, DMI - IPLab, Università degli Studi di Catania ANTONINO FURNARI, DMI - IPLab, Università degli Studi di Catania SEBASTIANO BATTIATO,
More informationPrivacy Preserving, Standard- Based Wellness and Activity Data Modelling & Management within Smart Homes
Privacy Preserving, Standard- Based Wellness and Activity Data Modelling & Management within Smart Homes Ismini Psychoula (ESR 3) De Montfort University Prof. Liming Chen, Dr. Feng Chen 24 th October 2017
More informationRecognizing Activities of Daily Living with a Wrist-mounted Camera Supplemental Material
Recognizing Activities of Daily Living with a Wrist-mounted Camera Supplemental Material Katsunori Ohnishi, Atsushi Kanehira, Asako Kanezaki, Tatsuya Harada Graduate School of Information Science and Technology,
More informationThe Key to Your Guitar Playing Success
Table Of Contents Become Aware of Who You Are as a Guitar Player 3 Remember Why You Were Supposed to be a Guitar Player in the First Place Stop Comparing Yourself to the Rest of the World 6 Set Realistic
More informationSome thoughts on safety of machine learning
Pattern Recognition and Applications Lab Some thoughts on safety of machine learning Fabio Roli HUML 2016, Venice, December 16th, 2016 Department of Electrical and Electronic Engineering University of
More informationInterview Starter Kit
Interview Starter Kit Great insights start with people so for our Ageing Challenge we encourage you to think about who you could talk with to build awareness around the issues we re focused on. We re eager
More informationVideo Title Generation
Video Title Generation Kuo-Hao Zeng! NTHU EE! Tseng-Hung Chen! NTHU EE! Juan Carlos Niebles! Stanford CS! Min Sun! NTHU EE! Present at! Motivation VSLab Non-edited! No description (e.g., video title)!
More informationarxiv: v1 [cs.cv] 19 Apr 2018
Survey of Face Detection on Low-quality Images arxiv:1804.07362v1 [cs.cv] 19 Apr 2018 Yuqian Zhou, Ding Liu, Thomas Huang Beckmann Institute, University of Illinois at Urbana-Champaign, USA {yuqian2, dingliu2}@illinois.edu
More informationLANDMARK recognition is an important feature for
1 NU-LiteNet: Mobile Landmark Recognition using Convolutional Neural Networks Chakkrit Termritthikun, Surachet Kanprachar, Paisarn Muneesawang arxiv:1810.01074v1 [cs.cv] 2 Oct 2018 Abstract The growth
More informationsensing opportunities
sensing opportunities for mobile health persuasion jonfroehlich@gmail.com phd candidate in computer science university of washington mobile health conference stanford university, 05.24.2010 design: use:
More informationRecognizing Personal Contexts from Egocentric Images
Recognizing Personal Contexts from Egocentric Images Antonino Furnari, Giovanni M. Farinella, Sebastiano Battiato Department of Mathematics and Computer Science - University of Catania Viale Andrea Doria,
More informationValue-added Applications with Deep Learning. src:
SMART TOURISM Value-added Applications with Deep Learning src: https://www.wttc.org/-/media/files/reports/economic-impact-research/countries-2017/thailand2017.pdf Somnuk Phon-Amnuaisuk, Minh-Son Dao, CIE,
More informationImplementation of Barcode Localization Technique using Morphological Operations
Implementation of Barcode Localization Technique using Morphological Operations Savreet Kaur Student, Master of Technology, Department of Computer Engineering, ABSTRACT Barcode Localization is an extremely
More informationWi-Fi Fingerprinting through Active Learning using Smartphones
Wi-Fi Fingerprinting through Active Learning using Smartphones Le T. Nguyen Carnegie Mellon University Moffet Field, CA, USA le.nguyen@sv.cmu.edu Joy Zhang Carnegie Mellon University Moffet Field, CA,
More informationFinding people in repeated shots of the same scene
Finding people in repeated shots of the same scene Josef Sivic C. Lawrence Zitnick Richard Szeliski University of Oxford Microsoft Research Abstract The goal of this work is to find all occurrences of
More informationReinventing movies How do we tell stories in VR? Diego Gutierrez Graphics & Imaging Lab Universidad de Zaragoza
Reinventing movies How do we tell stories in VR? Diego Gutierrez Graphics & Imaging Lab Universidad de Zaragoza Computer Graphics Computational Imaging Virtual Reality Joint work with: A. Serrano, J. Ruiz-Borau
More informationInterviews. The Four Interview Questions You Must be Able to Answer
An interview is a two-way exchange of information. While the interviewer is interested in learning more about what you have to offer, the interviewee (you!) should be interested in learning more about
More informationMultispectral Pedestrian Detection using Deep Fusion Convolutional Neural Networks
Multispectral Pedestrian Detection using Deep Fusion Convolutional Neural Networks Jo rg Wagner1,2, Volker Fischer1, Michael Herman1 and Sven Behnke2 1- Robert Bosch GmbH - 70442 Stuttgart - Germany 2-
More informationTechnology designed to empower people
Edition July 2018 Smart Health, Wearables, Artificial intelligence Technology designed to empower people Through new interfaces - close to the body - technology can enable us to become more aware of our
More informationSemantic Segmentation on Resource Constrained Devices
Semantic Segmentation on Resource Constrained Devices Sachin Mehta University of Washington, Seattle In collaboration with Mohammad Rastegari, Anat Caspi, Linda Shapiro, and Hannaneh Hajishirzi Project
More information