Understanding egocentric imagery, for fun and science

Size: px
Start display at page:

Download "Understanding egocentric imagery, for fun and science"

Transcription

1 Understanding egocentric imagery, for fun and science David Crandall School of Informa-cs and Compu-ng Indiana University Joint work with: Denise Anthony (Dartmouth), Apu Kapadia, Chen Yu; PhD Students: Sven Bambach, Roberto Hoyle, Mohammed Korayem, Stefan Lee, Robert Templeman; Undergrads: Steven Armes, Dennis Chen

2 - Google - Narra-ve - Autographer - Samsung - GoPro

3

4 June 2, 11:10 am July 2, 11:10 am Aug 2, 11:10 am Sept 2, 11:10 am

5

6 Daguerreotypes, 1839 The Kodak, 1888 Polaroid Land Camera, 1948 Digital camera, 1975 J-phone, 2000

7 Daguerreotypes, 1839 The Kodak, 1888 Polaroid Land Camera, 1948 Digital camera, 1975 J-phone, 2000

8

9

10 What if your device were hacked? Robert Templeman, Zahid Rahman, David Crandall, Apu Kapadia. PlaceRaider: Virtual The_ in Physical Spaces with Smartphones. NDSS, 2013.

11 R. Templeman, Z. Rahman, D. Crandall, A. Kapadia. PlaceRaider: Virtual The_ in Physical Spaces with Smartphones. NDSS, 2013.

12 Vision to the rescue! Could computer vision automa-cally help Organize and analysis egocentric image streams? Find the great photos amongst all the bad ones? Warn before sharing a photo with private data to Facebook? Block or censor private photos from being taken by the device, and/or uploaded to the cloud?

13 What makes an image sensi-ve? 37 undergrads, wearing life-logging cameras for a week, each day reviewing images and labeling them in various ways. R. Hoyle, R. Templeman, S. Armes, D. Anthony, D. Crandall, A. Kapadia, "Privacy behaviors of lifeloggers using wearable cameras," UBICOMP, 2014.

14 Ethical, legal, IRB considera-ons

15 Why not share the photo? Reason Responses No good reason to share it 36.0% Objects (other than people) in the photo 30.7% Where this photo was taken 22.6% People within the photo 18.4% Participant was in the photo 11.5% It had private information 11.5% It would have been embarrassing to share it 5.4% It would have violated someone else s privacy 3.8% It was a bad photo 1.5%

16 Why not share the photo? Reason Responses No good reason to share it 36.0% Objects (other than people) in the photo 30.7% Where this photo was taken 22.6% People within the photo 18.4% Participant was in the photo 11.5% It had private information 11.5% It would have been embarrassing to share it 5.4% It would have violated someone else s privacy 3.8% It was a bad photo 1.5% 16

17 Why not share the photo? Reason Responses No good reason to share it 36.0% Objects (other than people) in the photo 30.7% Where this photo was taken 22.6% People within the photo 18.4% Participant was in the photo 11.5% It had private information 11.5% It would have been embarrassing to share it 5.4% It would have violated someone else s privacy 3.8% It was a bad photo 1.5% 17

18 Why not share the photo? Reason Responses No good reason to share it 36.0% Objects (other than people) in the photo 30.7% Where this photo was taken 22.6% People within the photo 18.4% Participant was in the photo 11.5% It had private information 11.5% It would have been embarrassing to share it 5.4% It would have violated someone else s privacy 3.8% It was a bad photo 1.5% 18

19 Why not share the photo? Reason Responses No good reason to share it 36.0% Objects (other than people) in the photo 30.7% Where this photo was taken 22.6% People within the photo 18.4% Participant was in the photo 11.5% It had private information 11.5% It would have been embarrassing to share it 5.4% It would have violated someone else s privacy 3.8% It was a bad photo 1.5% 19

20 Place recogni-on in lifelogging images Given a stream of lifelogging photos, where was each photo taken? In which specific room? In which type of room? Use SVMs with standard image features HOG, GIST, LBP, SIFT, R. Templeman, M. Korayem, D. Crandall, A. Kapadia. PlaceAvoider: Steering first-person cameras away from sensi-ve spaces, NDSS 2014.

21 Classifying photo streams with HMMs ProbabiliHes with individual photo classifiers: Bathroom: Bedroom: Garage: Living: Office: ProbabiliHes aker applying HMM: Bathroom: Bedroom: Garage: Living: Office:

22 Evalua-on Tested on 5 realis-c lifelogging datasets Dataset Baseline Local features + HMM Global features + HMM Local+global features + HMM Local+global +human + HMM House % 89.2% 64.0% 89.2% 95.0% House % 55.0% 56.4% 74.6% 76.8% House % 97.4% 86.9% 98.7% 99.8% Workplace % 75.5% 89.2% 87.7% 91.0% Workplace % 92.3% 81.2% 98.7% 100.0% Average 28.5% 81.9% 74.8% 89.8% 92.5%

23 Sample results

24 Detec-ng screens CNNs, plus temporal smoothing, gives ~95% 2-way (monitor vs nomonitor) performance on real lifelogging data (vs ~73% baseline). M. Korayem, R. Templeman, D. Chen, D. Crandall, A. Kapadia. ScreenAvoider: Protec-ng computer screens from ubiquitous cameras. (Arxiv preprint), 2014.

25 Vannevar Bush, The Atlan:c, 1945

26 child s first person view parent s first person view from eye camera from eye camera from head camera from head camera

27 (Very broad) Research ques-ons How do children coordinate their hand movement, eye gaze, and head movements? How do these develop and change over -me? Are certain parerns predic-ve of later deficiencies? How do children and parents interact, and how does this differ across subjects? How do they jointly coordinate aren-on? Which interac-on parerns are most successful for learning?

28 Typical experiments 13 child-parent dyads (childrens mean age = 13 months, σ = 3.2 months) Parents told to engage child with toys, naturally, e.g.: Exchanging toys back and forth Joint ac-ons with toys ~5 minutes of video per trial In lab sexng, more recently in naturalis-c environments

29 Analysis Video data processing: Pixel-level visual saliency es-ma-on Saliency map models e.g. of Ix et al Object segmenta-on and recogni-on Posi-ons of toys, hands, and faces Object holding ac-vi-es MoHon data processing: Op-cal flow Head, hand iner-al sensors

30 Eye gaze within the visual field Children μ = (340, 231) σ x = 86, σ y = 65 N = 148,279 Parents μ = (361, 224) σ x = 53, σ y = 60 N = 148,279 S. Bambach, D. Crandall, C. Yu. Understanding embodied visual aren-on in child-parent interac-ons, IEEE Interna:onal Conference on Development and Learning, 2013.

31 Head-eye coordina-on Moving Head StaHonary Head Children µ = (333, 253) σ x = 80, σ y = 70 N = 15,626 µ = (335, 226) σ x = 84, σ y = 63 N = 71,388 Parents µ = (351, 229) σ x = 57, σ y = 62 N = 63,035 µ = (374, 210) σ x = 47, σ y = 56 N = 33,981

32 Eye-hand coordina-on Empty Hands Holding Toy Children µ = (332, 231) σ x = 84, σ y = 68 N = 55,704 µ = (301, 242) σ x = 85, σ y = 62 N = 10,937 Parents µ = (336, 244) σ x = 71, σ y = 70 N = 27,062 µ = (333, 270) σ x = 63, σ y = 73 N = 11,239

33 Saliency in first-person views Comparison of average saliency (based on 148,000 frames each, using Ix et al) à No significant difference in child and parent views Child Parent

34 Saliency in first-person views Comparison of average saliency within hotspot around gaze loca-on hotspot à Gaze predic-veness differs significantly Child Parent

35 Longer-term ques-on Can we jointly model head and hand pose, eye gaze, saliency, and ac-vity, both to berer perform egocentric computer vision, and to help explain human vision? Head Eye Hand

36 Why start with hands? Hands are in nearly every frame of egocentric video Hand configura-on reflects what we are doing and what we are paying aren-on to Detec-ng hands is a fundamental problem for both computers and people The feeling of ownership of our limbs is a fundamental aspect of self-consciousness. [Ehrsson 2004]

37 A. Fathi, X. Ren, J. Rehg, "Learning to recognize objects in egocentric activities," CVPR Pirsiavash, Hamed, and Ramanan. "Detecting activities of daily living in first-person camera views." CVPR 2012 A. Fathi, A. Farhadi, and J. Rehg. Understanding egocentric activities, ICCV 2011

38 Hand detec-on and disambigua-on In egocentric video of interac-ng people, locate: The observer s hands (my le_, my right) The other person s hands (your le_, your right) The other person s head (your head) Our approaches so far: Strong temporal models, weak spa-al/appearance models S. Bambach, J. Franchak, D. Crandall, C. Yu. Detec-ng hands in children s egocentric views to understand embodied aren-on during social interac-on. CogSci Strong spa-al/temporal models, weak appearance models, explicit camera (head) mo-on model S. Lee, S. Bambach, C. Yu, D. Crandall. This hand is my hand: A probabilis-c approach to hand disambigua-on in egocentric video, CVPR Egovision, Strong appearance models, weak spa-al/temporal models S. Bambach, S. Lee, D. Crandall, C. Yu. Lending a hand: Detec-ng hands and recognizing ac-vi-es in complex egocentric interac-ons. ICCV, 2015.

39 Strong appearance models: CNNs (of course) Mostly off-the-shelf Caffe, fine-tuned from ImageNet Generate candidates using domain-specific informa-on Sample from distribu-on over size, posi-on, shape of hand regions in training data, biased by skin color detector Other s right hand My le_ hand Other s le_ hand My right hand Coverage Direct Sampling Direct Sampling (only spatial) Sampling from Selective Search Selective Search Sampling from Objectness Objectness # Window Proposals

40 Hand detec-on & disambigua-on results 4 actors x 4 ac-vi-es x 3 loca-ons = 48 unique videos 15,053 pixel-level ground truth segmenta-ons Precision Precision ours AP: AUC: SS AP: AUC: obj. AP: AUC: Recall own left AP: AUC: own right AP: AUC: other left AP: AUC: other right AP: AUC: Recall

41 Hand detec-on & segmenta-on

42

43 Does hand pose alone reveal first person ac-vi-es? Different interac-ons afford different hand grasps [Napier 1965] Train and test CNNs for 4 ac-vi-es (puzzle, jenga, cards, chess) on masked out images Single frame accuracy: 66.4% with GT masks, 50.9% with automa-c masks (vs 25% baseline) With temporal cues (50 frames): 92.9% with GT masks, 73.4% with automa-c masks

44 Hand PosiHons during Sustained ASenHon, within child's field of view Parent s Right Hand Parent s Left Hand Child s Left Hand Child s Right Hand 6 child-parent dyads (children s mean age = 19 months, σ = 2.56 months) 67,913 frames (~38 minutes) of child-view video

45 Daguerreotypes, 1839 The Kodak, 1888 Polaroid Land Camera, 1948 Digital camera, 1975 J-phone, 2000

46 Future work Detec-ng and disambigua-ng hands Generalize hand-based ac-vity recogni-on to more ac-vi-es, finer-grained ac-ons More challenging social situa-ons (e.g. more than two interac-ng people, moving camera wearer) Applica-ons for egocentric video data

47 For more informahon: hsp://vision.soic.indiana.edu/ Based on joint work with: Faculty: Denise Anthony, Apu Kapadia, Chen Yu PhD Students: Sven Bambach, Roberto Hoyle, Mohammed Korayem, Stefan Lee, Robert Templeman Undergrad students: Steven Armes, Dennis Chen Sponsors: NSF (III CAREER, SaTC, EAGER, DIBBs), Intelligence Advanced Research Projects Ac-vity (IARPA), Air Force Office of Scien-fic Research (AFOSR), Google, Nvidia, IU FRSP, IUCRG, IU D2I Center, Lily Endowment

Exploring Wearable Cameras for Educational Purposes

Exploring Wearable Cameras for Educational Purposes 70 Exploring Wearable Cameras for Educational Purposes Jouni Ikonen and Antti Knutas Abstract: The paper explores the idea of using wearable cameras in educational settings. In the study, a wearable camera

More information

arxiv: v1 [cs.cr] 28 Nov 2014

arxiv: v1 [cs.cr] 28 Nov 2014 ScreenAvoider: Protecting Computer Screens from Ubiquitous Cameras Mohammed Korayem, Robert Templeman, Dennis Chen, David Crandall, Apu Kapadia arxiv:1412.0008v1 [cs.cr] 28 Nov 2014 School of Informatics

More information

Identifying First-person Camera Wearers in Third-person Videos

Identifying First-person Camera Wearers in Third-person Videos Identifying First-person Camera Wearers in Third-person Videos Chenyou Fan 1, Jangwon Lee 1, Mingze Xu 1, Krishna Kumar Singh 2, Yong Jae Lee 2, David J. Crandall 1 and Michael S. Ryoo 1 1 Indiana University

More information

Wearable Sensing for Understanding, Forecasting and Assisting Human Activity. Kris Kitani Assistant Research Professor Carnegie Mellon University

Wearable Sensing for Understanding, Forecasting and Assisting Human Activity. Kris Kitani Assistant Research Professor Carnegie Mellon University Wearable Sensing for Understanding, Forecasting and Assisting Human Activity Kris Kitani Assistant Research Professor Carnegie Mellon University camera Wearable sensing camera smartphone camera Wearable

More information

Going Deeper into First-Person Activity Recognition

Going Deeper into First-Person Activity Recognition Going Deeper into First-Person Activity Recognition Minghuang Ma, Haoqi Fan and Kris M. Kitani Carnegie Mellon University Pittsburgh, PA 15213, USA minghuam@andrew.cmu.edu haoqif@andrew.cmu.edu kkitani@cs.cmu.edu

More information

Presen ng the Early Warning Systems: MicroMesh + GlobalMesh

Presen ng the Early Warning Systems: MicroMesh + GlobalMesh Presen ng the Early Warning Systems: MicroMesh + GlobalMesh RF Detection: A Technology Brief concerning the detection of Drones Above + Phones Below, and everything in-between HSS Development 2 The Early

More information

DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition. ECE 289G: Paper Presentation #3 Philipp Gysel

DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition. ECE 289G: Paper Presentation #3 Philipp Gysel DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition ECE 289G: Paper Presentation #3 Philipp Gysel Autonomous Car ECE 289G Paper Presentation, Philipp Gysel Slide 2 Source: maps.google.com

More information

First Person Action Recognition Using Deep Learned Descriptors

First Person Action Recognition Using Deep Learned Descriptors First Person Action Recognition Using Deep Learned Descriptors Suriya Singh 1 Chetan Arora 2 C. V. Jawahar 1 1 IIIT Hyderabad, India 2 IIIT Delhi, India Abstract We focus on the problem of wearer s action

More information

Continuous Gesture Recognition Fact Sheet

Continuous Gesture Recognition Fact Sheet Continuous Gesture Recognition Fact Sheet August 17, 2016 1 Team details Team name: ICT NHCI Team leader name: Xiujuan Chai Team leader address, phone number and email Address: No.6 Kexueyuan South Road

More information

COACHES COopera-ve Autonomous Robots in Complex and Human EnvironmentS

COACHES COopera-ve Autonomous Robots in Complex and Human EnvironmentS COACHES COopera-ve Autonomous Robots in Complex and Human EnvironmentS Coordinator : Abdel- Illah Mouaddib Partners : University of Caen Basse- Normandie, France University of Sapienza Roma, Italy Vrije

More information

Driving Using End-to-End Deep Learning

Driving Using End-to-End Deep Learning Driving Using End-to-End Deep Learning Farzain Majeed farza@knights.ucf.edu Kishan Athrey kishan.athrey@knights.ucf.edu Dr. Mubarak Shah shah@crcv.ucf.edu Abstract This work explores the problem of autonomously

More information

The Distributed Camera

The Distributed Camera The Distributed Camera Noah Snavely Cornell University Microsoft Faculty Summit June 16, 2013 The Age of Exapixel Image Data Over a trillion photos available online Millions uploaded every hour Interconnected

More information

A Fuller Understanding of Fully Convolutional Networks. Evan Shelhamer* Jonathan Long* Trevor Darrell UC Berkeley in CVPR'15, PAMI'16

A Fuller Understanding of Fully Convolutional Networks. Evan Shelhamer* Jonathan Long* Trevor Darrell UC Berkeley in CVPR'15, PAMI'16 A Fuller Understanding of Fully Convolutional Networks Evan Shelhamer* Jonathan Long* Trevor Darrell UC Berkeley in CVPR'15, PAMI'16 1 pixels in, pixels out colorization Zhang et al.2016 monocular depth

More information

Automatic understanding of the visual world

Automatic understanding of the visual world Automatic understanding of the visual world 1 Machine visual perception Artificial capacity to see, understand the visual world Object recognition Image or sequence of images Action recognition 2 Machine

More information

CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS. Kuan-Chuan Peng and Tsuhan Chen

CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS. Kuan-Chuan Peng and Tsuhan Chen CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS Kuan-Chuan Peng and Tsuhan Chen Cornell University School of Electrical and Computer Engineering Ithaca, NY 14850

More information

Semantic Localization of Indoor Places. Lukas Kuster

Semantic Localization of Indoor Places. Lukas Kuster Semantic Localization of Indoor Places Lukas Kuster Motivation GPS for localization [7] 2 Motivation Indoor navigation [8] 3 Motivation Crowd sensing [9] 4 Motivation Targeted Advertisement [10] 5 Motivation

More information

Colorful Image Colorizations Supplementary Material

Colorful Image Colorizations Supplementary Material Colorful Image Colorizations Supplementary Material Richard Zhang, Phillip Isola, Alexei A. Efros {rich.zhang, isola, efros}@eecs.berkeley.edu University of California, Berkeley 1 Overview This document

More information

Video Object Segmentation with Re-identification

Video Object Segmentation with Re-identification Video Object Segmentation with Re-identification Xiaoxiao Li, Yuankai Qi, Zhe Wang, Kai Chen, Ziwei Liu, Jianping Shi Ping Luo, Chen Change Loy, Xiaoou Tang The Chinese University of Hong Kong, SenseTime

More information

TOWARDS A UNIFIED FRAMEWORK FOR HAND-BASED METHODS IN FIRST PERSON VISION.

TOWARDS A UNIFIED FRAMEWORK FOR HAND-BASED METHODS IN FIRST PERSON VISION. TOWARDS A UNFED FRAMEWORK FOR HAND-BASED METHODS N FRST PERSON VSON. Alejandro Betancourt 1,2, Pietro Morerio 1, Lucio Marcenaro 1, Emilia Barakova 2, Matthias Rauterberg 2, Carlo Regazzoni 1 1 nformation

More information

An FRN is a unique number used to identify an applicant in all transactions with the FCC. If you don t have an FRN

An FRN is a unique number used to identify an applicant in all transactions with the FCC. If you don t have an FRN How-to Guide: Registering Receive-Only Antennas FCC Form 312 and Schedule B Disclaimer: The materials provided herein are for informa onal purposes only and do not cons tute legal advice. You should contact

More information

FIREBREAK. Installa on Guide FIREBREAK. USA Distribu on by Mei Guo Interna onal, LLC USA - Interna onal Patent Pending.

FIREBREAK. Installa on Guide FIREBREAK. USA Distribu on by Mei Guo Interna onal, LLC USA - Interna onal Patent Pending. FIREBREAK USA Distribu on by Mei Guo Interna onal, LLC USA - Interna onal Patent Pending (1) Product Informa on Proprietary product made from a blend of high purity ceramic, alumina and zirconia cyrstal

More information

Face detection, face alignment, and face image parsing

Face detection, face alignment, and face image parsing Lecture overview Face detection, face alignment, and face image parsing Brandon M. Smith Guest Lecturer, CS 534 Monday, October 21, 2013 Brief introduction to local features Face detection Face alignment

More information

Recognizing Micro-Actions and Reactions from Paired Egocentric Videos

Recognizing Micro-Actions and Reactions from Paired Egocentric Videos Recognizing Micro-Actions and Reactions from Paired Egocentric Videos Ryo Yonetani The University of Tokyo Tokyo, Japan yonetani@iis.u-tokyo.ac.jp Kris M. Kitani Carnegie Mellon University Pittsburgh,

More information

TELLING STORIES OF VALUE WITH IOT DATA

TELLING STORIES OF VALUE WITH IOT DATA TELLING STORIES OF VALUE WITH IOT DATA VISUALIZATION BAREND BOTHA VIDEO TRANSCRIPT Tell me a little bit about yourself and your background in IoT. I came from a web development and design background and

More information

Machine Learning for Intelligent Transportation Systems

Machine Learning for Intelligent Transportation Systems Machine Learning for Intelligent Transportation Systems Patrick Emami (CISE), Anand Rangarajan (CISE), Sanjay Ranka (CISE), Lily Elefteriadou (CE) MALT Lab, UFTI September 6, 2018 ITS - A Broad Perspective

More information

Vishnu Nath. Usage of computer vision and humanoid robotics to create autonomous robots. (Ximea Currera RL04C Camera Kit)

Vishnu Nath. Usage of computer vision and humanoid robotics to create autonomous robots. (Ximea Currera RL04C Camera Kit) Vishnu Nath Usage of computer vision and humanoid robotics to create autonomous robots (Ximea Currera RL04C Camera Kit) Acknowledgements Firstly, I would like to thank Ivan Klimkovic of Ximea Corporation,

More information

Vision-based User-interfaces for Pervasive Computing. CHI 2003 Tutorial Notes. Trevor Darrell Vision Interface Group MIT AI Lab

Vision-based User-interfaces for Pervasive Computing. CHI 2003 Tutorial Notes. Trevor Darrell Vision Interface Group MIT AI Lab Vision-based User-interfaces for Pervasive Computing Tutorial Notes Vision Interface Group MIT AI Lab Table of contents Biographical sketch..ii Agenda..iii Objectives.. iv Abstract..v Introduction....1

More information

Multi-task Learning of Dish Detection and Calorie Estimation

Multi-task Learning of Dish Detection and Calorie Estimation Multi-task Learning of Dish Detection and Calorie Estimation Department of Informatics, The University of Electro-Communications, Tokyo 1-5-1 Chofugaoka, Chofu-shi, Tokyo 182-8585 JAPAN ABSTRACT In recent

More information

Mitigating Bystander Privacy Concerns in Egocentric Activity Recognition with Deep Learning and Intentional Image Degradation

Mitigating Bystander Privacy Concerns in Egocentric Activity Recognition with Deep Learning and Intentional Image Degradation Mitigating Bystander Privacy Concerns in Egocentric Activity Recognition with Deep Learning and Intentional Image Degradation MARIELLA DIMICCOLI*, University of Barcelona and Computer Vision Center, Spain

More information

DOTTORATO DI RICERCA IN INFORMATICA IX CICLO UNIVERSITA DEGLI STUDI DI SALERNO. Forensic Analysis for Digital Images.

DOTTORATO DI RICERCA IN INFORMATICA IX CICLO UNIVERSITA DEGLI STUDI DI SALERNO. Forensic Analysis for Digital Images. DOTTORATO DI RICERCA IN INFORMATICA IX CICLO UNIVERSITA DEGLI STUDI DI SALERNO Forensic Analysis for Digital Images Maurizio Cembalo November, 2010 PhD Program Chair Prof.ssa Margherita Napoli Supervisor

More information

KrishnaCam: Using a Longitudinal, Single-Person, Egocentric Dataset for Scene Understanding Tasks

KrishnaCam: Using a Longitudinal, Single-Person, Egocentric Dataset for Scene Understanding Tasks KrishnaCam: Using a Longitudinal, Single-Person, Egocentric Dataset for Scene Understanding Tasks Krishna Kumar Singh 1,3 Kayvon Fatahalian 1 Alexei A. Efros 2 1 Carnegie Mellon University 2 UC Berkeley

More information

Context Aware Computing

Context Aware Computing Context Aware Computing Context aware computing: the use of sensors and other sources of information about a user s context to provide more relevant information and services Context independent: acts exactly

More information

6. SENSE OF COMMUNITY INTRODUCTION. Direction for community connections, equity and inclusiveness, and culture and identity.

6. SENSE OF COMMUNITY INTRODUCTION. Direction for community connections, equity and inclusiveness, and culture and identity. 6. SENSE OF COMMUNITY Direction for community connections, equity and inclusiveness, and culture and identity. INTRODUCTION The concept of a sense of community is both abstract, and very relevant. It is

More information

ICARP III. 3 rd International Conference on Arctic Research Planning (ICARP III) Volker Rachold International Arctic Science Committee

ICARP III. 3 rd International Conference on Arctic Research Planning (ICARP III) Volker Rachold International Arctic Science Committee 3 rd International Conference on Arctic Research Planning (ICARP III) Volker Rachold International Arctic Science Committee ISIRA Toyama (Japan), 25 April 2015 FARO provides a framework to help: identify

More information

Curriculum Vitae. Computer Vision, Image Processing, Biometrics. Computer Vision, Vision Rehabilitation, Vision Science

Curriculum Vitae. Computer Vision, Image Processing, Biometrics. Computer Vision, Vision Rehabilitation, Vision Science Curriculum Vitae Date Prepared: 01/09/2016 (last updated: 09/12/2016) Name: Shrinivas J. Pundlik Education 07/2002 B.E. (Bachelor of Engineering) Electronics Engineering University of Pune, Pune, India

More information

Perceptual Characters of Photorealistic See-through Vision in Handheld Augmented Reality

Perceptual Characters of Photorealistic See-through Vision in Handheld Augmented Reality Perceptual Characters of Photorealistic See-through Vision in Handheld Augmented Reality Arindam Dey PhD Student Magic Vision Lab University of South Australia Supervised by: Dr Christian Sandor and Prof.

More information

AVA: A Large-Scale Database for Aesthetic Visual Analysis

AVA: A Large-Scale Database for Aesthetic Visual Analysis 1 AVA: A Large-Scale Database for Aesthetic Visual Analysis Wei-Ta Chu National Chung Cheng University N. Murray, L. Marchesotti, and F. Perronnin, AVA: A Large-Scale Database for Aesthetic Visual Analysis,

More information

Computer Vision, Lecture 3

Computer Vision, Lecture 3 Computer Vision, Lecture 3 Professor Hager http://www.cs.jhu.edu/~hager /4/200 CS 46, Copyright G.D. Hager Outline for Today Image noise Filtering by Convolution Properties of Convolution /4/200 CS 46,

More information

Mobile Cognitive Indoor Assistive Navigation for the Visually Impaired

Mobile Cognitive Indoor Assistive Navigation for the Visually Impaired 1 Mobile Cognitive Indoor Assistive Navigation for the Visually Impaired Bing Li 1, Manjekar Budhai 2, Bowen Xiao 3, Liang Yang 1, Jizhong Xiao 1 1 Department of Electrical Engineering, The City College,

More information

Virtual Reality Based Scalable Framework for Travel Planning and Training

Virtual Reality Based Scalable Framework for Travel Planning and Training Virtual Reality Based Scalable Framework for Travel Planning and Training Loren Abdulezer, Jason DaSilva Evolving Technologies Corporation, AXS Lab, Inc. la@evolvingtech.com, jdasilvax@gmail.com Abstract

More information

Domain Adaptation & Transfer: All You Need to Use Simulation for Real

Domain Adaptation & Transfer: All You Need to Use Simulation for Real Domain Adaptation & Transfer: All You Need to Use Simulation for Real Boqing Gong Tecent AI Lab Department of Computer Science An intelligent robot Semantic segmentation of urban scenes Assign each pixel

More information

Face Registration Using Wearable Active Vision Systems for Augmented Memory

Face Registration Using Wearable Active Vision Systems for Augmented Memory DICTA2002: Digital Image Computing Techniques and Applications, 21 22 January 2002, Melbourne, Australia 1 Face Registration Using Wearable Active Vision Systems for Augmented Memory Takekazu Kato Takeshi

More information

Ethical issues raised by big data and real world evidence projects. Dr Andrew Turner

Ethical issues raised by big data and real world evidence projects. Dr Andrew Turner Ethical issues raised by big data and real world evidence projects Dr Andrew Turner andrew.turner@oii.ox.ac.uk December 8, 2017 What is real world evidence and big data? Real world evidence is evidence

More information

Liangliang Cao *, Jiebo Luo +, Thomas S. Huang *

Liangliang Cao *, Jiebo Luo +, Thomas S. Huang * Annotating ti Photo Collections by Label Propagation Liangliang Cao *, Jiebo Luo +, Thomas S. Huang * + Kodak Research Laboratories *University of Illinois at Urbana-Champaign (UIUC) ACM Multimedia 2008

More information

Egocentric Field-of-View Localization Using First-Person Point-of-View Devices

Egocentric Field-of-View Localization Using First-Person Point-of-View Devices Egocentric Field-of-View Localization Using First-Person Point-of-View Devices Vinay Bettadapura 1,2 vinay@gatech.edu Irfan Essa 1,2 irfan@cc.gatech.edu 1 Google Inc., Mountain View, CA, USA Caroline Pantofaru

More information

A Spatiotemporal Approach for Social Situation Recognition

A Spatiotemporal Approach for Social Situation Recognition A Spatiotemporal Approach for Social Situation Recognition Christian Meurisch, Tahir Hussain, Artur Gogel, Benedikt Schmidt, Immanuel Schweizer, Max Mühlhäuser Telecooperation Lab, TU Darmstadt MOTIVATION

More information

IEEE SmartAg Ini-a-ve: Technology applied to the food supply chain John Verboncoeur, Chair IEEE SmartAg Ini>a>ve Michigan State University

IEEE SmartAg Ini-a-ve: Technology applied to the food supply chain John Verboncoeur, Chair IEEE SmartAg Ini>a>ve Michigan State University IEEE SmartAg Ini-a-ve: Technology applied to the food supply chain John Verboncoeur, Chair IEEE SmartAg Ini>a>ve Michigan State University 1 IEEE SmartAg Ini-a-ve Smart Technologies and Innovations Applied

More information

Lecture 7: Scene Text Detection and Recognition. Dr. Cong Yao Megvii (Face++) Researcher

Lecture 7: Scene Text Detection and Recognition. Dr. Cong Yao Megvii (Face++) Researcher Lecture 7: Scene Text Detection and Recognition Dr. Cong Yao Megvii (Face++) Researcher yaocong@megvii.com Outline Background and Introduction Conventional Methods Deep Learning Methods Datasets and Competitions

More information

Detection and Segmentation. Fei-Fei Li & Justin Johnson & Serena Yeung. Lecture 11 -

Detection and Segmentation. Fei-Fei Li & Justin Johnson & Serena Yeung. Lecture 11 - Lecture 11: Detection and Segmentation Lecture 11-1 May 10, 2017 Administrative Midterms being graded Please don t discuss midterms until next week - some students not yet taken A2 being graded Project

More information

Personal Driving Diary: Constructing a Video Archive of Everyday Driving Events

Personal Driving Diary: Constructing a Video Archive of Everyday Driving Events Proceedings of IEEE Workshop on Applications of Computer Vision (WACV), Kona, Hawaii, January 2011 Personal Driving Diary: Constructing a Video Archive of Everyday Driving Events M. S. Ryoo, Jae-Yeong

More information

GESTURE RECOGNITION WITH 3D CNNS

GESTURE RECOGNITION WITH 3D CNNS April 4-7, 2016 Silicon Valley GESTURE RECOGNITION WITH 3D CNNS Pavlo Molchanov Xiaodong Yang Shalini Gupta Kihwan Kim Stephen Tyree Jan Kautz 4/6/2016 Motivation AGENDA Problem statement Selecting the

More information

Effects of the Unscented Kalman Filter Process for High Performance Face Detector

Effects of the Unscented Kalman Filter Process for High Performance Face Detector Effects of the Unscented Kalman Filter Process for High Performance Face Detector Bikash Lamsal and Naofumi Matsumoto Abstract This paper concerns with a high performance algorithm for human face detection

More information

Part I New Sensing Technologies for Societies and Environment

Part I New Sensing Technologies for Societies and Environment Part I New Sensing Technologies for Societies and Environment Introduction New ICT-Mediated Sensing Opportunities Andreas Hotho, Gerd Stumme, and Jan Theunis During the last century, the application of

More information

Mobile Crowdsensing enabled IoT frameworks: harnessing the power and wisdom of the crowd

Mobile Crowdsensing enabled IoT frameworks: harnessing the power and wisdom of the crowd Mobile Crowdsensing enabled IoT frameworks: harnessing the power and wisdom of the crowd Malamati Louta Konstantina Banti University of Western Macedonia OUTLINE Internet of Things Mobile Crowd Sensing

More information

Understanding User Privacy in Internet of Things Environments IEEE WORLD FORUM ON INTERNET OF THINGS / 30

Understanding User Privacy in Internet of Things Environments IEEE WORLD FORUM ON INTERNET OF THINGS / 30 Understanding User Privacy in Internet of Things Environments HOSUB LEE AND ALFRED KOBSA DONALD BREN SCHOOL OF INFORMATION AND COMPUTER SCIENCES UNIVERSITY OF CALIFORNIA, IRVINE 2016-12-13 IEEE WORLD FORUM

More information

RUGGED. MARINIZED. LOW MAINTENANCE.

RUGGED. MARINIZED. LOW MAINTENANCE. RUGGED. MARINIZED. LOW MAINTENANCE. MWIR LWIR SWIR NIGHT VISION DAY / LOW LIGHT LASER DAZZLER / LRF FULL SPECTRUM EO / IR SYSTEMS Series NN 1000 NN 2000 NN 6000 NN 6000 NN 7000 MODEL NN 1045 NN HSC NN

More information

Real Time and Non-intrusive Driver Fatigue Monitoring

Real Time and Non-intrusive Driver Fatigue Monitoring Real Time and Non-intrusive Driver Fatigue Monitoring Qiang Ji and Zhiwei Zhu jiq@rpi rpi.edu Intelligent Systems Lab Rensselaer Polytechnic Institute (RPI) Supported by AFOSR and Honda Introduction Motivation:

More information

Nkxg"Ncwij"Nqxg"Ctv 423: Iudqfklvh#Rssruwxqlw #Sdfndjh. Rtqrtkgvct{"("Eqphkfgpvkcn"Nkxg"Ncwij"Nqxg"Ctv."NNE"423:

NkxgNcwijNqxgCtv 423: Iudqfklvh#Rssruwxqlw #Sdfndjh. Rtqrtkgvct{(EqphkfgpvkcnNkxgNcwijNqxgCtv.NNE423: Nkxg"Ncwij"Nqxg"Ctv 423: Iudqfklvh#Rssruwxqlw #Sdfndjh Rtqrtkgvct{"("Eqphkfgpvkcn"Nkxg"Ncwij"Nqxg"Ctv."NNE"423: Our Story What We Do Our Culture In 2010, during one of the most devasta ng mes in our economy,

More information

arxiv: v1 [cs.cv] 27 Nov 2016

arxiv: v1 [cs.cv] 27 Nov 2016 Real-Time Video Highlights for Yahoo Esports arxiv:1611.08780v1 [cs.cv] 27 Nov 2016 Yale Song Yahoo Research New York, USA yalesong@yahoo-inc.com Abstract Esports has gained global popularity in recent

More information

CS688/WST665 Student presentation Learning Fine-grained Image Similarity with Deep Ranking CVPR Gayoung Lee ( 이가영 )

CS688/WST665 Student presentation Learning Fine-grained Image Similarity with Deep Ranking CVPR Gayoung Lee ( 이가영 ) CS688/WST665 Student presentation Learning Fine-grained Image Similarity with Deep Ranking CVPR 2014 Gayoung Lee ( 이가영 ) Contents 1. Background knowledge 2. Proposed method 3. Experimental Result 4. Conclusion

More information

UNDERGRADUATE PROJECT REPORT

UNDERGRADUATE PROJECT REPORT UNIVERSITAT POLITÈCNICA DE CATALUNYA ESEIAAT EgoMon Gaze and Video Dataset for Visual Saliency Prediction June 2016 UNDERGRADUATE PROJECT REPORT Supervised by Xavier Giró and Cathal Gurrin Acknowledgements

More information

Making a Difference in 2017 IPOs & CES

Making a Difference in 2017 IPOs & CES SHAREHOLDER UPDATE JANUARY 2017 Making a Difference in 2017 IPOs & CES By Tom Astle IPOs - Will 2017 finally be the year? Our portfolio strategy of investing in later-stage growth companies was partially

More information

You don t want to be the next meme : College Students Workarounds to Manage Privacy in the Era of Pervasive Photography

You don t want to be the next meme : College Students Workarounds to Manage Privacy in the Era of Pervasive Photography You don t want to be the next meme : College Students Workarounds to Manage Privacy in the Era of Pervasive Photography Yasmeen Rashidi, Tousif Ahmed, Felicia Patel, Emily Fath, Apu Kapadia, Christena

More information

Learning Pixel-Distribution Prior with Wider Convolution for Image Denoising

Learning Pixel-Distribution Prior with Wider Convolution for Image Denoising Learning Pixel-Distribution Prior with Wider Convolution for Image Denoising Peng Liu University of Florida pliu1@ufl.edu Ruogu Fang University of Florida ruogu.fang@bme.ufl.edu arxiv:177.9135v1 [cs.cv]

More information

System of Recognizing Human Action by Mining in Time-Series Motion Logs and Applications

System of Recognizing Human Action by Mining in Time-Series Motion Logs and Applications The 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems October 18-22, 2010, Taipei, Taiwan System of Recognizing Human Action by Mining in Time-Series Motion Logs and Applications

More information

CS 131 Lecture 1: Course introduction

CS 131 Lecture 1: Course introduction CS 131 Lecture 1: Course introduction Olivier Moindrot Department of Computer Science Stanford University Stanford, CA 94305 olivierm@stanford.edu 1 What is computer vision? 1.1 Definition Two definitions

More information

A Glimpse of Human-Computer Interaction

A Glimpse of Human-Computer Interaction A Glimpse of Human-Computer Interaction Jim Hollan Co-Director Design Lab Department of Cognitive Science Department of Computer Science and Engineering Email: hollan@ucsd.edu Lab: Design Lab at UC San

More information

Wadehra Kartik, Kathpalia Mukul, Bahl Vasudha, International Journal of Advance Research, Ideas and Innovations in Technology

Wadehra Kartik, Kathpalia Mukul, Bahl Vasudha, International Journal of Advance Research, Ideas and Innovations in Technology ISSN: 2454-132X Impact factor: 4.295 (Volume 4, Issue 1) Available online at www.ijariit.com Hand Detection and Gesture Recognition in Real-Time Using Haar-Classification and Convolutional Neural Networks

More information

Classifying the Brain's Motor Activity via Deep Learning

Classifying the Brain's Motor Activity via Deep Learning Final Report Classifying the Brain's Motor Activity via Deep Learning Tania Morimoto & Sean Sketch Motivation Over 50 million Americans suffer from mobility or dexterity impairments. Over the past few

More information

What Is And How Will Machine Learning Change Our Lives. Fair Use Agreement

What Is And How Will Machine Learning Change Our Lives. Fair Use Agreement What Is And How Will Machine Learning Change Our Lives Raymond Ptucha, Rochester Institute of Technology 2018 Engineering Symposium April 24, 2018, 9:45am Ptucha 18 1 Fair Use Agreement This agreement

More information

Photo Selection for Family Album using Deep Neural Networks

Photo Selection for Family Album using Deep Neural Networks Photo Selection for Family Album using Deep Neural Networks ABSTRACT Sijie Shen The University of Tokyo shensijie@hal.t.u-tokyo.ac.jp Michi Sato Chikaku Inc. michisato@chikaku.co.jp The development of

More information

Deep Learning for Infrastructure Assessment in Africa using Remote Sensing Data

Deep Learning for Infrastructure Assessment in Africa using Remote Sensing Data Deep Learning for Infrastructure Assessment in Africa using Remote Sensing Data Pascaline Dupas Department of Economics, Stanford University Data for Development Initiative @ Stanford Center on Global

More information

Activities at SC 24 WG 9: An Overview

Activities at SC 24 WG 9: An Overview Activities at SC 24 WG 9: An Overview G E R A R D J. K I M, C O N V E N E R I S O J T C 1 S C 2 4 W G 9 Mixed and Augmented Reality (MAR) ISO SC 24 and MAR ISO-IEC JTC 1 SC 24 Have developed standards

More information

Forget Luminance Conversion and Do Something Better

Forget Luminance Conversion and Do Something Better Forget Luminance Conversion and Do Something Better Rang M. H. Nguyen National University of Singapore nguyenho@comp.nus.edu.sg Michael S. Brown York University mbrown@eecs.yorku.ca Supplemental Material

More information

Thermal glass tube processing with laser beam SINCE 1950

Thermal glass tube processing with laser beam SINCE 1950 Thermal glass tube processing with laser beam SINCE 1950 Target applica on Thermal glass tube processing with laser beam Labglass: Lamp glass: joining, forming, cu ng joining For the first me in public

More information

WHO. 6 staff people. Tel: / Fax: Website: vision.unipv.it

WHO. 6 staff people. Tel: / Fax: Website: vision.unipv.it It has been active in the Department of Electrical, Computer and Biomedical Engineering of the University of Pavia since the early 70s. The group s initial research activities concentrated on image enhancement

More information

ASIAN WOMEN IN LEADERSHIP SUMMIT-CHINA

ASIAN WOMEN IN LEADERSHIP SUMMIT-CHINA 亚洲女子领导峰会 - 中国 ASIAN WOMEN IN LEADERSHIP SUMMIT-CHINA Friday, March 9th, 2018 Shanghai Sponsored by #acceleratingchange Handmade to Delight ASIAN WOMEN IN LEADERSHIP SUMMIT CHINA The Annual Asian Women

More information

Local and Low-Cost White Space Detection

Local and Low-Cost White Space Detection Local and Low-Cost White Space Detection Ahmed Saeed*, Khaled A. Harras, Ellen Zegura*, and Mostafa Ammar* *Georgia Institute of Technology Carnegie Mellon University Qatar White Space Definition A vacant

More information

An Egocentric Perspec/ve on Ac/ve Vision and Visual Object Learning in Toddlers

An Egocentric Perspec/ve on Ac/ve Vision and Visual Object Learning in Toddlers An Egocentric Perspec/ve on Ac/ve Vision and Visual Object Learning in Toddlers S. Bambach, D. Crandall, L. Smith, C. Yu. ICDL 2017 Experiment presenters: Arjun, Ginevra Their Experiments Image source:

More information

UIC-ATC-ScalCom-CBDCom-IoP Hacking Health Behaviors through Wearable Sensing

UIC-ATC-ScalCom-CBDCom-IoP Hacking Health Behaviors through Wearable Sensing UIC-ATC-ScalCom-CBDCom-IoP 2015 Tutorial Talk I Hacking Health Behaviors through Wearable Sensing Guanling Chen Univ. of Massachusetts Lowell E-mail:glchen@cs.uml.edu http://www.cs.uml.edu/~glchen/ The

More information

Egocentric Visitors Localization in Cultural Sites

Egocentric Visitors Localization in Cultural Sites Egocentric Visitors Localization in Cultural Sites FRANCESCO RAGUSA, DMI - IPLab, Università degli Studi di Catania ANTONINO FURNARI, DMI - IPLab, Università degli Studi di Catania SEBASTIANO BATTIATO,

More information

Privacy Preserving, Standard- Based Wellness and Activity Data Modelling & Management within Smart Homes

Privacy Preserving, Standard- Based Wellness and Activity Data Modelling & Management within Smart Homes Privacy Preserving, Standard- Based Wellness and Activity Data Modelling & Management within Smart Homes Ismini Psychoula (ESR 3) De Montfort University Prof. Liming Chen, Dr. Feng Chen 24 th October 2017

More information

Recognizing Activities of Daily Living with a Wrist-mounted Camera Supplemental Material

Recognizing Activities of Daily Living with a Wrist-mounted Camera Supplemental Material Recognizing Activities of Daily Living with a Wrist-mounted Camera Supplemental Material Katsunori Ohnishi, Atsushi Kanehira, Asako Kanezaki, Tatsuya Harada Graduate School of Information Science and Technology,

More information

The Key to Your Guitar Playing Success

The Key to Your Guitar Playing Success Table Of Contents Become Aware of Who You Are as a Guitar Player 3 Remember Why You Were Supposed to be a Guitar Player in the First Place Stop Comparing Yourself to the Rest of the World 6 Set Realistic

More information

Some thoughts on safety of machine learning

Some thoughts on safety of machine learning Pattern Recognition and Applications Lab Some thoughts on safety of machine learning Fabio Roli HUML 2016, Venice, December 16th, 2016 Department of Electrical and Electronic Engineering University of

More information

Interview Starter Kit

Interview Starter Kit Interview Starter Kit Great insights start with people so for our Ageing Challenge we encourage you to think about who you could talk with to build awareness around the issues we re focused on. We re eager

More information

Video Title Generation

Video Title Generation Video Title Generation Kuo-Hao Zeng! NTHU EE! Tseng-Hung Chen! NTHU EE! Juan Carlos Niebles! Stanford CS! Min Sun! NTHU EE! Present at! Motivation VSLab Non-edited! No description (e.g., video title)!

More information

arxiv: v1 [cs.cv] 19 Apr 2018

arxiv: v1 [cs.cv] 19 Apr 2018 Survey of Face Detection on Low-quality Images arxiv:1804.07362v1 [cs.cv] 19 Apr 2018 Yuqian Zhou, Ding Liu, Thomas Huang Beckmann Institute, University of Illinois at Urbana-Champaign, USA {yuqian2, dingliu2}@illinois.edu

More information

LANDMARK recognition is an important feature for

LANDMARK recognition is an important feature for 1 NU-LiteNet: Mobile Landmark Recognition using Convolutional Neural Networks Chakkrit Termritthikun, Surachet Kanprachar, Paisarn Muneesawang arxiv:1810.01074v1 [cs.cv] 2 Oct 2018 Abstract The growth

More information

sensing opportunities

sensing opportunities sensing opportunities for mobile health persuasion jonfroehlich@gmail.com phd candidate in computer science university of washington mobile health conference stanford university, 05.24.2010 design: use:

More information

Recognizing Personal Contexts from Egocentric Images

Recognizing Personal Contexts from Egocentric Images Recognizing Personal Contexts from Egocentric Images Antonino Furnari, Giovanni M. Farinella, Sebastiano Battiato Department of Mathematics and Computer Science - University of Catania Viale Andrea Doria,

More information

Value-added Applications with Deep Learning. src:

Value-added Applications with Deep Learning. src: SMART TOURISM Value-added Applications with Deep Learning src: https://www.wttc.org/-/media/files/reports/economic-impact-research/countries-2017/thailand2017.pdf Somnuk Phon-Amnuaisuk, Minh-Son Dao, CIE,

More information

Implementation of Barcode Localization Technique using Morphological Operations

Implementation of Barcode Localization Technique using Morphological Operations Implementation of Barcode Localization Technique using Morphological Operations Savreet Kaur Student, Master of Technology, Department of Computer Engineering, ABSTRACT Barcode Localization is an extremely

More information

Wi-Fi Fingerprinting through Active Learning using Smartphones

Wi-Fi Fingerprinting through Active Learning using Smartphones Wi-Fi Fingerprinting through Active Learning using Smartphones Le T. Nguyen Carnegie Mellon University Moffet Field, CA, USA le.nguyen@sv.cmu.edu Joy Zhang Carnegie Mellon University Moffet Field, CA,

More information

Finding people in repeated shots of the same scene

Finding people in repeated shots of the same scene Finding people in repeated shots of the same scene Josef Sivic C. Lawrence Zitnick Richard Szeliski University of Oxford Microsoft Research Abstract The goal of this work is to find all occurrences of

More information

Reinventing movies How do we tell stories in VR? Diego Gutierrez Graphics & Imaging Lab Universidad de Zaragoza

Reinventing movies How do we tell stories in VR? Diego Gutierrez Graphics & Imaging Lab Universidad de Zaragoza Reinventing movies How do we tell stories in VR? Diego Gutierrez Graphics & Imaging Lab Universidad de Zaragoza Computer Graphics Computational Imaging Virtual Reality Joint work with: A. Serrano, J. Ruiz-Borau

More information

Interviews. The Four Interview Questions You Must be Able to Answer

Interviews. The Four Interview Questions You Must be Able to Answer An interview is a two-way exchange of information. While the interviewer is interested in learning more about what you have to offer, the interviewee (you!) should be interested in learning more about

More information

Multispectral Pedestrian Detection using Deep Fusion Convolutional Neural Networks

Multispectral Pedestrian Detection using Deep Fusion Convolutional Neural Networks Multispectral Pedestrian Detection using Deep Fusion Convolutional Neural Networks Jo rg Wagner1,2, Volker Fischer1, Michael Herman1 and Sven Behnke2 1- Robert Bosch GmbH - 70442 Stuttgart - Germany 2-

More information

Technology designed to empower people

Technology designed to empower people Edition July 2018 Smart Health, Wearables, Artificial intelligence Technology designed to empower people Through new interfaces - close to the body - technology can enable us to become more aware of our

More information

Semantic Segmentation on Resource Constrained Devices

Semantic Segmentation on Resource Constrained Devices Semantic Segmentation on Resource Constrained Devices Sachin Mehta University of Washington, Seattle In collaboration with Mohammad Rastegari, Anat Caspi, Linda Shapiro, and Hannaneh Hajishirzi Project

More information