Face2Mus: A Facial Emotion Based Internet Radio Tuner Application

Size: px
Start display at page:

Download "Face2Mus: A Facial Emotion Based Internet Radio Tuner Application"

Transcription

1 Face2Mus: A Facial Emotion Based Internet Radio Tuner Application Yara Rizk, Maya Safieddine, David Matchoulian, Mariette Awad Department of Electrical and Computer Engineering American University of Beirut Beirut, Lebanon {yar01, mhs36, dsm05, mariette.awad}@aub.edu.lb Abstract We propose in this paper, Face2Mus, a mobile application that streams music from online radio stations after identifying the user's emotions, without interfering with the device s usage. Face2Mus streams songs from online radio stations and classifies them into emotion classes based on audio features using an energy aware support vector machine (SVM) classifier. In parallel, the application captures images of the user s face using the smartphone or tablet s camera and classifying them into one of three emotions, using a multiclass SVM trained on facial geometric distances and wrinkles. The audio classification based on regular SVM achieved an overall testing accuracy of 99.83% when trained on the Million Song Dataset subset, whereas the energy aware SVM exhibited an average degradation of 1.93% when a 59% reduction in the number of support vectors (SV) is enforced. The image classification achieved an overall testing accuracy of 87.5% using leave one out validation on a home-made image database. The overall application requires 272KB of storage space, 12 to 24 MB of RAM and a startup time of approximately 2 minutes. Aside from its entertainment potentials, Face2Mus has possible usage in music therapy for improving people s well-being and emotional status. Keywords Affect recognition, Audio and image classification, mobile applications, Support Vector Machine I. INTRODUCTION The emergence of smartphones has rendered mobile applications the latest trend in the software development world, creating a multi-billion dollar industry [1] with music related apps attracting a large number of downloads [2, 3]. In addition, the use of music as a therapeutic tool has also emerged as a promising alternative in the medical field to help treat ailments such as stress, pain, and high blood pressure; it affects the mood and overall health of individuals [4]. Therefore, a smart emotion sensitive music app would be of high value for entertainment and medication purposes. In this paper, we introduce such an application as a new contribution to the currently popular affective computing research area. Face2Mus is a mobile application that streams music from online radio stations, based on the user s emotion. This emotion is deduced from image-based emotion recognition algorithms applied on the user s face, captured by the device s camera. No physiological sensors are needed, thus leading to a more accessible application with a simpler interface. An accuracy of 87.5% was achieved on a homemade image database with leave one out cross validation using a multiclass SVM. Song features were retrieved from a web service, simplifying the song classification preprocessing. Testing an energy aware support vector machine (SVM), an SVM with less SV than the regular SVM, on a subset of the Million Song Database gave an overall accuracy of 97.96% versus 99.83% for regular SVM while reducing the number of support vectors (SV) on average by 59%. In what follows, we provide a summary of related work done in the fields of affective computing and application development in Section 2. Section 3 presents a description of the proposed solution while Section 4 reports on the experimental results obtained. Finally, Section 5 concludes with follow on research potentials. II. LITERATURE REVIEW Many mobile applications have been developed tackling the user s emotion recognition [5, 6] and emotion based music classification [7, 8] separately, but none have combined them as suggested in this paper. Published work on this topic includes Affective DJ, a stand-alone device which chooses calming or energetic songs from a local playlist based on the individual s affective state deduced from skin conductivity [9]; unlike our proposed work, their method requires sensors not available in widely available smart phones. A lot of work has been done on image-based affect recognition. Classification using facial features such as the eyebrows, mouth, eyes, etc. [10, 11, 12, 13] achieved accuracies above 90%. For example, [11] reported a 93% average accuracy whereas [13] reached 96% for some emotion classes. Using physiological features such as skin conductivity [9], cardiac activity, muscle tension [10], etc. attained comparable results. In some cases, both facial features and body gestures were used [14] to achieve accuracies in the range of 82 to 96% [15]. Although physiological features were more indicative of a person s emotional state than facial expressions, they required special measuring equipment, which is not feasible for a mobile application. Several methods were implemented in the literature to tract facial features in images for emotion recognition. One /14/$ IEEE 257

2 method developed by Ekman, extracted action units (AUs) that describe a specific feature in the face such as a raised eyebrow [16]. Many papers used this method for emotion recognition [13, 17, 18]. Reference [17] achieved accuracies ranging from 64.29% to 100% on certain AUs; [13] realized comparable results. Another approach represented facial features by motion units (MUs) which describe the direction of motion of a feature and its concentration [18, 19], information not captured by AUs. Reference [19] used various classifiers, whose classification error ranged from 4.43 to 13.23%. However, this method is computationally expensive [20] and person dependent [18]; it does not generalize well. Hence, AUs would be potentially better for a mobile application with limited computational resources. Considering affect recognition in music, researchers have developed many theories to assign emotions or moods to songs based on their audio features. Hevner s theory, a categorical theory, divided moods into eight clusters of adjectives over an adjective circle [21]. On the other hand, Russell s valence vs. arousal axes [22], and Thayer s energy vs. stress axes [23] are examples of dimensional models that work over continuous multidimensional spaces. Several methods exist to identify the emotion in songs derived from the above mentioned theories. Some used audio features including mode, tempo, harmony and rhythm [24]; others analyzed the lyrics of the song using natural language processing and some combined both. Reference [25] extracted tempo, loudness and harmony using Short Time Fourier Transform (STFT) to build an audio spectrogram that implies the required information. Other sound descriptors are low-level audio descriptors (LLD) such as spectral descriptors, harmonic descriptors, and perceptual descriptors that included Mel-Frequency Cepstral Coefficients (MFCC), loudness, sharpness, spread, and roughness [25]. In addition to these descriptors, several open source libraries were available to extract audio features including Java MIR (jmir) [26] which extracts low level and high level audio features, in addition to cultural level features and C++ Library for Audio and Music (CLAM) which extracts features like mode and harmony [26, 27]. The Echonest is a music intelligence platform web service that analyzes music and offers APIs of audio analysis that return details related to tempo rhythm, time signature, loudness, mode, and key confidence [28]. Diverse classifiers were used to perform the classification tasks given the extracted features. Reference [29] implemented a hybrid music mood and theme classifier, which combined SVM with Radial Basic Function (RBF) kernel trained on audio features and Naïve Bayes Multinominal classifier trained on social tags. Using highest probability distribution, a weighted sum of the classifiers was adopted to maximize precision, recall, accuracy and F1 measure. III. FACE2MUS Face2Mus is an Android application that streams music from online radio stations based on the emotion of the app user deduced by emotion recognition techniques applied to their facial images. The workflow for Face2Mus is illustrated in Fig. 1. To infer the user s emotion, Face2Mus captures images of the user s face at an adaptive rate dependent on the sustainability of this emotion, using the mobile device s camera. Android s Native Face Detection localizes the face in the frame and crops out the face and locates the midpoint and distance between the eyes. The cropped image and extracted parameters are sent to a server running Matlab to perform the emotion recognition. On the server, a tracking algorithm is used to locate the points of interest (POIs) based on color and geometric properties of facial features. Then, wrinkles and distances between POIs are generated; these features are shown in Fig. 2.They include wrinkles on the forehead, between the eyebrows and on the cheeks and distances between points on the mouth and eyes to describe the curvature of the mouth and between the eyebrows and eyes. Finally, multiclass SVM categorizes the frame into one of three emotions (neutral, happy, or sad), based on these features. In parallel, the metadata of now-playing internet radio station songs, broadcasted by Streamfinder [30], is retrieved by the device. Based on its metadata, the song s audio features such as tempo, mode, loudness and time signature are retrieved from Echonest [28]. A variant of SVM classifier categorizes the songs into one of three classes: calm, happy, and sad. Since SVM s decision hyper plane is in terms of SV which can reach up to 50% of the training set, real time classification and low memory requirements, necessary for Face2Mus success, might be unattainable using a regular SVM. Therefore, LMSVM [31] was adopted to generate a decision hyper plane with a reduced SV set. The reduced SV set requires less computations and memory requirements leading to lower power consumption when classifying a new song, rendering LMSVM energy aware. This is achieved by clustering the training data, selecting heterogeneous clusters (clusters with points from different classes) and training on these boundary points to reduce the hyper plane s complexity. Finally, the Song Selector block uses the emotional state of the individual and the list of emotion-tagged songs to decide on an appropriate song to play. IV. EXPERIMENTAL RESULTS The application is tested for its overall functionality and classification accuracy. First, the training and testing results of the audio and image classifiers are presented separately, followed by the overall performance of Face2Mus. The application was tested on a Samsung Galaxy Tab 10.1 model, with a server running on a 64-bit Intel Core i5 processor. The application was developed using Android 3.1 API and the image processing and classification code was written in Matlab R2011b. 258

3 Fig. 1. Face2Mus overall block diagram. Table I. Most of the misclassified images were from the sad class. This was not surprising because the neutral and sad images were similar and difficult to distinguish. Furthermore, the feature set containing distances only performed better than the distances and wrinkles. The wrinkles contained relatively high noise and therefore did not contribute enough useful information to improve the classification accuracy. Table II reports on the testing accuracy of individual classifiers before aggregating them into a majority vote architecture. The test set is composed of 24 images that were not used while training the various classifiers; one image per person per class. The best classifiers were saved on the server and used to classify images sent to the server by the application running on the tablet. The size of the files saved on the server was approximately 72 KB. The process of identifying POIs, generating the features (distances only) and classifying an image needed approximately 0.25 seconds to complete. Including wrinkles, the run time increased to approximately 0.67 seconds. Since the wrinkles did not improve the classification accuracy and significantly worsened the run time for the image processing blocks, the wrinkles were removed from the feature set. C. Audio Database Classifier architectures were trained and tested on the Million Song Dataset [37] subset which included 10,000 instances and 4 features: mode, tempo, time signature and loudness. The instances were clustered into three classes using k-means clustering algorithm. The database was not balanced; 2683 were tagged as sad, 5765 were happy, and 1552 were calm. A. Image Database Fig. 2. Facial features. The image classifiers were trained on a home-made database of colored images captured by a Samsung Galaxy Tab 10.1, under similar lighting conditions. The database is composed of 120 images of eight people, 5 images for each acted emotion per person. The images were classified into 3 classes: neutral, sad and happy. For illustration purposes, a subset of these images for each class is displayed in Fig. 3. B. Image Classification Face2Mus was trained and tested on Neural network and SVM classifiers, from the Neural Network toolbox in Matlab and libsvm [36] library respectively. A grid search was performed to find the best classifier architecture for the task at hand. Majority vote classifiers given different feature sets (distances or distances and wrinkles) were compared. Finally, leave one out validation technique was used to validate the results of the classifiers. The majority vote architecture composed of one vs. one SVM classifiers, using RBF kernel on a feature set containing distances only, produced the best results. The results for some of the SVM architectures are reported in Fig. 3. Sample images from the database. TABLE I. IMAGE CLASSIFICATION RESULTS Aggregation Rule Features Accuracy (%) 1 vs. 1 Majority vote Distances vs. 1 Majority vote Distances + Wrinkles TABLE II. INDIVIDUAL CLASSIFIERS RESULTS Features Class SVM RBF Accuracy Regularization Sigma (%) Neutral vs. Happy Distances Neutral vs. Sad Sad vs. Happy Neutral vs. Happy 2 Distances Neutral vs. Sad 2 Wrinkles Sad vs. Happy

4 D. Audio Classification A neural network (NN) and a set of one vs. all LM-SVM classifiers were trained to classify the songs given the 4 features. A grid search was performed to find the optimal box constraint and kernel parameter values. Several kernels including the Gaussian and linear kernels were also tested. The training was validated using 5-fold, cross fold validation. The best results were produced by the multi-class SVM models; the accuracies of the individual one vs. all classifiers are reported in Table III. The classifiers were combined into a one vs. all hierarchical structure and achieved 99.83% accuracy when tested on 600 songs that were not included in the training set. The order of the classifiers did not affect the results of the classifier. The number of SV was 6.06, 4.60 and 5.18% for the Calm, Sad and Happy classes respectively when using regular SVM. SV set cardinality was reduced to 1.68, 2.35 and 2.29% with a minor reduction in prediction accuracy, as shown in Table III. Table IV shows the percentage of degradation in accuracy and the percentage of SV reduction obtained. Although the generalization capabilities of the classifier are slightly degraded, the gains in computational resources saved are significant. The best models obtained from the training of the various song classifiers were saved on the Android tablet. They were used to classify new songs obtained from the audio acquisition block. The computation time depends on the number of SV used by the model. The time to classify a new song on the Android device was measured to be approximately 4.1 seconds on average. The memory needed to save the classifier models also depends on the number of SVs; the size of these files was approximately 32kB. SVM LM-SVM TABLE III. AUDIO CLASSIFIER RESULTS Accuracy (%) Number of SVs (%) Calm Sad Happy Calm Sad Happy TABLE IV. PERCENTAGE DEGRADATION AND REDUCTION Degradation in Accuracy (%) Reduction in SV (%) Calm Sad Happy Average TABLE V. RUN TIME FOR EACH BLOCK OF FACE2MUS Run Time (seconds) Image Acquisition & Face Localization & Extraction 5.78 Transmit data to the server POI Localization 0.15 Feature Generation Image Classification Audio Acquisition 9.78 Audio Features Acquisition 1.00 Song Classification 4.10 E. Overall Face2Mus Performance The size of the application is approximately 272KB, considered small compared to similar applications and the standard built-in memory storage in tablets which is at least 16GB. The application uses between 12 to 24 MB of RAM out of the available 1GB on the device. It also needs an approximate of 2 minutes to start playing the first song. This is mainly due to the time required to obtain several songs from Streamfinder and classify them. However, this time can vary based on the speed of the internet connection. The run times for the individual blocks of Fig. 1 are included in Table V. The Audio Acquisition time is the average time to obtain one song from Streamfinder, given a satisfactory internet connection. Inadequate connectivity would result in many connection timeouts, significantly slowing down the application. The lack on information on the remaining play time of a song presents a challenge to the application which might store a link to a song which is almost complete. To reduce the startup time, the structure of the application can be modified to run the audio processing blocks on the server, instead of the device. This would result in a server continuously tagging songs from online radio stations and storing them in a playlist. Whenever a user logs in to the application, the application will simply need to identify the user s emotion and request a song corresponding to this emotion from the server. However, based on a limited number of trials performed using the application, the results were satisfactory, despite the shortcomings of some of the blocks mentioned above. V. CONCLUSION This work proposed an automated online song streaming application which plays songs based on the emotion of the user. The emotion is detected using facial feature identification, tracking and classification. The song classifiers achieved an average accuracy of 99.83% over all classes when using multiclass SVM and of 97.96% when using LM-SVM with an SV average reduction of 59%. The image classifiers achieved 87.5% accuracy using a one vs. one majority vote SVM classifier tested on a balanced test set of 24 images. The overall application needs 1 MB of memory, 12 to 24 MB of RAM and approximately 2 minutes to start up. Further improvements, and the focus of future work, can be made especially in the image classification blocks where accuracy can be enhanced by generating more robust features. Real time performance can be further improved by eliminating network latency for image classification and running the audio classification on a server. ACKNOWLEDGMENT This work was funded by the University Research Board at the American University of Beirut. REFERENCES [1] M. Walsh (2011, October 7). Mobile App Biz Soars: $12B by 2015 [Online]. 260

5 mediapost.com/publications/article/160163/mobileapp-biz-soars-12bby-2015.html [2] The Independent (2011, November 9). Mobile application trends for 2012: the top ten applications [Online]. independent.co.uk/life-style/gadgets-and-tech/news/mobileapplication-trends-for-2012-the-top-ten-applications html [3] J. Imam (2012, June 16). Young listeners opting to stream, not own music [Online]. edition.cnn.com/2012/06/15/tech/web/music-streaming/index.html [4] C.E. Guzzetta. Effects of relaxation and music therapy on patients in a coronary care unit with presumptive acute myocardial infraction. Heart & Lung: The journal of critical care, vol.18, issue 6, p.609, [5] Asanka Senavirathna (2012). Face Mood Detector [Online]. play.google.com/store/apps/details?id=com.wideapps.android.facemo oddetector [6] DSS (2012). Mood Scanner [Online]. play.google.com/store/apps/details?id=com.dikkar.moodscanner [7] Syntonetic (2011). Moodagent [Online]. itunes.apple.com/lb/app/moodagent/id ?mt=8 [8] JVC Kenwood Corporation (2012). Kenwood Music Control [Online]. play.google.com/store/apps/details?id=com.jvckenwood.kmc&hl=en [9] R. Picard et al., A new affect-perceiving interface and its application to personalized music selection, Proc. from the 1998 Workshop on Perceptual User Interfaces, [10] J.N. Bailenson et al., Real-time classification of evoked emotions using facial feature tracking and physiological responses, International journal of human-computer studies, vol. 66 issue 5, pp , [11] M.S. Bartlett, G. Littlewort, I. Fasel and J.R. Movellan, "Real Time Face Detection and Facial Expression Recognition: Development and Applications to Human Computer Interaction," in Conf. on Computer Vision and Pattern Recognition Workshop, 2003, pp.53. [12] N. Agarwal et al., Mood Detection: Implementing a facial expression recognition system. (CS229 project, 2009). [13] M.S. Bartlett et al., Measuring facial expressions by computer image analysis. Psychophysiology, 36: , [14] H. Gunes, M. Piccardi, and M. Pantic. From the lab to the real world: Affect recognition using multiple cues and modalities. in J. Or, editor, Affective Computing: Focus on Emotion Expression, Synthesis, and Recognition, pages Vienna, Austria, [15] P. Viola and M. Jones, Robust real-time object detection, International Journal of Computer Vision, [16] P. Ekman and W. Friesen, Facial Action Coding System. Consulting Psychologists Press, [17] Y. Tian et al., Recognizing action units for facial expression analysis, IEEE Trans. Pattern Anal. Mach. Intell. 23 (2) (2001) [18] P. Lucey et al. The extended Cohn-Kanade dataset (CK+): A complete dataset for action unit and emotion-specified expression, in Computer Vision and Pattern Recognition Workshop on Human- Communicative Behavior, [19] Y. Sun et al., Authentic emotion detection in real-time video, in Int l Workshop on Human-Computer Interaction, Lecture Notes in Computer Science, vol. 3058, Springer, 2004, pp [20] H. Tao and T. S. Huang. Connected vibrations: A modal analysis approach to non-rigid motion tracking, In CVPR, pp , [21] Hevner, K. Experimental studies of the elements of expression in music, American Journal of Psychology, vol. 48, pp , [22] Russell, J. A. A circumplex model of affect, Journal of Personality and Social Psychology, vol. 39, pp , [23] Thayer, R. E. The Biopsychology of Mood and Arousal. New York: Oxford University Press, [24] O. C. Meyers, A Mood-Based Music Classification and Exploration System, Master s thesis, Massachusetts Institute of Technology, [25] T. Jehan, Creating Music by Listening, PhD thesis, Massachusetts Institute of Technology, [26] C. McKay, Automatic Music Classification with jmir, PhD thesis, McGill University, [27] X. Amatriain, P. Arumi, D. Garcia, CLAM: A Framework for Efficient and Rapid Development of Cross platform Audio Applications. In Proc. of the 14th annual ACM int l conf. on Multimedia, pp , [28] The Echonest (2011). Dev Center [Online]. [29] K. Bischoff, C.S. Firan, R. Paiu, W. Nejdle, C. Laurier, M. Sordo, Music Mood and Theme Classification A Hybrid Approach. In Proc. of the 10th ISMIR (Int l Society for Music Information Retrieval) conf., pp , [30] StreamFinder (n.a.). Commercial Online Radio Station Data API Access [Online]. streamfinder.com/commercial-internetradio-api/ [31] Y. Rizk, N. Mitri, and M. Awad, A Local Mixture Based SVM for an Efficient Supervised Binary Classification, in International Joint Conference on Neural Networks, 2013, Dallas, TX. [32] Chih-Chung Chang and Chih-Jen Lin, LIBSVM : a library for support vector machines. ACM Transactions on Intelligent Systems and Technology, 2:27:1--27:27, [33] T. Bertin-Mahieux, D. Ellis, B. Whitman and P. Lamere, The Million Song Dataset, in Proc. 12th Int l Society for Music Information Retrieval Conf.,

Emotion Based Music Player

Emotion Based Music Player ISSN 2278 0211 (Online) Emotion Based Music Player Nikhil Zaware Tejas Rajgure Amey Bhadang D. D. Sapkal Professor, Department of Computer Engineering, Pune, India Abstract: Facial expression provides

More information

Session 2: 10 Year Vision session (11:00-12:20) - Tuesday. Session 3: Poster Highlights A (14:00-15:00) - Tuesday 20 posters (3minutes per poster)

Session 2: 10 Year Vision session (11:00-12:20) - Tuesday. Session 3: Poster Highlights A (14:00-15:00) - Tuesday 20 posters (3minutes per poster) Lessons from Collecting a Million Biometric Samples 109 Expression Robust 3D Face Recognition by Matching Multi-component Local Shape Descriptors on the Nasal and Adjoining Cheek Regions 177 Shared Representation

More information

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY A PATH FOR HORIZING YOUR INNOVATIVE WORK SMILE DETECTION WITH IMPROVED MISDETECTION RATE AND REDUCED FALSE ALARM RATE VRUSHALI

More information

Real-Time Face Detection and Tracking for High Resolution Smart Camera System

Real-Time Face Detection and Tracking for High Resolution Smart Camera System Digital Image Computing Techniques and Applications Real-Time Face Detection and Tracking for High Resolution Smart Camera System Y. M. Mustafah a,b, T. Shan a, A. W. Azman a,b, A. Bigdeli a, B. C. Lovell

More information

Combined Approach for Face Detection, Eye Region Detection and Eye State Analysis- Extended Paper

Combined Approach for Face Detection, Eye Region Detection and Eye State Analysis- Extended Paper International Journal of Engineering Research and Development e-issn: 2278-067X, p-issn: 2278-800X, www.ijerd.com Volume 10, Issue 9 (September 2014), PP.57-68 Combined Approach for Face Detection, Eye

More information

3D Face Recognition in Biometrics

3D Face Recognition in Biometrics 3D Face Recognition in Biometrics CHAO LI, ARMANDO BARRETO Electrical & Computer Engineering Department Florida International University 10555 West Flagler ST. EAS 3970 33174 USA {cli007, barretoa}@fiu.edu

More information

Sound Recognition. ~ CSE 352 Team 3 ~ Jason Park Evan Glover. Kevin Lui Aman Rawat. Prof. Anita Wasilewska

Sound Recognition. ~ CSE 352 Team 3 ~ Jason Park Evan Glover. Kevin Lui Aman Rawat. Prof. Anita Wasilewska Sound Recognition ~ CSE 352 Team 3 ~ Jason Park Evan Glover Kevin Lui Aman Rawat Prof. Anita Wasilewska What is Sound? Sound is a vibration that propagates as a typically audible mechanical wave of pressure

More information

Context Aware Computing

Context Aware Computing Context Aware Computing Context aware computing: the use of sensors and other sources of information about a user s context to provide more relevant information and services Context independent: acts exactly

More information

BIOMETRIC IDENTIFICATION USING 3D FACE SCANS

BIOMETRIC IDENTIFICATION USING 3D FACE SCANS BIOMETRIC IDENTIFICATION USING 3D FACE SCANS Chao Li Armando Barreto Craig Chin Jing Zhai Electrical and Computer Engineering Department Florida International University Miami, Florida, 33174, USA ABSTRACT

More information

RESEARCH AND DEVELOPMENT OF DSP-BASED FACE RECOGNITION SYSTEM FOR ROBOTIC REHABILITATION NURSING BEDS

RESEARCH AND DEVELOPMENT OF DSP-BASED FACE RECOGNITION SYSTEM FOR ROBOTIC REHABILITATION NURSING BEDS RESEARCH AND DEVELOPMENT OF DSP-BASED FACE RECOGNITION SYSTEM FOR ROBOTIC REHABILITATION NURSING BEDS Ming XING and Wushan CHENG College of Mechanical Engineering, Shanghai University of Engineering Science,

More information

S Pramod Kumar. Keywords Human emotion, physiological Signal, Emotion recognition, Hardwired logic, reprocessing.

S Pramod Kumar. Keywords Human emotion, physiological Signal, Emotion recognition, Hardwired logic, reprocessing. Volume 5, Issue 5, May 2015 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Human Emotion Recognition

More information

EFFICIENT ATTENDANCE MANAGEMENT SYSTEM USING FACE DETECTION AND RECOGNITION

EFFICIENT ATTENDANCE MANAGEMENT SYSTEM USING FACE DETECTION AND RECOGNITION EFFICIENT ATTENDANCE MANAGEMENT SYSTEM USING FACE DETECTION AND RECOGNITION 1 Arun.A.V, 2 Bhatath.S, 3 Chethan.N, 4 Manmohan.C.M, 5 Hamsaveni M 1,2,3,4,5 Department of Computer Science and Engineering,

More information

Music Mood Classification Using Audio Power and Audio Harmonicity Based on MPEG-7 Audio Features and Support Vector Machine

Music Mood Classification Using Audio Power and Audio Harmonicity Based on MPEG-7 Audio Features and Support Vector Machine Music Mood Classification Using Audio Power and Audio Harmonicity Based on MPEG-7 Audio Features and Support Vector Machine Johanes Andre Ridoean, Riyanarto Sarno, Dwi Sunaryo Department of Informatics

More information

3D Face Recognition System in Time Critical Security Applications

3D Face Recognition System in Time Critical Security Applications Middle-East Journal of Scientific Research 25 (7): 1619-1623, 2017 ISSN 1990-9233 IDOSI Publications, 2017 DOI: 10.5829/idosi.mejsr.2017.1619.1623 3D Face Recognition System in Time Critical Security Applications

More information

Real Time Multimodal Emotion Recognition System using Facial Landmarks and Hand over Face Gestures

Real Time Multimodal Emotion Recognition System using Facial Landmarks and Hand over Face Gestures Real Time Multimodal Emotion Recognition System using Facial Landmarks and Hand over Face Gestures Mahesh Krishnananda Prabhu and Dinesh Babu Jayagopi Abstract Over the last few years, emotional intelligent

More information

Drum Transcription Based on Independent Subspace Analysis

Drum Transcription Based on Independent Subspace Analysis Report for EE 391 Special Studies and Reports for Electrical Engineering Drum Transcription Based on Independent Subspace Analysis Yinyi Guo Center for Computer Research in Music and Acoustics, Stanford,

More information

Latest trends in sentiment analysis - A survey

Latest trends in sentiment analysis - A survey Latest trends in sentiment analysis - A survey Anju Rose G Punneliparambil PG Scholar Department of Computer Science & Engineering Govt. Engineering College, Thrissur, India anjurose.ar@gmail.com Abstract

More information

Audio Fingerprinting using Fractional Fourier Transform

Audio Fingerprinting using Fractional Fourier Transform Audio Fingerprinting using Fractional Fourier Transform Swati V. Sutar 1, D. G. Bhalke 2 1 (Department of Electronics & Telecommunication, JSPM s RSCOE college of Engineering Pune, India) 2 (Department,

More information

Student Attendance Monitoring System Via Face Detection and Recognition System

Student Attendance Monitoring System Via Face Detection and Recognition System IJSTE - International Journal of Science Technology & Engineering Volume 2 Issue 11 May 2016 ISSN (online): 2349-784X Student Attendance Monitoring System Via Face Detection and Recognition System Pinal

More information

MATLAB DIGITAL IMAGE/SIGNAL PROCESSING TITLES

MATLAB DIGITAL IMAGE/SIGNAL PROCESSING TITLES MATLAB DIGITAL IMAGE/SIGNAL PROCESSING TITLES -2018 S.NO PROJECT CODE 1 ITIMP01 2 ITIMP02 3 ITIMP03 4 ITIMP04 5 ITIMP05 6 ITIMP06 7 ITIMP07 8 ITIMP08 9 ITIMP09 `10 ITIMP10 11 ITIMP11 12 ITIMP12 13 ITIMP13

More information

Electric Guitar Pickups Recognition

Electric Guitar Pickups Recognition Electric Guitar Pickups Recognition Warren Jonhow Lee warrenjo@stanford.edu Yi-Chun Chen yichunc@stanford.edu Abstract Electric guitar pickups convert vibration of strings to eletric signals and thus direcly

More information

Music Recommendation using Recurrent Neural Networks

Music Recommendation using Recurrent Neural Networks Music Recommendation using Recurrent Neural Networks Ashustosh Choudhary * ashutoshchou@cs.umass.edu Mayank Agarwal * mayankagarwa@cs.umass.edu Abstract A large amount of information is contained in the

More information

Image Forgery Detection Using Svm Classifier

Image Forgery Detection Using Svm Classifier Image Forgery Detection Using Svm Classifier Anita Sahani 1, K.Srilatha 2 M.E. Student [Embedded System], Dept. Of E.C.E., Sathyabama University, Chennai, India 1 Assistant Professor, Dept. Of E.C.E, Sathyabama

More information

Emotion Based Music Player Using Facial Recognition

Emotion Based Music Player Using Facial Recognition Emotion Based Music Player Using Facial Recognition Prof. Vijaykumar R. Ghule, Abhijeet B. Benke, Shubham S. Jadhav, Swapnil A. Joshi Asst. Professor, Smt.Kashibai Navale College Of Engineering, Pune,

More information

Content Based Image Retrieval Using Color Histogram

Content Based Image Retrieval Using Color Histogram Content Based Image Retrieval Using Color Histogram Nitin Jain Assistant Professor, Lokmanya Tilak College of Engineering, Navi Mumbai, India. Dr. S. S. Salankar Professor, G.H. Raisoni College of Engineering,

More information

An Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi

An Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi An Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi Department of E&TC Engineering,PVPIT,Bavdhan,Pune ABSTRACT: In the last decades vehicle license plate recognition systems

More information

Accurate Emotion Detection of Digital Images Using Bezier Curves

Accurate Emotion Detection of Digital Images Using Bezier Curves Accurate Emotion Detection of Digital Images Using Bezier Curves C.Karuna Sharma, T.Aswini, A.Vinodhini, V.Selvi Abstract Image capturing and detecting the emotions of face that have unconstrained level

More information

Auto-tagging The Facebook

Auto-tagging The Facebook Auto-tagging The Facebook Jonathan Michelson and Jorge Ortiz Stanford University 2006 E-mail: JonMich@Stanford.edu, jorge.ortiz@stanford.com Introduction For those not familiar, The Facebook is an extremely

More information

An Hybrid MLP-SVM Handwritten Digit Recognizer

An Hybrid MLP-SVM Handwritten Digit Recognizer An Hybrid MLP-SVM Handwritten Digit Recognizer A. Bellili ½ ¾ M. Gilloux ¾ P. Gallinari ½ ½ LIP6, Université Pierre et Marie Curie ¾ La Poste 4, Place Jussieu 10, rue de l Ile Mabon, BP 86334 75252 Paris

More information

Study Impact of Architectural Style and Partial View on Landmark Recognition

Study Impact of Architectural Style and Partial View on Landmark Recognition Study Impact of Architectural Style and Partial View on Landmark Recognition Ying Chen smileyc@stanford.edu 1. Introduction Landmark recognition in image processing is one of the important object recognition

More information

INTRODUCTION TO DEEP LEARNING. Steve Tjoa June 2013

INTRODUCTION TO DEEP LEARNING. Steve Tjoa June 2013 INTRODUCTION TO DEEP LEARNING Steve Tjoa kiemyang@gmail.com June 2013 Acknowledgements http://ufldl.stanford.edu/wiki/index.php/ UFLDL_Tutorial http://youtu.be/ayzoubkuf3m http://youtu.be/zmnoatzigik 2

More information

SOUND SOURCE RECOGNITION AND MODELING

SOUND SOURCE RECOGNITION AND MODELING SOUND SOURCE RECOGNITION AND MODELING CASA seminar, summer 2000 Antti Eronen antti.eronen@tut.fi Contents: Basics of human sound source recognition Timbre Voice recognition Recognition of environmental

More information

Classification of Voltage Sag Using Multi-resolution Analysis and Support Vector Machine

Classification of Voltage Sag Using Multi-resolution Analysis and Support Vector Machine Journal of Clean Energy Technologies, Vol. 4, No. 3, May 2016 Classification of Voltage Sag Using Multi-resolution Analysis and Support Vector Machine Hanim Ismail, Zuhaina Zakaria, and Noraliza Hamzah

More information

Speech/Music Change Point Detection using Sonogram and AANN

Speech/Music Change Point Detection using Sonogram and AANN International Journal of Information & Computation Technology. ISSN 0974-2239 Volume 6, Number 1 (2016), pp. 45-49 International Research Publications House http://www. irphouse.com Speech/Music Change

More information

Understanding the city to make it smart

Understanding the city to make it smart Understanding the city to make it smart Roberta De Michele and Marco Furini Communication and Economics Department Universty of Modena and Reggio Emilia, Reggio Emilia, 42121, Italy, marco.furini@unimore.it

More information

Applications of Music Processing

Applications of Music Processing Lecture Music Processing Applications of Music Processing Christian Dittmar International Audio Laboratories Erlangen christian.dittmar@audiolabs-erlangen.de Singing Voice Detection Important pre-requisite

More information

Rhythmic Similarity -- a quick paper review. Presented by: Shi Yong March 15, 2007 Music Technology, McGill University

Rhythmic Similarity -- a quick paper review. Presented by: Shi Yong March 15, 2007 Music Technology, McGill University Rhythmic Similarity -- a quick paper review Presented by: Shi Yong March 15, 2007 Music Technology, McGill University Contents Introduction Three examples J. Foote 2001, 2002 J. Paulus 2002 S. Dixon 2004

More information

Live Hand Gesture Recognition using an Android Device

Live Hand Gesture Recognition using an Android Device Live Hand Gesture Recognition using an Android Device Mr. Yogesh B. Dongare Department of Computer Engineering. G.H.Raisoni College of Engineering and Management, Ahmednagar. Email- yogesh.dongare05@gmail.com

More information

Classification of ships using autocorrelation technique for feature extraction of the underwater acoustic noise

Classification of ships using autocorrelation technique for feature extraction of the underwater acoustic noise Classification of ships using autocorrelation technique for feature extraction of the underwater acoustic noise Noha KORANY 1 Alexandria University, Egypt ABSTRACT The paper applies spectral analysis to

More information

Keyword: Morphological operation, template matching, license plate localization, character recognition.

Keyword: Morphological operation, template matching, license plate localization, character recognition. Volume 4, Issue 11, November 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Automatic

More information

A Comparison Study of Image Descriptors on Low- Resolution Face Image Verification

A Comparison Study of Image Descriptors on Low- Resolution Face Image Verification A Comparison Study of Image Descriptors on Low- Resolution Face Image Verification Gittipat Jetsiktat, Sasipa Panthuwadeethorn and Suphakant Phimoltares Advanced Virtual and Intelligent Computing (AVIC)

More information

Performance study of Text-independent Speaker identification system using MFCC & IMFCC for Telephone and Microphone Speeches

Performance study of Text-independent Speaker identification system using MFCC & IMFCC for Telephone and Microphone Speeches Performance study of Text-independent Speaker identification system using & I for Telephone and Microphone Speeches Ruchi Chaudhary, National Technical Research Organization Abstract: A state-of-the-art

More information

The Automatic Classification Problem. Perceptrons, SVMs, and Friends: Some Discriminative Models for Classification

The Automatic Classification Problem. Perceptrons, SVMs, and Friends: Some Discriminative Models for Classification Perceptrons, SVMs, and Friends: Some Discriminative Models for Classification Parallel to AIMA 8., 8., 8.6.3, 8.9 The Automatic Classification Problem Assign object/event or sequence of objects/events

More information

CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS. Kuan-Chuan Peng and Tsuhan Chen

CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS. Kuan-Chuan Peng and Tsuhan Chen CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS Kuan-Chuan Peng and Tsuhan Chen Cornell University School of Electrical and Computer Engineering Ithaca, NY 14850

More information

AUTOMATIC EYE DETECTION IN FACIAL IMAGES WITH UNCONSTRAINED BACKGROUNDS

AUTOMATIC EYE DETECTION IN FACIAL IMAGES WITH UNCONSTRAINED BACKGROUNDS AUTOMATIC EYE DETECTION IN FACIAL IMAGES WITH UNCONSTRAINED BACKGROUNDS Dr John Cowell Dept. of Computer Science, De Montfort University, The Gateway, Leicester, LE1 9BH England, jcowell@dmu.ac.uk ABSTRACT

More information

Face Detection: A Literature Review

Face Detection: A Literature Review Face Detection: A Literature Review Dr.Vipulsangram.K.Kadam 1, Deepali G. Ganakwar 2 Professor, Department of Electronics Engineering, P.E.S. College of Engineering, Nagsenvana Aurangabad, Maharashtra,

More information

Pose Invariant Face Recognition

Pose Invariant Face Recognition Pose Invariant Face Recognition Fu Jie Huang Zhihua Zhou Hong-Jiang Zhang Tsuhan Chen Electrical and Computer Engineering Department Carnegie Mellon University jhuangfu@cmu.edu State Key Lab for Novel

More information

Campus Location Recognition using Audio Signals

Campus Location Recognition using Audio Signals 1 Campus Location Recognition using Audio Signals James Sun,Reid Westwood SUNetID:jsun2015,rwestwoo Email: jsun2015@stanford.edu, rwestwoo@stanford.edu I. INTRODUCTION People use sound both consciously

More information

Audio Similarity. Mark Zadel MUMT 611 March 8, Audio Similarity p.1/23

Audio Similarity. Mark Zadel MUMT 611 March 8, Audio Similarity p.1/23 Audio Similarity Mark Zadel MUMT 611 March 8, 2004 Audio Similarity p.1/23 Overview MFCCs Foote Content-Based Retrieval of Music and Audio (1997) Logan, Salomon A Music Similarity Function Based On Signal

More information

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES International Journal of Advanced Research in Engineering and Technology (IJARET) Volume 9, Issue 3, May - June 2018, pp. 177 185, Article ID: IJARET_09_03_023 Available online at http://www.iaeme.com/ijaret/issues.asp?jtype=ijaret&vtype=9&itype=3

More information

FACE VERIFICATION SYSTEM IN MOBILE DEVICES BY USING COGNITIVE SERVICES

FACE VERIFICATION SYSTEM IN MOBILE DEVICES BY USING COGNITIVE SERVICES International Journal of Intelligent Systems and Applications in Engineering Advanced Technology and Science ISSN:2147-67992147-6799 www.atscience.org/ijisae Original Research Paper FACE VERIFICATION SYSTEM

More information

Distance Estimation and Localization of Sound Sources in Reverberant Conditions using Deep Neural Networks

Distance Estimation and Localization of Sound Sources in Reverberant Conditions using Deep Neural Networks Distance Estimation and Localization of Sound Sources in Reverberant Conditions using Deep Neural Networks Mariam Yiwere 1 and Eun Joo Rhee 2 1 Department of Computer Engineering, Hanbat National University,

More information

An Accurate Algorithm for Generating a Music Playlist based on Facial Expressions

An Accurate Algorithm for Generating a Music Playlist based on Facial Expressions An Accurate Algorithm for Generating a Music Playlist based on Facial Expressions Anukriti Dureha Computer Science and Engineering Department Amity School of Engineering & Technology, Amity University,

More information

Moodify. A music search engine by. Rock, Saru, Vincent, Walter

Moodify. A music search engine by. Rock, Saru, Vincent, Walter Moodify A music search engine by Rock, Saru, Vincent, Walter Explore music through mood Create a Web App that recommends songs based on how the user is feeling - 7 supported moods Joy Love Sad Surprise

More information

SONG RETRIEVAL SYSTEM USING HIDDEN MARKOV MODELS

SONG RETRIEVAL SYSTEM USING HIDDEN MARKOV MODELS SONG RETRIEVAL SYSTEM USING HIDDEN MARKOV MODELS AKSHAY CHANDRASHEKARAN ANOOP RAMAKRISHNA akshayc@cmu.edu anoopr@andrew.cmu.edu ABHISHEK JAIN GE YANG ajain2@andrew.cmu.edu younger@cmu.edu NIDHI KOHLI R

More information

AUTOMATED MALARIA PARASITE DETECTION BASED ON IMAGE PROCESSING PROJECT REFERENCE NO.: 38S1511

AUTOMATED MALARIA PARASITE DETECTION BASED ON IMAGE PROCESSING PROJECT REFERENCE NO.: 38S1511 AUTOMATED MALARIA PARASITE DETECTION BASED ON IMAGE PROCESSING PROJECT REFERENCE NO.: 38S1511 COLLEGE : BANGALORE INSTITUTE OF TECHNOLOGY, BENGALURU BRANCH : COMPUTER SCIENCE AND ENGINEERING GUIDE : DR.

More information

An Automated Face Reader for Fatigue Detection

An Automated Face Reader for Fatigue Detection An Automated Face Reader for Fatigue Detection Haisong Gu Dept. of Computer Science University of Nevada Reno Haisonggu@ieee.org Qiang Ji Dept. of ECSE Rensselaer Polytechnic Institute qji@ecse.rpi.edu

More information

Mel Spectrum Analysis of Speech Recognition using Single Microphone

Mel Spectrum Analysis of Speech Recognition using Single Microphone International Journal of Engineering Research in Electronics and Communication Mel Spectrum Analysis of Speech Recognition using Single Microphone [1] Lakshmi S.A, [2] Cholavendan M [1] PG Scholar, Sree

More information

A Mathematical model for the determination of distance of an object in a 2D image

A Mathematical model for the determination of distance of an object in a 2D image A Mathematical model for the determination of distance of an object in a 2D image Deepu R 1, Murali S 2,Vikram Raju 3 Maharaja Institute of Technology Mysore, Karnataka, India rdeepusingh@mitmysore.in

More information

Convolutional Neural Networks: Real Time Emotion Recognition

Convolutional Neural Networks: Real Time Emotion Recognition Convolutional Neural Networks: Real Time Emotion Recognition Bruce Nguyen, William Truong, Harsha Yeddanapudy Motivation: Machine emotion recognition has long been a challenge and popular topic in the

More information

Book Cover Recognition Project

Book Cover Recognition Project Book Cover Recognition Project Carolina Galleguillos Department of Computer Science University of California San Diego La Jolla, CA 92093-0404 cgallegu@cs.ucsd.edu Abstract The purpose of this project

More information

Biometrics Final Project Report

Biometrics Final Project Report Andres Uribe au2158 Introduction Biometrics Final Project Report Coin Counter The main objective for the project was to build a program that could count the coins money value in a picture. The work was

More information

Main Subject Detection of Image by Cropping Specific Sharp Area

Main Subject Detection of Image by Cropping Specific Sharp Area Main Subject Detection of Image by Cropping Specific Sharp Area FOTIOS C. VAIOULIS 1, MARIOS S. POULOS 1, GEORGE D. BOKOS 1 and NIKOLAOS ALEXANDRIS 2 Department of Archives and Library Science Ionian University

More information

AI Application Processing Requirements

AI Application Processing Requirements AI Application Processing Requirements 1 Low Medium High Sensor analysis Activity Recognition (motion sensors) Stress Analysis or Attention Analysis Audio & sound Speech Recognition Object detection Computer

More information

A Novel Fuzzy Neural Network Based Distance Relaying Scheme

A Novel Fuzzy Neural Network Based Distance Relaying Scheme 902 IEEE TRANSACTIONS ON POWER DELIVERY, VOL. 15, NO. 3, JULY 2000 A Novel Fuzzy Neural Network Based Distance Relaying Scheme P. K. Dash, A. K. Pradhan, and G. Panda Abstract This paper presents a new

More information

Multimodal Face Recognition using Hybrid Correlation Filters

Multimodal Face Recognition using Hybrid Correlation Filters Multimodal Face Recognition using Hybrid Correlation Filters Anamika Dubey, Abhishek Sharma Electrical Engineering Department, Indian Institute of Technology Roorkee, India {ana.iitr, abhisharayiya}@gmail.com

More information

Determining Guava Freshness by Flicking Signal Recognition Using HMM Acoustic Models

Determining Guava Freshness by Flicking Signal Recognition Using HMM Acoustic Models Determining Guava Freshness by Flicking Signal Recognition Using HMM Acoustic Models Rong Phoophuangpairoj applied signal processing to animal sounds [1]-[3]. In speech recognition, digitized human speech

More information

A New Fake Iris Detection Method

A New Fake Iris Detection Method A New Fake Iris Detection Method Xiaofu He 1, Yue Lu 1, and Pengfei Shi 2 1 Department of Computer Science and Technology, East China Normal University, Shanghai 200241, China {xfhe,ylu}@cs.ecnu.edu.cn

More information

A SURVEY ON HAND GESTURE RECOGNITION

A SURVEY ON HAND GESTURE RECOGNITION A SURVEY ON HAND GESTURE RECOGNITION U.K. Jaliya 1, Dr. Darshak Thakore 2, Deepali Kawdiya 3 1 Assistant Professor, Department of Computer Engineering, B.V.M, Gujarat, India 2 Assistant Professor, Department

More information

Research on Hand Gesture Recognition Using Convolutional Neural Network

Research on Hand Gesture Recognition Using Convolutional Neural Network Research on Hand Gesture Recognition Using Convolutional Neural Network Tian Zhaoyang a, Cheng Lee Lung b a Department of Electronic Engineering, City University of Hong Kong, Hong Kong, China E-mail address:

More information

Dimension Reduction of the Modulation Spectrogram for Speaker Verification

Dimension Reduction of the Modulation Spectrogram for Speaker Verification Dimension Reduction of the Modulation Spectrogram for Speaker Verification Tomi Kinnunen Speech and Image Processing Unit Department of Computer Science University of Joensuu, Finland Kong Aik Lee and

More information

Demosaicing Algorithm for Color Filter Arrays Based on SVMs

Demosaicing Algorithm for Color Filter Arrays Based on SVMs www.ijcsi.org 212 Demosaicing Algorithm for Color Filter Arrays Based on SVMs Xiao-fen JIA, Bai-ting Zhao School of Electrical and Information Engineering, Anhui University of Science & Technology Huainan

More information

Song Shuffler Based on Automatic Human Emotion Recognition

Song Shuffler Based on Automatic Human Emotion Recognition Recent Advances in Technology and Engineering (RATE-2017) 6 th National Conference by TJIT, Bangalore International Journal of Science, Engineering and Technology An Open Access Journal Song Shuffler Based

More information

A Proposal for Security Oversight at Automated Teller Machine System

A Proposal for Security Oversight at Automated Teller Machine System International Journal of Engineering Research and Development e-issn: 2278-067X, p-issn: 2278-800X, www.ijerd.com Volume 10, Issue 6 (June 2014), PP.18-25 A Proposal for Security Oversight at Automated

More information

Hand Gesture Recognition System Using Camera

Hand Gesture Recognition System Using Camera Hand Gesture Recognition System Using Camera Viraj Shinde, Tushar Bacchav, Jitendra Pawar, Mangesh Sanap B.E computer engineering,navsahyadri Education Society sgroup of Institutions,pune. Abstract - In

More information

Design and Implementation of an Audio Classification System Based on SVM

Design and Implementation of an Audio Classification System Based on SVM Available online at www.sciencedirect.com Procedia ngineering 15 (011) 4031 4035 Advanced in Control ngineering and Information Science Design and Implementation of an Audio Classification System Based

More information

Multi-User Blood Alcohol Content Estimation in a Realistic Simulator using Artificial Neural Networks and Support Vector Machines

Multi-User Blood Alcohol Content Estimation in a Realistic Simulator using Artificial Neural Networks and Support Vector Machines Multi-User Blood Alcohol Content Estimation in a Realistic Simulator using Artificial Neural Networks and Support Vector Machines ROBINEL Audrey & PUZENAT Didier {arobinel, dpuzenat}@univ-ag.fr Laboratoire

More information

An Optimization of Audio Classification and Segmentation using GASOM Algorithm

An Optimization of Audio Classification and Segmentation using GASOM Algorithm An Optimization of Audio Classification and Segmentation using GASOM Algorithm Dabbabi Karim, Cherif Adnen Research Unity of Processing and Analysis of Electrical and Energetic Systems Faculty of Sciences

More information

Segmentation using Saturation Thresholding and its Application in Content-Based Retrieval of Images

Segmentation using Saturation Thresholding and its Application in Content-Based Retrieval of Images Segmentation using Saturation Thresholding and its Application in Content-Based Retrieval of Images A. Vadivel 1, M. Mohan 1, Shamik Sural 2 and A.K.Majumdar 1 1 Department of Computer Science and Engineering,

More information

Face Detection System on Ada boost Algorithm Using Haar Classifiers

Face Detection System on Ada boost Algorithm Using Haar Classifiers Vol.2, Issue.6, Nov-Dec. 2012 pp-3996-4000 ISSN: 2249-6645 Face Detection System on Ada boost Algorithm Using Haar Classifiers M. Gopi Krishna, A. Srinivasulu, Prof (Dr.) T.K.Basak 1, 2 Department of Electronics

More information

Image Extraction using Image Mining Technique

Image Extraction using Image Mining Technique IOSR Journal of Engineering (IOSRJEN) e-issn: 2250-3021, p-issn: 2278-8719 Vol. 3, Issue 9 (September. 2013), V2 PP 36-42 Image Extraction using Image Mining Technique Prof. Samir Kumar Bandyopadhyay,

More information

Text Emotion Detection using Neural Network

Text Emotion Detection using Neural Network International Journal of Engineering Research and Technology. ISSN 0974-3154 Volume 7, Number 2 (2014), pp. 153-159 International Research Publication House http://www.irphouse.com Text Emotion Detection

More information

An Un-awarely Collected Real World Face Database: The ISL-Door Face Database

An Un-awarely Collected Real World Face Database: The ISL-Door Face Database An Un-awarely Collected Real World Face Database: The ISL-Door Face Database Hazım Kemal Ekenel, Rainer Stiefelhagen Interactive Systems Labs (ISL), Universität Karlsruhe (TH), Am Fasanengarten 5, 76131

More information

Robust Hand Gesture Recognition for Robotic Hand Control

Robust Hand Gesture Recognition for Robotic Hand Control Robust Hand Gesture Recognition for Robotic Hand Control Ankit Chaudhary Robust Hand Gesture Recognition for Robotic Hand Control 123 Ankit Chaudhary Department of Computer Science Northwest Missouri State

More information

AN ANALYSIS OF SPEECH RECOGNITION PERFORMANCE BASED UPON NETWORK LAYERS AND TRANSFER FUNCTIONS

AN ANALYSIS OF SPEECH RECOGNITION PERFORMANCE BASED UPON NETWORK LAYERS AND TRANSFER FUNCTIONS AN ANALYSIS OF SPEECH RECOGNITION PERFORMANCE BASED UPON NETWORK LAYERS AND TRANSFER FUNCTIONS Kuldeep Kumar 1, R. K. Aggarwal 1 and Ankita Jain 2 1 Department of Computer Engineering, National Institute

More information

Realizing Human-Centricity: Data-Driven Services

Realizing Human-Centricity: Data-Driven Services Realizing Human-Centricity: Data-Driven Services Ajay Chander R&D Lead, Data Driven Life Innovations Fujitsu Laboratories of America January 22, 2014 INTERNAL USE ONLY Copyright 2014 FUJITSU LIMITED Context:

More information

Multimedia Forensics

Multimedia Forensics Multimedia Forensics Using Mathematics and Machine Learning to Determine an Image's Source and Authenticity Matthew C. Stamm Multimedia & Information Security Lab (MISL) Department of Electrical and Computer

More information

Classification in Image processing: A Survey

Classification in Image processing: A Survey Classification in Image processing: A Survey Rashmi R V, Sheela Sridhar Department of computer science and Engineering, B.N.M.I.T, Bangalore-560070 Department of computer science and Engineering, B.N.M.I.T,

More information

Support Vector Machine Classification of Snow Radar Interface Layers

Support Vector Machine Classification of Snow Radar Interface Layers Support Vector Machine Classification of Snow Radar Interface Layers Michael Johnson December 15, 2011 Abstract Operation IceBridge is a NASA funded survey of polar sea and land ice consisting of multiple

More information

Blue Eyes Technology with Electric Imp Explorer Kit Ankita Shaily*, Saurabh Anand I.

Blue Eyes Technology with Electric Imp Explorer Kit Ankita Shaily*, Saurabh Anand I. ABSTRACT 2018 IJSRST Volume 4 Issue6 Print ISSN: 2395-6011 Online ISSN: 2395-602X National Conference on Smart Computation and Technology in Conjunction with The Smart City Convergence 2018 Blue Eyes Technology

More information

Electronic disguised voice identification based on Mel- Frequency Cepstral Coefficient analysis

Electronic disguised voice identification based on Mel- Frequency Cepstral Coefficient analysis International Journal of Scientific and Research Publications, Volume 5, Issue 11, November 2015 412 Electronic disguised voice identification based on Mel- Frequency Cepstral Coefficient analysis Shalate

More information

How to build an autonomous anything

How to build an autonomous anything How to build an autonomous anything Loren Shure Application Engineering MathWorks 2015 The MathWorks, Inc. 1 2 3 4 5 6 7 Autonomous Technology 8 Autonomous Technology Having the power for self-governance

More information

Integrated Driving Aware System in the Real-World: Sensing, Computing and Feedback

Integrated Driving Aware System in the Real-World: Sensing, Computing and Feedback Integrated Driving Aware System in the Real-World: Sensing, Computing and Feedback Jung Wook Park HCI Institute Carnegie Mellon University 5000 Forbes Avenue Pittsburgh, PA, USA, 15213 jungwoop@andrew.cmu.edu

More information

Human Authentication from Brain EEG Signals using Machine Learning

Human Authentication from Brain EEG Signals using Machine Learning Volume 118 No. 24 2018 ISSN: 1314-3395 (on-line version) url: http://www.acadpubl.eu/hub/ http://www.acadpubl.eu/hub/ Human Authentication from Brain EEG Signals using Machine Learning Urmila Kalshetti,

More information

MUSICAL GENRE CLASSIFICATION OF AUDIO DATA USING SOURCE SEPARATION TECHNIQUES. P.S. Lampropoulou, A.S. Lampropoulos and G.A.

MUSICAL GENRE CLASSIFICATION OF AUDIO DATA USING SOURCE SEPARATION TECHNIQUES. P.S. Lampropoulou, A.S. Lampropoulos and G.A. MUSICAL GENRE CLASSIFICATION OF AUDIO DATA USING SOURCE SEPARATION TECHNIQUES P.S. Lampropoulou, A.S. Lampropoulos and G.A. Tsihrintzis Department of Informatics, University of Piraeus 80 Karaoli & Dimitriou

More information

Classification of Road Images for Lane Detection

Classification of Road Images for Lane Detection Classification of Road Images for Lane Detection Mingyu Kim minkyu89@stanford.edu Insun Jang insunj@stanford.edu Eunmo Yang eyang89@stanford.edu 1. Introduction In the research on autonomous car, it is

More information

A Smart Home Design and Implementation Based on Kinect

A Smart Home Design and Implementation Based on Kinect 2018 International Conference on Physics, Computing and Mathematical Modeling (PCMM 2018) ISBN: 978-1-60595-549-0 A Smart Home Design and Implementation Based on Kinect Jin-wen DENG 1,2, Xue-jun ZHANG

More information

A TWO-PART PREDICTIVE CODER FOR MULTITASK SIGNAL COMPRESSION. Scott Deeann Chen and Pierre Moulin

A TWO-PART PREDICTIVE CODER FOR MULTITASK SIGNAL COMPRESSION. Scott Deeann Chen and Pierre Moulin A TWO-PART PREDICTIVE CODER FOR MULTITASK SIGNAL COMPRESSION Scott Deeann Chen and Pierre Moulin University of Illinois at Urbana-Champaign Department of Electrical and Computer Engineering 5 North Mathews

More information

A Review of Related Work on Machine Learning in Semiconductor Manufacturing and Assembly Lines

A Review of Related Work on Machine Learning in Semiconductor Manufacturing and Assembly Lines A Review of Related Work on Machine Learning in Semiconductor Manufacturing and Assembly Lines DI Darko Stanisavljevic VIRTUAL VEHICLE DI Michael Spitzer VIRTUAL VEHICLE i-know 16 18.-19.10.2016, Graz

More information

Human-Computer Intelligent Interaction: A Survey

Human-Computer Intelligent Interaction: A Survey Human-Computer Intelligent Interaction: A Survey Michael Lew 1, Erwin M. Bakker 1, Nicu Sebe 2, and Thomas S. Huang 3 1 LIACS Media Lab, Leiden University, The Netherlands 2 ISIS Group, University of Amsterdam,

More information