HiPOP
A high performance point of purchase content deleiver system. This project explores the feasibility of a user centric delivery of point-of-purchase information using biometric data capture, intelligent analysis of facial data, height, weight, body type, age, gender and other forms of data that can be directly captured in a non-invasive manner


hipop


Such a system must have an inherent intelligence that is ambient, and ubiquitous – allowing for interpretation of a wide variety of stimuli and that can be easily collected. The systems intelligences must have offer a range of options that can be autonomously responsive and give meaningful responses to the visual and sensor cues.
The aim of this research is to study the feasibility of a system which is able to deliver user-aware and user-centric point-of-purchase information. The system must have an inherent intelligence that is ambient and ubiquitous. It must be able to give autonomous and meaningful responses to visual and sensor cues. Such a system prefigures an information delivery for advertising, social communication and even for forms emergency communications that may be needed in a public or social environment. 



The video above demonstrates an early stage of the HiPOP project. in this video we can see the abilty to detect gender and smile. This technique currently uses an image processing approach by identifying shapes within an image field using Viola and Jones

Open CV Haar-like features application [1], [2],[3] and a “feret” database [4] of facial image andsupport vector machine (LibSVM) [3] to classify the faces to glean attributes such as gender, or other individual characteristics. This system will be able to capture information such as facial data, height, weight, gender, race, location, gaze, product label etc. In runtime mode the system Intelligently analyses such data to provide better solution for advertising and social communication. The Image Progressing requires: Segmenting out face rectangles; Scaling to 24*24 grayscale image and Equalising the histogram to increase contrast. An OpenCV library is used to detect and segment faces from video images:1.Using a cascade of boosted classifiers working with haar-like features.2.Training classifiers by a database of face and non-face images.3.Input images are scanned at different scales to find regions that are likely to contain faces. Gender | Glasses | Smile Detection We use the SVM classifier method: data points are dealt with as a p-dimensional vector. We use the library LibSVM.

Gender Detection
1. Database  FERET http://www.nist.gov/humanid/colorferet  FA | FB | QR | QL | HL | HR 2. Rate of accuracy  FB(dvd2) : 246/268 = 91.791%

Glasses Detection
1. Database  FERET (FA | FB | QR | QL)  GENKI  WWW (from internet) 2. Rate of Accuracy Database  FB(dvd2) : 248/268 = 92.5373 %

Smile Detection
1. Database  GENKI http://mplab.ucsd.edu, The MPLab GENKI-4K Dataset
2. Rate of Accuracy  GENKI  File2113.jpg – file2440.jpg : 278/323 = 86.0681%  File2000.jpg – file2500.jpg : 417/492 = 84.7561%
3. Result is better when tested with webcams.


Future work: Age Estimation 1. Using PCA to get the age vector( facial features of every age )  2. Extract the wrinkle features using edge detection and then using classifier to separate people of different ages。 Further work will move towards the use of a 3D mesh vector template using elastic bunch graph matching that tags facial feature and performs a local feature analysis (LFA) of ordered matches, followed by a surface texture analysis (STA) that allows for detection of skin features [5][6]. The detection of basic emotional states, anger, disgust, fear, happiness, sadness, and surprise[7]. Additional work will focus on personal attribute cues such as hairstyles, clothing, and other cues that allow for estimation of demographic groupings and "tribe" associations.


1. Viola, P., & Jones, M. (2001). Robust real-time object detection. Paper presented at the Second International Workshop on Theories of Visual Modelling Learning, Computing, and Sampling

2. Bradski, G. and Kaehler, A., (2008). Learning OpenCV. OReilly.

3. Burges, C. J.C., (1998) A Tutorial on Support Vector Machines for Pattern Recognition. Data Mining and Knowledge Discovery 2, 121-167

4. http://www.nist.gov/huma nid/colorferet

5. Wiskott, L.; Fellous, J.-M.; Kuiger, N.; von der Malsburg, C. (1997) Face recognition by elastic bunch graph matching, Pattern Analysis and Machine Intelligence, IEEE Transactions on Machine Intelligence, Volume: 19 Issue:7 Pages 775 - 779

6. Bronstein, A. M.; Bronstein, M. M., and Kimmel, R. (2005). "Three-dimensional face recognition". International Journal of Computer Vision (IJCV) 64 (1): 5–30

7. Ekman, P., (1999), "Basic Emotions", in Dalgleish, T; Power, M, Handbook of Cognition and Emotion, Sussex, UK: John Wiley & Sons, http://www.paulekman.com/wp-content/uploads/2009/02/Basic-Emotions.pdf