Effective March 4, 2023
Dear user,
We want to provide you with transparency and clarity regarding our emotional artificial intelligence system. This document details how it works, the methods for extracting data, information about the sources of data, methods used for evaluating the system, as well as any known limitations or biases. We also provide information on potential risks and benefits of using the emotional AI system, as well as guidelines for safe and effective use.
The technology developed and distributed by Cynny S.p.A. (hereinafter “Cynny“), called “MorphCast® EMOTION AI HTML5 SDK” and/or the other available products, services and features that integrate it (e.g. MorphCast Web Apps or MorphCast Emotion AI Interactive Video Platform) is an artificial intelligence software capable of processing images from a video stream that allows you to establish, with varying degrees of probability and accuracy (see Annex A to this document), the emotional state of the framed subject and some other apparent characteristics such as, for example, the cut or color of the hair, the apparent age and gender, facial emotion expressions, in any case without any identification of the subject (hereinafter, the “Software“). For a complete list of features extracted from the Software, please refer to Annex B to this document.
This technology, which Cynny makes available to its users (qualified as “Business Users”, pursuant to the applicable “Terms of Use”, and hereinafter referred to as “Customers“) has been specifically designed and developed so as not to identify the framed subjects. These subjects are the end users of the Software chosen by the Customer (hereinafter, the “End Users“), with whom Cynny has no relationship whatsoever.
The Software can be used autonomously by the Customer or be activated, by the Customer’s choice, in the context of use of other services provided by Cynny, whose terms of use and privacy policies are available at this link: mission page. As regards the registration data of the Customers, the relative Privacy Policy will also apply, available at this link: Emotion AI – Privacy Policy.
How it works
The Software works as a Javascript library embedded in a web page usually belonging to the Customers’ domain, running in the browser of the End User device, which, as indicated above, has a relationship exclusively with the Customer (and not with Cynny).
The Software does not process biometric data that allow the univocal identification of a subject, but only an instant (volatile) processing of personal data during the processing of images (frames), read from the video stream (for example of a camera, which is used only as a sensor and not as an image and/or video recorder) provided by the End User’s browser. Cynny does not carry out any processing or control over such personal data as they are processed through the Software automatically and directly on the End User’s device.
Accordingly, the use of the Software is subject to the following terms of use, both of which are beyond Cynny’s control: (1) the terms of use of the End User’s browser in which the Software is incorporated and (2) any software or third party API required to allow the web page to function.
Notwithstanding the foregoing, as the supplier of the Software, Cynny (a) warrants that it has adopted all appropriate security measures in the design of the Software also with regard to the processing of personal data, taking into account privacy-by-design approach and ensuring adequate security of the software itself and; (b) makes itself available to collaborate with the Customer (who, in case it uses the extracted data autonomously, assumes the title of Data Controller pursuant to EU Regulation 2016/679, hereinafter “GDPR“) for any processing of personal data relating to the End User, showing and explaining all the functions of the Software, in order to enable it to carry out a precise risk analysis with respect to the fundamental rights and freedoms of the persons concerned (including any impact assessments, where necessary and / or appropriate, in the opinion of the Data Controller). In fact, it will be the Data Controller who will have to establish the purposes, methods and legal basis of the processing, the time and place (physical or virtual) for the storage of personal data relating to End Users, also providing them with specific online information. with the applicable legislation for the protection of personal data (including the GDPR).
According to the provisions of the Terms of Use of the Software (available here: Terms of Use), the same is aimed solely and exclusively at professional users, with the express exclusion of consumers and, therefore, of the use of the Software for domestic / personal purposes. In any case, whatever the type of processing carried out by the Customer, as Data Controller, the consequences deriving from the application of the applicable legislation, including that for the protection of personal data, are exclusively the responsibility of the Data Controller.
It is also specified that the Software operates integrated in a private corporate hardware / software context of the Customer’s End User (on end-point resources, including web pages belonging to the Customer’s domain and therefore under its control) equipped with security necessary to avoid the compromise of the data affected by the processing by the Software. In line with the foregoing, the consequent responsibilities relating to the processing of personal data of End Users by the Customers fall exclusively on the Customer, which therefore undertakes, in accordance with the provisions of the “Indemnification” section of the terms of use of the Software, to fully indemnify Cynny from any prejudicial consequence that may arise in this regard.
As per the emotional data made available to the Customers by the Software through the “Data Storage” module and the relative dashboard and also if the Customer uses the MorphCast Emotion AI Interactive Video Platform service by activating the “Statistic Dashboard” function, it is specified that such statistic data cannot be considered private since they are not associated with natural persons, they are anonymous and aggregated with daily granularity, furthermore they are made available only if the aggregation exceeds a minimum threshold to ensure that anybody cannot connect them to a natural person. Therefore, these data are to be considered not subject to privacy laws, including the GDPR.
For these reasons Cynny, to the extent necessary and to the maximum extent permitted by applicable law in accordance with the “Exclusive Remedy and Limitation of Liability” section of the terms of use of the Software, declines all responsibility for the content of the images and/or videos generated by the Customer, including through the use of the Software, as well as collected, stored and processed exclusively by the Data Controller and therefore cannot in any case be responsible for the correctness and/or legitimacy of such contents pursuant to the applicable legislation (including, by way of example, the regulations on privacy, copyright, morality, public order, etc.).
Risk assessment
Cynny, while providing the Software as a tool that allows the processing of personal data relating to End Users, is not in a position to know (i) how such Software will be integrated and used by its Customers and End Users; nor, even less, (ii) which categories of personal data will be processed. In this regard, Cynny therefore qualifies as a “third party” pursuant to the GDPR, which has not received from the Customer any legitimacy to process any personal data of the End Users, as this treatment is reserved exclusively to the Customer. In any case, as indicated above, Cynny, in compliance with the general principles for the protection of privacy, has been concerned with the security aspects of the Software, making available to the Customer a Software that integrates the security measures of the host environment, against the violation or compromise of personal data (with the express exclusion of biometric data, which, as indicated above, are in no way processed by the Software) represented by the images read by the camera stream.
In fact, the video stream, for example from the camera of the host device of the End User, is made accessible to the Software, at the request of the End User, directly on the browser of that device; such access is allowed only with the authorization and control of the End User. The Software then reads the images of the stream on the End User’s browser, only if granted by the End User himself, only when the application is active and only for the purposes declared according to the specific privacy policy that will be provided by the Customer to the Final user.
Each image of the video stream remains in the volatile memory of the device (e.g. smartphone, tablet, PC etc.) of the End User, accessible from the Software within the browser sandbox, only for the time strictly necessary for processing the result, estimated in about 100ms, after which this image is destroyed and overwritten by the next image. Even the last frame of the stream is destroyed as soon as it is processed; therefore, no image is stored either in the browser of the End User device or, even less, within tools under the control of Cynny.
Cynny therefore has no control over these images, which therefore can never be transmitted, archived or shared with others by the latter.
Constant commitment to conscious and ethical use of Emotion AI
Cynny is constantly committed to providing all the information needed for better understanding and use of Emotion AI technologies. Check our guidelines and policies for a responsible use of Emotion AI. Cynny has also adopted a code of ethics which you can find here.
Contacts
For any clarifications and/or problems related to the operation of the Software with respect to privacy, please contact [email protected].
Cynny has also appointed a Data Protection Officer, who may be reached at the email address [email protected].
Annex A: transparency on the degree of probability and accuracy
Emotion AI facial recognition is a form of AI that is designed to analyze facial expressions and identify emotions such as happiness, sadness, anger, and surprise. The technology used by MorphCast is a combination of computer vision and convolutional deep neural network algorithms to analyze images or videos of people’s faces, and identify patterns that correspond to specific emotions.
The accuracy of emotion AI facial recognition technology can be affected by a number of factors, including:
– Image or video quality: The technology is more likely to produce accurate results when the images or videos being analyzed are of high quality, with clear and well-lit faces. Factors such as very low resolution, poor lighting, or camera angle can make it more difficult for the algorithm to accurately identify emotions.
Algorithm: Different algorithms have different levels of accuracy. MorphCast algorithms use deep learning techniques like convolutional neural networks (CNNs) and are trained on large datasets, these are considered to be more accurate than the ones that use traditional machine learning techniques.
– Training data: The accuracy of emotion AI facial recognition technology can also be affected by the quality and diversity of the training data that the algorithm has been exposed to. MorphCast algorithms are trained on a diverse dataset of images and videos from people of different ages, genders, and ethnicities. They are likely to be more accurate than those that are trained on a more limited dataset.
It is also important to note that emotion recognition technology is still in its early stages, and in MorphCast there is continuously ongoing research to improve its accuracy. Furthermore, there is a degree of bias in these technologies, particularly with regards to race and gender. As a result, it is important to be aware of these biases and take steps to mitigate them. This is the commitment that MorphCast has given itself since its inception.
Precisely with regard to the Bias we want to cite a recent research on it:
Abstract: Emotional expressions are a language of social interaction. Guided by recent advances in the study of expression and intersectionality, the present investigation examined how gender, ethnicity, and social class influence the signaling and recognition of 34 states in dynamic full-body expressive behavior. One hundred fifty-five Asian, Latinx, and European Americans expressed 34 emotional states with their full bodies. We then gathered 22,174 individual ratings of these expressions. In keeping with recent studies, people can recognize up to 29 full-body multimodal expressions of emotion. Neither gender nor ethnicity influenced the signaling or recognition of emotion, contrary to hypothesis. Social class, however, did have an influence: in keeping with past studies, lower class individuals proved to be more reliable signalers of emotion, and more reliable judges of full body expressions of emotion. Discussion focused on intersectionality and emotion. (PsycInfo Database Record (c) 2022 APA, all rights reserved).
Additionally, many studies have shown that emotion recognition technology is more accurate for certain emotions (e.g. happiness, anger) than others (e.g. sadness, surprise), and there is a degree of variability across different algorithms, which makes it challenging to give a general number for the accuracy of these technologies. Despite this, the current emotion prediction accuracy can be considered between 70% and 80% for almost all algorithms on the market.
In this regard, MorphCast has been evaluated in its accuracy by a university study that compares the accuracy of different software, detected on the six basic emotions of Paul Ekman, with the perception of the human being. These are the results of the accuracy study:
Damien Dupré, Eva G. Krumhuber, Dennis Küster, & Gary McKeown by respectively Dublin City University, University College London, University of Bremen, Queen’s University Belfast published here: Emotion recognition in humans and machine using posed and spontaneous facial expression. The “Emotional Tracking (SDK v1.0) was developed by the company MorphCast founded in 2013. Emotional Tracking SDK has the distinction of being a javascript engine less than 1MB and works directly on mobile browsers (i.e, without remote server and API processing).”
–
–
–
Annex B: type of data extracted
MorphCast SDK has a modular architecture which allows you to load only what you need. Here a quick description of the modules available:
- FACE DETECTOR
- POSE
- AGE
- GENDER
- EMOTIONS
- AROUSAL VALENCE
- ATTENTION
- WISH
- OTHER FEATURES
FACE DETECTOR
It detects the presence of a face in the field of view of the webcam, or in the input image.
The tracked face is passed to the other modules and then analyzed. In case of multiple faces, the total number of faces is available but, then, only the main face can be analyzed.
POSE
It estimates the head-pose rotation angles expressed in radians as pitch, roll and yaw.
AGE
It estimates the likely age of the main face with a granularity of years, or within an age group for better numerical stability.
GENDER
It estimates the most likely gender of the main face, Male or Female.
EMOTIONS
It estimates the presence and the respective intensities of facial expressions in the format of seven core emotions – anger, disgust, fear, happiness, sadness, and surprise, plus the neutral expression – according to the Ekman discrete model.
AROUSAL VALENCE
It estimates the emotional arousal and valence intensity. According to the dimensional model of Russell. Arousal is the degree of engagement (positive arousal), or disengagement (negative arousal); valence is the degree of pleasantness (positive valence), or unpleasantness (negative valence).
It outputs an object containing the smoothed probabilities in the range 0.00, – 1.00 of the 98 emotional affects, computed from the points in the 2D (valence, arousal) emotional space, according to the mappings of Schrerer and Ahn:
Adventurous |
Afraid |
Alarmed |
Ambitious |
Amorous |
Amused |
Angry |
Annoyed |
Anxious |
Apathetic |
Aroused |
Ashamed |
Astonished |
At Ease |
Attentive |
Bellicose |
Bitter |
Bored |
Calm |
Compassionate |
Conceited |
Confident |
Conscientious |
Contemplative |
Contemptuous |
Content |
Convinced |
Courageous |
Defiant |
Dejected |
Delighted |
Depressed |
Desperate |
Despondent |
Determined |
Disappointed |
Discontented |
Disgusted |
Dissatisfied |
Distressed |
Distrustful |
Doubtful |
Droopy |
Embarrassed |
Enraged |
Enthusiastic |
Envious |
Excited |
Expectant |
Feel Guilt |
Feel Well |
Feeling Superior |
Friendly |
Frustrated |
Glad |
Gloomy |
Happy |
Hateful |
Hesitant |
Hopeful |
Hostile |
Impatient |
Impressed |
Indignant |
Insulted |
Interested |
Jealous |
Joyous |
Languid |
Light Hearted |
Loathing |
Longing |
Lusting |
Melancholic |
Miserable |
Passionate |
Peaceful |
Pensive |
Pleased |
Polite |
Relaxed |
Reverent |
Sad |
Satisfied |
Selfconfident |
Serene |
Serious |
Sleepy |
Solemn |
Startled |
Suspicious |
Taken Aback |
Tense |
Tired |
Triumphant |
Uncomfortable |
Wavering |
Worried |
Here a representation in the 2D (valence, arousal) emotional space:
The simplified 38 emotional affects output can be useful in several cases:
An output with the quadrant of the 2D emotional space is also provided. This gives information about the appraisal dimensions: goal conduciveness/obstructiveness and coping potential (control/power).
ATTENTION
It estimates the attention level of the user to the screen, considering whether the user’s face is in or out of the field of view of the webcam, head position and other emotional and mood behavior. The speed of attention fall and rise can be properly tuned in an independent manner according to the use case.
WISH
It estimates the value of the MorphCast® Face Wish index. This is a proprietary metric that, considering the interest and sentiment of a customer, summarizes in a holistic manner his/her experience about a particular content or product presented.
OTHER FEATURES
It estimates the presence of the following face features:
Arched Eyebrows |
Double Chin |
Narrow Eyes |
Attractive |
Earrings |
Necklace |
Bags Under Eyes |
Eyebrows Bushy |
Necktie |
Bald |
Eyeglasses |
No Beard |
Bangs |
Goatee |
Oval Face |
Beard 5 O’Clock Shadow |
Gray Hair |
Pale Skin |
Big Lips |
Hat |
Pointy Nose |
Big Nose |
Heavy Makeup |
Receding Hairline |
Black Hair |
High Cheekbones |
Rosy Cheeks |
Blond Hair |
Lipstick |
Sideburns |
Brown Hair |
Mouth Slightly Open |
Straight Hair |
Chubby |
Mustache |
Wavy Hair |
Bibliography
- Russell, J., Lewicka, M. & Niit, T. (1989). A cross-cultural study of a circumplex model of affect. Journal of personality and social psychology, 57, 848–856.
- Scherer, Klaus. (2005). Scherer KR. What are emotions? And how can they be measured? Soc Sci Inf 44: 695-729. Social Science Information. 44. 695-792.
- Ahn, Junghyun & Gobron, Stéphane & Silvestre, Quentin & Thalmann, Daniel. (2010). Asymmetrical Facial Expressions based on an Advanced Interpretation of Two-dimensional Russells Emotional Model.
- Paltoglou, G., & Thelwall, M. (2012). Seeing stars of valence and arousal in blog posts. IEEE Transactions on Affective Computing, 4(1), 116-123.