Beyond the Algorithm: Building Trustworthy Emotion AI at MorphCast

AI Custom Action Eye

Last Update May 28, 2024

Dear professional customer,

We want to provide you with transparency and clarity regarding our emotional artificial intelligence system. This document details how it works, the methods for extracting data, information about the sources of data, and methods used for evaluating the system. It also includes known limitations or biases, as well as an assessment of potential risks and benefits of using the emotional AI system. To align with the AI Act, we are implementing ongoing compliance activities. These include regular risk assessments, documentation outlining compliance measures, and a process for addressing potential biases within our system. We provide guidelines for safe and effective use, aiming to ensure responsible deployment of our AI system.

The technology developed and distributed by Cynny S.p.A. (hereinafter “Cynny“), called “MorphCast® EMOTION AI HTML5 SDK” and/or the other available products, services and features that integrate it (e.g. MorphCast Web Apps or MorphCast Emotion AI Interactive Media Platform) is an artificial intelligence software capable of processing images from a video stream that allows you to establish, with varying degrees of probability and accuracy as you can see below in this document, the emotional state of the framed subject and some other apparent characteristics such as, for example, the cut or color of the hair, the apparent age and gender, facial emotion expressions, in any case without any identification of the subject (hereinafter, the “Software“). For a complete list of features extracted from the Software, please refer below in this document.

This technology, which Cynny makes available to its users (qualified as “Business Users”, pursuant to the applicable “Terms of Use”, and hereinafter referred to as “Customers“) has been specifically designed and developed so as not to identify the framed subjects. These subjects are the end users of the Software chosen by the Customer (hereinafter, the “End Users“), with whom Cynny has no relationship whatsoever.

The Software can be used autonomously by the Customer or be activated, by the Customer’s choice, in the context of use of other services provided by Cynny, whose terms of use and privacy policies are available at this link: mission page. As regards the registration data of the Customers, the relative Privacy Policy will also apply, available at this link: Emotion AI – Privacy Policy.

How it works – Operation

The Software works as a Javascript library embedded in a web page usually belonging to the Customers’ domain, running in the browser of the End User device, which, as indicated above, has a relationship exclusively with the Customer (and not with Cynny).

The Software does not process biometric data that allow the univocal identification of a subject, but only an instant (volatile) processing of personal data during the processing of images (frames), read from the video stream (for example of a camera, which is used only as a sensor and not as an image and/or video recorder) provided by the End User’s browser. Cynny does not carry out any processing or control over such personal data as they are processed through the Software automatically and directly on the End User’s device browser or app.

Accordingly, the use of the Software is subject to the following terms of use, both of which are beyond Cynny’s control: (1) the terms of use of the End User’s browser in which the Software is incorporated and (2) any software or third party API required to allow the web page to function.

Notwithstanding the foregoing, as the supplier of the Software, Cynny (a) warrants that it has adopted all appropriate security measures in the design of the Software also with regard to the processing of personal data, taking into account privacy-by-design approach and ensuring adequate security of the software itself and; (b) makes itself available to collaborate with the Customer (who, in case it uses the extracted data autonomously, assumes the title of Data Controller pursuant to EU Regulation 2016/679, hereinafter “GDPR“) for any processing of personal data relating to the End User, showing and explaining all the functions of the Software, in order to enable it to carry out a precise risk analysis with respect to the fundamental rights and freedoms of the persons concerned (including any impact assessments, where necessary and / or appropriate, in the opinion of the Data Controller). In fact, it will be the Data Controller who will have to establish the purposes, methods and legal basis of the processing, the time and place (physical or virtual) for the storage of personal data relating to End Users, also providing them with specific online information. with the applicable legislation for the protection of personal data (including the GDPR).

According to the provisions of the Terms of Use of the Software (available here for all available applications: Terms of Use), the same is aimed solely and exclusively at professional users, with the express exclusion of consumers and, therefore, of the use of the Software for domestic / personal purposes. In any case, whatever the type of processing carried out by the Customer, as Data Controller, the consequences deriving from the application of the applicable legislation, including that for the protection of personal data, are exclusively the responsibility of the Data Controller.

It is also specified that the Software operates integrated in a private corporate hardware / software context of the Customer’s End User (on end-point resources, including web pages belonging to the Customer’s domain and therefore under its control) equipped with security necessary to avoid the compromise of the data affected by the processing by the Software. In line with the foregoing, the consequent responsibilities relating to the processing of personal data of End Users by the Customers fall exclusively on the Customer, which therefore undertakes, in accordance with the provisions of the “Indemnification” section of the terms of use of the Software, to fully indemnify Cynny from any prejudicial consequence that may arise in this regard.

As per the emotional data made available to the Customers by the Software through the “Data Storage” module and the relative dashboard and also if the Customer uses the MorphCast Emotion AI Interactive Media Platform service by activating the “Statistic Dashboard” function, it is specified that such statistic data cannot be considered private since they are not associated with natural persons. Therefore, these data are to be considered not subject to privacy laws, including the GDPR.

Important: To ensure you understand how your data and the data of your users is used, please review the privacy policy directly associated with each product. You can find summaries of all policies on this page.


Risk Assessment

MorphCast regularly conducts risk assessments, including audits by qualified third parties. For more information about audits, please refer to this documentation.

While Cynny provides Software that enables the processing of personal data relating to End Users, we cannot directly control how Customers and End Users integrate or utilize it, nor the specific categories of personal data processed. As such, Cynny operates as a “third party” under GDPR guidelines, without Customer-granted legitimacy to process End User data independently. Nonetheless, we are deeply committed to upholding privacy principles and have prioritized security within our Software. Cynny provides Customers with a solution that integrates within host environment security measures, with the express exclusion of biometric data processing.

The video stream accessed by our Software, such as an End User’s camera feed, is made available upon specific authorization and remains under End User control. Access occurs directly within the End User’s browser. When granted, Cynny processes images from the stream solely for declared purposes outlined by the Customer’s privacy policy. This processing occurs within the End User’s device and browser sandbox, with each image held temporarily (approximately 100ms) before deletion. We do not store images on the End User’s device or within Cynny-controlled tools.

While Cynny has no direct control over these images, we remain committed to facilitating vigilant and responsible use of our Software by our Customers. This includes providing technical measures for data protection and supporting Customers in ensuring their practices align with ethical guidelines and regulations.


Respecting Human Rights in the Age of AI

At MorphCast, we believe in the transformative power of AI to connect people and enhance communication. However, we also recognize the ethical responsibility that comes with harnessing such powerful technology. We are committed to upholding human rights principles in all aspects of our development and operation, ensuring that AI enhances human flourishing while respecting individual dignity and fundamental freedoms.

Our Guiding Principles:

  • Transparency and Explainability: We strive for transparency in our data collection, model development, and decision-making processes. We provide users with clear explanations of how MorphCast works and how their data is used.
  • Non-discrimination and Fairness: We design and train our systems to be fair and unbiased, avoiding any potential for discrimination based on protected characteristics. We actively monitor for and address any potential biases that may emerge.
  • Privacy and Data Protection: We respect user privacy by adhering to relevant data protection regulations and collecting only the minimum data necessary for service operation. We empower users with control over their data and prioritize its security and confidentiality.
  • Freedom of Expression and Opinion: We recognize the importance of freedom of expression and opinion. We do not censor user content and only restrict communication that violates our terms of service or promotes illegal activities.
  • Accountability and Oversight: We hold ourselves accountable for our actions and are open to feedback and scrutiny. We engage with stakeholders from diverse backgrounds to ensure our practices are ethical and responsible.

Putting Principles into Practice:

  • Bias Detection and Mitigation: We employ diverse datasets and utilize fairness metrics to identify and mitigate potential biases in our models. We regularly review and update our systems to ensure consistent fairness.
  • User Control and Consent: We provide users with clear options to control how their data is collected, used, and shared. We obtain explicit consent before processing any personal information.
  • Transparent Privacy Policy: Our privacy policy is easily accessible and clearly explains our data practices. We avoid complex legal jargon and ensure the information is understandable to all users.
  • Human Review and Oversight: We implement human review processes for sensitive content and algorithmic decisions, ensuring a balance between AI automation and human judgment.
  • Collaboration and Stakeholder Engagement: We actively engage with civil society organizations, legal experts, and ethicists to learn from diverse perspectives and ensure our practices align with evolving human rights standards.

By upholding these principles, MorphCast aims to be a leader in responsible AI development, fostering an environment where technology serves humanity and empowers individuals.

It is important to note that this is not a comprehensive list and the specific implementation of these principles will evolve as technology and regulations develop. We remain committed to ongoing learning and improvement, ensuring that MorphCast remains a force for good in the world.


User Profiling

At MorphCast, we recognize the potential of user profiling to personalize experiences and tailor our platform to individual needs. However, we also understand the ethical concerns surrounding user profiling and are committed to utilizing this technology responsibly and transparently.

Our Approach to User Profiling:

  • Limited and Necessary Data: We collect only the data essential for understanding user preferences and customizing their experience. We avoid unnecessary data collection and prioritize user privacy.
  • Transparency and Choice: We inform users about what data we collect and how it is used for profiling purposes. We empower users with clear options to manage their profile data and opt out of profiling altogether.
  • Aggregation and Anonymization: We prioritize data minimization and anonymization wherever possible. We aggregate user data for analysis and avoid personal identification whenever feasible.
  • Fairness and Non-discrimination: We design our profiling algorithms to be fair and unbiased, ensuring they do not disadvantage or discriminate against any particular group of users.
  • Human Oversight and Control: We implement human oversight mechanisms to review and validate profiling results, ensuring responsible decision-making and preventing algorithmic bias.

Benefits of User Profiling:

  • Personalized Recommendations: We leverage profiling to suggest relevant content, features, and connections to individual users, enhancing their experience and engagement.
  • Improved Content Discovery: By understanding user preferences, we can tailor content displays and search results to individual interests, promoting relevant and valuable information.
  • Enhanced Accessibility: Profiling can inform accessibility features and personalize interfaces to better meet the needs of diverse users.

Addressing Concerns:

  • Data Security and Privacy: We prioritize data security with robust measures like encryption and access controls. We comply with relevant data protection regulations and provide users with control over their data.
  • Algorithmic Bias: We actively test and monitor our profiling algorithms for potential biases and take corrective actions to ensure fairness and inclusivity.
  • Transparency and Control: We offer users clear explanations of how their data is used for profiling and provide them with options to manage their profiles and opt out.

By adhering to these principles, MorphCast strives to utilize user profiling responsibly and ethically. We believe that personalized experiences can enhance user engagement without compromising individual privacy or creating discriminatory outcomes.

It is important to note that this is an evolving area and we are committed to continuously learning and adapting our approach as technology and regulations change. We welcome feedback and discussion on user profiling to ensure we remain responsible and transparent in our practices.


Training Data

Our neural networks are developed with custom datasets that encompass a diverse range of faces, covering various ethnicities, ages, and genders. Due to confidentiality, details about the dataset remain proprietary. However, it’s important to acknowledge that like any AI system, ours is not immune to biases related to race, gender, and age. Nevertheless, we have evidence to support that such biases do not significantly impact our analysis results. Specifically, MorphCast SDK underwent rigorous evaluation on two benchmark tests by an independent research center, as documented in their publication ( The test was made on testbed BU-4DFE among different ages, genders and the following ethnicities: Asian, Black, Hispanic / Latino, African American and White population. This evaluation confirmed the SDK’s accuracy across different ages, genders, and ethnic groups.

We encourage users to be mindful of potential biases on a case-by-case basis, aiming to mitigate or eliminate their effects whenever possible. This approach ensures the most effective use of our technology, tailored to the specific objectives of each analysis.


Measures to Ensure Security and Reliability

At MorphCast, we understand the importance of trust. We harness the power of AI to help people connect and engage, but we do so with a firm commitment to security and reliability. Aligned with the AI Act’s Measures to Ensure Security and Reliability, here’s how we actively address key risks associated with our system:


  • Accuracy and Bias: We strive for unbiased results by using diverse data sources, implementing data quality checks, and employing bias detection methods. Regular data quality assessments, fairness audits, and differential privacy techniques are integral to our process.
  • Security: We prioritize data protection with robust measures like encryption, access controls, and regular penetration testing. Strong password policies, need-to-know access principles, and user security awareness training further elevate our data security posture.
  • Compliance: We operate with transparency and respect for user privacy. We obtain clear and informed consent, define data collection purposes explicitly, and adhere to data minimization principles. Our transparent privacy policy empowers users with control over their data, and we ensure compliance with relevant data protection regulations.


  • Model Fairness and Explainability: We employ diverse training data, leverage counterfactual explanations, and utilize interpretability techniques to ensure model fairness and explainability. Regular fairness metric measurements, key decision explanations, and the ability for users to challenge results uphold these commitments.
  • Vulnerability Management: We prioritize system security through regular vulnerability assessments, prompt security patch deployments, and the application of secure development practices. Secure coding, penetration testing, and intrusion detection/prevention systems bolster our defenses.
  • Reliability and Error Handling: Comprehensive testing, performance monitoring, and robust error handling mechanisms ensure seamless system operation. Thorough system testing, in-production performance monitoring, and well-defined error handling procedures contribute to a reliable user experience.

Societal Impact:

  • Fairness and Inclusivity: We actively promote fairness and inclusivity by utilizing diverse data, implementing model fairness measures, and establishing responsible use guidelines. Regular fairness audits, stakeholder consultations, and transparency regarding sensitive data handling are foundational elements of our approach.
  • Preventing Misuse and Manipulation: We safeguard against misuse and manipulation through various mechanisms. Empowering users with data control, offering educational resources on responsible AI use, and actively detecting and preventing misuse empower users and promote responsible AI usage.
  • Responsible AI Implementation: We analyze potential job displacement risks, offer reskilling and upskilling opportunities, and engage with stakeholders to develop ethical AI implementation guidelines. Impact assessments, workforce development programs, and ethical AI implementation principles guide our responsible AI deployment efforts.

Performance Monitoring and Updates:

We continuously monitor MorphCast’s accuracy, fairness, security, and societal impact through robust performance metrics and dashboards. Ongoing analysis helps us identify emerging risks and performance issues, leading to timely updates. Retraining with diverse data, implementing new security measures, or adjusting algorithmic parameters are part of our update process. We document our performance monitoring procedures, update decision-making processes, and maintain a changelog for all MorphCast updates, ensuring transparency and accountability.

This is just an overview of our commitment to security and reliability. By continuously evaluating and improving our efforts, we strive to earn and maintain your trust as we utilize AI responsibly to enhance human connection and engagement.


Principles of Cybersecurity and Data Protection

At MorphCast, we operate with the understanding that the power of AI comes with a significant responsibility – protecting user data and ensuring the security of our platform. We are committed to upholding the highest standards of cybersecurity and data protection, guided by the following key principles:

1. Security by Design: Security is embedded into every aspect of MorphCast, from system architecture to development practices. We employ industry-leading security measures like encryption, access controls, and regular security assessments to ensure data confidentiality, integrity, and availability.

2. Privacy-First Approach: We respect user privacy by collecting only the data essential for service operation and obtaining explicit consent before processing it. We adhere to data minimization principles, limit data retention periods, and empower users with control over their data through transparent privacy policies and clear opt-out mechanisms.

3. Transparency and Accountability: We believe in being transparent about how we collect, use, and protect user data. Our privacy policies are easily accessible, and we provide clear explanations regarding data processing purposes and potential risks. We are accountable for our actions and regularly review our practices to ensure compliance with relevant regulations and ethical standards.

4. User Education and Empowerment: We believe an informed user base is key to robust security and privacy. We provide users with accessible resources and educational materials to understand how MorphCast handles data and how to use it responsibly. We encourage users to report any suspicious activity and actively participate in our security improvement efforts.

5. Continuous Improvement: We understand that the cybersecurity and data protection landscape is constantly evolving. We are committed to staying ahead of the curve by conducting regular risk assessments, incorporating new security best practices, and actively engaging with relevant experts and stakeholders.

These principles are the foundation of our commitment to building a trustworthy and secure platform for AI-powered engagement. We strive to earn and maintain user trust by continuously demonstrating our dedication to responsible data practices and robust cybersecurity measures.

Additionally, we adhere to relevant regulations and frameworks such as:

  • The General Data Protection Regulation (GDPR)
  • The California Consumer Privacy Act (CCPA)
  • The Cybersecurity and Infrastructure Security Agency (CISA) guidelines

By upholding these principles and regulations, MorphCast aims to be a leader in responsible AI development and foster a trusted environment for human connection and engagement.


Damage or Discrimination Complaint System

At MorphCast, we strive to utilize AI responsibly and create a platform that fosters positive and inclusive interactions. However, we recognize that any use of technology can have unintended consequences. To ensure fairness and address potential concerns, we have established a clear and accessible Damage or Discrimination Complaint System.

Our Commitment:

  • We are committed to investigating all claims of damage or discrimination arising from the use of MorphCast promptly and comprehensively.
  • We believe in transparency and will communicate openly with complainants throughout the investigation process.
  • We will take appropriate action to address any identified issues, implementing remedial measures to prevent similar situations in the future.

What You Can Report:

You can submit a complaint if you believe you have suffered:

  • Discrimination: This includes any unfair treatment based on factors such as race, ethnicity, religion, gender, sexual orientation, disability, or age.
  • Emotional harm: This could involve distress, anxiety, or other negative emotions caused by the use of MorphCast.
  • Financial damage: This covers any direct financial losses resulting from the misuse of MorphCast.

How to Submit a Complaint:

What We Need from You:

To expedite the investigation process, please provide the following information in your complaint:

  • A clear description of the incident.
  • The date and time the incident occurred.
  • Any relevant screenshots or recordings.
  • Your contact information.

What Happens Next:

  • Upon receiving your complaint, we will acknowledge it within 24 hours.
  • A dedicated team will review your complaint and may request additional information if needed.
  • We will keep you informed of the progress of the investigation and communicate our findings and any remedial actions within a reasonable timeframe.

Additional Resources:

  • You can access our Terms of Service and Privacy Policy on our website for further information regarding our commitments.
  • We encourage you to contact us if you have any questions or concerns about our Damage or Discrimination Complaint System.


Transparency on the degree of probability and accuracy

Emotion AI facial recognition is a form of AI that is designed to analyze facial expressions and identify emotions such as happiness, sadness, anger, and surprise. The technology used by MorphCast is a combination of computer vision and convolutional deep neural network  algorithms to analyze images or videos of people’s faces, and identify patterns that correspond to specific emotions.

The accuracy of emotion AI facial recognition technology can be affected by a number of factors, including:

  • Image or video quality: The technology is more likely to produce accurate results when the images or videos being analyzed are of high quality, with clear and well-lit faces. Factors such as very low resolution, poor lighting, or camera angle can make it more difficult for the algorithm to accurately identify emotions.
  • Algorithm: Different algorithms have different levels of accuracy. MorphCast algorithms use deep learning techniques like convolutional neural networks (CNNs) and are trained on large datasets, these are considered to be more accurate than the ones that use traditional machine learning techniques.
  • Training data: The accuracy of emotion AI facial recognition technology can also be affected by the quality and diversity of the training data that the algorithm has been exposed to. MorphCast algorithms are trained on a diverse dataset of images and videos from people of different ages, genders, and ethnicities. They are likely to be more accurate than those that are trained on a more limited dataset.
  • lighting conditions: In addition to our commitment to diversity in training data, it’s crucial to address the role of environmental factors—particularly lighting conditions—on the accuracy of our AI technology. The interaction between light and skin, known as light refraction, varies significantly across different skin tones. This variation is not a matter of better or worse performance but depends on the context: in low-light conditions, the accuracy of facial recognition on lighter skin tones might be higher, whereas in brightly lit environments, darker skin tones may be detected with greater precision. These differences, attributable to the physics of light interaction, can have a more substantial impact on recognition accuracy than the minimal biases we’ve diligently minimized through training our AI with a vast and diverse dataset.

It is also important to note that emotion recognition technology is still in its early stages, and in MorphCast there is continuously ongoing research to improve its accuracy. Furthermore, there is a degree of bias in these technologies, particularly with regards to race and gender. As a result, it is important to be aware of these biases and take steps to mitigate them. This is the commitment that MorphCast has given itself since its inception. 

Precisely with regard to the Bias we want to cite a recent research on it:

Intersectionality in emotion signaling and recognition: The influence of gender, ethnicity, and social class.

Abstract: Emotional expressions are a language of social interaction. Guided by recent advances in the study of expression and intersectionality, the present investigation examined how gender, ethnicity, and social class influence the signaling and recognition of 34 states in dynamic full-body expressive behavior. One hundred fifty-five Asian, Latinx, and European Americans expressed 34 emotional states with their full bodies. We then gathered 22,174 individual ratings of these expressions. In keeping with recent studies, people can recognize up to 29 full-body multimodal expressions of emotion. Neither gender nor ethnicity influenced the signaling or recognition of emotion, contrary to hypothesis. Social class, however, did have an influence: in keeping with past studies, lower class individuals proved to be more reliable signalers of emotion, and more reliable judges of full body expressions of emotion. Discussion focused on intersectionality and emotion. (PsycInfo Database Record (c) 2022 APA, all rights reserved).

Emotional expressions

Additionally, many studies have shown that emotion recognition technology is more accurate for certain emotions (e.g. happiness, anger) than others (e.g. sadness, surprise), and there is a degree of variability across different algorithms, which makes it challenging to give a general number for the accuracy of these technologies. Despite this, the current emotion prediction accuracy can be considered between 70% and 80% for almost all algorithms on the market.

In this regard, MorphCast has been evaluated in its accuracy by a university study that compares the accuracy of different software, detected on the six basic emotions of Paul Ekman, with the perception of the human being. These are the results of the accuracy study:

Damien Dupré, Eva G. Krumhuber, Dennis Küster, & Gary McKeown by respectively Dublin City University, University College London, University of Bremen, Queen’s University Belfast published here: Emotion recognition in humans and machine using posed and spontaneous facial expression. The “Emotional Tracking (SDK v1.0) was developed by the company MorphCast founded in 2013. Emotional Tracking SDK has the distinction of being a javascript engine less than 1MB and works directly on mobile browsers (i.e, without remote server and API processing).”

Comparison of emotion recognition in humans and machine using posed and spontaneous facial expression (general)

Comparison of emotion recognition in humans and machine using posed and spontaneous facial expression (for each emotion)



Capabilities – Type of data extracted

MorphCast SDK has a modular architecture which allows you to load only what you need. Here a quick description of the modules available:

  • POSE
  • AGE
  • WISH


It detects the presence of a face in the field of view of the webcam, or in the input image.

The tracked face is passed to the other modules and then analyzed. In case of multiple faces, the total number of faces is available but, then, only the main face can be analyzed.


It estimates the head-pose rotation angles expressed in radians as pitch, roll and yaw.

poses and rotation of head recognized by emotion recognition ai
The ZERO point is when a face looks straight at the camera


It estimates the likely age of the main face with a granularity of years, or within an age group for better numerical stability. 


It estimates the most likely gender of the main face, Male or Female. 


It estimates the presence and the respective intensities of facial expressions in the format of seven core emotions – anger, disgust, fear, happiness, sadness, and surprise, plus the neutral expression – according to the Ekman discrete model


It estimates the emotional arousal and valence intensity. According to the dimensional model of Russell. Arousal is the degree of engagement (positive arousal), or disengagement (negative arousal); valence is the degree of pleasantness (positive valence), or unpleasantness (negative valence).

It outputs an object containing the smoothed probabilities in the range 0.00, – 1.00 of the 98 emotional affects, computed from the points in the 2D (valence, arousal) emotional space, according to the mappings of Schrerer and Ahn:


98 circumplex model of affects words

98 words Rassell Circumplex model of affects

The 98 emotional affects output 2D Quadrant

Russell 2D Arousal-Valence quadrant

The simplified 38 emotional affects output can be useful in several cases

Circumflex model of affects (Representation in the 2D emotional space with 38 words)



It estimates the attention level of the user to the screen, considering whether the user’s face is in or out of the field of view of the webcam, head position and other emotional and mood behavior. The speed of attention fall and rise can be properly tuned in an independent manner according to the use case.


It estimates the value of the MorphCast® Face Wish index. This is a proprietary metric that, considering the interest and sentiment of a customer, summarizes in a holistic manner his/her experience about a particular content or product presented on the screen. This output is very sensible to the variation of expressions, giving him a value in estimate a change of sentiment looking a product or a scene in a video.


It gauges the intensity of arousal and valence based on the 17-degree angle of the circumplex model of affect (Russell). This exclusive metric provides a comprehensive overview of an individual’s positivity, capturing facial expressions.

positivity meter in emotion recognition ai
α = 17 (customizable)

This output can be used for multiple types of detection by modifying the alpha angle on the whole model.


Several alarm are outputing by the AI engine to help developer to trigger reactions at possible cheating situations (NO FACE, MORE FACES, LOW ATTENTION…)


It estimates the presence of the following face features:

Arched Eyebrows

Double Chin

Narrow Eyes




Bags Under Eyes

Eyebrows Bushy




No Beard



Oval Face

Beard 5 O’Clock Shadow

Gray Hair

Pale Skin

Big Lips


Pointy Nose

Big Nose

Heavy Makeup

Receding Hairline

Black Hair

High Cheekbones

Rosy Cheeks

Blond Hair



Brown Hair

Mouth Slightly Open

Straight Hair



Wavy Hair




Constant commitment to conscious and ethical use of Emotion AI

Cynny is constantly committed to providing all the information needed for better understanding and use of Emotion AI technologies. Check our guidelines and policies for a responsible use of Emotion AI. Cynny has also adopted a code of ethics which you can find here.

Frequently Asked Questions

We also invite you to consult these frequently asked questions, which we continuously update, with details on the most common requests for clarification from our clients on the topics covered in this document.


If you experience any damage or abuse using our platform, please inform us using this form on our website.

For any clarifications and/or problems related to the operation of the Software with respect to privacy, please contact

Cynny has also appointed a Data Protection Officer, who may be reached at the email address