Open access peer-reviewed chapter

Perspective Chapter: Embracing the Complexity of Human Emotion

Written By

Saeed Albarhami Thabit

Reviewed: 31 July 2023 Published: 10 January 2024

DOI: 10.5772/intechopen.112701

From the Edited Volume

Advances in Sentiment Analysis - Techniques, Applications, and Challenges

Edited by Jinfeng Li

Chapter metrics overview

66 Chapter Downloads

View Full Metrics

Abstract

In this chapter, we delve into the multifaceted world of human emotions through the lens of advanced analysis techniques, aiming to unlock a deeper understanding of human behavior and decision-making processes in our digital landscape. We begin by illustrating the complexity of human emotions and the significance of accurate emotion detection across various applications, from marketing and customer relationship management to healthcare and social media monitoring. This context leads us to discuss state-of-the-art emotion detection methods, including transformer-based models, context-aware emotion detection, physiological signal recognition, and multimodal emotion analysis. Here, we adopt a systematic approach to emotion analysis, utilizing the transformer-based architecture fine-tuned on a tweets dataset. Our methodology achieves an accuracy of 82.53%, a precision of 82.79%, a recall of 82.53%, and an F1 score of 82.29% in predicting emotional categories. The chapter also scrutinizes challenges, limitations, and ethical considerations in this field, including ambiguity, subjectivity, and cross-cultural variations. Finally, we glance into the future of emotion analysis, focusing on integrating emotional intelligence into artificial intelligence systems and developing personalized techniques. We aim to spur further research and collaboration in this field, thus enriching our understanding of the dynamic role of human emotions in our interconnected world.

Keywords

  • emotion analysis
  • human emotions
  • emotional complexity
  • transformer-based models
  • multimodal emotion analysis
  • context-aware emotion detection
  • physiological signals
  • mental health monitoring
  • human behavior
  • decision-making processes
  • digital landscape
  • public opinion analysis
  • ethical considerations
  • cross-cultural variations
  • emotional intelligence
  • personalized emotion analysis
  • social media monitoring

1. Introduction

In our increasingly interconnected world, understanding the complexities of human emotions has become essential. In the digital landscape, the quest to understand and analyze complex human emotions has become more relevant through the lens of the most recent advanced emotion analysis techniques, setting the stage for a profound grasp of human behavior, communication, and decision-making processes. This chapter dives into the mysterious world of human emotions, employing advanced emotion analysis techniques as a powerful tool to unveil the underlying emotions that drive our actions and interactions.

We start by journeying into the labyrinth of human emotions, highlighting why understanding emotional complexity is crucial. The importance of accurate emotion analysis is discussed across various contexts, from marketing and customer relationship management, as emphasized in earlier research on the crucial role of polarity detection and emotion recognition in comprehending and forecasting customer service experiences [1, 2] to healthcare and social media monitoring, highlighted in past studies such as [2, 3, 4, 5, 6]. We then delve into cutting-edge techniques for complex emotion detection, including transformer-based models [7], multimodal emotion analysis [8, 9], context-aware emotion detection [10, 11]. We also discuss emotion recognition using physiological signals [4, 12, 13, 14] and address their potential to deepen our understanding of human emotions. The chapter also navigates emotion analysis’s challenges, limitations, and ethical aspects, including issues like ambiguity, subjectivity, and cross-cultural variations in emotional expression, emphasizing a call to action for further research and collaboration to fully comprehend the multi-dimensional nature of human emotions in our digital era.

We conclude with a glimpse into the future of emotion analysis, accentuating the integration of emotional intelligence in AI systems, ethical considerations, and the emergence of personalized emotion analysis techniques tailored to individual preferences and cultural backgrounds. By offering a comprehensive view of recent advancements and challenges in emotion analysis, this chapter aims to inspire further research and collaboration, fostering a more profound understanding of the multi-dimensional nature of human emotions in our digital era.

Advertisement

2. Background

The landscape of human emotions is a rich tapestry woven with a multitude of feelings, emotions, and sentiments, varying in depth, complexity, and expression. Emotions are multifaceted and rarely exist in isolation, with humans often experiencing a blend of emotions simultaneously. Human nature’s nuanced complexity offers intriguing possibilities and poses significant challenges for emotion analysis. Emotions are fundamentally subjective and deeply personal, making their accurate assessment challenging. To illustrate, consider the feeling of joy. One person’s expression of joy could be another’s expression of satisfaction, depending on their emotional and cultural background. This inherent subjectivity necessitates an intricate understanding of individual emotions in emotion analysis.

Emotions are also dynamic and ephemeral, changing rapidly in response to stimuli. The emotion behind words can change dramatically depending on the context or even the tone in which they are expressed. For instance, a person tweeting “I love this!” might convey genuine enthusiasm when discussing their favorite book but sarcasm when discussing a disliked policy. This demonstrates the necessity for emotion analysis models that consider the context and changes in emotion over time. Therefore, static emotion analysis models can often fall short, necessitating more dynamic and context-aware models that can adapt to the fluidity of human emotions.

Cultural and societal factors can profoundly influence emotional expression and interpretation. The same emotion can be expressed differently across cultures, and what might be perceived as a positive emotion in one culture could be seen as negative in another. Consider the example of a movie viewer watching a sad scene. One person might react with profound sadness, tears welling in their eyes, while another might feel a sense of nostalgic melancholy, yet another might be unmoved, deeming it overly sentimental. This variability in emotional reactions to the same stimulus underscores the personal nature of emotions, presenting a considerable challenge for emotion analysis. Consequently, understanding the cultural nuances of emotional expression and incorporating them into emotion analysis models is critical.

Furthermore, emotional complexity extends beyond verbal or textual expression. Non-verbal signals such as facial expressions and tone play an essential role in communicating emotions and are often more truthful than words. For example, when a person says “I am fine” with a neutral facial expression, their tone may reveal underlying sadness or frustration. This highlights the importance of multimodal emotion analysis, which integrates text, audio, and visual data to understand emotions comprehensively. Therefore, a comprehensive understanding of emotional complexity necessitates the incorporation of these non-verbal cues in emotion analysis.

The challenge of understanding emotional complexity underscores the significance of employing advanced emotion analysis techniques to navigate the complex emotional landscape, respect its intricacies, and accurately decode the underlying emotions. This challenging task of emotional understanding requires an amalgamation of linguistics, psychology, machine learning, and deep learning techniques. However, overcoming this challenge promises to revolutionize numerous fields, from marketing and customer relationship management to mental health monitoring and public opinion analysis.

Understanding emotional complexity is a vital prerequisite for the advancement of emotion analysis. By recognizing the multi-dimensional nature of human emotions and developing sophisticated techniques to capture these dimensions, we can better effectively and ethically harness the power of human emotions. This journey, as challenging as it is fascinating, provides immense opportunities for further research, collaboration, and innovation in emotion analysis.

As our understanding of emotional complexity deepens, numerous applications have started to benefit from this wealth of information. From personalized marketing to mental health support, emotion analysis allows us to tap into a human being’s most intimate element – their emotions.

2.1 Personalized marketing and advertising

Personalized marketing and advertising have truly revolutionized traditional business practices, giving businesses a unique opportunity to engage with their audience more intimately. Emotion analysis, often conducted via advanced artificial intelligence technologies and emotion analysis tools, plays a pivotal role in this revolution, serving as the backbone for understanding and engaging with the complex emotional landscape of consumers. The heart of emotion analysis lies in the interpretation of the emotional states and reactions of customers. This process examines various customer feedback forms such as product reviews, social media commentary, or customer service interactions. By investigating these mediums, businesses can understand how their products or services are being emotionally received and perceived by the customers [2].

Emotion analysis can help reveal patterns of customer opinion that go unnoticed. For instance, it can bring to light a widespread emotion of disappointment in a product feature or unveil a sense of joy associated with a particular service experience. This information can be further segmented by demographic groups, providing valuable insights into the emotional responses of different market segments.

Once these emotions are understood, businesses can adjust their marketing and advertising strategies accordingly. If the analysis reveals a negative emotion towards a product, the company may modify the marketing message to address and mitigate this negativity. On the other hand, if customers demonstrate a positive emotional response to a particular product feature or service, the company might amplify these emotions in their advertising campaigns, harnessing the power of positive affirmation to enhance customer loyalty and encourage repeat business.

Furthermore, emotion analysis can also be used to create more personalized and emotionally resonant marketing campaigns [10, 15, 16]. By understanding the specific emotions associated with a brand or product, companies can tailor their messaging to evoke similar emotions, creating a more profound and authentic connection with their audience. For example, a car company that finds its customers associate feelings of freedom and adventure with their products might develop advertising campaigns that evoke these emotions, resonating on a deeper, more personal level with their audience.

Emotion analysis in personalized marketing and advertising thus holds immense potential for fostering customer loyalty and engagement. As businesses refine their understanding of their customers’ emotional responses, they will be better equipped to respond to their needs, tailor their products, and shape their messaging to resonate more deeply with their target audience. As such, emotion analysis represents a powerful tool in the modern business arsenal that promises to continue shaping the landscape of personalized marketing and advertising in the years to come.

2.2 Social media monitoring

Social media is a goldmine for emotion analysis. Social media platforms have precipitated a paradigm shift in how businesses interact with consumers and understand and monitor public opinion [3, 17]. Such platforms serve as a rich repository of public opinion, and when mined intelligently, they can provide unprecedented insights into consumer attitudes, needs, and behaviors. In this context, sentiment and emotion analysis applied to social media monitoring can be invaluable. For instance, consider a company’s launch of a new product or service. By analyzing the social media discourse surrounding this launch—which could range from tweets to posts and photos to videos, a company can gain an in-depth understanding of public view towards the product or service. Moreover, sophisticated emotion analysis algorithms can uncover the polarity of the emotion (positive, negative, or neutral) and the nuances of the emotional responses elicited, such as excitement, disappointment, anticipation, confusion, or admiration.

Consider a tech company unveiling a new smartphone model as a hypothetical example. Emotion analysis of social media reactions could reveal that while the phone’s design elicits positive emotions and excitement, its price generates disappointment or frustration. This nuanced understanding can inform the company’s subsequent marketing strategies, pricing decisions, and design improvements for future models. The power of emotion analysis in social media monitoring extends beyond product launches. It can be employed to assess public reaction to advertising campaigns, gauge consumer satisfaction with customer service, monitor brand reputation, track emotion towards competitors, and identify emerging market trends or consumer needs.

By harnessing the power of sentiment and emotion analysis in social media monitoring, businesses can turn the tide of public opinion in their favor, make informed strategic decisions, and maintain a competitive edge in an increasingly digital marketplace.

2.3 Customer relationship management (CRM)

Within the domain of Customer Relationship Management (CRM), understanding emotional complexity allows businesses to respond to customer interactions promptly and empathetically. CRM has become a vital strategy in today’s business landscape, where the customer is at the center of all operations. Using sentiment and emotion analysis within CRM frameworks is steering in a new era of personalized and emotionally aligned customer service, driving customer satisfaction and loyalty [1, 2].

A key aspect of CRM is interaction management, encompassing all touchpoints between a business and its customers. Here, emotion analysis can offer critical insights into a customer’s mind. For instance, an email from a disgruntled customer might express disappointment or feeling undervalued. An advanced CRM system with emotion analysis capabilities can decipher these complex emotions, providing a nuanced understanding of the customer’s feeling. This understanding empowers customer service representatives to respond empathetically and effectively, thus enhancing the overall customer experience. For instance, suppose a customer sends an email complaint about a recently purchased product that did not meet their expectations. A typical response might address the complaint at face value, offering a refund or replacement. However, with sentiment and emotion analysis, the CRM system could reveal underlying disappointment due to high expectations from the brand or annoyance at the inconvenience caused. Thus, the customer service representative could tailor their response to acknowledge these emotions, apologize, and offer a goodwill gesture, such as a discount on the next purchase. This tailored response will likely transform a potentially negative customer experience into a positive one, fostering customer loyalty.

In a broader sense, emotion analysis in CRM can aid in proactive issue resolution, trend identification, and strategic decision-making. For example, consistently high levels of frustration or disappointment related to a specific product or service aspect could prompt a business to investigate and rectify the underlying issue, potentially preventing a multitude of similar complaints in the future. By integrating emotion analysis into CRM systems, businesses can respond to customers more effectively and preempt issues, improve their offerings, and ultimately enhance customer satisfaction and loyalty.

2.4 Healthcare and mental health support

In healthcare, understanding patients’ emotions can enhance patient-provider communication and potentially improve care delivery [2, 3, 4, 5]. Healthcare providers can use emotion analysis to assess patients’ feelings about their treatment, helping to adapt it better to suit their emotional needs. A patient’s complex emotions might include fear, confusion, or hope, which could significantly impact their treatment and recovery process. Each application shows how embracing emotional complexity provides a richer understanding of human feeling, fostering more authentic and effective connections and solutions. By continuing to improve and develop emotion analysis techniques, we open up a world of possibilities for greater emotional understanding in our increasingly digital era. Prior research [18] demonstrated that social media can effectively detect and diagnose major depressive disorder through behavioral cues, while [19] highlighted the feasibility and efficacy of conversational agents like Woebot in delivering self-help interventions for anxiety and depression, emphasizing their potential as engaging tools for proactive mental health care. This empathetic interaction brings mental health support to those who might otherwise not have access, providing comfort and understanding in a non-judgmental, AI-powered space.

Advertisement

3. Literature review

In this literature review section, we explore the frontiers of emotion detection, examining the pioneering techniques that have emerged in this arena in the past few years. Advanced emotion analysis methodologies, analogous to precision instruments, can unravel the complex network of human emotions intricately, unveiling profound insights into our behavioral patterns, decision-making processes, and interpersonal dynamics. These sophisticated techniques not only cater to academic interest but are also designed in a manner that can captivate the curiosity of an everyday reader, owing to the universality of the emotional experience they decode.

3.1 Transformer-based models

The field of emotion analysis has undergone a profound transformation with the advent of transformer-based models such as BERT (Bidirectional Encoder Representations from Transformers) [20], GPT-3 (Generative Pretrained Transformer 3) [21], T5 (Text-to-Text Transfer Transformer) [22], and LLaMA (Large Language Model Meta AI) [23]. BERT employs bidirectional training of transformers, facilitating a nuanced understanding of word context by referencing the surrounding text. GPT-3, an autoregressive language model, utilizes machine learning techniques to generate human-like text. T5 innovatively reformulates every natural language processing task as a text-to-text problem, thereby training the model on diverse tasks. LLaMA, a collection of foundation language models, enables the adaptation of large pre-trained models to specific tasks, negating the necessity for extensive fine-tuning.

All these models, including their various adaptations, employ self-attention mechanisms [7]. This fundamental component of transformer architectures empowers models to assess the relevance of words in a sentence according to their contextual interrelation rather than their standalone significance. Consequently, this mechanism supports an understanding of the broader context of a sentence, including crucial linguistic elements such as negation.

Moreover, transformer-based models can also learn subtle emotional nuances crucial in complex emotion detection. Consider a statement such as, “It is a fine day.” Here, the word “fine” may convey a cheerful or neutral emotion depending on the speaker’s tone, context, and individual language usage habits. A model like BERT, pre-trained on a large text corpus, might have encountered similar usage patterns and could more accurately predict the intended emotion.

3.2 Context-aware emotion analysis

Context plays an important role in understanding human emotions. The same statement can hold different emotional connotations in different scenarios. Consider a tweet saying, “The final season was mind-blowing!” Without context, it is impossible to determine whether the emotion is positive (excitement about a television show season final) or negative (criticism of a political leader’s final term). Context-aware emotion analysis approaches try to incorporate such context by considering additional information about the source, subject matter, or surrounding text or by using sophisticated models capable of learning contextual representations [10, 11].

Despite the advancement in LLMs (Large Language Models) that process vast amounts of text data using deep learning techniques to generate coherent and contextually relevant text resembling human language [24], such as transformer-based like “BERT, GPT-3, T5, LLaMA”, and others [20, 21, 22, 23] are designed to consider the context when processing text. The transformer architecture is built around self-attention [7], allowing the model to weigh the importance of each word in a sentence when trying to understand or generate a particular word. These models can take into account the broader context to a certain extent. Nevertheless, they only explicitly measure aspects like emotion or subjectivity if specifically trained to do so in tasks such as emotion analysis. However, their ability to account for context can benefit such tasks.

Context-aware emotion analysis is a more specialized task that explicitly seeks to understand the feeling in light of the broader context. So, in situations where understanding opinion is critical (like customer reviews and social media monitoring.), models specifically trained for context-aware emotion analysis could be more effective than a general-purpose transformer model. However, there is vastly active research in this area, and many of the latest transformer-based models are very good at tasks like emotion analysis, even in complex and nuanced situations. Nonetheless, they could be better, and there can still be instances where they might not fully capture the emotion, especially in cases where more profound domain knowledge or cultural understanding is required.

Consider the following hypothetical scenario: a sentence states,” The company’s latest yearly earnings report showed a decline in revenue but an increase in market share.” Trying to find whether the sentence represents satisfaction and happiness or disappointment and fear can be difficult without context and domain knowledge, as it might be either positive(satisfaction) or negative(disappointment); for instance, considering the company in a highly competitive industry, “the company’s latest yearly earnings report showed a decline in revenue but an increase in market share” might suggest that the company has successfully gained a larger market share than its competitors, indicating potential long-term growth and profitability, therefore, the overall message could be satisfaction and happiness. However, on the other hand, the company’s latest yearly earnings report showed a decline in revenue might raise concerns about a negative trend in sales. Despite an increase in market share, the decline in revenue suggests challenges in attracting customers or generating sufficient sales, potentially impacting the company’s overall financial health; therefore, the overall emotion could be disappointment and fear.

3.3 Emotion recognition using physiological signals

Recent advancements in wearable technology have opened up new avenues for emotion detection [9]. Wearable devices can capture physiological signals such as heart rate and skin temperature, galvanic skin response, and even brainwave patterns, which are demonstrably linked to emotional states. For example, a sudden spike in heart rate and skin temperature might indicate a state of excitement or stress. To illustrate, heart rate variability (HRV), the variation in time between each heartbeat, is a robust indicator of an individual’s emotional state [9, 25]. Numerous studies have substantiated the correlation between HRV and emotions; for instance, an elevated heart rate and reduced HRV often signify a heightened emotional state, such as excitement or tension. Similarly, skin temperature is another physiological signal that varies with emotional changes. Research reveals that skin temperature increases during periods of intense emotional arousal due to the activation of the sympathetic nervous system. A sudden spike in skin temperature indicates a state of excitement, fear, or anger.

Moreover, Galvanic Skin Response (GSR), which measures changes in the skin’s conductance, is highly sensitive to emotional arousal [16, 26]. An emotional event triggers the sweat glands, increasing skin conductance—a phenomenon that wearables can accurately measure, aiding in emotion recognition. Electroencephalogram (EEG) signals have also been used in emotion recognition, although more challenging to acquire outside clinical or research settings [16, 26, 27]. Brainwave patterns have been associated with different emotional states, opening up the possibility of emotion detection from EEG data.

However, it is important to note that these methods are relatively more invasive than emotion analysis based on text or speech, and their usage must comply with strict privacy and consent regulations. While these techniques are more intrusive and require user consent, analyzing physiological signals for emotion recognition has considerable potential for applications spanning various fields. These include but are not limited to health monitoring—where it can aid in diagnosing and treating mood disorders, stress management—by providing real-time biofeedback to users, and even the domain of personalized recommendations—where consumer emotional response to products can guide tailored marketing strategies.

3.4 Multimodal emotion analysis

While text is a critical channel for expressing emotions, it is not the sole medium through which emotions can be communicated or understood. Emotions are multi-dimensional and can be transmitted through a multitude of channels. Visual cues (like images or videos) and auditory signals (like tone or voice pitch) also carry crucial emotional information [28]. Multimodal emotion analysis incorporates these multiple data streams to derive a more holistic understanding of emotion.

Text-based models, including the most recent advanced transformer-based have indisputably revolutionized the field of emotion analysis concerning textual data. But, these models fall short in their capability to predict emotion expressed through visual or auditory mediums. In light of limitations identified by previous researchers [27, 29], which highlight the challenges of relying solely on text-based emotion analysis, implementing multimodal emotion analysis becomes critical. Multimodal emotion analysis, as detailed in [8, 9, 15, 30], is a method that offers a more precise and comprehensive overview of emotion by integrating insights from various data types such as text, audio, and video. For instance, during the evaluation of customer service calls, a customer might verbalize their satisfaction, yet underlying tones of frustration or disappointment might be discernable through their tone or hesitation. Likewise, in the analysis of video reviews, visual expressions prove essential in accurately conveying emotions, further underscoring the indispensable role of a multimodal approach in emotion analysis. Performing multimodal emotion analysis is a complex task that involves various steps and processes. The idea is to combine information from different modes (like text, audio, and video) to make predictions.

It is worth noting that aligning multimodal data can be a significant challenge, especially in scenarios where the data is collected independently or in an unstructured way. However, there are some strategies to ensure the correct association:

  • Timestamps: If each data source (text, audio, video) has an associated timestamp, you could align the data based on these timestamps.

  • Identifier Labels: If your data comes from a system that tags or labels each data entry with a unique identifier, you can use this to align your data.

  • Sequential Alignment: If your data comes in a sequential or chronological order (for instance, a video transcript), you can correlate the data based on the order.

  • Manual Alignment: Manual alignment can be performed if your dataset is not too large. This method can be very accurate but is time-consuming and requires many resources.

The aforementioned techniques are revolutionizing the way we understand and analyze human emotions. However, it is worth noting that each technique has strengths and weaknesses and may be more suited to particular applications than others. Furthermore, they all grapple with challenges such as ambiguity, subjectivity, and cultural variations in emotional expression, echoing the need for continuous research, refinement, and innovation in the field of emotion analysis.

Advanced emotion analysis techniques serve as our compass, guiding us towards a deeper and more nuanced understanding of our collective emotional landscape. Embracing emotional complexity is not an end goal but a continuous journey, like our understanding and exploration of human emotion.

Advertisement

4. Methodology

This analysis adopts a systematic approach to investigate human emotions within cloud providers’ services. The methodology encompasses data collection, cleaning, preprocessing, exploratory data analysis, data splitting, model Fine-tuning, testing, and evaluation utilizing the transformer-based architecture as illustrated in (Figure 1). The following subsections provide a comprehensive outline of each phase:

Figure 1.

Research methodology.

4.1 Data collection

We rely on Twitter’s Rest APIs as our primary data source to build a comprehensive and representative dataset, providing access to many tweets about cloud computing services. Twitter’s large user base and concise tweet format make it an ideal platform for capturing and analyzing human emotions.

We have deliberately collected tweets as our primary data source to capture the nuanced and often ambiguous expression of human emotions. By avoiding pre-made datasets with predefined emotion labels, we aim to observe how our model performs in real-world scenarios where emotions are often conveyed in a complex and uncertain manner. This approach allows us to assess the model’s ability to handle emotional expression’s inherent variability and subtleties in social media contexts.

Our data collection strategy focuses on retrieving English language tweets utilizing specific hashtags such as “Azure, azurecloud, azure, AWS, awscloud, amazoncloud, GoogleCloud, GCPCloud, googlecloud”. These hashtags are widely used in cloud computing discussions, ensuring the relevance of the collected data. We also exclude retweets to prioritize original content and capture authentic user emotion.

4.2 Data cleaning and preprocessing

The data cleaning and preprocessing phase in our research played a pivotal role in preparing our dataset of 9200 tweets for the subsequent stages of emotion analysis. We processed the raw tweets to remove extraneous elements such as URLs, user handles, hashtags, and certain punctuation marks. To minimize semantic discrepancies, we transformed the entire dataset into lowercase. Duplicate tweets were also identified and removed at this stage.

A crucial phase of our preprocessing involved identifying prospective emotion classes by applying the K-means clustering algorithm [31]. To ascertain the optimal number of clusters, we used the sum of squared errors across various cluster counts and capitalized on the elbow method, as shown in (Figure 2).

Figure 2.

Elbow analysis in K-means clustering.

The count of clusters was further substantiated by visualizing the groups generated by the K-means algorithm and analyzing the inflection point on the plot of the Sum of Squared Errors (SSE) against the number of clusters, as depicted in (Figure 3). This visualization distinctly represented data points and cluster centers; each data point was color-coded according to its assigned group. Upon thoroughly assessing these visualizations, coupled with the outcomes of the elbow method, we determined that three represented the optimal number of clusters.

Figure 3.

Multiple K value in K-means clustering.

Through these comprehensive preprocessing steps and the utilization of unsupervised learning techniques, we gleaned valuable insights about potential emotion classes within our dataset, structured the raw data, and refined it into a form that optimizes the effectiveness and accuracy of our machine learning model. These actions formed the foundation for our subsequent manual labeling process and emotion prediction efforts. However, to get explicit emotion labels like ‘anger’ and ‘joy,’ we manually read through the tweets and assign emotions to the tweets dataset.

The data classification involved a harmonious blend of unsupervised learning and manual labeling. This mixed-methods approach allowed us to delineate complex emotion categories within our dataset and build a foundation for our emotion prediction model.

4.3 Data exploration

Exploratory Data Analysis plays a pivotal role in comprehending the dataset and discerning the inherent characteristics of tweets about cloud providers’ services. Various visualization techniques were employed to explore the data effectively and gain valuable insights.

The following Table 1 presents statistics, such as the total number of tweets, average text length, and the number of unique words, offering a concise overview of the dataset.

Total number of tweetsAverage text lengthNumber of unique words
9200108.07347815958

Table 1.

Summary table.

The word frequency, illustrated in (Figure 4), played a significant role in pinpointing common terms, offering key insights into the dominant linguistic patterns captured in the tweets. This depiction highlights the words that appear most frequently and contributes to a comprehensive understanding of the linguistic traits embodied within the dataset.

Figure 4.

Terms frequency.

In parallel, the word cloud visualization (Figure 5) was instrumental in visually representing and illuminating prominent themes within the tweets dataset. The word cloud effectively highlighted the most significant terms by employing varying font sizes to denote word frequency or importance.

Figure 5.

Prominent terms.

The subsequent Table 2 summarizes the association between each tweet and its emotion. This provided an illustrative snapshot of the tweets dataset. This compilation, thus, forms a solid foundation for subsequent stages of our research, where these categories will be employed for further emotion analysis.

TweetEmotion
If it was e.g. SharePoint we could simply add a script to the master page and everything would be covered. Why cannot we do something similar with PowerApps?distress
Microsoft Azure DevOps server and team foundation server information disclosure vulnerabilitydistress
I love when we are professional in our explanationsjoy
Function-as-a-Service? Demystifying serverless deployments with by Chris Kipp. Let us use Zeit Now platform as a super simple example of how easy it is to get applications up and running!joy
Businesses use to obtain smarter and assure uninterrupted service delivery in environments. Find out how our for reduces risk and increases agilityneutral
Many games are already using Azure services for multiplayer gamesneutral

Table 2.

Tweets with corresponding emotions.

The bar chart, as portrayed in (Figure 6), was utilized to represent the diverse range of emotions in our dataset. Not only does this chart depict the distribution of each emotion, but it also shows their proportional representation relative to the entire data set. This allows for a clear visual comparison, highlighting the predominance or rarity of specific emotions.

Figure 6.

Emotions distribution.

The histogram, as visualized in (Figure 7), was systematically employed to uncover the underlying patterns and potential anomalies associated with the length of the tweets.

Figure 7.

The distribution of tweet text lengths.

4.4 Data splitting

The cleaned dataset was split into training, validation, and test sets through stratified sampling, ensuring each set contained a representative proportion of samples for each target emotion. This is an important consideration, especially for imbalanced datasets, ensuring that the balance of each emotion in the training, validation, and test sets mirrors that in the original dataset.

Following the data segregation, an analysis was conducted to understand the distribution of emotions across the different subsets. The distribution was visualized using a grouped bar chart (Figure 8), which exhibited the proportion of each emotion in the training, validation, and test datasets.

Figure 8.

Emotions distribution.

The following Table 3 presents a summary of the emotion distribution in the tweets dataset. Each row represents a specific emotion, its corresponding encoding, and the count of tweets associated with that emotion. The dataset consists of 4982 tweets labeled as “neutral” (encoding: 2), 2628 tweets labeled as “joy” (encoding: 1), and 1590 tweets labeled as “distress” (encoding: 0). This information provides an overview of the distribution of emotions within the dataset and serves as a foundation for further analysis and modeling.

LabelEncodingCount of tweets
Neutral24982
Joy12628
Distress01590

Table 3.

Summary of emotion distribution in tweets.

4.5 Feature engineering and selection

In the feature engineering step within our methodology, the unprocessed tweet corpus was converted into a structured format compatible with the RoBERTa-based model. This process involved tokenization of the tweets into subwords, adjusting these sequences to a fixed length, and creating attention masks. Each component is critical in ensuring our model can accurately interpret the data.

The tokenization stage uses a pre-trained RoBERTa tokenizer to convert each tweet into a sequence of subword tokens. Each token is represented as a unique integer identifier. This approach mitigates the limitation of a fixed vocabulary and helps to preserve meaningful linguistic nuances that could otherwise be lost.

The below Table 4 shows the properties of the token sequences and a statistical analysis of the lengths of these sequences post-tokenization. The summary statistics indicated that, from the corpus of 9200 tweets, the average token sequence length is approximately 24.48 tokens, with a standard deviation of 12.55. The shortest tokenized tweet consists of only three tokens, while the longest extends to 81 tokens. Notably, 50% of the tweets have a token length of 22 or fewer tokens, revealing a substantial skewness in the distribution towards shorter tweets.

StatisticValue
Count9200
Mean24.48
Standard deviation12.55
Min3
Median22
Max81

Table 4.

Statistics of token lengths.

Since transformer-based models require input data in uniform length, the sequences were padded or truncated to a pre-defined maximum length of 64 tokens. Padding is indispensable when the token length is less than the maximum defined length, filling the residual positions with a designated padding token. Conversely, for sequences exceeding the maximum size, truncation ensures the sequence is limited to the initial 64 tokens.

Upon achieving uniform token sequences, attention masks were generated for each tweet. An attention mask, essentially a binary tensor, indicates the positions containing actual content (denoted by 1 s) versus the padded positions (represented by 0 s). This plays a pivotal role in focusing the model’s attention on the substantive content of each sequence, disregarding the irrelevant padded elements during self-attention computations.

To visually explain the distribution of content and padding across tweets, we created a heatmap as shown in (Figure 9) of the attention masks. In this heatmap, the x-axis signifies the token positions across a tweet, while the y-axis corresponds to individual tweets. Each cell’s color intensity indicates whether the corresponding position is filled with content (darker shades) or padding (lighter shades). This visualization outlines the areas of real content in each tweet sequence, demonstrating the effectiveness of our feature engineering methodology.

Figure 9.

Attention masks.

4.6 Model fine-tuning

The RoBERTa transformer-based architecture [32] was selected for its effectiveness in natural language processing tasks. The model was fine-tuned using the preprocessed dataset, optimizing its ability to accurately predict emotions associated with the tweets dataset regarding the cloud provider’s services.

The model was trained over four epochs, concluding at 400 global steps with a final training loss of 0.47, as depicted in the following Table 5. Training loss consistently decreased, but an increase in validation loss after the fourth epoch signaled overfitting. The training process was halted after the fourth epoch to safeguard the model’s capacity for generalization on unfamiliar data. Subsequently, the model was explicitly fine-tuned over these optimal four epochs. The research presents results from this optimal point, avoiding overfitting bias.

StepTraining lossValidation loss
1000.9114000.597774
2000.5222000.583825
3000.4443000.595613
4000.3460000.592506

Table 5.

Model training.

Trained on a MacBook Pro with M2 Max processor (12-Core CPU, 38-Core GPU), and 32GB of unified memory. TrainOutput (global step = 400, training loss = 0.4781294107437134,total flos: 1125220975502400.0, epoch: 4.0).

4.7 Model validation and testing

The model testing process involves testing the model using the test dataset. The model’s predictions are obtained by performing inference on the test dataset, and these predictions are converted into class labels using a label mapping dictionary. The actual labels are also transformed into their corresponding emotion labels.

The following Table 6 displays a subset of the predictions to summarize the model’s performance. Each table row represents a tweet from the test dataset, with the “Tweet” column showing a truncated version of the tweet text. The “Token IDs” column represents a truncated version of the tokenized representation of the tweet. The “Predicted” column displays the model’s predicted emotion label for each tweet, while the “Actual” column indicates the ground truth emotion label for comparison.

TweetToken IDsPredictedActual
struggling with managing your cloud-comp…[0, 23543, 6149, 1527, 19], …distressdistress
certified with certifications yet? visi…[0, 25782, 3786, 19, 21045], …neutralneutral
ready to transform your business with bi…[0, 16402, 7, 7891, 110], …joyjoy
enterprise customers are on the move to…[0, 11798, 22627, 916, 32], …neutralneutral
wow! congratulations miles!! theyre l…[0, 34798, 27785, 24285, 1788], …joyjoy
hansis willhite earn academic all-distri…[0, 298, 1253, 354, 40], …joyneutral

Table 6.

Tweet predictions.

The table serves as a means to compare the predicted and actual emotion labels for a subset of the test dataset, providing insights into the model’s performance in accurately classifying emotions based on the input tweets. It offers valuable information regarding the model’s ability to discern and predict emotions within the context.

4.8 Model evaluation

The performance of this model was evaluated using several widely accepted metrics, including accuracy, precision, recall, and the F1 score.

As depicted in the subsequent Table 7, the model achieved an accuracy of 82.53%. This indicates that, across all predictions, approximately 82.53% of the model’s predictions matched the actual classes. This is a tangible result because our problem has more than two classes. Precision for our model stood at 82.79%. Precision measures how many of the model’s positive predictions were correct. In other words, when our model predicted an emotion, it was correct about 82.79% of the time. The model’s recall score was also 82.53%, matching the accuracy. Recall measures how well the model can find all the relevant cases within a dataset. The same recall and accuracy suggest a balanced distribution of class labels in the dataset, and the model is equally good at predicting all classes. The F1 score of the model, a harmonic mean of precision and recall, was 82.29%. The F1 score is a better metric when there are imbalanced classes, as it considers both false positives and false negatives.

AccuracyPrecisionRecallF1 Score
0.82536231884057980.82793770077871030.82536231884057980.8229926445758601

Table 7.

Model evaluation metrics.

The model showed varying performance in terms of per-class results as displayed in Table 8. The matrix illustrates the model’s performance in predicting each class and the instances where it was incorrect. For the ‘distress’ class, the model accurately predicted 153 instances out of the actual distress samples. However, it incorrectly classified 3 instances as ‘joy’ and 83 instances as ‘neutral.’ In the case of the ‘joy’ class, the model demonstrated a relatively higher accuracy, correctly identifying 335 instances. Nevertheless, it misclassified 1 instance as ‘distress’ and 58 instances as ‘neutral.’ Regarding the ‘neutral’ class, the model achieved accurate predictions for 651 instances. However, it erroneously classified 21 instances as ‘distress’ and 75 instances as ‘joy.’

DistressJoyNeutral
Distress153383
Joy133558
Neutral2175651

Table 8.

Confusion matrix.

The confusion matrix provides a comprehensive overview of the model’s performance across different classes, revealing both correct predictions and misclassifications. This analysis offers valuable insights into the model’s ability to distinguish between various emotional categories.

The confusion matrix (Figure 10) provides us with a perspective on model performance. Although the model has shown a reasonably good overall performance, it could be improved further, particularly in its ability to correctly predict ‘distress’ and ‘neutral,’ as indicated by the number of instances misclassified into these categories.

Figure 10.

Terms frequency bar chart.

4.9 Model deployment

Model deployment refers to making a trained model available in a production environment, where it can provide predictions to new input data. In the context of our study, the deployment of the emotion analysis model involves wrapping the model into an Application Programming Interface (API), which can serve as a standardized interface for other software components to communicate with the model. Specifically, the API receives raw tweet data as input, preprocesses the data in the same way as during the model training phase, passes the preprocessed data to the model, and finally outputs the model’s emotion predictions. The deployment of the model as an API offers several benefits. Firstly, it facilitates the integration of the model into existing systems or workflows, as these systems can interact with the model simply by making requests to the API. Secondly, it allows the model to be hosted on a server and concurrently provide predictions as a service to multiple users or systems.

4.10 Continuous improvement

For the continuous improvement of our model, we can leverage real-time feedback from users or fine-tune the model on new data. Users who disagree with the emotion prediction could submit the correct emotion, which can be stored alongside the original tweet. This valuable data can then be used for further fine-tuning our model. Furthermore, we should constantly monitor the model’s performance in production, as the input data distribution may change over time, which could affect the model’s performance. Regular retraining or fine-tuning of the model with recent data will be essential. Finally, ensuring that the API’s performance and uptime are satisfactory is essential, as this can directly impact the user experience. Following these practices ensures that our model continuously improves and stays robust and valuable over time.

Advertisement

5. Challenges, limitations and ethical considerations

While the journey of deciphering human emotions through emotion analysis has proven fruitful, it has also been fraught with challenges and limitations [27, 29], encased in layers of ethical considerations [33, 34]. This section examines these aspects, highlighting their inherent complexity and the ongoing need for meticulous attention and innovation in addressing them.

5.1 Challenges

The primary challenge is human emotions’ inherent ambiguity and subjectivity. Human language is replete with nuances, context-specific implications, and idiomatic expressions, making it challenging to assign a particular emotion accurately. For instance, sarcasm, an expressive form of emotion, often implies the opposite of the literal emotion, posing a significant challenge to emotion analysis algorithms. Another substantial challenge is dealing with the ever-evolving nature of language, especially in informal platforms like social media, where new slang and emoticons frequently appear. This dynamic environment necessitates continual learning and adaptation, pushing the boundaries of emotion analysis techniques.

5.2 Limitations

Despite rapid advancements, emotion analysis techniques are inherently limited in their ability to comprehend the full spectrum of human emotions due to their reliance on predefined categories and labels. While these techniques excel at identifying basic emotions such as joy, sadness, anger, and fear, they often falter when faced with more nuanced emotions such as sarcasm, irony, or mixed emotions. Moreover, most emotion analysis methods are primarily text-based, limiting their applicability in scenarios where emotions are conveyed through other means like tone, facial expressions, or body language. Even multimodal emotion analysis, which incorporates visual and auditory information, faces limitations due to the complexity and diversity of non-verbal emotional cues.

5.3 Ethical considerations

Emotion analysis, especially when applied at scale on social media and other digital platforms, raises several ethical questions. The first is the matter of privacy and consent. While public posts can be considered fair game, is it ethical to analyze a person’s emotional state without explicit consent? Additionally, how the results of emotion analysis are used also poses ethical concerns. For example, using emotion analysis to manipulate public opinion, target vulnerable individuals, or perpetrate discrimination is ethically questionable. Furthermore, the risk of reinforcing biases presents another ethical dilemma. If an emotion analysis model is trained on biased data, it can perpetuate harmful stereotypes, leading to unfair outcomes. For instance, the model could falsely associate certain dialects or speech patterns with negative emotions, leading to discriminatory practices.

While emotion analysis offers powerful tools for understanding human emotions, it is not without its challenges, limitations, and ethical considerations. Ambiguity and subjectivity, evolving language norms, and the difficulty of capturing the full range of human emotions underscore the complexities involved. Ethical issues, including privacy, consent, and potential misuse of analysis results, further complicate matters. Therefore, continued research, careful methodological refinement, and thoughtful ethical guidelines are crucial to advancing emotion analysis in a manner that respects and upholds our shared human values.

Advertisement

6. Conclusion

In summary, this chapter has navigated through the nuanced universe of human feelings, employing cutting-edge emotion analysis methodologies. It underscores the layered nature of emotions and the crucial role of precise emotion detection spanning diverse applications. Forefront approaches, such as transformer-based models, multimodal emotion analysis that amalgamates text, audio, and visual data, context-aware emotion analysis that considers situational variables, and emotion recognition using physiological signals, have been under the spotlight with their contributions towards a comprehensive understanding of human emotions.

Our implementation of the transformer-based model on a tweets dataset achieved an accuracy of 82.53%, a precision of 82.79%, a recall of 82.53%, and an F1 score of 82.29%. These results underscore the efficacy of such models in understanding and predicting emotional categories. Their practical applications span sectors like personalized marketing, social media analytics, healthcare services, and customer relationship management, demonstrating the potential of emotion analysis to delve into the emotional dimensions of human behavior, facilitating more genuine connections and effective outcomes.

The discourse also highlights potential roadblocks, constraints, and ethical quandaries, such as interpretation difficulties, subjective bias, cultural diversities, and privacy-related issues. Further research should address these challenges, explore ways to reduce bias, and improve accuracy across diverse cultural contexts.

As we gaze into the future, emotion analysis will move towards a more integrated approach by incorporating emotional intelligence with artificial intelligence systems and tailoring techniques to individual needs. By weaving in emotional subtleties and accounting for cultural contexts, emotion analysis can offer a holistic insight into human emotions in our globally linked society. These advancements will be critical in grasping the complexity of emotions, which is pivotal for the progress of emotion analysis and the ethical and practical application of human emotions’ power.

The expedition to decode the labyrinth of emotions fosters vast possibilities for more research, collaboration, and innovation. By tackling challenges, honing methodologies, and adhering to ethical norms, emotion analysis can sustain its evolution, paying heed to the complexities of human emotions while preserving our collective human values.

References

  1. 1. Cambria E, Schuller B, Xia Y, Havasi C. New avenues in opinion mining and sentiment analysis. IEEE Intelligent Systems. 2013;28:15-21. ISSN 15411672. DOI: 10.1109/MIS.2013.30
  2. 2. Bollen J, Mao H, Zeng X. Twitter mood predicts the stock market. Journal of Computational Science. 2011:1-8. ISSN 18777503;2. DOI: 10.1016/j.jocs.2010.12.007
  3. 3. Liu B. Sentiment analysis and opinion mining. In: Synthesis Lectures on Human Language Technologies. Vol. 5. California, USA: Morgan & Claypool Publishers; 2012. ISSN 19474040. DOI: 10.2200/S00416ED1V01Y201204HLT016
  4. 4. Shu L, Xie J, Yang M, Li Z, Li Z, Liao D, et al. A review of emotion recognition using physiological signals. Sensors. 2018;18:1-33. ISSN 14248220.
  5. 5. Lovejoy CA, Buch V, Maruthappu M. Technology and mental health: The role of artificial intelligence. European Psychiatry. 2019;55:1-3. ISSN 17783585.
  6. 6. Greaves F, Ramirez-Cano D, Millett C, Darzi A, Donaldson L. Harnessing the cloud of patient experience: Using social media to detect poor quality healthcare. BMJ Quality and Safety. 2013;22:183-186. ISSN 20445415.
  7. 7. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. Advances in Neural Information Processing Systems. 2017;2017:1-9
  8. 8. Soleymani M, Lichtenauer J, Pun T, Pantic M. A multimodal database for affect recognition and implicit tagging. IEEE Transactions on Affective Computing. 2012;3:42-55. ISSN 19493045. DOI: 10.1109/T-AFFC.2011.25
  9. 9. Poria S, Chaturvedi I, Cambria E, Hussain A. Convolutional mkl based multimodal emotion recognition and sentiment analysis. Proceedings - IEEE International Conference on Data Mining, ICDM. 2017:439-448. DOI: 10.1109/ICDM.2016.178
  10. 10. Hazarika D, Poria S, Zadeh A, Cambria E, Morency LP, Zimmermann R. Conversational memory network for emotion recognition in dyadic dialogue videos. Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - Proceedings of the Conference. 2018;1:2122-2132. DOI: 10.18653/v1/n18-1193
  11. 11. Zaheer M, Guruganesh G, Dubey A, Ainslie J, Alberti C, Ontanon S, et al. Big bird: Transformers for longer sequences. Advances in Neural Information Processing Systems. 2020;2020:93-110
  12. 12. Yin Z, Zhao M, Wang Y, Yang J, Zhang J. Recognition of emotions using multimodal physiological signals and an ensemble deep learning model. Computer Methods and Programs in Biomedicine. 2017;140:93-110. ISSN 18727565. DOI: 10.1016/j.cmpb.2016.12.005
  13. 13. Bota PJ, Wang C, Fred ALN, Da Silva HP. A review, current challenges, and future possibilities on emotion recognition using machine learning and physiological signals. IEEE Access. 2019;7:140990-141020. ISSN 21693536
  14. 14. Chen J, Wang C, Wang K, Yin C, Zhao C, Tao X, et al. Heu emotion: A large-scale database for multimodal emotion recognition in the wild. Neural Computing and Applications. 2021;33:8669-8685. ISSN 14333058. DOI: 10.1007/s00521-020-05616-w
  15. 15. Zadeh A, Chen M, Cambria E, Poria S, Morency LP. Tensor fusion network for multimodal sentiment analysis. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing. 2017;2017:1103-1114. DOI: 10.18653/v1/d17-1115
  16. 16. Subramanian R, Wache J, Abadi MK, Vieriu RL, Winkler S, Sebe N. Ascertain: Emotion and personality recognition using commercial sensors. IEEE Transactions on Affective Computing. 2018;9:147-160. ISSN 19493045. DOI: 10.1109/TAFFC.2016.2625250
  17. 17. Karamitsos I, Albarhami S, Apostolopoulos C. Tweet sentiment analysis (tsa) for cloud providers using classification algorithms and latent semantic analysis. Journal of Data Analysis and Information Processing. 2019;07:276-294. ISSN 2327-7211. DOI: 10.4236/jdaip.2019.74016
  18. 18. De Choudhury M, Gamon M, Counts S, Horvitz E. Predicting depression via social media. Proceedings of the 7th International Conference on Weblogs and Social Media. 2013;7:128-137. DOI: 10.1609/icwsm.v7i1.14432
  19. 19. Fitzpatrick KK, Darcy A, Vierhile M. Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (woebot): A randomized controlled trial. JMIR Mental Health. 2017;4:1-9. ISSN 23687959. DOI: 10.2196/mental.7785
  20. 20. Devlin J, Chang MW, Lee K, Toutanova K. Bert: Pre-training of deep bidirectional transformers for language understanding. Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT 2019). 2019;1:4171-4186
  21. 21. Brown TB, Mann B, Ryder N, Subbiah M, Kaplan J, Dhariwal P, et al. Language models are few-shot learners. Advances in Neural Information Processing Systems. 2020;2020:1-19
  22. 22. Raffel C, Shazeer N, Roberts A, Lee K, Narang S, Matena M, et al. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research. 2020;21:1-46. ISSN 15337928
  23. 23. Touvron H, Lavril T, Izacard G, Martinet X, Lachaux MA, Lacroix T et al. Llama: Open and efficient foundation language models. arXiv. 2023:1-12. eprint - 2302.13971
  24. 24. Radford A, Narasimhan K, Salimans T, Sutskever I, et al. Improving language understanding by generative pre-training. OpenAI. 2018:1-8
  25. 25. Appelhans BM Luecken LJ. Heart rate variability as an index of regulated emotional responding. Review of General Psychology. 2006;10:229-240. ISSN 10892680
  26. 26. Kwon YH, Shin SB, Kim SD. Electroencephalography based fusion two-dimensional (2d)-convolution neural networks (CNN) model for emotion recognition system. Sensors (Switzerland). 2018;18:1-12. ISSN 14248220. DOI: 10.3390/s18051383
  27. 27. Zhang J, Yin Z, Chen P, Nichele S. Emotion recognition using multi-modal data and machine learning techniques: A tutorial and review. Information Fusion. 2020;59:103-126. ISSN 15662535. DOI: 10.1016/j.inffus.2020.01.011
  28. 28. Hubert Tsai YH, Bai S, Paul Pu Liang J, Kolter Z, Morency LP, Salakhutdinov R. Multimodal transformer for unaligned multimodal language sequences. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics. 2020:6558-6569. DOI: 10.18653/v1/p19-1656
  29. 29. Poria S, Hazarika D, Majumder N, Mihalcea R. Beneath the tip of the iceberg: Current challenges and new directions in sentiment analysis research. IEEE Transactions on Affective Computing. 2023;14:108-132. ISSN 19493045. DOI: 10.1109/TAFFC.2020.3038167
  30. 30. Zadeh A, Liang PP, Vanbriesen J, Poria S, Tong E, Cambria E, et al. Multimodal Language Analysis in the Wild: Cmu-Mosei Dataset and Interpretable Dynamic Fusion Graph. Vol. 1. Copenhagen, Denmark: Association for Computational Linguistics (ACL); 2018. ISBN 9781948087322. pp. 2236-2246. DOI: 10.18653/v1/p18-1208
  31. 31. MacQueen J et al. Some methods for classification and analysis of multivariate observations. In: Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability. Vol. 1. Oakland, CA, USA: University of California Press; 1967. pp. 281-297
  32. 32. Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D et al. Roberta: A robustly optimized bert pretraining approach. Preprint Identifier: arXiv:1907.11692. 2019:1-13
  33. 33. Brent Daniel Mittelstadt and Luciano Floridi. The ethics of big data: Current and foreseeable issues in biomedical contexts. Science and Engineering Ethics. 2016;22:303-341. ISSN 14715546
  34. 34. Datta A, Sen S, Zick Y. Algorithmic transparency via quantitative input influence: Theory and experiments with learning systems. Proceedings - 2016 IEEE Symposium on Security and Privacy. 2016;2016:598-617. DOI: 10.1109/SP.2016.42

Written By

Saeed Albarhami Thabit

Reviewed: 31 July 2023 Published: 10 January 2024