56:
295:– a statistical Markov model in which the states and state transitions are not directly available to observation. Instead, the series of outputs dependent on the states are visible. In the case of affect recognition, the outputs represent the sequence of speech feature vectors, which allow the deduction of states' sequences through which the model progressed. The states can consist of various intermediate steps in the expression of an emotion, and each of them has a probability distribution over the possible output vectors. The states' sequences allow us to predict the affective state which we are trying to classify, and this is one of the most commonly used techniques within the area of speech affect detection.
727:, which produces a graph indicating blood flow through the extremities. The peaks of the waves indicate a cardiac cycle where the heart has pumped blood to the extremities. If the subject experiences fear or is startled, their heart usually 'jumps' and beats quickly for some time, causing the amplitude of the cardiac cycle to increase. This can clearly be seen on a photoplethysmograph when the distance between the trough and the peak of the wave has decreased. As the subject calms down, and as the body's inner core expands, allowing more blood to flow back to the extremities, the cycle will return to normal.
423:
databases, the participants are asked to display different basic emotional expressions, while in spontaneous expression database, the expressions are natural. Spontaneous emotion elicitation requires significant effort in the selection of proper stimuli which can lead to a rich display of intended emotions. Secondly, the process involves tagging of emotions by trained individuals manually which makes the databases highly reliable. Since perception of expressions and their intensity is subjective in nature, the annotation by experts is essential for the purpose of validation.
846:
ability, and then formulate reasonable teaching plans. At the same time, they can pay attention to students' inner feelings, which is helpful to students' psychological health. Especially in distance education, due to the separation of time and space, there is no emotional incentive between teachers and students for two-way communication. Without the atmosphere brought by traditional classroom learning, students are easily bored, and affect the learning effect. Applying affective computing in distance education system can effectively improve this situation.
22:
151:. The goal of most of these techniques is to produce labels that would match the labels a human perceiver would give in the same situation: For example, if a person makes a facial expression furrowing their brow, then the computer vision system might be taught to label their face as appearing "confused" or as "concentrating" or "slightly negative" (as opposed to positive, which it might say if they were smiling in a happy-appearing way). These labels may or may not correspond to what the person is actually feeling.
3991:
685:
the answer to a question, or they could be complex and meaningful as when communicating with sign language. Without making use of any object or surrounding environment, we can wave our hands, clap or beckon. On the other hand, when using objects, we can point at them, move, touch or handle these. A computer should be able to recognize these, analyze the context and respond in a meaningful way, in order to be efficiently used for Human–Computer
Interaction.
5555:
4001:
334:
as increasing the performance, which is particularly significant to real-time detection. The range of possible choices is vast, with some studies mentioning the use of over 200 distinct features. It is crucial to identify those that are redundant and undesirable in order to optimize the system and increase the success rate of correct emotion detection. The most common speech characteristics are categorized into the following groups.
4759:
4753:
325:
attempt to produce such database was the FAU Aibo
Emotion Corpus for CEICES (Combining Efforts for Improving Automatic Classification of Emotional User States), which was developed based on a realistic context of children (age 10–13) playing with Sony's Aibo robot pet. Likewise, producing one standard database for all emotional research would provide a method of evaluating and comparing different affect recognition systems.
4011:
768:
usually studied to detect emotion: The corrugator supercilii muscle, also known as the 'frowning' muscle, draws the brow down into a frown, and therefore is the best test for negative, unpleasant emotional response.↵The zygomaticus major muscle is responsible for pulling the corners of the mouth back when you smile, and therefore is the muscle used to test for a positive emotional response.
772:
749:
866:, as well as a growing number of robots used in health care benefit from emotional awareness because they can better judge users' and patient's emotional states and alter their actions/programming appropriately. This is especially important in those countries with growing aging populations and/or a lack of younger workers to address their needs.
642:
recognition, affect recognition), the accuracy of modeling and tracking has been an issue. As hardware evolves, as more data are collected and as new discoveries are made and new practices introduced, this lack of accuracy fades, leaving behind noise issues. However, methods for noise removal exist including neighborhood averaging,
776:
game where there is usually not much exciting game play, there is a high level of resistance recorded, which suggests a low level of conductivity and therefore less arousal. This is in clear contrast with the sudden trough where the player is killed as one is usually very stressed and tense as their character is killed in the game.
300:
classifier taken separately. It is compared with two other sets of classifiers: one-against-all (OAA) multiclass SVM with Hybrid kernels and the set of classifiers which consists of the following two basic classifiers: C5.0 and Neural
Network. The proposed variant achieves better performance than the other two sets of classifiers.
823:
converting the pixel color of the standard RGB color space to a color space such as oRGB color space or LMS channels perform better when dealing with faces. So, map the above vector onto the better color space and decompose into red-green and yellow-blue channels. Then use deep learning methods to find equivalent emotions.
792:
sympathetic branch of the autonomic nervous system. As the sweat glands are activated, even before the skin feels sweaty, the level of the EDA can be captured (usually using conductance) and used to discern small changes in autonomic arousal. The more aroused a subject is, the greater the skin conductance tends to be.
963:
improve computer-mediated interpersonal communication. It does not necessarily seek to map emotion into an objective mathematical model for machine interpretation, but rather let humans make sense of each other's emotional expressions in open-ended ways that might be ambiguous, subjective, and sensitive to context.
313:), which assumes the existence of six basic emotions (anger, fear, disgust, surprise, joy, sadness), the others simply being a mix of the former ones. Nevertheless, these still offer high audio quality and balanced classes (although often too few), which contribute to high success rates in recognizing emotions.
934:
One could also use affective state recognition in order to judge the impact of a TV advertisement through a real-time video recording of that person and through the subsequent study of his or her facial expression. Averaging the results obtained on a large group of subjects, one can tell whether that
813:
The surface of the human face is innervated with a large network of blood vessels. Blood flow variations in these vessels yield visible color changes on the face. Whether or not facial emotions activate facial muscles, variations in blood flow, blood pressure, glucose levels, and other changes occur.
767:
Facial electromyography is a technique used to measure the electrical activity of the facial muscles by amplifying the tiny electrical impulses that are generated by muscle fibers when they contract. The face expresses a great deal of emotion, however, there are two main facial muscle groups that are
333:
The complexity of the affect recognition process increases with the number of classes (affects) and speech descriptors used within the classifier. It is, therefore, crucial to select only the most relevant features in order to assure the ability of the model to successfully identify emotions, as well
299:
It is proved that having enough acoustic evidence available the emotional state of a person can be classified by a set of majority voting classifiers. The proposed set of classifiers is based on three main classifiers: kNN, C4.5 and SVM-RBF Kernel. This set achieves better performance than each basic
407:
processing or active appearance models. More than one modalities can be combined or fused (multimodal recognition, e.g. facial expressions and speech prosody, facial expressions and hand gestures, or facial expressions with speech and text for multimodal data and metadata analysis) to provide a more
324:
Despite the numerous advantages which naturalistic data has over acted data, it is difficult to obtain and usually has low emotional intensity. Moreover, data obtained in a natural context has lower signal quality, due to surroundings noise and distance of the subjects from the microphone. The first
251:
As of 2010, the most frequently used classifiers were linear discriminant classifiers (LDC), k-nearest neighbor (k-NN), Gaussian mixture model (GMM), support vector machines (SVM), artificial neural networks (ANN), decision tree algorithms and hidden Markov models (HMMs). Various studies showed that
966:
Picard's critics describe her concept of emotion as "objective, internal, private, and mechanistic". They say it reduces emotion to a discrete psychological signal occurring inside the body that can be measured and which is an input to cognition, undercutting the complexity of emotional experience.
775:
Here we can see a plot of skin resistance measured using GSR and time whilst the subject played a video game. There are several peaks that are clear in the graph, which suggests that GSR is a good method of differentiating between an aroused and a non-aroused state. For example, at the start of the
216:
Various changes in the autonomic nervous system can indirectly alter a person's speech, and affective technologies can leverage this information to recognize emotion. For example, speech produced in a state of fear, anger, or joy becomes fast, loud, and precisely enunciated, with a higher and wider
744:
It can be cumbersome to ensure that the sensor shining an infra-red light and monitoring the reflected light is always pointing at the same extremity, especially seeing as subjects often stretch and readjust their position while using a computer. There are other factors that can affect one's blood
697:
This could be used to detect a user's affective state by monitoring and analyzing their physiological signs. These signs range from changes in heart rate and skin conductance to minute contractions of the facial muscles and changes in facial blood flow. This area is gaining momentum and we are now
688:
There are many proposed methods to detect the body gesture. Some literature differentiates 2 different approaches in gesture recognition: a 3D model based and an appearance-based. The foremost method makes use of 3D information of key elements of the body parts in order to obtain several important
684:
Gestures could be efficiently used as a means of detecting a particular emotional state of the user, especially when used in conjunction with speech and face recognition. Depending on the specific action, gestures could be simple reflexive responses, like lifting your shoulders when you don't know
316:
However, for real life application, naturalistic data is preferred. A naturalistic database can be produced by observation and analysis of subjects in their natural context. Ultimately, such database should allow the system to recognize emotions based on their context as well as work out the goals
200:
The categorical approach tends to use discrete classes such as happy, sad, angry, fearful, surprise, disgust. Different kinds of machine learning regression and classification models can be used for having machines produce continuous or discrete labels. Sometimes models are also built that allow
914:
operations such as steering and maneuvering are used in various fields such as aviation, transportation and medicine. Integrating affective computing capabilities in this type of training systems, in accordance with the adaptive automation approach, has been found to be effective in improving the
791:
Galvanic skin response (GSR) is an outdated term for a more general phenomenon known as or EDA. EDA is a general phenomena whereby the skin's electrical properties change. The skin is innervated by the , so measuring its resistance or conductance provides a way to quantify small changes in the
159:
Another area within affective computing is the design of computational devices proposed to exhibit either innate emotional capabilities or that are capable of convincingly simulating emotions. A more practical approach, based on current technological capabilities, is the simulation of emotions in
822:
Approaches are based on facial color changes. Delaunay triangulation is used to create the triangular local areas. Some of these triangles which define the interior of the mouth and eyes (sclera and iris) are removed. Use the left triangular areas’ pixels to create feature vectors. It shows that
227:
Speech analysis is an effective method of identifying affective state, having an average reported accuracy of 70 to 80% in research from 2003 and 2006. These systems tend to outperform average human accuracy (approximately 60%) but are less accurate than systems which employ other modalities for
930:
One idea put forth by the
Romanian researcher Dr. Nicu Sebe in an interview is the analysis of a person's face while they are using a certain product (he mentioned ice cream as an example). Companies would then be able to use such analysis to infer whether their product will or will not be well
854:
The applications of sensory computing may contribute to improving road safety. For example, a car can monitor the emotion of all occupants and engage in additional safety measures, such as alerting other vehicles if it detects the driver to be angry. In addition, affective computing systems for
831:
Aesthetics, in the world of art and photography, refers to the principles of the nature and appreciation of beauty. Judging beauty and other aesthetic qualities is a highly subjective task. Computer scientists at Penn State treat the challenge of automatically inferring the aesthetic quality of
962:
Picard's focus is human–computer interaction, and her goal for affective computing is to "give computers the ability to recognize, express, and in some cases, 'have' emotions". In contrast, the interactional approach seeks to help "people to understand and experience their own emotions" and to
665:
The FACS combinations do not correspond in a 1:1 way with the emotions that the psychologists originally proposed (note that this lack of a 1:1 mapping also occurs in speech recognition with homophones and homonyms and many other sources of ambiguity, and may be mitigated by bringing in other
641:
As with every computational practice, in affect detection by facial processing, some obstacles need to be surpassed, in order to fully unlock the hidden potential of the overall algorithm or method employed. In the early days of almost every kind of AI-based detection (speech recognition, face
308:
The vast majority of present systems are data-dependent. This creates one of the biggest challenges in detecting emotions based on speech, as it implicates choosing an appropriate database used to train the classifier. Most of the currently possessed data was obtained from actors and is thus a
123:
that capture data about the user's physical state or behavior without interpreting the input. The data gathered is analogous to the cues humans use to perceive emotions in others. For example, a video camera might capture facial expressions, body posture, and gestures, while a microphone might
845:
Affection influences learners' learning state. Using affective computing technology, computers can judge the learners' affection and learning state by recognizing their facial expressions. In education, the teacher can use the analysis result to understand the student's learning and accepting
426:
Researchers work with three types of databases, such as a database of peak expression images only, a database of image sequences portraying an emotion from neutral to its peak, and video clips with emotional annotations. Many facial expression databases have been created and made public for
422:
Creation of an emotion database is a difficult and time-consuming task. However, database creation is an essential step in the creation of a system that will recognize human emotions. Most of the publicly available emotion databases include posed facial expressions only. In posed expression
970:
The interactional approach asserts that though emotion has biophysical aspects, it is "culturally grounded, dynamically experienced, and to some degree constructed in action and interaction". Put another way, it considers "emotion as a social and cultural product experienced through our
832:
pictures using their visual content as a machine learning problem, with a peer-rated on-line photo sharing website as a data source. They extract certain visual features based on the intuition that they can discriminate between aesthetically pleasing and displeasing images.
180:
includes an attempt to give these programs, which simulate humans, the emotional dimension as well, including reactions in accordance with the reaction that a real person would react in a certain emotionally stimulating situation as well as facial expressions and gestures.
445:
proposed the idea that facial expressions of emotion are not culturally determined, but universal. Thus, he suggested that they are biological in origin and can, therefore, be safely and correctly categorized. He therefore officially put forth six basic emotions, in 1972:
196:
In psychology, cognitive science, and in neuroscience, there have been two main approaches for describing how humans perceive and classify emotion: continuous or categorical. The continuous approach tends to use dimensions such as negative vs. positive, calm vs. aroused.
735:
Infra-red light is shone on the skin by special sensor hardware, and the amount of light reflected is measured. The amount of reflected and transmitted light correlates to the BVP as light is absorbed by hemoglobin which is found richly in the bloodstream.
745:
volume pulse. As it is a measure of blood flow through the extremities, if the subject feels hot, or particularly cold, then their body may allow more, or less, blood to flow to the extremities, all of this regardless of the subject's emotional state.
271:– is a probabilistic model used for representing the existence of subpopulations within the overall population. Each sub-population is described using the mixture distribution, which allows for classification of observations into the sub-populations.
869:
Affective computing is also being applied to the development of communicative technologies for use by people with autism. The affective component of a text is also increasingly gaining attention, particularly its role in the so-called emotional or
799:
electrodes placed somewhere on the skin and applying a small voltage between them. To maximize comfort and reduce irritation the electrodes can be placed on the wrist, legs, or feet, which leaves the hands fully free for daily activity.
564:
are action units (AU). They are, basically, a contraction or a relaxation of one or more muscles. Psychologists have proposed the following classification of six basic emotions, according to their action units ("+" here mean "and"):
217:
range in pitch, whereas emotions such as tiredness, boredom, or sadness tend to generate slow, low-pitched, and slurred speech. Some emotions have been found to be more easily computationally identified, such as anger or approval.
228:
emotion detection, such as physiological states or facial expressions. However, since many speech characteristics are independent of semantics or culture, this technique is considered to be a promising route for further research.
855:
monitoring the driver's stress may allow various interventions such as driver assistance systems adjusted according to the stress level and minimal and direct interventions to change the emotional state of the driver.
689:
parameters, like palm position or joint angles. On the other hand, appearance-based systems use images or videos to for direct interpretation. Hand gestures have been a common focus of body gesture detection methods.
927:, such as affective mirrors allowing the user to see how he or she performs; emotion monitoring agents sending a warning before one sends an angry email; or even music players selecting tracks based on mood.
2220:
1156:
289:– work based on following a decision tree in which leaves represent the classification outcome, and branches represent the conjunction of subsequent features that lead to the classification.
265:– Classification happens by locating the object in the feature space, and comparing it with the k nearest neighbors (training examples). The majority vote decides on the classification.
659:
Facial expressions do not always correspond to an underlying emotion that matches them (e.g. they can be posed or faked, or a person can feel emotions but maintain a "poker face").
317:
and outcomes of the interaction. The nature of this type of data allows for authentic real life implementation, due to the fact it describes states naturally occurring during the
653:
The fact that posed expressions, as used by most subjects of the various studies, are not natural, and therefore algorithms trained on these may not apply to natural expressions.
5240:
2051:
135:
Recognizing emotional information requires the extraction of meaningful patterns from the gathered data. This is done using machine learning techniques that process different
1257:
2787:
1755:
248:, broad enough to fit every need for its application, as well as the selection of a successful classifier which will allow for quick and accurate emotion identification.
1557:
Lee, C.M.; Narayanan, S.; Pieraccini, R., Recognition of
Negative Emotion in the Human Speech Signals, Workshop on Auto. Speech Recognition and Understanding, Dec 2001
662:
FACS did not include dynamics, while dynamics can help disambiguate (e.g. smiles of genuine happiness tend to have different dynamics than "try to look happy" smiles.)
656:
The lack of rotational movement freedom. Affect detection works very well with frontal use, but upon rotating the head more than 20 degrees, "there've been problems".
796:
259:– Classification happens based on the value obtained from the linear combination of the feature values, which are usually provided in the form of vector features.
252:
choosing the appropriate classifier can significantly enhance the overall performance of the system. The list below gives a brief description of each algorithm:
1323:"The Effect of Multimodal Emotional Expression on Responses to a Digital Human during a Self-Disclosure Conversation: a Computational Analysis of User Language"
2472:
2237:
2098:
752:
The corrugator supercilii muscle and zygomaticus major muscle are the 2 main muscles used for measuring the electrical activity, in facial electromyography.
2930:
2240:, Lecture Notes in Computer Science, vol. 3953, Proceedings of the European Conference on Computer Vision, Part III, pp. 288–301, Graz, Austria, May 2006.
1730:
1163:
556:
A system has been conceived by psychologists in order to formally categorize the physical expression of emotions on faces. The central concept of the
1801:
Singh, Premjeet; Saha, Goutam; Sahidullah, Md (2021). "Non-linear frequency warping using constant-Q transformation for speech emotion recognition".
1918:
220:
Emotional speech processing technologies recognize the user's emotional state using computational analysis of speech features. Vocal parameters and
2627:
Khandaker, M (2009). "Designing affective video games to support the social-emotional development of teenagers with autism spectrum disorders".
669:
Accuracy of recognition is improved by adding context; however, adding context and other modalities increases computational cost and complexity
106:. The machine should interpret the emotional state of humans and adapt its behavior to them, giving an appropriate response to those emotions.
1981:
2764:
277:– is a type of (usually binary) linear classifier which decides in which of the two (or more) possible classes, each input may fall into.
3027:
1947:
1507:
Dellaert, F., Polizin, t., and Waibel, A., Recognizing
Emotion in Speech", In Proc. Of ICSLP 1996, Philadelphia, PA, pp.1970–1973, 1996
1382:
283:– is a mathematical model, inspired by biological neural networks, that can better grasp possible non-linearities of the feature space.
6343:
136:
5648:
4047:
176:
that emotion is "not especially different from the processes that we call 'thinking.'" The innovative approach "digital humans" or
1605:
Yacoub, Sherif; Simske, Steve; Lin, Xiaofan; Burns, John (2003). "Recognition of
Emotions in Interactive Voice Response Systems".
184:
Emotion in machines often refers to emotion in computational, often AI-based, systems. As a result, the terms 'emotional AI' and '
2013:
1230:
2828:
1200:
6252:
5743:
3744:
3716:
955:
or "information model" concept of emotion has been criticized by and contrasted with the "post-cognitivist" or "interactional"
890:
that measure the pressure with which a button is pressed: this has been shown to correlate strongly with the players' level of
3769:
2996:
2611:
2379:
2301:"Associating Vehicles Automation With Drivers Functional State Assessment Systems: A Challenge for Road Safety in the Future"
2173:
Bratkova, Margarita; Boulos, Solomon; Shirley, Peter (2009). "oRGB: A Practical
Opponent Color Space for Computer Graphics".
1828:
1686:
1534:
5343:
3620:
309:
representation of archetypal emotions. Those so-called acted databases are usually based on the Basic
Emotions theory (by
6180:
5811:
5084:
4988:
3774:
3046:
2281:
1297:
5806:
3279:
2481:
1179:
The introduction of emotion to computer science was done by
Pickard (sic) who created the field of affective computing.
55:
2099:
Effects of positive and negative affect on electromyographic activity over zygomaticus major and corrugator supercilii
2029:
J. K. Aggarwal, Q. Cai, Human Motion Analysis: A Review, Computer Vision and Image Understanding, Vol. 73, No. 3, 1999
5198:
3926:
3754:
3284:
2412:
1878:
698:
seeing real products that implement the techniques. The four main physiological signs that are usually analyzed are
6232:
5000:
4014:
3108:
1017:
2485:
1281:
Heise, David (2004). "Enculturating agents with expressive role behavior". In Sabine Payr; Trappl, Robert (eds.).
5072:
3402:
1964:
6305:
6227:
5678:
4958:
3693:
3655:
3312:
3020:
2536:
2459:
944:
318:
66:
is the study and development of systems and devices that can recognize, interpret, process, and simulate human
6285:
1772:
6444:
6336:
6257:
6140:
5442:
5288:
4933:
3835:
3812:
3542:
3532:
1027:
347:
Contour slope – describes the tendency of the frequency change over time, it can be rising, falling or level.
1454:
6546:
6267:
5889:
5265:
4040:
3916:
3504:
3412:
3317:
3093:
3078:
895:
262:
1429:"A model of the perception of facial expressions of emotion by humans: Research overview and perspectives"
6295:
5188:
4004:
3739:
3237:
2553:
1927:
557:
551:
417:
224:
features such as pitch variables and speech rate can be analyzed through pattern recognition techniques.
144:
6290:
1852:
Caridakis, G.; Malatesta, L.; Kessous, L.; Amir, N.; Raouzaiou, A.; Karpouzis, K. (November 2–4, 2006).
935:
commercial (or movie) has the desired effect and what the elements which interest the watcher most are.
82:. While some core ideas in the field may be traced as far back as to early philosophical inquiries into
6551:
6090:
5388:
5218:
3976:
3625:
2656:"Stress-Adaptive Training: An Adaptive Psychomotor Training According to Stress Measured by Grip Force"
924:
441:
By doing cross-cultural research in Papua, New Guinea, on the Fore Tribesmen, at the end of the 1960s,
2772:
1903:. Nebraska Symposium on Motivation. Lincoln, Nebraska: University of Nebraska Press. pp. 207–283.
6222:
6214:
5844:
5578:
5454:
5298:
5193:
4995:
3994:
3921:
3896:
3759:
3407:
3013:
404:
280:
185:
2968:
2807:
1615:
1124:
6329:
5464:
5432:
5368:
5336:
5151:
5116:
4157:
3845:
3678:
3264:
3133:
2519:
1428:
952:
481:
However, in the 1990s Ekman expanded his list of basic emotions, including a range of positive and
2954:
Hudlicka, Eva (2003). "To feel or not to feel: The role of affect in human–computer interaction".
1978:
378:
Loudness – measures the amplitude of the speech waveform, translates to the energy of an utterance
6386:
6247:
5854:
5776:
5534:
5499:
5484:
5479:
5469:
5418:
5293:
5208:
5052:
4775:
4267:
4033:
3906:
3840:
3731:
3547:
3207:
1006:
762:
707:
167:
160:
conversational agents in order to enrich and facilitate interactivity between human and machine.
1378:"An analytical framework for studying attitude towards emotional AI: The three-pronged approach"
561:
6439:
6262:
6065:
5904:
5781:
5761:
5708:
5618:
5573:
5544:
5519:
5449:
5437:
5403:
5136:
5079:
4943:
4873:
4725:
4464:
4060:
3971:
3802:
3683:
3450:
3440:
3435:
2963:
2802:
2514:
1869:
Balomenos, T.; Raouzaiou, A.; Ioannou, S.; Drosopoulos, A.; Karpouzis, K.; Kollias, S. (2004).
1610:
786:
509:
436:
395:
The detection and processing of facial expression are achieved through various methods such as
286:
274:
268:
129:
99:
6242:
6237:
5922:
5864:
5816:
5643:
5608:
5583:
5474:
5363:
5223:
4868:
4673:
4102:
3941:
3911:
3901:
3797:
3711:
3587:
3527:
3494:
3484:
3367:
3332:
3322:
3259:
3128:
3103:
3098:
3063:
2043:
981:
898:. Affective games have been used in medical research to support the emotional development of
344:
Average pitch – description of how high/low the speaker speaks relative to the normal speech.
221:
201:
combinations across the categories, e.g. a happy-surprised face or a fearful-surprised face.
2765:"Mona Lisa: Smiling? Computer Scientists Develop Software That Evaluates Facial Expressions"
1944:
1568:
1519:
Proceedings of the Second International Conference on Automatic Face and Gesture Recognition
814:
Also, the facial color signal is independent from that provided by facial muscle movements.
646:, median filtering, or newer methods such as the Bacterial Foraging Optimization Algorithm.
560:, or FACS, as created by Paul Ekman and Wallace V. Friesen in 1978 based on earlier work by
353:
Pitch range – measures the spread between the maximum and minimum frequency of an utterance.
59:
Electronic devices such as robots are increasingly able to recognise and mimic human emotion
6479:
6429:
6398:
5849:
5786:
5688:
5598:
5428:
5413:
5235:
5178:
5168:
5156:
5067:
5062:
5047:
5032:
4948:
4898:
4893:
4838:
4698:
4449:
3701:
3673:
3645:
3640:
3469:
3445:
3397:
3380:
3375:
3357:
3347:
3342:
3304:
3254:
3249:
3166:
3112:
2667:
2568:
2127:
1517:
Roy, D.; Pentland, A. (1996-10-01). "Automatic spoken affect classification and analysis".
911:
724:
8:
6459:
5693:
5683:
5668:
5633:
5628:
5613:
5593:
5588:
5459:
5398:
5329:
5183:
5146:
5131:
5101:
5042:
5027:
4983:
4968:
4853:
4499:
3966:
3891:
3807:
3792:
3557:
3337:
3294:
3289:
3186:
3176:
3148:
1012:
1001:
679:
400:
292:
205:
172:
67:
2671:
2572:
2131:
2020:. Technology Review: The Authority on the Future of Technology. Retrieved 21 March 2011.
1870:
1853:
1404:
1377:
1217:, triggered an explosion of interest in the emotional side of computers and their users.
35:
Please help update this article to reflect recent events or newly available information.
6160:
6040:
5874:
5821:
5733:
5728:
5663:
5603:
5563:
5489:
5141:
5121:
5106:
5096:
5037:
5017:
4978:
4973:
4938:
4923:
4883:
4843:
4715:
4454:
4444:
4092:
3931:
3830:
3706:
3663:
3572:
3514:
3499:
3489:
3274:
3073:
2888:
2820:
2698:
2655:
2598:. Intelligent Technologies for Interactive Entertainment (INTETAIN). pp. 221–227.
2385:
2335:
2300:
2198:
2150:
2115:
2068:
1834:
1806:
1628:
1587:
1540:
1485:
1358:
1032:
959:
approach taken by Kirsten Boehner and others which views emotion as inherently social.
475:
245:
140:
103:
6280:
2977:
2010:
1858:. International Conference on Multimodal Interfaces (ICMI'06). Banff, Alberta, Canada.
1228:
1192:
52:
Area of research in computer science aiming to understand the emotional state of users
6520:
6489:
6403:
6185:
6175:
6100:
6075:
5756:
5713:
5638:
5539:
5524:
5408:
5373:
5126:
4953:
4928:
4905:
4878:
4693:
4683:
4540:
4530:
4397:
4324:
4257:
4227:
3951:
3881:
3860:
3822:
3630:
3597:
3577:
3269:
3181:
3055:
2992:
2703:
2685:
2636:
2607:
2532:
2474:
Affective Videogames and Modes of Affective Gaming: Assist Me, Challenge Me, Emote Me
2455:
2418:
2408:
2389:
2375:
2340:
2322:
2190:
2155:
1838:
1824:
1682:
1675:
1530:
1477:
1409:
1362:
1350:
1342:
1037:
991:
256:
204:
The following sections consider many of the kinds of input data used for the task of
79:
6310:
6300:
2892:
2824:
2202:
2114:
Benitez-Quiroz, Carlos F.; Srinivasan, Ramprakash; Martinez, Aleix M. (2018-03-19).
1820:
1632:
1544:
427:
expression recognition purpose. Two of the widely used databases are CK+ and JAFFE.
350:
Final lowering – the amount by which the frequency falls at the end of an utterance.
83:
6449:
6371:
6195:
6190:
6055:
6050:
5950:
5771:
5718:
5673:
5658:
5653:
5529:
5504:
5283:
5089:
5057:
4730:
4688:
4604:
4599:
3784:
3668:
3635:
3430:
3352:
3241:
3227:
3222:
3171:
3158:
3083:
3036:
2973:
2880:
2812:
2744:
2734:
2693:
2675:
2599:
2576:
2524:
2367:
2364:
Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems
2330:
2312:
2263:
2182:
2145:
2135:
2072:
2060:
1855:
Modeling naturalistic affective states via facial and vocal expressions recognition
1816:
1719:. Connexions – Sharing Knowledge and Building Communities. Retrieved 10 March 2011.
1620:
1591:
1579:
1522:
1469:
1399:
1391:
1334:
1247:
1238:
1229:
Garay, Nestor; Idoia Cearreta; Juan Miguel LĂłpez; Inmaculada Fajardo (April 2006).
1102:
1071:
986:
871:
703:
524:
482:
148:
71:
1979:"Bacterial Foraging Optimization Algorithm – Swarm Algorithms – Clever Algorithms"
1735:
Scientific and Technical Journal of Information Technologies, Mechanics and Optics
1583:
1489:
1213:
Rosalind Picard, a genial MIT professor, is the field's godmother; her 1997 book,
361:
Speech rate – describes the rate of words or syllables uttered over a unit of time
6376:
6165:
6130:
6070:
6020:
5509:
5245:
5213:
5111:
4918:
4910:
4848:
4735:
4354:
3855:
3749:
3721:
3615:
3567:
3552:
3537:
3392:
3387:
3327:
3217:
3191:
3143:
3088:
2986:
2816:
2603:
2580:
2359:
2044:"Visual Interpretation of Hand Gestures for Human–Computer Interaction: A Review"
2017:
1985:
1951:
1882:
1624:
1120:
948:
514:
485:
not all of which are encoded in facial muscles. The newly included emotions are:
87:
2884:
2405:
Heart of the Machine: Our Future in a World of Artificial Emotional Intelligence
1997:
1322:
1106:
364:
Stress frequency – measures the rate of occurrences of pitch accented utterances
236:
The process of speech/text affect detection requires the creation of a reliable
6515:
6510:
6352:
6205:
6125:
6120:
6085:
6060:
6030:
6025:
6010:
6005:
5985:
5980:
5960:
5751:
5270:
5203:
5022:
4703:
4663:
4589:
4237:
3961:
3865:
3764:
3610:
3582:
2911:
2868:
1803:
2021 International Conference on Computer Communication and Informatics (ICCCI)
1338:
1022:
241:
2739:
2722:
1473:
1395:
1252:
6540:
6469:
6110:
6095:
6080:
6035:
6000:
5965:
5945:
5935:
5930:
5909:
5869:
5801:
5703:
5698:
5278:
4963:
4647:
4513:
4365:
4334:
4262:
4167:
3850:
3138:
2689:
2596:
Affective Pacman: A Frustrating Game for Brain–Computer Interface Experiments
2422:
2326:
2317:
2223:, The IEEE International Conference on Computer Vision (ICCV), 2019, pp. 0–0.
1526:
1481:
1346:
643:
504:
384:
Pitch Discontinuity – describes the transitions of the fundamental frequency.
375:
Brilliance – describes the dominance of high or low frequencies In the speech
177:
163:
125:
2371:
2252:"Review of affective computing in education/Learning: Trends and challenges"
2221:
Micro Expression Classification using Facial Color and Deep Learning Methods
2140:
1756:
The repertoire of nonverbal behavior: Categories, origins, usage, and coding
170:, relates emotions to the broader issues of machine intelligence stating in
6505:
6474:
6381:
6366:
6155:
6150:
6145:
6115:
6105:
6045:
6015:
5995:
5894:
5791:
5494:
5378:
5250:
5163:
5007:
4459:
4387:
4274:
4137:
3946:
3605:
2929:
Hook, Kristina; Staahl, Anna; Sundstrom, Petra; Laaksolahti, Jarmo (2008).
2907:
2864:
2707:
2640:
2554:"Turning shortcomings into challenges: Brain–computer interfaces for games"
2447:
2435:
2344:
2216:
2194:
2159:
1413:
1354:
863:
396:
341:
Accent shape – affected by the rate of change of the fundamental frequency.
2654:
Sahar, Yotam; Wagner, Michael; Barel, Ariel; Shoval, Shraga (2022-11-01).
2528:
124:
capture speech. Other sensors detect emotional cues by directly measuring
98:. One of the motivations for the research is the ability to give machines
6434:
6200:
5990:
5975:
5940:
5796:
5723:
5423:
4817:
4790:
4629:
4567:
4439:
4407:
4329:
4309:
4202:
3936:
3562:
3474:
2358:
Balters, Stephanie; Bernstein, Madeline; Paredes, Pablo E. (2019-05-02).
2238:
Studying Aesthetics in Photographic Images Using a Computational Approach
883:
882:
Affective video games can access their players' emotional states through
723:
A subject's blood volume pulse (BVP) can be measured by a process called
529:
499:
381:
Pause Discontinuity – describes the transitions between sound and silence
2723:"Tune in to Your Emotions: A Robust Personalized Affective Music Player"
2251:
2186:
1716:
1455:"Recognition of Affective Communicative Intent in Robot-Directed Speech"
886:
devices. A particularly simple form of biofeedback is available through
6408:
6393:
6135:
5970:
5766:
5623:
5514:
5352:
5255:
4614:
4484:
4294:
4187:
4182:
4082:
4077:
3956:
3886:
3479:
3212:
3068:
2749:
2267:
1914:
1896:
996:
956:
442:
310:
75:
2680:
2282:"In-Car Facial Recognition Detects Angry Drivers To Prevent Road Rage"
2064:
6484:
6454:
6413:
6170:
5955:
5859:
4805:
4780:
4767:
4710:
4656:
4550:
4525:
4494:
4402:
4359:
4339:
4289:
4284:
4222:
4217:
4192:
4132:
4112:
4097:
4087:
3461:
3422:
2360:"On-road Stress Analysis for In-car Interventions During the Commute"
2116:"Facial color is an efficient mechanism to visually transmit emotion"
1672:
1075:
489:
465:
95:
2916:
Proceedings of the Aarhus Decennial Conference on Critical Computing
2513:. CHI '03 Extended Abstracts on Human Factors in Computing Systems.
1868:
1090:
6464:
5884:
5879:
5839:
4800:
4795:
4785:
4720:
4557:
4479:
4469:
4434:
4421:
4304:
4247:
4197:
4177:
3522:
3005:
2858:
2856:
2854:
2852:
2850:
2848:
1901:
Universals and Cultural Differences in Facial Expression of Emotion
1811:
915:
quality of training and shortening the required training duration.
534:
494:
237:
6321:
4025:
1062:
Tao, Jianhua; Tieniu Tan (2005). "Affective Computing: A Review".
114:
5230:
4810:
4758:
4752:
4678:
4638:
4624:
4619:
4609:
4520:
4380:
4279:
4242:
4212:
4207:
4172:
4152:
4142:
4127:
4069:
4056:
891:
887:
470:
455:
120:
2845:
2552:
Nijholt, Anton; Plass-Oude Bos, Danny; Reuderink, Boris (2009).
2250:
Wu, Chih-Hung; Huang, Yueh-Min; Hwang, Jan-Pan (November 2016).
2113:
2042:
Pavlovic, Vladimir I.; Sharma, Rajeev; Huang, Thomas S. (1997).
6525:
5554:
5260:
4594:
4582:
4577:
4572:
4426:
4373:
4232:
4147:
2233:
1283:
Agent Culture: Human-Agent Interaction in a Mutlicultural World
899:
5241:
Interactions between the emotional and executive brain systems
2928:
2551:
2452:
Computing Attitude and Affect in Text: Theory and Applications
2052:
IEEE Transactions on Pattern Analysis and Machine Intelligence
1321:
Loveys, Kate; Sagar, Mark; Broadbent, Elizabeth (2020-07-22).
5321:
4742:
4668:
4562:
4535:
4414:
4349:
4344:
4314:
4252:
4122:
4117:
2985:
Scherer, Klaus R; Bänziger, Tanja; Roesch, Etienne B (2010).
1851:
699:
539:
519:
450:
86:, the more modern branch of computer science originated with
2988:
A Blueprint for Affective Computing: A Sourcebook and Manual
2905:
2862:
1917:(1999). "Basic Emotions". In Dalgleish, T; Power, M (eds.).
748:
119:
Detecting emotional information usually begins with passive
5899:
4545:
4506:
4489:
4392:
4319:
4299:
4107:
2454:. Dordrecht: Springer Science & Business Media. p. 94.
1067:
938:
460:
771:
4474:
4162:
2084:
2082:
1945:"Facial Action Coding System (FACS) and the FACS Manual"
2721:
Janssen, Joris H.; van den Broek, Egon L. (July 2012).
2594:
Reuderink, Boris; Nijholt, Anton; Poel, Mannes (2009).
2511:
Affective gaming: Measuring emotion through the gamepad
2357:
1566:
2914:(2005). "Affection: From Information to Interaction".
2653:
2471:
Gilleade, Kiel Mark; Dix, Alan; Allanson, Jen (2005).
2011:"Better Face-Recognition Software – Technology Review"
1569:"Emotion recognition in spontaneous speech using GMMs"
1426:
2984:
2172:
2079:
1871:"Emotion Analysis in Man-Machine Interaction Systems"
1788:
1704:
1135:
1133:
408:
robust estimation of the subject's emotional state.
372:
Breathiness – measures the aspiration noise in speech
2720:
2593:
1784:
1782:
1731:"Extended speech emotion recognition and prediction"
1700:
1698:
1644:
1642:
1604:
1320:
2470:
2041:
1926:. Sussex, UK: John Wiley & Sons. Archived from
1800:
795:Skin conductance is often measured using two small
2088:Picard, Rosalind (1998). Affective Computing. MIT.
1674:
1673:Charles Osgood; William May; Murray Miron (1975).
1130:
923:Affective computing has potential applications in
90:'s 1995 paper on affective computing and her book
2785:
2407:. New York: Arcade Publishing. pp. 150–153.
1779:
1750:
1748:
1728:
1695:
1654:
1639:
1452:
636:
369:Voice quality parameters and energy descriptors:
6538:
4636:
2299:Collet, Christian; Musicant, Oren (2019-04-24).
1285:. Lawrence Erlbaum Associates. pp. 127–142.
1193:"The Love Machine; Building computers that care"
2956:International Journal of Human–Computer Studies
2873:International Journal of Human–Computer Studies
2788:"Co-experience: user experience as interaction"
2120:Proceedings of the National Academy of Sciences
1064:Affective Computing and Intelligent Interaction
166:, one of the pioneering computer scientists in
115:Detecting and recognizing emotional information
4511:
2298:
2091:
1766:
1764:
1745:
1677:Cross-Cultural Universals of Affective Meaning
1231:"Assistive Technology and Affective Mediation"
545:
411:
6337:
5337:
4041:
3045:Note: This template roughly follows the 2012
3021:
1988:. Clever Algorithms. Retrieved 21 March 2011.
1567:Neiberg, D; Elenius, K; Laskowski, K (2006).
1154:
1061:
4815:
4654:
4645:
4412:
2871:(2007). "How emotion is made and measured".
2629:Studies in Health Technology and Informatics
2249:
1862:
1845:
1453:Breazeal, Cynthia; Aryananda, Lijin (2002).
70:. It is an interdisciplinary field spanning
5314:indicate emotion names in foreign languages
4378:
2037:
2035:
1875:Machine Learning for Multimodal Interaction
1873:. In Bengio, Samy; Bourlard, Herve (eds.).
1761:
1516:
1503:
1501:
1499:
692:
6344:
6330:
5344:
5330:
4048:
4034:
3028:
3014:
2727:User Modeling and User-Adapted Interaction
2508:
2000:. Soft Computing. Retrieved 18 March 2011.
756:
390:
2967:
2806:
2786:Battarbee, Katja; Koskinen, Ilpo (2005).
2748:
2738:
2697:
2679:
2626:
2518:
2334:
2316:
2256:British Journal of Educational Technology
2149:
2139:
1810:
1614:
1427:Aleix, and Shichuan Du, Martinez (2012).
1403:
1251:
1155:Kleine-Cosack, Christian (October 2006).
835:
780:
430:
2953:
2032:
1954:. A Human Face. Retrieved 21 March 2011.
1666:
1660:
1648:
1496:
1436:The Journal of Machine Learning Research
1157:"Recognition and Simulation of Emotions"
939:Cognitivist vs. interactional approaches
770:
747:
54:
2175:IEEE Computer Graphics and Applications
2109:
2107:
1729:S.E. Khoruzhnikov; et al. (2014).
1190:
905:
6539:
3745:Knowledge representation and reasoning
2509:Sykes, Jonathan; Brown, Simon (2003).
1770:
1754:Ekman, P. & Friesen, W. V (1969).
1295:
1144:. Cambridge, MA: MIT Press. p. 1.
1139:
154:
6325:
5325:
4029:
3770:Philosophy of artificial intelligence
3009:
2402:
2226:
1913:
1895:
1280:
1088:
1055:
918:
713:
328:
3089:Energy consumption (Green computing)
3035:
2906:Boehner, Kirsten; DePaula, Rogerio;
2863:Boehner, Kirsten; DePaula, Rogerio;
2166:
2104:
2097:Larsen JT, Norris CJ, Cacioppo JT, "
894:; at the other end of the scale are
826:
15:
6351:
4055:
3775:Distributed artificial intelligence
3047:ACM Computing Classification System
2991:. Oxford: Oxford University Press.
2209:
1789:Scherer, Bänziger & Roesch 2010
1705:Scherer, Bänziger & Roesch 2010
1551:
931:received by the respective market.
211:
128:data, such as skin temperature and
13:
3280:Integrated development environment
2023:
1375:
1070:3784. Springer. pp. 981–995.
14:
6563:
3755:Automated planning and scheduling
3285:Software configuration management
1920:Handbook of Cognition and Emotion
1879:Lecture Notes in Computer Science
849:
6470:Augmented reality tabletop games
5553:
4757:
4751:
4009:
3999:
3990:
3989:
1263:from the original on 28 May 2008
1203:from the original on 18 May 2008
1191:Diamond, David (December 2003).
1018:Friendly artificial intelligence
817:
739:
20:
4000:
3403:Computational complexity theory
2922:
2899:
2779:
2771:. 1 August 2006. Archived from
2757:
2714:
2647:
2620:
2587:
2545:
2502:
2464:
2440:
2436:Projects in Affective Computing
2429:
2396:
2351:
2305:Frontiers in Human Neuroscience
2292:
2274:
2243:
2003:
1991:
1971:
1957:
1938:
1907:
1889:
1821:10.1109/ICCCI50826.2021.9402569
1794:
1771:Steidl, Stefan (5 March 2011).
1722:
1710:
1598:
1560:
1510:
1446:
1420:
1376:Ho, Manh-Tung (29 March 2023).
1369:
1314:
803:
673:
191:
6445:Live action role-playing games
5351:
3187:Network performance evaluation
2947:
2232:Ritendra Datta, Dhiraj Joshi,
1296:Restak, Richard (2006-12-17).
1289:
1274:
1222:
1184:
1148:
1114:
1082:
877:
730:
637:Challenges in facial detection
1:
5649:Industrial and organizational
3558:Multimedia information system
3543:Geographic information system
3533:Enterprise information system
3122:Computer systems organization
2978:10.1016/s1071-5819(03)00047-8
1584:10.21437/Interspeech.2006-277
1044:
1028:Multimodal sentiment analysis
858:
808:
710:, and facial color patterns.
231:
5890:Human factors and ergonomics
5266:Social emotional development
3917:Computational social science
3505:Theoretical computer science
3318:Software development process
3094:Electronic design automation
3079:Very Large Scale Integration
2817:10.1080/15710880412331289917
2604:10.1007/978-3-642-02315-6_23
2581:10.1016/j.entcom.2009.09.007
1625:10.21437/Eurospeech.2003-307
1049:
840:
303:
7:
4512:
3740:Natural language processing
3528:Information storage systems
2931:"Interactional empowerment"
2885:10.1016/j.ijhcs.2006.11.016
1681:. Univ. of Illinois Press.
1123:MIT Technical Report #321 (
974:
718:
558:Facial Action Coding System
552:Facial Action Coding System
546:Facial Action Coding System
418:Facial expression databases
412:Facial expression databases
149:facial expression detection
145:natural language processing
10:
6568:
3656:Human–computer interaction
3626:Intrusion detection system
3538:Social information systems
3523:Database management system
2446:Shanahan, James; Qu, Yan;
1775:. Pattern Recognition Lab.
1576:Proceedings of Interspeech
1339:10.1007/s10916-020-01624-4
1327:Journal of Medical Systems
945:human–computer interaction
925:human–computer interaction
784:
760:
677:
549:
434:
415:
338:Frequency characteristics
319:human–computer interaction
6498:
6422:
6359:
6276:
6213:
5920:
5830:
5742:
5579:Applied behavior analysis
5562:
5551:
5387:
5359:
5309:
4828:
4766:
4749:
4068:
3985:
3922:Computational engineering
3897:Computational mathematics
3874:
3821:
3783:
3730:
3692:
3654:
3596:
3513:
3459:
3421:
3366:
3303:
3236:
3200:
3157:
3121:
3054:
3043:
2740:10.1007/s11257-011-9107-7
1950:October 19, 2013, at the
1773:"FAU Aibo Emotion Corpus"
1607:Proceedings of Eurospeech
1396:10.1016/j.mex.2023.102149
1140:Picard, Rosalind (1997).
1107:10.1093/mind/os-IX.34.188
896:brain–computer interfaces
666:channels of information).
649:Other challenges include
644:linear Gaussian smoothing
29:This article needs to be
5199:in virtual communication
3932:Computational healthcare
3927:Differentiable computing
3846:Graphics processing unit
3265:Domain-specific language
3134:Computational complexity
2318:10.3389/fnhum.2019.00131
1965:"Spatial domain methods"
1899:(1972). Cole, J. (ed.).
1717:"Gaussian Mixture Model"
1527:10.1109/AFGR.1996.557292
693:Physiological monitoring
287:Decision tree algorithms
109:
6440:Alternate reality games
5855:Behavioral neuroscience
5419:Behavioral neuroscience
3907:Computational chemistry
3841:Photograph manipulation
3732:Artificial intelligence
3548:Decision support system
2561:Entertainment Computing
2403:Yonck, Richard (2017).
2372:10.1145/3290607.3312824
2141:10.1073/pnas.1716084115
1474:10.1023/a:1013215010749
1468:(1). Springer: 83–104.
1253:10.17011/ht/urn.2006159
1089:James, William (1884).
1007:Emotion Markup Language
763:Facial electromyography
757:Facial electromyography
708:facial electromyography
391:Facial affect detection
358:Time-related features:
168:artificial intelligence
5905:Psychology of religion
5845:Behavioral engineering
5782:Human subject research
5438:Cognitive neuroscience
5404:Affective neuroscience
4816:
4655:
4646:
4637:
4413:
4379:
3972:Educational technology
3803:Reinforcement learning
3553:Process control system
3451:Computational geometry
3441:Algorithmic efficiency
3436:Analysis of algorithms
3084:Systems on Chip (SoCs)
1758:. Semiotica, 1, 49–98.
836:Potential applications
797:silver-silver chloride
787:Galvanic skin response
781:Galvanic skin response
777:
753:
704:galvanic skin response
437:Emotion classification
431:Emotion classification
100:emotional intelligence
60:
6281:Wiktionary definition
5817:Self-report inventory
5812:Quantitative research
3942:Electronic publishing
3912:Computational biology
3902:Computational physics
3798:Unsupervised learning
3712:Distributed computing
3588:Information retrieval
3495:Mathematical analysis
3485:Mathematical software
3368:Theory of computation
3333:Software construction
3323:Requirements analysis
3201:Software organization
3129:Computer architecture
3099:Hardware acceleration
3064:Printed circuit board
2529:10.1145/765891.765957
2366:. ACM. pp. 1–6.
1121:"Affective Computing"
1111:Cited by Tao and Tan.
982:Affect control theory
774:
751:
58:
6460:Location-based games
6399:Ubiquitous computing
5807:Qualitative research
5762:Behavior epigenetics
5236:Group affective tone
3702:Concurrent computing
3674:Ubiquitous computing
3646:Application security
3641:Information security
3470:Discrete mathematics
3446:Randomized algorithm
3398:Computability theory
3376:Model of computation
3348:Software maintenance
3343:Software engineering
3305:Software development
3255:Programming language
3250:Programming paradigm
3167:Network architecture
2484:Conf. Archived from
1521:. pp. 363–367.
943:Within the field of
910:Training methods of
906:Psychomotor training
725:photoplethysmography
562:Carl-Herman Hjortsjö
520:Pride in achievement
401:hidden Markov models
6547:Affective computing
6286:Wiktionary category
5850:Behavioral genetics
5822:Statistical surveys
5679:Occupational health
5414:Behavioral genetics
5289:constructed emotion
4959:functional accounts
3977:Document management
3967:Operations research
3892:Enterprise software
3808:Multi-task learning
3793:Supervised learning
3515:Information systems
3338:Software deployment
3295:Software repository
3149:Real-time computing
2775:on 19 October 2007.
2672:2022Senso..22.8368S
2573:2009itie.conf..153N
2236:and James Z. Wang,
2187:10.1109/mcg.2009.13
2132:2018PNAS..115.3581B
2101:", (September 2003)
1977:Clever Algorithms.
1885:. pp. 318–328.
1302:The Washington Post
1215:Affective Computing
1142:Affective Computing
1013:Emotion recognition
1002:Character Computing
680:Gesture recognition
206:emotion recognition
173:The Emotion Machine
155:Emotion in machines
130:galvanic resistance
92:Affective Computing
64:Affective computing
6258:Schools of thought
6161:Richard E. Nisbett
6041:Donald T. Campbell
5719:Sport and exercise
5189:in decision-making
4430:(sense of purpose)
3760:Search methodology
3707:Parallel computing
3664:Interaction design
3573:Computing platform
3500:Numerical analysis
3490:Information theory
3275:Software framework
3238:Software notations
3177:Network components
3074:Integrated circuit
2268:10.1111/bjet.12324
2016:2011-06-08 at the
1984:2019-06-12 at the
1881:. Vol. 3361.
1298:"Mind Over Matter"
1033:Sentiment analysis
919:Other applications
778:
754:
714:Blood volume pulse
700:blood volume pulse
329:Speech descriptors
246:vector space model
188:' are being used.
141:speech recognition
61:
6552:Affective science
6534:
6533:
6490:Transreality game
6404:Context awareness
6319:
6318:
6296:Wikimedia Commons
6223:Counseling topics
6186:Ronald C. Kessler
6176:Shelley E. Taylor
6101:Lawrence Kohlberg
6076:Stanley Schachter
5875:Consumer behavior
5757:Archival research
5525:Psycholinguistics
5409:Affective science
5319:
5318:
4906:Appeal to emotion
4684:Social connection
4023:
4022:
3952:Electronic voting
3882:Quantum Computing
3875:Applied computing
3861:Image compression
3631:Hardware security
3621:Security services
3578:Digital marketing
3358:Open-source model
3270:Modeling language
3182:Network scheduler
2998:978-0-19-956670-9
2681:10.3390/s22218368
2613:978-3-642-02314-9
2381:978-1-4503-5971-9
2288:. 30 August 2018.
2126:(14): 3581–3586.
2065:10.1109/34.598226
1830:978-1-7281-5875-4
1688:978-94-007-5069-2
1536:978-0-8186-7713-7
1462:Autonomous Robots
1091:"What Is Emotion"
1038:Wearable computer
992:Affective haptics
827:Visual aesthetics
634:
633:
483:negative emotions
80:cognitive science
50:
49:
6559:
6450:Affective gaming
6430:Ubiquitous games
6346:
6339:
6332:
6323:
6322:
6253:Research methods
6196:Richard Davidson
6191:Joseph E. LeDoux
6066:George A. Miller
6056:David McClelland
6051:Herbert A. Simon
5951:Edward Thorndike
5772:Content analysis
5557:
5530:Psychophysiology
5346:
5339:
5332:
5323:
5322:
5294:discrete emotion
5194:in the workplace
5090:Empathy quotient
4821:
4761:
4755:
4660:
4651:
4642:
4517:
4418:
4384:
4050:
4043:
4036:
4027:
4026:
4013:
4012:
4003:
4002:
3993:
3992:
3813:Cross-validation
3785:Machine learning
3669:Social computing
3636:Network security
3431:Algorithm design
3353:Programming team
3313:Control variable
3290:Software library
3228:Software quality
3223:Operating system
3172:Network protocol
3037:Computer science
3030:
3023:
3016:
3007:
3006:
3002:
2981:
2971:
2942:
2941:
2935:
2926:
2920:
2919:
2903:
2897:
2896:
2860:
2843:
2842:
2840:
2839:
2833:
2827:. Archived from
2810:
2792:
2783:
2777:
2776:
2761:
2755:
2754:
2752:
2742:
2718:
2712:
2711:
2701:
2683:
2651:
2645:
2644:
2624:
2618:
2617:
2591:
2585:
2584:
2558:
2549:
2543:
2542:
2522:
2506:
2500:
2499:
2497:
2496:
2490:
2479:
2468:
2462:
2444:
2438:
2433:
2427:
2426:
2400:
2394:
2393:
2355:
2349:
2348:
2338:
2320:
2296:
2290:
2289:
2278:
2272:
2271:
2262:(6): 1304–1323.
2247:
2241:
2230:
2224:
2213:
2207:
2206:
2170:
2164:
2163:
2153:
2143:
2111:
2102:
2095:
2089:
2086:
2077:
2076:
2048:
2039:
2030:
2027:
2021:
2009:Williams, Mark.
2007:
2001:
1998:"Soft Computing"
1995:
1989:
1975:
1969:
1968:
1961:
1955:
1942:
1936:
1934:
1932:
1925:
1911:
1905:
1904:
1893:
1887:
1886:
1866:
1860:
1859:
1849:
1843:
1842:
1814:
1805:. pp. 1–4.
1798:
1792:
1786:
1777:
1776:
1768:
1759:
1752:
1743:
1742:
1726:
1720:
1714:
1708:
1702:
1693:
1692:
1680:
1670:
1664:
1658:
1652:
1646:
1637:
1636:
1618:
1602:
1596:
1595:
1573:
1564:
1558:
1555:
1549:
1548:
1514:
1508:
1505:
1494:
1493:
1459:
1450:
1444:
1443:
1433:
1424:
1418:
1417:
1407:
1373:
1367:
1366:
1318:
1312:
1311:
1309:
1308:
1293:
1287:
1286:
1278:
1272:
1271:
1269:
1268:
1262:
1255:
1239:Human Technology
1235:
1226:
1220:
1219:
1210:
1208:
1188:
1182:
1181:
1176:
1174:
1168:
1162:. Archived from
1161:
1152:
1146:
1145:
1137:
1128:
1118:
1112:
1110:
1086:
1080:
1079:
1076:10.1007/11573548
1059:
987:Affective design
872:emotive Internet
568:
567:
535:Sensory pleasure
212:Emotional speech
104:simulate empathy
72:computer science
45:
42:
36:
24:
23:
16:
6567:
6566:
6562:
6561:
6560:
6558:
6557:
6556:
6537:
6536:
6535:
6530:
6494:
6418:
6377:Performing arts
6355:
6353:Pervasive games
6350:
6320:
6315:
6272:
6248:Psychotherapies
6209:
6166:Martin Seligman
6131:Daniel Kahneman
6071:Richard Lazarus
6021:Raymond Cattell
5925:
5916:
5915:
5914:
5826:
5738:
5565:
5558:
5549:
5510:Neuropsychology
5390:
5383:
5355:
5350:
5320:
5315:
5305:
5246:Jealousy in art
4989:in conversation
4911:Amygdala hijack
4824:
4762:
4756:
4747:
4736:sense of wonder
4064:
4054:
4024:
4019:
4010:
3981:
3962:Word processing
3870:
3856:Virtual reality
3817:
3779:
3750:Computer vision
3726:
3722:Multiprocessing
3688:
3650:
3616:Security hacker
3592:
3568:Digital library
3509:
3460:Mathematics of
3455:
3417:
3393:Automata theory
3388:Formal language
3362:
3328:Software design
3299:
3232:
3218:Virtual machine
3196:
3192:Network service
3153:
3144:Embedded system
3117:
3050:
3039:
3034:
2999:
2969:10.1.1.180.6429
2950:
2945:
2933:
2927:
2923:
2912:Sengers, Phoebe
2904:
2900:
2869:Sengers, Phoebe
2861:
2846:
2837:
2835:
2831:
2808:10.1.1.294.9178
2790:
2784:
2780:
2763:
2762:
2758:
2719:
2715:
2652:
2648:
2625:
2621:
2614:
2592:
2588:
2556:
2550:
2546:
2539:
2507:
2503:
2494:
2492:
2488:
2477:
2469:
2465:
2445:
2441:
2434:
2430:
2415:
2401:
2397:
2382:
2356:
2352:
2297:
2293:
2280:
2279:
2275:
2248:
2244:
2231:
2227:
2214:
2210:
2171:
2167:
2112:
2105:
2096:
2092:
2087:
2080:
2046:
2040:
2033:
2028:
2024:
2018:Wayback Machine
2008:
2004:
1996:
1992:
1986:Wayback Machine
1976:
1972:
1963:
1962:
1958:
1952:Wayback Machine
1943:
1939:
1930:
1923:
1912:
1908:
1894:
1890:
1883:Springer-Verlag
1867:
1863:
1850:
1846:
1831:
1799:
1795:
1787:
1780:
1769:
1762:
1753:
1746:
1727:
1723:
1715:
1711:
1703:
1696:
1689:
1671:
1667:
1659:
1655:
1647:
1640:
1616:10.1.1.420.8158
1603:
1599:
1571:
1565:
1561:
1556:
1552:
1537:
1515:
1511:
1506:
1497:
1457:
1451:
1447:
1442:(1): 1589–1608.
1431:
1425:
1421:
1374:
1370:
1319:
1315:
1306:
1304:
1294:
1290:
1279:
1275:
1266:
1264:
1260:
1233:
1227:
1223:
1206:
1204:
1189:
1185:
1172:
1170:
1169:on May 28, 2008
1166:
1159:
1153:
1149:
1138:
1131:
1119:
1115:
1101:(34): 188–205.
1087:
1083:
1060:
1056:
1052:
1047:
1042:
977:
971:interactions".
949:Rosalind Picard
941:
921:
908:
880:
861:
852:
843:
838:
829:
820:
811:
806:
789:
783:
765:
759:
742:
733:
721:
716:
695:
682:
676:
639:
554:
548:
439:
433:
420:
414:
393:
331:
306:
234:
214:
194:
157:
117:
112:
102:, including to
88:Rosalind Picard
53:
46:
40:
37:
34:
25:
21:
12:
11:
5:
6565:
6555:
6554:
6549:
6532:
6531:
6529:
6528:
6523:
6518:
6516:Jane McGonigal
6513:
6511:Eric Zimmerman
6508:
6502:
6500:
6496:
6495:
6493:
6492:
6487:
6482:
6480:Treasure hunts
6477:
6472:
6467:
6462:
6457:
6452:
6447:
6442:
6437:
6432:
6426:
6424:
6420:
6419:
6417:
6416:
6411:
6406:
6401:
6396:
6391:
6390:
6389:
6379:
6374:
6369:
6363:
6361:
6357:
6356:
6349:
6348:
6341:
6334:
6326:
6317:
6316:
6314:
6313:
6308:
6303:
6298:
6293:
6288:
6283:
6277:
6274:
6273:
6271:
6270:
6265:
6260:
6255:
6250:
6245:
6240:
6235:
6230:
6225:
6219:
6217:
6211:
6210:
6208:
6206:Roy Baumeister
6203:
6198:
6193:
6188:
6183:
6178:
6173:
6168:
6163:
6158:
6153:
6148:
6143:
6141:Michael Posner
6138:
6133:
6128:
6126:Elliot Aronson
6123:
6121:Walter Mischel
6118:
6113:
6108:
6103:
6098:
6093:
6088:
6086:Albert Bandura
6083:
6078:
6073:
6068:
6063:
6061:Leon Festinger
6058:
6053:
6048:
6043:
6038:
6033:
6031:Neal E. Miller
6028:
6026:Abraham Maslow
6023:
6018:
6013:
6011:Ernest Hilgard
6008:
6006:Donald O. Hebb
6003:
5998:
5993:
5988:
5986:J. P. Guilford
5983:
5981:Gordon Allport
5978:
5973:
5968:
5963:
5961:John B. Watson
5958:
5953:
5948:
5943:
5938:
5933:
5928:
5926:
5921:
5918:
5917:
5913:
5912:
5907:
5902:
5897:
5892:
5887:
5882:
5877:
5872:
5867:
5862:
5857:
5852:
5847:
5842:
5836:
5835:
5834:
5832:
5828:
5827:
5825:
5824:
5819:
5814:
5809:
5804:
5799:
5794:
5789:
5784:
5779:
5774:
5769:
5764:
5759:
5754:
5752:Animal testing
5748:
5746:
5740:
5739:
5737:
5736:
5731:
5726:
5721:
5716:
5711:
5706:
5701:
5696:
5691:
5686:
5681:
5676:
5671:
5666:
5661:
5656:
5651:
5646:
5641:
5636:
5631:
5626:
5621:
5616:
5611:
5606:
5601:
5596:
5591:
5586:
5581:
5576:
5570:
5568:
5560:
5559:
5552:
5550:
5548:
5547:
5542:
5537:
5532:
5527:
5522:
5517:
5512:
5507:
5502:
5497:
5492:
5487:
5482:
5477:
5472:
5467:
5462:
5457:
5455:Cross-cultural
5452:
5447:
5446:
5445:
5435:
5426:
5421:
5416:
5411:
5406:
5401:
5395:
5393:
5385:
5384:
5382:
5381:
5376:
5371:
5366:
5360:
5357:
5356:
5349:
5348:
5341:
5334:
5326:
5317:
5316:
5310:
5307:
5306:
5304:
5303:
5302:
5301:
5299:somatic marker
5296:
5291:
5286:
5281:
5273:
5271:Stoic passions
5268:
5263:
5258:
5253:
5248:
5243:
5238:
5233:
5228:
5227:
5226:
5221:
5219:social sharing
5216:
5211:
5209:self-conscious
5206:
5201:
5196:
5191:
5186:
5181:
5173:
5172:
5171:
5161:
5160:
5159:
5154:
5152:thought method
5149:
5144:
5139:
5134:
5129:
5124:
5119:
5117:lateralization
5114:
5109:
5104:
5099:
5094:
5093:
5092:
5087:
5077:
5076:
5075:
5065:
5060:
5055:
5050:
5045:
5040:
5035:
5030:
5025:
5020:
5012:
5011:
5010:
5005:
5004:
5003:
4993:
4992:
4991:
4981:
4976:
4971:
4966:
4961:
4956:
4951:
4946:
4944:classification
4941:
4936:
4931:
4926:
4921:
4913:
4908:
4903:
4902:
4901:
4896:
4888:
4887:
4886:
4881:
4876:
4871:
4866:
4858:
4857:
4856:
4851:
4846:
4841:
4832:
4830:
4826:
4825:
4823:
4822:
4813:
4808:
4803:
4798:
4793:
4788:
4783:
4778:
4772:
4770:
4764:
4763:
4750:
4748:
4746:
4745:
4740:
4739:
4738:
4728:
4723:
4718:
4713:
4708:
4707:
4706:
4696:
4691:
4686:
4681:
4676:
4671:
4666:
4664:Sentimentality
4661:
4652:
4643:
4634:
4633:
4632:
4622:
4617:
4612:
4607:
4602:
4597:
4592:
4587:
4586:
4585:
4580:
4575:
4570:
4560:
4555:
4554:
4553:
4543:
4538:
4533:
4528:
4523:
4518:
4509:
4504:
4503:
4502:
4500:at first sight
4497:
4487:
4482:
4477:
4472:
4467:
4462:
4457:
4452:
4447:
4442:
4437:
4432:
4424:
4419:
4410:
4405:
4400:
4395:
4390:
4385:
4376:
4371:
4370:
4369:
4357:
4352:
4347:
4342:
4337:
4332:
4327:
4322:
4317:
4312:
4307:
4302:
4297:
4292:
4287:
4282:
4277:
4272:
4271:
4270:
4260:
4255:
4250:
4245:
4240:
4238:Disappointment
4235:
4230:
4225:
4220:
4215:
4210:
4205:
4200:
4195:
4190:
4185:
4180:
4175:
4170:
4165:
4160:
4155:
4150:
4145:
4140:
4135:
4130:
4125:
4120:
4115:
4110:
4105:
4100:
4095:
4090:
4085:
4080:
4074:
4072:
4066:
4065:
4053:
4052:
4045:
4038:
4030:
4021:
4020:
4018:
4017:
4007:
3997:
3986:
3983:
3982:
3980:
3979:
3974:
3969:
3964:
3959:
3954:
3949:
3944:
3939:
3934:
3929:
3924:
3919:
3914:
3909:
3904:
3899:
3894:
3889:
3884:
3878:
3876:
3872:
3871:
3869:
3868:
3866:Solid modeling
3863:
3858:
3853:
3848:
3843:
3838:
3833:
3827:
3825:
3819:
3818:
3816:
3815:
3810:
3805:
3800:
3795:
3789:
3787:
3781:
3780:
3778:
3777:
3772:
3767:
3765:Control method
3762:
3757:
3752:
3747:
3742:
3736:
3734:
3728:
3727:
3725:
3724:
3719:
3717:Multithreading
3714:
3709:
3704:
3698:
3696:
3690:
3689:
3687:
3686:
3681:
3676:
3671:
3666:
3660:
3658:
3652:
3651:
3649:
3648:
3643:
3638:
3633:
3628:
3623:
3618:
3613:
3611:Formal methods
3608:
3602:
3600:
3594:
3593:
3591:
3590:
3585:
3583:World Wide Web
3580:
3575:
3570:
3565:
3560:
3555:
3550:
3545:
3540:
3535:
3530:
3525:
3519:
3517:
3511:
3510:
3508:
3507:
3502:
3497:
3492:
3487:
3482:
3477:
3472:
3466:
3464:
3457:
3456:
3454:
3453:
3448:
3443:
3438:
3433:
3427:
3425:
3419:
3418:
3416:
3415:
3410:
3405:
3400:
3395:
3390:
3385:
3384:
3383:
3372:
3370:
3364:
3363:
3361:
3360:
3355:
3350:
3345:
3340:
3335:
3330:
3325:
3320:
3315:
3309:
3307:
3301:
3300:
3298:
3297:
3292:
3287:
3282:
3277:
3272:
3267:
3262:
3257:
3252:
3246:
3244:
3234:
3233:
3231:
3230:
3225:
3220:
3215:
3210:
3204:
3202:
3198:
3197:
3195:
3194:
3189:
3184:
3179:
3174:
3169:
3163:
3161:
3155:
3154:
3152:
3151:
3146:
3141:
3136:
3131:
3125:
3123:
3119:
3118:
3116:
3115:
3106:
3101:
3096:
3091:
3086:
3081:
3076:
3071:
3066:
3060:
3058:
3052:
3051:
3044:
3041:
3040:
3033:
3032:
3025:
3018:
3010:
3004:
3003:
2997:
2982:
2949:
2946:
2944:
2943:
2921:
2898:
2879:(4): 275–291.
2844:
2778:
2756:
2733:(3): 255–279.
2713:
2646:
2619:
2612:
2586:
2544:
2537:
2520:10.1.1.92.2123
2501:
2463:
2439:
2428:
2413:
2395:
2380:
2350:
2291:
2273:
2242:
2225:
2215:Hadas Shahar,
2208:
2165:
2103:
2090:
2078:
2059:(7): 677–695.
2031:
2022:
2002:
1990:
1970:
1956:
1937:
1933:on 2010-12-28.
1906:
1888:
1861:
1844:
1829:
1793:
1778:
1760:
1744:
1721:
1709:
1694:
1687:
1665:
1653:
1638:
1597:
1559:
1550:
1535:
1509:
1495:
1445:
1419:
1368:
1313:
1288:
1273:
1221:
1183:
1147:
1129:
1113:
1081:
1053:
1051:
1048:
1046:
1043:
1041:
1040:
1035:
1030:
1025:
1023:Kismet (robot)
1020:
1015:
1010:
1004:
999:
994:
989:
984:
978:
976:
973:
940:
937:
920:
917:
907:
904:
879:
876:
860:
857:
851:
850:Transportation
848:
842:
839:
837:
834:
828:
825:
819:
816:
810:
807:
805:
802:
785:Main article:
782:
779:
761:Main article:
758:
755:
741:
738:
732:
729:
720:
717:
715:
712:
694:
691:
678:Main article:
675:
672:
671:
670:
667:
663:
660:
657:
654:
638:
635:
632:
631:
628:
624:
623:
620:
616:
615:
612:
608:
607:
606:1+2+4+5+20+26
604:
600:
599:
596:
592:
591:
588:
584:
583:
580:
576:
575:
572:
550:Main article:
547:
544:
543:
542:
537:
532:
527:
522:
517:
512:
507:
502:
497:
492:
479:
478:
473:
468:
463:
458:
453:
435:Main article:
432:
429:
416:Main article:
413:
410:
405:neural network
392:
389:
388:
387:
386:
385:
382:
379:
376:
373:
367:
366:
365:
362:
356:
355:
354:
351:
348:
345:
342:
330:
327:
305:
302:
297:
296:
290:
284:
278:
272:
266:
260:
242:knowledge base
233:
230:
213:
210:
193:
190:
178:virtual humans
156:
153:
116:
113:
111:
108:
51:
48:
47:
28:
26:
19:
9:
6:
4:
3:
2:
6564:
6553:
6550:
6548:
6545:
6544:
6542:
6527:
6524:
6522:
6519:
6517:
6514:
6512:
6509:
6507:
6504:
6503:
6501:
6497:
6491:
6488:
6486:
6483:
6481:
6478:
6476:
6475:Serious games
6473:
6471:
6468:
6466:
6463:
6461:
6458:
6456:
6453:
6451:
6448:
6446:
6443:
6441:
6438:
6436:
6433:
6431:
6428:
6427:
6425:
6421:
6415:
6412:
6410:
6407:
6405:
6402:
6400:
6397:
6395:
6392:
6388:
6385:
6384:
6383:
6380:
6378:
6375:
6373:
6370:
6368:
6365:
6364:
6362:
6358:
6354:
6347:
6342:
6340:
6335:
6333:
6328:
6327:
6324:
6312:
6309:
6307:
6304:
6302:
6299:
6297:
6294:
6292:
6289:
6287:
6284:
6282:
6279:
6278:
6275:
6269:
6266:
6264:
6261:
6259:
6256:
6254:
6251:
6249:
6246:
6244:
6243:Psychologists
6241:
6239:
6236:
6234:
6233:Organizations
6231:
6229:
6226:
6224:
6221:
6220:
6218:
6216:
6212:
6207:
6204:
6202:
6199:
6197:
6194:
6192:
6189:
6187:
6184:
6182:
6181:John Anderson
6179:
6177:
6174:
6172:
6169:
6167:
6164:
6162:
6159:
6157:
6154:
6152:
6149:
6147:
6144:
6142:
6139:
6137:
6134:
6132:
6129:
6127:
6124:
6122:
6119:
6117:
6114:
6112:
6111:Ulric Neisser
6109:
6107:
6104:
6102:
6099:
6097:
6096:Endel Tulving
6094:
6092:
6089:
6087:
6084:
6082:
6081:Robert Zajonc
6079:
6077:
6074:
6072:
6069:
6067:
6064:
6062:
6059:
6057:
6054:
6052:
6049:
6047:
6044:
6042:
6039:
6037:
6036:Jerome Bruner
6034:
6032:
6029:
6027:
6024:
6022:
6019:
6017:
6014:
6012:
6009:
6007:
6004:
6002:
6001:B. F. Skinner
5999:
5997:
5994:
5992:
5989:
5987:
5984:
5982:
5979:
5977:
5974:
5972:
5969:
5967:
5966:Clark L. Hull
5964:
5962:
5959:
5957:
5954:
5952:
5949:
5947:
5946:Sigmund Freud
5944:
5942:
5939:
5937:
5936:William James
5934:
5932:
5931:Wilhelm Wundt
5929:
5927:
5924:
5923:Psychologists
5919:
5911:
5910:Psychometrics
5908:
5906:
5903:
5901:
5898:
5896:
5893:
5891:
5888:
5886:
5883:
5881:
5878:
5876:
5873:
5871:
5870:Consciousness
5868:
5866:
5863:
5861:
5858:
5856:
5853:
5851:
5848:
5846:
5843:
5841:
5838:
5837:
5833:
5829:
5823:
5820:
5818:
5815:
5813:
5810:
5808:
5805:
5803:
5802:Psychophysics
5800:
5798:
5795:
5793:
5790:
5788:
5785:
5783:
5780:
5778:
5775:
5773:
5770:
5768:
5765:
5763:
5760:
5758:
5755:
5753:
5750:
5749:
5747:
5745:
5744:Methodologies
5741:
5735:
5732:
5730:
5727:
5725:
5722:
5720:
5717:
5715:
5712:
5710:
5707:
5705:
5704:Psychotherapy
5702:
5700:
5699:Psychometrics
5697:
5695:
5692:
5690:
5687:
5685:
5682:
5680:
5677:
5675:
5672:
5670:
5667:
5665:
5662:
5660:
5657:
5655:
5652:
5650:
5647:
5645:
5642:
5640:
5637:
5635:
5632:
5630:
5627:
5625:
5622:
5620:
5617:
5615:
5612:
5610:
5607:
5605:
5602:
5600:
5597:
5595:
5592:
5590:
5587:
5585:
5582:
5580:
5577:
5575:
5572:
5571:
5569:
5567:
5561:
5556:
5546:
5543:
5541:
5538:
5536:
5533:
5531:
5528:
5526:
5523:
5521:
5518:
5516:
5513:
5511:
5508:
5506:
5503:
5501:
5498:
5496:
5493:
5491:
5488:
5486:
5483:
5481:
5478:
5476:
5473:
5471:
5468:
5466:
5465:Developmental
5463:
5461:
5458:
5456:
5453:
5451:
5448:
5444:
5441:
5440:
5439:
5436:
5434:
5430:
5427:
5425:
5422:
5420:
5417:
5415:
5412:
5410:
5407:
5405:
5402:
5400:
5397:
5396:
5394:
5392:
5386:
5380:
5377:
5375:
5372:
5370:
5367:
5365:
5362:
5361:
5358:
5354:
5347:
5342:
5340:
5335:
5333:
5328:
5327:
5324:
5313:
5308:
5300:
5297:
5295:
5292:
5290:
5287:
5285:
5282:
5280:
5277:
5276:
5274:
5272:
5269:
5267:
5264:
5262:
5259:
5257:
5254:
5252:
5249:
5247:
5244:
5242:
5239:
5237:
5234:
5232:
5229:
5225:
5222:
5220:
5217:
5215:
5212:
5210:
5207:
5205:
5202:
5200:
5197:
5195:
5192:
5190:
5187:
5185:
5182:
5180:
5177:
5176:
5174:
5170:
5167:
5166:
5165:
5162:
5158:
5155:
5153:
5150:
5148:
5145:
5143:
5140:
5138:
5135:
5133:
5130:
5128:
5125:
5123:
5120:
5118:
5115:
5113:
5110:
5108:
5105:
5103:
5100:
5098:
5095:
5091:
5088:
5086:
5083:
5082:
5081:
5078:
5074:
5071:
5070:
5069:
5066:
5064:
5061:
5059:
5056:
5054:
5053:dysregulation
5051:
5049:
5046:
5044:
5041:
5039:
5036:
5034:
5031:
5029:
5026:
5024:
5021:
5019:
5016:
5015:
5013:
5009:
5006:
5002:
5001:interpersonal
4999:
4998:
4997:
4994:
4990:
4987:
4986:
4985:
4982:
4980:
4977:
4975:
4972:
4970:
4967:
4965:
4962:
4960:
4957:
4955:
4952:
4950:
4947:
4945:
4942:
4940:
4937:
4935:
4932:
4930:
4927:
4925:
4922:
4920:
4917:
4916:
4914:
4912:
4909:
4907:
4904:
4900:
4897:
4895:
4892:
4891:
4889:
4885:
4882:
4880:
4877:
4875:
4872:
4870:
4867:
4865:
4862:
4861:
4859:
4855:
4854:in psychology
4852:
4850:
4847:
4845:
4842:
4840:
4839:consciousness
4837:
4836:
4834:
4833:
4831:
4827:
4820:
4819:
4814:
4812:
4809:
4807:
4804:
4802:
4799:
4797:
4794:
4792:
4789:
4787:
4784:
4782:
4779:
4777:
4774:
4773:
4771:
4769:
4765:
4760:
4754:
4744:
4741:
4737:
4734:
4733:
4732:
4729:
4727:
4724:
4722:
4719:
4717:
4714:
4712:
4709:
4705:
4702:
4701:
4700:
4697:
4695:
4692:
4690:
4687:
4685:
4682:
4680:
4677:
4675:
4672:
4670:
4667:
4665:
4662:
4659:
4658:
4653:
4650:
4649:
4648:Schadenfreude
4644:
4641:
4640:
4635:
4631:
4628:
4627:
4626:
4623:
4621:
4618:
4616:
4613:
4611:
4608:
4606:
4603:
4601:
4598:
4596:
4593:
4591:
4588:
4584:
4581:
4579:
4576:
4574:
4571:
4569:
4566:
4565:
4564:
4561:
4559:
4556:
4552:
4549:
4548:
4547:
4544:
4542:
4539:
4537:
4534:
4532:
4529:
4527:
4524:
4522:
4519:
4516:
4515:
4514:Mono no aware
4510:
4508:
4505:
4501:
4498:
4496:
4493:
4492:
4491:
4488:
4486:
4483:
4481:
4478:
4476:
4473:
4471:
4468:
4466:
4463:
4461:
4458:
4456:
4453:
4451:
4448:
4446:
4443:
4441:
4438:
4436:
4433:
4431:
4429:
4425:
4423:
4420:
4417:
4416:
4411:
4409:
4406:
4404:
4401:
4399:
4396:
4394:
4391:
4389:
4386:
4383:
4382:
4377:
4375:
4372:
4368:
4367:
4366:Joie de vivre
4363:
4362:
4361:
4358:
4356:
4353:
4351:
4348:
4346:
4343:
4341:
4338:
4336:
4335:Gratification
4333:
4331:
4328:
4326:
4323:
4321:
4318:
4316:
4313:
4311:
4308:
4306:
4303:
4301:
4298:
4296:
4293:
4291:
4288:
4286:
4283:
4281:
4278:
4276:
4273:
4269:
4266:
4265:
4264:
4263:Embarrassment
4261:
4259:
4256:
4254:
4251:
4249:
4246:
4244:
4241:
4239:
4236:
4234:
4231:
4229:
4226:
4224:
4221:
4219:
4216:
4214:
4211:
4209:
4206:
4204:
4201:
4199:
4196:
4194:
4191:
4189:
4186:
4184:
4181:
4179:
4176:
4174:
4171:
4169:
4168:Belongingness
4166:
4164:
4161:
4159:
4156:
4154:
4151:
4149:
4146:
4144:
4141:
4139:
4136:
4134:
4131:
4129:
4126:
4124:
4121:
4119:
4116:
4114:
4111:
4109:
4106:
4104:
4101:
4099:
4096:
4094:
4091:
4089:
4086:
4084:
4081:
4079:
4076:
4075:
4073:
4071:
4067:
4062:
4058:
4051:
4046:
4044:
4039:
4037:
4032:
4031:
4028:
4016:
4008:
4006:
3998:
3996:
3988:
3987:
3984:
3978:
3975:
3973:
3970:
3968:
3965:
3963:
3960:
3958:
3955:
3953:
3950:
3948:
3945:
3943:
3940:
3938:
3935:
3933:
3930:
3928:
3925:
3923:
3920:
3918:
3915:
3913:
3910:
3908:
3905:
3903:
3900:
3898:
3895:
3893:
3890:
3888:
3885:
3883:
3880:
3879:
3877:
3873:
3867:
3864:
3862:
3859:
3857:
3854:
3852:
3851:Mixed reality
3849:
3847:
3844:
3842:
3839:
3837:
3834:
3832:
3829:
3828:
3826:
3824:
3820:
3814:
3811:
3809:
3806:
3804:
3801:
3799:
3796:
3794:
3791:
3790:
3788:
3786:
3782:
3776:
3773:
3771:
3768:
3766:
3763:
3761:
3758:
3756:
3753:
3751:
3748:
3746:
3743:
3741:
3738:
3737:
3735:
3733:
3729:
3723:
3720:
3718:
3715:
3713:
3710:
3708:
3705:
3703:
3700:
3699:
3697:
3695:
3691:
3685:
3684:Accessibility
3682:
3680:
3679:Visualization
3677:
3675:
3672:
3670:
3667:
3665:
3662:
3661:
3659:
3657:
3653:
3647:
3644:
3642:
3639:
3637:
3634:
3632:
3629:
3627:
3624:
3622:
3619:
3617:
3614:
3612:
3609:
3607:
3604:
3603:
3601:
3599:
3595:
3589:
3586:
3584:
3581:
3579:
3576:
3574:
3571:
3569:
3566:
3564:
3561:
3559:
3556:
3554:
3551:
3549:
3546:
3544:
3541:
3539:
3536:
3534:
3531:
3529:
3526:
3524:
3521:
3520:
3518:
3516:
3512:
3506:
3503:
3501:
3498:
3496:
3493:
3491:
3488:
3486:
3483:
3481:
3478:
3476:
3473:
3471:
3468:
3467:
3465:
3463:
3458:
3452:
3449:
3447:
3444:
3442:
3439:
3437:
3434:
3432:
3429:
3428:
3426:
3424:
3420:
3414:
3411:
3409:
3406:
3404:
3401:
3399:
3396:
3394:
3391:
3389:
3386:
3382:
3379:
3378:
3377:
3374:
3373:
3371:
3369:
3365:
3359:
3356:
3354:
3351:
3349:
3346:
3344:
3341:
3339:
3336:
3334:
3331:
3329:
3326:
3324:
3321:
3319:
3316:
3314:
3311:
3310:
3308:
3306:
3302:
3296:
3293:
3291:
3288:
3286:
3283:
3281:
3278:
3276:
3273:
3271:
3268:
3266:
3263:
3261:
3258:
3256:
3253:
3251:
3248:
3247:
3245:
3243:
3239:
3235:
3229:
3226:
3224:
3221:
3219:
3216:
3214:
3211:
3209:
3206:
3205:
3203:
3199:
3193:
3190:
3188:
3185:
3183:
3180:
3178:
3175:
3173:
3170:
3168:
3165:
3164:
3162:
3160:
3156:
3150:
3147:
3145:
3142:
3140:
3139:Dependability
3137:
3135:
3132:
3130:
3127:
3126:
3124:
3120:
3114:
3110:
3107:
3105:
3102:
3100:
3097:
3095:
3092:
3090:
3087:
3085:
3082:
3080:
3077:
3075:
3072:
3070:
3067:
3065:
3062:
3061:
3059:
3057:
3053:
3048:
3042:
3038:
3031:
3026:
3024:
3019:
3017:
3012:
3011:
3008:
3000:
2994:
2990:
2989:
2983:
2979:
2975:
2970:
2965:
2962:(1–2): 1–32.
2961:
2957:
2952:
2951:
2939:
2932:
2925:
2917:
2913:
2909:
2908:Dourish, Paul
2902:
2894:
2890:
2886:
2882:
2878:
2874:
2870:
2866:
2865:Dourish, Paul
2859:
2857:
2855:
2853:
2851:
2849:
2834:on 2017-12-14
2830:
2826:
2822:
2818:
2814:
2809:
2804:
2800:
2796:
2789:
2782:
2774:
2770:
2766:
2760:
2751:
2746:
2741:
2736:
2732:
2728:
2724:
2717:
2709:
2705:
2700:
2695:
2691:
2687:
2682:
2677:
2673:
2669:
2665:
2661:
2657:
2650:
2642:
2638:
2634:
2630:
2623:
2615:
2609:
2605:
2601:
2597:
2590:
2582:
2578:
2574:
2570:
2566:
2562:
2555:
2548:
2540:
2534:
2530:
2526:
2521:
2516:
2512:
2505:
2491:on 2015-04-06
2487:
2483:
2476:
2475:
2467:
2461:
2457:
2453:
2449:
2448:Wiebe, Janyce
2443:
2437:
2432:
2424:
2420:
2416:
2414:9781628727333
2410:
2406:
2399:
2391:
2387:
2383:
2377:
2373:
2369:
2365:
2361:
2354:
2346:
2342:
2337:
2332:
2328:
2324:
2319:
2314:
2310:
2306:
2302:
2295:
2287:
2283:
2277:
2269:
2265:
2261:
2257:
2253:
2246:
2239:
2235:
2229:
2222:
2218:
2212:
2204:
2200:
2196:
2192:
2188:
2184:
2180:
2176:
2169:
2161:
2157:
2152:
2147:
2142:
2137:
2133:
2129:
2125:
2121:
2117:
2110:
2108:
2100:
2094:
2085:
2083:
2074:
2070:
2066:
2062:
2058:
2054:
2053:
2045:
2038:
2036:
2026:
2019:
2015:
2012:
2006:
1999:
1994:
1987:
1983:
1980:
1974:
1966:
1960:
1953:
1949:
1946:
1941:
1929:
1922:
1921:
1916:
1910:
1902:
1898:
1892:
1884:
1880:
1876:
1872:
1865:
1857:
1856:
1848:
1840:
1836:
1832:
1826:
1822:
1818:
1813:
1808:
1804:
1797:
1791:, p. 243
1790:
1785:
1783:
1774:
1767:
1765:
1757:
1751:
1749:
1740:
1736:
1732:
1725:
1718:
1713:
1707:, p. 241
1706:
1701:
1699:
1690:
1684:
1679:
1678:
1669:
1662:
1661:Hudlicka 2003
1657:
1650:
1649:Hudlicka 2003
1645:
1643:
1634:
1630:
1626:
1622:
1617:
1612:
1608:
1601:
1593:
1589:
1585:
1581:
1577:
1570:
1563:
1554:
1546:
1542:
1538:
1532:
1528:
1524:
1520:
1513:
1504:
1502:
1500:
1491:
1487:
1483:
1479:
1475:
1471:
1467:
1463:
1456:
1449:
1441:
1437:
1430:
1423:
1415:
1411:
1406:
1401:
1397:
1393:
1389:
1385:
1384:
1379:
1372:
1364:
1360:
1356:
1352:
1348:
1344:
1340:
1336:
1332:
1328:
1324:
1317:
1303:
1299:
1292:
1284:
1277:
1259:
1254:
1249:
1245:
1241:
1240:
1232:
1225:
1218:
1216:
1202:
1198:
1194:
1187:
1180:
1165:
1158:
1151:
1143:
1136:
1134:
1126:
1122:
1117:
1108:
1104:
1100:
1096:
1092:
1085:
1077:
1073:
1069:
1065:
1058:
1054:
1039:
1036:
1034:
1031:
1029:
1026:
1024:
1021:
1019:
1016:
1014:
1011:
1008:
1005:
1003:
1000:
998:
995:
993:
990:
988:
985:
983:
980:
979:
972:
968:
964:
960:
958:
954:
950:
946:
936:
932:
928:
926:
916:
913:
903:
901:
897:
893:
889:
885:
875:
873:
867:
865:
864:Social robots
856:
847:
833:
824:
815:
801:
798:
793:
788:
773:
769:
764:
750:
746:
740:Disadvantages
737:
728:
726:
711:
709:
705:
701:
690:
686:
681:
668:
664:
661:
658:
655:
652:
651:
650:
647:
645:
629:
626:
625:
621:
618:
617:
613:
610:
609:
605:
602:
601:
597:
594:
593:
589:
586:
585:
581:
578:
577:
574:Action units
573:
570:
569:
566:
563:
559:
553:
541:
538:
536:
533:
531:
528:
526:
523:
521:
518:
516:
513:
511:
508:
506:
505:Embarrassment
503:
501:
498:
496:
493:
491:
488:
487:
486:
484:
477:
474:
472:
469:
467:
464:
462:
459:
457:
454:
452:
449:
448:
447:
444:
438:
428:
424:
419:
409:
406:
402:
398:
383:
380:
377:
374:
371:
370:
368:
363:
360:
359:
357:
352:
349:
346:
343:
340:
339:
337:
336:
335:
326:
322:
320:
314:
312:
301:
294:
291:
288:
285:
282:
279:
276:
273:
270:
267:
264:
261:
258:
255:
254:
253:
249:
247:
243:
239:
229:
225:
223:
218:
209:
207:
202:
198:
189:
187:
182:
179:
175:
174:
169:
165:
164:Marvin Minsky
161:
152:
150:
146:
142:
138:
133:
131:
127:
126:physiological
122:
107:
105:
101:
97:
94:published by
93:
89:
85:
81:
77:
73:
69:
65:
57:
44:
32:
27:
18:
17:
6506:Blast Theory
6435:Mobile games
6382:Storytelling
6367:Role-playing
6156:Larry Squire
6151:Bruce McEwen
6146:Amos Tversky
6116:Jerome Kagan
6106:Noam Chomsky
6046:Hans Eysenck
6016:Harry Harlow
5996:Erik Erikson
5895:Intelligence
5792:Neuroimaging
5535:Quantitative
5500:Mathematical
5495:Intelligence
5485:Experimental
5480:Evolutionary
5470:Differential
5379:Psychologist
5311:
5251:Meta-emotion
5164:Emotionality
5137:responsivity
5085:and bullying
5080:intelligence
4890:Affectivity
4874:neuroscience
4863:
4844:in education
4427:
4388:Homesickness
4364:
4290:Enthrallment
4275:Emotion work
4138:Anticipation
3947:Cyberwarfare
3606:Cryptography
2987:
2959:
2955:
2937:
2924:
2915:
2901:
2876:
2872:
2836:. Retrieved
2829:the original
2798:
2794:
2781:
2773:the original
2769:ScienceDaily
2768:
2759:
2730:
2726:
2716:
2666:(21): 8368.
2663:
2659:
2649:
2632:
2628:
2622:
2595:
2589:
2567:(2): 85–94.
2564:
2560:
2547:
2510:
2504:
2493:. Retrieved
2486:the original
2473:
2466:
2451:
2442:
2431:
2404:
2398:
2363:
2353:
2308:
2304:
2294:
2285:
2276:
2259:
2255:
2245:
2228:
2217:Hagit Hel-Or
2211:
2181:(1): 42–55.
2178:
2174:
2168:
2123:
2119:
2093:
2056:
2050:
2025:
2005:
1993:
1973:
1959:
1940:
1928:the original
1919:
1909:
1900:
1891:
1874:
1864:
1854:
1847:
1802:
1796:
1738:
1734:
1724:
1712:
1676:
1668:
1663:, p. 25
1656:
1651:, p. 24
1606:
1600:
1575:
1562:
1553:
1518:
1512:
1465:
1461:
1448:
1439:
1435:
1422:
1387:
1381:
1371:
1330:
1326:
1316:
1305:. Retrieved
1301:
1291:
1282:
1276:
1265:. Retrieved
1246:(1): 55–83.
1243:
1237:
1224:
1214:
1212:
1205:. Retrieved
1196:
1186:
1178:
1171:. Retrieved
1164:the original
1150:
1141:
1116:
1098:
1094:
1084:
1066:. Vol.
1063:
1057:
969:
965:
961:
942:
933:
929:
922:
909:
881:
868:
862:
853:
844:
830:
821:
812:
804:Facial color
794:
790:
766:
743:
734:
722:
696:
687:
683:
674:Body gesture
648:
640:
555:
530:Satisfaction
480:
440:
425:
421:
397:optical flow
394:
332:
323:
315:
307:
298:
250:
235:
226:
219:
215:
203:
199:
195:
192:Technologies
183:
171:
162:
158:
134:
118:
91:
63:
62:
41:January 2023
38:
30:
6372:Persistence
6228:Disciplines
6201:Susan Fiske
6091:Roger Brown
5991:Carl Rogers
5976:Jean Piaget
5941:Ivan Pavlov
5797:Observation
5777:Experiments
5724:Suicidology
5619:Educational
5574:Anomalistic
5545:Theoretical
5520:Personality
5450:Comparative
5433:Cognitivism
5424:Behaviorism
5179:and culture
4984:recognition
4969:homeostatic
4869:forecasting
4818:Weltschmerz
4791:Misanthropy
4568:grandiosity
4450:Inspiration
4440:Infatuation
4408:Humiliation
4330:Frustration
4203:Contentment
3957:Video games
3937:Digital art
3694:Concurrency
3563:Data mining
3475:Probability
3208:Interpreter
2948:Works cited
2801:(1): 5–18.
2750:2066/103051
1915:Ekman, Paul
1897:Ekman, Paul
1609:: 729–732.
1009:(EmotionML)
953:cognitivist
912:psychomotor
884:biofeedback
878:Video games
818:Methodology
731:Methodology
500:Contentment
6541:Categories
6499:Developers
6485:Flash mobs
6455:Smart toys
6409:Crossmedia
6394:Gamemaster
6387:transmedia
6291:Wikisource
6136:Paul Ekman
5971:Kurt Lewin
5865:Competence
5787:Interviews
5767:Case study
5644:Humanistic
5624:Ergonomics
5609:Counseling
5584:Assessment
5566:psychology
5515:Perception
5475:Ecological
5391:psychology
5369:Philosophy
5353:Psychology
5256:Pathognomy
5157:well-being
5073:and gender
5068:expression
5063:exhaustion
5048:detachment
5033:competence
5014:Emotional
4996:regulation
4979:perception
4974:in animals
4924:and memory
4860:Affective
4768:Worldviews
4630:melancholy
4615:Resentment
4485:Loneliness
4460:Irritation
4445:Insecurity
4435:Indulgence
4310:Excitement
4295:Enthusiasm
4228:Depression
4188:Confidence
4183:Compassion
4158:Attraction
4083:Admiration
4078:Acceptance
4015:Glossaries
3887:E-commerce
3480:Statistics
3423:Algorithms
3381:Stochastic
3213:Middleware
3069:Peripheral
2940:: 647–656.
2838:2016-02-02
2538:1581136374
2495:2016-12-10
2460:1402040261
1812:2102.04029
1390:(102149).
1333:(9): 143.
1307:2008-05-13
1267:2008-05-12
1045:References
997:Chatterbot
957:pragmatist
902:children.
859:Healthcare
630:R12A+R14A
598:1+2+5B+26
510:Excitement
443:Paul Ekman
311:Paul Ekman
232:Algorithms
186:emotion AI
139:, such as
137:modalities
76:psychology
6465:Exergames
6414:Emergence
6311:Wikibooks
6301:Wikiquote
6171:Ed Diener
5956:Carl Jung
5860:Cognition
5689:Political
5599:Community
5429:Cognitive
5284:appraisal
5224:sociology
5175:Emotions
5147:symbiosis
5132:reasoning
5102:isolation
5043:contagion
5028:blackmail
4954:expressed
4949:evolution
4939:and sleep
4929:and music
4864:computing
4811:Reclusion
4806:Pessimism
4781:Defeatism
4711:Suffering
4657:Sehnsucht
4600:Rejection
4551:self-pity
4526:Nostalgia
4495:limerence
4465:Isolation
4403:Hostility
4360:Happiness
4340:Gratitude
4285:Emptiness
4268:vicarious
4218:Curiosity
4193:Confusion
4133:Annoyance
4113:Amusement
4103:Agitation
4098:Affection
4093:Aesthetic
4088:Adoration
3836:Rendering
3831:Animation
3462:computing
3413:Semantics
3104:Processor
2964:CiteSeerX
2938:Proc. CHI
2803:CiteSeerX
2690:1424-8220
2515:CiteSeerX
2423:956349457
2390:144207824
2327:1662-5161
1839:231846518
1741:(6): 137.
1611:CiteSeerX
1482:0929-5593
1363:220717084
1347:0148-5598
1050:Citations
841:Education
614:4+5+7+23
579:Happiness
490:Amusement
466:Happiness
304:Databases
96:MIT Press
6360:Concepts
6306:Wikinews
6263:Timeline
5885:Feelings
5880:Emotions
5840:Behavior
5831:Concepts
5709:Religion
5694:Positive
5684:Pastoral
5669:Military
5634:Forensic
5629:Feminist
5614:Critical
5604:Consumer
5594:Coaching
5589:Clinical
5564:Applied
5460:Cultural
5399:Abnormal
5142:security
5122:literacy
5107:lability
5097:intimacy
5038:conflict
5018:aperture
4915:Emotion
4899:negative
4894:positive
4884:spectrum
4849:measures
4801:Optimism
4796:Nihilism
4786:Fatalism
4776:Cynicism
4721:Sympathy
4716:Surprise
4558:Pleasure
4480:Kindness
4470:Jealousy
4455:Interest
4422:Hysteria
4305:Euphoria
4248:Distrust
4198:Contempt
4178:Calmness
4070:Emotions
4057:Emotions
3995:Category
3823:Graphics
3598:Security
3260:Compiler
3159:Networks
3056:Hardware
2918:: 59–68.
2893:15551492
2825:15296236
2795:CoDesign
2708:36366066
2641:19592726
2635:: 37–9.
2480:. Proc.
2450:(2006).
2345:31114489
2203:16690341
2195:19363957
2160:29555780
2014:Archived
1982:Archived
1948:Archived
1633:11671944
1545:23157273
1414:37091958
1405:10113835
1383:MethodsX
1355:32700060
1258:Archived
1201:Archived
1125:Abstract
975:See also
900:autistic
888:gamepads
809:Overview
719:Overview
627:Contempt
622:9+15+16
595:Surprise
495:Contempt
476:Surprise
238:database
222:prosodic
6521:Niantic
6238:Outline
5734:Traffic
5729:Systems
5664:Medical
5490:Gestalt
5364:History
5312:Italics
5275:Theory
5231:Feeling
5184:history
5169:bounded
5127:prosody
4934:and sex
4919:and art
4879:science
4835:Affect
4829:Related
4704:chronic
4679:Shyness
4639:Saudade
4625:Sadness
4620:Revenge
4610:Remorse
4541:Passion
4531:Outrage
4521:Neglect
4381:Hiraeth
4280:Empathy
4258:Ecstasy
4243:Disgust
4213:Cruelty
4208:Courage
4173:Boredom
4153:Arousal
4143:Anxiety
4128:Anguish
4005:Outline
2699:9654132
2668:Bibcode
2660:Sensors
2569:Bibcode
2336:6503868
2311:: 131.
2286:Gizmodo
2151:5889636
2128:Bibcode
2073:7185733
1592:5790745
1207:May 13,
1173:May 13,
1127:), 1995
892:arousal
619:Disgust
590:1+4+15
587:Sadness
571:Emotion
471:Sadness
456:Disgust
321:(HCI).
121:sensors
84:emotion
68:affects
31:updated
6526:Mojang
6423:Genres
6268:Topics
5714:School
5639:Health
5540:Social
5443:Social
5389:Basic
5374:Portal
5279:affect
5261:Pathos
5214:social
5058:eating
4731:Wonder
4699:Stress
4689:Sorrow
4605:Relief
4595:Regret
4583:vanity
4578:insult
4573:hubris
4428:Ikigai
4398:Horror
4374:Hatred
4233:Desire
4223:Defeat
4148:Apathy
2995:
2966:
2891:
2823:
2805:
2706:
2696:
2688:
2639:
2610:
2535:
2517:
2458:
2421:
2411:
2388:
2378:
2343:
2333:
2325:
2234:Jia Li
2201:
2193:
2158:
2148:
2071:
1837:
1827:
1685:
1631:
1613:
1590:
1543:
1533:
1490:459892
1488:
1480:
1412:
1402:
1361:
1353:
1345:
525:Relief
78:, and
6215:Lists
5674:Music
5659:Media
5654:Legal
5505:Moral
5204:moral
5112:labor
4964:group
4743:Worry
4726:Trust
4694:Spite
4674:Shock
4669:Shame
4563:Pride
4536:Panic
4415:Hygge
4355:Guilt
4350:Grief
4345:Greed
4315:Faith
4253:Doubt
4123:Angst
4118:Anger
4108:Agony
3408:Logic
3242:tools
2934:(PDF)
2889:S2CID
2832:(PDF)
2821:S2CID
2791:(PDF)
2557:(PDF)
2489:(PDF)
2482:DiGRA
2478:(PDF)
2386:S2CID
2199:S2CID
2069:S2CID
2047:(PDF)
1931:(PDF)
1924:(PDF)
1835:S2CID
1807:arXiv
1629:S2CID
1588:S2CID
1572:(PDF)
1541:S2CID
1486:S2CID
1458:(PDF)
1432:(PDF)
1359:S2CID
1261:(PDF)
1234:(PDF)
1197:Wired
1167:(PDF)
1160:(PDF)
611:Anger
582:6+12
540:Shame
515:Guilt
451:Anger
244:, or
147:, or
110:Areas
5900:Mind
5023:bias
5008:work
4590:Rage
4546:Pity
4507:Lust
4490:Love
4393:Hope
4325:Flow
4320:Fear
4300:Envy
4061:list
3240:and
3113:Form
3109:Size
2993:ISBN
2704:PMID
2686:ISSN
2637:PMID
2608:ISBN
2533:ISBN
2456:ISBN
2419:OCLC
2409:ISBN
2376:ISBN
2341:PMID
2323:ISSN
2191:PMID
2156:PMID
1825:ISBN
1683:ISBN
1531:ISBN
1478:ISSN
1410:PMID
1351:PMID
1343:ISSN
1209:2008
1175:2008
1095:Mind
1068:LNCS
603:Fear
461:Fear
293:HMMs
263:k-NN
4475:Joy
4163:Awe
2974:doi
2881:doi
2813:doi
2745:hdl
2735:doi
2694:PMC
2676:doi
2633:144
2600:doi
2577:doi
2525:doi
2368:doi
2331:PMC
2313:doi
2264:doi
2183:doi
2146:PMC
2136:doi
2124:115
2061:doi
1817:doi
1621:doi
1580:doi
1523:doi
1470:doi
1400:PMC
1392:doi
1335:doi
1248:doi
1103:doi
1072:doi
951:'s
281:ANN
275:SVM
269:GMM
257:LDC
6543::
3111:/
2972:.
2960:59
2958:.
2936:.
2910:;
2887:.
2877:65
2875:.
2867:;
2847:^
2819:.
2811:.
2797:.
2793:.
2767:.
2743:.
2731:22
2729:.
2725:.
2702:.
2692:.
2684:.
2674:.
2664:22
2662:.
2658:.
2631:.
2606:.
2575:.
2563:.
2559:.
2531:.
2523:.
2417:.
2384:.
2374:.
2362:.
2339:.
2329:.
2321:.
2309:13
2307:.
2303:.
2284:.
2260:47
2258:.
2254:.
2219:,
2197:.
2189:.
2179:29
2177:.
2154:.
2144:.
2134:.
2122:.
2118:.
2106:^
2081:^
2067:.
2057:19
2055:.
2049:.
2034:^
1877:.
1833:.
1823:.
1815:.
1781:^
1763:^
1747:^
1739:14
1737:.
1733:.
1697:^
1641:^
1627:.
1619:.
1586:.
1578:.
1574:.
1539:.
1529:.
1498:^
1484:.
1476:.
1466:12
1464:.
1460:.
1440:13
1438:.
1434:.
1408:.
1398:.
1388:10
1386:.
1380:.
1357:.
1349:.
1341:.
1331:44
1329:.
1325:.
1300:.
1256:.
1242:.
1236:.
1211:.
1199:.
1195:.
1177:.
1132:^
1097:.
1093:.
947:,
874:.
706:,
702:,
403:,
399:,
240:,
208:.
143:,
132:.
74:,
6345:e
6338:t
6331:v
5431:/
5345:e
5338:t
5331:v
4063:)
4059:(
4049:e
4042:t
4035:v
3049:.
3029:e
3022:t
3015:v
3001:.
2980:.
2976::
2895:.
2883::
2841:.
2815::
2799:1
2753:.
2747::
2737::
2710:.
2678::
2670::
2643:.
2616:.
2602::
2583:.
2579::
2571::
2565:1
2541:.
2527::
2498:.
2425:.
2392:.
2370::
2347:.
2315::
2270:.
2266::
2205:.
2185::
2162:.
2138::
2130::
2075:.
2063::
1967:.
1935:.
1841:.
1819::
1809::
1691:.
1635:.
1623::
1594:.
1582::
1547:.
1525::
1492:.
1472::
1416:.
1394::
1365:.
1337::
1310:.
1270:.
1250::
1244:2
1109:.
1105::
1099:9
1078:.
1074::
43:)
39:(
33:.
Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.