Chapter 3

Semantic Aspects in Sentiment Analysis

M. Nissima; V. Pattib    a University of Groningen, Groningen, The Netherlands
b University of Turin, Torino, Italy

Abstract

Semantics plays an important role in the accurate analysis of the context of a sentiment expression. We analyze this role from two perspectives: the way semantics is encoded in sentiment resources, such as lexica, corpora, and ontologies, and the way it is used by automatic systems that perform sentiment analysis on social media data. Specifically, we review and discuss state-of-the-art methods and tools that rely on semantic models and resources, possibly enabling reasoning, so as to deal with some key challenges for sentiment analysis. Examples of such challenges are context dependency and finer-grained sentiment detection, which can translate into the assigning of sentiment values to aspects of objects or into the design and use of a wider spectrum of affective labels, or the use of current techniques in finer-grained semantic processing. In our dealing with semantics, lexical aspects must be combined with pragmatic and cognitive issues, and other dimensions involved in conveying sentiment, such as emotions.

Keywords

Semantics; Affective lexica; Sentiment resources; Social media annotated corpora; Ontologies; Reasoning with Semantics; Semantic-aware sentiment analysis

1 Introduction

The fact that semantics must play a crucial role in the sentiment interpretation of text is rather obvious, as even just considering the plain meaning of words can be very indicative (“I liked it” vs. “I hated it”). However, things are not that simple or straightforward for at least two reasons: (1) meaning is not so easy to define, detect, and extract automatically, and (2) sentiment analysis is often not just a matter of distinguishing positive from negative opinions, especially in recent developments.

In the 2015 SemEval campaign, four shared tasks were organized within the Sentiment Analysis track: a rather general task on sentiment analysis in Twitter (task 10 [1], with four subtasks), a task focused on figurative language, entitled “Sentiment Analysis of Figurative Language in Twitter” (task 11 [2]), an aspect-based sentiment analysis task (task 12 [3]), where systems had to identify aspects of entities and the sentiment expressed for each aspect, and a rather different task focused on events’ polarity, entitled “CLIPEval Implicit Polarity of Events” (task 9 [4]). Within the ongoing SemEval 2016, there is also a task on detecting stance in tweets (task 61 )—that is, detecting the position of the author with respect to a given target (against/in favor/neutral)—and one on determining sentiment intensity (task 72 ). Some of these tasks provide datasets in more than one language. Additionally, a shared task on concept-level sentiment analysis has been organized recently in the context of the European Semantic Web Conference [5]. This fervent, current action on stimulating research, resources, and tools in this field by organizing more numerous and more complex tasks tells us not only that interest in sentiment analysis is growing but also that sentiment analysis is no longer just about detecting whether a given review or tweet is objective or subjective, and in the latter case it is whether positive or negative. Rather, it requires a more complex analysis and interpretation of messages that in turn must rely on deeper processing and understanding. Thus although it is true that semantics and semantic processing play a crucial role in this, we must see how this happens, from several points of view.

First, and following intuition, words are sentiment informative at a plain lexical semantics level (“good” is positive, “bad” is negative). This is reflected in the creation of sentiment and emotion lexica and corpora that can be used in system development, also for languages other than English. Second, deeper linguistic processing is required to perform finer-grained tasks. For instance, in aspect-based sentiment analysis, entities and aspects must be identified, as well as relations among them, and one cannot rely on lexical semantics only; also, in irony detection, systems must incorporate some module that deals with figurative language. Third, even deeper text processing might not suffice for the level of analysis required, and might need to be complemented by reasoning over concepts, which could be done by exploitation of semantic resources that the Semantic Web community has to offer in this sense, such as web ontologies and semantic networks. Fourth, sentiment analysis stretches out to, and intersects with, other related areas, such as emotion and personality detection, so the semantics of words and text has to be determined at different levels of affect interpretation.

In this chapter, we review a large collection of semantic resources for sentiment analysis and show how semantics plays various roles in the development of sentiment-aware tools and resources. Specifically, we discuss how state-of-the-art semantic processing is used and adapted to fit the requirements of progressively finer-grained tasks; for example, how semantic information is exploited in statistical models, how advances in semantic similarity models can be ported to sentiment analysis, and how automated reasoning and semantic metadata processing can be used in this field. Through this review we highlight the interaction of sentiment analysis with related affect resources and processing.

2 Semantic Resources for Sentiment Analysis

Affective information expressed in our texts is multifaceted. Both sentiment and emotion lexicons, and psycholinguistic resources available for English, refer to various affective models and capture different nuances of affect, such as sentiment polarity, emotional categories, and emotional dimensions. Such lexica are usually lists of words with which a positive or negative or emotion-related label (or score) is associated. Besides flat vocabularies of affective words, other resources include and model semantic, conceptual, and affective information associated with multiword natural language expressions, by enabling concept-level analysis of sentiment and emotions conveyed in texts. In our view all such resources represent a rich and varied lexical knowledge of affect, under different perspectives. Therefore we offer here a comprehensive description of such different resources and of their use in the context of sentiment analysis to distinguish between different opinions and sentiment.

2.1 Classical Resources on Sentiment

One of the first and most widely used resources is the Subjectivity Lexicon [6],3 which is a list of subjectivity clues compiled from several sources, annotated both manually and automatically. This lexicon is the core of OpinionFinder, one of the first systems for the automatic detection of polarity. Another widely used resource is the Opinion Lexicon4 compiled by Bing Liu. The list contains approximately 6800 English words that are classified as either positive or negative. Both resources were compiled manually and are thus quite accurate. However, they make two simplifications: first, they encode sentiment information in terms of a sharp division between a positive and negative sentiment value rather than providing a scale of positivity/negativity; second, they associate sentiment with words rather than with their senses. We will discuss the latter issue first, and the former later.

The lack of sense distinctions is quite a limitation from a semantic perspective. Indeed, lexical entries are often polysemous, so the same string might actually have a completely opposite sentiment depending on the context in which it is used. For example, “crazy” can be used in a negative as well as a positive way, which is strictly context dependent. Creating a flat list where words are assigned a binary polarity value will not account for the complexity of such semantic aspects.

One step further in creating sentiment resources is thus assigning polarity values at the sense level rather than the word level. This is the principle behind the annotation scheme developed in [7], which gave rise to the Subjectivity Sense Annotations,5 a sense-aware lexicon. This resource actually addresses sentiment at one level up with respect to polarity, as it classifies a given word sense as objective or subjective, without specifying, in the latter case, its polarity value. As a plus, though, this lexicon includes part-of-speech information. Apart from sense distinctions of exactly the same lexical entries, there can also be ambiguities related to parts of speech, as a given word could be, for example, a noun and also an adjective, and could exhibit different polarity features accordingly. A good example is “novel,” which has a neutral polarity when used as a noun but a rather positive one when used as an adjective.

The creation of SentiWordNet [8] addresses ambiguities both within and across parts of speech. SentiWordNet is an extension of WordNet [9], where three sentiment scores are assigned to each synset, thus to word senses: positivity, negativity, and objectivity. In addition to the assigning of sentiment to senses, one nice feature of SentiWordNet is that sentiment is a gradual concept rather than a categorical one, thus addressing the first limitation of the simplest resources highlighted above. The same synset can exhibit the three values (positivity, negativity, neutrality) with different degrees, summing up to 1. The main drawback of SentiWordNet is that its sentiment scores are for the most part automatically assigned, so the presence of noise is not trivial. It is also known to have a positive bias [10], and so is not so good for detecting negative opinions.

The very first sentiment-aware resource we know of—namely, General Inquirer6 [11], whose compilation was started in 1966—accounted for sense distinctions, although polarity was conceived as a binary value rather than a gradual concept.

Another way to indirectly address word sense disambiguation is to have domain-dependent lexica. Often a given word will exhibit only one sense in a specific domain (the classic example of “bank” is indicative here, as in a financial context it is unlikely this term would be used in its “river bank” sense). There are two large lexica that are built around very specific domains, exploiting contributors’ reviews: the Yelp Lexicon, with a focus on restaurant reviews, derived from the very large Yelp challenge dataset,7 and the Amazon Lexicon, built on laptop reviews posted on Amazon from June 1995 to March 2013 [12, 13]. Additionally, there are sentiment lists tailored for the financial domain, developed chiefly by Loughran and McDonald [14] after they noted that three quarters of the words with a negative polarity contained in standard, domain-independent resources did not actually have a negative connotation in the financial domain.

Among the resources that conceive polarity as a gradual rather than categorical value, we mention AFINN, which is a manually compiled list of 2477 English words and phrases rated with an integer between − 5 (very negative) and + 5 (very positive),8 and the Sentiment140 lexicon, which contains about 60,000 unigrams and 10 times more bigrams. Differently from the AFINN lexicon, Sentiment140 is compiled automatically by exploiting tweets with emoticons. This resource is part of a suite of lexica all automatically compiled by Said Mohammad and coworkers,9 under the NRC general label [15, 16], which also includes the NRC Hashtag Sentiment Lexicon, built by exploiting the sentiment of hashtags, and two more “double” resources built on the previous ones; namely, the NRC Hashtag Affirmative Context Sentiment Lexicon (and NRC Hashtag Negated Context Sentiment Lexicon), and the Sentiment140 Affirmative Context Lexicon (and Sentiment140 Negated Context Lexicon), where the average sentiment of a term occurring in an affirmative context is separate from the average sentiment of the same term occurring in a negated context.

Within the NRC tools there is also a small manually (crowdsourced) produced lexicon of about 1500 entries, the MaxDiff Twitter Sentiment Lexicon [17]. Handmade resources are always smaller than automatically built ones, as they are costly to create, but they are obviously more reliable. The ways of reducing noise in the automatic acquisition of resources have been explored by, for instance, the use of sentiment proxies such as emoticons [18] or, usually, the adoption of a semiautomatic approach, bootstrapping manually annotated sets of seeds [19, 20]. This is also the approach behind the original creation of SentiWordNet. Finally, using seeds, Chen and Skien [21] automatically produced sentiment lexica for more than 130 languages, reporting an overlap with existing resources of more than 90%.

For the same economical reasons we have just mentioned, and because tools for basic text processing are more advanced for English than for other languages, the creation of lexica for other languages can benefit from the automatic porting of data and techniques from already existing English resources [22]. An example is the transfer of SentiWordNet to other languages. Because English Wordnet is easily aligned to WordNets in other languages, porting SentiWordNet becomes rather straightforward, and it has been shown that sense transfer is robust across languages in about 90% of cases [23] (see also [24] for a WordNet-based, almost unsupervised approach for generating polarity lexica in multiple languages). Transfer of resources can be done in two ways: to create a stable, reusable resource, or “on the fly.” An example of the former is Sentix for Italian (via SentiWordNet), which is currently the most commonly used resource for this language in sentiment analysis systems [25, 26]. As for the latter, resources are translated and used in the context of a sentiment analysis system, but are not necessarily stabilized and distributed for further use; for example, Hernández Farías et al. [27] use a large number of English lexica/resources automatically translated into Italian to exploit information as features in their supervised system, but no specific evaluation or refining of such resources is performed.

2.2 Beyond the Polarity Valence: Emotion Lexica, Ontologies, and Psycholinguistic Resources

Recently, a variety of affective lexica have been proposed to offer information about affect expressed in text according to finer levels of granularity (eg, referring not simply to positive or negative sentiment polarity but to emotional categories such as joy, sadness, and fear). Moreover, a variety of psycholinguistic resources are available that can give some additional measure about the emotional disclosure in social media texts, according to different theoretical perspectives on emotions. All such affect-related resources could be useful with the purpose to increase the coverage of different aspects of affect in textual content. We organize the description of such resources into three groups: the first group is related to information about emotions by referring to a finer-grained categorization model; the second group includes psycholinguistic resources and other resources that refer to different perspectives on affect, according to dimensional approaches to emotion modeling; the third group includes knowledge-based resources and ontologies, which have been developed with the twofold aim to help sentiment analysis systems to grasp the conceptual and affective information associated with natural language opinions, and to use Semantic Web and linked data technologies to provide structured, reusable, and meaningful sentiment analysis results.

2.2.1 Emotion lexica: finer-grained affective lexica based on emotional categories

Theories in the nature of emotion suggest the existence of basic or fundamental emotions such as anger, fear, joy, sadness, and disgust. Different approaches propose different basic or fundamental sets, each having its own specific eliciting conditions and its own specific physiological, expressive, and behavioral reaction patterns. Accordingly, available resources refer to different models of emotions well grounded in psychology, such as the ones proposed by Plutchik [28] and Ekman [29].

One of the first resources referring to a finer-grained model of affect is WordNet-Affect,10 which was developed through the selection and labeling of WordNet synsets representing affective concepts [30]. A number of WordNet synsets are assigned to one or more affective labels (called a-labels). In particular, the affective concepts representing emotional states are individuated by synsets marked with the a-label emotion. There are also other a-labels for those concepts representing moods, traits, situations eliciting emotions, and so on. The newer version, WordNet-Affect 1.1, includes more than 900 synsets and proposes also a taxonomy of emotions, where the hierarchical structure is modeled on the WordNet hyperonym relation. Starting from WordNet-Affect, for task 14 at SemEval 2007, a new version of the resource has been provided [31]. This resource includes only a portion of WordNet-Affect, as it was reannotated at a finer-grained level with use of the six emotional category labels from [29]: joy, fear, anger, sadness, disgust, and surprise. The resource EmoLex11 has been developed as part of the NRC suite of lexica. It is a word-emotion association lexicon [32] built via crowdsourcing; the annotations were manually done through Amazon’s Mechanical Turk. It contains 14,182 words labeled according to Plutchik’s eight primary emotions [28]—joy, sadness, anger, fear, trust, surprise, disgust, and anticipation—and also annotations for negative and positive sentiments. Emolex was originally annotated at a word-sense level. Then the word-level lexicon was created by the taking of the union of emotions associated with all the senses of a word. Very recently, NRC Emolex has been provided also in more than 20 languages as a result of translation of the English terms with use of Google Translate. The underlying assumption is that, despite some cultural differences, affective norms are stable across languages.

Another sense-level affective lexicon is SentiSense,12 which attaches emotional meanings to concepts from the WordNet lexical database [33]. It is composed by a list of 2190 synsets tagged with emotional labels from a set of 14 emotional categories, which refer to a merger of models by Arnold [34], Plutchik, and Parrot [35]: joy, fear, surprise, anger, disgust, love, anticipation, hope, despair, sadness, calmness, like, hate, and ambiguous. Such emotional categories are also related via an antonym relationship. SentiSense has been developed semiautomatically with use of several semantic relations between synsets in WordNet.

2.2.2 Psycholinguistic resources and other accounts of affect

Psycholinguistic resources can be also helpful for capturing the emotional content of the text, allowing the mapping of words onto psychologically valid dimensions. One of the most well-known psycholinguistic resources is the Linguistic Inquiry and Word Count (LIWC)13 dictionary [36], which assigns one or more psychological categories, such as positive emotion and negative emotion, to individual words. The 2007 version of LIWC comprises almost 4500 words and word stems distributed in categories for analysis of psycholinguistic features in texts. For example, the word “happy” would be labeled with the categories positive emotion and affect. The most recent evolution, LIWC2015, is composed of almost 6400 words, word stems, and selected emoticons, distributed in a wider set of categories arranged hierarchically. The dictionary has been extended and new categories have been added. The new additions allow the user to better tackle with social media language.

Moreover, some psychological theories propose that the nature of an emotional state is determined by its position in a space of independent dimensions. According to such an approach, emotions can be defined as a coincidence of values on a number of different strategic dimensions. Therefore they are described not by marking a small set of discrete categories but rather by scoring properties such as valence (positive/negative) and arousal (active/passive) in a continuous range of values. Accordingly, some lexical resources have been built by reference to models with a dimensional view on affect-related phenomena.

For instance, the Dictionary of Affect in Language14 developed by Whissell [37] includes 8742 English words rated on a three-point scale along three dimensions: activation (degree of response that humans have under an emotional state); imagery (how difficult it is to form a mental picture of a given word); pleasantness (degree of pleasure produced by words). Instead, the Affective Norms for English Words15 [38] is a database developed by the rating of 1034 English words in terms of the Osgood, Suci, and Tannenbaum dimensional theory of emotions along three dimensions of the valence-arousal-dominance model [39]: valence or pleasure (the polarity of the emotional activation ranging from positive to negative); arousal or intensity (the degree of excitement or activation an individual feels toward a given stimulus, ranging from calm to exciting); dominance or control (the degree of control an individual feels over a specific stimulus, ranging from out of control to in control).

2.2.3 Concept-based resources and ontologies: toward a knowledge-based approach to sentiment analysis

Many researchers are currently devoting efforts to develop ontologies of emotions, in some cases also referring to the Semantic Web initiative [4043]. The use of structured knowledge via ontologies or semantic networks in the sentiment analysis tasks opens new opportunities for understanding opinions expressed in texts. In particular, such knowledge bases can include semantic information both on the sentiment domain and the emotion domain, or on the concepts and the context related to words used for expressing opinions, enabling interesting possibilities of reasoning on such knowledge (see Section 3.4). Again, these ontologies typically try to mirror models of emotions well established in psychology. However, as a difference from emotion lexica such as Emolex, where emotions are taken as a flat list of concepts, here a more sophisticated knowledge can be encoded. For instance, in ontologies the taxonomic relationships among the emotions can be represented, or other interesting relations such as intensity, similarity, or oppositions. The representation of such relationships between concepts in the ontology enables also interesting possibility to automatically reason on such knowledge about emotions. As we will see, Plutchik’s circumplex model [28], which inspired the development of the Emolex lexicon, has also been exploited as a reference of concept-level resources and ontology of emotions with important differences that we will highlight.

Concept-level resources use ontologies or semantic networks to enable semantic text analysis. One of the most used resources in this category is SenticNet16 [40], which aims to create a collection of commonly used polarity concepts (ie, commonsense concepts with relatively strong positive or negative polarity). Differently from other resources such as SentiWordNet, which also contains null-polarity terms, SenticNet does not contain concepts with neutral or almost neutral polarity. The current version includes 30,000 natural language concepts collected from the Open Mind corpus, and the resource is distributed in Resource Description Framework (RDF) XML format. Each concept is associated with emotion categorization values expressed in terms of the hourglass of emotions model [44], which organizes and blends 24 emotional categories from Plutchik’s model into four affective dimensions (pleasantness, attention, sensitivity, and aptitude). Moreover, in SenticNet a polarity value that lies in the interval [−1, 1] is associated with each concept. This value is calculated in terms of the four dimensions of the hourglass of emotions model, and specifies if (and to what extent) the input concept is positive or negative (eg, concepts such as “make a good impression” will have a polarity value close to 1, whereas concepts such as “being fired,” “leave behind,” or “lose control” will have a polarity value close to − 1. In this case the emotion categorization model supports comparison and aggregation among results of an emotional analysis of concepts into polarity values.

EmoSenticNet17 is another concept-based lexical resource [45] and was automatically built by the merging of WordNet-Affect and SenticNet, with the main aim to have a more complete resource containing not only quantitative polarity scores associated with each SenticNet concept but also qualitative affective labels. In particular, it assigns WordNet-Affect emotion labels related to the six Ekman basic emotions mentioned before (disgust, sadness, anger, joy, fear, and surprise) to SenticNet concepts. The whole list currently includes 13189 annotated entries.

In [46], an ontology of emotional categories based Plutchik’s circumplex model [28] was developed (ArsEmotica Ontology of Emotions), encoded in OWL 2 QL. The ontology has been defined in the context of the ArsEmotica project (see Section 3.4) but it is so generic that it might be used to analyze emotions in running text in any domain. It encodes the emotional categories of Plutchik’s model and links them with semantic relationships, which can be represented as a wheel of emotions. In particular, it accounts for various levels of intensity, the similarity and opposite relationships, and compositions of basic emotions (primary dyads). Overall, it distinguishes the 32 emotional concepts of the wheel of emotions. The emotional concepts have been semiautomatically linked to synsets from WordNet-Affect 1.1, and then to Italian lexical entries, by exploitation of MultiWordNet, an Italian WordNet strictly aligned with Princeton WordNet 1.6. The linkage between the language level (lexicon based) and the conceptual level representing the emotional concepts has been formalized by integration of the ontology framework Lexicon Model for Ontologies [47].

Finally, we mention a set of affect-related ontologies in the Semantic Web field that are not lexical resources but were created as a response to a more foundational need to represent all the main features of an emotion, and to standardize the knowledge of emotions, so as to support very broad semantic interoperability among affective computing applications, by allowing the mapping of concepts and properties belonging to different emotion representation models.

The Human Emotion Ontology [41] has been developed in this direction, as has the semantic vocabulary Onyx,18 which has been proposed to describe emotion analysis processes and results [43].

2.3 Social Media Corpora Annotated for Sentiment and Fine Emotion Categories

The growing interest in the development of automatic systems for sentiment analysis has also prompted the production of annotated corpora that could be exploited for the development of such systems in a learning fashion (see also Section 3).

The design of schemes for the annotation of corpora is always a task in the field of data classification, which leads to theoretical assumptions about the concepts to be annotated. It defines what kind of information must be annotated, the inventory of markers to be used, and the annotation’s granularity. As highlighted in [48], in annotated corpora for sentiment analysis this is especially challenging. Research in psychology outlines three main approaches to the modeling of emotions and sentiments: the categorical, the dimensional, and the appraisal-based approach. The most widespread are the categorical and the dimensional ones, which describe emotions by marking a small set of discrete categories and scoring properties in a continuous range of values. Accordingly, the kinds of knowledge usually annotated are the sentiment’s polarity (positive vs. negative), category (happiness vs. sadness), the source and target of the sentiment, and its intensity. Annotations can be based on simple broad polarity labels, possibly equipped with intensity ratings, which also helps us to classify texts where mixed sentiments are expressed. They can also be based on labels representing different emotions.

The resources developed within the SemEval evaluation campaigns (regularly since 2013) comprise tweet collections annotated for subjectivity, polarity, emotions, irony, and aspects/properties, and are not restricted to English [25, 26, 48, 49]. In addition to the SemEval related data, other collections have been constructed, especially on news and events or product reviews, which are obviously a very good source of opinionated texts, and a field of interest for any business-related application that works with users’ opinions. For example, a widely used dataset is the Movie Review Data,19 a collection of movie reviews [50]. Sentiment labels (positive vs. negative) are assigned to the global review, while in an extension of this corpus, annotation is done at the sentence level [51]. Another classic resource is the MPQA corpus,20 based on [52], which contains news articles from a wide variety of sources and which was manually annotated for opinions and other private states. This corpus has been expanded very recently with finer-grained information, and now includes entity and event target annotations [53], in line with the increasing interest in aspect-based sentiment analysis. A pioneer work in this sense is [54]: given a product, Hu and Liu [54] mine and summarize reviews related to specific features of that product, on which customers have expressed their positive or negative opinion, so that sentiment is associated with aspects of rather than the whole product. Extraction of opinions on aspects (specifically service, location, and rooms of hotels) is also the focus in [55], where the authors collected and processed 10,000 TripAdvisor reviews.

Interest in properties and aspect-based sentiment analysis is increasing also in Twitter-based work, as shown by recently organized evaluation campaigns on this [3]. Indeed, even if a lot shorter in characters than a product review, the same tweet can contain both positive and negative information, possibly relating to different entities that are mentioned. Niek Sanders’s Twitter dataset21 is not aspect annotated but a specific topic is assigned to each tweet, and this topic can be possibly used as a proxy for the entity with which the sentiment is associated, as is also done in [25], where specific hashtags are used as proxies.

Still in term of proxies, it is interesting to note that to save annotation effort in the creation of sentiment datasets, there have been experiments with distant learning, where class labels are not assigned manually but are rather derived from other information available. Performing sentiment analysis on tweets, Go et al. [56] trained a few classifiers using emoticons as noisy labels, and achieved an accuracy of about 80%. Recently, a new Twitter corpus was released by Mohammad et al. [57], where a multilayer scheme is applied. It contains a set of tweets with annotations concerning different aspects: sentiment (positive or negative), but also finer-grained annotation of emotions (anger, anticipation, disgust, fear, joy, sadness, surprise, and trust), and in addition purpose (to point out a mistake, to support, to ridicule, etc.) and style (simple statement, sarcasm, hyperbole, understatement). The corpus is not generic but focused on the political domain. Mohammad et al. collected tweets labeled with a set of hashtags pertaining to the 2012 US presidential election. The tweets were annotated manually by reliance on crowdsourcing platforms. This is the first dataset including a fine-grained annotation concerning stylistic features related to irony, and also a multilayer annotation concerning affect, since both annotations on sentiment polarity and on emotions are provided. For style, only 23% of the tweets were labeled with a style tag pertinent to the expression of irony, whereas most of them were annotated with the label simple statement, which can be interpreted as a tag for marking nonironic expressions.

Previous corpora annotated for fine-grained emotion categories, such as the one proposed for SemEval 2007 task14 [31], were not focused on social media data. The availability of such a new kind of datasets opens the way to the possibility of building machine learning systems that predict emotion and purpose labels in unseen tweets, as proposed by Mohammad et al. [57], who also presented the first results and a baseline for such new sentiment-related tasks on Twitter.

3 Using Semantics in Sentiment Analysis

Because of the task’s nature, semantics is obviously a crucial ingredient of any sentiment analysis system. However, depending on the system’s complexity, and depending also on the specific task that is undertaken (see Section 1), semantic information of different types can be accessed and incorporated in various different ways. In Section 3.1 we discuss how lexical information mostly derived by the resources discussed in Section 2 is used, while in Section 3.2 we briefly review how current semantic similarity models are adapted to sentiment processing. In Section 3.3 we explore the deeper semantic processing that is required by finer-grained tasks such as aspect-based sentiment analysis. Finally, in Section 3.4 we review systems that perform sentiment analysis by reasoning over semantic resources.

3.1 Lexical Information

Independently of the approach (rule based or statistical), virtually all systems for sentiment analysis rely on information derived from lexical resources. In machine learning such information is used as features, normally in combination with other features, too. The best performing system for subtask B of SemEval 2015’s task 10, which is the most standard message-level polarity task, is an ensemble classifier that builds on three well-performing classifiers at SemEval 2013 and one at SemEval 2014 [58]. Although the classifiers were handpicked with attention not only to performance but also to significant differences in feature combinations, all four classifiers employ polarity dictionaries of some sort and in some way. Some systems rely on one resource only; for example, Günther and Furrer [59] only use SentiWordNet, while other systems will try to use and combine information from all available sources [15, 60].

As a general strategy, lexical information is collected at the word level (eg, the categorical or gradual polarity of a given token) for each word of the whole tweet, and is then propagated via different combinations at the sentence/message level. Such combinations yield values that are then used as different features. For example, feature values can be the number of tokens in the text with a positive score or the number of those with a negative score, but also derived values such as the global polarity score of a tweet obtained as the average of the tokens’ polarities. Also, maximum and minimum scores, when available, are used as features.

Most statistical approaches are support vector machine models that implement such features and similar features, and Taboada et al. [10] highlight that information from basic unigrams appears to be the most useful. Although this shows that words in themselves are very informative for this task, the intrinsic limitations in the use of information from a plain dictionary lookup are quite evident if one thinks that words are used in a specific (syntactic) context, and their polarity can change substantially according to how they relate to other words in the text. Mainly because of the lack of tools that can deal with Twitter, few systems perform word sense disambiguation, so the richer sense-annotated lexica cannot be exploited to their full potential. A way of addressing polysemy without disambiguation was suggested by Basile and Nissim [25], who introduced the concept of polypathy (calculated as the standard deviation of the polarity scores of the possible senses of a lemma) as an indicator of variance of polarity scores across a lemma’s synsets, which can then be used as a feature or threshold value (tokens that have a too high polypathy could, for example, be ignored, as they provide conflicting information).

Another crucial limitation of not taking context into account is negation [61] (recall also the separate lexica for negated and affirmative contexts reviewed earlier). To cope with this aspect, Taboada et al. [10] incorporate information from what they call contextual valence shifters, showing an increase in performance. This is a first step toward deeper language processing, which is considered more and more necessary, even for short texts such as tweets.

3.2 Distributional Semantics

In the context of sentiment analysis the idea of exploiting the distributional hypothesis—namely, the assumption that words that occur in the same contexts tend to have similar meanings [62, 63]—simply boils down to the fact that similarity models that predict, for example, that “amazing” and “wonderful” are similar could be extended to predict that if “amazing” has a positive value, so too will “wonderful,” and will be at the opposite end of spectrum to, say, “terrible” and “awful.” However, general similarity models usually take into account the lexical and morphosyntactic context of a word but not necessarily the text’s polarity. Therefore similarity is potentially accurate at a syntactic and more general semantic level but not necessarily in a sentiment-aware way, so “good” and “bad” end up being very similar. This is true for classic distributional similarity models as well as for the more recent and successful distributed vector representations known as word embeddings, at least in their standard formulation [64, 65].

As a first attempt at directly incorporating sentiment in learning the distributional context of a word, so that words expressing similar sentiment do indeed end up having similar vector representations, Maas et al. [66] developed a model where they give a sentiment predictor function to each word. On the positive/negative classification of tweets, their system is shown to perform better than models that incorporate embeddings trained in a nonsentiment-aware fashion. An even more powerful and recent model, which outperforms that of Maas et al. [66], was proposed by Tang et al. [67], who train a neural network by associating each n-gram with the polarity of a sentence (thus beyond the word level, as in [66]), and show that sentiment-specific word embeddings effectively distinguish words with opposite sentiment polarity. This model performs better than other models that use generally learned embeddings.

3.3 Entities, Properties, and Relations

Interest in finer-grained sentiment analysis has also necessarily prompted the need for finer-grained semantic analysis, and therefore deeper language processing. For example, aspect-based sentiment analysis must rely on the identification of specific entities and/or properties of entities in reviews or tweets. To do this, standard techniques for entity detection and classification are employed, such as sequential taggers, possibly retrained for specific domains. Particular attention to (named) entities in sentiment analysis is also shown by the OpeNER EU-funded project,22 which focuses on named entity recognition within sentiment analysis.

Further, relations between the entities and events involved must be identified, so as to know what is said of which entity. An obvious way to do this is to exploit dependency relations, although deeper processing of tweets is not so simple because of the idiosyncratic and often ungrammatical language such short texts contain (although recent work based on learning neural knowledge graph embeddings shows an error reduction of more than 26% in semantic parsing of tweets [68]). Similar issues arise when one is developing systems to detect stance, as in order to assess the opinion of someone toward a given target, all relations between the entities involved must be correctly identified and associated with the sentiment expressed. However, a deeper linguistic analysis of text is also beneficial, if not necessary, for standard message- or text-level sentiment analysis, as it helps to treat the issue of contextual valence shifters mentioned in Section 3.1 by also accounting for word order and sentence structure. To this end, the Natural Language Processing Group at Stanford University developed a sentiment treebank.23 This treebank has been used to train a recursive neural network built on top of grammatical structures [69], achieving an increase of 5 percentage points on sentence polarity classification. On fine-grained sentiment level they obtained a 9.7% improvement over a bag-of-words baseline, and overall showed the ability to accurately capture the effects of negation and its scope at various levels in the tree structures.

3.4 Concept-Level Sentiment Analysis: Reasoning With Semantics

Approaches enabling concept- and context-based analysis can lead to a better understanding of opinions expressed in textual data, therefore reducing the gap between unstructured information and structured machine-processable data. Concept-level sentiment analysis exploits large semantic knowledge bases (eg, ontologies and semantic networks), together with natural language processing tools and techniques, thus stepping away from blind use of keywords and word co-occurrence counts. Rather, it relies on the implicit features associated with natural language concepts. Unlike purely syntactic techniques, concept-based approaches are able to detect also sentiments that are expressed in a subtle manner (eg, through the analysis of concepts that do not explicitly convey any emotion) but that are implicitly linked to other concepts that do so.

The accent on concept-level sentiment analysis is recent. We can observe different knowledge representations of the concepts related to words and of their affective load, as well as different ways in which systems that embrace this perspective on sentiment analysis exploit and reason about such knowledge, depending also on the specific task they address.

Also, this perspective on sentiment analysis offers a challenge to the Semantic Web community. A periodic shared task on concept-level sentiment analysis has been launched recently [5], and some of the systems described below explicitly rely on the use of Semantic Web and linked data resources and tools to enable automated reasoning (eg, ontological reasoning on the taxonomic structure of an ontology of affect) and semantic metadata processing.

Let us start with Sentilo,24 an unsupervised, domain-independent system that relies on a knowledge-based approach where existing tools for natural language processing and publicly available lexical/affective resources are combined with Semantic Web techniques for representing and reasoning about knowledge involved in opinion sentences. Sentilo exploits the semantic graph representation of a sentence enriched with opinion-related information (eg, opinion holder, topics, sentiment scores). Such a semantic representation allows Sentilo to address the finer-grained tasks of detecting holders and topics of an opinion, and additionally to identify and distinguish both main topics and subtopics. Specifically, given an input opinion sentence (natural language text), the system returns the corresponding semantic representation as a FRED graph (RDF/OWL graph) annotated with concepts from a semantic model of opinions. Annotations concerning the opinion holder, opinion features, and opinion topics are produced on the basis of a set of heuristic rules exploiting the FRED semantic representation and by use of lexical and affective resources, such as SentiWordNet and SenticNet. An evaluation of Sentilo as the opinion holder and main/subtopic detection tool has been presented [70]. The system is able to achieve good performance compared with the other tools, highlighting promising results on opinion holder (F1: 95%), subtopic (F1: 78%), and topic (F1: 68%) detection. Sentilo is also available as a REST service that returns RDF as output. This is important to make the output of the system machine processable and reusable.

In other systems the accent is on the use of conceptual and affective knowledge of words, the focus being on sophisticated representations of the affective space that can enable commonsense or ontological reasoning on networks or taxonomies of concepts. This is, for instance, the case for many systems exploiting the resources of the SenticNet suite of resources [71]. Another system that can be placed in this category is ArsEmotica,25 which is an example of a domain-dependent system focused on cultural heritage. It applies sentiment analysis to resources from online art collections, and exploits tags given by the visitors on social platforms as an additional source of information [46]. The extraction of the emotional semantics from social tags is driven by an ontology-based representation of the emotional categories presented in the previous section. Methods and tools from the Semantic Web and natural language processing provide the building blocks for the creation of a semantic social space to organize artworks according to the ontology of emotions. The output is encoded into World Wide Web Consortium ontology languages. The semantic representation is not limited to affect but concerns also aspects of the specific art domain. This gives the twofold advantage of enabling tractable reasoning on the relationships between emotions and other interesting dimensions of the domain (eg, artists and artworks), and fostering the interoperability and integration of tools developed in the Semantic Web and linked data community. The system has been evaluated against a dataset of tagged multimedia artworks from the ArsMeteo Italian art portal.26 A SPARQL end point has been implemented to explore the collection and to extract information about the relationships among emotions, artworks and authors by posing queries such as: “Give me the artworks stirring emotions similar to sadness and belonging to the music genre.” In some systems the focus on conceptual semantics, not limited to affect, is also combined with issues related to the contextual semantics of words. Such approaches do not offer fixed sentiment polarities but assign context-specific sentiment orientation to words. An interesting proposal in this line is SentiCircles, a lexicon-based approach for the detection of sentiment in Twitter posts at both entity level and tweet level [72]. SentiCircle builds a dynamic representation of context to tune the preassigned strength and polarity of words in the lexicon, and incorporates both contextual semantics (ie, semantics inferred from the co-occurrence of words) and conceptual semantics (ie, semantics extracted from background ontologies).

4 Conclusions

We started this contribution by pointing out how much interest and research in sentiment analysis have grown in the past few years. We have highlighted that in the recent SemEval campaigns tasks more complex than basic polarity classification have been proposed. And complexity here means finer-grained interpretations of sentiment, be it related to figurative uses such as irony or be it related to singling out which entities or entity properties are actually in focus in an opinionated statement. This is also reflected in resource creation, where we have seen that corpora have been extended in their annotation to include polarity at the sentence level, where before this was done at the global document level, or at the entity level, where before it was done at the message level.

Accordingly, sentiment analysis systems are required to be ever more accurate and more sophisticated at the same time, thus attempting to perform deeper semantic processing that is rather successful on language data cleaner than that from Twitter and for other natural language processing tasks but is still experimental with regard to noisy social media data, with all the subjectivity that sentiment analysis carries. We have provided a survey of the state-of-the-art approaches and resources in this sense, highlighting how semantics is being coded and used in this emerging and growing field. From what we see, sentiment analysis in social media appears to be expanding in a variety of directions, incorporating and adapting processing tools from natural language processing and reasoning tools from the Semantic Web community, as well as including possibly orthogonal aspects such as a whole range of affect-related concepts. What we also see is that semantics necessarily remains the core of this, in the use, coding, and processing of the meaning of words and larger phrases, so we can only hope that this survey serves as a starting point for much more to come.

Our expectations do not concern only the development of sentiment analysis per se, since related tasks, such as personality and irony detection [7375], or author profiling [76], also have interest in exploiting the lexical knowledge of affect encoded in the resources we have described.

References

[1] Rosenthal S., Nakov P., Kiritchenko S., Mohammad S., Ritter A., Stoyanov V. SemEval-2015 task 10: sentiment analysis in Twitter. In: Proceedings of the Ninth International Workshop on Semantic Evaluation (SemEval 2015). Denver, CO: Association for Computational Linguistics; 2015:451–463.

[2] Ghosh A., Li G., Veale T., Rosso P., Shutova E., Barnden J., Reyes A. SemEval-2015 task 11: sentiment analysis of figurative language in Twitter. In: Proceedings of the Ninth International Workshop on Semantic Evaluation (SemEval 2015). Denver, CO: Association for Computational Linguistics; 2015:470–478.

[3] Pontiki M., Galanis D., Papageorgiou H., Manandhar S., Androutsopoulos I. SemEval-2015 task 12: aspect based sentiment analysis. In: Proceedings of the Ninth International Workshop on Semantic Evaluation (SemEval 2015). Denver, CO: Association for Computational Linguistics; 2015:486–495.

[4] Russo I., Caselli T., Strapparava C. SemEval-2015 task 9: CLIPEval implicit polarity of events. In: Proceedings of the Ninth International Workshop on Semantic Evaluation (SemEval 2015). Denver, CO: Association for Computational Linguistics; 2015:443–450.

[5] Recupero D.R., Cambria E. ESWC’14 challenge on concept-level sentiment analysis. In: Communications in Computer and Information Science; Springer; 3–20. Semantic Web Evaluation Challenge—SemWebEval 2014 at ESWC 2014, Anissaras, Crete, Greece, May 25–29, 2014, Revised Selected Papers. 2014;475.

[6] Wilson T., Wiebe J., Hoffmann P. Recognizing contextual polarity in phrase-level sentiment analysis. In: Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing, HLT ’05. 2005:347–354 Vancouver, Canada.

[7] Wiebe J., Mihalcea R. Word sense and subjectivity. In: Proceedings of the 21st International Conference on Computational Linguistics and the 44th Annual Meeting of the Association for Computational Linguistics. 2006:1065–1072.

[8] Baccianella S., Esuli A., Sebastiani F. SentiWordNet 3.0: an enhanced lexical resource for sentiment analysis and opinion mining. In: Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC’10). Valletta, Malta: European Language Resources Association (ELRA); 2010:2200–2204.

[9] Fellbaum C. WordNet: An Electronic Lexical Database. Cambridge, MA: Bradford Books; 1998.

[10] Taboada M., Brooke J., Tofiloski M., Voll K., Stede M. Lexicon-based methods for sentiment analysis. Comput. Linguist. 2011;37(2):267–307.

[11] Stone P.J., Dunphy D.C., Smith M.S., Ogilvie D.M. The general inquirer: a computer approach to content analysis. Am. J. Sociol. 1966;73(5):634–635.

[12] McAuley J.J., Leskovec J. Hidden factors and hidden topics: understanding rating dimensions with review text. In: in: Proceedings of the 7th ACM Conference on Recommender Systems. 2013:165–172.

[13] Kiritchenko S., Zhu X., Cherry C., Mohammad S. NRC-Canada-2014: detecting aspects and sentiment in customer reviews. In: Proceedings of the Eighth International Workshop on Semantic Evaluation (SemEval 2014). Dublin, Ireland: Association for Computational Linguistics and Dublin City University; 2014:437–442.

[14] Loughran T., McDonald B. When is a liability not a liability? Textual analysis, dictionaries, and 10-ks. J. Financ. 2011;66(1):35–65.

[15] Mohammad S., Kiritchenko S., Zhu X. NRC-Canada: building the state-of-the art in sentiment analysis of tweets. In: Proceedings of the Seventh International Workshop on Semantic Evaluation Exercises (SemEval-2013). 2013:321–327 Atlanta, GA.

[16] Zhu X., Kiritchenko S., Mohammad S. NRC-Canada-2014: recent improvements in the sentiment analysis of tweets. In: Proceedings of the Eighth International Workshop on Semantic Evaluation (SemEval 2014). Dublin, Ireland: Association for Computational Linguistics and Dublin City University; 2014:443–447.

[17] Kiritchenko S., Zhu X., Mohammad S.M. Sentiment analysis of short informal texts. J. Artif. Intell. Res. 2014;50:723–762.

[18] Pak A., Paroubek P. Twitter for sentiment analysis: when language resources are not available. In: Proceedings of the 23rd International Workshop on Database and Expert Systems Applications. 2011:111–115.

[19] Pitel G., Grefenstette G. Semi-automatic building method for a multidimensional affect dictionary for a new language. In: Proceedings of the Sixth International Conference on Language Resources and Evaluation (LREC’08). Marrakech, Morocco: European Language Resources Association (ELRA); 2008.

[20] Smedt T.D., Daelemans W. “Vreselijk mooi!” (terribly beautiful): a subjectivity lexicon for Dutch adjectives. In: Calzolari N., Choukri K., Declerck T., Dogan M.U., Maegaard B., Mariani J., Odijk J., Piperidis S., eds. Proceedings of LREC 2012. European Language Resources Association (ELRA); 2012:3568–3572.

[21] Chen Y., Skiena S. Building sentiment lexicons for all major languages. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). Baltimore, Maryland: Association for Computational Linguistics; 2014:383–389.

[22] Banea C., Mihalcea R., Wiebe J. Porting multilingual subjectivity resources across languages. Trans. Affect. Comput. 2013;4(2):211–225.

[23] Banea C., Mihalcea R., Wiebe J. Sense-level subjectivity in a multilingual setting. Comput. Speech Lang. 2014;28(1):7–19.

[24] San Vicente I., Agerri R., Rigau G. Simple, robust and (almost) unsupervised generation of polarity lexicons for multiple languages. In: Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics (EACL2014). 2014:88–97 Gothenburg, Sweden.

[25] Basile V., Nissim M. Sentiment analysis on Italian tweets. In: Proceedings of the Fourth Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis. 2013:100–107 ACL, Atlanta, GA.

[26] Basile V., Bolioli A., Nissim M., Patti V., Rosso P. Overview of the Evalita 2014 SENTIment POLarity classification task. Proceedings of the Fourth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian (EVALITA 2014), Pisa, Italy. 2014.

[27] Hernández Farías D.I., Buscaldi D., Priego-Sánchez B. IRADABE: adapting English lexicons to the Italian sentiment polarity classification task. In: First Italian Conference on Computational Linguistics (CLiC-it 2014) and the Fourth International Workshop EVALITA 2014. 2014:75–81.

[28] Plutchik R. The nature of emotions. Am. Sci. 2001;89(4):344–350.

[29] Ekman P. An argument for basic emotions. Cogn. Emot. 1992;6(3–4):169–200.

[30] Strapparava C., Valitutti A. WordNet-Affect: an affective extension of WordNet. In: Proceedings of the International Conference on Language Resources and Evaluation (LREC’04). European Language Resources Association (ELRA); 1083–1086. 2004;4.

[31] Strapparava C., Mihalcea R. SemEval-2007 task 14: affective text. In: Proceedings of the Fourth International Workshop on Semantic Evaluations. Stroudsburg, PA: Association for Computational Linguistics; 2007:70–74.

[32] Mohammad S.M., Turney P.D. Crowdsourcing a word-emotion association lexicon. Comput. Intell. 2013;29(3):436–465.

[33] J. Carrillo de AlbornozPlaza L., Gervás P. SentiSense: an easily scalable concept-based affective lexicon for sentiment analysis. In: Proceedings of the Eight International Conference on Language Resources and Evaluation (LREC’12). Istanbul, Turkey: European Language Resources Association (ELRA); 2012:3562–3567.

[34] Arnold M.B. Emotion and Personality. New York, NY: Columbia University Press; . 1960;1.

[35] Parrot W.G. Emotions in Social Psychology: Essential Readings. Philadelphia: Psychology Press; 2001.

[36] Pennebaker J.W., Francis M.E., Booth R.J. Linguistic Inquiry and Word Count: LIWC 2001. Mahwah, NJ: Lawrence Erlbaum Associates; 2001.

[37] Whissell C. Using the revised dictionary of affect in language to quantify the emotional undertones of samples of natural languages. Psychol. Rep. 2009;2(105):509–521.

[38] Bradley M.M., Lang P.J. Affective norms for English words (ANEW): instruction manual and affective ratings. Gainesville, FL: Center for Research in Psychophysiology, University of Florida; 1999.

[39] Russell J.A., Mehrabian A. Evidence for a three-factor theory of emotions. J. Res. Pers. 1977;11(3):273–294.

[40] Cambria E., Olsher D., Rajagopal D. SenticNet 3: a common and common-sense knowledge base for cognition-driven sentiment analysis. In: Proceedings of AAAI Conference on Artificial Intelligence. 2014:1515–1521.

[41] Grassi M. Developing HEO human emotions ontology. In: Proceedings of the 2009 Joint COST 2101 and 2102 International Conference on Biometric ID Management and Multimodal Communication. Berlin, Heidelberg: Springer-Verlag; 2009:244–251.

[42] Francisco V., Gervas P., Peinado F. Ontological reasoning for improving the treatment of emotions in text. Knowl. Inform. Syst. 2010;25:421–443.

[43] Sánchez-Rada J.F., Iglesias C.A. Onyx: a linked data approach to emotion representation. Inform. Process. Manage. 2016;52(1):99–114.

[44] Cambria E., Livingstone A., Hussain A. The hourglass of emotions. In: Lecture Notes in Computer Science; Springer; 144–157. COST 2102 Training School, Revised Selected Papers. 2011;7403.

[45] Poria S., Gelbukh A., Hussain A., Howard N., Das D., Bandyopadhyay S. Enhanced SenticNet with affective labels for concept-based opinion mining. IEEE Intell. Syst. 2013;28(2):31–38.

[46] Bertola F., Patti V. Ontology-based affective models to organize artworks in the social semantic web. Inform. Process. Manage. 2016;52(1):139–162.

[47] McCrae J., Spohr D., Cimiano P. Linking lexical resources and ontologies on the semantic web with LEMON. In: New York, NY: Springer; 2011:245–259. The Semantic Web: Research and Applications..

[48] Bosco C., Patti V., Bolioli A. Developing corpora for sentiment analysis: the case of Irony and Senti-TUT. IEEE Intell. Syst. 2013;28(2):55–63.

[49] Steinberger J., Brychcın T., Konkol M. Aspect-level sentiment analysis in czech. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (ACL 2014). 2014:24.

[50] Pang B., Lee L. A sentimental education: sentiment analysis using subjectivity. In: Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics (ACL 2004). 2004:271–278.

[51] Pang B., Lee L. Seeing stars: exploiting class relationships for sentiment categorization with respect to rating scales. In: Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics (ACL 2005). 2005:115–124.

[52] Wiebe J., Wilson T., Cardie C. Annotating expressions of opinions and emotions in language. Lang. Resour. Eval. 2005;39(2–3):165–210.

[53] Deng L., Wiebe J. MPQA 3.0: an entity/event-level sentiment corpus. In: Mihalcea R., Chai J.Y., Sarkar A., eds. Proceedings of HLT-NAACL 2015. The Association for Computational Linguistics; 2015:1323–1328.

[54] Hu M., Liu B. Mining and summarizing customer reviews. In: Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. New York, NY: ACM; 2004:168–177.

[55] Titov I., McDonald R.T. A joint model of text and aspect ratings for sentiment summarization. In: Proceedings of the 46th Annual Meeting of the Association for Computational Linguistics (ACL 2008). 308–316. 2008;8.

[56] Go A., Bhayani R., Huang L. Twitter sentiment classification using distant supervision. CS224N Project Report, Stanford. 2009;1:12.

[57] Mohammad S.M., Zhu X., Kiritchenko S., Martin J. Sentiment, emotion, purpose, and style in electoral tweets. Inform. Process. Manage. 2015;51(4):480–499.

[58] Hagen M., Potthast M., Büchner M., Stein B. Webis: an ensemble for Twitter sentiment detection. In: Proceedings of the Ninth International Workshop on Semantic Evaluation (SemEval 2015). Denver, CO: Association for Computational Linguistics; 2015:582–589.

[59] Günther T., Furrer L. GU-MLT-LT: sentiment analysis of short messages using linguistic features and stochastic gradient descent. In: Proceedings of the Second Joint Conference on Lexical and Computational Semantics (*SEM), Volume 2: Proceedings of the Seventh International Workshop on Semantic Evaluation (SemEval 2013). Atlanta, GA: Association for Computational Linguistics; 2013:328–332.

[60] Miura Y., Sakaki S., Hattori K., Ohkuma T. Teamx: a sentiment analyzer with enhanced lexicon mapping and weighting scheme for unbalanced data. In: Proceedings of the Eighth International Workshop on Semantic Evaluation (SemEval 2014). 2014:628–632.

[61] Saurí R. A Factuality Profiler for Eventualities in Text. Ann Arbor, MI: ProQuest; 2008.

[62] Harris Z. Distributional structure. Word. 1954;10(23):146–162.

[63] Sahlgren M. The distributional hypothesis. Ital. J. Linguist. 2008;20(1):33–54.

[64] Mikolov T., Chen K., Corrado G., Dean J. Efficient estimation of word representations in vector space. In: Proceedings of the ICLR Workshop. 2013:1–12.

[65] Mikolov T., Sutskever I., Chen K., Corrado G.S., Dean J. Distributed representations of words and phrases and their compositionality. In: Advances in Neural Information Processing Systems (NIPS). 2013:3111–3119.

[66] Maas A.L., Daly R.E., Pham P.T., Huang D., Ng A.Y., Potts C. Learning word vectors for sentiment analysis. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies—Volume 1. Stroudsburg, PA: Association for Computational Linguistics; 2011:142–150.

[67] Tang D., Wei F., Yang N., Zhou M., Liu T., Qin B. Learning sentiment-specific word embedding for Twitter sentiment classification. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics. 1555–1565. 2014;1.

[68] Heck L., Huang H. Deep learning of knowledge graph embeddings for semantic parsing of Twitter dialogs. In: Proceedings of the Second IEEE Global Conference on Signal and Information Processing (DRAFT). IEEE Institute of Electrical and Electronics Engineers; 2014:597–601.

[69] Socher R., Perelygin A., Wu J., Chuang J., Manning C.D., Ng A., Potts C. Recursive deep models for semantic compositionality over a sentiment treebank. In: Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing. Seattle, WA: Association for Computational Linguistics; 2013:1631–1642.

[70] Gangemi A., Presutti V., Reforgiato Recupero D. Frame-based detection of opinion holders and topics: a model and a tool. IEEE Comp. Int. Mag. 2014;9(1):20–30.

[71] Cambria E., Schuller B.W., Liu B., Wang H., Havasi C. Knowledge-based approaches to concept-level sentiment analysis. IEEE Intell. Syst. 2013;28(2):12–14.

[72] Saif H., He Y., Fernandez M., Alani H. Contextual semantics for sentiment analysis of Twitter. Inform. Process. Manage. 2016;52(1):5–19.

[73] Poria S., Gelbukh A.F., Agarwal B., Cambria E., Howard N. Common sense knowledge based personality recognition from text. In: Advances in Soft Computing and Its Applications, Proceedings of the 12th Mexican International Conference on Artificial Intelligence, MICAI 2013, Part II. Springer; 2013:484–496.

[74] Hernández Farías D.I., Benedí J., Rosso P. Applying basic features from sentiment analysis for automatic irony detection. In: Lecture Notes in Computer Science; Springer; 337–344. Pattern Recognition and Image Analysis, Proceedings of IbPRIA 2015. 2015;vol. 9117.

[75] Sulis E., Farías D.I.H., Rosso P., Patti V., Ruffo G. Figurative messages and affect in Twitter: Differences between #irony, #sarcasm and #not. Knowledge-Based Systems, Elsevier. 2016. In Press http://dx.doi.org/10.1016/j.knosys.2016.05.035.

[76] Rangel F.M., Rosso P. On the impact of emotions on author profiling. Inform. Process. Manage. 2016;52(1):73–92.


..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset