Author: Alshaabi, Thayer; Oort, Colin Van; Fudolig, Mikaela; Arnold, Michael V.; Danforth, Christopher M.; Dodds, Peter Sheridan
Title: Augmenting semantic lexicons using word embeddings and transfer learning Cord-id: ods7yu18 Document date: 2021_9_18
ID: ods7yu18
Snippet: Sentiment-aware intelligent systems are essential to a wide array of applications including marketing, political campaigns, recommender systems, behavioral economics, social psychology, and national security. These sentiment-aware intelligent systems are driven by language models which broadly fall into two paradigms: 1. Lexicon-based and 2. Contextual. Although recent contextual models are increasingly dominant, we still see demand for lexicon-based models because of their interpretability and
Document: Sentiment-aware intelligent systems are essential to a wide array of applications including marketing, political campaigns, recommender systems, behavioral economics, social psychology, and national security. These sentiment-aware intelligent systems are driven by language models which broadly fall into two paradigms: 1. Lexicon-based and 2. Contextual. Although recent contextual models are increasingly dominant, we still see demand for lexicon-based models because of their interpretability and ease of use. For example, lexicon-based models allow researchers to readily determine which words and phrases contribute most to a change in measured sentiment. A challenge for any lexicon-based approach is that the lexicon needs to be routinely expanded with new words and expressions. Crowdsourcing annotations for semantic dictionaries may be an expensive and time-consuming task. Here, we propose two models for predicting sentiment scores to augment semantic lexicons at a relatively low cost using word embeddings and transfer learning. Our first model establishes a baseline employing a simple and shallow neural network initialized with pre-trained word embeddings using a non-contextual approach. Our second model improves upon our baseline, featuring a deep Transformer-based network that brings to bear word definitions to estimate their lexical polarity. Our evaluation shows that both models are able to score new words with a similar accuracy to reviewers from Amazon Mechanical Turk, but at a fraction of the cost.
Search related documents:
Co phrase search for related documents- absolute error and long short: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25
- academic research and long short: 1, 2, 3, 4
- accuracy improve and adequately represent: 1
- accuracy improve and local context: 1
- accuracy improve and long short: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13
- local context and long short: 1, 2
Co phrase search for related documents, hyperlinks ordered by date