Search results
(21 - 24 of 24)
Pages
- Title
- Data used to develop #Polar scores
- Creator
- Culotta, Aron, Hemphill, Libby, Heston, Matthew
- Date
- 2013, 2016
- Description
-
We present a new approach to measuring political polarization, including a novel algorithm and open source Python code, which leverages...
Show moreWe present a new approach to measuring political polarization, including a novel algorithm and open source Python code, which leverages Twitter content to produce measures of polarization for both users and hashtags. #Polar scores provide advantages over existing measures because they (1) can be calculated throughout the legislative cycle, (2) allow for easy differentiation between users with similar scores, (3) are chamber-agnostic, and (4) are a generic approach that can be applied beyond the U.S. Congress. #Polar scores leverage available information such as party labels, word frequency, and hashtags to create an accessible, straightforward algorithm for estimating polarity using text. (from the paper: Hemphill, L., Culotta, A., and Heston, M. (forthcoming) #Polar Scores: Measuring partisanship using social media content. Journal of Information Technology & Politics.)
The dataset contains one plain text TSV file with the following information for each of the 55,244 tweets used to develop #Polar scores : tweet_id, created_at, user_id, screen_name, tag, shortid, sex, party, state, chamber, name. The file contains one row per hashtag, and therefore tweets may appear more than once. The Python code for calculating #Polar scores is available here: http://doi.org/10.5281/zenodo.53888
Show less
- Title
- Descriptors and measurements of verbal violence in tweets: toxicity README
- Creator
- Guberman, Joshua, Hemphill, Libby
- Date
- 2016, 2016
- Description
-
Sponsorship: National Science Foundation Award # 1525662
The purpose of this data collection was to test a scale for detecting verbal violence...
Show moreSponsorship: National Science Foundation Award # 1525662
The purpose of this data collection was to test a scale for detecting verbal violence in Tweets. Workers at Mechanical Turk were first asked to complete a qualification test and then invited to code additional Tweets according to our scale. The qualification test involved a detailed explanation of each item of the scale, a walkthrough of a tweet that we had coded according to all 14 scale-items, a practice exercise, and a test. In the practice exercise, potential coders attempted to code a tweet on their own using our scale. After submitting their ratings, they were shown our own ratings for the same tweet and explanations for each of our ratings. The test component consisted of another coding task, in which coders were asked to code another tweet that we had already coded ourselves. The workers who, on test, with our ratings of that tweet on at least 11 out of the 14 items “passed” the test, earning the qualification that allowed them to participate in future coding tasks. Variables in the data include the ID of the Tweet (so that you may find it on Twitter; Twitter Terms of Service prohibit us from sharing the Tweets), the ID number we assigned to the coder, the rating that coder provided for each of the 14 items on our scale, the gender and age of the coder, and any comments the coder provided. APA (6th Edition) CITATION: Guberman, J. and Hemphill, L. (2016) Descriptors and measurements of verbal violence in tweets [data file and codebook]. doi: 10.6084/m9.figshare.3179368.
Show less
- Title
- Descriptors and measurements of verbal violence in tweets
- Creator
- Guberman, Joshua, Hemphill, Libby
- Date
- 2016, 2016
- Description
-
Sponsorship: National Science Foundation Award # 1525662
The purpose of this data collection was to test a scale for detecting verbal violence...
Show moreSponsorship: National Science Foundation Award # 1525662
The purpose of this data collection was to test a scale for detecting verbal violence in Tweets. Workers at Mechanical Turk were first asked to complete a qualification test and then invited to code additional Tweets according to our scale. The qualification test involved a detailed explanation of each item of the scale, a walkthrough of a tweet that we had coded according to all 14 scale-items, a practice exercise, and a test. In the practice exercise, potential coders attempted to code a tweet on their own using our scale. After submitting their ratings, they were shown our own ratings for the same tweet and explanations for each of our ratings. The test component consisted of another coding task, in which coders were asked to code another tweet that we had already coded ourselves. The workers who, on test, with our ratings of that tweet on at least 11 out of the 14 items “passed” the test, earning the qualification that allowed them to participate in future coding tasks. Variables in the data include the ID of the Tweet (so that you may find it on Twitter; Twitter Terms of Service prohibit us from sharing the Tweets), the ID number we assigned to the coder, the rating that coder provided for each of the 14 items on our scale, the gender and age of the coder, and any comments the coder provided. APA (6th Edition) CITATION: Guberman, J. and Hemphill, L. (2016) Descriptors and measurements of verbal violence in tweets [data file and codebook]. doi: 10.6084/m9.figshare.3179368.
Show less
- Title
- Descriptors and measurements of verbal violence in tweets: toxicity_first_steps
- Creator
- Guberman, Joshua, Hemphill, Libby
- Date
- 2016, 2016
- Description
-
Sponsorship: National Science Foundation Award # 1525662
The purpose of this data collection was to test a scale for detecting verbal violence...
Show moreSponsorship: National Science Foundation Award # 1525662
The purpose of this data collection was to test a scale for detecting verbal violence in Tweets. Workers at Mechanical Turk were first asked to complete a qualification test and then invited to code additional Tweets according to our scale. The qualification test involved a detailed explanation of each item of the scale, a walkthrough of a tweet that we had coded according to all 14 scale-items, a practice exercise, and a test. In the practice exercise, potential coders attempted to code a tweet on their own using our scale. After submitting their ratings, they were shown our own ratings for the same tweet and explanations for each of our ratings. The test component consisted of another coding task, in which coders were asked to code another tweet that we had already coded ourselves. The workers who, on test, with our ratings of that tweet on at least 11 out of the 14 items “passed” the test, earning the qualification that allowed them to participate in future coding tasks. Variables in the data include the ID of the Tweet (so that you may find it on Twitter; Twitter Terms of Service prohibit us from sharing the Tweets), the ID number we assigned to the coder, the rating that coder provided for each of the 14 items on our scale, the gender and age of the coder, and any comments the coder provided. APA (6th Edition) CITATION: Guberman, J. and Hemphill, L. (2016) Descriptors and measurements of verbal violence in tweets [data file and codebook]. doi: 10.6084/m9.figshare.3179368.
Show less