forked from derekgreene/twitter-jsonl-tools
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathjsonl-tweet-hashtags.py
executable file
·100 lines (87 loc) · 3.52 KB
/
jsonl-tweet-hashtags.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
#!/usr/bin/env python
"""
Provide list of the most frequently-appearing hashtags for a JSONL file, where each line contains a JSON-formatted tweet
as retrieved from the Twitter API.
Sample usage:
python jsonl-tweet-hashtags.py sample/sample-tweets-500.jsonl
"""
import sys, fileinput, operator
from datetime import datetime
from optparse import OptionParser
from collections import defaultdict
import logging as log
try:
import ujson as json
except:
import json
from prettytable import PrettyTable
# --------------------------------------------------------------
def parse_twitter_date( s, ignore_time_zones = True ):
# hack for cases where timezone is not supported by Python strptime
if ignore_time_zones:
parts = s.split(" ")
smodified =" ".join( parts[0:4] + [ parts[-1] ] )
return datetime.strptime(smodified,'%a %b %d %H:%M:%S %Y')
return datetime.strptime(s,'%a %b %d %H:%M:%S %z %Y')
def main():
parser = OptionParser(usage="usage: %prog [options] json_file1 json_file2 ...")
parser.add_option("-t", "--top", action="store", type="int", dest="top", help="number of top hashtags to display", default=10)
parser.add_option("--start_date", action="store", type="string", dest="start_date", help="separator character for output file (default is comma)", default="")
parser.add_option("--end_date", action="store", type="string", dest="end_date", help="separator character for output file (default is comma)", default="")
(options, args) = parser.parse_args()
if( len(args) < 1 ):
parser.error( "Must specify at least one JSONL file" )
log.basicConfig(level=20, format='%(message)s')
if options.start_date:
start_date = datetime.strptime(options.start_date, '%Y-%m-%d')
else:
start_date = None
if options.end_date:
end_date = datetime.strptime(options.end_date, '%Y-%m-%d')
else:
end_date = None
for tweets_path in args:
log.info("Loading tweets from %s ..." % tweets_path)
# Process every line as JSON data
hashtags = {}
num_tweets, num_failed, line_number, has_hashtags = 0, 0, 0, 0
counts = defaultdict(int)
for l in fileinput.input(tweets_path):
l = l.strip()
if len(l) == 0:
continue
try:
line_number += 1
tweet = json.loads(l)
sdate = parse_twitter_date(tweet["created_at"])
# If start and end date are specified and tweet date is out of this range - ignore this tweet
if (start_date and sdate < start_date) or (end_date and sdate > end_date):
continue
if "entities" in tweet:
if "hashtags" in tweet["entities"] and len(tweet["entities"]["hashtags"]) > 0:
has_hashtags += 1
for tag in tweet["entities"]["hashtags"]:
text = "#" + tag["text"].lower().strip()
counts[text] += 1
num_tweets += 1
if line_number % 50000 == 0:
log.info("Processed %d lines" % line_number)
except Exception as e:
log.error("Failed to parse tweet on line %d: %s" % ( line_number, e ) )
num_failed += 1
fileinput.close()
log.info("Found %d/%d tweets containing at least one hashtag. File has %d distinct hashtags" % ( has_hashtags, num_tweets, len(counts) ) )
# Display the top hashtags for this file
sx = sorted(counts.items(), key=operator.itemgetter(1), reverse=True)
log.info("Top %d hashtags appearing in tweets:" % min( len(sx), options.top ) )
tab = PrettyTable( ["Hashtag", "Count"] )
tab.align["Hashtag"] = "l"
tab.align["Count"] = "r"
for i, pair in enumerate(sx):
if i > options.top:
break
tab.add_row( pair )
log.info(tab)
# --------------------------------------------------------------
if __name__ == "__main__":
main()