Skip to content

mfriebel/tweet_bot

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Tweets Collection and Analysis Pipeline

This project implements a data pipeline using Docker. Tweets are streamed about sustainability are streamed via Tweepy Listener and stored in a MongoDB database. The ETL job performs live sentiment analysis (using VADER) on the stored tweets and loads them with the according score into a PostgreSQL database. In the end tweets with most positive sentiment are posted on Slack using a Webhook.

Pipeline

Docker Compose necessities

Setting up local environmental variables

Changing streaming filter

The file get_tweets.py contains the Tweepy Tweets Listener. The topic filter is found at the end of the file: stream.filter(track=['sustainable'], languages=['en'])

About

This project implements a data pipeline using Docker to extract, transform (sentiment analysis) and load twitter posts and finally post them on Slack.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published