Twitter has become an important communication channel in times of emergency.
The ubiquitousness of smartphones enables people to announce an emergency they’re observing in real-time.
Because of this, more agencies are interested in programatically monitoring Twitter (i.e. disaster relief organizations and news agencies).
-
Total approach towards the project can be seen on kaggle
- Machine Learning approach : https://www.kaggle.com/mohitnirgulkar/disaster-tweets-classification-using-ml
- Deep Learning approach : https://www.kaggle.com/mohitnirgulkar/disaster-tweets-classification-using-deep-learning
- Exploratory Data Analysis
- EDA after Data Cleaning
- Data Preprocessing using NLP
- Machine Learning models for classifying Tweets data
- Deep Learning approach for classifying Tweets data
- Model Deployment
- Packages : Pandas, Numpy, Matplotlib, Plotly, Word-cloud, Tensorflow, Scikit-Learn, Keras, Keras-tuner, Nltk etc.
- Dataset : https://www.kaggle.com/c/nlp-getting-started
- Word Embeddings : https://www.kaggle.com/danielwillgeorge/glove6b100dtxt
-
Visualising Target Variable of the Dataset
-
Visualising Length of Tweets
-
Visualising Average word lengths of Tweets
-
Visualising most common stop words in the text data
-
Visualising most common punctuations in the text data
-
We use Python Regex library and nltk lemmatizing methods for Data Cleaning
-