TimeLMs allows for easy access to models continuously trained on social media over regular intervals for researching language model degradation, as well as cultural shifts affecting language usage on social media.
Paper (ACL-2022 Demo): https://arxiv.org/pdf/2202.03829.pdf
Below we provide instructions for getting started with TimeLMs and a few usage examples. For a more detailed guide, please see our notebook demo. The demo is also available as a Google Colab notebook.
TimeLMs has also been integrated into the TweetNLP library.
List of models released in the scope of TimeLMs (available through the Hugging Face hub):
Model Name | # Tweets | Last Date | Link | Verified Users Set |
---|---|---|---|---|
twitter-roberta-base-2019-90m | 90.26M | 2019-12-31 | Hub Link | v310821 |
twitter-roberta-base-mar2020 | 94.46M | 2020-03-31 | Hub Link | v310821 |
twitter-roberta-base-jun2020 | 98.66M | 2020-06-30 | Hub Link | v310821 |
twitter-roberta-base-sep2020 | 102.86M | 2020-09-30 | Hub Link | v310821 |
twitter-roberta-base-dec2020 | 107.06M | 2020-12-31 | Hub Link | v310821 |
twitter-roberta-base-mar2021 | 111.26M | 2021-03-31 | Hub Link | v310821 |
twitter-roberta-base-jun2021 | 115.46M | 2021-06-30 | Hub Link | v310821 |
twitter-roberta-base-sep2021 | 119.66M | 2021-09-30 | Hub Link | v310821 |
twitter-roberta-base-dec2021 | 123.86M | 2021-12-31 | Hub Link | v310821 |
twitter-roberta-base-2021-124m | 123.86M | 2021-12-31 | Hub Link | v310821 |
twitter-roberta-base-mar2022 | 128.06M | 2022-03-31 | Hub Link | v050422 |
twitter-roberta-base-jun2022 | 132.26M | 2022-06-30 | Hub Link | v050422 |
twitter-roberta-base-mar2022-15M-incr | 138.86M | 2022-03-31 | Hub Link | v050422 |
twitter-roberta-base-jun2022-15M-incr | 153.86M | 2022-06-30 | Hub Link | v050422 |
twitter-roberta-base-sep2022 | 168.86M | 2022-09-30 | Hub Link | v050422 |
twitter-roberta-base-2022-154m | 154M | 2022-12-31 | Hub Link | v091122 |
twitter-roberta-large-2022-154m | 154M | 2022-12-31 | Hub Link | v091122 |
You may create a new environment using conda and install dependencies following the commands below. We assume you already have PyTorch with CUDA support installed (tested with torch==1.8.2+cu111 and CUDA 11.2).
$ conda create -n timelms python=3.7
$ conda activate timelms
$ pip install -r requirements.txt
You can load our interface simply with these two lines, importing the TimeLMs class from the timelms.py file in this repository.
from timelms import TimeLMs
tlms = TimeLMs(device='cuda:0')
TimeLMs currently supports the following temporal modes for determining which models are employed for different tweets.
- 'latest': using our most recently trained Twitter model.
- 'YYYY-MM' (custom): using the model closest to a custom date provided by the user (e.g., '2020-11').
- 'corresponding': using the model that was trained only until to each tweet's date (i.e., its specific quarter).
- 'quarterly': using all available models trained over time in quarterly intervals.
The corresponding
mode requires tweets with a created_at
field with dates under any format that begins with YYYY-MM.
tweets = [{'text': 'She is pure heart #SanaTheBBWinner', 'created_at': '2020-02-09T05:55:00.000Z'},
{'text': 'Looking forward to watching Squid Game tonight !', 'created_at': '2021-10-11T12:34:56.000Z'}]
pseudo_ppls = tlms.get_pseudo_ppl(tweets, mode='corresponding')
To get pseudo-perplexity (PPPL) scores for a set of tweets, you just need to pass a list of tweets to tlms.get_pseudo_ppl()
, specifying your desired mode. Depending on the chosen mode, you'll get a score from each applicable model (2 models for this example). Besides PPPL scores by model, this method also returns the input tweets with their specific pseudo-log likelihood (PLL) values.
tweets = [{"text": "So glad I'm <mask> vaccinated ."},
{"text": "Looking forward to watching <mask> Game tonight !"}]
preds = tlms.get_masked_predictions(tweets, mode='quarterly', top_k=3)
To get masked predictions using our models, you just need to pass a list of tweets to tlms.get_masked_predictions()
, specifying your desired mode and number of predictions.
In the example above, we're choosing the quarterly
mode, which does not require date fields.
tlms.eval_model('roberta-base', 'data/tweets/tweets-2020-2021-subset-rnd.jl')
We provide a method for evaluating other models supported by the Transformers package using PPPL. For evaluating over the periods of 2020 to 2021, we recommend retrieving the tweets used for our evaluation (we provide tweet ids here), or using the 50K subset provided in this repository as an alternative. For the time being, we only support models based on RoBERTa (most Twitter LMs).
Below you find instructions for using our scripts to retrieve and preprocess Twitter data. The same scripts were used for obtaining our training and testing corpora for TimeLMs.
$ python scripts/sampler_api.py 2020 01 35 # <YYYY> <MM> <MIN_MARK>
A generic sample of tweets from the Twitter API can be retrieved using the sampler_api.py script. By 'generic' we mean tweets that are not targetting any specific content (we use stopwords as query terms, more details in the paper).
The MIN_MARK variable is the specific minute passed to the API request. You should set this value according to your preference for the time difference between requests. In our paper, we used several calls to this script in increments of 5 minutes.
This script retrieves tweets for every hour of every day of the given YYYY-MM at the specified MIN_MARK.
Every response is stored as its own file in data/responses
. Requests for files already in that folder will be skipped.
Requires the API BEARER_TOKEN available as an environment variable. You can set that up with:
$ export 'BEARER_TOKEN'='<your_bearer_token>'
The script is set up to wait 7 seconds between requests of 500 results. In case of error, the script waits another 60 seconds before retrying (and increments time between requests by 0.01 seconds).
$ python scripts/combine.py tweets-2020-Q3.jl 2020-01 2020-02 2020-03 # <output_file> <months:YYYY-MM>
After populating data/responses
with tweets retrieved from the API, you can use the combine.py script to combine those responses into a single .jl file restricted to tweets for specified year-months.
This script also merges metrics and location info so that all data pertaining to a particular tweet is contained in a single-line JSON entry of the output .jl file.
You may specify any number of YYYY-MMs. If none are provided, the script will use all available tweets.
$ python scripts/preprocess.py --src tweets-2020-Q3.jl --out tweets-2020-Q3.cleaned.jl
Finally, the merged .jl file can be preprocessed using the preprocess.py script. This step requires the following additional packages:
$ pip install datasketch==1.5.3
$ pip install xxhash==2.0.2
This script removes duplicates, near duplicates and tweets from most frequent users (likely bots, details in the paper) besides replacing user mentions with '@user' for anonymization, except for popular users (i.e., verified users).
The set of verified users was determined using the get_verified.py script, producing the verified_users.v310821.txt file shared with this repository (and more recent versions).
If you use TimeLMs in your research, please use the following bib entry to cite the reference paper.
@inproceedings{loureiro-etal-2022-timelms,
title = "{T}ime{LM}s: Diachronic Language Models from {T}witter",
author = "Loureiro, Daniel and
Barbieri, Francesco and
Neves, Leonardo and
Espinosa Anke, Luis and
Camacho-collados, Jose",
booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: System Demonstrations",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.acl-demo.25",
doi = "10.18653/v1/2022.acl-demo.25",
pages = "251--260"
}
TimeLMs is released without any restrictions, but our scoring code is based on the https://github.com/awslabs/mlm-scoring repository, which is distributed under Apache License 2.0. We also refer users to Twitter regulations regarding use of our models and test sets.