Skip to content

Simulates a simple game of Dead by Daylight and streams the data to S3 via Apache Firehose. Visualized in AWS QuickSight.

Notifications You must be signed in to change notification settings

MPierre9/DBD_Game_Sim

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Dead By Daylight Simple Game Data Pipeline

Summary: This repo gives a brief overview of the process, architecture, and various components used to make a simple Dead By Daylight (DBD) game data stream. This was created primarily for learning purposes. In the simulated data we try to show a correlation between a players play time and the amount of kills they can get (higher play time more likely).

Day one: alt text As more data streams (stopped intermittently because of AWS Cost): alt text alt text

Components used

MySQL Database

This DB contains two tables USER and KILLER. The game simulator Python script will query the Database (or AWS Elasticache if the results are cached) to create a match. The DB contains 1000 users and 18 killers. alt text

AWS EC2

Two EC2 instances were created. One to run our backed Python scrip to simulate a game and a second for our MySQL DB instance. alt text

Docker

To test scalability Docker containers were used to simulate multiple instances of the data being generated. alt text

Steps involved

  1. Installing Docker on RHEL EC2 instance
  2. docker pull centos
  3. docker run centos
  4. docker run -it -d centos
  5. docker exec -it container_id /bin/bash
  6. yum install update -y
  7. yum install python3
  8. yum install nano
  9. yum install passwd
  10. useradd rm_log our service account user for MySQL
  11. passwd rm_log setting up user password
  12. su - rm_log
  13. Setup AWS Credentials
  14. mkdir ~/.aws
  15. nano ~/.aws/credentials
  16. Adding:
[default]
aws_access_key_id = YOUR_KEY
aws_secret_access_key = YOUR_SECRET
  1. Setup AWS Region Config
  2. nano ~/.aws/config
  3. Adding:
[default]
region=us-east-1
  1. Run curl icanhazip.com to get public IP and add to AWS security group for port 3306
  2. Grab https://github.com/MPierre9/DBD_Game_Sim/blob/master/dbd_game_sim.py and test in container
  3. Execute chmod +x dbd_game_sim.py
  4. Exit container and commit changes docker commit container_id centos_stream_1
  5. With the centos_stream_1 image create the desired amount of instances and execute.
  6. Execute `nohup python3 -u ./dbd_game_sim.py > ~/dbd_game_sim.log
  7. Ensure process is running in background with ps ax
  8. To kill run pkill -f dbd_game_sim.py

Checking Container Data Log alt text

Future state, creating a Dockerfile to has libraries and packages pre-installed

AWS Elasticache

The Elasticache cache is used to cache results from the game sim Python script to increase query performance and decrease DB strain. The idea is if multiple game sims are running then they could work off the cached player list and killers. alt text

AWS Kinesis Firehose

Once a game is complete two streams are sent via AWS Kinesis Firehose to be stored in S3. alt text

Player Stream

{"username": "cmenhenitthv", "play_time": 25, "killer": "", "killed": true, "match_id": "14a9b218-cc99-428e-932a-872acf357f99-02-03-2020-00-46-38"} {"username": "dinnettrm", "play_time": 70, "killer": "", "killed": true, "match_id": "14a9b218-cc99-428e-932a-872acf357f99-02-03-2020-00-46-38"} {"username": "dbaskeyfield87", "play_time": 93, "killer": "", "killed": true, "match_id": "14a9b218-cc99-428e-932a-872acf357f99-02-03-2020-00-46-38"} {"username": "jtabert6g", "play_time": 47, "killer": "the demogorgon", "killed": false, "match_id": "14a9b218-cc99-428e-932a-872acf357f99-02-03-2020-00-46-38"} {"username": "jantoniouez", "play_time": 148, "killer": "", "killed": true, "match_id": "14a9b218-cc99-428e-932a-872acf357f99-02-03-2020-00-46-38"}

Match Stream

{"match_id": "00778193-85d9-4256-9f52-e2c229eebcd0-02-03-2020-00-46-08", "match_time": "02-03-2020-00-46-08"}

AWS Lambda

A lambda function called BHVR_TRANSFORM decodes the base64 data to ascii and adds a newline character to make the JSON records look a bit nicer.

console.log('Loading function');
exports.handler = async (event, context, callback) => {
    /* Process the list of records and transform them */
    const output = event.records.map((record) => {
        /* This transformation is the "identity" transformation, the data is left intact */
        let entry = (new Buffer(record.data, 'base64')).toString('ascii');
        let result = entry + "\n";
        const payload = (new Buffer(result, 'ascii')).toString('base64');
        
        return {
            recordId: record.recordId,
            result: 'Ok',
            data: payload
        };
    });
    console.log('Processing completed.  Successful records ${output.length}.');
    callback(null,{ records: output });
};

AWS S3

AWS S3 is the destination for our JSON records streaming via Kinesis Firehose. There are two directories in S3 for the data. dbd_match_id which contains all the match_ids and the date the match occurred. dbd_player_match which contains all the player data for that match (survivor info, who was the killer, who did they kill, etc). alt text

AWS Glue

Glue is the best. In Glue we have two Crawlers BHVR_CRAWLER and BHVR_MATCH_ID_CRAWLER which connect to S3 to poll for additional match data and add any new data to tables dbd_match_id and dbd_player_match.

alt text

AWS Athena

Used to query the tables we have in the Data Catalog from AWS Glue. From there a pretty query was made to see the kill count for each respective killer in a match

select username, killer, play_time, (select count(*) from "bhvr"."dbd_player_match" where killed and match_id=b.match_id) as kill_count from "bhvr"."dbd_player_match" as a INNER JOIN "bhvr"."dbd_match_id" as b ON a.match_id=b.match_id WHERE a.match_id=b.match_id and killer <> '' GROUP BY username, killer, play_time, b.match_id

AWS Quicksight

Finally this data is visualized in AWS Quicksight. alt text

About

Simulates a simple game of Dead by Daylight and streams the data to S3 via Apache Firehose. Visualized in AWS QuickSight.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published