Skip to content

drexelai/pose-dection-nets

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 

Repository files navigation

Pose Detection:

Drexel AI 2023 Spring Project

This project is discontinued

Introduction:

Background

https://arxiv.org/pdf/1808.09568.pdf

Methods

Data

https://drexel0-my.sharepoint.com/:u:/g/personal/az548_drexel_edu/Ee-4nyzqx75JurbGzTv246EB_bcG4-A9827G9AwMuT5gyQ?e=Laa7Ls

https://cydar.ist.psu.edu/emotionchallenge/index.php

Models:

  • YOLO, MoveNet, PoseNet (Getting pose coordinates)
  • Transformer models (classification task)

Tasks:

April 17th

  • Literature review of current methods -> Outline important findings (Everyone)
    • Pose classification
    • Pose estimation
  • Understanding the BoLD dataset (Alex)
  • Look through examples of pose estimation and classification (Arjit, Bobby, Alisha)

April 24th

  • Steven will run Alex's model for pose estimation
  • Alisha will run MoveNet model for pose estimation
  • Look into Omnipose and RSN (figure out if it's pretrained, look into features) -> Yashoda
  • Multi-person pose estimation? -> Arijit
  • Alex will look into MediaPipe and more detailed pose estimation

Potential models for classification

  • Temporal CNN
  • Transformer

For next meeting:

  • Compare outputs from each of the models
  • If outputs are good, move onto classification

May 4th

Important: for every v that is 0, xyz are also 0

1d convolution on 2d grid to shrink temporal dimension to 1x1xfeature maps -> spacial convolution ->feedforward network

also could use attention mechanism & transformer instead of spacial

seperate vs sequential convolutions for temporal and spatial

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published