Skip to content
View teowu's full-sized avatar
🎯
Focusing
🎯
Focusing
  • Nanyang Technological University
  • 50 Nanyang Avenue, Singapore
  • 02:17 (UTC +08:00)
  • X @HaoningTimothy

Organizations

@VQAssessment @Q-Future

Block or report teowu

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
teowu/README.md
  • 👋 Hi, I’m Teo (Timothy) Wu, Final Year PhD Candidate in Nanyang Technological University 🇸🇬, Resume, Homepage
  • I am working on Large Multi-modal Models (LMMs), especially on videos and other long-form multi-modal inputs. See our Neurips 2024 D&B, LongVideoBench for the first-ever benchmark designed for long-context video-text interleaved inputs!

Prior to working on general LMMs, I am also the creator of Q-Future, a project that aims to utilize LMMs to boost low-level vision, visual evaluation, and related topics. Here are the two representative works:

  • Q-Align (Visual Scorer): Open in Huggingface Spaces
  • Co-Instruct (Pro-level Low-level Vision-Language Assistant/Chatbot): Open in Huggingface Spaces

See my top Repos:

Pinned Loading

  1. longvideobench/LongVideoBench longvideobench/LongVideoBench Public

    [Neurips 24' D&B] Official Dataloader and Evaluation Scripts for LongVideoBench.

    Python 58 2

  2. Q-Future/A-Bench Q-Future/A-Bench Public

    [LMM + AIGC] What do we expect from LMMs as AIGI evaluators and how do they perform?

    136 3

  3. Q-Future/Co-Instruct Q-Future/Co-Instruct Public

    ④[ECCV 2024 Oral, Comparison among Multiple Images!] A study on open-ended multi-image quality comparison: a dataset, a model and a benchmark.

    64 4

  4. Q-Future/Q-Align Q-Future/Q-Align Public

    ③[ICML2024] [IQA, IAA, VQA] All-in-one Foundation Model for visual scoring. Can efficiently fine-tune to downstream datasets.

    Python 250 16

  5. Q-Future/Q-Instruct Q-Future/Q-Instruct Public

    ②[CVPR 2024] Low-level visual instruction tuning, with a 200K dataset and a model zoo for fine-tuned checkpoints.

    Python 190 8

  6. Q-Future/Q-Bench Q-Future/Q-Bench Public

    ①[ICLR2024 Spotlight] (GPT-4V/Gemini-Pro/Qwen-VL-Plus+16 OS MLLMs) A benchmark for multi-modality LLMs (MLLMs) on low-level vision and visual quality assessment.

    Jupyter Notebook 238 12