-
University of Surrey
- https://www.linkedin.com/in/thomas-ng-kam-woh/
Highlights
- Pro
Lists (1)
Sort Name ascending (A-Z)
Stars
Multi-object image datasets with ground-truth segmentation masks and generative factors.
Stable Virtual Camera: Generative View Synthesis with Diffusion Models
This repo contains the code for our paper Compositor: Bottom-Up Clustering and Compositing for Robust Part and Object Segmentation
[ICLR'25] Official Implementation for Consistent Flow Distillation for Text-to-3D Generation
The implementation of "Prismer: A Vision-Language Model with Multi-Task Experts".
InstantSplat: Sparse-view SfM-free Gaussian Splatting in Seconds
Official repo for paper "Structured 3D Latents for Scalable and Versatile 3D Generation" (CVPR'25).
Official repository for our work on micro-budget training of large-scale diffusion models.
HaMeR: Reconstructing Hands in 3D with Transformers
Official implementation of FIFO-Diffusion: Generating Infinite Videos from Text without Training (NeurIPS 2024)
A diffuser implementation of Zero123. Zero-1-to-3: Zero-shot One Image to 3D Object (ICCV23)
Avatar Generation For Characters and Game Assets Using Deep Fakes
Official Implementation of paper "MonST3R: A Simple Approach for Estimating Geometry in the Presence of Motion"
🐸💬 - a deep learning toolkit for Text-to-Speech, battle-tested in research and production
Tencent Hunyuan3D-1.0: A Unified Framework for Text-to-3D and Image-to-3D Generation
[SIGGRAPH'24] 2D Gaussian Splatting for Geometrically Accurate Radiance Fields
Code for Paper 'Redefining Temporal Modeling in Video Diffusion: The Vectorized Timestep Approach'
Hallo2: Long-Duration and High-Resolution Audio-driven Portrait Image Animation
[CVPR 2025] EchoMimicV2: Towards Striking, Simplified, and Semi-Body Human Animation
TerDiT: Ternary Diffusion Models with Transformers
A lightweight tool for camera pose visualization
High-resolution models for human tasks.
ECCV2020 paper "Whole-Body Human Pose Estimation in the Wild"
Official Pytorch Implementation for "MultiDiffusion: Fusing Diffusion Paths for Controlled Image Generation" presenting "MultiDiffusion" (ICML 2023)
[ECCV 2024] MOFA-Video: Controllable Image Animation via Generative Motion Field Adaptions in Frozen Image-to-Video Diffusion Model.
Inpaint anything using Segment Anything and inpainting models.