OSMO: Open-vocabulary Self-eMOtion Tracking
OSMO introduces egocentric self-emotion tracking with a large-scale dataset, a multi-task benchmark, and OSIRIS, a multimodal model for coherent emotion timelines.
I enjoy making things. Here are a selection of projects that I have worked on over the years.
OSMO introduces egocentric self-emotion tracking with a large-scale dataset, a multi-task benchmark, and OSIRIS, a multimodal model for coherent emotion timelines.
OSKAR is a self-supervised multimodal foundation model that learns in the latent space by predicting masked multimodal features.
MaskCLR improves the robustness of transformer-based action recognition methods against noisy and incomplete skeletons.
S-JEPA is an instantiation of JEPA for self-supervised skeletal action recognition.
We take a step towards computer-aided waste detection and present the first in-the-wild industrial-grade waste detection and segmentation dataset, ZeroWaste.