-
Stability.ai, Eleuther.ai
- Seattle, WA
- http:https://dmarx.github.io
- @DigThatData
Block or Report
Block or report dmarx
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseML Video
PySlowFast: video understanding codebase from FAIR for reproducing state-of-the-art video models.
Code + pre-trained models for the paper Keeping Your Eye on the Ball Trajectory Attention in Video Transformers
The official pytorch implementation of our paper "Is Space-Time Attention All You Need for Video Understanding?"
"Object-Region Video Transformers”, Herzig et al., CVPR 2022
[CVPR2022] Bridge-Prompt: Towards Ordinal Action Understanding in Instructional Videos
A JAX implementation of Broaden Your Views for Self-Supervised Video Learning, or BraVe for short.
Align and Prompt: Video-and-Language Pre-training with Entity Prompts
Large-scale text-video dataset. 10 million captioned short videos.
[ICLR 2022] TAda! Temporally-Adaptive Convolutions for Video Understanding. This codebase provides solutions for video classification, video representation learning and temporal detection.
Video Autoencoder: self-supervised disentanglement of 3D structure and motion (ICCV 2021). Website: https://zlai0.github.io/VideoAutoencoder/
Frozen in Time: A Joint Video and Image Encoder for End-to-End Retrieval [ICCV'21]
Repository for "Space-Time Correspondence as a Contrastive Random Walk" (NeurIPS 2020)
Official code for "Towards An End-to-End Framework for Flow-Guided Video Inpainting" (CVPR2022)
Implementation of NÜWA, state of the art attention network for text to video synthesis, in Pytorch
This is an official implementation for "Video Swin Transformers".
OpenMMLab's Next Generation Video Understanding Toolbox and Benchmark
Text-to-video generation. The repo for ICLR2023 paper "CogVideo: Large-scale Pretraining for Text-to-Video Generation via Transformers"
An end-to-end PyTorch framework for image and video classification
[ICCV 2019] TSM: Temporal Shift Module for Efficient Video Understanding
A deep learning library for video understanding research.
Tooll 3 is an open source software to create realtime motion graphics.
This is the repo for my experiments with StyleGAN2. There are many like it, but this one is mine. Contains code for the paper Audio-reactive Latent Interpolations with StyleGAN.
The code for ECCV22 paper "Fast-Vid2Vid: Spatial-Temporal Compression for Video-to-Video Synthesis"
Official PyTorch implementation of TATS: A Long Video Generation Framework with Time-Agnostic VQGAN and Time-Sensitive Transformer (ECCV 2022)
A High-performance cross-platform Video Processing Python framework powerpacked with unique trailblazing features 🔥
Python bindings for FFmpeg - with complex filtering support
This repository contains the source code for the paper First Order Motion Model for Image Animation