-
MBZUAI
- Abu Dhabi, UAE
- https://muzairkhattak.github.io/
Highlights
- Pro
Block or Report
Block or report muzairkhattak
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
FLAIR: A Foundation LAnguage-Image model of the Retina for fundus image understanding.
Official code for Goldfish model for long video understanding and MiniGPT4-video for short video understanding
Official release of InternLM2.5 base and chat models. 1M context support
Official inference repo for FLUX.1 models
What do we learn from inverting CLIP models?
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use th…
Advanced AI Explainability for computer vision. Support for CNNs, Vision Transformers, Classification, Object detection, Segmentation, Image similarity and more.
Hacking sklearn's t-SNE implementation to animate embedding process
Official repository of Evolutionary Optimization of Model Merging Recipes
Robust recipes to align language models with human and AI preferences
Safe-CLIP: Removing NSFW Concepts from Vision-and-Language Models. ECCV 2024
CUBE is a benchmark to evaluate the Cultural Competence of T2I models
VidProM: A Million-scale Real Prompt-Gallery Dataset for Text-to-Video Diffusion Models
Generative Models by Stability AI
[ECCV 2024] - Improving Zero-shot Generalization of Learned Prompts via Unsupervised Knowledge Distillation
The official repository for "One Model to Rule them All: Towards Universal Segmentation for Medical Images with Text Prompts"
[ECCV 2024] Official PyTorch implementation of DreamLIP: Language-Image Pre-training with Long Captions
The official repository to build SAT-DS, a medical data collection of 72 public segmentation datasets, contains over 22K 3D images, 302K segmentation masks and 497 classes from 3 different modaliti…
Official Pytorch implementation of 'Facing the Elephant in the Room: Visual Prompt Tuning or Full Finetuning'? (ICLR2024)
Official implementation of paper titled "GroupMamba: Parameter-Efficient and Accurate Group Visual State Space Model"
Pure Python Spell Checking http:https://pyspellchecker.readthedocs.io/en/latest/
[NeurIPS 2023 Oral] Quilt-1M: One Million Image-Text Pairs for Histopathology.
Pathology Language and Image Pre-Training (PLIP) is the first vision and language foundation model for Pathology AI (Nature Medicine). PLIP is a large-scale pre-trained model that can be used to ex…