-
MORSE Corp
- Cambridge, MA
Lists (5)
Sort Name ascending (A-Z)
Stars
Official code for PuLID: Pure and Lightning ID Customization via Contrastive Alignment
DSPy: The framework for programming—not prompting—foundation models
Text-to-video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型
Open-source evaluation toolkit of large vision-language models (LVLMs), support ~100 VLMs, 40+ benchmarks
A modular graph-based Retrieval-Augmented Generation (RAG) system
Framework agnostic sliced/tiled inference + interactive ui + error analysis plots
[ECCV 2024] Elysium: Exploring Object-level Perception in Videos via MLLM
Turns Data and AI algorithms into production-ready web applications in no time.
E5-V: Universal Embeddings with Multimodal Large Language Models
An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.
Bag of Tricks: Benchmarking of Jailbreak Attacks on LLMs. Empirical tricks for LLM Jailbreaking.
Official Repository of paper VideoGPT+: Integrating Image and Video Encoders for Enhanced Video Understanding
fabric is an open-source framework for augmenting humans using AI. It provides a modular framework for solving specific problems using a crowdsourced set of AI prompts that can be used anywhere.
Composio equip's your AI agents & LLMs with 100+ high-quality integrations via function calling
Drop in a screenshot and convert it to clean code (HTML/Tailwind/React/Vue)
A curated list of papers and resources related to Described Object Detection, Open-Vocabulary/Open-World Object Detection and Referring Expression Comprehension. Updated frequently and pull request…
Extract clean markdown from PDFs, URLs, Word docs, slides, videos, and more, ready for any LLM. ⚡
Real-time and accurate open-vocabulary end-to-end object detection
A Survey on Vision-Language Geo-Foundation Models (VLGFMs)
Techniques for deep learning with satellite & aerial imagery
Automated prompt-based testing and evaluation of Gen AI applications
Depth Anything V2. A More Capable Foundation Model for Monocular Depth Estimation