-
Nanjing University
- 163 Xianlin Avenue, Nanjing, China
- [email protected]
- https://scholar.google.com/citations?user=V01xzWQAAAAJ&hl=zh-CN
Stars
A novel Multimodal Large Language Model (MLLM) architecture, designed to structurally align visual and textual embeddings.
AlignGPT Family: Support Various Large Language Models and Visual Backbones
the official repo for EMNLP 2024 (main) paper "EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language Models"
A curated list of awesome Multimodal studies.
Official repo for "AlignGPT: Multi-modal Large Language Models with Adaptive Alignment Capability"
A paper list for aspect based sentiment analysis.
The model, data and code for the visual GUI Agent SeeClick
Open-sourced codes for MiniGPT-4 and MiniGPT-v2 (https://minigpt-4.github.io, https://minigpt-v2.github.io/)
The code implementation of "M2DF: Multi-grained Multi-curriculum Denoising Framework for Multimodal Aspect-based Sentiment Analysis"
GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)
Additional code to construct the datasets of paper *DRIN: A Dynamic Relation Interactive Network for Multimodal Entity Linking*.
Code and data for *DRIN: A Dynamic Relation Interactive Network for Multimodal Entity Linking*
[ICLR'24 spotlight] Chinese and English Multimodal Large Model Series (Chat and Paint) | 基于CPM基础模型的中英双语多模态大模型系列
多模态中文LLaMA&Alpaca大语言模型(VisualCLA)
mPLUG-Owl: The Powerful Multi-modal Large Language Model Family
❓ Attention-based Visual Question Answering in Torch
Collect awesome best papers from top AI conferences.
Code for our ACL 2021 paper - ConSERT: A Contrastive Framework for Self-Supervised Sentence Representation Transfer
Pytorch reimplement of the paper "A Novel Cascade Binary Tagging Framework for Relational Triple Extraction" ACL2020. The original code is written in keras.
vime, an easy and structural config for (neo)vim users