-
Tongji University
- shanghai
- https://mic.tongji.edu.cn/main.htm
Lists (4)
Sort Name ascending (A-Z)
Starred repositories
image scene graph generation benchmark
Benchmarking Panoptic Scene Graph Generation (PSG), ECCV'22
A python toolkit for parsing captions (in natural language) into scene graphs (as symbolic representations).
A python wrapper for the Visual Genome API
医疗知识图谱构建实战,通过爬虫获取百度百科数据,使用Mongodb存储结构化三元组,并使用neo4j进行知识图谱的构建及可视化; Medical Knowledge Graph; Crawler; neo4j
A Chinese medical ChatGPT based on LLaMa, training from large-scale pretrain corpus and multi-turn dialogue dataset.
Medical NLP Competition, dataset, large models, paper
A tutorial and implement of disease centered Medical knowledge graph and qa system based on it。知识图谱构建,自动问答,基于kg的自动问答。以疾病为中心的一定规模医药领域知识图谱,并以该知识图谱完成自动问答与分析服务。
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
The codes about "Uni-MoE: Scaling Unified Multimodal Models with Mixture of Experts"
[Paper List] Papers integrating knowledge graphs (KGs) and large language models (LLMs)
Code for Navigating Connected Memories with a Task-oriented Dialog System
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
AMC: Adaptive Multi-expert Collaborative Network for Text-guided Image Retrieval
This is the official github repo of Think-on-Graph. If you are interested in our work or willing to join our research team in Shenzhen, please feel free to contact us by email ([email protected])
Implementation of our IJCAI2022 oral paper, ER-SAN: Enhanced-Adaptive Relation Self-Attention Network for Image Captioning.
Pre-trained Multimodal Large Language Model Enhances Dermatological Diagnosis using SkinGPT-4
Large Language-and-Vision Assistant for Biomedicine, built towards multimodal GPT-4 level capabilities.
X-VLM: Multi-Grained Vision Language Pre-Training (ICML 2022)
[CVPR 2024] Alpha-CLIP: A CLIP Model Focusing on Wherever You Want
[Paper][AAAI2024]Structure-CLIP: Towards Scene Graph Knowledge to Enhance Multi-modal Structured Representations
[Paper][Preprint 2024] MyGO: Discrete Modality Information as Fine-Grained Tokens for Multi-modal Knowledge Graph Completion
[CVPR'24] RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
Official implementation of our paper "Finetuned Multimodal Language Models are High-Quality Image-Text Data Filters".
[TIP2023] The code of “Plug-and-Play Regulators for Image-Text Matching”