- Republic of Korea, Seoul
-
04:50
(UTC +09:00)
Highlights
- Pro
Stars
Project Page for "LISA: Reasoning Segmentation via Large Language Model"
[MM2024, oral] "Self-Supervised Visual Preference Alignment" https://arxiv.org/abs/2404.10501
Paper Reproduction Google SCoRE(Training Language Models to Self-Correct via Reinforcement Learning)
"Repository for implementing the Back-in-Time Diffusion method for detecting medical deepfakes in CT and MRI scans, including training and evaluation tools."
Smooth Diffusion: Crafting Smooth Latent Spaces in Diffusion Models arXiv 2023 / CVPR 2024
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 and reasoning techniques.
g1: Using Llama-3.1 70b on Groq to create o1-like reasoning chains
Beyond Hallucinations: Enhancing LVLMs through Hallucination-Aware Direct Preference Optimization
A curated list of practical guide resources of Medical LLMs (Medical LLMs Tree, Tables, and Papers)
Train transformer language models with reinforcement learning.
This repository contains the code for SFT, RLHF, and DPO, designed for vision-based LLMs, including the LLaVA models and the LLaMA-3.2-vision models.
A new collection of medical VQA dataset based on MIMIC-CXR. Part of the work 'EHRXQA: A Multi-Modal Question Answering Dataset for Electronic Health Records with Chest X-ray Images, NeurIPS 2023 D&B'.
[ECCV 2024] Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMs
A Simple Latent Diffusion Approach for Panoptic Segmentation and Mask Inpainting [ECCV 2024]
[NeurIPS 2024] Calibrated Self-Rewarding Vision Language Models
Utilities intended for use with Llama models.
An interpretable large language model (LLM) for medical diagnosis.
Code for STaR: Bootstrapping Reasoning With Reasoning (NeurIPS 2022)
Benchmarking the medical calculation capabilities of large language models.
[EMNLP'24] RULE: Reliable Multimodal RAG for Factuality in Medical Vision Language Models
Repository for RadRevise, a dataset for instruction-based radiology report editing
Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI
MLLM-Bench: Evaluating Multimodal LLMs with Per-sample Criteria
[arXiv'24] EVA-X: A foundation model for general chest X-ray analysis with self-supervised learning