-
IDEA
- Shenzhen, China
-
06:23
(UTC -12:00) - https://rentainhe.github.io/
- @Tianhe_Ren
Block or Report
Block or report rentainhe
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseLists (10)
Sort Name ascending (A-Z)
Detection Transformer
detrex extension
detrex extension works
Generation
GAN, Diffusion, etc.IDEA-CVR work
Label-Convert-Tools
Convert the label into different format: yolo2coco, coco2yoloLLM-Engineer
Large Language ModelingOpen Vocabulary
Language
Sort by: Recently starred
Starred repositories
Code release for "Segment Anything without Supervision"
Implementation for "Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs"
[CVPR 2024] Exploring Orthogonality in Open World Object Detection
This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models?
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.
The most powerful and modular stable diffusion GUI, api and backend with a graph/nodes interface.
Object Recognition as Next Token Prediction (CVPR 2024)
[AAAI-23 Oral] Official implementation of the paper "Are Transformers Effective for Time Series Forecasting?"
PRESTO: Progressive Pretraining Enhances Synthetic Chemistry Outcomes
A Framework of Small-scale Large Multimodal Models
Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Flax.
Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.
Doing simple retrieval from LLM models at various context lengths to measure accuracy
DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence
From anything to mesh like human artists. Official impl. of "MeshAnything: Artist-Created Mesh Generation with Autoregressive Transformers"
[ICML 2024] Quest: Query-Aware Sparsity for Efficient Long-Context LLM Inference
🌋👵🏻 Yo'LLaVA: Your Personalized Language and Vision Assistant
🎉CUDA 笔记 / 大模型手撕CUDA / C++笔记,更新随缘: flash_attn、sgemm、sgemv、warp reduce、block reduce、dot product、elementwise、softmax、layernorm、rmsnorm、hist etc.
Depth Anything V2. A More Capable Foundation Model for Monocular Depth Estimation
Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.
Fast, correct Python JSON library supporting dataclasses, datetimes, and numpy
Ultra fast JSON decoder and encoder written in C with Python bindings
Official repo of our paper "SDXS: Real-Time One-Step Latent Diffusion Models with Image Conditions"
Autoregressive Model Beats Diffusion: 🦙 Llama for Scalable Image Generation
MLLM-Bench: Evaluating Multimodal LLMs with Per-sample Criteria