Block or Report
Block or report richjjj
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseLists (5)
Sort Name ascending (A-Z)
Stars
Language
Sort by: Recently starred
Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.
This repository is based on shouxieai/tensorRT_Pro, with adjustments to support YOLOv8.
Cartographer is a system that provides real-time simultaneous localization and mapping (SLAM) in 2D and 3D across multiple platforms and sensor configurations.
Autoware - the world's leading open-source software project for autonomous driving
跨平台的视频结构化(视频分析)框架,觉得有帮助的请给个星星 : ) 。**VideoPipe下一版本正在开发中,在保证跨平台、易上手的前提下,预计性能直逼deepstream等各硬件平台官方框架**。
Flash Attention in ~100 lines of CUDA (forward pass only)
A simple implementation of Tensorrt YOLOv8
[ICCV 2023] RepQ-ViT: Scale Reparameterization for Post-Training Quantization of Vision Transformers
The official implementation of the NeurIPS 2022 paper Q-ViT.
🚀🚀🚀This is an AI high-performance reasoning C++ library, Currently supports the deployment of yolov5, yolov7, yolov7-pose, yolov8, yolov8-seg, yolov8-pose, yolov8-obb, yolox, RTDETR, DETR, depth-an…
This repository serves as an example of deploying the YOLO models on Triton Server for performance and testing purposes
ONNX-compatible DeDoDe 🎶 Detect, Don't Describe - Describe, Don't Detect, for Local Feature Matching. Supports TensorRT 🚀
C++ application to perform computer vision tasks using Nvidia Triton Server for model inference
纯c++的全平台llm加速库,支持python调用,chatglm-6B级模型单卡可达10000+token / s,支持glm, llama, moss基座,手机端流畅运行
Deploying LLMs offline on the NVIDIA Jetson platform marks the dawn of a new era in embodied intelligence, where devices can function independently without continuous internet access.
A C++ implementation for UCMCTrack (SOTA in MOT17)
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.
A high-throughput and memory-efficient inference and serving engine for LLMs
TensorRT+YOLO系列的 多路 多卡 多实例 并行视频分析处理案例
gstreamer rtsp client support rockchip and jetson nx for C/C++ Python
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and…
A toolkit showing GPU's all-round capability in video processing
RTP-LLM: Alibaba's high-performance LLM inference engine for diverse applications.
提供多款 Shadowrocket 规则,拥有强劲的广告过滤功能。每日8时重新构建规则。
Official code implementation of Vary: Scaling Up the Vision Vocabulary of Large Vision Language Models.