-
Université Paris Cité
- Paris, France
- https://qtwang.github.io/
Block or Report
Block or report qtwang
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
[ICLR 2024] Official implementation of "TimeMixer: Decomposable Multiscale Mixing for Time Series Forecasting"
Collaborative Collection of C++ Best Practices. This online resource is part of Jason Turner's collection of C++ Best Practices resources. See README.md for more information.
An elegant \LaTeX\ résumé template. 大陆镜像 https://gods.coding.net/p/resume/git
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
Dumpy: A Compact and Adaptive Index for Large Data Series Collections (SIGMOD'23)
Open-source vector similarity search for Postgres
Seed guided neural metric learning approach for calculating trajectory similarities
Hackable and optimized Transformers building blocks, supporting a composable construction.
Running large language models on a single GPU for throughput-oriented scenarios.
Learning to Compress Prompts with Gist Tokens - https://arxiv.org/abs/2304.08467
[NeurIPS'23] H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models.
Examples from Programming in Parallel with CUDA
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
Running inference on the ZeroSCROLLS benchmark
Official repository for LongChat and LongEval
Unsupervised text tokenizer for Neural Network-based text generation.
StableLM: Stability AI Language Models
A toolkit for machine learning from time series
Fast and memory-efficient exact attention
This is the CUDA GPU implementation + Python interface (using PyTorch) of DCI. The paper can be found at https://arxiv.org/abs/1512.00442.
Official release of InternLM2 7B and 20B base and chat models. 200K context support
Python port of R's Comprehensive Dynamic Time Warp algorithms package