![awesome logo](https://raw.githubusercontent.com/github/explore/80688e429a7d4ef2fca1e82350fe8e3517d3494d/topics/awesome/awesome.png)
-
ByteDance
- Shanghai, China
-
13:53
(UTC +08:00) - https://www.zhihu.com/people/leon-6-17-37
Highlights
- Pro
Block or Report
Block or report LEON-gittech
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseLists (1)
Sort Name ascending (A-Z)
Language
Sort by: Recently starred
Starred repositories
An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.
Aligning pretrained language models with instruction data generated by themselves.
EMNLP'2023: Explore-Instruct: Enhancing Domain-Specific Instruction Coverage through Active Exploration
A curated, but incomplete, list of data-centric AI resources.
Collection of training data management explorations for large language models
FinGPT: Open-Source Financial Large Language Models! Revolutionize 🔥 We release the trained model on HuggingFace.
This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.
Lazy Predict help build a lot of basic models without much code and helps understand which models works better without any parameter tuning
Code for the paper "Evaluating Large Language Models Trained on Code"
[EMNLP 2021] SimCSE: Simple Contrastive Learning of Sentence Embeddings https://arxiv.org/abs/2104.08821
A Zotero plugin for syncing items and notes into Notion
A toolkit for building dense retrievers with deep language models.
SwissArmyTransformer is a flexible and powerful library to develop your own Transformer variants.
llama3 implementation one matrix multiplication at a time
🤗 Evaluate: A library for easily evaluating machine learning models and datasets.
Multilingual Sentence & Image Embeddings with BERT
A WebUI for Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
All available datasets for Instruction Tuning of Large Language Models
Fine tune a T5 transformer model using PyTorch & Transformers🤗
Learning adapter weights from task descriptions
Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"
Reproduce results and replicate training fo T0 (Multitask Prompted Training Enables Zero-Shot Task Generalization)
[ICML 2023] Exploring the Benefits of Training Expert Language Models over Instruction Tuning
Robust Speech Recognition via Large-Scale Weak Supervision
Official code for paper "Not All Tasks are Born Equal: Uniderstanding Zero-Shot Generalization"