Stars
本项目是一个面向小白开发者的大模型应用开发教程,在线阅读地址:https://datawhalechina.github.io/llm-universe/
This is a collection of research papers for Self-Correcting Large Language Models with Automated Feedback.
ToRA is a series of Tool-integrated Reasoning LLM Agents designed to solve challenging mathematical reasoning problems by interacting with tools [ICLR'24].
Data and Code for Program of Thoughts (TMLR 2023)
ClickPrompt - Streamline your prompt design, with ClickPrompt, you can easily view, share, and run these prompts with just one click. ClickPrompt 用于一键轻松查看、分享和执行您的 Prompt。
DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models
[ACL 2024] A Survey of Chain of Thought Reasoning: Advances, Frontiers and Future
arXiv 23 "Towards Improving Document Understanding: An Exploration on Text-Grounding via MLLMs"
Official PyTorch implementation for CVIU 23 "Progressive Recurrent Network for Shadow Removal"
Official PyTorch implementation for TCSVT 23 "Detect Any Shadow: Segment Anything for Video Shadow Detection"
Official PyTorch implementation for ACM MM22 "UDoc-GAN: Unpaired Document Illumination Correction with Background Light Prior"
Document Artifical Intelligence
The official code for “SimFIR: A Simple Framework for Fisheye Image Rectification with Self-supervised Representation Learning”, ICCV, 2023.
The official code for “Geometric Representation Learning for Document Image Rectification”, ECCV, 2022.
The official code for “DeepEraser: Deep Iterative Context Mining for Generic Text Eraser”, TMM, 2024.
Implementation of EMNLP 2023 Findings: Improving Question Generation with Multi-level Content Planning
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
LLMs interview notes and answers:该仓库主要记录大模型(LLMs)算法工程师相关的面试题和参考答案
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
Instruct-tune LLaMA on consumer hardware
Learning Transferable Pedestrian Representation from Multimodal Information Supervision
深度学习500问,以问答形式对常用的概率知识、线性代数、机器学习、深度学习、计算机视觉等热点问题进行阐述,以帮助自己及有需要的读者。 全书分为18个章节,50余万字。由于水平有限,书中不妥之处恳请广大读者批评指正。 未完待续............ 如有意合作,联系[email protected] 版权所有,违权必究 Tan 2018.06
LLaMa retrieval plugin script using OpenAI's retrieval plugin
Llama中文社区,Llama3在线体验和微调模型已开放,实时汇总最新Llama3学习资料,已将所有代码更新适配Llama3,构建最好的中文Llama大模型,完全开源可商用
《代码随想录》LeetCode 刷题攻略:200道经典题目刷题顺序,共60w字的详细图解,视频难点剖析,50余张思维导图,支持C++,Java,Python,Go,JavaScript等多语言版本,从此算法学习不再迷茫!🔥🔥 来看看,你会发现相见恨晚!🚀
Library for Knowledge Intensive Language Tasks
MTEB: Massive Text Embedding Benchmark