NEW HF π€ COURSE to help people dive into Computer Vision built by the HF community. Over the last 6 months the hugging face discord community has been hard at work developing a new computer vision course. Receive a Certificate of completion and share it on your socials π€.
Finally! My first post for the lovely community out there!
Here's a highly quantized finetuned version of gemma focused exclusively on Prompt Engineering. Write as ambiguous you want and leave the job to this model
Outperforms industry giants like GPT-4, Gemini, Meditron-70B, Med-PaLM-1, and Med-PaLM-2 in the biomedical domain. π₯π π
OpenBioLLM-70B delivers SOTA performance, setting a new state-of-the-art for models of its size. OpenBioLLM-8B model and even surpasses GPT-3.5, Gemini, and Meditron-70B! π
Today's release is just the beginning! In the coming months, we'll be introducing:
π₯ππ New Research Alert - YOCO! πππ₯ π Title: You Only Cache Once: Decoder-Decoder Architectures for Language Models π
π Description: YOCO is a novel decoder-decoder architecture for LLMs that reduces memory requirements, speeds up prefilling, and maintains global attention. It consists of a self-decoder for encoding KV caches and a cross-decoder for reusing these caches via cross-attention.