Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

Join Posts waitlist
AvelinaΒ 
posted an update 6 minutes ago
view post
Post
Found out my ECCV paper is getting rejected because of a LaTeX compile error :(
mattmdjagaΒ 
posted an update 25 minutes ago
zamalΒ 
posted an update about 1 hour ago
view post
Post
68
Finally!
My first post for the lovely community out there!

Here's a highly quantized finetuned version of gemma focused exclusively on Prompt Engineering. Write as ambiguous you want and leave the job to this model

zamal/gemma-7b-finetuned
samadplsΒ 
posted an update about 1 hour ago
aadityaΒ 
posted an update about 2 hours ago
view post
Post
82
Okay, I can post now, Yayy!!
Sharing the OpenBioLLM-70B as the first post :)

Introducing OpenBioLLM-Llama3-70B & 8B: The most capable openly available Medical-domain LLMs to date! πŸ©ΊπŸ’ŠπŸ§¬

Outperforms industry giants like GPT-4, Gemini, Meditron-70B, Med-PaLM-1, and Med-PaLM-2 in the biomedical domain. πŸ₯πŸ“ˆ 🌟

OpenBioLLM-70B delivers SOTA performance, setting a new state-of-the-art for models of its size. OpenBioLLM-8B model and even surpasses GPT-3.5, Gemini, and Meditron-70B! πŸš€

Today's release is just the beginning! In the coming months, we'll be introducing:

- Expanded medical domain coverage 🧠
- Longer context windows πŸ“œπŸ”
- Better benchmarks πŸ“ˆπŸ†
- Multimodal capabilitiesπŸ–₯οΈπŸ©ΊπŸ“ŠπŸ”¬

Medical-LLM Leaderboard: openlifescienceai/open_medical_llm_leaderboard

More detail : https://huggingface.co/blog/aaditya/openbiollm
  • 1 reply
Β·
CocoSunΒ 
posted an update about 2 hours ago
phenixrhyderΒ 
posted an update about 6 hours ago
view post
Post
590
Midjourney Ai
  • 1 reply
Β·
nkasmanoffΒ 
posted an update about 13 hours ago
sequelboxΒ 
posted an update about 15 hours ago
DmitryRyuminΒ 
posted an update about 16 hours ago
view post
Post
905
πŸ”₯πŸš€πŸŒŸ New Research Alert - YOCO! πŸŒŸπŸš€πŸ”₯
πŸ“„ Title: You Only Cache Once: Decoder-Decoder Architectures for Language Models πŸ”

πŸ“ Description: YOCO is a novel decoder-decoder architecture for LLMs that reduces memory requirements, speeds up prefilling, and maintains global attention. It consists of a self-decoder for encoding KV caches and a cross-decoder for reusing these caches via cross-attention.

πŸ‘₯ Authors: Yutao Sun et al.

πŸ“„ Paper: You Only Cache Once: Decoder-Decoder Architectures for Language Models (2405.05254)

πŸ“ Repository: https://github.com/microsoft/unilm/tree/master/YOCO

πŸ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

πŸ” Keywords: #YOCO #DecoderDecoder #LargeLanguageModels #EfficientArchitecture #GPUMemoryReduction #PrefillingSpeedup #GlobalAttention #DeepLearning #Innovation #AI