AI PC: Text Generation
Text generation LLMs that have been validated to run on the AI PC Intel® Core™ Ultra CPU and iGPU.
- Text Generation • Updated • 29 • 2
OpenVINO/mixtral-8x7b-instruct-v0.1-int4-ov
Text Generation • Updated • 55 • 3OpenVINO/phi-2-fp16-ov
Text Generation • Updated • 47 • 1OpenVINO/phi-2-int8-ov
Text Generation • Updated • 53OpenVINO/phi-2-int4-ov
Text Generation • Updated • 63 • 1OpenVINO/mistral-7b-instruct-v0.1-fp16-ov
Text Generation • Updated • 19OpenVINO/mistral-7b-instruct-v0.1-int8-ov
Text Generation • Updated • 899 • 1OpenVINO/mistral-7b-instruct-v0.1-int4-ov
Text Generation • Updated • 460OpenVINO/starcoder2-15b-fp16-ov
Text Generation • Updated • 21OpenVINO/starcoder2-15b-int8-ov
Text Generation • Updated • 19OpenVINO/starcoder2-15b-int4-ov
Text Generation • Updated • 21OpenVINO/neural-chat-7b-v3-3-fp16-ov
Text Generation • Updated • 27OpenVINO/neural-chat-7b-v3-3-int8-ov
Text Generation • Updated • 33OpenVINO/neural-chat-7b-v3-3-int4-ov
Text Generation • Updated • 56OpenVINO/mpt-7b-fp16-ov
Text Generation • Updated • 15OpenVINO/mpt-7b-int8-ov
Text Generation • Updated • 27OpenVINO/mpt-7b-int4-ov
Text Generation • Updated • 12OpenVINO/Phi-3-mini-128k-instruct-fp16-ov
Text Generation • Updated • 26OpenVINO/Phi-3-mini-128k-instruct-int8-ov
Text Generation • Updated • 26 • 2OpenVINO/Phi-3-mini-128k-instruct-int4-ov
Text Generation • Updated • 592 • 2OpenVINO/falcon-7b-instruct-fp16-ov
Text Generation • Updated • 20OpenVINO/falcon-7b-instruct-int8-ov
Text Generation • Updated • 24OpenVINO/falcon-7b-instruct-int4-ov
Text Generation • Updated • 22OpenVINO/open_llama_3b_v2-fp16-ov
Text Generation • Updated • 67OpenVINO/open_llama_3b_v2-int8-ov
Text Generation • Updated • 447 • 1OpenVINO/open_llama_3b_v2-int4-ov
Text Generation • Updated • 57OpenVINO/open_llama_7b_v2-fp16-ov
Text Generation • Updated • 29OpenVINO/open_llama_7b_v2-int8-ov
Text Generation • Updated • 17OpenVINO/open_llama_7b_v2-int4-ov
Text Generation • Updated • 24OpenVINO/gpt-j-6b-fp16-ov
Text Generation • Updated • 22OpenVINO/gpt-j-6b-int8-ov
Text Generation • Updated • 22OpenVINO/gpt-j-6b-int4-ov
Text Generation • Updated • 27OpenVINO/RedPajama-INCITE-7B-Chat-fp16-ov
Text Generation • Updated • 17OpenVINO/RedPajama-INCITE-7B-Chat-int8-ov
Text Generation • Updated • 16OpenVINO/RedPajama-INCITE-7B-Chat-int4-ov
Text Generation • Updated • 29OpenVINO/RedPajama-INCITE-7B-Instruct-fp16-ov
Text Generation • Updated • 20OpenVINO/RedPajama-INCITE-7B-Instruct-int8-ov
Text Generation • Updated • 18OpenVINO/RedPajama-INCITE-7B-Instruct-int4-ov
Text Generation • Updated • 26OpenVINO/dolly-v2-7b-fp16-ov
Text Generation • Updated • 15OpenVINO/dolly-v2-7b-int8-ov
Text Generation • Updated • 22OpenVINO/dolly-v2-7b-int4-ov
Text Generation • Updated • 20OpenVINO/Mistral-7B-Instruct-v0.2-fp16-ov
Text Generation • Updated • 26OpenVINO/Mistral-7B-Instruct-v0.2-int8-ov
Text Generation • Updated • 448 • 1OpenVINO/Mistral-7B-Instruct-v0.2-int4-ov
Text Generation • Updated • 587 • 1OpenVINO/Phi-3-medium-4k-instruct-fp16-ov
Text Generation • Updated • 28OpenVINO/Phi-3-medium-4k-instruct-int8-ov
Text Generation • Updated • 36OpenVINO/Phi-3-medium-4k-instruct-int4-ov
Text Generation • Updated • 414 • 3OpenVINO/pythia-1.4b-fp16-ov
Text Generation • Updated • 20OpenVINO/pythia-1.4b-int8-ov
Text Generation • Updated • 13OpenVINO/pythia-1.4b-int4-ov
Text Generation • Updated • 21OpenVINO/pythia-2.8b-fp16-ov
Text Generation • Updated • 23OpenVINO/pythia-2.8b-int8-ov
Text Generation • Updated • 15OpenVINO/pythia-2.8b-int4-ov
Text Generation • Updated • 25OpenVINO/pythia-6.9b-fp16-ov
Text Generation • Updated • 17OpenVINO/pythia-6.9b-int8-ov
Text Generation • Updated • 19OpenVINO/pythia-6.9b-int4-ov
Text Generation • Updated • 19OpenVINO/pythia-1b-fp16-ov
Text Generation • Updated • 18OpenVINO/pythia-1b-int8-ov
Text Generation • Updated • 16OpenVINO/pythia-1b-int4-ov
Text Generation • Updated • 17OpenVINO/neural-chat-7b-v1-1-fp16-ov
Text Generation • Updated • 18OpenVINO/neural-chat-7b-v1-1-int8-ov
Text Generation • Updated • 19OpenVINO/neural-chat-7b-v1-1-int4-ov
Text Generation • Updated • 53OpenVINO/persimmon-8b-chat-int4-ov
Text Generation • Updated • 27OpenVINO/persimmon-8b-chat-fp16-ov
Text Generation • Updated • 17OpenVINO/persimmon-8b-chat-int8-ov
Text Generation • Updated • 20
AhmedSSoliman/MarianCausalLM
Text Generation • Updated • 22Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
AurelPx/Pegasus-7b-slerp
Text Generation • Updated • 70 • 1Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
BAAI/Aquila-7B
Updated • 679 • 17Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
BAAI/Aquila2-7B
Text Generation • Updated • 770 • 6Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
BAAI/AquilaChat-7B
Updated • 378 • 48Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
BAAI/AquilaChat2-7B
Text Generation • Updated • 865 • 15Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
BigSalmon/GPT2Neo1.3BPoints
Text Generation • Updated • 15Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
EleutherAI/pythia-1.4b
Text Generation • Updated • 23.3k • 22Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
EleutherAI/pythia-12b
Text Generation • Updated • 80.4k • 131Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
EleutherAI/pythia-14m
Text Generation • Updated • 317k • 18Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
EleutherAI/pythia-160m
Text Generation • Updated • 94.5k • 25Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
EleutherAI/pythia-1b
Text Generation • Updated • 58.1k • 32Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
EleutherAI/pythia-2.8b
Text Generation • Updated • 20.3k • 28Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
EleutherAI/pythia-410m
Text Generation • Updated • 34.7k • 20Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
EleutherAI/pythia-6.9b
Text Generation • Updated • 26.6k • 47Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
EleutherAI/pythia-70m
Updated • 83.9k • 55Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
HuggingFaceH4/zephyr-7b-beta
Text Generation • Updated • 809k • • 1.6kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Intel/neural-chat-7b-v1-1
Text Generation • Updated • 40 • 23Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Intel/neural-chat-7b-v3-3
Text Generation • Updated • 53.4k • 75Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/CodeQwen1.5-7B-Chat
Text Generation • Updated • 48.5k • 312Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/Qwen-1_8B
Text Generation • Updated • 3.26k • 61Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/Qwen-1_8B-Chat
Text Generation • Updated • 58.5k • 108Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/Qwen-7B
Text Generation • Updated • 19k • 368Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/Qwen-7B-Chat
Text Generation • Updated • 63.5k • 751Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/Qwen1.5-0.5B
Text Generation • Updated • 401k • 143Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/Qwen1.5-0.5B-Chat
Text Generation • Updated • 331k • 74Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/Qwen1.5-1.8B
Text Generation • Updated • 354k • 43Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/Qwen1.5-1.8B-Chat
Text Generation • Updated • 12k • 45Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/Qwen1.5-4B
Text Generation • Updated • 12.8k • 33Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/Qwen1.5-4B-Chat
Text Generation • Updated • 6.18k • 37Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/Qwen1.5-7B
Text Generation • Updated • 203k • 45Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Qwen/Qwen1.5-7B-Chat
Text Generation • Updated • 31.8k • 162Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Salesforce/codegen-2B-multi
Text Generation • Updated • 11k • 35Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Salesforce/codegen-350M-mono
Text Generation • Updated • 14.1k • 86Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Salesforce/codegen-6B-multi
Text Generation • Updated • 4.47k • 19Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Salesforce/codegen2-1B_P
Text Generation • Updated • 2.58k • 38Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Salesforce/codegen2-3_7B_P
Text Generation • Updated • 129 • 15Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
Salesforce/codegen2-7B_P
Text Generation • Updated • 239 • 26Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
WizardLMTeam/WizardMath-7B-V1.1
Text Generation • Updated • 6.92k • 75Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
X-D-Lab/MindChat-Qwen2-4B
Text Generation • Updated • 15 • 4Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
adept/persimmon-8b-chat
Text Generation • Updated • 1.23k • 42Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
baichuan-inc/Baichuan2-13B-Chat
Text Generation • Updated • 93.8k • 421Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
baichuan-inc/Baichuan2-7B-Base
Text Generation • Updated • 1.89k • 77Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
baichuan-inc/Baichuan2-7B-Chat
Text Generation • Updated • 22.9k • 158Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
bigscience/bloom-560m
Text Generation • Updated • 222k • 346Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
bigscience/bloomz-1b1
Text Generation • Updated • 2.67k • 32Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
bigscience/bloomz-3b
Text Generation • Updated • 10.5k • 77Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
bigscience/bloomz-7b1-mt
Text Generation • Updated • 2.61k • 140Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
facebook/opt-1.3b
Text Generation • Updated • 18.6M • 151Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
facebook/opt-125m
Text Generation • Updated • 7.24M • 163Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
facebook/opt-13b
Text Generation • Updated • 19.8k • 65Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
facebook/opt-2.7b
Text Generation • Updated • 30.3k • 80Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
facebook/opt-350m
Text Generation • Updated • 226k • 129Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
facebook/opt-6.7b
Text Generation • Updated • 142k • 109Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
facebook/opt-iml-1.3b
Text Generation • Updated • 875 • 29Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
google/codegemma-1.1-2b
Text Generation • Updated • 111 • 17Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
google/codegemma-1.1-7b-it
Text Generation • Updated • 179 • 49Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
google/codegemma-2b
Text Generation • Updated • 5.75k • 72Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
google/codegemma-7b
Text Generation • Updated • 3.86k • 165Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
google/gemma-1.1-2b-it
Text Generation • Updated • 91.4k • 151Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
google/gemma-1.1-7b-it
Text Generation • Updated • 16.3k • • 263Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
google/gemma-2b
Text Generation • Updated • 495k • 908Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
google/gemma-2b-it
Text Generation • Updated • 95.8k • • 670Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
google/gemma-7b
Text Generation • Updated • 342k • • 3.05kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
google/gemma-7b-it
Text Generation • Updated • 467k • 1.14kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
ibm-granite/granite-3b-code-base-2k
Text Generation • Updated • 44.4k • 35Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
ibm-granite/granite-3b-code-instruct-2k
Text Generation • Updated • 13.2k • 31Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
ibm-granite/granite-8b-code-base-4k
Text Generation • Updated • 2.99k • 29Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
ibm-granite/granite-8b-code-instruct-4k
Text Generation • Updated • 9.51k • 108Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
internlm/internlm2-1_8b
Text Generation • Updated • 16.8k • 28Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
internlm/internlm2-7b
Text Generation • Updated • 12.7k • 40Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
internlm/internlm2-chat-1_8b
Text Generation • Updated • 8.94k • 29Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
internlm/internlm2-chat-7b
Text Generation • Updated • 15.9k • 82Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
internlm/internlm2-chat-7b-sft
Text Generation • Updated • 7.93k • 6Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
internlm/internlm2-math-7b
Text Generation • Updated • 259 • 27Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
internlm/internlm2-math-base-7b
Text Generation • Updated • 568 • 2Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
ise-uiuc/Magicoder-CL-7B
Text Generation • Updated • 96 • 21Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
ise-uiuc/Magicoder-DS-6.7B
Text Generation • Updated • 720 • 37Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
ise-uiuc/Magicoder-S-CL-7B
Text Generation • Updated • 1.93k • 44Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
ise-uiuc/Magicoder-S-DS-6.7B
Text Generation • Updated • 3.18k • 200Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
meta-llama/Llama-2-13b-chat-hf
Text Generation • Updated • 1.02M • 1.02kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
meta-llama/Llama-2-13b-hf
Text Generation • Updated • 190k • 573Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
meta-llama/Llama-2-7b-chat-hf
Text Generation • Updated • 830k • • 3.97kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
meta-llama/Llama-2-7b-hf
Text Generation • Updated • 1.28M • 1.77kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
meta-llama/Meta-Llama-3-8B
Text Generation • Updated • 664k • 5.81kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
meta-llama/Meta-Llama-3-8B-Instruct
Text Generation • Updated • 2.19M • • 3.59kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
meta-llama/Meta-Llama-Guard-2-8B
Text Generation • Updated • 12.6k • 281Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
microsoft/Phi-3-medium-4k-instruct
Text Generation • Updated • 42.1k • 211Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
microsoft/Phi-3-mini-128k-instruct
Text Generation • Updated • 584k • 1.6kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
microsoft/phi-2
Text Generation • Updated • 248k • 3.24kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
mistralai/Mistral-7B-Instruct-v0.2
Text Generation • Updated • 1.1M • • 2.57kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
mistralai/Mistral-7B-Instruct-v0.3
Text Generation • Updated • 484k • • 1.12kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
mistralai/Mistral-7B-v0.3
Text Generation • Updated • 288k • 387Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
mosaicml/mpt-7b
Text Generation • Updated • 42.7k • 1.16kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
mosaicml/mpt-7b-8k
Text Generation • Updated • 1.85k • 26Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
mosaicml/mpt-7b-8k-chat
Text Generation • Updated • 1.21k • 40Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
mosaicml/mpt-7b-chat
Text Generation • Updated • 8.3k • 512Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
mosaicml/mpt-7b-instruct
Text Generation • Updated • 8.05k • 467Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
mosaicml/mpt-7b-storywriter
Text Generation • Updated • 2.15k • 822Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
openai-community/gpt2
Text Generation • Updated • 15.9M • • 2.35kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
openbmb/MiniCPM-2B-sft-bf16
Text Generation • Updated • 1.86k • 118Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
openchat/openchat-3.6-8b-20240522
Text Generation • Updated • 14.6k • 149Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
stabilityai/stablelm-2-12b
Text Generation • Updated • 927 • 115Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
stabilityai/stablelm-2-12b-chat
Text Generation • Updated • 2.39k • 86Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
stabilityai/stablelm-2-1_6b
Text Generation • Updated • 4.39k • 185Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
stabilityai/stablelm-2-1_6b-chat
Text Generation • Updated • 4.54k • 31Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
stabilityai/stablelm-2-zephyr-1_6b
Text Generation • Updated • 8.98k • 180Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
stabilityai/stablelm-3b-4e1t
Text Generation • Updated • 16k • 309Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
stabilityai/stablelm-base-alpha-3b
Text Generation • Updated • 2.16k • 82Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
stabilityai/stablelm-tuned-alpha-7b
Text Generation • Updated • 4.15k • 357Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
stabilityai/stablelm-zephyr-3b
Text Generation • Updated • 12.8k • 247Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
stanford-crfm/BioMedLM
Text Generation • Updated • 2.67k • 394Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
tiiuae/falcon-11B
Text Generation • Updated • 17.9k • 212Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
tiiuae/falcon-7b
Text Generation • Updated • 105k • 1.08kNote This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
tiiuae/falcon-7b-instruct
Text Generation • Updated • 168k • • 918Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
togethercomputer/Pythia-Chat-Base-7B
Text Generation • Updated • 1.59k • 66Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
togethercomputer/RedPajama-INCITE-7B-Base
Text Generation • Updated • 1.09k • 94Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
togethercomputer/RedPajama-INCITE-7B-Chat
Text Generation • Updated • 1.71k • 92Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
togethercomputer/RedPajama-INCITE-7B-Instruct
Text Generation • Updated • 1.63k • 104Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
togethercomputer/RedPajama-INCITE-Chat-3B-v1
Text Generation • Updated • 8.52k • 152Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation
xverse/XVERSE-7B-Chat
Text Generation • Updated • 78 • 8Note This model was tested with OpenVINO version 2024.1.0, using the OVModelForCausalLM library with INT4 weight compression for the lowest memory footprint. To convert the model to OpenVINO, follow instructions at: https://huggingface.co/docs/optimum/main/en/intel/installation