A collection of models that are able to be run using onnxruntime-genai and can be served through embeddedllm library.
AI & ML interests
None defined yet.
Recent Activity
View all activity
Model Powered by Onnxruntime CPU GenAI
-
EmbeddedLLM/Phi-3-mini-4k-instruct-onnx-cpu-int4-rtn-block-32
Text Generation • Updated • 4 -
EmbeddedLLM/Phi-3-mini-4k-instruct-onnx-cpu-int4-rtn-block-32-acc-level-4
Text Generation • Updated • 3 -
EmbeddedLLM/Phi-3-mini-128k-instruct-onnx-cpu-int4-rtn-block-32
Text Generation • Updated • 3 -
EmbeddedLLM/Phi-3-mini-128k-instruct-onnx-cpu-int4-rtn-block-32-acc-level-4
Text Generation • Updated • 2
-
microsoft/Phi-3-mini-4k-instruct
Text Generation • 4B • Updated • 1.27M • 1.37k -
microsoft/Phi-3-mini-128k-instruct
Text Generation • 4B • Updated • 51.6k • 1.69k -
microsoft/Phi-3-medium-4k-instruct
Text Generation • 14B • Updated • 8.86k • 224 -
microsoft/Phi-3-medium-128k-instruct
Text Generation • 14B • Updated • 11.7k • 386
A collection of model that can be hosted using OpenVINO Model Server
Model Powered by Onnxruntime DirectML GenAI
-
EmbeddedLLM/Phi-3-mini-4k-instruct-onnx-directml
Text Generation • Updated • 3 -
EmbeddedLLM/Phi-3-mini-128k-instruct-onnx-directml
Text Generation • Updated • 2 -
EmbeddedLLM/Phi-3-medium-4k-instruct-onnx-directml
Text Generation • Updated • 2 -
EmbeddedLLM/Phi-3-medium-128k-instruct-onnx-directml
Text Generation • Updated • 2
-
microsoft/Phi-3-mini-4k-instruct
Text Generation • 4B • Updated • 1.27M • 1.37k -
microsoft/Phi-3-mini-128k-instruct
Text Generation • 4B • Updated • 51.6k • 1.69k -
microsoft/Phi-3-medium-4k-instruct
Text Generation • 14B • Updated • 8.86k • 224 -
microsoft/Phi-3-medium-128k-instruct
Text Generation • 14B • Updated • 11.7k • 386
A collection of models that are able to be run using onnxruntime-genai and can be served through embeddedllm library.
Model Powered by Onnxruntime DirectML GenAI
-
EmbeddedLLM/Phi-3-mini-4k-instruct-onnx-directml
Text Generation • Updated • 3 -
EmbeddedLLM/Phi-3-mini-128k-instruct-onnx-directml
Text Generation • Updated • 2 -
EmbeddedLLM/Phi-3-medium-4k-instruct-onnx-directml
Text Generation • Updated • 2 -
EmbeddedLLM/Phi-3-medium-128k-instruct-onnx-directml
Text Generation • Updated • 2
Model Powered by Onnxruntime CPU GenAI
-
EmbeddedLLM/Phi-3-mini-4k-instruct-onnx-cpu-int4-rtn-block-32
Text Generation • Updated • 4 -
EmbeddedLLM/Phi-3-mini-4k-instruct-onnx-cpu-int4-rtn-block-32-acc-level-4
Text Generation • Updated • 3 -
EmbeddedLLM/Phi-3-mini-128k-instruct-onnx-cpu-int4-rtn-block-32
Text Generation • Updated • 3 -
EmbeddedLLM/Phi-3-mini-128k-instruct-onnx-cpu-int4-rtn-block-32-acc-level-4
Text Generation • Updated • 2
-
microsoft/Phi-3-mini-4k-instruct
Text Generation • 4B • Updated • 1.27M • 1.37k -
microsoft/Phi-3-mini-128k-instruct
Text Generation • 4B • Updated • 51.6k • 1.69k -
microsoft/Phi-3-medium-4k-instruct
Text Generation • 14B • Updated • 8.86k • 224 -
microsoft/Phi-3-medium-128k-instruct
Text Generation • 14B • Updated • 11.7k • 386
-
microsoft/Phi-3-mini-4k-instruct
Text Generation • 4B • Updated • 1.27M • 1.37k -
microsoft/Phi-3-mini-128k-instruct
Text Generation • 4B • Updated • 51.6k • 1.69k -
microsoft/Phi-3-medium-4k-instruct
Text Generation • 14B • Updated • 8.86k • 224 -
microsoft/Phi-3-medium-128k-instruct
Text Generation • 14B • Updated • 11.7k • 386
A collection of model that can be hosted using OpenVINO Model Server