This article guides you through the process of sideloading models required for Procyon AI Text Generation Benchmark. Note that an active internet connection is still required to run the test. 

Models for ONNX Runtime 


ModelDownload Link
Phi-3.5-mini-instruct-onnx-int4

https://procyon-public-download.s3.us-east-1.amazonaws.com/zip_onnxdml_Phi-3.5-mini-instruct-onnx-int4.zip

Mistral-7B-Instruct-v0.2-onnx-int4https://procyon-public-download.s3.us-east-1.amazonaws.com/zip_onnxdml_mistral-7b-instruct-v0.2-ONNX.zip
llama-3_1-8b-instruct-onnx-int4https://procyon-public-download.s3.us-east-1.amazonaws.com/zip_onnxdml_llama-3_1-8b-instruct-onnx-int4.zip
Llama-2-13b-chat-hf-onnx-int4https://procyon-public-download.s3.us-east-1.amazonaws.com/zip_onnxdml_Llama-2-13b-chat-hf-onnx-int4.zip


Models for OpenVINO Runtime


ModelDownload Link
Phi-3.5-mini-instruct-ov-int4https://procyon-public-download.s3.us-east-1.amazonaws.com/zip_openvino_Phi-3.5-mini-instruct-ov-int4.zip
Mistral-7B-Instruct-v0.2-ov-int4https://procyon-public-download.s3.us-east-1.amazonaws.com/zip_openvino_Mistral-7B-Instruct-v0.2-ov-int4.zip
llama-3_1-8b-instruct-ov-int4https://procyon-public-download.s3.us-east-1.amazonaws.com/zip_openvino_llama-3_1-8b-instruct-ov-int4.zip
Llama-2-13b-chat-hf-ov-int4https://procyon-public-download.s3.us-east-1.amazonaws.com/zip_openvino_Llama-2-13b-chat-hf-ov-int4.zip

Installing the Models

You may choose to download individual models or all, based on your configuration and benchmarking needs. 

1. By default, the benchmark is installed in

%ProgramData%\UL\Procyon\chops\dlc\ai-textgeneration-benchmark\

2. Create a subfolder named "models", if it does not exist;

%ProgramData%\UL\Procyon\chops\dlc\ai-textgeneration-benchmark\models

3. Unzip the downloaded models and copy them over to the models folder. Note that the extracted folders should not contain the prefix "zip_". The models directory with all the downloaded models should look like this;