小满是什么季节| 牙痛用什么止痛| 以什么当什么| 番号是什么意思| 海马是什么类动物| 宫颈癌吃什么好| 亦什么意思| 肌钙蛋白低说明什么| 手电筒的金属外壳相当于电路中的什么| 思源名字的寓意是什么| 腺样体肥大挂什么科| 灭蚂蚁什么药最有效| 950是什么金| 天蓝色配什么颜色| 拔鼻毛有什么危害| 手腕痛什么原因| 胎盘内血池是什么| 消石灰是什么| 马齿苋能治什么病| 脖子淋巴结挂什么科| 梦见死人的场面是什么兆头| 请丧假需要什么证明| 1月19号是什么星座| 事无巨细是什么意思| 哂是什么意思| 牛肉炖什么菜好吃| 小本创业做什么生意好| 湿疹怎么治用什么药膏| 香火是什么意思| 鸡属相和什么属相最配| 郁郁寡欢什么意思| 漂亮的近义词是什么| 小孩心跳快是什么原因| 脾胃伏火是什么意思| 气血不足吃什么补最快| 2月25号是什么星座| 咖色配什么颜色好看| 股票pb是什么意思| 如何查自己是什么命格| 世界上最坚硬的东西是什么| 抽筋是什么病| 睾丸扭转是什么导致的| 喜欢是什么感觉| 为什么会眼压高| 丘疹是什么| 湖北有什么山| 最近爆发什么病毒感染| 二尖瓣微量反流什么意思| 贾蓉和王熙凤是什么关系| lycra是什么面料| 鹦鹉吃什么蔬菜| kitty什么意思| giuseppe是什么牌子| 什么叫做基本工资| 梦见胎死腹中预示什么| omega是什么牌子的手表| hm平方是什么单位| 喝枸杞子泡水有什么好处和坏处| ot什么意思| 臭菜是什么菜| 知青为什么要下乡| 你说到底为什么都是我的错| 2004年是什么年| 左后背发麻是什么原因| 田螺小子是什么意思| hk是什么意思| 准备好了吗时刻准备着是什么歌| 眼珠子疼是什么原因| 血小板低是什么引起的| 乳腺癌的症状是什么| 做亲子鉴定需要什么| 吃辣的胃疼吃什么药| 花生碎能做什么食物吃| 千岛酱是什么味道| 什么是袖珍人| 一什么饼干| 20度穿什么衣服合适| 打磨工是做什么的| 人的五官是什么| sage什么颜色| 茭白是什么植物| 体检前一天要注意什么| 体检需要带什么| 月经期间喝什么好排毒排污血| 基因突变是什么意思| 蝈蝈为什么不叫| 张少华什么时候去世的| 什么拜之交| 药店加盟需要什么条件| 高铁二等座是什么意思| 梦见妹妹是什么意思| 独占鳌头是什么意思| 官官相护是什么意思| 血脂稠吃什么食物好| 眼底出血用什么眼药水| 男性检查hpv挂什么科| 正常舌头是什么颜色| 薄荷长什么样| 为什么庙里不让孕妇去| 抗衰老吃什么| 形态各异是什么意思| 梦见猫吃老鼠什么意思| 爱是什么词| 尿黄是什么原因男性| 肝肾功能挂什么科| 淋巴结什么意思| mask是什么意思| 藿香正气水什么时候喝| 713是什么星座| 三亚是什么海| 九价是什么| 临床医学学什么| 戒掉手淫有什么好处| 水浒传有什么故事| 幸存者偏差是什么意思| 闲鱼转卖什么意思| csf是什么意思| 心率过快吃什么药好| 籍贯填写什么| 海棠果什么时候成熟| 与君共勉是什么意思| 肠胃不好拉肚子吃什么药| 性格内向的人适合做什么工作| 肛周脓肿挂什么科| 什么是陈皮| 刺梨根泡酒有什么功效| 热伤风流鼻涕吃什么药| 再创佳绩是什么意思| 老咳嗽是什么原因| 菠菜和豆腐为什么不能一起吃| 周长是什么意思| 网球肘用什么药| 眼泪多是什么原因| 守望先锋是什么类型的游戏| 抓手是什么意思| 180度是什么角| 什么是高脂血症| 农历11月25日是什么星座| 病毒性肝炎有什么症状| 酒不能和什么一起吃| 石乐读什么| 下午四点到五点是什么时辰| 怡的意思和含义是什么| 徐州二院全名叫什么| 伤口不愈合用什么药| 岁月静好浅笑安然什么意思| 尿急尿频尿不尽吃什么药| 一心一意什么意思| a4纸可以做什么手工| 耳膜炎是什么原因引起的| 美尼尔眩晕症吃什么药| 女人长期做俯卧撑有什么效果| 腰无力是什么原因| 纪梵希为什么不娶赫本| 仙逝是什么意思| 端着是什么意思| 刘庄为什么要灭了阴家| 脱脂乳是什么意思| 禾字五行属什么的| 天麻与什么煲汤最好| 虚病是什么意思| 为什么会莫名其妙的哭| 左侧肋骨下方疼痛是什么原因| 9月20号是什么星座| 胆囊炎可以吃什么| 女人更年期什么症状| 阴虚吃什么药效果最好| 为什么会有白头发| 阳阴阳是什么卦| 河水像什么| 空针是什么| 雪人是什么生肖| 刚生完孩子可以吃什么水果| 孤独的最高境界是什么| 蜻蜓是什么目| 义子是什么意思| 乙肝什么症状| 痔疮肛瘘是什么症状| 恶心想吐肚子疼是什么原因| 下午五点是什么时辰| 睡醒后口苦是什么原因| gm是什么牌子| 助听器什么价位| 寻麻疹吃什么药| 南京五行属什么| 总胆固醇高忌口什么| 拔火罐起水泡是什么原因| 产后第一次来月经是什么颜色| 蓝莓什么季节成熟| 一味是什么意思| 什么是透析| 肾精亏虚是什么意思| 脂溢性皮炎是什么原因引起的| 对线是什么意思| 做梦梦见地震是什么意思| 飞天奖是什么奖| 什么是凤凰男| 吃什么长头发又密又多| mixblu是什么牌子| 红线女是什么意思| 把握时机是指什么生肖| 宇宙的尽头是什么| 什么是一本| 做喉镜挂什么科| 拔罐后要注意什么| 淋巴结肿大是什么原因| c14呼气试验是检查什么的| 囊是什么意思| 怀孕吃什么菜最有营养| 为什么门牙突然有缝了| 良去掉一点读什么| 颈椎痛挂什么科| 鬓角长痘痘是什么原因| 梦到高考是什么意思| 胃胀气吃什么食物| 泥丸宫在什么位置| 肾衰竭吃什么好| 吹风样杂音见于什么病| 兰花叶子发黄是什么原因| 芝士是什么做的| 乳腺ca是什么意思| 沙参长什么样子图片| 情绪波动是什么意思| 身体发冷是什么原因| 白内障有什么症状表现| 胆囊壁固醇沉积是什么意思| 长瘊子是什么原因| 回奶吃什么快最有效的| 爱妃是什么意思| 碱性磷酸酶偏低是什么意思| 淋巴癌有什么症状| 太白金星是什么神| 丹参片和复方丹参片有什么区别| 一个日一个处一个口念什么| 用牛奶敷脸有什么好处和坏处| 轻断食什么意思| 做肠镜检查需要提前做什么准备| 溶栓治疗是什么意思| 计算机二级什么时候查成绩| 娘家人是什么意思| 王秋儿和王冬儿什么关系| 晚上睡觉出虚汗是什么原因| 尿血吃什么药最好| 男生回复嗯嗯代表什么| 什么咖啡最好| 静是什么意思| 辩证是什么意思| faye是什么意思| 痰中带血吃什么药| 血脂高什么意思| cold是什么意思| 218号是什么星座| 什么水果补肾| 荔枝与什么不能同吃| 什么牌子的辅酶q10好| 什么是大男子主义| 桫椤是什么植物| 梦见虫子是什么意思| 药引是什么意思| 蜻蜓点水是什么生肖| 小孩吐奶是什么原因| 17岁属什么| 玉米须能治什么病| 水痘疫苗什么时候打| 百度

Try to extend agent mode in VS Code!

减肥为什么会口臭

百度 央广网板块热点:标题:5G示范工程启动重庆最大窄带物联网NB-IoT商用重庆2018物联网生态高峰论坛暨重庆移动5G示范工程启动、NB-IoT商用发布在重庆举办,重庆移动董事长郭永宏做出上述表述。

Model conversion is an integrated development environment designed to help developers and AI engineers to convert, quantize, optimize and evaluate the pre-built machine learning models on your local Windows platform. It offers a streamlined, end-to-end experience for models converted from sources like Hugging Face, optimizing them and enabling inference on local devices powered by NPUs, GPUs, and CPUs.

Prerequisites

Create project

Creating a project in model conversion is the first step toward converting, optimizing, quantizing and evaluating machine learning models.

  1. Open the AI Toolkit view, and select Models > Conversion to launch model conversion

  2. Start a new project by selecting New Model Project

    Screenshot that shows view for creating model project, including Primary Side Bar and create project button.

  3. Choose a base model

    • Hugging Face Model: choose the base model with predefined recipes from the supported model list.
    • Model Template : if the model is not included in the base model, select an empty template for your customized recipes (advanced scenario).

    Screenshot that shows model list, such as bert, resnet, llama and so on.

  4. Enter project details: a unique Project Folder and a Project Name.

    A new folder with the specified project name is created in the location you selected for storing the project files.

Note

The first time you create a model project, it might take a while to set up the environment.

A README.md file is included in each project. If you close it, you can reopen it via the workspace. Screenshot that shows model readme.

Supported models

Model Conversion currently supports a growing list of models, including top Hugging Face models in PyTorch format.

LLM models

Model Name Hugging Face Path
Qwen2.5 1.5B Instruct Qwen/Qwen2.5-1.5B-Instruct
DeepSeek R1 Distill Qwen 1.5B deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
Meta LLaMA 3.2 1B Instruct meta-llama/Llama-3.2-1B-Instruct
Phi-3.5 Mini Instruct Phi-3.5-mini-instruct

Non-LLM models

Model Name Hugging Face Path
Intel BERT Base Uncased (MRPC) Intel/bert-base-uncased-mrpc
BERT Multilingual Cased google-bert/bert-base-multilingual-cased
ViT Base Patch16-224 google/vit-base-patch16-224
ResNet-50 resnet-50
CLIP ViT-B-32 (LAION) laion/CLIP-ViT-B-32-laion2B-s34B-b79K
CLIP ViT Base Patch16 clip-vit-base-patch16
CLIP ViT Base Patch32 clip-vit-base-patch32

(Optional) Add model into existing project

  1. Open the model project

  2. Select Models > Conversion, and then select Add Models on the right panel.

    Screenshot that shows how to add model. It contains a button to add models.

  3. Choose a base model or template, and then select Add.

    A folder containing the new model files is created in the current project folder.

(Optional) Create a new model project

  1. Open the model project

  2. Select Models > Conversion, and then select New Project on the right panel.

    Screenshot that shows how to create a new project. It contains a button to create a new project.

  3. Alternatively, close the current model project and create a new project from the start.

Run workflow

Running a workflow in model conversion is the core step that transform the pre-built ML model into an optimized and quantized ONNX model.

  1. Select File > Open Folder in VS Code to open the model project folder.

  2. Review the workflow configuration

    1. Select Models > Conversion
    2. Select the workflow template to view the conversion recipe.

    Screenshot that shows running a workflow. There is a workflow configuration section containing Conversion, Quantization and Evaluation.

    Conversion

    The workflow will always execute the conversion step, which transforms the model into ONNX format. This step cannot be disabled.

    Quantization

    This section enables you to configure the parameters for quantization.

    Important

    Hugging Face compliance alerts: During the quantization, we need the calibration datasets. You may be prompted to accept license terms before proceeding. If you missed the notification, the running process will be paused, waiting for your input. Please make sure notifications are enabled and that you accept the required licenses. Screenshot that shows disclaimer.

    • Activation Type: this is the data type used to represent the intermediate outputs (activations) of each layer in the neural network.

    • Weight Type: this is the data type used to represent the learned parameters (weights) of the model.

    • Quantization Dataset: calibration dataset used for quantization.

      If your workflow uses a dataset that requires license agreement approval on Hugging Face (e.g., ImageNet-1k), you’ll be prompted to accept the terms on the dataset page before proceeding. This is required for legal compliance.

      1. Select the HuggingFace Access Token button to get your Hugging Face Access Token.

        Screenshot that shows input token step 1: start to get Hugging Face Access Token.

      2. Select Open to open the Hugging Face website.

        Screenshot that shows input token step 2: open Hugging Face websites.

      3. Get your token on Hugging Face portal and paste it Quick Pick. Press Enter.

        Screenshot that shows input token step 3: input token on dropdown textbox.

    • Quantization Dataset Split: dataset could have different splits like validation, train and test.

    • Quantization Dataset Size: the number of data used to quantize the model.

    For more information about activation and weight type, please see Data type selection.

    You could also disable this section. In this case, the workflow will only convert the model to ONNX format but do not quantize the model.

    Evaluation

    In this section, you need to select the Execution Provider (EP) you want to use for evaluation, regardless of the platform on which the model was converted.

    • Evaluate on: the target device that you want to evaluate the model. Possible values are:
      • Qualcomm NPU: to use this, you need a compatible Qualcomm device.
      • AMD NPU: to use this, you need a device with a supported AMD NPU.
      • Intel NPU: to use this, you need a device with a supported Intel NPU.
      • CPU: any CPU could work.
    • Evaluation Dataset: dataset used for evaluation.
    • Evaluation Dataset Split: dataset could have different splits like validation, train and test.
    • Evaluation Dataset Size: the number of data used to evaluate the model.

    You could also disable this section. In this case, the workflow will only convert the model to ONNX format but do not evaluate the model.

  3. Run the workflow by selecting Run

    A default job name is generated using the workflow name and timestamp (e.g., bert_qdq_2025-08-04_20-45-00) for easy tracking.

    During the job running, you can Cancel the job by selecting the status indicator or the three-dot menu under Action in History board and select Stop Running.

    Hugging Face compliance alerts: During the quantization, we need the calibration datasets. You may be prompted to accept license terms before proceeding. If you missed the notification, the running process will be paused, waiting for your input. Please make sure notifications are enabled and that you accept the required licenses.

Note

Model conversion and quantization: you can run workflow on any device expect for LLM models. The Quantization configuration is optimized for NPU only. It's recommended to uncheck this step if the target system is not NPU.

LLM model quantization: If you want to quantize the LLM models, a Nvidia GPU is required.

If you want to quantize the model on another device with GPU, you can setup environment by yourselves, please refer ManualConversionOnGPU. Please note that only "Quantization" step need the GPU. After quantization, you can evaluate the model on NPU or CPU.

Tips for re-evaluation

After a model has been successfully converted, you could use the re-evaluate function to perform evaluation again without the model conversion.

Go to the History board and find the model run job. Select the three-dot menu under Action to Re-evaluate the model.

You can choose the different EPs or datasets for re-evaluation

Screenshot that shows re-evaluation. It contains configurations such as name, system and datasets settings.

Tips for failed jobs

If your job is canceled or failed, you can select job name to adjust the workflow and run job again. To avoid accidental overwrites, each execution creates a new history folder with its own configuration and results.

View results

The History Board in Conversion is your central dashboard for tracking, reviewing, and managing all workflow runs. Each time you run a model conversion and evaluation, a new entry is created in the History Board—ensuring full traceability and reproducibility.

  • Find the workflow run that you want to review. Each run is listed with a status indicator (e.g. Succeeded, Cancelled)
  • Select the run name to view the conversion configurations
  • Select the logs under Status indicator to view logs and detailed execution results
  • Once the model converted successfully, you can view the evaluation results under Metrics. Metrics such as accuracy, latency and throughput are displayed alongside each run

Screenshot that shows history, including name, time, parameters and so on.

Use sample notebook for model inference

  • Go to the History board. Select the three-dot menu under Action.

    Select Inference in Samples from the dropdown.

    Screenshot that shows actions, including inference, copy model path and re-evaluate.

  • Choose the Python environment

    • You'll be prompted to select a Python virtual environment. The default runtime is: C:\Users\{user_name}\.aitk\bin\model_lab_runtime\Python-WCR-win32-x64-3.12.9.
    • Note that the default runtime contains everything needed, otherwise, manually install the requirements.txt
  • The sample will launch in a Jupyter Notebook. You can customize the input data or parameters to test different scenarios.

Tip

Model compatibility: Ensure the converted model supports the specified EPs in the inference samples

Sample location: Inference samples are stored alongside the run artifacts in the history folder.

Export and share with others

Go to the History board. Select Export to share the model project with others. This copies the model project without history folder. If you want to share models with others, select the corresponding jobs. This copies the selected history folder containing the model and its configuration.

What you learned

In this article, you learned how to:

  • Create a model conversion project in AI Toolkit for VS Code.
  • Configure the conversion workflow, including quantization and evaluation settings.
  • Run the conversion workflow to transform a pre-built model into an optimized ONNX model.
  • View the results of the conversion, including metrics and logs.
  • Use the sample notebook for model inference and testing.
  • Export and share the model project with others.
  • Re-evaluate a model using different execution providers or datasets.
  • Handle failed jobs and adjust configurations for re-runs.
  • Understand the supported models and their requirements for conversion and quantization.

See also

县级市市长什么级别 穿旗袍配什么发型好看 六月十三是什么日子 乳房头疼是什么原因 糖化血红蛋白偏高是什么意思
包干价是什么意思 团县委是什么单位 骨折吃什么钙片 为什么会得风湿 冰激凌和冰淇淋有什么区别
卫生纸属于什么垃圾 什么山不能爬脑筋急转弯 榴莲不能和什么一起吃 无印良品属于什么档次 李约瑟难题是什么
7月14号是什么节日 e代表什么 结肠多发息肉是什么意思 马来西亚信仰什么教 牙龈肿了吃什么消炎药
用牛奶敷脸有什么好处和坏处hcv8jop3ns8r.cn 心什么诚服hcv7jop5ns1r.cn 手淫是什么意思hcv7jop5ns3r.cn 颈椎挂什么科室hcv9jop4ns3r.cn 猫咪疫苗什么时候打hcv9jop6ns5r.cn
狮子的天敌是什么动物naasee.com 英国全称是什么jiuxinfghf.com 尿酸偏高有什么危害sanhestory.com tb是什么hcv9jop0ns3r.cn 火奥念什么hcv8jop3ns3r.cn
疼痛门诊看什么病hcv8jop2ns1r.cn 系统性红斑狼疮不能吃什么hcv9jop6ns9r.cn 地米是什么药hcv8jop4ns6r.cn 蹭饭是什么意思hcv8jop1ns7r.cn 失眠是什么引起的zhongyiyatai.com
姓什么的人最多wzqsfys.com chick什么意思hcv8jop0ns0r.cn 为什么一饿就胃疼hcv9jop4ns5r.cn 日本是什么时候投降的hcv9jop5ns4r.cn 一月份什么星座hcv8jop4ns0r.cn
百度