mirror of
https://github.com/microsoft/vcpkg.git
synced 2024-12-19 07:57:48 +08:00
9c7ddcae77
* [vcpkg baseline][openvino] Disable find pybind11 * update version
177 lines
4.4 KiB
JSON
177 lines
4.4 KiB
JSON
{
|
|
"$schema": "https://raw.githubusercontent.com/microsoft/vcpkg-tool/main/docs/vcpkg.schema.json",
|
|
"name": "openvino",
|
|
"version-date": "2023-06-11",
|
|
"port-version": 2,
|
|
"maintainers": "OpenVINO Developers <openvino@intel.com>",
|
|
"summary": "This is a port for Open Visual Inference And Optimization toolkit for AI inference",
|
|
"description": [
|
|
"Intel® Distribution of OpenVINO™ toolkit is an open-source toolkit for optimizing ",
|
|
"and deploying AI inference. It can be used to develop applications and solutions based ",
|
|
"on deep learning tasks, such as: emulation of human vision, automatic speech recognition, ",
|
|
"natural language processing, recommendation systems, etc. It provides high-performance ",
|
|
"and rich deployment options, from edge to cloud"
|
|
],
|
|
"homepage": "https://github.com/openvinotoolkit/openvino",
|
|
"documentation": "https://docs.openvino.ai/latest/index.html",
|
|
"license": "Apache-2.0",
|
|
"dependencies": [
|
|
"ade",
|
|
{
|
|
"name": "pkgconf",
|
|
"host": true
|
|
},
|
|
"pugixml",
|
|
{
|
|
"name": "tbb",
|
|
"platform": "static | osx",
|
|
"version>=": "2021.9.0#2"
|
|
},
|
|
{
|
|
"name": "tbb",
|
|
"features": [
|
|
"hwloc"
|
|
],
|
|
"platform": "!static & !osx",
|
|
"version>=": "2021.9.0#2"
|
|
},
|
|
{
|
|
"name": "vcpkg-cmake",
|
|
"host": true
|
|
},
|
|
{
|
|
"name": "vcpkg-cmake-config",
|
|
"host": true
|
|
},
|
|
{
|
|
"name": "vcpkg-get-python-packages",
|
|
"host": true
|
|
},
|
|
{
|
|
"name": "xbyak",
|
|
"platform": "!(arm | uwp)",
|
|
"version>=": "6.69"
|
|
}
|
|
],
|
|
"default-features": [
|
|
"auto",
|
|
"auto-batch",
|
|
"default-features",
|
|
"hetero",
|
|
"ir",
|
|
"onnx",
|
|
"paddle",
|
|
"pytorch",
|
|
"tensorflow",
|
|
"tensorflow-lite"
|
|
],
|
|
"features": {
|
|
"auto": {
|
|
"description": "Enables Auto plugin for inference"
|
|
},
|
|
"auto-batch": {
|
|
"description": "Enables Auto Batch plugin for inference, useful for throughput mode"
|
|
},
|
|
"cpu": {
|
|
"description": "Enables CPU plugin for inference",
|
|
"supports": "!(windows & arm)"
|
|
},
|
|
"default-features": {
|
|
"description": "Enables all default features",
|
|
"dependencies": [
|
|
{
|
|
"name": "openvino",
|
|
"features": [
|
|
"cpu"
|
|
],
|
|
"platform": "!(windows & arm)"
|
|
},
|
|
{
|
|
"name": "openvino",
|
|
"features": [
|
|
"gpu"
|
|
],
|
|
"platform": "x64 & !(osx | uwp) & !static"
|
|
}
|
|
]
|
|
},
|
|
"gpu": {
|
|
"description": "Enables GPU plugin for inference",
|
|
"supports": "x64 & !(osx | uwp) & !static",
|
|
"dependencies": [
|
|
"opencl"
|
|
]
|
|
},
|
|
"hetero": {
|
|
"description": "Enables Hetero plugin for inference"
|
|
},
|
|
"ir": {
|
|
"description": "Enables IR frontend for reading models in OpenVINO IR format"
|
|
},
|
|
"onnx": {
|
|
"description": "Enables ONNX frontend for reading models in ONNX format",
|
|
"dependencies": [
|
|
{
|
|
"name": "onnx",
|
|
"version>=": "1.13.1"
|
|
},
|
|
{
|
|
"name": "protobuf",
|
|
"version>=": "3.20.3"
|
|
},
|
|
{
|
|
"name": "protobuf",
|
|
"host": true,
|
|
"version>=": "3.20.3"
|
|
}
|
|
]
|
|
},
|
|
"paddle": {
|
|
"description": "Enables PaddlePaddle frontend for reading models in PaddlePaddle format",
|
|
"dependencies": [
|
|
{
|
|
"name": "protobuf",
|
|
"version>=": "3.20.3"
|
|
},
|
|
{
|
|
"name": "protobuf",
|
|
"host": true,
|
|
"version>=": "3.20.3"
|
|
}
|
|
]
|
|
},
|
|
"pytorch": {
|
|
"description": "Enables PyTorch frontend to convert models in PyTorch format"
|
|
},
|
|
"tensorflow": {
|
|
"description": "Enables TensorFlow frontend for reading models in TensorFlow format",
|
|
"dependencies": [
|
|
{
|
|
"name": "protobuf",
|
|
"version>=": "3.20.3"
|
|
},
|
|
{
|
|
"name": "protobuf",
|
|
"host": true,
|
|
"version>=": "3.20.3"
|
|
},
|
|
"snappy"
|
|
]
|
|
},
|
|
"tensorflow-lite": {
|
|
"description": "Enables TensorFlow Lite frontend for reading models in TensorFlow Lite format",
|
|
"dependencies": [
|
|
{
|
|
"name": "flatbuffers",
|
|
"version>=": "2.0.6"
|
|
},
|
|
{
|
|
"name": "flatbuffers",
|
|
"host": true,
|
|
"version>=": "2.0.6"
|
|
}
|
|
]
|
|
}
|
|
}
|
|
}
|