From 10d042a34d00c91dfe1fe0b9c1508c4ef9c479f4 Mon Sep 17 00:00:00 2001 From: Ekaterina Aidova Date: Thu, 7 Mar 2024 08:49:33 +0400 Subject: [PATCH] prepare to 2024.0 (#1793) --- .docker/Pipfile | 4 +- .docker/Pipfile.lock | 17 ++--- README.md | 14 ++++- README_cn.md | 2 +- .../104-model-tools/104-model-tools.ipynb | 16 ++++- notebooks/108-gpu-device/108-gpu-device.ipynb | 63 ++++++++++++++++++- .../121-convert-to-openvino.ipynb | 25 +++++++- .../121-legacy-mo-convert-to-openvino.ipynb | 2 +- ...age-classification-nncf-quantization.ipynb | 14 ++++- .../208-optical-character-recognition.ipynb | 2 +- .../211-speech-to-text.ipynb | 2 +- .../222-vision-image-colorization.ipynb | 2 +- ...36-stable-diffusion-v2-text-to-image.ipynb | 2 +- .../246-depth-estimation-videpth.ipynb | 29 +++++++-- .../261-fast-segment-anything.ipynb | 2 +- .../270-sound-generation-audioldm2.ipynb | 3 +- ...mos2-multimodal-large-language-model.ipynb | 19 +++++- .../283-photo-maker/283-photo-maker.ipynb | 30 ++++++--- .../401-object-detection.ipynb | 2 +- .../403-action-recognition-webcam.ipynb | 2 +- .../406-3D-pose-estimation.ipynb | 2 +- .../407-person-tracking.ipynb | 2 +- requirements.txt | 4 +- 23 files changed, 212 insertions(+), 48 deletions(-) diff --git a/.docker/Pipfile b/.docker/Pipfile index e8a36a204a8..6cc9f302e30 100644 --- a/.docker/Pipfile +++ b/.docker/Pipfile @@ -24,11 +24,11 @@ matplotlib = ">=3.4" monai = ">=0.9.1,<1.0.0" nbval = "*" notebook = "<7.0.0" -nncf = "==2.8.0" +nncf = "==2.9.0" numpy = ">=1.21.0" onnx = ">=1.11.0" opencv-python = "*" -openvino-dev = {version = "==2023.3.0"} +openvino-dev = {version = "==2024.0.0"} openvino-telemetry = "==2023.2.1" ovmsclient = "*" Pillow = ">=8.3.2" diff --git a/.docker/Pipfile.lock b/.docker/Pipfile.lock index ac0fc262faa..43bbcbe534c 100644 --- a/.docker/Pipfile.lock +++ b/.docker/Pipfile.lock @@ -1,7 +1,7 @@ { "_meta": { "hash": { - "sha256": "bfd1026fcde32c0b31dd94d90b3c4f1f8fd9e0d3dda635d273fb8a49eab08236" + "sha256": "3b96808ffb9aaf977206dbe04c4267e01f2b27057f1bfcca7c096219507e5e8e" }, "pipfile-spec": 6, "requires": { @@ -2363,11 +2363,11 @@ }, "nncf": { "hashes": [ - "sha256:5affe2c33d00feef54f1fc80184e5b79c16afcb8e28f07d71e89aa10cc11f046", - "sha256:a8e1d3091b530810de598ba557dcfb40ab6cdb0127ff62ba3b173e219a9082b4" + "sha256:13d92cdda7f326af1ae1e4a31e4e5778c0b1712474c4b58a46d8bbe3b91ff3a7", + "sha256:aedd6f35038297218f23c2f799c0f3d29ac078126d65b79a9fdca1e5c49ecb49" ], "index": "pypi", - "version": "==2.8.0" + "version": "==2.9.0" }, "notebook": { "hashes": [ @@ -2530,15 +2530,16 @@ "sha256:cac7a91a2d0e073350c50cfecfdc6de748395597e22ec10285a8bd393786b5dd", "sha256:dfd4cf70ce0c16ede7e43903167db1c6e469987d2cc980339f3e79f2ce786dd4", "sha256:e1b59ce212697eedd09b042aff6a5cfde6851ed2cfea1e9348d152b847776574", - "sha256:1f4d0bb3ae9e763d5fe983e2396d798c7392dec10eb4b06d9e82ebf94a4cea97" + "sha256:1f4d0bb3ae9e763d5fe983e2396d798c7392dec10eb4b06d9e82ebf94a4cea97", + "sha256:39f6b2c099fedc9fa99fdddf7305f85aefe9addbf9fe406f1a6d4bf01d37a13f" ], - "version": "==2023.3.0" + "version": "==2024.0.0" }, "openvino-dev": { "hashes": [ - "sha256:0dacd775c0047f2dd234cdc1f465b0e801532371893d24977f699515c35f2c4a" + "sha256:bf27ba90359d66997c51f91f49e99df5b942d474c3075723b13ce76e358ff475" ], - "version": "==2023.3.0" + "version": "==2024.0.0" }, "openvino-telemetry": { "hashes": [ diff --git a/README.md b/README.md index cbd53e80fd6..71d8b1657d3 100644 --- a/README.md +++ b/README.md @@ -20,6 +20,7 @@ A collection of ready-to-run Jupyter notebooks for learning and experimenting wi - [Table of Contents](#table-of-contents) - [📝 Installation Guide](#-installation-guide) +- [🚀 Getting Started](#-getting-started) - [⚙️ System Requirements](#️-system-requirements) - [💻 Run the Notebooks](#-run-the-notebooks) - [To Launch a Single Notebook](#to-launch-a-single-notebook) @@ -42,6 +43,17 @@ OpenVINO Notebooks require Python and Git. To get started, select the guide for [![-----------------------------------------------------](https://user-images.githubusercontent.com/10940214/155750931-fc094349-b6ec-4e1f-9f9a-113e67941119.jpg)]() +
+ +## 🚀 Getting Started + +Explore Jupyter notebooks using this [page](https://openvinotoolkit.github.io/openvino_notebooks/), select one related to your needs or give them all a try. Good Luck! + +**NOTE: The main branch of this repository was updated to support the new OpenVINO 2024.0 release.** To upgrade to the new release version, please run `pip install --upgrade -r requirements.txt` in your `openvino_env` virtual environment. If you need to install for the first time, see the [Installation Guide](#-installation-guide) section below. If you wish to use the previous release version of OpenVINO, please checkout the [2023.3 branch](https://github.com/openvinotoolkit/openvino_notebooks/tree/2023.3). If you wish to use the previous Long Term Support (LTS) version of OpenVINO check out the [2022.3 branch](https://github.com/openvinotoolkit/openvino_notebooks/tree/2022.3). + +If you need help, please start a GitHub [Discussion](https://github.com/openvinotoolkit/openvino_notebooks/discussions). + + If you run into issues, please check the [troubleshooting section](#-troubleshooting), [FAQs](#-faq) or start a GitHub [discussion](https://github.com/openvinotoolkit/openvino_notebooks/discussions). Notebooks with ![binder logo](https://mybinder.org/badge_logo.svg) and ![colab logo](https://colab.research.google.com/assets/colab-badge.svg) buttons can be run without installing anything. [Binder](https://mybinder.org/) and [Google Colab](https://colab.research.google.com/) are free online services with limited resources. For the best performance, please follow the [Installation Guide](#-installation-guide) and run the notebooks locally. @@ -171,7 +183,7 @@ Made with [`contrib.rocks`](https://contrib.rocks). ## ❓ FAQ -* [Which devices does OpenVINO support?](https://docs.openvino.ai/2023.3/openvino_docs_OV_UG_supported_plugins_Supported_Devices.html#doxid-openvino-docs-o-v-u-g-supported-plugins-supported-devices) +* [Which devices does OpenVINO support?](https://docs.openvino.ai/2024/openvino-workflow/running-inference/inference-devices-and-modes.html) * [What is the first CPU generation you support with OpenVINO?](https://www.intel.com/content/www/us/en/developer/tools/openvino-toolkit/system-requirements.html) * [Are there any success stories about deploying real-world solutions with OpenVINO?](https://www.intel.com/content/www/us/en/internet-of-things/ai-in-production/success-stories.html) diff --git a/README_cn.md b/README_cn.md index abf16221249..223705f6883 100644 --- a/README_cn.md +++ b/README_cn.md @@ -366,7 +366,7 @@ jupyter lab notebooks ## ❓ 常见问题解答 -* [OpenVINO支持哪些设备?](https://docs.openvino.ai/2023.3/openvino_docs_OV_UG_supported_plugins_Supported_Devices.html#doxid-openvino-docs-o-v-u-g-supported-plugins-supported-devices) +* [OpenVINO支持哪些设备?](https://docs.openvino.ai/2024/openvino-workflow/running-inference/inference-devices-and-modes.html) * [OpenVINO支持的第一代CPU是什么?](https://www.intel.com/content/www/us/en/developer/tools/openvino-toolkit/system-requirements.html) * [在使用OpenVINO部署现实世界解决方案方面有没有成功的案例?](https://www.intel.com/content/www/us/en/internet-of-things/ai-in-production/success-stories.html) diff --git a/notebooks/104-model-tools/104-model-tools.ipynb b/notebooks/104-model-tools/104-model-tools.ipynb index fe1493bb85c..23196386f07 100644 --- a/notebooks/104-model-tools/104-model-tools.ipynb +++ b/notebooks/104-model-tools/104-model-tools.ipynb @@ -1,6 +1,7 @@ { "cells": [ { + "attachments": {}, "cell_type": "markdown", "id": "artificial-discretion", "metadata": {}, @@ -25,6 +26,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "c4dda33e", "metadata": {}, @@ -61,10 +63,11 @@ ], "source": [ "# Install openvino package\n", - "%pip install -q \"openvino-dev>=2023.1.0\"" + "%pip install -q \"openvino-dev>=2024.0.0\"" ] }, { + "attachments": {}, "cell_type": "markdown", "id": "fec90f61-08f3-4417-8a6f-ec61c9e5955b", "metadata": { @@ -94,6 +97,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "f0206a22-dc33-4666-ab2b-86386b97caca", "metadata": { @@ -129,6 +133,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "parental-assets", "metadata": { @@ -178,6 +183,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "judicial-preview", "metadata": {}, @@ -188,6 +194,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "rising-interval", "metadata": {}, @@ -258,6 +265,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "proprietary-checklist", "metadata": {}, @@ -344,6 +352,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "aa8d655f-215d-4e3c-adcb-e8fd4a2e8ab4", "metadata": {}, @@ -413,6 +422,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "7ea7e868-fd2d-4d11-9c87-7aa1f1301083", "metadata": {}, @@ -445,6 +455,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "54e01154-f700-479f-9111-147c95595d46", "metadata": {}, @@ -536,6 +547,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "75891996-cf53-4c76-ad3c-5fb468ccd7bb", "metadata": {}, @@ -545,6 +557,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "88d64dd7-789d-4536-ab8f-84999c73afaf", "metadata": {}, @@ -614,6 +627,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "3896e4bf-f7d0-4529-be97-921ef548de2a", "metadata": {}, diff --git a/notebooks/108-gpu-device/108-gpu-device.ipynb b/notebooks/108-gpu-device/108-gpu-device.ipynb index 26af218c1c5..00c43c33816 100644 --- a/notebooks/108-gpu-device/108-gpu-device.ipynb +++ b/notebooks/108-gpu-device/108-gpu-device.ipynb @@ -1,6 +1,7 @@ { "cells": [ { + "attachments": {}, "cell_type": "markdown", "id": "85bb873e-562a-4d60-9536-b6b88aecb5c2", "metadata": { @@ -10,7 +11,8 @@ "# Working with GPUs in OpenVINO™\n", "\n", "\n", - "#### Table of contents:\n\n", + "#### Table of contents:\n", + "\n", "- [Introduction](#Introduction)\n", " - [Install required packages](#Install-required-packages)\n", "- [Checking GPUs with Query Device](#Checking-GPUs-with-Query-Device)\n", @@ -44,6 +46,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "d6c8277d-29b3-4172-8600-fcbbbbdbafd2", "metadata": {}, @@ -54,6 +57,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "3e7253ed-d8a1-4475-8e83-263336639157", "metadata": { @@ -65,6 +69,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "ab1f9f13-63a4-4b2d-95e3-9a82c1e9cebd", "metadata": {}, @@ -75,6 +80,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "7a419978", "metadata": {}, @@ -101,7 +107,7 @@ } ], "source": [ - "%pip install -q \"openvino-dev>=2023.1.0\"\n", + "%pip install -q \"openvino-dev>=2024.0.0\"\n", "%pip install -q tensorflow\n", "\n", "# Fetch `notebook_utils` module\n", @@ -113,6 +119,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "d2f35f17-5209-43d6-b3b5-db35efb1f42e", "metadata": { @@ -124,6 +131,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "c84c37c2-b24b-4c9f-9ab8-d3a60b8393ec", "metadata": {}, @@ -132,6 +140,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "f0e91abc-6118-4532-9817-6ef44c51b0c2", "metadata": { @@ -143,6 +152,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "e76b6152-3915-415a-987b-c30935462007", "metadata": {}, @@ -180,6 +190,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "a95aef31-bf6f-4d7e-85a5-97dc6bead645", "metadata": {}, @@ -188,6 +199,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "4e266c79-7bb9-4907-9204-87b688087cbf", "metadata": {}, @@ -196,6 +208,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "c0491c07-0b6d-483f-a4d6-d3f7b5ec0c81", "metadata": { @@ -207,6 +220,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "54399a50-31cf-48bd-bf98-6dea0c0b1b20", "metadata": {}, @@ -215,6 +229,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "bf7509b6-67e3-46fb-bb23-5f46c58b0fc1", "metadata": {}, @@ -251,6 +266,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "aac3129a-129f-49aa-aba0-71ae1e892ada", "metadata": {}, @@ -322,6 +338,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "88546b71-2ae8-4519-b9b7-a54f444e204f", "metadata": { @@ -333,6 +350,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "6524096f-cc77-4dca-9e52-e5a3c2c0523b", "metadata": {}, @@ -351,6 +369,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "131bd0f9-9b1e-4fc1-ad09-3960e1d6c50f", "metadata": { @@ -362,6 +381,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "cea15935-2373-4029-8bc0-998558a6defe", "metadata": {}, @@ -370,6 +390,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "db4600db-faea-49cc-b9b9-93415223a38f", "metadata": {}, @@ -379,6 +400,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "821e29d1-6021-4552-a926-cb35bd6af776", "metadata": {}, @@ -387,6 +409,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "ad24a39d-8819-47a3-9c91-0f80e92649b6", "metadata": { @@ -487,6 +510,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "7ffaa8d2-724b-47f5-991d-22bbe6d1c994", "metadata": { @@ -552,6 +576,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "4cc043aa-8729-45bc-9e98-6b3daea9f271", "metadata": { @@ -563,6 +588,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "6999d5e8-3429-4e2b-a845-c5569ec1895f", "metadata": {}, @@ -586,6 +612,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "18c58745-69c3-4034-b221-bc4cf8a5b27b", "metadata": {}, @@ -594,6 +621,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "2747c592-29f1-49e9-9ed4-5bf1a39e5c8a", "metadata": {}, @@ -603,6 +631,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "22069d88-1d29-4978-ba32-2613f9d2143d", "metadata": {}, @@ -650,6 +679,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "f7915734-8138-43bc-9334-8c81a919fa26", "metadata": {}, @@ -693,6 +723,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "7328633a-ba7d-422d-9f62-446ca7db142e", "metadata": {}, @@ -701,6 +732,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "2aa916d7-4626-4f54-9b34-954eca19aafb", "metadata": { @@ -712,6 +744,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "a77cbc96-8832-4ed9-9109-30b0efce9da9", "metadata": {}, @@ -720,6 +753,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "7077b662-22f3-4c52-9c80-e5ac1309c482", "metadata": {}, @@ -743,6 +777,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "06589f38-ce35-457f-8395-a4a3f6327ea0", "metadata": {}, @@ -766,6 +801,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "b0a98815-669e-4868-a5f3-7104d6887fb3", "metadata": { @@ -777,6 +813,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "0da18381-edfc-4d3b-942a-87d6b0cbd5d3", "metadata": {}, @@ -799,6 +836,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "99351910-c54c-4f55-b1ce-eb74125a9dbd", "metadata": { @@ -810,6 +848,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "93b411e5-5c56-462a-82e3-22ba020f64c1", "metadata": {}, @@ -818,6 +857,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "f5921100-0f41-42d3-8f90-703c20cb11bd", "metadata": {}, @@ -917,6 +957,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "4c27f62b-6053-4a8a-a554-064b04fa2ac7", "metadata": {}, @@ -925,6 +966,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "9d8a5020-9e9c-48ae-8386-8078e2bf08a2", "metadata": {}, @@ -1094,6 +1136,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "9432edbe-3078-4534-a0cc-6bfe4bc00b28", "metadata": {}, @@ -1263,6 +1306,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "bd8e3e87-bdbf-4424-85aa-28c9b0c144f9", "metadata": {}, @@ -1434,6 +1478,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "d3fd1496-996d-49d8-80b9-3bfb3dda9729", "metadata": { @@ -1445,6 +1490,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "53a25247-a69f-40a6-be52-ea6de89bf470", "metadata": {}, @@ -1453,6 +1499,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "e1686090-2a2e-4ae6-8f22-7ef458ce221b", "metadata": {}, @@ -1498,6 +1545,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "eadfb30e-d6d3-4ed3-8e7e-f1a843ce2ec0", "metadata": {}, @@ -1541,6 +1589,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "083bce3d-647d-46d1-8540-4a2935bc829f", "metadata": {}, @@ -1603,6 +1652,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "7003286a-8349-46f6-9d38-c6bb959e892a", "metadata": {}, @@ -1641,6 +1691,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "0ad10b97-e32e-4885-af1d-0636e970b707", "metadata": { @@ -1652,6 +1703,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "91f6ed58-8b3f-415e-a733-a7517e200b12", "metadata": {}, @@ -1686,6 +1738,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "2d2f6aba-b3ba-478b-b02f-6b438b575c33", "metadata": {}, @@ -1712,6 +1765,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "fb47ae59-b86d-46c4-b219-e7beccd5f796", "metadata": {}, @@ -1761,6 +1815,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "7fd90b68-ae14-4fe9-98c7-e8d23146c82c", "metadata": {}, @@ -1857,6 +1912,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "58fb6bb6-0d1c-4b0b-8a0b-efb5667816ea", "metadata": { @@ -1868,6 +1924,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "dbbb4f28-2917-45ca-8fa0-8c670c0348bc", "metadata": {}, @@ -1933,4 +1990,4 @@ }, "nbformat": 4, "nbformat_minor": 5 -} \ No newline at end of file +} diff --git a/notebooks/121-convert-to-openvino/121-convert-to-openvino.ipynb b/notebooks/121-convert-to-openvino/121-convert-to-openvino.ipynb index 80eef5fe3b4..f92d1f61755 100644 --- a/notebooks/121-convert-to-openvino/121-convert-to-openvino.ipynb +++ b/notebooks/121-convert-to-openvino/121-convert-to-openvino.ipynb @@ -1,6 +1,7 @@ { "cells": [ { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -9,7 +10,8 @@ "This notebook shows how to convert a model from original framework format to OpenVINO Intermediate Representation (IR).\n", "\n", "\n", - "#### Table of contents:\n\n", + "#### Table of contents:\n", + "\n", "- [OpenVINO IR format](#OpenVINO-IR-format)\n", "- [Fetching example models](#Fetching-example-models)\n", "- [Conversion](#Conversion)\n", @@ -34,10 +36,11 @@ "# Required imports. Please execute this cell first.\n", "%pip install --upgrade pip\n", "%pip install -q --extra-index-url https://download.pytorch.org/whl/cpu \\\n", - "\"openvino-dev>=2023.2.0\" \"requests\" \"tqdm\" \"transformers[onnx]>=4.21.1\" \"torch\" \"torchvision\" \"tensorflow_hub\" \"tensorflow\"" + "\"openvino-dev>=2024.0.0\" \"requests\" \"tqdm\" \"transformers[onnx]>=4.21.1\" \"torch\" \"torchvision\" \"tensorflow_hub\" \"tensorflow\"" ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -119,6 +122,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -145,6 +149,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -189,6 +194,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -210,6 +216,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -247,6 +254,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -311,6 +319,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -366,6 +375,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -416,6 +426,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -469,6 +480,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -519,6 +531,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -558,6 +571,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -569,6 +583,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -620,6 +635,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -666,6 +682,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -712,6 +729,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -751,6 +769,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -804,4 +823,4 @@ }, "nbformat": 4, "nbformat_minor": 4 -} \ No newline at end of file +} diff --git a/notebooks/121-convert-to-openvino/121-legacy-mo-convert-to-openvino.ipynb b/notebooks/121-convert-to-openvino/121-legacy-mo-convert-to-openvino.ipynb index 2dcc10a54e5..d949619a5a5 100644 --- a/notebooks/121-convert-to-openvino/121-legacy-mo-convert-to-openvino.ipynb +++ b/notebooks/121-convert-to-openvino/121-legacy-mo-convert-to-openvino.ipynb @@ -35,7 +35,7 @@ "source": [ "# Required imports. Please execute this cell first.\n", "%pip install -q --extra-index-url https://download.pytorch.org/whl/cpu \\\n", - "\"openvino-dev>=2023.1.0\" \"requests\" \"tqdm\" \"transformers[onnx]>=4.21.1\" \"torch\" \"torchvision\"" + "\"openvino-dev>=2024.0.0\" \"requests\" \"tqdm\" \"transformers[onnx]>=4.21.1\" \"torch\" \"torchvision\"" ] }, { diff --git a/notebooks/127-big-transfer-quantization/127-tensorflow-bit-image-classification-nncf-quantization.ipynb b/notebooks/127-big-transfer-quantization/127-tensorflow-bit-image-classification-nncf-quantization.ipynb index 4effa7775d9..65caa7857a2 100644 --- a/notebooks/127-big-transfer-quantization/127-tensorflow-bit-image-classification-nncf-quantization.ipynb +++ b/notebooks/127-big-transfer-quantization/127-tensorflow-bit-image-classification-nncf-quantization.ipynb @@ -1,6 +1,7 @@ { "cells": [ { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -39,8 +40,7 @@ } ], "source": [ - "%pip uninstall -q -y openvino-dev openvino openvino-nightly\n", - "%pip install -q \"openvino-nightly\" \"nncf>=2.7.0\" \"tensorflow-hub>=0.15.0\" \"tensorflow_datasets\"\n", + "%pip install -q \"openvino>=2024.0.0\" \"nncf>=2.7.0\" \"tensorflow-hub>=0.15.0\" \"tensorflow_datasets\"\n", "%pip install -q \"scikit-learn>=1.3.2\"" ] }, @@ -122,6 +122,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -188,6 +189,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -280,6 +282,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -324,6 +327,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -355,6 +359,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -403,6 +408,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -458,6 +464,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -573,6 +580,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -609,6 +617,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -641,6 +650,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ diff --git a/notebooks/208-optical-character-recognition/208-optical-character-recognition.ipynb b/notebooks/208-optical-character-recognition/208-optical-character-recognition.ipynb index 6d7be38fd24..d2964d6c869 100644 --- a/notebooks/208-optical-character-recognition/208-optical-character-recognition.ipynb +++ b/notebooks/208-optical-character-recognition/208-optical-character-recognition.ipynb @@ -43,7 +43,7 @@ "outputs": [], "source": [ "# Install openvino-dev package\n", - "%pip install -q \"openvino-dev>=2023.1.0\" onnx" + "%pip install -q \"openvino-dev>=2024.0.0\" onnx" ] }, { diff --git a/notebooks/211-speech-to-text/211-speech-to-text.ipynb b/notebooks/211-speech-to-text/211-speech-to-text.ipynb index d0ee93ca34a..a9a9b130684 100644 --- a/notebooks/211-speech-to-text/211-speech-to-text.ipynb +++ b/notebooks/211-speech-to-text/211-speech-to-text.ipynb @@ -52,7 +52,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install -q \"librosa>=0.8.1\" \"matplotlib<3.8\" \"openvino-dev>=2023.1.0\" \"numpy<1.24\" \"ruamel.yaml\"" + "%pip install -q \"librosa>=0.8.1\" \"matplotlib<3.8\" \"openvino-dev>=2024.0.0\" \"numpy<1.24\" \"ruamel.yaml\"" ] }, { diff --git a/notebooks/222-vision-image-colorization/222-vision-image-colorization.ipynb b/notebooks/222-vision-image-colorization/222-vision-image-colorization.ipynb index 7f6f803dced..0612e571f57 100644 --- a/notebooks/222-vision-image-colorization/222-vision-image-colorization.ipynb +++ b/notebooks/222-vision-image-colorization/222-vision-image-colorization.ipynb @@ -46,7 +46,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install \"openvino-dev>=2023.1.0\"" + "%pip install \"openvino-dev>=2024.0.0\"" ] }, { diff --git a/notebooks/236-stable-diffusion-v2/236-stable-diffusion-v2-text-to-image.ipynb b/notebooks/236-stable-diffusion-v2/236-stable-diffusion-v2-text-to-image.ipynb index 58a7eb4882a..7f8dcca578f 100644 --- a/notebooks/236-stable-diffusion-v2/236-stable-diffusion-v2-text-to-image.ipynb +++ b/notebooks/236-stable-diffusion-v2/236-stable-diffusion-v2-text-to-image.ipynb @@ -96,7 +96,7 @@ ], "source": [ "%pip install -q \"diffusers>=0.14.0\" \"openvino>=2023.1.0\" \"datasets>=2.14.6\" \"transformers>=4.25.1\" gradio --extra-index-url https://download.pytorch.org/whl/cpu\n", - "%pip install -q \"git+https://github.com/openvinotoolkit/nncf@develop\"" + "%pip install -q \"nncf>=2.9.0\"" ] }, { diff --git a/notebooks/246-depth-estimation-videpth/246-depth-estimation-videpth.ipynb b/notebooks/246-depth-estimation-videpth/246-depth-estimation-videpth.ipynb index 992d64109c5..6cd7c433948 100644 --- a/notebooks/246-depth-estimation-videpth/246-depth-estimation-videpth.ipynb +++ b/notebooks/246-depth-estimation-videpth/246-depth-estimation-videpth.ipynb @@ -1,6 +1,7 @@ { "cells": [ { + "attachments": {}, "cell_type": "markdown", "id": "98466560-036d-41d2-87f7-957af6f64626", "metadata": { @@ -36,7 +37,8 @@ "We will be consulting the [VI-Depth repository](https://github.com/isl-org/VI-Depth) for the pre-processing, model transformations and basic utility code. A part of it has already been kept as it is in the [utils](utils) directory. At the same time we will learn how to perform [model conversion](https://docs.openvino.ai/2023.3/openvino_docs_OV_Converter_UG_prepare_model_convert_model_Convert_Model_From_PyTorch.html) for converting a model in a different format to the standard OpenVINO™ IR model representation *via* another format.\n", "\n", "\n", - "#### Table of contents:\n\n", + "#### Table of contents:\n", + "\n", "- [Imports](#Imports)\n", "- [Loading models and checkpoints](#Loading-models-and-checkpoints)\n", " - [Cleaning up the model directory](#Cleaning-up-the-model-directory)\n", @@ -58,6 +60,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "2818c1d3-cd03-40f8-a03d-2101a14571fe", "metadata": {}, @@ -89,8 +92,7 @@ " \n", "# Download the correct version of the PyTorch deep learning library associated with image models\n", "# alongside the lightning module\n", - "%pip uninstall -q -y openvino-dev openvino openvino-nightly\n", - "%pip install -q --extra-index-url https://download.pytorch.org/whl/cpu \"pytorch-lightning\" \"timm>=0.6.12\" \"openvino-nightly\"" + "%pip install -q \"openvino>=2024.0.0\" --extra-index-url https://download.pytorch.org/whl/cpu \"pytorch-lightning\" \"timm>=0.6.12\" \"openvino-nightly\"" ] }, { @@ -138,6 +140,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "ba7ff380-dc99-45b5-a5bc-18472dd7e99c", "metadata": {}, @@ -147,6 +150,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "625575d8-7179-4c04-87a7-e678d82ab13c", "metadata": {}, @@ -279,6 +283,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "da8e0c34-015c-4da2-aee7-3c0c6234c46e", "metadata": { @@ -312,6 +317,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "38058667-3fd9-4178-aa51-79e0c3c41694", "metadata": {}, @@ -371,6 +377,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "8eca36f1-c98e-44a6-9a38-c38c8279e290", "metadata": {}, @@ -469,6 +476,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "ab725795-45fa-4253-b927-3cdbfd0a4e62", "metadata": {}, @@ -507,6 +515,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "1b072354-cd1d-450c-9089-4d01ec0270f7", "metadata": {}, @@ -555,6 +564,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "960d51a2-e918-49a2-b607-41442cc7f987", "metadata": { @@ -642,6 +652,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "9946594b-6990-4511-abca-8859af7cbd2a", "metadata": {}, @@ -783,6 +794,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "5e8d08c4-0ccf-428e-ac83-bc2ee9835efc", "metadata": {}, @@ -798,6 +810,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "69b5fba3-9d2d-4499-940c-8e3e196e5d4d", "metadata": {}, @@ -887,6 +900,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "e0813452-a1e9-436c-8661-2b238cfe4802", "metadata": {}, @@ -924,6 +938,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "c783f5ad-78ff-4beb-9386-db55811220c0", "metadata": {}, @@ -1011,6 +1026,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "b870c7b4-d3d0-4eea-86dd-21bd21567a1b", "metadata": { @@ -1101,6 +1117,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "0a79770e-fd46-4719-a240-065c8add229d", "metadata": {}, @@ -1130,6 +1147,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "44a97b51-5d3e-4903-8c73-53aa881d7c35", "metadata": {}, @@ -1179,6 +1197,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "0102713e-fd0b-4e08-858b-a3d7fd50d51f", "metadata": {}, @@ -1251,6 +1270,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "3825aa74-b825-4502-999f-21f46bf86c50", "metadata": { @@ -1275,6 +1295,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "19806ccf-34f5-44c5-8b57-ec000f558601", "metadata": {}, @@ -1361,4 +1382,4 @@ }, "nbformat": 4, "nbformat_minor": 5 -} \ No newline at end of file +} diff --git a/notebooks/261-fast-segment-anything/261-fast-segment-anything.ipynb b/notebooks/261-fast-segment-anything/261-fast-segment-anything.ipynb index 5d66d0271db..800105763e4 100644 --- a/notebooks/261-fast-segment-anything/261-fast-segment-anything.ipynb +++ b/notebooks/261-fast-segment-anything/261-fast-segment-anything.ipynb @@ -63,7 +63,7 @@ ], "source": [ "%pip install -q \"ultralytics==8.0.200\" onnx --extra-index-url https://download.pytorch.org/whl/cpu\n", - "%pip install -q \"openvino-dev>=2023.1.0\"\n", + "%pip install -q \"openvino-dev>=2024.0.0\"\n", "%pip install -q \"nncf>=2.6.0\"\n", "%pip install -q \"gradio>=4.13\"" ] diff --git a/notebooks/270-sound-generation-audioldm2/270-sound-generation-audioldm2.ipynb b/notebooks/270-sound-generation-audioldm2/270-sound-generation-audioldm2.ipynb index a53201d3bf8..311f0799682 100644 --- a/notebooks/270-sound-generation-audioldm2/270-sound-generation-audioldm2.ipynb +++ b/notebooks/270-sound-generation-audioldm2/270-sound-generation-audioldm2.ipynb @@ -62,9 +62,8 @@ } ], "source": [ - "%pip uninstall -q -y \"openvino-dev\" \"openvino\" \"openvino-nightly\" \n", "%pip install -q accelerate \"diffusers>=0.21.0\" transformers torch gradio \"peft==0.6.2\" --extra-index-url https://download.pytorch.org/whl/cpu\n", - "%pip install -q \"openvino-nightly\"" + "%pip install -q \"openvino>=2024.0.0\"" ] }, { diff --git a/notebooks/281-kosmos2-multimodal-large-language-model/281-kosmos2-multimodal-large-language-model.ipynb b/notebooks/281-kosmos2-multimodal-large-language-model/281-kosmos2-multimodal-large-language-model.ipynb index bde35d5743c..d0d6f83045d 100644 --- a/notebooks/281-kosmos2-multimodal-large-language-model/281-kosmos2-multimodal-large-language-model.ipynb +++ b/notebooks/281-kosmos2-multimodal-large-language-model/281-kosmos2-multimodal-large-language-model.ipynb @@ -1,6 +1,7 @@ { "cells": [ { + "attachments": {}, "cell_type": "markdown", "id": "14186cec-1749-4be0-96b9-56ab5518ddb6", "metadata": { @@ -25,6 +26,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "663df1e2-6679-42e3-9b84-c55f5b4dc5a2", "metadata": {}, @@ -45,6 +47,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "1a75c7ce-6f94-43ba-8087-11d7ad717a6f", "metadata": {}, @@ -61,13 +64,13 @@ "outputs": [], "source": [ "%pip install --upgrade pip\n", - "%pip uninstall -q -y openvino-dev openvino openvino-nightly\n", - "%pip install -q \"openvino-nightly\"\n", + "%pip install -q \"openvino>=2024.0.0\"\n", "%pip install -q \"transformers>=4.35\" Pillow gradio opencv-python\n", "%pip install -q --extra-index-url https://download.pytorch.org/whl/cpu torch torchvision" ] }, { + "attachments": {}, "cell_type": "markdown", "id": "28152b98-b619-4f1b-9164-dd9c726ffb82", "metadata": {}, @@ -158,6 +161,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "75a5d679d53082a5", "metadata": { @@ -309,6 +313,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "ee206f85-101d-4276-8e31-50611c9a48ad", "metadata": { @@ -325,6 +330,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "9d6f6512-b155-4d92-8394-0662af9048bc", "metadata": {}, @@ -350,6 +356,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "aa115992-fca3-420f-89cc-bc684df11532", "metadata": {}, @@ -392,6 +399,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "1cf0e035-8cc4-49c8-b8c4-cb3d9f7aea2f", "metadata": {}, @@ -413,6 +421,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "11093a02-cb43-48de-a500-7cfb150c2180", "metadata": {}, @@ -450,6 +459,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "c84783d6-5a58-4cac-a11e-a33d925f38fc", "metadata": { @@ -585,6 +595,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "217e0425-e00e-4e6b-af5f-6e5dcc5a90dd", "metadata": {}, @@ -594,6 +605,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "1b58a25b-b114-4e52-b2e3-b97ad56e2744", "metadata": { @@ -645,6 +657,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "007b9693-dfd3-4413-81cf-4f69da0cffd8", "metadata": {}, @@ -920,6 +933,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "1bae26d3-2d18-44d7-9148-1a404fdfa08d", "metadata": {}, @@ -1000,6 +1014,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "dcd0b8fd2ac36d1a", "metadata": { diff --git a/notebooks/283-photo-maker/283-photo-maker.ipynb b/notebooks/283-photo-maker/283-photo-maker.ipynb index 377bd3ca9c3..0c09a964c69 100644 --- a/notebooks/283-photo-maker/283-photo-maker.ipynb +++ b/notebooks/283-photo-maker/283-photo-maker.ipynb @@ -1,6 +1,7 @@ { "cells": [ { + "attachments": {}, "cell_type": "markdown", "id": "ebc080c9-4580-4a8c-bbfb-e391d2871b10", "metadata": {}, @@ -11,9 +12,10 @@ "\n", "\n", "This notebook explores how to speed up PhotoMaker pipeline using OpenVINO.\n", - " ", + " \n", + "\n", + "#### Table of contents:\n", "\n", - "#### Table of contents:\n\n", "- [PhotoMaker pipeline introduction](#PhotoMaker-pipeline-introduction)\n", "- [Prerequisites](#Prerequisites)\n", "- [Load original pipeline and prepare models for conversion](#Load-original-pipeline-and-prepare-models-for-conversion)\n", @@ -31,13 +33,14 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "4a41b1c9", "metadata": {}, "source": [ "## PhotoMaker pipeline introduction\n", "[back to top ⬆️](#Table-of-contents:)\n", - "", + "\n", "\n", "For the proposed PhotoMaker,\n", "we first obtain the text embedding and image embeddings from `text encoder(s)` and `image(ID) encoder`, respectively. Then, we extract the\n", @@ -51,6 +54,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "7509c1ba-a099-49fd-a1c5-7118f2d0c4e6", "metadata": {}, @@ -75,6 +79,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "8648f92f", "metadata": {}, @@ -103,12 +108,12 @@ } ], "source": [ - "%pip uninstall -q -y openvino-dev openvino openvino-nightly\n", "%pip install -q --extra-index-url https://download.pytorch.org/whl/cpu\\\n", - "transformers diffusers gradio openvino-nightly torchvision" + "transformers diffusers gradio \"openvino>=2024.0.0\" torchvision" ] }, { + "attachments": {}, "cell_type": "markdown", "id": "5959be5b", "metadata": {}, @@ -134,6 +139,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "4ead4649-6aea-4c52-bec7-a87dd1a1251c", "metadata": {}, @@ -233,6 +239,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "7719901d-caed-4c4f-b89f-32eac02a4712", "metadata": {}, @@ -318,6 +325,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "fb4c8b72-3d6b-4643-ba56-7a3103725bae", "metadata": {}, @@ -369,6 +377,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "9c896614-c369-48fe-a04a-65456458c7dd", "metadata": {}, @@ -434,6 +443,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "49ae9c7e-15e6-4304-bd93-f493011503d9", "metadata": {}, @@ -520,6 +530,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "a74680d5-9bf1-438b-8d3b-bd91a8dc3802", "metadata": {}, @@ -572,6 +583,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "61c721a4-de35-4d62-b910-98927486580c", "metadata": {}, @@ -625,13 +637,14 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "c9c9170c", "metadata": {}, "source": [ "### Compile models and create their Wrappers for inference\n", "[back to top ⬆️](#Table-of-contents:)\n", - "", + "\n", "\n", "To access original PhotoMaker workflow, we have to create a new wrapper for each OpenVINO compiled model. For matching original pipeline, part of OpenVINO model wrapper's attributes should be reused from original model objects and inference output must be converted from numpy to `torch.tensor`.\n", "\n", @@ -768,6 +781,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "ed6bef08-e5c3-4d82-9541-627035413590", "metadata": {}, @@ -790,6 +804,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "dda95419", "metadata": {}, @@ -915,6 +930,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "235a6c7d-1f5c-4cb2-b3ad-829eee29327f", "metadata": {}, @@ -1090,4 +1106,4 @@ }, "nbformat": 4, "nbformat_minor": 5 -} \ No newline at end of file +} diff --git a/notebooks/401-object-detection-webcam/401-object-detection.ipynb b/notebooks/401-object-detection-webcam/401-object-detection.ipynb index f5aadb0c33b..726f2546cb9 100644 --- a/notebooks/401-object-detection-webcam/401-object-detection.ipynb +++ b/notebooks/401-object-detection-webcam/401-object-detection.ipynb @@ -45,7 +45,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install -q \"openvino-dev>=2023.1.0\"\n", + "%pip install -q \"openvino-dev>=2024.0.0\"\n", "%pip install -q tensorflow\n", "%pip install -q opencv-python requests tqdm\n", "\n", diff --git a/notebooks/403-action-recognition-webcam/403-action-recognition-webcam.ipynb b/notebooks/403-action-recognition-webcam/403-action-recognition-webcam.ipynb index 4cc45932e67..31ced3e5025 100644 --- a/notebooks/403-action-recognition-webcam/403-action-recognition-webcam.ipynb +++ b/notebooks/403-action-recognition-webcam/403-action-recognition-webcam.ipynb @@ -48,7 +48,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install -q \"openvino-dev>=2023.1.0\"" + "%pip install -q \"openvino-dev>=2024.0.0\"" ] }, { diff --git a/notebooks/406-3D-pose-estimation-webcam/406-3D-pose-estimation.ipynb b/notebooks/406-3D-pose-estimation-webcam/406-3D-pose-estimation.ipynb index 530c8fb0dd8..0cf64636750 100644 --- a/notebooks/406-3D-pose-estimation-webcam/406-3D-pose-estimation.ipynb +++ b/notebooks/406-3D-pose-estimation-webcam/406-3D-pose-estimation.ipynb @@ -51,7 +51,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install pythreejs \"openvino-dev>=2023.1.0\"" + "%pip install pythreejs \"openvino-dev>=2024.0.0\"" ] }, { diff --git a/notebooks/407-person-tracking-webcam/407-person-tracking.ipynb b/notebooks/407-person-tracking-webcam/407-person-tracking.ipynb index 78542fd1f88..85adeca101a 100644 --- a/notebooks/407-person-tracking-webcam/407-person-tracking.ipynb +++ b/notebooks/407-person-tracking-webcam/407-person-tracking.ipynb @@ -71,7 +71,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install -q \"openvino-dev>=2023.1.0\"\n", + "%pip install -q \"openvino-dev>=2024.0.0\"\n", "%pip install -q opencv-python matplotlib requests scipy" ] }, diff --git a/requirements.txt b/requirements.txt index 8c64c033c18..2adaa4d58f2 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,7 +1,7 @@ # openvino and its dependencies -openvino-dev>=2023.3.0 +openvino-dev>=2024.0.0 openvino-telemetry>=2023.2.1 -nncf==2.8.0 +nncf==2.9.0 # deep learning frameworks