{"id":87359,"date":"2023-11-13T17:08:33","date_gmt":"2023-11-13T09:08:33","guid":{"rendered":"https:\/\/www.seeedstudio.com\/blog\/?p=87359"},"modified":"2024-01-05T03:36:13","modified_gmt":"2024-01-05T03:36:13","slug":"deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training","status":"publish","type":"post","link":"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/","title":{"rendered":"Deploy CLIP Zero-shot Classification on Jetson Orin &#8211; Fast Locate and Categorize Data without Training"},"content":{"rendered":"\n<p>The traditional method of image classification is quite time and resource-consuming. It requires more than millions of labeled images as a huge dataset preparation, which is highly consistent with your targeted object and can be used to train the supervised classification model for the next step.<\/p>\n\n\n\n<p>However, supervised training can not always deliver good performance in general usage. Suppose you will test the model with another image in a different domain or include a similar object within an unfamiliar application scenario. In that case, it may identify the object as the unexpected class. That&#8217;s where CLIP(Contrastive Language\u2013Image Pretraining) unleashes the power of multimodal learning at a fast speed.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Understand of CLIP<\/h2>\n\n\n\n<p>So, what is <a href=\"https:\/\/openai.com\/research\/clip\">CLIP<\/a>? CLIP, which stands for Contrastive Language-Image Pretraining, is a computer vision model developed by OpenAI. It excels in understanding the relationship between images and their corresponding textual descriptions, built on a large body of work on zero-shot transfer, natural language supervision, and multimodal learning.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Basic Architecture <\/h3>\n\n\n\n<h4 class=\"wp-block-heading\">Contrastive learning in pre-training<\/h4>\n\n\n\n<p>CLIP utilizes the Vision Transformer (ViT) architecture to process images, enabling the model to capture global relationships among image features. Meanwhile, for text processing, CLIP employs a transformer-based architecture, tokenizing and processing text through transformer layers to understand semantic relationships within textual descriptions without predicting texts word for word.<\/p>\n\n\n\n<p>The model is trained using pairs of images and text, aiming to minimize the distance between positive pairs (comprising an image and its corresponding description) and maximize the distance between negative pairs (comprising an image and a randomly chosen description). This contrastive learning objective encourages the model to learn a meaningful representation where related content is close and unrelated content is separated.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-large is-resized\"><img fetchpriority=\"high\" decoding=\"async\" src=\"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2024\/01\/WeChatWorkScreenshot_2996bae5-d5ec-4468-a7b4-26264f9306d2-1030x723.png\" alt=\"\" class=\"wp-image-90705\" width=\"515\" height=\"362\" srcset=\"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2024\/01\/WeChatWorkScreenshot_2996bae5-d5ec-4468-a7b4-26264f9306d2-1030x723.png 1030w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2024\/01\/WeChatWorkScreenshot_2996bae5-d5ec-4468-a7b4-26264f9306d2-300x211.png 300w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2024\/01\/WeChatWorkScreenshot_2996bae5-d5ec-4468-a7b4-26264f9306d2-768x539.png 768w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2024\/01\/WeChatWorkScreenshot_2996bae5-d5ec-4468-a7b4-26264f9306d2-1024x719.png 1024w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2024\/01\/WeChatWorkScreenshot_2996bae5-d5ec-4468-a7b4-26264f9306d2.png 1194w\" sizes=\"(max-width: 515px) 100vw, 515px\" \/><figcaption class=\"wp-element-caption\">Image from Paper &#8220;<a href=\"https:\/\/arxiv.org\/pdf\/2103.00020.pdf\">Learning Transferable Visual Models From Natural Language Supervision<\/a>&#8220;<\/figcaption><\/figure><\/div>\n\n\n<h4 class=\"wp-block-heading\">Fine-tune for downstream tasks<\/h4>\n\n\n\n<p>After pre-training, CLIP can be fine-tuned on specific downstream tasks with task-specific datasets. You can easily tailor CLIP&#8217;s behavior to specific tasks and domains with prompt engineering. By utilizing a prompt template or even prompt ensembling, which uses multiple prompt templates with more contextual modifications added to match possible situations, the accuracy of similarity matching can be improved.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Why is CLIP important as a foundation model?<\/h3>\n\n\n\n<p>Unlike traditional models that require extensive labeled datasets for training, CLIP leverages a unique approach\u2014pretraining on a vast dataset containing image-text pairs from the Internet, enabling it to perform a wide array of tasks without the need for task-specific training data. The model&#8217;s ability to connect vision and language makes it adept at tasks such as image classification and even generating textual descriptions for images.<\/p>\n\n\n\n<p>You only need to define the possible prompts or descriptions the objects of the scene belong to, then CLIP will help you predict the most probable class for the given image or video based on its extensive pertaining.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-large is-resized\"><img decoding=\"async\" src=\"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/WeChatWorkScreenshot_256a071a-6c26-4f77-acb3-e7224bcc377e-1030x304.png\" alt=\"\" class=\"wp-image-87411\" width=\"773\" height=\"228\" srcset=\"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/WeChatWorkScreenshot_256a071a-6c26-4f77-acb3-e7224bcc377e-1030x304.png 1030w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/WeChatWorkScreenshot_256a071a-6c26-4f77-acb3-e7224bcc377e-300x88.png 300w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/WeChatWorkScreenshot_256a071a-6c26-4f77-acb3-e7224bcc377e-768x226.png 768w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/WeChatWorkScreenshot_256a071a-6c26-4f77-acb3-e7224bcc377e-1536x453.png 1536w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/WeChatWorkScreenshot_256a071a-6c26-4f77-acb3-e7224bcc377e-2048x604.png 2048w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/WeChatWorkScreenshot_256a071a-6c26-4f77-acb3-e7224bcc377e-1024x302.png 1024w\" sizes=\"(max-width: 773px) 100vw, 773px\" \/><\/figure><\/div>\n\n\n<p>In this guide, we are going to show application insights and running performance using CLIP on the Edge.   Basically, CLIP can be deployed for:<\/p>\n\n\n\n<p><strong>Advertisement search engine: <\/strong>classify video advertisement&#8217;s category, featured by shooting scenarios or specific main objects.<\/p>\n\n\n\n<p><strong>Content Moderation:<\/strong> identify and flag potentially inappropriate or harmful content in images and videos.<\/p>\n\n\n\n<p><strong>SNS video recommendation: <\/strong>classify the video content and deliver recommendations based on audiences&#8217; interest or viewing history.<\/p>\n\n\n\n<p><strong>Pinpoint important periods<\/strong> <strong>of the event in a long video:<\/strong> improve the efficiency for police to find the crime evidence on time.<\/p>\n\n\n\n<p><strong>Automation of the process in<\/strong> <strong>QSR<\/strong>: check what ingredients and sauces should be put in the delivered pizza, and provide a production list according to the visual comparison<\/p>\n\n\n\n<p><strong>Interior home design reference<\/strong>: help you discover visually similar interior design ideas based on your provided preferred decorating style images.<\/p>\n\n\n\n<p>We&#8217;ll see that the industry and application direction of CLIP is unlimited! So now, let&#8217;s dive deep into how to deploy CLIP at the edge to solve your image classification and other project tasks.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Deploy CLIP on the Site for Classification<\/h2>\n\n\n\n<p>Here are some of the interesting demos that we have deployed CLIP models on the <a href=\"https:\/\/www.seeedstudio.com\/reComputer-J4012-p-5586.html?queryID=89e05bcd3e3afa67b506f8373a2129ed&amp;objectID=5586&amp;indexName=bazaar_retailer_products\">reComputer J4012 based on NVIDIA Jetson Orin NX 16GB<\/a> using the <a href=\"https:\/\/inference.roboflow.com\/\">Roboflow Inference Server<\/a>. You can take reference of <a href=\"https:\/\/blog.roboflow.com\/roboflow-inference-server-clip\/\">Roboflow&#8217;s blog<\/a> for step-by-step guidance, however, just need to change the first Inference server installation part to the following code as adding TensorRT support, since we are using GPU: <\/p>\n\n\n\n<pre class=\"wp-block-code\"><code>```sh\ngit clone https:\/\/github.com\/roboflow\/inference\ncd inference\nsed -i '\/ENV OPENBLAS_CORETYPE=ARMV8\/a ENV ONNXRUNTIME_EXECUTION_PROVIDERS=TensorrtExecutionProvider' docker\/dockerfiles\/Dockerfile.onnx.jetson.5.1.1\ndocker build \\\n    -f docker\/dockerfiles\/Dockerfile.onnx.jetson.5.1.1 \\\n    -t roboflow\/roboflow-inference-server-jetson-5.1.1:seeed1 .\n```\n\n- Run the container \n\n```sh\ndocker run --privileged --net=host --runtime=nvidia roboflow\/roboflow-inference-server-jetson-5.1.1:seeed1\n```<\/code><\/pre>\n\n\n\n<h3 class=\"wp-block-heading\">Image-Prompt pair<\/h3>\n\n\n\n<p>Now you can easily classify images into different categories without training your model in advance. <\/p>\n\n\n\n<p>Start the Roboflow Inference Server Docker container on the Jetson Edge device, define the Roboflow API key, and run a demo script to start classifying different images based on the prompts you define. We have used Jetson Orin NX 16GB for this demo and it was able to achieve around 22 FPS performance with TensorRT FP16 precision!<\/p>\n\n\n\n<p>All you need to consider is prompt engineering &#8211; to find a more accurate or proper prompt to describe the scene of the image clearly for better classification performance, which means you need to tell CLIP what it should recognize in the image. This whole process of finding the correct prompt could be a long-term trial and error.<\/p>\n\n\n\n<p>If you also want to show the inferencing speed as our demo performed, feel free to add this part easily in the result out print part:<\/p>\n\n\n\n<pre class=\"wp-block-code\"><code>for file in sorted(os.listdir(IMAGE_DIR)):\n    image = f\"{IMAGE_DIR}\/{file}\"\n    predictions = classify_image(image)\n    print(get_highest_prediction(predictions&#91;\"similarity\"]), image)\n    print(1\/(predictions&#91;'time']))<\/code><\/pre>\n\n\n\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" src=\"https:\/\/files.seeedstudio.com\/products\/NVIDIA-Jetson\/CLIP-classify-photo.gif\" alt=\"\"\/><\/figure>\n\n\n\n<h3 class=\"wp-block-heading\">Video theme classification<\/h3>\n\n\n\n<p>Understanding a video scenario involves breaking down a video into individual frames, applying CLIP independently to each frame for image understanding, and then integrating temporal information across frames. <\/p>\n\n\n\n<p>By fusing features extracted from both individual frames and their temporal context, the model creates a representation that captures the overall content and context of the video. The fused features are then used for scenario classification, predicting the depicted activity or scenario. Fine-tuning a dataset containing video frames and scenario labels may be necessary to adapt CLIP to the specific requirements of the video task. The performance of the model is evaluated using standard metrics, and adjustments are made as needed for optimal results.<\/p>\n\n\n\n<p>In this demo, we use CLIP to identify the video as a scene containing a package. We can classify the scenario type, at what timestamps the package first appears, and how long it&#8217;s visible on the scene. It could be a perfect experiment for preventing package delivery from being stolen.&nbsp;Simply follow <a href=\"https:\/\/blog.roboflow.com\/how-to-analyze-and-classify-video-with-clip\/\">Roboflow&#8217;s blog<\/a> to classify the video step-by-step, also remember to import the supervision library at the beginning since we are going to use supervision to split up our video into frames.<\/p>\n\n\n\n<pre class=\"wp-block-code\"><code>import supervision as sv<\/code><\/pre>\n\n\n\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" src=\"https:\/\/files.seeedstudio.com\/products\/NVIDIA-Jetson\/CLIP-classify-video.gif\" alt=\"\"\/><\/figure>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<h3 class=\"wp-block-heading\">Seeed:&nbsp;NVIDIA&nbsp;Jetson Ecosystem Partner<\/h3>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter is-resized\"><img decoding=\"async\" src=\"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/08\/nvidia-elite-partner-badge-rgb-for-screen.jpg\" alt=\"\" class=\"wp-image-82789\" width=\"233\" height=\"109\" srcset=\"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/08\/nvidia-elite-partner-badge-rgb-for-screen.jpg 465w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/08\/nvidia-elite-partner-badge-rgb-for-screen-300x140.jpg 300w\" sizes=\"(max-width: 233px) 100vw, 233px\" \/><\/figure><\/div>\n\n\n<p>Seeed is an Elite partner for edge AI in the&nbsp;<a href=\"https:\/\/www.nvidia.com\/en-us\/about-nvidia\/partners\/\"><u>NVIDIA Partner Network<\/u><\/a>. Explore more carrier boards, full system devices, customization services, use cases, and developer tools on&nbsp;<a href=\"https:\/\/www.seeedstudio.com\/nvidia-jetson.html\"><u>Seeed\u2019s&nbsp;NVIDIA Jetson ecosystem<\/u><\/a>&nbsp;page.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter is-resized\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/09\/WeChatWorkScreenshot_0fa3f479-5b1c-4e23-9896-b989d6bde69c-3-1030x599.png\" alt=\"\" class=\"wp-image-84528\" width=\"773\" height=\"449\" srcset=\"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/09\/WeChatWorkScreenshot_0fa3f479-5b1c-4e23-9896-b989d6bde69c-3-1030x599.png 1030w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/09\/WeChatWorkScreenshot_0fa3f479-5b1c-4e23-9896-b989d6bde69c-3-300x174.png 300w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/09\/WeChatWorkScreenshot_0fa3f479-5b1c-4e23-9896-b989d6bde69c-3-768x446.png 768w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/09\/WeChatWorkScreenshot_0fa3f479-5b1c-4e23-9896-b989d6bde69c-3-1536x893.png 1536w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/09\/WeChatWorkScreenshot_0fa3f479-5b1c-4e23-9896-b989d6bde69c-3-1024x595.png 1024w, https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/09\/WeChatWorkScreenshot_0fa3f479-5b1c-4e23-9896-b989d6bde69c-3.png 1958w\" sizes=\"(max-width: 773px) 100vw, 773px\" \/><\/figure><\/div>\n\n\n<p>Join the forefront of AI innovation with us! Harness the power of cutting-edge hardware and technology to revolutionize the deployment of machine learning in the real world across industries. Be a part of our mission to provide developers and enterprises with the best ML solutions available. Check out our successful&nbsp;<a href=\"https:\/\/files.seeedstudio.com\/wiki\/NVIDIA\/NVIDIA_Jetson_example-Success_cases_and_examples_with_NVIDIA_Jetson.pdf\">case study catalog<\/a>&nbsp;to discover more edge AI possibilities!<\/p>\n\n\n\n<p>Take the first step and send us an email at&nbsp;<a href=\"mailto:edgeai@seeed.cc\">edgeai@seeed.cc<\/a>&nbsp;to become a part of this exciting journey!&nbsp;<\/p>\n\n\n\n<p>Download our latest&nbsp;<a href=\"https:\/\/files.seeedstudio.com\/wiki\/Seeed_Jetson\/Seeed-NVIDIA_Jetson_Catalog_V1.4.pdf\">Jetson Catalog<\/a>&nbsp;to find one option that suits you well. If you can\u2019t find the off-the-shelf Jetson hardware solution for your needs, please check out our&nbsp;<a href=\"https:\/\/www.seeedstudio.com\/odm\">customization services<\/a>, and submit a new product inquiry to us at&nbsp;odm@seeed.cc&nbsp;for evaluation.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>The traditional method of image classification is quite time and resource-consuming. It requires more than<\/p>\n","protected":false},"author":3606,"featured_media":87404,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_lmt_disableupdate":"","_lmt_disable":"","_price":"","_stock":"","_tribe_ticket_header":"","_tribe_default_ticket_provider":"","_tribe_ticket_capacity":"0","_ticket_start_date":"","_ticket_end_date":"","_tribe_ticket_show_description":"","_tribe_ticket_show_not_going":false,"_tribe_ticket_use_global_stock":"","_tribe_ticket_global_stock_level":"","_global_stock_mode":"","_global_stock_cap":"","_tribe_rsvp_for_event":"","_tribe_ticket_going_count":"","_tribe_ticket_not_going_count":"","_tribe_tickets_list":"[]","_tribe_ticket_has_attendee_info_fields":false,"iawp_total_views":0,"footnotes":""},"categories":[1,4393],"tags":[],"class_list":["post-87359","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-news","category-tech"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v24.0 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Deploy CLIP Zero-shot Classification on Jetson Orin &#8211; Fast Locate and Categorize Data without Training - Latest News from Seeed Studio<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Deploy CLIP Zero-shot Classification on Jetson Orin &#8211; Fast Locate and Categorize Data without Training - Latest News from Seeed Studio\" \/>\n<meta property=\"og:description\" content=\"The traditional method of image classification is quite time and resource-consuming. It requires more than\" \/>\n<meta property=\"og:url\" content=\"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/\" \/>\n<meta property=\"og:site_name\" content=\"Latest News from Seeed Studio\" \/>\n<meta property=\"article:published_time\" content=\"2023-11-13T09:08:33+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2024-01-05T03:36:13+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"1200\" \/>\n\t<meta property=\"og:image:height\" content=\"628\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Jennie Wang\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Jennie Wang\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"7 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/\",\"url\":\"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/\",\"name\":\"Deploy CLIP Zero-shot Classification on Jetson Orin &#8211; Fast Locate and Categorize Data without Training - Latest News from Seeed Studio\",\"isPartOf\":{\"@id\":\"https:\/\/www.seeedstudio.com\/blog\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg\",\"datePublished\":\"2023-11-13T09:08:33+00:00\",\"dateModified\":\"2024-01-05T03:36:13+00:00\",\"author\":{\"@id\":\"https:\/\/www.seeedstudio.com\/blog\/#\/schema\/person\/21041ae3908bbb4d44533f2b3b115fd1\"},\"breadcrumb\":{\"@id\":\"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/#primaryimage\",\"url\":\"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg\",\"contentUrl\":\"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg\",\"width\":1200,\"height\":628},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/www.seeedstudio.com\/blog\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Deploy CLIP Zero-shot Classification on Jetson Orin &#8211; Fast Locate and Categorize Data without Training\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/www.seeedstudio.com\/blog\/#website\",\"url\":\"https:\/\/www.seeedstudio.com\/blog\/\",\"name\":\"Latest News from Seeed Studio\",\"description\":\"Emerging IoT, AI and Autonomous Applications on the Edge\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/www.seeedstudio.com\/blog\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/www.seeedstudio.com\/blog\/#\/schema\/person\/21041ae3908bbb4d44533f2b3b115fd1\",\"name\":\"Jennie Wang\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/www.seeedstudio.com\/blog\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/b8fdf0c9ad5c32ab4f3981bb35a10566?s=96&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/b8fdf0c9ad5c32ab4f3981bb35a10566?s=96&r=g\",\"caption\":\"Jennie Wang\"},\"description\":\"Seeed Studio AIoT Marketing and Partnership Always coffee always alive \u2615\ufe0f\",\"sameAs\":[\"www.linkedin.com\/in\/jialinwang1215\"],\"url\":\"https:\/\/www.seeedstudio.com\/blog\/author\/jennie-wang\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Deploy CLIP Zero-shot Classification on Jetson Orin &#8211; Fast Locate and Categorize Data without Training - Latest News from Seeed Studio","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/","og_locale":"en_US","og_type":"article","og_title":"Deploy CLIP Zero-shot Classification on Jetson Orin &#8211; Fast Locate and Categorize Data without Training - Latest News from Seeed Studio","og_description":"The traditional method of image classification is quite time and resource-consuming. It requires more than","og_url":"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/","og_site_name":"Latest News from Seeed Studio","article_published_time":"2023-11-13T09:08:33+00:00","article_modified_time":"2024-01-05T03:36:13+00:00","og_image":[{"width":1200,"height":628,"url":"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg","type":"image\/jpeg"}],"author":"Jennie Wang","twitter_card":"summary_large_image","twitter_misc":{"Written by":"Jennie Wang","Est. reading time":"7 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/","url":"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/","name":"Deploy CLIP Zero-shot Classification on Jetson Orin &#8211; Fast Locate and Categorize Data without Training - Latest News from Seeed Studio","isPartOf":{"@id":"https:\/\/www.seeedstudio.com\/blog\/#website"},"primaryImageOfPage":{"@id":"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/#primaryimage"},"image":{"@id":"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/#primaryimage"},"thumbnailUrl":"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg","datePublished":"2023-11-13T09:08:33+00:00","dateModified":"2024-01-05T03:36:13+00:00","author":{"@id":"https:\/\/www.seeedstudio.com\/blog\/#\/schema\/person\/21041ae3908bbb4d44533f2b3b115fd1"},"breadcrumb":{"@id":"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/#primaryimage","url":"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg","contentUrl":"https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg","width":1200,"height":628},{"@type":"BreadcrumbList","@id":"https:\/\/www.seeedstudio.com\/blog\/2023\/11\/13\/deploy-clip-zero-shot-classification-on-jetson-orin-fast-locate-and-categorize-data-without-training\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/www.seeedstudio.com\/blog\/"},{"@type":"ListItem","position":2,"name":"Deploy CLIP Zero-shot Classification on Jetson Orin &#8211; Fast Locate and Categorize Data without Training"}]},{"@type":"WebSite","@id":"https:\/\/www.seeedstudio.com\/blog\/#website","url":"https:\/\/www.seeedstudio.com\/blog\/","name":"Latest News from Seeed Studio","description":"Emerging IoT, AI and Autonomous Applications on the Edge","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/www.seeedstudio.com\/blog\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Person","@id":"https:\/\/www.seeedstudio.com\/blog\/#\/schema\/person\/21041ae3908bbb4d44533f2b3b115fd1","name":"Jennie Wang","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.seeedstudio.com\/blog\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/b8fdf0c9ad5c32ab4f3981bb35a10566?s=96&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/b8fdf0c9ad5c32ab4f3981bb35a10566?s=96&r=g","caption":"Jennie Wang"},"description":"Seeed Studio AIoT Marketing and Partnership Always coffee always alive \u2615\ufe0f","sameAs":["www.linkedin.com\/in\/jialinwang1215"],"url":"https:\/\/www.seeedstudio.com\/blog\/author\/jennie-wang\/"}]}},"modified_by":"Jennie Wang","views":9947,"featured_image_urls":{"full":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg",1200,628,false],"thumbnail":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01-80x80.jpg",80,80,true],"medium":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01-300x157.jpg",300,157,true],"medium_large":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01-768x402.jpg",640,335,true],"large":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01-1030x539.jpg",640,335,true],"1536x1536":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg",1200,628,false],"2048x2048":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg",1200,628,false],"visody_icon":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg",32,17,false],"magazine-7-slider-full":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01.jpg",1200,628,false],"magazine-7-slider-center":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01-936x628.jpg",936,628,true],"magazine-7-featured":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01-1024x536.jpg",1024,536,true],"magazine-7-medium":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01-720x380.jpg",720,380,true],"magazine-7-medium-square":["https:\/\/www.seeedstudio.com\/blog\/wp-content\/uploads\/2023\/11\/01-675x450.jpg",675,450,true]},"author_info":{"display_name":"Jennie Wang","author_link":"https:\/\/www.seeedstudio.com\/blog\/author\/jennie-wang\/"},"category_info":"<a href=\"https:\/\/www.seeedstudio.com\/blog\/category\/news\/\" rel=\"category tag\">News<\/a> <a href=\"https:\/\/www.seeedstudio.com\/blog\/category\/tech\/\" rel=\"category tag\">Tech<\/a>","tag_info":"Tech","comment_count":"0","_links":{"self":[{"href":"https:\/\/www.seeedstudio.com\/blog\/wp-json\/wp\/v2\/posts\/87359","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.seeedstudio.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.seeedstudio.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.seeedstudio.com\/blog\/wp-json\/wp\/v2\/users\/3606"}],"replies":[{"embeddable":true,"href":"https:\/\/www.seeedstudio.com\/blog\/wp-json\/wp\/v2\/comments?post=87359"}],"version-history":[{"count":21,"href":"https:\/\/www.seeedstudio.com\/blog\/wp-json\/wp\/v2\/posts\/87359\/revisions"}],"predecessor-version":[{"id":90706,"href":"https:\/\/www.seeedstudio.com\/blog\/wp-json\/wp\/v2\/posts\/87359\/revisions\/90706"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.seeedstudio.com\/blog\/wp-json\/wp\/v2\/media\/87404"}],"wp:attachment":[{"href":"https:\/\/www.seeedstudio.com\/blog\/wp-json\/wp\/v2\/media?parent=87359"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.seeedstudio.com\/blog\/wp-json\/wp\/v2\/categories?post=87359"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.seeedstudio.com\/blog\/wp-json\/wp\/v2\/tags?post=87359"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}