{"id":1016,"date":"2019-11-05T08:00:52","date_gmt":"2019-11-05T07:00:52","guid":{"rendered":"http:\/\/cmagazine.test\/understanding-ai-inference\/"},"modified":"2021-12-23T13:02:21","modified_gmt":"2021-12-23T12:02:21","slug":"understanding-ai-inference","status":"publish","type":"post","link":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/","title":{"rendered":"Understanding AI: Inference"},"content":{"rendered":"<p><strong>Inferencing<\/strong> is the second phase of <strong><span id=\"urn:batch-analysis-0d46a94d-7c5e-4d3d-875f-7c77a0aa9d15\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/machine_learning\">machine learning<\/span><\/strong>, following on from the initial <strong>training<\/strong> phase. During the training phase, the algorithm generates a new model or repurposes a pre-trained model for a specific application and helps the model learn its parameters. During the inferencing phase, predictions and decisions on new <span id=\"urn:batch-analysis-951f5ac2-24e2-4127-873d-45a93d6dad85\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/data\">data<\/span> are made \u2013 based on the learned parameters.<\/p>\n<p>Learning requires a significant amount of time, computation power and electricity. In contrast, the inferencing phase requires less processing and draws less power too. However, the traditional way of computing in the central cloud may be just too resource-intensive for IoT devices. Each IoT node residing at the edge collects large datasets, making edge-to-cloud (and conversely cloud-to-edge) data transfer expensive and slow. Instead of relying on the cloud-based servers to do all the processing, \u201c<strong><span id=\"urn:batch-analysis-c0ead682-354e-472f-9943-49b4af92aee9\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/computing\">computing<\/span> at the edge<\/strong>\u201d performs most calculations directly and only transfers relevant information back to the cloud (and vice versa) when completely necessary. While computing at the edge reduces data transfer costs and time, this model also has certain drawbacks. For example, the need for IoT devices to be power-efficient runs contrary to the hefty amount of processing power that learning and inferencing demand. This is a problem that accelerators for AI <span id=\"urn:batch-analysis-e4548f33-bbd7-4077-a0a3-eee0d177c114\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/edge_computing\">edge computing<\/span> can potentially address.<\/p>\n<h2>AI Accelerators<\/h2>\n<p>Both hardware- and software-based AI accelerators expedite <span id=\"urn:batch-analysis-e78023c3-9ec7-4d89-a975-21b4fc0afdb5\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/machine_learning\">machine learning<\/span>. Hardware acceleration can target training, inferencing or possibly both. In some instances, the hardware can reduce the power requirement. In other cases, the hardware can improve the processing capacity. Several main types of chips, or processing units, exist for hardware acceleration. These include central processing units (CPUs), graphics processing units (GPUs), field-programmable gate arrays (<a href=\"https:\/\/www.codemotion.com\/magazine\/dev-hub\/machine-learning-dev\/fpgas-why-use-them\/\" target=\"_blank\" rel=\"noopener\">FPGAs)<\/a>, system-on-chips (SoCs), application-specific integrated circuits (ASICs), vision processing units (VPCs) and neuromorphic ICs. In addition to hardware acceleration, solutions on the market also comprise software-based approaches \u2013 like <span id=\"urn:batch-analysis-961a4509-bdde-4b47-b539-d959f258e6d7\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/machine_learning\">machine learning<\/span> frameworks for improving AI software development and optimizing system performance.<\/p>\n<h2>CPUs and GPUs<\/h2>\n<p>CPUs are what AI traditionally uses. While CPUs are designed to be all-purpose, they are often inadequate in supporting the massive calculations used in model generation and inferencing. In response, <span id=\"urn:batch-analysis-f08e8602-648e-49cc-9eeb-80d49be31462\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/ember_company\">companies<\/span> including ARM (with its DynamIQ product offering) and Samsung (with its Exynos 9 series), have started making AI-specific chips. While ARM and Samsung have chosen to stick with AI-specific CPUs, others are shifting toward GPUs.<\/p>\n<p>Originating in the video gaming industry and built for processing massive datasets, GPUs are a good match for <span id=\"urn:batch-analysis-7173c941-8a3d-4f05-927f-3e1ea7ca72d3\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/machine_learning\">machine learning<\/span>. Because GPUs have more processing units per chip and higher throughput, plus more parallel processing capability than CPUs, they cut down computation time significantly. In addition, a GPU\u2019s single processing unit weighs less than the multiple units CPUs use, making GPUs a better fit for constrained IoT devices, which require small and nimble components. The <span id=\"urn:batch-analysis-52c75fdc-a88b-4e3c-a543-503baa687a80\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/ember_company\">companies<\/span> that are making AI-specific GPUs include AMD (Radeon Instinct), NEC (SX-Aurora), NVIDIA (DGX) and Qualcomm (Adreno).<\/p>\n<h2>FPGAs<\/h2>\n<p>While CPUs and GPUs have considerable processing power at their disposal and are effective for accelerating learning and inferencing, they spend a lot of time and energy moving <span id=\"urn:batch-analysis-07b37c54-4c3d-4dcf-af24-795cf7c444bb\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/data\">data<\/span> between memory and processing. Since CPUs and GPUs are densely packed with circuits, they can often overheat and cause system failures. For remotely located IoT devices, the combination of high energy consumption and potential system failures is far from ideal. It makes sense to find a way to offload some tasks to more energy-efficient hardware.<\/p>\n<p>Based on programmable logic, FPGAs are a type of IC that can be reconfigured by customers or designers in the field after production. While generally not as powerful as CPUs or GPUs, FPGAs offer fast processing for some calculations (such as multiplication, addition, integration, differentiation and exponentials) by computing inside the chip instead of transferring <span id=\"urn:batch-analysis-23d0980f-1c0c-4a88-9d4f-9859ff8a217e\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/data\">data<\/span>. Although an FPGA offers more flexibility, it tends to be quite bulky, so miniaturization for IoT devices is a challenge for this type of chip. The major <span id=\"urn:batch-analysis-d524f9a9-9a00-429b-821a-0cb9ca1bae59\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/ember_company\">companies<\/span> that offer AI-targeted inference chips include NVIDIA (TensorRT) and Xilinx. Also, Microsoft is using FPGA chips to accelerate inference, and Intel is currently expanding its <a href=\"https:\/\/eu.mouser.com\/Search\/Refine?Ntk=P_MarCom&amp;Ntt=175224723\" target=\"_blank\" rel=\"noopener nofollow noreferrer\">FPGA portfolio<\/a>.<\/p>\n<h2>SoCs<\/h2>\n<p>SoCs contains can contain a combination of electronic components (microprocessors, microcontrollers, digital signal processors, on-chip memory, hardware accelerators, etc.). Due to the integration of the components onto a single semiconductor substrate, a SoC is more powerful than a microcontroller chip. In a <span id=\"urn:batch-analysis-697e4603-e4d9-4515-bb91-bde72af01aab\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/smartphone\">smartphone<\/span>, the SoC might integrate video, audio and image processing capabilities. ARM has developed its <span id=\"urn:batch-analysis-c3061f55-435c-4c3d-b30e-7064276aa79c\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/machine_learning\">Machine Learning<\/span> Processor and its Object Detection Processor \u2013 and these will be incorporated into SoCs in the future. HiSilicon, a Huawei-backed company, has licensed the IP from ARM to to make SoCs that are seeing preliminary utilization in phone handsets and tablets. Also, HiSilicon is making the Ascend chips for Huawei. Another big player in the SoC space is Arteris, which is developing a network-on-chip interconnect fabric technology (FlexNoC) that many mobile and wireless <span id=\"urn:batch-analysis-6ee7d290-a621-4c9f-9d70-fb79a9924e5c\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/ember_company\">companies<\/span> are using. Because Arteris holds a dominant position in IP, it has a bird\u2019s-eye view of the space. Other <span id=\"urn:batch-analysis-02d5e968-298d-4fff-b150-0e4315839244\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/ember_company\">companies<\/span> likely to soon be making a play in the AI SoC market include Intel (via its Movidius subsidiary), NXP, Renesas, Toshiba, Texas Instruments and STMicroelectronics.<\/p>\n<h2>ASICs, VPUs and Neuromorphic Chips<\/h2>\n<p>ASICs are specifically built for accelerating the training of deep learning algorithms, with examples including <span id=\"urn:batch-analysis-a9c69aa3-d112-40c8-a711-33fb46730c41\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/google\">Google<\/span>\u2019s Edge TPU and Intel\u2019s Nervana. A vision processing unit (VPU) is designed to accelerate machine vision tasks and run machine vision algorithms, such as convolutional neural networks (CNNs) \u2013 so VPU video processing capabilities differ from those of a GPU, which does not offer the same type of task-specific processing. Examples of VPUs include Intel\u2019s Movidius Myriad chips, <span id=\"urn:batch-analysis-97923a9a-004c-4351-8138-d6763f16e98c\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/google\">Google<\/span>\u2019s Pixel Visual Core, Microsoft\u2019s HoloLens, Inuitive\u2019s NU series and Mobileye\u2019s EyeQ.<\/p>\n<p>Digital chips and analog chips have their respective deficiencies: digital circuitry is precise but gobbles energy, while analog circuitry keeps both latency and energy consumption low but lacks precision. Therefore, researchers are looking for ways to combine the technical advantages of digital and analog chips while sidestepping the weaknesses. Inspired by the human brain, neuromorphic chips are designed to adhere to what is essentially a digital architecture, but use analog circuitry for mixed-<span id=\"urn:batch-analysis-4161b0b7-3859-4e54-b0e4-7cf783c23024\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/bandwidth_signal_processing\">signal processing<\/span>. <span id=\"urn:batch-analysis-86efc13b-7667-4cc8-8099-1ce85e10cda7\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/ibm\">IBM<\/span>\u2019s TrueNorth is a neuromorphic processor targeting sensor data pattern recognition and intelligence tasks. Also, Columbia University, Stanford University\u2019s \u2018Brains in Silicon\u2019 project, and the DARPA-backed University of Michigan IC Lab are all working on various aspects of neuromorphic system implementation.<\/p>\n<h2>Machine Learning Frameworks<\/h2>\n<p>AI accelerators also include <span id=\"urn:batch-analysis-823b9d5d-e8cc-4f9c-acdf-7825b50455f8\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/cogito_software\">software<\/span>. For example, <span id=\"urn:batch-analysis-37b9c1c7-d3ab-4c39-8097-a50dbf960ddc\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/machine_learning\">machine learning<\/span> frameworks, which can be interfaces, libraries or tools, help reduce the <span id=\"urn:batch-analysis-99623cfb-5c48-4424-861c-99c14774fc74\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/complexity\">complexity<\/span> associated with <span id=\"urn:batch-analysis-df156556-cdff-4475-a18f-b5d5d7f809f0\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/machine_learning\">machine learning<\/span> so that developers can build models and optimize <span id=\"urn:batch-analysis-e3cb00ef-a435-40db-a5fd-481557b23bbe\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/performance\">performance<\/span> more quickly and easily. Such frameworks are built to specific languages, like Python or <span id=\"urn:batch-analysis-61ce9f40-324a-4adc-8580-95cbc5f3ccee\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/java\">Java<\/span>. Some of the most popular open-source machine learning frameworks come from Amazon (AWS), Apache, Caffe2, Keras, Theano, <span id=\"urn:batch-analysis-04bfb2d5-b9b1-4d3b-bef9-d8e4bfe25ded\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/microsoft_azure\">Microsoft (Azure<\/span>) and <span id=\"urn:batch-analysis-e7c7a2da-11b5-4cef-966c-affcd6f31701\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/google\">Google<\/span> (TensorFlow). Also, some <span id=\"urn:batch-analysis-d696ab55-1240-497a-8563-2f90fa53a8f5\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/ember_company\">companies<\/span> offer in-house platforms. For example, Intel\u2019s OpenVINO toolkit is a software and hardware accelerator that optimizes inference with CNN models. In addition, Qualcomm\u2019s Snapdragon is a mobile platform and a software accelerator, <span id=\"urn:batch-analysis-f346005a-5abc-4a39-bd14-98243333d6b3\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/ibm\">IBM<\/span> has its Watson machine learning accelerator platform, and Huawei has recently launched its MindSpore AI framework.<\/p>\n<h2>First Steps with AI<\/h2>\n<p>Mouser now offers various items of hardware that can form the initial building blocks for AI implementation. Intel\u2019s plug-and-play <a href=\"https:\/\/eu.mouser.com\/new\/Intel\/intel-neural-compute-stick-2\/\" target=\"_blank\" rel=\"noopener nofollow noreferrer\">Neural&nbsp;Compute Stick 2<\/a> can aid engineers with early prototyping of deep neural networks. It relies on the <span id=\"urn:batch-analysis-204b86bd-d939-4082-bfa9-1bbadee36f88\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/ember_company\">company<\/span>\u2019s Movidius X VPU to deliver a compelling mix of power efficiency and <span id=\"urn:batch-analysis-587b57a6-a578-4db5-84e7-29467e3212fe\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/performance\">performance<\/span> \u2013 attaining 4TOPS. Targeted at industrial <span id=\"urn:batch-analysis-4ca75eae-f1bf-4614-bc80-6de67cbe5ebd\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/computing\">computing<\/span>, the highly compact <a href=\"https:\/\/eu.mouser.com\/new\/aaeon-up\/aaeon-up-ai-core\/\" target=\"_blank\" rel=\"noopener nofollow noreferrer\">AAEON UP AI Core processing module<\/a> is based on the mini-PCI Express format. It also features an Intel&nbsp;Movidius VPU (this time the Myriad 2 2450 \u2013 with 512MBytes of DDR memory, plus 12 VUW programmable SHAVE cores and dedicated vision accelerators all built in). The Gumstix <a href=\"https:\/\/eu.mouser.com\/new\/gumstix\/gumstix-nvidia-jetson-aerocore-2\/\" target=\"_blank\" rel=\"noopener nofollow noreferrer\">Aerocore 2 board<\/a> employs an array of NVIDIA Jetson TX1 and TX2 CUDA cores to give it strong parallel processing capabilities, along with an&nbsp;ARM Cortex-M4 microcontroller and numerous peripherals. It is particularly well suited to object recognition, production line inspection and various other kinds of machine vision.<\/p>\n<h2>Looking to the Future<\/h2>\n<p>With NVIDIA remaining dominant in industrial AI applications, most newcomers are focusing on the IoT AI space. GreenWave and Reduced Energy Microsystems are in the low-power chip arena, while Mythic and Syntiant are developing battery-powered processors. Similarly, Wiliot is making a Bluetooth chip that can be powered by ambient <span id=\"urn:batch-analysis-ca04720e-230b-4522-8cff-fa5c6901ab78\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/radio_frequency\">radio frequencies<\/span>. In the massive parallel data processing space, there are Vathys, Graphcore, Cerebras and Wave Computing. Meanwhile, Hailo Technologies and Horizon Robotics are working on specialized chips for autonomous vehicles. In the deep learning space, BrainChip has made the first spiking neural processor, Thinci has rolled out a streaming graph processor, and Gyrfalcon is developing a deep learning processor with proprietary AI processing in memory (APiM) technology. Lastly, at Groq, the ex-Googlers who designed <span id=\"urn:batch-analysis-dde0d249-5ff2-4d7f-9a54-ef316fa9b1f7\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/google\">Google<\/span>\u2019s TPU are developing a chip with ultra-low latency. As the field of <span id=\"urn:batch-analysis-355a88c9-59c6-4f71-bdee-ddd00c292702\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/machine_learning\">machine learning<\/span> witnesses astonishing progress, many technical challenges remain for IoT edge computing \u2013 with hardware and <span id=\"urn:batch-analysis-7de25718-4cb3-42c2-806a-25b29ac20554\" class=\"textannotation disambiguated wl-no-link wl-other\" itemid=\"http:\/\/data.wordlift.io\/wl01770\/entity\/software_developer\">software developers<\/span> continuing to reach for a superior processing performance\/energy efficiency balance.<\/p>\n<p>If you enjoyed this article and would like to read more on AI, other technology topics, new products, and more <a class=\"ek-link\" href=\"https:\/\/codemotion.typeform.com\/to\/TG9UMz\" target=\"_blank\" rel=\"noopener nofollow noreferrer\">SUBSCRIBE HERE.<\/a><\/p>\n<p><strong>Mouser Electronics<\/strong> is a worldwide leading authorised distributor of semiconductors and electronic components for over 800 industry-leading manufacturers. They specialise in the rapid introduction of new products and technologies for design engineers and buyers. Mouser Electronics extensive product offering includes semiconductors, interconnects, passives, and electromechanical components.<\/p>\n<h3>About the author<\/h3>\n<p><strong>Mark Patrick<\/strong> joined Mouser Electronics in July 2014 having previously held senior marketing roles at RS Components. Prior to RS, Mark spent 8 years at Texas Instruments in Applications Support and Technical Sales roles and holds a first class Honours Degree in Electronic Engineering from Coventry University.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Inferencing is the second phase of machine learning, following on from the initial training phase. During the training phase, the algorithm generates a new model or repurposes a pre-trained model for a specific application and helps the model learn its parameters. During the inferencing phase, predictions and decisions on new data are made \u2013 based&#8230; <a class=\"more-link\" href=\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/\">Read more<\/a><\/p>\n","protected":false},"author":83,"featured_media":2216,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_editorskit_title_hidden":false,"_editorskit_reading_time":6,"_editorskit_is_block_options_detached":false,"_editorskit_block_options_position":"{}","_uag_custom_page_level_css":"","_genesis_hide_title":false,"_genesis_hide_breadcrumbs":false,"_genesis_hide_singular_image":false,"_genesis_hide_footer_widgets":false,"_genesis_custom_body_class":"","_genesis_custom_post_class":"","_genesis_layout":"","footnotes":""},"categories":[46],"tags":[],"collections":[],"class_list":{"0":"post-1016","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-ai-ml","8":"entry"},"yoast_head":"<!-- This site is optimized with the Yoast SEO Premium plugin v26.9 (Yoast SEO v26.9) - https:\/\/yoast.com\/product\/yoast-seo-premium-wordpress\/ -->\n<title>Understanding AI: Inference - Codemotion Magazine<\/title>\n<meta name=\"description\" content=\"An overview about which IoT devices can support the process of inference in any AI\/Machine Learning application, and why some devices are better than others\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Understanding AI: Inference\" \/>\n<meta property=\"og:description\" content=\"An overview about which IoT devices can support the process of inference in any AI\/Machine Learning application, and why some devices are better than others\" \/>\n<meta property=\"og:url\" content=\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/\" \/>\n<meta property=\"og:site_name\" content=\"Codemotion Magazine\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/Codemotion.Italy\/\" \/>\n<meta property=\"article:published_time\" content=\"2019-11-05T07:00:52+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2021-12-23T12:02:21+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png\" \/>\n\t<meta property=\"og:image:width\" content=\"1146\" \/>\n\t<meta property=\"og:image:height\" content=\"675\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"Mark Patrick, Mouser Electronics\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@CodemotionIT\" \/>\n<meta name=\"twitter:site\" content=\"@CodemotionIT\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Mark Patrick, Mouser Electronics\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"8 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/\"},\"author\":{\"name\":\"Mark Patrick, Mouser Electronics\",\"@id\":\"https:\/\/www.codemotion.com\/magazine\/#\/schema\/person\/664e4da6990fc1344a2299435a542654\"},\"headline\":\"Understanding AI: Inference\",\"datePublished\":\"2019-11-05T07:00:52+00:00\",\"dateModified\":\"2021-12-23T12:02:21+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/\"},\"wordCount\":1639,\"publisher\":{\"@id\":\"https:\/\/www.codemotion.com\/magazine\/#organization\"},\"image\":{\"@id\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png\",\"articleSection\":[\"AI\/ML\"],\"inLanguage\":\"en-US\"},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/\",\"url\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/\",\"name\":\"Understanding AI: Inference - Codemotion Magazine\",\"isPartOf\":{\"@id\":\"https:\/\/www.codemotion.com\/magazine\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png\",\"datePublished\":\"2019-11-05T07:00:52+00:00\",\"dateModified\":\"2021-12-23T12:02:21+00:00\",\"description\":\"An overview about which IoT devices can support the process of inference in any AI\/Machine Learning application, and why some devices are better than others\",\"breadcrumb\":{\"@id\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#primaryimage\",\"url\":\"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png\",\"contentUrl\":\"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png\",\"width\":1146,\"height\":675,\"caption\":\"AI Inference\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/www.codemotion.com\/magazine\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"AI\/ML\",\"item\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/\"},{\"@type\":\"ListItem\",\"position\":3,\"name\":\"Machine Learning\",\"item\":\"https:\/\/www.codemotion.com\/magazine\/ai-ml\/machine-learning\/\"},{\"@type\":\"ListItem\",\"position\":4,\"name\":\"Understanding AI: Inference\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/www.codemotion.com\/magazine\/#website\",\"url\":\"https:\/\/www.codemotion.com\/magazine\/\",\"name\":\"Codemotion Magazine\",\"description\":\"We code the future. Together\",\"publisher\":{\"@id\":\"https:\/\/www.codemotion.com\/magazine\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/www.codemotion.com\/magazine\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/www.codemotion.com\/magazine\/#organization\",\"name\":\"Codemotion\",\"url\":\"https:\/\/www.codemotion.com\/magazine\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/www.codemotion.com\/magazine\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/codemotionlogo.png\",\"contentUrl\":\"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/codemotionlogo.png\",\"width\":225,\"height\":225,\"caption\":\"Codemotion\"},\"image\":{\"@id\":\"https:\/\/www.codemotion.com\/magazine\/#\/schema\/logo\/image\/\"},\"sameAs\":[\"https:\/\/www.facebook.com\/Codemotion.Italy\/\",\"https:\/\/x.com\/CodemotionIT\"]},{\"@type\":\"Person\",\"@id\":\"https:\/\/www.codemotion.com\/magazine\/#\/schema\/person\/664e4da6990fc1344a2299435a542654\",\"name\":\"Mark Patrick, Mouser Electronics\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/www.codemotion.com\/magazine\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/0d35fad9fee01e991637b67f54ae7cb8b001b5d2c1e4f7c1942b2105dad5a9bf?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/0d35fad9fee01e991637b67f54ae7cb8b001b5d2c1e4f7c1942b2105dad5a9bf?s=96&d=mm&r=g\",\"caption\":\"Mark Patrick, Mouser Electronics\"},\"url\":\"https:\/\/www.codemotion.com\/magazine\/author\/mark-patrick\/\"}]}<\/script>\n<!-- \/ Yoast SEO Premium plugin. -->","yoast_head_json":{"title":"Understanding AI: Inference - Codemotion Magazine","description":"An overview about which IoT devices can support the process of inference in any AI\/Machine Learning application, and why some devices are better than others","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/","og_locale":"en_US","og_type":"article","og_title":"Understanding AI: Inference","og_description":"An overview about which IoT devices can support the process of inference in any AI\/Machine Learning application, and why some devices are better than others","og_url":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/","og_site_name":"Codemotion Magazine","article_publisher":"https:\/\/www.facebook.com\/Codemotion.Italy\/","article_published_time":"2019-11-05T07:00:52+00:00","article_modified_time":"2021-12-23T12:02:21+00:00","og_image":[{"width":1146,"height":675,"url":"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png","type":"image\/png"}],"author":"Mark Patrick, Mouser Electronics","twitter_card":"summary_large_image","twitter_creator":"@CodemotionIT","twitter_site":"@CodemotionIT","twitter_misc":{"Written by":"Mark Patrick, Mouser Electronics","Est. reading time":"8 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#article","isPartOf":{"@id":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/"},"author":{"name":"Mark Patrick, Mouser Electronics","@id":"https:\/\/www.codemotion.com\/magazine\/#\/schema\/person\/664e4da6990fc1344a2299435a542654"},"headline":"Understanding AI: Inference","datePublished":"2019-11-05T07:00:52+00:00","dateModified":"2021-12-23T12:02:21+00:00","mainEntityOfPage":{"@id":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/"},"wordCount":1639,"publisher":{"@id":"https:\/\/www.codemotion.com\/magazine\/#organization"},"image":{"@id":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#primaryimage"},"thumbnailUrl":"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png","articleSection":["AI\/ML"],"inLanguage":"en-US"},{"@type":"WebPage","@id":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/","url":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/","name":"Understanding AI: Inference - Codemotion Magazine","isPartOf":{"@id":"https:\/\/www.codemotion.com\/magazine\/#website"},"primaryImageOfPage":{"@id":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#primaryimage"},"image":{"@id":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#primaryimage"},"thumbnailUrl":"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png","datePublished":"2019-11-05T07:00:52+00:00","dateModified":"2021-12-23T12:02:21+00:00","description":"An overview about which IoT devices can support the process of inference in any AI\/Machine Learning application, and why some devices are better than others","breadcrumb":{"@id":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#primaryimage","url":"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png","contentUrl":"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png","width":1146,"height":675,"caption":"AI Inference"},{"@type":"BreadcrumbList","@id":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/understanding-ai-inference\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/www.codemotion.com\/magazine\/"},{"@type":"ListItem","position":2,"name":"AI\/ML","item":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/"},{"@type":"ListItem","position":3,"name":"Machine Learning","item":"https:\/\/www.codemotion.com\/magazine\/ai-ml\/machine-learning\/"},{"@type":"ListItem","position":4,"name":"Understanding AI: Inference"}]},{"@type":"WebSite","@id":"https:\/\/www.codemotion.com\/magazine\/#website","url":"https:\/\/www.codemotion.com\/magazine\/","name":"Codemotion Magazine","description":"We code the future. Together","publisher":{"@id":"https:\/\/www.codemotion.com\/magazine\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/www.codemotion.com\/magazine\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/www.codemotion.com\/magazine\/#organization","name":"Codemotion","url":"https:\/\/www.codemotion.com\/magazine\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.codemotion.com\/magazine\/#\/schema\/logo\/image\/","url":"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/codemotionlogo.png","contentUrl":"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/codemotionlogo.png","width":225,"height":225,"caption":"Codemotion"},"image":{"@id":"https:\/\/www.codemotion.com\/magazine\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/Codemotion.Italy\/","https:\/\/x.com\/CodemotionIT"]},{"@type":"Person","@id":"https:\/\/www.codemotion.com\/magazine\/#\/schema\/person\/664e4da6990fc1344a2299435a542654","name":"Mark Patrick, Mouser Electronics","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.codemotion.com\/magazine\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/0d35fad9fee01e991637b67f54ae7cb8b001b5d2c1e4f7c1942b2105dad5a9bf?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/0d35fad9fee01e991637b67f54ae7cb8b001b5d2c1e4f7c1942b2105dad5a9bf?s=96&d=mm&r=g","caption":"Mark Patrick, Mouser Electronics"},"url":"https:\/\/www.codemotion.com\/magazine\/author\/mark-patrick\/"}]}},"featured_image_src":"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3-600x400.png","featured_image_src_square":"https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3-600x600.png","author_info":{"display_name":"Mark Patrick, Mouser Electronics","author_link":"https:\/\/www.codemotion.com\/magazine\/author\/mark-patrick\/"},"uagb_featured_image_src":{"full":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png",1146,675,false],"thumbnail":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3-150x150.png",150,150,true],"medium":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3-300x177.png",300,177,true],"medium_large":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3-768x452.png",768,452,true],"large":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3-1024x603.png",1024,603,true],"1536x1536":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png",1146,675,false],"2048x2048":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png",1146,675,false],"small-home-featured":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3.png",100,59,false],"sidebar-featured":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3-180x128.png",180,128,true],"genesis-singular-images":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3-896x504.png",896,504,true],"archive-featured":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3-400x225.png",400,225,true],"gb-block-post-grid-landscape":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3-600x400.png",600,400,true],"gb-block-post-grid-square":["https:\/\/www.codemotion.com\/magazine\/wp-content\/uploads\/2019\/11\/UnderstandingAI-InferenceFigA3_3-600x600.png",600,600,true]},"uagb_author_info":{"display_name":"Mark Patrick, Mouser Electronics","author_link":"https:\/\/www.codemotion.com\/magazine\/author\/mark-patrick\/"},"uagb_comment_info":0,"uagb_excerpt":"Inferencing is the second phase of machine learning, following on from the initial training phase. During the training phase, the algorithm generates a new model or repurposes a pre-trained model for a specific application and helps the model learn its parameters. During the inferencing phase, predictions and decisions on new data are made \u2013 based&#8230;&hellip;","lang":"en","_links":{"self":[{"href":"https:\/\/www.codemotion.com\/magazine\/wp-json\/wp\/v2\/posts\/1016","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.codemotion.com\/magazine\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.codemotion.com\/magazine\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.codemotion.com\/magazine\/wp-json\/wp\/v2\/users\/83"}],"replies":[{"embeddable":true,"href":"https:\/\/www.codemotion.com\/magazine\/wp-json\/wp\/v2\/comments?post=1016"}],"version-history":[{"count":6,"href":"https:\/\/www.codemotion.com\/magazine\/wp-json\/wp\/v2\/posts\/1016\/revisions"}],"predecessor-version":[{"id":14334,"href":"https:\/\/www.codemotion.com\/magazine\/wp-json\/wp\/v2\/posts\/1016\/revisions\/14334"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.codemotion.com\/magazine\/wp-json\/wp\/v2\/media\/2216"}],"wp:attachment":[{"href":"https:\/\/www.codemotion.com\/magazine\/wp-json\/wp\/v2\/media?parent=1016"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.codemotion.com\/magazine\/wp-json\/wp\/v2\/categories?post=1016"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.codemotion.com\/magazine\/wp-json\/wp\/v2\/tags?post=1016"},{"taxonomy":"collections","embeddable":true,"href":"https:\/\/www.codemotion.com\/magazine\/wp-json\/wp\/v2\/collections?post=1016"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}