Determining the potential of your AI algorithm begins with measurement

Author : Jeff Harris | Vice President, Global Corporate & Portfolio Marketing | Keysight Technologies

01 November 2022

At the core of every AI algorithm, explains Jeff Harris, Vice President, Global Corporate & Portfolio Marketing at electronics T&M expert, Keysight Technologies, are three basic ingredients: 1) the ability to measure; 2) knowing how much of what you measure needs to be processed; and, of course, 3) the ability to process more than one input at a time.

This article was originally featured in EPDT's H2 2022 Test & Measurement supplement, included in the November 2022 issue of EPDT magazine [read the digital issue]. And sign up to receive your own copy each month.

To what depth a system can measure can be thought of as its potential. Determining what aspects of those measurements must be sent to the processor can be thought of as delivering that potential. Finally, knowing how to combine the salient parts of those measurements in the correct proportions, known as sensor fusion, is the key to exploring an algorithm’s IQ or reasoning potential. Augment that sensor fusion with a feedback loop and the algorithm will have the ability to check and course-correct its logic, a necessary ingredient in machine learning...

These three attributes are the key to understanding the depth of an AI (artificial intelligence) algorithm’s unique power. And like many things, the more you cultivate and calibrate these foundational elements, the better the AI algorithm will perform in the long term. Now that we understand the three areas to explore, let’s dive into the first component, measurement depth, and how it’s critical to the foundation of building a robust, high-performing AI algorithm…

Measurement depth

Metrology is the study of measurement science, and measurement depth plays a crucial role in building a robust algorithm. The Gagemaker’s Rule, also often known as the ‘rule of tens’ or the 10:1 rule, states that a measurement device must be 10x more precise than the desired measurement. The reason that measurement depth is so critical is that it determines the possible level of precision and therefore sets the algorithm’s maximum potential. The more precision you have in any given measurement, the greater the AI algorithm’s potential.

Metrology focuses on the deep understanding of a particular measurement. That measurement can be as simple and distinct as voltage, ground or temperature, as multi-modal as the functioning of aircraft control surfaces, or as complex as maximising throughput on a manufacturing assembly line.

Whether you are measuring a single parameter or several, the depth of each measurement determines the level of programmability that’s possible. For instance, measuring a 3 Volt system to 1/10th of a volt is not as insightful as measuring to 1/1000th of a volt. Depending on the system that voltage is powering, the extra precision may be critical for battery life – or it may be a distraction. Maximising the potential of any algorithm requires matching the entire end-to-end measurement needs to the required depth. This is true no matter what’s being measured – even data systems, which may not be as immediately intuitive, so let’s look at one of those examples.

How to optimise measurement

Enterprise IT (information technology) stacks are now a complex web of interconnected data systems, each exchanging information aimed at tuning an organisation’s operations. These technology stacks include an array of software, such as CRM (customer relationship management), ERP (enterprise resource planning), databases, order fulfillment – and each with unique data formats and custom application programming interfaces (APIs). According to sales & marketing automation firm, Salesforce, the average company has over 900 applications in its tech stack, many of them cloud-based and all of them experiencing software updates that can have ripple impacts. Identifying and isolating problems, much less optimising performance across multiple intersecting software applications, is akin to finding a needle in a collection of interconnected haystacks.

Each software application in a tech stack usually has a different sponsor in an enterprise – finance, human resources (HR), sales, marketing, supply chain – and that primary org’s considerations are top of mind for IT. Every enterprise has custom workflows and integrations with numerous applications and backend systems, and user journeys often span various paths and are rarely linear. Therefore, even if two enterprises used identical applications in their tech stack, mapping all the exchange points and validating the end-to-end operation would be unique. If there were ever an application in need of AI, this would be it. The measurements, in this case, could be the intersystem data input points, the intrasystem data exchange points, and the data display points.

Understanding how an AI algorithm would operate in a system like this would start with understanding how it measures data points in three key areas:  

1. Measuring how users interface with the application, regardless of the operating system, which in some cases involves employing robotic process automation (RPA) when button pushes are required

2. Measuring the data exchanges between and command APIs that link the systems in a complex technology stack to ensure they are occurring correctly

3. Measuring on-screen information across omni platforms (desktops and mobile), such as images, text and logos, as a human would, to see how they render

Evaluating the measurement efficacy starts with its ability to measure, regardless of operating system, software versions, devices or interface mechanisms. The more conditions under which the AI cannot measure, the less impactful it will be in operation.


Whenever you assess the potential of anything, start with the foundation. At the foundation of every AI system is its ability to measure. The more it can measure, the more impactful it has the potential of being. Look at what it is capable of measuring and, more importantly, where it is not capable. Limited sensing results in limited AI algorithm potential. The old adage from Lord Kelvin stands true today that “if you cannot measure it, you cannot improve it”. To understand the true power of any AI, make sure to start by analysing its measurement breadth and depth.

More information...

Contact Details and Archive...

Print this page | E-mail this page