B2B software sales and marketing teams love hearing the term “artificial intelligence” (AI). AI has a smoke and mirrors effect. It sounds impressive. But, when we say “AI is doing this,” our buyers often know so little about AI that they don’t ask the hard questions.
In industries like the DevTools space, it is crucial that buyers understand both what products do and what their limitations are to ensure that these products meet their needs. If the purpose of AI is to make good decisions for humans, to accept that “AI is doing this” is to accept that we don’t really know how the product works or if it is making good decisions for us.
When we’re in the buyer role, we often don’t hold ourselves responsible for understanding AI and machine learning (ML) products because these technologies are intimidating. They’re incredibly complex.
This article addresses the limitations of AI and ML, so software buyers can ask the right questions to understand what they are buying.
One limitation of some AI or ML products is that for certain applications of the technology, there is no source of absolute truth to compare against the accuracy of the output. For example, neither humans nor machines know how to produce the perfect set of end-to-end tests for any given application. This is the test oracle problem: there is no objective standard of truth. No one wants to introduce this kind of uncertainty into their sales process. Yet, our buyers deserve well-informed answers about our products.
As a buyer, you need to understand the intended advantage of your seller’s AI product before making a purchase decision. Is it meant to make a decision that is more accurate—against an objective standard—than a human? Is it meant to make a faster decision with less cost? Or introduce an alternative methodology that uses new data in a new way? Answers to these questions influence how you will use the product and what value it provides.
Though AI is commonly accepted as “any machine that uses math to make decisions,” true AI is self-taught. AI has a neural net that mimics neurons in a human brain which allows it to teach, update and evolve itself. Because of this, true AI is difficult to build and is often experimental rather than commercial.
More often, what’s being described when we say AI is actually ML. ML is human-taught: Machines learn through human feedback using a probabilistic decision-making process that improves via ongoing correction. Machines take in data, run algorithms against it and output a decision — or series of assertions — based on probabilities. Humans correct the machine by telling it whether it was accurate in its assessment, and the machine updates. As it receives accuracy feedback, machines learn to make better decisions. And because ML is based on probabilities, it will sometimes make the wrong decisions.
Based on how you plan to use a product, you need to determine how rigorous its accuracy needs to be. How often a machine can make the wrong decisions and still serve its purpose is application-specific. Self-driving vehicles must be nearly perfectly accurate to be adopted. Paralegal ML toolsets likely need to be less accurate. How accurate does your product need to be?
Regardless of how you plan to use a product, it’s important to ask the right questions to understand the product and build resiliency around its accuracy levels. The next time a seller tells you “AI is doing this,” you can ask the following:
Not only is there a lot of “AI” that isn’t AI, but there is also algorithmic technology that isn’t ML. It is thus essential for buyers to know enough to ask the right questions and understand how these products make decisions.
There are limitations to all ML products, though the limitations differ by product and the way the product is applied. When a product’s accuracy levels are unknown, all you can do is ask if its methodology is valid for decision-making: Does it have access to better data than humans? Can it make smarter and faster decisions than humans with this data? If the answer is yes, you should consider buying the product rather than having your people do the work.
The data used to train AI models needs to reflect the production environments where applications are deployed.
Looking for a DevOps job? Look at these openings at NBC Universal, BAE, UBS, and other companies with three-letter abbreviations.
Tricentis is adding AI assistants to make it simpler for DevOps teams to create tests.
Redis is taking it in the chops, as both maintainers and customers move to the Valkey Redis fork.
GitLab Duo Chat is a natural language interface which helps generate code, create tests and access code summarizations.
Expect attacks on the open source software supply chain to accelerate, with attackers automating attacks in common open source software…