Q: How far along are corporate efforts to achieve fairness and eliminate bias in AI results?   Boezeman:  Trying to determine bias or skew in AI is a very difficult problem and requires a lot of extra care, services, and financial investment to be able to not only detect, but then fix and compensate those issues. Many corporations have unintentionally used biased or incomplete data in different models; understanding that and changing this behavior requires cultural changes and careful planning within a company. 
Those that operate under defined data ethics principles will be well-positioned to avoid bias in AI, or at least be able to detect and remedy it if and when it’s identified.  Q: Are companies doing enough to regularly review their AI results? What’s the best way to do this?   Boezeman:  As new tools are provided around the auditability of AI, we’ll see a lot more companies regularly reviewing their AI results. Today, many companies either buy a product that has an AI feature or capability embedded or it’s part of the proprietary feature of that product, which doesn’t expose the auditability. 
Companies may also stand up the basic AI capabilities for a specific use case, usually in that AI discover level of usage. However, in each of these cases the auditing is usually limited. Where auditing really becomes important is in “recommend” and “action” levels of AI. In these two phases, it’s important to use an auditing tool to not introduce bias and skew the results. 
One of the best ways to help with auditing AI is to use one of the bigger cloud service providers’ AI and ML services. Many of those vendors have tools and tech stacks that allow you to track this information. Also key is for identifying bias or bias-like behavior to be part of the training for data scientists and AI and ML developers. The more people are educated on what to look out for, the more prepared companies will be to identify and mitigate AI bias.    Q: Should IT leaders and staff receive more training and awareness to alleviate AI bias?   Boezeman:  Definitely. Both the data scientists and AI/ML developers need training on bias and skew, but it’s also important to expand this training to product managers, executives, marketing managers, and merchandisers. 
It’s easy to fall into the trap of doing what you’ve always done, or to only go after a bias-centric approach like many industries have done in the past. But with training around alleviating AI bias, staff across your organization will be able to identify bias rather than trusting that everything AI produces is fact. From there, your company can help mitigate its impact.  
Q: AI and machine learning initiatives have been underway for several years now. What lessons have enterprises been learning in terms of most productive adoption and deployment?   Boezeman: AI is not a panacea to solve everything. I have seen many attempts to throw AI at any use case, independent if AI is the right use case, all to enable a marketing story without providing real value. The trick to successful deployment of an AI solution is a combination of the quality of the data and the quality of the models and algorithms driving the decisioning. Simply put, if you put junk in, you’ll get junk out. The most successful deployments have a crisp use case, and well-defined data to operate with.  
Q: What areas of the organization are seeing the most success with AI?   Boezeman:  There are many different stages in AI, but mostly they can be boiled down to three basic states: discover, recommend, and automatic action. Right now, the places I see it mostly used is in discover – insights, alerts, notifications – space. This is where the system tells you something is going on abnormal or outside of known patterns, or something is trending in a direction you should care about. People trust this kind of interaction and model, and can easily collaborate if they want proof. 
Marketers leverage AI in the discover space to determine how successful their campaigns are, for example. Another example is a merchandiser that may deploy an AI-powered solution to detect fraud or issues with the customer journey. 
Where I still see a lot of hesitation is in the recommend and action states. I used to own a product that calculated the best price for a product and order to display them in a web storefront, based on many data points, from quantity, to profitability, to time to markdown, to storage space used in warehouse. And even this product could, if you turned it on, automatically take action. What we found is many merchandisers like seeing the recommendation, but they personally wanted to take action, and also wanted to see multiple options, and finally, they wanted to see the decision tree on why the system recommended an option. When we first launched it, we didn’t have the “Why did the system recommend XYZ?” functionality. Until we provided a way to allow the merchandiser the ability to see what the recommendation was based on, they didn’t trust it.  
Q: What technologies or technology approaches are making the most difference?