In Part I of this series, we described the technology behind ChatGPT, how it works, its applications and limitations that organizations should be aware of before jumping on the hype train. As you try to discern opportunities this technology presents to you organization, you will likely face a fog of uncertainty in terms of:
听
Which of our business problems should we solve with ChatGPT, and more broadly, AI? (product-market fit)
Which of our business problems can we solve ChatGPT and AI? (problem-solution fit)
What is the effort and cost involved in leveraging or creating such AI capabilities?
What are the associated risks, and how can we reduce them?
How can we deliver value rapidly and reliably?
In this article, we address these questions by taking an听expansive view of AI听and by sharing five recommendations that we鈥檝e distilled from our experience in delivering AI solutions.
听
听
1. Product-market fit: Start with the customer problem, not the tool
听
With the excitement around ChatGPT, it is tempting to fall into the trap of a tech-first approach 鈥 we have a shiny hammer, what can we hit with it? A common business mistake with AI projects is to start with available data or the AI techniques du jour. Instead,听start with a specific customer problem.听
听
Without a clear and compelling problem that is backed by the听voice of the customer, we will find ourselves in a vacuum that is quickly filled with 鈥渆xpert鈥 but unsubstantiated assumptions. Pressure from inferring threats from competitor media claims and leadership cultures that value 鈥渒nowing鈥 over experimentation can lead to months of wasted investment in engineering. As Peter Drucker famously said, 鈥渢here is nothing so useless as doing efficiently something that should not have been done at all.鈥澨
听
There are several practices that we can apply to improve our odds of betting on the right thing. One such practice is听, which helps us develop clarity in the customer struggle, our vision, the problem space, value propositions, and high-value use cases. This investment of a few weeks with the right people at the table 鈥 customers, product, business stakeholders, (rather than just data scientists and engineers) 鈥 can help us focus on ideas that will bring value to customers and to the business and avoid wasting people鈥檚 time on efforts without outcomes.
Tools such as the听听and听hypothesis canvas听are useful for assessing the value proposition and feasibility of using AI to solve our most compelling problems (see Figure 1).


Figure 1: The hypothesis canvas helps us articulate and frame testable assumptions (i.e. hypotheses) (Image from听Data-driven Hypothesis Development)
2. Problem-solution fit: Choose the right AI tools to create value
听
This recommendation is about expanding your toolkit to find听the most suitable and cost-effective approaches for your high-value use case, and avoid wasting effort in unnecessarily complex AI techniques. Recognize that while ChatGPT might trigger conversations, LLMs might not be the right solution to your problem.听
听
Beyond LLMs, there are other AI techniques that, in our experience, can effectively augment businesses鈥 ability to make optimal day-to-day operational and strategic decisions, and with far less data and compute resources (see Figure 2).


For example, in the top left quadrant, 魅影直播 worked with Marimekko 鈥 a Finnish lifestyle design company 鈥 and applied reinforcement learning to create an听adaptive and personalized shopping experience for customers. We built a decision factory that learns from users鈥 behavior in real-time, and does not require past data assets nor expert data science skills to be scaled across the company's digital platforms.听Within a few hours of release, the decision factory created a 41% lift in front page clickthrough, and after five weeks, a 24% increase in average revenue per user.
听
In the middle of the top two quadrants, we applied听operations research techniques听to create a model that helped听Kittil盲 Airport听to relieve the heavy pressure on airport infrastructure and resources during peak travel season. This reduced planning time from 3 hours to 30 seconds (a 99.7% reduction), and reduced the share of airport-related flight delays by 61% (even though the number of flights increased by 12% from previous year). The decrease in delays resulted in an estimated听monthly cost savings of 鈧500,000 and the airport鈥檚 Net Promoter Score (NPS) score increased by 20 points.
听
These are just some examples of the untapped potential of the family of mature听AI Augmented techniques. Choosing the right AI technique for your problem allows you to be smarter with the data you curate, and in so doing, significantly reduces the cost, complexity and risk involved in curating reams of historical data.
听
听
3. Effort: With bigger models come greater error surfaces


Artwork by听
听
A lesser known contemporary of ChatGPT is Meta鈥檚听, which was trained on 48 million research papers and intended to support scientific writing. However,听听because it produced misinformation and pseudoscience, all while sounding highly authoritative and convincing.听
听
There are two takeaways from this cautionary tale. First, throwing mountains of data at a model doesn鈥檛 mean we will get outcomes aligned with our intent and expectations. Second, creating and productionizing ML models without comprehensive testing comes at a significant cost to humans and society. It also increases the risk of reputational and financial damage to a business. In a more recent story, a Google spokesperson echoed this when they said that听听underscored 鈥渢he importance of a rigorous testing process.鈥
听
We can and must test AI models. We can do so by using听an array of ML testing techniques听to uncover sources of error and harm before any models are released. By defining model听fitness functions听鈥 objective measures of 鈥済ood enough鈥 or 鈥渂etter than before鈥 鈥 we can test our models and catch issues before they cause problems in production. If we struggle to articulate these model fitness functions, then we鈥檒l likely eventually discover that it鈥檚听. For generative AI applications, this will be a non-trivial effort that should be factored into decisions about which opportunities to pursue.
听
We must also test the system from a security perspective and conduct threat modeling exercises to identify potential failure modes (e.g., adversarial attacks and听). The absence of tests is a recipe for endless toil and production incidents for machine learning practitioners. A comprehensive test strategy is essential if you are to ensure that your investments lead to a high quality and delightful product.
听
听
4. Risk: Governance and ethics need to be a guiding framework, not an after-thought
听
鈥淎I ethics is not separate from other ethics, siloed off into its own much sexier space. Ethics is ethics, and even AI ethics is ultimately about how we treat others and how we protect human rights, particularly of the most vulnerable.鈥 鈥撎
听
The 魅影直播-sponsored听MIT State of Responsible Technology Report听observed that responsible technology is not a feel-good platitude, but a tangible organizational characteristic that contributes to better customer acquisition and retention, improved brand perception and prevention of听.听
听
Responsible AI is a听听field, and there are assessment techniques 鈥 such as data ethics canvas, failure modes and effects analysis, among others 鈥 that you can employ to assess the ethical risks of your product. It鈥檚 always beneficial to 鈥渟hift ethics left鈥 (moving ethical considerations earlier in the process) by involving the relevant stakeholders 鈥 spanning product, engineering, legal, delivery, security, governance, test users, etc. 鈥 to identify:
听
Failure modes and sources of harm of a product
Actors who may compromise or abuse the product, and how
Segments who/which are vulnerable to adverse impacts
Corresponding mitigation strategies for each risk
听
The risks (where risk = likelihood x impact) we identify must be actively managed along with other delivery risks on an ongoing basis. These assessments are not once-off, check-box exercises; they should form part of an organization鈥檚听data governance听and听听framework, with consideration of how governance can be听lightweight and actionable.
听
听
5. Execution: Accelerate experimentation and delivery with Lean product engineering practices
听
Many organizations and teams start their AI/ML journey with high hopes, but almost inevitably struggle to realize the potential of AI due to unforeseen time sinks and unanticipated detours 鈥斕 the devil is in the data detail. In 2019, it was reported that听. In 2021, even among companies who have successfully deployed ML models in production,听, an increase from 56% in 2020.听
听
These impediments in delivering value frustrates all involved 鈥 executives, investment sponsors, ML practitioners and product teams, among others. The good news is that it doesn鈥檛 have to be this way.听
听
In our experience,听Lean delivery practices听have consistently helped us to iterate towards building the right thing by:
听
Focusing on the voice of the customer
Continuously improving our processes to increase the flow of value
Reduce waste when building AI solutions
听
Continuous delivery for machine learning (CD4ML) can also help us听improve business outcomes听by听accelerating experimentation听and improving reliability. Executives and engineering leaders can help steer the organization towards these desirable outcomes by听advocating for effective engineering and delivery practices.
听
听
Parting thoughts
Amidst the prevailing media zeitgeist of 鈥淎I taking over the world鈥, our experience nudges us towards a more balanced view: which is that humans remain agents of change in our world and that AI is best suited to听augment, not replace, humans. But without care, intention and integrity, the systems that some groups create can听. As technologists, we have the ability and duty to design responsible, human-centric, AI-enabled systems to improve the outcomes for one another.
Disclaimer: The statements and opinions expressed in this article are those of the author(s) and do not necessarily reflect the positions of 魅影直播.