Large Language Models (LLMs): The Path to Differentiation in AI
Large language models (LLMs) are quickly becoming mainstream. Just one year after the debut of ChatGPT, launching an AI assistant has become more straightforward: integrate GPT-4, connect to a vector database, and leverage APIs based on user inputs. However, if you follow this formula alone, don't expect your app to gain traction in the crowded market.
Beyond Technology: Finding Sustainable Advantage
Technology is no longer a unique selling point for AI products, especially as the barriers to entry continue to diminish. With many companies having access to similar LLMs, any new technical breakthroughs are rapidly duplicated by competitors. The real game-changer lies in the application layer. Firms that effectively identify and solve real user problems will greatly enhance their chances of success. Solutions could take the shape of another chatbot or something entirely different.
Innovating through product design and experimentation is a crucial, yet often overlooked, path.
TikTok: More Than Just an Algorithm
While TikTok isn’t a generative AI application, it exemplifies how innovative product design can drive success. Many attribute TikTok's phenomenal growth solely to its algorithm. However, numerous other recommendation engines also deliver impressive results (just ask former YouTube product managers).
At their core, these systems adhere to foundational principles: recommending content similar to what users already enjoy (content-based filtering) and suggesting items based on what users with similar interests prefer (collaborative filtering).
TikTok revolutionized its algorithm by presenting it within a novel interface: an endless stream that allows seamless interaction through user swipes. With a focus on short-form video content, TikTok can swiftly learn user preferences and adapt its recommendations.
Moreover, TikTok provides top-notch creator tools. Anyone can easily film and edit videos using just a smartphone—no advanced video production skills are required.
In today’s landscape, competing in the short-form video arena revolves around the ecosystem each platform offers. While having an effective algorithm is essential, cultivating a loyal user base, creator revenue sharing, content moderation, and other integral features are vital for differentiation.
The Quest for Product-Market Fit in Generative AI
Understanding product-market fit (PMF) is often described as a gut feeling; it's the quality of a product that users adore and feel compelled to use repeatedly. Practically, products that grow rapidly and retain users usually indicate PMF.
Currently, most generative AI applications are far from achieving PMF, with the primary reason being their failure to address real user challenges. User needs have historically driven product development, but the excitement surrounding new technologies like LLMs often leads to the temptation of applying them inappropriately.
In the past year, nearly every major company has experimented with incorporating AI into their core offerings. Countless AI-driven startups are capitalizing on this trend, but many might not discover PMF. Some may succeed through sheer chance, highlighting the importance of innovation.
To improve your odds of finding PMF, let’s analyze one of generative AI's most notable successes: GitHub Copilot.
GitHub Copilot: A Case Study in Success
With over $100 million in annual recurring revenue (ARR), GitHub Copilot is arguably the most successful generative AI product to date, aside from ChatGPT. According to Retool’s State of AI in 2023 report, GitHub Copilot is favored by 68% of tech professionals.
The same report indicated a 58% decline in Stack Overflow usage compared to 2022, primarily due to the impact of Copilot and ChatGPT. This shift offers compelling evidence of Copilot's PMF. It effectively addresses a clear pain point for developers: writing and debugging code—an area where Copilot provides exceptional ease and accuracy.
The thoughtful design of Copilot extends deeply into both the product and engineering realms. Copilot isn’t just a reimagined version of ChatGPT; users typically interact with it through code completion suggestions in their text editors, not through a chat interface.
Copilot's utility surpasses traditional code writing; it also aids in refactoring, documentation, and code explanation. With empathy informing its design, the team at GitHub understands user journeys and enhances user experience beyond standard offerings. For example, they devised a technique known as “neighboring tabs,” which allows the LLM to access all open files for context, not just the immediate cursor area.
This level of investment in product and engineering blurs boundaries. Another aspect is Copilot’s focus on prompt engineering—an intricate method of optimizing LLM behavior through precise product knowledge.
Additionally, Copilot’s robust platform includes enterprise features, straightforward billing, and a commitment to privacy—all made meaningful by the core product's established PMF. This combination places GitHub in a compelling market position.
Incorporating Product Thinking into AI Development
To develop a successful AI application, it's essential to address legitimate user problems—no amount of engineering can compensate for a lack of this understanding. Engage with users or potential customers to discern their goals and the obstacles they face. Generative AI may or may not be the ideal solution.
If you are confident you’re building in a suitable area and LLMs are appropriate, you must still differentiate your product from ChatGPT. This differentiation could happen behind the scenes through advanced prompt engineering or retrieval algorithms, or it could manifest in a unique user interface. Ultimately, you need to consider: What unmet needs does your target audience have that existing solutions do not fulfill?
An iterative approach grounded in data often proves effective. Take prompt engineering as an example: a common mistake is to revise LLM prompts based on limited test inputs, deploy if outputs seem improved, and then proceed without thorough testing. This insufficient methodology doesn't ensure the LLM’s broad capability across various inputs. What does it truly mean for outputs to be 'better'? Ideally, it goes beyond intuition.
Most AI product builders closely follow their instincts to gauge PMF. However, companies that implement robust analytics gain a considerable competitive edge.
Ironically, LLM applications possess unique access to invaluable user feedback from amassed natural language data. Unlike traditional software, LLM products provide a transparent view of user inquiries, yet many companies fail to leverage this rich data effectively.
Analytics foster a crucial feedback loop between product development and engineering, bridging the two domains. Start with a hypothesis around a user problem, launch early to gather feedback, and then synthesize insights through analytics. This analytical approach uncovers critical product insights, enabling you to design a strategic engineering roadmap that solidifies your AI product’s success.