You are currently viewing How the Hell Will the companies Behind ChatGPT, Bard, and LLaMA Make Money?
image

The dizzying explosion of generative artificial intelligence platforms has been the big business story of the past year, but how they’ll make money and how smart companies can use them wisely are the questions that will dominate the next 12 months.

“Students and executives are no longer asking whether we should adopt AI—but rather, when and how to do so,” says Andy Wu, the Arjun and Minoo Melwani Family Associate Professor of Business Administration at Harvard Business School.

Wu’s recent case study and background note, AI Wars and the Generative AI Value Chain, offer a crash course in ChatGPT, Bard, and other AI chatbots—as well as the dueling tech titans behind them—and probe the strategic dilemmas ahead for innovators and users. The public’s fascination with the human-like aspects of chatbots may be overshadowing more fundamental questions about how companies can profit from AI, Wu says.

“I think the basic economics of a generative AI are being overlooked.”

In an interview, Wu discusses the challenging economics of AI, how business models are likely to differ from traditional software models, and some of the potentially painful tradeoffs ahead for companies such as Google, Microsoft, and others. Wu collaborated on the case study with HBS research associate Matt Higgins; HBS doctoral student Miaomiao Zhang; and Massachusetts Institute of Technology doctoral student Hang Jiang.

Ben Rand: What did you find most surprising in preparing this case and why?

Andy Wu: I think the basic economics of a generative AI are being overlooked. There are significant unanswered questions in terms of how people will actually make money with this technology. Google and OpenAI and others can’t lose money in perpetuity. But it’s not yet obvious to anyone exactly how this will be monetized. At minimum, I can tell you that we are going to need new business models, and the integration of generative AI is going to transform how we monetize software and the business model.

Rand: How so?

Wu: Our notions of fixed cost and variable costs are different here than they were for any other form of computing we’ve lived through in the past. The key insight is that the variable cost of delivering generative AI to an end user is not zero… which means we can’t necessarily be handing out future software-as-a-service applications containing generative AI for free to anyone or even as a paid subscription without usage limits as we are used to today. Usage pricing is going to be much more important.

A second distinction is that a significant portion of the core technology is open source, and a lot of the data being used to train these models is public data and may be copyrighted but is publicly available online. The barriers to entry for AI are not as high as it may seem. So many companies will be in the game, at least for specific vertical AI models and applications.

Rand: Is it too soon to tell which business model will emerge?

Wu: The companies are still trying to figure it out. But I think by their actions, we can get some hints about the direction we’re going to go. The generative AI companies out there are actually pricing on a usage model, which says to me that they don’t think they can make the subscription model work economically today.

Rand: Which companies are in the best position right now?

Wu: A real standout right now is Meta, in terms of fighting hard for a prominent position on the open-source side with their LLaMA model. Prior to last year, many would have assumed that Google would have been the putative leader in the open-source part of the market. Microsoft also deserves a lot of credit for making a decision to work with OpenAI and getting access to leading technology that they can integrate both into their applications and as a way to sell cloud computing services.

But what’s interesting here is that none of the big tech providers are in the business of selling the actual model itself. Amazon largely offers its cloud customers the open-source models that others have made. Meta is largely handing its model out for free (with some limits), and Microsoft outsourced much of the core technology to OpenAI. Looking at these decisions together, they are making a real, albeit subtle, statement about what to avoid, which is actually trying to directly monetize the core technology—the AI model—itself.

The challenge we face right now with AI is it’s very possible that the actual invention of the technology itself is not what people will make money on. It will transform the world, but the money isn’t made on the thing that enabled the transformation.

“The issue is, we normally think of intellectual property as being copyrighted or not copyrighted.”

Rand: What role will regulation play, do you think?

Wu: I understand the interest of regulators, given the risks of this technology. But it’s going to be very difficult for regulators to come up with a comprehensive policy that controls things in the way that they would aspire to control them. That comes down to one principal factor, and that is that the barrier to entry is not that high. There are already a significant amount of open-source models that you or I could build on.

So, let’s say we want to block AI from generating hate speech. To the extent that there is a market for hate speech, some entrepreneur will be able to build that model. It’s hard to exactly figure out how you would block it. If there is a market, someone will figure out how to do it.

Rand: Are there some areas where regulation may be useful?

Wu: Copyright law is one area they can address. The issue is, we normally think of intellectual property as being copyrighted or not copyrighted. But I like to teach my students that we’re in a new world. There’s copyrighted and not copyrighted, and then there’s also public and private. And so, the issue right now is that you can have copyrighted data that is also public. For instance, any newspaper publisher that allows their news articles to be indexed by search engines has put their intellectual property in this situation of being copyrighted but also public.

What do those creators do about their data? You can say it’s copyrighted, and say other people can’t use it, but you can’t really run around proving that all these different models are being trained using your data. This is something that regulators will have to clarify and also something that companies are taking note of themselves, particularly in the music space and image space.

“To the extent that maybe you don’t want an AI offering now, but you want one five years from now, the effort of building a centralized data store for all that data will be important.”

Rand: With so much to consider, how can managers stay on top of the developments in AI? They seem to be changing so fast.

Wu: I would advise managers not to play for the next year. Play for the next 10 years. The idea is that you want to have people inside your company who are on top of the different technologies and experimenting with different things. You need to give them a pathway to communicate with the CEO and top management team about which technologies to invest in.

The corollary to that is the integration of data across business units in a company. It’s not being done well enough right now, based on my experience. Companies already have a fairly complicated portfolio of different databases and enterprise products that store their data. That data increasingly needs to be kept track of, and, ideally, integrated. And so, to the extent that maybe you don’t want an AI offering now, but you want one five years from now, the effort of building a centralized data store for all that data will be important.

Rand: What are some best practices of things companies need to be doing now?

Wu: If you’re a company thinking about implementing AI, there are different levels of sophistication to consider. You could wait for other companies to develop the relevant applications, or you could buy an API and build your own application, or you could actually train your own model and then build your own application. And I think companies need to begin the process now of identifying what level of sophistication they want. For example, one early leader in this process is Bloomberg, where they have already gone ahead and built BloombergGPT, a large-language model tailored for financial tasks. They used their own proprietary data to finetune an open-source model. For a company like Bloomberg, providing financial insights is mission critical, and so they cannot wait around for someone else to develop that AI model and application.

HBS Working Knowledge

“Harvard Business School is the graduate business school of Harvard University, a private research university in Boston, Massachusetts. It is consistently ranked among the top business schools in the world and offers a large full-time MBA program, management-related doctoral programs, and executive education programs.”

 

Please visit the firm link to site


You can also contribute and send us your Article.


Interested in more? Learn below.