Is It Time for an Open LLM To Be Added to the LAMP Stack?

Patrik Backman knows a thing or two about open source software and the impact of the LAMP stack on the early web. He was a co-founder of MariaDB and before that the director of software engineering at MySQL, the original “M” in the LAMP stack — although many argue that MariaDB is the “M” these days. Backman is now a general partner at European VC firm, OpenOcean, and continues to be involved in open source technologies (he’s still on the board of MariaDB). Recently, he’s been on the lookout for open source generative AI products that could be added to the LAMP stack.
According to Patrik Backman, “collaborative ecosystem building” is what made LAMP successful.
The LAMP stack, in case you’ve been living under an online rock, is based on an acronym for Linux, Apache, MySQL, PHP/Perl/Python. It was coined at the end of 1998 to denote a set of open source solutions to build web applications. The LAMP stack was what led to the Dot Com boom and the Web 2.0 revolution that followed after, because it enabled startups to leverage low-cost, open technologies to build new products.
Backman, who was along for the ride with MySQL and then MariaDB, told me that “collaborative ecosystem building” was a key part of what made the LAMP stack successful.
“In the early days, [we] thought that it’s natural to have all these open source events and so forth, to both meet with the key people from these other initiatives and projects, but also collaborate on packaging things well together, open documentation, open development principles, open bug reports, and development plans.”
Perhaps the biggest question around the LAMP stack now, and whether AI fits into this template, is whether the current open source solutions for AI are even capable of community building and collaboration. For example, before we even think about adding Meta’s LLaMA models to the LAMP stack — LAMPL? — we first need to figure out whether they’re open enough. And secondly, can Meta play well with the open source community?
Can Developers Trust Open Source LLMs?
Meta’s LLaMA models aren’t the only open source large language models available. Some of the other prominent ones include Mistral, Falcon, Google’s T5, BLOOM, and Databricks’ MPT models (previously known as MosaicML MPT). There are many more listed on Hugging Face’s open LLM leaderboard.
“I think there will be many open source players competing, which also restricts what Meta / Facebook can do with LLaMA.”
– Patrik Backman, OpenOcean
Meta’s models are probably the most well-known and well-used open source LLM, but they’re also the most controversial. Some have claimed that the LLaMA models aren’t truly open source, because of limited access to training data and other issues — such as the licensing terms encouraging development within Meta’s ecosystem (rather than allowing unrestricted use and modification, as is typical of open source software). Meta also imposes commercial restrictions on LLaMA model use. I asked Backman what he thinks of Meta’s LLaMA.
Firstly, Backman agrees with Mark Zuckerberg that the future of AI will be open source. “The technologies around AI are moving at an enormous speed and to innovate in the best way, the open source approach is the winning one,” he told me.
But does he think there is a danger of startups building on LLaMA, given the limitations of the license and Meta’s control over it?
Backman replied that although there are such risks, ultimately he thinks there is enough open source LLM competition — he specifically mentioned Mistral AI from France — and so startups have plenty of choices.
“I think there will be many open source players competing, which also restricts what Meta / Facebook can do with LLaMA,” he said. “I do think that […] for it to really succeed, they [Meta] will maybe need to be a bit clearer on the freedom and the openness of the initiative. But, of course, the community needs everyone participating and working […] to put tougher requirements and requests on the openness side.”
On a podcast earlier this year, Backman made it clear that he thinks companies running open source projects are entitled to create whatever license they feel is best for their business. “I think what should be recognized, and understood, and respected is that, indeed, anyone building software, even providing a free and open-source software, they should be allowed to have a credible and sustainable business model in order for them to be able to maintain and further develop it,” he told the Open Source Underdogs podcast. “But, then, it’s about what kind of balance you want to strike with what is free and what is commercial, and that’s where the license comes in.”
Compute an Issue in 2024
“…for building advanced AI models, the compute side is way more demanding than before.”
Even if we added an open LLM to the LAMP stack, there is a big question about whether startups can easily access this technology, due to the massive amount of compute that LLMs require.
“Of course, now for building advanced AI models, the compute side is way more demanding than before,” said Backman. He suggested that access to compute, and the ability to afford it, makes it trickier for a startup than (for example) standing up a free Apache server on a Linux machine and plugging in MySQL, back in the day.
I asked whether startups he deals with tend to use managed services from the big internet companies for their compute needs. He replied that yes, they do, but the pattern has changed from five-to-ten years ago. Previously, startups tended to pick one cloud provider over the others, he said, but these days they will pit the providers against each other.
“Now, there’s kind of a knowledge that, hey, you have to be able to support them [different cloud providers] in parallel, to be able to pit them against each other and [have them] compete with each other, so that you keep prices low, you get credits, and this and that support from them.”
An Example: MindsDB
“It’s a balance of providing a great [open] stack […] while having a commercial cloud offering that we build a business around.”
One of the startups Backman’s VC firm has invested in is MindsDB, a self-described “AI data automation platform.” Like a lot of recent enterprise-focused AI products, the goal of MindsDB is to help organizations meld generative AI technology with their own in-house data.
“If you develop AI models that use your own proprietary data, then it’s the best platform out there,” said Backman, adding that MindsDB has had over 300,000 deployments so far and is “growing very fast.”
Backman contends that MindsDB is a good example of a startup striking the balance between open source technologies and commercial offerings.
“It’s a balance of providing a great stack — open source, freely available, good documentation, good use cases, good examples, and so forth — while [also] having a commercial cloud offering that we […] then build a business around.”
The LAMP Question
Returning to the question I posed at the start of this article: should we add an open LLM to the LAMP stack?
Based on what I learned from Patrik Backman and what we currently know about the LLaMA license, I think it’s too early to give Meta the benefit of the doubt. But let’s give it another year or two to see if LLaMA matures as an open source solution, or whether one of its competitors emerges to become a LAMP-worthy product.