✨ Check out this awesome post from Hacker News 📖
📂 **Category**:
📌 **What You’ll Learn**:

Generative artificial intelligence (GenAI) tools such as ChatGPT and Claude have two superpowers. The first
superpower is a boon: they can dramatically increase human productivity. I use them on a regular basis to answer
questions, learn new skills, write code, create images, and much more, all at a rate of speed and quality that was
science fiction just a few years ago. The second superpower is a bane: GenAI is quietly destroying the very
ecosystems that made it possible in the first place.
Under the hood, GenAI is built on large language models (LLMs), which are able to extract patterns, structure, and
statistical relationships from massive data sets. These data sets consist primarily of content created by human beings:
books, blog posts, articles, forum discussions, open source code, art, photography, and so on. LLMs are able to extract
value from this content at an unprecedented scale, but all that value is captured by the GenAI company and its users.
If you’re a content creator, you get nothing: no attribution, no referral traffic, no revenue share. Not even a
thank-you.
This feels unsustainable to me, a bit like a snake eating its own tail. In this blog post, I’ll go through three
examples of how GenAI is destroying the very ecosystems it relies on, and then discuss possible solutions that may
give everyone (users, GenAI companies, and content creators) more value.
Example #1: online communities
For many years, StackOverflow was the most popular Q&A site for programmers. Any time you
hit a weird error while coding, you’d do a search on Google, and more often than not, find a good answer on
StackOverflow. But now, in large part due to GenAI, StackOverflow is nearly dead:

Although StackOverflow’s decline started before GenAI went mainstream (ChatGPT was first released in 2022), GenAI
accelerated that decline considerably. That’s because nowadays, instead of searching around for an answer on a Q&A site,
and working to adapt that answer to your own codebase, you can ask GenAI tools to generate your code, fix any errors
you hit, answer any questions you run into, and so on. As a result, you’re considerably more productive.
But it doesn’t seem sustainable. A big part of why GenAI tools can answer programming questions and fix errors in your
code is because those tools were trained on StackOverflow data. So you as a programmer and the GenAI tool now
get much more value from that data, but StackOverflow gets none. If people stop asking and answering questions, what
will GenAI train on in the future?
I also get the impression that StackOverflow is not the only online community where this is happening. For example,
Quora seems largely dead. Wikipedia is
facing more threats than ever.
And although Reddit’s traffic numbers don’t show it, I (and many other Redditors) get the impression that it’s also
dying.
Example #2: open source
Tailwind CSS is a popular library programmers use to style and decorate their websites. In
fact, according to the 2025 State of CSS Survey, Tailwind is the
most popular CSS library, by far:

And Tailwind usage only seems to be growing:

Despite that, just a couple of weeks ago, the Tailwind team had to lay off 75% of its
staff. Why? The company behind
Tailwind CSS makes its money by selling a premium upgrade to the open source library called
Tailwind UI, which gives you a set of reusable, pre-built, professionally designed
components for building out your website. This was a great offering in the past, but in the age of GenAI, it’s more
problematic:
- Less traffic. Traffic to the Tailwind docs site is down by more than 40%, so fewer developers are discovering
that a paid Tailwind UI library exists. This is probably because developers are using GenAI to write their code,
and therefore, rarely have to look at docs anymore. - Less need. Whereas in the past, buying Tailwind UI would be a huge win in terms of speed (not having to build
those components from scratch) and quality (not everyone has access to top-notch designers), these days, you can get
GenAI to generate CSS components quickly and with high quality, at no additional cost. In fact, GenAI tools
can rapidly copy almost anyone’s work—not just a library like Tailwind UI, but almost any design, any SaaS tool,
and so on—while giving no attribution back to the original creators.
So developers are getting lots of value from Tailwind, GenAI is getting lots of value from Tailwind, but the creators
of Tailwind are getting crushed. I suspect something similar will happen with many other open source projects.
Example #3: books and blogs

When my latest book, Fundamentals of DevOps and Software Delivery, came
out last year, a friend of mine asked me an interesting question:
In the age of LLMs, will people still use books to learn the fundamentals?
I’m an avid reader, and still believe books play a key role in learning, but there’s no doubt that LLMs provide a new
method of learning that is incredibly compelling. GenAI has two remarkable qualities that make it a great teacher:
- It has infinite patience.
- It never judges you.
You can ask it all the questions you want, with no fear of sounding dumb. You can repeat the same question over and
over again if something isn’t clicking. You can request it to explain things in different ways: via text, via audio,
via diagrams. I’ve used GenAI to learned dozens of new things over just the last few months: how to do DIY
projects around the house, how to rehab a minor injury, how to cook eggs without giving them a sulfuric taste/smell,
and much more. And I learned most of these without reading anyone’s book or blog.
And that’s a problem. Much of the content I got from the GenAI tools was extracted from books and blog posts, with
no attribution. Even worse, some of this content was extracted illegally. Last year, Anthropic agreed to pay $1.5B
to settle a class-action
lawsuit for training their
LLMs on over 500,000 pirated books. That included several of my previous books!
This class-action lawsuit might sound like a win for authors, but it’s actually a disaster. There are likely many
AI companies training on pirated data who haven’t been caught. And even if they are caught, they might not
be sued. And even if they are sued, they might not lose or settle. And even if they do, they will just see it as the
cost of doing business. Anthropic recently raised $13B, reported revenue at $5B per year, and is valued at
$183B; OpenAI is trying
to raise $100B, with reported revenue of $20B per year, and a valuation of
$830B. And all these
numbers are growing fast. A $1.5B fine is just a drop in the bucket for companies like this. It’s a risk worth taking.
And I’m guessing all the GenAI companies are taking that risk. In another lawsuit, OpenAI
argued that it’s ‘impossible’ to create AI tools like ChatGPT without copyrighted
material. If that’s
what it takes to get to an $830B valuation, you better believe they are all going to steal and pirate as much
content as possible. And when they do, the creators of that content will get nothing. Nada. $0.
The GenAI model is broken
There are countless other examples where GenAI is benefiting from content, while giving nothing back to the content
creators: e.g., art, music, design, movies, copywriting, and so on. At its root, the GenAI model is broken:
- Create. Someone creates valuable content.
- Extract. A GenAI bot crawls the web, exacting all the valuable content.
- Subscribe. Users pay money to use a GenAI tool. This provides value to the GenAI company.
- Prompt. Users enter a prompt into a GenAI tool, which generates a response based on the content it extracted. The
response provides value to the user.
Did you notice what’s missing? The user has a way to get value (step 4), the GenAI company has a way to get value
(step 3), but the content creator gets nothing. Compare this to the search engine model (e.g., Google Search), which is
what we all used before GenAI came along:
- Create. Someone creates valuable content.
- Extract. A search bot crawls your website to extract the content.
- Prompt. Users enter a prompt into the search engine, and the search engine shows links to websites that match
the prompt, with the most relevant websites at the top. This provides value to the user. - Bid. The search engine also allows companies to bid on search terms. The highest bids get to show their links at
the top of the search results, clearly marked as a “sponsored result.” The money companies pay for bids provide
value to the search engine company. - Referral. The user clicks on a link, ending up on the content creator’s website. This provides value to the
content creator.
The search engine model was not perfect, but it at least created the opportunity for all parties in this three-sided
marketplace to capture value: the user in step 3, the search engine company in step 4, and the content creator in
step 5.
In short, the current GenAI model destroys the incentives to create new content. I’ve heard this referred to as “the
great content collapse.” Will it lead to a world where, after the 2020s, there’s little-to-no content created by humans?
Will the state of knowledge and creativity stagnate as a result?
To be clear, I’m not an innocent party in this. As I mentioned numerous times in this post, I use GenAI regularly.
There’s no doubt that it makes me more productive. I even used GenAI to create the cover image for this blog post!
But each time I use ChatGPT or Claude, I feel a bit guilty, as it doesn’t feel sustainable. The snake can’t keep
eating its own tail indefinitely.
So the question is, what do we do? If we want to avoid the great content collapse, we need a model of GenAI usage that
creates opportunities for all parties (user, GenAI company, content creator) to capture value. Below are two ideas for
how we might accomplish this.
Possible solution #1: pay-per-crawl

The only attempted solution I’ve heard about so far is CloudFlare’s pay-per-crawl
model, which seems to work as follows:
- Block GenAI crawlers by default. CloudFlare claims to be able to detect and block GenAI crawlers, so that,
by default, they can’t extract content from your website. - Charge GenAI crawlers for each crawl. You can then selectively allow parts of your website to be accessed by
GenAI crawlers in exchange for payment.
On the one hand, it’s fantastic to see a major company try to do something about this problem. On the other
hand, this approach seems to address the wrong part of the problem. The real value isn’t in crawling the data, it’s
in using it. For every one crawl, an LLM might use the data thousands or millions of times. If creators are only paid
per crawl, then the GenAI company still captures 99.999% of the value, and the creator gets next to nothing. Moreover,
this model only seems to work for websites (it’s not clear how you adapt it to books, art, music, etc.), and it
creates an incentive for GenAI companies to only crawl free content, which means paid content is less likely to ever be
discovered (which disproportionally benefits those with pockets deep enough to keep their content free).
Possible solution #2: pay-per-use
I came across a clever solution that felt directionally correct from this LinkedIn post by Tyrone
Joel
where he took a PDF of my book Terraform: Up & Running, uploaded it into
a GenAI tool, and asked the tool to follow the guidance in the book to generate Terraform code. This feels like it has
all the ingredients of a model of GenAI usage that is sustainable: the user gets value from the GenAI tool’s responses,
the GenAI company gets value from the user paying for a subscription, and the content creator gets value from the user
paying for their content (in this case, buying my book). This works fine for a single, specific piece of content, but
how do you make it work at scale, across all the content that is consumed by an LLM?
Here’s a rough proposal for what I’ll call the pay-per-use model:
- Create. Someone creates valuable content.
- Extract. A GenAI bot crawls the web, exacting all the valuable content.
- Subscribe. Users pay money to use a GenAI tool. This provides value to the GenAI company.
- Prompt. Users enter a prompt into a GenAI tool. The GenAI tool generates a response based on the content it
extracted. The response provides value to the user. - Referral. For each response, the GenAI tool lists the sources from which it extracted that content, perhaps
formatted as a list of links back to the content creators, sorted by relevance, similar to a search engine (see
the mock-up below). This provides value to the content creator. It also gives users additional value, as now you
can check the sources of the information the GenAI tool is feeding you. Did that health advice come from a
peer-reviewed article written by an expert or a random hoax someone posted from 4chan? Is that generated piece
of code based on a hardened, mature library that’s been proven in production at thousands of companies or a
totally untested snippet that’s full of security vulnerabilities that was thrown together by a junior
programmer? Wouldn’t it be great to know? - Revenue share. GenAI tools could offer a revenue sharing model similar to the one used by subscription services
such as Netflix, Spotify, and LinkedIn Learning: the more your content is used in GenAI responses (the more it shows
up in the list of sources, as per the previous item), relative to all other content creators, the higher a cut of
revenue you get. Content creators can prove they own various pieces of content (e.g., prove you own a certain
domain name, open source project, or book) to claim their revenue cut.

This model works for not only websites, but other types of content too, including copyrighted content. As a content
creator (e.g., author, musician, designer, etc.), you could opt into sharing your copyrighted content with a GenAI
company in exchange for getting referrals and revenue sharing each time your content is used. It might even help with
making open source more sustainable, as open source creators could earn revenue and referrals each time a GenAI tool
uses their code.
Conclusion
It’s critical that we find a more sustainable model as soon as possible. The snake can’t eat its own tail indefinitely.
And the snake—GenAI—isn’t going away. We can’t put the genie back in the bottle. In fact, it’s only going to get
better, more ubiquitous, and to provide more and more value to users and GenAI companies. But if we can’t find a way to
provide value to content creators too, then this will all fall apart.
That said, I don’t know enough about LLMs to say if a pay-per-use model is actually possible. Can LLMs track the source
of the content they consumed? Will GenAI companies be willing to do a revenue sharing model? Will they be willing to
be transparent about their sources and usage? What do you think? Let me know in the comments.
⚡ **What’s your take?**
Share your thoughts in the comments below!
#️⃣ **#tools #ChatGPT #Claude #destroying #ecosystems #rely**
🕒 **Posted on**: 1769020223
🌟 **Want more?** Click here for more info! 🌟
