I was reading this article about LLMs making bad citations. I found it pretty interesting, so I decided to try to replicate it with ChatGPT.

cjr.org/tow_center/we-compared

I tried it with a document I wrote, FEP 5711. It's an enhancement proposal for ActivityPub, adding some inverse relationships for important properties.

codeberg.org/fediverse/fep/src

Show thread

Anyway, I took a paragraph out of the document and asked ChatGPT to identify the URL, publisher, publication date, and title. It failed. You can see the transcript here:

chatgpt.com/share/68573fa9-b34

Show thread

I was surprised to see that it had really no visibility of the FEPs. After a while, I realized that codeberg.org, the hosting service for FEPs, has ChatGPT blocked.

codeberg.org/robots.txt

Show thread

I understand the goal; many people don't want their code to be used by LLM code generators. But it also means that this document repository isn't visible for people who use LLMs like a search engine. Numbers vary, but afaict somewhere around 10% of people use LLMs as their primary search engine, and about 50% of people use LLMs some of the time for search.

Show thread

I guess there's maybe some justification like, those people are bad, and they don't deserve nice things like Fediverse Enhancement Proposals? Or, maybe, we have to take a principled stand against LLMs by not providing any training data for them? Such that, perhaps, people disappointed by not having good results in LLMs will return to using traditional search engines like Google or Bing, which are more ethical because reasons.

Show thread
Follow

@evan No, the reason is simply the server load. The AI crawlers have so excessively crawled @Codeberg that their main service, to host a git server, was often very slow.

@elshid @Codeberg that's interesting! Also, really self-destructive.

@evan @elshid @Codeberg AI crawlers descend on online resources like locust and consume then until they are dead. When they recover, they do it again.

@evan @Codeberg Addendum: My source (apart from a matrix channel): blog.codeberg.org/letter-from-

"Codeberg was running smoothly overall, except for one thing: Occassionally, AI scrapers crawled our site too much and caused downtimes."

@elshid @Codeberg @evan yes, absolutely agree - as long as LLM scraper bots don't follow simple rules (like following robots.txt files) and act not like a destruction swarm on services, I'm not sorry for LLM users not finding the information.

I've seen LLM-bot swarms hitting forges and the only way to get the forges running for the intended audience was blocking . Sad thing is the scrapers then try to circumvent the blocks aggressively, even dropping their IDs so the blocking gets more complicated again. It's an arms race, and legit users of LLMs are not in the focus of any of the racing parties.

@tobias @elshid @Codeberg well, that's hardcore, but the problem is that when people can't find your information on their chosen search engine, they don't go find another search engine. They don't even know they're missing your info.

@tobias @elshid @Codeberg @evan Yes. But I think the LLM companies are to blame here and not the ones who run servers that are hit by a near "denial of service" level of requests who also don't follow any rules about content that should be ignored.
Sign in to participate in the conversation
Librem Social

Librem Social is an opt-in public network. Messages are shared under Creative Commons BY-SA 4.0 license terms. Policy.

Stay safe. Please abide by our code of conduct.

(Source code)

image/svg+xml Librem Chat image/svg+xml