Protecting content from LLMs, AI, etc.?

@matt Versara is a very interesting approach based on poisoning AI scraper responses:

This duality—poisoned content for scrapers, legitimate content for humans—is achieved by exploiting the differences in how humans and bots interact with web pages.

[…]

This approach is simple, but extremely powerful. Versara is detrimental to not only AI search tools, but every single scraper that we’ve tried.

If Versara doesn’t interfere with RSS readers it’s worth serious consideration. And it’s a lot of fun :grinning:

3 Likes

It’s also interesting to read the discussion on Versara on Lobsters.

This is really cool, thanks for sharing it! I’ve joined the waitlist so we can experiment with it here.

1 Like

As for progress, I have the basic ability here developed. I just need to make sure all bases are covered, including any potential downsides from this, before launching it for everyone. I have other priorities I need to address right now, but hope to revisit this in February.

2 Likes

I like the idea of this. But looking at their code example, I get a little nervous. They seem to separate humans from bots, but I suspect this interferes with assistive devices. Specifically screen readers.

@PaoloAmoroso, do you know anything about that?

Anyway, @matt, thank you for your diligence on this. It really is important that an anti-AI solution doesn’t have an inadvertent/negative impact.

1 Like

@markwyner I have no further details on Versara other than their website. Possible interference with assistive devices, RSS feed readers, and other benign scrapers is indeed a concern and I recommend reading the discussion on Versara on Lobsters.

1 Like

I was just looking at ways of blocking scraping the other day as I’ve started writing more actively. Thank you for continuing to work on this, it is very much appreciated!

1 Like