I don't think this is true, unless adding 'pdf' to the end of the name is very specialized. Even "Xanthar's Guide to Everything" gives me the full text as the second result, after D&DBeyond.
Yes, adding pdf is specialized, because 5e books are not typically sold as pdfs. The only way to obtain a 5e pdf is through scanning it in. And sure, if you
just google "Xanathar's Guide to Everything," you get a link to D&D beyond... where you can
buy the book for $29.95 (digital only--
not a pdf). If you google "Xanathar's Guide to Everything pdf" you get the Internet Archive, Anyflip, Scribd, and other sites where you obtain a pdf of the book without paying for it.
But when I type in "savage worlds pdf," I get a link to the actual PEG website, because those books are also sold in pdf form.
As Scribe said, I'm not weeping if people do this with Xannie's because Hasbro can afford to eat the loss. But that doesn't mean it's
right.
Do you think that Gemini had access to that material because it has been trained on copywritten data, or is it sufficient to have access to the internet? That is, if it never actually saw the Xanathar's book but was able to trawl Enworld, Reddit, the D&DBeyond forums, and anywhere else, would it be able to reconstruct those mechanics?
We know that Meta, at least, was trained on copyrighted data. We have actual proof of this. There have been lawsuits filed. Apparently (
as of two days ago) Meta is claiming that the 7 million books it pirated had "no economic value" and that they're protected under "
fair use" because, they claim, they don't reproduce the entire book.
Now, I got Gemini to pretty much reproduce the entirety of the grave cleric, which is not OGL. Which means that the idea that AI won't reproduce copyrighten material is bogus. Maybe some AIs won't, but others will.
So we have two options here:
(1) Gemini, like Meta, was trained on stolen material. This is morally, ethically, and financially wrong, no matter what (generic) you think about WotC or Hasbro.
(2) Gemini wasn't specifically trained on stolen data, but got its info from the internet. In this case, Gemini is dangerous because it's grabbing things randomly from online, which means its going to be giving false or even harmful information... such as in the case of the
poisonous mushrooms. (I admit I don't know which AI wrote that book.) While a D&D character isn't going to kill someone, it could--since it's picking information randonly instead being trained--give the wrong class information, which could lead to in-game problems.
And unless you have some evidence that Meta, ChatGPT, or whatever other AI you're using is better trained, smarter, or whatever, I see no reason not to assume that that other AI programs are trained in the same way.