Sure, but what happens when the article is updated at a later date, or rescinded, etc.? Should the LLMs be trained to repeat the article verbatim, or to say "according to this article[0], blah blah blah" with links to the sources?
Wikipedia works because we can update it in real time in response to changes. LLMs that need to constantly recrawl every time a page on the internet is updated, and that properly contextualize the content of that page, is a huge ask. Because at that point, it stops being an LLM and starts being a very energy-hungry search engine.
Well, it's just a bot, so no need for it to instantly react to any and every update.
I also have my doubts on whether it is possible to implement efficiently (or at all). I suspect that just yanking in the article and all the sources is non-feasible, and any smaller chunking would be missing too much context. Plus LLM logical capabilities are questionable too, so I don't know how well the comparison would work...
Wikipedia works because we can update it in real time in response to changes. LLMs that need to constantly recrawl every time a page on the internet is updated, and that properly contextualize the content of that page, is a huge ask. Because at that point, it stops being an LLM and starts being a very energy-hungry search engine.