this post was submitted on 22 Jun 2025
781 points (94.4% liked)

Technology

71843 readers
4256 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.

Then retrain on that.

Far too much garbage in any foundation model trained on uncorrected data.

Source.

More Context

Source.

Source.

you are viewing a single comment's thread
view the rest of the comments
[–] hansolo@lemmy.today 7 points 2 days ago* (last edited 2 days ago)

Musk probably heard about "synthetic data" training, which is where you use machine learning to create thousands of things that are typical-enough to be good training data. Microsoft uses it to take documents users upload to Office365, train the ML model, and then use that ML output to train an LLM so they can technically say "no, your data wasn't used to train an LLM." Because it trained the thing that trained the LLM.

However, you can't do that with LLM output and stuff like... History. WTF evidence and documents are the basis for the crap he wants to add? The hallucinations will just compound because who's going to cross-check this other than Grok anyway?