[plug] [a bit OT but OSS] Building LLM with expert datasets

Harry McNally harrymc at decisions-and-designs.com.au
Sun Feb 2 12:46:38 AWST 2025


Hello

I read an O'Reilly (I think) Radar or other newsletter an item linked to 
discussion about creating federated sets of LLMs (hosted by various companies) 
that are selected based on the query.

This would have to be of interest to O'Reilly because their published books 
and other resources would be higher quality training data.

When I spoke to James Bromberger about this his feedback was that, as Perl 
evolved, he had to buy each of the new O'Reilly books as they were 
significantly different with each new Perl version. Fair comment but it would 
be necessary for any programming LLM to have to not have "version" 
hallucination (or it isn't useful).

I have searched through emails and can't re-discover the reference or 
discussion. If anyone read a similar reference to federated sets of LLM that 
would be helpful to me.

I restarted thinking about this having got distracted today reading:

https://techstartups.com/2025/01/31/deepseek-r1-reproduced-for-30-berkeley-researchers-replicate-deepseek-r1-for-30-casting-doubt-on-h100-claims-and-controversy/

that led through TinyZero on GitHub:

https://github.com/Jiayi-Pan/TinyZero

to the veRL project and this talk (three months ago so old news to others I 
guess):

https://www.youtube.com/watch?v=MrhMcXkXvJU

and paper:

https://arxiv.org/abs/2409.19256

The whole flurry and upset seems to be because: "We haven't got the money, so 
we've got to think." https://www.azquotes.com/author/12817-Ernest_Rutherford

or in this case "We have lesser GPU's, so we've got to optimise".

All the best
Harry







More information about the plug mailing list