[plug] [a bit OT but OSS] Building LLM with expert datasets
    Harry McNally 
    harrymc at decisions-and-designs.com.au
       
    Sun Feb  2 12:46:38 AWST 2025
    
    
  
Hello
I read an O'Reilly (I think) Radar or other newsletter an item linked to 
discussion about creating federated sets of LLMs (hosted by various companies) 
that are selected based on the query.
This would have to be of interest to O'Reilly because their published books 
and other resources would be higher quality training data.
When I spoke to James Bromberger about this his feedback was that, as Perl 
evolved, he had to buy each of the new O'Reilly books as they were 
significantly different with each new Perl version. Fair comment but it would 
be necessary for any programming LLM to have to not have "version" 
hallucination (or it isn't useful).
I have searched through emails and can't re-discover the reference or 
discussion. If anyone read a similar reference to federated sets of LLM that 
would be helpful to me.
I restarted thinking about this having got distracted today reading:
https://techstartups.com/2025/01/31/deepseek-r1-reproduced-for-30-berkeley-researchers-replicate-deepseek-r1-for-30-casting-doubt-on-h100-claims-and-controversy/
that led through TinyZero on GitHub:
https://github.com/Jiayi-Pan/TinyZero
to the veRL project and this talk (three months ago so old news to others I 
guess):
https://www.youtube.com/watch?v=MrhMcXkXvJU
and paper:
https://arxiv.org/abs/2409.19256
The whole flurry and upset seems to be because: "We haven't got the money, so 
we've got to think." https://www.azquotes.com/author/12817-Ernest_Rutherford
or in this case "We have lesser GPU's, so we've got to optimise".
All the best
Harry
    
    
More information about the plug
mailing list