Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I'm mostly a layman with ML stuff, so I might be doing something wrong, but I've not been impressed with Llama even at higher levels. I've run the 35B model in my home lab and it gave some pretty nonsensical responses. The 13B did better though, so could very well be user error.


There’s this gold rush going on, you’re right, any B without RLHF is meh.

The things getting published as “on device LLM” focus on bitcrushing the lowest B model with minimal RLHF and then pronouncing we have on device LLMs. We’ll definitely get there but signal >>> noise currently.

First person to admit this and write their blog post with A / B tests vs. a Markov chain deserves the gold.


Have you tried Alpaca yet? It's a massive improvement on base LLaMA.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: