Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Oh I definitely agree that there are multiple levels of AI research that are valuable. Huge supporter of open source, and not meaning to talk down to anyone working on AI projects.

It's just that at the moment I'm finding the open source LLM community hard to contextualize from an outside perspective. Maybe it's because things are moving so fast (probably a good thing).

I just know that personally, I'm not going to be exploring any projects until I know they're near or exceeding GPT-4 performance level. And it's hard to develop an interest in anything else other than GPT-4 when comparison is so tough to begin with.



I'd suggest reading the recently leaked Google memo for some context about why open source LLMs are important (and are disruptive from the perspective of a large company). It gives a good insight into why closed source models like GPT-4 might be overtaken by open source even if they can't directly compete at the moment.

Typical reasons are highly specialised models that are cheap and fast to train, lack of censorship, lack of API and usage restrictions, lightweight variants and so on. The reason there's a lot of excitement right now is indeed how fast the space is moving.

https://www.semianalysis.com/p/google-we-have-no-moat-and-ne...


This was great. The timeline was the perfect catch-up I needed. Thank you!




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: