๐Ÿ‘คgeorgehill๐Ÿ•‘2y๐Ÿ”ผ289๐Ÿ—จ๏ธ115

(Replying to PARENT post)

I would care more about LLaMA architecture when I can get hands on, honestly this project is more interesting and lighting fast on even a 2060 laptop https://github.com/BlinkDL/RWKV-LM
๐Ÿ‘คamrb๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

I don't really understand the benchmarking aspect researchers are touting. The public never cared about LLMs until they had a proper conversation with one. You can beat GPT3 at any benchmark you'd like, but if you can't get people that "feeling" when chatting with your model, is it worth anything?

In the future there's going to have to be a way to benchmark the "human-ness" or "intrigue" or "feistiness" of a model to show us if its getting better at what we want

๐Ÿ‘คshmatt๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

What is the purpose of this? The model from meta is not available to public. Neither this open source "LLaMA-based ChatGPT" nor the "open source" LLaMA can be downloaded or actually used by public because it would required the actual trained model.
๐Ÿ‘คSakiWatanabe๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

I'm as much a META hater as anyone - their policies have consistently disappointed me in almost every aspect of their business - but their stance on this LLaMA project I must say I am happy with and seems to mark a turn for the better.

If they follow through on their promise of making the weights available and share source code that is a big step in the right direction for democratising this technology

๐Ÿ‘คagolio๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

๐Ÿ‘คgeorgehill๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

I am very far from an expert on this, but I think domain specific conversational AI would be much more useful than these large models. It's fun to ask an AI to compose a fresh 600bpm hip hop song about the relationship between materials science and the breeding habits of mosquitoes, but an open-source medical AI, application support AI, or many other applications would be much more practical, if they could be accurate enough. And especially if they could run "standalone." They could also consult with each other, as a network of specialized AI. Is work inching closer to more specific, more accurate applications? Or is this just a big gimmick/distraction phase around a maybe not so great idea of AI?
๐Ÿ‘คdavidy123๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

In what way is this a ChatGPT implementation or equivalent? Seems like a chatbot based on a different backend, therefore it has absolutely zero link to ChatGPT.
๐Ÿ‘คlevesque๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Have we got any details on the benchmarks that show LLaMa's 13B architecture outperforming GPT-3? Because that seems kindsof fantastical. Is it just a product of a very specific benchmark or does it reflect real world performance?
๐Ÿ‘คdidntreadarticl๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

This whole debate - if a 13B model can really be as good as GPT3 - would have been settled if we had a live demo. I am not sure their licence allows running public demos, even if you get the weights.
๐Ÿ‘คvisarga๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Looks like they are making ChatGPT clone that would be possible to run a single GPU. HN dream come true!
๐Ÿ‘คrnosov๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Fake title riding on ChatGPT popularity. I think that it should be updated to something like:

Open source implementation for LLaMA-based chat bot*

Open source implementation for LLaMA-based ChatGPT alternative*

๐Ÿ‘คvoytec๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

I don't have a decent gpu at my disposal... has anyone tried to run LLaMA on an EC2 GPU instance? If so, which instance type? (I don't wanna overpay)
๐Ÿ‘คholtkam2๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Are LLaMA's weights generally available/floating around yet?
๐Ÿ‘คjstsch๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

open-assistant from LAION is in the process of creating an OSS RLHF dataset for a personal assistant, may be useful for this project
๐Ÿ‘คbethecloud๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Can someone leak the weights please
๐Ÿ‘คthreevox๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

This obsession with locking up model weights behind a gate-keeping application form and calling it open source is weird. I don't know who the high priests are trying to fool.

If your model is really that good, unleash it into the open so that others can truly evaluate it-warts and all-and help improve it by identifying the flaws.

๐Ÿ‘คvivegi๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

The difference between invention & innovation is that innovation is when you ship your product to the masses. Can I query llama in a line of code? No
๐Ÿ‘คlouis030195๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Is the trained model available to download anywhere?
๐Ÿ‘คgersh๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

> LLaMA is creating a lot of excitement because it is smaller than GPT-3 but has better performance. For example, LLaMA's 13B architecture outperforms GPT-3 despite being 10 times smaller.

Exactly. Best part is that it is open-source.

That is worth getting excited about. Not a AI SaaS API owned by a so-called pseudo-non profit company which struggles on API uptime and availablity, just like GitHub.

This is the 'revolution' you are looking for that changes everything. Not ChatGPT.

๐Ÿ‘คrvz๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

Is it of any good?
๐Ÿ‘คkarmasimida๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0

(Replying to PARENT post)

This seems like a great first step to a truly open source LLM
๐Ÿ‘คJack5500๐Ÿ•‘2y๐Ÿ”ผ0๐Ÿ—จ๏ธ0