Hacker Newsnew | past | comments | ask | show | jobs | submit | gliched_robot's commentslogin

If anyone wants to try this out, here is the hugging-face spaces demo: https://huggingface.co/spaces/google/paligemma


This is far more superior than SORA, there is no comparison.


lol


Lmsys devs have all the answers, I am not sure how this has not leaked yet. They must a strong NDAs.


Why do they have any answers if they're just calling an API?


I do not understand the taught process here. They are regulating it so fast. It's almost like regulating car before even engine is invented.


ChatGPT exists.


GPU server locations, maybe?


LLM chat is so compute heavy and not bandwidth heavy that anywhere with reliable fiber and cheap electricity is suitable. Ping is lower than average keystroke delay for most who haven't undergone explicit speed typing training (we're talking 60~120 WPM for between intercontinental to pathological (other end of the world) servers). Bandwidth matters a bit more for multimodal interaction, but it's still rather minor.


Inference speed is not a great metric given the horizontal scalability of LLMs.


Disagree on Nvidia, most folks fine-tune model. Proof: there are about 20k models in huggingface derived from llama 2, all of them trained on Nvidia GPUs.


Fine tuning can take a fraction of the resources required for training, so I think the original point stands.


Maybe in isolation when only considering a single fine tune. But if you look at it in aggregate I am not so sure.


Maybe a typo?


I suppose it could be hallucinations about itself.

I suppose it's perfectly fair for large language models not necessarily to know these things, but as far as manual fine tuning, I think it would be reasonable to build models that are capable of answering questions about which model they are, their training date, their number of training parameters, and how they are different from other models, etc. Seems like it would be helpful for it to know and not have to try to do its best guess and potentially hallucinate. Although in my experience Llama 3 seemed to know what it was, but generally speaking it seems like this is not necessarily always the case.


This llama model some made it run on an iphone. https://x.com/1littlecoder/status/1781076849335861637?s=46


I see what you did here <q> carrying the "torch" <q>. LOL


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: