Hacker News new | past | comments | ask | show | jobs | submit login

  I love the original DeepSeek model, but the distilled versions are too dumb usually.
Apart from being dumber, they also don't know as much as R1. I can see how fine-tuning can improve reasoning capability (by showing examples of good CoT) but there's no reason that would improve the knowledge of facts (relative to the Qwen or Llama model on which the finetuning was based).



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: