Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The papers from Anthropic on interpretability are pretty good. They look at how certain concepts are encoded within the LLM.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: