Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

If you don't mind me asking, what sentence embeddings model (bert/roberta/etc) did you have the best luck with for your classifier? I like the quick retrain that can be done with an approach like this, though I have found that if you throw too many different SPAM profiles at a classifier it starts to degrade, and you might have to build multiple and ensemble them. The embedding backend can help a lot with that.


Tried bert but didn't get the proper result, probably wasn't working with it properly.

Here's the old source I have on my computer that did the training

https://gist.github.com/Alir3z4/6b26353928633f7db59f40f71c8f...

This was doing the early work and later changed more to fit other cases.

Pretty basic stuff.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: