Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It's such a wide range of model sizes that I could see why they compare with Llama 3 70b as well as Llama 3 8b (tables 12, 13). I agree that the Phi-3 series is a stronger competitor for knowledge extraction/summarizing and would make a good comparison. My current favorite for such tasks, on a VRAM-limited workstation, is Phi-3 medium (phi3:14b-instruct).




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: