Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Sampling is already constrained with temperature, top_k, top_p, top_a, typical_p, min_p, entropy_penalty, smoothing etc. – filtering tokens to valid ones according to grammar is just yet another alternative. It does make sense and can be used for producing programming language output as well – what's the point in generating/bothering with up front know, invalid output? Better to filter it out and allow valid completions only.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: