Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The tool is testing your ability to predict whether or not GPT-4 can get a task done correctly. You are supposed to provide your confidence level that it can do a task or not.

If you answer all the questions, at the end the tool will tell you how well calibrated your beliefs about GPT-4s capabilities are, and how that calibration compares to other users of the tool.



Even though this webpage seems like it is actually sending to GPT-4 it seems that it is not as I tried a few and get very different answers.

Many that the webpage claims are wrong GPT-4 actually gets right. Maybe it's some recent changes, but the flight time question for example, I tried many and was never able to get GPT4 to return any incorrect answer.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: