This is extremely interesting: The authors look at features (like making poetry, or calculating) of LLM production, make hypotheses about internal strategies to achieve the result, and experiment with these hypotheses.
I wonder if there is somewhere an explanation linking the logical operations made on a on dataset, are resulting in those behaviors?
I wonder if there is somewhere an explanation linking the logical operations made on a on dataset, are resulting in those behaviors?