Why does o1 keep cutting the response short? It just stops as if it ran out of tokens.. Does reasoning_tokens eat into the completion tokens?
usage: {
prompt_tokens: 942,
completion_tokens: 4096,
total_tokens: 5038,
completion_tokens_details: { reasoning_tokens: 2972 }
},