I have the same problem, also using Langchain (Langgraph to be precise), but with a GPT model from OpenAI. The problem must be with how Langchain streams tokens from their chains. I would also be happy to know if there is a fix for this, which does not invovle using a custom function to manually add space between alphabetic tokens and numeric tokens.