It kept truncating files only about 600 lines long. It also seems to rewrite the entire file each time instead of just sending diffs like aider making it super slow.
oh, I see your point now. Its weird that they are not doing the search and replace style editing.
Altho now that OpenAI also has Predicted Output, I think this will improve and it won't make mistakes while rewriting longer files.
The 600 line limit might be due to the output token limit on the LLM (not sure what they are using for the code rewriting)
It's not nearly as helpful as Claude.ai - it seems to only want to do the minimum required. On top of that it will quite regularly ignore what you've asked, give you back the exact code you gave it, or even generate syntactically invalid code.
It's amazing how much difference the prompt must make because using it is like going back to gpt3.5 yet it's the same model.