-
|
Hey, thanks for the great app.everything works fine with a non reasoning model like Claude 3.7 Sonnet, but when using Gemini 2.5 Pro (all via OpenRouter) its response is cut off to around 20 characters. Increasing the generated tokens does allow the response to be longer, but I have to have generated tokens set quite high for a full response, who h reduces the context significantly... I've seen a few issues like this on GH, but simply increasing the generated tokens slider doesn't seem like it actually solves the issue, so I'm not sure what to do here Thanks! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
Iirc OR model context lengths can be set in the app. Is there any reason why you can't set your Max Context to say, 32k and response length to 1k? Testing this on my device via OR w/ Gemini 2.5 Pro it seems to work fine. |
Beta Was this translation helpful? Give feedback.
Iirc OR model context lengths can be set in the app. Is there any reason why you can't set your Max Context to say, 32k and response length to 1k? Testing this on my device via OR w/ Gemini 2.5 Pro it seems to work fine.