OpenRouter ▷ #announcements (1 messages):
Kimi K2, Crashloop, Issue Resolution
- Kimi K2 Thinking Crashloop Crisis Averted: Kimi K2 Thinking experienced issues with both providers due to a crashloop triggered by a specific prompt.
- The issue has been resolved after collaborative efforts.
- Prompt-Induced Crashloop Plagues Kimi K2: A crashloop, induced by a problematic prompt, caused issues for Kimi K2 across multiple providers.
- Teams are actively collaborating to pinpoint and squash the gremlin causing the outage.
OpenRouter ▷ #app-showcase (7 messages):
Orchid AI Assistant, Release Date Estimation, The nature of work
- Orchid AI Assistant ETA: 2-48 Months!: The estimated release date for Orchid AI Assistant is projected to be within the next 2 to 48 months.
- A member reacted to this long and vague estimate with the word “crazy”.
- Contemplating the nature of ‘work’: A member expressed a dislike for “working,” suggesting that AI development aims to address this sentiment.
- The statement implies a desire to automate or alleviate the burdens associated with traditional labor through AI technologies.
OpenRouter ▷ #general (569 messages🔥🔥🔥):
OpenRouter video support, Polaris Alpha mini model, OpenAI adult content handling, Kimi K2 leaderboard rankings, Gemini 2.5 token usage
- OR may support videos in the future: A user expressed a wish for OpenRouter to support videos and text-to-speech (TTS) functionality, as shared in this tweet.
- Polaris Alpha possibly not a mini model: There is speculation that Polaris Alpha might not be a mini model, contrasting with the approach OpenAI took with GPT-5 as outlined in the GPT-5 System Card.
- OpenAI going adult - impacts OpenRouter: There is a question of how OpenRouter will handle OpenAI allowing adult content for users over 18, and whether users will need to bring their own API keys.
- Gemini 2.5 Flash chews through tokens: A user found that a 24-second, 900x600 video uploaded to Gemini 2.5 Flash consumed over 800k input tokens, contrary to Google’s documentation stating fixed rates of 263 tokens per second for video as mentioned in the token documentation.
- Cerebras Mandatory Reasoning: Users reported issues with the Cerebras model, where disabling reasoning caused errors; documentation confirms reasoning is mandatory.
- One workaround suggested was to omit the reasoning parameter altogether, after finding that
enableshould beenabledin the parameters.
- One workaround suggested was to omit the reasoning parameter altogether, after finding that
OpenRouter ▷ #new-models (2 messages):
“
- No New Models Discussion: There was no discussion about new models in the provided messages.
- The channel appears to be empty or the messages are not relevant to the topic.
- Absence of Relevant Content: No specific details or links related to model updates or technical discussions were found.
- The content might be missing or requires more context to generate meaningful summaries.
OpenRouter ▷ #discussion (29 messages🔥):
OpenRouter Model Node on n8n, OR Show Technical Segment, GPT-4 Regression, Chatroom Memory Setting, Automated Capabilities Scanning
- OpenRouter Model Node Inquiries Spark Curiosity: Members inquired whether the OpenRouter model node on n8n was created by the OpenRouter team or by an external entity.
- Another member suggested including a brief technical segment in the OR Show, such as a screen recording with a short discussion.
- GPT-4 Regression Troubles Users: Users reported a regression from GPT-4, with one noting that they were surprised to see the issue, and another saying Claude found two other discrepancies.
- The thread included attached images documenting discrepancies between different models on the platform.
- Chatroom ‘Memory’ Setting Misunderstood: A user inquired about the ‘chat history’ setting in the Chatroom, renamed as ‘Memory’, wondering what happened to it since the default value is 8.
- Another user clarified its location at the bottom, noting it was previously in the top left tab button, and another that they thought that would actually limit the $120/mtok output somehow.
- Automated Capabilities Scanning Proposed: A member suggested implementing some kind of automated capabilities scanning to detect changes in models/providers over time.
- They linked to an article on Cursor as an example, describing how a basic getWeather tool call could be used to check for functionality changes.
- GPT-5 Excels, Gemini Falls Flat: A user shared their positive experience using GPT-5 for creating schedules with nomenclature and filename structures, while noting their negative experience being locked out of Gemini code assist due to quota issues.
- They also mentioned needing to use DS3.1 for john-the-ripper help because Kimi refused, praising Meta’s under-the-radar AI projects and linking to a post on X to illustrate their point.