Skip to content

[Bug]: wasting token #76

@DevFlex-AI

Description

@DevFlex-AI

What happened?

the ai keeps faliling to run basic commands

Image

Steps to reproduce

1.build an app
2.try runnign basic commands

Error output

client:495 [vite] connecting...
logger.ts:188 INFO LLMManager Registered 22 providers: AmazonBedrock, Anthropic, Cerebras, Cohere, Deepseek, Fireworks, Github, Google, Groq, HuggingFace, Hyperbolic, LMStudio, Mistral, Moonshot, Ollama, OpenAILike, OpenAI, OpenRouter, Perplexity, Together, xAI, Z.ai
client:618 [vite] connected.
logger.ts:188 INFO SettingsStore Loaded env key status from session cache
logger.ts:188 INFO Debug logger initialized
react-dom_client.js?v=98215f87:32136 Download the React DevTools for a better development experience: https://react.dev/link/react-devtools
entry.client.tsx:24 [Sentry] SENTRY_DSN not set — client-side error monitoring disabled.
(anonymous) @ entry.client.tsx:24
logger.ts:188 INFO AutoBackup Auto-backup started (every 5min)
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] Forced reflow while executing JavaScript took 49ms
logger.ts:188 INFO RuntimeClient Runtime booted for project "2" at /home/project
logger.ts:188 INFO Runtime Runtime booted for project "2"
react-dom_client.js?v=98215f87:25560 [Violation] 'success' handler took 206ms
react-dom_client.js?v=98215f87:25560 [Violation] 'success' handler took 256ms
logger.ts:188 INFO ChatHistory Auto-rebuild: Installing dependencies...
react-dom_client.js?v=98215f87:25560 [Violation] 'success' handler took 300ms
framer-motion.js?v=98215f87:475 [Violation] 'requestAnimationFrame' handler took 71ms
[Violation] 'input' handler took <N>ms
[Violation] 'input' handler took <N>ms
[Violation] 'input' handler took <N>ms
[Violation] 'input' handler took <N>ms
[Violation] 'input' handler took <N>ms
logger.ts:188 ERROR ChatHistory Auto-rebuild: npm install failed with exit code 1
react-dom_client.js?v=98215f87:25560 [Violation] 'keydown' handler took 190ms
logger.ts:188 INFO StreamEventRouter Structured streaming started
logger.ts:188 WARN VersionsStore All capture attempts returned placeholders — giving up
[Violation] Forced reflow while executing JavaScript took 30ms
[Violation] Forced reflow while executing JavaScript took 61ms
 and terminal out put is: + html2canvas 1.4.1
+ ignore 6.0.2
+ isbot 4.4.0
+ isomorphic-git 1.32.1
+ istextorbinary 9.5.0
+ js-cookie 3.0.5
+ jspdf 4.2.0
+ jszip 3.10.1
+ nanostores 0.10.3
+ ollama-ai-provider 0.15.2
+ path-browserify 1.0.1
+ react 19.2.4
+ react-chartjs-2 5.3.0
+ react-dnd 16.0.1
+ react-dnd-html5-backend 16.0.1
+ react-dom 19.2.4
+ react-markdown 10.1.0
+ react-qrcode-logo 3.0.0
+ react-resizable-panels 2.1.9
+ react-router 7.13.1
+ react-window 2.2.7
+ rehype-raw 7.0.0
+ rehype-sanitize 6.0.0
+ remark-gfm 4.0.1
+ shiki 1.29.2
+ sonner 2.0.7
+ tailwind-merge 2.6.1
+ unist-util-visit 5.0.0
+ vite-plugin-node-polyfills 0.22.0
+ ws 8.19.0
+ zod 3.25.76

devDependencies:
+ @blitz/eslint-plugin 0.1.4
+ @commitlint/cli 20.5.0
+ @commitlint/config-conventional 20.5.0
+ @iconify-json/ph 1.2.2
+ @iconify-json/vscode-icons 1.2.43
+ @iconify/types 2.0.0
+ @playwright/test 1.58.2
+ @react-router/dev 7.13.1
+ @testing-library/jest-dom 6.9.1
+ @testing-library/react 16.3.2
+ @types/diff 5.2.3
+ @types/dom-speech-recognition 0.0.4
+ @types/file-saver 2.0.7
+ @types/js-cookie 3.0.6
+ @types/path-browserify 1.0.3
+ @types/react 19.2.14
+ @types/react-dom 19.2.3
+ @types/react-window 2.0.0
+ @types/ws 8.18.1
+ @vitejs/plugin-react 4.7.0
+ @vitest/coverage-v8 2.1.9
+ drizzle-kit 0.31.9
+ eslint-config-prettier 10.1.8
+ eslint-plugin-prettier 5.5.5
+ husky 9.1.7
+ knip 5.86.0
+ lint-staged 16.3.3
+ node-fetch 3.3.2
+ playwright 1.58.2
+ pnpm 9.15.9
+ prettier 3.8.1
+ rollup-plugin-visualizer 7.0.0
+ sass-embedded 1.98.0
+ typescript 5.9.3
+ unified 11.0.5
+ unocss 0.61.9
+ vite 5.4.21
+ vite-plugin-optimize-css-modules 1.2.0
+ vite-tsconfig-paths 4.3.2
+ vitest 2.1.9

> devonz@1.0.0 prepare /Users/ahmadazizi/Downloads/Devonz
> husky


Done in 36.7s

> devonz@1.0.0 dev /Users/ahmadazizi/Downloads/Devonz
> node pre-start.cjs  && react-router dev


★═══════════════════════════════════════★
           D E V O N Z
         ⚡️  Welcome  ⚡️
★═══════════════════════════════════════★

📍 Current Version Tag: v"1.0.0"
📍 Current Commit Version: "ad1b0d66"
  Please wait until the URL appears here
★═══════════════════════════════════════★
  ➜  Local:   http://localhost:5173/
  ➜  Network: use --host to expose
  ➜  press h + enter to show help
 INFO   InitDecryptor  Cookie decryptor registered
 INFO   LLMManager  Registered 22 providers: Anthropic, Cerebras, Cohere, Deepseek, Fireworks, Google, Groq, HuggingFace, Hyperbolic, Mistral, Moonshot, Ollama, OpenAI, OpenRouter, OpenAILike, Perplexity, xAI, Together, LMStudio, AmazonBedrock, Github, Z.ai
 INFO   Database  Initializing SQLite database at file:./data/devonz.db
 INFO   Database  Database migrations applied successfully
 WARN   PreviewsStore  Runtime not available, skipping init
 WARN   FilesStore  Runtime not available, skipping init
11:07:21 AM [vite] ✨ new dependencies optimized: vite-plugin-node-polyfills/shims/buffer, vite-plugin-node-polyfills/shims/global, vite-plugin-node-polyfills/shims/process, @sentry/react, js-cookie, chalk
11:07:21 AM [vite] ✨ optimized dependencies changed. reloading
 INFO   InitDecryptor  Cookie decryptor registered
 INFO   LLMManager  Registered 22 providers: Anthropic, Cerebras, Cohere, Deepseek, Fireworks, Google, Groq, HuggingFace, Hyperbolic, Mistral, Moonshot, Ollama, OpenAI, OpenRouter, OpenAILike, Perplexity, xAI, Together, LMStudio, AmazonBedrock, Github, Z.ai
 INFO   Database  Initializing SQLite database at file:./data/devonz.db
 INFO   Database  Database migrations applied successfully
 WARN   PreviewsStore  Runtime not available, skipping init
 WARN   FilesStore  Runtime not available, skipping init
11:07:45 AM [vite] ✨ new dependencies optimized: sonner, clsx, tailwind-merge, class-variance-authority, istextorbinary, @octokit/rest, zod, @ai-sdk/openai, @ai-sdk/anthropic, @ai-sdk/cerebras, @ai-sdk/cohere, @ai-sdk/deepseek, @ai-sdk/fireworks, @ai-sdk/google, @ai-sdk/mistral, ollama-ai-provider, @openrouter/ai-sdk-provider, @ai-sdk/amazon-bedrock, ai
11:07:45 AM [vite] ✨ optimized dependencies changed. reloading
 INFO   InitDecryptor  Cookie decryptor registered
 INFO   LLMManager  Registered 22 providers: Anthropic, Cerebras, Cohere, Deepseek, Fireworks, Google, Groq, HuggingFace, Hyperbolic, Mistral, Moonshot, Ollama, OpenAI, OpenRouter, OpenAILike, Perplexity, xAI, Together, LMStudio, AmazonBedrock, Github, Z.ai
 INFO   Database  Initializing SQLite database at file:./data/devonz.db
 INFO   Database  Database migrations applied successfully
 WARN   PreviewsStore  Runtime not available, skipping init
 WARN   FilesStore  Runtime not available, skipping init
 WARN   CsrfService  CSRF validation failed: cookie token missing
 WARN   Security  CSRF validation failed [object Object]
 ERROR   Models  Models loader failed TypeError: Cannot read properties of undefined (reading 'enabled')
 ERROR   LLMManager  Error getting dynamic models LMStudio : TypeError: fetch failed
 INFO   LLMManager  Caching 45 dynamic models for Google
11:08:15 AM [vite] ✨ new dependencies optimized: react-dnd, react-dnd-html5-backend, date-fns, isomorphic-git/http/web, react-resizable-panels
11:08:15 AM [vite] ✨ optimized dependencies changed. reloading
 INFO   InitDecryptor  Cookie decryptor registered
 INFO   LLMManager  Registered 22 providers: Anthropic, Cerebras, Cohere, Deepseek, Fireworks, Google, Groq, HuggingFace, Hyperbolic, Mistral, Moonshot, Ollama, OpenAI, OpenRouter, OpenAILike, Perplexity, xAI, Together, LMStudio, AmazonBedrock, Github, Z.ai
 INFO   Database  Initializing SQLite database at file:./data/devonz.db
 INFO   Database  Database migrations applied successfully
 WARN   PreviewsStore  Runtime not available, skipping init
 WARN   FilesStore  Runtime not available, skipping init
 ERROR   LLMManager  Error getting dynamic models LMStudio : TypeError: fetch failed
 INFO   LLMManager  Caching 45 dynamic models for Google
From https://github.com/zebbern/Devonz
 * branch              main       -> FETCH_HEAD
 ERROR   LLMManager  Error getting dynamic models LMStudio : TypeError: fetch failed
 INFO   GitClone  Cloning https://github.com/devonz-templates/3d-vector-visualization-system.git → _clone_b9c85ddd
 INFO   GitClone  Clone complete: 77 files read from https://github.com/devonz-templates/3d-vector-visualization-system.git
 INFO   GitClone  Finalized clone: _clone_b9c85ddd → 1
 INFO   LocalRuntime  Runtime booted for project "1" at /Users/ahmadazizi/.devonz/projects/1
 INFO   LocalRuntime  Created runtime for project "1"
11:09:45 AM [vite] ✨ new dependencies optimized: react-markdown, remark-gfm, rehype-sanitize, rehype-raw, unist-util-visit
11:09:45 AM [vite] ✨ optimized dependencies changed. reloading
 INFO   InitDecryptor  Cookie decryptor registered
 INFO   LLMManager  Registered 22 providers: Anthropic, Cerebras, Cohere, Deepseek, Fireworks, Google, Groq, HuggingFace, Hyperbolic, Mistral, Moonshot, Ollama, OpenAI, OpenRouter, OpenAILike, Perplexity, xAI, Together, LMStudio, AmazonBedrock, Github, Z.ai
 INFO   Database  Initializing SQLite database at file:./data/devonz.db
 INFO   Database  Database migrations applied successfully
 WARN   PreviewsStore  Runtime not available, skipping init
 WARN   FilesStore  Runtime not available, skipping init
AbortError: This operation was aborted
    at new DOMException (node:internal/per_context/domexception:53:5)
    at AbortController.abort (node:internal/abort_controller:391:18)
    at ServerResponse.<anonymous> (file:///Users/ahmadazizi/Downloads/Devonz/node_modules/.pnpm/@remix-run+node-fetch-server@0.13.0/node_modules/@remix-run/node-fetch-server/dist/lib/request-listener.js:90:41)
    at Object.onceWrapper (node:events:638:28)
    at ServerResponse.emit (node:events:524:28)
    at emitCloseNT (node:_http_server:1021:10)
    at Socket.onServerResponseClose (node:_http_server:277:5)
    at Socket.emit (node:events:536:35)
    at TCP.<anonymous> (node:net:343:12)
    at TCP.callbackTrampoline (node:internal/async_hooks:130:17)
AbortError: This operation was aborted
    at new DOMException (node:internal/per_context/domexception:53:5)
    at AbortController.abort (node:internal/abort_controller:391:18)
    at ServerResponse.<anonymous> (file:///Users/ahmadazizi/Downloads/Devonz/node_modules/.pnpm/@remix-run+node-fetch-server@0.13.0/node_modules/@remix-run/node-fetch-server/dist/lib/request-listener.js:90:41)
    at Object.onceWrapper (node:events:638:28)
    at ServerResponse.emit (node:events:524:28)
    at emitCloseNT (node:_http_server:1021:10)
    at Socket.onServerResponseClose (node:_http_server:277:5)
    at Socket.emit (node:events:536:35)
    at TCP.<anonymous> (node:net:343:12)
    at TCP.callbackTrampoline (node:internal/async_hooks:130:17)
 INFO   LocalRuntime  Runtime booted for project "1" at /Users/ahmadazizi/.devonz/projects/1
 INFO   LocalRuntime  Created runtime for project "1"
 ERROR   LLMManager  Error getting dynamic models LMStudio : TypeError: fetch failed
 INFO   LLMManager  Caching 45 dynamic models for Google
 INFO   GitManager  Auto-commit: 4d7fe51 — Project Setup
 INFO   CommandSafety  [AUDIT] [terminal] project=1 cmd="/bin/bash --login -i"
 INFO   create-summary  Summary cache MISS — calling LLM (hash: cca61492…)
 INFO   select-context  Context cache MISS — calling LLM (hash: c1507673…)
 WARN   api.chat  createSummary failed — continuing without summary: Error: Missing API key for AmazonBedrock provider
 WARN   api.chat  selectContext failed — falling back to all files: Error: Missing API key for AmazonBedrock provider
 INFO   api.chat  ⏱ Total context optimization (parallel): 8ms
 INFO   stream-text  Token limits for model anthropic.claude-3-5-sonnet-20241022-v2:0: maxTokens=8192, maxTokenAllowed=200000, maxCompletionTokens=undefined
 INFO   stream-text  Sending llm call to AmazonBedrock with model anthropic.claude-3-5-sonnet-20241022-v2:0
 INFO   stream-text  Model "anthropic.claude-3-5-sonnet-20241022-v2:0" is reasoning model: false, using maxTokens: 8192
 INFO   stream-text  DEBUG STREAM: Options filtering for model "anthropic.claude-3-5-sonnet-20241022-v2:0": {
  "isReasoning": false,
  "originalOptions": {
    "supabaseConnection": {
      "isConnected": false,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  },
  "filteredOptions": {
    "supabaseConnection": {
      "isConnected": false,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  },
  "originalOptionsKeys": [
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish"
  ],
  "filteredOptionsKeys": [
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish"
  ],
  "removedParams": []
}
 WARN   stream-recovery  Stream timeout detected
 INFO   stream-recovery  Attempting stream recovery (attempt 1)
 WARN   api.chat  Stream timeout - attempting recovery
[unocss] failed to load icon "ph:server"
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-recovery  Stream timeout detected
 INFO   stream-recovery  Attempting stream recovery (attempt 2)
 WARN   api.chat  Stream timeout - attempting recovery
 INFO   create-summary  Summary cache MISS — calling LLM (hash: a89af8f4…)
 INFO   select-context  Context cache MISS — calling LLM (hash: c1507673…)
 WARN   api.chat  selectContext failed — falling back to all files: AI_APICallError: models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
 WARN   api.chat  createSummary failed — continuing without summary: AI_APICallError: models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
 INFO   api.chat  ⏱ Total context optimization (parallel): 394ms
 INFO   stream-text  Token limits for model gemini-2.5-pro-preview-05-06: maxTokens=65536, maxTokenAllowed=1048576, maxCompletionTokens=65536
 INFO   stream-text  Sending llm call to Google with model gemini-2.5-pro-preview-05-06
 INFO   stream-text  Model "gemini-2.5-pro-preview-05-06" is reasoning model: false, using maxTokens: 65536
 INFO   stream-text  DEBUG STREAM: Options filtering for model "gemini-2.5-pro-preview-05-06": {
  "isReasoning": false,
  "originalOptions": {
    "supabaseConnection": {
      "isConnected": false,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  },
  "filteredOptions": {
    "supabaseConnection": {
      "isConnected": false,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  },
  "originalOptionsKeys": [
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish"
  ],
  "filteredOptionsKeys": [
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish"
  ],
  "removedParams": []
}
 INFO   stream-text  DEBUG STREAM: Final streaming params for model "gemini-2.5-pro-preview-05-06": {
  "hasTemperature": false,
  "hasMaxTokens": true,
  "hasMaxCompletionTokens": false,
  "paramKeys": [
    "maxTokens",
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish",
    "onChunk"
  ],
  "streamParams": {
    "maxTokens": 65536,
    "supabaseConnection": {
      "isConnected": false,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  }
}
 ERROR   stream-text  Primary model Google/gemini-2.5-pro-preview-05-06 failed (model_not_found): models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
 INFO   stream-text  Trying fallback candidate: Google/deep-research-pro-preview-12-2025
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/deep-research-pro-preview-12-2025 failed: The input token count exceeds the maximum number of tokens allowed (131072).
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
Please retry in 32.012347233s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash-001
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
Please retry in 22.329205505s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash-lite
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-recovery  Stream timeout detected
 ERROR   stream-recovery  Max retries reached for stream recovery
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash-lite failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
Please retry in 11.233145886s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash-lite-001
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash-lite-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
Please retry in 786.670055ms.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.5-computer-use-preview-10-2025
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-recovery  Stream timeout detected
 INFO   stream-recovery  Attempting stream recovery (attempt 1)
 WARN   api.chat  Stream timeout - attempting recovery
 WARN   stream-text  Fallback candidate Google/gemini-2.5-computer-use-preview-10-2025 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
Please retry in 50.16257745s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.5-flash
 INFO   LLMManager  Found 45 cached models for Google
 INFO   stream-text  Fallback succeeded: Google/gemini-2.5-flash (primary Google/gemini-2.5-pro-preview-05-06 failed with model_not_found)
 INFO   api.chat  ⏱ streamText completed in 60695ms
 ERROR   LLMManager  Error getting dynamic models LMStudio : TypeError: fetch failed
 INFO   api.llmcall  Generating response Provider: Google, Model: gemini-2.5-pro-preview-05-06
 INFO   LLMConstants  Enabling Google thinking for gemini-2.5-pro-preview-05-06 (budget: 16384 tokens)
 INFO   api.llmcall  Extended thinking enabled for Google/gemini-2.5-pro-preview-05-06 in llmcall
 ERROR   api.llmcall  AI_APICallError: models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
 INFO   stream-text  Token limits for model gemini-2.5-pro-preview-05-06: maxTokens=65536, maxTokenAllowed=1048576, maxCompletionTokens=65536
 INFO   stream-text  Sending llm call to Google with model gemini-2.5-pro-preview-05-06
 INFO   stream-text  Model "gemini-2.5-pro-preview-05-06" is reasoning model: false, using maxTokens: 65536
 INFO   stream-text  DEBUG STREAM: Options filtering for model "gemini-2.5-pro-preview-05-06": {
  "isReasoning": false,
  "originalOptions": {
    "supabaseConnection": {
      "isConnected": false,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  },
  "filteredOptions": {
    "supabaseConnection": {
      "isConnected": false,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  },
  "originalOptionsKeys": [
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish"
  ],
  "filteredOptionsKeys": [
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish"
  ],
  "removedParams": []
}
 INFO   stream-text  DEBUG STREAM: Final streaming params for model "gemini-2.5-pro-preview-05-06": {
  "hasTemperature": false,
  "hasMaxTokens": true,
  "hasMaxCompletionTokens": false,
  "paramKeys": [
    "maxTokens",
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish",
    "onChunk"
  ],
  "streamParams": {
    "maxTokens": 65536,
    "supabaseConnection": {
      "isConnected": false,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  }
}
 INFO   GitManager  Git repo initialized at /Users/ahmadazizi/.devonz/projects/2
 INFO   LocalRuntime  Runtime booted for project "2" at /Users/ahmadazizi/.devonz/projects/2
 INFO   LocalRuntime  Created runtime for project "2"
 ERROR   stream-text  Primary model Google/gemini-2.5-pro-preview-05-06 failed (model_not_found): models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
 INFO   stream-text  Trying fallback candidate: Google/deep-research-pro-preview-12-2025
 INFO   LLMManager  Found 45 cached models for Google
 INFO   CommandSafety  [AUDIT] [terminal] project=2 cmd="/bin/bash --login -i"
 WARN   stream-text  Fallback candidate Google/deep-research-pro-preview-12-2025 failed: This model only supports Interactions API.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
Please retry in 56.722578406s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash-001
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
Please retry in 48.762281478s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash-lite
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash-lite failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
Please retry in 41.881256221s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash-lite-001
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash-lite-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
Please retry in 34.784434065s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.5-computer-use-preview-10-2025
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.5-computer-use-preview-10-2025 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: computer-use-preview
Please retry in 26.16960235s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.5-flash
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-recovery  Stream timeout detected
 INFO   stream-recovery  Attempting stream recovery (attempt 1)
 WARN   api.chat  Stream timeout - attempting recovery
 INFO   stream-text  Fallback succeeded: Google/gemini-2.5-flash (primary Google/gemini-2.5-pro-preview-05-06 failed with model_not_found)
 INFO   api.chat  ⏱ streamText completed in 83086ms
 INFO   api.chat  Skipping context optimization for short chat (3 messages ≤ 3) — using all files
 INFO   stream-text  Token limits for model gemini-2.5-pro-preview-05-06: maxTokens=65536, maxTokenAllowed=1048576, maxCompletionTokens=65536
 INFO   stream-text  Sending llm call to Google with model gemini-2.5-pro-preview-05-06
 INFO   stream-text  Model "gemini-2.5-pro-preview-05-06" is reasoning model: false, using maxTokens: 65536
 INFO   stream-text  DEBUG STREAM: Options filtering for model "gemini-2.5-pro-preview-05-06": {
  "isReasoning": false,
  "originalOptions": {
    "supabaseConnection": {
      "isConnected": false,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  },
  "filteredOptions": {
    "supabaseConnection": {
      "isConnected": false,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  },
  "originalOptionsKeys": [
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish"
  ],
  "filteredOptionsKeys": [
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish"
  ],
  "removedParams": []
}
 INFO   stream-text  DEBUG STREAM: Final streaming params for model "gemini-2.5-pro-preview-05-06": {
  "hasTemperature": false,
  "hasMaxTokens": true,
  "hasMaxCompletionTokens": false,
  "paramKeys": [
    "maxTokens",
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish",
    "onChunk"
  ],
  "streamParams": {
    "maxTokens": 65536,
    "supabaseConnection": {
      "isConnected": false,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  }
}
 ERROR   stream-text  Primary model Google/gemini-2.5-pro-preview-05-06 failed (model_not_found): models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
 INFO   stream-text  Trying fallback candidate: Google/deep-research-pro-preview-12-2025
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/deep-research-pro-preview-12-2025 failed: This model only supports Interactions API.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
Please retry in 534.971821ms.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash-001
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
Please retry in 52.150566881s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash-lite
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash-lite failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
Please retry in 44.486285675s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash-lite-001
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash-lite-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
Please retry in 36.223944537s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.5-computer-use-preview-10-2025
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.5-computer-use-preview-10-2025 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: computer-use-preview
Please retry in 27.745361147s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.5-flash
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-recovery  Stream timeout detected
 INFO   stream-recovery  Attempting stream recovery (attempt 1)
 WARN   api.chat  Stream timeout - attempting recovery
 INFO   stream-text  Fallback succeeded: Google/gemini-2.5-flash (primary Google/gemini-2.5-pro-preview-05-06 failed with model_not_found)
 INFO   api.chat  ⏱ streamText completed in 50106ms
 INFO   GitManager  Auto-commit: e159802 — Run Lovable.dev Clone
 INFO   CommandSafety  [AUDIT] [terminal] project=2 cmd="/bin/bash --login -i"
 INFO   LocalRuntime  Idle timeout: tearing down runtime for "1" (inactive for 656s)
 INFO   LocalRuntime  Tearing down runtime for project "1"
 ERROR   LLMManager  Error getting dynamic models LMStudio : TypeError: fetch failed
 INFO   LocalRuntime  Cleaning 2 orphaned session(s) for "2"
 WARN   LocalRuntime  Session 05cfa8dd-1dfb-4ab0-891b-2d64274b4f07 did not exit within 5000ms — proceeding
 WARN   LocalRuntime  Session 08ee0e1f-17b8-4404-979d-6dd96da93689 did not exit within 5000ms — proceeding
 INFO   CommandSafety  [AUDIT] [terminal] project=2 cmd="/bin/bash --login -i"
 ERROR   LLMManager  Error getting dynamic models LMStudio : TypeError: fetch failed
 INFO   LocalRuntime  Cleaning 1 orphaned session(s) for "2"
 WARN   LocalRuntime  Session ce49abe4-1715-435a-8198-4342ba4aae01 did not exit within 5000ms — proceeding
 INFO   CommandSafety  [AUDIT] [terminal] project=2 cmd="/bin/bash --login -i"
 INFO   LocalRuntime  Idle timeout: tearing down runtime for "1" (inactive for 882s)
 INFO   LocalRuntime  Tearing down runtime for project "1"
 INFO   LocalRuntime  Idle timeout: tearing down runtime for "2" (inactive for 878s)
 INFO   LocalRuntime  Tearing down runtime for project "2"
 WARN   LocalRuntime  Session 5c260dd5-cf24-4ad0-90a2-ae0af247e131 did not exit within 5000ms during teardown — proceeding
 WARN   LocalRuntime  Session 1cc5584b-43bd-43a7-ba43-58c17099153f did not exit within 5000ms during teardown — proceeding
 ERROR   LLMManager  Error getting dynamic models LMStudio : TypeError: fetch failed
 INFO   LocalRuntime  Runtime booted for project "2" at /Users/ahmadazizi/.devonz/projects/2
 INFO   LocalRuntime  Created runtime for project "2"
 INFO   CommandSafety  [AUDIT] [terminal] project=2 cmd="/bin/bash --login -i"
 INFO   create-summary  Summary cache MISS — calling LLM (hash: 38a6f8e4…)
 INFO   select-context  Context cache MISS — calling LLM (hash: 42e87b94…)
 WARN   api.chat  selectContext failed — falling back to all files: AI_APICallError: models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
 WARN   api.chat  createSummary failed — continuing without summary: AI_APICallError: models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
 INFO   api.chat  ⏱ Total context optimization (parallel): 681ms
 INFO   stream-text  Token limits for model gemini-2.5-pro-preview-05-06: maxTokens=65536, maxTokenAllowed=1048576, maxCompletionTokens=65536
 INFO   stream-text  Sending llm call to Google with model gemini-2.5-pro-preview-05-06
 INFO   stream-text  Model "gemini-2.5-pro-preview-05-06" is reasoning model: false, using maxTokens: 65536
 INFO   stream-text  DEBUG STREAM: Options filtering for model "gemini-2.5-pro-preview-05-06": {
  "isReasoning": false,
  "originalOptions": {
    "supabaseConnection": {
      "isConnected": true,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  },
  "filteredOptions": {
    "supabaseConnection": {
      "isConnected": true,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  },
  "originalOptionsKeys": [
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish"
  ],
  "filteredOptionsKeys": [
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish"
  ],
  "removedParams": []
}
 INFO   stream-text  DEBUG STREAM: Final streaming params for model "gemini-2.5-pro-preview-05-06": {
  "hasTemperature": false,
  "hasMaxTokens": true,
  "hasMaxCompletionTokens": false,
  "paramKeys": [
    "maxTokens",
    "supabaseConnection",
    "toolChoice",
    "tools",
    "maxSteps",
    "agentMode",
    "onStepFinish",
    "onFinish",
    "onChunk"
  ],
  "streamParams": {
    "maxTokens": 65536,
    "supabaseConnection": {
      "isConnected": true,
      "hasSelectedProject": false,
      "credentials": {}
    },
    "toolChoice": "auto",
    "tools": {},
    "maxSteps": 5,
    "agentMode": false
  }
}
 ERROR   stream-text  Primary model Google/gemini-2.5-pro-preview-05-06 failed (model_not_found): models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
 INFO   stream-text  Trying fallback candidate: Google/deep-research-pro-preview-12-2025
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/deep-research-pro-preview-12-2025 failed: This model only supports Interactions API.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
Please retry in 58.259935932s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash-001
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
Please retry in 49.656320514s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash-lite
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash-lite failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
Please retry in 41.080654724s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.0-flash-lite-001
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.0-flash-lite-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
Please retry in 32.81503548s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.5-computer-use-preview-10-2025
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-text  Fallback candidate Google/gemini-2.5-computer-use-preview-10-2025 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: computer-use-preview
Please retry in 25.149524246s.
 INFO   stream-text  Trying fallback candidate: Google/gemini-2.5-flash
 INFO   LLMManager  Found 45 cached models for Google
 WARN   stream-recovery  Stream timeout detected
 INFO   stream-recovery  Attempting stream recovery (attempt 1)
 WARN   api.chat  Stream timeout - attempting recovery
 INFO   stream-text  Fallback succeeded: Google/gemini-2.5-flash (primary Google/gemini-2.5-pro-preview-05-06 failed with model_not_found)
 INFO   api.chat  ⏱ streamText completed in 62672ms

What should have happened?

it should have worked

Environment info

node 20 mac os 15

Anything else?

Image

Metadata

Metadata

Assignees

Labels

bugSomething isn't workingtriageNeeds triage and prioritization

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions