client:495 [vite] connecting...
logger.ts:188 INFO LLMManager Registered 22 providers: AmazonBedrock, Anthropic, Cerebras, Cohere, Deepseek, Fireworks, Github, Google, Groq, HuggingFace, Hyperbolic, LMStudio, Mistral, Moonshot, Ollama, OpenAILike, OpenAI, OpenRouter, Perplexity, Together, xAI, Z.ai
client:618 [vite] connected.
logger.ts:188 INFO SettingsStore Loaded env key status from session cache
logger.ts:188 INFO Debug logger initialized
react-dom_client.js?v=98215f87:32136 Download the React DevTools for a better development experience: https://react.dev/link/react-devtools
entry.client.tsx:24 [Sentry] SENTRY_DSN not set — client-side error monitoring disabled.
(anonymous) @ entry.client.tsx:24
logger.ts:188 INFO AutoBackup Auto-backup started (every 5min)
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] 'message' handler took <N>ms
[Violation] Forced reflow while executing JavaScript took 49ms
logger.ts:188 INFO RuntimeClient Runtime booted for project "2" at /home/project
logger.ts:188 INFO Runtime Runtime booted for project "2"
react-dom_client.js?v=98215f87:25560 [Violation] 'success' handler took 206ms
react-dom_client.js?v=98215f87:25560 [Violation] 'success' handler took 256ms
logger.ts:188 INFO ChatHistory Auto-rebuild: Installing dependencies...
react-dom_client.js?v=98215f87:25560 [Violation] 'success' handler took 300ms
framer-motion.js?v=98215f87:475 [Violation] 'requestAnimationFrame' handler took 71ms
[Violation] 'input' handler took <N>ms
[Violation] 'input' handler took <N>ms
[Violation] 'input' handler took <N>ms
[Violation] 'input' handler took <N>ms
[Violation] 'input' handler took <N>ms
logger.ts:188 ERROR ChatHistory Auto-rebuild: npm install failed with exit code 1
react-dom_client.js?v=98215f87:25560 [Violation] 'keydown' handler took 190ms
logger.ts:188 INFO StreamEventRouter Structured streaming started
logger.ts:188 WARN VersionsStore All capture attempts returned placeholders — giving up
[Violation] Forced reflow while executing JavaScript took 30ms
[Violation] Forced reflow while executing JavaScript took 61ms
and terminal out put is: + html2canvas 1.4.1
+ ignore 6.0.2
+ isbot 4.4.0
+ isomorphic-git 1.32.1
+ istextorbinary 9.5.0
+ js-cookie 3.0.5
+ jspdf 4.2.0
+ jszip 3.10.1
+ nanostores 0.10.3
+ ollama-ai-provider 0.15.2
+ path-browserify 1.0.1
+ react 19.2.4
+ react-chartjs-2 5.3.0
+ react-dnd 16.0.1
+ react-dnd-html5-backend 16.0.1
+ react-dom 19.2.4
+ react-markdown 10.1.0
+ react-qrcode-logo 3.0.0
+ react-resizable-panels 2.1.9
+ react-router 7.13.1
+ react-window 2.2.7
+ rehype-raw 7.0.0
+ rehype-sanitize 6.0.0
+ remark-gfm 4.0.1
+ shiki 1.29.2
+ sonner 2.0.7
+ tailwind-merge 2.6.1
+ unist-util-visit 5.0.0
+ vite-plugin-node-polyfills 0.22.0
+ ws 8.19.0
+ zod 3.25.76
devDependencies:
+ @blitz/eslint-plugin 0.1.4
+ @commitlint/cli 20.5.0
+ @commitlint/config-conventional 20.5.0
+ @iconify-json/ph 1.2.2
+ @iconify-json/vscode-icons 1.2.43
+ @iconify/types 2.0.0
+ @playwright/test 1.58.2
+ @react-router/dev 7.13.1
+ @testing-library/jest-dom 6.9.1
+ @testing-library/react 16.3.2
+ @types/diff 5.2.3
+ @types/dom-speech-recognition 0.0.4
+ @types/file-saver 2.0.7
+ @types/js-cookie 3.0.6
+ @types/path-browserify 1.0.3
+ @types/react 19.2.14
+ @types/react-dom 19.2.3
+ @types/react-window 2.0.0
+ @types/ws 8.18.1
+ @vitejs/plugin-react 4.7.0
+ @vitest/coverage-v8 2.1.9
+ drizzle-kit 0.31.9
+ eslint-config-prettier 10.1.8
+ eslint-plugin-prettier 5.5.5
+ husky 9.1.7
+ knip 5.86.0
+ lint-staged 16.3.3
+ node-fetch 3.3.2
+ playwright 1.58.2
+ pnpm 9.15.9
+ prettier 3.8.1
+ rollup-plugin-visualizer 7.0.0
+ sass-embedded 1.98.0
+ typescript 5.9.3
+ unified 11.0.5
+ unocss 0.61.9
+ vite 5.4.21
+ vite-plugin-optimize-css-modules 1.2.0
+ vite-tsconfig-paths 4.3.2
+ vitest 2.1.9
> devonz@1.0.0 prepare /Users/ahmadazizi/Downloads/Devonz
> husky
Done in 36.7s
> devonz@1.0.0 dev /Users/ahmadazizi/Downloads/Devonz
> node pre-start.cjs && react-router dev
★═══════════════════════════════════════★
D E V O N Z
⚡️ Welcome ⚡️
★═══════════════════════════════════════★
📍 Current Version Tag: v"1.0.0"
📍 Current Commit Version: "ad1b0d66"
Please wait until the URL appears here
★═══════════════════════════════════════★
➜ Local: http://localhost:5173/
➜ Network: use --host to expose
➜ press h + enter to show help
INFO InitDecryptor Cookie decryptor registered
INFO LLMManager Registered 22 providers: Anthropic, Cerebras, Cohere, Deepseek, Fireworks, Google, Groq, HuggingFace, Hyperbolic, Mistral, Moonshot, Ollama, OpenAI, OpenRouter, OpenAILike, Perplexity, xAI, Together, LMStudio, AmazonBedrock, Github, Z.ai
INFO Database Initializing SQLite database at file:./data/devonz.db
INFO Database Database migrations applied successfully
WARN PreviewsStore Runtime not available, skipping init
WARN FilesStore Runtime not available, skipping init
11:07:21 AM [vite] ✨ new dependencies optimized: vite-plugin-node-polyfills/shims/buffer, vite-plugin-node-polyfills/shims/global, vite-plugin-node-polyfills/shims/process, @sentry/react, js-cookie, chalk
11:07:21 AM [vite] ✨ optimized dependencies changed. reloading
INFO InitDecryptor Cookie decryptor registered
INFO LLMManager Registered 22 providers: Anthropic, Cerebras, Cohere, Deepseek, Fireworks, Google, Groq, HuggingFace, Hyperbolic, Mistral, Moonshot, Ollama, OpenAI, OpenRouter, OpenAILike, Perplexity, xAI, Together, LMStudio, AmazonBedrock, Github, Z.ai
INFO Database Initializing SQLite database at file:./data/devonz.db
INFO Database Database migrations applied successfully
WARN PreviewsStore Runtime not available, skipping init
WARN FilesStore Runtime not available, skipping init
11:07:45 AM [vite] ✨ new dependencies optimized: sonner, clsx, tailwind-merge, class-variance-authority, istextorbinary, @octokit/rest, zod, @ai-sdk/openai, @ai-sdk/anthropic, @ai-sdk/cerebras, @ai-sdk/cohere, @ai-sdk/deepseek, @ai-sdk/fireworks, @ai-sdk/google, @ai-sdk/mistral, ollama-ai-provider, @openrouter/ai-sdk-provider, @ai-sdk/amazon-bedrock, ai
11:07:45 AM [vite] ✨ optimized dependencies changed. reloading
INFO InitDecryptor Cookie decryptor registered
INFO LLMManager Registered 22 providers: Anthropic, Cerebras, Cohere, Deepseek, Fireworks, Google, Groq, HuggingFace, Hyperbolic, Mistral, Moonshot, Ollama, OpenAI, OpenRouter, OpenAILike, Perplexity, xAI, Together, LMStudio, AmazonBedrock, Github, Z.ai
INFO Database Initializing SQLite database at file:./data/devonz.db
INFO Database Database migrations applied successfully
WARN PreviewsStore Runtime not available, skipping init
WARN FilesStore Runtime not available, skipping init
WARN CsrfService CSRF validation failed: cookie token missing
WARN Security CSRF validation failed [object Object]
ERROR Models Models loader failed TypeError: Cannot read properties of undefined (reading 'enabled')
ERROR LLMManager Error getting dynamic models LMStudio : TypeError: fetch failed
INFO LLMManager Caching 45 dynamic models for Google
11:08:15 AM [vite] ✨ new dependencies optimized: react-dnd, react-dnd-html5-backend, date-fns, isomorphic-git/http/web, react-resizable-panels
11:08:15 AM [vite] ✨ optimized dependencies changed. reloading
INFO InitDecryptor Cookie decryptor registered
INFO LLMManager Registered 22 providers: Anthropic, Cerebras, Cohere, Deepseek, Fireworks, Google, Groq, HuggingFace, Hyperbolic, Mistral, Moonshot, Ollama, OpenAI, OpenRouter, OpenAILike, Perplexity, xAI, Together, LMStudio, AmazonBedrock, Github, Z.ai
INFO Database Initializing SQLite database at file:./data/devonz.db
INFO Database Database migrations applied successfully
WARN PreviewsStore Runtime not available, skipping init
WARN FilesStore Runtime not available, skipping init
ERROR LLMManager Error getting dynamic models LMStudio : TypeError: fetch failed
INFO LLMManager Caching 45 dynamic models for Google
From https://github.com/zebbern/Devonz
* branch main -> FETCH_HEAD
ERROR LLMManager Error getting dynamic models LMStudio : TypeError: fetch failed
INFO GitClone Cloning https://github.com/devonz-templates/3d-vector-visualization-system.git → _clone_b9c85ddd
INFO GitClone Clone complete: 77 files read from https://github.com/devonz-templates/3d-vector-visualization-system.git
INFO GitClone Finalized clone: _clone_b9c85ddd → 1
INFO LocalRuntime Runtime booted for project "1" at /Users/ahmadazizi/.devonz/projects/1
INFO LocalRuntime Created runtime for project "1"
11:09:45 AM [vite] ✨ new dependencies optimized: react-markdown, remark-gfm, rehype-sanitize, rehype-raw, unist-util-visit
11:09:45 AM [vite] ✨ optimized dependencies changed. reloading
INFO InitDecryptor Cookie decryptor registered
INFO LLMManager Registered 22 providers: Anthropic, Cerebras, Cohere, Deepseek, Fireworks, Google, Groq, HuggingFace, Hyperbolic, Mistral, Moonshot, Ollama, OpenAI, OpenRouter, OpenAILike, Perplexity, xAI, Together, LMStudio, AmazonBedrock, Github, Z.ai
INFO Database Initializing SQLite database at file:./data/devonz.db
INFO Database Database migrations applied successfully
WARN PreviewsStore Runtime not available, skipping init
WARN FilesStore Runtime not available, skipping init
AbortError: This operation was aborted
at new DOMException (node:internal/per_context/domexception:53:5)
at AbortController.abort (node:internal/abort_controller:391:18)
at ServerResponse.<anonymous> (file:///Users/ahmadazizi/Downloads/Devonz/node_modules/.pnpm/@remix-run+node-fetch-server@0.13.0/node_modules/@remix-run/node-fetch-server/dist/lib/request-listener.js:90:41)
at Object.onceWrapper (node:events:638:28)
at ServerResponse.emit (node:events:524:28)
at emitCloseNT (node:_http_server:1021:10)
at Socket.onServerResponseClose (node:_http_server:277:5)
at Socket.emit (node:events:536:35)
at TCP.<anonymous> (node:net:343:12)
at TCP.callbackTrampoline (node:internal/async_hooks:130:17)
AbortError: This operation was aborted
at new DOMException (node:internal/per_context/domexception:53:5)
at AbortController.abort (node:internal/abort_controller:391:18)
at ServerResponse.<anonymous> (file:///Users/ahmadazizi/Downloads/Devonz/node_modules/.pnpm/@remix-run+node-fetch-server@0.13.0/node_modules/@remix-run/node-fetch-server/dist/lib/request-listener.js:90:41)
at Object.onceWrapper (node:events:638:28)
at ServerResponse.emit (node:events:524:28)
at emitCloseNT (node:_http_server:1021:10)
at Socket.onServerResponseClose (node:_http_server:277:5)
at Socket.emit (node:events:536:35)
at TCP.<anonymous> (node:net:343:12)
at TCP.callbackTrampoline (node:internal/async_hooks:130:17)
INFO LocalRuntime Runtime booted for project "1" at /Users/ahmadazizi/.devonz/projects/1
INFO LocalRuntime Created runtime for project "1"
ERROR LLMManager Error getting dynamic models LMStudio : TypeError: fetch failed
INFO LLMManager Caching 45 dynamic models for Google
INFO GitManager Auto-commit: 4d7fe51 — Project Setup
INFO CommandSafety [AUDIT] [terminal] project=1 cmd="/bin/bash --login -i"
INFO create-summary Summary cache MISS — calling LLM (hash: cca61492…)
INFO select-context Context cache MISS — calling LLM (hash: c1507673…)
WARN api.chat createSummary failed — continuing without summary: Error: Missing API key for AmazonBedrock provider
WARN api.chat selectContext failed — falling back to all files: Error: Missing API key for AmazonBedrock provider
INFO api.chat ⏱ Total context optimization (parallel): 8ms
INFO stream-text Token limits for model anthropic.claude-3-5-sonnet-20241022-v2:0: maxTokens=8192, maxTokenAllowed=200000, maxCompletionTokens=undefined
INFO stream-text Sending llm call to AmazonBedrock with model anthropic.claude-3-5-sonnet-20241022-v2:0
INFO stream-text Model "anthropic.claude-3-5-sonnet-20241022-v2:0" is reasoning model: false, using maxTokens: 8192
INFO stream-text DEBUG STREAM: Options filtering for model "anthropic.claude-3-5-sonnet-20241022-v2:0": {
"isReasoning": false,
"originalOptions": {
"supabaseConnection": {
"isConnected": false,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
},
"filteredOptions": {
"supabaseConnection": {
"isConnected": false,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
},
"originalOptionsKeys": [
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish"
],
"filteredOptionsKeys": [
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish"
],
"removedParams": []
}
WARN stream-recovery Stream timeout detected
INFO stream-recovery Attempting stream recovery (attempt 1)
WARN api.chat Stream timeout - attempting recovery
[unocss] failed to load icon "ph:server"
INFO LLMManager Found 45 cached models for Google
WARN stream-recovery Stream timeout detected
INFO stream-recovery Attempting stream recovery (attempt 2)
WARN api.chat Stream timeout - attempting recovery
INFO create-summary Summary cache MISS — calling LLM (hash: a89af8f4…)
INFO select-context Context cache MISS — calling LLM (hash: c1507673…)
WARN api.chat selectContext failed — falling back to all files: AI_APICallError: models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
WARN api.chat createSummary failed — continuing without summary: AI_APICallError: models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
INFO api.chat ⏱ Total context optimization (parallel): 394ms
INFO stream-text Token limits for model gemini-2.5-pro-preview-05-06: maxTokens=65536, maxTokenAllowed=1048576, maxCompletionTokens=65536
INFO stream-text Sending llm call to Google with model gemini-2.5-pro-preview-05-06
INFO stream-text Model "gemini-2.5-pro-preview-05-06" is reasoning model: false, using maxTokens: 65536
INFO stream-text DEBUG STREAM: Options filtering for model "gemini-2.5-pro-preview-05-06": {
"isReasoning": false,
"originalOptions": {
"supabaseConnection": {
"isConnected": false,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
},
"filteredOptions": {
"supabaseConnection": {
"isConnected": false,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
},
"originalOptionsKeys": [
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish"
],
"filteredOptionsKeys": [
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish"
],
"removedParams": []
}
INFO stream-text DEBUG STREAM: Final streaming params for model "gemini-2.5-pro-preview-05-06": {
"hasTemperature": false,
"hasMaxTokens": true,
"hasMaxCompletionTokens": false,
"paramKeys": [
"maxTokens",
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish",
"onChunk"
],
"streamParams": {
"maxTokens": 65536,
"supabaseConnection": {
"isConnected": false,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
}
}
ERROR stream-text Primary model Google/gemini-2.5-pro-preview-05-06 failed (model_not_found): models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
INFO stream-text Trying fallback candidate: Google/deep-research-pro-preview-12-2025
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/deep-research-pro-preview-12-2025 failed: The input token count exceeds the maximum number of tokens allowed (131072).
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash
WARN stream-text Fallback candidate Google/gemini-2.0-flash failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
Please retry in 32.012347233s.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash-001
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.0-flash-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
Please retry in 22.329205505s.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash-lite
INFO LLMManager Found 45 cached models for Google
WARN stream-recovery Stream timeout detected
ERROR stream-recovery Max retries reached for stream recovery
WARN stream-text Fallback candidate Google/gemini-2.0-flash-lite failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
Please retry in 11.233145886s.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash-lite-001
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.0-flash-lite-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
Please retry in 786.670055ms.
INFO stream-text Trying fallback candidate: Google/gemini-2.5-computer-use-preview-10-2025
INFO LLMManager Found 45 cached models for Google
WARN stream-recovery Stream timeout detected
INFO stream-recovery Attempting stream recovery (attempt 1)
WARN api.chat Stream timeout - attempting recovery
WARN stream-text Fallback candidate Google/gemini-2.5-computer-use-preview-10-2025 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
Please retry in 50.16257745s.
INFO stream-text Trying fallback candidate: Google/gemini-2.5-flash
INFO LLMManager Found 45 cached models for Google
INFO stream-text Fallback succeeded: Google/gemini-2.5-flash (primary Google/gemini-2.5-pro-preview-05-06 failed with model_not_found)
INFO api.chat ⏱ streamText completed in 60695ms
ERROR LLMManager Error getting dynamic models LMStudio : TypeError: fetch failed
INFO api.llmcall Generating response Provider: Google, Model: gemini-2.5-pro-preview-05-06
INFO LLMConstants Enabling Google thinking for gemini-2.5-pro-preview-05-06 (budget: 16384 tokens)
INFO api.llmcall Extended thinking enabled for Google/gemini-2.5-pro-preview-05-06 in llmcall
ERROR api.llmcall AI_APICallError: models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
INFO stream-text Token limits for model gemini-2.5-pro-preview-05-06: maxTokens=65536, maxTokenAllowed=1048576, maxCompletionTokens=65536
INFO stream-text Sending llm call to Google with model gemini-2.5-pro-preview-05-06
INFO stream-text Model "gemini-2.5-pro-preview-05-06" is reasoning model: false, using maxTokens: 65536
INFO stream-text DEBUG STREAM: Options filtering for model "gemini-2.5-pro-preview-05-06": {
"isReasoning": false,
"originalOptions": {
"supabaseConnection": {
"isConnected": false,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
},
"filteredOptions": {
"supabaseConnection": {
"isConnected": false,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
},
"originalOptionsKeys": [
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish"
],
"filteredOptionsKeys": [
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish"
],
"removedParams": []
}
INFO stream-text DEBUG STREAM: Final streaming params for model "gemini-2.5-pro-preview-05-06": {
"hasTemperature": false,
"hasMaxTokens": true,
"hasMaxCompletionTokens": false,
"paramKeys": [
"maxTokens",
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish",
"onChunk"
],
"streamParams": {
"maxTokens": 65536,
"supabaseConnection": {
"isConnected": false,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
}
}
INFO GitManager Git repo initialized at /Users/ahmadazizi/.devonz/projects/2
INFO LocalRuntime Runtime booted for project "2" at /Users/ahmadazizi/.devonz/projects/2
INFO LocalRuntime Created runtime for project "2"
ERROR stream-text Primary model Google/gemini-2.5-pro-preview-05-06 failed (model_not_found): models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
INFO stream-text Trying fallback candidate: Google/deep-research-pro-preview-12-2025
INFO LLMManager Found 45 cached models for Google
INFO CommandSafety [AUDIT] [terminal] project=2 cmd="/bin/bash --login -i"
WARN stream-text Fallback candidate Google/deep-research-pro-preview-12-2025 failed: This model only supports Interactions API.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash
WARN stream-text Fallback candidate Google/gemini-2.0-flash failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
Please retry in 56.722578406s.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash-001
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.0-flash-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
Please retry in 48.762281478s.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash-lite
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.0-flash-lite failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
Please retry in 41.881256221s.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash-lite-001
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.0-flash-lite-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
Please retry in 34.784434065s.
INFO stream-text Trying fallback candidate: Google/gemini-2.5-computer-use-preview-10-2025
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.5-computer-use-preview-10-2025 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: computer-use-preview
Please retry in 26.16960235s.
INFO stream-text Trying fallback candidate: Google/gemini-2.5-flash
INFO LLMManager Found 45 cached models for Google
WARN stream-recovery Stream timeout detected
INFO stream-recovery Attempting stream recovery (attempt 1)
WARN api.chat Stream timeout - attempting recovery
INFO stream-text Fallback succeeded: Google/gemini-2.5-flash (primary Google/gemini-2.5-pro-preview-05-06 failed with model_not_found)
INFO api.chat ⏱ streamText completed in 83086ms
INFO api.chat Skipping context optimization for short chat (3 messages ≤ 3) — using all files
INFO stream-text Token limits for model gemini-2.5-pro-preview-05-06: maxTokens=65536, maxTokenAllowed=1048576, maxCompletionTokens=65536
INFO stream-text Sending llm call to Google with model gemini-2.5-pro-preview-05-06
INFO stream-text Model "gemini-2.5-pro-preview-05-06" is reasoning model: false, using maxTokens: 65536
INFO stream-text DEBUG STREAM: Options filtering for model "gemini-2.5-pro-preview-05-06": {
"isReasoning": false,
"originalOptions": {
"supabaseConnection": {
"isConnected": false,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
},
"filteredOptions": {
"supabaseConnection": {
"isConnected": false,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
},
"originalOptionsKeys": [
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish"
],
"filteredOptionsKeys": [
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish"
],
"removedParams": []
}
INFO stream-text DEBUG STREAM: Final streaming params for model "gemini-2.5-pro-preview-05-06": {
"hasTemperature": false,
"hasMaxTokens": true,
"hasMaxCompletionTokens": false,
"paramKeys": [
"maxTokens",
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish",
"onChunk"
],
"streamParams": {
"maxTokens": 65536,
"supabaseConnection": {
"isConnected": false,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
}
}
ERROR stream-text Primary model Google/gemini-2.5-pro-preview-05-06 failed (model_not_found): models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
INFO stream-text Trying fallback candidate: Google/deep-research-pro-preview-12-2025
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/deep-research-pro-preview-12-2025 failed: This model only supports Interactions API.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash
WARN stream-text Fallback candidate Google/gemini-2.0-flash failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
Please retry in 534.971821ms.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash-001
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.0-flash-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
Please retry in 52.150566881s.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash-lite
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.0-flash-lite failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
Please retry in 44.486285675s.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash-lite-001
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.0-flash-lite-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
Please retry in 36.223944537s.
INFO stream-text Trying fallback candidate: Google/gemini-2.5-computer-use-preview-10-2025
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.5-computer-use-preview-10-2025 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: computer-use-preview
Please retry in 27.745361147s.
INFO stream-text Trying fallback candidate: Google/gemini-2.5-flash
INFO LLMManager Found 45 cached models for Google
WARN stream-recovery Stream timeout detected
INFO stream-recovery Attempting stream recovery (attempt 1)
WARN api.chat Stream timeout - attempting recovery
INFO stream-text Fallback succeeded: Google/gemini-2.5-flash (primary Google/gemini-2.5-pro-preview-05-06 failed with model_not_found)
INFO api.chat ⏱ streamText completed in 50106ms
INFO GitManager Auto-commit: e159802 — Run Lovable.dev Clone
INFO CommandSafety [AUDIT] [terminal] project=2 cmd="/bin/bash --login -i"
INFO LocalRuntime Idle timeout: tearing down runtime for "1" (inactive for 656s)
INFO LocalRuntime Tearing down runtime for project "1"
ERROR LLMManager Error getting dynamic models LMStudio : TypeError: fetch failed
INFO LocalRuntime Cleaning 2 orphaned session(s) for "2"
WARN LocalRuntime Session 05cfa8dd-1dfb-4ab0-891b-2d64274b4f07 did not exit within 5000ms — proceeding
WARN LocalRuntime Session 08ee0e1f-17b8-4404-979d-6dd96da93689 did not exit within 5000ms — proceeding
INFO CommandSafety [AUDIT] [terminal] project=2 cmd="/bin/bash --login -i"
ERROR LLMManager Error getting dynamic models LMStudio : TypeError: fetch failed
INFO LocalRuntime Cleaning 1 orphaned session(s) for "2"
WARN LocalRuntime Session ce49abe4-1715-435a-8198-4342ba4aae01 did not exit within 5000ms — proceeding
INFO CommandSafety [AUDIT] [terminal] project=2 cmd="/bin/bash --login -i"
INFO LocalRuntime Idle timeout: tearing down runtime for "1" (inactive for 882s)
INFO LocalRuntime Tearing down runtime for project "1"
INFO LocalRuntime Idle timeout: tearing down runtime for "2" (inactive for 878s)
INFO LocalRuntime Tearing down runtime for project "2"
WARN LocalRuntime Session 5c260dd5-cf24-4ad0-90a2-ae0af247e131 did not exit within 5000ms during teardown — proceeding
WARN LocalRuntime Session 1cc5584b-43bd-43a7-ba43-58c17099153f did not exit within 5000ms during teardown — proceeding
ERROR LLMManager Error getting dynamic models LMStudio : TypeError: fetch failed
INFO LocalRuntime Runtime booted for project "2" at /Users/ahmadazizi/.devonz/projects/2
INFO LocalRuntime Created runtime for project "2"
INFO CommandSafety [AUDIT] [terminal] project=2 cmd="/bin/bash --login -i"
INFO create-summary Summary cache MISS — calling LLM (hash: 38a6f8e4…)
INFO select-context Context cache MISS — calling LLM (hash: 42e87b94…)
WARN api.chat selectContext failed — falling back to all files: AI_APICallError: models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
WARN api.chat createSummary failed — continuing without summary: AI_APICallError: models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
INFO api.chat ⏱ Total context optimization (parallel): 681ms
INFO stream-text Token limits for model gemini-2.5-pro-preview-05-06: maxTokens=65536, maxTokenAllowed=1048576, maxCompletionTokens=65536
INFO stream-text Sending llm call to Google with model gemini-2.5-pro-preview-05-06
INFO stream-text Model "gemini-2.5-pro-preview-05-06" is reasoning model: false, using maxTokens: 65536
INFO stream-text DEBUG STREAM: Options filtering for model "gemini-2.5-pro-preview-05-06": {
"isReasoning": false,
"originalOptions": {
"supabaseConnection": {
"isConnected": true,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
},
"filteredOptions": {
"supabaseConnection": {
"isConnected": true,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
},
"originalOptionsKeys": [
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish"
],
"filteredOptionsKeys": [
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish"
],
"removedParams": []
}
INFO stream-text DEBUG STREAM: Final streaming params for model "gemini-2.5-pro-preview-05-06": {
"hasTemperature": false,
"hasMaxTokens": true,
"hasMaxCompletionTokens": false,
"paramKeys": [
"maxTokens",
"supabaseConnection",
"toolChoice",
"tools",
"maxSteps",
"agentMode",
"onStepFinish",
"onFinish",
"onChunk"
],
"streamParams": {
"maxTokens": 65536,
"supabaseConnection": {
"isConnected": true,
"hasSelectedProject": false,
"credentials": {}
},
"toolChoice": "auto",
"tools": {},
"maxSteps": 5,
"agentMode": false
}
}
ERROR stream-text Primary model Google/gemini-2.5-pro-preview-05-06 failed (model_not_found): models/gemini-2.5-pro-preview-05-06 is not found for API version v1beta, or is not supported for generateContent. Call ListModels to see the list of available models and their supported methods.
INFO stream-text Trying fallback candidate: Google/deep-research-pro-preview-12-2025
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/deep-research-pro-preview-12-2025 failed: This model only supports Interactions API.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash
WARN stream-text Fallback candidate Google/gemini-2.0-flash failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
Please retry in 58.259935932s.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash-001
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.0-flash-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash
Please retry in 49.656320514s.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash-lite
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.0-flash-lite failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
Please retry in 41.080654724s.
INFO stream-text Trying fallback candidate: Google/gemini-2.0-flash-lite-001
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.0-flash-lite-001 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: gemini-2.0-flash-lite
Please retry in 32.81503548s.
INFO stream-text Trying fallback candidate: Google/gemini-2.5-computer-use-preview-10-2025
INFO LLMManager Found 45 cached models for Google
WARN stream-text Fallback candidate Google/gemini-2.5-computer-use-preview-10-2025 failed: Failed after 3 attempts. Last error: You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits. To monitor your current usage, head to: https://ai.dev/rate-limit.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests, limit: 0, model: computer-use-preview
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_input_token_count, limit: 0, model: computer-use-preview
Please retry in 25.149524246s.
INFO stream-text Trying fallback candidate: Google/gemini-2.5-flash
INFO LLMManager Found 45 cached models for Google
WARN stream-recovery Stream timeout detected
INFO stream-recovery Attempting stream recovery (attempt 1)
WARN api.chat Stream timeout - attempting recovery
INFO stream-text Fallback succeeded: Google/gemini-2.5-flash (primary Google/gemini-2.5-pro-preview-05-06 failed with model_not_found)
INFO api.chat ⏱ streamText completed in 62672ms
What happened?
the ai keeps faliling to run basic commands
Steps to reproduce
1.build an app
2.try runnign basic commands
Error output
What should have happened?
it should have worked
Environment info
Anything else?