chore: Speed up api through concurrency#385
chore: Speed up api through concurrency#385DaniilSmirnov wants to merge 4 commits intoAsyncFuncAI:mainfrom
Conversation
Summary of ChangesHello @DaniilSmirnov, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed! This pull request focuses on enhancing the overall performance and responsiveness of the API and frontend generation processes. By transitioning to asynchronous file operations, increasing API worker count, and boosting frontend concurrency, the system is now better equipped to handle higher loads and deliver a smoother user experience. Highlights
Using Gemini Code AssistThe full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips. Invoking Gemini You can request assistance from Gemini at any point by creating a comment using either
Customization To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a Limitations & Feedback Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here. You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension. Footnotes
|
There was a problem hiding this comment.
Code Review
This pull request aims to improve API performance by refactoring synchronous file operations to an asynchronous implementation using aiofiles. The changes are well-executed for the most part, correctly replacing os functions with their aioos counterparts for file existence checks, reading, writing, and removal. Increasing the number of uvicorn workers and client-side concurrent requests is also a sensible step towards better performance.
However, I've identified a critical issue where a synchronous os.walk() call remains within an async function, which will block the event loop and counteract the intended performance gains. Additionally, there's a potential startup crash issue related to how the number of workers is determined. My review includes detailed feedback and suggestions to address these points.
Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
…/reduce-api-response-time
Hi!
In this PR i refactor os using in api.py to async implementation
Also, i increase number of workers and threads for api and generation