Skip to main content
Glama

autonomous-frontend-browser-tools

FUTURE_PLANS.md6.38 kB
#### If you are interested in the future of this project 1. Give this repo a ⭐️ only if you liked it after using it. 2. Give me a follow on LInkedIn and X.com for future updates [https://www.linkedin.com/in/meet-limbani/](https://www.linkedin.com/in/meet-limbani/) [https://x.com/winds_ai](https://x.com/winds_ai). 3. Go to [https://windsurf.canny.io/feature-requests](https://windsurf.canny.io/feature-requests), search for these requests and upvote them so we can build upon this tool more. Feature Requests: - Show memories retrieved based on prompt - Switch the tool flow from predefined tools to code execution tool - Add custom model selection for tool call - Prompt field is resetted - Add a wait time tool call that has a check button to continue when I see fit # Future Plans for Autonomous Frontend Browser Tools This document outlines the future direction and planned enhancements for the Autonomous Frontend Browser Tools project. The goal is to further empower AI coding assistants by providing them with richer context and more intuitive tools, ultimately leading to more autonomous and efficient frontend development. ## Enhancing Tool Synergy and LLM Understanding Currently, I have all the tools set up, but LLMs still struggle to understand how to use these tools together in sync with each other and code autonomously. For now, when I prompt, I give it instructions on how to do this separately. To overcome that, I have to update the names and descriptions of tools so that the LLM can understand that it can use these tools in conjunction with each other and other tools and models needs to be improved to support using more and complex tools properly with their agentic behaviours. Gpt-5 works well with these tools. ## Automated Project Context Setup (Memories/Rules) Next, I currently have to manually create memories in Windsurf and rules files in Cursor so that the AI has full context of the project. I plan to introduce a new tool that can utilize the agents to understand the structure of the project and then call this tool with some arguments to automatically create rules files, i'm working on such prompts. My initial thought is that the agent will locate all the necessary files like authentication context, API call structure setup, utilities and helpers, how the pages and routing are structured along with navigation, and then make a tool call with paths of these files as arguments. Then, this tool will make a dynamic prompt based on the project setup and return that prompt. Now, the agent will take over again by understanding that prompt, and it will utilize its built-in tools to access those files and set up memories automatically. The main key will be how the prompt is set up. _This will require a lot of credits for researching and optimizing, so I am postponing this until I gather some money to spend on this one._ ## Advanced FRD Ingestion and Intelligent Retrieval OR I can shift to keyword based retrieval ( without a need of a Vector DB )on the FRD document ( sounds simple but it has potential in my initial testing) Another future set of tools that I plan to introduce are tools that ingest Functional Requirement Documents (FRDs) into a vector database by properly chunking them. We will be able to check its status, and the IDE agent can call this tool to understand what the requirements are and work accordingly. An FRD is a complex and big document with a lot of interdependencies and a complex layout and structure, especially a poorly written FRD. So, I have some ideas on how to solve this by making an agent instead of simple semantic-based retrieval. Initial thoughts are that this will either use Roaming RAG or an agent that can augment the query asked into multiple queries and keywords. This agent will have context of all the modules present in the FRD, so using this, it can craft queries that can identify potential interdependencies between my query and other modules present. This is to ensure that we don't miss any part from the FRD because of wordings and some expected interdependencies. There is a new research paper that I saw today discusses reasoning RAG, and that will be useful here. _This also requires a lot of credits and funds to create the agent, so I have to postpone this also._ ## Giving LLM Component and UI context Till now, I have focused on API integration and to work you have to have components setup to get the exact code and UI you want to get But the Reality is that we don't want that, we want our code to be based on our rules and components so that we can debug it anytime and make chnages even if we are not using LLM. If we give all that control to LLM then the output will be un-deterministic. For this, I think we need to make something like a directory of components accessible to the LLM, like how shadcn uses it's npx commnads to sever this network of components. What if we make an traversable file that has list of all the components and their configurable parameters like custom styles, props, etc. and then we can call this tool for the component and then it will return that data. so it would be better if all components library like MUI, antd, chakra, etc. have their own MCP tool that can do this. For now, I will try to make this tool by myself, let's see how this goes. Also to get accurate UI we need to maintain this information of whcih type of component are we using for which type of work. At the end it all comes down to memory efficiency and effectiveness ## Custom MCP Client Development Another significant limitation is that current IDE clients like Cursor, Windsurf, and others don't provide full MCP server functionality, particularly around advanced features like prompts, resource management, and extended tool capabilities. To unlock the full potential of our Autonomous Frontend Browser Tools, we might need to fork Cline and build our own custom MCP client. This would involve creating a specialized client with enhanced MCP protocol support, better tool orchestration, and seamless integration with our browser automation tools for truly autonomous frontend development workflows. --- I believe these future enhancements will significantly improve the capabilities of AI-assisted frontend development. Suggest me More Ideas if you come up with any even if it is small, it can make big impacts with other ideas.

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/Winds-AI/Frontend-development-MCP-tools-public'

If you have feedback or need assistance with the MCP directory API, please join our Discord server