Stores and manages scraped LLM inference pricing data from various providers in a local database for querying and comparison.
Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@LLM Inference Pricing Research Servercompare cloudrift and deepinfra pricing for llama 3.1"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
In this project, you are going to make a chatbot to scrape LLM Inference Serving websites to research costs of serving various LLMs. You will do this by writing an MCP Server that hooks up to Firecrawl's API and saving the data in a SQLite Database. You should use the following websites to scrape:
"cloudrift": "https://www.cloudrift.ai/inference"
"deepinfra": "https://deepinfra.com/pricing"
"fireworks": "https://fireworks.ai/pricing#serverless-pricing"
"groq": "https://groq.com/pricing"
Make a venv with uv
Sync venv with pyproject.toml (
uv sync)Make an API Key on Anthropic and Firecrawl
Complete the 2 tool calls in
starter_server.pyChange the
server_config.jsonto point to your server fileComplete any section in
starter_client.pythat has "#complete".Test using any methods taught in the course
Use the following prompts in your chatbot but play around with all the LLM providers in the list above:
"How much does cloudrift ai (https://www.cloudrift.ai/inference) charge for deepseek v3?"
"How much does deepinfra (https://deepinfra.com/pricing) charge for deepseek v3"
"Compare cloudrift ai and deepinfra's costs for deepseek v3"