Perplexity MCP Server
by PoliTwit1984
- src
- mcp_server_openai
import logging
from openai import AsyncOpenAI
logger = logging.getLogger(__name__)
class LLMConnector:
def __init__(self, openai_api_key: str):
self.client = AsyncOpenAI(api_key=openai_api_key)
async def ask_openai(self, query: str, model: str = "gpt-4", temperature: float = 0.7, max_tokens: int = 500) -> str:
try:
response = await self.client.chat.completions.create(
messages=[
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": query}
],
model=model,
temperature=temperature,
max_tokens=max_tokens
)
return response.choices[0].message.content
except Exception as e:
logger.error(f"Failed to query OpenAI: {str(e)}")
raise