Spaces:
Configuration error
Configuration error
""" | |
Translate from OpenAI's `/v1/chat/completions` to Perplexity's `/v1/chat/completions` | |
""" | |
from typing import Optional, Tuple | |
import litellm | |
from litellm._logging import verbose_logger | |
from litellm.secret_managers.main import get_secret_str | |
from ...openai.chat.gpt_transformation import OpenAIGPTConfig | |
class PerplexityChatConfig(OpenAIGPTConfig): | |
def custom_llm_provider(self) -> Optional[str]: | |
return "perplexity" | |
def _get_openai_compatible_provider_info( | |
self, api_base: Optional[str], api_key: Optional[str] | |
) -> Tuple[Optional[str], Optional[str]]: | |
api_base = api_base or get_secret_str("PERPLEXITY_API_BASE") or "https://api.perplexity.ai" # type: ignore | |
dynamic_api_key = ( | |
api_key | |
or get_secret_str("PERPLEXITYAI_API_KEY") | |
or get_secret_str("PERPLEXITY_API_KEY") | |
) | |
return api_base, dynamic_api_key | |
def get_supported_openai_params(self, model: str) -> list: | |
""" | |
Perplexity supports a subset of OpenAI params | |
Ref: https://docs.perplexity.ai/api-reference/chat-completions | |
Eg. Perplexity does not support tools, tool_choice, function_call, functions, etc. | |
""" | |
base_openai_params = [ | |
"frequency_penalty", | |
"max_tokens", | |
"max_completion_tokens", | |
"presence_penalty", | |
"response_format", | |
"stream", | |
"temperature", | |
"top_p", | |
"max_retries", | |
"extra_headers", | |
] | |
try: | |
if litellm.supports_reasoning( | |
model=model, custom_llm_provider=self.custom_llm_provider | |
): | |
base_openai_params.append("reasoning_effort") | |
except Exception as e: | |
verbose_logger.debug(f"Error checking if model supports reasoning: {e}") | |
return base_openai_params | |