"""
Registry data for lazy imports.

This module contains all the name tuples and import maps used by the lazy import system.
Separated from the handler functions for better organization.
"""

# Cost calculator names that support lazy loading via _lazy_import_cost_calculator
COST_CALCULATOR_NAMES = (
    "completion_cost",
    "cost_per_token",
    "response_cost_calculator",
)

# Litellm logging names that support lazy loading via _lazy_import_litellm_logging
LITELLM_LOGGING_NAMES = (
    "Logging",
    "modify_integration",
)

# Utils names that support lazy loading via _lazy_import_utils
UTILS_NAMES = (
    "exception_type",
    "get_optional_params",
    "get_response_string",
    "token_counter",
    "create_pretrained_tokenizer",
    "create_tokenizer",
    "supports_function_calling",
    "supports_web_search",
    "supports_url_context",
    "supports_response_schema",
    "supports_parallel_function_calling",
    "supports_vision",
    "supports_audio_input",
    "supports_audio_output",
    "supports_system_messages",
    "supports_reasoning",
    "get_litellm_params",
    "acreate",
    "get_max_tokens",
    "get_model_info",
    "register_prompt_template",
    "validate_environment",
    "check_valid_key",
    "register_model",
    "encode",
    "decode",
    "_calculate_retry_after",
    "_should_retry",
    "get_supported_openai_params",
    "get_api_base",
    "get_first_chars_messages",
    "ModelResponse",
    "ModelResponseStream",
    "EmbeddingResponse",
    "ImageResponse",
    "TranscriptionResponse",
    "TextCompletionResponse",
    "get_provider_fields",
    "ModelResponseListIterator",
    "get_valid_models",
    "timeout",
    "get_llm_provider",
    "remove_index_from_tool_calls",
)

# Token counter names that support lazy loading via _lazy_import_token_counter
TOKEN_COUNTER_NAMES = ("get_modified_max_tokens",)

# LLM client cache names that support lazy loading via _lazy_import_llm_client_cache
LLM_CLIENT_CACHE_NAMES = (
    "LLMClientCache",
    "in_memory_llm_clients_cache",
)

# Bedrock type names that support lazy loading via _lazy_import_bedrock_types
BEDROCK_TYPES_NAMES = ("COHERE_EMBEDDING_INPUT_TYPES",)

# Common types from litellm.types.utils that support lazy loading via
# _lazy_import_types_utils
TYPES_UTILS_NAMES = (
    "ImageObject",
    "BudgetConfig",
    "all_litellm_params",
    "_litellm_completion_params",
    "CredentialItem",
    "PriorityReservationDict",
    "StandardKeyGenerationConfig",
    "SearchProviders",
    "GenericStreamingChunk",
)

# Caching / cache classes that support lazy loading via _lazy_import_caching
CACHING_NAMES = (
    "Cache",
    "DualCache",
    "RedisCache",
    "InMemoryCache",
)

# HTTP handler names that support lazy loading via _lazy_import_http_handlers
HTTP_HANDLER_NAMES = (
    "module_level_aclient",
    "module_level_client",
)

# Dotprompt integration names that support lazy loading via _lazy_import_dotprompt
DOTPROMPT_NAMES = (
    "global_prompt_manager",
    "global_prompt_directory",
    "set_global_prompt_directory",
)

# LLM config classes that support lazy loading via _lazy_import_llm_configs
LLM_CONFIG_NAMES = (
    "AmazonConverseConfig",
    "OpenAILikeChatConfig",
    "GaladrielChatConfig",
    "GithubChatConfig",
    "AzureAnthropicConfig",
    "BytezChatConfig",
    "CompactifAIChatConfig",
    "EmpowerChatConfig",
    "MinimaxChatConfig",
    "AiohttpOpenAIChatConfig",
    "HuggingFaceChatConfig",
    "HuggingFaceEmbeddingConfig",
    "OobaboogaConfig",
    "MaritalkConfig",
    "OpenrouterConfig",
    "DataRobotConfig",
    "AnthropicConfig",
    "AnthropicTextConfig",
    "GroqSTTConfig",
    "TritonConfig",
    "TritonGenerateConfig",
    "TritonInferConfig",
    "TritonEmbeddingConfig",
    "HuggingFaceRerankConfig",
    "DatabricksConfig",
    "DatabricksEmbeddingConfig",
    "PredibaseConfig",
    "ReplicateConfig",
    "SnowflakeConfig",
    "CohereRerankConfig",
    "CohereRerankV2Config",
    "AzureAIRerankConfig",
    "InfinityRerankConfig",
    "JinaAIRerankConfig",
    "DeepinfraRerankConfig",
    "HostedVLLMRerankConfig",
    "NvidiaNimRerankConfig",
    "NvidiaNimRankingConfig",
    "VertexAIRerankConfig",
    "FireworksAIRerankConfig",
    "VoyageRerankConfig",
    "IBMWatsonXRerankConfig",
    "ClarifaiConfig",
    "AI21ChatConfig",
    "LlamaAPIConfig",
    "TogetherAITextCompletionConfig",
    "CloudflareChatConfig",
    "NovitaConfig",
    "PetalsConfig",
    "OllamaChatConfig",
    "OllamaConfig",
    "SagemakerConfig",
    "SagemakerChatConfig",
    "CohereChatConfig",
    "AnthropicMessagesConfig",
    "AmazonAnthropicClaudeMessagesConfig",
    "TogetherAIConfig",
    "NLPCloudConfig",
    "VertexGeminiConfig",
    "GoogleAIStudioGeminiConfig",
    "VertexAIAnthropicConfig",
    "VertexAILlama3Config",
    "VertexAIAi21Config",
    "AmazonCohereChatConfig",
    "AmazonBedrockGlobalConfig",
    "AmazonAI21Config",
    "AmazonInvokeNovaConfig",
    "AmazonQwen2Config",
    "AmazonQwen3Config",
    # Aliases for backwards compatibility
    "VertexAIConfig",  # Alias for VertexGeminiConfig
    "GeminiConfig",  # Alias for GoogleAIStudioGeminiConfig
    "AmazonAnthropicConfig",
    "AmazonAnthropicClaudeConfig",
    "AmazonCohereConfig",
    "AmazonLlamaConfig",
    "AmazonDeepSeekR1Config",
    "AmazonMistralConfig",
    "AmazonMoonshotConfig",
    "AmazonTitanConfig",
    "AmazonTwelveLabsPegasusConfig",
    "AmazonInvokeConfig",
    "AmazonBedrockOpenAIConfig",
    "AmazonStabilityConfig",
    "AmazonStability3Config",
    "AmazonNovaCanvasConfig",
    "AmazonTitanG1Config",
    "AmazonTitanMultimodalEmbeddingG1Config",
    "CohereV2ChatConfig",
    "BedrockCohereEmbeddingConfig",
    "TwelveLabsMarengoEmbeddingConfig",
    "AmazonNovaEmbeddingConfig",
    "OpenAIConfig",
    "MistralEmbeddingConfig",
    "OpenAIImageVariationConfig",
    "DeepInfraConfig",
    "DeepgramAudioTranscriptionConfig",
    "TopazImageVariationConfig",
    "OpenAITextCompletionConfig",
    "GroqChatConfig",
    "BedrockMantleChatConfig",
    "A2AConfig",
    "GenAIHubOrchestrationConfig",
    "VoyageEmbeddingConfig",
    "VoyageContextualEmbeddingConfig",
    "InfinityEmbeddingConfig",
    "PerplexityEmbeddingConfig",
    "AzureAIStudioConfig",
    "MistralConfig",
    "OpenAIResponsesAPIConfig",
    "AzureOpenAIResponsesAPIConfig",
    "AzureOpenAIOSeriesResponsesAPIConfig",
    "XAIResponsesAPIConfig",
    "LiteLLMProxyResponsesAPIConfig",
    "HostedVLLMResponsesAPIConfig",
    "VolcEngineResponsesAPIConfig",
    "PerplexityResponsesConfig",
    "DatabricksResponsesAPIConfig",
    "OpenRouterResponsesAPIConfig",
    "GoogleAIStudioInteractionsConfig",
    "OpenAIOSeriesConfig",
    "AnthropicSkillsConfig",
    "BaseSkillsAPIConfig",
    "GradientAIConfig",
    # Alias for backwards compatibility
    "OpenAIO1Config",  # Alias for OpenAIOSeriesConfig
    "OpenAIGPTConfig",
    "OpenAIGPT5Config",
    "OpenAIWhisperAudioTranscriptionConfig",
    "OpenAIGPTAudioTranscriptionConfig",
    "OpenAIGPTAudioConfig",
    "NvidiaNimConfig",
    "NvidiaNimEmbeddingConfig",
    "FeatherlessAIConfig",
    "CerebrasConfig",
    "BasetenConfig",
    "SambanovaConfig",
    "SambaNovaEmbeddingConfig",
    "FireworksAIConfig",
    "FireworksAITextCompletionConfig",
    "FireworksAIAudioTranscriptionConfig",
    "FireworksAIEmbeddingConfig",
    "FriendliaiChatConfig",
    "JinaAIEmbeddingConfig",
    "XAIChatConfig",
    "ZAIChatConfig",
    "AIMLChatConfig",
    "VolcEngineChatConfig",
    "CodestralTextCompletionConfig",
    "AzureOpenAIAssistantsAPIConfig",
    "HerokuChatConfig",
    "CometAPIConfig",
    "AzureOpenAIConfig",
    "AzureOpenAIGPT5Config",
    "AzureOpenAITextConfig",
    "HostedVLLMChatConfig",
    "HostedVLLMEmbeddingConfig",
    # Alias for backwards compatibility
    "VolcEngineConfig",  # Alias for VolcEngineChatConfig
    "LlamafileChatConfig",
    "LiteLLMProxyChatConfig",
    "VLLMConfig",
    "DeepSeekChatConfig",
    "LMStudioChatConfig",
    "LmStudioEmbeddingConfig",
    "NscaleConfig",
    "PerplexityChatConfig",
    "AzureOpenAIO1Config",
    "IBMWatsonXAIConfig",
    "IBMWatsonXChatConfig",
    "IBMWatsonXEmbeddingConfig",
    "GenAIHubEmbeddingConfig",
    "IBMWatsonXAudioTranscriptionConfig",
    "GithubCopilotConfig",
    "GithubCopilotResponsesAPIConfig",
    "ChatGPTConfig",
    "ChatGPTResponsesAPIConfig",
    "ManusResponsesAPIConfig",
    "GithubCopilotEmbeddingConfig",
    "NebiusConfig",
    "WandbConfig",
    "GigaChatConfig",
    "GigaChatEmbeddingConfig",
    "DashScopeChatConfig",
    "MoonshotChatConfig",
    "DockerModelRunnerChatConfig",
    "V0ChatConfig",
    "OCIChatConfig",
    "MorphChatConfig",
    "RAGFlowConfig",
    "LambdaAIChatConfig",
    "HyperbolicChatConfig",
    "VercelAIGatewayConfig",
    "OVHCloudChatConfig",
    "OVHCloudEmbeddingConfig",
    "CometAPIEmbeddingConfig",
    "LemonadeChatConfig",
    "SnowflakeEmbeddingConfig",
    "AmazonNovaChatConfig",
)

# Types that support lazy loading via _lazy_import_types
TYPES_NAMES = (
    "GuardrailItem",
    "DefaultTeamSSOParams",
    "LiteLLM_UpperboundKeyGenerateParams",
    "KeyManagementSystem",
    "PriorityReservationSettings",
    "CustomLogger",
    "LoggingCallbackManager",
    "DatadogLLMObsInitParams",
    # Note: LlmProviders is NOT lazy-loaded because it's imported during import time
    # in multiple places including openai.py (via main import)
    # Note: KeyManagementSettings is NOT lazy-loaded because _key_management_settings
    # is accessed during import time in secret_managers/main.py
)

# LLM provider logic names that support lazy loading via _lazy_import_llm_provider_logic
LLM_PROVIDER_LOGIC_NAMES = (
    "get_llm_provider",
    "remove_index_from_tool_calls",
)

# Utils module names that support lazy loading via _lazy_import_utils_module
# These are attributes accessed from litellm.utils module
UTILS_MODULE_NAMES = (
    "encoding",
    "BaseVectorStore",
    "CredentialAccessor",
    "exception_type",
    "get_error_message",
    "_get_response_headers",
    "get_llm_provider",
    "_is_non_openai_azure_model",
    "get_supported_openai_params",
    "LiteLLMResponseObjectHandler",
    "_handle_invalid_parallel_tool_calls",
    "convert_to_model_response_object",
    "convert_to_streaming_response",
    "convert_to_streaming_response_async",
    "get_api_base",
    "ResponseMetadata",
    "_parse_content_for_reasoning",
    "LiteLLMLoggingObject",
    "redact_message_input_output_from_logging",
    "CustomStreamWrapper",
    "BaseGoogleGenAIGenerateContentConfig",
    "BaseOCRConfig",
    "BaseSearchConfig",
    "BaseTextToSpeechConfig",
    "BedrockModelInfo",
    "CohereModelInfo",
    "MistralOCRConfig",
    "Rules",
    "AsyncHTTPHandler",
    "HTTPHandler",
    "get_num_retries_from_retry_policy",
    "reset_retry_policy",
    "get_secret",
    "get_coroutine_checker",
    "get_litellm_logging_class",
    "get_set_callbacks",
    "get_litellm_metadata_from_kwargs",
    "map_finish_reason",
    "process_response_headers",
    "delete_nested_value",
    "is_nested_path",
    "_get_base_model_from_litellm_call_metadata",
    "get_litellm_params",
    "_ensure_extra_body_is_safe",
    "get_formatted_prompt",
    "get_response_headers",
    "update_response_metadata",
    "executor",
    "BaseAnthropicMessagesConfig",
    "BaseAudioTranscriptionConfig",
    "BaseBatchesConfig",
    "BaseContainerConfig",
    "BaseEmbeddingConfig",
    "BaseImageEditConfig",
    "BaseImageGenerationConfig",
    "BaseImageVariationConfig",
    "BasePassthroughConfig",
    "BaseRealtimeConfig",
    "BaseRerankConfig",
    "BaseVectorStoreConfig",
    "BaseVectorStoreFilesConfig",
    "BaseVideoConfig",
    "ANTHROPIC_API_ONLY_HEADERS",
    "AnthropicThinkingParam",
    "RerankResponse",
    "ChatCompletionDeltaToolCallChunk",
    "ChatCompletionToolCallChunk",
    "ChatCompletionToolCallFunctionChunk",
    "LiteLLM_Params",
)

# Import maps for registry pattern - reduces repetition
_UTILS_IMPORT_MAP = {
    "exception_type": (".utils", "exception_type"),
    "get_optional_params": (".utils", "get_optional_params"),
    "get_response_string": (".utils", "get_response_string"),
    "token_counter": (".utils", "token_counter"),
    "create_pretrained_tokenizer": (".utils", "create_pretrained_tokenizer"),
    "create_tokenizer": (".utils", "create_tokenizer"),
    "supports_function_calling": (".utils", "supports_function_calling"),
    "supports_web_search": (".utils", "supports_web_search"),
    "supports_url_context": (".utils", "supports_url_context"),
    "supports_response_schema": (".utils", "supports_response_schema"),
    "supports_parallel_function_calling": (
        ".utils",
        "supports_parallel_function_calling",
    ),
    "supports_vision": (".utils", "supports_vision"),
    "supports_audio_input": (".utils", "supports_audio_input"),
    "supports_audio_output": (".utils", "supports_audio_output"),
    "supports_system_messages": (".utils", "supports_system_messages"),
    "supports_reasoning": (".utils", "supports_reasoning"),
    "get_litellm_params": (".utils", "get_litellm_params"),
    "acreate": (".utils", "acreate"),
    "get_max_tokens": (".utils", "get_max_tokens"),
    "get_model_info": (".utils", "get_model_info"),
    "register_prompt_template": (".utils", "register_prompt_template"),
    "validate_environment": (".utils", "validate_environment"),
    "check_valid_key": (".utils", "check_valid_key"),
    "register_model": (".utils", "register_model"),
    "encode": (".utils", "encode"),
    "decode": (".utils", "decode"),
    "_calculate_retry_after": (".utils", "_calculate_retry_after"),
    "_should_retry": (".utils", "_should_retry"),
    "get_supported_openai_params": (".utils", "get_supported_openai_params"),
    "get_api_base": (".utils", "get_api_base"),
    "get_first_chars_messages": (".utils", "get_first_chars_messages"),
    "ModelResponse": (".utils", "ModelResponse"),
    "ModelResponseStream": (".utils", "ModelResponseStream"),
    "EmbeddingResponse": (".utils", "EmbeddingResponse"),
    "ImageResponse": (".utils", "ImageResponse"),
    "TranscriptionResponse": (".utils", "TranscriptionResponse"),
    "TextCompletionResponse": (".utils", "TextCompletionResponse"),
    "get_provider_fields": (".utils", "get_provider_fields"),
    "ModelResponseListIterator": (".utils", "ModelResponseListIterator"),
    "get_valid_models": (".utils", "get_valid_models"),
    "timeout": (".timeout", "timeout"),
    "get_llm_provider": (
        "litellm.litellm_core_utils.get_llm_provider_logic",
        "get_llm_provider",
    ),
    "remove_index_from_tool_calls": (
        "litellm.litellm_core_utils.core_helpers",
        "remove_index_from_tool_calls",
    ),
}

_COST_CALCULATOR_IMPORT_MAP = {
    "completion_cost": (".cost_calculator", "completion_cost"),
    "cost_per_token": (".cost_calculator", "cost_per_token"),
    "response_cost_calculator": (".cost_calculator", "response_cost_calculator"),
}

_TYPES_UTILS_IMPORT_MAP = {
    "ImageObject": (".types.utils", "ImageObject"),
    "BudgetConfig": (".types.utils", "BudgetConfig"),
    "all_litellm_params": (".types.utils", "all_litellm_params"),
    "_litellm_completion_params": (".types.utils", "all_litellm_params"),  # Alias
    "CredentialItem": (".types.utils", "CredentialItem"),
    "PriorityReservationDict": (".types.utils", "PriorityReservationDict"),
    "StandardKeyGenerationConfig": (".types.utils", "StandardKeyGenerationConfig"),
    "SearchProviders": (".types.utils", "SearchProviders"),
    "GenericStreamingChunk": (".types.utils", "GenericStreamingChunk"),
}

_TOKEN_COUNTER_IMPORT_MAP = {
    "get_modified_max_tokens": (
        "litellm.litellm_core_utils.token_counter",
        "get_modified_max_tokens",
    ),
}

_BEDROCK_TYPES_IMPORT_MAP = {
    "COHERE_EMBEDDING_INPUT_TYPES": (
        "litellm.types.llms.bedrock",
        "COHERE_EMBEDDING_INPUT_TYPES",
    ),
}

_CACHING_IMPORT_MAP = {
    "Cache": ("litellm.caching.caching", "Cache"),
    "DualCache": ("litellm.caching.caching", "DualCache"),
    "RedisCache": ("litellm.caching.caching", "RedisCache"),
    "InMemoryCache": ("litellm.caching.caching", "InMemoryCache"),
}

_LITELLM_LOGGING_IMPORT_MAP = {
    "Logging": ("litellm.litellm_core_utils.litellm_logging", "Logging"),
    "modify_integration": (
        "litellm.litellm_core_utils.litellm_logging",
        "modify_integration",
    ),
}

_DOTPROMPT_IMPORT_MAP = {
    "global_prompt_manager": (
        "litellm.integrations.dotprompt",
        "global_prompt_manager",
    ),
    "global_prompt_directory": (
        "litellm.integrations.dotprompt",
        "global_prompt_directory",
    ),
    "set_global_prompt_directory": (
        "litellm.integrations.dotprompt",
        "set_global_prompt_directory",
    ),
}

_TYPES_IMPORT_MAP = {
    "GuardrailItem": ("litellm.types.guardrails", "GuardrailItem"),
    "DefaultTeamSSOParams": (
        "litellm.types.proxy.management_endpoints.ui_sso",
        "DefaultTeamSSOParams",
    ),
    "LiteLLM_UpperboundKeyGenerateParams": (
        "litellm.types.proxy.management_endpoints.ui_sso",
        "LiteLLM_UpperboundKeyGenerateParams",
    ),
    "KeyManagementSystem": (
        "litellm.types.secret_managers.main",
        "KeyManagementSystem",
    ),
    "PriorityReservationSettings": (
        "litellm.types.utils",
        "PriorityReservationSettings",
    ),
    "CustomLogger": ("litellm.integrations.custom_logger", "CustomLogger"),
    "LoggingCallbackManager": (
        "litellm.litellm_core_utils.logging_callback_manager",
        "LoggingCallbackManager",
    ),
    "DatadogLLMObsInitParams": (
        "litellm.types.integrations.datadog_llm_obs",
        "DatadogLLMObsInitParams",
    ),
}

_LLM_PROVIDER_LOGIC_IMPORT_MAP = {
    "get_llm_provider": (
        "litellm.litellm_core_utils.get_llm_provider_logic",
        "get_llm_provider",
    ),
    "remove_index_from_tool_calls": (
        "litellm.litellm_core_utils.core_helpers",
        "remove_index_from_tool_calls",
    ),
}

_LLM_CONFIGS_IMPORT_MAP = {
    "AmazonConverseConfig": (
        ".llms.bedrock.chat.converse_transformation",
        "AmazonConverseConfig",
    ),
    "OpenAILikeChatConfig": (".llms.openai_like.chat.handler", "OpenAILikeChatConfig"),
    "GaladrielChatConfig": (
        ".llms.galadriel.chat.transformation",
        "GaladrielChatConfig",
    ),
    "GithubChatConfig": (".llms.github.chat.transformation", "GithubChatConfig"),
    "AzureAnthropicConfig": (
        ".llms.azure_ai.anthropic.transformation",
        "AzureAnthropicConfig",
    ),
    "BytezChatConfig": (".llms.bytez.chat.transformation", "BytezChatConfig"),
    "CompactifAIChatConfig": (
        ".llms.compactifai.chat.transformation",
        "CompactifAIChatConfig",
    ),
    "EmpowerChatConfig": (".llms.empower.chat.transformation", "EmpowerChatConfig"),
    "MinimaxChatConfig": (".llms.minimax.chat.transformation", "MinimaxChatConfig"),
    "AiohttpOpenAIChatConfig": (
        ".llms.aiohttp_openai.chat.transformation",
        "AiohttpOpenAIChatConfig",
    ),
    "HuggingFaceChatConfig": (
        ".llms.huggingface.chat.transformation",
        "HuggingFaceChatConfig",
    ),
    "HuggingFaceEmbeddingConfig": (
        ".llms.huggingface.embedding.transformation",
        "HuggingFaceEmbeddingConfig",
    ),
    "OobaboogaConfig": (".llms.oobabooga.chat.transformation", "OobaboogaConfig"),
    "MaritalkConfig": (".llms.maritalk", "MaritalkConfig"),
    "OpenrouterConfig": (".llms.openrouter.chat.transformation", "OpenrouterConfig"),
    "DataRobotConfig": (".llms.datarobot.chat.transformation", "DataRobotConfig"),
    "AnthropicConfig": (".llms.anthropic.chat.transformation", "AnthropicConfig"),
    "AnthropicTextConfig": (
        ".llms.anthropic.completion.transformation",
        "AnthropicTextConfig",
    ),
    "GroqSTTConfig": (".llms.groq.stt.transformation", "GroqSTTConfig"),
    "TritonConfig": (".llms.triton.completion.transformation", "TritonConfig"),
    "TritonGenerateConfig": (
        ".llms.triton.completion.transformation",
        "TritonGenerateConfig",
    ),
    "TritonInferConfig": (
        ".llms.triton.completion.transformation",
        "TritonInferConfig",
    ),
    "TritonEmbeddingConfig": (
        ".llms.triton.embedding.transformation",
        "TritonEmbeddingConfig",
    ),
    "HuggingFaceRerankConfig": (
        ".llms.huggingface.rerank.transformation",
        "HuggingFaceRerankConfig",
    ),
    "DatabricksConfig": (".llms.databricks.chat.transformation", "DatabricksConfig"),
    "DatabricksEmbeddingConfig": (
        ".llms.databricks.embed.transformation",
        "DatabricksEmbeddingConfig",
    ),
    "PredibaseConfig": (".llms.predibase.chat.transformation", "PredibaseConfig"),
    "ReplicateConfig": (".llms.replicate.chat.transformation", "ReplicateConfig"),
    "SnowflakeConfig": (".llms.snowflake.chat.transformation", "SnowflakeConfig"),
    "CohereRerankConfig": (".llms.cohere.rerank.transformation", "CohereRerankConfig"),
    "CohereRerankV2Config": (
        ".llms.cohere.rerank_v2.transformation",
        "CohereRerankV2Config",
    ),
    "AzureAIRerankConfig": (
        ".llms.azure_ai.rerank.transformation",
        "AzureAIRerankConfig",
    ),
    "InfinityRerankConfig": (
        ".llms.infinity.rerank.transformation",
        "InfinityRerankConfig",
    ),
    "JinaAIRerankConfig": (".llms.jina_ai.rerank.transformation", "JinaAIRerankConfig"),
    "DeepinfraRerankConfig": (
        ".llms.deepinfra.rerank.transformation",
        "DeepinfraRerankConfig",
    ),
    "HostedVLLMRerankConfig": (
        ".llms.hosted_vllm.rerank.transformation",
        "HostedVLLMRerankConfig",
    ),
    "NvidiaNimRerankConfig": (
        ".llms.nvidia_nim.rerank.transformation",
        "NvidiaNimRerankConfig",
    ),
    "NvidiaNimRankingConfig": (
        ".llms.nvidia_nim.rerank.ranking_transformation",
        "NvidiaNimRankingConfig",
    ),
    "VertexAIRerankConfig": (
        ".llms.vertex_ai.rerank.transformation",
        "VertexAIRerankConfig",
    ),
    "FireworksAIRerankConfig": (
        ".llms.fireworks_ai.rerank.transformation",
        "FireworksAIRerankConfig",
    ),
    "VoyageRerankConfig": (".llms.voyage.rerank.transformation", "VoyageRerankConfig"),
    "IBMWatsonXRerankConfig": (".llms.watsonx.rerank.transformation", "IBMWatsonXRerankConfig"),
    "ClarifaiConfig": (".llms.clarifai.chat.transformation", "ClarifaiConfig"),
    "AI21ChatConfig": (".llms.ai21.chat.transformation", "AI21ChatConfig"),
    "LlamaAPIConfig": (".llms.meta_llama.chat.transformation", "LlamaAPIConfig"),
    "TogetherAITextCompletionConfig": (
        ".llms.together_ai.completion.transformation",
        "TogetherAITextCompletionConfig",
    ),
    "CloudflareChatConfig": (
        ".llms.cloudflare.chat.transformation",
        "CloudflareChatConfig",
    ),
    "NovitaConfig": (".llms.novita.chat.transformation", "NovitaConfig"),
    "PetalsConfig": (".llms.petals.completion.transformation", "PetalsConfig"),
    "OllamaChatConfig": (".llms.ollama.chat.transformation", "OllamaChatConfig"),
    "OllamaConfig": (".llms.ollama.completion.transformation", "OllamaConfig"),
    "SagemakerConfig": (".llms.sagemaker.completion.transformation", "SagemakerConfig"),
    "SagemakerChatConfig": (
        ".llms.sagemaker.chat.transformation",
        "SagemakerChatConfig",
    ),
    "CohereChatConfig": (".llms.cohere.chat.transformation", "CohereChatConfig"),
    "AnthropicMessagesConfig": (
        ".llms.anthropic.experimental_pass_through.messages.transformation",
        "AnthropicMessagesConfig",
    ),
    "AmazonAnthropicClaudeMessagesConfig": (
        ".llms.bedrock.messages.invoke_transformations.anthropic_claude3_transformation",
        "AmazonAnthropicClaudeMessagesConfig",
    ),
    "TogetherAIConfig": (".llms.together_ai.chat", "TogetherAIConfig"),
    "NLPCloudConfig": (".llms.nlp_cloud.chat.handler", "NLPCloudConfig"),
    "VertexGeminiConfig": (
        ".llms.vertex_ai.gemini.vertex_and_google_ai_studio_gemini",
        "VertexGeminiConfig",
    ),
    "GoogleAIStudioGeminiConfig": (
        ".llms.gemini.chat.transformation",
        "GoogleAIStudioGeminiConfig",
    ),
    "VertexAIAnthropicConfig": (
        ".llms.vertex_ai.vertex_ai_partner_models.anthropic.transformation",
        "VertexAIAnthropicConfig",
    ),
    "VertexAILlama3Config": (
        ".llms.vertex_ai.vertex_ai_partner_models.llama3.transformation",
        "VertexAILlama3Config",
    ),
    "VertexAIAi21Config": (
        ".llms.vertex_ai.vertex_ai_partner_models.ai21.transformation",
        "VertexAIAi21Config",
    ),
    "AmazonCohereChatConfig": (
        ".llms.bedrock.chat.invoke_handler",
        "AmazonCohereChatConfig",
    ),
    "AmazonBedrockGlobalConfig": (
        ".llms.bedrock.common_utils",
        "AmazonBedrockGlobalConfig",
    ),
    "AmazonAI21Config": (
        ".llms.bedrock.chat.invoke_transformations.amazon_ai21_transformation",
        "AmazonAI21Config",
    ),
    "AmazonInvokeNovaConfig": (
        ".llms.bedrock.chat.invoke_transformations.amazon_nova_transformation",
        "AmazonInvokeNovaConfig",
    ),
    "AmazonQwen2Config": (
        ".llms.bedrock.chat.invoke_transformations.amazon_qwen2_transformation",
        "AmazonQwen2Config",
    ),
    "AmazonQwen3Config": (
        ".llms.bedrock.chat.invoke_transformations.amazon_qwen3_transformation",
        "AmazonQwen3Config",
    ),
    # Aliases for backwards compatibility
    "VertexAIConfig": (
        ".llms.vertex_ai.gemini.vertex_and_google_ai_studio_gemini",
        "VertexGeminiConfig",
    ),  # Alias
    "GeminiConfig": (
        ".llms.gemini.chat.transformation",
        "GoogleAIStudioGeminiConfig",
    ),  # Alias
    "AmazonAnthropicConfig": (
        ".llms.bedrock.chat.invoke_transformations.anthropic_claude2_transformation",
        "AmazonAnthropicConfig",
    ),
    "AmazonAnthropicClaudeConfig": (
        ".llms.bedrock.chat.invoke_transformations.anthropic_claude3_transformation",
        "AmazonAnthropicClaudeConfig",
    ),
    "AmazonCohereConfig": (
        ".llms.bedrock.chat.invoke_transformations.amazon_cohere_transformation",
        "AmazonCohereConfig",
    ),
    "AmazonLlamaConfig": (
        ".llms.bedrock.chat.invoke_transformations.amazon_llama_transformation",
        "AmazonLlamaConfig",
    ),
    "AmazonDeepSeekR1Config": (
        ".llms.bedrock.chat.invoke_transformations.amazon_deepseek_transformation",
        "AmazonDeepSeekR1Config",
    ),
    "AmazonMistralConfig": (
        ".llms.bedrock.chat.invoke_transformations.amazon_mistral_transformation",
        "AmazonMistralConfig",
    ),
    "AmazonMoonshotConfig": (
        ".llms.bedrock.chat.invoke_transformations.amazon_moonshot_transformation",
        "AmazonMoonshotConfig",
    ),
    "AmazonTitanConfig": (
        ".llms.bedrock.chat.invoke_transformations.amazon_titan_transformation",
        "AmazonTitanConfig",
    ),
    "AmazonTwelveLabsPegasusConfig": (
        ".llms.bedrock.chat.invoke_transformations.amazon_twelvelabs_pegasus_transformation",
        "AmazonTwelveLabsPegasusConfig",
    ),
    "AmazonInvokeConfig": (
        ".llms.bedrock.chat.invoke_transformations.base_invoke_transformation",
        "AmazonInvokeConfig",
    ),
    "AmazonBedrockOpenAIConfig": (
        ".llms.bedrock.chat.invoke_transformations.amazon_openai_transformation",
        "AmazonBedrockOpenAIConfig",
    ),
    "AmazonStabilityConfig": (
        ".llms.bedrock.image_generation.amazon_stability1_transformation",
        "AmazonStabilityConfig",
    ),
    "AmazonStability3Config": (
        ".llms.bedrock.image_generation.amazon_stability3_transformation",
        "AmazonStability3Config",
    ),
    "AmazonNovaCanvasConfig": (
        ".llms.bedrock.image_generation.amazon_nova_canvas_transformation",
        "AmazonNovaCanvasConfig",
    ),
    "AmazonTitanG1Config": (
        ".llms.bedrock.embed.amazon_titan_g1_transformation",
        "AmazonTitanG1Config",
    ),
    "AmazonTitanMultimodalEmbeddingG1Config": (
        ".llms.bedrock.embed.amazon_titan_multimodal_transformation",
        "AmazonTitanMultimodalEmbeddingG1Config",
    ),
    "CohereV2ChatConfig": (".llms.cohere.chat.v2_transformation", "CohereV2ChatConfig"),
    "BedrockCohereEmbeddingConfig": (
        ".llms.bedrock.embed.cohere_transformation",
        "BedrockCohereEmbeddingConfig",
    ),
    "TwelveLabsMarengoEmbeddingConfig": (
        ".llms.bedrock.embed.twelvelabs_marengo_transformation",
        "TwelveLabsMarengoEmbeddingConfig",
    ),
    "AmazonNovaEmbeddingConfig": (
        ".llms.bedrock.embed.amazon_nova_transformation",
        "AmazonNovaEmbeddingConfig",
    ),
    "OpenAIConfig": (".llms.openai.openai", "OpenAIConfig"),
    "MistralEmbeddingConfig": (".llms.openai.openai", "MistralEmbeddingConfig"),
    "OpenAIImageVariationConfig": (
        ".llms.openai.image_variations.transformation",
        "OpenAIImageVariationConfig",
    ),
    "DeepInfraConfig": (".llms.deepinfra.chat.transformation", "DeepInfraConfig"),
    "DeepgramAudioTranscriptionConfig": (
        ".llms.deepgram.audio_transcription.transformation",
        "DeepgramAudioTranscriptionConfig",
    ),
    "TopazImageVariationConfig": (
        ".llms.topaz.image_variations.transformation",
        "TopazImageVariationConfig",
    ),
    "OpenAITextCompletionConfig": (
        "litellm.llms.openai.completion.transformation",
        "OpenAITextCompletionConfig",
    ),
    "GroqChatConfig": (".llms.groq.chat.transformation", "GroqChatConfig"),
    "BedrockMantleChatConfig": (".llms.bedrock_mantle.chat.transformation", "BedrockMantleChatConfig"),
    "A2AConfig": (".llms.a2a.chat.transformation", "A2AConfig"),
    "GenAIHubOrchestrationConfig": (
        ".llms.sap.chat.transformation",
        "GenAIHubOrchestrationConfig",
    ),
    "VoyageEmbeddingConfig": (
        ".llms.voyage.embedding.transformation",
        "VoyageEmbeddingConfig",
    ),
    "VoyageContextualEmbeddingConfig": (
        ".llms.voyage.embedding.transformation_contextual",
        "VoyageContextualEmbeddingConfig",
    ),
    "InfinityEmbeddingConfig": (
        ".llms.infinity.embedding.transformation",
        "InfinityEmbeddingConfig",
    ),
    "PerplexityEmbeddingConfig": (
        ".llms.perplexity.embedding.transformation",
        "PerplexityEmbeddingConfig",
    ),
    "AzureAIStudioConfig": (
        ".llms.azure_ai.chat.transformation",
        "AzureAIStudioConfig",
    ),
    "MistralConfig": (".llms.mistral.chat.transformation", "MistralConfig"),
    "OpenAIResponsesAPIConfig": (
        ".llms.openai.responses.transformation",
        "OpenAIResponsesAPIConfig",
    ),
    "AzureOpenAIResponsesAPIConfig": (
        ".llms.azure.responses.transformation",
        "AzureOpenAIResponsesAPIConfig",
    ),
    "AzureOpenAIOSeriesResponsesAPIConfig": (
        ".llms.azure.responses.o_series_transformation",
        "AzureOpenAIOSeriesResponsesAPIConfig",
    ),
    "XAIResponsesAPIConfig": (
        ".llms.xai.responses.transformation",
        "XAIResponsesAPIConfig",
    ),
    "LiteLLMProxyResponsesAPIConfig": (
        ".llms.litellm_proxy.responses.transformation",
        "LiteLLMProxyResponsesAPIConfig",
    ),
    "HostedVLLMResponsesAPIConfig": (
        ".llms.hosted_vllm.responses.transformation",
        "HostedVLLMResponsesAPIConfig",
    ),
    "VolcEngineResponsesAPIConfig": (
        ".llms.volcengine.responses.transformation",
        "VolcEngineResponsesAPIConfig",
    ),
    "ManusResponsesAPIConfig": (
        ".llms.manus.responses.transformation",
        "ManusResponsesAPIConfig",
    ),
    "PerplexityResponsesConfig": (
        ".llms.perplexity.responses.transformation",
        "PerplexityResponsesConfig",
    ),
    "DatabricksResponsesAPIConfig": (
        ".llms.databricks.responses.transformation",
        "DatabricksResponsesAPIConfig",
    ),
    "OpenRouterResponsesAPIConfig": (
        ".llms.openrouter.responses.transformation",
        "OpenRouterResponsesAPIConfig",
    ),
    "GoogleAIStudioInteractionsConfig": (
        ".llms.gemini.interactions.transformation",
        "GoogleAIStudioInteractionsConfig",
    ),
    "OpenAIOSeriesConfig": (
        ".llms.openai.chat.o_series_transformation",
        "OpenAIOSeriesConfig",
    ),
    "AnthropicSkillsConfig": (
        ".llms.anthropic.skills.transformation",
        "AnthropicSkillsConfig",
    ),
    "BaseSkillsAPIConfig": (
        ".llms.base_llm.skills.transformation",
        "BaseSkillsAPIConfig",
    ),
    "GradientAIConfig": (".llms.gradient_ai.chat.transformation", "GradientAIConfig"),
    # Alias for backwards compatibility
    "OpenAIO1Config": (
        ".llms.openai.chat.o_series_transformation",
        "OpenAIOSeriesConfig",
    ),  # Alias
    "OpenAIGPTConfig": (".llms.openai.chat.gpt_transformation", "OpenAIGPTConfig"),
    "OpenAIGPT5Config": (".llms.openai.chat.gpt_5_transformation", "OpenAIGPT5Config"),
    "OpenAIWhisperAudioTranscriptionConfig": (
        ".llms.openai.transcriptions.whisper_transformation",
        "OpenAIWhisperAudioTranscriptionConfig",
    ),
    "OpenAIGPTAudioTranscriptionConfig": (
        ".llms.openai.transcriptions.gpt_transformation",
        "OpenAIGPTAudioTranscriptionConfig",
    ),
    "OpenAIGPTAudioConfig": (
        ".llms.openai.chat.gpt_audio_transformation",
        "OpenAIGPTAudioConfig",
    ),
    "NvidiaNimConfig": (".llms.nvidia_nim.chat.transformation", "NvidiaNimConfig"),
    "NvidiaNimEmbeddingConfig": (".llms.nvidia_nim.embed", "NvidiaNimEmbeddingConfig"),
    "FeatherlessAIConfig": (
        ".llms.featherless_ai.chat.transformation",
        "FeatherlessAIConfig",
    ),
    "CerebrasConfig": (".llms.cerebras.chat", "CerebrasConfig"),
    "BasetenConfig": (".llms.baseten.chat", "BasetenConfig"),
    "SambanovaConfig": (".llms.sambanova.chat", "SambanovaConfig"),
    "SambaNovaEmbeddingConfig": (
        ".llms.sambanova.embedding.transformation",
        "SambaNovaEmbeddingConfig",
    ),
    "FireworksAIConfig": (
        ".llms.fireworks_ai.chat.transformation",
        "FireworksAIConfig",
    ),
    "FireworksAITextCompletionConfig": (
        ".llms.fireworks_ai.completion.transformation",
        "FireworksAITextCompletionConfig",
    ),
    "FireworksAIAudioTranscriptionConfig": (
        ".llms.fireworks_ai.audio_transcription.transformation",
        "FireworksAIAudioTranscriptionConfig",
    ),
    "FireworksAIEmbeddingConfig": (
        ".llms.fireworks_ai.embed.fireworks_ai_transformation",
        "FireworksAIEmbeddingConfig",
    ),
    "FriendliaiChatConfig": (
        ".llms.friendliai.chat.transformation",
        "FriendliaiChatConfig",
    ),
    "JinaAIEmbeddingConfig": (
        ".llms.jina_ai.embedding.transformation",
        "JinaAIEmbeddingConfig",
    ),
    "XAIChatConfig": (".llms.xai.chat.transformation", "XAIChatConfig"),
    "ZAIChatConfig": (".llms.zai.chat.transformation", "ZAIChatConfig"),
    "AIMLChatConfig": (".llms.aiml.chat.transformation", "AIMLChatConfig"),
    "VolcEngineChatConfig": (
        ".llms.volcengine.chat.transformation",
        "VolcEngineChatConfig",
    ),
    "CodestralTextCompletionConfig": (
        ".llms.codestral.completion.transformation",
        "CodestralTextCompletionConfig",
    ),
    "AzureOpenAIAssistantsAPIConfig": (
        ".llms.azure.azure",
        "AzureOpenAIAssistantsAPIConfig",
    ),
    "HerokuChatConfig": (".llms.heroku.chat.transformation", "HerokuChatConfig"),
    "CometAPIConfig": (".llms.cometapi.chat.transformation", "CometAPIConfig"),
    "AzureOpenAIConfig": (".llms.azure.chat.gpt_transformation", "AzureOpenAIConfig"),
    "AzureOpenAIGPT5Config": (
        ".llms.azure.chat.gpt_5_transformation",
        "AzureOpenAIGPT5Config",
    ),
    "AzureOpenAITextConfig": (
        ".llms.azure.completion.transformation",
        "AzureOpenAITextConfig",
    ),
    "HostedVLLMChatConfig": (
        ".llms.hosted_vllm.chat.transformation",
        "HostedVLLMChatConfig",
    ),
    "HostedVLLMEmbeddingConfig": (
        ".llms.hosted_vllm.embedding.transformation",
        "HostedVLLMEmbeddingConfig",
    ),
    # Alias for backwards compatibility
    "VolcEngineConfig": (
        ".llms.volcengine.chat.transformation",
        "VolcEngineChatConfig",
    ),  # Alias
    "LlamafileChatConfig": (
        ".llms.llamafile.chat.transformation",
        "LlamafileChatConfig",
    ),
    "LiteLLMProxyChatConfig": (
        ".llms.litellm_proxy.chat.transformation",
        "LiteLLMProxyChatConfig",
    ),
    "VLLMConfig": (".llms.vllm.completion.transformation", "VLLMConfig"),
    "DeepSeekChatConfig": (".llms.deepseek.chat.transformation", "DeepSeekChatConfig"),
    "LMStudioChatConfig": (".llms.lm_studio.chat.transformation", "LMStudioChatConfig"),
    "LmStudioEmbeddingConfig": (
        ".llms.lm_studio.embed.transformation",
        "LmStudioEmbeddingConfig",
    ),
    "NscaleConfig": (".llms.nscale.chat.transformation", "NscaleConfig"),
    "PerplexityChatConfig": (
        ".llms.perplexity.chat.transformation",
        "PerplexityChatConfig",
    ),
    "AzureOpenAIO1Config": (
        ".llms.azure.chat.o_series_transformation",
        "AzureOpenAIO1Config",
    ),
    "IBMWatsonXAIConfig": (
        ".llms.watsonx.completion.transformation",
        "IBMWatsonXAIConfig",
    ),
    "IBMWatsonXChatConfig": (
        ".llms.watsonx.chat.transformation",
        "IBMWatsonXChatConfig",
    ),
    "IBMWatsonXEmbeddingConfig": (
        ".llms.watsonx.embed.transformation",
        "IBMWatsonXEmbeddingConfig",
    ),
    "GenAIHubEmbeddingConfig": (
        ".llms.sap.embed.transformation",
        "GenAIHubEmbeddingConfig",
    ),
    "IBMWatsonXAudioTranscriptionConfig": (
        ".llms.watsonx.audio_transcription.transformation",
        "IBMWatsonXAudioTranscriptionConfig",
    ),
    "GithubCopilotConfig": (
        ".llms.github_copilot.chat.transformation",
        "GithubCopilotConfig",
    ),
    "GithubCopilotResponsesAPIConfig": (
        ".llms.github_copilot.responses.transformation",
        "GithubCopilotResponsesAPIConfig",
    ),
    "GithubCopilotEmbeddingConfig": (
        ".llms.github_copilot.embedding.transformation",
        "GithubCopilotEmbeddingConfig",
    ),
    "ChatGPTConfig": (".llms.chatgpt.chat.transformation", "ChatGPTConfig"),
    "ChatGPTResponsesAPIConfig": (
        ".llms.chatgpt.responses.transformation",
        "ChatGPTResponsesAPIConfig",
    ),
    "NebiusConfig": (".llms.nebius.chat.transformation", "NebiusConfig"),
    "WandbConfig": (".llms.wandb.chat.transformation", "WandbConfig"),
    "GigaChatConfig": (".llms.gigachat.chat.transformation", "GigaChatConfig"),
    "GigaChatEmbeddingConfig": (
        ".llms.gigachat.embedding.transformation",
        "GigaChatEmbeddingConfig",
    ),
    "DashScopeChatConfig": (
        ".llms.dashscope.chat.transformation",
        "DashScopeChatConfig",
    ),
    "MoonshotChatConfig": (".llms.moonshot.chat.transformation", "MoonshotChatConfig"),
    "DockerModelRunnerChatConfig": (
        ".llms.docker_model_runner.chat.transformation",
        "DockerModelRunnerChatConfig",
    ),
    "V0ChatConfig": (".llms.v0.chat.transformation", "V0ChatConfig"),
    "OCIChatConfig": (".llms.oci.chat.transformation", "OCIChatConfig"),
    "MorphChatConfig": (".llms.morph.chat.transformation", "MorphChatConfig"),
    "RAGFlowConfig": (".llms.ragflow.chat.transformation", "RAGFlowConfig"),
    "LambdaAIChatConfig": (".llms.lambda_ai.chat.transformation", "LambdaAIChatConfig"),
    "HyperbolicChatConfig": (
        ".llms.hyperbolic.chat.transformation",
        "HyperbolicChatConfig",
    ),
    "VercelAIGatewayConfig": (
        ".llms.vercel_ai_gateway.chat.transformation",
        "VercelAIGatewayConfig",
    ),
    "OVHCloudChatConfig": (".llms.ovhcloud.chat.transformation", "OVHCloudChatConfig"),
    "OVHCloudEmbeddingConfig": (
        ".llms.ovhcloud.embedding.transformation",
        "OVHCloudEmbeddingConfig",
    ),
    "CometAPIEmbeddingConfig": (
        ".llms.cometapi.embed.transformation",
        "CometAPIEmbeddingConfig",
    ),
    "LemonadeChatConfig": (".llms.lemonade.chat.transformation", "LemonadeChatConfig"),
    "SnowflakeEmbeddingConfig": (
        ".llms.snowflake.embedding.transformation",
        "SnowflakeEmbeddingConfig",
    ),
    "AmazonNovaChatConfig": (
        ".llms.amazon_nova.chat.transformation",
        "AmazonNovaChatConfig",
    ),
}

# Import map for utils module lazy imports
_UTILS_MODULE_IMPORT_MAP = {
    "encoding": ("litellm.main", "encoding"),
    "BaseVectorStore": (
        "litellm.integrations.vector_store_integrations.base_vector_store",
        "BaseVectorStore",
    ),
    "CredentialAccessor": (
        "litellm.litellm_core_utils.credential_accessor",
        "CredentialAccessor",
    ),
    "exception_type": (
        "litellm.litellm_core_utils.exception_mapping_utils",
        "exception_type",
    ),
    "get_error_message": (
        "litellm.litellm_core_utils.exception_mapping_utils",
        "get_error_message",
    ),
    "_get_response_headers": (
        "litellm.litellm_core_utils.exception_mapping_utils",
        "_get_response_headers",
    ),
    "get_llm_provider": (
        "litellm.litellm_core_utils.get_llm_provider_logic",
        "get_llm_provider",
    ),
    "_is_non_openai_azure_model": (
        "litellm.litellm_core_utils.get_llm_provider_logic",
        "_is_non_openai_azure_model",
    ),
    "get_supported_openai_params": (
        "litellm.litellm_core_utils.get_supported_openai_params",
        "get_supported_openai_params",
    ),
    "LiteLLMResponseObjectHandler": (
        "litellm.litellm_core_utils.llm_response_utils.convert_dict_to_response",
        "LiteLLMResponseObjectHandler",
    ),
    "_handle_invalid_parallel_tool_calls": (
        "litellm.litellm_core_utils.llm_response_utils.convert_dict_to_response",
        "_handle_invalid_parallel_tool_calls",
    ),
    "convert_to_model_response_object": (
        "litellm.litellm_core_utils.llm_response_utils.convert_dict_to_response",
        "convert_to_model_response_object",
    ),
    "convert_to_streaming_response": (
        "litellm.litellm_core_utils.llm_response_utils.convert_dict_to_response",
        "convert_to_streaming_response",
    ),
    "convert_to_streaming_response_async": (
        "litellm.litellm_core_utils.llm_response_utils.convert_dict_to_response",
        "convert_to_streaming_response_async",
    ),
    "get_api_base": (
        "litellm.litellm_core_utils.llm_response_utils.get_api_base",
        "get_api_base",
    ),
    "ResponseMetadata": (
        "litellm.litellm_core_utils.llm_response_utils.response_metadata",
        "ResponseMetadata",
    ),
    "_parse_content_for_reasoning": (
        "litellm.litellm_core_utils.prompt_templates.common_utils",
        "_parse_content_for_reasoning",
    ),
    "LiteLLMLoggingObject": (
        "litellm.litellm_core_utils.redact_messages",
        "LiteLLMLoggingObject",
    ),
    "redact_message_input_output_from_logging": (
        "litellm.litellm_core_utils.redact_messages",
        "redact_message_input_output_from_logging",
    ),
    "CustomStreamWrapper": (
        "litellm.litellm_core_utils.streaming_handler",
        "CustomStreamWrapper",
    ),
    "BaseGoogleGenAIGenerateContentConfig": (
        "litellm.llms.base_llm.google_genai.transformation",
        "BaseGoogleGenAIGenerateContentConfig",
    ),
    "BaseOCRConfig": ("litellm.llms.base_llm.ocr.transformation", "BaseOCRConfig"),
    "BaseSearchConfig": (
        "litellm.llms.base_llm.search.transformation",
        "BaseSearchConfig",
    ),
    "BaseTextToSpeechConfig": (
        "litellm.llms.base_llm.text_to_speech.transformation",
        "BaseTextToSpeechConfig",
    ),
    "BedrockModelInfo": ("litellm.llms.bedrock.common_utils", "BedrockModelInfo"),
    "CohereModelInfo": ("litellm.llms.cohere.common_utils", "CohereModelInfo"),
    "MistralOCRConfig": ("litellm.llms.mistral.ocr.transformation", "MistralOCRConfig"),
    "Rules": ("litellm.litellm_core_utils.rules", "Rules"),
    "AsyncHTTPHandler": ("litellm.llms.custom_httpx.http_handler", "AsyncHTTPHandler"),
    "HTTPHandler": ("litellm.llms.custom_httpx.http_handler", "HTTPHandler"),
    "get_num_retries_from_retry_policy": (
        "litellm.router_utils.get_retry_from_policy",
        "get_num_retries_from_retry_policy",
    ),
    "reset_retry_policy": (
        "litellm.router_utils.get_retry_from_policy",
        "reset_retry_policy",
    ),
    "get_secret": ("litellm.secret_managers.main", "get_secret"),
    "get_coroutine_checker": (
        "litellm.litellm_core_utils.cached_imports",
        "get_coroutine_checker",
    ),
    "get_litellm_logging_class": (
        "litellm.litellm_core_utils.cached_imports",
        "get_litellm_logging_class",
    ),
    "get_set_callbacks": (
        "litellm.litellm_core_utils.cached_imports",
        "get_set_callbacks",
    ),
    "get_litellm_metadata_from_kwargs": (
        "litellm.litellm_core_utils.core_helpers",
        "get_litellm_metadata_from_kwargs",
    ),
    "map_finish_reason": (
        "litellm.litellm_core_utils.core_helpers",
        "map_finish_reason",
    ),
    "process_response_headers": (
        "litellm.litellm_core_utils.core_helpers",
        "process_response_headers",
    ),
    "delete_nested_value": (
        "litellm.litellm_core_utils.dot_notation_indexing",
        "delete_nested_value",
    ),
    "is_nested_path": (
        "litellm.litellm_core_utils.dot_notation_indexing",
        "is_nested_path",
    ),
    "_get_base_model_from_litellm_call_metadata": (
        "litellm.litellm_core_utils.get_litellm_params",
        "_get_base_model_from_litellm_call_metadata",
    ),
    "get_litellm_params": (
        "litellm.litellm_core_utils.get_litellm_params",
        "get_litellm_params",
    ),
    "_ensure_extra_body_is_safe": (
        "litellm.litellm_core_utils.llm_request_utils",
        "_ensure_extra_body_is_safe",
    ),
    "get_formatted_prompt": (
        "litellm.litellm_core_utils.llm_response_utils.get_formatted_prompt",
        "get_formatted_prompt",
    ),
    "get_response_headers": (
        "litellm.litellm_core_utils.llm_response_utils.get_headers",
        "get_response_headers",
    ),
    "update_response_metadata": (
        "litellm.litellm_core_utils.llm_response_utils.response_metadata",
        "update_response_metadata",
    ),
    "executor": ("litellm.litellm_core_utils.thread_pool_executor", "executor"),
    "BaseAnthropicMessagesConfig": (
        "litellm.llms.base_llm.anthropic_messages.transformation",
        "BaseAnthropicMessagesConfig",
    ),
    "BaseAudioTranscriptionConfig": (
        "litellm.llms.base_llm.audio_transcription.transformation",
        "BaseAudioTranscriptionConfig",
    ),
    "BaseBatchesConfig": (
        "litellm.llms.base_llm.batches.transformation",
        "BaseBatchesConfig",
    ),
    "BaseContainerConfig": (
        "litellm.llms.base_llm.containers.transformation",
        "BaseContainerConfig",
    ),
    "BaseEmbeddingConfig": (
        "litellm.llms.base_llm.embedding.transformation",
        "BaseEmbeddingConfig",
    ),
    "BaseImageEditConfig": (
        "litellm.llms.base_llm.image_edit.transformation",
        "BaseImageEditConfig",
    ),
    "BaseImageGenerationConfig": (
        "litellm.llms.base_llm.image_generation.transformation",
        "BaseImageGenerationConfig",
    ),
    "BaseImageVariationConfig": (
        "litellm.llms.base_llm.image_variations.transformation",
        "BaseImageVariationConfig",
    ),
    "BasePassthroughConfig": (
        "litellm.llms.base_llm.passthrough.transformation",
        "BasePassthroughConfig",
    ),
    "BaseRealtimeConfig": (
        "litellm.llms.base_llm.realtime.transformation",
        "BaseRealtimeConfig",
    ),
    "BaseRerankConfig": (
        "litellm.llms.base_llm.rerank.transformation",
        "BaseRerankConfig",
    ),
    "BaseVectorStoreConfig": (
        "litellm.llms.base_llm.vector_store.transformation",
        "BaseVectorStoreConfig",
    ),
    "BaseVectorStoreFilesConfig": (
        "litellm.llms.base_llm.vector_store_files.transformation",
        "BaseVectorStoreFilesConfig",
    ),
    "BaseVideoConfig": (
        "litellm.llms.base_llm.videos.transformation",
        "BaseVideoConfig",
    ),
    "ANTHROPIC_API_ONLY_HEADERS": (
        "litellm.types.llms.anthropic",
        "ANTHROPIC_API_ONLY_HEADERS",
    ),
    "AnthropicThinkingParam": (
        "litellm.types.llms.anthropic",
        "AnthropicThinkingParam",
    ),
    "RerankResponse": ("litellm.types.rerank", "RerankResponse"),
    "ChatCompletionDeltaToolCallChunk": (
        "litellm.types.llms.openai",
        "ChatCompletionDeltaToolCallChunk",
    ),
    "ChatCompletionToolCallChunk": (
        "litellm.types.llms.openai",
        "ChatCompletionToolCallChunk",
    ),
    "ChatCompletionToolCallFunctionChunk": (
        "litellm.types.llms.openai",
        "ChatCompletionToolCallFunctionChunk",
    ),
    "LiteLLM_Params": ("litellm.types.router", "LiteLLM_Params"),
}

# Export all name tuples and import maps for use in _lazy_imports.py
__all__ = [
    # Name tuples
    "COST_CALCULATOR_NAMES",
    "LITELLM_LOGGING_NAMES",
    "UTILS_NAMES",
    "TOKEN_COUNTER_NAMES",
    "LLM_CLIENT_CACHE_NAMES",
    "BEDROCK_TYPES_NAMES",
    "TYPES_UTILS_NAMES",
    "CACHING_NAMES",
    "HTTP_HANDLER_NAMES",
    "DOTPROMPT_NAMES",
    "LLM_CONFIG_NAMES",
    "TYPES_NAMES",
    "LLM_PROVIDER_LOGIC_NAMES",
    "UTILS_MODULE_NAMES",
    # Import maps
    "_UTILS_IMPORT_MAP",
    "_COST_CALCULATOR_IMPORT_MAP",
    "_TYPES_UTILS_IMPORT_MAP",
    "_TOKEN_COUNTER_IMPORT_MAP",
    "_BEDROCK_TYPES_IMPORT_MAP",
    "_CACHING_IMPORT_MAP",
    "_LITELLM_LOGGING_IMPORT_MAP",
    "_DOTPROMPT_IMPORT_MAP",
    "_TYPES_IMPORT_MAP",
    "_LLM_CONFIGS_IMPORT_MAP",
    "_LLM_PROVIDER_LOGIC_IMPORT_MAP",
    "_UTILS_MODULE_IMPORT_MAP",
]
