Ë
    Ýªgû  ã                   óJ   — d Z ddlmZ ddlmZ ddlmZ dededeeef   fd„Z	y	)
zU
Helper util for handling anthropic-specific cost calculation
- e.g.: prompt caching
é    )ÚTuple©Úgeneric_cost_per_token)ÚUsageÚmodelÚusageÚreturnc                 ó   — t        | |d¬«      S )aO  
    Calculates the cost per token for a given model, prompt tokens, and completion tokens.

    Input:
        - model: str, the model name without provider prefix
        - usage: LiteLLM Usage block, containing anthropic caching information

    Returns:
        Tuple[float, float] - prompt_cost_in_usd, completion_cost_in_usd
    Ú	anthropic)r   r   Úcustom_llm_providerr   )r   r   s     ú\/var/www/openai/venv/lib/python3.12/site-packages/litellm/llms/anthropic/cost_calculation.pyÚcost_per_tokenr      s   € ô "Ø˜5°kôð ó    N)
Ú__doc__Útypingr   Ú.litellm.litellm_core_utils.llm_cost_calc.utilsr   Úlitellm.types.utilsr   ÚstrÚfloatr   © r   r   Ú<module>r      s8   ðñõ
 å QÝ %ð˜#ð  eð °°e¸U°lÑ0Cô r   