llmclient 0.2.0

Rust LLM client - Gemini, GPT, Claude, Mistral, Groq
Documentation
Gemini:

[{\n  \"error\": {\n    \"code\": 400,\n    \"message\": \"98192 is out of supported range [1, 8193);  for max output tokens parameter.\",\n    \"status\": \"INVALID_ARGUMENT\"\n  }\n}\n]

Gpt :

"{\n  \"error\": {\n    \"message\": \"'messages' must contain the word 'json' in some form, to use 'response_format' of type 'json_object'.\",\n    \"type\": \"invalid_request_error\",\n    \"param\": \"messages\",\n    \"code\": null\n  }\n}\n"
thread 'gpt::tests::test_call_gpt_basic' panicked at src/gpt.rs:336:66:

Claude:

"{\"type\":\"error\",\"error\":{\"type\":\"rate_limit_error\",\"message\":\"Number of concurrent connections has exceeded your rate limit. Please try again later or contact sales at https://www.anthropic.com/contact-sales to discuss your options for a rate limit increase.\"}}"

// Mistral
    Error:
      type: object
      properties:
        type:
          type: string
          nullable: false
        message:
          type: string
          nullable: false
        param:
          type: string
          nullable: true
        code:
          type: string
          nullable: true
      required:
        - type
        - message
        - param
        - code
    ErrorResponse:
      type: object
      properties:
        error:
          $ref: '#/components/schemas/Error'
      required:
        - error