Skip to content

Releases: huggingface/huggingface_hub

[v0.33.4] [Tiny-Agent]: Fix schema validation error for default MCP tools

11 Jul 12:35
Compare
Choose a tag to compare

[v0.33.3] [Tiny-Agent]: Update tiny-agents example

11 Jul 08:55
8709168
Compare
Choose a tag to compare

[v0.33.2] [Tiny-Agent]: Switch to VSCode MCP format

02 Jul 06:29
8c5770f
Compare
Choose a tag to compare

Full Changelog: v0.33.1...v0.33.2

  • [Tiny-Agent] Switch to VSCode MCP format + fix headers handling #3166 by @Wauplin

Breaking changes:

  • no more config nested mapping => everything at root level
  • headers at root level instead of inside options.requestInit
  • updated the way values are pulled from ENV (based on input id)

Example of agent.json:

{
  "model": "Qwen/Qwen2.5-72B-Instruct",
  "provider": "nebius",
  "inputs": [
    {
      "type": "promptString",
      "id": "hf-token",
      "description": "Token for Hugging Face API access",
      "password": true
    }
  ],
  "servers": [
    {
      "type": "http",
      "url": "https://huggingface.co/mcp",
      "headers": {
        "Authorization": "Bearer ${input:hf-token}"
      }
    }
  ]
}

Find more examples in https://huggingface.co/datasets/tiny-agents/tiny-agents

[v0.33.1]: Inference Providers Bug Fixes, Tiny-Agents Message handling Improvement, and Inference Endpoints Health Check Update

25 Jun 12:18
Compare
Choose a tag to compare

Full Changelog: v0.33.0...v0.33.1

This release introduces bug fixes for chat completion type compatibility and feature extraction parameters, enhanced message handling in tiny-agents, and updated inference endpoint health check:

  • [Tiny agents] Add tool call to messages #3159 by @NielsRogge
  • fix: update payload preparation to merge parameters into the output dictionary #3160 by @mishig25
  • fix(inference_endpoints): use GET healthRoute instead of GET / to check status #3165 by @mfuntowicz
  • Recursive filter_none in Inference Providers #3178 by @Wauplin

[v0.33.0]: Welcoming Featherless.AI and Groq as Inference Providers!

11 Jun 14:14
d5dff4e
Compare
Choose a tag to compare

⚡ New provider: Featherless.AI

Featherless AI is a serverless AI inference provider with unique model loading and GPU orchestration abilities that makes an exceptionally large catalog of models available for users. Providers often offer either a low cost of access to a limited set of models, or an unlimited range of models with users managing servers and the associated costs of operation. Featherless provides the best of both worlds offering unmatched model range and variety but with serverless pricing. Find the full list of supported models on the models page.

from huggingface_hub import InferenceClient

client = InferenceClient(provider="featherless-ai")

completion = client.chat.completions.create(
    model="deepseek-ai/DeepSeek-R1-0528", 
    messages=[
        {
            "role": "user",
            "content": "What is the capital of France?"
        }
    ], 
)

print(completion.choices[0].message)
  • ✨ Support for Featherless.ai as inference provider by @pohnean in #3081

⚡ New provider: Groq

At the heart of Groq's technology is the Language Processing Unit (LPU™), a new type of end-to-end processing unit system that provides the fastest inference for computationally intensive applications with a sequential component, such as Large Language Models (LLMs). LPUs are designed to overcome the limitations of GPUs for inference, offering significantly lower latency and higher throughput. This makes them ideal for real-time AI applications.

Groq offers fast AI inference for openly-available models. They provide an API that allows developers to easily integrate these models into their applications. It offers an on-demand, pay-as-you-go model for accessing a wide range of openly-available LLMs.

from huggingface_hub import InferenceClient

client = InferenceClient(provider="groq")

completion = client.chat.completions.create(
    model="meta-llama/Llama-4-Scout-17B-16E-Instruct",
    messages=[
        {
            "role": "user",
            "content": [
                {"type": "text", "text": "Describe this image in one sentence."},
                {
                    "type": "image_url",
                    "image_url": {"url": "https://vagabundler.com/wp-content/uploads/2019/06/P3160166-Copy.jpg"},
                },
            ],
        }
    ],
)

print(completion.choices[0].message)

🤖 MCP and Tiny-agents

It is now possible to run tiny-agents using a local server e.g. llama.cpp. 100% local agents are right behind the corner!

Fixing some DX issues in the tiny-agents CLI.

📚 Documentation

New translation from the Hindi-speaking community, for the community!

  • Added Hindi translation for git_vs_http.md in concepts section by @february-king in #3156

🛠️ Small fixes and maintenance

😌 QoL improvements

🐛 Bug and typo fixes

🏗️ internal

Significant community contributions

The following contributors have made significant changes to the library over the last release:

  • @pohnean
    • ✨ Support for Featherless.ai as inference provider (#3081)
  • @february-king
    • Added Hindi translation for git_vs_http.md in concepts section (#3156)

[v0.32.6] [Upload large folder] fix for wrongly saved upload_mode/remote_oid

11 Jun 08:18
f498b42
Compare
Choose a tag to compare

[v0.32.5] [Tiny-Agents] inject environment variables in headers

10 Jun 16:04
8dfb199
Compare
Choose a tag to compare
  • Inject env var in headers + better type annotations #3142

Full Changelog: v0.32.4...v0.32.5

[v0.32.4]: Bug fixes in `tiny-agents`, and fix input handling for question-answering task.

03 Jun 10:04
Compare
Choose a tag to compare

Full Changelog: v0.32.3...v0.32.4

This release introduces bug fixes to tiny-agents and InferenceClient.question_answering:

[v0.32.3]: Handle env variables in `tiny-agents`, better CLI exit and handling of MCP tool calls arguments

30 May 08:29
Compare
Choose a tag to compare

Full Changelog: v0.32.2...v0.32.3

This release introduces some improvements and bug fixes to tiny-agents:

  • [tiny-agents] Handle env variables in tiny-agents (Python client) #3129
  • [Fix] tiny-agents cli exit issues #3125
  • Improve Handling of MCP Tool Call Arguments #3127

[v0.32.2]: Add endpoint support in Tiny-Agent + fix `snapshot_download` on large repos

27 May 09:24
6dd0164
Compare
Choose a tag to compare

Full Changelog: v0.32.1...v0.32.2

  • [MCP] Add local/remote endpoint inference support #3121
  • Fix snapshot_download on very large repo (>50k files) #3122