Close Menu
Versa AI hub
  • AI Ethics
  • AI Legislation
  • Business
  • Cybersecurity
  • Media and Entertainment
  • Content Creation
  • Art Generation
  • Research
  • Tools
  • Resources

Subscribe to Updates

Subscribe to our newsletter and stay updated with the latest news and exclusive offers.

What's Hot

Anthropic usage statistics paint a detailed picture of AI success

January 24, 2026

YouTube vows to fight ‘AI slop’ in 2026

January 23, 2026

Spreading real-time interactive video with Overworld

January 23, 2026
Facebook X (Twitter) Instagram
Versa AI hubVersa AI hub
Saturday, January 24
Facebook X (Twitter) Instagram
Login
  • AI Ethics
  • AI Legislation
  • Business
  • Cybersecurity
  • Media and Entertainment
  • Content Creation
  • Art Generation
  • Research
  • Tools
  • Resources
Versa AI hub
Home»Tools»Scaleway hugging face reasoning provider
Tools

Scaleway hugging face reasoning provider

versatileaiBy versatileaiSeptember 20, 2025No Comments4 Mins Read
Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
#image_title
Share
Facebook Twitter LinkedIn Pinterest Email

I’m excited to share that Scaleway is a supported reasoning provider for Hug Face Hub! Scaleway joins a growing ecosystem and directly enhances the breadth and capabilities of serverless inference on the hub’s model page. Inference providers are seamlessly integrated into the client SDK (both JS and Python), making it easy to use different models using preferred providers.

With this launch, popular open weight models like the GPT-Oss, QWEN3, Deepseek R1, and Gemma 3 are easier to access than ever. You can browse your Scaleway organization in the hub at https://huggingface.co/scaleway and try out supported models at https://huggingface.co/models?inference_provider =scaleway&sort=treending.

The Scaleway Generative API is a fully managed serverless service that provides access to the frontier AI models of major research labs through simple API calls. The service offers competitive per token pricing starting at 0.20 euros per million tokens.

The service runs on a secure infrastructure in the European data center (Paris, France) and ensures data sovereignty and low latency for European users. The platform supports advanced features such as structured output, function calls and multimodal features for both text and image processing.

Built for production, Scaleway’s inference infrastructure provides sub-200ms response time for the first token, making it ideal for interactive applications and agent workflows. This service supports both text generation and embedded models. Find out more about Scaleway’s platform and infrastructure at https://www.scaleway.com/en/generative-apis/.

Learn more about using Scaleway as an inference provider on our dedicated documentation page.

See the list of supported models here.

How it works

In the website UI

In User Account Settings, you set your own API key for the provider you signed up for. If no custom key is configured, the request is routed through HF. Order a provider if you like. This applies to model page widgets and code snippets.

Inference provider

As mentioned before, when calling an inference provider there are two modes: a custom key (the call goes directly to the inference provider, using the corresponding inference provider’s own API key) (in that case no tokens are required from the provider.

Inference provider

The model page introduces third-party inference providers (compatible with current models sorted by user preferences)

Inference provider

From the client SDK

I’m using Huggingface_hub from Python

The following example shows how to use Swiss AI’s Apertus-70B using Scaleway as the inference provider: Automatic routing through a hugging face can be used with a hugging face token or your own Scaleway API key if you have one.

Note: This requires using a recent version of Huggingface_hub (>=0.34.6).

Import OS
from huggingface_hub Import Inference client=Inference client(provider=“Scaleway”,api_key = os.environ(“HF_TOKEN”) ) message = ({
“role”: “user”,
“content”: “Writing poetry in Shakespeare’s style”
}) complete = client.chat.completions.create(model =“Openai/gpt-oss-120b”message = message,)

printing(complete.choices)0). message)

From JS using @huggingface/Incerence

Import { inference } from “@Huggingface/Inference”;

const Client= new inference(process.Env.hf_token);

const chatcompletion = wait client.ChatCompletion({
Model: “Openai/gpt-oss-120b”,
message:({
role: “user”,
content: “Writing poetry in Shakespeare’s style”,},),
Provider: “Scaleway”,});

console.log(ChatCompletion.Choices(0).message);

Request

Here’s how billing works:

For direct requests, i.e. when using keys from inference providers, the corresponding provider will be billed. For example, if you are using a Scaleway API key, your Scaleway account will be billed.

For routed requests, i.e. when authenticating through a facehub that hugs, you only pay the standard provider API rate. There is no additional markup from us. It simply passes the provider’s costs directly. (In the future, we may establish a revenue sharing agreement with our provider partners.)

Important Memopia users get $2 worth of inference credits each month. You can use them between providers. 🔥

Subscribe to our Hugging Face Pro plan for access to inference credits, Zerogpu, Spaces Dev Mode, 20x high limits and more.

We also infer small allocations for sign-in free users for free, but upgrade to Pro if possible!

Feedback and next steps

We want to get your feedback! Share your thoughts and comments here: https://huggingface.co/spaces/huggingface/huggingdiscussions/discussions/49

author avatar
versatileai
See Full Bio
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Previous ArticleTrust Paradox: Colorado AI Law and Local Government Challenges – PA Times Online
Next Article Waves Bazaar 2.0: Strengthen the power of digital content creators
versatileai

Related Posts

Tools

Anthropic usage statistics paint a detailed picture of AI success

January 24, 2026
Tools

Spreading real-time interactive video with Overworld

January 23, 2026
Tools

D4RT: Integrated fast 4D scene reconstruction and tracking

January 23, 2026
Add A Comment

Comments are closed.

Top Posts

Gemini achieves gold medal level at International University Programming Contest World Finals

January 21, 20266 Views

U.S. House committee to vote on bill giving Congress authority over AI chip exports | WIN 98.5 Your Country | WNWN-FM

January 21, 20266 Views

Things security leaders need to know

July 9, 20256 Views
Stay In Touch
  • YouTube
  • TikTok
  • Twitter
  • Instagram
  • Threads
Latest Reviews

Subscribe to Updates

Subscribe to our newsletter and stay updated with the latest news and exclusive offers.

Most Popular

Gemini achieves gold medal level at International University Programming Contest World Finals

January 21, 20266 Views

U.S. House committee to vote on bill giving Congress authority over AI chip exports | WIN 98.5 Your Country | WNWN-FM

January 21, 20266 Views

Things security leaders need to know

July 9, 20256 Views
Don't Miss

Anthropic usage statistics paint a detailed picture of AI success

January 24, 2026

YouTube vows to fight ‘AI slop’ in 2026

January 23, 2026

Spreading real-time interactive video with Overworld

January 23, 2026
Service Area
X (Twitter) Instagram YouTube TikTok Threads RSS
  • About Us
  • Contact Us
  • Privacy Policy
  • Terms and Conditions
  • Disclaimer
© 2026 Versa AI Hub. All Rights Reserved.

Type above and press Enter to search. Press Esc to cancel.

Sign In or Register

Welcome Back!

Login to your account below.

Lost password?