Close Menu
Versa AI hub
  • AI Ethics
  • AI Legislation
  • Business
  • Cybersecurity
  • Media and Entertainment
  • Content Creation
  • Art Generation
  • Research
  • Tools

Subscribe to Updates

Subscribe to our newsletter and stay updated with the latest news and exclusive offers.

What's Hot

StarCoder2 and Stack V2

July 4, 2025

Intel®Gaudi®2AI Accelerator Text Generation Pipeline

July 3, 2025

CAC has announced AI-powered business registration portal – thisdaylive

July 3, 2025
Facebook X (Twitter) Instagram
Versa AI hubVersa AI hub
Friday, July 4
Facebook X (Twitter) Instagram
Login
  • AI Ethics
  • AI Legislation
  • Business
  • Cybersecurity
  • Media and Entertainment
  • Content Creation
  • Art Generation
  • Research
  • Tools
Versa AI hub
Home»Tools»Hub fireworks
Tools

Hub fireworks

By February 17, 2025No Comments2 Mins Read
Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
Share
Facebook Twitter LinkedIn Pinterest Email




Julian Chaumonde's avatar


Following the recent announcement on Hub’s inference provider, we are excited to share the fireworks.

fireworks.ai directly provides fiery serverless inference across the HF ecosystem of model pages and libraries and tools, making inferences to your favorite models easier than ever to execute.

Fireworks.ai is supported as a face-hugging reasoning provider

Of course, from now on, serverless inference can be performed on the following models via fireworks.

More lists can be found here for the complete list.

Light up your project with today’s fireworks.

How it works

In the website UI

fireworks.ai Inference Provider UI

Find all models supported for fireworks in HF here.

From the client SDK

I’m using Huggingface_hub from Python

The following example shows how to use Fireworks.ai as an inference provider: Automatic routing through a hugging face can be used with a hugging face token, or if you have your own fireworks, you can use your own.

Install huggingface_hub from source:

PIP Install git+https://github.com/huggingface/huggingface_hub

Use the Huggingface_Hub Python library to define provider parameters and call the Fireworks.ai endpoint.

from huggingface_hub Import Inference client=Inference client(provider=“firework”,api_key =「xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
) Message = ({
“role”: “user”,
“content”: “What is the capital of France?”
}) complete = client.chat.completions.create(model =“deepseek-ai/deepseek-r1”message = message, max_tokens =500
))

printing(complete.choices)0). message)

From JS using @huggingface/Incerence

Import { hfinference } from “@Huggingface/Inference”;

const Client= new hfinference(「xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx);

const chatcompletion = wait client.ChatCompletion({
Model: “deepseek-ai/deepseek-r1”,
message:({
role: “user”,
content: “How do you make a very spicy mayonnaise?”
}),
Provider: “firework”,
max_tokens: 500
});

console.log(ChatCompletion.Choices(0).message);

From an HTTP call

Here’s how to call Llama-3.3-70B-Instruct as an inference provider via Curl using Fireworks.ai:

curl ‘https://router.huggingface.co/fireworks-ai/v1/chat/completions’ \ -H ‘Authorization: Bearer xxxxxxxxx’ \ -H ‘Content-Type: application/json’ \ –data ‘{ ” Model “:” accounts/fireworks/models/llama-v3-70b-instruct “, “messages”: ({” role “:” user “, ” content “:” what does life mean if you were a dog? Is it? “}),” max_tokens “:500, “stream”:false}’

Request

If you make a direct request, that is, if you use a fireworks key, you will be billed directly to your fireworks account.

For routed requests, i.e. when authenticating through the hub, you only pay the standard fireworks API rate. There is no additional markup. Pass the provider’s costs directly. (In the future, we may establish a revenue sharing agreement with our provider partners.)

Important Memopia users get $2 worth of inference credits each month. You can use them between providers. 🔥

Subscribe to our Hugging Face Pro plan for access to inference credits, Zerogpu, Spaces Dev Mode, 20x high limits and more.

author avatar
See Full Bio
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Previous ArticleUK launches AI Security Institute to protect national security and fight crime
Next Article I’m focusing too much on AI and removing “social” from social media

Related Posts

Tools

StarCoder2 and Stack V2

July 4, 2025
Tools

Intel®Gaudi®2AI Accelerator Text Generation Pipeline

July 3, 2025
Tools

Research shows that AI can reduce global carbon emissions

July 3, 2025
Add A Comment
Leave A Reply Cancel Reply

Top Posts

Impact International | EU AI ACT Enforcement: Business Transparency and Human Rights Impact in 2025

June 2, 20251 Views

Presight plans to expand its AI business internationally

April 14, 20251 Views

PlanetScale Vectors GA: MySQL and AI Database Game Changer

April 14, 20251 Views
Stay In Touch
  • YouTube
  • TikTok
  • Twitter
  • Instagram
  • Threads
Latest Reviews

Subscribe to Updates

Subscribe to our newsletter and stay updated with the latest news and exclusive offers.

Most Popular

Impact International | EU AI ACT Enforcement: Business Transparency and Human Rights Impact in 2025

June 2, 20251 Views

Presight plans to expand its AI business internationally

April 14, 20251 Views

PlanetScale Vectors GA: MySQL and AI Database Game Changer

April 14, 20251 Views
Don't Miss

StarCoder2 and Stack V2

July 4, 2025

Intel®Gaudi®2AI Accelerator Text Generation Pipeline

July 3, 2025

CAC has announced AI-powered business registration portal – thisdaylive

July 3, 2025
Service Area
X (Twitter) Instagram YouTube TikTok Threads RSS
  • About Us
  • Contact Us
  • Privacy Policy
  • Terms and Conditions
  • Disclaimer
© 2025 Versa AI Hub. All Rights Reserved.

Type above and press Enter to search. Press Esc to cancel.

Sign In or Register

Welcome Back!

Login to your account below.

Lost password?