AI & ML interests
None defined yet.

freddyaboultonย
posted
an
update
about 16 hours ago

freddyaboultonย
posted
an
update
16 days ago
Post
481
Time is running out! โฐ
Less than 24 hours to participate in the MCP Hackathon and win thousands of dollars in prizes! Don't miss this opportunity to showcase your skills.
Visit Agents-MCP-Hackathon/AI-Marketing-Content-Creator to register!
Less than 24 hours to participate in the MCP Hackathon and win thousands of dollars in prizes! Don't miss this opportunity to showcase your skills.
Visit Agents-MCP-Hackathon/AI-Marketing-Content-Creator to register!

freddyaboultonย
posted
an
update
16 days ago
Post
350
๐จ NotebookLM Dethroned?! ๐จ
Meet Fluxions vui: The new open-source dialogue generation model.
๐คฏ 100M Params, 40k hours audio!
๐๏ธ Multi-speaker audio
๐ Non-speech sounds (like [laughs]!)
๐ MIT License
Is this the future of content creation? Watch the video and decide for yourself!
https://huggingface.co/spaces/fluxions/vui-spacehttps://huggingface.co/fluxions/vui
Meet Fluxions vui: The new open-source dialogue generation model.
๐คฏ 100M Params, 40k hours audio!
๐๏ธ Multi-speaker audio
๐ Non-speech sounds (like [laughs]!)
๐ MIT License
Is this the future of content creation? Watch the video and decide for yourself!
https://huggingface.co/spaces/fluxions/vui-spacehttps://huggingface.co/fluxions/vui
Post
1838
The Gradio x Agents x MCP hackathon keeps growing! We now have more $1,000,000 in credit for participants and and >$16,000 in cash prizes for winners.
We've kept registration open until the end of this week, so join and let's build cool stuff together as a community: https://huggingface.co/spaces/ysharma/gradio-hackathon-registration-2025
We've kept registration open until the end of this week, so join and let's build cool stuff together as a community: https://huggingface.co/spaces/ysharma/gradio-hackathon-registration-2025
Post
4999
HOW TO ADD MCP SUPPORT TO ANY ๐ค SPACE
Gradio now supports MCP! If you want to convert an existing Space, like this one hexgrad/Kokoro-TTS, so that you can use it with Claude Desktop / Cursor / Cline / TinyAgents / or any LLM that supports MCP, here's all you need to do:
1. Duplicate the Space (in the Settings Tab)
2. Upgrade the Gradio
3. Set
4. (Optionally) add docstrings to the function so that the LLM knows how to use it, like this:
That's it! Now your LLM will be able to talk to you ๐คฏ
Gradio now supports MCP! If you want to convert an existing Space, like this one hexgrad/Kokoro-TTS, so that you can use it with Claude Desktop / Cursor / Cline / TinyAgents / or any LLM that supports MCP, here's all you need to do:
1. Duplicate the Space (in the Settings Tab)
2. Upgrade the Gradio
sdk_version
to 5.28
(in the README.md
)3. Set
mcp_server=True
in launch()
4. (Optionally) add docstrings to the function so that the LLM knows how to use it, like this:
def generate(text, speed=1):
"""
Convert text to speech audio.
Parameters:
text (str): The input text to be converted to speech.
speed (float, optional): Playback speed of the generated speech.
That's it! Now your LLM will be able to talk to you ๐คฏ
Post
2733
Hi folks! Excited to share a new feature from the Gradio team along with a tutorial.
If you don't already know, Gradio is an open-source Python library used to build interfaces for machine learning models. Beyond just creating UIs, Gradio also exposes API capabilities and now, Gradio apps can be launched Model Context Protocol (MCP) servers for LLMs.
If you already know how to use Gradio, there are only two additional things you need to do:
* Add standard docstrings to your function (these will be used to generate the descriptions for your tools for the LLM)
* Set
Here's a complete example (make sure you already have the latest version of Gradio installed):
This is a very simple example, but you can add the ability to generate Ghibli images or speak emotions to any LLM that supports MCP. Once you have an MCP running locally, you can copy-paste the same app to host it on [Hugging Face Spaces](https://huggingface.co/spaces/) as well.
All free and open-source of course! Full tutorial: https://www.gradio.app/guides/building-mcp-server-with-gradio
If you don't already know, Gradio is an open-source Python library used to build interfaces for machine learning models. Beyond just creating UIs, Gradio also exposes API capabilities and now, Gradio apps can be launched Model Context Protocol (MCP) servers for LLMs.
If you already know how to use Gradio, there are only two additional things you need to do:
* Add standard docstrings to your function (these will be used to generate the descriptions for your tools for the LLM)
* Set
mcp_server=True
in launch()
Here's a complete example (make sure you already have the latest version of Gradio installed):
import gradio as gr
def letter_counter(word, letter):
"""Count the occurrences of a specific letter in a word.
Args:
word: The word or phrase to analyze
letter: The letter to count occurrences of
Returns:
The number of times the letter appears in the word
"""
return word.lower().count(letter.lower())
demo = gr.Interface(
fn=letter_counter,
inputs=["text", "text"],
outputs="number",
title="Letter Counter",
description="Count how many times a letter appears in a word"
)
demo.launch(mcp_server=True)
This is a very simple example, but you can add the ability to generate Ghibli images or speak emotions to any LLM that supports MCP. Once you have an MCP running locally, you can copy-paste the same app to host it on [Hugging Face Spaces](https://huggingface.co/spaces/) as well.
All free and open-source of course! Full tutorial: https://www.gradio.app/guides/building-mcp-server-with-gradio

hystsย
updated
a
Space
2 months ago
Post
3848
JOURNEY TO 1 MILLION DEVELOPERS
5 years ago, we launched Gradio as a simple Python library to let researchers at Stanford easily demo computer vision models with a web interface.
Today, Gradio is used by >1 million developers each month to build and share AI web apps. This includes some of the most popular open-source projects of all time, like Automatic1111, Fooocus, Oobaboogaโs Text WebUI, Dall-E Mini, and LLaMA-Factory.
How did we get here? How did Gradio keep growing in the very crowded field of open-source Python libraries? I get this question a lot from folks who are building their own open-source libraries. This post distills some of the lessons that I have learned over the past few years:
1. Invest in good primitives, not high-level abstractions
2. Embed virality directly into your library
3. Focus on a (growing) niche
4. Your only roadmap should be rapid iteration
5. Maximize ways users can consume your library's outputs
1. Invest in good primitives, not high-level abstractions
When we first launched Gradio, we offered only one high-level class (gr.Interface), which created a complete web app from a single Python function. We quickly realized that developers wanted to create other kinds of apps (e.g. multi-step workflows, chatbots, streaming applications), but as we started listing out the apps users wanted to build, we realized what we needed to do:
Read the rest here: https://x.com/abidlabs/status/1907886
5 years ago, we launched Gradio as a simple Python library to let researchers at Stanford easily demo computer vision models with a web interface.
Today, Gradio is used by >1 million developers each month to build and share AI web apps. This includes some of the most popular open-source projects of all time, like Automatic1111, Fooocus, Oobaboogaโs Text WebUI, Dall-E Mini, and LLaMA-Factory.
How did we get here? How did Gradio keep growing in the very crowded field of open-source Python libraries? I get this question a lot from folks who are building their own open-source libraries. This post distills some of the lessons that I have learned over the past few years:
1. Invest in good primitives, not high-level abstractions
2. Embed virality directly into your library
3. Focus on a (growing) niche
4. Your only roadmap should be rapid iteration
5. Maximize ways users can consume your library's outputs
1. Invest in good primitives, not high-level abstractions
When we first launched Gradio, we offered only one high-level class (gr.Interface), which created a complete web app from a single Python function. We quickly realized that developers wanted to create other kinds of apps (e.g. multi-step workflows, chatbots, streaming applications), but as we started listing out the apps users wanted to build, we realized what we needed to do:
Read the rest here: https://x.com/abidlabs/status/1907886

freddyaboultonย
posted
an
update
3 months ago
Post
2143
Ever wanted to share your AI creations with friends? โจ
Screenshots are fine, but imagine letting others play with your ACTUAL model!
Introducing Gradio deep links ๐ - now you can share interactive AI apps, not just images.
Add a gr.DeepLinkButton to any app and get shareable URLs that let ANYONE experiment with your models.
Screenshots are fine, but imagine letting others play with your ACTUAL model!
Introducing Gradio deep links ๐ - now you can share interactive AI apps, not just images.
Add a gr.DeepLinkButton to any app and get shareable URLs that let ANYONE experiment with your models.

freddyaboultonย
posted
an
update
4 months ago
Post
2005
Privacy matters when talking to AI! ๐
We've just added a microphone mute button to FastRTC in our latest update (v0.0.14). Now you control exactly what your LLM hears.
Plus lots more features in this release! Check them out:
https://github.com/freddyaboulton/fastrtc/releases/tag/0.0.14
We've just added a microphone mute button to FastRTC in our latest update (v0.0.14). Now you control exactly what your LLM hears.
Plus lots more features in this release! Check them out:
https://github.com/freddyaboulton/fastrtc/releases/tag/0.0.14

freddyaboultonย
posted
an
update
4 months ago
Post
3316
Getting WebRTC and Websockets right in python is very tricky. If you've tried to wrap an LLM in a real-time audio layer then you know what I'm talking about.
That's where FastRTC comes in! It makes WebRTC and Websocket streams super easy with minimal code and overhead.
Check out our org: hf.co/fastrtc
That's where FastRTC comes in! It makes WebRTC and Websocket streams super easy with minimal code and overhead.
Check out our org: hf.co/fastrtc
Post
14882
I was thinking i need to step up my game on training Flux LoRas models, time to have some fun ! โ๏ธ
Expect a new drop per week on aesthetics that catched my attention, here are 3 of them that worked really well !
fffiloni/cute-comic-800
fffiloni/carbo-800
fffiloni/oniric-750
Expect a new drop per week on aesthetics that catched my attention, here are 3 of them that worked really well !
fffiloni/cute-comic-800
fffiloni/carbo-800
fffiloni/oniric-750
Post
3578
Explain like i'm 5 the last take from
@thomwolf
on X about Dario's essay on DeepSeek:
โโบ Open-source AI is like a big cookbook that everyone can read and improve. Instead of a few chefs keeping their recipes secret, anyone can cook, test, and invent new things.
If only one company controls AI, everything stops if they have a problemโlike when the internet goes down. With open-source, many people can help, making sure it keeps running smoothly.
AI isnโt just a race between two countries; itโs a team effort around the world. By sharing, we move faster and create safer technology for everyone.
โ
๐ค
โโบ Open-source AI is like a big cookbook that everyone can read and improve. Instead of a few chefs keeping their recipes secret, anyone can cook, test, and invent new things.
If only one company controls AI, everything stops if they have a problemโlike when the internet goes down. With open-source, many people can help, making sure it keeps running smoothly.
AI isnโt just a race between two countries; itโs a team effort around the world. By sharing, we move faster and create safer technology for everyone.
โ
๐ค
Post
21875
Google drops Gemini 2.0 Flash Thinking
a new experimental model that unlocks stronger reasoning capabilities and shows its thoughts. The model plans (with thoughts visible), can solve complex problems with Flash speeds, and more
now available in anychat, try it out: https://huggingface.co/spaces/akhaliq/anychat
a new experimental model that unlocks stronger reasoning capabilities and shows its thoughts. The model plans (with thoughts visible), can solve complex problems with Flash speeds, and more
now available in anychat, try it out: https://huggingface.co/spaces/akhaliq/anychat

freddyaboultonย
posted
an
update
6 months ago
Post
1839
Just created a Gradio space for playing with the new OAI realtime voice API!
freddyaboulton/openai-realtime-voice
freddyaboulton/openai-realtime-voice

freddyaboultonย
posted
an
update
6 months ago
Post
1024
Gemini can talk ๐ฃ๏ธ
Check out the new multimodal API from Google on @akhaliq 's anychat or my space. It's very fast and smart ๐
https://huggingface.co/spaces/freddyaboulton/gemini-voicehttps://huggingface.co/spaces/akhaliq/anychat
Check out the new multimodal API from Google on @akhaliq 's anychat or my space. It's very fast and smart ๐
https://huggingface.co/spaces/freddyaboulton/gemini-voicehttps://huggingface.co/spaces/akhaliq/anychat

freddyaboultonย
posted
an
update
7 months ago
Post
2625
Version 0.0.21 of gradio-pdf now properly loads chinese characters!

freddyaboultonย
posted
an
update
7 months ago
Post
1664
Hello Llama 3.2! ๐ฃ๏ธ๐ฆ
Build a Siri-like coding assistant that responds to "Hello Llama" in 100 lines of python! All with Gradio, webRTC ๐
freddyaboulton/hey-llama-code-editor
Build a Siri-like coding assistant that responds to "Hello Llama" in 100 lines of python! All with Gradio, webRTC ๐
freddyaboulton/hey-llama-code-editor

freddyaboultonย
posted
an
update
7 months ago
Post
1206
Just created a cookbook of real time audio/video spaces created using Gradio and WebRTC โก๏ธ
Use this and the [docs](https://freddyaboulton.github.io/gradio-webrtc/) to get started building the next gen of AI apps!
freddyaboulton/gradio-webrtc-cookbook-6758ba7745aeca7b1be7de0f
Use this and the [docs](https://freddyaboulton.github.io/gradio-webrtc/) to get started building the next gen of AI apps!
freddyaboulton/gradio-webrtc-cookbook-6758ba7745aeca7b1be7de0f
Post
21580
QwQ-32B-Preview is now available in anychat
A reasoning model that is competitive with OpenAI o1-mini and o1-preview
try it out: https://huggingface.co/spaces/akhaliq/anychat
A reasoning model that is competitive with OpenAI o1-mini and o1-preview
try it out: https://huggingface.co/spaces/akhaliq/anychat