A Python package that makes it easy for developers to create machine learning apps powered by OpenAI, Google's Gemini models, Anthropic's Claude, LumaAI, CrewAI, and XAI's Grok.
You can install ai-gradio
with different providers:
# Install with OpenAI support
pip install 'ai-gradio[openai]'
# Install with Gemini support
pip install 'ai-gradio[gemini]'
# Install with CrewAI support
pip install 'ai-gradio[crewai]'
# Install with Anthropic support
pip install 'ai-gradio[anthropic]'
# Install with LumaAI support
pip install 'ai-gradio[lumaai]'
# Install with XAI support
pip install 'ai-gradio[xai]'
# Install with Cohere support
pip install 'ai-gradio[cohere]'
# Install with SambaNova support
pip install 'ai-gradio[sambanova]'
# Install with all providers
pip install 'ai-gradio[all]'
# Installation additions:
pip install 'ai-gradio[fireworks]'
pip install 'ai-gradio[together]'
pip install 'ai-gradio[qwen]'
pip install 'ai-gradio[hyperbolic]'
# Install with DeepSeek support
pip install 'ai-gradio[deepseek]'
First, set your API key in the environment:
For OpenAI:
export OPENAI_API_KEY=<your token>
For Gemini:
export GEMINI_API_KEY=<your token>
For Anthropic:
export ANTHROPIC_API_KEY=<your token>
For LumaAI:
export LUMAAI_API_KEY=<your token>
For XAI:
export XAI_API_KEY=<your token>
For Cohere:
export COHERE_API_KEY=<your token>
For SambaNova:
export SAMBANOVA_API_KEY=<your token>
For DeepSeek:
export DEEPSEEK_API_KEY=<your token>
Then in a Python file:
import gradio as gr
from ai_gradio import registry
# Create a Gradio interface
interface = gr.load(
name='gpt-4-turbo', # or 'gemini-pro' for Gemini, or 'xai:grok-beta' for Grok
src=registry,
title='AI Chat',
description='Chat with an AI model'
).launch()
Basic text chat is supported for all text models. The interface provides a chat-like experience where you can have conversations with the AI model.
Voice chat is supported for OpenAI realtime models. You can enable it in two ways:
# Using a realtime model
interface = gr.load(
name='gpt-4o-realtime-preview-2024-10-01',
src=registry
).launch()
# Or explicitly enabling voice chat with any realtime model
interface = gr.load(
name='gpt-4o-mini-realtime-preview-2024-12-17',
src=registry,
enable_voice=True
).launch()
For voice chat functionality, you'll need:
- OpenAI API key (required):
export OPENAI_API_KEY=<your OpenAI token>
- Twilio credentials (recommended for better WebRTC performance):
export TWILIO_ACCOUNT_SID=<your Twilio account SID>
export TWILIO_AUTH_TOKEN=<your Twilio auth token>
You can get Twilio credentials by:
- Creating a free account at Twilio
- Finding your Account SID and Auth Token in the Twilio Console
Without Twilio credentials, voice chat will still work but might have connectivity issues in some network environments.
Video chat is supported for Gemini models. You can enable it by setting enable_video=True
:
interface = gr.load(
name='gemini-pro',
src=registry,
enable_video=True
).launch()
DeepSeek models support text generation and coding assistance:
interface = gr.load(
name='deepseek:deepseek-chat',
src=registry,
title='DeepSeek Chat',
description='Chat with DeepSeek'
).launch()
# For code assistance
interface = gr.load(
name='deepseek:deepseek-coder',
src=registry,
title='DeepSeek Coder',
description='Get coding help from DeepSeek'
).launch()
# For vision tasks
interface = gr.load(
name='deepseek:deepseek-vision',
src=registry,
title='DeepSeek Vision',
description='Visual understanding with DeepSeek'
).launch()
Anthropic's Claude models are supported for text generation:
interface = gr.load(
name='anthropic:claude-3-opus-20240229',
src=registry,
title='Claude Chat',
description='Chat with Claude'
).launch()
LumaAI support allows you to generate videos and images from text prompts:
# For video generation
interface = gr.load(
name='lumaai:dream-machine',
src=registry,
title='LumaAI Video Generation'
).launch()
# For image generation
interface = gr.load(
name='lumaai:photon-1',
src=registry,
title='LumaAI Image Generation'
).launch()
CrewAI support allows you to create teams of AI agents that work together to solve complex tasks. Enable it by using the CrewAI provider:
interface = gr.load(
name='crewai:gpt-4-turbo',
src=registry,
title='AI Team Chat',
description='Chat with a team of specialized AI agents'
).launch()
The CrewAI integration supports different specialized agent teams:
-
support
: A team of support agents that help answer questions, including:- Senior Support Representative
- Support Quality Assurance Specialist
-
article
: A team of content creation agents, including:- Content Planner
- Content Writer
- Editor
You can specify the crew type when creating the interface:
interface = gr.load(
name='crewai:gpt-4-turbo',
src=registry,
crew_type='article', # or 'support'
title='AI Writing Team',
description='Create articles with a team of AI agents'
).launch()
When using the support
crew type, you can provide a documentation URL that the agents will reference when answering questions. The interface will automatically show a URL input field.
When loading a model, you can specify the provider explicitly using the format provider:model_name
.
# Explicit provider
interface = gr.load(
name='gemini:gemini-pro',
src=registry
).launch()
You can customize the interface by adding examples, changing the title, or adding a description:
interface = gr.load(
name='gpt-4-turbo',
src=registry,
title='Custom AI Chat',
description='Chat with an AI assistant',
examples=[
"Explain quantum computing to a 5-year old",
"What's the difference between machine learning and AI?"
]
).launch()
You can combine multiple models in a single interface using Gradio's Blocks:
import gradio as gr
from ai_gradio import registry
with gr.Blocks() as demo:
with gr.Tab("GPT-4"):
gr.load('gpt-4-turbo', src=registry)
with gr.Tab("Gemini"):
gr.load('gemini-pro', src=registry)
with gr.Tab("Claude"):
gr.load('anthropic:claude-3-opus-20240229', src=registry)
with gr.Tab("LumaAI"):
gr.load('lumaai:dream-machine', src=registry)
with gr.Tab("CrewAI"):
gr.load('crewai:gpt-4-turbo', src=registry)
with gr.Tab("Grok"):
gr.load('xai:grok-beta', src=registry)
demo.launch()
- gpt-4-turbo
- gpt-4
- gpt-3.5-turbo
- gemini-pro
- gemini-pro-vision
- gemini-2.0-flash-exp
- claude-3-opus-20240229
- claude-3-sonnet-20240229
- claude-3-haiku-20240307
- claude-2.1
- claude-2.0
- claude-instant-1.2
- dream-machine (video generation)
- photon-1 (image generation)
- photon-flash-1 (fast image generation)
- crewai:gpt-4-turbo
- crewai:gpt-4
- crewai:gpt-3.5-turbo
- grok-beta
- grok-vision-beta
- command
- command-light
- command-nightly
- command-r
- llama2-70b-chat
- llama2-13b-chat
- llama2-7b-chat
- mixtral-8x7b-chat
- mistral-7b-chat
- whisper-v3
- whisper-v3-turbo
- f1-preview
- f1-mini
- meta-llama/Llama-Vision-Free
- meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo
- meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo
- meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo
- meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo
- meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo
- meta-llama/Meta-Llama-3-8B-Instruct-Turbo
- meta-llama/Meta-Llama-3-70B-Instruct-Turbo
- meta-llama/Llama-3.2-3B-Instruct-Turbo
- meta-llama/Meta-Llama-3-8B-Instruct-Lite
- meta-llama/Meta-Llama-3-70B-Instruct-Lite
- qwen-turbo-latest
- qwen-turbo
- qwen-plus
- qwen-max
- qwen1.5-110b-chat
- qwen1.5-72b-chat
- qwen1.5-32b-chat
- qwen1.5-14b-chat
- qwen1.5-7b-chat
- qwq-32b-preview
- qvq-72b-preview
- meta-llama/llama-3.3-70b
- Qwen/QwQ-32B-Preview
- Qwen/qwen2.5-coder-32b
- meta-llama/llama-3.2-3b
- Qwen/qwen2.5-72b
- deepseek/deepseek-v2.5
- meta-llama/llama-3-70b
- hermes/hermes-3-70b
- meta-llama/llama-3.1-405b
- meta-llama/llama-3.1-70b
- meta-llama/llama-3.1-8b
- deepseek-chat
- deepseek-coder
- deepseek-vision
- Python 3.10 or higher
- gradio >= 5.9.1
Additional dependencies are installed based on your chosen provider:
- OpenAI:
openai>=1.58.1
- Gemini:
google-generativeai
- CrewAI:
crewai>=0.1.0
,langchain>=0.1.0
,langchain-openai>=0.0.2
,crewai-tools>=0.0.1
- Anthropic:
anthropic>=1.0.0
- LumaAI:
lumaai>=0.0.3
- XAI:
xai>=0.1.0
- Cohere:
cohere>=5.0.0
- DeepSeek:
openai>=1.58.1
If you get a 401 authentication error, make sure your API key is properly set. You can set it manually in your Python session:
import os
# For OpenAI
os.environ["OPENAI_API_KEY"] = "your-api-key"
# For Gemini
os.environ["GEMINI_API_KEY"] = "your-api-key"
# For Anthropic
os.environ["ANTHROPIC_API_KEY"] = "your-api-key"
# For LumaAI
os.environ["LUMAAI_API_KEY"] = "your-api-key"
# For XAI
os.environ["XAI_API_KEY"] = "your-api-key"
# For Cohere
os.environ["COHERE_API_KEY"] = "your-api-key"
# For SambaNova
os.environ["SAMBANOVA_API_KEY"] = "your-api-key"
# Environment variables additions:
export FIREWORKS_API_KEY=<your token>
export TOGETHER_API_KEY=<your token>
export QWEN_API_KEY=<your token>
export HYPERBOLIC_API_KEY=<your token>
# Additional troubleshooting environment variables:
os.environ["FIREWORKS_API_KEY"] = "your-api-key"
os.environ["TOGETHER_API_KEY"] = "your-api-key"
os.environ["QWEN_API_KEY"] = "your-api-key"
os.environ["HYPERBOLIC_API_KEY"] = "your-api-key"
os.environ["DEEPSEEK_API_KEY"] = "your-api-key"
### No Providers Error
If you see an error about no providers being installed, make sure you've installed the package with the desired provider:
```bash
# Install with OpenAI support
pip install 'ai-gradio[openai]'
# Install with Gemini support
pip install 'ai-gradio[gemini]'
# Install with CrewAI support
pip install 'ai-gradio[crewai]'
# Install with Anthropic support
pip install 'ai-gradio[anthropic]'
# Install with LumaAI support
pip install 'ai-gradio[lumaai]'
# Install with XAI support
pip install 'ai-gradio[xai]'
# Install with Cohere support
pip install 'ai-gradio[cohere]'
# Install all providers
pip install 'ai-gradio[all]'
For voice chat functionality:
- gradio-webrtc
- numba==0.60.0
- pydub
- librosa
- websockets
- twilio
- gradio_webrtc[vad]
- numpy
For video chat functionality:
- opencv-python
- Pillow