The blazing-fast AI inference client for Python
Built with Rust for performance. Unified interface for OpenAI, Anthropic, Gemini, Groq, and more.
Same simple API, multiple AI providers
Most popular AI model
1from bhumi.base_client import BaseLLMClient, LLMConfig
2import asyncio
3
4async def main():
5 config = LLMConfig(
6 api_key="YOUR_OPENAI_API_KEY",
7 model="openai/gpt-4o",
8 max_tokens=1000
9 )
10
11 client = BaseLLMClient(config)
12
13 response = await client.completion([
14 {"role": "user", "content": "Explain quantum computing"}
15 ])
16
17 print(response['text'])
18
19asyncio.run(main())
Get started with your first AI call in seconds
Simple async client usage
1from bhumi.base_client import BaseLLMClient, LLMConfig
2import asyncio
3
4async def main():
5 config = LLMConfig(
6 api_key="YOUR_OPENAI_API_KEY",
7 model="openai/gpt-4o",
8 max_tokens=1000
9 )
10
11 client = BaseLLMClient(config)
12
13 response = await client.completion([
14 {"role": "user", "content": "Hello, world!"}
15 ])
16
17 print(response['text'])
18
19asyncio.run(main())
Benchmarked against popular alternatives
Built for developers who need reliable, fast AI inference
Rust-powered performance with async Python interface
Unified API for all major AI providers
Built-in rate limiting and error handling
Install Bhumi and make your first AI call in under 2 minutes