Helicone is an open-source observability platform for Language Learning Models (LLMs). It offers the following features:
-
๐ Logs all of your requests to OpenAI in a user-friendly UI
-
๐พ Caching, custom rate limits, and retries
-
๐ Track costs and latencies by users and custom properties
-
๐ฎ Every log is a playground: iterate on prompts and chat conversations in a UI
-
๐ Share results and collaborate with your friends or teammates
-
๐ (Coming soon) APIs to log feedback and evaluate results
Get your API key by signing up here.
export HELICONE_API_KEY=<your API key>
pip install helicone
from helicone.openai_proxy import openai
response = openai.Completion.create(
model="text-davinci-003",
prompt="What is Helicone?",
user="alice@bob.com",
# Optional Helicone features:
cache=True,
properties={"conversation_id": 12},
rate_limit_policy={"quota": 100, "time_window": 60, "segment": "user"}
)
๐ Then view your logs at Helicone.
Helicone's cloud offering is deployed on Cloudflare and ensures the lowest latency add-on to your API requests.
To get started locally, Helicone is comprised of four services:
- Frontend (Node)
- The proxy worker (Wrangler)
- Application database (Supabase)
- Analytics database (ClickHouse)
If you have any questions, contact help@helicone.ai or join discord.
nvm install 18.11.0
nvm use 18.11.0
npm install -g wrangler
npm install -g yarn
Install Supabase
brew install supabase/tap/supabase
# This will start clickhouse locally
python3 clickhouse/ch_hcone.py --start
cd web
# start supabase to log all the db stuff...
supabase start
# start frontend
yarn
yarn dev
# start worker (simulates oai.hconeai.com)
# in another terminal
cd worker
yarn
wrangler dev --local
# Make your request to local host
curl --request POST \
--url http://127.0.0.1:8787/v1/chat/completions \
--header 'Authorization: Bearer <KEY>' \
--data '{
"model": "gpt-3.5-turbo",
"messages": [
{
"role": "user",
"content": "Can you give me a random number?"
}
],
"temperature": 1,
"max_tokens": 7
}'
# Now you can go to localhost:3000 and create an account and see your request.
# When creating an account on localhost, you will automatically be signed in.
Make sure your .env file is in web/.env
. Here is an example:
NEXT_PUBLIC_STRIPE_PUBLISHABLE_KEY=""
STRIPE_SECRET_KEY=""
NEXT_PUBLIC_HELICONE_BILLING_PORTAL_LINK=""
NEXT_PUBLIC_HELICONE_CONTACT_LINK="https://calendly.com/d/x5d-9q9-v7x/helicone-discovery-call"
STRIPE_PRICE_ID=""
STRIPE_STARTER_PRICE_ID=""
STRIPE_ENTERPRISE_PRODUCT_ID=""
STRIPE_STARTER_PRODUCT_ID=""
DATABASE_URL="postgresql://postgres:postgres@localhost:54322/postgres"
NEXT_PUBLIC_SUPABASE_ANON_KEY="eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJzdXBhYmFzZS1kZW1vIiwicm9sZSI6ImFub24iLCJleHAiOjE5ODM4MTI5OTZ9.CRXP1A7WOeoJeXxjNni43kdQwgnWNReilDMblYTn_I0"
NEXT_PUBLIC_SUPABASE_URL="http://localhost:54321"
SUPABASE_URL="http://localhost:54321"
SUPABASE_SERVICE_KEY="eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJzdXBhYmFzZS1kZW1vIiwicm9sZSI6InNlcnZpY2Vfcm9sZSIsImV4cCI6MTk4MzgxMjk5Nn0.EGIM96RAZx35lJzdJsyH-qQwv8Hdp7fsn3W0YpN81IU"
Name | Docs |
---|---|
nextjs-chat-app | Docs |
langchain | Docs |
langchainjs | Docs |
ModelFusion | Docs |
We are extremely open to contributors on documentation, integrations, and feature requests.
- If you have an idea for how Helicone can be better, create a GitHub issue or vote on the roadmap
- Join discord to ask questions
Helicone is licensed under the MIT License.