merlinn-co / merlinn

Open source AI on-call developer πŸ§™β€β™‚οΈ Get relevant context & root cause analysis in seconds about production incidents and make on-call engineers 10x better 🏎️

Home Page:https://merlinn.co/?utm_source=github

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Merlinn - open-source AI on-call developer


Docs Β· Demo Β· Report Bug Β· Feature Request Β· Blog Β· Slack

Overview πŸ’«

Merlinn is an AI-powered on-call engineer. It can automatically jump into incidents & alerts with you, and provide you useful & contextual insights and RCA in real time.

Why ❓

Most people don't like to do on-call shifts. It requires engineers to be swift and solve problems quickly. Moreover, it takes time to reach to the root cause of the problem. That's why we developed Merlinn. We believe Gen AI can help on-call developrs solve issues faster.

Table of Contents

Key features 🎯

  • Automatic RCA: Merlinn automatically listens to production incidents/alerts and automatically investigates them for you.
  • Slack integration: Merlinn lives inside your Slack. Simply connect it and enjoy an on-call engineer that never sleeps.
  • Integrations: Merlinn integrates with popular observability/incident management tools such as Datadog, Coralogix, Opsgenie and Pagerduty. It also integrates to other tools as GitHub, Notion, Jira and Confluence to gain insights on incidents.
  • Intuitive UX: Merlinn offers a familiar experience. You can talk to it and ask follow-up questions.
  • Secure: Self-host Merlinn and own your data. Always.
  • Open Source: We love open-source. Self-host Merlinn and use it for free.

Demo πŸŽ₯

Checkout our demo video to see Merlinn in action.

Getting started πŸš€

In order to run Merlinn, you need to clone the repo & run the app using Docker Compose.

Prerequisites πŸ“œ

Ensure you have the following installed:

  • Docker & Docker Compose - The app works with Docker containers. To run it, you need to have Docker Desktop, which comes with Docker CLI, Docker Engine and Docker Compose.

Quick installation 🏎️

You can find the installation video here.

  1. Clone the repository:

    git clone git@github.com:merlinn-co/merlinn.git && cd merlinn
  2. Configure LiteLLM Proxy Server:

    We use LiteLLM Proxy Server to interact with 100+ of LLMs in a unified interface (OpenAI interface).

    1. Copy the example files:

      cp config/litellm/.env.example config/litellm/.env
      cp config/litellm/config.example.yaml config/litellm/config.yaml
    2. Define your OpenAI key and place it inside config/litellm/.env as OPENAI_API_KEY. You can get your API key here. Rest assured, you won't be charged unless you use the API. For more details on pricing, check here.

    3. (Optional) Define custom endpoints. If you want to use other vendors (AWS Bedrock, Azure OpenAI, Anthropic, Hugging Face models, etc), checkout LiteLLM Proxy documentation. You simply need to change config/litellm/.env & config/litellm/config.yaml. Checkout the comments there & LiteLLM's documentation. Note You have to use a vendor that supports function calling.

  3. Copy the .env.example file:

    cp .env.example .env
  4. Open the .env file in your favorite editor (vim, vscode, emacs, etc):

    vim .env # or emacs or vscode or nano
  5. Update these variables:

    • SLACK_BOT_TOKEN, SLACK_APP_TOKEN and SLACK_SIGNING_SECRET - These variables are needed in order to talk to Merlinn on Slack. Please follow this guide to create a new Slack app in your organization.

    • (Optional) SMTP_CONNECTION_URL - This variable is needed in order to invite new members to your Merlinn organization via email and allow them to use the bot. It's not mandatory if you just want to test Merlinn and play with it. If you do want to send invites to your team members, you can use a service like SendGrid/Mailgun. Should follow this pattern: smtp://username:password@domain:port.

  6. Launch the project:

    docker compose up -d

That's it. You should be able to visit Merlinn's dashboard in http://localhost:5173. Simply create a user (with the same e-mail as the one in your Slack user) and start to configure your organization. If something does not work for you, please checkout our troubleshooting or reach out to us via our support channels.

The next steps are to configure your organization a bit more (connect incident management tools, build a knowledge base, etc). Head over to the connect & configure section in our docs for more information πŸ’«

Using DockerHub images

If you want, you can pull our Docker images from DockerHub instead of cloning the repo & building from scratch.

In order to do that, follow these steps:

  1. Download configuration files:

    curl https://raw.githubusercontent.com/merlinn-co/merlinn/main/tools/scripts/download_env_files.sh | sh
  2. Follow steps 2 and 5 above to configure LiteLLM Proxy and your .env file respectively. Namely, you'd need to configure your OpenAI key at config/litellm/.env and configure your Slack credentials in the root .env.

  3. Spin up the environment using docker compose:

    curl https://raw.githubusercontent.com/merlinn-co/merlinn/main/tools/scripts/start.sh | sh

That's it πŸ’« You should be able to visit Merlinn's dashboard in http://localhost:5173.

Updating Merlinn πŸ§™β€β™‚οΈ

  1. Pull the latest changes:

    git pull
  2. Rebuild images:

    docker-compose up --build -d

Visualize Knowledge Base πŸ—ΊοΈ

We use ChromaDB as our vector DB. We also use vector admin in order to see the ingested documents. To use vector admin, simply run this command:

docker compose up vector-admin -d

This command starts vector-admin at port 3001. Head over to http://localhost:3001 and configure your local ChromaDB. Note: Since vector-admin runs inside a docker container, in the "host" field make sure to insert http://host.docker.internal:8000 instead of http://localhost:8000. This is because "localhost" doesn't refer to the host inside the container itself.

Moreover, in the "API Header & Key", you'd need to put "X-Chroma-Token" as the header and the value you have inside .env CHROMA_SERVER_AUTHN_CREDENTIALS as the value.

To learn how to use VectorAdmin, visit the docs.

Support and feedback πŸ‘·β€β™€οΈ

In order of preference the best way to communicate with us:

  • GitHub Discussions: Contribute ideas, support requests and report bugs (preferred as there is a static & permanent for other community members).
  • Slack: community support. Click here to join.
  • Privately: contact at support@merlinn.co

Contributing to Merlinn ⛑️

If you're interested in contributing to Merlinn, checkout our CONTRIBUTING.md file πŸ’« πŸ§™β€β™‚οΈ

Troubleshooting βš’οΈ

If you encounter any problems/errors/issues with Merlinn, checkout our troubleshooting guide. We try to update it regularly, and fix some of the urgent problems there as soon as possible.

Moreover, feel free to reach out to us at our support channels.

Telemetry πŸ”’

By default, Merlinn automatically sends basic usage statistics from self-hosted instances to our server via PostHog.

This allows us to:

  • Understand how Merlinn is used so we can improve it.
  • Track overall usage for internal purposes and external reporting, such as for fundraising.

Rest assured, the data collected is not shared with third parties and does not include any sensitive information. We aim to be transparent, and you can review the specific data we collect here.

If you prefer not to participate, you can easily opt-out by setting TELEMETRY_ENABLED=false inside your .env.

License πŸ“ƒ

This project is licensed under the Apache 2.0 license - see the LICENSE file for details

Learn more πŸ”

Check out the official website at https://merlinn.co for more information.

Contributors ✨

Built with ❀️ by Dudu & Topaz

Dudu: Github, Twitter

Topaz: Github, Twitter

About

Open source AI on-call developer πŸ§™β€β™‚οΈ Get relevant context & root cause analysis in seconds about production incidents and make on-call engineers 10x better 🏎️

https://merlinn.co/?utm_source=github

License:Apache License 2.0


Languages

Language:TypeScript 82.0%Language:Python 16.4%Language:Dockerfile 0.5%Language:Shell 0.4%Language:CSS 0.2%Language:JavaScript 0.2%Language:HTML 0.2%Language:HCL 0.0%Language:Jsonnet 0.0%