The simplest and most intuitive open-source implementation of an open source perplexity.ai, a large language model(LLM) service which cites information from Google. No fancy GUI or LLM agents are involved, just 200 lines of python code. Please check out this video for more explanations!
Check out the conversations nanoPerplexityAI has generated
- Get the user query
- LLM checks the user query, decides whether to execute a Google search, and if searching, reformulates the user query into a Google-suited query to find relevant webpage URLs and fetch texts. (In practice, PerplexityAI searches its already indexed sources)
- Build a prompt using
system prompt + webpage context + user query
- Call the LLM API to generate an answer
- As LLM perform stream completion, save the LLM response into a markdown file for better visualization.
#PerplexityAI does not reformat the search results and therefore not all search results are used and cited in the LLM response. This is because they prioritize displaying search results quickly and streaming LLM completion for a better user experience.
pip install googlesearch-python requests beautifulsoup4 lxml backoff openai
export OPENAI_API_KEY=<Your OpenAI API KEY>
python nanoPerplexityAI.py
The script will prompt you to type your question, then it will generate an answer in playground.md
You can type a key s for [s]ave and q for [q]uit
You can utilise Visual Studio Code to replicate the simpler version of PerplexityAI GUI. Open Preview of playground.md
as you run python nanoPerplexityAI.py
and you will see the real time generation!
Other ways involve opening in Markdown Playground or pushing the output markdown files to your github repo for displaying markdown
Thank you perplexity.ai for the amazing idea and clarity-ai and Perplexica for coding inspirations on the open-source implementation of perplexity.ai.