Chat with Website Example¶
Build a pipeline that answers questions about one or more websites. Uses fetching, cleaning and an LLM to generate answers, and supports streaming via OpenAI-compatible chat endpoints when implemented in the wrapper.
Where is the code?¶
- Wrapper example directory: examples/pipeline_wrappers/chat_with_website_streaming
- See the main docs for
PipelineWrapperbasics and OpenAI compatibility
Deploy¶
hayhooks pipeline deploy-files -n chat_with_website examples/pipeline_wrappers/chat_with_website_streaming
Run¶
- API mode:
curl -X POST http://localhost:1416/chat_with_website/run \
-H 'Content-Type: application/json' \
-d '{"query": "What is this website about?", "urls": ["https://python.org"]}'
- Chat (OpenAI-compatible), when
run_chat_completion/_asyncis implemented:
curl -X POST http://localhost:1416/v1/chat/completions \
-H 'Content-Type: application/json' \
-d '{
"model": "chat_with_website",
"messages": [{"role": "user", "content": "Tell me about https://github.com"}]
}'
Development Tips
- For development, use
--overwriteto redeploy a changed wrapper:hayhooks pipeline deploy-files -n chat_with_website --overwrite <dir> - Some examples may require extra Python packages (e.g.,
trafilatura). Install as needed.
Related¶
- General guide: Main docs
- Examples index: Examples Overview