Skip to content

πŸ” LLM orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.

License

Notifications You must be signed in to change notification settings

viveksilimkhan1/haystack

Β 
Β 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Haystack
CI/CD Tests Docker image release Schemas code style - Black types - Mypy Coverage Status
Docs Sync docs with Readme Website
Package PyPI PyPI - Downloads PyPI - Python Version GitHub License Compliance
Meta Discord Twitter Follow

Haystack is an end-to-end NLP framework that enables you to build applications powered by LLMs, Transformer models, vector search and more. Whether you want to perform question answering, answer generation, semantic document search, or build tools that are capable of complex decision-making and query resolution, you can use the state-of-the-art NLP models with Haystack to build end-to-end NLP applications solving your use case.

Quickstart

Haystack is built around the concept of pipelines. A pipeline is a powerful structure that performs an NLP task. It's made up of components connected together. For example, you can connect a Retriever and a PromptNode to build a Generative Question Answering pipeline that uses your own data.

Try out how Haystack answers questions about Game of Thrones using the Retrieval Augmented Generation (RAG) approach πŸ‘‡

First, run the minimal Haystack installation:

pip install farm-haystack

Then, index your data to the DocumentStore, build a RAG pipeline, and ask a question on your data:

from haystack.document_stores import InMemoryDocumentStore
from haystack.utils import build_pipeline, add_example_data, print_answers

# We are model agnostic :) Here, you can choose from: "anthropic", "cohere", "huggingface", and "openai".
provider = "openai"
API_KEY = "sk-..." # ADD YOUR KEY HERE

# We support many different databases. Here, we load a simple and lightweight in-memory database.
document_store = InMemoryDocumentStore(use_bm25=True)

# Download and add Game of Thrones TXT articles to Haystack DocumentStore.
# You can also provide a folder with your local documents.
add_example_data(document_store, "data/GoT_getting_started")

# Build a pipeline with a Retriever to get relevant documents to the query and a PromptNode interacting with LLMs using a custom prompt.
pipeline = build_pipeline(provider, API_KEY, document_store)

# Ask a question on the data you just added.
result = pipeline.run(query="Who is the father of Arya Stark?")

# For details, like which documents were used to generate the answer, look into the <result> object
print_answers(result, details="medium")

The output of the pipeline will reference the documents used to generate the answer:

'Query: Who is the father of Arya Stark?'
'Answers:'
[{'answer': 'The father of Arya Stark is Lord Eddard Stark of '
                'Winterfell. [Document 1, Document 4, Document 5]'}]

Congratulations, you have just built your first Haystack app!

Core Concepts

πŸƒβ€β™€οΈ Pipelines: This is the standard Haystack structure that builds on top of your data to perform various NLP tasks such as retrieval augmented generation, question answering and more. The data in a Pipeline flows from one Node to the next. You define how Nodes interact with each other, and how one Node pushes data to the next.

An example pipeline would consist of one Retriever Node and one PromptNode. When the pipeline runs with a query, the Retriever first retrieves the relevant context to the query from your data, and then the PromptNode uses an LLM to generate the final answer.

βš›οΈ Nodes: Each Node achieves one thing. Such as preprocessing documents, retrieving documents, using language models to answer questions and so on.

πŸ•΅οΈ Agent: (since 1.15) An Agent is a component that is powered by an LLM, such as GPT-3. It can decide on the next best course of action so as to get to the result of a query. It uses the Tools available to it to achieve this. While a pipeline has a clear start and end, an Agent is able to decide whether the query has been resolved or not. It may also make use of a Pipeline as a Tool.

πŸ› οΈ Tools: You can think of a Tool as an expert, that is able to do something really well. Such as a calculator, good at mathematics. Or a WebRetriever, good at retrieving pages from the internet. A Node or pipeline in Haystack can also be used as a Tool. A Tool is a component that is used by an Agent, to resolve complex queries.

πŸ—‚οΈ DocumentStores: A DocumentStore is database where you store your text data for Haystack to access. Haystack DocumentStores are available with ElasticSearch, Opensearch, Weaviate, Pinecone, FAISS and more. For a full list of available DocumentStores, check out our documentation.

What to Build with Haystack

  • Build retrieval augmented generation (RAG) by making use of one of the available vector databases and customizing your LLM interaction, the sky is the limit πŸš€
  • Perform Question Answering in natural language to find granular answers in your documents.
  • Perform semantic search and retrieve documents according to meaning.
  • Build applications that can make complex decisions making to answer complex queries: such as systems that can resolve complex customer queries, do knowledge search on many disconnected resources and so on.
  • Use off-the-shelf models or fine-tune them to your data.
  • Use user feedback to evaluate, benchmark, and continuously improve your models.

Features

  • Latest models: Haystack allows you to use and compare models available from OpenAI, Cohere and Hugging Face, as well as your own local models or models hosted on SageMaker. Use the latest LLMs or Transformer-based models (for example: BERT, RoBERTa, MiniLM).
  • Modular: Multiple choices to fit your tech stack and use case. A wide choice of DocumentStores to store your data, file conversion tools and more
  • Open: Integrated with Hugging Face's model hub, OpenAI, Cohere and various Azure services.
  • Scalable: Scale to millions of docs using retrievers and production-scale components like Elasticsearch and a fastAPI REST API.
  • End-to-End: All tooling in one place: file conversion, cleaning, splitting, training, eval, inference, labeling, and more.
  • Customizable: Fine-tune models to your domain or implement your custom Nodes.
  • Continuous Learning: Collect new training data from user feedback in production & improve your models continuously.

Resources

πŸ“’ Docs Components, Pipeline Nodes, Guides, API Reference
πŸ’Ύ Installation How to install Haystack
πŸŽ“ Tutorials See what Haystack can do with our Notebooks & Scripts
πŸŽ‰Β Haystack Extras A repository that lists extra Haystack packages and components that can be installed separately.
πŸ”° Demos A repository containing Haystack demo applications with Docker Compose and a REST API
πŸ–– Community Discord, Twitter, Stack Overflow, GitHub Discussions
πŸ’™ Contributing We welcome all contributions!
πŸ“Š Benchmarks Speed & Accuracy of Retriever, Readers and DocumentStores
πŸ”­ Roadmap Public roadmap of Haystack
πŸ“° Blog Learn about the latest with Haystack and NLP
☎️ Jobs We're hiring! Have a look at our open positions

πŸ’Ύ Installation

For a detailed installation guide see the official documentation. There you’ll find instructions for custom installations handling Windows and Apple Silicon.

Basic Installation

Use pip to install a basic version of Haystack's latest release:

pip install farm-haystack

This command installs everything needed for basic Pipelines that use an in-memory DocumentStore and external LLM provider (e.g. OpenAI).

Full Installation

To use more advanced features, like certain DocumentStores, inference with local transformer models, FileConverters, OCR, or Ray, you need to install further dependencies. The following command installs the latest release of Haystack and all its dependencies:

pip install 'farm-haystack[all]' ## or 'all-gpu' for the GPU-enabled dependencies

If you want to install only the dependencies needed for model inference on your local hardware (not remote API endpoints), such as torch and sentence-transformers, you can use the following command:

pip install 'farm-haystack[inference]' ## installs torch, sentence-transformers, sentencepiece, and huggingface-hub

If you want to try out the newest features that are not in an official release yet, you can install the unstable version from the main branch with the following command:

pip install git+https://github.com/deepset-ai/haystack.git@main#egg=farm-haystack

To be able to make changes to Haystack code, first of all clone this repo:

git clone https://github.com/deepset-ai/haystack.git

Then move into the cloned folder and install the project with pip, including the development dependencies:

cd haystack && pip install -e '.[dev]'

If you want to contribute to the Haystack repo, check our Contributor Guidelines first.

See the list of dependencies to check which ones you want to install (for example, [all], [dev], or other).

Installing the REST API

Haystack comes packaged with a REST API so that you can deploy it as a service. Run the following command from the root directory of the Haystack repo to install REST_API:

pip install rest_api/

You can find out more about our PyPi package on our PyPi page.

πŸ”°Demos

You can find some of our hosted demos with instructions to run them locally too on our haystack-demos repository

πŸ’« Reduce Hallucinations with Retrieval Augmentation - Generative QA with LLMs

πŸ₯ Should I follow? - Summarizing tweets with LLMs

🌎 Explore The World - Extractive Question Answering

πŸ–– Community

If you have a feature request or a bug report, feel free to open an issue in Github. We regularly check these and you can expect a quick response. If you'd like to discuss a topic, or get more general advice on how to make Haystack work for your project, you can start a thread in Github Discussions or our Discord channel. We also check Twitter and Stack Overflow.

πŸ’™ Contributing

We are very open to the community's contributions - be it a quick fix of a typo, or a completely new feature! You don't need to be a Haystack expert to provide meaningful improvements. To learn how to get started, check out our Contributor Guidelines first.

Who Uses Haystack

Here's a list of projects and companies using Haystack. Want to add yours? Open a PR, add it to the list and let the world know that you use Haystack!

About

πŸ” LLM orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 98.7%
  • HTML 1.2%
  • HCL 0.1%