Skip to main content

A hybrid chatbot.

Project description

WAFL 0.0.80 TestsDocs

Introduction

Logo

WAFL is a framework for home assistants. It is designed to combine Large Language Models and rules to create a predictable behavior. Specifically, instead of organising the work of an LLM into a chain of thoughts, WAFL intends to organise its behavior into inference trees.

WAFL is a work in progress. The current version requires the user to specify the rules to follow. While it is ready to play with, it might not ready for production depending on your use-case.

Installation

In this version, WAFL is a two-part system. Both can be installed on the same machine.

The two parts of WAFL

Interface side

The first part is local to your machine and needs to have access to a microphone and speaker. To install it, run the following commands:

$ sudo apt-get install portaudio19-dev ffmpeg
$ pip install wafl

After installing the requirements, you can initialize the interface by running the following command:

$ wafl init

which creates a config.json file that you can edit to change the default settings. A standard rule file is also created as wafl.rules. Please see the examples in the following chapters.

LLM side (needs a GPU)

The second part (LLM side) is a model server for the speech-to-text model, the LLM, the embedding system, and the text-to-speech model.

Installation

In order to quickly run the LLM side, you can use the following installation commands:

pip install wafl-llm
wafl-llm start

which will use the default models and start the server on port 8080.

Docker

A docker image can be used to run it as in the following:

$ docker run -p8080:8080 --env NVIDIA_DISABLE_REQUIRE=1 --gpus all fractalego/wafl-llm:0.80

The interface side has a config.json file that needs to be filled with the IP address of the LLM side. The default is localhost. Alternatively, you can run the LLM side by cloning this repository.

Running WAFL

This document contains a few examples of how to use the wafl CLI. There are four modes in which to run the system

$ wafl run-audio

This is the main mode of operation. It will run the system in a loop, waiting for the user to speak a command. The activation word is the name defined in config.json. The default name is "computer", but you can change it to whatever you want.

$ wafl run-server

It runs a local web server that listens for HTTP requests on port 8889. The server will act as a chatbot, executing commands and returning the result as defined in the rules.

$ wafl run-cli

This command works as for the run-server command, but it will listen for commands on the command line. It does not run a webserver and is useful for testing purposes.

$ wafl run-tests

This command will run all the tests defined in the file testcases.txt.

Documentation

The documentation can be found at wafl.readthedocs.io.

Mastodon

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

wafl-0.0.83-py3-none-any.whl (319.3 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page