Skip to content

Latest commit

 

History

History
73 lines (54 loc) · 2.69 KB

README.md

File metadata and controls

73 lines (54 loc) · 2.69 KB
 _    _  _____  _     ______  _____   ___   _      _   __
| |  | ||_   _|| |    |  _  \|_   _| / _ \ | |    | | / /
| |  | |  | |  | |    | | | |  | |  / /_\ \| |    | |/ / 
| |/\| |  | |  | |    | | | |  | |  |  _  || |    |    \ 
\  /\  / _| |_ | |____| |/ /   | |  | | | || |____| |\  \
 \/  \/  \___/ \_____/|___/    \_/  \_| |_/\_____/\_| \_/

WILDTALK

The wildtalk module module generates responses for given input using the given model.

In order to change the model used for generation, change this option:

Option Description
model one of "convai_gpt", "gpt2", "parlai"

If no separate wildtalk server is running, this module starts a wildtalk server which can be used by multiple ravestate instances.

Available Models

ConvAI GPT

This model is based on transfer-learning-conv-ai

It can be used by setting model to "convai_gpt"

Option Description
temperature higher value -> more variation in output
max_length maximal length of generated output
top_k <=0: no filtering, >0: keep only top k tokens with highest probability.
top_p <=0.0 no filtering, >0.0: keep smallest subset whose total probability mass >= top_p
max_history maximal number of previous dialog turns to be used for output generation

GPT2

This model uses the gpt2 transformer and model from pytorch_transformers. It uses the gpt2-medium dataset.

It can be used by setting model to "gpt2"

Option Description
temperature higher value -> more variation in output
max_length maximal length of generated output
top_k <=0: no filtering, >0: keep only top k tokens with highest probability.

Parlai

This model uses roboy-parlai.

It can be used by setting model to "parlai"

Option Description
temperature higher value -> more variation in output
max_length maximal length of generated output
top_k <=0: no filtering, >0: keep only top k tokens with highest probability.

Separate Server

A separate server can be used for running the wildtalk generation. This can be configured with the following options:

Option Description
server_address Address under which the server is accessible
server_port Port of the server

To start a standalone server for wildtalk generation, execute this in the /modules folder of ravestate:

python -c "from ravestate_wildtalk import server; server.run(port=<PORTNUMBER>, model=<MODELNAME>)

Note that the model used on server startup will be the model used when accessed from ravestate.