Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Extension: homestar w/ llama2 build & model + prompt chain component/effect + example #628

Open
1 of 4 tasks
zeeshanlakhani opened this issue Mar 25, 2024 · 0 comments
Open
1 of 4 tasks
Assignees
Labels
enhancement New feature or request

Comments

@zeeshanlakhani
Copy link
Contributor

zeeshanlakhani commented Mar 25, 2024

Summary

Add functionality to Homestar to allow for the user-driven execution of an LLM Chain within a sandboxed environment (Wasm) as a workflow composed of a series of steps as prompts (akin to a series of step functions). The outcome of this feature is for the inference to operate locally on a trained model (e.g. Llama 2, Minstral) privately provided by the host platform it's executing on.

The learning goals of this feature addition are to experiment with working with LLMs locally on hosts where the training data remains private and only computationally derived information can be shared with other users/peers for consumption, allowing for working with AI computation in ways not tied to any specific vendor or large cloud provider. Frankly, this work will showcase everything against what IEEE Spectrum's Open-Source AI Is Uniquely Dangerous article scrutinizes. Incorporating ways for users to chain LLM steps together while controlling what inference gets exhibited without the infrastructure concerns or data risks typically associated with external cloud services, presents a unique opportunity to democratize AI capabilities. By ensuring that users can interact with and execute complex AI workflows with ease, this feature aims to bridge the gap between advanced AI technologies and non-technical end users.

Components

  • build llama.cpp bindings and leverage llm-chain
  • enable CUDA support for os's that can use it
  • associated Wasm module
  • example app
@zeeshanlakhani zeeshanlakhani added the enhancement New feature or request label Mar 25, 2024
@zeeshanlakhani zeeshanlakhani self-assigned this Mar 25, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant