New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Rest API for inference locally #1563
Comments
An extensive gradio API exists, see: See readme_client.md and examples via test code like And a full chat OpenAI API that is REST capable exists, but no upload of file or other things exists yet. Is that what you are looking for? |
What I am looking for is a fastapi rest API for the different ingestion techniques and a rag completion API so I can use H2OGPT as a backend rag for my frontend webUI. Also, I wish you included JSON metadata for filtering in ingestion and rag completion so we can choose the files to chat with. |
Am currently building something exactly like this. its still in development tho. U can certainly fork the repo or make PR's. the foundation is there. The project extends the official FastAPI Template so scalling and deploying wont really much of a husle. check it out here: https://github.com/abuyusif01/h2ogpt-fast-api/tree/main/backend/app/h2ogpt there's still alot things need to be done. Including a proper README and support Streaming the Response (I planed to get this done in this weekend) Here is what we currently support:
|
hi @abuyusif01 |
@mohamed-alired Subsequently, i restructure the repo, write a readme and containerize the app. Its now easy to setup + extend @pseudotensor |
hi
I have installed h2ogpt locally, but I want to build a frontend app using it, so I was wondering if there's an API that I can consume, like one for ingestion and another for inference.
The text was updated successfully, but these errors were encountered: