Skip to content

marswen/ThreeSensei

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ThreeSensei

中文 | English

This project attempts to use the approach of the LLM agent to solve complex reasoning problems in the clinical field. Taking the consultation scenario as an example, it theoretically supports other clinical decision support issues such as differential diagnosis, nursing care plan development, and risk assessment.

How it works

graph TB
    A(Analyze relevant departments) --> B(Generate persona prompts for doctors in each department)
    B(Generate persona prompts for doctors in each department) --> C(Each virtual doctor gives diagnostic opinions on medical records)
    C(Each virtual doctor gives diagnostic opinions on medical records) --> D(Summarize all diagnostic opinions)
    D(Summarize all diagnostic opinions) --> E(Each virtual doctor provides additional suggestions on the discussion results)
    E(Each virtual doctor provides additional suggestions on the diagnostic opinions) --> F{Any additional suggestions?}
    F{Any additional suggestions?} -- No --> H(Summarize medical records)
    F{Any additional suggestions?} -- Yes --> G(Collect additional suggestions)
    G(Collect additional suggestions) --> I(Improve diagnostic opinions according to the additional suggestions)
    I(Improve diagnostic opinions according to the additional suggestions) --> E(Each virtual doctor provides additional suggestions on the discussion results)
    H(Summarize medical records) --> J(Generate consultation report)

How to use

Configuration

Supports local models Llama, ChatGLM, Baichuan, and OpenAI API.

If using a local model, modify MODEL_PATH to the huggingface model name or the local cache directory in the .env file.

If using OpenAI API, modify OPENAI_API_TOKEN and OPENAI_MODEL_NAME in the .env file.

Model accuracy and inference parameters are also configured in the .env file. Local model loading accuracy supports full precision, float16, and int8.

Running

Loading medical record data from FILE_PATH as an example, and parameter lang can be "en" or "zh":

cd src
python cli.py --file_path=FILE_PATH --lang=en

The running logs are saved in log/*.log, and the consultation report results are saved in report/*.txt.

About

Leverage LLM agent to conduct clinical consultation

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages