Summary
This is the stack for LFM-3B-JP.
How to run for the first time
- Download
Source code (zip)
below. - Unzip the file into an empty folder.
- Run
launch.sh
.
Models
Currently, each on-prem stack can only run one model at a time. The launch script runs lfm-3b-jp
by default. To switch models, change MODEL_NAME
and MODEL_IMAGE
in the .env
file according to table below, and run ./launch.sh
again.
Model Name | Model Image |
---|---|
lfm-3b-jp |
liquidai/lfm-3b-jp:0.0.1-e |
lfm-3b-ichikara |
liquidai/lfm-3b-ichikara:0.0.1-e |
Update
To update the stack, change STACK_VERSION
and MODEL_IMAGE
in the .env
file and run the launch script again.
How to test
- After running
launch.sh
, wait up to 2 min for model initialization, and runtest-api.sh
.- This script will trigger a smoke test to verify that the inference server is running correctly.
- Visit
0.0.0.0:3000
and chat with the model in a web UI.