mirror of
https://github.com/DrHo1y/ezrknn-llm.git
synced 2026-03-24 01:26:44 +07:00
RKLLM-Server Demo
Before Run
Before running the demo, you need to prepare the following files:
- The transformed RKLLM model file in board.
- check the IP address of the board with 'ifconfig' command.
RKLLM-Server-Flask Demo
Build
You can run the demo with the only command:
# ./build_rkllm_server_flask.sh [target_platform:rk3588/rk3576] [RKLLM-Server workshop] [transformed_rkllm_model_path in borad]
./build_rkllm_server_flask.sh rk3588 /user/data/rkllm_server /user/data/rkllm_server/model.rkllm
Access with API
After building the RKLLM-Server-Flask, You can use ‘chat_api_flask.py’ to access the RKLLM-Server-Flask and get the answser of RKLLM models.
Attention: you should check the IP address of the board with 'ifconfig' command and replace the IP address in the ‘chat_api_flask.py’.
RKLLM-Server-Gradio Demo
Build
You can run the demo with the only command:
# ./build_rkllm_server_gradio.sh [target_platform:rk3588/rk3576] [RKLLM-Server workshop] [transformed_rkllm_model_path in borad]
./build_rkllm_server_gradio.sh rk3588 /user/data/rkllm_server /user/data/rkllm_server/model.rkllm
Access the Server
After running the demo, You can access the RKLLM-Server-Gradio with two ways:
- Just Start your browser and access the URL: ‘http://[board_ip]:8080/’. You can chat with the RKLLM models in visual interface.
- Use the 'chat_api_gradio.py'(you need fix the IP address in the code previously) and get the answser of RKLLM models.