Project Roadmap#
A breakdown of the functionality we aim to provide in stages:
Command-line deployment of the LLM
Deployment of the LLM via web-based graphical user interface (GUI)
Providing easily accessible graphical user interfaces to interact with the LLM once it is running
Status#
Deployment |
Interaction |
|
---|---|---|
CLI |
testing |
development |
GUI |
tbd |
not started |
Command-Line Deployment#
Status: Testing
We provide a command-line tool for generating SLURM jobs that deploy a pre-configured LM server on a compute node, with helpful utilities for forwarding the port served by the LM and querying its API.
GUI-based Deployment#
Status: tbd
We provide a web-based graphical interface (like MyDella) with textboxes/dropdown menus to choose an open-source LM and deploy it on Della.
The deployment interface can be made with Open OnDemand or a simple web interface
GUIs to interact with the LM#
Status: not started
We provide several “supported” graphical interfaces for users that the LM server will automatically be connected to after they request the job through the MyDella-like interface.
Target GUIs:
Chat
Completion
Batch zero/few-shot