-
Notifications
You must be signed in to change notification settings - Fork 261
Open
Labels
Description
Rather than serving all models with generic model.file
filename, ramalama should provide more information about the currently served or loaded models.
Also, could allow passing a model-config file to allow to switch easily between models with a single instance of a server.
https://llama-cpp-python.readthedocs.io/en/latest/server/#configuration-and-multi-model-support
Here, I've renamed the file ramalama downloaded to granite-code
so it shows at llamacpp:port/models.
Currently, any model served by ramalama is listed as model.file
vpavlin, aviallon, mrwsl, ElSamhaa, craychee and 2 more