Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Document the host flag in ramalama.conf file #447

Merged
merged 1 commit into from
Nov 12, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion docs/ramalama-serve.1.md
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ Generate specified configuration format for running the AI Model as a service
show this help message and exit

#### **--host**="0.0.0.0"
ip address to listen
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Doesn't block merge, but IP address to listen made more sense to me, we will have at least 2 other servers soon llama-cpp-python and vllm

IP address for llama.cpp to listen on.

#### **--name**, **-n**
Name of the container to run the Model in.
Expand Down
4 changes: 4 additions & 0 deletions docs/ramalama.conf
Original file line number Diff line number Diff line change
Expand Up @@ -35,6 +35,10 @@
#
#image = "quay.io/ramalama/ramalama:latest"

# IP address for llama.cpp to listen on.
#
#host = "0.0.0.0"

# Specify default port for services to listen on
#
#port = "8080"
Expand Down
8 changes: 6 additions & 2 deletions docs/ramalama.conf.5.md
Original file line number Diff line number Diff line change
Expand Up @@ -70,8 +70,12 @@ Image to be used when building and pushing --type=car models
**engine**="podman"

Run RamaLama using the specified container engine.
Valid options (Podman, Docker)
RAMALAMA_CONTAINER_ENGINE environment variable overrides this field.
Valid options are: Podman and Docker
This field can be overridden by the RAMALAMA_CONTAINER_ENGINE environment variable.

**host**="0.0.0.0"

IP address for llama.cpp to listen on.

**image**="quay.io/ramalama/ramalama:latest"

Expand Down
2 changes: 1 addition & 1 deletion ramalama/cli.py
Original file line number Diff line number Diff line change
Expand Up @@ -614,7 +614,7 @@ def serve_parser(subparsers):
parser = subparsers.add_parser("serve", help="serve REST API on specified AI Model")
parser.add_argument("--authfile", help="path of the authentication file")
parser.add_argument("-d", "--detach", action="store_true", dest="detach", help="run the container in detached mode")
parser.add_argument("--host", default=config.get('host', "0.0.0.0"), help="ip address to listen")
parser.add_argument("--host", default=config.get('host', "0.0.0.0"), help="IP address to listen")
parser.add_argument("-n", "--name", dest="name", help="name of container in which the Model will be run")
parser.add_argument(
"-p", "--port", default=config.get('port', "8080"), help="port for AI Model server to listen on"
Expand Down
Loading