API Reference
GoServe provides a REST API for model management and inference.
Server Information
GET /
Returns information about the server.
Response:
{
"name": "GoServe",
"version": "0.1.0-dev",
"message": "Welcome to GoServe - Fast ML Model Server"
}
Health Checks
GET /health
Liveness check. Returns 200 OK if the server is running.
GET /ready
Readiness check. Returns 200 OK if the server is ready to accept requests.
Monitoring
GET /metrics
Exposes Prometheus metrics.
Response: Prometheus text format.
Model Management
GET /v1/models
Lists all loaded models.
Response:
{
"models": [
{
"name": "my_model",
"path": "models/my_model.onnx",
"inputs": [{"name": "input_0", "shape": [-1, 30], "type": "float32"}],
"outputs": [{"name": "output_0", "shape": [-1, 2], "type": "float32"}]
}
]
}
POST /v1/models
Loads a new model.
Request Body:
Inference
POST /v1/models/{model_name}/infer
Executes inference using the specified model.
Generic Request Format (Recommended):
Array Request Format:
Response Format:
Type Support
Currently supports FLOAT32 and INT64 tensor types. Work is in progress for full support of all ONNX types.