Interface ServingEndpointsService


  • @Generated
    public interface ServingEndpointsService
    The Serving Endpoints API allows you to create, update, and delete model serving endpoints.

    You can use a serving endpoint to serve models from the Databricks Model Registry or from Unity Catalog. Endpoints expose the underlying models as scalable REST API endpoints using serverless compute. This means the endpoints and associated compute resources are fully managed by Databricks and will not appear in your cloud account. A serving endpoint can consist of one or more MLflow models from the Databricks Model Registry, called served models. A serving endpoint can have at most ten served models. You can configure traffic settings to define how requests should be routed to your served models behind an endpoint. Additionally, you can configure the scale of resources that should be applied to each served model.

    This is the high-level interface, that contains generated methods.

    Evolving: this interface is under development. Method signatures may change.

    • Method Detail

      • buildLogs

        BuildLogsResponse buildLogs​(BuildLogsRequest buildLogsRequest)
        Retrieve the logs associated with building the model's environment for a given serving endpoint's served model.

        Retrieves the build logs associated with the provided served model.

      • exportMetrics

        void exportMetrics​(ExportMetricsRequest exportMetricsRequest)
        Retrieve the metrics associated with a serving endpoint.

        Retrieves the metrics associated with the provided serving endpoint in either Prometheus or OpenMetrics exposition format.

      • logs

        ServerLogsResponse logs​(LogsRequest logsRequest)
        Retrieve the most recent log lines associated with a given serving endpoint's served model.

        Retrieves the service logs associated with the provided served model.

      • updateConfig

        ServingEndpointDetailed updateConfig​(EndpointCoreConfigInput endpointCoreConfigInput)
        Update a serving endpoint with a new config.

        Updates any combination of the serving endpoint's served models, the compute configuration of those served models, and the endpoint's traffic config. An endpoint that already has an update in progress can not be updated until the current update completes or fails.