Yummy mlflow models serving

Performant rust implementation of the models server allows to deploy mlflow models as Rest API. The server expose the same /invocations endpoint.

See: benchmark to analyse response times and resource consumption in comparison to the mlflow server.

Yummy mlflow

The MLflow rust wrapper currently supports models: [x] lightgbm [x] catboost (only binary classification)

The implementation currently supports MLflow models kept on local path.

pip3 install yummy[mlflow]

To run the model run:

yummy models serve -h -p 8080 -m /tmp/binary_lightgbm/

The yummy-mlflow will expose HTTP server. The request response is compatible with MLflow model serving API.



curl -X POST "http://localhost:8080/invocations" \
-H "Content-Type: application/json" \
-d '{
    "columns": ["0","1","2","3","4","5","6","7","8","9","10",
    "data": [
     [ 0.913333, -0.598156, -0.425909, -0.929365,  1.281985,
       0.488531,  0.874184, -1.223610,  0.050988,  0.342557,
      -0.164303,  0.830961,  0.997086,


[[0.9849612333276241, 0.008531186707393178, 0.006507579964982725]]