Graphsignal Logger
Project description
Graphsignal Logger
Overview
Graphsignal is a machine learning model monitoring platform. It helps ML engineers and data scientists address data issues and analyze model performance in production. Learn more at graphsignal.com.
Model Monitoring
- Data monitoring. Monitor offline and online predictions for data validity and anomalies, data drift, model drift, exceptions, and more.
- Model performance monitoring. Monitor model performance for binary, categorical and numeric models and data segments.
- Automatic issue detection. Graphsignal automatically detects and notifies on issues with data and models, no need to manually setup and maintain complex rules.
- Model framework and deployment agnostic. Monitor models serving online, in streaming apps, accessed via APIs or offline, running batch predictions.
- Any scale and data size. Graphsignal logger only sends data statistics allowing it to scale with your application and data.
- Data privacy. No raw data is sent to Graphsignal cloud, only data statistics and metadata.
Dashboards and Alerting
Data Analysis
Model Performance
Automatic Alerts
Documentation
See full documentation at graphsignal.com/docs.
Getting Started
1. Installation
Install the Python logger by running
pip install graphsignal
Or clone and install the GitHub repository.
git clone https://github.com/graphsignal/graphsignal.git
python setup.py install
Import the package in your application
import graphsignal
2. Configuration
Configure the logger by specifying your API key.
graphsignal.configure(api_key='my_api_key')
To get an API key, sign up for a free account at graphsignal.com. The key can then be found in your account's Settings / API Keys page.
3. Logging session
Get logging session for a deployed model identified by deployment name. Multiple sessions can be used in parallel in case of multi-model scrips or servers.
sess = graphsignal.session(deployment_name='model1_prod')
Log any model metadata such as model version or deployment information.
sess.log_metadata('key1', 'val1')
4. Prediction Logging
Log single or batch model prediction/inference data. Computed data statistics are uploaded at certain intervals and on process exit.
Log single prediction.
sess.log_prediction(
features={'feat1': 1.2, 'feat2': 'XX'},
prediction=True)
Log prediction batch. Pass prediction data using list
, dict
, numpy.ndarray
or pandas.DataFrame
.
sess.log_prediction_batch(
features=[[1.2, 70], [3.5, 40]],
predictions=[[0.5], [0.75]])
Log prediction exceptions and errors.
sess.log_exception(
message='wrong format',
extra_info={'feature': 'F1'})
See logging API reference for full documentation.
5. Ground Truth Logging
Log ground truth when it becomes available. Because ground truth is usually available at a later point, ground truth logging is independent from prediction logging. Prediction logging is not required for model performance monitoring and visualization.
sess.log_ground_truth(
label=True,
prediction=False)
See logging API reference for full documentation.
Model output type is inferred from label and prediction types. Model performance metrics such as accuracy, F1-score, MSE, etc. are computed based on the model output type.
To additionally visualize and monitor performance metrics for various data segments, a segments
list can be provided.
sess.log_ground_truth(
label=True,
prediction=False,
segments=['seg1', 'seg2'])
6. Dashboards and Alerting
After logging is setup, sign in to Graphsignal to check out various dashboards and set up alerting for automatically detected issues.
Examples
Logging model serving predictions.
from tensorflow import keras
import json
from flask import Flask
from flask import request
import graphsignal
graphsignal.configure(api_key='my_key')
# Get Graphsignal logging session for deployed model
sess = graphsignal.session(deployment_name='fraud_detection_prod')
sess.log_metadata('model version', '1.0')
model = keras.models.load_model('fraud_model.h5')
app = Flask(__name__)
@app.route('/predict_fraud', methods = ['POST'])
def predict_digit():
try:
features = request.get_json()
# feature extraction code here...
output_data = model.predict([input_data])
# Log prediction
sess.log_prediction(
features=features,
prediction=output_data[0])
except:
sess.log_exception(exc_info=True)
return json.dumps(output_data.tolist())
app.run(port=8090)
Logging ground truth.
import graphsignal
graphsignal.configure(api_key='my_key')
# Get Graphsignal logging session for deployed model
sess = graphsignal.session(deployment_name='job_recommender_prod')
sess.log_metadata('model version', 'v1.2')
...
sess.log_ground_truth(
label=True,
prediction=False)
See more examples.
Performance
Graphsignal logger uses streaming algorithms for computing data statistics to ensure production-level performance and memory usage. Data statistics are computed for time windows and sent to Graphsignal by the background thread.
Since only data statistics is sent to our servers, there is no limitation on logged data size.
Security and Privacy
Graphsignal logger can only open outbound connections to log-api.graphsignal.com
and send data, no inbound connections or commands are possible.
No raw data is sent to Graphsignal cloud, only data statistics and metadata.
Troubleshooting
To enable debug logging, add debug_mode=True
to configure()
. If the debug log doesn't give you any hints on how to fix a problem, please report it to our support team via your account.
In case of connection issues, please make sure outgoing connections to https://log-api.graphsignal.com
are allowed.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.