Logging utilities for SpaCy
Starting with spaCy v3.2, alternate loggers are moved into a separate package so that they can be added and updated independently from the core spaCy library.
spacy-loggers
currently provides loggers for:
spacy-loggers
also provides additional utility loggers to facilitate interoperation
between individual loggers.
If you'd like to add a new logger or logging option, please submit a PR to this repo!
spacy-loggers
should be installed automatically with spaCy v3.2+, so you
usually don't need to install it separately. You can install it with pip
or
from the conda channel conda-forge
:
pip install spacy-loggers
conda install -c conda-forge spacy-loggers
This logger requires wandb
to be installed and configured:
pip install wandb
wandb login
spacy.WandbLogger.v5
is a logger that sends the results of each training step
to the dashboard of the Weights & Biases tool. To use
this logger, Weights & Biases should be installed, and you should be logged in.
The logger will send the full config file to W&B, as well as various system
information such as memory utilization, network traffic, disk IO, GPU
statistics, etc. This will also include information such as your hostname and
operating system, as well as the location of your Python executable.
spacy.WandbLogger.v4
and below automatically call the default console logger.
However, starting with spacy.WandbLogger.v5
, console logging must be activated
through the use of the ChainLogger. This allows the user to configure
the console logger's parameters according to their preferences.
Note that by default, the full (interpolated)
training config is sent over to the
W&B dashboard. If you prefer to exclude certain information such as path
names, you can list those fields in "dot notation" in the
remove_config_values
parameter. These fields will then be removed from the
config before uploading, but will otherwise remain in the config file stored
on your local system.
[training.logger]
@loggers = "spacy.WandbLogger.v5"
project_name = "monitor_spacy_training"
remove_config_values = ["paths.train", "paths.dev", "corpora.train.path", "corpora.dev.path"]
log_dataset_dir = "corpus"
model_log_interval = 1000
Name | Type | Description |
---|---|---|
project_name |
str |
The name of the project in the Weights & Biases interface. The project will be created automatically if it doesn't exist yet. |
remove_config_values |
List[str] |
A list of values to exclude from the config before it is uploaded to W&B (default: [] ). |
model_log_interval |
Optional[int] |
Steps to wait between logging model checkpoints to the W&B dasboard (default: None ). Added in spacy.WandbLogger.v2 . |
log_dataset_dir |
Optional[str] |
Directory containing the dataset to be logged and versioned as a W&B artifact (default: None ). Added in spacy.WandbLogger.v2 . |
entity |
Optional[str] |
An entity is a username or team name where you're sending runs. If you don't specify an entity, the run will be sent to your default entity, which is usually your username (default: None ). Added in spacy.WandbLogger.v3 . |
run_name |
Optional[str] |
The name of the run. If you don't specify a run name, the name will be created by the wandb library (default: None ). Added in spacy.WandbLogger.v3 . |
log_best_dir |
Optional[str] |
Directory containing the best trained model as saved by spaCy (by default in training/model-best ), to be logged and versioned as a W&B artifact (default: None ). Added in spacy.WandbLogger.v4 . |
log_latest_dir |
Optional[str] |
Directory containing the latest trained model as saved by spaCy (by default in training/model-latest ), to be logged and versioned as a W&B artifact (default: None ). Added in spacy.WandbLogger.v4 . |
log_custom_stats |
Optional[List[str]] |
A list of regular expressions that will be applied to the info dictionary passed to the logger (default: None ). Statistics and metrics that match these regexps will be automatically logged. Added in spacy.WandbLogger.v5 . |
This logger requires mlflow
to be installed and configured:
pip install mlflow
spacy.MLflowLogger.v2
is a logger that tracks the results of each training step
using the MLflow tool. To use
this logger, MLflow should be installed. At the beginning of each model training
operation, the logger will initialize a new MLflow run and set it as the active
run under which metrics and parameters wil be logged. The logger will then log
the entire config file as parameters of the active run. After each training step,
the following actions are performed:
score
.loss_
prefix.output_path
argument is provided during the training pipeline initialization phase.By default, the tracking API writes data into files in a local ./mlruns
directory.
spacy.MLflowLogger.v1
and below automatically call the default console logger.
However, starting with spacy.MLflowLogger.v2
, console logging must be activated
through the use of the ChainLogger. This allows the user to configure
the console logger's parameters according to their preferences.
Note that by default, the full (interpolated)
training config is sent over to
MLflow. If you prefer to exclude certain information such as path
names, you can list those fields in "dot notation" in the
remove_config_values
parameter. These fields will then be removed from the
config before uploading, but will otherwise remain in the config file stored
on your local system.
[training.logger]
@loggers = "spacy.MLflowLogger.v2"
experiment_id = "1"
run_name = "with_fast_alignments"
nested = False
remove_config_values = ["paths.train", "paths.dev", "corpora.train.path", "corpora.dev.path"]
Name | Type | Description |
---|---|---|
run_id |
Optional[str] |
Unique ID of an existing MLflow run to which parameters and metrics are logged. Can be omitted if experiment_id and run_id are provided (default: None ). |
experiment_id |
Optional[str] |
ID of an existing experiment under which to create the current run. Only applicable when run_id is None (default: None ). |
run_name |
Optional[str] |
Name of new run. Only applicable when run_id is None (default: None ). |
nested |
bool |
Controls whether run is nested in parent run. True creates a nested run (default: False ). |
tags |
Optional[Dict[str, Any]] |
A dictionary of string keys and values to set as tags on the run. If a run is being resumed, these tags are set on the resumed run. If a new run is being created, these tags are set on the new run (default: None ). |
remove_config_values |
List[str] |
A list of values to exclude from the config before it is uploaded to MLflow (default: [] ). |
log_custom_stats |
Optional[List[str]] |
A list of regular expressions that will be applied to the info dictionary passed to the logger (default: None ). Statistics and metrics that match these regexps will be automatically logged. Added in spacy.MLflowLogger.v2 . |
This logger requires clearml
to be installed and configured:
pip install clearml
clearml-init
spacy.ClearMLLogger.v2
is a logger that tracks the results of each training step
using the ClearML tool. To use
this logger, ClearML should be installed and you should have initialized (using the command above).
The logger will send all the gathered information to your ClearML server, either the hosted free tier
or the open source self-hosted server. This logger captures the following information, all of which is visible in the ClearML web UI:
score
.loss_
prefix.In addition to the above, the following artifacts can also be optionally captured:
spacy.ClearMLLogger.v1
and below automatically call the default console logger.
However, starting with spacy.ClearMLLogger.v2
, console logging must be activated
through the use of the ChainLogger. This allows the user to configure
the console logger's parameters according to their preferences.
Note that by default, the full (interpolated)
training config is sent over to
ClearML. If you prefer to exclude certain information such as path
names, you can list those fields in "dot notation" in the
remove_config_values
parameter. These fields will then be removed from the
config before uploading, but will otherwise remain in the config file stored
on your local system.
[training.logger]
@loggers = "spacy.ClearMLLogger.v2"
project_name = "Hello ClearML!"
task_name = "My spaCy Task"
model_log_interval = 1000
log_best_dir = training/model-best
log_latest_dir = training/model-last
log_dataset_dir = corpus
remove_config_values = ["paths.train", "paths.dev", "corpora.train.path", "corpora.dev.path"]
Name | Type | Description |
---|---|---|
project_name |
str |
The name of the project in the ClearML interface. The project will be created automatically if it doesn't exist yet. |
task_name |
str |
The name of the ClearML task. A task is an experiment that lives inside a project. Can be non-unique. |
remove_config_values |
List[str] |
A list of values to exclude from the config before it is uploaded to ClearML (default: [] ). |
model_log_interval |
Optional[int] |
Steps to wait between logging model checkpoints to the ClearML dasboard (default: None ). Will have no effect without also setting log_best_dir or log_latest_dir . |
log_dataset_dir |
Optional[str] |
Directory containing the dataset to be logged and versioned as a ClearML Dataset (default: None ). |
log_best_dir |
Optional[str] |
Directory containing the best trained model as saved by spaCy (by default in training/model-best ), to be logged and versioned as a ClearML artifact (default: None ) |
log_latest_dir |
Optional[str] |
Directory containing the latest trained model as saved by spaCy (by default in training/model-last ), to be logged and versioned as a ClearML artifact (default: None ) |
log_custom_stats |
Optional[List[str]] |
A list of regular expressions that will be applied to the info dictionary passed to the logger (default: None ). Statistics and metrics that match these regexps will be automatically logged. Added in spacy.ClearMLLogger.v2 . |
This logger requires torch
to be installed:
pip install torch
spacy.PyTorchLogger.v1
is different from the other loggers above in that it does not act as a bridge between spaCy and
an external framework. Instead, it is used to query PyTorch-specific metrics and make them available to other loggers.
Therefore, it's primarily intended to be used with ChainLogger.
Whenever a logging checkpoint is reached, it queries statistics from the PyTorch backend and stores them in the dictionary passed to it. Downstream loggers can thereafter lookup the statistics and log them to their preferred framework.
The following PyTorch statistics are currently supported:
[training.logger]
@loggers = "spacy.ChainLogger.v1"
logger1 = {"@loggers": "spacy.PyTorchLogger.v1", "prefix": "pytorch", "device": "0", "cuda_mem_metric": "current"}
# Alternatively, you can use any other logger that provides the `log_custom_stats` parameter.
logger2 = {"@loggers": "spacy.LookupLogger.v1", "patterns": ["pytorch"]}
Name | Type | Description |
---|---|---|
prefix |
str |
All metric names are prefixed with this string using dot notation, e.g: <prefix>.<metric> (default: pytorch ). |
device |
int |
The identifier of the CUDA device (default: 0 ). |
cuda_mem_pool |
str |
One of the memory pool values specified in the PyTorch docs: all , large_pool , small_pool (default: all ). |
cuda_mem_metric |
str |
One of the memory metric values specified in the PyTorch docs: current , peak , allocated , freed . To log all metrics, use all instead (default: all ). |
This logger requires cupy
to be installed:
pip install cupy
Similar to PyTorchLogger
, spacy.CupyLogger.v1
does not act as a bridge between spaCy and an external framework
but rather is used with the ChainLogger to facilitate the flow of metrics to other loggers.
The CupyLogger
queries statistics from the CuPy backend and stores them in the info dictionary passed to it. Downstream
loggers can thereafter lookup the statistics and log them to their preferred framework.
The following CuPy statistics are currently supported:
[training.logger]
@loggers = "spacy.ChainLogger.v1"
logger1 = {"@loggers": "spacy.CupyLogger.v1", "prefix": "cupy"}
# Alternatively, you can use any other logger that provides the `log_custom_stats` parameter.
logger2 = {"@loggers": "spacy.LookupLogger.v1", "patterns": ["cupy"]}
Name | Type | Description |
---|---|---|
prefix |
str |
All metric names are prefixed with this string using dot notation, e.g: <prefix>.<metric> (default: "cupy" ). |
This logger can be used to daisy-chain multiple loggers and execute them in-order. Loggers that are executed earlier in the chain can pass information to those that come later by adding it to the dictionary that is passed to them.
Currently, up to 10 loggers can be chained together.
[training.logger]
@loggers = "spacy.ChainLogger.v1"
logger1 = {"@loggers": "spacy.PyTorchLogger.v1"}
logger2 = {"@loggers": "spacy.ConsoleLogger.v1", "progress_bar": "true"}
Name | Type | Description |
---|---|---|
logger1 |
Optional[Callable] |
The first logger in the chain (default: None ). |
logger2 |
Optional[Callable] |
The second logger in the chain (default: None ). |
logger3 |
Optional[Callable] |
The third logger in the chain (default: None ). |
logger4 |
Optional[Callable] |
The fourth logger in the chain (default: None ). |
logger5 |
Optional[Callable] |
The fifth logger in the chain (default: None ). |
logger6 |
Optional[Callable] |
The sixth logger in the chain (default: None ). |
logger7 |
Optional[Callable] |
The seventh logger in the chain (default: None ). |
logger8 |
Optional[Callable] |
The eighth logger in the chain (default: None ). |
logger9 |
Optional[Callable] |
The ninth logger in the chain (default: None ). |
logger10 |
Optional[Callable] |
The tenth logger in the chain (default: None ). |
This logger can be used to lookup statistics in the info dictionary and print them to stdout
. It is primarily
intended to be used as a tool when developing new loggers.
[training.logger]
@loggers = "spacy.ChainLogger.v1"
logger1 = {"@loggers": "spacy.PyTorchLogger.v1", "prefix": "pytorch"}
logger2 = {"@loggers": "spacy.LookupLogger.v1", "patterns": ["^[pP]ytorch"]}
Name | Type | Description |
---|---|---|
patterns |
List[str] |
A list of regular expressions. If a statistic's name matches one of these, it's printed to stdout . |
Please use spaCy's issue tracker to report a bug, or open a new thread on the discussion board for any other issue.