[ZK Lab] #3: Platform - Weights & Biases

Weights & Biases is the machine learning platform for developers to build better models faster.

Use W&B’s lightweight, interoperable tools to

  • track experiments and version quickly
  • iterate on datasets
  • evaluate model performance
  • reproduce models
  • visualize results
  • spot regressions
  • share findings with colleagues

Experiment Tracking

  • Use the wandb Python library to track machine learning experiments with a few lines of code.

  • Have lightweight integrations for PyTorch and Keras.

Integrations

Weights & Biases integrations make it fast and easy to set up experiment tracking and data versioning inside existing projects.

Useful Links:

  • Examples: GitHub repo with working, end-to-end code examples for all of our integrations
  • Colab: Try out W&B inside different frameworks, such as PyTorch Lightning, in an interactive notebook – no installation required
  • Video Tutorials: Learn to use W&B with YouTube videos for PyTorch, Keras, and more.

Collaborative Reports

Reports let you organize and embed visualizations, describe your findings, share updates with collaborators, and more.

Typical use cases for reports

Notes: Add a visualization with a quick summary

Capture an important observation, an idea for future work, or a milestone reached in the development of a project. All experiment runs in your report will link to their parameters, metrics, logs, and code, so you can save the full context of your work.

Labeling and Crowdsourcing

Collaboration: Share findings with your colleagues

Explain how to get started with a project, share what you’ve observed so far, and synthesize the latest findings. Your colleagues can make suggestions or discuss details using comments on any panel or at the end of the report.

Labeling and Crowdsourcing

Work log: Track what you’ve tried and plan next steps

Write down your thoughts on experiments, your findings, and any gotchas and next steps as you work through a project, keeping everything organized in one place. This lets you “document” all the important pieces beyond your scripts.

Labeling and Crowdsourcing

Data + Model Versioning

Open In Colab

Use W&B Artifacts for dataset versioning, model versioning, and tracking dependencies and results across machine learning pipelines. Think of an artifact as a versioned folder of data. You can store entire datasets directly in artifacts, or use artifact references to point to data in other systems like S3, GCP, or your own system.

Data Visualization

Open In Colab

Use W&B Tables to log, query, and analyze tabular data. Understand your datasets, visualize model predictions, and share insights in a central dashboard.

  • Compare changes precisely across models, epochs, or individual examples
  • Understand higher-level patterns in your data
  • Capture and communicate your insights with visual samples

Hyperparameter Tuning

Open In Colab

Use Weights & Biases Sweeps to automate hyperparameter optimization and explore the space of possible models.

Model Management

A Model Registry is a system of record for organizing ML Models - often serving as an interface between model producers and consumers. This guide will show you how to use W&B as a Model Registry to track and report on the complete workflow of developing a model:

  • Model Catalog & Versioning: Save and restore every version of your model & learned parameters - organize versions by use case and objective.
  • Model Metadata: Track training metrics, assign custom metadata, and document rich markdown descriptions of your models.
  • Model Lineage Tracking & Reproducibility: Track the exact code, hyperparameters, & training dataset used to produce the model.
  • Model Lifecycle: Promote promising models to positions like “staging” or “production” - allowing downstream users to fetch the best model automatically. Model Reporting: Create a Report dashboard to summarize model progression and performance over time

© 2022 Jiawei Lu. All rights reserved.