Skip to content

W&B Logging support for Finetuning#815

Merged
psinger-prior merged 18 commits intomainfrom
psi/finetuning_v2
Apr 7, 2026
Merged

W&B Logging support for Finetuning#815
psinger-prior merged 18 commits intomainfrom
psi/finetuning_v2

Conversation

@psinger-prior
Copy link
Copy Markdown
Contributor

@psinger-prior psinger-prior commented Mar 11, 2026

Issue

Closes #810

Motivation and Context

Tracking training runs is important, implemented a logging class that can be expanded for new logger support. First support is for W&B.

This PR builds on top of #812 and should only merged afterwards.

Public API Changes

  • No Public API changes
  • Yes, Public API changes (Details below)

How Has This Been Tested?


Checklist

  • The changes have been tested locally.
  • Documentation has been updated (if the public API or usage changes).
  • A changelog entry has been added (see changelog/README.md), or "no changelog needed" label requested.
  • The code follows the project's style guidelines.
  • I have considered the impact of these changes on the public API.

image

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request introduces two major features: multi-GPU training support using DistributedDataParallel (DDP) and experiment tracking with Weights & Biases. The DDP implementation is well-structured, handling process group initialization, data sampling, and metric synchronization correctly. The introduction of a logging protocol with a W&B implementation is a great addition for experiment tracking. I've identified a critical issue regarding DDP support for multiple models, which will cause a crash. I've also made a couple of medium-severity suggestions to improve code clarity and documentation.

Comment thread src/tabpfn/finetuning/finetuned_base.py
Comment thread changelog/812.added.md Outdated
Comment thread src/tabpfn/finetuning/finetuned_base.py
@egeonur egeonur mentioned this pull request Mar 11, 2026
4 tasks
@noahho noahho requested review from a team and priorjulien and removed request for a team March 31, 2026 14:40
@noahho
Copy link
Copy Markdown
Collaborator

noahho commented Mar 31, 2026

@psinger-prior whats the state of this PR?

@noahho noahho removed the request for review from priorjulien March 31, 2026 14:40
@psinger-prior psinger-prior marked this pull request as ready for review April 7, 2026 10:48
@psinger-prior psinger-prior requested a review from a team as a code owner April 7, 2026 10:48
@psinger-prior psinger-prior requested review from priorphil and removed request for a team April 7, 2026 10:48
@chatgpt-codex-connector
Copy link
Copy Markdown

Codex usage limits have been reached for code reviews. Please check with the admins of this repo to increase the limits by adding credits.
Credits must be used to enable repository wide code reviews.

Copy link
Copy Markdown
Collaborator

@priorphil priorphil left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks!

@psinger-prior psinger-prior enabled auto-merge April 7, 2026 12:29
@psinger-prior psinger-prior added this pull request to the merge queue Apr 7, 2026
Merged via the queue into main with commit 0fba68c Apr 7, 2026
12 checks passed
ggprior pushed a commit that referenced this pull request Apr 8, 2026
ggprior pushed a commit that referenced this pull request Apr 8, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Experiment tracking/logging for Finetuning

3 participants