Skip to content

Conversation

@sonhm3029
Copy link

@sonhm3029 sonhm3029 commented Sep 2, 2025

πŸ“ Description

  • Provide a clear summary of the changes and the issue that has been addressed.
  • πŸ› οΈ Fixes # (issue number)

✨ Changes

Select what type of change your PR is:

  • πŸš€ New feature (non-breaking change which adds functionality)
  • 🐞 Bug fix (non-breaking change which fixes an issue)
  • πŸ”„ Refactor (non-breaking change which refactors the code base)
  • ⚑ Performance improvements
  • 🎨 Style changes (code style/formatting)
  • πŸ§ͺ Tests (adding/modifying tests)
  • πŸ“š Documentation update
  • πŸ“¦ Build system changes
  • 🚧 CI/CD configuration
  • πŸ”§ Chore (general maintenance)
  • πŸ”’ Security update
  • πŸ’₯ Breaking change (fix or feature that would cause existing functionality to not work as expected)

βœ… Checklist

Before you submit your pull request, please make sure you have completed the following steps:

  • πŸ“š I have made the necessary updates to the documentation (if applicable).
  • πŸ§ͺ I have written tests that support my changes and prove that my fix is effective or my feature works (if applicable).
  • 🏷️ My PR title follows conventional commit format.

For more information about code review checklists, see the Code Review Checklist.

@rajeshgangireddy
Copy link
Contributor

References
Discussion : #2749
Issue : #2923

self.add_state("target", default=[], dist_reduce_fx="cat")
self.register_buffer("fpr_limit", torch.tensor(fpr_limit))
# Determine the device dynamically
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Will still work for devices that are neither cuda or cpu. For example, when we have intel GPUs, the device is called xpu. It would be nice to solve this such that it works independent of the device used.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants