Skip to content

Conversation

@atalman
Copy link
Contributor

@atalman atalman commented Jun 5, 2025

Move non inductor workflows cuda 12.6->cuda 12.8

@pytorch-bot
Copy link

pytorch-bot bot commented Jun 5, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/155234

Note: Links to docs will display an error until the docs builds have been completed.

❌ 4 New Failures, 1 Cancelled Job, 7 Unrelated Failures

As of commit bd33e48 with merge base 8892b78 (image):

NEW FAILURES - The following jobs have failed:

CANCELLED JOB - The following job was cancelled. Please retry:

FLAKY - The following jobs failed but were likely due to flakiness present on trunk:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@pytorch-bot pytorch-bot bot added the topic: not user facing topic category label Jun 5, 2025
@atalman atalman added ci-no-td Do not run TD on this PR ciflow/pull ciflow/periodic Trigger jobs ran periodically on master (periodic.yml) on the PR ciflow/slow labels Jun 5, 2025
@atalman atalman marked this pull request as ready for review June 5, 2025 16:36
@atalman atalman requested review from a team and jeffdaily as code owners June 5, 2025 16:36
@atalman atalman force-pushed the cuda12_8_migration branch from 3e6a325 to 875345a Compare June 6, 2025 18:40
@atalman
Copy link
Contributor Author

atalman commented Jun 9, 2025

Observing these failures on trunk:
test_nn.py::TestNNDeviceTypeCUDA::test_variable_sequence_cuda_float32 GH job link HUD commit link

inductor/test_inductor_freezing.py::FreezingGpuTests::test_cpp_wrapper_cuda GH job link HUD commit link

test_expanded_weights.py::TestExpandedWeightModuleCUDA::test_module_nn_GRU_eval_mode_cuda_float32 GH job link HUD commit link

test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 GH job link HUD commit link

Looks like errors on this issue are pre-existing on trunk this Friday

@atalman atalman force-pushed the cuda12_8_migration branch from 875345a to 24d5cb3 Compare June 9, 2025 20:55
@atalman
Copy link
Contributor Author

atalman commented Jun 9, 2025

@pytorchmergebot merge -f "lint is green all other tests have been already run"

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged immediately since you used the force (-f) flag, bypassing any CI checks (ETA: 1-5 minutes). Please use -f as last resort and instead consider -i/--ignore-current to continue the merge ignoring current failures. This will allow currently pending tests to finish and report signal before the merge.

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@clee2000
Copy link
Contributor

@pytorchbot revert -m "causing a bunch of tests to fail? ex test_nn.py::TestNNDeviceTypeCUDA::test_variable_sequence_cuda_float32 GH job link HUD commit link, some of the failures attributed to broken trunk on friday seem real?" -c ignoredsignal

@pytorchmergebot
Copy link
Collaborator

@pytorchbot successfully started a revert job. Check the current status here.
Questions? Feedback? Please reach out to the PyTorch DevX Team

@pytorchmergebot
Copy link
Collaborator

@atalman your PR has been successfully reverted.

@atalman
Copy link
Contributor Author

atalman commented Jun 11, 2025

This are existing issues:

test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_complex128 GH job link HUD commit link

@atalman atalman added the keep-going Don't stop on first failure, keep running tests until the end label Jun 11, 2025
@atalman
Copy link
Contributor Author

atalman commented Jun 12, 2025

@pytorchmergebot merge -f "all previously failed workflows are passing now"

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged immediately since you used the force (-f) flag, bypassing any CI checks (ETA: 1-5 minutes). Please use -f as last resort and instead consider -i/--ignore-current to continue the merge ignoring current failures. This will allow currently pending tests to finish and report signal before the merge.

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ci-no-td Do not run TD on this PR ciflow/periodic Trigger jobs ran periodically on master (periodic.yml) on the PR ciflow/pull ciflow/slow keep-going Don't stop on first failure, keep running tests until the end Merged Reverted topic: not user facing topic category

Projects

None yet

Development

Successfully merging this pull request may close these issues.

7 participants