Skip to content

Conversation

@fengyuan14
Copy link
Collaborator

@fengyuan14 fengyuan14 commented Jan 11, 2024

Stack from ghstack (oldest at bottom):

e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan [email protected]

@pytorch-bot
Copy link

pytorch-bot bot commented Jan 11, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/117234

Note: Links to docs will display an error until the docs builds have been completed.

❌ 1 New Failure

As of commit d0f7eaa with merge base b830751 (image):

NEW FAILURE - The following job has failed:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@pytorch-bot pytorch-bot bot added the release notes: sparse release notes category label Jan 11, 2024
fengyuan14 added a commit that referenced this pull request Jan 11, 2024
…upcoming SYCL kernels

e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

ghstack-source-id: 0db78a6
Pull Request resolved: #117234
…upcoming SYCL kernels

e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <[email protected]>

[ghstack-poisoned]
fengyuan14 added a commit that referenced this pull request Jan 12, 2024
…upcoming SYCL kernels

e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

ghstack-source-id: 9cafc7c
Pull Request resolved: #117234
…reusing in upcoming SYCL kernels"

e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

[ghstack-poisoned]
…reusing in upcoming SYCL kernels"


e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

[ghstack-poisoned]
…reusing in upcoming SYCL kernels"


e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

[ghstack-poisoned]
…reusing in upcoming SYCL kernels"


e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

[ghstack-poisoned]
…reusing in upcoming SYCL kernels"


e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

[ghstack-poisoned]
fengyuan14 added a commit that referenced this pull request Jan 24, 2024
…upcoming SYCL kernels

e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

ghstack-source-id: 9634e48
Pull Request resolved: #117234
…reusing in upcoming SYCL kernels"


e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

[ghstack-poisoned]
…reusing in upcoming SYCL kernels"


e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

[ghstack-poisoned]
@gujinghui
Copy link
Collaborator

@pytorchbot rebase

@pytorchmergebot
Copy link
Collaborator

@pytorchbot started a rebase job onto refs/remotes/origin/viable/strict. Check the current status here

…reusing in upcoming SYCL kernels"


e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

[ghstack-poisoned]
@pytorchmergebot
Copy link
Collaborator

Successfully rebased gh/arthuryuan1987/2/orig onto refs/remotes/origin/viable/strict, please pull locally before adding more changes (for example, via ghstack checkout https://github.com/pytorch/pytorch/pull/117234)

pytorchmergebot pushed a commit that referenced this pull request Jan 31, 2024
…upcoming SYCL kernels

e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

ghstack-source-id: e619ad4
Pull Request resolved: #117234
fengyuan14 added a commit that referenced this pull request Feb 5, 2024
…upcoming SYCL kernels

e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

ghstack-source-id: 8943663
Pull Request resolved: #117234
…reusing in upcoming SYCL kernels"


e.g. OffsetCalculator, IntegerDivider, MemoryAccess utilities, ElementwiseInvoke.
feature: #114835

Abstracted helpers are about integer arithmetic and C++ standard template helpers for load, store and custom functor invoke.
They are general for any kind of kernel language, which supports basic interger arithmetic and C++ standard template, except for,
when we have some backend specific intrinsic, like CUDA __umulhi. Using backend specific macro to isolate divergency.

They are irrelevant to concurrency algorithm of kernel, backend specific data type (e.g. CUDA half2)
and backend specific logic concurrent resource configuration (e.g. CUDA block size, thread work size).

Signed-off-by: Feng Yuan <feng1.yuanintel.com>

[ghstack-poisoned]
@github-actions
Copy link
Contributor

github-actions bot commented Apr 5, 2024

Looks like this PR hasn't been updated in a while so we're going to go ahead and mark this as Stale.
Feel free to remove the Stale label if you feel this was a mistake.
If you are unable to remove the Stale label please contact a maintainer in order to do so.
If you want the bot to never mark this PR stale again, add the no-stale label.
Stale pull requests will automatically be closed after 30 days of inactivity.

@github-actions github-actions bot added the Stale label Apr 5, 2024
@github-actions github-actions bot closed this May 5, 2024
@github-actions github-actions bot deleted the gh/arthuryuan1987/2/head branch June 5, 2024 01:52
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants