Skip to content

Question: Inference from CUDA allocated memory #10180

@IvensaMDH

Description

@IvensaMDH

Hi,

Is it possible to run inference (using CUDA provider) )from memory already allocated on CUDA without moving it from GPU to CPU and vice versa?

Could you provide an example of such implementation in C#?

Thanks,

/M

Metadata

Metadata

Assignees

Labels

apiissues related to all other APIs: C, C++, Python, etc.staleissues that have not been addressed in a while; categorized by a bot

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions