Skip to content

[Web] WebGPU backend fails to load some model due to exception during initialization inside transpose optimizer #15869

@gegogi

Description

@gegogi

Describe the issue

I am trying to load a model on WebGPU backend env.
I could load the model downloaded from:
https://github.com/onnx/models/blob/main/vision/classification/mobilenet/model/mobilenetv2-12.onnx
But I couldn't load the following model:
https://huggingface.co/runwayml/stable-diffusion-v1-5/tree/onnx/vae_encoder
Both models can be loaded using Python onnxruntime.

To reproduce

Download the model from:
https://huggingface.co/runwayml/stable-diffusion-v1-5/tree/onnx/vae_encoder
and run the following code:

const ort = require('onnxruntime-web/webgpu');
async function main() {
        const modelPath = './models/sd15_vae_encoder_model.onnx';
        const session = await ort.InferenceSession.create(modelPath, {executionProviders: ['webgpu']});
}

Urgency

No response

ONNX Runtime Installation

Released Package

ONNX Runtime Version or Commit ID

[email protected]

Execution Provider

Other / Unknown

Metadata

Metadata

Assignees

No one assigned

    Labels

    model:transformerissues related to a transformer model: BERT, GPT2, Hugging Face, Longformer, T5, etc.platform:webissues related to ONNX Runtime web; typically submitted using template

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions