..
|
attention.prototxt
|
attention_dynamic_shapes.prototxt
|
attention_extra_add.prototxt
|
attention_mask_index_1.prototxt
|
attention_mask_index_2.prototxt
|
attention_mask_index_3.prototxt
|
attention_mask_index_4.prototxt
|
attention_past.prototxt
|
attention_qkv_hidden_sizes.prototxt
|
attention_unidirectional.prototxt
|
bias_gelu.prototxt
|
dynamic_quantize_matmul.prototxt
|
dynamic_quantize_matmul_bias.prototxt
|
embed_layer_normalization.prototxt
|
embed_layer_normalization_dynamic_shapes.prototxt
|
embed_layer_normalization_dynamic_shapes_with_position_ids.prototxt
|
embed_layer_normalization_with_segment_embedding.prototxt
|
embed_layer_normalization_with_segment_embedding_and_mask.prototxt
|
fused_conv_clip.prototxt
|
fused_conv_hard_sigmoid.prototxt
|
fused_conv_leaky_relu.prototxt
|
fused_conv_relu.prototxt
|
fused_conv_relu_z_input.prototxt
|
fused_conv_sigmoid.prototxt
|
fused_conv_tanh.prototxt
|
fusedgemm.prototxt
|
fusedmatmul_2D.prototxt
|
fusedmatmul_trans_and_transbatch_enabled.prototxt
|
gather_nd.prototxt
|
matmul_integer_to_float.prototxt
|
matmulnbits_3x17.prototxt
|
matmulnbits_3x4.prototxt
|
pad_1d.prototxt
|
pad_2d.prototxt
|
q_linear_add.prototxt
|
q_linear_leaky_relu.prototxt
|
q_linear_mul.prototxt
|
q_linear_sigmoid.prototxt
|
quick_gelu.prototxt
|
range_with_delta.prototxt
|
range_without_delta.prototxt
|
simplified_layer_normalization.prototxt
|
simplified_layer_normalization_2x2x8.prototxt
|
simplified_layer_normalization_3x8.prototxt
|
skip_layer_normalization_dynamic_shapes.prototxt
|
skip_layer_normalization_with_gamma.prototxt
|
skip_layer_normalization_with_gamma_beta.prototxt
|
skip_layer_normalization_with_gamma_beta_bias.prototxt
|
skip_simplified_layer_normalization.prototxt
|
trilu_lower.prototxt
|