diff --git "a/original/compiled/TextEncoder.mlmodelc/model.mil" "b/original/compiled/TextEncoder.mlmodelc/model.mil" new file mode 100644--- /dev/null +++ "b/original/compiled/TextEncoder.mlmodelc/model.mil" @@ -0,0 +1,1642 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3304.5.2"}, {"coremlc-version", "3304.6.2"}, {"coremltools-component-torch", "2.1.2"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "7.1"}})] +{ + func main(tensor input_ids) { + tensor var_5 = const()[name = tensor("op_5"), val = tensor(-1)]; + tensor var_6 = const()[name = tensor("op_6"), val = tensor(false)]; + tensor cast_1_dtype_0 = const()[name = tensor("cast_1_dtype_0"), val = tensor("int32")]; + tensor inputs_embeds_axis_0 = const()[name = tensor("inputs_embeds_axis_0"), val = tensor(0)]; + tensor inputs_embeds_batch_dims_0 = const()[name = tensor("inputs_embeds_batch_dims_0"), val = tensor(0)]; + tensor text_encoder_text_model_embeddings_token_embedding_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_embeddings_token_embedding_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor cast_239 = cast(dtype = cast_1_dtype_0, x = input_ids)[name = tensor("cast_239")]; + tensor inputs_embeds_cast_fp16 = gather(axis = inputs_embeds_axis_0, batch_dims = inputs_embeds_batch_dims_0, indices = cast_239, x = text_encoder_text_model_embeddings_token_embedding_weight_to_fp16)[name = tensor("inputs_embeds_cast_fp16")]; + tensor position_embeddings_to_fp16 = const()[name = tensor("position_embeddings_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101187712)))]; + tensor input_3_cast_fp16 = add(x = inputs_embeds_cast_fp16, y = position_embeddings_to_fp16)[name = tensor("input_3_cast_fp16")]; + tensor hidden_states_1_axes_0 = const()[name = tensor("hidden_states_1_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101345472)))]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101347584)))]; + tensor var_15_to_fp16 = const()[name = tensor("op_15_to_fp16"), val = tensor(0x1.5p-17)]; + tensor hidden_states_1_cast_fp16 = layer_norm(axes = hidden_states_1_axes_0, beta = text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("hidden_states_1_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101349696)))]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103446912)))]; + tensor linear_0_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16, x = hidden_states_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; + tensor var_129_to_fp16 = const()[name = tensor("op_129_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_5_cast_fp16 = mul(x = linear_0_cast_fp16, y = var_129_to_fp16)[name = tensor("tensor_5_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103449024)))]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105546240)))]; + tensor linear_1_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16, x = hidden_states_1_cast_fp16)[name = tensor("linear_1_cast_fp16")]; + tensor var_134 = const()[name = tensor("op_134"), val = tensor([1, -1, 16, 64])]; + tensor var_135_cast_fp16 = reshape(shape = var_134, x = linear_1_cast_fp16)[name = tensor("op_135_cast_fp16")]; + tensor var_136_perm_0 = const()[name = tensor("op_136_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105548352)))]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107645568)))]; + tensor linear_2_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16, x = hidden_states_1_cast_fp16)[name = tensor("linear_2_cast_fp16")]; + tensor var_141 = const()[name = tensor("op_141"), val = tensor([1, -1, 16, 64])]; + tensor var_142_cast_fp16 = reshape(shape = var_141, x = linear_2_cast_fp16)[name = tensor("op_142_cast_fp16")]; + tensor var_143_perm_0 = const()[name = tensor("op_143_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_150 = const()[name = tensor("op_150"), val = tensor([1, 77, 16, 64])]; + tensor var_151_cast_fp16 = reshape(shape = var_150, x = tensor_5_cast_fp16)[name = tensor("op_151_cast_fp16")]; + tensor var_152_perm_0 = const()[name = tensor("op_152_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_154 = const()[name = tensor("op_154"), val = tensor([16, -1, 64])]; + tensor transpose_113 = transpose(perm = var_152_perm_0, x = var_151_cast_fp16)[name = tensor("transpose_113")]; + tensor query_states_1_cast_fp16 = reshape(shape = var_154, x = transpose_113)[name = tensor("query_states_1_cast_fp16")]; + tensor var_156 = const()[name = tensor("op_156"), val = tensor([16, -1, 64])]; + tensor transpose_115 = transpose(perm = var_136_perm_0, x = var_135_cast_fp16)[name = tensor("transpose_115")]; + tensor key_states_3_cast_fp16 = reshape(shape = var_156, x = transpose_115)[name = tensor("key_states_3_cast_fp16")]; + tensor var_158 = const()[name = tensor("op_158"), val = tensor([16, -1, 64])]; + tensor transpose_114 = transpose(perm = var_143_perm_0, x = var_142_cast_fp16)[name = tensor("transpose_114")]; + tensor value_states_3_cast_fp16 = reshape(shape = var_158, x = transpose_114)[name = tensor("value_states_3_cast_fp16")]; + tensor var_161_perm_0 = const()[name = tensor("op_161_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_1_transpose_x_0 = const()[name = tensor("attn_weights_1_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_1_transpose_y_0 = const()[name = tensor("attn_weights_1_transpose_y_0"), val = tensor(false)]; + tensor transpose_112 = transpose(perm = var_161_perm_0, x = key_states_3_cast_fp16)[name = tensor("transpose_112")]; + tensor attn_weights_1_cast_fp16 = matmul(transpose_x = attn_weights_1_transpose_x_0, transpose_y = attn_weights_1_transpose_y_0, x = query_states_1_cast_fp16, y = transpose_112)[name = tensor("attn_weights_1_cast_fp16")]; + tensor var_163 = const()[name = tensor("op_163"), val = tensor([1, 16, 77, 77])]; + tensor var_164_cast_fp16 = reshape(shape = var_163, x = attn_weights_1_cast_fp16)[name = tensor("op_164_cast_fp16")]; + tensor var_57_to_fp16 = const()[name = tensor("op_57_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107647680)))]; + tensor attn_weights_3_cast_fp16 = add(x = var_164_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_3_cast_fp16")]; + tensor var_169 = const()[name = tensor("op_169"), val = tensor([16, 77, 77])]; + tensor input_5_cast_fp16 = reshape(shape = var_169, x = attn_weights_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; + tensor input_7_cast_fp16 = softmax(axis = var_5, x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; + tensor attn_output_1_transpose_x_0 = const()[name = tensor("attn_output_1_transpose_x_0"), val = tensor(false)]; + tensor attn_output_1_transpose_y_0 = const()[name = tensor("attn_output_1_transpose_y_0"), val = tensor(false)]; + tensor attn_output_1_cast_fp16 = matmul(transpose_x = attn_output_1_transpose_x_0, transpose_y = attn_output_1_transpose_y_0, x = input_7_cast_fp16, y = value_states_3_cast_fp16)[name = tensor("attn_output_1_cast_fp16")]; + tensor var_174 = const()[name = tensor("op_174"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_3_cast_fp16 = reshape(shape = var_174, x = attn_output_1_cast_fp16)[name = tensor("attn_output_3_cast_fp16")]; + tensor attn_output_5_perm_0 = const()[name = tensor("attn_output_5_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_177 = const()[name = tensor("op_177"), val = tensor([1, 77, 1024])]; + tensor transpose_111 = transpose(perm = attn_output_5_perm_0, x = attn_output_3_cast_fp16)[name = tensor("transpose_111")]; + tensor input_9_cast_fp16 = reshape(shape = var_177, x = transpose_111)[name = tensor("input_9_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107659648)))]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109756864)))]; + tensor linear_3_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16, x = input_9_cast_fp16)[name = tensor("linear_3_cast_fp16")]; + tensor input_11_cast_fp16 = add(x = input_3_cast_fp16, y = linear_3_cast_fp16)[name = tensor("input_11_cast_fp16")]; + tensor input_13_axes_0 = const()[name = tensor("input_13_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109758976)))]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109761088)))]; + tensor input_13_cast_fp16 = layer_norm(axes = input_13_axes_0, beta = text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109763200)))]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(118151872)))]; + tensor linear_4_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16, x = input_13_cast_fp16)[name = tensor("linear_4_cast_fp16")]; + tensor input_17_mode_0 = const()[name = tensor("input_17_mode_0"), val = tensor("EXACT")]; + tensor input_17_cast_fp16 = gelu(mode = input_17_mode_0, x = linear_4_cast_fp16)[name = tensor("input_17_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(118160128)))]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126548800)))]; + tensor linear_5_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16, x = input_17_cast_fp16)[name = tensor("linear_5_cast_fp16")]; + tensor input_19_cast_fp16 = add(x = input_11_cast_fp16, y = linear_5_cast_fp16)[name = tensor("input_19_cast_fp16")]; + tensor hidden_states_7_axes_0 = const()[name = tensor("hidden_states_7_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126550912)))]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126553024)))]; + tensor hidden_states_7_cast_fp16 = layer_norm(axes = hidden_states_7_axes_0, beta = text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16, x = input_19_cast_fp16)[name = tensor("hidden_states_7_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126555136)))]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128652352)))]; + tensor linear_6_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16, x = hidden_states_7_cast_fp16)[name = tensor("linear_6_cast_fp16")]; + tensor var_216_to_fp16 = const()[name = tensor("op_216_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_11_cast_fp16 = mul(x = linear_6_cast_fp16, y = var_216_to_fp16)[name = tensor("tensor_11_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128654464)))]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130751680)))]; + tensor linear_7_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16, x = hidden_states_7_cast_fp16)[name = tensor("linear_7_cast_fp16")]; + tensor var_221 = const()[name = tensor("op_221"), val = tensor([1, -1, 16, 64])]; + tensor var_222_cast_fp16 = reshape(shape = var_221, x = linear_7_cast_fp16)[name = tensor("op_222_cast_fp16")]; + tensor var_223_perm_0 = const()[name = tensor("op_223_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130753792)))]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132851008)))]; + tensor linear_8_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16, x = hidden_states_7_cast_fp16)[name = tensor("linear_8_cast_fp16")]; + tensor var_228 = const()[name = tensor("op_228"), val = tensor([1, -1, 16, 64])]; + tensor var_229_cast_fp16 = reshape(shape = var_228, x = linear_8_cast_fp16)[name = tensor("op_229_cast_fp16")]; + tensor var_230_perm_0 = const()[name = tensor("op_230_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_237 = const()[name = tensor("op_237"), val = tensor([1, 77, 16, 64])]; + tensor var_238_cast_fp16 = reshape(shape = var_237, x = tensor_11_cast_fp16)[name = tensor("op_238_cast_fp16")]; + tensor var_239_perm_0 = const()[name = tensor("op_239_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_241 = const()[name = tensor("op_241"), val = tensor([16, -1, 64])]; + tensor transpose_108 = transpose(perm = var_239_perm_0, x = var_238_cast_fp16)[name = tensor("transpose_108")]; + tensor query_states_3_cast_fp16 = reshape(shape = var_241, x = transpose_108)[name = tensor("query_states_3_cast_fp16")]; + tensor var_243 = const()[name = tensor("op_243"), val = tensor([16, -1, 64])]; + tensor transpose_110 = transpose(perm = var_223_perm_0, x = var_222_cast_fp16)[name = tensor("transpose_110")]; + tensor key_states_7_cast_fp16 = reshape(shape = var_243, x = transpose_110)[name = tensor("key_states_7_cast_fp16")]; + tensor var_245 = const()[name = tensor("op_245"), val = tensor([16, -1, 64])]; + tensor transpose_109 = transpose(perm = var_230_perm_0, x = var_229_cast_fp16)[name = tensor("transpose_109")]; + tensor value_states_7_cast_fp16 = reshape(shape = var_245, x = transpose_109)[name = tensor("value_states_7_cast_fp16")]; + tensor var_248_perm_0 = const()[name = tensor("op_248_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_7_transpose_x_0 = const()[name = tensor("attn_weights_7_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_7_transpose_y_0 = const()[name = tensor("attn_weights_7_transpose_y_0"), val = tensor(false)]; + tensor transpose_107 = transpose(perm = var_248_perm_0, x = key_states_7_cast_fp16)[name = tensor("transpose_107")]; + tensor attn_weights_7_cast_fp16 = matmul(transpose_x = attn_weights_7_transpose_x_0, transpose_y = attn_weights_7_transpose_y_0, x = query_states_3_cast_fp16, y = transpose_107)[name = tensor("attn_weights_7_cast_fp16")]; + tensor var_250 = const()[name = tensor("op_250"), val = tensor([1, 16, 77, 77])]; + tensor var_251_cast_fp16 = reshape(shape = var_250, x = attn_weights_7_cast_fp16)[name = tensor("op_251_cast_fp16")]; + tensor attn_weights_9_cast_fp16 = add(x = var_251_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_9_cast_fp16")]; + tensor var_256 = const()[name = tensor("op_256"), val = tensor([16, 77, 77])]; + tensor input_21_cast_fp16 = reshape(shape = var_256, x = attn_weights_9_cast_fp16)[name = tensor("input_21_cast_fp16")]; + tensor input_23_cast_fp16 = softmax(axis = var_5, x = input_21_cast_fp16)[name = tensor("input_23_cast_fp16")]; + tensor attn_output_7_transpose_x_0 = const()[name = tensor("attn_output_7_transpose_x_0"), val = tensor(false)]; + tensor attn_output_7_transpose_y_0 = const()[name = tensor("attn_output_7_transpose_y_0"), val = tensor(false)]; + tensor attn_output_7_cast_fp16 = matmul(transpose_x = attn_output_7_transpose_x_0, transpose_y = attn_output_7_transpose_y_0, x = input_23_cast_fp16, y = value_states_7_cast_fp16)[name = tensor("attn_output_7_cast_fp16")]; + tensor var_261 = const()[name = tensor("op_261"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_9_cast_fp16 = reshape(shape = var_261, x = attn_output_7_cast_fp16)[name = tensor("attn_output_9_cast_fp16")]; + tensor attn_output_11_perm_0 = const()[name = tensor("attn_output_11_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_264 = const()[name = tensor("op_264"), val = tensor([1, 77, 1024])]; + tensor transpose_106 = transpose(perm = attn_output_11_perm_0, x = attn_output_9_cast_fp16)[name = tensor("transpose_106")]; + tensor input_25_cast_fp16 = reshape(shape = var_264, x = transpose_106)[name = tensor("input_25_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132853120)))]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134950336)))]; + tensor linear_9_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16, x = input_25_cast_fp16)[name = tensor("linear_9_cast_fp16")]; + tensor input_27_cast_fp16 = add(x = input_19_cast_fp16, y = linear_9_cast_fp16)[name = tensor("input_27_cast_fp16")]; + tensor input_29_axes_0 = const()[name = tensor("input_29_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134952448)))]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134954560)))]; + tensor input_29_cast_fp16 = layer_norm(axes = input_29_axes_0, beta = text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16, x = input_27_cast_fp16)[name = tensor("input_29_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134956672)))]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143345344)))]; + tensor linear_10_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16, x = input_29_cast_fp16)[name = tensor("linear_10_cast_fp16")]; + tensor input_33_mode_0 = const()[name = tensor("input_33_mode_0"), val = tensor("EXACT")]; + tensor input_33_cast_fp16 = gelu(mode = input_33_mode_0, x = linear_10_cast_fp16)[name = tensor("input_33_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143353600)))]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151742272)))]; + tensor linear_11_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16, x = input_33_cast_fp16)[name = tensor("linear_11_cast_fp16")]; + tensor input_35_cast_fp16 = add(x = input_27_cast_fp16, y = linear_11_cast_fp16)[name = tensor("input_35_cast_fp16")]; + tensor hidden_states_13_axes_0 = const()[name = tensor("hidden_states_13_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151744384)))]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151746496)))]; + tensor hidden_states_13_cast_fp16 = layer_norm(axes = hidden_states_13_axes_0, beta = text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16, x = input_35_cast_fp16)[name = tensor("hidden_states_13_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151748608)))]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153845824)))]; + tensor linear_12_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16, x = hidden_states_13_cast_fp16)[name = tensor("linear_12_cast_fp16")]; + tensor var_303_to_fp16 = const()[name = tensor("op_303_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_17_cast_fp16 = mul(x = linear_12_cast_fp16, y = var_303_to_fp16)[name = tensor("tensor_17_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153847936)))]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155945152)))]; + tensor linear_13_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16, x = hidden_states_13_cast_fp16)[name = tensor("linear_13_cast_fp16")]; + tensor var_308 = const()[name = tensor("op_308"), val = tensor([1, -1, 16, 64])]; + tensor var_309_cast_fp16 = reshape(shape = var_308, x = linear_13_cast_fp16)[name = tensor("op_309_cast_fp16")]; + tensor var_310_perm_0 = const()[name = tensor("op_310_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155947264)))]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158044480)))]; + tensor linear_14_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16, x = hidden_states_13_cast_fp16)[name = tensor("linear_14_cast_fp16")]; + tensor var_315 = const()[name = tensor("op_315"), val = tensor([1, -1, 16, 64])]; + tensor var_316_cast_fp16 = reshape(shape = var_315, x = linear_14_cast_fp16)[name = tensor("op_316_cast_fp16")]; + tensor var_317_perm_0 = const()[name = tensor("op_317_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_324 = const()[name = tensor("op_324"), val = tensor([1, 77, 16, 64])]; + tensor var_325_cast_fp16 = reshape(shape = var_324, x = tensor_17_cast_fp16)[name = tensor("op_325_cast_fp16")]; + tensor var_326_perm_0 = const()[name = tensor("op_326_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_328 = const()[name = tensor("op_328"), val = tensor([16, -1, 64])]; + tensor transpose_103 = transpose(perm = var_326_perm_0, x = var_325_cast_fp16)[name = tensor("transpose_103")]; + tensor query_states_5_cast_fp16 = reshape(shape = var_328, x = transpose_103)[name = tensor("query_states_5_cast_fp16")]; + tensor var_330 = const()[name = tensor("op_330"), val = tensor([16, -1, 64])]; + tensor transpose_105 = transpose(perm = var_310_perm_0, x = var_309_cast_fp16)[name = tensor("transpose_105")]; + tensor key_states_11_cast_fp16 = reshape(shape = var_330, x = transpose_105)[name = tensor("key_states_11_cast_fp16")]; + tensor var_332 = const()[name = tensor("op_332"), val = tensor([16, -1, 64])]; + tensor transpose_104 = transpose(perm = var_317_perm_0, x = var_316_cast_fp16)[name = tensor("transpose_104")]; + tensor value_states_11_cast_fp16 = reshape(shape = var_332, x = transpose_104)[name = tensor("value_states_11_cast_fp16")]; + tensor var_335_perm_0 = const()[name = tensor("op_335_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_13_transpose_x_0 = const()[name = tensor("attn_weights_13_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_13_transpose_y_0 = const()[name = tensor("attn_weights_13_transpose_y_0"), val = tensor(false)]; + tensor transpose_102 = transpose(perm = var_335_perm_0, x = key_states_11_cast_fp16)[name = tensor("transpose_102")]; + tensor attn_weights_13_cast_fp16 = matmul(transpose_x = attn_weights_13_transpose_x_0, transpose_y = attn_weights_13_transpose_y_0, x = query_states_5_cast_fp16, y = transpose_102)[name = tensor("attn_weights_13_cast_fp16")]; + tensor var_337 = const()[name = tensor("op_337"), val = tensor([1, 16, 77, 77])]; + tensor var_338_cast_fp16 = reshape(shape = var_337, x = attn_weights_13_cast_fp16)[name = tensor("op_338_cast_fp16")]; + tensor attn_weights_15_cast_fp16 = add(x = var_338_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_15_cast_fp16")]; + tensor var_343 = const()[name = tensor("op_343"), val = tensor([16, 77, 77])]; + tensor input_37_cast_fp16 = reshape(shape = var_343, x = attn_weights_15_cast_fp16)[name = tensor("input_37_cast_fp16")]; + tensor input_39_cast_fp16 = softmax(axis = var_5, x = input_37_cast_fp16)[name = tensor("input_39_cast_fp16")]; + tensor attn_output_13_transpose_x_0 = const()[name = tensor("attn_output_13_transpose_x_0"), val = tensor(false)]; + tensor attn_output_13_transpose_y_0 = const()[name = tensor("attn_output_13_transpose_y_0"), val = tensor(false)]; + tensor attn_output_13_cast_fp16 = matmul(transpose_x = attn_output_13_transpose_x_0, transpose_y = attn_output_13_transpose_y_0, x = input_39_cast_fp16, y = value_states_11_cast_fp16)[name = tensor("attn_output_13_cast_fp16")]; + tensor var_348 = const()[name = tensor("op_348"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_15_cast_fp16 = reshape(shape = var_348, x = attn_output_13_cast_fp16)[name = tensor("attn_output_15_cast_fp16")]; + tensor attn_output_17_perm_0 = const()[name = tensor("attn_output_17_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_351 = const()[name = tensor("op_351"), val = tensor([1, 77, 1024])]; + tensor transpose_101 = transpose(perm = attn_output_17_perm_0, x = attn_output_15_cast_fp16)[name = tensor("transpose_101")]; + tensor input_41_cast_fp16 = reshape(shape = var_351, x = transpose_101)[name = tensor("input_41_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158046592)))]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160143808)))]; + tensor linear_15_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16, x = input_41_cast_fp16)[name = tensor("linear_15_cast_fp16")]; + tensor input_43_cast_fp16 = add(x = input_35_cast_fp16, y = linear_15_cast_fp16)[name = tensor("input_43_cast_fp16")]; + tensor input_45_axes_0 = const()[name = tensor("input_45_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160145920)))]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160148032)))]; + tensor input_45_cast_fp16 = layer_norm(axes = input_45_axes_0, beta = text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16, x = input_43_cast_fp16)[name = tensor("input_45_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160150144)))]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168538816)))]; + tensor linear_16_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16, x = input_45_cast_fp16)[name = tensor("linear_16_cast_fp16")]; + tensor input_49_mode_0 = const()[name = tensor("input_49_mode_0"), val = tensor("EXACT")]; + tensor input_49_cast_fp16 = gelu(mode = input_49_mode_0, x = linear_16_cast_fp16)[name = tensor("input_49_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168547072)))]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176935744)))]; + tensor linear_17_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16, x = input_49_cast_fp16)[name = tensor("linear_17_cast_fp16")]; + tensor input_51_cast_fp16 = add(x = input_43_cast_fp16, y = linear_17_cast_fp16)[name = tensor("input_51_cast_fp16")]; + tensor hidden_states_19_axes_0 = const()[name = tensor("hidden_states_19_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176937856)))]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176939968)))]; + tensor hidden_states_19_cast_fp16 = layer_norm(axes = hidden_states_19_axes_0, beta = text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16, x = input_51_cast_fp16)[name = tensor("hidden_states_19_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176942080)))]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179039296)))]; + tensor linear_18_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16, x = hidden_states_19_cast_fp16)[name = tensor("linear_18_cast_fp16")]; + tensor var_390_to_fp16 = const()[name = tensor("op_390_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_23_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_390_to_fp16)[name = tensor("tensor_23_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179041408)))]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181138624)))]; + tensor linear_19_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16, x = hidden_states_19_cast_fp16)[name = tensor("linear_19_cast_fp16")]; + tensor var_395 = const()[name = tensor("op_395"), val = tensor([1, -1, 16, 64])]; + tensor var_396_cast_fp16 = reshape(shape = var_395, x = linear_19_cast_fp16)[name = tensor("op_396_cast_fp16")]; + tensor var_397_perm_0 = const()[name = tensor("op_397_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181140736)))]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183237952)))]; + tensor linear_20_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16, x = hidden_states_19_cast_fp16)[name = tensor("linear_20_cast_fp16")]; + tensor var_402 = const()[name = tensor("op_402"), val = tensor([1, -1, 16, 64])]; + tensor var_403_cast_fp16 = reshape(shape = var_402, x = linear_20_cast_fp16)[name = tensor("op_403_cast_fp16")]; + tensor var_404_perm_0 = const()[name = tensor("op_404_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_411 = const()[name = tensor("op_411"), val = tensor([1, 77, 16, 64])]; + tensor var_412_cast_fp16 = reshape(shape = var_411, x = tensor_23_cast_fp16)[name = tensor("op_412_cast_fp16")]; + tensor var_413_perm_0 = const()[name = tensor("op_413_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_415 = const()[name = tensor("op_415"), val = tensor([16, -1, 64])]; + tensor transpose_98 = transpose(perm = var_413_perm_0, x = var_412_cast_fp16)[name = tensor("transpose_98")]; + tensor query_states_7_cast_fp16 = reshape(shape = var_415, x = transpose_98)[name = tensor("query_states_7_cast_fp16")]; + tensor var_417 = const()[name = tensor("op_417"), val = tensor([16, -1, 64])]; + tensor transpose_100 = transpose(perm = var_397_perm_0, x = var_396_cast_fp16)[name = tensor("transpose_100")]; + tensor key_states_15_cast_fp16 = reshape(shape = var_417, x = transpose_100)[name = tensor("key_states_15_cast_fp16")]; + tensor var_419 = const()[name = tensor("op_419"), val = tensor([16, -1, 64])]; + tensor transpose_99 = transpose(perm = var_404_perm_0, x = var_403_cast_fp16)[name = tensor("transpose_99")]; + tensor value_states_15_cast_fp16 = reshape(shape = var_419, x = transpose_99)[name = tensor("value_states_15_cast_fp16")]; + tensor var_422_perm_0 = const()[name = tensor("op_422_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_19_transpose_x_0 = const()[name = tensor("attn_weights_19_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_19_transpose_y_0 = const()[name = tensor("attn_weights_19_transpose_y_0"), val = tensor(false)]; + tensor transpose_97 = transpose(perm = var_422_perm_0, x = key_states_15_cast_fp16)[name = tensor("transpose_97")]; + tensor attn_weights_19_cast_fp16 = matmul(transpose_x = attn_weights_19_transpose_x_0, transpose_y = attn_weights_19_transpose_y_0, x = query_states_7_cast_fp16, y = transpose_97)[name = tensor("attn_weights_19_cast_fp16")]; + tensor var_424 = const()[name = tensor("op_424"), val = tensor([1, 16, 77, 77])]; + tensor var_425_cast_fp16 = reshape(shape = var_424, x = attn_weights_19_cast_fp16)[name = tensor("op_425_cast_fp16")]; + tensor attn_weights_21_cast_fp16 = add(x = var_425_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_21_cast_fp16")]; + tensor var_430 = const()[name = tensor("op_430"), val = tensor([16, 77, 77])]; + tensor input_53_cast_fp16 = reshape(shape = var_430, x = attn_weights_21_cast_fp16)[name = tensor("input_53_cast_fp16")]; + tensor input_55_cast_fp16 = softmax(axis = var_5, x = input_53_cast_fp16)[name = tensor("input_55_cast_fp16")]; + tensor attn_output_19_transpose_x_0 = const()[name = tensor("attn_output_19_transpose_x_0"), val = tensor(false)]; + tensor attn_output_19_transpose_y_0 = const()[name = tensor("attn_output_19_transpose_y_0"), val = tensor(false)]; + tensor attn_output_19_cast_fp16 = matmul(transpose_x = attn_output_19_transpose_x_0, transpose_y = attn_output_19_transpose_y_0, x = input_55_cast_fp16, y = value_states_15_cast_fp16)[name = tensor("attn_output_19_cast_fp16")]; + tensor var_435 = const()[name = tensor("op_435"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_21_cast_fp16 = reshape(shape = var_435, x = attn_output_19_cast_fp16)[name = tensor("attn_output_21_cast_fp16")]; + tensor attn_output_23_perm_0 = const()[name = tensor("attn_output_23_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_438 = const()[name = tensor("op_438"), val = tensor([1, 77, 1024])]; + tensor transpose_96 = transpose(perm = attn_output_23_perm_0, x = attn_output_21_cast_fp16)[name = tensor("transpose_96")]; + tensor input_57_cast_fp16 = reshape(shape = var_438, x = transpose_96)[name = tensor("input_57_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183240064)))]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185337280)))]; + tensor linear_21_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16, x = input_57_cast_fp16)[name = tensor("linear_21_cast_fp16")]; + tensor input_59_cast_fp16 = add(x = input_51_cast_fp16, y = linear_21_cast_fp16)[name = tensor("input_59_cast_fp16")]; + tensor input_61_axes_0 = const()[name = tensor("input_61_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185339392)))]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185341504)))]; + tensor input_61_cast_fp16 = layer_norm(axes = input_61_axes_0, beta = text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("input_61_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185343616)))]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193732288)))]; + tensor linear_22_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16, x = input_61_cast_fp16)[name = tensor("linear_22_cast_fp16")]; + tensor input_65_mode_0 = const()[name = tensor("input_65_mode_0"), val = tensor("EXACT")]; + tensor input_65_cast_fp16 = gelu(mode = input_65_mode_0, x = linear_22_cast_fp16)[name = tensor("input_65_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193740544)))]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202129216)))]; + tensor linear_23_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16, x = input_65_cast_fp16)[name = tensor("linear_23_cast_fp16")]; + tensor input_67_cast_fp16 = add(x = input_59_cast_fp16, y = linear_23_cast_fp16)[name = tensor("input_67_cast_fp16")]; + tensor hidden_states_25_axes_0 = const()[name = tensor("hidden_states_25_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202131328)))]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202133440)))]; + tensor hidden_states_25_cast_fp16 = layer_norm(axes = hidden_states_25_axes_0, beta = text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16, x = input_67_cast_fp16)[name = tensor("hidden_states_25_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202135552)))]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204232768)))]; + tensor linear_24_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16, x = hidden_states_25_cast_fp16)[name = tensor("linear_24_cast_fp16")]; + tensor var_477_to_fp16 = const()[name = tensor("op_477_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_29_cast_fp16 = mul(x = linear_24_cast_fp16, y = var_477_to_fp16)[name = tensor("tensor_29_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204234880)))]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206332096)))]; + tensor linear_25_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16, x = hidden_states_25_cast_fp16)[name = tensor("linear_25_cast_fp16")]; + tensor var_482 = const()[name = tensor("op_482"), val = tensor([1, -1, 16, 64])]; + tensor var_483_cast_fp16 = reshape(shape = var_482, x = linear_25_cast_fp16)[name = tensor("op_483_cast_fp16")]; + tensor var_484_perm_0 = const()[name = tensor("op_484_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206334208)))]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208431424)))]; + tensor linear_26_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16, x = hidden_states_25_cast_fp16)[name = tensor("linear_26_cast_fp16")]; + tensor var_489 = const()[name = tensor("op_489"), val = tensor([1, -1, 16, 64])]; + tensor var_490_cast_fp16 = reshape(shape = var_489, x = linear_26_cast_fp16)[name = tensor("op_490_cast_fp16")]; + tensor var_491_perm_0 = const()[name = tensor("op_491_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_498 = const()[name = tensor("op_498"), val = tensor([1, 77, 16, 64])]; + tensor var_499_cast_fp16 = reshape(shape = var_498, x = tensor_29_cast_fp16)[name = tensor("op_499_cast_fp16")]; + tensor var_500_perm_0 = const()[name = tensor("op_500_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_502 = const()[name = tensor("op_502"), val = tensor([16, -1, 64])]; + tensor transpose_93 = transpose(perm = var_500_perm_0, x = var_499_cast_fp16)[name = tensor("transpose_93")]; + tensor query_states_9_cast_fp16 = reshape(shape = var_502, x = transpose_93)[name = tensor("query_states_9_cast_fp16")]; + tensor var_504 = const()[name = tensor("op_504"), val = tensor([16, -1, 64])]; + tensor transpose_95 = transpose(perm = var_484_perm_0, x = var_483_cast_fp16)[name = tensor("transpose_95")]; + tensor key_states_19_cast_fp16 = reshape(shape = var_504, x = transpose_95)[name = tensor("key_states_19_cast_fp16")]; + tensor var_506 = const()[name = tensor("op_506"), val = tensor([16, -1, 64])]; + tensor transpose_94 = transpose(perm = var_491_perm_0, x = var_490_cast_fp16)[name = tensor("transpose_94")]; + tensor value_states_19_cast_fp16 = reshape(shape = var_506, x = transpose_94)[name = tensor("value_states_19_cast_fp16")]; + tensor var_509_perm_0 = const()[name = tensor("op_509_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_25_transpose_x_0 = const()[name = tensor("attn_weights_25_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_25_transpose_y_0 = const()[name = tensor("attn_weights_25_transpose_y_0"), val = tensor(false)]; + tensor transpose_92 = transpose(perm = var_509_perm_0, x = key_states_19_cast_fp16)[name = tensor("transpose_92")]; + tensor attn_weights_25_cast_fp16 = matmul(transpose_x = attn_weights_25_transpose_x_0, transpose_y = attn_weights_25_transpose_y_0, x = query_states_9_cast_fp16, y = transpose_92)[name = tensor("attn_weights_25_cast_fp16")]; + tensor var_511 = const()[name = tensor("op_511"), val = tensor([1, 16, 77, 77])]; + tensor var_512_cast_fp16 = reshape(shape = var_511, x = attn_weights_25_cast_fp16)[name = tensor("op_512_cast_fp16")]; + tensor attn_weights_27_cast_fp16 = add(x = var_512_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_27_cast_fp16")]; + tensor var_517 = const()[name = tensor("op_517"), val = tensor([16, 77, 77])]; + tensor input_69_cast_fp16 = reshape(shape = var_517, x = attn_weights_27_cast_fp16)[name = tensor("input_69_cast_fp16")]; + tensor input_71_cast_fp16 = softmax(axis = var_5, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; + tensor attn_output_25_transpose_x_0 = const()[name = tensor("attn_output_25_transpose_x_0"), val = tensor(false)]; + tensor attn_output_25_transpose_y_0 = const()[name = tensor("attn_output_25_transpose_y_0"), val = tensor(false)]; + tensor attn_output_25_cast_fp16 = matmul(transpose_x = attn_output_25_transpose_x_0, transpose_y = attn_output_25_transpose_y_0, x = input_71_cast_fp16, y = value_states_19_cast_fp16)[name = tensor("attn_output_25_cast_fp16")]; + tensor var_522 = const()[name = tensor("op_522"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_27_cast_fp16 = reshape(shape = var_522, x = attn_output_25_cast_fp16)[name = tensor("attn_output_27_cast_fp16")]; + tensor attn_output_29_perm_0 = const()[name = tensor("attn_output_29_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_525 = const()[name = tensor("op_525"), val = tensor([1, 77, 1024])]; + tensor transpose_91 = transpose(perm = attn_output_29_perm_0, x = attn_output_27_cast_fp16)[name = tensor("transpose_91")]; + tensor input_73_cast_fp16 = reshape(shape = var_525, x = transpose_91)[name = tensor("input_73_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208433536)))]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210530752)))]; + tensor linear_27_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16, x = input_73_cast_fp16)[name = tensor("linear_27_cast_fp16")]; + tensor input_75_cast_fp16 = add(x = input_67_cast_fp16, y = linear_27_cast_fp16)[name = tensor("input_75_cast_fp16")]; + tensor input_77_axes_0 = const()[name = tensor("input_77_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210532864)))]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210534976)))]; + tensor input_77_cast_fp16 = layer_norm(axes = input_77_axes_0, beta = text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16, x = input_75_cast_fp16)[name = tensor("input_77_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210537088)))]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218925760)))]; + tensor linear_28_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("linear_28_cast_fp16")]; + tensor input_81_mode_0 = const()[name = tensor("input_81_mode_0"), val = tensor("EXACT")]; + tensor input_81_cast_fp16 = gelu(mode = input_81_mode_0, x = linear_28_cast_fp16)[name = tensor("input_81_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218934016)))]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227322688)))]; + tensor linear_29_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16, x = input_81_cast_fp16)[name = tensor("linear_29_cast_fp16")]; + tensor input_83_cast_fp16 = add(x = input_75_cast_fp16, y = linear_29_cast_fp16)[name = tensor("input_83_cast_fp16")]; + tensor hidden_states_31_axes_0 = const()[name = tensor("hidden_states_31_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227324800)))]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227326912)))]; + tensor hidden_states_31_cast_fp16 = layer_norm(axes = hidden_states_31_axes_0, beta = text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("hidden_states_31_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227329024)))]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(229426240)))]; + tensor linear_30_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16, x = hidden_states_31_cast_fp16)[name = tensor("linear_30_cast_fp16")]; + tensor var_564_to_fp16 = const()[name = tensor("op_564_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_35_cast_fp16 = mul(x = linear_30_cast_fp16, y = var_564_to_fp16)[name = tensor("tensor_35_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(229428352)))]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231525568)))]; + tensor linear_31_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16, x = hidden_states_31_cast_fp16)[name = tensor("linear_31_cast_fp16")]; + tensor var_569 = const()[name = tensor("op_569"), val = tensor([1, -1, 16, 64])]; + tensor var_570_cast_fp16 = reshape(shape = var_569, x = linear_31_cast_fp16)[name = tensor("op_570_cast_fp16")]; + tensor var_571_perm_0 = const()[name = tensor("op_571_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231527680)))]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233624896)))]; + tensor linear_32_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16, x = hidden_states_31_cast_fp16)[name = tensor("linear_32_cast_fp16")]; + tensor var_576 = const()[name = tensor("op_576"), val = tensor([1, -1, 16, 64])]; + tensor var_577_cast_fp16 = reshape(shape = var_576, x = linear_32_cast_fp16)[name = tensor("op_577_cast_fp16")]; + tensor var_578_perm_0 = const()[name = tensor("op_578_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_585 = const()[name = tensor("op_585"), val = tensor([1, 77, 16, 64])]; + tensor var_586_cast_fp16 = reshape(shape = var_585, x = tensor_35_cast_fp16)[name = tensor("op_586_cast_fp16")]; + tensor var_587_perm_0 = const()[name = tensor("op_587_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_589 = const()[name = tensor("op_589"), val = tensor([16, -1, 64])]; + tensor transpose_88 = transpose(perm = var_587_perm_0, x = var_586_cast_fp16)[name = tensor("transpose_88")]; + tensor query_states_11_cast_fp16 = reshape(shape = var_589, x = transpose_88)[name = tensor("query_states_11_cast_fp16")]; + tensor var_591 = const()[name = tensor("op_591"), val = tensor([16, -1, 64])]; + tensor transpose_90 = transpose(perm = var_571_perm_0, x = var_570_cast_fp16)[name = tensor("transpose_90")]; + tensor key_states_23_cast_fp16 = reshape(shape = var_591, x = transpose_90)[name = tensor("key_states_23_cast_fp16")]; + tensor var_593 = const()[name = tensor("op_593"), val = tensor([16, -1, 64])]; + tensor transpose_89 = transpose(perm = var_578_perm_0, x = var_577_cast_fp16)[name = tensor("transpose_89")]; + tensor value_states_23_cast_fp16 = reshape(shape = var_593, x = transpose_89)[name = tensor("value_states_23_cast_fp16")]; + tensor var_596_perm_0 = const()[name = tensor("op_596_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_31_transpose_x_0 = const()[name = tensor("attn_weights_31_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_31_transpose_y_0 = const()[name = tensor("attn_weights_31_transpose_y_0"), val = tensor(false)]; + tensor transpose_87 = transpose(perm = var_596_perm_0, x = key_states_23_cast_fp16)[name = tensor("transpose_87")]; + tensor attn_weights_31_cast_fp16 = matmul(transpose_x = attn_weights_31_transpose_x_0, transpose_y = attn_weights_31_transpose_y_0, x = query_states_11_cast_fp16, y = transpose_87)[name = tensor("attn_weights_31_cast_fp16")]; + tensor var_598 = const()[name = tensor("op_598"), val = tensor([1, 16, 77, 77])]; + tensor var_599_cast_fp16 = reshape(shape = var_598, x = attn_weights_31_cast_fp16)[name = tensor("op_599_cast_fp16")]; + tensor attn_weights_33_cast_fp16 = add(x = var_599_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_33_cast_fp16")]; + tensor var_604 = const()[name = tensor("op_604"), val = tensor([16, 77, 77])]; + tensor input_85_cast_fp16 = reshape(shape = var_604, x = attn_weights_33_cast_fp16)[name = tensor("input_85_cast_fp16")]; + tensor input_87_cast_fp16 = softmax(axis = var_5, x = input_85_cast_fp16)[name = tensor("input_87_cast_fp16")]; + tensor attn_output_31_transpose_x_0 = const()[name = tensor("attn_output_31_transpose_x_0"), val = tensor(false)]; + tensor attn_output_31_transpose_y_0 = const()[name = tensor("attn_output_31_transpose_y_0"), val = tensor(false)]; + tensor attn_output_31_cast_fp16 = matmul(transpose_x = attn_output_31_transpose_x_0, transpose_y = attn_output_31_transpose_y_0, x = input_87_cast_fp16, y = value_states_23_cast_fp16)[name = tensor("attn_output_31_cast_fp16")]; + tensor var_609 = const()[name = tensor("op_609"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_33_cast_fp16 = reshape(shape = var_609, x = attn_output_31_cast_fp16)[name = tensor("attn_output_33_cast_fp16")]; + tensor attn_output_35_perm_0 = const()[name = tensor("attn_output_35_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_612 = const()[name = tensor("op_612"), val = tensor([1, 77, 1024])]; + tensor transpose_86 = transpose(perm = attn_output_35_perm_0, x = attn_output_33_cast_fp16)[name = tensor("transpose_86")]; + tensor input_89_cast_fp16 = reshape(shape = var_612, x = transpose_86)[name = tensor("input_89_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233627008)))]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235724224)))]; + tensor linear_33_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16, x = input_89_cast_fp16)[name = tensor("linear_33_cast_fp16")]; + tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_33_cast_fp16)[name = tensor("input_91_cast_fp16")]; + tensor input_93_axes_0 = const()[name = tensor("input_93_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235726336)))]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235728448)))]; + tensor input_93_cast_fp16 = layer_norm(axes = input_93_axes_0, beta = text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("input_93_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235730560)))]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244119232)))]; + tensor linear_34_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16, x = input_93_cast_fp16)[name = tensor("linear_34_cast_fp16")]; + tensor input_97_mode_0 = const()[name = tensor("input_97_mode_0"), val = tensor("EXACT")]; + tensor input_97_cast_fp16 = gelu(mode = input_97_mode_0, x = linear_34_cast_fp16)[name = tensor("input_97_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244127488)))]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(252516160)))]; + tensor linear_35_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16, x = input_97_cast_fp16)[name = tensor("linear_35_cast_fp16")]; + tensor input_99_cast_fp16 = add(x = input_91_cast_fp16, y = linear_35_cast_fp16)[name = tensor("input_99_cast_fp16")]; + tensor hidden_states_37_axes_0 = const()[name = tensor("hidden_states_37_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(252518272)))]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(252520384)))]; + tensor hidden_states_37_cast_fp16 = layer_norm(axes = hidden_states_37_axes_0, beta = text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16, x = input_99_cast_fp16)[name = tensor("hidden_states_37_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(252522496)))]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254619712)))]; + tensor linear_36_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16, x = hidden_states_37_cast_fp16)[name = tensor("linear_36_cast_fp16")]; + tensor var_651_to_fp16 = const()[name = tensor("op_651_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_41_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_651_to_fp16)[name = tensor("tensor_41_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254621824)))]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(256719040)))]; + tensor linear_37_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16, x = hidden_states_37_cast_fp16)[name = tensor("linear_37_cast_fp16")]; + tensor var_656 = const()[name = tensor("op_656"), val = tensor([1, -1, 16, 64])]; + tensor var_657_cast_fp16 = reshape(shape = var_656, x = linear_37_cast_fp16)[name = tensor("op_657_cast_fp16")]; + tensor var_658_perm_0 = const()[name = tensor("op_658_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(256721152)))]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258818368)))]; + tensor linear_38_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16, x = hidden_states_37_cast_fp16)[name = tensor("linear_38_cast_fp16")]; + tensor var_663 = const()[name = tensor("op_663"), val = tensor([1, -1, 16, 64])]; + tensor var_664_cast_fp16 = reshape(shape = var_663, x = linear_38_cast_fp16)[name = tensor("op_664_cast_fp16")]; + tensor var_665_perm_0 = const()[name = tensor("op_665_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_672 = const()[name = tensor("op_672"), val = tensor([1, 77, 16, 64])]; + tensor var_673_cast_fp16 = reshape(shape = var_672, x = tensor_41_cast_fp16)[name = tensor("op_673_cast_fp16")]; + tensor var_674_perm_0 = const()[name = tensor("op_674_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_676 = const()[name = tensor("op_676"), val = tensor([16, -1, 64])]; + tensor transpose_83 = transpose(perm = var_674_perm_0, x = var_673_cast_fp16)[name = tensor("transpose_83")]; + tensor query_states_13_cast_fp16 = reshape(shape = var_676, x = transpose_83)[name = tensor("query_states_13_cast_fp16")]; + tensor var_678 = const()[name = tensor("op_678"), val = tensor([16, -1, 64])]; + tensor transpose_85 = transpose(perm = var_658_perm_0, x = var_657_cast_fp16)[name = tensor("transpose_85")]; + tensor key_states_27_cast_fp16 = reshape(shape = var_678, x = transpose_85)[name = tensor("key_states_27_cast_fp16")]; + tensor var_680 = const()[name = tensor("op_680"), val = tensor([16, -1, 64])]; + tensor transpose_84 = transpose(perm = var_665_perm_0, x = var_664_cast_fp16)[name = tensor("transpose_84")]; + tensor value_states_27_cast_fp16 = reshape(shape = var_680, x = transpose_84)[name = tensor("value_states_27_cast_fp16")]; + tensor var_683_perm_0 = const()[name = tensor("op_683_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_37_transpose_x_0 = const()[name = tensor("attn_weights_37_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_37_transpose_y_0 = const()[name = tensor("attn_weights_37_transpose_y_0"), val = tensor(false)]; + tensor transpose_82 = transpose(perm = var_683_perm_0, x = key_states_27_cast_fp16)[name = tensor("transpose_82")]; + tensor attn_weights_37_cast_fp16 = matmul(transpose_x = attn_weights_37_transpose_x_0, transpose_y = attn_weights_37_transpose_y_0, x = query_states_13_cast_fp16, y = transpose_82)[name = tensor("attn_weights_37_cast_fp16")]; + tensor var_685 = const()[name = tensor("op_685"), val = tensor([1, 16, 77, 77])]; + tensor var_686_cast_fp16 = reshape(shape = var_685, x = attn_weights_37_cast_fp16)[name = tensor("op_686_cast_fp16")]; + tensor attn_weights_39_cast_fp16 = add(x = var_686_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_39_cast_fp16")]; + tensor var_691 = const()[name = tensor("op_691"), val = tensor([16, 77, 77])]; + tensor input_101_cast_fp16 = reshape(shape = var_691, x = attn_weights_39_cast_fp16)[name = tensor("input_101_cast_fp16")]; + tensor input_103_cast_fp16 = softmax(axis = var_5, x = input_101_cast_fp16)[name = tensor("input_103_cast_fp16")]; + tensor attn_output_37_transpose_x_0 = const()[name = tensor("attn_output_37_transpose_x_0"), val = tensor(false)]; + tensor attn_output_37_transpose_y_0 = const()[name = tensor("attn_output_37_transpose_y_0"), val = tensor(false)]; + tensor attn_output_37_cast_fp16 = matmul(transpose_x = attn_output_37_transpose_x_0, transpose_y = attn_output_37_transpose_y_0, x = input_103_cast_fp16, y = value_states_27_cast_fp16)[name = tensor("attn_output_37_cast_fp16")]; + tensor var_696 = const()[name = tensor("op_696"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_39_cast_fp16 = reshape(shape = var_696, x = attn_output_37_cast_fp16)[name = tensor("attn_output_39_cast_fp16")]; + tensor attn_output_41_perm_0 = const()[name = tensor("attn_output_41_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_699 = const()[name = tensor("op_699"), val = tensor([1, 77, 1024])]; + tensor transpose_81 = transpose(perm = attn_output_41_perm_0, x = attn_output_39_cast_fp16)[name = tensor("transpose_81")]; + tensor input_105_cast_fp16 = reshape(shape = var_699, x = transpose_81)[name = tensor("input_105_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(258820480)))]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260917696)))]; + tensor linear_39_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16, x = input_105_cast_fp16)[name = tensor("linear_39_cast_fp16")]; + tensor input_107_cast_fp16 = add(x = input_99_cast_fp16, y = linear_39_cast_fp16)[name = tensor("input_107_cast_fp16")]; + tensor input_109_axes_0 = const()[name = tensor("input_109_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260919808)))]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260921920)))]; + tensor input_109_cast_fp16 = layer_norm(axes = input_109_axes_0, beta = text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16, x = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260924032)))]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269312704)))]; + tensor linear_40_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16, x = input_109_cast_fp16)[name = tensor("linear_40_cast_fp16")]; + tensor input_113_mode_0 = const()[name = tensor("input_113_mode_0"), val = tensor("EXACT")]; + tensor input_113_cast_fp16 = gelu(mode = input_113_mode_0, x = linear_40_cast_fp16)[name = tensor("input_113_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269320960)))]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(277709632)))]; + tensor linear_41_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16, x = input_113_cast_fp16)[name = tensor("linear_41_cast_fp16")]; + tensor input_115_cast_fp16 = add(x = input_107_cast_fp16, y = linear_41_cast_fp16)[name = tensor("input_115_cast_fp16")]; + tensor hidden_states_43_axes_0 = const()[name = tensor("hidden_states_43_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(277711744)))]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(277713856)))]; + tensor hidden_states_43_cast_fp16 = layer_norm(axes = hidden_states_43_axes_0, beta = text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("hidden_states_43_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(277715968)))]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279813184)))]; + tensor linear_42_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16, x = hidden_states_43_cast_fp16)[name = tensor("linear_42_cast_fp16")]; + tensor var_738_to_fp16 = const()[name = tensor("op_738_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_47_cast_fp16 = mul(x = linear_42_cast_fp16, y = var_738_to_fp16)[name = tensor("tensor_47_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279815296)))]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(281912512)))]; + tensor linear_43_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16, x = hidden_states_43_cast_fp16)[name = tensor("linear_43_cast_fp16")]; + tensor var_743 = const()[name = tensor("op_743"), val = tensor([1, -1, 16, 64])]; + tensor var_744_cast_fp16 = reshape(shape = var_743, x = linear_43_cast_fp16)[name = tensor("op_744_cast_fp16")]; + tensor var_745_perm_0 = const()[name = tensor("op_745_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(281914624)))]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284011840)))]; + tensor linear_44_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16, x = hidden_states_43_cast_fp16)[name = tensor("linear_44_cast_fp16")]; + tensor var_750 = const()[name = tensor("op_750"), val = tensor([1, -1, 16, 64])]; + tensor var_751_cast_fp16 = reshape(shape = var_750, x = linear_44_cast_fp16)[name = tensor("op_751_cast_fp16")]; + tensor var_752_perm_0 = const()[name = tensor("op_752_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_759 = const()[name = tensor("op_759"), val = tensor([1, 77, 16, 64])]; + tensor var_760_cast_fp16 = reshape(shape = var_759, x = tensor_47_cast_fp16)[name = tensor("op_760_cast_fp16")]; + tensor var_761_perm_0 = const()[name = tensor("op_761_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_763 = const()[name = tensor("op_763"), val = tensor([16, -1, 64])]; + tensor transpose_78 = transpose(perm = var_761_perm_0, x = var_760_cast_fp16)[name = tensor("transpose_78")]; + tensor query_states_15_cast_fp16 = reshape(shape = var_763, x = transpose_78)[name = tensor("query_states_15_cast_fp16")]; + tensor var_765 = const()[name = tensor("op_765"), val = tensor([16, -1, 64])]; + tensor transpose_80 = transpose(perm = var_745_perm_0, x = var_744_cast_fp16)[name = tensor("transpose_80")]; + tensor key_states_31_cast_fp16 = reshape(shape = var_765, x = transpose_80)[name = tensor("key_states_31_cast_fp16")]; + tensor var_767 = const()[name = tensor("op_767"), val = tensor([16, -1, 64])]; + tensor transpose_79 = transpose(perm = var_752_perm_0, x = var_751_cast_fp16)[name = tensor("transpose_79")]; + tensor value_states_31_cast_fp16 = reshape(shape = var_767, x = transpose_79)[name = tensor("value_states_31_cast_fp16")]; + tensor var_770_perm_0 = const()[name = tensor("op_770_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_43_transpose_x_0 = const()[name = tensor("attn_weights_43_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_43_transpose_y_0 = const()[name = tensor("attn_weights_43_transpose_y_0"), val = tensor(false)]; + tensor transpose_77 = transpose(perm = var_770_perm_0, x = key_states_31_cast_fp16)[name = tensor("transpose_77")]; + tensor attn_weights_43_cast_fp16 = matmul(transpose_x = attn_weights_43_transpose_x_0, transpose_y = attn_weights_43_transpose_y_0, x = query_states_15_cast_fp16, y = transpose_77)[name = tensor("attn_weights_43_cast_fp16")]; + tensor var_772 = const()[name = tensor("op_772"), val = tensor([1, 16, 77, 77])]; + tensor var_773_cast_fp16 = reshape(shape = var_772, x = attn_weights_43_cast_fp16)[name = tensor("op_773_cast_fp16")]; + tensor attn_weights_45_cast_fp16 = add(x = var_773_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_45_cast_fp16")]; + tensor var_778 = const()[name = tensor("op_778"), val = tensor([16, 77, 77])]; + tensor input_117_cast_fp16 = reshape(shape = var_778, x = attn_weights_45_cast_fp16)[name = tensor("input_117_cast_fp16")]; + tensor input_119_cast_fp16 = softmax(axis = var_5, x = input_117_cast_fp16)[name = tensor("input_119_cast_fp16")]; + tensor attn_output_43_transpose_x_0 = const()[name = tensor("attn_output_43_transpose_x_0"), val = tensor(false)]; + tensor attn_output_43_transpose_y_0 = const()[name = tensor("attn_output_43_transpose_y_0"), val = tensor(false)]; + tensor attn_output_43_cast_fp16 = matmul(transpose_x = attn_output_43_transpose_x_0, transpose_y = attn_output_43_transpose_y_0, x = input_119_cast_fp16, y = value_states_31_cast_fp16)[name = tensor("attn_output_43_cast_fp16")]; + tensor var_783 = const()[name = tensor("op_783"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_45_cast_fp16 = reshape(shape = var_783, x = attn_output_43_cast_fp16)[name = tensor("attn_output_45_cast_fp16")]; + tensor attn_output_47_perm_0 = const()[name = tensor("attn_output_47_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_786 = const()[name = tensor("op_786"), val = tensor([1, 77, 1024])]; + tensor transpose_76 = transpose(perm = attn_output_47_perm_0, x = attn_output_45_cast_fp16)[name = tensor("transpose_76")]; + tensor input_121_cast_fp16 = reshape(shape = var_786, x = transpose_76)[name = tensor("input_121_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284013952)))]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286111168)))]; + tensor linear_45_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("linear_45_cast_fp16")]; + tensor input_123_cast_fp16 = add(x = input_115_cast_fp16, y = linear_45_cast_fp16)[name = tensor("input_123_cast_fp16")]; + tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286113280)))]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286115392)))]; + tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286117504)))]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294506176)))]; + tensor linear_46_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16, x = input_125_cast_fp16)[name = tensor("linear_46_cast_fp16")]; + tensor input_129_mode_0 = const()[name = tensor("input_129_mode_0"), val = tensor("EXACT")]; + tensor input_129_cast_fp16 = gelu(mode = input_129_mode_0, x = linear_46_cast_fp16)[name = tensor("input_129_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294514432)))]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302903104)))]; + tensor linear_47_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("linear_47_cast_fp16")]; + tensor input_131_cast_fp16 = add(x = input_123_cast_fp16, y = linear_47_cast_fp16)[name = tensor("input_131_cast_fp16")]; + tensor hidden_states_49_axes_0 = const()[name = tensor("hidden_states_49_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302905216)))]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302907328)))]; + tensor hidden_states_49_cast_fp16 = layer_norm(axes = hidden_states_49_axes_0, beta = text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16, x = input_131_cast_fp16)[name = tensor("hidden_states_49_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(302909440)))]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305006656)))]; + tensor linear_48_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16, x = hidden_states_49_cast_fp16)[name = tensor("linear_48_cast_fp16")]; + tensor var_825_to_fp16 = const()[name = tensor("op_825_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_53_cast_fp16 = mul(x = linear_48_cast_fp16, y = var_825_to_fp16)[name = tensor("tensor_53_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305008768)))]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307105984)))]; + tensor linear_49_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16, x = hidden_states_49_cast_fp16)[name = tensor("linear_49_cast_fp16")]; + tensor var_830 = const()[name = tensor("op_830"), val = tensor([1, -1, 16, 64])]; + tensor var_831_cast_fp16 = reshape(shape = var_830, x = linear_49_cast_fp16)[name = tensor("op_831_cast_fp16")]; + tensor var_832_perm_0 = const()[name = tensor("op_832_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307108096)))]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309205312)))]; + tensor linear_50_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16, x = hidden_states_49_cast_fp16)[name = tensor("linear_50_cast_fp16")]; + tensor var_837 = const()[name = tensor("op_837"), val = tensor([1, -1, 16, 64])]; + tensor var_838_cast_fp16 = reshape(shape = var_837, x = linear_50_cast_fp16)[name = tensor("op_838_cast_fp16")]; + tensor var_839_perm_0 = const()[name = tensor("op_839_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_846 = const()[name = tensor("op_846"), val = tensor([1, 77, 16, 64])]; + tensor var_847_cast_fp16 = reshape(shape = var_846, x = tensor_53_cast_fp16)[name = tensor("op_847_cast_fp16")]; + tensor var_848_perm_0 = const()[name = tensor("op_848_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_850 = const()[name = tensor("op_850"), val = tensor([16, -1, 64])]; + tensor transpose_73 = transpose(perm = var_848_perm_0, x = var_847_cast_fp16)[name = tensor("transpose_73")]; + tensor query_states_17_cast_fp16 = reshape(shape = var_850, x = transpose_73)[name = tensor("query_states_17_cast_fp16")]; + tensor var_852 = const()[name = tensor("op_852"), val = tensor([16, -1, 64])]; + tensor transpose_75 = transpose(perm = var_832_perm_0, x = var_831_cast_fp16)[name = tensor("transpose_75")]; + tensor key_states_35_cast_fp16 = reshape(shape = var_852, x = transpose_75)[name = tensor("key_states_35_cast_fp16")]; + tensor var_854 = const()[name = tensor("op_854"), val = tensor([16, -1, 64])]; + tensor transpose_74 = transpose(perm = var_839_perm_0, x = var_838_cast_fp16)[name = tensor("transpose_74")]; + tensor value_states_35_cast_fp16 = reshape(shape = var_854, x = transpose_74)[name = tensor("value_states_35_cast_fp16")]; + tensor var_857_perm_0 = const()[name = tensor("op_857_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_49_transpose_x_0 = const()[name = tensor("attn_weights_49_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_49_transpose_y_0 = const()[name = tensor("attn_weights_49_transpose_y_0"), val = tensor(false)]; + tensor transpose_72 = transpose(perm = var_857_perm_0, x = key_states_35_cast_fp16)[name = tensor("transpose_72")]; + tensor attn_weights_49_cast_fp16 = matmul(transpose_x = attn_weights_49_transpose_x_0, transpose_y = attn_weights_49_transpose_y_0, x = query_states_17_cast_fp16, y = transpose_72)[name = tensor("attn_weights_49_cast_fp16")]; + tensor var_859 = const()[name = tensor("op_859"), val = tensor([1, 16, 77, 77])]; + tensor var_860_cast_fp16 = reshape(shape = var_859, x = attn_weights_49_cast_fp16)[name = tensor("op_860_cast_fp16")]; + tensor attn_weights_51_cast_fp16 = add(x = var_860_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_51_cast_fp16")]; + tensor var_865 = const()[name = tensor("op_865"), val = tensor([16, 77, 77])]; + tensor input_133_cast_fp16 = reshape(shape = var_865, x = attn_weights_51_cast_fp16)[name = tensor("input_133_cast_fp16")]; + tensor input_135_cast_fp16 = softmax(axis = var_5, x = input_133_cast_fp16)[name = tensor("input_135_cast_fp16")]; + tensor attn_output_49_transpose_x_0 = const()[name = tensor("attn_output_49_transpose_x_0"), val = tensor(false)]; + tensor attn_output_49_transpose_y_0 = const()[name = tensor("attn_output_49_transpose_y_0"), val = tensor(false)]; + tensor attn_output_49_cast_fp16 = matmul(transpose_x = attn_output_49_transpose_x_0, transpose_y = attn_output_49_transpose_y_0, x = input_135_cast_fp16, y = value_states_35_cast_fp16)[name = tensor("attn_output_49_cast_fp16")]; + tensor var_870 = const()[name = tensor("op_870"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_51_cast_fp16 = reshape(shape = var_870, x = attn_output_49_cast_fp16)[name = tensor("attn_output_51_cast_fp16")]; + tensor attn_output_53_perm_0 = const()[name = tensor("attn_output_53_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_873 = const()[name = tensor("op_873"), val = tensor([1, 77, 1024])]; + tensor transpose_71 = transpose(perm = attn_output_53_perm_0, x = attn_output_51_cast_fp16)[name = tensor("transpose_71")]; + tensor input_137_cast_fp16 = reshape(shape = var_873, x = transpose_71)[name = tensor("input_137_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309207424)))]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311304640)))]; + tensor linear_51_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16, x = input_137_cast_fp16)[name = tensor("linear_51_cast_fp16")]; + tensor input_139_cast_fp16 = add(x = input_131_cast_fp16, y = linear_51_cast_fp16)[name = tensor("input_139_cast_fp16")]; + tensor input_141_axes_0 = const()[name = tensor("input_141_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311306752)))]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311308864)))]; + tensor input_141_cast_fp16 = layer_norm(axes = input_141_axes_0, beta = text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16, x = input_139_cast_fp16)[name = tensor("input_141_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311310976)))]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319699648)))]; + tensor linear_52_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16, x = input_141_cast_fp16)[name = tensor("linear_52_cast_fp16")]; + tensor input_145_mode_0 = const()[name = tensor("input_145_mode_0"), val = tensor("EXACT")]; + tensor input_145_cast_fp16 = gelu(mode = input_145_mode_0, x = linear_52_cast_fp16)[name = tensor("input_145_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319707904)))]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328096576)))]; + tensor linear_53_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16, x = input_145_cast_fp16)[name = tensor("linear_53_cast_fp16")]; + tensor input_147_cast_fp16 = add(x = input_139_cast_fp16, y = linear_53_cast_fp16)[name = tensor("input_147_cast_fp16")]; + tensor hidden_states_55_axes_0 = const()[name = tensor("hidden_states_55_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328098688)))]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328100800)))]; + tensor hidden_states_55_cast_fp16 = layer_norm(axes = hidden_states_55_axes_0, beta = text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16, x = input_147_cast_fp16)[name = tensor("hidden_states_55_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328102912)))]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(330200128)))]; + tensor linear_54_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16, x = hidden_states_55_cast_fp16)[name = tensor("linear_54_cast_fp16")]; + tensor var_912_to_fp16 = const()[name = tensor("op_912_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_59_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_912_to_fp16)[name = tensor("tensor_59_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(330202240)))]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332299456)))]; + tensor linear_55_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16, x = hidden_states_55_cast_fp16)[name = tensor("linear_55_cast_fp16")]; + tensor var_917 = const()[name = tensor("op_917"), val = tensor([1, -1, 16, 64])]; + tensor var_918_cast_fp16 = reshape(shape = var_917, x = linear_55_cast_fp16)[name = tensor("op_918_cast_fp16")]; + tensor var_919_perm_0 = const()[name = tensor("op_919_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332301568)))]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334398784)))]; + tensor linear_56_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16, x = hidden_states_55_cast_fp16)[name = tensor("linear_56_cast_fp16")]; + tensor var_924 = const()[name = tensor("op_924"), val = tensor([1, -1, 16, 64])]; + tensor var_925_cast_fp16 = reshape(shape = var_924, x = linear_56_cast_fp16)[name = tensor("op_925_cast_fp16")]; + tensor var_926_perm_0 = const()[name = tensor("op_926_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_933 = const()[name = tensor("op_933"), val = tensor([1, 77, 16, 64])]; + tensor var_934_cast_fp16 = reshape(shape = var_933, x = tensor_59_cast_fp16)[name = tensor("op_934_cast_fp16")]; + tensor var_935_perm_0 = const()[name = tensor("op_935_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_937 = const()[name = tensor("op_937"), val = tensor([16, -1, 64])]; + tensor transpose_68 = transpose(perm = var_935_perm_0, x = var_934_cast_fp16)[name = tensor("transpose_68")]; + tensor query_states_19_cast_fp16 = reshape(shape = var_937, x = transpose_68)[name = tensor("query_states_19_cast_fp16")]; + tensor var_939 = const()[name = tensor("op_939"), val = tensor([16, -1, 64])]; + tensor transpose_70 = transpose(perm = var_919_perm_0, x = var_918_cast_fp16)[name = tensor("transpose_70")]; + tensor key_states_39_cast_fp16 = reshape(shape = var_939, x = transpose_70)[name = tensor("key_states_39_cast_fp16")]; + tensor var_941 = const()[name = tensor("op_941"), val = tensor([16, -1, 64])]; + tensor transpose_69 = transpose(perm = var_926_perm_0, x = var_925_cast_fp16)[name = tensor("transpose_69")]; + tensor value_states_39_cast_fp16 = reshape(shape = var_941, x = transpose_69)[name = tensor("value_states_39_cast_fp16")]; + tensor var_944_perm_0 = const()[name = tensor("op_944_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_55_transpose_x_0 = const()[name = tensor("attn_weights_55_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_55_transpose_y_0 = const()[name = tensor("attn_weights_55_transpose_y_0"), val = tensor(false)]; + tensor transpose_67 = transpose(perm = var_944_perm_0, x = key_states_39_cast_fp16)[name = tensor("transpose_67")]; + tensor attn_weights_55_cast_fp16 = matmul(transpose_x = attn_weights_55_transpose_x_0, transpose_y = attn_weights_55_transpose_y_0, x = query_states_19_cast_fp16, y = transpose_67)[name = tensor("attn_weights_55_cast_fp16")]; + tensor var_946 = const()[name = tensor("op_946"), val = tensor([1, 16, 77, 77])]; + tensor var_947_cast_fp16 = reshape(shape = var_946, x = attn_weights_55_cast_fp16)[name = tensor("op_947_cast_fp16")]; + tensor attn_weights_57_cast_fp16 = add(x = var_947_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_57_cast_fp16")]; + tensor var_952 = const()[name = tensor("op_952"), val = tensor([16, 77, 77])]; + tensor input_149_cast_fp16 = reshape(shape = var_952, x = attn_weights_57_cast_fp16)[name = tensor("input_149_cast_fp16")]; + tensor input_151_cast_fp16 = softmax(axis = var_5, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; + tensor attn_output_55_transpose_x_0 = const()[name = tensor("attn_output_55_transpose_x_0"), val = tensor(false)]; + tensor attn_output_55_transpose_y_0 = const()[name = tensor("attn_output_55_transpose_y_0"), val = tensor(false)]; + tensor attn_output_55_cast_fp16 = matmul(transpose_x = attn_output_55_transpose_x_0, transpose_y = attn_output_55_transpose_y_0, x = input_151_cast_fp16, y = value_states_39_cast_fp16)[name = tensor("attn_output_55_cast_fp16")]; + tensor var_957 = const()[name = tensor("op_957"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_57_cast_fp16 = reshape(shape = var_957, x = attn_output_55_cast_fp16)[name = tensor("attn_output_57_cast_fp16")]; + tensor attn_output_59_perm_0 = const()[name = tensor("attn_output_59_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_960 = const()[name = tensor("op_960"), val = tensor([1, 77, 1024])]; + tensor transpose_66 = transpose(perm = attn_output_59_perm_0, x = attn_output_57_cast_fp16)[name = tensor("transpose_66")]; + tensor input_153_cast_fp16 = reshape(shape = var_960, x = transpose_66)[name = tensor("input_153_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334400896)))]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336498112)))]; + tensor linear_57_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16, x = input_153_cast_fp16)[name = tensor("linear_57_cast_fp16")]; + tensor input_155_cast_fp16 = add(x = input_147_cast_fp16, y = linear_57_cast_fp16)[name = tensor("input_155_cast_fp16")]; + tensor input_157_axes_0 = const()[name = tensor("input_157_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336500224)))]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336502336)))]; + tensor input_157_cast_fp16 = layer_norm(axes = input_157_axes_0, beta = text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16, x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336504448)))]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344893120)))]; + tensor linear_58_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16, x = input_157_cast_fp16)[name = tensor("linear_58_cast_fp16")]; + tensor input_161_mode_0 = const()[name = tensor("input_161_mode_0"), val = tensor("EXACT")]; + tensor input_161_cast_fp16 = gelu(mode = input_161_mode_0, x = linear_58_cast_fp16)[name = tensor("input_161_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344901376)))]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353290048)))]; + tensor linear_59_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16, x = input_161_cast_fp16)[name = tensor("linear_59_cast_fp16")]; + tensor input_163_cast_fp16 = add(x = input_155_cast_fp16, y = linear_59_cast_fp16)[name = tensor("input_163_cast_fp16")]; + tensor hidden_states_61_axes_0 = const()[name = tensor("hidden_states_61_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353292160)))]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353294272)))]; + tensor hidden_states_61_cast_fp16 = layer_norm(axes = hidden_states_61_axes_0, beta = text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("hidden_states_61_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353296384)))]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(355393600)))]; + tensor linear_60_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16, x = hidden_states_61_cast_fp16)[name = tensor("linear_60_cast_fp16")]; + tensor var_999_to_fp16 = const()[name = tensor("op_999_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_65_cast_fp16 = mul(x = linear_60_cast_fp16, y = var_999_to_fp16)[name = tensor("tensor_65_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(355395712)))]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357492928)))]; + tensor linear_61_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16, x = hidden_states_61_cast_fp16)[name = tensor("linear_61_cast_fp16")]; + tensor var_1004 = const()[name = tensor("op_1004"), val = tensor([1, -1, 16, 64])]; + tensor var_1005_cast_fp16 = reshape(shape = var_1004, x = linear_61_cast_fp16)[name = tensor("op_1005_cast_fp16")]; + tensor var_1006_perm_0 = const()[name = tensor("op_1006_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357495040)))]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359592256)))]; + tensor linear_62_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16, x = hidden_states_61_cast_fp16)[name = tensor("linear_62_cast_fp16")]; + tensor var_1011 = const()[name = tensor("op_1011"), val = tensor([1, -1, 16, 64])]; + tensor var_1012_cast_fp16 = reshape(shape = var_1011, x = linear_62_cast_fp16)[name = tensor("op_1012_cast_fp16")]; + tensor var_1013_perm_0 = const()[name = tensor("op_1013_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1020 = const()[name = tensor("op_1020"), val = tensor([1, 77, 16, 64])]; + tensor var_1021_cast_fp16 = reshape(shape = var_1020, x = tensor_65_cast_fp16)[name = tensor("op_1021_cast_fp16")]; + tensor var_1022_perm_0 = const()[name = tensor("op_1022_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1024 = const()[name = tensor("op_1024"), val = tensor([16, -1, 64])]; + tensor transpose_63 = transpose(perm = var_1022_perm_0, x = var_1021_cast_fp16)[name = tensor("transpose_63")]; + tensor query_states_21_cast_fp16 = reshape(shape = var_1024, x = transpose_63)[name = tensor("query_states_21_cast_fp16")]; + tensor var_1026 = const()[name = tensor("op_1026"), val = tensor([16, -1, 64])]; + tensor transpose_65 = transpose(perm = var_1006_perm_0, x = var_1005_cast_fp16)[name = tensor("transpose_65")]; + tensor key_states_43_cast_fp16 = reshape(shape = var_1026, x = transpose_65)[name = tensor("key_states_43_cast_fp16")]; + tensor var_1028 = const()[name = tensor("op_1028"), val = tensor([16, -1, 64])]; + tensor transpose_64 = transpose(perm = var_1013_perm_0, x = var_1012_cast_fp16)[name = tensor("transpose_64")]; + tensor value_states_43_cast_fp16 = reshape(shape = var_1028, x = transpose_64)[name = tensor("value_states_43_cast_fp16")]; + tensor var_1031_perm_0 = const()[name = tensor("op_1031_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_61_transpose_x_0 = const()[name = tensor("attn_weights_61_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_61_transpose_y_0 = const()[name = tensor("attn_weights_61_transpose_y_0"), val = tensor(false)]; + tensor transpose_62 = transpose(perm = var_1031_perm_0, x = key_states_43_cast_fp16)[name = tensor("transpose_62")]; + tensor attn_weights_61_cast_fp16 = matmul(transpose_x = attn_weights_61_transpose_x_0, transpose_y = attn_weights_61_transpose_y_0, x = query_states_21_cast_fp16, y = transpose_62)[name = tensor("attn_weights_61_cast_fp16")]; + tensor var_1033 = const()[name = tensor("op_1033"), val = tensor([1, 16, 77, 77])]; + tensor var_1034_cast_fp16 = reshape(shape = var_1033, x = attn_weights_61_cast_fp16)[name = tensor("op_1034_cast_fp16")]; + tensor attn_weights_63_cast_fp16 = add(x = var_1034_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_63_cast_fp16")]; + tensor var_1039 = const()[name = tensor("op_1039"), val = tensor([16, 77, 77])]; + tensor input_165_cast_fp16 = reshape(shape = var_1039, x = attn_weights_63_cast_fp16)[name = tensor("input_165_cast_fp16")]; + tensor input_167_cast_fp16 = softmax(axis = var_5, x = input_165_cast_fp16)[name = tensor("input_167_cast_fp16")]; + tensor attn_output_61_transpose_x_0 = const()[name = tensor("attn_output_61_transpose_x_0"), val = tensor(false)]; + tensor attn_output_61_transpose_y_0 = const()[name = tensor("attn_output_61_transpose_y_0"), val = tensor(false)]; + tensor attn_output_61_cast_fp16 = matmul(transpose_x = attn_output_61_transpose_x_0, transpose_y = attn_output_61_transpose_y_0, x = input_167_cast_fp16, y = value_states_43_cast_fp16)[name = tensor("attn_output_61_cast_fp16")]; + tensor var_1044 = const()[name = tensor("op_1044"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_63_cast_fp16 = reshape(shape = var_1044, x = attn_output_61_cast_fp16)[name = tensor("attn_output_63_cast_fp16")]; + tensor attn_output_65_perm_0 = const()[name = tensor("attn_output_65_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1047 = const()[name = tensor("op_1047"), val = tensor([1, 77, 1024])]; + tensor transpose_61 = transpose(perm = attn_output_65_perm_0, x = attn_output_63_cast_fp16)[name = tensor("transpose_61")]; + tensor input_169_cast_fp16 = reshape(shape = var_1047, x = transpose_61)[name = tensor("input_169_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359594368)))]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(361691584)))]; + tensor linear_63_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16, x = input_169_cast_fp16)[name = tensor("linear_63_cast_fp16")]; + tensor input_171_cast_fp16 = add(x = input_163_cast_fp16, y = linear_63_cast_fp16)[name = tensor("input_171_cast_fp16")]; + tensor input_173_axes_0 = const()[name = tensor("input_173_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(361693696)))]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(361695808)))]; + tensor input_173_cast_fp16 = layer_norm(axes = input_173_axes_0, beta = text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16, x = input_171_cast_fp16)[name = tensor("input_173_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(361697920)))]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(370086592)))]; + tensor linear_64_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("linear_64_cast_fp16")]; + tensor input_177_mode_0 = const()[name = tensor("input_177_mode_0"), val = tensor("EXACT")]; + tensor input_177_cast_fp16 = gelu(mode = input_177_mode_0, x = linear_64_cast_fp16)[name = tensor("input_177_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(370094848)))]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378483520)))]; + tensor linear_65_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16, x = input_177_cast_fp16)[name = tensor("linear_65_cast_fp16")]; + tensor input_179_cast_fp16 = add(x = input_171_cast_fp16, y = linear_65_cast_fp16)[name = tensor("input_179_cast_fp16")]; + tensor hidden_states_67_axes_0 = const()[name = tensor("hidden_states_67_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378485632)))]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378487744)))]; + tensor hidden_states_67_cast_fp16 = layer_norm(axes = hidden_states_67_axes_0, beta = text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16, x = input_179_cast_fp16)[name = tensor("hidden_states_67_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378489856)))]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380587072)))]; + tensor linear_66_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16, x = hidden_states_67_cast_fp16)[name = tensor("linear_66_cast_fp16")]; + tensor var_1086_to_fp16 = const()[name = tensor("op_1086_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_71_cast_fp16 = mul(x = linear_66_cast_fp16, y = var_1086_to_fp16)[name = tensor("tensor_71_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380589184)))]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382686400)))]; + tensor linear_67_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16, x = hidden_states_67_cast_fp16)[name = tensor("linear_67_cast_fp16")]; + tensor var_1091 = const()[name = tensor("op_1091"), val = tensor([1, -1, 16, 64])]; + tensor var_1092_cast_fp16 = reshape(shape = var_1091, x = linear_67_cast_fp16)[name = tensor("op_1092_cast_fp16")]; + tensor var_1093_perm_0 = const()[name = tensor("op_1093_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382688512)))]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384785728)))]; + tensor linear_68_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16, x = hidden_states_67_cast_fp16)[name = tensor("linear_68_cast_fp16")]; + tensor var_1098 = const()[name = tensor("op_1098"), val = tensor([1, -1, 16, 64])]; + tensor var_1099_cast_fp16 = reshape(shape = var_1098, x = linear_68_cast_fp16)[name = tensor("op_1099_cast_fp16")]; + tensor var_1100_perm_0 = const()[name = tensor("op_1100_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1107 = const()[name = tensor("op_1107"), val = tensor([1, 77, 16, 64])]; + tensor var_1108_cast_fp16 = reshape(shape = var_1107, x = tensor_71_cast_fp16)[name = tensor("op_1108_cast_fp16")]; + tensor var_1109_perm_0 = const()[name = tensor("op_1109_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1111 = const()[name = tensor("op_1111"), val = tensor([16, -1, 64])]; + tensor transpose_58 = transpose(perm = var_1109_perm_0, x = var_1108_cast_fp16)[name = tensor("transpose_58")]; + tensor query_states_23_cast_fp16 = reshape(shape = var_1111, x = transpose_58)[name = tensor("query_states_23_cast_fp16")]; + tensor var_1113 = const()[name = tensor("op_1113"), val = tensor([16, -1, 64])]; + tensor transpose_60 = transpose(perm = var_1093_perm_0, x = var_1092_cast_fp16)[name = tensor("transpose_60")]; + tensor key_states_47_cast_fp16 = reshape(shape = var_1113, x = transpose_60)[name = tensor("key_states_47_cast_fp16")]; + tensor var_1115 = const()[name = tensor("op_1115"), val = tensor([16, -1, 64])]; + tensor transpose_59 = transpose(perm = var_1100_perm_0, x = var_1099_cast_fp16)[name = tensor("transpose_59")]; + tensor value_states_47_cast_fp16 = reshape(shape = var_1115, x = transpose_59)[name = tensor("value_states_47_cast_fp16")]; + tensor var_1118_perm_0 = const()[name = tensor("op_1118_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_67_transpose_x_0 = const()[name = tensor("attn_weights_67_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_67_transpose_y_0 = const()[name = tensor("attn_weights_67_transpose_y_0"), val = tensor(false)]; + tensor transpose_57 = transpose(perm = var_1118_perm_0, x = key_states_47_cast_fp16)[name = tensor("transpose_57")]; + tensor attn_weights_67_cast_fp16 = matmul(transpose_x = attn_weights_67_transpose_x_0, transpose_y = attn_weights_67_transpose_y_0, x = query_states_23_cast_fp16, y = transpose_57)[name = tensor("attn_weights_67_cast_fp16")]; + tensor var_1120 = const()[name = tensor("op_1120"), val = tensor([1, 16, 77, 77])]; + tensor var_1121_cast_fp16 = reshape(shape = var_1120, x = attn_weights_67_cast_fp16)[name = tensor("op_1121_cast_fp16")]; + tensor attn_weights_69_cast_fp16 = add(x = var_1121_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_69_cast_fp16")]; + tensor var_1126 = const()[name = tensor("op_1126"), val = tensor([16, 77, 77])]; + tensor input_181_cast_fp16 = reshape(shape = var_1126, x = attn_weights_69_cast_fp16)[name = tensor("input_181_cast_fp16")]; + tensor input_183_cast_fp16 = softmax(axis = var_5, x = input_181_cast_fp16)[name = tensor("input_183_cast_fp16")]; + tensor attn_output_67_transpose_x_0 = const()[name = tensor("attn_output_67_transpose_x_0"), val = tensor(false)]; + tensor attn_output_67_transpose_y_0 = const()[name = tensor("attn_output_67_transpose_y_0"), val = tensor(false)]; + tensor attn_output_67_cast_fp16 = matmul(transpose_x = attn_output_67_transpose_x_0, transpose_y = attn_output_67_transpose_y_0, x = input_183_cast_fp16, y = value_states_47_cast_fp16)[name = tensor("attn_output_67_cast_fp16")]; + tensor var_1131 = const()[name = tensor("op_1131"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_69_cast_fp16 = reshape(shape = var_1131, x = attn_output_67_cast_fp16)[name = tensor("attn_output_69_cast_fp16")]; + tensor attn_output_71_perm_0 = const()[name = tensor("attn_output_71_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1134 = const()[name = tensor("op_1134"), val = tensor([1, 77, 1024])]; + tensor transpose_56 = transpose(perm = attn_output_71_perm_0, x = attn_output_69_cast_fp16)[name = tensor("transpose_56")]; + tensor input_185_cast_fp16 = reshape(shape = var_1134, x = transpose_56)[name = tensor("input_185_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384787840)))]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(386885056)))]; + tensor linear_69_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16, x = input_185_cast_fp16)[name = tensor("linear_69_cast_fp16")]; + tensor input_187_cast_fp16 = add(x = input_179_cast_fp16, y = linear_69_cast_fp16)[name = tensor("input_187_cast_fp16")]; + tensor input_189_axes_0 = const()[name = tensor("input_189_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(386887168)))]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(386889280)))]; + tensor input_189_cast_fp16 = layer_norm(axes = input_189_axes_0, beta = text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("input_189_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(386891392)))]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(395280064)))]; + tensor linear_70_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16, x = input_189_cast_fp16)[name = tensor("linear_70_cast_fp16")]; + tensor input_193_mode_0 = const()[name = tensor("input_193_mode_0"), val = tensor("EXACT")]; + tensor input_193_cast_fp16 = gelu(mode = input_193_mode_0, x = linear_70_cast_fp16)[name = tensor("input_193_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(395288320)))]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(403676992)))]; + tensor linear_71_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16, x = input_193_cast_fp16)[name = tensor("linear_71_cast_fp16")]; + tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_71_cast_fp16)[name = tensor("input_195_cast_fp16")]; + tensor hidden_states_73_axes_0 = const()[name = tensor("hidden_states_73_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(403679104)))]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(403681216)))]; + tensor hidden_states_73_cast_fp16 = layer_norm(axes = hidden_states_73_axes_0, beta = text_encoder_text_model_encoder_layers_12_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_12_layer_norm1_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("hidden_states_73_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(403683328)))]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405780544)))]; + tensor linear_72_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_q_proj_weight_to_fp16, x = hidden_states_73_cast_fp16)[name = tensor("linear_72_cast_fp16")]; + tensor var_1173_to_fp16 = const()[name = tensor("op_1173_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_77_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1173_to_fp16)[name = tensor("tensor_77_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405782656)))]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407879872)))]; + tensor linear_73_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_k_proj_weight_to_fp16, x = hidden_states_73_cast_fp16)[name = tensor("linear_73_cast_fp16")]; + tensor var_1178 = const()[name = tensor("op_1178"), val = tensor([1, -1, 16, 64])]; + tensor var_1179_cast_fp16 = reshape(shape = var_1178, x = linear_73_cast_fp16)[name = tensor("op_1179_cast_fp16")]; + tensor var_1180_perm_0 = const()[name = tensor("op_1180_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407881984)))]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409979200)))]; + tensor linear_74_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_v_proj_weight_to_fp16, x = hidden_states_73_cast_fp16)[name = tensor("linear_74_cast_fp16")]; + tensor var_1185 = const()[name = tensor("op_1185"), val = tensor([1, -1, 16, 64])]; + tensor var_1186_cast_fp16 = reshape(shape = var_1185, x = linear_74_cast_fp16)[name = tensor("op_1186_cast_fp16")]; + tensor var_1187_perm_0 = const()[name = tensor("op_1187_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1194 = const()[name = tensor("op_1194"), val = tensor([1, 77, 16, 64])]; + tensor var_1195_cast_fp16 = reshape(shape = var_1194, x = tensor_77_cast_fp16)[name = tensor("op_1195_cast_fp16")]; + tensor var_1196_perm_0 = const()[name = tensor("op_1196_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1198 = const()[name = tensor("op_1198"), val = tensor([16, -1, 64])]; + tensor transpose_53 = transpose(perm = var_1196_perm_0, x = var_1195_cast_fp16)[name = tensor("transpose_53")]; + tensor query_states_25_cast_fp16 = reshape(shape = var_1198, x = transpose_53)[name = tensor("query_states_25_cast_fp16")]; + tensor var_1200 = const()[name = tensor("op_1200"), val = tensor([16, -1, 64])]; + tensor transpose_55 = transpose(perm = var_1180_perm_0, x = var_1179_cast_fp16)[name = tensor("transpose_55")]; + tensor key_states_51_cast_fp16 = reshape(shape = var_1200, x = transpose_55)[name = tensor("key_states_51_cast_fp16")]; + tensor var_1202 = const()[name = tensor("op_1202"), val = tensor([16, -1, 64])]; + tensor transpose_54 = transpose(perm = var_1187_perm_0, x = var_1186_cast_fp16)[name = tensor("transpose_54")]; + tensor value_states_51_cast_fp16 = reshape(shape = var_1202, x = transpose_54)[name = tensor("value_states_51_cast_fp16")]; + tensor var_1205_perm_0 = const()[name = tensor("op_1205_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_73_transpose_x_0 = const()[name = tensor("attn_weights_73_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_73_transpose_y_0 = const()[name = tensor("attn_weights_73_transpose_y_0"), val = tensor(false)]; + tensor transpose_52 = transpose(perm = var_1205_perm_0, x = key_states_51_cast_fp16)[name = tensor("transpose_52")]; + tensor attn_weights_73_cast_fp16 = matmul(transpose_x = attn_weights_73_transpose_x_0, transpose_y = attn_weights_73_transpose_y_0, x = query_states_25_cast_fp16, y = transpose_52)[name = tensor("attn_weights_73_cast_fp16")]; + tensor var_1207 = const()[name = tensor("op_1207"), val = tensor([1, 16, 77, 77])]; + tensor var_1208_cast_fp16 = reshape(shape = var_1207, x = attn_weights_73_cast_fp16)[name = tensor("op_1208_cast_fp16")]; + tensor attn_weights_75_cast_fp16 = add(x = var_1208_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_75_cast_fp16")]; + tensor var_1213 = const()[name = tensor("op_1213"), val = tensor([16, 77, 77])]; + tensor input_197_cast_fp16 = reshape(shape = var_1213, x = attn_weights_75_cast_fp16)[name = tensor("input_197_cast_fp16")]; + tensor input_199_cast_fp16 = softmax(axis = var_5, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; + tensor attn_output_73_transpose_x_0 = const()[name = tensor("attn_output_73_transpose_x_0"), val = tensor(false)]; + tensor attn_output_73_transpose_y_0 = const()[name = tensor("attn_output_73_transpose_y_0"), val = tensor(false)]; + tensor attn_output_73_cast_fp16 = matmul(transpose_x = attn_output_73_transpose_x_0, transpose_y = attn_output_73_transpose_y_0, x = input_199_cast_fp16, y = value_states_51_cast_fp16)[name = tensor("attn_output_73_cast_fp16")]; + tensor var_1218 = const()[name = tensor("op_1218"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_75_cast_fp16 = reshape(shape = var_1218, x = attn_output_73_cast_fp16)[name = tensor("attn_output_75_cast_fp16")]; + tensor attn_output_77_perm_0 = const()[name = tensor("attn_output_77_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1221 = const()[name = tensor("op_1221"), val = tensor([1, 77, 1024])]; + tensor transpose_51 = transpose(perm = attn_output_77_perm_0, x = attn_output_75_cast_fp16)[name = tensor("transpose_51")]; + tensor input_201_cast_fp16 = reshape(shape = var_1221, x = transpose_51)[name = tensor("input_201_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409981312)))]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412078528)))]; + tensor linear_75_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_out_proj_weight_to_fp16, x = input_201_cast_fp16)[name = tensor("linear_75_cast_fp16")]; + tensor input_203_cast_fp16 = add(x = input_195_cast_fp16, y = linear_75_cast_fp16)[name = tensor("input_203_cast_fp16")]; + tensor input_205_axes_0 = const()[name = tensor("input_205_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412080640)))]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412082752)))]; + tensor input_205_cast_fp16 = layer_norm(axes = input_205_axes_0, beta = text_encoder_text_model_encoder_layers_12_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_12_layer_norm2_weight_to_fp16, x = input_203_cast_fp16)[name = tensor("input_205_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412084864)))]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(420473536)))]; + tensor linear_76_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_mlp_fc1_weight_to_fp16, x = input_205_cast_fp16)[name = tensor("linear_76_cast_fp16")]; + tensor input_209_mode_0 = const()[name = tensor("input_209_mode_0"), val = tensor("EXACT")]; + tensor input_209_cast_fp16 = gelu(mode = input_209_mode_0, x = linear_76_cast_fp16)[name = tensor("input_209_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(420481792)))]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(428870464)))]; + tensor linear_77_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_mlp_fc2_weight_to_fp16, x = input_209_cast_fp16)[name = tensor("linear_77_cast_fp16")]; + tensor input_211_cast_fp16 = add(x = input_203_cast_fp16, y = linear_77_cast_fp16)[name = tensor("input_211_cast_fp16")]; + tensor hidden_states_79_axes_0 = const()[name = tensor("hidden_states_79_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(428872576)))]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(428874688)))]; + tensor hidden_states_79_cast_fp16 = layer_norm(axes = hidden_states_79_axes_0, beta = text_encoder_text_model_encoder_layers_13_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_13_layer_norm1_weight_to_fp16, x = input_211_cast_fp16)[name = tensor("hidden_states_79_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(428876800)))]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430974016)))]; + tensor linear_78_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_q_proj_weight_to_fp16, x = hidden_states_79_cast_fp16)[name = tensor("linear_78_cast_fp16")]; + tensor var_1260_to_fp16 = const()[name = tensor("op_1260_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_83_cast_fp16 = mul(x = linear_78_cast_fp16, y = var_1260_to_fp16)[name = tensor("tensor_83_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430976128)))]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433073344)))]; + tensor linear_79_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_k_proj_weight_to_fp16, x = hidden_states_79_cast_fp16)[name = tensor("linear_79_cast_fp16")]; + tensor var_1265 = const()[name = tensor("op_1265"), val = tensor([1, -1, 16, 64])]; + tensor var_1266_cast_fp16 = reshape(shape = var_1265, x = linear_79_cast_fp16)[name = tensor("op_1266_cast_fp16")]; + tensor var_1267_perm_0 = const()[name = tensor("op_1267_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433075456)))]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435172672)))]; + tensor linear_80_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_v_proj_weight_to_fp16, x = hidden_states_79_cast_fp16)[name = tensor("linear_80_cast_fp16")]; + tensor var_1272 = const()[name = tensor("op_1272"), val = tensor([1, -1, 16, 64])]; + tensor var_1273_cast_fp16 = reshape(shape = var_1272, x = linear_80_cast_fp16)[name = tensor("op_1273_cast_fp16")]; + tensor var_1274_perm_0 = const()[name = tensor("op_1274_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1281 = const()[name = tensor("op_1281"), val = tensor([1, 77, 16, 64])]; + tensor var_1282_cast_fp16 = reshape(shape = var_1281, x = tensor_83_cast_fp16)[name = tensor("op_1282_cast_fp16")]; + tensor var_1283_perm_0 = const()[name = tensor("op_1283_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1285 = const()[name = tensor("op_1285"), val = tensor([16, -1, 64])]; + tensor transpose_48 = transpose(perm = var_1283_perm_0, x = var_1282_cast_fp16)[name = tensor("transpose_48")]; + tensor query_states_27_cast_fp16 = reshape(shape = var_1285, x = transpose_48)[name = tensor("query_states_27_cast_fp16")]; + tensor var_1287 = const()[name = tensor("op_1287"), val = tensor([16, -1, 64])]; + tensor transpose_50 = transpose(perm = var_1267_perm_0, x = var_1266_cast_fp16)[name = tensor("transpose_50")]; + tensor key_states_55_cast_fp16 = reshape(shape = var_1287, x = transpose_50)[name = tensor("key_states_55_cast_fp16")]; + tensor var_1289 = const()[name = tensor("op_1289"), val = tensor([16, -1, 64])]; + tensor transpose_49 = transpose(perm = var_1274_perm_0, x = var_1273_cast_fp16)[name = tensor("transpose_49")]; + tensor value_states_55_cast_fp16 = reshape(shape = var_1289, x = transpose_49)[name = tensor("value_states_55_cast_fp16")]; + tensor var_1292_perm_0 = const()[name = tensor("op_1292_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_79_transpose_x_0 = const()[name = tensor("attn_weights_79_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_79_transpose_y_0 = const()[name = tensor("attn_weights_79_transpose_y_0"), val = tensor(false)]; + tensor transpose_47 = transpose(perm = var_1292_perm_0, x = key_states_55_cast_fp16)[name = tensor("transpose_47")]; + tensor attn_weights_79_cast_fp16 = matmul(transpose_x = attn_weights_79_transpose_x_0, transpose_y = attn_weights_79_transpose_y_0, x = query_states_27_cast_fp16, y = transpose_47)[name = tensor("attn_weights_79_cast_fp16")]; + tensor var_1294 = const()[name = tensor("op_1294"), val = tensor([1, 16, 77, 77])]; + tensor var_1295_cast_fp16 = reshape(shape = var_1294, x = attn_weights_79_cast_fp16)[name = tensor("op_1295_cast_fp16")]; + tensor attn_weights_81_cast_fp16 = add(x = var_1295_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_81_cast_fp16")]; + tensor var_1300 = const()[name = tensor("op_1300"), val = tensor([16, 77, 77])]; + tensor input_213_cast_fp16 = reshape(shape = var_1300, x = attn_weights_81_cast_fp16)[name = tensor("input_213_cast_fp16")]; + tensor input_215_cast_fp16 = softmax(axis = var_5, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; + tensor attn_output_79_transpose_x_0 = const()[name = tensor("attn_output_79_transpose_x_0"), val = tensor(false)]; + tensor attn_output_79_transpose_y_0 = const()[name = tensor("attn_output_79_transpose_y_0"), val = tensor(false)]; + tensor attn_output_79_cast_fp16 = matmul(transpose_x = attn_output_79_transpose_x_0, transpose_y = attn_output_79_transpose_y_0, x = input_215_cast_fp16, y = value_states_55_cast_fp16)[name = tensor("attn_output_79_cast_fp16")]; + tensor var_1305 = const()[name = tensor("op_1305"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_81_cast_fp16 = reshape(shape = var_1305, x = attn_output_79_cast_fp16)[name = tensor("attn_output_81_cast_fp16")]; + tensor attn_output_83_perm_0 = const()[name = tensor("attn_output_83_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1308 = const()[name = tensor("op_1308"), val = tensor([1, 77, 1024])]; + tensor transpose_46 = transpose(perm = attn_output_83_perm_0, x = attn_output_81_cast_fp16)[name = tensor("transpose_46")]; + tensor input_217_cast_fp16 = reshape(shape = var_1308, x = transpose_46)[name = tensor("input_217_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435174784)))]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437272000)))]; + tensor linear_81_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_out_proj_weight_to_fp16, x = input_217_cast_fp16)[name = tensor("linear_81_cast_fp16")]; + tensor input_219_cast_fp16 = add(x = input_211_cast_fp16, y = linear_81_cast_fp16)[name = tensor("input_219_cast_fp16")]; + tensor input_221_axes_0 = const()[name = tensor("input_221_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437274112)))]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437276224)))]; + tensor input_221_cast_fp16 = layer_norm(axes = input_221_axes_0, beta = text_encoder_text_model_encoder_layers_13_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_13_layer_norm2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("input_221_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437278336)))]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445667008)))]; + tensor linear_82_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_mlp_fc1_weight_to_fp16, x = input_221_cast_fp16)[name = tensor("linear_82_cast_fp16")]; + tensor input_225_mode_0 = const()[name = tensor("input_225_mode_0"), val = tensor("EXACT")]; + tensor input_225_cast_fp16 = gelu(mode = input_225_mode_0, x = linear_82_cast_fp16)[name = tensor("input_225_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445675264)))]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454063936)))]; + tensor linear_83_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_mlp_fc2_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("linear_83_cast_fp16")]; + tensor input_227_cast_fp16 = add(x = input_219_cast_fp16, y = linear_83_cast_fp16)[name = tensor("input_227_cast_fp16")]; + tensor hidden_states_85_axes_0 = const()[name = tensor("hidden_states_85_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454066048)))]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454068160)))]; + tensor hidden_states_85_cast_fp16 = layer_norm(axes = hidden_states_85_axes_0, beta = text_encoder_text_model_encoder_layers_14_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_14_layer_norm1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("hidden_states_85_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454070272)))]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456167488)))]; + tensor linear_84_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_q_proj_weight_to_fp16, x = hidden_states_85_cast_fp16)[name = tensor("linear_84_cast_fp16")]; + tensor var_1347_to_fp16 = const()[name = tensor("op_1347_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_89_cast_fp16 = mul(x = linear_84_cast_fp16, y = var_1347_to_fp16)[name = tensor("tensor_89_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456169600)))]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(458266816)))]; + tensor linear_85_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_k_proj_weight_to_fp16, x = hidden_states_85_cast_fp16)[name = tensor("linear_85_cast_fp16")]; + tensor var_1352 = const()[name = tensor("op_1352"), val = tensor([1, -1, 16, 64])]; + tensor var_1353_cast_fp16 = reshape(shape = var_1352, x = linear_85_cast_fp16)[name = tensor("op_1353_cast_fp16")]; + tensor var_1354_perm_0 = const()[name = tensor("op_1354_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(458268928)))]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(460366144)))]; + tensor linear_86_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_v_proj_weight_to_fp16, x = hidden_states_85_cast_fp16)[name = tensor("linear_86_cast_fp16")]; + tensor var_1359 = const()[name = tensor("op_1359"), val = tensor([1, -1, 16, 64])]; + tensor var_1360_cast_fp16 = reshape(shape = var_1359, x = linear_86_cast_fp16)[name = tensor("op_1360_cast_fp16")]; + tensor var_1361_perm_0 = const()[name = tensor("op_1361_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1368 = const()[name = tensor("op_1368"), val = tensor([1, 77, 16, 64])]; + tensor var_1369_cast_fp16 = reshape(shape = var_1368, x = tensor_89_cast_fp16)[name = tensor("op_1369_cast_fp16")]; + tensor var_1370_perm_0 = const()[name = tensor("op_1370_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1372 = const()[name = tensor("op_1372"), val = tensor([16, -1, 64])]; + tensor transpose_43 = transpose(perm = var_1370_perm_0, x = var_1369_cast_fp16)[name = tensor("transpose_43")]; + tensor query_states_29_cast_fp16 = reshape(shape = var_1372, x = transpose_43)[name = tensor("query_states_29_cast_fp16")]; + tensor var_1374 = const()[name = tensor("op_1374"), val = tensor([16, -1, 64])]; + tensor transpose_45 = transpose(perm = var_1354_perm_0, x = var_1353_cast_fp16)[name = tensor("transpose_45")]; + tensor key_states_59_cast_fp16 = reshape(shape = var_1374, x = transpose_45)[name = tensor("key_states_59_cast_fp16")]; + tensor var_1376 = const()[name = tensor("op_1376"), val = tensor([16, -1, 64])]; + tensor transpose_44 = transpose(perm = var_1361_perm_0, x = var_1360_cast_fp16)[name = tensor("transpose_44")]; + tensor value_states_59_cast_fp16 = reshape(shape = var_1376, x = transpose_44)[name = tensor("value_states_59_cast_fp16")]; + tensor var_1379_perm_0 = const()[name = tensor("op_1379_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_85_transpose_x_0 = const()[name = tensor("attn_weights_85_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_85_transpose_y_0 = const()[name = tensor("attn_weights_85_transpose_y_0"), val = tensor(false)]; + tensor transpose_42 = transpose(perm = var_1379_perm_0, x = key_states_59_cast_fp16)[name = tensor("transpose_42")]; + tensor attn_weights_85_cast_fp16 = matmul(transpose_x = attn_weights_85_transpose_x_0, transpose_y = attn_weights_85_transpose_y_0, x = query_states_29_cast_fp16, y = transpose_42)[name = tensor("attn_weights_85_cast_fp16")]; + tensor var_1381 = const()[name = tensor("op_1381"), val = tensor([1, 16, 77, 77])]; + tensor var_1382_cast_fp16 = reshape(shape = var_1381, x = attn_weights_85_cast_fp16)[name = tensor("op_1382_cast_fp16")]; + tensor attn_weights_87_cast_fp16 = add(x = var_1382_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_87_cast_fp16")]; + tensor var_1387 = const()[name = tensor("op_1387"), val = tensor([16, 77, 77])]; + tensor input_229_cast_fp16 = reshape(shape = var_1387, x = attn_weights_87_cast_fp16)[name = tensor("input_229_cast_fp16")]; + tensor input_231_cast_fp16 = softmax(axis = var_5, x = input_229_cast_fp16)[name = tensor("input_231_cast_fp16")]; + tensor attn_output_85_transpose_x_0 = const()[name = tensor("attn_output_85_transpose_x_0"), val = tensor(false)]; + tensor attn_output_85_transpose_y_0 = const()[name = tensor("attn_output_85_transpose_y_0"), val = tensor(false)]; + tensor attn_output_85_cast_fp16 = matmul(transpose_x = attn_output_85_transpose_x_0, transpose_y = attn_output_85_transpose_y_0, x = input_231_cast_fp16, y = value_states_59_cast_fp16)[name = tensor("attn_output_85_cast_fp16")]; + tensor var_1392 = const()[name = tensor("op_1392"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_87_cast_fp16 = reshape(shape = var_1392, x = attn_output_85_cast_fp16)[name = tensor("attn_output_87_cast_fp16")]; + tensor attn_output_89_perm_0 = const()[name = tensor("attn_output_89_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1395 = const()[name = tensor("op_1395"), val = tensor([1, 77, 1024])]; + tensor transpose_41 = transpose(perm = attn_output_89_perm_0, x = attn_output_87_cast_fp16)[name = tensor("transpose_41")]; + tensor input_233_cast_fp16 = reshape(shape = var_1395, x = transpose_41)[name = tensor("input_233_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(460368256)))]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(462465472)))]; + tensor linear_87_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_out_proj_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("linear_87_cast_fp16")]; + tensor input_235_cast_fp16 = add(x = input_227_cast_fp16, y = linear_87_cast_fp16)[name = tensor("input_235_cast_fp16")]; + tensor input_237_axes_0 = const()[name = tensor("input_237_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(462467584)))]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(462469696)))]; + tensor input_237_cast_fp16 = layer_norm(axes = input_237_axes_0, beta = text_encoder_text_model_encoder_layers_14_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_14_layer_norm2_weight_to_fp16, x = input_235_cast_fp16)[name = tensor("input_237_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(462471808)))]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470860480)))]; + tensor linear_88_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_mlp_fc1_weight_to_fp16, x = input_237_cast_fp16)[name = tensor("linear_88_cast_fp16")]; + tensor input_241_mode_0 = const()[name = tensor("input_241_mode_0"), val = tensor("EXACT")]; + tensor input_241_cast_fp16 = gelu(mode = input_241_mode_0, x = linear_88_cast_fp16)[name = tensor("input_241_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470868736)))]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(479257408)))]; + tensor linear_89_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_mlp_fc2_weight_to_fp16, x = input_241_cast_fp16)[name = tensor("linear_89_cast_fp16")]; + tensor input_243_cast_fp16 = add(x = input_235_cast_fp16, y = linear_89_cast_fp16)[name = tensor("input_243_cast_fp16")]; + tensor hidden_states_91_axes_0 = const()[name = tensor("hidden_states_91_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(479259520)))]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(479261632)))]; + tensor hidden_states_91_cast_fp16 = layer_norm(axes = hidden_states_91_axes_0, beta = text_encoder_text_model_encoder_layers_15_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_15_layer_norm1_weight_to_fp16, x = input_243_cast_fp16)[name = tensor("hidden_states_91_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(479263744)))]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481360960)))]; + tensor linear_90_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_q_proj_weight_to_fp16, x = hidden_states_91_cast_fp16)[name = tensor("linear_90_cast_fp16")]; + tensor var_1434_to_fp16 = const()[name = tensor("op_1434_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_95_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_1434_to_fp16)[name = tensor("tensor_95_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481363072)))]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483460288)))]; + tensor linear_91_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_k_proj_weight_to_fp16, x = hidden_states_91_cast_fp16)[name = tensor("linear_91_cast_fp16")]; + tensor var_1439 = const()[name = tensor("op_1439"), val = tensor([1, -1, 16, 64])]; + tensor var_1440_cast_fp16 = reshape(shape = var_1439, x = linear_91_cast_fp16)[name = tensor("op_1440_cast_fp16")]; + tensor var_1441_perm_0 = const()[name = tensor("op_1441_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483462400)))]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485559616)))]; + tensor linear_92_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_v_proj_weight_to_fp16, x = hidden_states_91_cast_fp16)[name = tensor("linear_92_cast_fp16")]; + tensor var_1446 = const()[name = tensor("op_1446"), val = tensor([1, -1, 16, 64])]; + tensor var_1447_cast_fp16 = reshape(shape = var_1446, x = linear_92_cast_fp16)[name = tensor("op_1447_cast_fp16")]; + tensor var_1448_perm_0 = const()[name = tensor("op_1448_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1455 = const()[name = tensor("op_1455"), val = tensor([1, 77, 16, 64])]; + tensor var_1456_cast_fp16 = reshape(shape = var_1455, x = tensor_95_cast_fp16)[name = tensor("op_1456_cast_fp16")]; + tensor var_1457_perm_0 = const()[name = tensor("op_1457_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1459 = const()[name = tensor("op_1459"), val = tensor([16, -1, 64])]; + tensor transpose_38 = transpose(perm = var_1457_perm_0, x = var_1456_cast_fp16)[name = tensor("transpose_38")]; + tensor query_states_31_cast_fp16 = reshape(shape = var_1459, x = transpose_38)[name = tensor("query_states_31_cast_fp16")]; + tensor var_1461 = const()[name = tensor("op_1461"), val = tensor([16, -1, 64])]; + tensor transpose_40 = transpose(perm = var_1441_perm_0, x = var_1440_cast_fp16)[name = tensor("transpose_40")]; + tensor key_states_63_cast_fp16 = reshape(shape = var_1461, x = transpose_40)[name = tensor("key_states_63_cast_fp16")]; + tensor var_1463 = const()[name = tensor("op_1463"), val = tensor([16, -1, 64])]; + tensor transpose_39 = transpose(perm = var_1448_perm_0, x = var_1447_cast_fp16)[name = tensor("transpose_39")]; + tensor value_states_63_cast_fp16 = reshape(shape = var_1463, x = transpose_39)[name = tensor("value_states_63_cast_fp16")]; + tensor var_1466_perm_0 = const()[name = tensor("op_1466_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_91_transpose_x_0 = const()[name = tensor("attn_weights_91_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_91_transpose_y_0 = const()[name = tensor("attn_weights_91_transpose_y_0"), val = tensor(false)]; + tensor transpose_37 = transpose(perm = var_1466_perm_0, x = key_states_63_cast_fp16)[name = tensor("transpose_37")]; + tensor attn_weights_91_cast_fp16 = matmul(transpose_x = attn_weights_91_transpose_x_0, transpose_y = attn_weights_91_transpose_y_0, x = query_states_31_cast_fp16, y = transpose_37)[name = tensor("attn_weights_91_cast_fp16")]; + tensor var_1468 = const()[name = tensor("op_1468"), val = tensor([1, 16, 77, 77])]; + tensor var_1469_cast_fp16 = reshape(shape = var_1468, x = attn_weights_91_cast_fp16)[name = tensor("op_1469_cast_fp16")]; + tensor attn_weights_93_cast_fp16 = add(x = var_1469_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_93_cast_fp16")]; + tensor var_1474 = const()[name = tensor("op_1474"), val = tensor([16, 77, 77])]; + tensor input_245_cast_fp16 = reshape(shape = var_1474, x = attn_weights_93_cast_fp16)[name = tensor("input_245_cast_fp16")]; + tensor input_247_cast_fp16 = softmax(axis = var_5, x = input_245_cast_fp16)[name = tensor("input_247_cast_fp16")]; + tensor attn_output_91_transpose_x_0 = const()[name = tensor("attn_output_91_transpose_x_0"), val = tensor(false)]; + tensor attn_output_91_transpose_y_0 = const()[name = tensor("attn_output_91_transpose_y_0"), val = tensor(false)]; + tensor attn_output_91_cast_fp16 = matmul(transpose_x = attn_output_91_transpose_x_0, transpose_y = attn_output_91_transpose_y_0, x = input_247_cast_fp16, y = value_states_63_cast_fp16)[name = tensor("attn_output_91_cast_fp16")]; + tensor var_1479 = const()[name = tensor("op_1479"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_93_cast_fp16 = reshape(shape = var_1479, x = attn_output_91_cast_fp16)[name = tensor("attn_output_93_cast_fp16")]; + tensor attn_output_95_perm_0 = const()[name = tensor("attn_output_95_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1482 = const()[name = tensor("op_1482"), val = tensor([1, 77, 1024])]; + tensor transpose_36 = transpose(perm = attn_output_95_perm_0, x = attn_output_93_cast_fp16)[name = tensor("transpose_36")]; + tensor input_249_cast_fp16 = reshape(shape = var_1482, x = transpose_36)[name = tensor("input_249_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485561728)))]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(487658944)))]; + tensor linear_93_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_out_proj_weight_to_fp16, x = input_249_cast_fp16)[name = tensor("linear_93_cast_fp16")]; + tensor input_251_cast_fp16 = add(x = input_243_cast_fp16, y = linear_93_cast_fp16)[name = tensor("input_251_cast_fp16")]; + tensor input_253_axes_0 = const()[name = tensor("input_253_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(487661056)))]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(487663168)))]; + tensor input_253_cast_fp16 = layer_norm(axes = input_253_axes_0, beta = text_encoder_text_model_encoder_layers_15_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_15_layer_norm2_weight_to_fp16, x = input_251_cast_fp16)[name = tensor("input_253_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(487665280)))]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(496053952)))]; + tensor linear_94_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_mlp_fc1_weight_to_fp16, x = input_253_cast_fp16)[name = tensor("linear_94_cast_fp16")]; + tensor input_257_mode_0 = const()[name = tensor("input_257_mode_0"), val = tensor("EXACT")]; + tensor input_257_cast_fp16 = gelu(mode = input_257_mode_0, x = linear_94_cast_fp16)[name = tensor("input_257_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(496062208)))]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504450880)))]; + tensor linear_95_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_mlp_fc2_weight_to_fp16, x = input_257_cast_fp16)[name = tensor("linear_95_cast_fp16")]; + tensor input_259_cast_fp16 = add(x = input_251_cast_fp16, y = linear_95_cast_fp16)[name = tensor("input_259_cast_fp16")]; + tensor hidden_states_97_axes_0 = const()[name = tensor("hidden_states_97_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504452992)))]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504455104)))]; + tensor hidden_states_97_cast_fp16 = layer_norm(axes = hidden_states_97_axes_0, beta = text_encoder_text_model_encoder_layers_16_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_16_layer_norm1_weight_to_fp16, x = input_259_cast_fp16)[name = tensor("hidden_states_97_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504457216)))]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506554432)))]; + tensor linear_96_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_q_proj_weight_to_fp16, x = hidden_states_97_cast_fp16)[name = tensor("linear_96_cast_fp16")]; + tensor var_1521_to_fp16 = const()[name = tensor("op_1521_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_101_cast_fp16 = mul(x = linear_96_cast_fp16, y = var_1521_to_fp16)[name = tensor("tensor_101_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506556544)))]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(508653760)))]; + tensor linear_97_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_k_proj_weight_to_fp16, x = hidden_states_97_cast_fp16)[name = tensor("linear_97_cast_fp16")]; + tensor var_1526 = const()[name = tensor("op_1526"), val = tensor([1, -1, 16, 64])]; + tensor var_1527_cast_fp16 = reshape(shape = var_1526, x = linear_97_cast_fp16)[name = tensor("op_1527_cast_fp16")]; + tensor var_1528_perm_0 = const()[name = tensor("op_1528_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(508655872)))]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510753088)))]; + tensor linear_98_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_v_proj_weight_to_fp16, x = hidden_states_97_cast_fp16)[name = tensor("linear_98_cast_fp16")]; + tensor var_1533 = const()[name = tensor("op_1533"), val = tensor([1, -1, 16, 64])]; + tensor var_1534_cast_fp16 = reshape(shape = var_1533, x = linear_98_cast_fp16)[name = tensor("op_1534_cast_fp16")]; + tensor var_1535_perm_0 = const()[name = tensor("op_1535_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1542 = const()[name = tensor("op_1542"), val = tensor([1, 77, 16, 64])]; + tensor var_1543_cast_fp16 = reshape(shape = var_1542, x = tensor_101_cast_fp16)[name = tensor("op_1543_cast_fp16")]; + tensor var_1544_perm_0 = const()[name = tensor("op_1544_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1546 = const()[name = tensor("op_1546"), val = tensor([16, -1, 64])]; + tensor transpose_33 = transpose(perm = var_1544_perm_0, x = var_1543_cast_fp16)[name = tensor("transpose_33")]; + tensor query_states_33_cast_fp16 = reshape(shape = var_1546, x = transpose_33)[name = tensor("query_states_33_cast_fp16")]; + tensor var_1548 = const()[name = tensor("op_1548"), val = tensor([16, -1, 64])]; + tensor transpose_35 = transpose(perm = var_1528_perm_0, x = var_1527_cast_fp16)[name = tensor("transpose_35")]; + tensor key_states_67_cast_fp16 = reshape(shape = var_1548, x = transpose_35)[name = tensor("key_states_67_cast_fp16")]; + tensor var_1550 = const()[name = tensor("op_1550"), val = tensor([16, -1, 64])]; + tensor transpose_34 = transpose(perm = var_1535_perm_0, x = var_1534_cast_fp16)[name = tensor("transpose_34")]; + tensor value_states_67_cast_fp16 = reshape(shape = var_1550, x = transpose_34)[name = tensor("value_states_67_cast_fp16")]; + tensor var_1553_perm_0 = const()[name = tensor("op_1553_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_97_transpose_x_0 = const()[name = tensor("attn_weights_97_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_97_transpose_y_0 = const()[name = tensor("attn_weights_97_transpose_y_0"), val = tensor(false)]; + tensor transpose_32 = transpose(perm = var_1553_perm_0, x = key_states_67_cast_fp16)[name = tensor("transpose_32")]; + tensor attn_weights_97_cast_fp16 = matmul(transpose_x = attn_weights_97_transpose_x_0, transpose_y = attn_weights_97_transpose_y_0, x = query_states_33_cast_fp16, y = transpose_32)[name = tensor("attn_weights_97_cast_fp16")]; + tensor var_1555 = const()[name = tensor("op_1555"), val = tensor([1, 16, 77, 77])]; + tensor var_1556_cast_fp16 = reshape(shape = var_1555, x = attn_weights_97_cast_fp16)[name = tensor("op_1556_cast_fp16")]; + tensor attn_weights_99_cast_fp16 = add(x = var_1556_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_99_cast_fp16")]; + tensor var_1561 = const()[name = tensor("op_1561"), val = tensor([16, 77, 77])]; + tensor input_261_cast_fp16 = reshape(shape = var_1561, x = attn_weights_99_cast_fp16)[name = tensor("input_261_cast_fp16")]; + tensor input_263_cast_fp16 = softmax(axis = var_5, x = input_261_cast_fp16)[name = tensor("input_263_cast_fp16")]; + tensor attn_output_97_transpose_x_0 = const()[name = tensor("attn_output_97_transpose_x_0"), val = tensor(false)]; + tensor attn_output_97_transpose_y_0 = const()[name = tensor("attn_output_97_transpose_y_0"), val = tensor(false)]; + tensor attn_output_97_cast_fp16 = matmul(transpose_x = attn_output_97_transpose_x_0, transpose_y = attn_output_97_transpose_y_0, x = input_263_cast_fp16, y = value_states_67_cast_fp16)[name = tensor("attn_output_97_cast_fp16")]; + tensor var_1566 = const()[name = tensor("op_1566"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_99_cast_fp16 = reshape(shape = var_1566, x = attn_output_97_cast_fp16)[name = tensor("attn_output_99_cast_fp16")]; + tensor attn_output_101_perm_0 = const()[name = tensor("attn_output_101_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1569 = const()[name = tensor("op_1569"), val = tensor([1, 77, 1024])]; + tensor transpose_31 = transpose(perm = attn_output_101_perm_0, x = attn_output_99_cast_fp16)[name = tensor("transpose_31")]; + tensor input_265_cast_fp16 = reshape(shape = var_1569, x = transpose_31)[name = tensor("input_265_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510755200)))]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(512852416)))]; + tensor linear_99_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_out_proj_weight_to_fp16, x = input_265_cast_fp16)[name = tensor("linear_99_cast_fp16")]; + tensor input_267_cast_fp16 = add(x = input_259_cast_fp16, y = linear_99_cast_fp16)[name = tensor("input_267_cast_fp16")]; + tensor input_269_axes_0 = const()[name = tensor("input_269_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(512854528)))]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(512856640)))]; + tensor input_269_cast_fp16 = layer_norm(axes = input_269_axes_0, beta = text_encoder_text_model_encoder_layers_16_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_16_layer_norm2_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("input_269_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(512858752)))]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(521247424)))]; + tensor linear_100_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_mlp_fc1_weight_to_fp16, x = input_269_cast_fp16)[name = tensor("linear_100_cast_fp16")]; + tensor input_273_mode_0 = const()[name = tensor("input_273_mode_0"), val = tensor("EXACT")]; + tensor input_273_cast_fp16 = gelu(mode = input_273_mode_0, x = linear_100_cast_fp16)[name = tensor("input_273_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(521255680)))]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529644352)))]; + tensor linear_101_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_mlp_fc2_weight_to_fp16, x = input_273_cast_fp16)[name = tensor("linear_101_cast_fp16")]; + tensor input_275_cast_fp16 = add(x = input_267_cast_fp16, y = linear_101_cast_fp16)[name = tensor("input_275_cast_fp16")]; + tensor hidden_states_103_axes_0 = const()[name = tensor("hidden_states_103_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529646464)))]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529648576)))]; + tensor hidden_states_103_cast_fp16 = layer_norm(axes = hidden_states_103_axes_0, beta = text_encoder_text_model_encoder_layers_17_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_17_layer_norm1_weight_to_fp16, x = input_275_cast_fp16)[name = tensor("hidden_states_103_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529650688)))]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531747904)))]; + tensor linear_102_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_q_proj_weight_to_fp16, x = hidden_states_103_cast_fp16)[name = tensor("linear_102_cast_fp16")]; + tensor var_1608_to_fp16 = const()[name = tensor("op_1608_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_107_cast_fp16 = mul(x = linear_102_cast_fp16, y = var_1608_to_fp16)[name = tensor("tensor_107_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531750016)))]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(533847232)))]; + tensor linear_103_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_k_proj_weight_to_fp16, x = hidden_states_103_cast_fp16)[name = tensor("linear_103_cast_fp16")]; + tensor var_1613 = const()[name = tensor("op_1613"), val = tensor([1, -1, 16, 64])]; + tensor var_1614_cast_fp16 = reshape(shape = var_1613, x = linear_103_cast_fp16)[name = tensor("op_1614_cast_fp16")]; + tensor var_1615_perm_0 = const()[name = tensor("op_1615_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(533849344)))]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535946560)))]; + tensor linear_104_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_v_proj_weight_to_fp16, x = hidden_states_103_cast_fp16)[name = tensor("linear_104_cast_fp16")]; + tensor var_1620 = const()[name = tensor("op_1620"), val = tensor([1, -1, 16, 64])]; + tensor var_1621_cast_fp16 = reshape(shape = var_1620, x = linear_104_cast_fp16)[name = tensor("op_1621_cast_fp16")]; + tensor var_1622_perm_0 = const()[name = tensor("op_1622_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1629 = const()[name = tensor("op_1629"), val = tensor([1, 77, 16, 64])]; + tensor var_1630_cast_fp16 = reshape(shape = var_1629, x = tensor_107_cast_fp16)[name = tensor("op_1630_cast_fp16")]; + tensor var_1631_perm_0 = const()[name = tensor("op_1631_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1633 = const()[name = tensor("op_1633"), val = tensor([16, -1, 64])]; + tensor transpose_28 = transpose(perm = var_1631_perm_0, x = var_1630_cast_fp16)[name = tensor("transpose_28")]; + tensor query_states_35_cast_fp16 = reshape(shape = var_1633, x = transpose_28)[name = tensor("query_states_35_cast_fp16")]; + tensor var_1635 = const()[name = tensor("op_1635"), val = tensor([16, -1, 64])]; + tensor transpose_30 = transpose(perm = var_1615_perm_0, x = var_1614_cast_fp16)[name = tensor("transpose_30")]; + tensor key_states_71_cast_fp16 = reshape(shape = var_1635, x = transpose_30)[name = tensor("key_states_71_cast_fp16")]; + tensor var_1637 = const()[name = tensor("op_1637"), val = tensor([16, -1, 64])]; + tensor transpose_29 = transpose(perm = var_1622_perm_0, x = var_1621_cast_fp16)[name = tensor("transpose_29")]; + tensor value_states_71_cast_fp16 = reshape(shape = var_1637, x = transpose_29)[name = tensor("value_states_71_cast_fp16")]; + tensor var_1640_perm_0 = const()[name = tensor("op_1640_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_103_transpose_x_0 = const()[name = tensor("attn_weights_103_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_103_transpose_y_0 = const()[name = tensor("attn_weights_103_transpose_y_0"), val = tensor(false)]; + tensor transpose_27 = transpose(perm = var_1640_perm_0, x = key_states_71_cast_fp16)[name = tensor("transpose_27")]; + tensor attn_weights_103_cast_fp16 = matmul(transpose_x = attn_weights_103_transpose_x_0, transpose_y = attn_weights_103_transpose_y_0, x = query_states_35_cast_fp16, y = transpose_27)[name = tensor("attn_weights_103_cast_fp16")]; + tensor var_1642 = const()[name = tensor("op_1642"), val = tensor([1, 16, 77, 77])]; + tensor var_1643_cast_fp16 = reshape(shape = var_1642, x = attn_weights_103_cast_fp16)[name = tensor("op_1643_cast_fp16")]; + tensor attn_weights_105_cast_fp16 = add(x = var_1643_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_105_cast_fp16")]; + tensor var_1648 = const()[name = tensor("op_1648"), val = tensor([16, 77, 77])]; + tensor input_277_cast_fp16 = reshape(shape = var_1648, x = attn_weights_105_cast_fp16)[name = tensor("input_277_cast_fp16")]; + tensor input_279_cast_fp16 = softmax(axis = var_5, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; + tensor attn_output_103_transpose_x_0 = const()[name = tensor("attn_output_103_transpose_x_0"), val = tensor(false)]; + tensor attn_output_103_transpose_y_0 = const()[name = tensor("attn_output_103_transpose_y_0"), val = tensor(false)]; + tensor attn_output_103_cast_fp16 = matmul(transpose_x = attn_output_103_transpose_x_0, transpose_y = attn_output_103_transpose_y_0, x = input_279_cast_fp16, y = value_states_71_cast_fp16)[name = tensor("attn_output_103_cast_fp16")]; + tensor var_1653 = const()[name = tensor("op_1653"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_105_cast_fp16 = reshape(shape = var_1653, x = attn_output_103_cast_fp16)[name = tensor("attn_output_105_cast_fp16")]; + tensor attn_output_107_perm_0 = const()[name = tensor("attn_output_107_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1656 = const()[name = tensor("op_1656"), val = tensor([1, 77, 1024])]; + tensor transpose_26 = transpose(perm = attn_output_107_perm_0, x = attn_output_105_cast_fp16)[name = tensor("transpose_26")]; + tensor input_281_cast_fp16 = reshape(shape = var_1656, x = transpose_26)[name = tensor("input_281_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535948672)))]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(538045888)))]; + tensor linear_105_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_out_proj_weight_to_fp16, x = input_281_cast_fp16)[name = tensor("linear_105_cast_fp16")]; + tensor input_283_cast_fp16 = add(x = input_275_cast_fp16, y = linear_105_cast_fp16)[name = tensor("input_283_cast_fp16")]; + tensor input_285_axes_0 = const()[name = tensor("input_285_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(538048000)))]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(538050112)))]; + tensor input_285_cast_fp16 = layer_norm(axes = input_285_axes_0, beta = text_encoder_text_model_encoder_layers_17_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_17_layer_norm2_weight_to_fp16, x = input_283_cast_fp16)[name = tensor("input_285_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(538052224)))]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(546440896)))]; + tensor linear_106_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_mlp_fc1_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("linear_106_cast_fp16")]; + tensor input_289_mode_0 = const()[name = tensor("input_289_mode_0"), val = tensor("EXACT")]; + tensor input_289_cast_fp16 = gelu(mode = input_289_mode_0, x = linear_106_cast_fp16)[name = tensor("input_289_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(546449152)))]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554837824)))]; + tensor linear_107_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_mlp_fc2_weight_to_fp16, x = input_289_cast_fp16)[name = tensor("linear_107_cast_fp16")]; + tensor input_291_cast_fp16 = add(x = input_283_cast_fp16, y = linear_107_cast_fp16)[name = tensor("input_291_cast_fp16")]; + tensor hidden_states_109_axes_0 = const()[name = tensor("hidden_states_109_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554839936)))]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554842048)))]; + tensor hidden_states_109_cast_fp16 = layer_norm(axes = hidden_states_109_axes_0, beta = text_encoder_text_model_encoder_layers_18_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_18_layer_norm1_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("hidden_states_109_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554844160)))]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556941376)))]; + tensor linear_108_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_q_proj_weight_to_fp16, x = hidden_states_109_cast_fp16)[name = tensor("linear_108_cast_fp16")]; + tensor var_1695_to_fp16 = const()[name = tensor("op_1695_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_113_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_1695_to_fp16)[name = tensor("tensor_113_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556943488)))]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559040704)))]; + tensor linear_109_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_k_proj_weight_to_fp16, x = hidden_states_109_cast_fp16)[name = tensor("linear_109_cast_fp16")]; + tensor var_1700 = const()[name = tensor("op_1700"), val = tensor([1, -1, 16, 64])]; + tensor var_1701_cast_fp16 = reshape(shape = var_1700, x = linear_109_cast_fp16)[name = tensor("op_1701_cast_fp16")]; + tensor var_1702_perm_0 = const()[name = tensor("op_1702_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559042816)))]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(561140032)))]; + tensor linear_110_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_v_proj_weight_to_fp16, x = hidden_states_109_cast_fp16)[name = tensor("linear_110_cast_fp16")]; + tensor var_1707 = const()[name = tensor("op_1707"), val = tensor([1, -1, 16, 64])]; + tensor var_1708_cast_fp16 = reshape(shape = var_1707, x = linear_110_cast_fp16)[name = tensor("op_1708_cast_fp16")]; + tensor var_1709_perm_0 = const()[name = tensor("op_1709_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1716 = const()[name = tensor("op_1716"), val = tensor([1, 77, 16, 64])]; + tensor var_1717_cast_fp16 = reshape(shape = var_1716, x = tensor_113_cast_fp16)[name = tensor("op_1717_cast_fp16")]; + tensor var_1718_perm_0 = const()[name = tensor("op_1718_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1720 = const()[name = tensor("op_1720"), val = tensor([16, -1, 64])]; + tensor transpose_23 = transpose(perm = var_1718_perm_0, x = var_1717_cast_fp16)[name = tensor("transpose_23")]; + tensor query_states_37_cast_fp16 = reshape(shape = var_1720, x = transpose_23)[name = tensor("query_states_37_cast_fp16")]; + tensor var_1722 = const()[name = tensor("op_1722"), val = tensor([16, -1, 64])]; + tensor transpose_25 = transpose(perm = var_1702_perm_0, x = var_1701_cast_fp16)[name = tensor("transpose_25")]; + tensor key_states_75_cast_fp16 = reshape(shape = var_1722, x = transpose_25)[name = tensor("key_states_75_cast_fp16")]; + tensor var_1724 = const()[name = tensor("op_1724"), val = tensor([16, -1, 64])]; + tensor transpose_24 = transpose(perm = var_1709_perm_0, x = var_1708_cast_fp16)[name = tensor("transpose_24")]; + tensor value_states_75_cast_fp16 = reshape(shape = var_1724, x = transpose_24)[name = tensor("value_states_75_cast_fp16")]; + tensor var_1727_perm_0 = const()[name = tensor("op_1727_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_109_transpose_x_0 = const()[name = tensor("attn_weights_109_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_109_transpose_y_0 = const()[name = tensor("attn_weights_109_transpose_y_0"), val = tensor(false)]; + tensor transpose_22 = transpose(perm = var_1727_perm_0, x = key_states_75_cast_fp16)[name = tensor("transpose_22")]; + tensor attn_weights_109_cast_fp16 = matmul(transpose_x = attn_weights_109_transpose_x_0, transpose_y = attn_weights_109_transpose_y_0, x = query_states_37_cast_fp16, y = transpose_22)[name = tensor("attn_weights_109_cast_fp16")]; + tensor var_1729 = const()[name = tensor("op_1729"), val = tensor([1, 16, 77, 77])]; + tensor var_1730_cast_fp16 = reshape(shape = var_1729, x = attn_weights_109_cast_fp16)[name = tensor("op_1730_cast_fp16")]; + tensor attn_weights_111_cast_fp16 = add(x = var_1730_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_111_cast_fp16")]; + tensor var_1735 = const()[name = tensor("op_1735"), val = tensor([16, 77, 77])]; + tensor input_293_cast_fp16 = reshape(shape = var_1735, x = attn_weights_111_cast_fp16)[name = tensor("input_293_cast_fp16")]; + tensor input_295_cast_fp16 = softmax(axis = var_5, x = input_293_cast_fp16)[name = tensor("input_295_cast_fp16")]; + tensor attn_output_109_transpose_x_0 = const()[name = tensor("attn_output_109_transpose_x_0"), val = tensor(false)]; + tensor attn_output_109_transpose_y_0 = const()[name = tensor("attn_output_109_transpose_y_0"), val = tensor(false)]; + tensor attn_output_109_cast_fp16 = matmul(transpose_x = attn_output_109_transpose_x_0, transpose_y = attn_output_109_transpose_y_0, x = input_295_cast_fp16, y = value_states_75_cast_fp16)[name = tensor("attn_output_109_cast_fp16")]; + tensor var_1740 = const()[name = tensor("op_1740"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_111_cast_fp16 = reshape(shape = var_1740, x = attn_output_109_cast_fp16)[name = tensor("attn_output_111_cast_fp16")]; + tensor attn_output_113_perm_0 = const()[name = tensor("attn_output_113_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1743 = const()[name = tensor("op_1743"), val = tensor([1, 77, 1024])]; + tensor transpose_21 = transpose(perm = attn_output_113_perm_0, x = attn_output_111_cast_fp16)[name = tensor("transpose_21")]; + tensor input_297_cast_fp16 = reshape(shape = var_1743, x = transpose_21)[name = tensor("input_297_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(561142144)))]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563239360)))]; + tensor linear_111_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_out_proj_weight_to_fp16, x = input_297_cast_fp16)[name = tensor("linear_111_cast_fp16")]; + tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_111_cast_fp16)[name = tensor("input_299_cast_fp16")]; + tensor input_301_axes_0 = const()[name = tensor("input_301_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563241472)))]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563243584)))]; + tensor input_301_cast_fp16 = layer_norm(axes = input_301_axes_0, beta = text_encoder_text_model_encoder_layers_18_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_18_layer_norm2_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("input_301_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563245696)))]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571634368)))]; + tensor linear_112_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_mlp_fc1_weight_to_fp16, x = input_301_cast_fp16)[name = tensor("linear_112_cast_fp16")]; + tensor input_305_mode_0 = const()[name = tensor("input_305_mode_0"), val = tensor("EXACT")]; + tensor input_305_cast_fp16 = gelu(mode = input_305_mode_0, x = linear_112_cast_fp16)[name = tensor("input_305_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571642624)))]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580031296)))]; + tensor linear_113_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_mlp_fc2_weight_to_fp16, x = input_305_cast_fp16)[name = tensor("linear_113_cast_fp16")]; + tensor input_307_cast_fp16 = add(x = input_299_cast_fp16, y = linear_113_cast_fp16)[name = tensor("input_307_cast_fp16")]; + tensor hidden_states_115_axes_0 = const()[name = tensor("hidden_states_115_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580033408)))]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580035520)))]; + tensor hidden_states_115_cast_fp16 = layer_norm(axes = hidden_states_115_axes_0, beta = text_encoder_text_model_encoder_layers_19_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_19_layer_norm1_weight_to_fp16, x = input_307_cast_fp16)[name = tensor("hidden_states_115_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(580037632)))]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(582134848)))]; + tensor linear_114_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_q_proj_weight_to_fp16, x = hidden_states_115_cast_fp16)[name = tensor("linear_114_cast_fp16")]; + tensor var_1782_to_fp16 = const()[name = tensor("op_1782_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_119_cast_fp16 = mul(x = linear_114_cast_fp16, y = var_1782_to_fp16)[name = tensor("tensor_119_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(582136960)))]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584234176)))]; + tensor linear_115_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_k_proj_weight_to_fp16, x = hidden_states_115_cast_fp16)[name = tensor("linear_115_cast_fp16")]; + tensor var_1787 = const()[name = tensor("op_1787"), val = tensor([1, -1, 16, 64])]; + tensor var_1788_cast_fp16 = reshape(shape = var_1787, x = linear_115_cast_fp16)[name = tensor("op_1788_cast_fp16")]; + tensor var_1789_perm_0 = const()[name = tensor("op_1789_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584236288)))]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(586333504)))]; + tensor linear_116_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_v_proj_weight_to_fp16, x = hidden_states_115_cast_fp16)[name = tensor("linear_116_cast_fp16")]; + tensor var_1794 = const()[name = tensor("op_1794"), val = tensor([1, -1, 16, 64])]; + tensor var_1795_cast_fp16 = reshape(shape = var_1794, x = linear_116_cast_fp16)[name = tensor("op_1795_cast_fp16")]; + tensor var_1796_perm_0 = const()[name = tensor("op_1796_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1803 = const()[name = tensor("op_1803"), val = tensor([1, 77, 16, 64])]; + tensor var_1804_cast_fp16 = reshape(shape = var_1803, x = tensor_119_cast_fp16)[name = tensor("op_1804_cast_fp16")]; + tensor var_1805_perm_0 = const()[name = tensor("op_1805_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1807 = const()[name = tensor("op_1807"), val = tensor([16, -1, 64])]; + tensor transpose_18 = transpose(perm = var_1805_perm_0, x = var_1804_cast_fp16)[name = tensor("transpose_18")]; + tensor query_states_39_cast_fp16 = reshape(shape = var_1807, x = transpose_18)[name = tensor("query_states_39_cast_fp16")]; + tensor var_1809 = const()[name = tensor("op_1809"), val = tensor([16, -1, 64])]; + tensor transpose_20 = transpose(perm = var_1789_perm_0, x = var_1788_cast_fp16)[name = tensor("transpose_20")]; + tensor key_states_79_cast_fp16 = reshape(shape = var_1809, x = transpose_20)[name = tensor("key_states_79_cast_fp16")]; + tensor var_1811 = const()[name = tensor("op_1811"), val = tensor([16, -1, 64])]; + tensor transpose_19 = transpose(perm = var_1796_perm_0, x = var_1795_cast_fp16)[name = tensor("transpose_19")]; + tensor value_states_79_cast_fp16 = reshape(shape = var_1811, x = transpose_19)[name = tensor("value_states_79_cast_fp16")]; + tensor var_1814_perm_0 = const()[name = tensor("op_1814_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_115_transpose_x_0 = const()[name = tensor("attn_weights_115_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_115_transpose_y_0 = const()[name = tensor("attn_weights_115_transpose_y_0"), val = tensor(false)]; + tensor transpose_17 = transpose(perm = var_1814_perm_0, x = key_states_79_cast_fp16)[name = tensor("transpose_17")]; + tensor attn_weights_115_cast_fp16 = matmul(transpose_x = attn_weights_115_transpose_x_0, transpose_y = attn_weights_115_transpose_y_0, x = query_states_39_cast_fp16, y = transpose_17)[name = tensor("attn_weights_115_cast_fp16")]; + tensor var_1816 = const()[name = tensor("op_1816"), val = tensor([1, 16, 77, 77])]; + tensor var_1817_cast_fp16 = reshape(shape = var_1816, x = attn_weights_115_cast_fp16)[name = tensor("op_1817_cast_fp16")]; + tensor attn_weights_117_cast_fp16 = add(x = var_1817_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_117_cast_fp16")]; + tensor var_1822 = const()[name = tensor("op_1822"), val = tensor([16, 77, 77])]; + tensor input_309_cast_fp16 = reshape(shape = var_1822, x = attn_weights_117_cast_fp16)[name = tensor("input_309_cast_fp16")]; + tensor input_311_cast_fp16 = softmax(axis = var_5, x = input_309_cast_fp16)[name = tensor("input_311_cast_fp16")]; + tensor attn_output_115_transpose_x_0 = const()[name = tensor("attn_output_115_transpose_x_0"), val = tensor(false)]; + tensor attn_output_115_transpose_y_0 = const()[name = tensor("attn_output_115_transpose_y_0"), val = tensor(false)]; + tensor attn_output_115_cast_fp16 = matmul(transpose_x = attn_output_115_transpose_x_0, transpose_y = attn_output_115_transpose_y_0, x = input_311_cast_fp16, y = value_states_79_cast_fp16)[name = tensor("attn_output_115_cast_fp16")]; + tensor var_1827 = const()[name = tensor("op_1827"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_117_cast_fp16 = reshape(shape = var_1827, x = attn_output_115_cast_fp16)[name = tensor("attn_output_117_cast_fp16")]; + tensor attn_output_119_perm_0 = const()[name = tensor("attn_output_119_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1830 = const()[name = tensor("op_1830"), val = tensor([1, 77, 1024])]; + tensor transpose_16 = transpose(perm = attn_output_119_perm_0, x = attn_output_117_cast_fp16)[name = tensor("transpose_16")]; + tensor input_313_cast_fp16 = reshape(shape = var_1830, x = transpose_16)[name = tensor("input_313_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(586335616)))]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588432832)))]; + tensor linear_117_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_out_proj_weight_to_fp16, x = input_313_cast_fp16)[name = tensor("linear_117_cast_fp16")]; + tensor input_315_cast_fp16 = add(x = input_307_cast_fp16, y = linear_117_cast_fp16)[name = tensor("input_315_cast_fp16")]; + tensor input_317_axes_0 = const()[name = tensor("input_317_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588434944)))]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588437056)))]; + tensor input_317_cast_fp16 = layer_norm(axes = input_317_axes_0, beta = text_encoder_text_model_encoder_layers_19_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_19_layer_norm2_weight_to_fp16, x = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588439168)))]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596827840)))]; + tensor linear_118_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_mlp_fc1_weight_to_fp16, x = input_317_cast_fp16)[name = tensor("linear_118_cast_fp16")]; + tensor input_321_mode_0 = const()[name = tensor("input_321_mode_0"), val = tensor("EXACT")]; + tensor input_321_cast_fp16 = gelu(mode = input_321_mode_0, x = linear_118_cast_fp16)[name = tensor("input_321_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(596836096)))]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(605224768)))]; + tensor linear_119_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_mlp_fc2_weight_to_fp16, x = input_321_cast_fp16)[name = tensor("linear_119_cast_fp16")]; + tensor input_323_cast_fp16 = add(x = input_315_cast_fp16, y = linear_119_cast_fp16)[name = tensor("input_323_cast_fp16")]; + tensor hidden_states_121_axes_0 = const()[name = tensor("hidden_states_121_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(605226880)))]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(605228992)))]; + tensor hidden_states_121_cast_fp16 = layer_norm(axes = hidden_states_121_axes_0, beta = text_encoder_text_model_encoder_layers_20_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_20_layer_norm1_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("hidden_states_121_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(605231104)))]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(607328320)))]; + tensor linear_120_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_q_proj_weight_to_fp16, x = hidden_states_121_cast_fp16)[name = tensor("linear_120_cast_fp16")]; + tensor var_1869_to_fp16 = const()[name = tensor("op_1869_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_125_cast_fp16 = mul(x = linear_120_cast_fp16, y = var_1869_to_fp16)[name = tensor("tensor_125_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(607330432)))]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(609427648)))]; + tensor linear_121_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_k_proj_weight_to_fp16, x = hidden_states_121_cast_fp16)[name = tensor("linear_121_cast_fp16")]; + tensor var_1874 = const()[name = tensor("op_1874"), val = tensor([1, -1, 16, 64])]; + tensor var_1875_cast_fp16 = reshape(shape = var_1874, x = linear_121_cast_fp16)[name = tensor("op_1875_cast_fp16")]; + tensor var_1876_perm_0 = const()[name = tensor("op_1876_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(609429760)))]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(611526976)))]; + tensor linear_122_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_v_proj_weight_to_fp16, x = hidden_states_121_cast_fp16)[name = tensor("linear_122_cast_fp16")]; + tensor var_1881 = const()[name = tensor("op_1881"), val = tensor([1, -1, 16, 64])]; + tensor var_1882_cast_fp16 = reshape(shape = var_1881, x = linear_122_cast_fp16)[name = tensor("op_1882_cast_fp16")]; + tensor var_1883_perm_0 = const()[name = tensor("op_1883_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1890 = const()[name = tensor("op_1890"), val = tensor([1, 77, 16, 64])]; + tensor var_1891_cast_fp16 = reshape(shape = var_1890, x = tensor_125_cast_fp16)[name = tensor("op_1891_cast_fp16")]; + tensor var_1892_perm_0 = const()[name = tensor("op_1892_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1894 = const()[name = tensor("op_1894"), val = tensor([16, -1, 64])]; + tensor transpose_13 = transpose(perm = var_1892_perm_0, x = var_1891_cast_fp16)[name = tensor("transpose_13")]; + tensor query_states_41_cast_fp16 = reshape(shape = var_1894, x = transpose_13)[name = tensor("query_states_41_cast_fp16")]; + tensor var_1896 = const()[name = tensor("op_1896"), val = tensor([16, -1, 64])]; + tensor transpose_15 = transpose(perm = var_1876_perm_0, x = var_1875_cast_fp16)[name = tensor("transpose_15")]; + tensor key_states_83_cast_fp16 = reshape(shape = var_1896, x = transpose_15)[name = tensor("key_states_83_cast_fp16")]; + tensor var_1898 = const()[name = tensor("op_1898"), val = tensor([16, -1, 64])]; + tensor transpose_14 = transpose(perm = var_1883_perm_0, x = var_1882_cast_fp16)[name = tensor("transpose_14")]; + tensor value_states_83_cast_fp16 = reshape(shape = var_1898, x = transpose_14)[name = tensor("value_states_83_cast_fp16")]; + tensor var_1901_perm_0 = const()[name = tensor("op_1901_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_121_transpose_x_0 = const()[name = tensor("attn_weights_121_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_121_transpose_y_0 = const()[name = tensor("attn_weights_121_transpose_y_0"), val = tensor(false)]; + tensor transpose_12 = transpose(perm = var_1901_perm_0, x = key_states_83_cast_fp16)[name = tensor("transpose_12")]; + tensor attn_weights_121_cast_fp16 = matmul(transpose_x = attn_weights_121_transpose_x_0, transpose_y = attn_weights_121_transpose_y_0, x = query_states_41_cast_fp16, y = transpose_12)[name = tensor("attn_weights_121_cast_fp16")]; + tensor var_1903 = const()[name = tensor("op_1903"), val = tensor([1, 16, 77, 77])]; + tensor var_1904_cast_fp16 = reshape(shape = var_1903, x = attn_weights_121_cast_fp16)[name = tensor("op_1904_cast_fp16")]; + tensor attn_weights_123_cast_fp16 = add(x = var_1904_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_123_cast_fp16")]; + tensor var_1909 = const()[name = tensor("op_1909"), val = tensor([16, 77, 77])]; + tensor input_325_cast_fp16 = reshape(shape = var_1909, x = attn_weights_123_cast_fp16)[name = tensor("input_325_cast_fp16")]; + tensor input_327_cast_fp16 = softmax(axis = var_5, x = input_325_cast_fp16)[name = tensor("input_327_cast_fp16")]; + tensor attn_output_121_transpose_x_0 = const()[name = tensor("attn_output_121_transpose_x_0"), val = tensor(false)]; + tensor attn_output_121_transpose_y_0 = const()[name = tensor("attn_output_121_transpose_y_0"), val = tensor(false)]; + tensor attn_output_121_cast_fp16 = matmul(transpose_x = attn_output_121_transpose_x_0, transpose_y = attn_output_121_transpose_y_0, x = input_327_cast_fp16, y = value_states_83_cast_fp16)[name = tensor("attn_output_121_cast_fp16")]; + tensor var_1914 = const()[name = tensor("op_1914"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_123_cast_fp16 = reshape(shape = var_1914, x = attn_output_121_cast_fp16)[name = tensor("attn_output_123_cast_fp16")]; + tensor attn_output_125_perm_0 = const()[name = tensor("attn_output_125_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1917 = const()[name = tensor("op_1917"), val = tensor([1, 77, 1024])]; + tensor transpose_11 = transpose(perm = attn_output_125_perm_0, x = attn_output_123_cast_fp16)[name = tensor("transpose_11")]; + tensor input_329_cast_fp16 = reshape(shape = var_1917, x = transpose_11)[name = tensor("input_329_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(611529088)))]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(613626304)))]; + tensor linear_123_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_out_proj_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("linear_123_cast_fp16")]; + tensor input_331_cast_fp16 = add(x = input_323_cast_fp16, y = linear_123_cast_fp16)[name = tensor("input_331_cast_fp16")]; + tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(613628416)))]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(613630528)))]; + tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = text_encoder_text_model_encoder_layers_20_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_20_layer_norm2_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(613632640)))]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(622021312)))]; + tensor linear_124_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_mlp_fc1_weight_to_fp16, x = input_333_cast_fp16)[name = tensor("linear_124_cast_fp16")]; + tensor input_337_mode_0 = const()[name = tensor("input_337_mode_0"), val = tensor("EXACT")]; + tensor input_337_cast_fp16 = gelu(mode = input_337_mode_0, x = linear_124_cast_fp16)[name = tensor("input_337_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(622029568)))]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(630418240)))]; + tensor linear_125_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_mlp_fc2_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("linear_125_cast_fp16")]; + tensor input_339_cast_fp16 = add(x = input_331_cast_fp16, y = linear_125_cast_fp16)[name = tensor("input_339_cast_fp16")]; + tensor hidden_states_127_axes_0 = const()[name = tensor("hidden_states_127_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(630420352)))]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(630422464)))]; + tensor hidden_states_127_cast_fp16 = layer_norm(axes = hidden_states_127_axes_0, beta = text_encoder_text_model_encoder_layers_21_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_21_layer_norm1_weight_to_fp16, x = input_339_cast_fp16)[name = tensor("hidden_states_127_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(630424576)))]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(632521792)))]; + tensor linear_126_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_q_proj_weight_to_fp16, x = hidden_states_127_cast_fp16)[name = tensor("linear_126_cast_fp16")]; + tensor var_1956_to_fp16 = const()[name = tensor("op_1956_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_131_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_1956_to_fp16)[name = tensor("tensor_131_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(632523904)))]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(634621120)))]; + tensor linear_127_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_k_proj_weight_to_fp16, x = hidden_states_127_cast_fp16)[name = tensor("linear_127_cast_fp16")]; + tensor var_1961 = const()[name = tensor("op_1961"), val = tensor([1, -1, 16, 64])]; + tensor var_1962_cast_fp16 = reshape(shape = var_1961, x = linear_127_cast_fp16)[name = tensor("op_1962_cast_fp16")]; + tensor var_1963_perm_0 = const()[name = tensor("op_1963_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(634623232)))]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(636720448)))]; + tensor linear_128_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_v_proj_weight_to_fp16, x = hidden_states_127_cast_fp16)[name = tensor("linear_128_cast_fp16")]; + tensor var_1968 = const()[name = tensor("op_1968"), val = tensor([1, -1, 16, 64])]; + tensor var_1969_cast_fp16 = reshape(shape = var_1968, x = linear_128_cast_fp16)[name = tensor("op_1969_cast_fp16")]; + tensor var_1970_perm_0 = const()[name = tensor("op_1970_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1977 = const()[name = tensor("op_1977"), val = tensor([1, 77, 16, 64])]; + tensor var_1978_cast_fp16 = reshape(shape = var_1977, x = tensor_131_cast_fp16)[name = tensor("op_1978_cast_fp16")]; + tensor var_1979_perm_0 = const()[name = tensor("op_1979_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1981 = const()[name = tensor("op_1981"), val = tensor([16, -1, 64])]; + tensor transpose_8 = transpose(perm = var_1979_perm_0, x = var_1978_cast_fp16)[name = tensor("transpose_8")]; + tensor query_states_43_cast_fp16 = reshape(shape = var_1981, x = transpose_8)[name = tensor("query_states_43_cast_fp16")]; + tensor var_1983 = const()[name = tensor("op_1983"), val = tensor([16, -1, 64])]; + tensor transpose_10 = transpose(perm = var_1963_perm_0, x = var_1962_cast_fp16)[name = tensor("transpose_10")]; + tensor key_states_87_cast_fp16 = reshape(shape = var_1983, x = transpose_10)[name = tensor("key_states_87_cast_fp16")]; + tensor var_1985 = const()[name = tensor("op_1985"), val = tensor([16, -1, 64])]; + tensor transpose_9 = transpose(perm = var_1970_perm_0, x = var_1969_cast_fp16)[name = tensor("transpose_9")]; + tensor value_states_87_cast_fp16 = reshape(shape = var_1985, x = transpose_9)[name = tensor("value_states_87_cast_fp16")]; + tensor var_1988_perm_0 = const()[name = tensor("op_1988_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_127_transpose_x_0 = const()[name = tensor("attn_weights_127_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_127_transpose_y_0 = const()[name = tensor("attn_weights_127_transpose_y_0"), val = tensor(false)]; + tensor transpose_7 = transpose(perm = var_1988_perm_0, x = key_states_87_cast_fp16)[name = tensor("transpose_7")]; + tensor attn_weights_127_cast_fp16 = matmul(transpose_x = attn_weights_127_transpose_x_0, transpose_y = attn_weights_127_transpose_y_0, x = query_states_43_cast_fp16, y = transpose_7)[name = tensor("attn_weights_127_cast_fp16")]; + tensor var_1990 = const()[name = tensor("op_1990"), val = tensor([1, 16, 77, 77])]; + tensor var_1991_cast_fp16 = reshape(shape = var_1990, x = attn_weights_127_cast_fp16)[name = tensor("op_1991_cast_fp16")]; + tensor attn_weights_129_cast_fp16 = add(x = var_1991_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_129_cast_fp16")]; + tensor var_1996 = const()[name = tensor("op_1996"), val = tensor([16, 77, 77])]; + tensor input_341_cast_fp16 = reshape(shape = var_1996, x = attn_weights_129_cast_fp16)[name = tensor("input_341_cast_fp16")]; + tensor input_343_cast_fp16 = softmax(axis = var_5, x = input_341_cast_fp16)[name = tensor("input_343_cast_fp16")]; + tensor attn_output_127_transpose_x_0 = const()[name = tensor("attn_output_127_transpose_x_0"), val = tensor(false)]; + tensor attn_output_127_transpose_y_0 = const()[name = tensor("attn_output_127_transpose_y_0"), val = tensor(false)]; + tensor attn_output_127_cast_fp16 = matmul(transpose_x = attn_output_127_transpose_x_0, transpose_y = attn_output_127_transpose_y_0, x = input_343_cast_fp16, y = value_states_87_cast_fp16)[name = tensor("attn_output_127_cast_fp16")]; + tensor var_2001 = const()[name = tensor("op_2001"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_129_cast_fp16 = reshape(shape = var_2001, x = attn_output_127_cast_fp16)[name = tensor("attn_output_129_cast_fp16")]; + tensor attn_output_131_perm_0 = const()[name = tensor("attn_output_131_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2004 = const()[name = tensor("op_2004"), val = tensor([1, 77, 1024])]; + tensor transpose_6 = transpose(perm = attn_output_131_perm_0, x = attn_output_129_cast_fp16)[name = tensor("transpose_6")]; + tensor input_345_cast_fp16 = reshape(shape = var_2004, x = transpose_6)[name = tensor("input_345_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(636722560)))]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(638819776)))]; + tensor linear_129_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_out_proj_weight_to_fp16, x = input_345_cast_fp16)[name = tensor("linear_129_cast_fp16")]; + tensor input_347_cast_fp16 = add(x = input_339_cast_fp16, y = linear_129_cast_fp16)[name = tensor("input_347_cast_fp16")]; + tensor input_349_axes_0 = const()[name = tensor("input_349_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(638821888)))]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(638824000)))]; + tensor input_349_cast_fp16 = layer_norm(axes = input_349_axes_0, beta = text_encoder_text_model_encoder_layers_21_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_21_layer_norm2_weight_to_fp16, x = input_347_cast_fp16)[name = tensor("input_349_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(638826112)))]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(647214784)))]; + tensor linear_130_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_mlp_fc1_weight_to_fp16, x = input_349_cast_fp16)[name = tensor("linear_130_cast_fp16")]; + tensor input_353_mode_0 = const()[name = tensor("input_353_mode_0"), val = tensor("EXACT")]; + tensor input_353_cast_fp16 = gelu(mode = input_353_mode_0, x = linear_130_cast_fp16)[name = tensor("input_353_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(647223040)))]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655611712)))]; + tensor linear_131_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_mlp_fc2_weight_to_fp16, x = input_353_cast_fp16)[name = tensor("linear_131_cast_fp16")]; + tensor input_355_cast_fp16 = add(x = input_347_cast_fp16, y = linear_131_cast_fp16)[name = tensor("input_355_cast_fp16")]; + tensor hidden_states_133_axes_0 = const()[name = tensor("hidden_states_133_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655613824)))]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655615936)))]; + tensor hidden_states_133_cast_fp16 = layer_norm(axes = hidden_states_133_axes_0, beta = text_encoder_text_model_encoder_layers_22_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_22_layer_norm1_weight_to_fp16, x = input_355_cast_fp16)[name = tensor("hidden_states_133_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655618048)))]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(657715264)))]; + tensor linear_132_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_q_proj_weight_to_fp16, x = hidden_states_133_cast_fp16)[name = tensor("linear_132_cast_fp16")]; + tensor var_2043_to_fp16 = const()[name = tensor("op_2043_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_cast_fp16 = mul(x = linear_132_cast_fp16, y = var_2043_to_fp16)[name = tensor("tensor_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(657717376)))]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(659814592)))]; + tensor linear_133_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_k_proj_weight_to_fp16, x = hidden_states_133_cast_fp16)[name = tensor("linear_133_cast_fp16")]; + tensor var_2048 = const()[name = tensor("op_2048"), val = tensor([1, -1, 16, 64])]; + tensor var_2049_cast_fp16 = reshape(shape = var_2048, x = linear_133_cast_fp16)[name = tensor("op_2049_cast_fp16")]; + tensor var_2050_perm_0 = const()[name = tensor("op_2050_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(659816704)))]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(661913920)))]; + tensor linear_134_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_v_proj_weight_to_fp16, x = hidden_states_133_cast_fp16)[name = tensor("linear_134_cast_fp16")]; + tensor var_2055 = const()[name = tensor("op_2055"), val = tensor([1, -1, 16, 64])]; + tensor var_2056_cast_fp16 = reshape(shape = var_2055, x = linear_134_cast_fp16)[name = tensor("op_2056_cast_fp16")]; + tensor var_2057_perm_0 = const()[name = tensor("op_2057_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2064 = const()[name = tensor("op_2064"), val = tensor([1, 77, 16, 64])]; + tensor var_2065_cast_fp16 = reshape(shape = var_2064, x = tensor_cast_fp16)[name = tensor("op_2065_cast_fp16")]; + tensor var_2066_perm_0 = const()[name = tensor("op_2066_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2068 = const()[name = tensor("op_2068"), val = tensor([16, -1, 64])]; + tensor transpose_3 = transpose(perm = var_2066_perm_0, x = var_2065_cast_fp16)[name = tensor("transpose_3")]; + tensor query_states_cast_fp16 = reshape(shape = var_2068, x = transpose_3)[name = tensor("query_states_cast_fp16")]; + tensor var_2070 = const()[name = tensor("op_2070"), val = tensor([16, -1, 64])]; + tensor transpose_5 = transpose(perm = var_2050_perm_0, x = var_2049_cast_fp16)[name = tensor("transpose_5")]; + tensor key_states_cast_fp16 = reshape(shape = var_2070, x = transpose_5)[name = tensor("key_states_cast_fp16")]; + tensor var_2072 = const()[name = tensor("op_2072"), val = tensor([16, -1, 64])]; + tensor transpose_4 = transpose(perm = var_2057_perm_0, x = var_2056_cast_fp16)[name = tensor("transpose_4")]; + tensor value_states_cast_fp16 = reshape(shape = var_2072, x = transpose_4)[name = tensor("value_states_cast_fp16")]; + tensor var_2075_perm_0 = const()[name = tensor("op_2075_perm_0"), val = tensor([0, 2, 1])]; + tensor attn_weights_133_transpose_x_0 = const()[name = tensor("attn_weights_133_transpose_x_0"), val = tensor(false)]; + tensor attn_weights_133_transpose_y_0 = const()[name = tensor("attn_weights_133_transpose_y_0"), val = tensor(false)]; + tensor transpose_2 = transpose(perm = var_2075_perm_0, x = key_states_cast_fp16)[name = tensor("transpose_2")]; + tensor attn_weights_133_cast_fp16 = matmul(transpose_x = attn_weights_133_transpose_x_0, transpose_y = attn_weights_133_transpose_y_0, x = query_states_cast_fp16, y = transpose_2)[name = tensor("attn_weights_133_cast_fp16")]; + tensor var_2077 = const()[name = tensor("op_2077"), val = tensor([1, 16, 77, 77])]; + tensor var_2078_cast_fp16 = reshape(shape = var_2077, x = attn_weights_133_cast_fp16)[name = tensor("op_2078_cast_fp16")]; + tensor attn_weights_135_cast_fp16 = add(x = var_2078_cast_fp16, y = var_57_to_fp16)[name = tensor("attn_weights_135_cast_fp16")]; + tensor var_2083 = const()[name = tensor("op_2083"), val = tensor([16, 77, 77])]; + tensor input_357_cast_fp16 = reshape(shape = var_2083, x = attn_weights_135_cast_fp16)[name = tensor("input_357_cast_fp16")]; + tensor input_359_cast_fp16 = softmax(axis = var_5, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; + tensor attn_output_133_transpose_x_0 = const()[name = tensor("attn_output_133_transpose_x_0"), val = tensor(false)]; + tensor attn_output_133_transpose_y_0 = const()[name = tensor("attn_output_133_transpose_y_0"), val = tensor(false)]; + tensor attn_output_133_cast_fp16 = matmul(transpose_x = attn_output_133_transpose_x_0, transpose_y = attn_output_133_transpose_y_0, x = input_359_cast_fp16, y = value_states_cast_fp16)[name = tensor("attn_output_133_cast_fp16")]; + tensor var_2088 = const()[name = tensor("op_2088"), val = tensor([1, 16, 77, 64])]; + tensor attn_output_135_cast_fp16 = reshape(shape = var_2088, x = attn_output_133_cast_fp16)[name = tensor("attn_output_135_cast_fp16")]; + tensor attn_output_perm_0 = const()[name = tensor("attn_output_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2091 = const()[name = tensor("op_2091"), val = tensor([1, 77, 1024])]; + tensor transpose_1 = transpose(perm = attn_output_perm_0, x = attn_output_135_cast_fp16)[name = tensor("transpose_1")]; + tensor input_361_cast_fp16 = reshape(shape = var_2091, x = transpose_1)[name = tensor("input_361_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(661916032)))]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(664013248)))]; + tensor linear_135_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_out_proj_weight_to_fp16, x = input_361_cast_fp16)[name = tensor("linear_135_cast_fp16")]; + tensor input_363_cast_fp16 = add(x = input_355_cast_fp16, y = linear_135_cast_fp16)[name = tensor("input_363_cast_fp16")]; + tensor input_365_axes_0 = const()[name = tensor("input_365_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(664015360)))]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(664017472)))]; + tensor input_365_cast_fp16 = layer_norm(axes = input_365_axes_0, beta = text_encoder_text_model_encoder_layers_22_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_22_layer_norm2_weight_to_fp16, x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(664019584)))]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(672408256)))]; + tensor linear_136_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_mlp_fc1_weight_to_fp16, x = input_365_cast_fp16)[name = tensor("linear_136_cast_fp16")]; + tensor input_369_mode_0 = const()[name = tensor("input_369_mode_0"), val = tensor("EXACT")]; + tensor input_369_cast_fp16 = gelu(mode = input_369_mode_0, x = linear_136_cast_fp16)[name = tensor("input_369_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(672416512)))]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(680805184)))]; + tensor linear_137_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_mlp_fc2_weight_to_fp16, x = input_369_cast_fp16)[name = tensor("linear_137_cast_fp16")]; + tensor input_cast_fp16 = add(x = input_363_cast_fp16, y = linear_137_cast_fp16)[name = tensor("input_cast_fp16")]; + tensor last_hidden_state_axes_0 = const()[name = tensor("last_hidden_state_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_final_layer_norm_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_final_layer_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(680807296)))]; + tensor text_encoder_text_model_final_layer_norm_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_final_layer_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(680809408)))]; + tensor last_hidden_state_cast_fp16 = layer_norm(axes = last_hidden_state_axes_0, beta = text_encoder_text_model_final_layer_norm_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_final_layer_norm_weight_to_fp16, x = input_cast_fp16)[name = tensor("last_hidden_state_cast_fp16")]; + tensor last_hidden_state_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("last_hidden_state_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor var_2116 = const()[name = tensor("op_2116"), val = tensor([0])]; + tensor var_2118 = reduce_argmax(axis = var_5, keep_dims = var_6, x = cast_239)[name = tensor("op_2118")]; + tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(1)]; + tensor stack_0 = stack(axis = stack_0_axis_0, values = (var_2116, var_2118))[name = tensor("stack_0")]; + tensor var_2120_transpose_batch_dims_0 = const()[name = tensor("op_2120_transpose_batch_dims_0"), val = tensor(0)]; + tensor var_2120_transpose_cast_fp16 = gather_nd(batch_dims = var_2120_transpose_batch_dims_0, indices = stack_0, x = last_hidden_state_cast_fp16)[name = tensor("op_2120_transpose_cast_fp16")]; + tensor var_2120_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_2120_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor pooled_outputs = cast(dtype = var_2120_cast_fp16_to_fp32_dtype_0, x = var_2120_transpose_cast_fp16)[name = tensor("cast_237")]; + tensor last_hidden_state = cast(dtype = last_hidden_state_cast_fp16_to_fp32_dtype_0, x = last_hidden_state_cast_fp16)[name = tensor("cast_238")]; + } -> (last_hidden_state, pooled_outputs); +} \ No newline at end of file