SUimeModelTraner/export.record

113 lines
78 KiB
Plaintext
Raw Permalink Blame History

This file contains invisible Unicode characters

This file contains invisible Unicode characters that are indistinguishable to humans but may be processed differently by a computer. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

📁 输出目录: /home/songsenand/Project/SUimeModelTraner/exported_models
📦 加载checkpoint: /home/songsenand/下载/20260416best_model.pt
Downloading Model from https://www.modelscope.cn to directory: /home/songsenand/.cache/modelscope/hub/models/iic/nlp_structbert_backbone_lite_std
Loading weights: 0%| | 0/103 [00:00<?, ?it/s]
Loading weights: 1%|█▏ | 1/103 [00:00<00:00, 29746.84it/s, Materializing param=embeddings.LayerNorm.bias]
Loading weights: 1%|█▏ | 1/103 [00:00<00:00, 9177.91it/s, Materializing param=embeddings.LayerNorm.bias]
Loading weights: 2%|██▎ | 2/103 [00:00<00:00, 9915.61it/s, Materializing param=embeddings.LayerNorm.weight]
Loading weights: 2%|██▎ | 2/103 [00:00<00:00, 7854.50it/s, Materializing param=embeddings.LayerNorm.weight]
Loading weights: 3%|███ | 3/103 [00:00<00:00, 9225.01it/s, Materializing param=embeddings.position_embeddings.weight]
Loading weights: 3%|███ | 3/103 [00:00<00:00, 7953.80it/s, Materializing param=embeddings.position_embeddings.weight]
Loading weights: 4%|████ | 4/103 [00:00<00:00, 8986.19it/s, Materializing param=embeddings.token_type_embeddings.weight]
Loading weights: 4%|████ | 4/103 [00:00<00:00, 8093.21it/s, Materializing param=embeddings.token_type_embeddings.weight]
Loading weights: 5%|█████▎ | 5/103 [00:00<00:00, 9023.89it/s, Materializing param=embeddings.word_embeddings.weight]
Loading weights: 5%|█████▎ | 5/103 [00:00<00:00, 8365.19it/s, Materializing param=embeddings.word_embeddings.weight]
Loading weights: 6%|█████▌ | 6/103 [00:00<00:00, 9174.56it/s, Materializing param=encoder.layer.0.attention.output.LayerNorm.bias]
Loading weights: 6%|█████▌ | 6/103 [00:00<00:00, 8589.02it/s, Materializing param=encoder.layer.0.attention.output.LayerNorm.bias]
Loading weights: 7%|██████▍ | 7/103 [00:00<00:00, 9229.84it/s, Materializing param=encoder.layer.0.attention.output.LayerNorm.weight]
Loading weights: 7%|██████▍ | 7/103 [00:00<00:00, 8722.56it/s, Materializing param=encoder.layer.0.attention.output.LayerNorm.weight]
Loading weights: 8%|███████▊ | 8/103 [00:00<00:00, 9302.59it/s, Materializing param=encoder.layer.0.attention.output.dense.bias]
Loading weights: 8%|███████▊ | 8/103 [00:00<00:00, 8862.77it/s, Materializing param=encoder.layer.0.attention.output.dense.bias]
Loading weights: 9%|████████▌ | 9/103 [00:00<00:00, 9327.58it/s, Materializing param=encoder.layer.0.attention.output.dense.weight]
Loading weights: 9%|████████▌ | 9/103 [00:00<00:00, 8921.94it/s, Materializing param=encoder.layer.0.attention.output.dense.weight]
Loading weights: 10%|██████████ | 10/103 [00:00<00:00, 9402.16it/s, Materializing param=encoder.layer.0.attention.self.key.bias]
Loading weights: 10%|██████████ | 10/103 [00:00<00:00, 9039.45it/s, Materializing param=encoder.layer.0.attention.self.key.bias]
Loading weights: 11%|██████████▊ | 11/103 [00:00<00:00, 9475.73it/s, Materializing param=encoder.layer.0.attention.self.key.weight]
Loading weights: 11%|██████████▊ | 11/103 [00:00<00:00, 9123.46it/s, Materializing param=encoder.layer.0.attention.self.key.weight]
Loading weights: 12%|███████████▊ | 12/103 [00:00<00:00, 9510.89it/s, Materializing param=encoder.layer.0.attention.self.query.bias]
Loading weights: 12%|███████████▊ | 12/103 [00:00<00:00, 9206.45it/s, Materializing param=encoder.layer.0.attention.self.query.bias]
Loading weights: 13%|████████████▍ | 13/103 [00:00<00:00, 9567.64it/s, Materializing param=encoder.layer.0.attention.self.query.weight]
Loading weights: 13%|████████████▍ | 13/103 [00:00<00:00, 9277.85it/s, Materializing param=encoder.layer.0.attention.self.query.weight]
Loading weights: 14%|█████████████▋ | 14/103 [00:00<00:00, 9548.01it/s, Materializing param=encoder.layer.0.attention.self.value.bias]
Loading weights: 14%|█████████████▋ | 14/103 [00:00<00:00, 9269.18it/s, Materializing param=encoder.layer.0.attention.self.value.bias]
Loading weights: 15%|██████████████▍ | 15/103 [00:00<00:00, 9564.39it/s, Materializing param=encoder.layer.0.attention.self.value.weight]
Loading weights: 15%|██████████████▍ | 15/103 [00:00<00:00, 9313.78it/s, Materializing param=encoder.layer.0.attention.self.value.weight]
Loading weights: 16%|███████████████▉ | 16/103 [00:00<00:00, 9617.21it/s, Materializing param=encoder.layer.0.intermediate.dense.bias]
Loading weights: 16%|███████████████▉ | 16/103 [00:00<00:00, 9383.23it/s, Materializing param=encoder.layer.0.intermediate.dense.bias]
Loading weights: 17%|████████████████▋ | 17/103 [00:00<00:00, 9669.54it/s, Materializing param=encoder.layer.0.intermediate.dense.weight]
Loading weights: 17%|████████████████▋ | 17/103 [00:00<00:00, 9434.13it/s, Materializing param=encoder.layer.0.intermediate.dense.weight]
Loading weights: 17%|██████████████████▎ | 18/103 [00:00<00:00, 9705.29it/s, Materializing param=encoder.layer.0.output.LayerNorm.bias]
Loading weights: 17%|██████████████████▎ | 18/103 [00:00<00:00, 9491.76it/s, Materializing param=encoder.layer.0.output.LayerNorm.bias]
Loading weights: 18%|███████████████████ | 19/103 [00:00<00:00, 9753.00it/s, Materializing param=encoder.layer.0.output.LayerNorm.weight]
Loading weights: 18%|███████████████████ | 19/103 [00:00<00:00, 9554.22it/s, Materializing param=encoder.layer.0.output.LayerNorm.weight]
Loading weights: 19%|█████████████████████▏ | 20/103 [00:00<00:00, 9803.21it/s, Materializing param=encoder.layer.0.output.dense.bias]
Loading weights: 19%|█████████████████████▏ | 20/103 [00:00<00:00, 9614.45it/s, Materializing param=encoder.layer.0.output.dense.bias]
Loading weights: 20%|█████████████████████▊ | 21/103 [00:00<00:00, 9852.39it/s, Materializing param=encoder.layer.0.output.dense.weight]
Loading weights: 20%|█████████████████████▊ | 21/103 [00:00<00:00, 9669.60it/s, Materializing param=encoder.layer.0.output.dense.weight]
Loading weights: 21%|████████████████████▎ | 22/103 [00:00<00:00, 9896.47it/s, Materializing param=encoder.layer.1.attention.output.LayerNorm.bias]
Loading weights: 21%|████████████████████▎ | 22/103 [00:00<00:00, 9716.19it/s, Materializing param=encoder.layer.1.attention.output.LayerNorm.bias]
Loading weights: 22%|████████████████████▊ | 23/103 [00:00<00:00, 9927.86it/s, Materializing param=encoder.layer.1.attention.output.LayerNorm.weight]
Loading weights: 22%|████████████████████▊ | 23/103 [00:00<00:00, 9756.17it/s, Materializing param=encoder.layer.1.attention.output.LayerNorm.weight]
Loading weights: 23%|███████████████████████ | 24/103 [00:00<00:00, 9956.80it/s, Materializing param=encoder.layer.1.attention.output.dense.bias]
Loading weights: 23%|███████████████████████ | 24/103 [00:00<00:00, 9762.71it/s, Materializing param=encoder.layer.1.attention.output.dense.bias]
Loading weights: 24%|███████████████████████▌ | 25/103 [00:00<00:00, 9932.52it/s, Materializing param=encoder.layer.1.attention.output.dense.weight]
Loading weights: 24%|███████████████████████▌ | 25/103 [00:00<00:00, 9767.82it/s, Materializing param=encoder.layer.1.attention.output.dense.weight]
Loading weights: 25%|██████████████████████████ | 26/103 [00:00<00:00, 9941.83it/s, Materializing param=encoder.layer.1.attention.self.key.bias]
Loading weights: 25%|██████████████████████████ | 26/103 [00:00<00:00, 9781.32it/s, Materializing param=encoder.layer.1.attention.self.key.bias]
Loading weights: 26%|██████████████████████████▍ | 27/103 [00:00<00:00, 9951.34it/s, Materializing param=encoder.layer.1.attention.self.key.weight]
Loading weights: 26%|██████████████████████████▍ | 27/103 [00:00<00:00, 9800.62it/s, Materializing param=encoder.layer.1.attention.self.key.weight]
Loading weights: 27%|███████████████████████████▍ | 28/103 [00:00<00:00, 9968.64it/s, Materializing param=encoder.layer.1.attention.self.query.bias]
Loading weights: 27%|███████████████████████████▍ | 28/103 [00:00<00:00, 9825.19it/s, Materializing param=encoder.layer.1.attention.self.query.bias]
Loading weights: 28%|███████████████████████████▊ | 29/103 [00:00<00:00, 9989.72it/s, Materializing param=encoder.layer.1.attention.self.query.weight]
Loading weights: 28%|███████████████████████████▊ | 29/103 [00:00<00:00, 9846.58it/s, Materializing param=encoder.layer.1.attention.self.query.weight]
Loading weights: 29%|█████████████████████████████▏ | 30/103 [00:00<00:00, 10007.09it/s, Materializing param=encoder.layer.1.attention.self.value.bias]
Loading weights: 29%|█████████████████████████████▍ | 30/103 [00:00<00:00, 9873.60it/s, Materializing param=encoder.layer.1.attention.self.value.bias]
Loading weights: 30%|█████████████████████████████▍ | 31/103 [00:00<00:00, 10031.12it/s, Materializing param=encoder.layer.1.attention.self.value.weight]
Loading weights: 30%|█████████████████████████████▊ | 31/103 [00:00<00:00, 9900.51it/s, Materializing param=encoder.layer.1.attention.self.value.weight]
Loading weights: 31%|███████████████████████████████▋ | 32/103 [00:00<00:00, 10054.52it/s, Materializing param=encoder.layer.1.intermediate.dense.bias]
Loading weights: 31%|███████████████████████████████▉ | 32/103 [00:00<00:00, 9921.48it/s, Materializing param=encoder.layer.1.intermediate.dense.bias]
Loading weights: 32%|████████████████████████████████ | 33/103 [00:00<00:00, 10067.79it/s, Materializing param=encoder.layer.1.intermediate.dense.weight]
Loading weights: 32%|████████████████████████████████▎ | 33/103 [00:00<00:00, 9942.68it/s, Materializing param=encoder.layer.1.intermediate.dense.weight]
Loading weights: 33%|██████████████████████████████████▎ | 34/103 [00:00<00:00, 10086.74it/s, Materializing param=encoder.layer.1.output.LayerNorm.bias]
Loading weights: 33%|██████████████████████████████████▋ | 34/103 [00:00<00:00, 9967.59it/s, Materializing param=encoder.layer.1.output.LayerNorm.bias]
Loading weights: 34%|██████████████████████████████████▋ | 35/103 [00:00<00:00, 10075.54it/s, Materializing param=encoder.layer.1.output.LayerNorm.weight]
Loading weights: 34%|███████████████████████████████████ | 35/103 [00:00<00:00, 9954.61it/s, Materializing param=encoder.layer.1.output.LayerNorm.weight]
Loading weights: 35%|█████████████████████████████████████▋ | 36/103 [00:00<00:00, 10089.87it/s, Materializing param=encoder.layer.1.output.dense.bias]
Loading weights: 35%|██████████████████████████████████████ | 36/103 [00:00<00:00, 9978.52it/s, Materializing param=encoder.layer.1.output.dense.bias]
Loading weights: 36%|██████████████████████████████████████ | 37/103 [00:00<00:00, 10110.71it/s, Materializing param=encoder.layer.1.output.dense.weight]
Loading weights: 36%|██████████████████████████████████████ | 37/103 [00:00<00:00, 10000.60it/s, Materializing param=encoder.layer.1.output.dense.weight]
Loading weights: 37%|██████████████████████████████████▋ | 38/103 [00:00<00:00, 10131.81it/s, Materializing param=encoder.layer.2.attention.output.LayerNorm.bias]
Loading weights: 37%|██████████████████████████████████▋ | 38/103 [00:00<00:00, 10020.97it/s, Materializing param=encoder.layer.2.attention.output.LayerNorm.bias]
Loading weights: 38%|██████████████████████████████████▊ | 39/103 [00:00<00:00, 10141.85it/s, Materializing param=encoder.layer.2.attention.output.LayerNorm.weight]
Loading weights: 38%|██████████████████████████████████▊ | 39/103 [00:00<00:00, 10027.45it/s, Materializing param=encoder.layer.2.attention.output.LayerNorm.weight]
Loading weights: 39%|██████████████████████████████████████ | 40/103 [00:00<00:00, 10130.56it/s, Materializing param=encoder.layer.2.attention.output.dense.bias]
Loading weights: 39%|██████████████████████████████████████ | 40/103 [00:00<00:00, 10022.83it/s, Materializing param=encoder.layer.2.attention.output.dense.bias]
Loading weights: 40%|██████████████████████████████████████▏ | 41/103 [00:00<00:00, 10131.17it/s, Materializing param=encoder.layer.2.attention.output.dense.weight]
Loading weights: 40%|██████████████████████████████████████▏ | 41/103 [00:00<00:00, 10025.45it/s, Materializing param=encoder.layer.2.attention.output.dense.weight]
Loading weights: 41%|█████████████████████████████████████████▌ | 42/103 [00:00<00:00, 10135.83it/s, Materializing param=encoder.layer.2.attention.self.key.bias]
Loading weights: 41%|█████████████████████████████████████████▌ | 42/103 [00:00<00:00, 10033.65it/s, Materializing param=encoder.layer.2.attention.self.key.bias]
Loading weights: 42%|█████████████████████████████████████████▋ | 43/103 [00:00<00:00, 10143.70it/s, Materializing param=encoder.layer.2.attention.self.key.weight]
Loading weights: 42%|█████████████████████████████████████████▋ | 43/103 [00:00<00:00, 10045.40it/s, Materializing param=encoder.layer.2.attention.self.key.weight]
Loading weights: 43%|██████████████████████████████████████████▋ | 44/103 [00:00<00:00, 10155.14it/s, Materializing param=encoder.layer.2.attention.self.query.bias]
Loading weights: 43%|██████████████████████████████████████████▋ | 44/103 [00:00<00:00, 10058.28it/s, Materializing param=encoder.layer.2.attention.self.query.bias]
Loading weights: 44%|██████████████████████████████████████████▊ | 45/103 [00:00<00:00, 10164.45it/s, Materializing param=encoder.layer.2.attention.self.query.weight]
Loading weights: 44%|██████████████████████████████████████████▊ | 45/103 [00:00<00:00, 10052.93it/s, Materializing param=encoder.layer.2.attention.self.query.weight]
Loading weights: 45%|████████████████████████████████████████████▋ | 46/103 [00:00<00:00, 10143.42it/s, Materializing param=encoder.layer.2.attention.self.value.bias]
Loading weights: 45%|████████████████████████████████████████████▋ | 46/103 [00:00<00:00, 10049.90it/s, Materializing param=encoder.layer.2.attention.self.value.bias]
Loading weights: 46%|████████████████████████████████████████████▋ | 47/103 [00:00<00:00, 10149.95it/s, Materializing param=encoder.layer.2.attention.self.value.weight]
Loading weights: 46%|████████████████████████████████████████████▋ | 47/103 [00:00<00:00, 10060.34it/s, Materializing param=encoder.layer.2.attention.self.value.weight]
Loading weights: 47%|███████████████████████████████████████████████▌ | 48/103 [00:00<00:00, 10160.83it/s, Materializing param=encoder.layer.2.intermediate.dense.bias]
Loading weights: 47%|███████████████████████████████████████████████▌ | 48/103 [00:00<00:00, 10074.89it/s, Materializing param=encoder.layer.2.intermediate.dense.bias]
Loading weights: 48%|███████████████████████████████████████████████▌ | 49/103 [00:00<00:00, 10175.31it/s, Materializing param=encoder.layer.2.intermediate.dense.weight]
Loading weights: 48%|███████████████████████████████████████████████▌ | 49/103 [00:00<00:00, 10089.39it/s, Materializing param=encoder.layer.2.intermediate.dense.weight]
Loading weights: 49%|██████████████████████████████████████████████████▍ | 50/103 [00:00<00:00, 10187.77it/s, Materializing param=encoder.layer.2.output.LayerNorm.bias]
Loading weights: 49%|██████████████████████████████████████████████████▍ | 50/103 [00:00<00:00, 10104.81it/s, Materializing param=encoder.layer.2.output.LayerNorm.bias]
Loading weights: 50%|██████████████████████████████████████████████████▌ | 51/103 [00:00<00:00, 10200.74it/s, Materializing param=encoder.layer.2.output.LayerNorm.weight]
Loading weights: 50%|██████████████████████████████████████████████████▌ | 51/103 [00:00<00:00, 10118.23it/s, Materializing param=encoder.layer.2.output.LayerNorm.weight]
Loading weights: 50%|██████████████████████████████████████████████████████▌ | 52/103 [00:00<00:00, 10211.33it/s, Materializing param=encoder.layer.2.output.dense.bias]
Loading weights: 50%|██████████████████████████████████████████████████████▌ | 52/103 [00:00<00:00, 10115.19it/s, Materializing param=encoder.layer.2.output.dense.bias]
Loading weights: 51%|██████████████████████████████████████████████████████▌ | 53/103 [00:00<00:00, 10203.71it/s, Materializing param=encoder.layer.2.output.dense.weight]
Loading weights: 51%|██████████████████████████████████████████████████████▌ | 53/103 [00:00<00:00, 10121.94it/s, Materializing param=encoder.layer.2.output.dense.weight]
Loading weights: 52%|█████████████████████████████████████████████████▎ | 54/103 [00:00<00:00, 10209.26it/s, Materializing param=encoder.layer.3.attention.output.LayerNorm.bias]
Loading weights: 52%|█████████████████████████████████████████████████▎ | 54/103 [00:00<00:00, 10128.45it/s, Materializing param=encoder.layer.3.attention.output.LayerNorm.bias]
Loading weights: 53%|█████████████████████████████████████████████████▏ | 55/103 [00:00<00:00, 10210.99it/s, Materializing param=encoder.layer.3.attention.output.LayerNorm.weight]
Loading weights: 53%|█████████████████████████████████████████████████▏ | 55/103 [00:00<00:00, 10131.17it/s, Materializing param=encoder.layer.3.attention.output.LayerNorm.weight]
Loading weights: 54%|█████████████████████████████████████████████████████▎ | 56/103 [00:00<00:00, 10194.05it/s, Materializing param=encoder.layer.3.attention.output.dense.bias]
Loading weights: 54%|█████████████████████████████████████████████████████▎ | 56/103 [00:00<00:00, 10112.41it/s, Materializing param=encoder.layer.3.attention.output.dense.bias]
Loading weights: 55%|█████████████████████████████████████████████████████▏ | 57/103 [00:00<00:00, 10190.33it/s, Materializing param=encoder.layer.3.attention.output.dense.weight]
Loading weights: 55%|█████████████████████████████████████████████████████▏ | 57/103 [00:00<00:00, 10112.31it/s, Materializing param=encoder.layer.3.attention.output.dense.weight]
Loading weights: 56%|█████████████████████████████████████████████████████████▍ | 58/103 [00:00<00:00, 10191.44it/s, Materializing param=encoder.layer.3.attention.self.key.bias]
Loading weights: 56%|█████████████████████████████████████████████████████████▍ | 58/103 [00:00<00:00, 10116.00it/s, Materializing param=encoder.layer.3.attention.self.key.bias]
Loading weights: 57%|█████████████████████████████████████████████████████████▎ | 59/103 [00:00<00:00, 10188.31it/s, Materializing param=encoder.layer.3.attention.self.key.weight]
Loading weights: 57%|█████████████████████████████████████████████████████████▎ | 59/103 [00:00<00:00, 10113.78it/s, Materializing param=encoder.layer.3.attention.self.key.weight]
Loading weights: 58%|██████████████████████████████████████████████████████████▎ | 60/103 [00:00<00:00, 10190.24it/s, Materializing param=encoder.layer.3.attention.self.query.bias]
Loading weights: 58%|██████████████████████████████████████████████████████████▎ | 60/103 [00:00<00:00, 10117.73it/s, Materializing param=encoder.layer.3.attention.self.query.bias]
Loading weights: 59%|██████████████████████████████████████████████████████████ | 61/103 [00:00<00:00, 10189.67it/s, Materializing param=encoder.layer.3.attention.self.query.weight]
Loading weights: 59%|██████████████████████████████████████████████████████████ | 61/103 [00:00<00:00, 10116.75it/s, Materializing param=encoder.layer.3.attention.self.query.weight]
Loading weights: 60%|████████████████████████████████████████████████████████████▏ | 62/103 [00:00<00:00, 10190.32it/s, Materializing param=encoder.layer.3.attention.self.value.bias]
Loading weights: 60%|████████████████████████████████████████████████████████████▏ | 62/103 [00:00<00:00, 10120.92it/s, Materializing param=encoder.layer.3.attention.self.value.bias]
Loading weights: 61%|███████████████████████████████████████████████████████████▉ | 63/103 [00:00<00:00, 10193.70it/s, Materializing param=encoder.layer.3.attention.self.value.weight]
Loading weights: 61%|███████████████████████████████████████████████████████████▉ | 63/103 [00:00<00:00, 10124.57it/s, Materializing param=encoder.layer.3.attention.self.value.weight]
Loading weights: 62%|███████████████████████████████████████████████████████████████▍ | 64/103 [00:00<00:00, 10197.37it/s, Materializing param=encoder.layer.3.intermediate.dense.bias]
Loading weights: 62%|███████████████████████████████████████████████████████████████▍ | 64/103 [00:00<00:00, 10130.02it/s, Materializing param=encoder.layer.3.intermediate.dense.bias]
Loading weights: 63%|███████████████████████████████████████████████████████████████ | 65/103 [00:00<00:00, 10198.63it/s, Materializing param=encoder.layer.3.intermediate.dense.weight]
Loading weights: 63%|███████████████████████████████████████████████████████████████ | 65/103 [00:00<00:00, 10130.79it/s, Materializing param=encoder.layer.3.intermediate.dense.weight]
Loading weights: 64%|██████████████████████████████████████████████████████████████████▋ | 66/103 [00:00<00:00, 10187.84it/s, Materializing param=encoder.layer.3.output.LayerNorm.bias]
Loading weights: 64%|██████████████████████████████████████████████████████████████████▋ | 66/103 [00:00<00:00, 10120.06it/s, Materializing param=encoder.layer.3.output.LayerNorm.bias]
Loading weights: 65%|██████████████████████████████████████████████████████████████████▎ | 67/103 [00:00<00:00, 10189.21it/s, Materializing param=encoder.layer.3.output.LayerNorm.weight]
Loading weights: 65%|██████████████████████████████████████████████████████████████████▎ | 67/103 [00:00<00:00, 10124.23it/s, Materializing param=encoder.layer.3.output.LayerNorm.weight]
Loading weights: 66%|███████████████████████████████████████████████████████████████████████▎ | 68/103 [00:00<00:00, 10192.36it/s, Materializing param=encoder.layer.3.output.dense.bias]
Loading weights: 66%|███████████████████████████████████████████████████████████████████████▎ | 68/103 [00:00<00:00, 10129.73it/s, Materializing param=encoder.layer.3.output.dense.bias]
Loading weights: 67%|███████████████████████████████████████████████████████████████████████ | 69/103 [00:00<00:00, 10198.65it/s, Materializing param=encoder.layer.3.output.dense.weight]
Loading weights: 67%|███████████████████████████████████████████████████████████████████████ | 69/103 [00:00<00:00, 10136.14it/s, Materializing param=encoder.layer.3.output.dense.weight]
Loading weights: 68%|███████████████████████████████████████████████████████████████▉ | 70/103 [00:00<00:00, 10204.41it/s, Materializing param=encoder.layer.4.attention.output.LayerNorm.bias]
Loading weights: 68%|███████████████████████████████████████████████████████████████▉ | 70/103 [00:00<00:00, 10137.82it/s, Materializing param=encoder.layer.4.attention.output.LayerNorm.bias]
Loading weights: 69%|███████████████████████████████████████████████████████████████▍ | 71/103 [00:00<00:00, 10201.27it/s, Materializing param=encoder.layer.4.attention.output.LayerNorm.weight]
Loading weights: 69%|███████████████████████████████████████████████████████████████▍ | 71/103 [00:00<00:00, 10136.69it/s, Materializing param=encoder.layer.4.attention.output.LayerNorm.weight]
Loading weights: 70%|████████████████████████████████████████████████████████████████████▌ | 72/103 [00:00<00:00, 10197.19it/s, Materializing param=encoder.layer.4.attention.output.dense.bias]
Loading weights: 70%|████████████████████████████████████████████████████████████████████▌ | 72/103 [00:00<00:00, 10136.27it/s, Materializing param=encoder.layer.4.attention.output.dense.bias]
Loading weights: 71%|████████████████████████████████████████████████████████████████████ | 73/103 [00:00<00:00, 10201.04it/s, Materializing param=encoder.layer.4.attention.output.dense.weight]
Loading weights: 71%|████████████████████████████████████████████████████████████████████ | 73/103 [00:00<00:00, 10141.91it/s, Materializing param=encoder.layer.4.attention.output.dense.weight]
Loading weights: 72%|█████████████████████████████████████████████████████████████████████████▎ | 74/103 [00:00<00:00, 10205.12it/s, Materializing param=encoder.layer.4.attention.self.key.bias]
Loading weights: 72%|█████████████████████████████████████████████████████████████████████████▎ | 74/103 [00:00<00:00, 10147.07it/s, Materializing param=encoder.layer.4.attention.self.key.bias]
Loading weights: 73%|████████████████████████████████████████████████████████████████████████▊ | 75/103 [00:00<00:00, 10210.75it/s, Materializing param=encoder.layer.4.attention.self.key.weight]
Loading weights: 73%|████████████████████████████████████████████████████████████████████████▊ | 75/103 [00:00<00:00, 10153.41it/s, Materializing param=encoder.layer.4.attention.self.key.weight]
Loading weights: 74%|█████████████████████████████████████████████████████████████████████████▊ | 76/103 [00:00<00:00, 10207.73it/s, Materializing param=encoder.layer.4.attention.self.query.bias]
Loading weights: 74%|█████████████████████████████████████████████████████████████████████████▊ | 76/103 [00:00<00:00, 10122.80it/s, Materializing param=encoder.layer.4.attention.self.query.bias]
Loading weights: 75%|█████████████████████████████████████████████████████████████████████████▎ | 77/103 [00:00<00:00, 10170.41it/s, Materializing param=encoder.layer.4.attention.self.query.weight]
Loading weights: 75%|█████████████████████████████████████████████████████████████████████████▎ | 77/103 [00:00<00:00, 10113.72it/s, Materializing param=encoder.layer.4.attention.self.query.weight]
Loading weights: 76%|███████████████████████████████████████████████████████████████████████████▋ | 78/103 [00:00<00:00, 10171.80it/s, Materializing param=encoder.layer.4.attention.self.value.bias]
Loading weights: 76%|███████████████████████████████████████████████████████████████████████████▋ | 78/103 [00:00<00:00, 10117.07it/s, Materializing param=encoder.layer.4.attention.self.value.bias]
Loading weights: 77%|███████████████████████████████████████████████████████████████████████████▏ | 79/103 [00:00<00:00, 10176.60it/s, Materializing param=encoder.layer.4.attention.self.value.weight]
Loading weights: 77%|███████████████████████████████████████████████████████████████████████████▏ | 79/103 [00:00<00:00, 10122.50it/s, Materializing param=encoder.layer.4.attention.self.value.weight]
Loading weights: 78%|███████████████████████████████████████████████████████████████████████████████▏ | 80/103 [00:00<00:00, 10181.59it/s, Materializing param=encoder.layer.4.intermediate.dense.bias]
Loading weights: 78%|███████████████████████████████████████████████████████████████████████████████▏ | 80/103 [00:00<00:00, 10128.42it/s, Materializing param=encoder.layer.4.intermediate.dense.bias]
Loading weights: 79%|██████████████████████████████████████████████████████████████████████████████▋ | 81/103 [00:00<00:00, 10187.98it/s, Materializing param=encoder.layer.4.intermediate.dense.weight]
Loading weights: 79%|██████████████████████████████████████████████████████████████████████████████▋ | 81/103 [00:00<00:00, 10135.70it/s, Materializing param=encoder.layer.4.intermediate.dense.weight]
Loading weights: 80%|██████████████████████████████████████████████████████████████████████████████████▊ | 82/103 [00:00<00:00, 10190.91it/s, Materializing param=encoder.layer.4.output.LayerNorm.bias]
Loading weights: 80%|██████████████████████████████████████████████████████████████████████████████████▊ | 82/103 [00:00<00:00, 10138.64it/s, Materializing param=encoder.layer.4.output.LayerNorm.bias]
Loading weights: 81%|██████████████████████████████████████████████████████████████████████████████████▏ | 83/103 [00:00<00:00, 10196.75it/s, Materializing param=encoder.layer.4.output.LayerNorm.weight]
Loading weights: 81%|██████████████████████████████████████████████████████████████████████████████████▏ | 83/103 [00:00<00:00, 10146.23it/s, Materializing param=encoder.layer.4.output.LayerNorm.weight]
Loading weights: 82%|████████████████████████████████████████████████████████████████████████████████████████ | 84/103 [00:00<00:00, 10204.23it/s, Materializing param=encoder.layer.4.output.dense.bias]
Loading weights: 82%|████████████████████████████████████████████████████████████████████████████████████████ | 84/103 [00:00<00:00, 10153.65it/s, Materializing param=encoder.layer.4.output.dense.bias]
Loading weights: 83%|███████████████████████████████████████████████████████████████████████████████████████▍ | 85/103 [00:00<00:00, 10210.96it/s, Materializing param=encoder.layer.4.output.dense.weight]
Loading weights: 83%|███████████████████████████████████████████████████████████████████████████████████████▍ | 85/103 [00:00<00:00, 10158.88it/s, Materializing param=encoder.layer.4.output.dense.weight]
Loading weights: 83%|██████████████████████████████████████████████████████████████████████████████▍ | 86/103 [00:00<00:00, 10214.37it/s, Materializing param=encoder.layer.5.attention.output.LayerNorm.bias]
Loading weights: 83%|██████████████████████████████████████████████████████████████████████████████▍ | 86/103 [00:00<00:00, 10156.56it/s, Materializing param=encoder.layer.5.attention.output.LayerNorm.bias]
Loading weights: 84%|█████████████████████████████████████████████████████████████████████████████▋ | 87/103 [00:00<00:00, 10202.27it/s, Materializing param=encoder.layer.5.attention.output.LayerNorm.weight]
Loading weights: 84%|█████████████████████████████████████████████████████████████████████████████▋ | 87/103 [00:00<00:00, 10151.46it/s, Materializing param=encoder.layer.5.attention.output.LayerNorm.weight]
Loading weights: 85%|███████████████████████████████████████████████████████████████████████████████████▋ | 88/103 [00:00<00:00, 10203.71it/s, Materializing param=encoder.layer.5.attention.output.dense.bias]
Loading weights: 85%|███████████████████████████████████████████████████████████████████████████████████▋ | 88/103 [00:00<00:00, 10154.58it/s, Materializing param=encoder.layer.5.attention.output.dense.bias]
Loading weights: 86%|██████████████████████████████████████████████████████████████████████████████████▉ | 89/103 [00:00<00:00, 10207.07it/s, Materializing param=encoder.layer.5.attention.output.dense.weight]
Loading weights: 86%|██████████████████████████████████████████████████████████████████████████████████▉ | 89/103 [00:00<00:00, 10154.59it/s, Materializing param=encoder.layer.5.attention.output.dense.weight]
Loading weights: 87%|█████████████████████████████████████████████████████████████████████████████████████████▏ | 90/103 [00:00<00:00, 10207.60it/s, Materializing param=encoder.layer.5.attention.self.key.bias]
Loading weights: 87%|█████████████████████████████████████████████████████████████████████████████████████████▏ | 90/103 [00:00<00:00, 10160.89it/s, Materializing param=encoder.layer.5.attention.self.key.bias]
Loading weights: 88%|████████████████████████████████████████████████████████████████████████████████████████▎ | 91/103 [00:00<00:00, 10214.13it/s, Materializing param=encoder.layer.5.attention.self.key.weight]
Loading weights: 88%|████████████████████████████████████████████████████████████████████████████████████████▎ | 91/103 [00:00<00:00, 10167.33it/s, Materializing param=encoder.layer.5.attention.self.key.weight]
Loading weights: 89%|█████████████████████████████████████████████████████████████████████████████████████████▎ | 92/103 [00:00<00:00, 10219.17it/s, Materializing param=encoder.layer.5.attention.self.query.bias]
Loading weights: 89%|█████████████████████████████████████████████████████████████████████████████████████████▎ | 92/103 [00:00<00:00, 10158.64it/s, Materializing param=encoder.layer.5.attention.self.query.bias]
Loading weights: 90%|████████████████████████████████████████████████████████████████████████████████████████▍ | 93/103 [00:00<00:00, 10209.66it/s, Materializing param=encoder.layer.5.attention.self.query.weight]
Loading weights: 90%|████████████████████████████████████████████████████████████████████████████████████████▍ | 93/103 [00:00<00:00, 10164.17it/s, Materializing param=encoder.layer.5.attention.self.query.weight]
Loading weights: 91%|███████████████████████████████████████████████████████████████████████████████████████████▎ | 94/103 [00:00<00:00, 10215.96it/s, Materializing param=encoder.layer.5.attention.self.value.bias]
Loading weights: 91%|███████████████████████████████████████████████████████████████████████████████████████████▎ | 94/103 [00:00<00:00, 10170.90it/s, Materializing param=encoder.layer.5.attention.self.value.bias]
Loading weights: 92%|██████████████████████████████████████████████████████████████████████████████████████████▍ | 95/103 [00:00<00:00, 10222.14it/s, Materializing param=encoder.layer.5.attention.self.value.weight]
Loading weights: 92%|██████████████████████████████████████████████████████████████████████████████████████████▍ | 95/103 [00:00<00:00, 10177.75it/s, Materializing param=encoder.layer.5.attention.self.value.weight]
Loading weights: 93%|███████████████████████████████████████████████████████████████████████████████████████████████ | 96/103 [00:00<00:00, 10228.19it/s, Materializing param=encoder.layer.5.intermediate.dense.bias]
Loading weights: 93%|███████████████████████████████████████████████████████████████████████████████████████████████ | 96/103 [00:00<00:00, 10184.73it/s, Materializing param=encoder.layer.5.intermediate.dense.bias]
Loading weights: 94%|██████████████████████████████████████████████████████████████████████████████████████████████▏ | 97/103 [00:00<00:00, 10213.06it/s, Materializing param=encoder.layer.5.intermediate.dense.weight]
Loading weights: 94%|██████████████████████████████████████████████████████████████████████████████████████████████▏ | 97/103 [00:00<00:00, 10165.34it/s, Materializing param=encoder.layer.5.intermediate.dense.weight]
Loading weights: 95%|██████████████████████████████████████████████████████████████████████████████████████████████████▉ | 98/103 [00:00<00:00, 10214.25it/s, Materializing param=encoder.layer.5.output.LayerNorm.bias]
Loading weights: 95%|██████████████████████████████████████████████████████████████████████████████████████████████████▉ | 98/103 [00:00<00:00, 10171.78it/s, Materializing param=encoder.layer.5.output.LayerNorm.bias]
Loading weights: 96%|██████████████████████████████████████████████████████████████████████████████████████████████████ | 99/103 [00:00<00:00, 10221.70it/s, Materializing param=encoder.layer.5.output.LayerNorm.weight]
Loading weights: 96%|██████████████████████████████████████████████████████████████████████████████████████████████████ | 99/103 [00:00<00:00, 10179.10it/s, Materializing param=encoder.layer.5.output.LayerNorm.weight]
Loading weights: 97%|███████████████████████████████████████████████████████████████████████████████████████████████████████▉ | 100/103 [00:00<00:00, 10228.26it/s, Materializing param=encoder.layer.5.output.dense.bias]
Loading weights: 97%|███████████████████████████████████████████████████████████████████████████████████████████████████████▉ | 100/103 [00:00<00:00, 10184.30it/s, Materializing param=encoder.layer.5.output.dense.bias]
Loading weights: 98%|██████████████████████████████████████████████████████████████████████████████████████████████████████▉ | 101/103 [00:00<00:00, 10232.48it/s, Materializing param=encoder.layer.5.output.dense.weight]
Loading weights: 98%|██████████████████████████████████████████████████████████████████████████████████████████████████████▉ | 101/103 [00:00<00:00, 10191.13it/s, Materializing param=encoder.layer.5.output.dense.weight]
Loading weights: 99%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▊ | 102/103 [00:00<00:00, 10231.23it/s, Materializing param=pooler.dense.bias]
Loading weights: 99%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▊ | 102/103 [00:00<00:00, 10189.81it/s, Materializing param=pooler.dense.bias]
Loading weights: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 103/103 [00:00<00:00, 10238.50it/s, Materializing param=pooler.dense.weight]
Loading weights: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 103/103 [00:00<00:00, 10198.37it/s, Materializing param=pooler.dense.weight]
Loading weights: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 103/103 [00:00<00:00, 10135.69it/s, Materializing param=pooler.dense.weight]
BertModel LOAD REPORT from: /home/songsenand/.cache/modelscope/hub/models/iic/nlp_structbert_backbone_lite_std
Key | Status | Details
-------------------------------------------+------------+--------
cls.predictions.bias | UNEXPECTED |
cls.seq_relationship.bias | UNEXPECTED |
cls.predictions.transform.dense.weight | UNEXPECTED |
cls.predictions.decoder.bias | UNEXPECTED |
cls.seq_relationship.weight | UNEXPECTED |
cls.predictions.transform.dense.bias | UNEXPECTED |
bert.embeddings.position_ids | UNEXPECTED |
cls.predictions.transform.LayerNorm.weight | UNEXPECTED |
cls.predictions.decoder.weight | UNEXPECTED |
cls.predictions.transform.LayerNorm.bias | UNEXPECTED |
Notes:
- UNEXPECTED :can be ignored when loading from different task/architecture; not ok if you expect identical arch.
/home/songsenand/Project/SUimeModelTraner/export_onnx.py:84: UserWarning: # 'dynamic_axes' is not recommended when dynamo=True, and may lead to 'torch._dynamo.exc.UserError: Constraints violated.' Supply the 'dynamic_shapes' argument instead if export is unsuccessful.
torch.onnx.export(
W0417 14:31:59.817000 710675 .venv/lib/python3.12/site-packages/torch/onnx/_internal/exporter/_compat.py:133] Setting ONNX exporter to use operator set version 18 because the requested opset_version 14 is a lower version than we have implementations for. Automatic version conversion will be performed, which may not be successful at converting to the requested version. If version conversion is unsuccessful, the opset version of the exported model will be kept at 18. Please consider setting opset_version >=18 to leverage latest ONNX features
W0417 14:32:00.100000 710675 .venv/lib/python3.12/site-packages/torch/onnx/_internal/exporter/_registration.py:110] torchvision is not installed. Skipping torchvision::nms
W0417 14:32:00.101000 710675 .venv/lib/python3.12/site-packages/torch/onnx/_internal/exporter/_registration.py:110] torchvision is not installed. Skipping torchvision::roi_align
W0417 14:32:00.101000 710675 .venv/lib/python3.12/site-packages/torch/onnx/_internal/exporter/_registration.py:110] torchvision is not installed. Skipping torchvision::roi_pool
/home/songsenand/.local/share/uv/python/cpython-3.12.12-linux-x86_64-gnu/lib/python3.12/contextlib.py:144: UserWarning: The tensor attributes self.pinyin_lstm._flat_weights[0], self.pinyin_lstm._flat_weights[1], self.pinyin_lstm._flat_weights[2], self.pinyin_lstm._flat_weights[3], self.pinyin_lstm._flat_weights[4], self.pinyin_lstm._flat_weights[5], self.pinyin_lstm._flat_weights[6], self.pinyin_lstm._flat_weights[7], self.pinyin_lstm._flat_weights[8], self.pinyin_lstm._flat_weights[9], self.pinyin_lstm._flat_weights[10], self.pinyin_lstm._flat_weights[11], self.pinyin_lstm._flat_weights[12], self.pinyin_lstm._flat_weights[13], self.pinyin_lstm._flat_weights[14], self.pinyin_lstm._flat_weights[15] were assigned during export. Such attributes must be registered as buffers using the `register_buffer` API (https://pytorch.org/docs/stable/generated/torch.nn.Module.html#torch.nn.Module.register_buffer).
next(self.gen)
/home/songsenand/.local/share/uv/python/cpython-3.12.12-linux-x86_64-gnu/lib/python3.12/copyreg.py:99: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead.
return cls.__new__(cls, *args)
The model version conversion is not supported by the onnxscript version converter and fallback is enabled. The model will be converted using the onnx C API (target version: 14).
Failed to convert the model to the target version 14 using the ONNX C API. The model was not modified
Traceback (most recent call last):
File "/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/onnxscript/version_converter/__init__.py", line 120, in call
converted_proto = _c_api_utils.call_onnx_api(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/onnxscript/version_converter/_c_api_utils.py", line 65, in call_onnx_api
result = func(proto)
^^^^^^^^^^^
File "/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/onnxscript/version_converter/__init__.py", line 115, in _partial_convert_version
return onnx.version_converter.convert_version(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/onnx/version_converter.py", line 39, in convert_version
converted_model_str = C.convert_version(model_str, target_version)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: /github/workspace/onnx/version_converter/adapters/no_previous_version.h:24: adapt: Assertion `false` failed: No Previous Version of LayerNormalization exists
/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/torch/onnx/_internal/exporter/_onnx_program.py:487: UserWarning: # The axis name: batch_size will not be used, since it shares the same shape constraints with another axis: batch_size.
rename_mapping = _dynamic_shapes.create_rename_mapping(
/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/torch/onnx/_internal/exporter/_onnx_program.py:487: UserWarning: # The axis name: seq_len will not be used, since it shares the same shape constraints with another axis: seq_len.
rename_mapping = _dynamic_shapes.create_rename_mapping(
/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/torch/nn/modules/transformer.py:531: UserWarning: The PyTorch API of nested tensors is in prototype stage and will change in the near future. We recommend specifying layout=torch.jagged when constructing a nested tensor, as this layout receives active development, has better operator coverage, and works with torch.compile. (Triggered internally at /pytorch/aten/src/ATen/NestedTensorImpl.cpp:178.)
output = torch._nested_tensor_from_mask(
/home/songsenand/Project/SUimeModelTraner/export_onnx.py:155: UserWarning: # 'dynamic_axes' is not recommended when dynamo=True, and may lead to 'torch._dynamo.exc.UserError: Constraints violated.' Supply the 'dynamic_shapes' argument instead if export is unsuccessful.
torch.onnx.export(
W0417 14:32:05.304000 710675 .venv/lib/python3.12/site-packages/torch/onnx/_internal/exporter/_compat.py:133] Setting ONNX exporter to use operator set version 18 because the requested opset_version 14 is a lower version than we have implementations for. Automatic version conversion will be performed, which may not be successful at converting to the requested version. If version conversion is unsuccessful, the opset version of the exported model will be kept at 18. Please consider setting opset_version >=18 to leverage latest ONNX features
W0417 14:32:05.492000 710675 .venv/lib/python3.12/site-packages/torch/onnx/_internal/exporter/_registration.py:110] torchvision is not installed. Skipping torchvision::nms
W0417 14:32:05.492000 710675 .venv/lib/python3.12/site-packages/torch/onnx/_internal/exporter/_registration.py:110] torchvision is not installed. Skipping torchvision::roi_align
W0417 14:32:05.492000 710675 .venv/lib/python3.12/site-packages/torch/onnx/_internal/exporter/_registration.py:110] torchvision is not installed. Skipping torchvision::roi_pool
/home/songsenand/.local/share/uv/python/cpython-3.12.12-linux-x86_64-gnu/lib/python3.12/copyreg.py:99: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead.
return cls.__new__(cls, *args)
The model version conversion is not supported by the onnxscript version converter and fallback is enabled. The model will be converted using the onnx C API (target version: 14).
Failed to convert the model to the target version 14 using the ONNX C API. The model was not modified
Traceback (most recent call last):
File "/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/onnxscript/version_converter/__init__.py", line 120, in call
converted_proto = _c_api_utils.call_onnx_api(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/onnxscript/version_converter/_c_api_utils.py", line 65, in call_onnx_api
result = func(proto)
^^^^^^^^^^^
File "/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/onnxscript/version_converter/__init__.py", line 115, in _partial_convert_version
return onnx.version_converter.convert_version(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/onnx/version_converter.py", line 39, in convert_version
converted_model_str = C.convert_version(model_str, target_version)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: /github/workspace/onnx/version_converter/adapters/no_previous_version.h:24: adapt: Assertion `false` failed: No Previous Version of LayerNormalization exists
/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/torch/onnx/_internal/exporter/_onnx_program.py:487: UserWarning: # The axis name: batch_size will not be used, since it shares the same shape constraints with another axis: batch_size.
rename_mapping = _dynamic_shapes.create_rename_mapping(
/home/songsenand/Project/SUimeModelTraner/.venv/lib/python3.12/site-packages/torch/onnx/_internal/exporter/_onnx_program.py:487: UserWarning: # The axis name: seq_len will not be used, since it shares the same shape constraints with another axis: seq_len.
rename_mapping = _dynamic_shapes.create_rename_mapping(
📊 模型配置: {'learning_rate': 1e-06, 'weight_decay': 0.05, 'warmup_ratio': 0.1, 'label_smoothing': 0.1, 'total_steps': 781250}
正在导出上下文编码器到: exported_models/context_encoder.onnx
Applied 77 of general pattern rewrite rules.
✅ 上下文编码器导出完成
✅ ONNX模型验证通过
正在导解码器到: exported_models/decoder.onnx
Applied 21 of general pattern rewrite rules.
✅ 解码器导出完成
✅ ONNX模型验证通过
✅ 示例输入已保存到: exported_models/example_inputs.npz
✅ PyTorch示例输入已保存到: exported_models/example_inputs.pt
✅ 推理示例脚本已保存到: exported_models/inference_example.py
============================================================
🎉 ONNX导出完成
============================================================
生成的模型文件:
- exported_models/context_encoder.onnx
- exported_models/decoder.onnx
- exported_models/example_inputs.npz
- exported_models/example_inputs.pt
- exported_models/inference_example.py
使用方法:
1. 检查模型: python -m onnx.checker exported_models/context_encoder.onnx
2. 运行推理示例: cd exported_models && python inference_example.py
3. 集成到您的应用: 参考inference_example.py中的ONNXInference类
注意:
- 请确保安装了onnxruntime: pip install onnxruntime
- GPU推理需要onnxruntime-gpu: pip install onnxruntime-gpu
- 束搜索算法需要根据实际需求进行调整