Upload model.pt
5c8f85d
-
1.52 kB
initial commit
-
79 Bytes
Upload 2 files
model.pt
Detected Pickle imports (62)
- "__torch__.torch.nn.modules.conv.___torch_mangle_15.Conv1d",
- "__torch__.layers.Transformer_EncDec.EncoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_31.LayerNorm",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_0.Dropout",
- "__torch__.torch.nn.modules.padding.ReplicationPad1d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_11.Linear",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_36.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_34.LayerNorm",
- "__torch__.torch.nn.modules.normalization.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_35.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_21.Dropout",
- "__torch__.torch.nn.modules.linear.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_5.Conv1d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_12.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_24.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_23.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_32.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_8.Dropout",
- "__torch__.torch.nn.modules.conv.___torch_mangle_28.Conv1d",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.layers.Embed.PositionalEmbedding",
- "__torch__.torch.nn.modules.container.ModuleList",
- "__torch__.torch.nn.modules.linear.___torch_mangle_3.Linear",
- "__torch__.layers.Embed.PatchEmbedding",
- "__torch__.layers.SelfAttention_Family.___torch_mangle_14.AttentionLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_18.LayerNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_29.Conv1d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_1.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_6.LayerNorm",
- "__torch__.PatchTST.Model",
- "__torch__.layers.SelfAttention_Family.FullAttention",
- "__torch__.layers.SelfAttention_Family.AttentionLayer",
- "__torch__.torch.nn.modules.conv.___torch_mangle_16.Conv1d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_25.Linear",
- "__torch__.torch.nn.modules.flatten.Flatten",
- "__torch__.layers.Transformer_EncDec.Encoder",
- "__torch__.torch.nn.modules.linear.___torch_mangle_26.Linear",
- "__torch__.layers.SelfAttention_Family.___torch_mangle_9.FullAttention",
- "__torch__.torch.nn.modules.dropout.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_2.Linear",
- "__torch__.layers.Transformer_EncDec.___torch_mangle_20.EncoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_30.LayerNorm",
- "__torch__.PatchTST.FlattenHead",
- "__torch__.layers.SelfAttention_Family.___torch_mangle_27.AttentionLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_4.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_10.Linear",
- "__torch__.layers.Transformer_EncDec.___torch_mangle_33.EncoderLayer",
- "__torch__.layers.SelfAttention_Family.___torch_mangle_22.FullAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_19.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_13.Linear",
- "__torch__.torch.nn.modules.conv.Conv1d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_17.LayerNorm",
- "torch.DoubleStorage",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.FloatStorage",
- "torch.jit._pickle.restore_type_tag",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict"
How to fix it?
4.54 MB
Upload model.pt