download
history
blame
contribute
delete
16.4 MB
Detected Pickle imports (110)
- "__torch__.torch.nn.modules.linear.___torch_mangle_65.Linear",
- "__torch__.torch.nn.modules.container.ModuleList",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_43.LayerNorm",
- "__torch__.transformers.models.bart.modeling_bart.BartEncoderLayer",
- "__torch__.transformers.activations.___torch_mangle_69.GELUActivation",
- "__torch__.torch.nn.modules.linear.___torch_mangle_2.Linear",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_63.BartDecoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_76.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_22.LayerNorm",
- "__torch__.transformers.models.bart.modeling_bart.BartForConditionalGeneration",
- "__torch__.transformers.activations.___torch_mangle_52.GELUActivation",
- "__torch__.torch.nn.modules.linear.___torch_mangle_4.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_11.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_31.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_45.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_67.Linear",
- "__torch__.torch.nn.modules.normalization.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_24.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_49.Linear",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_58.BartSdpaAttention",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.linear.___torch_mangle_41.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_64.Linear",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_10.BartSdpaAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_55.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_59.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_13.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_1.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_37.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_38.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_15.LayerNorm",
- "__torch__.transformers.models.bart.modeling_bart.BartDecoderLayer",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_16.BartEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_72.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_70.LayerNorm",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_27.BartEncoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_82.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_3.Linear",
- "__torch__.torch.nn.modules.sparse.Embedding",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_80.BartDecoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_19.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_61.Linear",
- "__torch__.transformers.models.bart.modeling_bart.BartSdpaAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_57.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_50.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_8.Linear",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_35.BartSdpaAttention",
- "collections.OrderedDict",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_68.BartSdpaAttention",
- "__torch__.transformers.models.bart.modeling_bart.BartLearnedPositionalEmbedding",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_21.BartSdpaAttention",
- "__torch__.transformers.models.bart.modeling_bart.BartEncoder",
- "__torch__.torch.nn.modules.linear.___torch_mangle_6.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_54.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_83.Linear",
- "__torch__.ExportMiniBartG2P",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_53.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_40.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_33.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_7.Linear",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_42.BartSdpaAttention",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_30.BartLearnedPositionalEmbedding",
- "__torch__.torch.nn.modules.linear.___torch_mangle_78.Linear",
- "__torch__.transformers.activations.___torch_mangle_36.GELUActivation",
- "__torch__.torch.nn.modules.linear.___torch_mangle_14.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_56.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_5.LayerNorm",
- "__torch__.transformers.models.bart.modeling_bart.BartModel",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_28.LayerNorm",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_51.BartSdpaAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_79.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_18.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_20.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_32.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_60.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_39.Linear",
- "__torch__.transformers.activations.___torch_mangle_23.GELUActivation",
- "__torch__.torch.nn.modules.container.___torch_mangle_81.ModuleList",
- "__torch__.transformers.activations.GELUActivation",
- "__torch__.torch.nn.modules.linear.___torch_mangle_73.Linear",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_29.BartScaledWordEmbedding",
- "__torch__.torch.nn.modules.linear.___torch_mangle_71.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_9.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_46.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_62.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_44.Linear",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.linear.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_77.Linear",
- "__torch__.transformers.models.bart.modeling_bart.___torch_mangle_75.BartSdpaAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_34.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_25.Linear",
- "__torch__.transformers.models.bart.modeling_bart.BartScaledWordEmbedding",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_26.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_0.Linear",
- "__torch__.transformers.models.bart.modeling_bart.BartDecoder",
- "__torch__.torch.nn.modules.linear.___torch_mangle_66.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_17.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_48.Linear",
- "__torch__.transformers.activations.___torch_mangle_12.GELUActivation",
- "__torch__.torch.nn.modules.linear.___torch_mangle_74.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_47.Linear",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.LongStorage",
- "torch.DoubleStorage",
- "torch._utils._rebuild_tensor_v2",
- "torch.jit._pickle.restore_type_tag",
- "torch.LongStorage",
- "collections.OrderedDict"
Git LFS Details
- SHA256: e2adccd65e63a86a12f527de9c37cc29e4fc9a32ff16f074f5c42fe6214e9d75
- Pointer size: 133 Bytes
- Size of remote file: 16.4 MB
Git Large File Storage (LFS) replaces large files with text pointers inside Git, while storing the file contents on a remote server. More info.