download history blame contribute delete
Detected Pickle imports (155)
- "__torch__.___torch_mangle_27.TransformerBlock",
- "__torch__.torch.nn.modules.linear.___torch_mangle_41.Linear",
- "__torch__.GPTPyTorchNoCache",
- "__torch__.torch.nn.modules.linear.___torch_mangle_30.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_23.Linear",
- "__torch__.___torch_mangle_54.MultiHeadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_6.Linear",
- "__torch__.___torch_mangle_32.MultiHeadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_0.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_28.Linear",
- "torch.LongStorage",
- "__torch__.torch.nn.modules.linear.___torch_mangle_44.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_4.Linear",
- "__torch__.___torch_mangle_21.MultiHeadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_40.Linear",
- "__torch__.___torch_mangle_24.FeedForward",
- "__torch__.___torch_mangle_10.MultiHeadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_62.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_61.LayerNorm",
- "torch.ByteStorage",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_48.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_5.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_26.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_58.LayerNorm",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.linear.___torch_mangle_45.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_19.Linear",
- "__torch__.___torch_mangle_35.FeedForward",
- "__torch__.torch.nn.modules.linear.___torch_mangle_3.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_12.Linear",
- "__torch__.torch.nn.modules.sparse.Embedding",
- "__torch__.torch.nn.modules.linear.___torch_mangle_50.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_39.Linear",
- "__torch__.MultiHeadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_37.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_53.Linear",
- "__torch__.___torch_mangle_38.TransformerBlock",
- "__torch__.___torch_mangle_57.FeedForward",
- "__torch__.torch.nn.modules.linear.___torch_mangle_55.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_17.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_2.Linear",
- "__torch__.___torch_mangle_60.TransformerBlock",
- "__torch__.torch.nn.modules.linear.___torch_mangle_51.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_15.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_1.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_9.Linear",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.linear.___torch_mangle_8.Linear",
- "__torch__.LearnedPositionalEmbedding",
- "__torch__.GPTPyTorch",
- "__torch__.torch.nn.modules.normalization.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_47.LayerNorm",
- "__torch__.___torch_mangle_13.FeedForward",
- "__torch__.torch.nn.modules.linear.___torch_mangle_20.Linear",
- "__torch__.___torch_mangle_16.TransformerBlock",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_59.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_11.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_29.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_42.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_7.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_18.Linear",
- "__torch__.TransformerBlock",
- "__torch__.torch.nn.modules.linear.___torch_mangle_56.Linear",
- "__torch__.___torch_mangle_49.TransformerBlock",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_36.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_52.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_25.LayerNorm",
- "__torch__.___torch_mangle_43.MultiHeadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_31.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_22.Linear",
- "__torch__.torch.nn.modules.linear.Linear",
- "__torch__.FeedForward",
- "__torch__.___torch_mangle_46.FeedForward",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.linear.___torch_mangle_33.Linear",
- "__torch__.torch.nn.modules.container.ModuleList",
- "__torch__.torch.nn.modules.linear.___torch_mangle_34.Linear",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "collections.OrderedDict",
- "torch.LongStorage",
- "torch.FloatStorage",
- "torch.DoubleStorage",
- "torch._utils._rebuild_tensor_v2",
- "torch.jit._pickle.restore_type_tag",
- "torch.LongStorage",
- "collections.OrderedDict",
- "torch._utils._rebuild_tensor_v2"
350 MB
- Xet hash:
- f2e7468488492bb17b83ec600a4275b22abb5898d74f06f3db9cece1846b6a09
- Size of remote file:
- 350 MB
- SHA256:
- d044329c05b9a0e75fdc38fd85afdab9d6c2d4b233114b9fcb641284a635011d
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.