download
history
blame
contribute
delete
361 MB
Detected Pickle imports (119)
- "__torch__.torch.nn.modules.normalization.___torch_mangle_530.LayerNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_514.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_446.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_498.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_440.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_464.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_515.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_444.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_443.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_449.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_501.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_519.GroupNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_487.Conv2d",
- "__torch__.diffusion_outJITnative.diffusion_out",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.linear.___torch_mangle_521.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_524.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_537.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_457.LayerNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_523.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_460.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_442.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_448.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_473.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_461.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_435.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_522.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_534.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_438.MultiheadAttention",
- "__torch__.torch.nn.modules.activation.___torch_mangle_496.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_454.LayerNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_474.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_507.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_497.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_436.LayerNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_482.Conv2d",
- "__torch__.torch.nn.modules.activation.___torch_mangle_441.MultiheadAttention",
- "__torch__.torch.nn.modules.activation.___torch_mangle_459.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_504.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_536.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_518.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_502.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_434.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_481.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_485.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_535.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_532.MultiheadAttention",
- "__torch__.torch.nn.modules.activation.___torch_mangle_493.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_471.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_486.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_430.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_465.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_462.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_513.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_529.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_527.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_503.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_509.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_525.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_455.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_512.LayerNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_450.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_472.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_445.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_500.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_492.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_437.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_520.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_528.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_476.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_480.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_463.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_526.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_428.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_516.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_533.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_531.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_467.GroupNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_447.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_478.LayerNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_511.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_479.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_494.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_431.GroupNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_429.Conv2d",
- "__torch__.torch.nn.modules.conv.___torch_mangle_451.Conv2d",
- "__torch__.torch.nn.modules.conv.___torch_mangle_468.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_453.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_484.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_517.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_488.Conv2d",
- "__torch__.torch.nn.modules.conv.___torch_mangle_433.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_466.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_508.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_538.Conv2d",
- "__torch__.torch.nn.modules.activation.___torch_mangle_456.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_452.GroupNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_469.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_470.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_489.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_490.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_477.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_483.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_499.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_510.NonDynamicallyQuantizableLinear",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.conv.___torch_mangle_505.Conv2d",
- "__torch__.torch.nn.modules.conv.___torch_mangle_432.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_491.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_458.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_439.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_475.LayerNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_506.Conv2d",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.linear.___torch_mangle_495.NonDynamicallyQuantizableLinear",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.LongStorage",
- "torch.DoubleStorage"
Git LFS Details
- SHA256: be4be631eaf00a7076f53d39815eed0d60d3b71f1eb4a1e8e3c2ef33a95c30a5
- Pointer size: 134 Bytes
- Size of remote file: 361 MB
Git Large File Storage (LFS) replaces large files with text pointers inside Git, while storing the file contents on a remote server. More info.