ResNeXt
Implementation of a classification framework
...Each branch is a small transformation (e.g. bottleneck block) and their outputs are summed—this enables richer representation without excessive parameter blowup. The design is modular and homogeneous, making it relatively easy to scale (by tuning cardinality, width, depth) and adopt in existing residual frameworks. The official repository offers a Torch (Lua) implementation with code for training, evaluation, and pretrained models on ImageNet. In practice, ResNeXt models often outperform standard ResNet models of comparable complexity.