Compression Team
Team delivers the methods of Deep Learning models compression
BUSINESS CASES
RAM, Energy, CPU/GPU lower consumption
On-device models transfer
Models usage on low-bit CPU
Speeding up calculations
TASKS CLASSES
Low-bit and Post-Training Quantization for complex architectures (Img-to-img, Transformer)
Knowledge Distillation & Pruning
Neural Architecture Search & SuperNets
CURRENT CHALLENGES
Methods for complex architectures such as Transformers and img2img
Unification of methods usability, transition to end-to-end optimization and on-device transfer
ACHIEVEMENTS
START COOPERATION
Made on
Tilda