Large-scale machine learning and deep learning models are increasingly common. For instance, GPT-3 is trained on 570 GB of text and consists of 175 billion parameters. However, whilst training large models helps improve state-of-the-art performance, deploying such cumbersome models especially on edge devices is not straightforward. Additionally, the majority of data science modeling work focuses […]
The post Knowledge Distillation: Principles, Algorithms, Applications appeared first on neptune.ai.
from Planet SciPy
read more
No comments:
Post a Comment