+Model compression is the art of shrinking large [Machine Learning](/wiki/machine_learning) models into more efficient forms. This process reduces model size and computational cost, enabling faster inference and deployment on resource-constrained devices, often through techniques like [Quantization](/wiki/quantization).
+## See also
+- [Pruning](/wiki/pruning)
+- [Deep Learning](/wiki/deep_learning)
+- [Edge AI](/wiki/edge_ai)
... 1 more lines