On Distributed Quantization for Classification. Hanna, O. A., Ezzeldin, Y. H., Sadjadpour, T., Fragouli, C., & Diggavi, S. IEEE Journal on Selected Areas in Information Theory, 1(1):237-249, May, 2020.
On Distributed Quantization for Classification [link]Arxiv  doi  abstract   bibtex   2 downloads  
We consider the problem of distributed feature quantization, where the goal is to enable a pretrained classifier at a central node to carry out its classification on features that are gathered from distributed nodes through communication constrained channels. We propose the design of distributed quantization schemes specifically tailored to the classification task: unlike quantization schemes that help the central node reconstruct the original signal as accurately as possible, our focus is not reconstruction accuracy, but instead correct classification. Our work does not make any a priori distributional assumptions on the data, but instead uses training data for the quantizer design. Our main contributions include: we prove NP-hardness of finding optimal quantizers in the general case; we design an optimal scheme for a special case; we propose quantization algorithms, that leverage discrete neural representations and training data, and can be designed in polynomial-time for any number of features, any number of classes, and arbitrary division of features across the distributed nodes. We find that tailoring the quantizers to the classification task can offer significant savings: as compared to alternatives, we can achieve more than a factor of two reduction in terms of the number of bits communicated, for the same classification accuracy.
@article{9061031,
 abstract = {We consider the problem of distributed feature quantization, where the goal is to enable a pretrained classifier at a central node to carry out its classification on features that are gathered from distributed nodes through communication constrained channels. We propose the design of distributed quantization schemes specifically tailored to the classification task: unlike quantization schemes that help the central node reconstruct the original signal as accurately as possible, our focus is not reconstruction accuracy, but instead correct classification. Our work does not make any a priori distributional assumptions on the data, but instead uses training data for the quantizer design. Our main contributions include: we prove NP-hardness of finding optimal quantizers in the general case; we design an optimal scheme for a special case; we propose quantization algorithms, that leverage discrete neural representations and training data, and can be designed in polynomial-time for any number of features, any number of classes, and arbitrary division of features across the distributed nodes. We find that tailoring the quantizers to the classification task can offer significant savings: as compared to alternatives, we can achieve more than a factor of two reduction in terms of the number of bits communicated, for the same classification accuracy.},
 author = {O. A. {Hanna} and Y. H. {Ezzeldin} and T. {Sadjadpour} and C. {Fragouli} and S. {Diggavi}},
 doi = {10.1109/JSAIT.2020.2986467},
 issn = {2641-8770},
 journal = {IEEE Journal on Selected Areas in Information Theory},
 keywords = {computational complexity;optimisation;quantisation (signal);signal classification;signal reconstruction;signal representation;distributed feature quantization;pretrained classifier;central node;distributed nodes;communication constrained channels;classification task;quantizer design;quantization algorithms;neural representations;training data;distributed quantization;NP-hardness;signal reconstruction;Quantization (signal);Image reconstruction;Training;Testing;Information theory;Task analysis;Training data;Distributed quantization;inference;communication-bounded inference;quantization with deep learning},
 month = {May},
 number = {1},
 pages = {237-249},
 tags = {journal,DML,CEDL},
 title = {On Distributed Quantization for Classification},
 type = {2},
 url_arxiv = {https://arxiv.org/abs/1911.00216},
 volume = {1},
 year = {2020}
}

Downloads: 2