Near Optimal Sampling to Discover and Bound the Communication Rate of a Discrete Memoryless Channel
Links to Files
Permanent Link
Author/Creator
Author/Creator ORCID
Date
Type of Work
Department
Computer Science and Electrical Engineering
Program
Engineering, Electrical
Citation of Original Publication
Rights
This item may be protected under Title 17 of the U.S. Copyright Law. It is made available by UMBC for non-commercial research and education. For permission to publish or reproduce, please see http://aok.lib.umbc.edu/specoll/repro.php or contact Special Collections at speccoll(at)umbc.edu
Distribution Rights granted to UMBC by the author.
Distribution Rights granted to UMBC by the author.
Abstract
The aim of this research is to introduce, develop, and evaluate active learning algorithms that provide near-optimal high-probability bounds on the channel capacity across a discrete memoryless channel (DMC), where the only information about the channel internals (i.e., the channel law) is from a set of observed input-output sample pairs (i.e., channel probes). Channel capacity is the average maximum rate (for each value sent into the channel) at which information can flow across the channel with an arbitrarily low error rate. Extending results on the discrete memoryless compound channel, which is comprised of deterministic constraints on the channel law uncertainty, we first develop an offline algorithm to process an observed set of N input-output sample pairs (probes) to compute lower and upper bounds on channel capacity (such that the bounds hold with high probability). A key ingredient of this result is our new probability approximately correct (PAC) sublevel set bound that improves the convergence rate of Sanov's Theorem from O(ln(N)/N) to O(ln(ln(N)/N) (where the convergence rate is the Ôbig OhÕ BachmannÐLandau limiting behavior). The experimental (simulation) results match the analytical results. Adapting techniques from pure-exploration multi-arm bandit problems, we further develop an online algorithm aimed at minimizing the sample complexity (i.e., the number of channel probes or samples of the DMC) required to establish a lower-bound channel rate RL that satisfies several stopping criteria, where the lower bound rate RL approaches the channel capacity with high probability. We show that if the sample complexity of any optimal online algorithm (including a non-feasible clairvoyant algorithm that is given side information to simplify the analysis) is O(N) samples, then our online algorithm requires less than O(N*ln(ln(N))) sequential channel probes. Thus, we prove that our algorithm is within a sub-logarithmic factor of optimal (i.e., it is near optimal). The results of this research may provide a deeper general understanding of active machine learning algorithms.
