Next Article in Journal
Visual Secure Image Encryption Scheme Based on Compressed Sensing and Regional Energy
Previous Article in Journal
An Entropy-Based Approach to Measurement of Stock Market Depth
Previous Article in Special Issue
Time-Limited Codewords over Band-Limited Channels: Data Rates and the Dimension of the W-T Space
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Class Cost-Constrained Random Coding for Correlated Sources over the Multiple-Access Channel

by
Arezou Rezazadeh
1,*,†,
Josep Font-Segura
2,
Alfonso Martinez
2 and
Albert Guillén i Fàbregas
2,3,4
1
Department of Electrical Engineering, Chalmers University of Technology, SE-412 96 Gothenburg, Sweden
2
Department of Information and Communication Technologies, Universitat Pompeu Fabra, 08018 Barcelona, Spain
3
Institució Catalana de Recerca i Estudis Avançats, 08010 Barcelona, Spain
4
Department of Engineering, University of Cambridge, Cambridge CB2 1PZ, UK
*
Author to whom correspondence should be addressed.
This work was carried out when the first author was with the Department of Information and Communication Technologies at Universitat Pompeu Fabra.
Entropy 2021, 23(5), 569; https://doi.org/10.3390/e23050569
Submission received: 12 April 2021 / Revised: 30 April 2021 / Accepted: 1 May 2021 / Published: 3 May 2021
(This article belongs to the Special Issue Finite-Length Information Theory)

Abstract

:
This paper studies a generalized version of multi-class cost-constrained random-coding ensemble with multiple auxiliary costs for the transmission of N correlated sources over an N-user multiple-access channel. For each user, the set of messages is partitioned into classes and codebooks are generated according to a distribution depending on the class index of the source message and under the constraint that the codewords satisfy a set of cost functions. Proper choices of the cost functions recover different coding schemes including message-dependent and message-independent versions of independent and identically distributed, independent conditionally distributed, constant-composition and conditional constant composition ensembles. The transmissibility region of the scheme is related to the Cover-El Gamal-Salehi region. A related family of correlated-source Gallager source exponent functions is also studied. The achievable exponents are compared for correlated and independent sources, both numerically and analytically.

1. Introduction

In information theory, the fundamental problem of communication over a channel is studied from two complementary perspectives. First, one characterizes the transmissibility conditions, namely the circumstances under which the error probability asymptically vanishes as the blocklength goes to infinity. Second, one describes by means of error exponents the speed at which this error probability vanishes; the larger the exponent, the faster the error probability tends to zero. Since finding an exact expression for error probability is very difficult, a large body of work has investigated upper and lower bounds on the average error probability, or equivalently lower and upper bounds for the error exponent. In point-to-point, that is, single-user communication, using separate source-channel random coding [1,2], possibly with expurgation [1] (Eq. 5.7.10), yields lower bounds on the error exponent. In contrast, finding an upper bound to the error exponent satisfied by every code is more challenging. Generally, the hypothesis-testing method [3] is employed to derive upper bounds for the error exponent. Two well-known upper bounds to the error exponent are the sphere-packing exponent [4] and the minimum-distance exponent [5]. In fact, for rates greater than critical rate [1] (Sec. 5.6), the random-coding and sphere-packing bounds coincide with each other, while the expurgated and minimum-distance bounds coincide at rate zero.
For point-to-point communication, it was shown in ref. [1] (Prob. 5.16) that joint source-channel coding leads in general to a larger exponent than separate source-channel coding. Additionally, using codewords with a composition dependent on the source message leads to a larger exponent than the case where codewords are drawn according to a fixed product distribution [6,7]. Moreover, a scheme where source messages are assigned to disjoint classes and encoded by codes that depend on the class index, attains the sphere-packing exponent in those cases where it is tight [8].
Many works have been devoted to studying the transmissibility and the error exponent for a two-user multiple-access channel (MAC) [9,10,11]. Separate source-channel coding for the MAC with independent sources was studied in refs. [9,12]. In ref. [13], a universal exponent for the MAC was derived by considering separate source-channel coding. In ref. [14], a transmissible region is derived for the MAC under mismatched decoding, where the decoding rule is fixed and possibly suboptimal. In ref. [15], it was shown that using structure coding can improve the error exponent of the MAC. The maximum-error-probability criterion and the impact of feedback for the MAC were studied in ref. [16]. By considering separate source-channel coding, lower and upper bounds for the error exponent of the MAC were respectively obtained in refs. [17,18]. For the MAC with independent sources, the idea of exploting the dependency between messages and codewords was studied in ref. [19]. In ref. [20], an achievable exponent for the MAC with independent sources was given in the dual domain, that is, as a lower dimensional problem over parameters in terms of Gallager functions. For the MAC with correlated sources, it was shown in ref. [11] that considering statistical dependency between messages and codewords for the MAC with correlated sources leads to a larger transmissible region. However, an example presented in ref. [21] shows that one can reliably transmit information through the MAC without satisfying the reliable transmission obtained in ref. [11]. In another line of work, superposition coding with Gacs Körner Witsenhausen (GKW) common part is used in ref. [22] to to describe the sufficient conditions lossless recoverability.
In contrast to single-user communication, the problem of reliable transmission of two correlated sources has not been solved yet and just the sufficient conditions of a reliable transmission has been derived. In ref. [23], by applying coding techniques, a new set of sufficient conditions were proposed. Moreover, in ref. [24] new sufficient conditions for the three-user MAC with correlated sources were studied. In ref. [25], an achievable exponent derived was presented in the primal domain, that is, as a multi-dimensional optimization problem over distributions that is generally difficult to analyze.
In this paper, we examine how statistical dependency between the messages and codewords improves the exponent, as well as its impact on the transmissibility region. In view of refs. [1] (Ch. 7) and [26], we study a generalized message-dependent cost-constrained random-coding ensemble with multiple cost functions. By choosing the proper cost functions, the multi-class cost-constrained ensemble subsumes multiple ensembles previously considered in the literature and recovers the transmissibility region in ref. [11].
The paper is organized as follows—in Section 2, we present the problem of transmission of N correlated sources over an N-input discrete memoryless multiple-access channel and provide the key definitions of error probability, transmissibility, random-coding ensemble, and achievable exponent. In Section 3, we review the existing random-coding ensembles, define a novel generalized multi-class cost-constraint ensemble and characterize its achievable exponent. In the discussion Section 4, we characterize the transmissibility region for our error exponent, relate the exponent to standard Gallager source and channel functions, and provide numerical results and formulas that allow us to rank the exponents attained by the various standard random-coding ensembles.

2. Problem Formulation

We study the simultaneous transmission of N correlated, discrete, memoryless sources over a channel; users are indexed by ν N = { 1 , 2 , , N } . The source messages u ν of user ν have n symbols drawn from the alphabet U ν . We denote by u σ the ordered vector of source messages for all users in a set σ 2 N , i.e., a subset of the set of all user indices, and similarly by U σ the Cartesian product of the source alphabets in the set σ . When σ = N , u N and U N denote the ordered vector of source messages for all users and the Cartesian product of the all source alphabets respectively. The sources are memoryless and are characterized by the joint probability distribution P N
P N ( u N ) = t = 1 n P N ( u N , t ) ,
and by the symbol joint probability distribution P N . The source message and symbol marginal distributions of user ν N are denoted by P ν and P ν respectively. Assuming that the sources are independent, the marginal distributions induce new joint (mismatched) probability distributions of sets of users σ 2 N . The induced independent-message and -symbol probabilities, denoted by P σ ind and P σ ind , are given by
P σ ind ( u σ ) = ν σ P ν ( u ν ) ,
and similarly for P σ ind .
Each user ν has an encoder that maps, without cooperation with the other users, the source message u ν onto a codeword x ν ( u ν ) also of length n and with symbols drawn from the alphabet X ν . We denote the codebook of user ν by C n ν . We denote by x σ X σ n the vector of codewords for all users in a set σ 2 N . Both terminals simultaneously send these codewords over a discrete memoryless multiple access channel with output alphabet Y . The symbolwise transition probability is denoted by W, and the channel is characterized by a conditional probability distribution
W ( y | x N ) = t = 1 n W ( y t | x N , t ) ,
where y is the received sequence of length n.
Based on y , a joint decoder estimates all transmitted source messages u N according to the maximum a posteriori criterion:
u ^ N = arg max u N U N n P N ( u N ) W y | x N ( u N ) ,
where U N n denotes the set of all possible source messages u N . An error occurs if the decoded messages u ^ N differ from the transmitted u N ; we refer to u ^ N u N as an error event. The error probability for a given set of codebooks, P e ( C n N ) , is thus given by
P e ( C n N ) Pr U ^ N U N .
In our analysis, it will prove convenient to split the error event into 2 N 1 distinct types of error events indexed by the non-empty subsets in the power set of the user indices 2 N \ , for example, τ { { 1 } , { 2 } , { 1 , 2 } } for N = 2 . More precisely, the error event of type τ corresponds to the conditions u ^ ν u ν for all ν τ and u ^ ν = u ν for all ν τ c , where τ c is the complement of τ in the power set of the user indices.
We are interested in the asymptotics of the error probability for sufficiently large n, namely whether the error probability vanishes and how fast this probability tends to zero as it vanishes. The sources U N are said to be transmissible over the channel if there exists a sequence of codebooks C n N such that lim n P e ( C n N ) = 0 . To characterize the speed at which the error probability vanishes, we use the notion of exponent. An exponent E is said to be achievable if there exists a sequence of codebooks such that
lim inf n 1 n log P e ( C n N ) E .
Source transmissibility and error-exponent achievability are typically studied by means of random coding. With random coding, one generates and studies sequences of ensembles of codebooks whose codewords are randomly drawn from a distribution Q ν ( x ν | u ν ) independently for each user; as indicated by the notation, this distribution may possibly depend on the source message u ν . The random-coding probability distribution for the channel input Q N ( x N | u N ) combined for all users is given by
Q N ( x N | u N ) = ν N Q ν ( x ν | u ν ) .
The use of random coding allows us to study how the error probability averaged over the ensemble, denoted by P ¯ e vanishes as n grows. More importantly, it shows the existence of good codes in the ensemble such that their error probability vanishes. For the point-to-point and the multiple-access channels, a number of such random-coding ensembles have been studied in the literature, as reviewed in the following section, where we also present a multi-class cost-constrained ensemble subsuming all these ensembles and characterize the achievable exponent and transmissibility region of this ensemble.

Summary of Notation Used in the Paper

Sets are usually denoted by calligraphic upper case letter, e.g., X , and the n-Cartesian product set of X is denoted by X n . The cardinality of a set such as X is denoted by | X | . The indicator function representing an error event or that an element x belongs to a set X is denoted by 1 { x X } .
The number of users is denoted by N and user indices are typically represented by ν . The set of all users is denoted by N . The power set of all subsets of N is denoted by 2 N and the complement of a subset σ 2 N is denoted by σ c ; sets in the power set of users are denoted that by Greek letters, for example, τ and σ . The number of source-message classes and of cost functions for user ν are respectively denoted by K ν and L ν ; the sets of such classes are functions are respectively denoted by K ν and L ν . Indices for source classes and cost functions are typically denoted by i ν and ν respectively.
Subscripts and superscripts in a quantity A may represent sets of user indices σ . Depending on the context, the quantity represents a list or a suitable product of variables for all elements in the set σ . For instance, for σ = { 1 , 2 } , A σ = ( A 1 , A 2 ) or A σ = ( A 1 , A 2 ) . If the quantity is a probability distribution, its value for σ represents the probability distribution of the sequence, for example, Q σ i σ ( x σ ) = ν σ Q ν i ν ( x ν ) . If the quantity is a set, its value for σ is the Cartesian product, for example, U σ = U 1 × U 2 for σ = { 1 , 2 } . If σ = , then A σ = A σ = 0 . If σ is a singleton, for example, σ = { 2 } , we simply write A 2 or A 2 . We denote the operation that merges and sorts two lists A σ 1 and A σ 2 with σ 1 σ 2 = into an ordered list containing all users in the union σ 1 σ 2 by [ A σ 1 , A σ 2 ] . For sets of user indices, we denote such merging operation by [ σ 1 , σ 2 ] and we have [ σ , σ c ] = N .
Scalar random variables are denoted by capital letters, for example, X and lowercase letters represent a particular realisation, for example, x X . Capital bold letter denotes random vectors or sequences, for example, X , while small bold letter x X n denote deterministic vectors or sequences. Probability distributions for vectors or sequences, typically of length n, (resp. for symbols) are represented by text-style letters, for example, P , Q , W (resp. math-style letters, for example, P, Q, W). Sequences symbols are usually affixed a subscript to indicate a user index; the t-th symbol in the sequence x ν is denoted by x ν , t .
The source-symbol distribution for user ν is denoted by P ν ( u ν ) . The joint distribution for users σ is denoted by P σ ( u σ ) ; the joint distribution, computed as if the sources were independent, is denoted by P σ ind ( u σ ) . The conditional source distribution for users σ 1 given another set σ 2 is denoted by P σ 1 | σ 2 ( u σ 1 | u σ 2 ) . Vector or sequence distributions are defined analogously with P replaced by P . Channel input distributions are denoted by Q ν ( x ν ) , Q ν i ν ( x ν ) , or Q ν , u ν i ν ( x ν ) , where i ν denotes the index of the class source message and Q ν , u ν ( x ν ) is a shorthand for the conditional distribution Q ν ( x ν | u ν ) . Cost functions are similarly denoted by a ν ( x ν ) , a ν i ν ( x ν ) , or a ν , u ν i ν ( x ν ) . Vector or sequence distributions are defined analogously with Q or a respectively replaced by Q or a . The conditional distribution for the channel output symbol (resp. sequence) is denoted by W ( y | x N ) (resp. W ( y | x N ) ).

3. Multi-Class Cost-Constrained Ensemble with Statistical Dependency

3.1. Review of Random-Coding Ensembles

The simplest and oldest random-coding ensemble is the independent, identically distributed (iid) [1,12,17,27], where the symbols x ν , t in all codewords x ν of a given user ν are generated independently according to the same input distributions Q ν ( x ν , t ) for all source messages u ν . Throughout the paper, we shall identify ensembles by hyphenated acronyms, where the first part indicates the possible dependence of the codeword on the source message and the second part describes the generation of symbols in a codeword. This first ensemble is thus the message-independent iid (mi-iid) ensemble, since codewords have the same distribution for all source messages and symbols are independent of each other and independent of the source message symbols too. For the mi-iid ensemble, the random-coding distribution is given by
Q ν mi - iid ( x ν | u ν ) = t = 1 n Q ν ( x ν , t ) .
In the message-independent, independent-conditionally-distributed (mi-icd) ensemble, the codewords x ν of user ν are generated identically for all source messages u ν , independently of the full message u ν , and with symbols according to a set of | U ν | conditional probability distributions Q ν , u ν ( x ν ) Q ν ( x ν | u ν ) . To this end, let I u ν ( u ν ) denote the set of positions where the symbol u ν U appears in the sequence u ν , namely
I u ν ( u ν ) = t { 1 , 2 , , n } : u ν , t = u ν .
Within each subsequence of u ν where u ν , t = u ν , represented by u ν I u ν ( u ν ) , symbols are drawn independently according to Q ν , u ν ( x ν ) . For this mi-icd ensemble, codewords are generated according to
Q ν mi - icd ( x ν | u ν ) = u ν U ν t I u ν ( u ν ) Q ν , u ν ( x ν , t )
= t = 1 n Q ν , u ν , t ( x ν , t ) .
Compared to the mi-iid ensemble, the mi-icd ensemble can lead to a larger transmissible region for the multiple-access channel with correlated sources [11,21]. An example of generation of three codewords x ν ( 1 ) , x ν ( 2 ) and x ν ( 3 ) in the mi-icd ensemble is shown in Figure 1, for a given source sequence u ν = ( α , β , β , γ , β , γ , γ , α , β , α ) with source alphabet U = { α , β , γ } . To generate each codeword x ν with alphabet X = { a , c , e } , three subcodewords x ν ( I α ( u ν ) , x ν ( I β ( u ν ) and x ν ( I γ ( u ν ) are pairwise-independently generated with i. i. d. distributions Q ν , α = ( 1 \ 3 , 1 \ 3 , 1 \ 3 ) , Q ν , β = ( 1 \ 2 , 1 \ 4 , 1 \ 4 ) and Q ν , γ = ( 1 \ 3 , 2 \ 3 , 0 ) , respectively. Symbols generated according to Q ν , α , Q ν , β and Q ν , γ are respectively represented as green circles, blue boxes and red diamonds in the figure. In the example, I α ( u ν ) = { 1 , 8 , 10 } , I β ( u ν ) = { 2 , 3 , 5 , 9 } and I γ ( u ν ) = { 4 , 6 , 7 } . For instance, the subcodeword x ν ( 1 ) ( I γ ( u ν ) has three symbols, each generated independently from Q ν , γ , leading to the red-diamond symbols x ν ( 1 ) ( I γ ( u ν ) = ( a , a , a ) .
Next, we have the message-dependent iid (md-iid) ensemble [6,8,19,25,28], where codewords for each user are generated with i. i. d. symbols according to different distributions Q ν i ν ( x ν ) that depend on the full source message through the class index i ν of the class the source message belongs to. More precisely, for user ν = 1 , 2 with source marginal distribution P ν , the i ν -th class A ν i ν , where i ν K ν = { 1 , , K ν } , is defined as the set of all source messages whose probability P ν ( u ν ) is within a given interval, that is,
A ν i ν = u ν U ν n : γ ν , i ν n < P ν ( u ν ) γ ν , i ν 1 n ,
where the thresholds γ ν , j are K ν + 1 non-negative numbers, ordered from higher to lower, such that 0 = γ ν , K ν γ ν , K ν 1 γ ν , 1 < γ ν , 0 = 1 , and min u ν P ν ( u ν ) < γ ν , K ν 1 and γ ν , 1 max u ν P ν ( u ν ) . The md-iid random-coding distribution is given by
Q ν md - iid ( x ν | u ν ) = t = 1 n Q ν i ν ( u ν ) ( x ν , t ) .
The exponent of this md-iid ensemble can be larger than that of the mi-iid ensemble for joint source-channel coding [8,20,28].
In the message-dependent, independent conditional symbol distributions (md-icd) ensemble, messages in the class i ν for user ν are encoded with codewords whose symbols are generated independently according to the conditional input distribution Q ν , u ν i ν ( x ν ) . The random-coding distribution of the md-icd ensemble is thus given by
Q ν md - icd ( x ν | u ν ) = u ν U ν t I u ν ( u ν ) Q ν , u ν i ν ( u ν ) ( x ν , t ) .
In the message-independent, constant-composition (mi-cc) ensemble [29,30], codewords x ν are drawn independently with an empirical distribution Q ^ ν ( x ν ) close to a given Q ν ( x ν ) , independently of the source message u ν . For each user, codewords x ν are randomly picked from T ν n ( Q ν ) , the set of all sequences whose empirical distribution has a variational distance to Q ν of at most 1 \ n , that is
T ν n ( Q ν ) = x ν X ν n : max x ν Q ^ ν ( x ν ) Q ν ( x ν ) < 1 n .
For this mi-cc ensemble, the random-coding distribution is given by
Q ν mi - cc ( x ν | u ν ) = 1 | T ν n ( Q ν ) | 1 x ν T ν n ( Q ν ) .
While the mi-cc and mi-iid ensembles lead to identical transmissibility conditions, the former may achieve strictly larger exponents for suboptimal input distributions already in single-user settings [29].
The message-independent, conditional constant-composition (mi-ccc) ensemble combines features of the mi-icd and mi-cc ensembles. For each subsequence u ν I u ν ( u ν ) , the corresponding subcodewords x ν I u ν ( u ν ) are drawn independently from the set T ν | I u ν ( u ν ) | ( Q ν , u ν ) of subsequences with empirical distribution close to Q ν , u ν ( x ν ) , namely
T ν | I u ν ( u ν ) | ( Q ν , u ν ) = x ν X ν | I u ν ( u ν ) | : max x ν x ν Q ^ ν ( x ν ) Q ν , u ν ( x ν ) < 1 | I u ν ( u ν ) | .
The random-coding distribution of the mi-ccc ensemble is given by
Q ν mi - ccc ( x ν | u ν ) = u ν U ν 1 T ν | I u ν ( u ν ) | ( Q ν , u ν ) 1 x ν I u ν ( u ν ) T ν | I u ν ( u ν ) | ( Q ν , u ν ) .
An example of the generation of three codewords x ν ( 4 ) , x ν ( 5 ) and x ν ( 6 ) in the mi-ccc ensemble is also shown in Figure 1 as a comparison to the md-iid ensemble, for the same source sequence u ν , source alphabet U = { α , β , γ } and input alphabet X = { a , c , e } . Now, to generate each codeword x ν , three subcodewords x ν ( I α ( u ν ) ) , x ν ( I β ( u ν ) ) and x ν ( I γ ( u ν ) ) are pairwise-independently, uniformly drawn in the type classes with empirical distributions Q ^ ν , α , Q ^ ν , β and Q ^ ν , γ that are closest to Q ν , α , Q ν , β and Q ν , γ , respectively. Since in the example | I α ( u ν ) | = 3 , | I β ( u ν ) | = 4 and | I γ ( u ν ) | = 3 , it follows that Q ^ ν , α = ( 1 \ 3 , 1 \ 3 , 1 \ 3 ) , Q ^ ν , β = ( 1 \ 2 , 1 \ 4 , 1 \ 4 ) and Q ^ ν , γ = ( 1 \ 3 , 2 \ 3 , 0 ) . Symbols generated according to Q ^ ν , α , Q ^ ν , β and Q ^ ν , γ are respectively represented as green doubled circles, blue doubled boxes and red doubled diamonds in the figure. For instance, all subcodewords x ν ( j ) ( I γ ( u ν ) ) , for j = 4 , 5 , 6 , have three symbols jointly generated from the constant-composition type Q ^ ν , γ , that is, exactly one a and two cs.
The message-dependent, constant-composition (md-cc) ensemble combines the features of having different distributions for different messages with constant-composition random coding. For messages in the class i ν { 1 , , K ν } for user ν , codewords are drawn from the set of sequences with empirical distribution close to Q ν i ν ( x ν ) . For this ensemble, the random-coding distribution is given by
Q ν md - cc ( x ν | u ν ) = 1 T ν n Q ν i ν ( u ν ) 1 x ν T ν n Q ν i ν ( u ν ) ) } .
Finally, the message-dependent, conditional constant-composition (md-ccc) ensemble combines several of the ensembles listed above. For a given message u ν = ( u ν , 1 , , u ν , n ) in the i ν -th class, that is, u ν A ν i ν , the subsequence of u ν having the same symbol u ν , that is, u ν I u ν ( u ν ) , is encoded with pairwise-independent codewords generated from the set of codewords with empirical distribution very close to Q ν , u ν i ν ( x ν ) . The random-coding distribution of the md-ccc ensemble is thus given by
Q ν md - ccc ( x ν | u ν ) = u ν U ν 1 T ν | I u ν ( u ν ) | ( Q ν , u ν i ν ( u ν ) ) 1 x ν I u ν ( u ν ) T ν | I u ν ( u ν ) | ( Q ν , u ν i ν ( u ν ) ) .

3.2. Generalized Multi-Class Cost-Constrained Ensemble

Motivated by the ensembles listed in the previous section, and inspired by refs. [1] (Ch. 7) and [26] (Sec. II), we study a generalized message-dependent multi-class cost-constrained random-coding ensemble with multiple auxiliary costs.
For each user, we partition the set of source messages into K ν disjoint classes with thresholds on the message probabilities as in Equation (12). Let the source message be in the i ν -th class, that is, i ν ( u ν ) = i ν . Given the source message u ν and the source symbol u ν , we consider the subsequence u ν I u ν ( u ν ) , where I u ν ( u ν ) is defined in Equation (9), and we denote the corresponding source subsequence and subcodeword by u ν I u ν ( u ν ) and x ν I u ν ( u ν ) respectively. For each user ν , class index i ν , and source message symbol u ν , the subcodeword x ν I u ν ( u ν ) is drawn according to a symbolwise i. i. d. distribution Q ν , u ν i ν ( x ν ) conditioned on a set of cost constraints being satisfied. We consider L ν additive cost functions a ν , u ν i ν , ν ( x ν ) , ν L ν = { 1 , , L ν } . The total cost a ν , u ν i ν , ν x ν I u ν ( u ν ) of the subcodeword x ν I u ν ( u ν ) is given by the sum of the symbol costs a ν , u ν i ν , ν , namely
a ν , u ν i ν , ν x ν I u ν ( u ν ) = j I u ν ( u ν ) a ν , u ν i ν , ν ( x ν , j ) .
We assume that the average cost ϕ ν , u ν i ν , ν under the conditional distribution Q ν , u ν i ν is zero:
ϕ ν , u ν i ν , ν = x ν X ν Q ν , u ν i ν ( x ν ) a ν , u ν i ν , ν ( x ν ) = 0 .
Finally, fix some parameters δ ν > 0 and let D ν i ν be the set of codewords for which the average empirical cost of its constituent subcodewords 1 | I u ν ( u ν ) | a ν , u ν i ν , ν x ν I u ν ( u ν ) is close to the statistical mean ϕ ν , u ν i ν , ν = 0 for all cost functions and source symbols, i.e.,
D ν , u ν i ν x ν X ν n : 1 | I u ν ( u ν ) | a ν , u ν i ν , ν x ν I u ν ( u ν ) δ ν | I u ν ( u ν ) | , u ν U ν , ν L ν .
Codewords x ν are the combination of subcodewords x ν I u ν ( u ν ) with respective positions in I u ν ( u ν ) . For this multi-class cost-constrained ensemble, the random-coding distribution is thus given by
Q ν cost ( x ν | u ν ) = 1 Ξ ν u ν U ν t I u ν ( u ν ) Q ν , u ν i ν ( x ν , t ) 1 x ν D ν , u ν i ν
= 1 Ξ ν t = 1 n Q ν , u ν , t i ν ( x ν , t ) 1 x ν D ν , u ν i ν ,
where Ξ ν is a normalizing constant and the class index is determined by the source message, i ν = i ν ( u ν ) .
The multi-class cost-constrained ensemble subsumes all the ensembles described in Section 3.1. First of all, the iid and icd ensembles are recovered by setting L ν = 0 and choosing the appropriate number of classes K ν and random-coding distributions Q ν , Q ν , u ν , Q ν i ν and Q ν , u ν i ν . For all these cases, the set D ν , u ν i ν includes all generated codewords and the normalizing constant is Ξ ν = 1 .
To recover the constant-composition ensembles, for which constraints force the subcodewords to belong to some set T ν n ( Q ν ) or T ν | I u ν ( u ν ) | ( Q ν , u ν i ν ) , for each of the K ν classes for user ν we set δ ν < 1 , L ν = | X ν | and bijectively map the channel input symbols to cost function indices ν ( x ν ) so that
a ν , u ν i ν , ν ( x ν ) = 1 x ν = ν Q ν , u ν i ν ( ν ) .
In case the ensemble does not depend on either i ν or u ν , these symbols are dropped from Equation (26). For example, for the md-cc ensemble, we have a ν , ν i ν ( ν ) = 1 x ν = ν Q ν i ν ( x ν ) . In addition, the codeword set D ν , u ν i ν in Equation (23) is simplified as
D ν , u ν i ν = x ν X ν n : 1 n t = 1 n 1 x ν , t = x Q ν i ν ( x ) 1 n , x X ν ,
which is the same as T n ( Q ν i ν ) given a version of Equation (15) where Q ν may depend on i ν .
Again, choosing the right number of classes K ν and random-coding distributions Q ν , Q ν , u ν , Q ν i ν , and Q ν , u ν i ν recovers the various constant-composition ensembles. By construction, the set D ν , u ν i ν includes only the (sub)codewords with empirical distribution close to respectively Q ν , Q ν , u ν , Q ν i ν , and Q ν , u ν i ν , and the normalizing constant Ξ ν is the probability of the corresponding type set (or product thereof). As an example, for the md-ccc ensemble, choosing the cost functions in Equation (26) as follows
a ν , u ν i ν , ν x ν I u ν ( u ν ) = j I u ν ( u ν ) 1 x ν , j = ν Q ν , u ν i ν ( ν )
yields the following cost-constraint set, which is equivalent to Equation (17),
D ν , u ν i ν = x ν X ν n : | j I u ( u ν ) 1 x ν , j = x I u ( u ν ) Q ν , u ν i ν ( x ) | 1 I u ( u ν ) , u U ν , x X ν .

3.3. Exponent for the Generalized Multi-Class Cost-Constrained Ensemble

Theorem 1.
For the transmission of N correlated memorlyess sources with joint distribution P N , where N = { 1 , 2 , , N } , over a channel with input x N over a memoryless channel with transition probabilitiy W ( y | x N ) , consider a random-coding multi-class cost-constrained ensemble where source messages for each user ν N are allocated, depending on their probabilities, into K ν classes with thresholds { γ ν , 0 , γ ν , 1 , , γ ν , K ν } , as in Equation (12), and encoded onto codewords randomly generated with a distribution Q ν i ν ( x ν | u ν ) that depends on the source message according to Equation (24) through symbol distributions Q ν , u ν i ν that possibly depend on the source-message class index i ν and source symbol u ν and L ν cost functions a ν , u ν i ν , ν , ν { 1 , 2 , , L ν } . This random-coding ensemble attains the following exponent E cos t
E cost = min τ 2 N \ , i N K N max 0 ρ 1 max λ N L , U 0 , r N u N N R E τ i N ρ , λ N L , U , r N u N N ,
where the Gallager function E τ i N ρ , λ N L , U , r N u N N is given by
E τ i N ρ , λ N L , U , r N u N N = log u τ c , x τ c , y u τ , x τ P N ( u N ) 1 1 + ρ Λ N i N ( u N ) Q τ , u τ i τ ( x τ ) R N , u N i N ( x N ) Q τ c , u τ c i τ c ( x τ c ) W ( y | x N ) 1 1 + ρ 1 + ρ ,
and the functions Λ σ i σ ( u σ ) and R σ , u σ i σ ( x σ ) are respectively given by
Λ σ i σ ( u σ ) = ν σ P ν ( u ν ) γ ν , i ν λ ν L γ ν , i ν 1 P ν ( u ν ) λ ν U ,
R σ , u σ i σ ( x σ ) = ν σ ν L ν e r ν u ν ν a ν , u ν i ν , ν ( x ν ) ,
and implicitly depende on the set of optimization parameters λ N L , U , r N u N N .
Proof. 
This result is proved in Appendix A. □
The random-coding exponent in Equation (30) depends on the partitioning of the source-message set into classes, the channel input distributions, and the codeword cost-constraint functions. The best possible generalized cost-constraint exponent is obtained by optimizing over the multi-class partitioning, the cost constraints and the input distributions. We briefly discuss the optimization w. r. t. the thresholds of the source messages partitioning in Appendix B. In the next section, we provide some numerical examples where we compute the optimal exponents for either independent or correlated sources, and find that the optimal number of classes is two. In ref. [31] (Sec. 3.2.1.1), we provide some indications of why this optimality of only two classes is harder to establish in multi-user scenarios, compared to the single-user case. In the next section, we use Equations (31) and (30) to respectively obtain the source and channel Gallager functions of the various ensembles in Section 3.1 and rank their achievable exponents and transmissibility regions.

4. Discussion

4.1. Gallager Functions for Correlated Sources

In this section, we evaluate the generalized Gallager function E τ i N ρ , λ N L , U , r N u N N of the multi-class cost-constrained ensemble in Equation (31) for the various ensembles described in Section 3.1. In the cases where it is possible, we relate this Gallager function to the well-known [1] correlated-source and channel Gallager functions, respectively given by:
E s , σ ( ρ , P N ) = log u σ c u σ P N ( u N ) 1 1 + ρ 1 + ρ ,
E 0 ( ρ , Q , W ) = log y x Q ( x ) W ( y | x ) 1 1 + ρ 1 + ρ ,
where σ 2 N . Using that [ u σ , u σ c ] = u N , the standard Gallager source function is given by E s ( ρ , P N ) = E s , N ( ρ , P N ) , with N = { 1 , , N } the set of user indices.
For the simple mi-iid ensemble, with only one source class and no cost constraints, K ν = 1 and L ν = 0 for all ν N , and Λ σ i σ ( u σ ) = R σ , u σ i σ ( x σ ) = 1 for all σ 2 N . With no statistical dependency between messages and codewords, Q ν , u ν ( x ν ) = Q ν ( x ν ) . Setting i N = 1 and λ N L , U = r N u N N = 0 in Equation (31) gives the Gallager function E τ mi - iid ( ρ , P N , Q N , W ) ,
E τ mi - iid ( ρ , P N , Q N , W ) = log u τ c , x τ c , y u τ , x τ P N ( u N ) 1 1 + ρ Q τ ( x τ ) Q τ c ( x τ c ) W ( y | x N ) 1 1 + ρ 1 + ρ .
Isolating the summations over u τ c and u τ , we can split the Gallager function as
E τ mi - iid ( ρ , P N , Q N , W ) = E 0 ( ρ , Q τ , Q τ c W ) E s , τ ( ρ , P N ) ,
where Q τ c W is a shorthand for Q τ c ( x τ c ) W ( y | x N ) , the transition probability of a channel with input x τ and output ( x τ c , y ) .
For the mi-icd ensemble, we have a similar set-up as for the mi-iid ensemble, where Q ν , u ν ( x ν ) now may depend on u ν . In this case, the Gallager function E τ mi - icd ( · ) is given by Equation (36) with Q σ ( x σ ) replaced by Q σ , u σ ( x σ ) , for σ { τ , τ c } :
E τ mi - icd ( ρ , P N , Q N , U , W ) = log u τ c , x τ c , y u τ , x τ P N ( u N ) 1 1 + ρ Q τ , u τ ( x τ ) Q τ c , u τ c ( x τ c ) W ( y | x N ) 1 1 + ρ 1 + ρ .
As the summations over u τ c and u τ are not independent from the rest, the Gallager function does not split into source and channel functions unless the sources are independent, in which case one can find an mi-iid ensemble with a tilted unconditional input distribution and identical exponent. To this end, and for a given conditional input distribution Q ν , u ν ( x ν ) , let us define a tilted distribution Q ν ρ ( x ν ) as
Q ν ρ ( x ν ) = u ν P ν ( u ν ) 1 1 + ρ u ¯ ν P ν ( u ¯ ν ) 1 1 + ρ Q ν , u ν ( x ν ) .
From this equation, we have the following equality:
Q ν ρ ( x ν ) u ¯ ν P ν ( u ¯ ν ) 1 1 + ρ = u ν P ν ( u ν ) 1 1 + ρ Q ν , u ν ( x ν ) .
Substituting this identity together with P N ( u N ) = P τ ( u τ ) P τ c ( u τ c ) in Equation (38) and rearranging the result, we obtain the following Gallager function for independent sources:
E τ mi - icd ( ρ , P N , Q N , U , W ) = E 0 ( ρ , Q τ ρ , W Q τ c ) E s ( ρ , P τ )
= E τ mi - iid ( ρ , P N , [ Q τ ρ , Q τ c ] , W ) .
For the md-iid and md-icd ensembles, there are K ν source classes per user and no cost constraints, i.e., L ν = 0 and R σ , u σ i σ ( x σ ) = 1 for ν N and σ 2 N . Settting r N u N N = 0 in Equation (31) gives the Gallager function E τ , i N md - icd ( · ) for generic i N [31] (Eq. (4.36)),
E τ , i N md - icd ( ρ , P N , Q N , U i N , W ) = log u τ c , x τ c , y u τ , x τ P N ( u N ) 1 1 + ρ Λ N i N ( u N ) Q τ , u τ i τ ( x τ ) Q τ c , u τ c i τ c ( x τ c ) W ( y | x N ) 1 1 + ρ 1 + ρ .
The Gallager function E τ , i N md - iid ( · ) for the md-iid ensemble is obtained by setting Q σ , u σ ( x σ ) = Q σ ( x σ ) , independent of u ν , for σ { τ , τ c } in Equation (43). As the summations over u τ c and u τ are now independent from the rest, the Gallager function splits as
E τ , i N md - iid ( ρ , P N , Q N i N , W ) = E 0 ( ρ , Q τ i τ , Q τ c i τ c W ) E s , τ i N ( ρ , P N ) ,
where we defined E s , τ i N ( ρ , P N ) , a modified Gallager E s -function, as
E s , τ i N ( ρ , P N ) = log u τ c u τ P N ( u N ) 1 1 + ρ Λ N i N ( u N ) 1 + ρ .
The maximization w.r.t. λ N L , U in Equation (30) only affects the second term in the r. h. s. of Equation (44), since the function Λ N i N only appears in the source part of the exponent. In Appendix C, we discuss the properties of Equation (45) after the maximization w.r.t. λ N L , U as a function of ρ , and establish some connections to the Gallager source function (34) and to the source functions for the single-user md-iid ensemble in ref. [8].
The Gallager functions for the constant-composition ensembles differ from the ones considered so far in the presence of L ν = | X ν | cost functions a ν , u ν i ν , ν ( x ν ) , given in Equation (26), for each input distribution Q ν , u ν i ν ( x ν ) . These cost functions appear in the Gallager functions through the factors R σ , u σ i σ ( x σ ) , for σ { τ , τ c } that multiply each appearance of Q σ , u σ i σ ( x σ ) in the function, and through their associated optimization parameters r N u N N . The expressions of the Gallager functions for these constant-composition ensembles can be easily inferred from this obversation, so we focus on the factor R σ , u σ i σ ( x σ ) itself.
For the mi-cc and md-cc ensembles, the cost functions a ν , u ν i ν , ν ( x ν ) , factor R σ , u σ i σ ( x σ ) , and associated optimization parameter r ν u ν ν are independent of u ν , we thus write a ν i ν , ν ( x ν ) , R σ i σ ( x σ ) , and r ν ν . The expressions in Equations (26) and (33) for L ν = X ν give
R ν i ν ( x ν ) = e ν X ν r ν ν 1 { x ν = ν } Q ν i ν ( ν ) .
The exponent in Equation (46) can be evaluated as
ν X ν r ν ν 1 x ν = ν Q ν i ν ( ν ) = r ν x ν ν X ν r ν ν Q ν i ν ( ν )
= α τ , ν i ν ( x ν ) ,
where we have defined a function α τ , ν i ν ( x ν ) that depends on τ and i ν through the optimization parameters r ν ν . We can be easily verify that α τ , ν i ν has zero mean, in other words, x ν α τ , ν i ν ( x ν ) Q ν i ν ( x ν ) = 0 . At this point, the parameters r ν ν may be replaced by the equivalent real-valued functions α τ , ν i ν ( x ν ) . We obtain the mi-cc Gallager function E τ mi - cc ( · ) by setting i N = 1 and λ N L , U = 0 in Equation (31),
E τ mi - cc ( ρ , α τ , N , P N , Q N , W )
= log u τ c , x τ c , y u τ , x τ P N ( u N ) 1 1 + ρ Q τ ( x τ ) e α τ , N ( x N ) Q τ c ( x τ c ) W ( y | x N ) 1 1 + ρ 1 + ρ
= log x τ c , y x τ Q τ ( x τ ) e α τ , N ( x N ) Q τ c ( x τ c ) W ( y | x N ) 1 1 + ρ 1 + ρ E s , τ ( ρ , P N ) ,
where we split the Gallager function into channel and source terms in analogy to Equation (37).
In ref. [31] (Eq. (4.49)), the md-cc ensemble was studied for N = 2 users in both the primal and dual domains. The md-cc Gallager function E τ md - cc ( · ) for N users is obtained by combining the derivation of Equation (50) with that of Equation (44) to yield
E τ md - cc ( ρ , α τ , N i N , P N , Q N i N , W ) = log x τ c , y x τ Q τ ( x τ ) e α τ , N i N ( x N ) Q τ c ( x τ c ) W ( y | x N ) 1 1 + ρ 1 + ρ log u τ c u τ P N ( u N ) 1 1 + ρ Λ N i N ( u N ) 1 + ρ .
As in previous cases, the exponent is obtained after maximization over α τ , N i N .
Concluding our list, the cost functions a ν , u ν i ν , ν ( x ν ) , factors R σ , u σ i σ ( x σ ) , and parameters r ν u ν ν for the mi-ccc and md-ccc ensembles do depend on u ν . In analogy to Equation (48), we define a zero-mean function β τ , ν , u ν i ν ( x ν ) as
β τ , ν , u ν i ν ( x ν ) = r ν u ν x ν ν X ν r ν u ν ν Q ν , u ν i ν ( ν ) ,
and similarly for β τ , ν , u ν ( x ν ) for the mi-ccc ensemble. The Gallager function for the mi-ccc ensemble E τ mi - ccc ( · ) is obtained by combining the derivations of Equation (50) and of Equation (38),
E τ mi - ccc ( ρ , β τ , N , u N , P N , Q N , U , W ) = log u τ c , x τ c , y u τ , x τ P N ( u N ) 1 1 + ρ Q τ , u τ ( x τ ) e β τ , N , u N ( x N ) Q τ c , u τ c ( x τ c ) W ( y | x N ) 1 1 + ρ 1 + ρ .
Similarly, for the md-ccc ensemble, and in agreeement with the 2-user case studied in ref. [31] (Eq. (4.45)), combining the derivations of Equations (50) and (43), yields
E τ , i N md - ccc ( ρ , β τ , N , u N i N , P N , Q N , U i N , W ) = log u τ c , x τ c , y u τ , x τ P N ( u N ) 1 1 + ρ Λ N i N ( u N ) Q τ , u τ i τ ( x τ ) e β τ , N , u N i N ( x N ) Q τ c , u τ c i τ c ( x τ c ) W ( y | x N ) 1 1 + ρ 1 + ρ .

4.2. Transmissibility

We may obtain the transmissibility conditions from the achievable exponents derived in Section 4.1, following the random-coding method described in ref. [1] (Th. 5.6.4). The analysis extends the transmissibility condition for joint source-channel coding in ref. [1] (Prob. 5.16), to account for statistical dependency of the codeword on the source message in the multiuser set-up. As mentioned above, the source U N is transmissible over the channel W if there exists a sequence of codes with vanishing error probability, or equivalently, with strictly positive achievable error exponent E c o s t in Equation (30). As an example, we present the derivation for the mi-icd ensemble where the class and cost functions in Equations (32) and (33) are inactive, namely Λ σ i σ ( u σ ) = R σ , u σ i σ ( x σ ) = 1 for all σ 2 N , and leave the general case of K ν > 1 classes and cost-constrained codewords as an open problem.
For the mi-icd case, and similarly to Gallager’s E 0 -function [1] (Th. 5.6.3), the Gallager function E τ mi - icd ( · ) in Equation (38) is concave (∩) with respect to ρ and satisfies E τ mi - icd ( ρ = 0 , · ) = 0 . For every τ 2 N \ , let ρ ^ τ be the optimizer given by
ρ ^ τ = arg max 0 ρ 1 E τ mi - icd ( ρ , P N , Q N , U , W ) .
Therefore, the achievable exponent is strictly positive, namely E τ mi - icd ( ρ ^ τ , · ) > 0 , as far as the slope of the E τ mi - icd ( ρ , · ) function is strictly positive at ρ = 0 , that is
ρ E τ mi - icd ( ρ , P N , Q N , U , W ) ρ = 0 > 0 .
Taking the derivative with respect to ρ at both sides of Equation (38), after some algebraic manipulations, we find that (56) is equivalent to
u τ c P τ c ( u τ c ) x τ c , y u τ , x τ P τ | τ c ( u τ | u τ c ) Q τ , u τ ( x τ ) Q τ c , u τ c ( x τ c ) W ( y | x N ) × × log P τ | τ c ( u τ | u τ c ) Q τ c , u τ c ( x τ c ) W ( y | x N ) u ¯ τ , x ¯ τ P τ | τ c ( u ¯ τ | u τ c ) Q τ , u ¯ τ ( x ¯ τ ) Q τ c , u τ c ( x τ c ) W ( y | [ x ¯ τ , x τ c ] ) > 0 .
We next write the expression in the left hand-side of the inequality (57) in terms of entropy and mutual information. We denote as H ( P ) the entropy of a source with distribution P [32] (Eq. (2.1)) and by I ( Q , W ) the mutual information of a channel W with input distribution Q [32] (Eq. (2.28)). For σ 2 N , we define a channel input distribution Q τ | σ , that is conditioned to the source messages u σ , as
Q τ | σ ( x τ | u σ ) = u τ U τ P τ | σ ( u τ | u σ ) Q τ , u τ ( x τ ) .
Therefore, the transmissibility condition (57) can be compactly expressed as
H P τ | τ c < I Q τ | τ c , W | P τ c Q τ c | τ c , τ 2 N \ .
As it is, Q τ c | τ c is “transparent”, as it cancels inside the fraction, and the channel law may also be written as Q τ c | τ c W , removing the conditioning in the mutual information. With N = { 1 , 2 } in Equation (59), we recover the achievable Cover-El Gamal-Salehi region [11] (Eq. (3)).

4.3. Numerical Examples

In this section, we present two simple examples showing that the exponent of the md-iid ensemble can be larger than that of the mi-iid ensemble with only two classes (and associated input distributions) for each user. First, we consider two correlated discrete memoryless sources, N = 2 and N = { 1 , 2 } , with alphabet U ν = { 0 , 1 } for both users ν N , and probability distribution P N ( u 1 , u 2 ) given in matrix form as
P N = 0.0005 0.0095 0.0005 0.9895 .
The sources are sent over a discrete memoryless multiple-access channel with input alphabets X 1 = X 2 = { 1 , 2 , 3 , 4 , 5 , 6 } and output alphabet Y = { 1 , 2 , 3 , 4 } . The channel transition probabilites are given by a 36 × 4 matrix W, such that W ( y | x 1 , x 2 ) is the row x 1 + 6 ( x 2 1 ) . The transition matrix W is given by
W = W 1 W 2 W 3 W 4 W 5 W 6 ,
where the 6 × 4 submatrices W , = 1 , , 6 are given as follows. First, the submatrix W 1 corresponds to the point-to-point channel discussed in ref. [8] (Sec. IV.C), given by
W 1 = 1 3 k 1 k 1 k 1 k 1 k 1 1 3 k 1 k 1 k 1 k 1 k 1 1 3 k 1 k 1 k 1 k 1 k 1 1 3 k 1 0.5 k 2 0.5 k 2 k 2 k 2 k 2 k 2 0.5 k 2 0.5 k 2 ,
for k 1 = 0.045 and k 2 = 0.01 . Let the m-th row of matrix W 1 is denoted by W 1 ( m ) . The matrix W 2 (resp. W 3 ) is a 6 × 4 matrix whose rows are all W 1 ( 5 ) (resp. W 1 ( 6 ) ). The matrices W 4 , W 5 and W 6 are respectively given by
W 4 = W 1 ( 2 ) W 1 ( 3 ) W 1 ( 4 ) W 1 ( 1 ) W 1 ( 6 ) W 1 ( 5 ) , W 5 = W 1 ( 3 ) W 1 ( 4 ) W 1 ( 1 ) W 1 ( 2 ) W 1 ( 5 ) W 1 ( 6 ) , W 6 = W 1 ( 4 ) W 1 ( 1 ) W 1 ( 2 ) W 1 ( 3 ) W 1 ( 6 ) W 1 ( 5 ) .
The optimal achievable exponent [8] (Sec. IV.C) for the single-user channel W 1 in Equation (62) is related to two different distributions Q and Q , given in vector form by
Q = 0 , 0 , 0 , 0 , 1 \ 2 , 1 \ 2 ,
Q = 1 \ 4 , 1 \ 4 , 1 \ 4 , 1 \ 4 , 0 , 0 .
We let each user employ these distributions in the md-iid ensemble with input distribution in Equation (13) according to the source message partitioning in Equation (12) with K ν = 2 classes per user and thresholds γ N = ( γ 1 , γ 2 ) . Since we consider two input distributions for each user, the channel Gallager function max ρ [ 0 , 1 ] E 0 ( ρ , Q τ i τ , W Q τ c i τ c ) is not concave in ρ [8]. To find the md-iid exponent E md - iid , we optimize over the class thresholds following the method in Appendix B with the Gallager function in Equation (44), exploit the properties of the source function in Equation (45) in Appendix C, and also find the optimal input distribution assignment of Q ν i ν for each ν { 1 , 2 } . In our setting, we have four possible assignments, namely
Ω 1 : Q 1 1 = Q 2 1 = Q , Q 1 2 = Q 2 2 = Q ,
Ω 2 : Q 1 1 = Q 2 2 = Q , Q 1 2 = Q 2 1 = Q ,
Ω 3 : Q 1 2 = Q 2 1 = Q , Q 1 1 = Q 2 2 = Q ,
Ω 4 : Q 1 2 = Q 2 2 = Q , Q 1 1 = Q 2 1 = Q .
We start our numerical discussion by assessing which of the possible four assignments in Equations (66)–(69) leads to a higher error exponent. For each possible pair of thresholds ( γ 1 , γ 2 ) , we numerically calculate the optimal assignment Ω ( γ N ) given by
Ω ( γ N ) = arg max Ω j min i N min τ E τ i N ( γ N ) ,
and the corresponding achievable error exponent E m d i i d ( γ N ) as
E c o s t ( γ N ) = max Ω j min i N min τ E τ i N ( γ N ) ,
where the exponent function E τ i N ( γ N ) is given in Equation (A55). Figure 2 and Figure 3 respectively show Ω ( γ N ) and E c o s t ( γ N ) for the valid range of γ N . For most pair of thresholds ( γ 1 , γ 2 ) , assignments Ω 1 and Ω 3 lead to the highest exponent among the possible assignments, while assignments Ω 2 and Ω 4 are optimal only for a marginal region. Using this information, and combined with the values of the achievable exponents in Figure 3, we determine the message-dependent exponent
E m d i i d = max γ N E c o s t ( γ N ) .
In this example, we obtained the achievable exponent E md - iid = 0.2611 , corresponding to the input distribution assignment Ω 1 in Equation (66) and optimal source message partitioning γ 1 = 0.8469 and γ 2 = 0.6581 . The optimal point γ N is shown by a white (black) bullet in Figure 2 (Figure 3).
Alternatively, we may first optimize over γ N and then over the assignments Ω j . To do so, we solve the system of Equation (A58) in Appendix B to numerically determine the optimal thresholds γ N , and compute the exponent E cost ( Ω j ) as
E cost ( Ω j ) = min i N min τ E τ i N ( γ N ) ,
where the exponent function E τ i N ( γ N ) is given in Equation (A55). We provide in Table 1 the values of the optimal thresholds γ N and exponents E τ i N ( γ N ) under the different assignment Ω j , for the three types of error τ and the four possible user classes i N . For each assignment, the minimum over i N and τ as in Equation (73) is highlighted in gray, leading to the exponent E cos t ( Ω j ) . The message-dependent exponent is then
E md - iid = max j E cos t ( Ω j ) ,
recovering the error exponent E md - iid = 0.2611 for input distribution assignment Ω 1 obtained using the previous method in Equation (71).
In the second example, we consider the transmission of two independent discrete memoryless sources with identical source alphabets U ν = { 0 , 1 } with distributions induced by the marginals of Equation (60), given by P 1 ( 0 ) = 0.01 and P 2 ( 0 ) = 0.001 . These sources are transmitted over the multiple-access channel with transition probability given by Equation (61), and are encoded using the md-iid ensemble with the input distribution assignments Ω j in Equations (66)–(69). Following the footsteps of the correlated sources case, in Table 2 we calculate optimal thresholds γ N and exponents E τ i N ( γ N ) for the possible input distribution assignments and determine the exponent of the md-iid ensemble using Equations (73) and (74). In this case, the optimal assignment is again Ω 1 , with optimal source message partitioning specified by the thresholds γ 1 = 0.8779 and γ 2 = 0.6933 , achieving an exponent of E md - iid = 0.2458 , slightly smaller than that of correlated sources.
For the sake of completeness and purpose of comparison, we also calculate the exponent for the mi-iid ensemble described in Equation (8). In the absence of message dependence, for a given assignment Ω j , the mi-iid exponent is given by
E no - cost ( Ω j ) = min τ E τ ,
where the exponent function E τ is given by E τ = max ρ E τ mi - iid ( ρ , P N , Q N , W ) and E τ mi - iid is the Gallager function in Equation (37), described in the previous subsection. For both the correlated and independent sources described above, Table 3 presents the achievable exponents E τ for each type of error τ and input distribution assignment ( Q 1 , Q 2 ) , where Q 1 and Q 2 are either of Q and Q in Equations (64) and (65). In our numerical example for correlated sources, the assignment with highest exponent is ( Q 1 , Q 2 ) = ( Q , Q ) , giving an exponent of E mi - iid = 0.2503 , slightly smaller than that of the md-iid ensemble. In contrast, the mi-iid exponent for independent sources, according to the second part of Table 3 is found to be E mi iid = 0.2367 with input distribution ( Q 1 , Q 2 ) = ( Q , Q ) . In this case, the md-iid exponent E md iid is around 4 % larger that the mi-iid; this situation is in contrast with to-point communication, where the gain in exponent achieved by an ensemble with two distributions is typically smaller, for example, 1 % in ref. [8]. Hence, message-dependent random coding with two class distributions, compared to iid random coding, may lead to a higher error exponent gain in the MAC than in point-to-point communication.

4.4. Comparison of the Random-Coding Achievable Error Exponents

From the numerical results presented in Section 4.3, as well as from refs. [8,20,28,31], the message-dependent ensembles attain in general a larger exponent than their message-independent counterparts. We now compare the random-coding exponents for the ensembles presented in Section 3.1, whose Gallager functions were obtained in Section 4.1.
For independent sources, we found in Equation (42) that for a given conditional input distribution Q ν , u ν ( x ν ) and ρ , there exists an iiid distribution Q ν , ρ given by Equation (39) with identical Gallager function. Thus, the mi-iid and mi-icd ensembles attains the same exponent, after maximization over the input distributions. Similarly, we conclude that md-iid and md-icd-ensembles attain the same exponent.
In ref. [31] (Prop. 2.9), it was proved that for point-to-point communication, the exponent of the mi-ccc ensemble may be lower than that of the mi-cc ensemble. The same steps actually prove the same result for the MAC with independent sources. Thus, for the MAC with independent sources we have
E m i c c c E m i c c E m d c c , E m d c c c E m d c c ,
E m i i i d E m i c c E m d c c , E m d i i d E m d c c ,
and E m d c c is thus largest among the ensembles in Section 3.1 for an arbitrary input distribution. As discussed in ref. [29] (Th. 4), for optimal input distributions both E m d c c and E m d i i d may coincide.
Concerning the optimal partitioning into message classes, for point-to-point communication it is known that partitioning the source-message set into two classes is sufficient to attain the optimal error exponent [8,31] (Prop. 2.7). However, the proof of ref. [31] (Prop. 2.7) cannot be easily generalized to the MAC with independent sources. At the same time, we could not find an example showing that assigning more than two input distributions leads to a larger exponent. Hence, finding the sufficient number of input distributions is for the message-dependent exponent is an open problem.
The comparisons in Equations (76) and (77) for correlated sources require, in general, a more sophisticated machinery and we consider here two simple cases. For the message-dependent md-icd and md-ccc ensembles, we observe that compared to E τ , i N m d i c d in Equation (43) the E τ , i N m d c c c exponent in Equation (54) contains an additional term β τ , N , u N i N ( x N ) to guarantee the constant-composition distribution as in Equation (52). This allows to recover E τ , i N m d c c c by setting β τ , N , u N i N ( x N ) = 0 in E τ , i N m d i c d and to prove that E m d i c d E m d c c c after maximizing w. r. t. β τ , N , u N i N ( x N ) . Similarly for the ensembles with statistical independence between messages and codewords, we observe that the constant-composition exponent E τ , i N m d c c in Equation (51) also contains the additional term α τ , N i N ( x N ) compared to its iid counterpart E τ , i N m d i i d in Equation (44), yielding E m d i i d E m d c c . Put together, for correlated sources it holds that
E m d i c d E m d c c c , E m d i i d E m d c c ,
suggesting that, as in the case of single-user communication, the use of constant-composition input distributions may lead to higher exponents than the symbol-wise independent distributions when transmitting correlated sources over the MAC.
Summarizing, proper choices of the cost functions recover the different coding schemes considered in Section 3.1, including message-dependent and message-independent versions of iid, independent conditionally distributed, constant-composition, and conditional constant composition ensembles. Thanks to the flexibility of the generalized cost-constraint random-coding ensemble, the achievable exponents of the various ensembles can be compared and ranked, both numerically and analytically.

Author Contributions

Conceptualization, A.M. and A.G.i.F.; methodology J.F.-S., software, A.R.; formal analysis, A.R.; witting A.R., J.F.-S. and A.M. All authors have equally contributed to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been funded in part by the European Research Council under grant 725411, and by the Spanish Ministry of Economy and Competitiveness under grant TEC2016-78434-C3-1-R.

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Proof of Theorem 1

We start by bounding the average error probability over the generalized cost-constraint ensemble, P ¯ e . Counting ties as errors, the random coding union bound [2] (Th. 16) for joint source-channel is
P ¯ e u N , x N , y P N ( u N ) Q N ( x N | u N ) W ( y | x N ) min 1 , u ^ N u N Pr P N ( u ^ N ) W y | X ^ N P N ( u N ) W ( y | x N ) 1 ,
where Q N ( x N | u N ) is given by Equation (7), with every user using the generalized cost-constrained input distribution Q ν ( x ν | u ν ) as in Equation (24), and x ^ N has the same distribution as x N but conditioned on u ^ N rather than u N , i.e., Q N ( x N | u ^ N ) . The summation over u ^ N u N can be split into 2 N 1 distinct types of error events indexed by the non-empty subsets in the power set of the user indices 2 N \ , e.g., τ { { 1 } , { 2 } , { 1 , 2 } } for N = 2 , such that u ^ τ c = u τ c and u ^ ν u ν for all ν τ .
Since min { 1 , a + b } min { 1 , a } + min { 1 , b } , we bound P ¯ e as
P ¯ e τ 2 N \ P ¯ e τ ,
where P ¯ e τ is in turn given by
P ¯ e τ = u N , x N , y P N ( u N ) Q N ( x N | u N ) W ( y | x N ) min 1 , u ^ N : u ^ τ c = u τ c , u ^ ν u ν , ν τ Pr P N ( u ^ N ) W y | [ x τ c , X ^ τ ] P N ( u N ) W ( y | x N ) 1 ,
where the inner probability is computed according to the distribution Q τ ( x τ | u τ ) , including only users u τ in the set τ as x ^ τ c = x τ c . We recall that [ x τ c , X ^ τ ] is the sorted merger of the channel inputs for users in the sets τ c and τ , in this case x τ c and X ^ τ respectively.
Next, we split the summation over u N in Equation (A3) into classes i N K N defined by Equation (12), summing then over the messages belonging to the Cartesian product of the sets A N i N . We note that codewords are generated according to distributions that depend on the class index of the sources. Let D N , u N i N be the Cartesian product of the sets of codewords D ν , u ν i ν in Equation (23) for ν = 1 , 2 , , N , and define
Q N i N ( x N | u N ) = ν N Q ν i ν ( x ν | u ν ) ,
where Q ν i ν ( x ν | u ν ) is given by either Equation (24) or Equation (25). Then, the double outer summation of Equation (A3) over u N and x N can be written as
u N , x N P N ( u N ) Q N ( x N | u N ) = i N K N u N A N i N x N D N , u N i N P N ( u N ) Q N i N ( x N | u N )
= i τ c K τ c u τ c A τ c i τ c x τ c D τ c , u τ c i τ c P τ c ( u τ c ) Q τ c i τ c ( x τ c | u τ c ) i τ K τ u τ A τ i τ x τ D τ , u τ i τ P τ | τ c ( u τ | u τ c ) Q τ i τ ( x τ | u τ ) ,
where we split the summations over u N and x N into separate summations over u τ c and u τ , similarly with x τ c and x τ with the corresponding rearrangements in the probabilities, and written the term Q τ c i τ c ( x τ c | u τ c ) in a similar way to Equation (A4). The inner summation of Equation (A3) can be split in an analogous manner based on the classes to which u ^ τ belongs, now indexed by the variable j τ K τ . Applying this fact together with Markov’s inequality
Pr { A 1 } min s 0 E [ A s ]
to upper bound the probability with a parameter s 0 that implicitly depends on the error-event type τ and indices i τ c , i τ , and j τ . We bound the inner summation of Equation (A3) as
u ^ N : u ^ τ c = u τ c , u ^ ν u ν , ν τ Pr P N ( u ^ N ) W y | [ x τ c , X ^ τ ] P N ( u N ) W ( y | x N ) 1 j τ K τ min s 0 u ^ τ A τ j τ x ^ τ D τ , u ^ τ j τ Q τ j τ ( x ^ τ | u ^ τ ) P τ | τ c ( u ^ τ | u τ c ) W y | [ x τ c , x ^ τ ] P τ | τ c ( u τ | u τ c ) W y | [ x τ c , x τ ] s ,
where we also used that P N ( u ^ N ) = P τ c ( u ^ τ c ) P τ | τ c ( u ^ τ | u ^ τ c ) = P τ c ( u τ c ) P τ | τ c ( u ^ τ | u τ c ) to rewrite the message probabilities in Equation (A8) and we expressed the codeword x N as [ x τ c , x τ ] .
Inserting Equations (A6) and (A8) into Equation (A3) and using the following inequality for A 0 ,
min { 1 , A } min ρ [ 0 , 1 ] A ρ ,
where ρ [ 0 , 1 ] , we further bound P ¯ e τ as
P ¯ e τ i τ c K τ c i τ K τ j τ K τ min s 0 min ρ [ 0 , 1 ] P ¯ e , i τ c i τ τ , j τ ,
where after some minor rearrangements P ¯ e , i τ c i τ τ , j τ is in turn given by
P ¯ e , i τ c i τ τ , j τ = u τ c A τ c i τ c x τ c D τ c , u τ c i τ c P τ c ( u τ c ) Q τ c i τ c ( x τ c | u τ c ) y Y n u τ A τ i τ x τ D τ , u τ i τ P τ | τ c ( u τ | u τ c ) Q τ i τ ( x τ | u τ ) W y | [ x τ c , x τ ] u ^ τ A τ j τ x ^ τ D τ , u ^ τ j τ Q τ j τ ( x ^ τ | u ^ τ ) P τ | τ c ( u ^ τ | u τ c ) W y | [ x τ c , x ^ τ ] P τ | τ c ( u τ | u τ c ) W y | [ x τ c , x τ ] s ρ .
Note that, for some conveniently chosen variables z 0 and z i τ , sets Z 0 and Z i τ , as well as functions f 0 ( z 0 ) and f i τ s ( z 0 , z i τ ) , with i τ K τ , we can express P ¯ e , i τ c i τ τ , j τ as
P ¯ e , i τ c i τ τ , j τ = z 0 Z 0 f 0 ( z 0 ) z i τ Z i τ f i τ 1 s i τ , j τ ρ i τ , j τ ( z 0 , z i τ ) z j τ Z j τ f j τ s i τ , j τ ( z 0 , z j τ ) ρ i τ , j τ .
In Equation (A12), the variable z 0 stands for the triplet ( u τ c , x τ c , y ) , the alphabet Z 0 for the Cartesian product A τ c i τ c × D τ c , u τ c i τ c × Y n and the function f 0 ( z 0 ) is given by P τ c ( u τ c ) Q τ c i τ c ( x τ c | u τ c ) . The variable z i τ stands for the pair ( u τ , x τ ) , the alphabet Z i τ for the Cartesian product A τ i τ × D τ , u τ i τ and the function f i τ s ( z 0 , z i τ ) is given by P τ | τ c ( u τ | u τ c ) s Q τ i τ ( x τ | u τ ) W y | [ x τ c , x τ ] s .
The optimization parameters s and ρ in Equation (A10) implictly depend on the error-event type τ and the indices i τ c , i τ , and j τ . For new parameters ρ ¯ τ [ 0 , 1 ] , τ K τ , setting
s i τ , j τ = 1 1 + ρ ¯ j τ ,
ρ i τ , j τ = ρ ¯ i τ ( 1 + ρ ¯ j τ ) 1 + ρ ¯ i τ ,
In Equation (A12), we obtain the following partial upper bound in Equation (A10) as
min s 0 min ρ [ 0 , 1 ] P ¯ e , i τ c i τ τ , j τ min ρ ¯ τ [ 0 , 1 ] K τ z 0 Z 0 f 0 ( z 0 ) z i τ Z i τ f i τ 1 1 + ρ ¯ i τ ( z 0 , z i τ ) z j τ Z j τ f j τ 1 1 + ρ ¯ j τ ( z 0 , z j τ ) ρ ¯ i τ ( 1 + ρ ¯ j τ ) 1 + ρ ¯ i τ .
Here, we have kept implicit the dependence on τ and i τ c of the optimization parameter ρ ¯ τ . Now, applying Hölder’s inequality [33] (Th. 13) in the form
i K α i a i b i i K α i a i p 1 p i K α i b i p p 1 p 1 p , for p [ 1 , ) ,
to the expression in Equation (A15) with p i τ , j τ = 1 + ρ ¯ i τ 1 , we obtain
min s 0 min ρ [ 0 , 1 ] P ¯ e , i τ c i τ τ , j τ min ρ ¯ τ [ 0 , 1 ] K τ z 0 Z 0 f 0 ( z 0 ) z i τ Z i τ f i τ 1 1 + ρ ¯ i τ ( z 0 , z i τ ) 1 + ρ ¯ i τ 1 1 + ρ ¯ i τ z 0 Z 0 f 0 ( z 0 ) z j τ Z j τ f j τ 1 1 + ρ ¯ j τ ( z 0 , z j τ ) 1 + ρ ¯ j τ ρ ¯ i τ 1 + ρ ¯ i τ .
Next, putting Equation (A17) back in Equation (A10) and using the following inequality, proved in Appendix A.1, for A i 0 and 0 s i 1
i , j K A i s i A j 1 s i 2 | K | i K A i
in the double summation over i τ and j τ in Equation (A10), the following upper bound holds
i τ K τ j τ K τ min s 0 min ρ [ 0 , 1 ] P ¯ e , i τ c i τ τ , j τ 2 K τ i τ K τ min ρ [ 0 , 1 ] P ¯ e , i τ c i τ τ ,
where we have moved the optimization over ρ ¯ τ inside the summation over i τ and renamed ρ ¯ τ as ρ , with the dependence on the index i τ kept implicit. Moreover, the expression for P ¯ e , i τ c i τ τ is in fact given by P ¯ e , i τ c i τ τ , j τ in Equation (A11) after setting i τ = j τ , s = 1 1 + ρ and rearranging terms, that is,
P ¯ e , i τ c i τ τ = u τ c A τ c i τ c x τ c D τ c , u τ c i τ c P τ c ( u τ c ) Q τ c i τ c ( x τ c | u τ c ) y Y n u τ A τ i τ x τ D τ , u τ i τ P τ | τ c ( u τ | u τ c ) 1 1 + ρ Q τ i τ ( x τ | u τ ) W ( y | x N ) 1 1 + ρ 1 + ρ .
It remains to factorize Equation (A20) into a product of symbol distributions in order to obtain a single-letter expression for the exponent. We start by upper bounding the summations over the input messages u τ c and u τ . For a list of users σ with corresponding messages u σ , list of class indices i σ and some function p σ i σ ( u σ ) , we have that
u σ A σ i σ p σ i σ ( u σ ) = u σ U σ n p σ i σ ( u σ ) 1 { u σ A σ i σ } ,
where we used the definition of the message sets A σ i σ in Equation (12) and the identity
i K f i = i N f i 1 { i K } .
Using the upper bound
1 { a < b c } min λ L , λ U 0 b a λ L c b λ U
for a , b , c > 0 with λ L , λ U 0 , together with the fact that the source-message classes are defined separately for each user to express the source message probabilities in terms of P σ ind ( u σ ) = ν σ P ν ( u ν ) similarly to Equation (2), we upper bound the r. h. s. of Equation (A21) as
u σ A σ i σ p σ i σ ( u σ ) min λ σ L , U 0 u σ U σ n p σ i σ ( u σ ) P σ ind ( u σ ) γ σ , i σ n λ σ L γ σ , i σ 1 n P σ ind ( u σ ) λ σ U ,
where we jointly wrote λ σ L and λ σ U as λ σ L , U . Definining
Λ σ i σ ( u σ ) = P σ ind ( u σ ) γ σ , i σ λ σ L γ σ , i σ 1 P σ ind ( u σ ) λ σ U .
and taking into account that the sources are memoryless, we obtain that the summations w. r. t. the source messages u τ and u τ c in Equation (A20) are upper bounded as
u σ A σ i σ p σ i σ i σ ( u σ ) min λ σ L , U u σ U σ n p σ i σ c i σ ( u σ ) t = 1 n Λ σ i σ ( u σ , t ) .
respecively for σ = τ and σ = τ c .
We proceed in a similar manner for the summations w. r. t. the codewords x τ and x τ c in Equation (A20). For a list of users σ and some function q σ i σ ( u σ , x σ ) implicitly defined, the summation over channel codewords x σ D σ , u σ i σ can be upper bounded as:
x σ D σ , u σ i σ q σ i σ ( u σ , x σ ) = x σ X σ n q σ i σ ( u σ , x σ ) 1 { x σ D σ , u σ i σ }
= x σ X σ n q σ i σ ( u σ , x σ ) ν σ υ ν U ν ν L ν 1 a ν , υ ν i ν , ν x ν I υ ν ( u ν ) δ ν
min r σ υ σ σ , r ¯ σ υ σ σ x σ X σ n q σ i σ ( u σ , x σ ) υ σ U σ σ L σ e r σ υ σ σ a σ , υ σ i σ , σ ( x σ ( I υ σ ( u σ ) ) ) + r ¯ σ υ σ σ δ σ ,
where we used Equation (A22) in Equation (A27), the fact that the codeword ensembles are defined separately for each user together with the definition of the ensemble cost constraints in Equation (23) and subcodewords x ν I u ν ( u ν ) in Equation (A28), and a variant of Equation (A23) proved in Appendix A.2,
1 { | a | δ } min r , r ¯ e r a + r ¯ δ
for r R and r ¯ 0 , in Equation (A29) for each indicator function of Equation (A28) and combined the product of exponentials over σ as a single exponential using the list notation. We continue by rewriting the double product over υ σ and σ in Equation (A29) as follows
υ σ U σ σ L σ e r σ υ σ σ a σ , υ σ i σ , σ ( x σ ( I υ σ ( u σ ) ) ) + r ¯ σ υ σ σ δ σ = υ σ U σ σ L σ e r ¯ σ υ σ σ δ σ t I υ σ ( u σ ) e r σ υ σ σ a σ , υ σ i σ , σ ( x σ , t ( I υ σ ( u σ ) ) )
= β σ t = 1 n R σ , u τ , t i τ ( x τ , t ) ,
where in Equation (A31) we wrote the cost function in terms of the symbol costs and in Equation (A32) we rearranged terms and introduced a factor β σ that depends on the list { r ¯ σ υ σ σ } and a function R σ , u σ i σ ( x σ ) that depends on the list { r σ υ σ σ } and are respectively given by
β σ = υ σ U σ σ L σ e r ¯ σ υ σ σ δ σ ,
R σ , u σ i σ ( x σ ) = σ L σ e r σ u σ σ a σ , u σ i σ , σ ( x σ ) .
Replacing Equation (A32) back into Equation (A29), we obtain that the summations over the codewords are upper bounded as
x σ D σ , u σ i σ q σ i σ ( u σ , x σ ) min r σ u σ σ , r ¯ σ u σ σ x σ X σ n q σ i σ ( u σ , x σ ) β σ t = 1 n R σ , u σ , t i σ ( x τ , t ) .
for both user lists σ = τ and σ = τ c .
We now combine Equations (A26) and (A35) for σ = τ to bound the summation inside the parenthesis in Equation (A20) as
u τ A τ i τ x τ D τ , u τ i τ P τ | τ c ( u τ | u τ c ) 1 1 + ρ Q τ i τ ( x τ | u τ ) W ( y | x N ) 1 1 + ρ u τ U τ n x τ X τ n P τ | τ c ( u τ | u τ c ) 1 1 + ρ min λ τ L , U r τ u τ τ , r ¯ τ u τ τ β τ Ξ τ t = 1 n Λ τ i τ ( u τ , t ) Q τ , u τ , t i τ ( x τ , t ) R τ , u τ , t i τ ( x τ , t ) W ( y | x N ) 1 1 + ρ ,
where we expressed the distribution Q τ i τ ( x τ , u τ ) in terms of the symbol-wise iid distribution Q τ , u τ i τ ( x ) as in Equation (25). Since both source and channel are memoryless, we may now factorize and rearrange the expression in Equation (A36) into single-letter, symbolwise factors as
u τ A τ i τ x τ D τ , u τ i τ P τ | τ c ( u τ | u τ c ) 1 1 + ρ Q τ i τ ( x τ | u τ ) W ( y | x N ) 1 1 + ρ min λ τ L , U r τ u τ τ , r ¯ τ u τ τ β τ Ξ τ t = 1 n g τ i τ ( u τ c , t , x τ c , t , y t ) ,
where, for a list of users σ , the function g σ i σ ( u σ c , x σ c , y ) is defined as
g σ i σ ( u σ c , x σ c , y ) = u σ U σ x σ X σ P σ | σ c ( u σ | u σ c ) 1 1 + ρ Λ σ i σ ( u σ ) Q σ , u σ i σ ( x σ ) R σ , u σ i σ ( x σ ) W ( y | x N ) 1 1 + ρ .
Although not explicitly, the function g τ i τ ( u τ c , x τ c , y ) in Equation (A37) depends on several optimization parameters, namely ρ , λ τ L , U , r τ u τ τ , r ¯ τ u τ τ , which depend in turn on the error-event type τ and class indices i τ and i τ c .
Again, we use Equations (A26) and (A35) for σ = τ c and the fact that the source is memoryless to upper bound the summation outside the parenthesis in Equation (A20) as
u τ c A τ c i τ c x τ c D τ c , u τ c i τ c P τ c ( u τ c ) Q τ c i τ c ( x τ c | u τ c ) min λ τ c L , U r τ c u τ c τ c , r ¯ τ c u τ c τ c u τ c U τ c n x τ c X τ c n β τ c Ξ τ c t = 1 n P τ c ( u τ c , t ) Λ τ c i τ c ( u τ c , t ) Q τ c , u τ c , t i τ c ( x τ c , t ) R τ c , u τ c , t i τ c ( x τ c , t ) .
Substituting Equations (A37) and (A39) in Equation (A20), the resulting expression back into Equation (A19) and then into Equation (A10), we get
P ¯ e τ 2 K τ i τ c K τ c i τ K τ min ρ , λ τ c L , U r τ c u τ c τ c , r ¯ τ c u τ c τ c u τ c U τ c n x τ c X τ c n β τ c Ξ τ c t = 1 n P τ c ( u τ c , t ) Λ τ c i τ c ( u τ c , t ) Q τ c , u τ c , t i τ c ( x τ c , t ) R τ c , u τ c , t i τ c ( x τ c , t ) y Y n β τ Ξ τ 1 + ρ min λ τ L , U r τ u τ τ , r ¯ τ u τ τ t = 1 n g τ i τ ( u τ c , t , x τ c , t , y t ) 1 + ρ .
Let us define now the function h σ i σ , i σ c of the user set σ and the class indices i σ and i σ c as
h σ i σ , i σ c = u σ U σ , x σ X σ , y Y P σ ( u σ ) Λ σ i σ ( u σ ) Q σ , u σ i σ ( x σ ) R σ , u σ i σ ( x σ ) g σ c i σ c ( u σ , x σ , y ) 1 + ρ .
With this definition, we can rewrite Equation (A40) in a compact manner as
P ¯ e τ 2 K τ i N K N min ρ , λ N L , U , r N u N N , r ¯ N u N N β τ c Ξ τ c β τ Ξ τ 1 + ρ t = 1 n h τ c i τ c , i τ ,
where we have also combined the complementary sets i τ c and i τ into i N , and similarly for λ N L , λ N U , r N u N N , and r ¯ N u N N . Finally, substituting Equation (A42) into Equation (A10) and then back into Equation (A2), taking (minus) the logarithm of the bound on P ¯ e , dividing the result by n, and the limit as n , we obtain a lower bound E KL cost to the exponent of the generalized cost-constrained ensemble E KL cost , namely
E K cost L = min τ , i N max ρ , λ N L , U , r N u N N log h τ c i τ c , i τ ,
where we have used that as n , the quantities 2 K τ , β τ c Ξ τ c , and β τ Ξ τ 1 + ρ are subexponential in the blocklength n and do not contribute to the exponent, accordingly removed r ¯ τ c u τ c τ c and r ¯ τ u τ τ from the optimization parameter list, and finally used that the exponential decay of the error probability in Equation (A2) will be dominated by the worst error type τ and the worst classes assignment i N . It will prove convenient to the express the exponent in terms of a Gallager function E τ i N ρ , λ N L , U , r N u N N , defined as
E τ i N ρ , λ N L , U , r N u N N = log h τ c i τ c , i τ .
Substituted the expression for h τ c i τ c , i τ in Equation (A40), where Λ τ i τ ( u τ ) and R ν , u ν i ν ( x ν ) are respectively given by Equations (A25) and (A34), we may express E τ i N ρ , λ N L , U , r N u N N as
E τ i N ( ρ , λ N L , U , r N u N N ) = log u τ c , x τ c , y P τ c ( u τ c ) Λ τ c i τ c ( u τ c ) Q τ c , u τ c i τ c ( x τ c ) R τ c , u τ c i τ c ( x τ c ) = log ( u τ , x τ P τ | τ c ( u τ | u τ c ) 1 1 + ρ Λ τ i τ ( u τ ) Q τ , u τ i τ ( x τ ) R τ , u τ i τ ( x τ ) W ( y | x N ) 1 1 + ρ 1 + ρ ,
or equivalently in the alternative form
E τ i N ρ , λ N L , U , r N u N N = log u τ c , x τ c , y ( u τ , x τ P N ( u N ) 1 1 + ρ Λ N i N ( u N ) Q τ , u τ i τ ( x τ ) R N , u N i N ( x N ) Q τ c , u τ c i τ c ( x τ c ) W ( y | x N ) 1 1 + ρ 1 + ρ ,
where in Equation (A46) we have moved the product
P τ c ( u τ c ) Λ τ c i τ c ( u τ c ) Q τ c , u τ c i τ c ( x τ c ) R τ c , u τ c i τ c ( x τ c )
inside the parenthesis and merged terms in τ and τ c as done above, as well as redefined the optimization parameters λ τ c L 1 + ρ , λ τ c L 1 + ρ , and r τ c u τ c τ c 1 + ρ as λ τ c L , λ τ c U , and r τ c u τ c τ c respectively.

Appendix A.1. Proof of Equation (A18)

A sketch of the proof of the inequality in Equation (A18) proceeds as follows:
i , j K A i s i A j 1 s i i , j K s i A i + ( 1 s i ) A j
i , j K ( A i + A j )
= 2 | K | i K A i
where Equation (A47) follows from the inequality between arithmetic and geometric means and in Equation (A48) we used that 0 s 1 .

Appendix A.2. Proof of Equation (A30)

We have the following
1 { | a | δ } = 1 { δ a } 1 { a δ }
= 1 { e δ e a } 1 { e a e + δ }
e r ( a + δ ) e r + ( δ a )
= e r a + r ¯ δ
for r , r + 0 or equivalently r = r r + R and r ¯ = r + r + 0 . The bound in Equation (A53) can be optimized w. r. t. r and r ^ .

Appendix B. Computation of the Optimum Multi-Class Thresholds

In this section we find some conditions describing the optimum partitioning of the source-message set into classes for the optimization of the exponent in Equation (30). For simplicity, let each user ν N have two classes, K ν = 2 .
From the class definition in Equation (12) with K ν = 2 , we have that γ ν , 2 = 0 and γ ν , 0 = 1 , so we need find just one optimum γ ν , 1 for each user, which redefine as γ ν . Optimizing the exponent in Equation (30) over γ N gives
max r N u N N 0 γ N 1 E cost = max r N u N N 0 γ N 1 min r N u N N i N min r N u N N τ max ρ , λ N L , U , r N u N N E τ i N ρ , λ N L , U , r N u N N ,
where one of the parameters λ N L or λ N U is zero for each i N , as the corresponding constraint is absent. For each γ N , we have a minimization over 2 N assignments i N . Following the same steps as in refs. [31] (Sec. 4.1.2) and [31] (Lemma 4.3), we find that E τ i N ( γ N ) defined, with some abuse of notation, as
E τ i N ( γ N ) = max ρ , λ N L , U , r N u N N E τ i N ρ , λ N L , U , r N u N N ,
is a non-decreasing (resp. non-increasing) function with respect to γ ν for i N = [ i ν , i ν c ] with i ν = 1 (resp. i ν = 2 ), irrespective of the values of i ν c and of ν . For the sake of completeness, we present an independent proof of this fact here. Let i ν = 1 and τ be arbitrary. Using Equation (31), the function E τ i N ( γ N , ρ ) has the form log z f 1 ( z ) \ γ ν λ ν L ) for some function f 1 ( z ) , as all γ N are independent from each other, regardless the value of i ν c . Since λ ν L 0 , the function E τ i N ( γ N , ρ ) in Equation (A55) is non-decreasing with respect to γ ν . When i ν = 2 , this function E τ i N ( γ N , ρ ) has the form log z f 2 ( z ) γ ν λ ν U ) for some f 2 ( z ) , and is therefore non-increasing. This behavior will not change after taking maximization over ρ . As the minimum of monotonic functions is monotonic, the function E τ i N ( γ N ) is non-decreasing (non-increasing) with respect to γ ν , when i ν = 1 ( i ν = 2 ).
For any ν and fixed γ ν c , we may write the optimization problem in Equation (A54) as
max γ ν c max γ ν min i ν min i ν c min τ E τ [ i ν , i ν c ] [ γ ν , γ ν c ] .
The optimization problem max γ ν min i ν min i ν c min τ E τ [ i ν , i ν c ] [ γ ν , γ ν c ] satisfies the following lemma, proved in Appendix B.1, with γ = γ ν , i = i ν , and k i ( γ ) = min i ν c min τ E τ [ i ν , i ν c ] [ γ ν , γ ν c ] .
Lemma A1.
Let k 1 ( γ ) and k 2 ( γ ) be respectively continuous non-decreasing and non-increasing functions with respect to γ [ 0 , 1 ] . The optimal γ maximizing min i = 1 , 2 k i ( γ ) satisfies the following equation
k 1 ( γ ) = k 2 ( γ ) .
When Equation (A57) does not have any solution, we have γ = 0 if k 1 ( 0 ) > k 2 ( 0 ) , and γ = 1 otherwise.
Therefore, the optimal γ ν satisfies
min i ν c min τ E τ [ 1 , i ν c ] [ γ ν * , γ ν c ] = min i ν c min τ E τ [ 2 , i ν c ] [ γ ν * , γ ν c ] ,
if Equation (A58) has a solution. If not, the inequality min i ν c min τ E τ [ 1 , i ν c ] [ 0 , γ ν c ] > min i ν c min τ E τ [ 1 , i ν c ] [ 0 , γ ν c ] is satisfied, we have γ ν = 0 or γ ν = 1 otherwise. Since Equation (A58) holds for any ν , evaluating it for each ν gives a system of equations for the computation of the optimal thresholds.
In ref. [31] (Sec. 3.2.1.1), we give a graphical interpretation of the solutions to Equation (A58) and outline the relevant differences with the single-user case. We observe a strong coupling between the exponent and the thresholds that prevents to find the optimal number of classes, suggesting that, unlike the single-user case, two classes might not be sufficient.

Proof of Lemma A1

The relative behaviour of a non-decreasing function with a non-increasing function can be categorized in three cases.
  • If k 1 ( 0 ) < k 2 ( 0 ) and k 1 ( 1 ) > k 2 ( 1 ) , there exists a γ such that k 1 ( γ ) = k 2 ( γ ) . In this case, the function min i k i ( γ ) is non-decreasing from [ 0 , γ ) , and non-increasing from ( γ , 1 ] . Thus, the maximum over γ of min i k i ( γ ) occurs at γ = γ .
  • If k 1 ( 0 ) < k 2 ( 0 ) and k 1 ( 1 ) < k 2 ( 1 ) , k 1 ( γ ) and k 2 ( γ ) do not cross in γ [ 0 , 1 ] . Hence, we have min i k i ( γ ) = k 1 ( γ ) and obviously since it is an non-decreasing function the maximum over γ occurs at γ = γ = 1 .
  • When k 1 ( 0 ) k 2 ( 0 ) , we have min i k i ( γ ) = k 2 ( γ ) and hence γ = 0 .

Appendix C. Properties of the Modified Gallager Source Function

In this appendix, we study the modified Gallager source function E s , τ i N in Equation (45) involved in the achievable exponent for the md-iid ensemble. For the sake of simplicity, we consider the rather illustrative case of N = 2 users, each having a K ν = 2 -class partition of the source messages with thresholds i N where N = { 1 , 2 } . From the definition of the sets A ν i ν in Equation (12) with γ ν , 0 = 1 , γ ν , 1 = γ ν and γ ν , 2 = 0 , the two message sets
A ν 1 ( γ ν ) = u ν U ν n : P ν ( u ν ) γ ν n ,
A ν 2 ( γ ν ) = u ν U ν n : P ν ( u ν ) < γ ν n ,
are specified using a single threshold γ ν for each user ν { 1 , 2 } . With some abuse of notation, we include the optimization w. r. t. λ N L , U and make explicit the dependence on the thresholds γ N in the expression of the source function E s , τ i N in Equation (45), namely
E s , τ i N ( ρ , P N , γ N ) = min λ N L , U 0 log u τ c u τ P N ( u N ) 1 1 + ρ Λ N i N ( u N ) 1 + ρ .
For i ν = 1 , the set A ν 1 ( γ ν ) in Equation (A59) has no upper threshold, hence we find that the optimal parameter λ ν U in this case is λ ^ ν U = 0 . Similarly for i ν = 2 , we obtain that λ ^ ν L = 0 . As a consequence and without any loss of generality, we define λ ν = λ ν L for i ν = 1 , and λ ν = λ ν U for i ν = 2 , and further simplify Equation (A61) to the following optimization problem
E s , τ i N ( ρ , P N , γ N ) = min λ N 0 log u τ c u τ P N ( u N ) 1 1 + ρ γ 1 P 1 ( u 1 ) ( 1 ) i 1 λ 1 γ 2 P 2 ( u 2 ) ( 1 ) i 2 λ 2 1 + ρ ,
where we also used the definition of the functions Λ σ i σ in Equation (32) with σ = { 1 , 2 } . We recall that P 1 and P 2 are the marginal distributions for users ν = 1 and ν = 2 , respectively, and the indices i ν { 1 , 2 } indicate that user ν transmits a source message selected from the class A ν i ν ( γ ν ) in Equations (A59) and (A60). It can be shown that the objective function in the r. h. s. of Equation (A62) is convex w. r. t. both λ 1 and λ 2 . Hence, the minimizers λ ^ 1 and λ ^ 2 in the source function E s , τ i N ( ρ , P N , γ N ) are respectively given by λ ^ 1 = max { λ 1 , 0 } and λ ^ 2 = max { λ 2 , 0 } , where λ 1 and λ 2 are the unique solution after setting the partial derivatives of the r. h. s. of Equation (A62) to zero. Two special cases can be obtained from Equation (A62).
The first case is when γ ν = 1 for ν { 1 , 2 } , implying that no message partition happens whatsoever. In such a case, we have that λ ^ 1 = λ ^ 2 = 0 and Equation (A62) reduces to the joint source-channel coding source function for correlated-sources in Equation (34), i.e.,
E s , τ ( ρ , P N ) = log u τ c u τ P N u N 1 1 + ρ 1 + ρ .
The second one is the case of independent sources. Substituting P N = P 1 P 2 in Equation (A62), after some algebra, we obtain that E s , τ i N ( ρ , P N , γ N ) can be split into two terms as
E s , τ i N ( ρ , P N , γ N ) = E s i τ ( ρ , P τ , γ τ ) + E s i τ c ( 0 , P τ c , γ τ c ) ,
where we defined the function E s i ( ρ , P , γ ) as
E s i ( ρ , P , γ ) = min λ 0 log u P ( u ) 1 1 + ρ γ P ( u ) ( 1 ) i λ 1 + ρ
for arbitrary class index i { 1 , 2 } , source distribution P and threshold γ . First, we find that the unique solution after setting the derivative of the r. h. s. of Equation (A65) to zero, denoted as λ , is implicitly given by
u P ( u ) 1 1 + α log P ( u ) u P ( u ) 1 1 + α = log ( γ ) ,
where we made the convenient change of variable
1 1 + α = 1 1 + ρ ( 1 ) i λ .
Although not made explicit, λ depends on the triplet ( i , P , ρ , γ ) . When λ < 0 , or equivalently when
( 1 ) i 1 1 + ρ 1 1 + α < 0
we have that λ ^ = max ( 0 , λ ) = 0 , implying that Equation (A65) simplifies to
E s i ( ρ , P , γ ) = E s ( ρ , P ) ,
where E s ( ρ , P ) is the Gallager source function
E s ( ρ , P ) = log u P ( u ) 1 1 + ρ 1 + ρ .
Otherwise, when λ ^ = λ 0 , a regime given by the following inequality
( 1 ) i 1 1 + ρ 1 1 + α 0
we may substitute λ = λ in the objective function in Equation (A65) to obtain
E s i ( ρ , P , γ ) = ( 1 + ρ ) log u P ( u ) 1 1 + α + α ρ 1 + α log ( γ ) ,
where we wrote the expression in terms of α . Using Equation (A66) into Equation (A72) to replace log ( γ ) , we get
E s i ( ρ , P , γ ) = ( 1 + ρ ) log u P ( u ) 1 1 + α + α ρ 1 + α u P ( u ) 1 1 + α log P ( u ) u P ( u ) 1 1 + α .
After some algebra, we are able to express the former equation in terms of the derivative of the E s -function in Equation (A70), given by
E s ( ρ , P ) = log u P ( u ) 1 1 + ρ 1 1 + ρ u P ( u ) 1 1 + ρ log P ( u ) u P ( u ) 1 1 + ρ ,
and the E s -function itself, as
E s i ( ρ , P , γ ) = E s ( α , P ) + ( ρ α ) E s ( α , P ) .
We may finally combine Equations (A69) and (A75), with the respective ranges in Equations (A68) and (A71) to write the E s i ( ρ , P , γ ) function in Equation (A65) piecewise as
E s 1 ( ρ , P , γ ) = E s ( ρ , P ) 1 1 + ρ 1 1 + α , E s ( α , P ) + E s ( α ) ( ρ α ) 1 1 + ρ < 1 1 + α ,
and
E s 2 ( ρ , P , γ ) = E s ( ρ , P ) 1 1 + ρ < 1 1 + α , E s ( α , P ) + E s ( α ) ( ρ α ) 1 1 + ρ 1 1 + α .
where α is the solution to the implicit Equation (A66), hence recovering the source error exponent functions of the md-iid ensemle described in ref. [8] (Lemma 1). The source functions E s 1 ( ρ , P , γ ) and E s 2 ( ρ , P , γ ) follow the Gallager function in Equation (A70) for a certain interval of ρ , and are the straight-line tangent beyond that interval. The tangent point α is a function of the distribution P and of the multi-class threshold γ .
Once E s i ( ρ , P , γ ) in Equation (A65) is fully characterized, we may now discuss the correlated-sources error function E s , τ i N ( ρ , P N , γ N ) in Equation (A64) in terms of the error type τ . We start with the third error type τ = { 1 , 2 } , for which since τ c = , we have that
E s , τ i N ( ρ , P N , γ N ) = E s i 1 ( ρ , P 1 , γ 1 ) + E s i 2 ( ρ , P 2 , γ 2 ) ,
namely the superposition of two E s i functions as the ones in Equations (A76) and (A77), one for each user. For the remaining of this appendix, we consider the more informative error types τ = { 1 } and τ = { 2 } for the four possible pairs of class indices i 1 and i 2 in Equations (A59) and (A60), since in this case E s , τ i N in Equation (A64) is either directly an E s ( ρ , P τ ) function or the straight-line tangent to it, in both cases shifted by a constant term given by E s i τ c ( 0 , P τ c , γ τ c ) .
Figure A1 shows the family of E s , τ i N source functions respectively for independent and correlated sources, as a function of ρ where P N given by Equation (60) and τ = { 1 } . For independent sources, we observe that the source functions E s , τ 1 , 1 and E s , τ 2 , 1 follow the solid blue line depicting E s ( ρ , P τ ) as in Equation (A70) for a certain interval of ρ , and then take the tangent line beyond. A similar behavior is observed for the sources functions E s , τ 1 , 2 and E s , τ 2 , 2 , which in this case follow or are tangent to the solid black line, the solid blue Gallager’s source function shifted by the constant function E s i τ c ( 0 , P τ c , γ τ c ) as in Equation (A64).
For correlated sources, the source functions E s , τ 1 , 1 and E s , τ 2 , 1 follow the generalized Gallager’s source function given by Equation (A63) for a certain interval, but unlike independent sources they are not straight lines but a curve tangent to E s , τ beyond that interval. Some intuition about this fact can be gained from the primal form of the source function E s , τ i N . Consider, for instance, the source function E s , τ 2 , 1 in Figure A1 for correlated sources, for which i 1 = 2 and i 2 = 1 . The primal form of this source function E s , τ 2 , 1 can be obtained as a constrained optimization problem w. r. t. some auxiliary joint distribution P ^ N . The interval in ρ where E s , τ 2 , 1 does not follow E s , τ in the dual form (approximately for ρ 0.5 in the figure) corresponds to the case where only one of the two constraints on the auxiliary distribution P ^ N is actually active in the primal form, where the constraint is given by u N P ^ N ( u N ) log P ν ( u ν ) = log ( γ ν ) . This implies that, unlike the case of independent sources where each source has its auxiliary distribution P ^ 1 and P ^ 2 constrained, for correlated sources the joint auxiliary distribution P ^ N is not fully constrained but is the union of joint distributions with one constrained marginal distribution. This partial constraint manifests as a curve in ρ , rather than a straight line, in the dual form. A similar behavior is observed for E s , τ 1 , 2 and E s , τ 2 , 2 , which instead of following the source function for joint source-channel coding in Equation (A63) for some intervals of ρ , they follow the curve
min λ ν 0 log u τ c u τ P N ( u N ) 1 1 + ρ P ν ( u ν ) γ ν ( 1 ) i ν λ ν 1 + ρ 1 + ρ ,
corresponding to Equation (A62) when the constraint for one source is not active, i.e., λ ^ ν c = 0 .
Figure A1. Example of the source functions E s , τ i N in Equation (A62) for independent and correlated sources and error type τ { 1 } , { 2 } .
Figure A1. Example of the source functions E s , τ i N in Equation (A62) for independent and correlated sources and error type τ { 1 } , { 2 } .
Entropy 23 00569 g0a1

References

  1. Gallager, R. Information Theory and Reliable Communication; John Wiley & Sons: Hoboken, NJ, USA, 1968. [Google Scholar]
  2. Polyanskiy, Y.; Poor, H.V.; Verdú, S. Channel coding rate in the finite blocklength regime. IEEE Trans. Inf. Theory 2010, 56, 2307–2359. [Google Scholar] [CrossRef]
  3. Jelinek, F. Probabilistic Information Theory-Discrete and Memoryless Models; McGraw-Hill Book Company: New York, NY, USA, 1968. [Google Scholar]
  4. Shannon, C.; Gallager, R.; Berlekamp, E. Lower bounds to error probability for coding on discrete memoryless channels. I. Inf. Control 1967, 10, 65–103. [Google Scholar] [CrossRef] [Green Version]
  5. Shannon, C.; Gallager, R.; Berlekamp, E. Lower bounds to error probability for coding on discrete memoryless channels. II. Inf. Control 1967, 10, 522–552. [Google Scholar] [CrossRef] [Green Version]
  6. Csiszár, I. Joint source-channel error exponent. Probl. Control Inf. Theory 1980, 9, 315–328. [Google Scholar]
  7. Zhong, Y.; Alajaji, F.; Campbell, L.L. On the joint source-channel coding error exponent for discrete memoryless systems. IEEE Trans. Inf. Theory 2006, 52, 1450–1468. [Google Scholar] [CrossRef] [Green Version]
  8. Tauste Campo, A.; Vazquez-Vilar, G.; Guillén i Fàbregas, A.; Martinez, A.; Koch, T. A derivation of the source-channel error exponent using nonidentical product distributions. IEEE Trans. Inf. Theory 2014, 60, 3209–3217. [Google Scholar] [CrossRef] [Green Version]
  9. Ahlswede, R. Multi-way communication channels. In Proceedings of the 2nd International Symposium on Information Theory, Tsaghkadzor, Armenia, USSR, 2–8 September 1971. [Google Scholar]
  10. Slepian, D.; Wolf, J.K. A coding theorem for multiple access channel with correlated sources. Bell Syst. Tech. J. 1973, 52, 1037–1076. [Google Scholar] [CrossRef]
  11. Cover, T.M.; El Gamal, A.; Salehi, M. Multiple access channels with arbitrarily correlated sources. IEEE Trans. Inf. Theory 1980, 26, 648–657. [Google Scholar] [CrossRef] [Green Version]
  12. Gallager, R. A perspective on multiple access channels. IEEE Trans. Inf. Theory 1985, 31, 124–142. [Google Scholar] [CrossRef]
  13. Liu, Y.S.; Hughes, B. A new universal random coding bound for the multiple-access channel. IEEE Trans. Inf. Theory 1996, 42, 376–386. [Google Scholar]
  14. Lapidoth, A. Mismatched decoding and the multiple-access channel. IEEE Trans. Inf. Theory 1996, 42, 1439–1452. [Google Scholar] [CrossRef]
  15. Haim, E.; Kochman, Y.; Erez, U. Improving the MAC error exponent using distributed structure. In Proceedings of the 2011 IEEE International Symposium on Information Theory (ISIT), Saint Petersburg, Russia, 31 July–5 August 2011. [Google Scholar]
  16. Cai, N. The maximum error probability criterion, random encoder, and feedback, in multiple input channels. Entropy 2014, 16, 1211–1242. [Google Scholar] [CrossRef] [Green Version]
  17. Nazari, A.; Anastasopoulos, A.; Pradhan, S.S. Error exponent for multiple-access channels: Lower bounds. IEEE Trans. Inf. Theory 2014, 60, 5095–5115. [Google Scholar] [CrossRef] [Green Version]
  18. Nazari, A.; Pradhan, S.S.; Anastasopoulos, A. Error exponent for multiple access channels: Upper bounds. IEEE Trans. Inf. Theory 2015, 61, 3605–3621. [Google Scholar] [CrossRef] [Green Version]
  19. Farkas, L.; Kói, T. Random access and source-channel coding error exponents for multiple access channels. IEEE Trans. Inf. Theory 2015, 61, 3029–3040. [Google Scholar] [CrossRef] [Green Version]
  20. Rezazadeh, A.; Font-Segura, J.; Martinez, A.; Guillén i Fàbregas, A. Multiple-Access Channel with Independent Sources: Error Exponent Analysis. In Proceedings of the 2018 IEEE Information Theory Workshop (ITW), Guangzhou, China, 25–29 November 2018. [Google Scholar]
  21. Dueck, G. A note on the multiple access channel with correlated sources (Corresp.). IEEE Trans. Inf. Theory 1981, 27, 232–235. [Google Scholar] [CrossRef]
  22. Padakandla, A. Communicating correlated sources over MAC and interference channels II: Joint source-channel coding. IEEE Trans. Inf. Theory 2021. [Google Scholar] [CrossRef]
  23. Padakandla, A. Communicating correlated sources over a MAC. In Proceedings of the 2017 IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, 25–30 June 2017. [Google Scholar]
  24. Heidari, M.; Shirani, F.; Pradhan, S.S. New sufficient conditions for multiple-access channel with correlated sources. In Proceedings of the 2016 IEEE International Symposium on Information Theory (ISIT), Barcelona, Spain, 10–15 July 2016. [Google Scholar]
  25. Rezazadeh, A.; Font-Segura, J.; Martinez, A.; Guillén i Fàbregas, A. An achievable error exponent for the multiple access channel with correlated sources. In Proceedings of the 2017 IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, 25–30 June 2017. [Google Scholar]
  26. Scarlett, J.; Martinez, A.; Guillén i Fàbregas, A. Mismatched decoding: Error exponents, second-order rates and saddlepoint approximations. IEEE Trans. Inf. Theory 2014, 60, 2647–2666. [Google Scholar] [CrossRef] [Green Version]
  27. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. Journal 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  28. Rezazadeh, A.; Font-Segura, J.; Martinez, A.; Guillén i Fàbregas, A. Joint source-channel coding for the multiple-access channel with correlated sources. In Proceedings of the 2019 International Symposium on Information Theory (ISIT), Paris, France, 7–12 July 2019. [Google Scholar]
  29. Gallager, R. Fixed Composition Arguments and Lower Bounds to Error Probability. Available online: http://web.mit.edu/gallager/www/notes/notes5.pdf (accessed on 3 May 2021).
  30. Csiszár, I.; Körner, J. Information Theory: Coding Theorems for Discrete Memoryless Systems; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  31. Rezazadeh, A. Error Exponent Analysis for the Multiple-Access Channel with Correlated Sources. Ph.D. Thesis, Universitat Pompeu Fabra, Barcelona, Spain, 2019. [Google Scholar]
  32. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  33. Hardy, G.H.; Littlewood, J.E.; Pólya, G. Inequalities, 2nd ed.; Cambridge University Press: Cambridge, UK, 1934. [Google Scholar]
Figure 1. Example of codewords x ν ( 1 ) , x ν ( 2 ) and x ν ( 3 ) in the mi-icd ensemble and x ν ( 4 ) , x ν ( 5 ) and x ν ( 6 ) in the mi-ccc ensemble, for a given source sequence u ν .
Figure 1. Example of codewords x ν ( 1 ) , x ν ( 2 ) and x ν ( 3 ) in the mi-icd ensemble and x ν ( 4 ) , x ν ( 5 ) and x ν ( 6 ) in the mi-ccc ensemble, for a given source sequence u ν .
Entropy 23 00569 g001
Figure 2. Correlated-sources optimal assignment Ω ( γ N ) in Equation (70) for all pairs of thresholds ( γ 1 , γ 2 ) .
Figure 2. Correlated-sources optimal assignment Ω ( γ N ) in Equation (70) for all pairs of thresholds ( γ 1 , γ 2 ) .
Entropy 23 00569 g002
Figure 3. Correlated-sources error exponent E c o s t ( γ N ) in Equation (71) for all pairs of thresholds ( γ 1 , γ 2 ) .
Figure 3. Correlated-sources error exponent E c o s t ( γ N ) in Equation (71) for all pairs of thresholds ( γ 1 , γ 2 ) .
Entropy 23 00569 g003
Table 1. Correlated-sources optimal thresholds γ N and exponents E τ i N ( γ N ) in Equation (73) for assignments Ω j in Equations (66)–(69). For each assignment, the minimum over i N and τ is highlighted in gray.
Table 1. Correlated-sources optimal thresholds γ N and exponents E τ i N ( γ N ) in Equation (73) for assignments Ω j in Equations (66)–(69). For each assignment, the minimum over i N and τ is highlighted in gray.
Assignment Ω 1 Assignment Ω 2
γ 1 = 0.8469 γ 2 = 0.6581 γ 1 = 1 γ 2 = 1
( i 1 , i 2 ) ( 1 , 1 ) ( 1 , 2 ) ( 2 , 1 ) ( 2 , 2 ) ( 1 , 1 ) ( 1 , 2 ) ( 2 , 1 ) ( 2 , 2 )
τ = { 1 } 0.31310.27350.31200.2611 0.06420.32680.10050.3604
τ = { 2 } 0.39860.43690.26110.4119 0.39590.39860.43230.3110
τ = { 1 , 2 } 0.26110.29720.26300.2883 0.21080.21080.23600.2637
Assignment Ω 3 Assignment Ω 4
γ 1 = 0.5605 γ 2 = 0.6709 γ 1 = 0.6985 γ 2 = 0.9033
( i 1 , i 2 ) ( 1 , 1 ) ( 1 , 2 ) ( 2 , 1 ) ( 2 , 2 ) ( 1 , 1 ) ( 1 , 2 ) ( 2 , 1 ) ( 2 , 2 )
τ = { 1 } 0.31200.25030.27630.2897 0.08790.36050.08790.3112
τ = { 2 } 0.25030.38980.56750.5731 0.36640.25030.47200.4684
τ = { 1 , 2 } 0.26300.28160.25030.3012 0.23600.26320.20970.2097
Table 2. Independent-sources md-iid optimal thresholds γ N and exponents E τ i N ( γ N ) in Equation (73) for assignments Ω j in Equations (66)–(69). For each assignment, the minimum over i N and τ is highlighted in gray.
Table 2. Independent-sources md-iid optimal thresholds γ N and exponents E τ i N ( γ N ) in Equation (73) for assignments Ω j in Equations (66)–(69). For each assignment, the minimum over i N and τ is highlighted in gray.
Assignment Ω 1 Assignment Ω 2
γ 1 = 0.8779 γ 2 = 0.6933 γ 1 = 0.8776 γ 2 = 1
( i 1 , i 2 ) ( 1 , 1 ) ( 1 , 2 ) ( 2 , 1 ) ( 2 , 2 ) ( 1 , 1 ) ( 1 , 2 ) ( 2 , 1 ) ( 2 , 2 )
τ = { 1 } 0.33430.24580.30890.2458 0.09130.33410.09130.3089
τ = { 2 } 0.38500.39870.24580.3788 0.45550.38500.43570.2459
τ = { 1 , 2 } 0.27300.28700.26850.2863 0.34300.27280.29560.2685
Assignment Ω 3 Assignment Ω 4
γ 1 = 0.61 γ 2 = 0.7043 γ 1 = 0.7092 γ 2 = 1
( i 1 , i 2 ) ( 1 , 1 ) ( 1 , 2 ) ( 2 , 1 ) ( 2 , 2 ) ( 1 , 1 ) ( 1 , 2 ) ( 2 , 1 ) ( 2 , 2 )
τ = { 1 } 0.30890.23670.26810.3078 0.09130.31170.09130.2648
τ = { 2 } 0.23670.36720.54250.5538 0.42690.23670.53930.4683
τ = { 1 , 2 } 0.26850.28110.23670.3133 0.30060.26850.27400.2164
Table 3. Mi-iid exponents E τ in Equation (75) for two correlated and two independent sources vs several input distribution assigments ( Q 1 , Q 2 ) . For each assignment, the minimum over τ is highlighted in gray.
Table 3. Mi-iid exponents E τ in Equation (75) for two correlated and two independent sources vs several input distribution assigments ( Q 1 , Q 2 ) . For each assignment, the minimum over τ is highlighted in gray.
Correlated Sources
( Q 1 , Q 2 ) ( Q , Q ) ( Q , Q ) ( Q , Q ) ( Q , Q )
τ = { 1 } 0.26820.06420.31200.0879
τ = { 2 } 0.39860.39860.25030.3696
τ = { 1 , 2 } 0.20970.20970.26300.2360
Independent Sources
( Q 1 , Q 2 ) ( Q , Q ) ( Q , Q ) ( Q , Q ) ( Q , Q )
τ = { 1 } 0.26480.30890.06270.0865
τ = { 2 } 0.38500.23670.38500.3559
τ = { 1 , 2 } 0.21640.26850.21640.2421
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rezazadeh, A.; Font-Segura, J.; Martinez, A.; Guillén i Fàbregas, A. Multi-Class Cost-Constrained Random Coding for Correlated Sources over the Multiple-Access Channel. Entropy 2021, 23, 569. https://doi.org/10.3390/e23050569

AMA Style

Rezazadeh A, Font-Segura J, Martinez A, Guillén i Fàbregas A. Multi-Class Cost-Constrained Random Coding for Correlated Sources over the Multiple-Access Channel. Entropy. 2021; 23(5):569. https://doi.org/10.3390/e23050569

Chicago/Turabian Style

Rezazadeh, Arezou, Josep Font-Segura, Alfonso Martinez, and Albert Guillén i Fàbregas. 2021. "Multi-Class Cost-Constrained Random Coding for Correlated Sources over the Multiple-Access Channel" Entropy 23, no. 5: 569. https://doi.org/10.3390/e23050569

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop