hybrid noncoherent network coding

15
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 59, NO. 6, JUNE 2013 3317 Hybrid Noncoherent Network Coding Vitaly Skachek, Olgica Milenkovic, Member, IEEE, and Angelia Nedić, Member, IEEE Abstract—We describe a novel extension of subspace codes for noncoherent networks, suitable for use when the network is viewed as a communication system that introduces both dimension and symbol errors. We show that when symbol erasures occur in a signicantly large number of different basis vectors transmitted through the network and when the min-cut of the network is much smaller than the length of the transmitted codewords, the new family of codes outperforms their subspace code counterparts. For the proposed coding scheme, termed hybrid network coding, we derive two upper bounds on the size of the codes. These bounds represent a variation of the Singleton and of the sphere-packing bound. We show that a simple concatenated scheme that consists of subspace codes and Reed–Solomon codes is asymptotically optimal with respect to the Singleton bound. Finally, we describe two efcient decoding algorithms for concatenated subspace codes that in certain cases have smaller complexity than their subspace decoder counterparts. Index Terms—Error correction, noncoherent network coding, subspace codes, symbol errors. I. INTRODUCTION N ETWORK coding is a scheme introduced by Ahlswede et al. [1] for efcient communication over networks with transmission bottlenecks. The authors of [1] showed that under a broadcast scenario in networks, the maximal theoretically achievable communication rate—called the capacity of the network—can be characterized by minimal cuts in the network and achieved by appropriate coding methods. In the last decade, network coding became a focal point of re- search in coding theory. There exists a variety of network coding solutions currently used in practice: random network coding ap- proach was rst proposed in [9]; algebraic coding was shown to achieve the capacity of a class of networks in [15] and [16]; non- linear approaches were also studied in [3]. Manuscript received August 02, 2011; revised September 23, 2012; accepted November 23, 2012. Date of publication January 30, 2013; date of current ver- sion May 15, 2013. This work was supported in part by the Air Force Ofce for Scientic Research Grant AF FA9550-09-1-0612 and in part by the Na- tional Science Foundation Grants CCF 0939370 and CCF 1117980. This paper was presented in part at the 2012 IEEE International Symposium on Network Coding. V. Skachek was with the Coordinated Science Laboratory, University of Illi- nois at Urbana-Champaign, Urbana, IL 61801 USA. He is now with the Institute of Computer Science, Faculty of Mathematics and Computer Science, Univer- sity of Tartu, Tartu 50409, Estonia (e-mail: [email protected]). O. Milenkovic and A. Nedić are with the Coordinated Science Laboratory, University of Illinois at Urbana-Champaign, Urbana, IL 61801 USA (e-mail: [email protected]; [email protected]). Communicated by C. Fragouli, Associate Editor for Communication Net- works. Digital Object Identier 10.1109/TIT.2013.2243899 The use of network coding for error-correction was rst proposed in [2]. When the network topology is not known, or when it changes with time, it was suggested in [14] to use sub- space coding for joint error-correction and network coding, and suitable codes were constructed therein. Subspace codes are closely related to rank-metric codes, also extensively studied in the coding literature [6], [19], [8]. The parameters of the codes in [14] were further improved in a series of subsequent papers, including [4], [7], [13], [17], [22], and [23]. Bounds on the parameters of subspace codes were derived in [5] and [24]. It should also be mentioned that the subspace codes, which were proposed in [14] for noncoherent network coding, were studied independently in the area of cryptography under the name authentication codes [26]. In our work, we follow the line of research started in [14]. More specically, we consider error-correction for a special case of network coding, suitable for practical applications in which the topology of the network is not known or changes with time. This type of scheme is, as already mentioned, known as coding for noncoherent networks. Currently, the only known approach for noncoherent network coding utilizes subspace codes. Subspace codes for noncoherent network coding are based on the idea that the transmitted data vectors can be associated with linear vector subspaces. Linear network coding does not change the information about the subspaces, since it only allows for linear combining of the transmitted basis vectors. Hence, if there are no errors, the receiver obtains uncompromised in- formation regarding the transmitted subspace. The transmitted subspaces can only be modied within the network through the introduction of errors. In order to protect the transmitted infor- mation, one has to add carefully structured redundancy into the subspace messages. In the context of the work [14], the errors are modeled as di- mension gains and dimension losses. These notions, although of theoretical value, may appear rather abstract in certain net- working applications, where packets (symbols or collections of symbols) are subjected to erasures or substitution errors. One fundamental question remains: how is one to interpret the no- tion of dimension gains and losses in terms of symbol errors and erasures, and what kind of errors and erasures constitute dimen- sion gains and losses? We propose a hybrid approach to noncoherent network coding, which attempts to connect the notions of dimension loss and gain with those of individual symbol errors and era- sures. The crux of our approach is to consider network coding where dimension gains and losses, in addition to individual symbol errors and erasures, are all possible. This allows us to study the tradeoffs between the required overhead in the network layer aimed at correcting dimension gains/losses, and 0018-9448/$31.00 © 2013 IEEE

Upload: angelia

Post on 09-Feb-2017

213 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Hybrid Noncoherent Network Coding

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 59, NO. 6, JUNE 2013 3317

Hybrid Noncoherent Network CodingVitaly Skachek, Olgica Milenkovic, Member, IEEE, and Angelia Nedić, Member, IEEE

Abstract—We describe a novel extension of subspace codesfor noncoherent networks, suitable for use when the network isviewed as a communication system that introduces both dimensionand symbol errors. We show that when symbol erasures occur ina significantly large number of different basis vectors transmittedthrough the network and when the min-cut of the network is muchsmaller than the length of the transmitted codewords, the newfamily of codes outperforms their subspace code counterparts.For the proposed coding scheme, termed hybrid network coding,we derive two upper bounds on the size of the codes. These boundsrepresent a variation of the Singleton and of the sphere-packingbound. We show that a simple concatenated scheme that consistsof subspace codes and Reed–Solomon codes is asymptoticallyoptimal with respect to the Singleton bound. Finally, we describetwo efficient decoding algorithms for concatenated subspace codesthat in certain cases have smaller complexity than their subspacedecoder counterparts.

Index Terms—Error correction, noncoherent network coding,subspace codes, symbol errors.

I. INTRODUCTION

N ETWORK coding is a scheme introduced by Ahlswedeet al. [1] for efficient communication over networks with

transmission bottlenecks. The authors of [1] showed that undera broadcast scenario in networks, the maximal theoreticallyachievable communication rate—called the capacity of thenetwork—can be characterized by minimal cuts in the networkand achieved by appropriate coding methods.In the last decade, network coding became a focal point of re-

search in coding theory. There exists a variety of network codingsolutions currently used in practice: random network coding ap-proach was first proposed in [9]; algebraic coding was shown toachieve the capacity of a class of networks in [15] and [16]; non-linear approaches were also studied in [3].

Manuscript received August 02, 2011; revised September 23, 2012; acceptedNovember 23, 2012. Date of publication January 30, 2013; date of current ver-sion May 15, 2013. This work was supported in part by the Air Force Officefor Scientific Research Grant AF FA9550-09-1-0612 and in part by the Na-tional Science Foundation Grants CCF 0939370 and CCF 1117980. This paperwas presented in part at the 2012 IEEE International Symposium on NetworkCoding.V. Skachek was with the Coordinated Science Laboratory, University of Illi-

nois at Urbana-Champaign, Urbana, IL 61801 USA. He is nowwith the Instituteof Computer Science, Faculty of Mathematics and Computer Science, Univer-sity of Tartu, Tartu 50409, Estonia (e-mail: [email protected]).O. Milenkovic and A. Nedić are with the Coordinated Science Laboratory,

University of Illinois at Urbana-Champaign, Urbana, IL 61801 USA (e-mail:[email protected]; [email protected]).Communicated by C. Fragouli, Associate Editor for Communication Net-

works.Digital Object Identifier 10.1109/TIT.2013.2243899

The use of network coding for error-correction was firstproposed in [2]. When the network topology is not known, orwhen it changes with time, it was suggested in [14] to use sub-space coding for joint error-correction and network coding, andsuitable codes were constructed therein. Subspace codes areclosely related to rank-metric codes, also extensively studiedin the coding literature [6], [19], [8]. The parameters of thecodes in [14] were further improved in a series of subsequentpapers, including [4], [7], [13], [17], [22], and [23]. Bounds onthe parameters of subspace codes were derived in [5] and [24].It should also be mentioned that the subspace codes, whichwere proposed in [14] for noncoherent network coding, werestudied independently in the area of cryptography under thename authentication codes [26].In our work, we follow the line of research started in [14].

More specifically, we consider error-correction for a specialcase of network coding, suitable for practical applications inwhich the topology of the network is not known or changeswith time. This type of scheme is, as already mentioned, knownas coding for noncoherent networks. Currently, the only knownapproach for noncoherent network coding utilizes subspacecodes.Subspace codes for noncoherent network coding are based

on the idea that the transmitted data vectors can be associatedwith linear vector subspaces. Linear network coding does notchange the information about the subspaces, since it only allowsfor linear combining of the transmitted basis vectors. Hence,if there are no errors, the receiver obtains uncompromised in-formation regarding the transmitted subspace. The transmittedsubspaces can only be modified within the network through theintroduction of errors. In order to protect the transmitted infor-mation, one has to add carefully structured redundancy into thesubspace messages.In the context of the work [14], the errors are modeled as di-

mension gains and dimension losses. These notions, althoughof theoretical value, may appear rather abstract in certain net-working applications, where packets (symbols or collections ofsymbols) are subjected to erasures or substitution errors. Onefundamental question remains: how is one to interpret the no-tion of dimension gains and losses in terms of symbol errors anderasures, and what kind of errors and erasures constitute dimen-sion gains and losses?We propose a hybrid approach to noncoherent network

coding, which attempts to connect the notions of dimensionloss and gain with those of individual symbol errors and era-sures. The crux of our approach is to consider network codingwhere dimension gains and losses, in addition to individualsymbol errors and erasures, are all possible. This allows usto study the tradeoffs between the required overhead in thenetwork layer aimed at correcting dimension gains/losses, and

0018-9448/$31.00 © 2013 IEEE

Page 2: Hybrid Noncoherent Network Coding

3318 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 59, NO. 6, JUNE 2013

the overhead in the physical layer designated to correctingsymbol erasures and errors.Our main result shows that by incorporating symbol error-

correcting mechanism into subspace codes, one can increasethe number of tolerable dimension gains and losses, withoutcompromising the network throughput. Hence, the proposed ap-proach leads to an increase in the overall number of correctableerrors in the subspace-based scheme akin to [14].In order to illustrate our approach, consider the following

straightforward example. Assume the case of a noncoherentlycoded network in which arbitrary (unknown) ten symbols areerased from the basis vectors representing the message. The firstquestion is how many dimension losses should be considered inthe model of [14]? One reasonable way to look at it is to assumethe worst case scenario where each symbol erasure introducesone dimension loss, and each error introduces a simultaneousdimension loss and gain. Consequently, ten symbol erasureswould amount to ten dimension losses. However, if there werean alternative way to correct some of these symbol erasures orerrors, the effective number of dimension losses and gains maybecome significantly smaller. In the example, correcting fivesymbol erasures would, in the best case, reduce the burden ofsubspace codes in terms of dimension loss recovery by five di-mensions. And, at least at first glance, correcting symbol errorsappears to be a task easier to accomplish than correcting dimen-sion errors.We, therefore, pose the following questions: what are the fun-

damental performance bounds for noncoherent network codingschemes, consequently termed hybrid network codes, capableof correcting symbol erasures and errors on one side, and di-mension gains and losses on the other side? What is the optimalrate allocation scheme for hybrid network codes with respect todimension losses/gains and symbol errors/erasures? What is theoptimal ratio between the two allocation rates and how can it beachieved practically? How does one efficiently correct errors inthis new scheme? The work in this paper is aimed at answeringthese questions.There are various potential applications for hybrid network

codes [11]. Hybrid codes can be useful in networks where nolink-layer error-correction is performed. Such networks includesensor networks for which the computational power of “inter-mediate nodes” is not sufficiently large. This prevents error-cor-rection to be performed before the errors propagate through thenetwork. Hybrid codes can also be used in networks for whicha physical layer packet is very small, the network layer packetconsists of many physical layer packets, and the packet may beregarded as a single symbol. In this case, if an error in the packetcannot be decoded, a symbol error is declared which is subse-quently “transferred” into a dimension loss/gain.We would also like to point out that upon publication of the

preliminary results on our hybrid network coding, two otherinteresting directions related to this model were proposed inthe literature. A concatenation of subspace codes and algebraiccodes was studied in the context of coding for distributed datastorage in [20]. More specifically, it was suggested in [20] touse a concatenation of Gabidulin codes [6] with classical max-imum distance separable code for correction of adversarial er-rors in distributed storage systems. In [27], the authors studied

symbol-level error-correction in random codes used over bothcoherent and noncoherent networks. However, in that work, noexplicit constructions of codes were presented.This paper is organized as follows. The notation and prior

work are discussed in Section II. In the sections that follow,we define hybrid codes that can be used for simultaneouscorrection of dimension losses/gains and symbol erasuresin noncoherent networks. More specifically, the basic coderequirements and parameters are presented and described inSection III. Two upper bounds on the size of hybrid codes, theSingleton bound and the sphere-packing bound, are presentedin Section IV. A straightforward concatenated code construc-tion appears in Section V-A. The analysis of code parametersand the comparison with known subspace code constructionsappear in Section V. The decoding algorithm for the proposedcodes is presented in Section VI. In Section VII, we show thatthe same codes can also be used for simultaneous correctionof dimension losses and symbol erasures/errors, and statesome results analogous to those in Sections III–VI. Finally, wediscuss some results related to simultaneous correction of bothdimension losses/gains and symbol erasures/errors.

II. NOTATION AND PRIOR WORK

Let be a vector space over a finite field , where isa power of a prime number. For a set of vectors , weuse to denote the linear span of the vectors in . We alsouse the notation for the linear span of the setof vectors . Let be the set of positive integernumbers. We write to denote the all-zero vector of length, for any . When the value of is clear from the

context, we sometimes write rather than . We also denote

by a unity vector which has a

one in position and zeros in all other positions. The lengthof the vector will be clear from the context.Let be linear subspaces of . We use the notation

for the dimension of . We denote the sum of twosubspaces and as .If , then for any , there is a uniquerepresentation in terms of the sum of two vectors ,where and . In this case, we say that isa direct sum, and denote it by . It is easy to check that

.Let . For , we define a projection of

onto , denoted by , as follows:

Similarly, we denote the projection of the vector onto by.

For two vectors, and , we write to denote their scalarproduct. Let and assume that1) for some ;2) For all , , it holds (i.e., andare orthogonal).

In that case, is uniquely defined by and . For everyvector , we have if

Page 3: Hybrid Noncoherent Network Coding

SKACHEK et al.: HYBRID NONCOHERENT NETWORK CODING 3319

. Then, we can define as the subspace ofobtained from by removing all zero entries from the vectorsin coordinates . In that case, we will write

. Observe that there is a natural bijection from the setof vectors in onto the set of vectors in , which is defined byremoving all zeros in coordinates . In the sequel,sometimes, we associate the vector in with its preimage inunder this bijection (or, simply speaking, sometimes we ignorethe zero coordinates as above). Thus, by slightlyabusing the notation, we may also write .Assume that . We use the notation for

the set of all subspaces of of dimension , and forthe set of all subspaces of of any dimension. The number of-dimensional subspaces of , , is given by the -aryGaussian coefficient (see [25, Ch. 24]):

For , let

be a distance measure between and in the Grassmanianmetric (see [14]). We use the notation for the Hammingdistance between two vectors and of the same length.We say that the code is an subspace

code, if it represents a set of subspaces in an ambient spaceover , and satisfies the following conditions:1) is a vector space over with ;2) for all , ;3) ;4) for all , , it holds that

, so that consequently .In [14], an subspace code wasconstructed by using an approach akin to Reed-Solomon (RS)codes. That code will henceforth be denoted by . We refer thereader to [14] for a detailed study of the code .To formalize the network model, the authors of [14] also

introduced the operator channel and erasure operator as fol-lows. Let be an integer. Given a subspace , if

, the stochastic erasure operator returnssome random -dimensional subspace of . Otherwise, it re-turns itself. Then, for any subspace in , it is always pos-sible to write , where is a realization of

, , and is a subspace of . In partic-ular, if , then is called a dimension loss.Similarly, if for some subspace ,then the corresponding operation is called a dimension gain.Decoding algorithms for the code were presented in [14]

and [21]. Suppose that is transmitted over the operatorchannel. Suppose also that an -dimensional subspaceof is received, where . Here,denotes the number of dimension losses when modifying

the subspace to , while similarly denotes the numberof dimension gains needed to transform into . Note that

, where is given in the decomposition of . The

decoders, presented in [14] and [21], are able to recover a singlewhenever . We denote hereafter a decoder

for the code described in [14] and [21] by . Note that thedecoding complexity of is polynomial both in the dimensionof the ambient vector space and the subspace distance .We find the following lemma useful in our subsequent deriva-

tions.Lemma II.1: Let be a vector space over ,

and let be two vector subspaces. Then

In other words, projections do not increase the subspace distance.Proof: By definition, we have

Let and . Taketo be a basis of and

to be linearly independent vectors in . Then,and jointly constitute a

basis of .Next, consider

Clearly, is a subspace of and of . There-fore

Note that the inclusion in the above relation may be strict.On the other hand, together with

spans . Wethus have that

(1)

Similarly to (1), it can be shown that

(2)From (1) and (2), we obtain that

This completes the proof of the claimed result.

Page 4: Hybrid Noncoherent Network Coding

3320 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 59, NO. 6, JUNE 2013

III. HYBRID CODING FOR SYMBOL ERASURES AND DIMENSIONGAINS/LOSSES

A. Motivation

Noncoherent network coding makes the topology of the net-work transparent to the code designer, and it has strong theo-retical foundations. Nevertheless, there are some practical is-sues that remain to be taken into account when applying thiscoding scheme. First, the notion of “dimension loss” is fairlyabstract since in networks only symbols (packets) can be erasedor subjected to errors. It is reasonable to assume that a dimen-sion loss corresponds to a number of symbol erasures/errorswithin the same message, although it is not clear how largethis number is supposed to be. In the worst case scenario, evenone symbol erasure may lead to the change of one dimension.Second, the achievable throughput of the scheme and the under-lying decoding complexity may be significantly inferior to thoseachievable only through classical network coding. Of course,this claim only holds if the error-correcting scheme can be inte-grated with a linear network coding method.Let denote the space for some and let be

a set of subspaces of of dimension . Assume thatis transmitted over a noncoherent network. Assume that whilepropagating through the network, the vectors of were sub-jected to symbol errors and symbol erasures.Denote by the subspace spanned by the vectors obtained at

the destination. Each of these vectors has, in the worst case sce-nario, at most symbol errors and symbol erasures. Indeed,this can be justified as follows. If some vector was transmittedin the network, and an erasure (or error) occurred in its th entry,in the worst case scenario, this erasure (error) only affects theth coordinates in all vectors in , causing this coordinate to beerased (or altered, respectively) in all of them. This is true forany network topology. Such erasure (or error) does not affectany other entries in the vectors observed by the receiver.We illustrate this concept by a simple example in Fig. 1. In

that example, one -entry is erased in one vector. In the net-work-coding approach proposed in [14], in the worst case sce-nario, this erasure is treated by disregarding the whole vector.In the proposed approach, in the worst case scenario, the corre-sponding entry in all vectors will be erased.This observation motivates the following definitions.Definition III.1: Consider a vector space . Write

for some and for some subspace. A symbol error in coordinate of is a mapping fromto , such that

Observe that, in general, one may havein Definition III.1.Definition III.2: Let and assume that

for some and for somesubspace . A symbol erasure in coordinate of is amapping from to such that

Fig. 1. (a) Basis vectors of the transmitted subspace : one vector has oneentry erased. (b) In the approach of [14], this causes a dimension loss: the cor-responding vector is discarded. (c) In the proposed alternative approach, thecorresponding entry is erased in all vectors.

The subspace can be naturally associated within the following simple way:

By slightly abusing the terminology, sometimes, we say that(rather than ) is obtained from by an erasure in coordinate. Strictly speaking, such is not a vector space, since there areno mathematical operations defined for the symbol “ ”. How-ever, it will be natural to define for all andthat if and only if , andthat if and only if .To this end, we remark that there are four potential types of

data errors in a network that are not necessarily incurred inde-pendently:1) Symbol erasures;2) Symbol errors;3) Dimension losses;4) Dimension gains.Below, we generalize the operator channel as follows.Definition III.3: Let be a subspace of . The sto-

chastic operator channel with symbol errors and erasures re-turns a random subspace such that

(3)

Page 5: Hybrid Noncoherent Network Coding

SKACHEK et al.: HYBRID NONCOHERENT NETWORK CODING 3321

for some , where all , , are one of thefollowing:1) a symbol error , for some ;2) a symbol erasure , for some , and there are exactly

’s that are symbol erasures;3) a dimension loss;4) a dimension gain.Furthermore, without loss of generality, we assume that allsymbol errors and erasures are preceded by all dimension gainsand losses.1 We reiterate this statement throughout the paper,since it is of importance in our subsequent derivations.We now on focus on two important cases of Definition III.3.Definition III.4: The operator channel in Definition III.3 is

called an operator channel with symbol erasures if the numberof symbol errors is always zero.For an operator channel with symbol erasures, for each output, we may and will always assume that dimension losses and

gains have occurred first, followed by symbol erasures. Morespecifically, let be the set of erasedcoordinates in , and let . Then

where , . Here, one first createsfrom via dimension errors only (losses and gains). Subse-

quently, is created from by erasing coordinates in . Letand . We say that is the number

of dimension losses, is the number of dimension gains,and is the number of symbol erasures.Definition III.5: The operator channel in Definition III.3 is

called an erasure-operator channel with symbol errors and era-sures if the number of dimension gains is always zero.For the operator channel in Definition III.5, we also assume

that dimension losses have occurred first, followed by symbolerrors and erasures. We define the numbers of dimension losses,symbol errors, and symbol erasures analogous to the case of theoperator channel with symbol erasures.In the forthcoming sections, we first concentrate on designing

codes that are able to handle simultaneously symbol erasures,dimension losses, and dimension gains. We postpone the dis-cussion about how to handle symbol errors to Sections VIIand VIII.

B. Code Definition

We start the development of our approach with the followingdefinition.Definition III.6: A subspace code (a set of

subspaces in of dimension ) is called a code capable ofcorrecting dimension errors (either losses or gains) and

symbol erasures, or more succinctly, a -hybrid code,if it satisfies the following properties.1) For any , .2) For any ,

.

1Even if the errors appear in a different order, the resulting subspace maystill be generated by first applying dimension errors and then subsequently in-troducing symbol errors/erasures.

3) Let . Let be the subspace obtained from viasymbol erasures, where . Then,and the space is the only preimage of in under thegiven symbol erasures.

4) Let . Let be obtained from and ,respectively, via symbol erasures, where. Here, both and have erasures in the same set ofcoordinates. Then,.

We explain next why the class of hybrid codes, satis-fying properties 1–4, is capable of correcting dimensionerrors and symbol erasures.Theorem III.7: Let be a code satisfying prop-

erties 1–4. Then, is capable of correcting any error pattern ofdimension errors and symbol erasures.

Proof: Suppose that is transmitted through the op-erator channel, and that the subspace isreceived, where dimension errors and symbol era-sures occurred. Note that here is not necessarily equal to .Recall the assumption that the dimension errors occurred first

and are followed by symbol erasures. As pointed out before, theorder in which dimensional errors occur is irrelevant.More formally, let

be a set of erased coordinates in , and let, for some subspace . Then

where , , and

(4)

We show that if satisfies properties 1–4, then it is possibleto recover from . Indeed, consider the following set of sub-spaces:

Take any . By property 3, ,and by property 4, .Therefore, is a subspace code.It is capable of correcting of up to dimension errors in

.Denote . Then, from Lemma II.1 and the

bound in (4)

We conclude that there exists a (not necessarily efficient)bounded-distance subspace decoder for the code that iscapable of recovering from .Finally, observe that is obtained from by erasing

coordinates indexed by . From property 3, the preimage ofunder these erasures is unique. Therefore, can be recoveredfrom .Remark III.8: The intuition behind the definition of hybrid

codes is that dimension losses may and actually occur as aconsequence of symbol erasures or errors. Symbol erasures are“easier” to correct than dimension losses, and upon correcting anumber of symbol erasures one expects to reduce the number of

Page 6: Hybrid Noncoherent Network Coding

3322 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 59, NO. 6, JUNE 2013

dimension losses. These claims are more rigorously formulatedin Section V-C.Throughout the remainder of the paper, we use the notation

2 to denote a hybrid codewith the following properties:1) ;2) for all , ;3) ;4) is a code capable of correcting dimension errorsand symbol erasures.

IV. BOUNDS ON THE PARAMETERS OF HYBRID CODES

In this section, we derive the Singleton and the sphere-packing bound for hybrid codes handling dimensionlosses/gains and symbol erasures simultaneously.

A. Singleton Bound

Assume that a vector space over has dimension , andlet be a subspace code. In what follows, we use apuncturing of the code .Definition IV.1: The puncturing of a code at

position is a set of subspaces of given by

(5)

Remark IV: In general, Definition IV.1 is different from thedefinition of puncturing in [14, Sec. IV.C]. In particular, thepuncturing in Definition IV.1 does not necessarily decrease thedimension of . On the other hand, Definition IV.1 is similar tothe first part of the definition of -coordinate puncturing in [4,Sec. 5.A].Theorem IV.3: Let be a code with parameters

in the ambient space . If ,then coordinate puncturing at coordinate yields a code withparameters .

Proof: Let be a code obtained by puncturing of theth coordinate in all vectors spaces in , as in (5). Clearly,the dimension of the ambient space decreases by one underthis puncturing, and so the resulting ambient space satis-fies .Let . Since , by property (3) in Definition III.6,

and has a unique preimage. Therefore,.The fact that puncturing does not change the subspace dis-

tance follows from the property that is a code capable ofcorrecting dimension errors and symbol errors.Thus, , whereand are obtained by puncturing of and , respectively.Since each subspace in is obtained from its preimage in byan erasure in the th coordinate, the codes’ Hamming distanceresulting from puncturing is at least .Theorem IV.4: The size of the

code satisfies

2Whenever it is apparent from the context, we omit the subscript .

where stands for the size of the largest sub-space code .

Proof: We apply coordinate puncturings to . Theresulting code is a subspace code.Indeed, it has the same number of codewords as , and it is a setof -dimensional subspaces in a -dimensional ambientspace, whose pairwise intersection is of dimension . Inparticular, its size is upper bounded by .Corollary IV.5: From the Singleton bound in [14], the sizeof the code satisfies

(6)We use the following result from [14].Lemma IV.6 (see [14, Lemma 4]): The Gaussian coefficientsatisfies

We also use the following definition of the rate of the sub-space code.Definition IV.7: The rate of the subspace code is defined

as .Next, let

Thus, by using the bound in (6), an asymptotic version of thelatter bound reads as follows.Corollary IV.8: The rate of a code

satisfies

Proof: We start with the first expression in the right-handside of (6), namely

From Lemma IV.6, we obtain that

Taking of both sides yields

and by dividing the last inequality by , we obtain

Page 7: Hybrid Noncoherent Network Coding

SKACHEK et al.: HYBRID NONCOHERENT NETWORK CODING 3323

By using the second expression in the right-hand side of (6),we similarly obtain that

B. Sphere-Packing Bound

We start with the following definition.Definition IV.9: Amatrix over is said to be in a reduced

row echelon form if the following conditions hold:1) Each nonzero row in has more leading zeros than theprevious row.

2) The leftmost nonzero entry in each row in is one.3) Every leftmost nonzero entry in each row is the onlynonzero entry in its column.

It is well known that any -dimensional subspace of canbe uniquely represented by an matrix over in reducedrow echelon form.Let be the ambient space , and let . Fix

two integers , and . Two vector spacesare called -adjacent if there exists a set

of coordinates , , and a vectorspace such that

and

Note that the adjacency relation is symmetric with respect to theorder of , namely and are -adjacent if and onlyif and are -adjacent.Assume that the hybrid code is

used over a network. Let be the transmitted subspace,and let be the received subspace, for some

, and for some as above, as a result ofdimension losses and gains, and symbol erasures. Then,and are -adjacent. If there is no other codewordsuch that and are -adjacent, then the decoder, whichis able to correct dimension erasures/gains and symbol era-sures, can recover from . This observation motivates thefollowing definition.Definition IV.10: Let be a vector space , and let

. The sphere around is defined as

are adjacent

Now, we recall the following result from [14].Theorem IV.11 (see [14, Th. 5]): For any , and

any

We generalize this theorem in the following way.

Theorem IV.12: Let be an code.For any , any , and any

Proof: Take a set ofcardinality . Let , and take a vector space givenby

Fix some and consider an arbitrary , such that

and are -adjacent. Define and

. Then, by the definition of -adjacency

(7)

Therefore, for a given , the number of subspaces satisfying(7) is given by

Next, we estimate the number of different subspacessuch that for a given . Consider the reduced rowechelon form matrix of dimension representing. In order to obtain this matrix, we need to establish the values

of the last entries in every row of , while the first entriesare equal to their counterparts in . There are rows in ,and each entry can take one of values. As each yieldsdifferent choices of , the claimed result follows.Note that the inequality in the statement of Theorem IV.12 is

due to a specific (rather restrictive) selection of the set . Thefollowing example further illustrates the idea of the proof.Example IV.1: Assume that , , , and, and that the reduced row echelon form of is given by

the following matrix over :

Then, the reduced row echelon form of is given by thefirst five columns of the above matrix. Consider such that

. Such a can be obtained when one of therows in the reduced row echelon form of is replaced by adifferent vector. One possible reduced row echelon form ofis the submatrix formed by the first 5 columns of the matrixbelow. Then, in order to find all possible options for , weneed to fill in the values of the black dots in the lastcolumns of the following matrix:

This can be done in ways.

Page 8: Hybrid Noncoherent Network Coding

3324 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 59, NO. 6, JUNE 2013

From Theorem IV.12, the following sphere-packing bound isimmediate.Corollary IV.13: Let be a code that corrects

dimension losses/gains and symbol erasures. Then,for all , the spheres are disjoint.Therefore

(8)

Now, we turn to an asymptotic analysis of the bound (8).From Lemma IV.6, we have

If , then the dominant term inis obtained when . In

that case, one has

where means that the two expressions andare asymptotically equal.

By taking the base- logarithm and dividing both sides of theabove expression by , we obtain the following result.Corollary IV.14: Let be a code that corrects

dimension losses/gains and symbol erasures. Then,its rate satisfies

V. CODE CONSTRUCTION

Next, we construct hybrid codes capable of correcting dimen-sion losses/gains and symbol erasures simultaneously. We showthat these codes are asymptotically optimal with respect to theSingleton bound. We also provide some examples comparinghybrid codes to subspace codes.

A. Code Construction

Let be a vector space of dimension , and letbe a subspace code in . In other

words, is a set of subspaces of of dimension , such thatfor any , , . We fix a basisof and denote its vectors by . We denotethe decoder for the subspace metric and by .

Let be an generator matrix of theGeneralized Reed–Solomon (GRS) code over of length

, given by

......

.... . .

...

. . .

Here, , , denote distinct nonzero fieldelements, while , , denote arbitrary nonzeroelements (see [18, Ch. 5] for more details).We use the notation for the th row of , for

. The code is capable of correcting anyerror pattern of errors and erasures given that .In this section, we are particularly interested in the case when

.Denote by a decoder for the code , which corrects any

error pattern of errors and erasures, whenever .Denote by the linear space . Let be amatrix over such that

and therefore

Clearly, such an exists since and are two differentbases for .We define a linear mapping as follows. For

an arbitrary vector

This mapping can be naturally extended to the mapping(with the slight abuse of notation), where

stands for a set of all linear subcodes of . For any ,we have

It is easy to see that is a linear mapping, and that the imageof the linear space is a linear space. Moreover, it is straight-forward to show that this mapping, when applied to subspacesof , is one-to-one. Thus, for any

(9)

One can check that for any , it holds

(10)

Next, we define a code as

Page 9: Hybrid Noncoherent Network Coding

SKACHEK et al.: HYBRID NONCOHERENT NETWORK CODING 3325

Theorem V.1: The code is a hybrid code over , withparameters .

Proof: It is straightforward to verify that has the firsttwo parameters as stated. The third parameter follows from thefact that is the number of subspaces in , and two differentsubspaces are mapped onto different subspaces under .Next, we show that is a code capable of correcting

dimension gains/losses and symbol erasures. It suffices toshow the following two properties.1) Let . Let be the subspace obtained from by any

symbol erasures, such that . Then,and the space is the only preimage of in .

2) Let . Let be obtained from and ,respectively, by symbol erasures, such that .Then, .

Indeed, that these two conditions are satisfied can be shown asfollows.1) Let and be obtained from by symbolerasures, such that . Let bea basis of , and let be a set of corre-sponding vectors obtained by symbol erasures. Then, forany , not all of which are zero

is a vector of the Hamming weight . Therefore, afterapplying symbol erasures, the Hamming weight of theresulting vector

is at least , for any ,not all of which are zero. Therefore, the vectors

are linearly independent, and thus.

Next, take a vector . Since the min-imum distance of a code is , and thus, the code cancorrect any pattern of up to symbol erasures, theonly possible preimage of under any symbol erasures,

, is . Therefore, eachhas a unique preimage.

2) Let and let be obtained from and ,respectively, by symbol erasures, such that .From part (1) of the proof, . It issufficient to show that . As-sume, on the contrary, that .This means that there exists and , ,such that by applying symbol erasures to these vectors,one obtains resulting vectors and that are equal. Re-call, however, that , and therefore .We hence arrive to a contradiction, and thus,

.

The following generalization of Property 1 above holds.Corollary V.2: Let be a subcode of (of any dimension).

Let be obtained from by arbitrary symbol erasures, such

that . Then, and the space isthe only preimage of in .The proof follows along the same lines of the proof of Prop-

erty 1.

B. Asymptotic Optimality

In Section IV, we derived upper bounds on the size of generalhybrid codes. These bounds imply upper bounds on the size ofthe code .Moreover, the code is asymptotically optimal withrespect to one of these bounds, as will be shown below.Consider the code constructed from a subspace code with

parameters and a classical GRS codewith parameters , , as de-scribed in the previous section. The resulting code is an

code. The number of codewords of the codeis . If we take as described in [14], with , then the-ary logarithm of the number of the codewords is given by

Therefore, the log of the cardinality of is given by

(11)

Hence, from Lemma IV.6, we have the following result.Corollary V.3:

Thus, the code is asymptotically order-optimal (i.e., optimalup to a constant factor) with respect to the Singleton bound (6).

C. Hybrid Codes Versus Subspace Codes

In this section, we analyze the error-correcting ability ofhybrid codes and subspace codes. We show that in some sce-narios, hybrid codes achieve significant improvement in coderate, while having similar error-correcting capability, whencompared to their subspace codes counterparts.

Examples

Let be the code defined as in [14]. When the code isused over a noncoherent network, in the worst case scenario,each symbol erasure translates into the loss of one dimension,and each symbol error translates into one dimension loss andone erroneous dimension gain. This may happen when all theerasures and errors occur in linearly independent vectors. In ad-dition, note that requiring each linearly independent vector tobe able to correct up to and including erasures is some-what restrictive, since it imposes an individual, rather than jointconstraint, on the total number of erasures in the transmittedsubspace.We show next the advantage of using the code for the case

when all data errors in the noncoherent network take form ofsymbol erasures. These symbol erasures are the cause of di-mension losses. The code has more codewords than whilehaving the same overall error-correcting capability.Example V.1: Consider the code with parameters

. This code can correct upto and including two dimension losses. If the symbol erasures

Page 10: Hybrid Noncoherent Network Coding

3326 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 59, NO. 6, JUNE 2013

happen in linearly independent vectors, the result is a loss of twodimensions, and the code can provably correct such two symbolerasures. Alternatively, one symbol error results in one dimen-sion loss and one dimension gain, which is also correctable bythis code. However, this code is not able to correct any com-bination of three symbol erasures that occur in different basisvectors. Note that the code contains (subspaces) codewords.Now, let and consider the set , where

. Fix some basis for . Let be GRS code(for ). Define the mapping as in Section V-A.The resulting code has

codewords (subspaces), for all . It has parameters. Since has a minimum distance

3, can correct any two symbol erasures. If those appear indifferent basis vectors, the dimension loss error correctingcapability matches that of the previously described subspacecode. But the number of codewords in the code is strictly largerthan that in .The increase in the number of codewords achieved through

hybrid coding in the above scenario is negligible for large fieldorders. Furthermore, even these modest advantages are presentonly in cases when the symbol erasures (or errors) do not appearin bursts within a few linearly independent vectors.However, the advantage of the new construction is signifi-

cantly more pronounced when the gap between and is large.This gap becomes of interest when the length of data transmittedin the network is much higher than the dimension of the sub-spaces used in coding, or in other words, when the min-cut ofthe network is small compared to the length of the coded vec-tors.Example V.2: Take the code with parameters

. This code can correctup to and including two dimension losses and it containscodewords.For comparison, take and consider the set

, where . Let be a GRS code,with . Define the mapping as before.

The resulting code has parameters

. Since has a minimum distance 3, can correct any twosymbol erasures.The number of codewords in the code equals

This number is strictly larger than (for all ), which isan upper bound on the size of any subspace code.

Comparison of Dimension Losses and Symbol Erasures

The examples described above motivate the following ques-tion: how many symbol erasures should be counted towards onedimension loss for the case that the subspace and hybrid codeshave the same number of codewords?To arrive at the desired conclusion, we use an upper bound on

the size of any constant-dimension subspace code, which was

derived in [14]. Therefore, our findings are also valid for thecodes constructed in [4], [14], and [22], as well as for any otherpossible subspace code.Throughout the section, we use to denote an arbitrary

subspace code. Let us fix the values of theparameters and . Recall that in the worst case, each symbolerasure can cause one dimension loss. We use the Singletonbound on the size of the code [14, Th. 9]. Any such code iscapable of correcting dimension losses, so in the worstcase scenario, it can provably correct only up to symbolerasures. From [14, Th. 9], we have

In comparison, the number of codewords in thecode constructed in Section V-A,

when is taken as in [14], is given by

In order to achieve the same erasure-correcting capability, weset . The underlying assumption isthat symbol erasures are corrected as dimensional losses,while the remaining erasures are handled as simple erasures. Werequire that, for small ,

This is equivalent to

or

which reduces to

(12)

The latter inequality holds for any choice of and ,when , for some small . When inequality (12)is satisfied, hybrid codes correct more symbol erasures than anyconstant-dimension subspace code, designed to correct dimen-sion errors only.Next, we consider maximizing the number of codewords inunder the constraints that , and

, , where is fixed and , are allowed to vary.Recall that

(13)

Let so that , where is aconstant. We aim at maximizing the function

(14)

By taking the first derivative of the expression with respect toand by setting it to zero, we find that . Therefore,the value of that maximizes the number of codewords equals

Page 11: Hybrid Noncoherent Network Coding

SKACHEK et al.: HYBRID NONCOHERENT NETWORK CODING 3327

If , then under the given constraints, the optimalvalue of equals , i.e., it is better to put all error-correcting capability on symbol erasure correction.Consider the expression for the number of codewords in (13).

There are two types of subspace and symbol errors considered:dimension losses and symbol erasures. A combination of sucherrors is subsequently termed an error pattern.Assume that for a specific code , correcting a dimension

loss is on average equivalent to correcting symbol erasures,for some . We consider an optimal selection procedure forthe parameters of for two different error patterns.If the error pattern consists of no dimension losses and

symbol erasures, then (13) becomes . Incomparison, if the error pattern consists of dimensionlosses and no symbol erasures, then (13) becomes

. Since each dimension loss is on average equivalentto symbol erasures, we have

and so

After applying some simple algebra, we obtain that

Therefore, vaguely speaking, it is as hard to correct one dimen-sion loss as it is to correct symbol erasures.

D. Hybrid Codes Versus Coding Vectors

We next compare hybrid codes with coding schemes usedin randomized network coding [9]. In the latter approach, theinformation is presented by vectors rather than vector spaces.Special headers are appended to the vectors in order to keeptrack of which linear combinations were applied to the originalpackets. As a result, the coding vectors have an overhead of atleast bits, where is the number of transmittedpackets. This overhead becomes of less significance if the lengthof the data payload is large.When “lifted” subspace codes, such as [21], are employed

over the regular operator channel, the vector subspace is usuallytransmitted by using its basis vectors. Each such basis vector hasa unique leading bit having value “1”. The collection of such bitscan be viewed as a header, used in coding vectors. When linearcombinations are received at the destination, the original vectorscan be recovered by applying the inverse linear transformation.Therefore, the coding schemes based on codes in [21] can alsobe viewed as a scheme based on coding vectors. Since the codesin [21] are asymptotically optimal with respect to the Singletonbound [14, Th. 9], we conclude that schemes based on codingvectors are asymptotically optimal. (See a related discussion in[10].)Next, consider a hybrid code constructed as in Section V-A,

where the matrix is a systematic generating matrix of a GRScode. Then, the transmitted basis vectors are the rows of ,and so each vector has a unique leading bit “1.” Therefore,the corresponding coding scheme may be viewed as a codingvector scheme with headers of size . By

Fig. 2. Decoder for dimension errors.

Corollary V.3, this scheme is asymptotically optimal with re-spect to the corresponding Singleton bound. Hence, asymptoti-cally optimal hybrid subspace codes have a performance com-parable to that of asymptotically optimal coding vector schemes.The advantages and drawbacks of the schemes based on

coding vectors and on the subspace codes were discussed in[10]. The analysis carried out there extends to hybrid codes, asthey may be constructed around subspace codes.

VI. CORRECTING DIMENSION ERRORS ANDSYMBOL ERASURES

We proceed to present an efficient decoding procedure whichhandles dimension losses, dimension gains, and symbol era-sures. Note that the proposed decoding method may fail in thecase that symbol errors are also present. This issue is discussedin more detail in Section VIII.As before, assume that is transmitted over a nonco-

herent network. Assume also that , where isnot necessarily equal to , was received.Let denote the vector space ,

where all erased coordinates are deleted, . Sim-ilarly, let denote the code where all coordinates erased inare deleted.We first compute , the intersection ofwith the subspace . Assume that are the

basis vectors of when the erased coordinates are marked as“ ”, and so . We apply the erasure-correctingGRS decoder of the code to each so as to obtain . Let

. We proceed by applying the inverse ofthe mapping , denoted by , to . The resulting subspaceis a subspace of , on which we now run the decoder for the

code .The algorithm described above is summarized in Fig. 2.This decoder can correct any combination of dimension

losses and dimension gains such that , andat most symbol erasures. This is proved in the followingtheorem.Theorem VI.1: The decoder in Fig. 2 can correct any error

pattern of up to dimension errors and up to symbolerasures in .

Proof: Suppose that is transmitted through the op-erator channel and that is received, wheresymbol erasures and dimension errors have occurred,

such that and .

Page 12: Hybrid Noncoherent Network Coding

3328 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 59, NO. 6, JUNE 2013

As before, we assume that dimension errors have occurredfirst, followed by symbol erasures. More specifically, let

be the set of erased coordinates in ,and let . Then

where , , and. In other words, first is obtained from by applying

only dimension errors (losses and gains). Then, is obtainedfrom by erasing coordinates in . Note that here we assumethat vectors in contain entries marked with “ ”, and so

. By contrast, is obtained by removingthose erased entries from all vectors in (or, equivalently, fromvectors in ).Denote . Then, by Lemma II.1

We have , . Recallthat . Therefore, since , we have

We, consequently, obtain

Observe that from Corollary V.2, it follows thatand . Moreover, can be

obtained from by symbol erasures. Then, according toCorollary V.2, . We conclude that

.Finally, due to (9) and (10), we have that

. Therefore, the decoderfor the code is able to recover , as claimed.We now turn to estimating the time complexity of hybrid de-

coding. The algorithms in Fig. 2 consists of the following mainsteps:1) Computation of the vector space .Observe that and , where

, since otherwise the decoder cannotcorrect dimension errors. Therefore, this computationcan be done by solving a system of at most equationsover with unknowns. By using Gaussianeliminations, this can be done in time

.2) applications of the decoder .Note that . This requires opera-tions over .

3) One application of the mapping .As before, this step is equivalent to multiplying anmatrix representing the basis of by an trans-formation matrix representing the mapping . This

Fig. 3. Decoder for symbol errors.

step requires opera-tions over .

4) One application of the decoder This takesoperations over (see [12, Ch. 5]).

By summing up all the quantities, we arrive at an expression forthe total complexity of the presented algorithm of the form

operations over .We note that the most time-consuming step in the decoding

process for various choices of the parameters is decoding ofa constant-dimension subspace code, which requires

operations over . However, if the error pattern in a spe-cific network contains a large number of symbol erasures, wecan design hybrid codes such that is fairly small (say, somesmall constant). This reduces the complexity of the overall de-coder, which represents another advantage of hybrid codes overclassical subspace codes.

VII. CORRECTING DIMENSIONS LOSSES AND SYMBOL ERRORS

We describe next how to use the code defined inSection V-A for correction of error patterns that consist ofdimension losses, symbol erasures, and symbol substitutions.More specifically, we show that the code is capable ofcorrecting any error pattern of up to dimension losses,symbol errors, and symbol erasures, wheneverand . However, we note that if, in additionto dimension losses, one also encounters dimension gains, thedecoder for the code may fail. This issue is elaborated on inSection VIII.

A. Decoding

Henceforth, we assume that is transmitted over a non-coherent network and that of dimension , whereis not necessarily equal to , is received.Suppose that , , are the basis

vectors in . We can also view the vectors as vectors in. We apply the GRS decoder for the code

on all these vectors. This decoder produces the vectors. We denote by the span of these

vectors. Then, we apply the inverse of the mapping , denotedby , to . The resulting subspace is a subspace of , onwhich the decoder for the code is applied.The decoding algorithm is summarized in Fig. 3.Analysis of the Decoding Algorithm: We analyze next the al-

gorithm in Fig. 3. The main result of this section is the followingtheorem.

Page 13: Hybrid Noncoherent Network Coding

SKACHEK et al.: HYBRID NONCOHERENT NETWORK CODING 3329

Theorem VII.1: The decoder in Fig. 3 can correct any errorpattern in which consists of dimension losses, symbolerrors and symbol erasures, whenever and

.Proof: Suppose that

is transmitted through an operator channel, and that, , is obtained by

the receiver. Assume that dimension losses, symbol errorsand symbol erasures have occurred.Let be an arbitrary received vector, . Then, can

be obtained from a unique vector by at most symbolerrors and at most symbol erasures, where

and , . Since is a linear combination ofvectors in , it follows that . Therefore, the decoderis able to recover from . By using the structure of the algo-rithm, we conclude that , and sois a subspace of .Since dimension losses occurred, ,

and . Due to (9) and (10), wehave that . Therefore, the decoder

is able to recover from , as claimed.Decoding TimeComplexity: The decoding algorithm consists

of the following computational steps:1) applications of an RS decoder, for codes of length

.By using Berlekamp–Massey type decoders, each de-coding round can be performed with opera-tions over . Thus, this step has a total complexity of

.2) One application of the mapping .First, we have to find a basis for . Gaussian elimina-tion requires operations over . The mapping

is equivalent to multiplying an matrix rep-resenting the basis of by an transforma-tion matrix representing the mapping . The compu-tation of this transformation matrix is done only once inthe preprocessing step, and so we may assume that thismatrix is known. We hence conclude that this step takes

operations over .3) One application of the decoder .This takes operations over (see [12, Ch.5]).

The total complexity of the presented algorithm equals

operations over .The number of operations depends on the dimension of the

received subspace, . It would, hence, be desirable to derivean upper bound on . However, since each linearly indepen-dent vector can carry a different pattern of symbol errors, theresulting dimension of , , may be rather large. However, ifwe assume that each link to the receiver carries only one vector,

can be bounded from above by the in-degree of the receiver.Note that the same issue arises in the context of classical sub-space coding, although it was not previously addressed in theliterature.

VIII. FOUR TYPES OF ERRORS AND DECODING FAILURE

A. Failure Example

The following example illustrates that the decoder in Fig. 3may fail in the presence of both symbol errors and dimensiongains.Let , , be a standard basis. Let, and let be a subspace code with .

The code is able to correct up to and including two dimensionlosses and/or gains. Additionally, letbe a basis of a GRS code . The code is able tocorrect one symbol error. Assume, without loss of generality,that is a codeword of a minimalweight in .Assume that the sender wants to transmit the space

to the receiver. According to the algorithm,the sender encodes this space as ,and sends the vectors , , through the network. As-sume that the vector is removed, and the erroneous vector

, is injected instead. At thispoint, the corresponding vector space under transmission is

. Then, it is plausible that , , and propagatefurther through the network due to network coding. To thisend, assume that the receiver receives the following linearcombinations, and . Assume also that during thelast transmission, the vector is subjected to a symbol error,resulting in .The receiver applies the decoder on these three vectors,

resulting in

The subspace received at the destination is

and so the corresponding preimage under is given by

Observe that and that , so thatthe subspace distance between and is four. Therefore, thesubspace decoder may fail when decoding from . Thissituation is illustrated in Fig. 4.

B. Decoding Strategies for Dimension Gains and SymbolErrors

To illustrate the difficulty of performing combined symboland dimension gain error decoding of the code , below wediscuss some alternative decoding strategies. We mention why

Page 14: Hybrid Noncoherent Network Coding

3330 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 59, NO. 6, JUNE 2013

Fig. 4. Situation when the decoder fails. The dimension error vector shouldbe decoded into . However, a symbol error changes into , whichis decoded into . This ambiguity increases the dimension of the errorspace, and causes decoder failure.

these strategies, when applied to the problem at hand, do notwork.

Gaussian eliminations on the orthogonal space.Assume that the vector space is transmitted and isobtained by a combination of symbol and dimension errors,including dimension gains. Then, can be represented as

, where is some error space. If therewere no symbol errors, the dimension of would equalthe number of dimension gains, which is small. However,if symbol errors are present, takes a more involved form.One can try to represent the space in a particular basis,for example one in which the symbol errors have lowweight. Ideally, each symbol error would correspond to avector of weight one in that space. If one could accomplishthis task, then it may be possible to find all the low weightvectors and remove them, or to puncture the correspondingcoordinates. After such a procedure, one would ideally beleft with only dimension gain vectors.A particular difficulty in this scenario is that there are toomany different bases for , and it is not immediately clearwhich basis should be selected. And, while in the rightbasis the symbol error vector will have weight one, in mostof the other bases this weight will be large. Moreover, thespace can be viewed as a dual code of . However, thenthe problem of finding low-weight vectors becomes similarto the problem of finding the smallest weight codeword inthe dual code, which is known to be NP hard. Therefore, itis likely that finding the right basis in is difficult, too.Using list-decoding for RS codes.One can think about using list-decoding for the code .Since the covering radius of RS codes is , it mayhappen that the dimension error transforms the codewordinto a vector at distance from any codeword. Then,even a single symbol error can move this codeword to adifferent ball of radius around a codeword, similarlyto the situation depicted in Fig. 4. Since list-decoding cancorrect only less than errors, list-decoding cannot recoverthe original codeword.The second problem associated with list decoding is as fol-lows. Even if one could construct a polynomial-size listof all possible codewords before the dimension error tookplace, there would be a different list for each receivedvector. Since there could be as many as different lists,

an exhaustive approach for choosing the right codewordsfrom all the lists can require a time exponential in .

IX. CONCLUSION

We introduced a new class of subspace codes capable of cor-recting both dimension errors and symbol errors, termed hy-brid codes. For these codes, we derived upper bounds on thesize of the codes and presented an asymptotically constant-op-timal concatenated code design method. We presented polyno-mial-time decoding algorithms which are capable of correctingthe following error patterns:1) dimension losses/gains and symbol erasures;2) dimension losses and symbol erasures/errors.We also discussed correction of error patterns that consist of

all four types of errors: dimension losses/gains and symbol era-sures/errors. As we illustrated by an example, the correspondingtask is difficult and is left as an open problem.

ACKNOWLEDGMENT

The authors are grateful to Danilo Silva for providing usefuland insightful comments about the work in the manuscript, toPascal Vontobel and the reviewers for many comments that im-proved the presentation of the work, and to the Associate Editor,Christina Fragouli, for insightful suggestions and for handlingthe manuscript.

REFERENCES[1] R. Ahlswede, N. Cai, S. Y. R. Li, and R. W. Yeung, “Network infor-

mation flow,” IEEE Trans. Inf. Theory, vol. 46, no. 4, pp. 1204–1216,Jul. 2000.

[2] N. Cai and R. W. Yeung, “Network coding and error-correction,” inProc. IEEE Inf. Theory Workshop, Bangalore, India, Oct. 2002, pp.119–122.

[3] R. Dougherty, C. Freiling, and K. Zeger, “Insufficiency of linear codingin network information flow,” IEEE Trans. Inf. Theory, vol. 51, no. 8,pp. 2745–2759, Aug. 2005.

[4] T. Etzion and N. Silberstein, “Error-correcting codes in projectivespaces via rank-metric codes and Ferrers diagrams,” IEEE Trans. Inf.Theory, vol. 55, no. 7, pp. 2909–2919, Jul. 2009.

[5] T. Etzion and A. Vardy, “Error-correcting codes in projective space,”IEEE Trans. Inf. Theory, vol. 57, no. 2, pp. 1165–1173, Feb. 2011.

[6] E.M. Gabidulin, “Theory of codes with maximal rank distance,” Probl.Inf. Transmiss., vol. 21, pp. 1–12, Jul. 1985.

[7] E. M. Gabidulin and M. Bossert, “Codes for network coding,” in Proc.IEEE Int. Symp. Inf. Theory, Toronto, ON, Canada, Jul. 2008, pp.867–870.

[8] M. Gadouleau and Z. Yan, “Constant-rank codes and their connectionto constant-dimension codes,” IEEE Trans. Inf. Theory, vol. 56, no. 7,pp. 3207–3216, Jul. 2010.

[9] T. Ho, R. Kötter, M. Médard, D. R. Karger, and M. Effros, “The bene-fits of coding over routing in a randomized setting,” in Proc. IEEE Int.Symp. Inf. Theory, Yokohama, Japan, Jun./Jul. 2003, p. 442.

[10] M. J. Siavoshani, S. Mohajer, C. Fragouli, and S. Diggavi, “On thecapacity of non-coherent network coding,” IEEE Trans. Inf. Theory,vol. 57, no. 2, pp. 1046–1066, Feb. 2011.

[11] S. Katti, D. Katabi, H. Balakrishnan, and M. Médard, “Symbol-levelnetwork coding for wireless mesh networks,” in Proc. ACM SIG-COMM Conf. Data Commun., Seattle, WA, USA, 2008, pp. 401–412.

[12] A. Khaleghi, D. Silva, and F. R. Kschischang, “Subspace codes,” Lec-ture Notes Comput. Sci., vol. 5921, pp. 1–21, 2009.

[13] A. Kohnert and S. Kurz, “Construction of large constant dimensioncodes with a prescribed minimum distance,” Lecture Notes Comput.Sci., vol. 5393, pp. 31–42, 2008.

[14] R. Kötter and F. R. Kschischang, “Coding for errors and erasures inrandom network coding,” IEEE Trans. Inf. Theory, vol. 54, no. 8, pp.3579–3591, Aug. 2008.

[15] R. Kötter and M. Médard, “An algebraic approach to network coding,”IEEE/ACM Trans. Netw., vol. 11, no. 5, pp. 782–795, Oct. 2003.

[16] S. Li, R. Yeung, and N. Cai, “Linear network coding,” IEEE Trans. Inf.Theory, vol. 49, no. 2, pp. 371–381, Feb. 2003.

Page 15: Hybrid Noncoherent Network Coding

SKACHEK et al.: HYBRID NONCOHERENT NETWORK CODING 3331

[17] F. Manganiello, E. Gorla, and J. Rosenthal, “Spread codes and spreaddecoding in network coding,” in Proc. IEEE Int. Symp. Inf. Theory,Toronto, ON, Canada, Jul. 2008, pp. 881–885.

[18] R. M. Roth, Introduction to Coding Theory. Cambridge, U.K.: Cam-bridge Univ. Press, 2006.

[19] R. M. Roth, “Maximum-rank array codes and their application to criss-cross error-correction,” IEEE Trans. Inf. Theory, vol. 37, no. 2, pp.328–336, Mar. 1991.

[20] N. Silberstein, A. S. Rawat, and S. Vishwanath, Error resilience in dis-tributed storage via rank-metric codes [Online]. Available: http://arxiv.org/abs/1202.0800

[21] D. Silva, F. R. Kschischang, and R. Kötter, “A rank-metric approachto error-control in random network coding,” IEEE Trans. Inf. Theory,vol. 54, no. 9, pp. 3951–3967, Sep. 2008.

[22] V. Skachek, “Recursive code construction for random networks,” IEEETrans. Inf. Theory, vol. 56, no. 3, pp. 1378–1382, Mar. 2010.

[23] A.-L. Trautmann and J. Rosenthal, “New improvements on the Ech-elon-Ferrers construction,” in Proc. 19th Int. Symp. Math. TheoryNetw. Syst., Budapest, Hungary, 2010, pp. 405–408.

[24] S. T. Xia and F. W. Fu, “Johnson type bounds on constant dimensioncodes,” Designs, Codes Cryptogr., vol. 50, pp. 163–172, Feb. 2009.

[25] J. H. van Lint and R.M. Wilson, A Course in Combinatorics, 2nd ed.Cambridge, U.K.: Cambridge Univ. Press, 2001.

[26] H. Wang, C. Xing, and R. Safavi-Naini, “Linear authentication codes:Bounds and constructions,” IEEE Trans. Inf. Theory, vol. 49, no. 4, pp.866–873, Apr. 2003.

[27] Q. Wang, S. Jaggi, and S. Y. R. Li, “Binary error correcting networkcodes,” in Proc. Inf. Theory Workshop, Paraty, Brazil, 2011, pp.498–502.

Vitaly Skachek received the B.A. (Cum Laude), M.Sc. and Ph.D. degrees incomputer science from the Technion—Israel Institute of Technology, in 1994,1998 and 2007, respectively.In the summer of 2004, he visited the Mathematics of Communications De-

partment at Bell Laboratories under the DIMACS Special Focus Program inComputational Information Theory and Coding. During 2007–2012, he held re-search positions with the Claude Shannon Institute, University College Dublin,with the School of Physical andMathematical Sciences, Nanyang TechnologicalUniversity, Singapore, with the Coordinated Science Laboratory, University ofIllinois at Urbana-Champaign, Urbana, and with the Department of Electricaland Computer Engineering, McGill University, Montréal. He is now a SeniorLecturer with the Institute of Computer Science, University of Tartu.Dr. Skachek is a recipient of the Permanent Excellent Faculty Instructor

award, given by Technion.

OlgicaMilenkovic (M’04) received herM.Sc. degree inMathematics and Ph.D.in Electrical Engineering from the University of Michigan Ann Arbor, in 2001ad 2002, respectively.In 2002 she joined the faculty of University of Colorado, Boulder. After a

visiting professor appointment at UCSD, she moved to University of Illinois,Urbana-Champaign in 2007, where she is now Associate Professor in the De-partment of Electrical and Computer Engineering. Her research interest includecoding theory, analysis of algorithms, bioinformatics, and signal processing.She served as editor in Chief of the Special Issue of the IEEE TRANSACTIONSOF INFORMATION THEORY on Molecular Biology and Neuroscience, andassociate editor of the IEEE TRANSACTIONS ON COMMUNICATIONS, the IEEETRANSACTIONS ON SIGNAL PROCESSING, and the IEEE TRANSACTIONS ONINFORMATION THEORY.For her work, she was rewarded the NSF Career Award, DARPA Young Fac-

ulty Awards, and several best conference paper awards in coding theory andbioinformatics. She is a Center for Advanced Studies Associate at UIUC.

Angelia Nedić (M’07) received her B.S. degree from the University of Mon-tenegro (1987) and M.S. degree from the University of Belgrade (1990), bothin Mathematics. She received her Ph.D. degrees from Moscow State Univer-sity (1994) in Mathematics and Mathematical Physics, and from Massachu-setts Institute of Technology in Electrical Engineering and Computer Science(2002). She has been at the BAE Systems Advanced Information Technologyfrom 2002–2006. She is currently an Associate professor at the Department ofIndustrial and Enterprise Systems Engineering at the University of Illinois atUrbana-Champaign, USA, which she joined in fall 2006. She is the recipientof the NSF CAREER Award 2008 in Operations Research for her work in dis-tributed multi-agent optimization.Her general interest is in optimization including fundamental theory, models,

algorithms, and applications. Her current research interest is focused on large-scale convex optimization, distributed multi-agent optimization, stochastic ap-proximations in optimization and variational inequalities with applications insignal processing, machine learning, and control.