Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!

 



Forgot your password?
typodupeerror
×
Networking The Internet Science

Rerouting the Networks 108

prostoalex writes "Scientific American looks at a new approach to clearing networking jams, in research funded by the US military. Instead of using routers to route the packets from point A and point B, thus making some hop in the sequence critical for delivering the message, researchers are exploring a new approach called 'network coding.' (Here is the illustration cited in the article.)" Quoting: "[Four researchers] then at the University of Hong Kong published groundbreaking work that introduced a new approach to distributing information across shared networks. In... network coding, routers are replaced by coders, which transmit evidence about messages instead of sending the messages themselves. When receivers collect the evidence, they deduce the original information from the assembled clues. Although this method may sound counterintuitive, network coding, which is still under study, has the potential to dramatically speed up and improve the reliability of all manner of communications systems and may well spark the next revolution in the field. Investigators are, of course, also exploring additional avenues for improving efficiency; as far as we know, though, those other approaches generally extend existing methods.'"
This discussion has been archived. No new comments can be posted.

Rerouting the Networks

Comments Filter:
  • by SpaceLifeForm ( 228190 ) on Monday May 21, 2007 @07:41PM (#19215673)
    All of that new NSA equipment would have to be re-worked.

    I doubt that the taxpayers would approve.
    • don't worry (Score:5, Funny)

      by geedra ( 1009933 ) on Monday May 21, 2007 @07:49PM (#19215717)
      we'll cut costs by outsourcing to china ;)
      • by elrous0 ( 869638 ) *
        Well, who knows routers better than the Chinese? This system has the potential to let them block the word "democracy" at a MUCH faster rate.
    • All of that new NSA equipment would have to be re-worked. I doubt that the taxpayers would approve.

      I am posting in this thread because I'm fresh out of mod points... But even if I weren't, there isn't any "+1, Deftly Ironic" moderation or I'd set you up with one of those...
      • by grcumb ( 781340 ) on Monday May 21, 2007 @08:48PM (#19216119) Homepage Journal

        All of that new NSA equipment would have to be re-worked. I doubt that the taxpayers would approve.
        I am posting in this thread because I'm fresh out of mod points... But even if I weren't, there isn't any "+1, Deftly Ironic" moderation or I'd set you up with one of those...

        It was removed because people kept using it when they should have chosen the "+1, Biting Satire" option instead.

        (For reference, this post should be modded "+1, Deadpan".)

  • by idontgno ( 624372 ) on Monday May 21, 2007 @07:49PM (#19215719) Journal
    the information transmission equivalent of holography. On one level, it seems sensible and obvious. On another level, it's rather novel (at least, the application of holographic distribution of data to a network architecture). At least, I've never heard of it, either in real-world technology research or even (to my recollection) SF technology.
    • Re: (Score:2, Informative)

      this already happens on really lossy transmissions -- especially things like satellite tv or, say, deep-ish space missions don't want to waste time retransmitting (in satellite tv, if you didn't get it, you can't ask for it again, your picture just looks terrible for a split second).

      so, they use error correcting codes, like, for example, parity (like raid 4,5) or reed-solomon (i think that's used in raid 6 or 7 or something), or turbo-codes.

      fyi, for the latter i don't actually know how it works.

      anyway, this
    • by PaulBu ( 473180 )
      Yep, good analogy. To me, it sounded like a digital equivalent of (analog) spread-spectrum communications. Or, as another poster replying to you, a version of distributed ECC. Sounds cool, will have to read the actual paper... :)

      Paul B.
  • by khasim ( 1285 ) <brandioch.conner@gmail.com> on Monday May 21, 2007 @07:53PM (#19215747)
    The numbers they give do not correspond to the diagram.

    Suppose we additionally send Amy's missive to Carl along link 1 and Ben's to Dana along link 4. Sending these two messages uses network resources (links 1 and 4) that the routing system could not usefully employ for meeting Amy's and Ben's demands. Carl's node receives Amy's transmission and knows for each instant (from link 6) whether the number of 1s in the pair of messages issued by Amy and Ben is even or odd. If Carl's node is programmed to also "know" the rule used by the coders at the start of link 5 or if it can infer the rule from the evidence itself, the collected evidence will enable it to decipher the message sent by Ben. And Dana's node will similarly uncover Amy's message.

    So they had originally started with one channel from A to D and that channel was also used by B sending to C.

    They said that it was inefficient because A to D would have to wait at times for B to C.

    Their "solution" seems to be to add a SECOND CHANNEL that will transmit a bit AS FAST AS THE PRIMARY CHANNEL and so both messages will get to their destination sooner.

    Why not just use that secondary channel to transmit the data in the first place?

    Did I miss something?
    • That's what it was looking like to me...

      It sounds the article's original hypothesis is 'Frames go from Router A to Router B.' So what they do to solve it is...add more routers. The diagram simply looks like 1 router that can route to 6 further routers (A-F) which then route to a further 20 (G-Z).

      I'm probably missing something too...that seems too easy.
      • The top router can only handle 6 outbound connections at 1 bit per second (total 6 bits per second).

        The second rank of routers can handle TEN outbound connections at 1 bit per second (total 10 bits per second).

        And they aren't adding any time delays for the merging of the data streams.

        I think I can see what they're trying to do the pictures. And it doesn't look right to me (but who am I anyway). Their choke point (the first router) needs a LOT of information about the second rank of routers.

        In essence, they'
        • The top router can only handle 6 outbound connections at 1 bit per second (total 6 bits per second). The second rank of routers can handle TEN outbound connections at 1 bit per second (total 10 bits per second).
          I think you are off by several orders of magnatude.
          • I think you didn't RTFA.. though you aren't missing much. It vaguely states a method of compressing information together without having any reasonable way to uncompress it again. That's what they need to work out before this will be useful. And the fact that he says things like "the startling truth" before stating things that aren't at all startling makes him sound like either an idiot or a salesman..
    • Re: (Score:2, Informative)

      What happens when the second node goes down?
      If you automatically broadcast the knowledge of the message across all available channels, if it gets to the other end its won.

      This is the not putting all your eggs in one basket scenario which attempts to address bandwidth limitation issues.

      It was only recently I was thinking about global communications on a wireless mesh without ISP involvement (reroute the backbone across home networks), I wonder if this could be the kickstart it needs?
    • by markov_chain ( 202465 ) on Monday May 21, 2007 @08:05PM (#19215819)
      You're right, their example is not very clear. It works best when there are extra resources available to the system that *can't* be used for other purposes; the best example of this is wireless communication, where all transmissions are broadcast. For example, if A sends a packet to B, but C also hears it, we can't possibly use the A-C transmission for some other data.
      • by khasim ( 1285 )
        For their system to work, it seems that A has to send to B and C

        which blocks C from sending anything

        but C must be sending to D at the same time so that the messages can be merged at the "coder" point.

        And the coder would then use TWO transmissions to send the "hint" and the merged data.

        It's still sending two data streams (one for merged stream and one for hints) when they were originally complaining about how it was a single channel.
        • by markov_chain ( 202465 ) on Monday May 21, 2007 @08:42PM (#19216095)
          The standard example in wireless networks goes like this. Suppose we have a 3-node chain A-B-C, where A has stuff to send to C, and vice versa, i.e. the communication is bidirectional. A and C are too far from each other, so they need to go through B.

          Without any coding this is how time is spent:
          1. A->B: pAB
          2. B->C: pAB
          3. C->B: pBA
          4. B->A: pBA

          Note than in steps 2 and 4, thanks to the nature of wireless channels both A and C get the packet transmitted by B. One of these receptions is thus wasted. With coding, it can be used to send the same data in 25% less time:

          1. A->B: pAB
          2. C->B: pBA
          3. B->A, B->C: pAB xor pBA

          To decode, A xors the received data with pAB, C with pBA.
          • by SurturZ ( 54334 ) on Monday May 21, 2007 @08:50PM (#19216127) Homepage Journal
            What about bad packets? Wouldn't error correction stuff the whole thing up? i.e. one re-sent packet and you lose out?
            • by jsn13 ( 734418 )
              [i didn't RTFA].
              Packet loss/retransmits doesn't matter much, IMHO. In GP example, host B could use some explicit out-of-band signaling to tell A and C what packets are used as xor context. Like, via IP options / IP ID field or something.
              • Re:No. (Score:4, Informative)

                by SurturZ ( 54334 ) on Monday May 21, 2007 @11:47PM (#19217321) Homepage Journal
                I worked it out, coding is still better.

                For the example before:
                1. A->B: pAB
                2. C->B: pBA
                3. B->A, B->C: pAB xor pBA

                Assume that at time=3 a bird flies between B & C. The sequence becomes:
                1. A->B: pAB
                2. C->B: pBA
                3. B->A, B->C: pAB xor pBA [B->A SUCCESS, B->C ERROR]
                4. B->C: pAB xor pBA [retry]

                B resends the packet, but addresses it only to C. This is still quicker than routing, which would take 5 time slices.
          • Where is "D"? (Score:3, Interesting)

            by khasim ( 1285 )
            Their discussion isn't about
            A to B to C
            and
            C to B to A

            It's about A sending to D while at the same time B is sending to C.

            You've left off "D".

            And you failed to account for how B would know ahead of time that C would be sending a message. Which fails completely when you try to account for "D" in the equation. You need to account for the packets telling B which points wish to transmit.

            In your wireless example, it would be easier to just skip B and have A broadcast its message to all and sundry and then C can br
            • You've left off "D".
              I wasn't explaining their example, I was giving you a different example where network coding can be put to good use with a clear reason for the existence of the additional unused resources. There are more good examples in wikipedia. [wikipedia.org]

              And you failed to account for how B would know ahead of time that C would be sending a message.
              That's an implementation detail. It's been solved by buffering some packets at B. There are good papers on this out there, like xor in the air. [mit.edu]


              In your wi
              • I wasn't explaining their example, I was giving you a different example where network coding can be put to good use with a clear reason for the existence of the additional unused resources.

                So while the discussion was about their example (did you notice the part where I said that the numbers in the diagram didn't match up with their example) you decide that you should be talking about something completely different.

                Yes, I'm glad you managed to work your way through the simple A B C problem.

                It's just a shame

                • Like the guy below said, check out the butterfly network on the wikipedia page, that explains it more clearly. The trick is that the transmissions are multicast, not unicast, and that the side links would go unused without coding.
          • Ahm... no.

            In your example, C's response dos not depend on what C receives from A. Also, B is magically able to simutanously send two messages (no serial device can do that).

            If the responce (C -> A) is indeed independent from the first message (A -> C), then A and C can sart sending in parallel, thus eliminating the gain in speed. Also, you assume that the message sent from A to C is identical to the message sent from C to A. Assuming that, I'd offen the even more efficient startegy of using A -> A,
    • by kidquantum ( 956549 ) on Monday May 21, 2007 @08:24PM (#19215959)
      There is a good explaination in this reference [arxiv.org]. Although that paper mostly treats the case of quantum networks (which is ubber cool) it gives the background. The standard example (not explained well in TFA) is the http://en.wikipedia.org/wiki/Network_coding [wikipedia.org]>butter fly network and the wikipedia article does a reasonable job explaining it. Basically, even with the three different routes, there is still a bottleneck if you just reroute information. But by coding the information, you get it all through. The example of the butterfly network is very simple and worth learning -- just google it if the wikipedia entry or the arxiv paper doesn't help.
    • by maop ( 309499 )

      Did I miss something?

      The original submitter missed something. The illustration in the summary has nothing to do with the example in the article. The illustration is a completely separate example illustrating multicasting using network coding.

      The example in the article has 6 nodes (A,B,...,F) and 7 links (1,2,...,7). You can reconstruct the topology based on the textual information in the article. The links are 1:(A,C) 2:(A,E) 3:(B,E) 4:(B,D) 5:(E,F) 6:(F,C) 7:(F,D).

  • to design communication systems that can route around nuclear attacks.
  • Hey, as long as I get my torrents faster, I'm all for it.
    • Avalanche was about applying network coding to an overlay network; this research is about applying network coding at the network core itself. So while they share the same technology, expect fundamentally different challenges.

      - sm
  • In the article, I need to transmit all three messages to all receivers and the transmitter/intermediate node link has been fabricated as an obvious bottleneck. How often, in a real network, are the receivers going to have that much connectivity and bandwidth? Also, where is the knowledge that all receiver links are up? The loss of any receiver link means that it will get NO messages, since no complete message is routed of any of its links. If I add enough parity data to allow reconstruction of partial m
    • Re: (Score:2, Insightful)

      by johndoe42 ( 179131 )
      Not at all. This idea is cooler than it sounds in the article.

      Rather than using the silly parity schemes as mentioned in the article, you'd use a long code (i.e. a code that could extend beyond a single transmission). Then, if you lose part of the transmission, you could request enough additional symbols to reconstruct the message. Any decent (i.e. maximum distance separable) erasure code has the property that the amount of data you need does not exceed the size of the message, even if you lose a subset
      • by mrogers ( 85392 )

        I imagine that the current research is on even better codes that wouldn't require as much feedback to work well.

        Yup, rateless codes (e.g. LT, Raptor and online codes) produce an endless stream of output symbols, any sufficiently large subset of which can be used to reconstruct the input.

    • Not to mention anything encrypted via AH+ESP is actually useless. Oh and tunneling? How often would they drop at the expense of one router failing. There is no mention either about trust factors. Would these nodes+receivers be configurable to accept certifications, preshared keys, MD5 (or better) checksums. This has gotten to be the biggest *cough* rip off of stating "Hey I invented P2P!... But its for routers! And uh, its connectionless so don't expect error control or recovery! Isn't that cool!"
  • by jihadist ( 1088389 ) on Monday May 21, 2007 @08:06PM (#19215821) Homepage Journal
    The problem with metadata, as Microsoft found out [slashdot.org] (note: not an MS-bashing post, just an example), is that by making information about information a higher source of "truth" than the information itself, you create an opportunity for any who control the metadata source (mass media, religion, network controllers) to re-define what is true.

    From that, they can stage any number of man-in-the-middle attacks -- the least potent but most widespread of which is convincing a clueless electorate^W userbase that they are certain sources of acclaimed truth, and manipulating them for their own narrow or evil ends.

    Philosophers write about this in inspiring 5-page screeds like "On Truth and Lies in a Non-Moral Sense" [anus.com] by Friedrich Nietzsche. That theory in itself is the foundation of postmodernism and some more dangerous philosophies as well. Could it be that philosophy applies to computer science?

    • Yes, because computer science relies heavily on logic, and logic is itself a philosophy (as well as the basis for most higher level philosophies).
    • by jamesh ( 87723 )
      One has to wonder if a link to 'anus.com' is SFW...
      • by LarsG ( 31008 )
        - This is /.
        - The topic is MitM attacks
        - Someone posts a link to anus.com

        Any decent content filter would mark that one as suspect.
    • by Vo1t ( 1079521 )
      BTW: the site where Nietzsche's document is hosted has a self-explanatory domain name. They are nihilists indeed.
    • by slew ( 2918 )
      Essentially, all so called science (including computer science) is subject to man-in-the-middle attacks.

      Do you trust the university?
      Do you trust the textbook/course notes/literature? (since usually you have not performed the experiments yourself, nor walked through the supporting proofs and/or math)
      Do you trust the professor/teacher?

      And more specific to computer science...

      Do you trust the algorithm?
      Do you trust the compiler?
      Do you trust the computer hardware?

      What about "post-modernism" do you find problemat
  • They could just start dropping 2 out of ever 3 packets. That would give lots more capacity!
  • by Animats ( 122034 ) on Monday May 21, 2007 @08:12PM (#19215871) Homepage

    Note that this is a distribution network for multicasting. Nothing wrong with that, but it doesn't do anything for ordinary point to point Internet communications. The cable TV people might find it useful, though.

    • This is also a hot topic in wireless networks. In multi-hop wireless networks it can enable nodes to forward many packets in a single transmission. A neat paper about this is here [mit.edu].

      Also the coding people are going crazy about this too. There were a couple papers showing that network coding is a simple extension of linear block codes.
  • Firstly, this might work for P2P, DHCP, home based (l)users, but it would never be functional in a real world business network. For one, lets take into consideration security. How would this network carry IPSec tunnel information. Those packet headers need to stay in tact not come from ranDumb address. Not only that, they're introducing n+r number of failures where n = number of nodes and r = number of receivers. Secondly sequencing... Would be a nightmare. How would each node know sequencing. What happens
    • Parent should not be moderated 'troll'.

      SSL would not work for example because it could not be trusted.

      It would only be usable for Multicast where you don't care about the integrity of the data (ex: IPTV).

      In... network coding, routers are replaced by coders, which transmit evidence about messages instead of sending the messages themselves. When receivers collect the evidence, they deduce the original information from the assembled clues.

      It's a solution in search of a problem. We already have Multi

  • ... "from the assembled clues"?

    Congratulations. You've just hardwired a rumor mill. Everyone knows how fast those things travel.

  • by Anonymous Coward
    Its a pitty that 99% of the internet's traffic is unicast not multicast so this won't actually help. Plus this gets really hard computationally quickly.
    • by Spy Hunter ( 317220 ) on Monday May 21, 2007 @09:00PM (#19216197) Journal
      Actually, most Internet traffic (BitTorrent) is multicast mapped onto unicast in a terribly inefficient manner. We are forced to use BitTorrent because ISPs refuse to implement multicast (which makes it hardly surprising that there is no multicast traffic on the Internet).
      • Switch to someone that provides IPv6 (or use a 4-to-6 connection), IPv6 mandates support for multicast.
        • by glwtta ( 532858 ) on Monday May 21, 2007 @09:49PM (#19216555) Homepage
          Switch to someone that provides IPv6 (or use a 4-to-6 connection), IPv6 mandates support for multicast.

          Hmm, that's like telling someone who's complaining that horses are hard to find to get a unicorn instead.
        • Just to be clear, "mandates suppport" means neither "is implemented" nor "works."

          Besides, if you're using a tunneling method, you'll by definition be creating a unicast stream to the v6 endpoint, which is probably topologically further away than the actual source of the content. Now, if you were able to buy a native IPv6 connection from some provider which didn't rely on IPv4 somewhere, you might be in luck. Then again, it would probably be easier to get most providers to implement IPv4 multicast than nat
      • We are forced to use BitTorrent because ISPs refuse to implement multicast

        If by "implement" you mean "turn on", you are correct. Virtually all networking hardware still in use will have multicast support built in. The ISPs just don't want to turn it on because they don't know how to make it fit their existing billing model. Think about it, right now, they accept one packet from a peer, they know they're only going to have to deliver at most one packet to another peer. With a (typical) multicast packet, they could have to deliver thousands of packets, but there isn't an effici

  • Don't do routers already 'efficiently devide' amongst the intermediate nodes, if necessary, by monitoring used/free bandwidth? Isn't the effect the same?

    I might be wrong; I just though this already sort of happens.
    • by zolf13 ( 941799 )
      Using link load as routing metric is like asking someone to shoot yourself in the foot.
  • by BinarySkies ( 920189 ) on Monday May 21, 2007 @09:11PM (#19216297) Homepage
    Point A: (sends a picture of a ladybug)
    Point B: 'Okay... is it an animal, vegetable, or mineral?'
    Coder: 'It's an animal'
    Point B: 'Is it... red?'

    Sounds efficient!
    • by grumling ( 94709 )
      It was Ms. Scarlett in the conservatory with the candlestick.
    • by aj50 ( 789101 )
      We joked about using something similar in our Software Engineering project.

      It was a flight booking system. People always complain that computers are hard to use so we'll link it to something they understand in the real world. Pretty much everyone knows the concept of higher or lower games so we considered a system which would tell you an available flight and you could decide to take it or ask for one earlier or later.

  • "route the packets from point A and point B, thus making some hop in the sequence critical for delivering the message"

    In a well designed network, nothing is dependent on one hop, no matter where it is. We call this a multi homed network and most ISPs utilize more than one path to a certain area of the net.

    Another thing to consider is that link state routing protocols are capable of quickly seeing that a route is down and picking another route from the topology table, and inserting that route into the routin
    • "route the packets from point A and point B, thus making some hop in the sequence critical for delivering the message"

      In a well designed network, nothing is dependent on one hop, no matter where it is. We call this a multi homed network and most ISPs utilize more than one path to a certain area of the net.


      I seem to remember this coming up a few years back. ISP A has connections to backbone provider B using various paths C, D, E, and F--which while seeming redundant actually all pass through point G wh
      • by Comen ( 321331 )
        If all your connection go through G, and a backhoe cuts that fiber etc... then your down, I don't care how you code anything, the original poster seems right to me, you have to manage physical links no matter what.
        Every area of most properly engineered networks are multi homed via at least 2 physical paths.
        In our CO's each different physical path comes in to a separate fiber patch panel, leaves out different paths out of the CO, and should be riding on different sides of the street etc, usually in different
    • by maop ( 309499 )
      The point is that network coding more fully utilizes network capacity and it is just as reliable or more reliable than traditional network routing.
  • by postbigbang ( 761081 ) on Monday May 21, 2007 @09:45PM (#19216525)
    Consider using the described method instead of routers, which use the ISO/OSI stack. The message uses an inference, in some ways like how striping works in a RAID 5 drive failure; the failed drive can be inferred from data on the surviving striped drives. Not a big deal. Chunks/nibbles of the messages have a low occupancy time within a cloud, but many receivers must evaluate many messages-- not all are intended for them and must be discarded or validated (imagine spoofing in such a concept as described in the diagrams!).

    But in the message diagram, there's a talker, and a receiver that gets messages through inference. Between the two, the cloud between them becomes unbelievably saturated with diffuse messages-- the same reason that ATM looks good on paper with its 53-byte packets, but in reality can't deal with traffic jams. This is why MPLS and other deterministic routing methods were invented-- to qualify transmission routes, not just shoot them like a shotgun into a cloud, hoping that the intended receiver gets the message and deciphers it through receive-side heuristics.

    This has been done before, and it didn't work. I'd love to hear comments on why this should succeed.
    • Mod parent up! (Score:3, Informative)

      by khasim ( 1285 )
      Exactly. Their diagram is pretty and has unused resources in the overly simplified "before" case ... but it does NOT scale when your have multiple senders and not every receiver wants every message.

      So each point in their diagram will either have to be aware of every other point (you think routing tables are bad right now) or they have to send MASSIVE amounts of extra data.

      Just from their diagram, they went from 3 lines to 6 lines at the first router. Then they fed those into 6 routers each with TEN lines.

      Th
  • Network coding (Score:4, Informative)

    by gardarh ( 1105395 ) on Monday May 21, 2007 @10:00PM (#19216617)
    Network coding is far from a brand new idea. It is introduced in a paper by Ahlswede http://pdos.csail.mit.edu/decouto/papers/ahlswede0 0.pdf [mit.edu] published in 2000 and has ever since been a very popular research topic in the networking world (http://www.ifp.uiuc.edu/~koetter/NWC/index.html [uiuc.edu]). These "clues" are linear equations where actual packets can be retrieved by applying a gaussian reduction on the equations. Its most obvious applications are with multicasting where utilization of network links can indeed be increased in an informational theoretical perspective. The tradeoffs are increased CPU load on the intermediate and end nodes. The research so far is a bit from getting into the practical stages but it has promise. As for "99% of internet traffic being unicast"; even though that might be true, one needs to think outside the box. If it turns out that multicasting will be much cheaper than now (multicasting is now in most cases basically multiple unicasting), broadcasting TV through packet networks might become much more efficient and this proportion could change. Finally, don't forget that research is still in early stages. I believe there are some years, even decades (if ever), until we will see any of this in practice. Maybe it turns out to be useless for computer networks, who knows. Even so, the basic principle might still prove useful for other applications such as routing inside solid state chips.
    • A simple form of it is already in use today with great results.

      It's Parchive [wikipedia.org] and it is crucial for the transmission of large data files through the usenet network. Usenet is essentially a multicast network, like that in the article and the development of parchive error correction files made it significantly more efficient, essentially bringing the need for retransmits down to nearly zero at a slight cost of 5-10% more data transmitted than would be necessary in a perfectly lossless network.
  • by Anonymous Coward on Monday May 21, 2007 @10:11PM (#19216691)
    First of all, this particular algorithm is really only suited for multicasting the same data to many nodes at the same time. It does not help at all with point-to-point unicasting, which is what the majority of Internet traffic looks like. Secondly, the only real trick here is that you can XOR the data together to transmit more information. So if you have equal length messages A, B, and C, instead of transmitting all three to any particular node, you only transmit XORs of some subset of them (A^B and B^C, or A^C and B^C, etc.) and from those XOR'd versions you can then extract the original 3 messages. Magic! Transmit the equivalent of 2 messages but get 3 messages worth of data! Not groundbreaking; not revolutionary; it's actually a pretty old trick. I've thought for a long time about what real world situation this might be practical for, and have come up with pretty much nothing. It's cute, though.
    • by ezrec ( 29765 )

      XOR would be a bad choice.

      Assume you have A^B, B^C, and A^C

      A^B ^ B^C = A^C

      A^B ^ A^C = B^C

      B^C ^ A^C = A^B

      No matter what you do, without *one original message*, you cannot reconstruct the originals. (Now, this is cool for things like one time pad encryption, but it won't work for network coding).

      However, if you *do* have an original message, they all fall out (assume A^B, B^C, and you have, say, 'A')

      A^B ^ A = B (that's one!)

      B^C ^ B = C (that's the other!)

      So to reconstruct all three me

  • Essentially what they are using are erasure codes such as tornado
    and digital fountains. The problem with these codes are that even
    though they near Shannon's limit, they are not nice towards other
    forms of data transmission such as TCP (aka bandwidth hungry).

    By using such codes the round-trip overheads that are evident in
    protocols such as TCP are eliminated.

    These codes are mainly used for massive multi-cast and stateless
    loss transmissions.

    But using these codes doesn't mean you get rid of routers, it just
    means
  • If this works, it could have several implications that is not immediately visible:

    When the data locally, between two nodes, look like gibberish, does it make it harder to charge traffic by the content in it? Like how a provider may in the future, with IPV6 charge you more per megabyte for say, downloading streaming video than for websurfing. Unlike IPV6 encryption, even the sender and receiver identities would be obfuscated.

    If network coding ends at my ISP, it could still charge me. Websites could also char
  • The concept is pretty interesting although the article is disapointing (and illustrate the priciple using another slashdot lile lame care analogy).

    However, even though we would design a workable technology from this idea, I expect huge resistance from router vendors but also from some Internet designer at IETF.

    Embedding such an advanced function within the network would violate the dogma where Network needs to be kept stupid and most of the function are to be supported by terminals.

    Of course, Internet is no
  • the usual tradeoff of squashing data and causing it to take up less space (and move faster) - but at the expense of compute complexity on the send or receive sides?

    I do NOT want end nodes to have to work harder than they do today. and routers already do their thing very well. adding MORE complexity to save line bandwidth seems silly to me unless you are still dial-up bound.

  • The hypothetical six-node digital network depicted in the box on these two pages can help clarify those options.
    I looked at the diagram link in the submission, and for the above referenced diagram that correlated to the example in the sentences that followed, and didn't find them. Anyone else encounter that problem?
  • ...you surely mean Repiping.
  • With my ECIP protocol we were doing that for video distribution in 1996 and 1997 with the largest network of adult video.
    www.ecip.com

    I called is server based routing, were a cluster of servers would keep tabs on each others status and network communication quality.
    How much latency and loss between each node.
    Then when a message was to be sent, they could try direct or go around the blockage by reflecting packets off a series of servers.
    Packets would also be split of across mul
  • It strikes me that network coding seems similar to the way information is propagated in the brain.
  • and I'm faster than ever!

"All the people are so happy now, their heads are caving in. I'm glad they are a snowman with protective rubber skin" -- They Might Be Giants

Working...