Clay Shirky Defends P2P 113
richard writes: "Clay Shirky has responded to Jon Katz's article, Does P2P Suck?, (and a WSJ article published the same day) in an article titled "Backlash!" on OpenP2P.com. Shirky says: "P2P means many things to many people. PC users don't have to be second-class citizens. PCs can be woven directly into the Internet. Content can be provided from the edges of the network just as surely as from the center. Millions of small computers can be more reliable than one giant server. Millions of small CPUs can do the work of a supercomputer. ... These are sloppy ideas, ideas that don't describe a technology or a business model, but they are also big ideas, and they are also good ideas.""
Apparently P2P doesn't suck.... (Score:1)
P2P (Score:1)
fucking crap...
P2P makes millions for idiots (Score:1)
Why P2P is pointless (Score:1)
Human society has been able to make "free" stuff since forever. Where is all this "free" stuff then? Nowhere. You only get stuff made when it can make a return on the time/money/effort invested.
Don't worry, artists & creators won't go broke. They'll just make stuff that defies digital reproduction.
When p2p comes it will come like the web (Score:1)
A single company will come up with something that actually works and works well and most importantly is easy to use. And it will be safe and secure. The programs will run in a sandbox on the users computers and be totally isolated from the actual computing resources. The file sharing will only see the directory tree that it is pointed at and no other.
I think that the key to the storage is the directory service. Being able to find the data and then download it is crucial. We need a self organizing heirachy. Don't ask me how to do it, if I knew I would be famous already.
The distributed processing and distributed hardware already works well, it will just be icing on the cake by adding these to the distributed storage system.
And don't forget that a distributed database is a combination of distributed storage and distributed processing.
Now, imagine a company with 1,000 desktops. They all run this software and join together in a single computer that has 5 TB of storage, and can run 1,000 different processes at 400MHz each. This is more powerful than most supercomputers and the company gets it for free. Of course, a lot of the storage is going to have to be redundant so that the loss of a single computer will have no effect.
Cool!
Re:Shirky is a weak writer (Score:2)
As far as IMG being the GUI, without question it laid the way for GUI approaches to the net. Witness clickable image maps, the server side of which followed pretty closely on IMG's tail. Without IMG there was no need for GUI approaches so I for one don't mind them being synonymous.
Re:Good Points (Score:2)
I think that's more than a mindset; I think that's what most of the public wants to be.
In fact, that could be the disconnect at a lot of the dot-com woes. And possibly at the stalling of a lot of the "systems" in first- and second-world countries.
As /. users, we are almost all active participants. If there's a conversation going on, we want to be a part of it. If something new is on the horizon, we want to hear about it. We want our thoughts to be feedback.
Joe and Jane Sixpack don't. They don't want Tivo to help them program their own network; they want to switch on the box and have the network programmed for them. They don't want to worry about schools; they want to send their kid to school and have him or her come back educated. They don't want power generation in their backyard; they want the switch the light switch and to have the light come on.
They make a few decisions about where they live, what they do with their time, etc. but for the most part, they don't want to be heavily involved. They know the schools are broken, but they don't want to be the ones to fix them. They know politics is broken, but their reaction is to drop out and not vote. They just want things to work.
The active 10% is out here trying to bring new approaches to them, but they don't really care. Too much change is interesting to us, a whirlwind world we enjoy. We have an advantage in it, because we understand a lot more of it. But change is a problem to them. They don't care for instability; even if it's a pain in the ass, they want to get in their car and go pick up groceries and haul them home, because that's what they've always done and doing what they've always done is appealing in a topsy-turvy world. The more we, the active, push for innovation, the more the passives resist.
A friend of mine is a passive. He's even a computer guy. Back in 1986 I asked him why he didn't get a PC. "I use computers all day at work, I don't want one at home," he said. In 1993 I showed him Usenet. "It's a lot of trouble to read all that stuff, and then have to write back," he said. It was like more work to him. Now he has Internet access, I asked him why he didn't check out /. and other such high-profile sites. But I already knew: he's a passive. A good guy, even a competent, intelligent guy, but he wants life to come to him.
If we can understand these people, we can improve their lives (and they can pay us dearly to do that), but we can't expect them to take an active role. And so when you say to a passive "...all you have to do to run Linux is..." don't be surprised if the answer you get is "No, I don't have to do anything." So the challenge of not only the open source world, but the entire world, is to create paths of least resistance. If they have to do something, they won't. You have to figure out how to make them want to do something.
Ugh, I'll stop now before I become Katz.
Re:Good Points (Score:3)
Part of the problem is the public's mindset of 'consumer'. The Internet has the ability to promote the average person to an active role, but is being strangled by those who believe it is nothing more than television with a keyboard.
A lot of the community oriented services such as IRC and email have begun to erode that image, but there really haven't been any new truly successful technologies stressing community in years. The public has been sucked into this thing called 'the web' and been taught that's all there is to this Internet thing.
Re:So what in heck is this P2P thing? (Score:3)
Of course, with every major advance we make there is always an increasing risk of mass destruction. There is an end to all things. Do nothing and an outside force will destroy everything. Do something and at least you are in control.
So what in heck is this P2P thing? (Score:5)
P2P, Distributed Aggregation and Distributed Computing are three separate but related things.
Peer-to-peer is simply a type of network where all nodes on the system are on equal standing with each other. There are no dedicated server machines, no dedicated client machines, but rather everyone is both a server and a client and they communicate with eachother as equals.
This type of system lends itself to a very interesting change in the way someone finds information. Instead of going to a place (e.g. slashdot.org) to get information, you go to the information to get a place.
Distributed aggregation is a method of intelligently locating and well, aggregating resources distributed among nodes across a network. Whether these resources are files, CPU time or disk space, the method of aggregation should remain basically the same. This fits in very well with the peer-to-peer model to provide each node with a simple way of locating resources on other machines.
Distributed computing is a method of using resources distributed among nodes across a network. Distributed aggregation can be thought of a part of distributed computing as you have to be able to find the resources to use them, but not all distributed computing systems provide or even need a method of handling dynamic changes in the network. Of course, distributed computing systems are not typically peer-to-peer. Individual nodes on the network rarely communicate with each other to share information, but instead handle jobs in batch fashion and push the results up to a central server.
Many have argued that peer-to-peer has existed on the Internet since time began and that all things are basically peer-to-peer. This is quite true in some respects. At the protocol level, machines communicate with other machines in a manner that can be considered peer-to-peer, but historically at the application level there have been a very clear line between servers and clients.
We currently live in a world where the majority of computers are nothing more than glorified dumb terminals utilizing only a small fraction of their computing power. My hope is that one day, the average person won't "use" the Internet, but instead "be" the Internet.
Of course, that's just my opinion. I could be wrong.
Re:P2P and NAT (Score:1)
But I think that raw UDP is probably a better solution. Ask a Gnutella user on a 56k modem how "reliable" TCP is.
Re:P2P and NAT (Score:1)
I don't follow. Why would a fully P2P system require that the open port be fixed? Each system is going to have be enumerated on the network in some fashion...some other system(s) is(are) going to at least have to have your IP address. Why can't it also have a port number. From a firewall standpoint, each system behind the firewall could have its own forwarded port on the external interface.
Re:Direct connections? (Score:1)
I have to concur. I know a little about networking, but mostly my experience is in high-performance computing. In parallel programming, avoiding unnecessary communication is a way of life. Once I looked at the P2P query problem, I hit upon an interesting approach to remove the broadcast requirement by using an approximate query routing scheme. This has the potential to fix P2P network scaling problems at a fundamental level. So I wrote up a paper [homestead.com] outlining the approach. I politely sent off email to the Clip2, OpenP2P, etc. briefly explaining the idea, directing them to the paper, and asking for comments. I've got nothing. Now, maybe it is not much of an idea. I am an outsider in that community, but I really wonder if anyone bothered to read it.
In particular, everyone seems so focused on the "super-peer" concept with Reflector that they ignore underlying scaling problem that will still be there with super-peers. As a network becomes larger, it will not be able to tolerate broadcast queries...even with super-peers. Here [darkridge.com] is a nice explanation why.
Re:Direct connections? (Score:1)
Though I recognize the deficiency, I do not view it as unassailable, which seems to be Ridder's conclusion. In my paper, you will also see a reference to Clip2's analysis pointing to the same fundamental scaling problems with Gnutella. Like Clip2, I believe it is fixable. I simply offer a different approach then they are pursuing. And one that I believe would compliment ongoing P2P development efforts.
My point is that there is a problem with the query strategy, or the parallel algorithm, if you will permit me to borrow terminology from my area of expertise. It is unnecessary to impose such sizable bandwidth requirements on the network when the same query can be effectively processed with a more efficient algorithm. You don't buy a Origin2000 to allow you to run BubbleSort faster. You implement QuickSort and run in on your PC.
And, yes, I have read the specification. There is very little there that has anything to do with query acceleration at an algorithmic level. The lack of respect that you have shown me by flatly accusing me of ignorance without even bothering to examine my paper simply underscores the point I made earlier in this thread--the developers at work here are too caught up in religious zeoltry and clannishness to open their minds to new ideas and constructive criticism.
Should you bother to read my paper, you will see that I am not motivated by any political calling. I honestly don't have time to use Napster, Gnutella, or any of the rest. This is merely an interesting problem. My purpose is solely to add to the discussion and encourage the P2P development community to question some of the assumptions that they have made. Judge my ideas if you care to. My motivation is simply to see P2P technology advance and contribute if I am able.
Re:Direct connections? (Score:1)
Re:UDP? (Score:2)
Uhh
Uhh, oh hell it actually makes the problem worse.
Re:Revolution (Score:1)
The hope and strategy of those interests which want the internet to be the next cable TV is to set up the architecture to favor broadcasting and inhibit peer to peer connectivity. To unlock more of the potential of the internet we desparately need to force IPv6 as the norm. Both Apple and Microsoft support it in their new OSes and Unix OS's like BSD have supported it for some time. It appears the bottleneck right now is with the ISP's who are loathe to spend money to offer something that people don't even know they will want. In fact since ISP's are generally opposed to users running server applications at all it shouldn't be surprising that they aren't leading the charge to update the infrastructure to enable precisely that. There are many other important characteristics of IPv6 but for the future of P2P the vastly larger address space (64 bit rather than 32 bit) is the most important.
Re:Direct connections? (Score:1)
Here is a paper about scaling and the gnutella protocol written by someone who actually bothers with the salient details: Flow Control by S. Osokine [vadem.com]. I haven't finished reading it yet, but I know Sergei well enough to trust he won't go off on meaningless tangents unlike some other papers.
The reason why you might not have gotten any comment yet is that it is an incredibly busy time for gnutella developers. There is a lot of potential in the original protocol that requires the slogging of lots of code (UI in particular). If a sustained effort is not made to improve a client, one risks losing the necessary traction to achieve a scale worth addressing. Personally I want to abstract out a layer to allow for deployment of other applications on top of the gnutella protocol. But time is limited and coding opportunities infinite. Choices and compromises have to be made. Don't get discouraged if your ideas haven't received the attention they deserve yet. Maintain your web page and engage in conversations on the gdf mailing list and eventually it might get read.
Re:Good Points (Score:2)
Good Points (Score:3)
I like the last one in particular:
As a testement to that my main "workstation" is a K6 233. Workstation is an exageration. I use the 233 mostly for www, and email. I have a more powerfull machine, but it is for games. What do you mean by "be the Internet"? I thought at first that it mean that each person can serve information directly to others. I think that we are pretty close to that now. Many of those people who want to, run good presonal web sites. So I have decided that "to be" the Internet must be something greater. Something that the cyberpunk authors haven't predicted yet. I think that in order "to be" the Internet our interfaces (mostly www browers) have to change completely. In my mind "to be" anything is an active role. Currently I find much of the Internet to be passive in nature. I think our interfaces encourage that.just some random after bed time thoughts
Re:Shirky is a weak writer (Score:2)
Disclaimer: I really don't know who the heck Clay Shirky is and I haven't read much of his writing. This is just an observation.
Re:P2P (Score:1)
To a small degree when applied to internet based systems. But I think we will still find major server types used in businesses for many long years to come. Although it sounds nice where everyone shares with everyone, in a business this ammounts to mass chaos. And concidering that companies would have a harder time backing up files from all over their network, where when it is in a centralized place it keeps the confustion down.
The internet can change rappidly and often, but most companies don't chang that often or that drasticly.
Re:Shirky is a weak writer (Score:1)
That said, an image doth not a GUI make, and Clay should know better.
Re:Shirky is a weak writer (Score:1)
-clay
Re:Shirky is a weak writer (Score:2)
Its was in fact Mosaic that created "Web interface as GUI", by making clickable pictures.
-clay
Re:P2P and NAT (Score:2)
1) Use a proxy that is not behind a NAT firewal. This opens up problems all its own, which is fairly obvious.
2) Use UDP. For a number of reasons, a lot of people hate this idea as well.
The only hope (currently) is one of two things. First, that NAT firewall vendors will implement a suitable solution to bypass this problem (unlikely) and second, that we all get fixed IP addresses out the ass (about as unlikely).
So, in short, break out the wallet for those beefy relay servers!
(or use UDP
Re:Direct connections? (Score:2)
There are a few tweaks which improve the efficiency of this type of network, such as a reputation/affinity value attached to peers to keep you connected to the best, while quickly filtering out the worst or dissimilar.
The communication is multiplexed over a single UDP port and can handle hundreds of thousands of concurrent connections at the lowest layer. (higher level ALPINE connections require more overhead, and are restricted to 10,000 to 100,000 depending on user preference)
At any rate, my point is that you can use a simple packet routing architecture like IP to accomplish a flat, large, directly connected network that is usable.
If you want higher performance, more efficiency, and greater throughput you would need to start experimenting with some of the advanced network architectures you mention. However, the chance of such a network reaching the masses any time soon is pretty slim.
Re:P2P and NAT (Score:2)
1) Peer_A opens a UDP socket. Sends a packet to a well known server, or servers, that simply send a reply that contains the source IP and port of the packet.
2) Peer_A records this source IP and Port, as it is what the NAT gateway is masquerading its connection as.
3) Peer_B does the same. It know has its masqueraded IP and PORT from its NAT gateway.
4) Peer_A and Peer_B can now send packets to each other at their respective masqueraded IP and PORTs.
Re:P2P (Score:1)
Juno's a step ahead of you (Score:2)
They recently changed their terms of service mandating that Juno can use its customers' computers for that very purpose. They can terminate your service if you don't leave your computer on 24x7 so that it can do the processing and dial in to Juno at whim for more data. The fun details are at www.byte.com/column/BYT20010222S0004 [byte.com].
Naturally they portray this as a benevolent thing and a chance to be part of their "Virtual Supercomputing Project," which claims to be completely voluntary, despite the fact that their Terms of Service directly contradict this:
Happy computing! :)
Cheers,
Why is anyone surprised? (Score:2)
Katz is a tabloid journalist. He writes whatever will get the best ratings^W^Wmost comments. People respond accordingly. Now someone from the side he's attacked defends themselves, and we have the same reactions; some defend their position, some attack it.
I fail to see how any of this is really that remarkable. So p2p is the latest buzzword. So what? So long as we have marketroids who have to make quotas or journalists with deadlines, there'll be buzzwords. And wherever there are buzzwords there will be people to attack or defend them.
Personally, I'm going back to play with some cool bleeding edge stuff that might just be involved with a buzzword or two. I don't care, since I think it's cool in and of its own right.
Direct connections? (Score:1)
No topology that comes to mind seems plausible for direct connections for everyone. Currently, the star - substar topology works well, where local nodes are connected to local routers in a star topology (hub, or a bunch of dialup users calling in to one center) which is conneted to its upstream provider in a similar setup (with an ATM bus system at the top level, but this is irrelevant). This system, though theoretically not as resiliant as direct connections, keeps routing tables small and paths relatively small. The idea is to reduce complexity. Most users currently lack the bandwidth to be a user and a router at the same time! When the dream of "broadband in the home" spreads everywhere, perhaps this might be more likely, but right now current topologies satisfy. The only major obstacle is that provider-to-provider connections are not as plentiful as we'd all like. My route to AOL goes through AT&T! You'd think that with a provider my size, they wouldn't have the intermediary netowrk.
Yet I digress.
Re:Direct connections? (Score:1)
I, however, unlike you, will not rant about your lack of thought prior to posting. I had thought about Wireless LANs when I posted, which is why I did not include that.
Re:Get a CLUE, faggot. (Score:1)
[Note for the writer to whom I am responding: The above is sarcastic. I felt the need to explain this since I did not feel you would understand it without an explanation.]
Re:Direct connections? (Score:1)
Moreover, this brings up a philosophical point. Should everyone with an opinion post in response to something? Or only those people who think others care? Everyone should express themselves, IMHO, if nothing more than to have a question answered (or an idea corrected). What is your opinion on this? Please, this time, respond with something more substantial than one line berating my cluelessness.
Re:MOD THIS STUPID GARBAGE DOWN (Score:1)
Re:grow a brain asshole (Score:1)
Re:grow a brain asshole (Score:1)
Re:grow a brain asshole (Score:1)
Are 11th graders there like 11th graders here (I can play too!)
Re:grow a brain asshole (Score:1)
You are right.. but the thing is.. (Score:4)
You and I and lots of others know how the internet works.. we don't like the 'centralized broadcast' way it's starting to be used.. and don't like how people insist that p2p is something 'new'.. but think about this.
For mom & pop jones out there.. it IS something new. Sure they could have always done it.. but are just now realizing it. To them, it's NEW. The applications are new... everything is new. So it'
s good to have articles like this....
Re:fringe now, but commercialized soon (Score:2)
Re:Revolution (Score:2)
Nice, I never thought of it that way
Revolution (Score:3)
It will succeed, however, simply because it's gained enough momentum that it cannot be stopped. And because it cannot be controlled.
Re:P2P (Score:2)
Clay Shirky is a savvy guy, and he has a point that P2P is a good idea, albeit a sloppy one just now. The main thesis of his article is that power resides not only in centered servers (this is Sun's doomed wishful thinking, exactly), but also exists and is growing on the edges, as widespread use of easily acquired and highly capable software triggers the law of large numbers and a shift to topple a tipping point, to utterly overwhelm those few evil hegemonists who seek to exert centralized _control_ in order to extract artificial scarcity based revenues from the large mass of networked connected people. Your anger is better directed at those who charge $15 per CD, $9 per movie.
Re:P2P (Score:2)
Re:Direct connections? (Score:2)
Difficult, but far from impossible. Mesh routing - which is really what you're talking about here - is a much-studied field and pretty reasonable solutions to all of the major problems are known. Check out Routing in the Internet by Christian Huitema for a pretty good overview of the relevant theory and practice.
IMO one of the big problems with P2P is that too many P2P implementors are either ignorant or disdainful of related work in this and other areas - usually both, which is a bad combination. I went to the O'Reilly P2P conference in SF a couple of months ago, and overall it was fantastic, but I did notice one thing. Everyone there seemed very sophisticated about crypto and security etc. but at the same time most were stunningly ignorant about routing, protocol design, performance management, and a bunch of other fields. There were exceptions, of course, don't get me wrong, but there's still way too much fad-following in the P2P community and not enough solid science or engineering.
Re:Direct connections? (Score:2)
That looks like a nice paper. Thanks!
Re:Humbug (Score:2)
I beg to differ. One of the major things differentiating supercomputers from anything else is the presence of *huge* internal memory and I/O bandwidth. That will never be duplicated by a distributed group of machines, and so for a certain very large and important class of programs distributed computation will never be a substitute for supercomputers. Fortunately, there are enough problems out there that *do* partition very easily, so that distributed computing is still worthwhile.
The importance of roles (Score:3)
I think one of the things people need to get straight in their heads wrt P2P is the difference between temporary and permanent roles. Obviously, if node A is always and forever a server, and node B is always and forever a client, and they're totally incapable of switching roles (e.g. neither even has the code to do so) then that's not P2P. Just as obviously, if there's not even a notion of client and server, if every node is necessarily able to perform in either role at any instant, that's about as P2P as you can get. (Note, however, that with respect to an individual transaction there is still one node acting as client/requester and another acting as server/responder).
The real battleground is the area in between, where a node may be a server one moment, or a client the next, changing according to the needs of the network (e.g. nodes entering and leaving). Is that P2P? The P2P purists ("peerier than thou", to use Dr. Shirky's term) would say no. More practical people would say yes, or close enough to yes that it doesn't matter. One trend that more and more people are noticing is that many P2P protocols/applications are developing ideas of "supernodes" or "reflectors" or "defenders" (my candidate for stupidest term yet) that, because of their superior resources, are given additional responsibilities. In other cases, certain functions have been partly or completely centralized within a mostly P2P framework, because nobody could figure out how to make that particular piece - usually a location, searching, or indexing piece - scale within a pure P2P paradigm.
The important thing about P2P is not "oh my god, there's a server, we must eliminate such heresy from our design!" What's important is decentralization and automatic reconfiguration, to avoid bottlenecks and single points of failure. Those are the problems we're trying to solve, remember? If the system is flexible so that work can be redistributed seamlessly from one place to handle either overload or failure, that's "P2P enough for me" even if a picture of the system at any one point in time shows some nodes in server roles and others in client roles. That generally means that each node must be capable of performing the different roles - i.e. the code must be present, the protocols must support it - but whether a given node actually does ever perform a given role doesn't matter.
Re:Revolution (Score:1)
impressive.
Re:Revolution (Score:3)
SuperID
Free Database Hosting [freesql.org]
Humbug (Score:3)
Millions of CPU doing a job of a supercomputer...what do you think a supercomputer is? It is thousands of CPU clustered. Only difference is that a supercomputer is pre-packaged. To have million "regular" PC clusters, it simply won't be able to scale for most applications (with notable exception being SETI@HOME), since the cost of routing information quickly overwelms the useful information that is being passed arround.
He also claims that PCs are second class citizens and they need to be servers. But does he have any idea what this will entail? Think of the security issues...think of the privacy issues...think of the performance issue it will bring to the entire internet. Even if the cost of routing scales linearly, that still sucks because number of computers that are being connected is increasing exponentially.
I'm not here to dismiss p2p - I think p2p will have a great future for some applications. But I think the author needs to think hard about his statements - which I think has very little meaning at all. The article seems too much hot air religious fervor about p2p.
PCs as first-class citizens (Score:4)
Re: Your .sig (even more OT) (Score:1)
***
Re:P2P and NAT (Score:2)
The only solution I can see is to add a layer between TCP and IP. Call it NATCP (or NCP, if you like nested TLAs). The NATCP data would be a chain of DWORD used to map back to a machine on the network. In the simplest implementations, it would be the IP address of the machine on the inner network, or be completely separate from the local IP address. 0.0.0.0 would be reserved for termination.
For those that fear that this takes away from the protection that NAT give the machines on the inner network: that is not the intention of NAT. NAT is used to give many machines access to the internet from a single IP address. Protection is provided by the firewall (which will often be you NAT server).
Note, that while I put this bewteen TCP and IP, it could easily go on top of TCP (although it more logically goes between the two).
For those that say, "why not just switch to IPv6", IPv6 doesn't solve this problem. ISPs will still only dole out a single IP address, because they want you to pay for each machine you connect at home (or at work). In this case, you get a single IP address, and you chain the NATCP address together, so even if you ISP is behind a NAT, their NATCP address gets prepended to yours. You can thus chain NAT networks with ease. This may be a better solution to the IP address shortage and allocation issues than IPv6.
P2P and NAT (Score:3)
Not File Sharing (Score:1)
In the 90's the net (via the web) morphed into the broadcast model. Yahoo or Cnet supplies, I receive. All about content. Imagine if all you were able to use a phone for was to listen to the news or buy something. Kind of sub-optimized.
The P2P (yuck) stuff is for giving people the ability to communicate directly. Sometimes that communication is about content, sometimes its just shooting the shit. The point is, they decide for themselves what they want to use it for, or if they use it at all.
If the technology is useful, then people will figure out how to make money with it. All the bullshit about nobody has a business model is almost totally irellevant to anyone that does not have money invested in any of these startups.
If the technology is usefull, people will use it. If not, they won't.
So a lack of good business models won't have much effect on adoption. Likewise brillaint business models won't save it if the technology is useless.
Re:Juno's a step ahead of you (Score:1)
Re:Revolution (Score:1)
Re:Napster.. (Score:2)
1. I am sure there are many terabytes of movies, music and books available that are no longer protected by copyright (they have been liberated) [danny.oz.au] laws. Distribution via a server based method might be better, more reliable etc, however, with no possibility for profit it would not exist.
2. It might be possible to use some sort of P2P system to distribute the traffic from high traffic websites more evenly around the internet. (what i am thinking about is kind of like caching, however each web browser would become a cache for machines around it... don't know if it is possible)
I am sure there are more...
Re:Revolution (Score:1)
Oh really? That's news to me. Actually there's very little money going into P2P right now. For example, early P2P company Infrasearch just sold out to Sun at a firesale price because they couldn't raise money, even with Marc Andreessen opening VC doors for them.
Since your 8 year old knows about dot bombs, ask him/her to explain current market conditions to you.
details detract from verisimilitude of article (Score:1)
I distinctly remember having to learn the IP stack in 1981. And isn't it Vince Cerf [rl.ac.uk]?
Re:Juno's a step ahead of you (Score:1)
Re:P2P (Score:1)
So remember this the next time your bluetooth palmphone slows down because your ISP is running their billing on it.
Sheesh.
PCs as "First Class Citizens" (Score:4)
Nathaniel P. Wilkerson
Domain Names for $13
Napster as Killer App (Score:2)
Enter Napster. Suddenly, everyone has a reason to think of their computer as no different from all the other computers (even if they are going through a centralized server). It becomes clear that there is great utility in being connected, and having access to other machines, both upstream and downstream. Now that the populus has gotten a taste of this, I doubt they'll go back. Napster will be re-implemented as Espra [espra.net] over Freenet [freenetproject.org], and given the much more generalized architecture, Peer to Peer networks will branch out into all kinds of new spheres of influence. I can't wait to watch it happen!
---
Re:Revolution (Score:5)
The very concept of peer-to-peer has been around since the early days (very early days) of the Internet. ARPANET was originally intended to be decentralized, in case of Global Thermonuclear Warfare (sorry, just finished watching a DVD of War Games
"P2P" (God I hate that stupid buzzword) is just a commercially friendly term to describe something that's already been around for 30 years. I fail to see the revolution here.
With the obvious exceptions of the RIAA and the MPAA, what businesses are trying to destroy the peer-to-peer concept? Hell, not many businesses are even getting into the concept. Why? Because, thanks in part to Napster, businesses don't see a lot of worth in the concept, unless they want to trade MP3s (or porn or movies). Granted, Napster is moving along to a subscription-based service, but there's still no guarantees that, in the end, it'll be successful. Maybe if someone develops a peer-to-peer service (yes, I'm aware of Freenet) that isn't being utilized by people trading Metallica MP3s, Jenna Jameson pictures, and Quicktime files of Gladiator, then maybe one of the big players will jump in.
As I see it, "P2P" (Did I mention that I fucking hate that buzzword?) is just a fad. The concept already exists people, just ask the people who worked on the ARPANET. Giving the concept a hip new acronym and a few evangelizers doesn't make it any bigger of a revolution than it already is.
--
Re:Direct connections? (Score:2)
Re:Juno's a step ahead of you (Score:1)
So, if say I conencted my laptop, which is legal noitced to forbid non corporate usage and I use a Juno accoutn, would that mean Juno are legally liable for illegal access to my machine?
2nd class citizens? (Score:1)
Seperate P2P Network (Score:1)
--
Re:fringe now, but commercialized soon (Score:1)
Juno is already considering using distributed computing to pay for their "free" internet service.
---
The AOL-Time Warner-Microsoft-Intel-CBS-ABC-NBC-Fox corporation:
Computing vs Storage (Score:1)
On a completely different note, P2P is an entirely new method of data havening (a la freenet). I would challenge the admin of any college campus to try to get a somewhat popular mp3 or zip off of a resident network. People have been doing this in colleges since the PC became prevalent in the home.
Re:Revolution (Score:3)
Zog has shells
Og trade Zog
This is the most natural model, and very old.
Then Og moves other side of mountain. Og and tribe appoint a representative to trade with Zog and tribe.
Then smart caveman becomes profesional representative, calls self merchent.
Many merchants ally and become company.
Og and Zog get dsl. Don't need Merchant anymore.
Everything back to normal.
Re:Juno's a step ahead of you (Score:1)
Re:Revolution (Score:2)
In my opinion, there is a war going on between people and corporations over the use and rights inherent in the Internet. People have no clout to stop the corps from doing whatever they want with the net (except for thousands and thousand of posts, which sometimes actually works), but corps with their gaggle of lawyers are quick and eager to point out and punish those who take the net and use it in innovative, if costly ways.
What started out as 1) a wonderful system of free information dissemination has devolved into either 2a) a way to harness each and every transaction into a money-making proposition, or 2b) a way to get really really neat stuff for free.
Of course, it will end up as neither, and the only people once again, who will realize the full potential of the net will be lawyers, as the net becomes the equivalent of modern television. I hope I'm wrong, but P2P may be reduced to nothing because there is no money to be made without that broker between every txn...
Re:Revolution (Score:1)
1. Find you
2. Ensure you have what I want
3. Request you to send it to me
This is a very laborious process. The point about computers is that the do these processes for us. The new "P2P" systems are automating these steps away. Older systems don't do that.
So, P2P is probably the wrong name, but I can't think of a better one
Re:P2P and NAT (Score:1)
This works best in a quasi-P2P system like Napster, which employs central servers, so you can let everyone know which redirected port to connect to. If you have multiple participating machines behind your firewall, you allocate them their own ports and configure the software to give out that port number.
In a fully P2P system, which would require a fixed port for incoming transactions, you can only have one participating machine behind the firewall. The participating machine would then act as a server for the P2P system for the rest of the network. For home systems, this isn't a problem, unless you want to decentralise your family/flatmates too
Re:P2P and NAT (Score:1)
A 'fully peer to peer system' would require that NO system is 'enumerated on the network in some fashion'. I presume you are implying some kind of central database of participants in the peer-to-peer system.
Without such a central database, there are two options for joining the system:
+ You try out candidates to see if they are participants
+ Someone tells you a participant by some offline method
I was thinking along the lines that, in the first case, it would be considerably more efficient to only have IP addresses to 'try' rather than IP addresses and port numbers.
In the second case the system becomes 'members only'.
Hmmm. I think I'll go and read the Freenet documentation again....
Re:P2P (Score:1)
Games like Starcraft work on an almost entirely peer-to-peer basis. The server's only role is in giving the peers a place to match up with each other - much like napster. And I think starcraft qualifies as a commercial success [gamespy.com], partly because of the online multiplayer scene.
--
Re: Your .sig (even more OT) - myoff-topic reply (Score:1)
I doubt it...but it does still crack me up. Anyway, the game is called BattleRangers, and I found the image zanyvg.overclocked.org [overclocked.org]. Most of the stuff featured there is boring but I should give credit since that's where I stole the image from myself. I put the image on my roommate's computer so I could get a shorter url that would fit within the 120 char sig limit.
--
Re:Revolution (OT) (Score:2)
Actually that was Joseph Kennedy, who sold in the summer of 1929, supposedly because a shoeshine boy gave him a stock tip. And the story is probably totally apocryphal - Kennedy himself denied it [nctimes.com].
--
Shirky is a weak writer (Score:4)
A few paragraphs later his theme is "big, sloppy ideas." Yeah, fine, he gets big and sloppy about his ideas of the past, and then parallels that distortion to a present he doesn't begin to define, and this passes for analysis? In a really vague way he may be waving his arms in the right direction, but why are we even trying to listen to someone whose prattling skirts close to the edge of intellectual dishonesty? It's like those old "make millions from the Internet" spams. Sure, you could make millions back in the day, by not by following the advice in those missives. It's because the likes of Shirky have been listened to by too many VCs and editors that the tech economy is so shakey now - false intelligence is more dangerous than ignorance.
Re:P2P and NAT (Score:2)
Re:P2P and NAT (Score:2)
I meant that Napster is not entirely P2P, of course.
....Joe User... (Score:1)
Don't get me wrong, but just bundle such services with the latest version of Windows and turn them "on" by default. That's how a lot of things considered "standard on a PC" became "standard". The Joe User won't even notice.
Actually, be honest: Linux distribs do this all the time too, I yet have to see a Linux distribution that doesn't come with a webserver enabled by default (Even the small 150 Meg distrib I use has one in it, which I disabled of course).
Re:P2P (Score:2)
Not to be rude, but I would say that that defines a revolution for almost anything. Once anyone can publish books (printing press) or own a telephone etc it does become a sort of revolution eventually changing society. Of course, plenty of stuff becomes super widely available but doesn't really change our lives, but...
Exactly (Score:1)
Re:Shirky is a weak writer (Score:1)
Re:So what in heck is this P2P thing? (Score:1)
Re:Revolution (Score:1)
I guess NextPage is thinking EXACTLY the opposite.
Re:Napster.. (Score:2)
I dunno - for all the the high-brow talk, is there another use for distributed file sharing? Gnutella and Freenet, as far as I can tell, consist of illegally shared MP3s and the same porn that's been passed around Hotline for years. (How many copies of gym.mpg do I need?) Besides being slow, buggy and unstable.
To my mind, if you have text documents you want to share, you put them on a web page. If you want to distribute demo MP3's for your band, a web page is definitely the way to go. If you're a political dissident, it seems to me that sharing from your computer is the last thing you do. You send the files to someone in a free country to make them available -- on a web page or FTP site.
Am I missing something? Of course, everyone here had all sorts of pious explanations for what they were using Napster for. Uh, yeah right.
If there are interesting, non-warez uses for P2P file sharing that are better than server-based methods, please enlighten me! I'm going to bed now, though, so flames and accusations of being a paid RIAA agent will go unnoticed (unless you're Freddy Krueger).
Unsettling MOTD at my ISP.
IRC, ICQ, Napster, ... are they P2P? (Score:1)
I did not look up the definition, but for me _true_ P2P would be a network of computers without need of a dedicated central server to accomplish the task they are up to.
Re:Computing vs Storage (Score:1)
Napster.. (Score:2)
New? What's new? (Score:2)
I fail to understand the hype about this "new" peer-to-peer stuff. In 1983/84, I set up a group of Apollo workstations. There was no server. Symbolic links allowed the hard disks on the workstations to act like one big disk, so everything appeared local to everyone. There was no server. That was over 17 years ago. What's "new" about it?
Oh, goody. PCs won't be "second-class citizens" anymore. Hmm. In 1996, the PCs (mostly Pentiums, but including a 486 or two) in my small company were all connected to the net. One of the Win95 machines ran an IRC server. Linux boxes ran FTP and various Java client and server apps and bots. The only tasks reserved for the "big servers" (Suns) were DNS, RealAudio (which I later ran from a Pentium as well), and the main Web site. Everything else was distributed, and mostly to PCs.
So it's not new, and it's not innovative. What's the big deal?
p2p = higher function of the internet brain (Score:3)
Please Remove Head from Ass (Score:2)
amount Van Gogh received from sales of all but 1 painting = $0.00
copyright fees collected by Bach = $0.00
amount they received in sampling/licensing fees = $0.00
amount (currency converted to USD and adjusted for inflation)
Re:That should read: (Score:2)