Slashdot is powered by your submissions, so send in your scoop

 



Forgot your password?
typodupeerror
×
GUI Operating Systems Programming Software Unix IT Technology

Rob Pike Responds 284

He starts by clearing up my error in saying he was a Unix co-creator in the original Call For Questions. From there he goes on to answer your questions both completely and lucidly. A refreshing change from the politicians and executives we've talked to so much recently, no doubt about it.


Pike:
First, let me clear up a misstatement. I am not a co-creator of Unix. I suppose I am described that way because I am co-author (with Brian Kernighan) of a book about Unix, but neither Brian nor I would want to take credit for creating Unix. Ken Thompson and Dennis Ritchie created Unix and deserve all the credit, and more. I joined their group - the Computing Science Research Center of Bell Labs - after 7th Edition Unix had come out.

1) Innovation and patents - by Zocalo
With so many of your ideas being used with such ubiquity in modern operating systems, what is your stance on the issue of patenting of software and other "intellectual property" concepts? Assuming that business isn't going to let IP patents go away as they strive to build patent stockpiles reminiscent of the nuclear arms buildup during the cold war, how would you like to see the issue resolved?


Pike:
Comparing patents to nuclear weapons is a bit extreme.

2) Systems research - by asyncster
In your paper, systems software research is irrelevant, you claim that there is little room for innovation in systems programming, and that all energy is devoted to supporting existing standards. Do you still feel this way now that you're working at Google?

Pike:
I was very careful to define my terms in that talk (it was never a paper). I was speaking primarily about operating systems and most of what I said then (early 2000) is still true.

Here at Google the issues are quite different. The scale of the problem we're trying to solve is so vast there are always challenges. I find it interesting that the slide in that talk about 'Things to Build' is a close match to the stuff we're doing at Google, if you squint a bit. To summarize:

GUI: Google put the cleanest, prettiest UI on the internet and work continues to find new ways to present data and make it easy to explore.

Component architectures: We use a number of big (BIG!) piece parts like the Google File System (GFS) and MapReduce (see the paper by Jeff Dean and Sanjay Ghemawat in the upcoming OSDI http://labs.google.com/papers/mapreduce.html) to build massive engines for processing data. Using those pieces we can harness zillions of machines with a few keystrokes to attack a problem like indexing the entire internet. (OK, it's not quite that easy, but it's still amazing.) I have a daily test job I run to monitor the health of one of the systems I'm developing; it uses a week of CPU time but runs for only a few minutes of real time.

Languages for distributed computing: I'm part of a team working on something along those lines that we hope to write up soon.

Bringing data to the user instead of the other way around: Those damn browsers are still in the way, but other ways of connecting to data are starting to appear, things like the Google API. However, the surface is barely scratched on this topic.

System administration: Google's production people are phenomenal at keeping all those machines humming and ready for your queries. They demonstrated that there was real progress to be made in the field of system administration, and they continue to push forward.

3) Back in The Day - by Greyfox
Were programmers treated as hot-pluggable resources as they are today? There seems to be a mystique to the programmer prior to about 1995.

From reading the various netnews posts and recollections of older programmers, it seems like the programmer back then was viewed as something of a wizard without whom all the computers he was responsible for would immediately collapse. Has anything really changed or was it the same back then as it is now? I'm wondering how much of what I've read is simply nostalgia.


Pike:
Isn't it just that today there are a lot more computers, a lot more programmers, and most people are familiar with what computers and programmers do? I'm not sure I understand your reference to 1995, but twenty or thirty years ago, computers were big expensive temples of modernity and anyone who could control their power was almost by definition a wizard. Today, even musicians can use computers (hi gary).

4) What are you doing... - by Mark Wilkinson
Google employees are apparently allowed to work on their own projects 20% of the time. Given that you probably can't comment on what you're doing for Google, what are you doing to fill the other 20%?


Pike:
One of the most interesting projects out there, one I am peripherally (but only peripherally) involved with, is the Large Synoptic Survey Telescope http://www.lsst.org, which will scan the visible sky to very high angular precision, in multiple colors, many times a year. It's got an 8.4 meter aperture and 10 square degree field, taking an image every 20 seconds with its 3 gigapixel (sic) camera. The resulting data set will be many petabytes of image and catalog data, a data miner's dream. The software for the telescope is as big a challenge as the instrument itself; just the real-time pixel pipeline on the mountain will make today's compute clusters look wimpy.

5) Database filesystems - by defile
The buzz around filesystems research nowadays is making the UNIX filesystem more database-ish. The buzz around database research nowadays is making the relational database more OOP-ish.

This research to me sounds like the original designers growing tired of the limitations of their "creations" now that they're commodities and going back to the drawing board to "do things right this time". I predict the reinvented versions will never catch on because they'll be too complex and inaccessible.

Of course, this second system syndrome isn't just limited to systems. It happens to bands, directors, probably in every creative art.

I think what we've got in the modern filesystem and RDBMS is about as good as it gets and we should move on. What do you think?


Pike:
This is not the first time databases and file systems have collided, merged, argued, and split up, and it won't be the last. The specifics of whether you have a file system or a database is a rather dull semantic dispute, a contest to see who's got the best technology, rigged in a way that neither side wins. Well, as with most technologies, the solution depends on the problem; there is no single right answer.

What's really interesting is how you think about accessing your data. File systems and databases provide different ways of organizing data to help find structure and meaning in what you've stored, but they're not the only approaches possible. Moreover, the structure they provide is really for one purpose: to simplify accessing it. Once you realize it's the access, not the structure, that matters, the whole debate changes character.

One of the big insights in the last few years, through work by the internet search engines but also tools like Udi Manber's glimpse, is that data with no meaningful structure can still be very powerful if the tools to help you search the data are good. In fact, structure can be bad if the structure you have doesn't fit the problem you're trying to solve today, regardless of how well it fit the problem you were solving yesterday. So I don't much care any more how my data is stored; what matters is how to retrieve the relevant pieces when I need them.

Grep was the definitive Unix tool early on; now we have tools that could be characterized as `grep my machine' and `grep the Internet'. GMail, Google's mail product, takes that idea and applies it to mail: don't bother organizing your mail messages; just put them away for searching later. It's quite liberating if you can let go your old file-and-folder-oriented mentality. Expect more liberation as searching replaces structure as the way to handle data.

6) Thoughts on Bell Labs - by geeber
Plan 9, Unix and so many other great things came out of Bell Labs. Since the crash of the internet bubble, telecom companies have suffered immensely. One of the results of this is that Lucent has systematically dismantled one of the world's greatest industrial research facilities. You spent a great part of your career at Bell Labs. What are your thoughts about the history and future (if any) of Bell Labs, and how did the culture of the Labs influence the growth of Unix?


Pike:
It's unfair to say `systematically dismantled', as though it was a deliberate process and there's nothing left. A more honest assessment might be that changes in the market and in government regulation made it harder to keep a freewheeling research lab thriving at the scale of the old Bell Labs. Bell Labs Research is much smaller these days, but there are still some very bright people working there and they're doing great stuff. I hope one day to see Bell Labs restored to its former glory, but the world has changed enough that that may never happen.

I could go on for pages about the old Bell Labs culture, but I must be brief. When I arrived, in 1980, the Computing Science Research Center, also known as 127 (later 1127; therein lies a tale) had recently launched 7th Edition Unix and the Center, after a long period of essentially zero growth, was just entering a period of rapid expansion. That expansion brought in a lot of new people with new ideas. I was a graphics guy then, and I hooked up with Bart Locanthi, another graphics guy, and we brought graphics to Research Unix with the Blit. Other folks brought in new languages, novel hardware, networking; all kinds of stuff. That period in the early 80s generated a lot of ideas that influenced Unix both within the Labs and in the outside community. I believe the fact that the Center was growing was a big part of its success. The growth not only provided new ideas, it also generated a kind of enthusiasm that doesn't exist in the steady state or in a shrinking group. Universities harness a variant of that energy with the continuous flow of graduate students; in industrial research you need to create it in other ways.

One odd detail that I think was vital to how the group functioned was a result of the first Unix being run on a clunky minicomputer with terminals in the machine room. People working on the system congregated in the room - to use the computer, you pretty much had to be there. (This idea didn't seem odd back then; it was a natural evolution of the old hour-at-a-time way of booking machines like the IBM 7090.) The folks liked working that way, so when the machine was moved to a different room from the terminals, even when it was possible to connect from your private office, there was still a `Unix room' with a bunch of terminals where people would congregate, code, design, and just hang out. (The coffee machine was there too.) The Unix room still exists, and it may be the greatest cultural reason for the success of Unix as a technology. More groups could profit from its lesson, but it's really hard to add a Unix-room-like space to an existing organization. You need the culture to encourage people not to hide in their offices, you need a way of using systems that makes a public machine a viable place to work - typically by storing the data somewhere other than the 'desktop' - and you need people like Ken and Dennis (and Brian Kernighan and Doug McIlroy and Mike Lesk and Stu Feldman and Greg Chesson and ...) hanging out in the room, but if you can make it work, it's magical.

When I first started at the Labs, I spent most of my time in the Unix room. The buzz was palpable; the education unparalleled.

(And speaking of Doug, he's the unsung hero of Unix. He was manager of the group that produced it and a huge creative force in the group, but he's almost unknown in the Unix community. He invented a couple of things you might have heard of: pipes and - get this - macros. Well, someone had to do it and that someone was Doug. As Ken once said when we were talking one day in the Unix room, "There's no one smarter than Doug.")

7) Languages - by btlzu2

Hello!

Maybe this is an overly-asked question, but I still often ponder it. Does object-oriented design negate or diminish the future prospects of Unix's continuing popularity?

I've developed in C (which I still love), but lately, I've been doing a lot of purely object-oriented development in Java. Using things like delegation and reusable classes have made life so much easier in many respects. Since the *nixes are so dependent upon C, I was wondering what future you see in C combined with Unix. Like I said, I love C and still enjoy developing in Unix, but there has to be a point where you build on your progress and the object-oriented languages, in my opinion, seem to be doing that.

Thank you for all your contributions!!!


Pike:
The future does indeed seem to have an OO hue. It may have bearing on Unix, but I doubt it; Unix in all its variants has become so important as the operating system of the internet that whatever the Java applications and desktop dances may lead to, Unix will still be pushing the packets around for a quite a while.

On a related topic, let me say that I'm not much of a fan of object-oriented design. I've seen some beautiful stuff done with OO, and I've even done some OO stuff myself, but it's just one way to approach a problem. For some problems, it's an ideal way; for others, it's not such a good fit.

Here's an analogy. If you want to make some physical artifact, you might decide to build it purely in wood because you like the way the grain of the wood adds to the beauty of the object. In fact many of the most beautiful things in the world are made of wood. But wood is not ideal for everything. No amount of beauty of the grain can make wood conduct electricity, or support a skyscraper, or absorb huge amounts of energy without breaking. Sometimes you need metal or plastic or synthetic materials; more often you need a wide range of materials to build something of lasting value. Don't let the fact that you love wood blind you to the problems wood has as a material, or to the possibilities offered by other materials.

The promoters of object-oriented design sometimes sound like master woodworkers waiting for the beauty of the physical block of wood to reveal itself before they begin to work. "Oh, look; if I turn the wood this way, the grain flows along the angle of the seat at just the right angle, see?" Great, nice chair. But will you notice the grain when you're sitting on it? And what about next time? Sometimes the thing that needs to be made is not hiding in any block of wood.

OO is great for problems where an interface applies naturally to a wide range of types, not so good for managing polymorphism (the machinations to get collections into OO languages are astounding to watch and can be hellish to work with), and remarkably ill-suited for network computing. That's why I reserve the right to match the language to the problem, and even - often - to coordinate software written in several languages towards solving a single problem.

It's that last point - different languages for different subproblems - that sometimes seems lost to the OO crowd. In a typical working day I probably use a half dozen languages - C, C++, Java, Python, Awk, Shell - and many more little languages you don't usually even think of as languages - regular expressions, Makefiles, shell wildcards, arithmetic, logic, statistics, calculus - the list goes on.

Does object-oriented design have much to say to Unix? Sure, but no more than functions or concurrency or databases or pattern matching or little languages or....

Regardless of what I think, though, OO design is the way people are taught to think about computing these days. I guess that's OK - the work does seem to get done, after all - but I wish the view was a little broader.

8) One tool for one job? - by sczimme
Given the nature of current operating systems and applications, do you think the idea of "one tool doing one job well" has been abandoned? If so, do you think a return to this model would help bring some innovation back to software development?

(It's easier to toss a small, single-purpose app and start over than it is to toss a large, feature-laden app and start over.)


Pike:
Those days are dead and gone and the eulogy was delivered by Perl.

9) Emacs or Vi? - by Neil Blender

Pike:

Neither.

When I was a lad, I hacked up the 6th Edition ed with Tom Duff, Hugh Redelmeier, and David Tilbrook to resuscitate qed, the editor Ken Thompson wrote for CTSS that was the inspiration for the much slimmer ed. (Children must learn these things for themselves.) Dennis Ritchie has a nice history of qed at http://cm.bell-labs.com/cm/cs/who/dmr/qed.html> .

I liked qed for one key reason: it was really good at editing a number of files simultaneously. Ed only handled one file at a time.

Ed and qed were command-driven line editors designed for printing terminals, not full-screen displays. After I got to Bell Labs, I tried out vi but it could only handle one file at a time, which I found too limiting. Then I tried emacs, which handled multiple files but much more clumsily than qed. But the thing that bothered me most about vi and emacs was that they gave you a two-dimensional display of your file but you had only a one-dimensional input device to talk to them. It was like giving directions with a map on the table, but being forced to say "up a little, right, no back down, right there, yes turn there that's the spot" instead of just putting your finger on the map.

(Today, emacs and vi support the mouse, but back in 1980 the versions I had access to had no support for mice. For that matter, there weren't really many mice yet.)

So as soon as the Blit started to work, it was time to write an editor that used the mouse as an input device. I used qed (mostly) and emacs (a little) to write the first draft of jim, a full-screen editor that showed you text you could point to with a mouse. Jim handled multiple files very smoothly, and was really easy to use, but it was not terribly powerful. (Similar editors had been at Xerox PARC and other research labs but, well, children must learn these things for themselves.)

A few years later I took the basic input idea of jim and put a new ed-like command language underneath it and called it sam, a locally popular editor that still has its adherents today. To me, the proof of sam's success was that it was the first full screen editor Ken Thompson liked. (He's still using it.) Here's the SP&E paper about sam from 1987: http://plan9.bell-labs.com/sys/doc/sam/sam.pdf.

A few years later, I decided the pop-up menu model for commanding an editor with a mouse was too restrictive, so I started over and built the much more radical Acme, which I'm using to write these answers. Here's the Acme paper: http://plan9.bell-labs.com/sys/doc/acme/acme.pdf

I don't expect any Slashdot readers to switch editors after reading these papers (although the code is available for most major platforms), but I think it's worth reading about them to see that there are ways of editing - and working - that span a much larger gamut than is captured by the question, 'Emacs or vi?'

10) Biggest problem with Unix - by akaina
Recently on the Google Labs Aptitude Test there was a question: "What's broken with Unix? How would you fix it?"

What would you have put?


Pike:
Ken Thompson and I started Plan 9 as an answer to that question. The major things we saw wrong with Unix when we started talking about what would become Plan 9, back around 1985, all stemmed from the appearance of a network. As a stand-alone system, Unix was pretty good. But when you networked Unix machines together, you got a network of stand-alone systems instead of a seamless, integrated networked system. Instead of one big file system, one user community, one secure setup uniting your network of machines, you had a hodgepodge of workarounds to Unix's fundamental design decision that each machine is self-sufficient.

Nothing's really changed today. The workarounds have become smoother and some of the things we can do with networks of Unix machines are pretty impressive, but when ssh is the foundation of your security architecture, you know things aren't working as they should.

Looking at things from a lower altitude:

I didn't use Unix at all, really, from about 1990 until 2002, when I joined Google. (I worked entirely on Plan 9, which I still believe does a pretty good job of solving those fundamental problems.) I was surprised when I came back to Unix how many of even the little things that were annoying in 1990 continue to annoy today. In 1975, when the argument vector had to live in a 512-byte-block, the 6th Edition system would often complain, 'arg list too long'. But today, when machines have gigabytes of memory, I still see that silly message far too often. The argument list is now limited somewhere north of 100K on the Linux machines I use at work, but come on people, dynamic memory allocation is a done deal!

I started keeping a list of these annoyances but it got too long and depressing so I just learned to live with them again. We really are using a 1970s era operating system well past its sell-by date. We get a lot done, and we have fun, but let's face it, the fundamental design of Unix is older than many of the readers of Slashdot, while lots of different, great ideas about computing and networks have been developed in the last 30 years. Using Unix is the computing equivalent of listening only to music by David Cassidy.

11) Re: Plan9 - by Spyffe

Rob,

Right now, there are a large number of research kernels. Plan 9, Inferno, AtheOS, Syllable, K42, Mach, L4, etc. all have their own ideas about the future of the kernel. But they all end up implementing a POSIX interface because the UNIX userland is the default.

The kernel space needs to be invigorated using a new userland that demands new and innovative functionality from the underlying system. Suppose you were to design a user environment for the next 30 years. What would the central abstractions be? What sort of applications would it support?


Pike:
At the risk of contradicting my last answer a little, let me ask you back: Does the kernel matter any more? I don't think it does. They're all the same at some level. I don't care nearly as much as I used to about the what the kernel does; it's so easy to emulate your way back to a familiar state.

Applications - web browsers, MP3 players, games, all that jazz - and networks are where the action is today, and aside from irritating little incompatibilities, the kernel has become a commodity. Almost all the programs I care about can run above Windows, Unix, Plan 9, and on PCs, Macs, palmtops and more. And that, of course, is why these all have a POSIX interface: so they can support those applications.

And then there's the standard network protocols to glue things together. It's all a uniform sea of interoperability (and bugs).

I think the future lies in new hardware as much as in new software. A generation from now machines will be so much more portable than they are now, so much more powerful, so much more interactive that we haven't begun to think about the changes they will bring. This may be the biggest threat to Microsoft: the PC, the desktop, the laptop, will all go the way of the slide rule. As one example, when flexible organic semiconductor displays roll out in a few years, the transformation in how and where people use computers and other devices will be amazing. It's going to be a wild ride.

===============
This discussion has been archived. No new comments can be posted.

Rob Pike Responds

Comments Filter:
  • Re:Damn. (Score:2, Insightful)

    by Anonymous Coward on Monday October 18, 2004 @12:11PM (#10556822)
    Maybe because he's getting tired of this issue? Maybe he wants to focus on actual code instead of politics?
  • by fireman sam ( 662213 ) on Monday October 18, 2004 @12:19PM (#10556869) Homepage Journal
    "Using Unix is the computing equivalent of listening only to music by David Cassidy"

    To continue the musical comparison. Windows, 15 different variations of the same mass produced pop song whos only existance is to make money for a company that already has a lot of money.

    I'll take David Cassidy, even if he has a CLI only.

  • by Black-Man ( 198831 ) on Monday October 18, 2004 @12:19PM (#10556875)
    C'mon... comparing corporate IP/Patents to the nuclear arms race? That kind of flawed reasoning works on slashdot, but not with anyone out in the real world.

    He gave an appropriate response to a STUPID analogy.

  • flamebait, robin (Score:0, Insightful)

    by Anonymous Coward on Monday October 18, 2004 @12:21PM (#10556889)
    refreshing change from the politicians and executives we've talked to so much recently

    Except that Slashdot refuses to confront the real business and political issues facing programmers.
    The editors' story selection betrays their biases toward ignorance and denial. There's too many puff pieces on Apple and KDE and not enough on how industry lobbyists have stacked the deck against American labor.

  • Re:Damn. (Score:1, Insightful)

    by Anonymous Coward on Monday October 18, 2004 @12:24PM (#10556914)

    Well that was a complete and total ignoring of the intent of the patent question on the basis of not agreeing with a minor portion of the question.

    You seem to be forgetting: with every question comes a set of givens. It simply appears that Rob didn't agree with the givens upon which the question is based.

  • by Anonymous Coward on Monday October 18, 2004 @12:24PM (#10556915)
    stacked the deck against american labor? whatever. we have a higher standard of living here then China does. it is more expensive to employ Americans.

  • Is that sarcasm? (Score:4, Insightful)

    by Anonymous Coward on Monday October 18, 2004 @12:24PM (#10556918)
    Hmm...The summary claims:

    From there he goes on to answer your questions both completely and lucidly. A refreshing change from the politicians and executives we've talked to so much recently, no doubt about it.

    The actual interview says:

    Pike:
    Comparing patents to nuclear weapons is a bit extreme.


    Clearly the summary is being sarcastic...
  • Re:Damn. (Score:5, Insightful)

    by jd ( 1658 ) <imipak@yahoGINSBERGo.com minus poet> on Monday October 18, 2004 @12:26PM (#10556930) Homepage Journal
    I have to agree. Sure, I'll go along with one of the ACs who said that some things don't dignify a reply, but software patents are a major element in today's world.


    The one-click patent is a symptom of the problem, but the K5 debate on applying the logic of the DMCA to patents is a symptom of the attitudes. Attitudes won't change, just because we're tired of them. They'll change when those who ARE tired of them propose a workable, viable alternative that meets the needs of industry and inventors.


    It's obvious enough that he knows that walking the walk is important - that's one reason he developed Plan 9! Kevin Mitnick proved, very conclusively, that computer security and data integrity are vulnerable to the foolishness of mere mortals. Redesigning on this scale is more than just rewriting some code. the sort of redesign Plan 9 represents is about seeing what doesn't work, and replacing it with something that does.


    Attitudes are broken. They need patching or replacing. Keep them the same, and all the software fixes in the world won't secure a single computer.

  • Re:Nice (Score:5, Insightful)

    by jd ( 1658 ) <imipak@yahoGINSBERGo.com minus poet> on Monday October 18, 2004 @12:29PM (#10556952) Homepage Journal
    I agree. A nuclear weapon can only be used to attack someone once and the fallout is confined to a few thousands of square miles.
  • Re:Damn. (Score:2, Insightful)

    by tjic ( 530860 ) on Monday October 18, 2004 @12:30PM (#10556955) Homepage
    It seemed quite clear to me that he was saying "the question was idiotic. I'm not going to call you an idiot, but I suggest that you reevaluate your axioms." Seems like a reasonable response to me.
  • by FreeUser ( 11483 ) on Monday October 18, 2004 @12:31PM (#10556961)
    Maybe because he's getting tired of this issue? Maybe he wants to focus on actual code instead of politics?

    And how, pray tell, is he going to do that when all but the most trivial code runs afoul of patents and is vulnerable to litigation? (According to many analysts, this is already the case.)

    Refusing to answer the question and using disagreement with the analogy used by the questioner as cover is an exceedingly political answer (and a tried and true method of dodging uncomfortable questions used by virtually every political candidate for office in recent years, as alluded to the "is he running for office" comment) ... dismissing the issue on such a weak pretense clearly amounts to taking sides on the issue, namely the side of the status quo, i.e. pro software patents.

    Hardly a non-political stance, merely a disingenuous one.
  • by Anoraknid the Sartor ( 9334 ) on Monday October 18, 2004 @12:34PM (#10556980)
    I took the point to be that some companies may feel they have to build up a patent portfolio merely so that they have something to wave at another company that attacks them with THEIR patent portfolio.

    A kind of "mutually assured destruction" stance...

    As such, the analogy with the reasoning that lay behind the nuclear arms race seems quite apt.

    The dismissal of the question does rather suggest that the speaker did not want to address the point at issue.
  • by postbigbang ( 761081 ) on Monday October 18, 2004 @12:36PM (#10557001)
    There's no congruity between IP issues and nuclear weapon stockpiling. Nuclear weapons are mass destruction devices. IP protection embues certain rights under various juridictions. There might be very important issues for the questioner in IP, but the question was worded poorly and was presumed to foster a baited answer. The context was poorly set, and the answer put the question in the nebulous context by which it was asked. Good answer.
  • The Unix Room (Score:5, Insightful)

    by joelethan ( 782993 ) on Monday October 18, 2004 @12:37PM (#10557010) Journal
    I was particularly struck by the story of the Unix Room where all the Unix people hung out.

    These days, developers seem to have their accommodation organised by blind chance, or worse, corporate whim.

    Many of my colleagues left their 6-12 man offices to join a 70 desk open-plan floor. The six of us architects (yeah, right) were pretty miffed to be shunted into a 1980's room just for six with beige vinyl on the walls and phones straight out of Flash Gordon. Now, two months later, we appreciate the working community that is our office.

    Good call Mr. Pike: humans function well in small self-organising or randomly-organised groups of up to 8. I'll rue the day we have to move out.

    /JE

  • by FreeUser ( 11483 ) on Monday October 18, 2004 @12:41PM (#10557043)
    Perhaps Pike's background with non-academic (i.e. commercial) research centers causes him to think about patents in a different light than, say, Stallman's background in working at MIT in the '70s (an academic research center, among other things).

    Well, since he dodged the question with a disingenuous slam of the questioner, using his disagreement with the questioner's analogy as cover to do so, we really don't know the answer to that. Based on his unwillingness to answer the question and defend his point of view (which one may surmise based on previous behavior and his dismissal of software patents as an issue worthy of addressing, is pro-software patent) we can guess that his perspective does differ from most in both the industry and academia (including Stallman), but with his refusal to answer the question we really don't know.
  • by Exmet Paff Daxx ( 535601 ) on Monday October 18, 2004 @12:44PM (#10557066) Homepage Journal
    Large companies are stockpiling software patents in exactly the same mutually-assured-destruction mindset anticipated in the cold war: If you sue me to death, I'll sue you to death. They even have the same peace treaties: I promise not to sue you with my patents if you promise the same. You could call them Patent Noproliferation Pacts.

    The fact that that question was sent to the interviewee meant that Slashdot's readers wanted to know his opinion of the patent system. He could have answered it in any manner he chose, but he chose to sidestep it instead because his employer (Google) believes in using patents aggressively in a mutually-assured-destruction way, even if it means the end of Linux. That is why he didn't answer, and your faux-objective pseudointellectual babble isn't fooling anyone.
  • by MattW ( 97290 ) <matt@ender.com> on Monday October 18, 2004 @12:45PM (#10557067) Homepage

    Pike:
    Comparing patents to nuclear weapons is a bit extreme.


    No it isn't. The comparison is drawn often, because both large patent portfolios as well as large nuclear arms stockpiles create a situaiton of Mutually Assured Destruction. Once the nukes start flying, nobody wins. Likewise, once the lawyers start slinging patent lawsuits, only the lawyers win.

    So the answer may be, "I have no idea", but the comparison is legitimate.
  • Re:Damn. (Score:3, Insightful)

    by ProfKyne ( 149971 ) on Monday October 18, 2004 @12:47PM (#10557091)

    I wondered about this too, surely he doesn't think the OP was suggesting patents are comparable to nukes -- the question referenced the way in which large corps gobble down patents, often with no immediate designs to follow through on a business plan or implementation.

    Pike does, however, mention that he works at Google, so maybe he interpreted the question as a shot against his employer or was simply advised by Google's PR not to answer the question.

  • by Forbman ( 794277 ) on Monday October 18, 2004 @12:48PM (#10557102)
    Actually, the analogy isn't, in a business sense.

    The goal of Microsoft, IBM, etc. (and a few smaller players and non-entities) is partly to gain as much IP control as possible, to not only avoid having to be beholden to some other IP-holding company, but also to use it to control competition, if required.

    The end-game is different, but what ends up is a business equivalent of nuclear detente, where the major players have enough weapons to counter any attack in court.

    Nuclear detente would have been the better analogy, not nuclear war.
  • by Anonymous Coward on Monday October 18, 2004 @12:59PM (#10557200)
    You got a real answer. He said that era is dead.

    You may not like his answer and may disagree but that does not make his answer any less valid. He said that one tool for one job is dead. If you want a different answer then ask someone else. Usually if you ask enough people you can find one that agrees with you. Hey, when that happens you can then say that you theory is correct becasue X agrees with you and ignore the droves of others who disagree. Comeom, all the kids are doing it. manipulate the data to make yourself valid. It is what the kewl kids do.

    Ok enough peer preesure. (that was a joke by the way) but please don't say he did not give a real answer, just say you disagree. One invalidates his answer the other refutes it. Big difference.

  • by GoofyBoy ( 44399 ) on Monday October 18, 2004 @01:06PM (#10557272) Journal
    >I'm not so sure that the analogy to MAD ... is all that extreme.

    But it leads the question. You put patents in the light of "pure evil" and so come off as having your own agenda and not an honestly-interested-in-the-answer question.

    Its like bringing up Nazi's in a converstation.
  • by graveyhead ( 210996 ) <fletchNO@SPAMfletchtronics.net> on Monday October 18, 2004 @01:09PM (#10557302)
    OK I have to agree with you.

    My first response was a bit over-emotional.

    I simply meant this: his answer seemes very glib to me. It would simply be nice if he had elaborated a bit :/
  • Re:The Unix Room (Score:4, Insightful)

    by nthomas ( 10354 ) on Monday October 18, 2004 @01:12PM (#10557351)
    I was particularly struck by the story of the Unix Room where all the Unix people hung out.

    Fascinating.

    I was at Columbia University last week for a meeting sponsored by the local ACM chapter [columbia.edu] and LXNY [lxny.org], the speaker was Stephen Bourne [eldoradoventures.com] (he who is sh [wikipedia.org]).

    At some point during his excellent talk on the history of Unix and his place in it, someone asked what he thought was the reason for the success of the operating system, and without hesitating, he talked about the room where all the terminals were located (he never specifically referred to it as the "Unix room" though) and how when you released software it was used immediately by those in the room and if something broke, you were called "idiot" (and probably worse) by your peers -- it was in your best interest to make sure you didn't put out junk as you really didn't have that dilution of responsibility that engineers have in a large corporation where the design team is in one wing of the building, the coders are in another, and the testers in yet another location, etc.

    It was a great speech, anyone who hasn't seen Dr. Bourne speak should do so, he is an excellent source of insight into the early years of Unix and software engineering in general. He is now working for a venture capital firm and roughly a third of his talk was spent talking about that, it's a testament to his great speaking skills that most of the people in the room didn't lose interest when he switched topics like that (I'm convinced that most hackers suffer from ADD).

    Thomas

  • by null etc. ( 524767 ) on Monday October 18, 2004 @01:17PM (#10557395)
    Well that was a complete and total ignoring of the intent of the patent question on the basis of not agreeing with a minor portion of the question.
    Thus, today's lesson is: don't insert your own stupid analogies into the question just to appear intelligent.

    I would have loved to see his response to the same question without the analogy. He would have been forced to answer, or explicitly acknowledge his dodging, if the submitter had merely posed the question by itself.

    Obviously, as an employee of a corporation with major intellectual property interests, he's not going to kill his career by speaking out against software patents. You gave him an easy out.
  • Re:Damn. (Score:1, Insightful)

    by bat'ka makhno ( 207538 ) on Monday October 18, 2004 @01:27PM (#10557494)
    advised by Google's PR not to answer the question

    My guess is that's exactly what happened. As a large technology company, Google has a vested interest in perpetuating the perversion that passes for our patent system. As the company's representative, Pike could not express an opinion that would have either 1) gone against the company's stance on patents or 2) plainly stated the company's support for abusive patent legislation, thereby diminishing the company's mindshare among slashdot readership.

    While not a surprise, I would've prefered Pike not to answer the question at all, rather than dismiss the excellent point with a snarky, immature remark.

  • Containment (Score:2, Insightful)

    by tepples ( 727027 ) <tepples.gmail@com> on Monday October 18, 2004 @01:28PM (#10557495) Homepage Journal

    Formalized interfaces are good, and components are better optimized than launching a whole separate process.

    Not in all cases. It's often easier for a program to contain a misbehaving component if the component runs in a separate process. For instance, if a web browser plug-in segfaults, do you want it to destroy the data you've entered into a form on another page?

  • Re:Damn. (Score:5, Insightful)

    by jd ( 1658 ) <imipak@yahoGINSBERGo.com minus poet> on Monday October 18, 2004 @01:40PM (#10557620) Homepage Journal
    Real programmers should focus on the whole of a problem, not just on the little piece they happen to want to solve. Trying to over-focus on specific areas is one reason Microsoft is notoriously weak on security, Unix is horrible on the desktop, Linux is weak on the hardware support, etc.


    None of these flaws are necessary, all of them are serious impedements and each in turn is likely to be the reason users will migrate away when viable alternatives exist.


    I wasn't overly impressed by Rob Pike's answer to the idea of specialist Unix tools, as opposed to more generalized software that can handle many different types of task, but it illustrates a blind-spot that could prove troublesome for Rob.


    The blind-spot is thus - if extreme specializing and narrow focussing is a dead philosophy in coding, it must also be a dead philosophy in software engineering in general. If the logic doesn't hold true any more, then it should be dispensed with completely. Dropping it only in one or two narrow areas is, in itself, an application of the very philosophy that is being rejected.


    (ie: It is an extreme specialization, rather than the general application, of a change in attitude.)


    Of course, this argument only holds true if my central beliefs are correct, which are that:


    • The distinction between a specification and an implementation of that specification is artificial and has no basis in fact, and
    • The distinction between algorithms implemented as programmatic solutions, physical solutions, social solutions or political solutions is also artificial, in that they are merely different forms of expression where the expression itself has no intrinsic form


    What could Rob say about patents? Well, for a start, he could have said that they are the existing method of solving the complex problem of fairly compensating people for their work, but that the solution is probably not the best and may need to be replaced with something better. He doesn't need to produce a working flow-chart on what the politicians should come up with. Nobody asked him to actually invent a better method.


    By not really answering the question, he sounds like he cares more about what his pay-masters would think than with giving an honest answer. Now, that's not horribly unreasonable, but how much extra effort does it take to say "that's not a question I can really answer"? At the very least, it would be an answer, and therefore respectful of the questioner.


    Personally, I don't subscribe to the notion that the one who pays the piper calls the tune. You could pay me - or anyone else - whatever money you liked, but not a single one of us could change the laws of physics, violate Pythagoras' Theorum, or make 1+1=3. Some things can't be changed for love or money.


    Programming is one such area. A problem is computable or it isn't. If it is computable, you can solve it with a computer program in finite time. If it is not computable, no general algorithmic solution exists. An act of Congress won't change this. If you assume God to be constrained by logic, then even an act of God woudn't change it.


    If it is possible to patent an algorithm, then it is possible to create patents that cover the ONLY workable solution to some set of problems. The owner of such a patent is claiming ownership of not just that specific solution, but the entirity of that class of problems. Since these are the only patents worth having (you can circumvent any others, because there are other algorithms which do the same thing) it follows that the system is inherently unstable and self-destructive.


    Creating or supporting bug-ridden implementations is bringing computing into disrepute, whether those implementations are in C on a computer, or in english at a patent office. Bugs are bugs are bugs, whatever the form, whoever the implementor. And bugs are never a Good Thing.

  • by Anonymous Coward on Monday October 18, 2004 @01:51PM (#10557713)

    I asked in my +5 interesting post why modern OS's are written in the lowest-level practical modern language (C) instead of the highest-level ones? I noted that at the time C was one of the highest-level languages and that was behind UNIX's success.

    Instead he wastes his space completely dodging a decent question on patents, and takes some softball OS question so he can spout off about how his baby Plan 9 is maginally better (you can pass any number of parameters on the command line). Wtf?

    Also his analogy about OO being like a wood craftsman is so opposite of reality. The fact is that modern chairs are not crafted but made out of components. Modern anything is made out of interchangeable parts (aka components, objects). There's nothing about C or alef, perl, scripting, etc that is based on components or plugging existing parts together. You can create a beautiful chair in an OO language, but nobody really cares since in an object-oriented language you sit, push in, stand on, whatever the chair. I mean come on, the guy creates his own editor in Alef (aka C) and he's talking as if that's not the craftsman side? wtf is he smoking? Serious.

    P.S. don't give me any of that BS that OSs in LISP existed; lisp is the functional-language equivalent of assembly language.

  • Pike:
    "The promoters of object-oriented design sometimes sound like master woodworkers waiting for the beauty of the physical block of wood to reveal itself before they begin to work."


    Comparing programming to woodworking is a bit extreme.
  • GIGO (Score:4, Insightful)

    by Doc Ruby ( 173196 ) on Monday October 18, 2004 @02:23PM (#10557952) Homepage Journal
    He answered the question by pointing out that it's nonsensical, as posed. He could have answered any number of other questions that weren't asked. A better question is why the moderator picked that poorly constructed question, rather than any of the answerable ones that weren't asked.
  • by Martin Blank ( 154261 ) on Monday October 18, 2004 @02:28PM (#10558009) Homepage Journal
    Slashdot folks also rate random flames at Microsoft fairly highly, too. While such moderation is sometimes deserved, that doesn't mean that the flames always have a valid point. I agree that Rob kind of ducked the question here, but just because Slashdotters rated something well doesn't automatically make it a good thing.
  • by HyperbolicParabaloid ( 220184 ) on Monday October 18, 2004 @02:31PM (#10558038) Journal
    Maybe he considers Arms control an issue more significant by many orders of magnitude than patents. I reasonable person might think that even discussing nuclear weapons and IP in the same sentence trivializes the significance that the arms race played in our lives for decades.
    Are software patents important? Yes. Do they threaten the very survival of our species? No.
  • by JohnsonJohnson ( 524590 ) on Monday October 18, 2004 @05:38PM (#10559602)

    I mentioned in another post that SQL products are NOT RDBMS

    Ah, a disciple of Date. If we're going to switch vocabularies that's fine. In the wider world DB2, Oracle, SQL Server etc. are synonymous with RDBMS and the terms are used interchangeably. You are of course correct, that technically that's incorrect, but the technical point has largely been ignored by the computer industy, users and developers. I can work with your definition as well.

    PageRank is an algorithm of popularity and not an algorithm of relevancy and as such, it really bears little relevance to implementation of relevancy algorithms as we are discussing. Of course, relevancy algorithms could contain page rank as a heuristic. See http://www.google.com/technology/

    My fault for being imprecise, PageRank is only part of Google's search algorithm and I used it to refer to the whole. The point about scalability still stands (I'll get to why Oracle et al. are a propos even though they aren't RDBMSes in a bit).

    In order for you to create a document (in the New World Order there really is no such thing as a web page any more) about Atlantic Slave Trade, you would have to have some sort of schema that defines it (by definition, it would require one).

    I think this is a case of worse is better. The New World Order may never be imposed because ignorant neophytes have gone and ran with HTML in directions the high priests never intended. Lacking any armies with which to enforce compliance it's an open question whether the high priests will ever be able to control the chaos again. Technically superior solutions (and I actually agree with you here in the beauty of a real RMBS as opposed to what's marketed as such) may have aesthetic superiority, but just as real RDBMSes will never replace Oracle, DB2 etc. until either the replacement cost is 0 or there is some orders of magnitude greater functionality not available in other systems, so those waiting for the NWO may still be waiting on their dying day.

    How does Google know that "my trip" refers to the trip that you took from 01-OCT-1995 to 20-OCT-1995 with three friends? What about the content of the pictures?

    It doesn't, and it doesn't have to. The query, my name Versailles pictures is probably good enough to find those pictures (and 10000 unrelated items, but the one I'm interested in is easily identifiable). And there's the problem with replacing Google: it's good enough for so many tasks that a replacement would have to be orders of magnitude better to displace it. As for when the pictures were taken, or what their contents are, I get that information the same way I get it now, from the file metadata, assuming it's correct. On the other hand, every photo organization program available (iPhoto etc.) allows the user to add all kinds of meta data to search for content. And guess what? Most user's don't, because the last thing you want to do after taking 200 pictures on a trip is spend 2 days typing captions for all of them. People still identify pictures they way they did with old fashioned photo albums: by context, comparing them with pictures of similar scenes or with time information (clothing style, etc.) to fill in details that may have been forgotten. This is easy enough for human intelligence to do, and so far beyond current AI ability, that I don't see any application improving on this model any time soon, although at some point someone may.

    Finally, there is no requirement for a 'formal language' - when you do a Google search do you have to specify a formal language? That is a matter of implementation

    But there are no implementations that don't have some form of formality involved. The problem is I (and the vast majority of people) have many bytes of free form text that Google indexes just fine. Simple Vector Quantization (I use a Mac) works fine for searching for documents locally on my machine, and my guess is that Google's algorithms are some form of vector quantiza

  • by JohnsonJohnson ( 524590 ) on Monday October 18, 2004 @06:59PM (#10560243)

    Your sentence illustrates the myth that, in the IT industry, technically superior products will rise to the top.

    I think I actually implied otherwise, that there are a host of reasons why products and ideas succeed in the marketplace, technical superiority being only one of them. Time to market, capitalization, flexibility of the developer (often products succeed in markets other than the original target, Java for example) etc. As an example consider Ted Nelson's Xanadu, arguably a superior system to HTML, but except for a very few afficionados, having no users to speak of.

    It is good enough only in the micro. There is a statistic which mentions the geometric (maybe even exponential) rate at which we are creating and storing new data. Sure, for your current family album this level of granularity may suffice - but I suspect in the future our family albums will be composed of video, audio, stills, etc. at a magnitude that makes getting 10,000 results impossible to sort through by hand. You'll require more accurate search results and will want to ask more precise questions. The RDBMS is the way to get this

    Or better heuristics, getting 10000 results isn't a problem as long as my target is in the first 10-30 results.

    the solution for tagging metadata is obviously not a 'solved problem' yet - this is mostly because no one has seriously tried to study it

    I beg to disagree, I think most people, on looking at the problem realize how hard it is to solve, and given the existence of competitors like Sherlock, Google, whatever Microsoft implements in Longhorn, the risk of getting no pay off at all, decide that there are other ways to attack the problem, or more profitable avenues of research. I'd refer to Pike's comments vis a vis kernels: there's plenty of research on kernels, but there's little reason for a user to switch kernels since there's not enough of a benefit to the applications they use compared to the pain of switching.

    The real question is not if or when we will have this but whether or not companies like Google will implement it correctly the first time or after the technically inferior solutions collapse under their own weight.

    Not being on the inside of Google myself I don't know if they think they're running into scaling limits, but outside evidence is that there is no problem. As for correct implementations, it seems to me that applying the query as metadata to search results would solve the problem of metadata generation without requiring a radical breakthrough in AI. It may be less intellectually or technically satisfying, but like a lot of half way solutions, like the Web itself, it can serve as a platform for further development without requiring an earthshattering breakthrough.

I've noticed several design suggestions in your code.

Working...