1439 posts • joined Friday 15th June 2007 09:17 GMT
I don't know how many of the readers remember much about the Digital Equipment Corporation (DEC), but they were involved very early on in the definition of many of the fundamentals that cloud computing is based on.
They were one of the people involved in creating Ethernet and the Internet (although they eschewed TCP/IP for their own DECNet as the preferred network for many years)
They were early adopters of the concept of mobile workloads spread across several machines (DEC-Cluster and VAX-Cluster).
They had network shared storage before almost anybody else (HSC devices) and things like LAVC (Local Area VAX Cluster).
They were one of the early pioneers of clustered desktop machines (DEC ALL-IN-1 and Pathworks) including network booted diskless PC's
And, least I forget it, UNIX was INVENTED on DEC machines (PDP11 systems)
I'm puzzled by the statement Ken Olsen made about UNIX, because DEC had commercialized UNIX in it's software portfolio for years. UNIX V7/11M was a port of Version 7 available through DEC in the early '80s on PDP11s, they did a System V port onto the VAX for AT&T (and I believe it was available to other companies as well), Ultrix was available from DEC in the early 80's on VAXen, you had DEC-Station MIPS based UNIX workstations in the '90s, OSF/1 was available as a supported OS, and later morphed into Tru64 UNIX on Alpha based systems later in the same decade. I can't think of a company that had as long a history of UNIX at the time DEC was subsumed into Compaq.
Maybe Ken thought VMS was the only OS needed, but fortunately other people in DEC did not agree. And others thought they had been daft to drop TOPS-20!
Just because there are text books does not mean that the way of working is correct or to everybody's liking. Take sociology for example......
I'm sure that there are many things that are completely insane that I can make rational sounding arguments to support. Try reading Douglas Adams' books for rational absurd reasoning (although, yes, I know he was a Mac User, but I'll forgive him that because of his genius)
The Mac way of working is fine if you use single or small numbers of applications. Not for many applications on a screen, like I use all the time.
And the argument about power users using key combinations is crazy. In my world, where I use Windows, CDE on UNIX, KDE, GNOME, and (god forbid) Unity, you just cannot learn every one of the myriad of key sequences. And in case you ask, I am an Emacs user, so am used to quite complex sets of key strokes.
Barclay's used to be very diligent about their failover tests. I was involved in several tests over the years I worked there. But that was when they actually had people in the UK, and did not rely on it being run from Pune or Singapore.
It used to be a big issue if one of the tests failed, and they generally had a second test pencilled in when the initial one was being planned just for this situation.
Still, it's been over 5 years since I last worked there, so who knows what has happened in that time.
Re: At last
What I wan to know is what happened between June last year and early April this year that caused his caps lock key to stick down. There's a gap in his posting history (which actually includes two moderator deleted posts).
He's been registered since 2010, but up until April this year had only entered a handful of comments. Since then, something has woken him up, and caused him to SHOUT about everything he's commented on.
I actually used one for a University assignment
This was in 1979 or 1980 (I can't remember exactly - getting old).
Was my first experience of 6502 machine code, which became very useful when I got my BEEB a few years later. Had to code a sine-wave generator using an attached D-A converter. Real pain putting the opcodes directly into the keyboard, with no means of storing the program.
I think that this one must be a later one, because the ones that Durham had not only used a calculator keyboard, but also a tiny calculator display as well, mounted in what I remember to be the top half of a Commodore 8 digit calculator. My memory may be playing up though...
Re: MAC Address -AC@11:46 8th May
"..you probably aren't gaining a lot over what a typical "current" ADSL router will provide.."
You think not?
Double NAT, not relying on ISPs router firmware to leak information, capture of packet headers and GBs of log files, multiple DMZs, intrusion detection log, control of inbound connections using SSH to give access to printing and filestorage in my home (you can really do a huge amount through SSH tunnels, including CIFS and lpd), configurable DDNS (I've tried the DDNS support in routers, and given up), not needing a syslog server to capture the logs that are too large to be held in the device, traffic logging from individual systems within the home environment (useful for determining who is the traffic hog), a proper user interface (shell) to diagnose network problems, tcpdump available, serial line attached to my RS/6000 to allow me to remotely power on and off from the Internet. You want me to go on, because I don't think this list is complete.
I don't use SmoothGuardian, SmoothWarrior or any of the other paid plugins, because I do not run my own SMTP server or multisite VPNs. I find Smoothwall Express quite capable enough for my needs, and have been using Smoothwall to protect my network for over 10 years, log before ADSL routers were as sophisticated that they are now.
Expense. A 10 year old 700Mhz Pentium 3 laptop, extra USB Ethernet adapter left lying around from god knows when, and a couple of Ethernet cables. Total outlay, nothing, zero, zilch. Burns about 20 watts of power with the screen off, so is not very expensive in energy either.
And, of course, my time.
Why do you assume I download Torrents? There is enough content in iPlayer, Sky Anytime, 4OD, Demand 5, YouTube, as well as the rest of the Internet, and my kids use Steam, Wii and Xbox games a lot. There's plenty of legal content lying around on the Internet. I'm not blocking the RIAA or MPIAA scanning my systems specifically, I'm trying to keep my home network safe from anybody who might want to do damage to it. I do not want ANYBODY snooping my network out of principle.
If I download torrents, it's only fan-subbed anime for series that are not available in the UK. There's a lot of non-H series that have never been available in the US or the UK, so is very difficult to get to see without some form of copyright infringement. If it were available, I would probably buy it rather than download it.
My library of purchased downloads, DVDs, CDs and videos is quite extensive, and I do buy almost all of the content that I have, although some of it is second hand. I take exception to the implication that I am any more a copyright infringer (even with my admission about anime) than anybody else, and ask whether you live in a greenhouse? At least I post under my own name!
@me - incomplete sentence.
"But interestingly, it is possible for the MAC addresses of machines connected to a single router device performing both border routing to the ADSL or cable network, and also DHCP and/or Wireless routing"
is not complete. It should also have "to leak" appended.
Re: MAC Address
But interestingly, it is possible for the MAC addresses of machines connected to a single router device performing both border routing to the ADSL or cable network, and also DHCP and/or Wireless routing.
What runs on the router is only as good as the firmware, and as we have seen with BT and their powerline Ethernet devices for BT Vision, it would appear that some ISP's modify the firmware to allow some remote discovery. And I'm not sure I fully trust uPnP not to leak service information externally. So we could see internal MAC addresses (that the router has to know in order to function), internal IP addresses (from DHCP), and possibly system types and function available to whatever runs in the router's firmware.
Maybe I'm paranoid, but I have a ADSL router which was not supplied by my ISP (and runs NAT), with a Linux based firewall (Smoothwall, which also runs NAT), and then a wireless hub inside the firewall. DHCP is run by the firewall, not by any of the appliances. This way, I believe that it is almost impossible for anything from the broadband side to get information from inside my network. Now that I'm not relying on wireless as much (I'm using a mixture of direct Cat 5 and, I'm afraid, powerline Ethernet for most network access now - and yes, I generate my own keys), I'm toying with the idea of putting the wireless on a separate DMZ just to give most of my network protection from wireless crackers. Just need to get another Ethernet port in the firewall.
My wife thinks I'm mad, having so much kit 'just to provide the internet', but then I believe (and I check!) that we've been completely clear of intrusion type attacks since I set this up.
I get good DVD playback on my EeePC 701 using a USB DVD drive running Ubuntu 10.10, and that is really an underpowered PC, being a Celeron clocked at less than 700MHz.
Methinks you need to look at the graphics options. Sounds like you've either not installed the Nvidia restricted drivers (which would be strange, as if that adapter was in the system when Ubuntu was installed, it should pick it up automatically), or something has disabled hardware rendering, and the system is using software rendering. Try installing and using the Nvidia driver settings tool from the Ubuntu repository (no, that's no more difficult than installing drivers from CD that came delivered with your graphics card).
Re: IBM ROMP vs. ARM
@starsilk. Thanks for the correction. I certainly knew about the multiply-add being missing, but I deliberately avoided talking about the multiply instruction being missing, because I just could not remember.
Re: Whatever is left of Digital Research
And I think that the Digital Equipment Corporation might have something to say about Intergalactic Digital Research's products as well! (CP/M was essentially a functional copy of RT/11).
IBM ROMP vs. ARM
The IBM ROMP chip (aka the 801) was never intended to be a general purpose RISC processor. It was intended to power an office automation product (think of a hardware word-processor like WANG used to sell).
As a result, although it could function as a General Purpose CPU, it was not really that suited for it. It was never a success because at the time, IBM could not see justification for entering the pre-Open Systems UNIX world. RT 6150 and the 6151 were intended as niche systems mainly for education, although they did surface as channel attached display front ends for CADAM and CATIA run on mainframes (and could actually run at least CATIA themselves). This changed completely with the RIOS RISC System/6000 architecture, where IBM was determined to have a creditable product, and invested heavily.
In comparison, the ARM was designed from the ground up as a general purpose CPU. Roger Wilson (as he was then) greatly admired the simplicity and orthogonality of the 6502 instruction set (it is rather elegant IMHO), and designed the instruction set for the ARM in a similar manner. Because the instruction set was orthogonal (like the 6502, the PDP11, and the NS320XX family), it makes the instruction decoding almost trivial. It also made modelling the ARM on an econet of BBC micro's (in BBC Basic, no less) much easier, which allowed them to debug the instruction set before committing anything to silicon.
They had to make some concessions on what they wanted. There was no multiply-add instruction, which appeared to be a hot item in RISC design at the time, and to keep it simple and within the transistor budget, all they could do was a shift-add, (the barrel shifter), which although useful, was a barrier to ultimate performance, but great for multi-byte graphics operations.
It was also simple enough so that they could design the interface and the support chips (MEMC, VIDC and IOC) themselves, achieving early machines with low chip counts.
This is all from memory of articles in Acorn User, PC World, Byte and other publications. Feel free to correct me if my recollections are wrong.
Re: Am I missing something?
What they have said is that you can't copyright something that says (using the example of another recent story) "produce a process that takes sea water as an input, and produces fresh water and brine as outputs" (which is a functional specification).
You can patent the method for doing this (reverse osmosis, for example) but that does not prevent someone from using evaporation or distillation to have the same effect.
I know that this would be a patent rather than copyright in this example, but the concept is the same.
Thus the code you write for your product is protected, but the description of what it does isn't. This has been fundamental in the concept of black-box testing and modular design for many decades, and changing this would break almost all modern industrial processes.
Just imagine not being able to replace Oracle with DB2, because the function of J/ODBC was subject to copyright, or even worse, not be able to port from UNIX to Linux because the interface to the C library was subject to copyright.
Nice reference to the original Dungeon!
It's supposed to take you to Y2 though IIRC, not generate a CRC.
I always wonder how much of the water that actually leaks from the pipes actually makes it's way back into the ground water reserves (especially in London), and thus is available again.
Anybody any ideas?
This is a very defeatist attitude. It assumes that all teachers and all students decide in the same year to do next to nothing.
If this does not happen, then all those teachers and students do is to make sure that they will fall behind the ones that do try. As what I was envisioning was competition, this is unlikely to happen.
Human beings are competitive, especially kids. Watch them play. They race, they throw, they compete in games of skill (marbles, conkers, hopscotch, computer games). It's coded into our make-up. You just need to engage their competitive nature in school to ensure that the best can be achieve. You also need to make sure that lesser grades than 'A' still have merit.
On a side note. I heard a news item about a boat builder who was complaining at the number of kids who are now sucked into the academic stream, who would have previously gone into some form of apprenticeship. He said that we needed bright kids to be the skilled artisans of the future, and all he was seeing after the competent ones had gone to university were the kids who were unable to master his skill. Was a very fair point well made.
Re: Why do we have a set pass mark for grades?
Marking to the curve is a double edged sword, and I accept that it makes comparing marks year-on-year more difficult, but you have to ask what the point of the exams actually are?
When I was doing my 'A' levels in the late '70s, the primary reason was so that you could be selected for further education. As there were many fewer university places available, the marking was set so that you could tell who was 'the best' from that year's student population. If less that 10% of the students got an A, these people, who would be the most likely to excel in that subject, got streamed to the best Universities. The next tier down could select from the remainder, and on downward through the Polytechnic system, aiming at people who would excel at HND qualifications, but may not be up to a full degree.
It did not matter whether there was grade comparison between years, it would be accepted that the best people would always get better marks than the weaker candidates, so the streaming would still work, and the 'right' people would always get to the establishment that best suited them.
Quite often, it was not the grades that determined what type of work someone ended up in, it was how far they went in the education system. Students who had got to University and completed a degree course had demonstrated by that fact that they were worth employing.
It is only now that the 'A' levels that are intended to give an absolute measure of how someone's worth that this problem occurs. Since schools have been measured by result, and the curve has been discarded, it has completely devalued them as a mechanism for selecting the best students. Governments and schools each have an interest in 'improving' the results.
Part of the problem is also political. Educationalists in the '70s and '80s became convinced that non-competitive grading was the only way to avoid stigmatization of kids (abolition of the 11+ and Grammar schools is an example). Schools were not allowed to say to kids "look, you are never going to succeed in becoming a theoretical Physicist, best do some vocational training". All children are given unrealistic expectations by being told that they can achieve anything, and in order to persist this myth, the exams are set so that they think they are good at a subject, when in fact they could be only mediocre.
This is just dumb. Life is competitive, and that is never going to change. When you go for a job, the best candidate wins (unless the recruitment process is also dumbed down, but that is another rant!) And people not suited or without an aptitude for a particular job will never get it, regardless of how much they want it.
Setting kids up with realistic expectations, and giving them some taste of reaching their ceiling by allowing some of them to experience disappointment is a required life skill that they have to learn at some point, and my view is that it should be part of the school experience, instead of a post University kick in the teeth.
Re: Doubling CPU cores is also doubling transistors
One of the problems that chip designers have is how to use the vast number of transistors that can be fitted onto the large die-sizes at the smallest scale.
They got to the point where more registers, more cache and more instructions units in a single core was not making for faster processors, so they then started using the still increasing transistor budget to put multiple cores on a single die.
There is a lot to be said for a large number of cores on a single die, but this has it's own problems with access to memory, cache coherency between cores and I/O.
Another avenue is putting disparate processors (like GPUs) on the same die, or even System on a Chip (SoC), where all of the functional elements (I/O, Graphics, memory etc) of a complete system appear on a single piece of silicon (think what is going into 'phones and tablets).
In my view, to make use of the vast scale of integration, it's about time we had a fundamental rethink about how processors work. I don't have any new ideas, but I think that listening to some of the people with outlandish ideas might be worthwhile in coming up with a completely new direction to investigate.
@Kebabbert Re: Hmm...
I was not clear about entitlement in my earlier post. There were Linux only Power 5 systems back in 2005 or so. What I was trying to say was that they were the same systems with the AIX and the IBM i entitlements turned off. They were also significantly cheaper, and also made it easier to use non-IBM branded disks.
My views about proprietary UNIX being on the downward curve has not changed. I have felt this way for most of the last decade. I still see Power having a place for many years to come.
Intel becoming predominant is much more about them having volume and critical mass in the processor market than speed or technology. PowerPC is still a relatively well architected processor, but for many companies developing products, it makes sense for them to use what is fast becoming a commodity product (Intel) rather than something that they have to put significant design effort into. A high-end PowerPC SoC would be interesting, but I don't think IBM would be interested in creating one of these for the server market.
Re: @Z Eden - Games should stay on Windows
In theory, DRM is not against the Linux way of doing things. If you are careful to make sure that you only use LGPL (not GPL) code in your DRM system, then you do not 'pollute' Linux by adding a DRM API above the OS, and you don't have to publish the details of your DRM. The rest of Linux works just swell.
The main reason why this has not been done to date is that the content providers do not trust that the OS cannot be hacked below the DRM API to gain access to their content, whether it is a game, music or a film.
Re: @DryBones yes, yes, that's all very well
Huh. DryBones deleted their post! Oh well, just as relevant to the OP of the "yes, yes, that's all very well" thread.
@DryBones Re: yes, yes, that's all very well
The problem with mp3, the DVD formats and many, many other restricted formats is that they are, well, restricted.
The very nature of Free Software, whether you are talking about free-as-the-air or free beer, is that it is either free of restrictions or free of charge. This means that the distro suppliers won't (in the case of as-the-air) or can't (as in beer because they can't afford it) put the support for restricted formats by default.
Blame the people who foist the restricted formats onto us all for this problem, not the distro suppliers.
Of course, earlier releases of Ubuntu often would tell you exactly what you had to do so that you could make the personal decision to break the licensing conditions or patents on the codecs that the distro supplier cannot make without opening themselves up to being dragged through the courts.
It was not that long ago that Canonical were being slated by the '-as-the-air' community for paying for licences for H.264 just so they could include it for people like you.
I know that this does not help you, but that is the nature of the world we live in. Would you pay for a version of Ubuntu or any other distro (so that the supplier could pay the license fees) that included all the codecs you need?
I'll give it a go
I will put this on a partition of my laptop, and boot into it occasionally to see how it is doing. If I can cope with Unity (although from current experience, I won't), I will switch over.
But the problems I had with Lucid stopped me from switching permanently from Hardy until they pulled support from the desktop release of Hardy. Even now, there are significant things that don't work on Lucid, despite defects being open in the Ubuntu fault tracking system.
I have Unity as the presentation layer on a netbook running 10.10, and also on a desktop running 11.10. Later releases may work better than earlier ones, but that does not alter the fact that I believe that it is less suited than Gnome 2 for people who work with multiple overlapping windows on several desktops. I can see it working well for the Mac OSX generation (single application occupying the whole screen most of the time), but that's not me. That way of working is just alien to the way I have worked since twm on X.10 or SunView. I want drop down menus attached to the window I am working on, not up at the top of the screen.
As for HUD, I've not played with it. It may be helpful, but it sounds to me like it will tie applications into the Window Manager in ways that will be detrimental to application portability, which can never be a good thing.
Re: I got one for automation
And the sad fact is that the people who are made to leave are often those that understood the automation, so as soon as something changes, the automation breaks and nobody knows how to fix it, so it becomes a manual process again.
About 7 years ago, I was part of a project automating the build of servers (IBM Power 5 servers running AIX) in a server farm. Could deploy an OS image on a virtualised machine with all patches, management and security software (and some frequently used application as well if required) installed and registered in about 40 minutes from bare metal to hand it over to the application installation team. Did all the work from base packages, no Golden Image in sight. Brilliant (and also stunned IBM when they came to see what we were doing!)
Came back to the company a year and a bit later, to find that the people running the process were all low skill process monkeys who had reverted to manual processes when new machine types came along, and they did not know how to tweak the process (even though it was fully documented!).
Broke my heart!
You needed a BitStik!
Put thrust on the twist of the joystick and had three buttons. Made complex manoeuvres less like shaking hands with an octopus!
And when you ran Elite on a 6502 second processor (if you bought a BitStik, then you probably had one of these as well to run the CAD software), you got Mode 1 graphics and none of the 'mode change' interrupt tear when it switched from Mode 4 and Mode 5 3/4 of the way down the screen.
A PDP 11/84 was a single-chip PDP 11 processor (J11?) in a minicomputer rack (it had a UNIBUS rather than a QBUS which made it a proper PDP 11 rather than a micro PDP 11 like the 11/83).
It was definitely *NOT* a mainframe, but a 16 bit minicomputer with address extension. IIRC, it probably was the most powerful of the whole PDP/11 family (I mean real PDP 11 rather than a VAX 11).
This 'new' ability to only run Linux is not new. If you have access to a Power 6 or Power 7 system, and look in ASMI or on the HMC (and I presume SDMC and IVM) at the entitlements section, there has been an entitlement for both AIX and IBM I for several years. Linux has been an officially endorsed OS by IBM on PowerPC for at least 7 years (they have had agreements with Slackware and SuSE), and there are official distributions of RedHat and Ubuntu from those companies.
This makes this a re-announcement of an existing policy, probably to remind some existing PowerPC shops that they can stick with Power rather than moving to another processor, even if they are switching OS. I very much doubt that the product announced will significantly differ from other systems that will still run AIX and IBM I.
This does not give any new reinforcement that policy that you bring up in every discussion about PowerPC or AIX. Both AIX and PowerPC will be here for some time still. That link lookes older each time I look at it.
Now. I'm not going to argue with the fact that AIX (along with all proprietary UNIX systems) is on the downward side of the popularity curve, and I do not think that PowerPC development is in a good place at the moment. It's expensive to build new generations of any processor, and I think that IBM is really thinking hard about what to do with the PowerPC line, at least in high end servers. Sometimes I wonder whether IBM really wants to remain in the hardware business at all (products that have been sold include their printer division, their storage division, the desktop and laptop PC business, and most recently their ATM and PoS business).
This policy may extend to their server systems as well. Power7+ is late according to previous product roadmaps, and there is strangely very little pre-announcement information about Power8. IBM has also made statements that their previously loss-leading HPC work has to become more commercial (probably one of the reasons why IBM pulled out of Blue Waters), which means that future generations of IH HPC systems are at risk.
But one of the effects of there being a creditably competitor to Intel processors is that makes Intel aggressively pursue new processors. Once they are only competing with themselves (remember, AMD need chipfabs like IBM to create their products, because they cannot fabricate processors themselves), then the rate of product development will slow significantly, as Intel would want to get more return on their investment.
I am really not looking forward to a point where the only processor game in town is X86 derived, and that is looking like a possibility within a decade unless ARM moves upward.
What! No mention of the Plus One
This plugged into the expansion slot, and provided the Microdrive interfaces, along with a joystick port, a serial port and some strange network which allowed you to link several similar systems together in a peer network, sharing the microdrives.
My father bought an early 48K system (I had a bought my own BBC model B), and it did indeed have light grey keys like the picture. In addition, it had the 32K add-on board, and also had a heat sink that ran the entire width of the system under the keyboard, leading to a warm programming experience.
I never really liked the Spectrum, it was too slow, had poor sound, the screen attributes just felt clunky, and that keyboard!
My Beeb, although supposedly lacking in memory, was just a class machine, and ended up being used for things you just could not consider using a spectrum for. OK, it was not suited to large dungeon type games, but I would contend that Snapper, Planetoid, Meteors and Arcadians were great copies of arcade games that the Speccy could not hope to match, and Freefall, Starship Command and especially Elite showed what you could actually do even with a supposed lack of memory.
But the Spectrum was an influential machine, no doubt.
Re: where the buck stops - AC
it depends. If you have decided to include a software product that needs escalated privilege (root or admin), then that is not Microsoft, and you must take some responsibility yourself, and should also blame the vendor of that package.
If it is software that does not require escalated privilege, and can get it using the package, then that would of course implicate Microsoft as well.
But in your example, it would be better to ask if Microsoft should take any responsibility for something they include from a third party as part of a windows installation (such as the CD and DVD burning code licensed from Roxio), as this is more like what Linux Distributions do.
Am I alone in remembering Red Box...
a company formed by Hermann Hauser in 1985. I know it's since disappeared, but it marketed a through-the-mains control module solution.
I can find almost no references to this in Google, Wikipedia etc. It's amazing how something used to be able to disappear almost without trace before the Internet.
Dell and terminals
"never made a machine that needed a dumb terminal" - this is untrue.
Dell had a brief foray into the UNIX on Intel world in the late '80s and early '90s with systems running SVR3 and SVR4. These systems were shipped with multi-port serial cards, so would have used terminals of the type produced by Wyse.
I can't remember what they used to call them, but I attended an interview for their UK support team. I also can't remember what the outcome of the interview was, but bearing in mind that the team was wrapped up not that long afterwards, it was probably better that I did not work for them.
Re: Surely this is illegal
Illegal? No, certainly not under the Data Protection Act. The employers are asking their prospective employees to volunteer their facebook account details. If they agree, then it is a private agreement between the individual and a company. This is exactly the same as a loan company asking for copies of your bank statement before offering a loan.
It may be counted as discrimination if it can be proved that the individual did not get the job because they refused to hand over details, but that would be a completely different issue.
I immediately thought that the employers were going to turn an applicant down if they actually DID give their login details over, because that would indicate a lack of understanding about on-line security! Ho hum.
@Chemist Re: So what /should/ have been done?
Reminds me of the old IBM PC error "Keyboard error - Press F1 to continue"
The old ones are the best!
11/780 was the base.
When the IBM PC was launched, remember it was a 16 bit processor in an 8 bit system (8088 had an 8 bit multiplexed data bus needing two cycles to store a 16 bit word), and was only clocked at 4.77 MHz. In the Personal Computer World BASIC benchmarks, the BBC micro could whip the ass off the IBM PC in performance terms, although this should not be a suggestion that Linpack results would be the same.
I always regarded an original 6MHz PC/AT as about the same processing power as a PDP 11/34, although that was only on a subjective feeling, and a VAX 11/780 was much more powerful than my 11/34.
A real supercomputer is a lot more than just processing power.
The current systems I am working with (still on the top 500 - just) are split (very approximately) equally cost-wise between processing, networking and storage.
The interlink is important for massively parallel jobs, and there is no point in crunching numbers if you can't store the results. Linpack can be a very misleading benchmark.
Re: Re: Server-Side Copy
Yes, if you are able to make one NFS server talk directly do another, without using a client computer. But I think that if you use SCP with two remote locations, the data still travels through your local machine, in-and-out.
I miss being able to go into an arcade and spend an hour or so bashing steel balls around a table. It's a great way of letting off steam. I used to be able to get a replay on my regular tables almost every game, which made it quite cheap! Oh, the feel of the flippers and the buffers and kickers...
Every now and then I will find a table in good enough nick to be worth playing, and I still have some of my skill. I went into a pub a few years ago with some friends, saw a table that I had never played before, and still managed to get two replays from the 50p I put in. Pissed off my mates who thought I must have been in before.
I live in a tourist oriented seaside town and I cannot find a single table worth playing here. It's sad really.
200 miles per day
I used to commute 96 miles each way daily for over three years. It's not that unusual. I currently have a 'short' commute of about 45 miles each way. And when was the last time that you left on a journey without any means of refuelling with less that a 25% margin or error? Who knows when you will get stuck in traffic/diverted? I certainly would not want to do more than a 150 mile journey in a car with a maximum range of 200.
Also, 200 is when it is new. How will it work after a years worth of daily recharging?
Re: make greater use of probabilistic information in their weather forecasts
The way the weather is presented in the media is controlled by the media, not the Met Office.
It's true that some, like the BBC, actually use Met Office forecasters, but the presentation style is normally controlled by the broadcaster, not the forecasters.
If you want to see isobars, look at the Met Office web site (try the link http://www.metoffice.gov.uk/weather/uk/surface_pressure.html directly), not the forecast on the telly.
Re: Helps if you actually forecast
Try the localised detailed forecast from the Met Office's web site, rather than relying on some summary forecast where the poor forecaster has to cram some approximation of the weather for a region the size of several counties into a few 10's of seconds.
You might as well try to describe the colour of London as seen from Google Earth in 10 words or less.
Version numbering and catchup
I agree with your comments on version numbering. There is not that difference in the outward appearance of AIX 5.3 and AIX 7.1. Under the covers, there have been quite a lot of changes, including dynamic partition migration, more control of WPAR isolation, support of more logical CPUs, USB storage support and up to 4 threads per core.
A lot of the changes are targeted to the very largest systems, but this is not really a problem because what features that you need are not in AIX already? I'm not including things like Gnome and KDE, because they are not really part of the OS. DBUS maybe? Remember that AIX has never really been a desktop OS, and much of that Linux innovation has been in things that are really applicable to personal systems rather than servers.
I don't rate your comparison of AIX and Ubuntu LTS. In the same time I have been working with AIX 5.3, I've gone from Ubuntu Dapper Drake (6.06), Hardy Hedgehog (8.04), and Lucid Lynx (10.04).
IBM have always clear about the lifetime of it's OS products. End of marketing is always announced at least a year before it is actually withdrawn (and normally soon after the +2 version is announced), and there is normally at least a year of support from End-of-Marketing to End-of-Support, and then there is always extended support for customers prepared to pay. And after that, the mature AIX product (which after so long in support is likely to have had all of the serious problems fixed) will have the fixes available on fixcentral for a couple of years more.
Re: Related to ongoing contracts that are late and run on AIX 5.3 perhaps
PTF stands for Program Temporary Fix.
AIX 5.3 was actually released in 2004, and the product lifetime cycle published on Fix Central has indicated that TL12 would go out-of-general support later this year as early as this time last year.
This means that AIX 5.3 will have had a lifetime of over seven years, and if you take into account extended support, will be more like 9 years.
TPM's article has several errors in it to do with dates and functionality of earlier versions of AIX (like LPARs being introduced in AIX 5.1). I've sent a correction, so we will have to see whether the article is fixed.
SciFi got there first
Larry Niven based the whole of one of his literary universes around a society who 'harvest' body parts from convicted felons. His conclusion was that eventually people would end up being broken up for offences as serious as jay-walking and tax evasion.
Look up "Flatlander" or "Gil the Arm" if you are interested in reading the stories.
Only those destined for the Chinese internal market. Anything sold in other countries would be subject to cases in each country in question.
Of course, in China anything could happen, given that their legal system is very different. But I think that many companies would re-consider using China as a manufacturing base if the Chinese authorities were to start confiscating goods destined for sale outside of China.