2348 posts • joined 10 Jun 2009
Re: Faster access to components?!
Funny, on the PCs I maintain today I've never needed anyting other than a #1 cross-head.
I do remember that early Dells were assembled with a nightmare mix of Philips, flat, Torx, hex-head, and OMG tamper-proof. But that was a LONG time ago.
Faster access to components?!
Why is it that Dell (and other "corporate" PC vendors) hold to the idea that it's quicker to get at a machine's innards if you don't need a screwdriver? My experience has always been that by the time you've worked out which tricky little catches to release and what to slide in which direction, you could have replaced a disk drive in a drive bay in a well-designed machine where it's held in place with screws.
And shortly after you've worked it out, you'll notice blood dripping from a finger, and have to take a break to find a sticking-plaster.
(Are there really any service engineers out there so dim that they can't use a screwdriver? )
Re: Why save the PC?
I also think MS has too much negative sentiment associated with its brand to really excite large portions of the market.
And when did IBM last excite large sections of the market? What percentage of people outside IT even know what IBM does? And ... so what?
Microsoft's bread and butter is desktop PCs used by businesses or for business. For content creation not content consumption. If they carry on down the road they are on, someone is going to eat their core business the way Linux(*) is eating the iPhone.
By the way, I don't care. I almost hope Microsoft does go the same way as Digital. Those whom the gods would destroy, they first make mad. Telling your customers that they're wrong and should <go away> is a symptom of corporate madness. It's also what Digital did more and more, until their fall.
(*) Android. It's Linux under the skin. Though the moral would be the same even if it weren't.
Re: Not really fixing any of the problems
One danger MS has now is they've split their customers into two groups, those who prefer a traditional desktop environment and those who prefer Metro.
Linux solved that problem ages ago. When you log in you select the UI you prefer from a list of those installed on your system. If your UI of preference isn't on the list and you have admin access to your machine, you download and install the one you want. I bitched like hell when Gnome 3 destroyed Gnome 2 rather than being able to install them both side by side ... that's NOT THE LINUX WAY ... but the open-source community fixed this brain damage both possible ways, with Mate (a fork of the original Gnome 2) and Cinnamon (a new but similar UI running atop Gnome 3 libraries).
But Microsoft think there must be one and only one UI, that they tear up and replace with something utterly different on a whim. And the Windows 8 UI is frankly about as loveble as Jimmy Saville. The more one learns ....
At the moment I have five windows visible on my screen. That's how I want to work. I don't want anything to ever go full-screen unless I click on the full-screen button - which I never do, because no application I run needs 1920 x 1200 pixels. Several more windows are hidden but I can bring them to the foreground with one click and return them to underneath with another one click. I don't have to touch the keyboard to context-switch.
Microsoft think I should be working differently.
If Windows 7 ever goes away, I'll find a job that doesn't involve ever having to interact with a Microsoft system at all.
Re: A solution looking for a problem.
My thought too. For now, USB3 is good enough for almost everything, and eSATA good enough for the exception.
But maybe Intel are looking a long way ahead. When memristor technology arrives (dirt-cheap randomly addressible nonvolatile RAM - OK I'm an optimist), how are you going to connect your 1Tb memory stick in the year 2020? Hint: USB3 will offer only a small fraction of the attainable speed.
Also think about why USB2 beat Firewire (which was faster).
Encryption is the real danger
If anyone wanted to read a non-encrypted document in some old format or other, I doubt that they'd need to be qualified to work at GCHQ in order to "break" the (non-)code. It's just reverse-engineering something that actually isn't designed to conceal.
Strong encryption, on the other hand, means that once the decryption key is gone, so is the document.
Colour temperature is a stupid attempt to sum up a spectrum in one number. What you really need is a spectroscope. You'll then see that an incandescent bulb of any sort emits a nice smooth spectrum (which can indeed be derived from the filament's temperature). What comes out of a CFL is mostly green and all lumps and bumps, and what comes out of an LED is a rather smoother two-humped distribution with an unnaturally high amount of blue.
Our eyes are evolved to work best with a single-hump spectrum a.k.a. sunlight. A halogen bulb comes closest, but is not as hot as the sun so it's a bit deficient in blue and violet. An old-fashioned "yellow" filament bulb generates less green and virtually no blue and violet, but the spectrum is still smooth.
There's also mounting evidence that blue and violet light is involved in maintaining our bodies' circadian rhythms. Yellowish lighting for night-time use was probably a very sound non-choice, and the medical impact of switching to CFLs and LEDs may yet to be appreciated. If anyone is suffering from insomnia, I'd suggest it's worth a try to kick out all the CFL and LED bulbs from your house, and run halogen bulbs on a dimmer in the evening to suppress their blue output. Conversely on grey winter mornings, a full-on halogen bulb or even the bluer LED illumination may be a very good thing.
Re: Why not intelligent sockets?
The difference in wattage between CFL or LED and incandescent is so great that there's no way they aren't saving electricity. Less than is advertized, though, for two reasons, one of which you'd not spot with any measurement of electricity going into the bulb.
1. CFL light spectrum is awful, and we compensate by upping the brightness. CFL warm-up time likewise. LED is far better and zero warm-up time.
2. "Waste heat" isn't all wasted. In winter, part heats the room, and the rest the space above. If you eliminate this heat source, the central heating has to supply the noticeably missing part(s). And in UK homes, lights are turned on for a much greater time in winter than in summer (short vs. long daylight hours).
Energy-saving bulbs make far more sense in the tropics, where they also save on airconditioning bills (i.e. extra electricity being used to pump the completely unwanted waste heat out of one's living quarters).
Anyway, give LEDs a few more years' development, and they'll probably become at least as good as halogen incandescent lighting.
It ought to be possible to produce a cheap version of powerline networking, with a bandwidth of a few kilobits per second, for the purpose of controlling domestic equipment. Get it ISO standardised, get the price down to a few pennies per chip, and build them into every appliance right down to the light bulb level. (Or into the ceiling roses or into bayonet to ES adapters! )
Just give us a physical disable mechanism, so technophobes or safety-critical devices can't be hacked!
Re: "There's no need to change your light fittings - you can buy screw to bayonet adapters"
For incandescent bulbs, I preferred screw. The problem is that the heat makes the plastic of the lampholder become brittle, and in my experience the bayonet holders are far more prone to breaking when you try to remove a blown bulb. Should be a thing of the past with flourescents and LEDs. Screw-fit bulbs working loose was probably caused by thermal cycling, so that is also probably a historical problem.
I also wonder how long it will be before designers realize that if an LED bulb will last 25 years, why make it a replaceable unit? Isn't it better to integrate the low-voltage PSU and the LEDs permanently into a metal light fitting? The light fitting would also provide a better heatsink for the LEDs, extending their life and/or increasing their brightness.
I've now got to ask - how do Germans indicate a diaresis, given that the usual method of doing so looks exactly like an umlaut?
Re: This thread is full of words that look like variable names I've seen.
Awful code? Or just code written by a person whose first language is not English, whose idea of a good variable name is therefore not the same as yours?
Re: There is even an IT angle
Maybe with more CPU cycles, they could start analyzing words with prefixes and suffixes, and put a wiggly brown line under words that might be OK but aren't actually in its dictionary. (Ie, words that break down as common prefixes and suffixes and something that is a known word in the middle.)
German (and Turkish and others) basically don't put spaces in a noun phrase, so there's no real significance to the length of a "word" and you're free to invent your own. Worth noting that spaces are actually a relatively modern development in writing. Theancientsdidn'tusespacesatall.
My thought too. Unlike most lengthy detritus in the dictionary, this word is actually useful (and has been so for centuries). If it didn't exist, political discourse would require that it be invented.
Re: look ma, no hidden variables
Does superluminal signalling automatically imply time travel?
I've read that it does. Interestingly, it's the IT angle of time travel (even if restricted to information only) which seems to me the greatest paradox. The problem is that if you can send even as little as one bit back in time, you can arrive at the result of any convergent iterative process in the time taken to compute a single iterative step (by sending the result of that iteration back in time to replace the initial approximation).
It's only a slight stretch to claim that such a computer would inevitably transform line noise into a strongly superhuman, perhaps ultimate, intelligence. So akin to the Fermi paradox, where is IT?
Am I the only person surprised that Intel hasn't used a bigLITTLE design? (ie, one with a much-simplified core for housekeeping when there's very little going on, sharing state with a much faster core to which it would hand over when things get too busy). Can they dynamically shut down so much of a core that they don't actually need to use silicon real-estate for a separate housekeeper-core architecture?
Re: In an ideal world...
I'm trying to think of any reason why HTML5 needs to contain the DRM standard, rather than just having a separate standardised DRM.
Surely if the media lobby wants a single world-standard DRM they should decide on one themselves and put it forwards to ISO. Once standardised, users would download just one DRM entity to handle all standards-conforming media sites, and the (non-)problem of twenty incompatible DRM systems would be solved. Some browsers might even choose to implement ISOxxxx natively rather than as a separate user-requested plugin.
The only conclusion I can draw is that the media lobby has an ulterior motive in trying to get DRM into html5. Tell them to sod off and standardize it amongst themselves without fouling the HTML standard!
You could ask the same question about Gold, or any other fungible store of value. Whatever your nation enacts as law, is probably as good an answer as you'll get. In the UK, the rules are different for Sovereigns (=UK currency), Krugerrands (=SA currency) and gold bullion bars (commodity, not currency).
The irony is that Bitcoin is exactly the sort of thing that libertarian-leaning right-wingers ought to be welcoming with open arms and trying to protect. It's early days, but the intention is to create a form of currency that's (a) as "hard" as gold, (b) electronically tradeable, and (c) anonymous, just like fiat paper used to be, until circa 1970.
Shows how far the Republican party has fallen, that it will shoot itself in both feet just to bash the Democrats. Do they WANT the state to be able to trace every financial transaction by every "free" citizen? Do they WANT the state to be able to hyper-inflate everybody's savings out of existence? I feat that both of these things will be visited on the USA, and probably the UK, within my lifetime.
Re: it rapidly fills to a point of becoming unnavigable
It's not a design flaw. It's an easily user-configurable option. It's also perfectly usable, if non-optimal, straight out of the box, because it's self-documenting. A program can't read your mind about where to insert itself, so it puts itself under its manufacturer's name and leaves it up to you to move or copy its entry to your place of choice. (Some programs do ask; more should). It also remembers the things you use most often and offers them to you at the top level (which is probably why most people never bother re-organising it).
If there's any problem, it's remembering how to change language, if the usual user of the PC is Finnish, or Arabic, or Chinese. But that's a general problem not specific to the start menu. Would have been so much more useful if there was a button (called Babelfish? ) on the keyboard, instead of a Windows button.
Re: I'm Sick Of This
It's possible to run KDE on Windows (and possibly other Linux desktops).
Reference please! It can't be any more shit than those tiles!
Re: Confort zone
No, that's not how it is.
While one is learning to use one's tools, one has to think about the tools, and one's productivity is low. When one has mastered them, one doesn't really think about them at all. They are just a part of one's environment, while one thinks about the actual useful work that one is accomplishing.
It's the difference between learning to ride a bicycle (if you can remember that far back) and riding one. You don't think about how to move the handlebars or when and how to change gears, you think about where you want to be and about what the other road users are up to. You can even day-dream, and you won't fall off, though failing to keep your mind on the other road-users can be fatal for other reasons.
Windows was wired into me back in the days of Windows 98. Thinking back, DecWindows and he AIX UI weren't so hugely different. (What? say the youngsters. Windows pre-Microsoft). XP SP3 was probably the best UI, with Gnome 2 a close second. Windows 7 was a tolerable irritation, like learning to drive a new car that you didn't actually want but couldn't refuse. Windows 8 is not a tolerable change (and neither was Gnome 3, but the open-source community rapidly fixed that stupidity, whereas Microsoft are very hard of hearing and have a monopoly on the source code).
Re: Bunch of nancies
Try configuring a PC for a university, or even a school. Large numbers of different programs. Students in each department and/or year use a different small subset of them. A menu is a perfect way to organise them. 1000+ tiles is random (or even defined) order isn't.
I'd be perfectly happy if those stupid tiles were a new option that an experienced sysadmin could turn OFF. But no, Microsoft insists on ramming them down everyone's throat. And that's only a small part of what's wrong with the Win 8 UI. It also insists on going full-screen at every opportunity. What I want to do is have several windows open on my nice large 1920x1080 screen, and switch between them with one mouse-click. That's WHY I have a large monitor (and they're cheap and flat these days, so it's no great expense - some of my colleagues have TWO).
If they don't fix it by the time Windows 7 is EOLed, I'll say goodbye to Microsoft even if I have to find a new job to do so. For now I'll just say no to 8. They've still got a while for the penny to drop, though it's looking a lot as if the board is going to have to eject Ballmer while there's still time to save the company. I've seen other big companies go down the plug-hole, because the boss's ego was too big to make the U turn that the customers were demanding.
Re: Since when did Microsoft start taking design cues...
Nice to see a mention of Cinnamon.
When the Gnome folks messed up with "3", the open-source community bitched about it a lot and then got on with doing what Microsoft are signally failing to do. One lot wrote a user-friendly environment to run on top of the new Graphics libraries (Cinnamon). The other lot rescued the old interface, warts and all (Mate). And of course, KDE, Gnome 2, XFCE, and dozens of others never actually got taken away.
I tried KDE, Mate and Cinnamon, decided I liked Cinnamon most, made "yum install Cinnamon" a standard part of my set-up repertoire, and stopped bitching.
Microsoft's "start button" is no Cinnamon. It's a fig-leaf with very serious caterpillar damage.
Re: PC buying slump and zero 'Comfy Chairs' at retailers...
Just tell them to buy one running Windows 7, because if they get "8" they'll hate it and you won't be able to do anything to help them.
If they're hard of hearing you get to say "told you so" and remind them that you said you couldn't help.
What if there are 1000+ apps installed?
We're a university. Our computers have abnormally large numbers of prograpms installed, because they are shared by students from many different departments, each of which needs a few different programs to the crowd.
A heirarchical menu is the perfect way to organise these. 1000+ tiles is not. This isn't the return of the start menu, it's a sop thrown in our faces. I wonder if it'll mollify enough people for Microsoft to get its misguided (or evil?) way with us?
There is evidence of sudden level changes in the form of certain estuaries. Sea level fell quite fast, geologically speaking, creating a new steep run for the river down into the sea. The resulting fast-moving water caused rapid erosion of a steep-sided valley. Then sea-level rose again, and the valley flooded with seawater, preserving its steep underwater form because erosion below the low-tide level is very slow. Take a look at Dartmouth or the Wye. It's very unlikely that the level of the UK land-mass could change fast enough to account for this.. The UK is geologically pretty stable (large earthquakes are rare, for example). The Andes are an example of rapidly-changing land levels - frequent catastrophic earthquakes, river-flows reversed in their valleys, and beaches raised tens of metres in mere thousands of years, as noted by Darwin. BTW glaciation creates a straight-ish U-shaped valley (or fjord), river erosion creates a wiggly V-shaped valley, so we can tell these southern estuaries aren't the result of glacier erosion.
One can also calculate what sea-level will be if all the ice-caps melt. At the very least, that should put you off investing in any land or property anywhere near to present sea-level! More seriously, the worst-case scenario is so bad that I do think we ought to heed the precautionary principle before it's too late.
Worst-case is permafrost thawing releasing methane causing global warming causing more permafrost thawing ... a positive feedback loop that won't stop until all the permafrost has thawed. There is geological evidence that such runaway warming events have happened several times in recent geological history. There is also geological evidence that the long-term stable (say 30Myear-average) situation for the Earth is with no large ice-caps at all and a MUCH warmer climate regulated by percentage cloud cover. Ice causes instability and positive feedback loops. We live in climatologically interesting times.
Laffer discredited? It's surely bleeding obvious that the receipts from a tax set at 100% are zero, because you'd have to be mad to earn under such a regime. So there must be a curve from zero income at 0% tax to zero income at 100% tax. The argument is over what shape that curve is and how it varies with time, for tax rates between the extremes, and the effects of changing the rates. This, like many interesting things in economics, cannot be measured with useful precision. (Though gut reaction says that if a government is taking more than half of what one earns, it's time to start thinking about leaving! )
A lot to be said for 0% corporation tax
Since so many companies can avoid paying this tax by relocating, why not scrap it?
There would then be no incentive to leave the UK and every incentive to relocate to the UK. More employment, so less benefits paid by the government to the unemployed. More purchasing by companies in the UK, so more indirect taxes raised. Maybe the government might find that it didn't need to raise taxes to make up any shortfall, maybe even the opposite. Worth a try?
Re: First of all ...
One reason I don't use smart devices that can't run Firefox with Adblock-plus. I *know* I can't stop their systems correlating everything they know about my browsing and targeting me with laughably useless advertizing. But at least I never have to look at it!
Good advice to all. ALWAYS be a buyer, NEVER become a sellee.
Re: Unknown brand
Brand X is how it should be. You should be able to have the hardware of your choice running the software of your choice. As phone CPUs get more powerful, the arguments for software that runs only on a particular phone, or only on a particular manufacturer's phones, will look weaker and weaker. Lock-in? No thanks.
Samsung has just released a pure-Android version of the Galaxy. Looks as if they understand, and they should concentrate on what hey are good at. At present, they're much better at hardware than software.
Grow your own!
People seem to be under the mistaken impression that the only natural colour for a tomato is red. Actually, they can be red, orange, yellow, green(but ripe!), purple, or near-black, to say nothing of striped or blotched in various combinations of these colours. This is just using the normal techniques of selective breeding to create new varieties. Check out a good seed catalogue.
The tastiest tomato variety I've ever tried is called "Black Russian" and is green-purple-black with much the same aesthetics as a really bad bruise. Unsurprisingly you won't find it in supermarkets. I'm not sure if it's grown commercially at all.
The uniform red globes you get in supermarkets are bred for high yield, visual appearance, long shelf life, and resistance to bruising in transit, with flavour in a rather distant fifth place. (One commercially-grown variety is called "Moneymaker", which rather gives the game away)
Re: What it's for
few governments are actually evil. And most are trying their best to do the right thing, incompetence permitting.
I'd tend to agree, except that they've convinved themselves that they are doing good in a circumstance where there may be nothing good to be done. We're now the subjects of an unprecedented experiment in economics. (If you think the 1930s were the precedent, remember that led to WW2, so I very much hope not). Power corrupts. Government is forever taking more power. We know what is proverbially paved with good intentions.
UK govt. has "printed" £375 billion by QE in recent years. That's over £6,000 per citizen. Some day, it will come flooding out into circulation. When that happens, I fear that the 1970s will be seen as a mere dress rehearsal for what is still to come.
I can remember the 1970s. I hope I'm wrong.
Same as other curencies!
It's all well and good to use progressively more difficult computational work to underpin bitcoin but that work uses not just real world time but energy.
The main reason gold is a store of value, is that mining more of it is expensive.
As for government paper, what is the energy cost of all the enforcement mechanisms needed to prevent forgers and fraudsters devaluing the currency? (Fraud squads, auditors, bankers all consume energy). Much harder to quantify, but it would be fair to say that if it wasn't substantial, the currency would be sunk pretty fast.
What it's for
Like any currency, different things to different folks.
It's the seed of a currency that is not under the control of a government or any other human agency. in that respect it's a hard currency like Gold. But physical gold is expensive to safeguard and validate and impossible to use for e-transactions, and "paper" or e-Gold is vulnerable to fraudsters (who basically sell the real gold out of whatever vault it's supposed to be in, leaving a hollow shell of un-backed paper gold.
As for who is using Bitcoins: all sorts of people, but I fear Bitcoins are of greatest utility to those with most to hide. Apparently there are shops out ther in cyberspace on the Tor network, that supply illegal substances. You buy using Bitcoins, untraceably, and a packet may later arrive in the post which may contain your choice of illegal drug. Buyer beware, but E-bay has proved that the straightforward sort of fraud is not profitable. You can't maintain a reputation while ripping off many of your customers. Especially not for low-value transactions that promise repeat purchases. It's better to cultivate the repeat business. What works for legal products appears to work for illegal ones. (All allegedly ... I have no idea beyond what I've read about it).
It seems to me that the more governments gain access to our transaction histories and employ "big data" computing, the more attractive the Bitcoin will become. I don't mind Tesco having a record of everything I've ever bought at Tesco (except the cash no card transactions!), but a world where the government's computers analyze every single purchase and financial transaction that I've ever made is not apealing. Enter many more Bitcoin users?
Which is one reason governments hate Bitcoins and would like to kill them off if they could. Their propaganda (with more than a grain of truth) is that Bitcoins aid money launderers and tax evaders. Their other reason is that if Bitcoins flourish, it will be at the expense of all governments' ability to print their own money, and thereby rob (or tax) us all by inflation.
We live in interesting times. (Smiley face just worked out that the circle on the inside of the empty envelope is the bit he's supposed to eat).
Re: OK. What's wrong with XP?
Quite a lot is wrong, but the faults are well-hidden "under the hood" to use a car analogy. Your mechanic (IT staffer) deals with the bad bits. You use the UI, which is actually rather good (and I say that as a 75% Linux user). You don't know anything of the bad bits until it breaks down.
Windows 7 is much better under the hood, but regrettably changed the UI in a lot of ways which didn't make it better, just gratuitously different. It was an annoyance, but soon enough one got used to it. Same level of annoyance as buying a new car from a different manufacturer. Major stuff the same, minor controls all different but same basic functions once you find them.
Windows 8 is slightly better than Windows 7 under the hood, but the UI sucks mightily. Square steering wheel, pedals swapped around, instruments that require you to take both hands off the steering wheel to use them ... like someone resurrected the controls from a Model T Ford and piled a load of 21st-century stuff around them. Couldn't happen with a car, because that sort of change would kill people.
MS can still rescue itself. It should give XP at least one year longer before ceasing bugfixes, and develop a smooth migration path from both XP and 7 to Windows 9 for techies and unskilled users alike. But they are walking ever closer to a cliff edge with their eyes shut.
Re: Real Solutions
The solution is not to use Classic Shell or Start8. It is to not buy Windows 8. At. All.
The problem with that as a corporate user is that we have downgrade rights, and that the hardware sold with Windows 7 costs MORE than the same hardware with Windows 8. And since we always blow away whatever crapware a PC ships with by re-installing our customized image, we're "buying" Windows 8 if MS cares to view it that way.
However, I'm sure that they can count activations and work out the truth of the above. If they want to. MS has (maybe had) an "Emperor's new clothes" problem.
Re: Smart watches are still pretty dumb
Yes, a watch needs to be readable without pressing any buttons.
Is it beyond our ken to make a screen (LCD?) that de-energises to transparent? Then put a conventional watch with hands behind it.
The other way would be if it could sense the characteristic gesture of turning the back of one's hand to face one's face. I doubt whether the tech to do that on a watch-sized battery is yet available.
Re: AMD senior manager of client, server, and embedded products Gary Silcot
Excludes mobile products? (embedded may mean machinery and vehicles, excluding smartphones). Or does it indicate that AMD has been in distress, and two or three posts got merged?
38 per cent preferring Windows 8 ...
Compared to 35 percent for Windows 7. I presume the other 27 percent declined to answer the silly question because there were no tick-boxes for [ Windows XP, Macintosh, several flavours of Linux Desktop, Android, ...]
I.e. they asked would you prefer to be (a) boiled alive or (b) eaten by fire ants. Or didn't ask at all and just made up some numbers.
Re: It's no joke
If they're attracted to relays and motors (?motors with brushes only?) then I'd hazard a guess that it's Ozone or
Nitrogen Oxides that attract them.
Re: Where's the advantage over free space optics?
How much do lasers that can be modulated at 40Gbits/second cost, compared to this transmitter? That may well be the main advantage.
And does a 250GHz carrier go through or around a bird? Line of sight optical won't.
Re: You can call me AI
In the Culture universe, there's no competition for resources, expecially not between Minds and humans.
I expect that if we ever get as far as AI in our own universe, something similar will happen. Once we've accepted that AIs deserve to be treated as autonomous thinking creatures with "human" rights, it will become apparent that silicon-based intelligence is much better-suited to vacuum than to moist oxidizing atmospheres. So the AI-expansionist-tendency will expand outwards, leaving a few human-loving AIs to get along with the bio-life that can't breathe vacuum.
They''d also be much better-suited to the deep time needed for interstellar travel at less than light-speed. Somewhat ironically, the way to make ten-thousand-year journeys tolerable is to slow down one's clock-rate, thereby greatly reducing the subjective span of time.
Re: "the incumbent always misses the next wave"
Have you never noticed that the railways tend to follow the canal routes?
Thi is not accidental. It's because they both had the same underlying need: an optimally un-hilly route from A to B. And so the railway companies bought out the canals for their rights-of-way, or the canal owners moved themselves into the railway business. Not sure if there's anything analagous in computer tech.
Re: PC shipments ffailing for ONE reason
I'd prefer new things to upgraded things.
Try comparing a 3-year-old PC upgraded to 8Gb RAM and an SSD, against a new bog-standard PC with a conventional hard disk. You may change your mind. Admittedly, this presupposes that you don't actually need more than 250Gb of storage (i.e. that the 500Gb or 1Tb hard disk will never get anywhere near full).
At the place I work, there are going to be a lot of SSD upgrades in the near future. (It holds the PC's software, all the user data lives on a server).
Re: PC shipments ffailing for ONE reason
Oh, they'll need replacing ... just not so often. It used to be standard to replace PCs every 3 years, because the technology was advancing so fast that a 3-year-old PC was obsolete and verging on unable to run today's software. That cycle is slowing. Now a 5-year-old PC is still usable, and today's state-of-the-art systems may be usable in seven years. Also (in the West anyway) everyone who needs a PC has got a PC. That's why the tablet market looks so healthy by comparison: lots of folks haven't got one yet. But it won't be long until the market for tablets is also saturated.
As for making them to fail as fast as possible: a manufacturer that does that will soon be recognised as selling short-lived crap. So they might get a short-term boost from making them cheaper than the competition and selling them at the same price, but you can only fool most folks once. Alfa Romeo never really recovered from making a car that rusted to death in three years instead of ten (ten was then acceptable ... not any more!)
Re: Going, going, gone.
Thinking along those lines, it might be better if I stay with Sky for a while (3 months?) and THEN tell Murdoch where to stick his service. This way BE gets to walk off with more of Murdoch's cash and therefore Murdoch has less of it.
Also my reason for leaving is made better-known to the cause thereof.
There are recent instances of new languages arising quite spontaneously, when groups of people without a common language find it necessary to communicate. Tok Pisin (Papua New Guinea, "pidgin english") is perhaps the newest. English itself was once such a language, born of mutual incomprehension between Norman French invaders and Anglo-Saxon natives. It's since evolved in a different direction to both parents, most noticeably by progressively jettisoning its formal grammar.
There's also a new sign language, born of deaf children being dumped in orphanages and left to rot (i.e. given no guidance on existing sign or other languages. So they invented one.)
- NASA boffin: RIDDLE of odd BULGE FOUND on MOON is SOLVED
- SOULLESS machine-intelligence ROBOT cars to hit Blighty in 2015
- BuzzGasm! Thirteen Astonishing True Facts You Never Knew About SCREWS
- Worstall on Wednesday YES, iPhones ARE getting slower with each new release of iOS
- Microsoft's Euro cloud darkens: Redmond must let feds into foreign servers