1871 posts • joined 15 Jun 2007
No, it's not magic
Apart from the stored heat, there is also tidal energy in the mantle, caused mostly by the gravity drag as a result of the Moon.
There are also some people who think that there are fission reactions happening at various points in the mantle and core generating heat.
Not current accounts. I doubt they are *that* stupid.
What makes you think that multiple heads per platter is new?
I saw a magnetic drum device (swap on an IBM 360-65) which had something like 8 rows of heads spaced around the drum to increase speed.
Deskstar and Ultrastar
I don't know whether this is relevant, but I have used IBM and Hitachi Ultrastar (the enterprise version which actually shares much of the mechanical components), and generally they have been quite reliable. I have 3 in an IBM RS6000 44P-170.
Thanks for the suggestion. Unfortunately, most people who end up using powerline ethernet have already tried WiFi, and have moved on because WiFi could not cut the mustard for reasons including thick walls, inconvenient location of the ADSL that the WiFi plugs into and congestion.
Am I ever glad about the decision
I've always said that cached thumbnails were a potential danger to anyone browsing the Internet or using a web-based mail client. I hope that this will set precedent.
I agree that the victim has been awarded a pitiful settlement, and I would be very interested in finding out in 6 months time what a CRB check shows for him. I think that if I were his lawyer, I would go for interim damages pending their clients reputation being restored. If this appears in an adverse way on his record at some time in the future, then they would be able to go for lifetime damages that should be enough to support this person for the rest of their natural life.
And it would be even better if the claim would have to be contributed to by all of the members of the prosecution rather than the tax payers.
I think we all have to agree...
Here are my cards. I've used Unity on the 10.10 netbook release, and IMHO it sucks big time. I've not tried 11.04 (yet) and admit that there is a good chance that it will work better but I don't normally run Betas, as I have enough to do in my life without having to fight bugs that other people take it upon themselves to find.
I am in "the desk top is there to control as many windows as I need" camp, and as a result, Unity is completely against the way I work.
But, and this is a big but, I am not a typical modern non-technical user. I work everyday with people who maximise whatever they are doing on their 22" 16x9 screen, work with tabbed terminal sessions in Konsole rather than multiple windows, and generally do one thing at a time. These are either OSX and iPad generation, or (bizarrely) people who have come from the green-screen single session terminal age and who have never felt comfortable in a windowed environment.
Having seen both of the above struggle to switch from one application to another, I believe that my multiple windows spread in a consistent manner across several desktops is better. But that's my opinion. I believe that I work faster than them, but again, that's my opinion.
I think that it's fair to say that Unity works well for those people who work in the way that Unity works, and it does not work for those who don't. By hey! This is a big world, and not everyone is the same.
I think it is necessary to explore new interfaces, as I am sure that I would not necessarily want Gnome or KDE as they are now on a touch-screen device. But having also been an Android user for around 6 months now, I'm not sure I would be happy with that on a tablet either, at least not on one that is more powerful than my netbook.
So I am happy that there is a new interface being tried, and even that it is the default (as long as they get the bugs out and make the alternative easy to choose). I just don't want to be forced to use it without jumping through hoops of fire if it doesn't suit me, and I would be grateful if others accepted that this is a perfectly valid viewpoint.
It's not that it's purple that bugs me...
...it's the blotches of other colours that make me want to reach for a screen degausser, until I remember that that has not been a problem for 20 years on CRT screens, and has never been a problem on flat-panel monitors.
And at least on the login screen, it is not obvious how to change it.
Yes, I've done it now, so I don't need anyone to tell me, but it should not be difficult on an OS aimed at ordinary users.
Yes, but there are legal streaming/download services available, so it is possible to use all that capacity without breaching copyright.
While it is technically possible that you may own the copyright to some CDs, this would have to be by contract, agreement or by actually being the person who recorded the material. As CD as a media is only about 30 years old, it is a fairly safe bet that nearly all material on any CD is still in copyright.
OK, ok, I know that there are compilation CDs out there that you can buy that contain already-copyright-expired material (almost any song recorded in the 1950's or earlier will fall into this category), and I know that EMI are getting twitchy about the Beatles back catalogue, but at present this is a minority of CDs.
One thing I want to throw out there is whether a re-master of a work recorded more than 50 years ago (in the UK) resets the copyright date. In theory, it could count as new material, but as IANAL, I do not know whether that is the case.
Any thoughts, anyone?
For those who don't recognise it
Take the flute riff. Work out where the bars are (clap 4 beats to the bar, along with the rhythm) such that you get 8 bars or 32 beats for the whole riff - it's a middle 8 after all) and it is the third and fourth bars of the flute riff, which matches the first line of the kookaburra song. It is pretty much note for note, with the same rhythm as well.
It's quite clear that it is the same, but I ask you, how can 2 bars, 11 notes in total be worth 5% of the revenue of the song. Especially when the original song is, I'm sure, meant to be onomatopoeic. At least it's the writers revenue they are going after. That should limit the payout significantly.
I'm absolutely certain that this ruling will lead to Australian lawyers finding two bars of matching melody in different songs, and then using this ruling as the basis for a copyright infringement claim. I'm sure that it could be done by computer if you had access to the right MIDI files.
There is at least one song in the UK chart at the moment, which every time I hear it, I say to myself "That's a song already", but I just can't put a name to it. I must try to remember to put more effort into working it out.
I feel that this should be dismissed exactly as EMI put it, as an unconscious tribute to the original song. Otherwise we will have to have melody searches against all songs currently in copyright before a new tune is published, in the same way as we have to have patent searches now. This will stop amateur writers dead in their tracks in the same way as patent searches stops small inventors now!
Re. MS, you also forgot
FUD campaigns, illegal (or at least immoral) discounting deals to PC manufacturers if competitor products were not installed, pricing models that penalised OEMs if they offered systems without an OS or with an alternate OS installed, including open-sourced code in products without acknowledging it's origin, random buy-up and shutdown acquisitions of competitor products, committee stuffing to undermine genuine open standards, participating in what amounts to patent cartels, unnecessary license purchases/ financial loans/bailouts with strings attached to control struggling competitors and ... oh well, I could go on, but it's all a matter of history.
One wonders what would have happened if Apple had not won it's case against Digital Research and GEM, which assured that MS had a significant clear-run at a graphical user interface in the Intel PC market. In my view GEM was significantly better to the versions of MS Windows that were available in the same time-frame.
I think you meant
Windows 98 SE and Windows ME, not Windows 98 ME. They were two different versions (ME following 98 SE), and while 98 SE was seen as good, ME was regarded as pants by pretty much everybody.
Didn't Apple say something about OSX servers recently....
Oh yes. They're stopping making them.
So we'll have racks full of Mac Minis then.
I'd just like to point out
that AT&T up until the late '80s used to run a large part of their environment around mainframes, many running UNIX! And you probably ought to look up other non-IBM OS's for 370 architecture systems as well. One of my personal favourites was MTS. I saw a demonstration of access to ARPANET (you know, a forerunner of the Internet) from this OS in the very early '80s. Also, for all it's problems, the influential OS Multics was a mainframe OS, and this set features that would appear in UNIX, VMS and a host of other OS's long forgotten.
I was involved with installing and running a channel-based Ethernet device running TCP/IP on a mainframe linking it to Sun and VAX systems in the later '80s (again, under UNIX).
I think that one needs to separate the hardware from the software, as there is a significant difference.
Mind you, if you look as some of the innovations, such as virtual addressing, virtualised systems, key based page level memory protection, I/O offload, multi-processor systems, distributed processing, hierarchical storage controllers, DMA, memory cache, multi-user and multi-tasking, use of ASCII (one of your benchmarks, ASCII was mandated by US government contracts in 1968, and before this was a COMMUNICATION standard, not a COMPUTING one), microcode, solid-state electronics and a host of more minor things, mainframe was often one of the first systems to implement them (often because the features were so expensive to implement, only mainframe-class machines could benefit).
Whilst many of these were not invented on the 360/370....zSeries systems (now the only real mainframe architecture remaining), they were almost all pioneered on mainframe-class systems like Atlas, KDF/9, Cyber/CDC, UNIVAC and others.
Damn. Forgot that the Reg. stripped multiple spaces out of comments. All of the note changes happen on syllables.
@AC re Calculators
I remembered having to work in base 12, and base 20 for £sd and base 16 and 14 for imperial weight as I hit submit.
Even though I was well versed in Maths at the age of 10/11, decimalisation still caused me problems when my one shilling of pocket money became 5 new pence.
"Decimalisation, decimalise, decimalisation will change your lives"
g a b cb g g a b c f g a a#af f f f g
Key might be wrong, and the a# should probably be written as a b flat, but I can't seem to see a flat symbol on the keyboard!
@AC re Joke - again
And you can't convert decimal integers to octal either!
99 (decimal) actually equals 143 (octal)!
For this one, you're getting the pedantic Maths teacher!
@AC re. Joke
No. I was serious and yes, I know that 8 is 010 and 9 is 011.
I was presuming that the person was using a calculator which worked in octal and decimal (and thus had 8 and 9 and point keys) but which was in octal mode, so that when they were typing in something like 18.49, the calculator actually registered 14 (neither the 8, the 9 or the decimal point would have registered). Would get the sums very wrong.
If you had actually bothered to think of the mechanics of it, you would have understood.
By the way. I think that your floating point octal to decimal is incorrect.
When writing non integer octals to one significant digit, the numbering would be
0.1 octal, which is 1/8 (0.125 decimal)
0.2 octal, which is 2/8 (0.25 decimal)
0.3 octal, which is 3/8 (0.375 decimal)
0.4 octal, which is 4/8 (0.5 decimal)
0.5 octal, which is 5/8 (0.675 decimal)
0.6 octal, which is 6/8 (0.75 decimal)
0.7 octal, which is 7/8 (0.875 decimal)
1.0, which is 8/8 (1.0 decimal)
So in Octal 0.5 + 0.2 + 0.1 will equal 1.0, which it needs to do in order for arithmetic to work.
The first significant digit after the octal point (geddit) is 1/8th's, the second is 1/64ths, the third is 1/512th's and so on.
This means that by casual inspection, 0.44 octal HAS to be larger than 0.5 decimal.
By my calculations 12.44 (octal) is (1x8) plus (2x1) plus (4/8) plus (4/64), which makes it 10.5625 (decimal) or 10.56 rounded to decimal pence.
I can't see how you got 10.14. Even if you had worked in pence, 1244 octal is 676 decimal.
You got the 0.95 decimal correct, however.
You could do the exact arithmetic if you worked in pence or cents. Non-integer arithmetic in any base other than 10 hurts my head.
If you look at physical calculators that work in octal (rather than 'soft' calculators on PC's and smartphones that can change the keyboard layout according to the mode), they normally do have an 8 and a 9 key (and a decimal point key), because they normally also work in decimal and hexadecimal.
These keys are normally disabled when the calculator is in octal mode, so you could press them, but I think I would notice that they had not registered. Also, as far as I am aware, nobody has produced a calculator that does non-integer arithmetic in anything other than base 10 (what a mind-bending concept that would be!).
For reference, look up the Texas Instruments Programmer Calculator that was available in the '80s, and any number of modern scientific calculators from makers like TI and Casio that also work in different bases including octal.
BTW. I was working in Octal on systems before PC's were invented, so I do understand it. I learnt clock arithmetic in bases other than 10 when I was about 8 in the 1960's, when they actually taught Maths in junior (primary) school.
Did you realise that Humans were meant to have thirteen fingers?
It's obvious, because in the HHGTTG, the ultimate question and answer is "What do you get when you multiply 6 by 9". Answer Forty two.
This is indeed the case if you work in base 13.
The only reason we work in base 10 is because we have 10 fingers. In some instances, it would actually make better sense to work in base 6, because you could then use one hand for 0-5 and the other as a carry. This enables you to count up to 35 with your two hands.
Gawd. I think up such crap!
Octal for checkbook balencing
I find this unlikely, as so many prices have 8's and 9's in them. Think of all the things that are one penny (cent) lower than a round number of pounds, euros or dollars. I guess that it depends on what your calculator does when you hit 8 or 9.
IBM trademarked the name TrackPoint
but other names include TouchStyk. Other names can be found in the relevant Wikipedia article, and many other good information sources.
IBM's name was the original, though, as they developed it to a product even if they did not invent it.
that when you clicked on a video in YouTube from Firefox on Ubuntu (at least, up to 10.4), it pops up a box that says that you have to install Flash, and then directs you to an Adobe page that asks which package you want to load. Once decided (and working out which format you need is about the only tricky bit), it downloads it and asks how you want to open it. Selecting the default will open whichever package manager is the default (IIRC it was gdebi last time I did it) and install it on your system.
OK, this will not be the version from the Ubuntu repository, but it should work.
As far as I remember, this is pretty much the same for Firefox on Windows.
Again, I believe that Totem will ask you to enable various options (lame and dvdcss) the first time it encounters a .mp3 or a DVD.
If you had a proper multi-user, multi-access OS with proper security deployed, you could get away from this whole 'personal real/virtual' machine and software deployment model with all of the associated security problems that is blighting us.
Oh, how 20th century of me to bring up diskless shared image and thin client access UNIX systems that were being done in the 1980s and 1990s. It was not without fault, but was far better than using a crowbar to squeeze multiple virtual systems onto a single piece of hardware, with all of the associated duplication and waste that this entails.
One of my mantras has been "There is no place for a personal computer in Business" for a long time, and I believe that it is as true now as it has ever been.
It really makes me want to go back in time an nuke Redmond even more.
I understand making data accessible. And I also understand that having relationships between data items makes a lot of sense. But I really doubt URIs are the way to do it.
My concern is that by using URIs (at least the way I understand them to work) will effectively hardcode location and shape information into the datasets in the same way that a schema does in a relational database, but with a fixed location. Unless someone can indicate otherwise, I believe that this makes the data almost completely non-portable.
OK, in a web-centric world this may make sense, but unless someone puts some clever caching technique, it means that you will only be able to use the data when you ate connected.
Sometimes you want to take a fixed snapshot, or make sure that the data in your thesis does not change between you writing it, and it being read by your moderator.
I'm all for making data easily useable (god knows I've spent enough time massaging data over the years), but to tie it to the Internet should obviously be stupid to anyone unless they are from the facebook generation.
I'm also not certain that it is reasonable for the person who originally structures and creates the relations in the data to be able to anticipate how that data will need to be used in the future. Today's data mining systems are all about making assiciations between data-sets that were never imagined when the data was recorded.
Something like an encapsulated schema in the data set would be a great advantage, but you would have to have someway of normalising not only the data sets, but the schema's to allow automated queries.
Probably not a single supercomputer.
That seems like an awful lot of hardware for the budget, and the dispersed nature means that it is far more likely to act as a group of smaller clusters that talk together than a single super-computer.
You will never be able to drive the WAN links at sufficient speed to spread anything other than encapsulated data type problems (like SETI@Home but larger) to the remote sites.
I would be interested in seeing how the power spreads around the six sites, because although the total amount of compute power may seem high, chances are that the power in any single part of the environment will be a fraction significantly under half of the total. That should put it much further down the top 500 that the 'low 30's' quoted in the article.
Also, by the time it is delivered, there will be new systems springing up in China, USA, Germany and even the UK!
Oh well. I have contracted for Fujitsu in South Wales before. Maybe I ought to dust off the old CV. Might be interesting to do some non-AIX work for a while, and I now have Infiniband experience.
Come on. It's really not monotone
I found most of a single scale in the 'song', at least seven notes, although most of the song stays on the three notes of a major chord.
I think many of the readers here ought to listen to the 'singles' chart nowadays, because they will be appalled by what is counted as music by the people who actually buy it in volume. I must admit, however, that I was intrigued yesterday to hear two versions of Adele's Someone like you (the normal version and the Brits version - both head and shoulders better than much of the rest) on the chart at the same time.
Autotune, whether it is required or not, is added to the vocal in it's most intrusive, buzzy manner for effect on so many songs now. I hesitate to say this, but JLS who obviously can sing a bit (no autotune allowed on X-factor live performances after all) have it on most of their songs now.
I do wonder what a 13 year old girl will be doing while "gettin it down" and which makes "we so excited" (sic) while "partying" which is legal! Sex and drugs and alcohol should all be out.
Anyway, I have to decide whether to listen to Planet Rock or Radio 3 on my way home to flush this meaningless and annoying fluff from my mental musical cache.
Whether the control rods are above or below depends on the design. In BWR instances, the rods are below the core - see the wikipedia article on boiling water reactors that actually have a diagram of the Fukushima type reactor, which is why I used the terms "inserted" and "removed" rather than "raised" and "lowered".
The simple fact is control rods in, reactor slowed. Control rods out, reactor quickened.
There are also different types of rods in some other types of reactor. There are moderator rods, whose purpose is to slow fast neutrons be become slow neutrons, which will actually speed up the reactor, and then there are the control rods, which are intended to quench the neutron flow to stop the reactor.
In a BWR type reactor, the whole core is immersed in water, and the water itself is a neutron moderator. There are only one type of rod, and these are all control rods. This is very different from PWR and AGR type reactors.
Again, this is what I understand from years of casual study, so I am not an expert.
The control rods form a part of the control system. They are not normally either completely in or completely out, they are normally partially inserted to control the speed of the reactor and thus the energy output. Whether they are above or below the core depends on the reactor design. These are apparently BWR (boiling water) reactors, and the rods are below the core, and held against hydraulic pressure by electromagnets or similar, such that should there be an interruption in electrical power, the rods will be automatically inserted by the pressure. This is a fail-safe system.
The rods allows the operators to 'damp down' (insert the rods) the reactor in times of low power demand or maintenance, and open it up (withdraw the rods) during periods of high demand. Under normal operation, you would never completely insert the rods, because that would stop the critical reaction, and effectively stop the reactor.
In the case of a serious event (such as an earthquake), it would be normal to completely insert the rods as a precautionary measure. This would effectively make the reactor subcritical, which will cause it to cool and eventually shutdown. This does not make the reactor immediately safe, but will remove any chance of it melting down. Most of the residual energy in the core will come from decay products of the U235 fission reaction that are themselves radioactive with short half-lives, and thus will spontaneously breakdown, releasing energy in the form of heat. These will breakdown naturally over a matter of days to the point where the reactor will generate less heat that it will loose through convection or conduction, and thus become 'cold'. This is what I think is meant by 'cooling fuel'.
It is this gradual breakdown of the decay products that requires cooling until a sufficient amount of them have decayed to the point that natural cooling will be greater than the heating effect.
Conversely, during startup, removing the control rods will allow the neutron flow to increase (U235 will always spontaneously decay and produce neutrons even in a non-critical reactor) until the critical point is reached, and the reactor becomes self-sustaining. Looking at sources, it appears that for a completely shutdown or new reactor (one with no uranium decay products in the fuel rods), a source of neutrons can be used as a 'starter' to speed up the build up of the neutron flux to achieve a critical reaction more quickly.
For anybody who is worried by the term 'critical', this is not being used as in 'dangerous', but as in a tipping point, in this case where the nuclear reaction becomes self-sustaining.
If you trust it, there are very good articles on nuclear reactors, BWR type reactors, control rods, and nuclear starters in Wikipedia. These appear quite objective and appear to me to be trustworthy, at least they do not conflict with other sources I have read.
..."It has taken major efforts by humans to keep them from going critical."
Please be careful with your use of 'critical'. As far as operating nuclear reactors are concerned, 'critical' is normal. Misusing the term may lead those who do not understand the terminology from becoming needlessly alarmed.
I admit that a reactor being shutdown should not be critical once the control rods are inserted, but I seriously doubt that in this case, the cores would have become critical in the nuclear sense even if the cooling had completely failed and they were damaged by heat.
The design is such that if a complete meltdown could occur, that resultant puddle of radioactive mess would be distributed over a large enough area such that a critical mass would not pool in any one place to allow an uncontrolled nuclear reaction to happen.
@Steve the Cynic
I'm a UNIX person through and through, and the first time I looked at this was in about 1987, with SVR2, which had code for DST, but had the cut-over dates to and from DST hard-coded in libc.
There is a configuration option that allows you to vary whether the clock is localtime or UTC without having to re-compile the kernel (a bit heavy handed in this day and age). It is based around setting UTC=yes early in the boot process (in /etc/default/rcS). The initial setting for a new install is supposed to be queried during the install process. I suppose I may have set it wrong, but I don't think that I would have made such an error, bearing in mind I was aware of the problem. Maybe I'll install again from the original media I had to see whether there was a flaw in the install process.
In my (biased, I admit) view, it's wrong to run a system on localtime, but I am in the UK, and winter time is the same as UTC (well, GMT anyway), so I have never had to worry about anything other than the Daylight Saving Time change. I guess that other locations have it harder.
Any reasonable system should run it's internal clock on UTC all the time, regardless of timezone and whether DST is in effect, and just alter the presentation of local time according to it's location, so 'adjusting' the clock should never be required. You should never have have the hardware clock changed on a correctly configured UNIX or Linux system, except to take into account leap-seconds or clock-drift.
This has always bugged the hell out of me when using a dual boot Linux/other-OS PC. Linux worked just swell changing the presented time according to DST, and then you boot your other operating system that ALTERS THE FREAKING UNDERLYING CLOCK!!!
When I put Ubuntu Lucid Lynx (10.04) on my Thinkpad last year, I was expecting this, but found that some misguided bright spark has added code to Linux (probably somewhere else other than Canonical) that actually expects the underlying clock to be changed incorrectly by the other-OS, and then 'breaks' the working tradditional UNIX/Linux time support to take this incorrect clock setting into account. Talk about working around someone else's errors.
Good for all the people who want it to just work and regularly boot both OS's, bad for anybody who actually understands what should be going on. Drove me nuts for an hour or two.
Hang on a sec. Back to iOS. Isn't it a UNIX/BSD derived OS?......
The trojan is an ELF executable, presumably for whichever processor runs in the Dlink router, but the vector to get it in there would appear to be a compromised MS Windows system that then attempts to brute-force access to the router. So there are actually two components, one of which infects a windows system, and the second of which is installed on the router by the first.
Thanks for explaining, although I did post this as a springboard to get replies.
As I have supported diskless UNIX systems for several years in the past (and will be again very shortly), I do understand about sharing a system image (which, incidental, on windows breaks a whole host of software unless you jump through hoops to redirect stuff away from the C: drive, which will be read-only, somewhere else - personal experience of pain here), and also identical hardware on the desktop. It's not a new technology except to Windows shops.
Citrix, VMWare and Microsoft are waaaaaaay behind the curve here compared to UNIX, both in diskless and remote display, and I have to feel that bending current windows to make it fit into a diskless/remote display model is the wrong way to go about it. Better would be to have made a 'new' windows with native thin client support and some compatibility with 'old' windows, than using a crowbar on the existing models. After all, MS did product switch before with NT. Maybe Longhorn should have been this, but they apparently could not get it to work without ex DEC system architects and IBM's assistance (WinNT history 101).
And I did talk about de-duplication, which is effectively what shared image is all about, and I did also talk about low power, diskless desktop display systems, but after a quick search, the only people I could find selling them was Wyse, who sell a diskless system running Windows CE for about the same price (once you factor peripherals in) as a basic PC. Many people in the past tried diskless PC's, and almost all of them are now NOT doing it (the earliest I remember was DEC Pathworks, which had diskless DOS systems with a network filesystem).
My closing comments about having been here before with other architectures still stand IMHO. I still think we have been here before, and I also still think that the current in-vogue implementations are flawed and designed to maximise revenue for suppliers rather than provide a good environment for customers.
So let me get this straight
You put all of your PC desktop images into large servers held in the data centre.
You then use something on each desktop to run a virtual session to those large servers.
What are those devices on the desktop? Oh yes, PC's.
I know that the devices on the desktop will be cheap/low power PC's, but bearing in mind how powerful even a basic PC is nowadays, where is the saving?
If you were to sell it to me as an administrative saving, or a deployment cost saving, or even as a data de-duplication saving, then I may be interested. But as a power saving?
Of course, if the desktop devices were diskless, low power consumption (ARM type power) real thin clients then this may make sense, but we've been here before, and commodity PC's always undercut specialist net devices (where are Tektronix, Oracle, NCD et. al. with their thin-clients, Netstations and X-terminals now. Oh yes, out of that business). The cost ends up being the screen, keyboard and I/O devices, not the PC itself.
Where savings are being made at the moment is that older low-power PC's are being used as the access devices, but this is unlikely to give you a power saving, and is not going to be a model for phase 2 and later roll-outs!
Does not affect LTS
LTS releases will still be 12.04 and probably 14.04. No difference.
6.06, 8.04 and 10.04 (Dapper, Hardy and Lucid) were for all versions (server, desktop, netbook), and that will not change.
@HMB - Word of advice
Check the native resolution of the display panel in the TV (it should be printed on the box or in the instruction manual). Too many TV's (and not always just the cheap ones) on the market today claim and will accept a 1080p signal, but will then downscale it to 1388x768 or 1680x1040 or whatever their native display panel can do.
I had a big argument with a major on-line retailer about this when the published resolution for a TV I bought from them was wrong on their website, and they were extremely slow to accept the fault. Even then, I needed to go through their onerous RMA process, which takes about 2 weeks, before they would refund.
I actually went through forcing the driver to override the EDID value read from the TV to prove the case, and actually at the end of the day concluded that using the VGA port rather than the HDMI or DVI port was far more flexible and gave more control.
I've just re-read my post, and it's odd.
I've written about fictional things, set in the future, but described them in the past tense!
It seemed right at the time, but it looks feels funny now. I think the common oxymoron is "a future history".
ST:NG Enterprise-B, C and E not lead ships
NCC1701-B was an Excelsior (updated) class,
NCC1701-C was an Ambassador class,
NCC1701-E was a Sovereign class.
Only the NX-01 (ST:E) was the lead ship of it's class.
Anorak please. Yes, with the Star Trek patches sown on the left breast.
...there is no guarantee that the overhaul that Obama wants is the one that most of the commenters here want. He just make it easier for Big Business to get their patents through. It depends on whether the problem he perceives is that the patents are not working, or whether they take too long to grant.
I suspect the latter (him being lobbied by the people with the deep pockets), which will probably make things worse from our perspective, not better.
"I always get the impression the utilities were written by clever people who just happen to be lazy and can't spell properly"
Actually, it was more like the Teletype ASR-33's in common use as terminals back when UNIX was being formed being sooooooooooooo sloooooooow, that people abbreviated commands and flags so that they could work reasonably quickly. It is also the reason why ed is incredibly terse, but ultimately very powerful as a line editor. I recommend to every serious UNIX/Linux user that they learn ed, just so they can use vi and sed effectively.
Another explanation is that they were lazy, a positive attribute for all good system admins. (do what's needed efficiently and with the least effort). At least, that's my opinion.
The downside of the VMS DCL command line model was that for every command, you had to add entries to the DCL command dictionary, whatever that was called (it's been a long, long time since I did any DCL configuration, and much of that was on RSX/11M not VMS). All of the command line parsing and in-line help that allowed abbreviations was done by the DCL command processor, although I do believe that it passed any unprocessed arguments through to the program as a last resort.
One of the most useful and irritating features (both at the same time) is that UNIX-like operating systems left that to the commands , although if you look at ancient UNIXes, there were very strong conventions that should have been followed (if you look back at the USENET archives, there are many quite animated discussions about whether the shell should do more command line processing than it did/does). This meant that you could easily add commands to UNIX, and that internal shell commands, functions, aliases and external programs appeared almost seamless.
One of the quirks, however, was that some of the ancient commands, many still used today, never did abide by the conventions (the most obvious of examples of this are the "dd" and "find" commands, that have been there almost since Epoch began, and never did conform).
Over time and UNIX versions, the conventions broke down. One character flags gave way to words, and as soon as that happened, you have to code around whether the arguments like "-Fart" are equivalent to "-F -a -r -t", or whether there is a wordy argument which actually tells the program to break wind (this is a quick example off the top of my head from the "ls" command. There may be better ones).
Then the BSD/GNU people got involved, and introduced what I regard as the abomination of the -- (double minus) argument, supposedly to allow the syntax to be extended, but actually misused my many post-linux developers to completely replace the original UNIX flag processing implemented by getopts.
Couple this with the fact that often the GNU and GNU-like commands were never documented by proper man pages (or even their supposed native "info" command), or even by informative usage strings, and that there was effectively no controlling influence (a flaw in the Open Source contributory model itself) and you get to today's mess. I can totally sympathise with you with it being difficult and awkward to use (I actually count VAX/VMS as being the OS I would prefer to use after UNIX)
I look back fondly to V7 and SVR2 and SVR3 days, when conventions meant something, and people bothered to use them.
My coat is the sad old gabardine raincoat with the Lyons annotated UNIX source in the pocket (which, in case anybody bothers to read to the bottom of my past comments, I lost, but have found again).
I agree completely with your analysis and your sentiments, and even that Ubuntu peaked around Hardy Heron, but what I have difficulty with is whether limiting shareholding to small investors could be achieved, and whether it would work.
I also have a problem in how you would keep the shares with small investors, unless they were bought and sold through a market other than one of the stock markets, or managed completely internally making it a privately held company with an internalised share market (and I don't know how that would be organised or regulated for a large number of investors).
Also, having shares is a method of raising capital to invest. It is not a way of attracting a revenue stream, regardless of how it is implemented.
With shares, there is always the expectation that the shareholder could get their money back. I think that you are eluding to a micro-donation scheme, where you give them money and expect to gain some influence as a result. This is more like a subscription.
I've briefly looked at the charity shares model and I must admit that I don't understand how it works (probably my bad). It looks to me if you effectively lend your money to a charity for them to use and get financial value, and you get the satisfaction of knowing that you are helping and getting value back in the way of ethical satisfaction rather than financial gain. I can't see how that can be a business model for an organisation like Canonical. They still need to get some revenue stream in.
Of course, I may have misunderstood what you are saying (in fact, that is probably quite likely).
@AC re. Maybe it's just diesels
"lazily cruise along singing along to Radio 4"
What do you listen to on Radio 4 that you can sing along to?
The only thing I can think of is Desert Island Disks, and you don't even get the whole tracks on that. Or maybe it is the music rounds of "I'm sorry, I havn't a clue" such as "Pick up Song", which I admit it is difficult not to sing along with, but which only lasts for a minute or two!
I find (especially listening to the Today program interviewing politicians) that I get angry, and end up shouting, impotantly, at the radio!
What alternatives and why?
Maybe MS are trying to make sure that older OS's are retired as well. I'm sure that IE8 could be made to run on XP, but why give users the choice to not pay MS more money by replacing the OS.
Interestingly, I still run a system that I boot on occasion which is Win 2000 Pro. Firing up IE6 on this (IE 7 cannot be installed) already says "Install another browser", with a link to IE8 which then tells me it can't run on that OS.
I know that Win 2000 is out of support now, so I have no cause for complaint (this system has some legacy payroll software that I can't re-install because of expired license key restrictions that allow me to run it in query mode, but not re-install it on another system - and the supplier no longer exists!), which I have to keep running for 3.5 more years to satisfy HMRC data retention requirements. It gets started about once every 6 months just to prove it still works.
@AC re Although....
Although I am in danger of resurrecting a long dead argument, I dispute that the ZX81 had more compute power than the BBC micro.
Although the BEEB's CPU only ran at 2MHz, whereas the ZX81 ran at 3.75MHz, the BEEB's CPU was a 6502 that executed most machine instructions in a single clock tick, whereas the Z80 in the ZX81 averaged four clock ticks per instruction, and some of them required up to 13.
This was the subject of endless controversy between Sinclair/Amstrad owners and BEEB/Apple/VIC20/C64 owners at the time.
In general, the Z80 instruction set was more advanced that the 6502, containing more instructions, more addressing modes, and even some proto 16 bit arithmetic instructions (by treating pairs of 8 bit registers as a 16 bit register). The 6502 contained enough instructions to do what was required though, and was considerably easier to program (the bible of both processors, written by Rodney Zak, "Programming the Z80" was at least twice as fat as his "Programming the 6502", and had smaller print to boot). This made the debate a forerunner of the CICS vs. RISC argument, which hinged around similar concepts.
Remember, though, that BBC BASIC was blindingly fast for the time, and it remained the fastest BASIC available well into the advent of 16 bit micros, as documented by PCW's BASIC benchmark that they ran on all the systems that they reviewed.
And the BBC micro had a built in assembler, and a means of passing arguments between BASIC and your machine code, and also documented all of the OS I/O, sound and graphics calls that you could make from your machine code. And of course, the display in the BBC was totally hardware driven, freeing the CPU up to run your programs.
I used to write both Z80 code on a ZX81 and Spectrum, and 6502 on a BBC, and believe me, 6502 was easier, and for basic data manipulation, faster.
Actually, they were not thermal at all. The paper was covered with aluminium, which conducted electricity, and was 'written' by a wire that passed a current through the paper as it wizzed round on a rubber belt. Where the 'spark' hit the paper, the aluminium vaporized, letting the black paper below show through. Crude, noisy, and completely incompatible with listening to the radio. That is why the 'paper' was silver with black print, and also why you got a new high power supply for the ZX81 if you bought the printer.
I believe that they were not allowed to sell the printer in the US, because (surprise, surprise) it contravened the US electrical interference regulations.
This was an example of innovative thinking that made Britain good at creating ideas, but pretty crap at exploiting them!
@ForthIsNotDead re. sound.
Although out of the box, the ZX81 did not have sound, there were a number of third party add-ons that gave you sound.
I had a Quicksilver board that had an AY8910 on it, feeding a secondary modulator to add sound to the TV signal.
Quicksilver also had a number of other accessories including a high resolution graphics board and a programmable character board. You needed an interface board that sat between the ZX81 and the RAM pack, which provided two interface slots for the add-ons. Ugly as sin, and made the RAM pack wobble problem different, but as I added an external keyboard to mine, complete with power switch and reset button, I did not have to touch my ZX81 at all.
As I've said before on these forums, my ZX81 actually had 18K of memory, the 16K RAM Pack, the 1K of internal static memory re-mapped to a different address when the RAM Pack was plugged in, and another 1K of static RAM on the ULA side of the data bus isolating resistors to hold programmable characters that were accessible by changing the contents of the Z80's I register, which was used to hold the base address of the character generation table, normally in the ROM. Happy days!
It was not called FPGA, although the technology was similar. The real name was a ULA, or Uncommitted Logic Array, which was a bleading edge technology in 1980/81, invented by Ferranti, a British company.
The ZX81, Spectrum and BBC Micro all had ULA's in them, to consolidate the function of dozens of 7400 TTL chips into a single large chip.
Unfortunately, the technology was still immature, and the production problems that Ferranti had were a large part of the shipping problems for all of the systems mentioned. I waited for nearly 6 months to get my BBC model B that had been ordered as soon as Acorn would take orders.
By the time later systems came along, it was possible to have your own design of chip fabricated moderately cheaply, so the ULA died an ignominious death
@stucs201 re. MK14
Are you sure it was a MK14? Many universities used a KIM-1 (produced by Commodore, at least at the end), which was a similar product but used a MOSTEK 6502 rather than the Nat. Semi SC/MP which is what I believe was in the MK14.
I remember having to write a sine-wave generator on a KIM-1, and I managed to get a higher resolution wave than everyone else by having a lookup table with just 1/4 of the whole cycle in the lookup table, whereas everybody else had at least a half cycle (and some of them stupidly coded the whole cycle). Fitting it into 512 BYTES was a real challenge. But then I also managed to write a simple lunar lander on the Sinclair Cambridge programmable calculator in just 32 program steps!
Nuclear fission sustainable?
I'm playing devils advocate here, because I'm pro-nuclear, but I would suggest that the amount of fissionable material on Earth is limited just like any other earthbound energy resource.
This means that in the long term, you cannot use 'sustainable' in connection to nuclear.
The way I look at it is that the Earth has many energy resources, but because of entropy, they are all limited to one extent or another. Fossil fuels are stored energy from the Sun, nuclear fission is heavy elements (from the death of older stars) acquired during the formation of the Solar system), nuclear fusion is light elements probably from the formation of the Sun, wind and ocean currents are driven mainly by solar energy from the Sun, tide is gravitational (mainly from the Moon, and tidal drag is causing the Moon to slow down and approach the Earth so even this is finite), geothermal is (probably) natural nuclear fission (see above) combined with tidal effects from the moon, biofuels are capturing energy from the Sun and direct solar is (obviously) from the Sun.
So if you discount total matter conversion (and boy would that be useful), and fusion of hydrogen electrolysed from water (finite on Earth but lots of it), all energy except direct and indirect solar energy is limited. And the Sun won't last forever!
- +Comment Anti-Facebook Ello: Here's why we're still in beta. SPAMGASM!
- Analysis Windows 10: One for the suits, right Microsoft? Or so one THOUGHT
- Vid+Pics Microsoft WINDOWS 10: Seven ATE Nine. Or Eight did really
- Xbox hackers snared US ARMY APACHE GUNSHIP ware - Feds
- George Clooney, WikiLeaks' lawyer wife hand out burner phones to wedding guests