1732 posts • joined 15 Jun 2007
Re: Mere tissue of a story
Looking at the staff writers remaining, then I think that your last statement is probably true.
Re: Big fish and small fry - yum!
One wonders whether this includes any single person contracting companies that many IT contractors work through.
If it does, then the figure is mightily misleading, because it will not indicate any change in anything other than how it is being counted.
...whether the PoS tills run an embedded version of Windows, or one of the full-fat versions?
Re: Linux support... well, who can say? @Jason
Ah. I'd forgotten the difference between SAS and SATA. I work so much with SAS that the restrictions in SATA compatibility flew past me while I wasn't thinking.
Linux support... well, who can say?
I'm making a (possibly erroneous) assumption that this thing is put together using industry standards, which may be wrong, but...
If this is two disks, with a 2 port SAS expander built in, then plugging it into a laptop will show 2 drives under Linux. They should just work.
What won't work is any fancy Acronis drive imaging software. But, boot from a live CD, attach your old drive via USB and then use gparted to partition and copy the data around. Only problem you may have is writing the boot record, but that should be relatively easy from Grub.
Anybody fancy giving me one to test this assumption?
And in the cloud, the storage is?
Well, I guess spinning rust and tape.
So not so clear cut at all.
I'm still dubious of the longevity of data stored on flash RAM, especially if the flash is stored 'cold', i.e. without power. Until this is proved, I would be dubious about using it for information that legally has to be kept for years, which is the traditional domain of archive and long-term backup.
And that's not to mention the security implications of having the data ultimately stored out of your control ('binding' contracts are only as good as the people who wrote them, and nothing like having physical security surrounding your data). If a cloud provider goes bust, or is taken over by another company whose modus operandi is not acceptable to you, how do you extract and export the terabytes of information they've been holding for you to move somewhere else, and ensure that they've destroyed all copies of the data.
The kit is not 30 years old. The design is.
OK the design probably needs to be updated, but the way this is written suggests that the exchange is still running on kit bought when the IBM PC/AT was the benchmark PC!
Re: Something not yet considered
Actually, it's mainly a matter of momentum. They are still providing the banking services for me mostly successfully, and it looks as if I was not too badly affected by the problems (I managed to get money out of an ATM at just past 19:00, in the middle of the supposed problems). I was not affected at all by the previous problems, but that is probably because I get paid at a different time from most people (it was credits into the accounts that were affected).
Also, it's easy moving an account, it's less easy moving a (reasonably priced) overdraft facility!
Something not yet considered
One point that has not been considered is that RBS are in an on-going transformation. As an RBS customer in England, I was given an outline of the Santander abortive sale, and recently information about the Williams and Glynn split (RBS are splitting most of their Scottish Nat-West branches, and most of their English RBS branches out into a separate bank with the revived name of Williams and Glynn).
I've already been moved onto a different URL for on-line banking and banking-as-an-app, and we've had our debit cards re-issued twice in the last 18 months with different numbers, presumably with a different bank code hashed in the long number.
I suspect that much of the $2bn+ spend promised next year is already earmarked for that split to happen
It is possible that it was some of the on-going work for this which caused the problems, although Cyber Monday would be a really bad day to plan it. More probably, someone just screwed up.
Re: How many are waiting for Windows 8 to be "retired"? @Steve Knox
Oops. For the memory in my first PC, read 16MB, not 16GB!
Wow. How much would that have cost in 1996!
10 minutes is obviously not long enough to spot such errors.
Re: Lets try to look at the facts @Denarius
I suggest that you take your Android tablet, attach a OTG USB cable to a small USB hub, and plug a proper keyboard and mouse in.
Android works fine like this, without the need to touch the screen at all. I really find that for desk related activities this improves the usability of the tablet.
The only thing it is missing for general use is applications-in-windows (although I know lots and lots of PC users who juat maximise the current application they are using and don't really need windows) and I can see the definite abillity to use a tablet-like device with some cloud services displacing PCs on desks.
If you have a modern Android phone, try adding proper display as well, and see whether you could conceive of using your phone as your only computing device!
Re: How many are waiting for Windows 8 to be "retired"? @Steve Knox
Steve, I respectfully disagree.
If you had said 2001 or 2002, the introduction of XP, I would tend to agree.
When the PC was introduced in 1981/1982, they were very clearly business only devices. Even in the US they were too expensive for a home purchase unless driven by a specific need, and this was even more the case outside of the US where a basic PC at introduction equated to about half of my yearly salary as it was at the time.
I would say that the whole area of media consumption is completely new from that time. You could not eve use a PC (and here I am talking about an IBM compatible PC) to play music off any domestic media available at the time, and that is the easiest media (unless you count books, which wern't distributed to be read on PCs).
The home market was better served in the US by Apple, Commodore and Atari kit, and by the plethora of UK manufacturers including Acorn, Sinclair etc in the UK, and Japanese companies for the rest of the world in the early '80s, and I would say that you cannot claim that a modern PC has anything more than a passing relationship to any of these devices.
It took until the mid '90s for the PC to become an attractive home purchase. I bought my first IBM compatible at that time (I was a committed BBC Micro user), for about £1000, and it came with a 100MHz Pentium, 16GB of memory, a 1.2GB hard disk, a CD Reader, an ATI Mach64 display adapter with 14" monitor providing 256 colour 800x600 resolution SVGA display, and a sound-blaster compatible sound card, running Windows 95 (and Linux - although this was a real effort getting it to work). This could be counted as a 'multimedia PC', and whilst you could listen to a CD, you would not (and probably could not) watch a film on it with any degree of enjoyment.
And to bring it back to the point, the majority of modern PCs in the home are used as media consumption and social media devices, which a tablet, chromebook or convertable will do as well or better than a PC. PCs (especially desktop PCs) will become niche devices for people who have a need for more storage or processing power than a low-power device can provide.
In business, PCs are mostly an alternative to form-filling, paperwork and performing data-lookups, and I suspect we will see PCs being displaced by thin clients based around the same technologies as a tablet-with-a-keyboard or small laptop (yes, really this time) because of the efficiencies in the administration and cost savings of large numbers of such devices.
A thin client deployed on top of Android on an Arm device, built into the screen (effectively a tablet), with a keyboard and pointing device, and limited amounts of local storage, connected to a server estate using a remote desktop technology or cloud service will be a very desireable device for many businesses, and probably cheaper than an equivalent environment built around desktop PCs. Deploy a secured WiFi, and you don't even need to cable the desks for data!
I can see devices like this retailing for around £100 per desk in volume, plus some extra for the backend services, in the near future. All we need are the applications, and I'm pretty sure that OpenStack or Azure, deployed either as a local or remote cloud will get sufficuent traction for serious software to appear and displace PC software.
It's all a bit bleak for the PC market, quite honestly.
Bloody hell. I've just argued myself into thinking seriously about SaaS cloud services! Maybe it's not all BS!
Re: How many are waiting for Windows 8 to be "retired"? @Steve Knox
I think that modern PCs are used for more than those from 1981.
Remember back then. No Internet (including facebook, twitter et. al.) no common graphics standard (and nothing capable of displaying a photograph), no digital cameras, connectivity to the outside world by modem and BBS if you were really advanced, no email, no audio hardware, no writable optical drives to rip/write CDs, no DVDs at all, no video, no mice. Monochrome dot matrix printers if you could afford $/£1000 otherwise no hardcopy.
If you are really talking 1981, then no hard disks. Everything from 5.25" 360KB floppy disks!
No porn! (at least none that was interesting - remembering animated ASCII art of a flasher that I remember doing the rounds by sneakernet).
And, and and. A PC would set you back $/£2500 (I'm actually looking at an IBM PC advert from 1982 in an old magazine at the moment), which was more than most people paid for their cars!
OK, you still had people writing letters and doing basic spreadsheet and simple database, and you did have text and basic block graphics games but that is about all that is common between then and now.
But now, there is very little that you can think of that a PC made in the last 5 years or so cannot do. Even new technologies such as 3D printing is well within the capabillities (with a suitable printer of course) of a modern PC.
What is happening is that the common sub-set of technologies that Jane and Joe Public need can now be done from a device that is more like a tablet than a computer. Only nerds like you and I use some of the other stuff.
I mean, really. A Chromebook or a tablet or even a modern phone with the option of an external keyboard and connectivity to a telly will do all of the media consumption and social media that your average non-technical user will ever need.
I can see nothing that will prevent the decline of the PC into the technical niche that it emerged from. It'll never disappear completely, but will become to a tablet as a tape recorder is to an MP3 player.
Re: OSs? @AC
If you want to be taken seriously, post this as something other than an AC.
BTW. My everyday system (systems, really, because I move/clone the disk from laptop to laptop as I change machines, upgrading the distro whenever appropriate - but still with the same home directory and machine identity) has been Linux only for about 8 years. It works very well, actually, and I make a habit of not fiddling with it under the covers, because I don't want to break it. Almost all of the installation, admin and maintenance is done using the GUI tools provided.
The problem here is that MS will probably stop updating Security Essentials immediately, and then pressure the other anti-virus companies to stop providing updates for XP (they have various contractual tricks they can use to force software suppliers who use MS application development tools and libraries to stop providing updates).
You will also find that MS will be indirectly pushing for companies like banks, who need decent security, to change their web-sites to stop accepting connections from IE8 or earlier, for "the user's own security and on-line safety". Of course you can run Firefox or Chrome, but Google have a history of stopping providing updates for Chrome on an OS once the OS goes unsupported.
Once you have an unsupported OS with no up-to-date AV, and a limited choice of browsers, I suspect that you may think again about whether the box may be more useful with Linux on it.
When it comes to XP on older kit
You absolutely have to keep XP SP3 off of it.
I'm convinced that MS added an extra "feature" to SP3, which was "make it run so slow that the user wants to dump it and buy a new computer".
My dad had a Thinkpad T43 (Pentium M Mobile, 1.83GHz, 1GB memory) running XP SP2 with auto-update turned off, quite well. Some MS social engineering trick (click here to fix this) got him to turn on auto-update, and now it is barely usable.
I know, I know. There's other security fixes in SP3 as well, but I've repeated the exercise several times on other machines running XP, and it really is the case that SP3 increases the OS footprint and loads heavyweight services that make an older machine really sluggish.
You should see how fast even a quite modest machine (by today's standards) runs with a fresh install from a 2002 XP install disk! (my first XP machine had an AMD Athlon XP1700 - 1.5GHz, 128MB memory, and was pretty fast at the time). Just don't connect it directly to the Internet, or browse any dodgy web sites!
Re: The overwhelming message I get from these ads
Microsoft also had an advertising campaign in Japan based around the recent Ghost in the Shell: Arise anime.
It showed various members of Section 9 doing what they do holding and passing around a Surface, which contained some important data or something.
The problem with this, as anybody who is familiar with GITS knows, is that the people shown (I remember at least Motoko and Batau in the ads) are cyberised, i.e. they are cyborgs with cyber-brains (implanted computers) together with some kick-ass comms. They had absolutely no need for a Microsoft surface to do the things that they were supposed to be using it for!
Just showed that either MS or their advertising agency really did not know what they were doing. I suspect that the animators probably felt a bit dirty to have done the ads, but only until the money hit the bank!
I think that the ads are still knocking around on YouTube if you want to see them yourself.
Re: Expansion @John Tserkezis
the Dark Lord was commenting about moving new kit into the machine room (normally this involves rolling it across the floor, which on a suspended floor would cause significant vibration, certainly more than having the disks powered up and the head moving.
I would actually have thought that the main reason why the disks were powered down was because of power consumption and temperature, rather than vibration. Disks are not that fragile.
480 drives in a rack is not that dense. 384 disks in a 30" rack mounted 4U enclosure is a much higher density (I have a rack with 5 of these disk enclosures in each of the HPC's I look after, totalling 1920 disks in 20U of space - about half a rack), and all of these are spinning all the time.
@Destroy All Monsters
My knowledge of telco machine rooms may be rather dated, but it used to be that almost all telephone exchanges put the kit directly on a concrete floor because of the weight (a practice evolved from having vast and very heavy mechanical exchanges). With a solid floor for load bearing, it made sense to take the cables up to the ceiling. Old habits die hard, and many modern exchanges were installed in old buildings.
It may be that modern electronic exchanges more closely resemble computer machine rooms, but in this case, you can see from the picture that it is a solid floor, with the cabling to the ceiling.
Re: Expansion @The Dark Lord
These are telco style machine rooms, no suspended floor and wiring from above.
The floors are solid sealed concrete, so probably don't vibrate too much.
And there you have PERCS. If you look at an IBM 9125-F2C (Power7 775 cluster), they are very dense, are water cooled (CPU, I/O Hub, memory and power components) with integrated electro/optical network interconnects eliminating external switches, and storage moved into very dense arrays of disks in separate racks.
When where I work moved from Power 6 575 clusters (which were themselves quite dense), they kept to approximately the same power budget, increased the compute power by between three and five times, doubled the disk capacity, all in about one third of the floor footprint of the older systems. And to cap it all, they actually cool the ambient temperature of the machine room.
But these systems proved to be too expensive for most customers, and IBM was a bit ambitious about the delivery timetable. Take this with a contraction in the finances of many companies, and IBM failed to sell enough of them to keep them in volume production. But they are very impressive pieces of hardware.
Replacing them with a 'next' generation of machines is going to be hard.
Re: Tellies can handle 60Hz input @AC
I understand what you are saying, you've not understood what I've said. But anyway, using linear light is introducing an additional motion blur component, as you've effectively got to interpolate intermediate frames that do not exist at the re-sample point, and they will always be in one way or another a guess. Also, doing it in near real-time may require more compute power for HD video than is in the Xbox.
What I said would still work, although as I also said, it is impractical.
Re: Tellies can handle 60Hz input @Mage
I'm assuming that it was you who down-voted me.
I was not suggesting frame conversion. I was suggesting that you used a frame rate between the Xbox and the TV that allowed exact timing of both frame rates to prevent the need to re-sample. This is why I chose 300 fps, as that is an integer multiple of both 50 and 60. This allows an EXACT number of frames for each of the different video sources. For the 50 fps source, you would leave the image up for 6x300HZ frames, and for the 60 fps source, you would leave an image up for 5x300HZ frames. A perfect fit, with no resampling, allowing both videos to be side-by-side at their native frame rates.
Of course it's completely impractical as well, and would only work for these two frame rates (or other divisors thereof).
If you assume both videos are interlaced, you could probably take that down to 150 fps, but that is a big assumption.
Re: Tellies can handle 60Hz input @Brangdon
Well, that's it. The Xbox is trying to harmonise the frame rates for two different video sources. It's not really a surprise that one or the other will be affected.
In order to be able to simultaneously display a 50 fps and a 60 fps picture perfectly, you would need to output from the Xbox to the TV at 300 fps (so the 50 fps image would appear on 6 consecutive frames, and the 60 fps image would appear on 5 consecutive frames).
This would be beyond most TVs, even modern ones.
Apart from the obvious power cable (and check the voltage ratings on the label on the back of the telly as well, although most European countries use between 220 and 250V), you have the problem of the DVB-T format, although most European tellies do DVB-T2, which is backward compatible with DVB-T.
You may have to tell it to scan different frequencies, and sometimes this is in a hidden menu. It depends where you are coming from.
If you are just using external video sources (DVD, consoles, set-top boxes etc), things should just work.
Don't understand this!
Modern flat panel televisions just do not have the old mains frequency lock or problems with the 'flyback' frequency that old CRTs have.
I very much doubt that there is any difference in the hardware for a Korean or Chinese television destined for the UK or for the US.
Tellies have a frame buffer (or two). The frame buffer is painted, and the picture is displayed. This can be asynchronous from any other timing signal external to the TV. As long as the hardware can keep up with fastest frame rate, it should be able to sync with any slower rate without any difficulty.
However, if the XBox is re-sampling the frame rate of an external video source as it passes through, then this could conceivably cause missed frames (50Hz->60Hz means some frames will be sampled twice). Anybody who has played around with frame rate when transcoding video will have experienced this, although I suspect that most people who believe they have done this probably used 'canned' settings rather than really experimenting.
So I suspect that the XBox is re-sampling at 60Hz, or possibly screwing around with the de-interlace settings (Sky broadcast HD at 1080i), rather than it being a problem between the XBox and the TV.
Re: Hmmm. Extract from the lawsuit.
It was the scale of the claim. "billions of images" and "near instantaneously" that I was mocking.
I'm sure that there are tools which will look at images and spot similarities, but I'm also sure that they're not instant. Lets assume the images are 100KB each, and there are "a billion" of them. That's 1x1014 bytes (hey, lookie what a silver badge allows me to do!), or approximately 100TB of image data. If they can read that and process it "near instantaneously" then they have a better system than the top 100 HPC system that I'm looking after at the moment.
Hmmm. Extract from the lawsuit.
Claim 29: ...which can scan billions of images nearly instantaneously......
Gosh. I really could do with some of these systems that Match.com must have. Near infinite disk bandwidth, and very sophisticated image hashing and analysis tools.
With that technology, I wonder why they're in the dating business. They ought to be coining it in from the application of this technology.
Re: New universities
I totally agree about your comments about 'New' Universities/Polytechnics.I think that giving them the option of becoming Universities was the worst thing that could possibly have happened to the Polytechnic sector.
I agree that most Poly's had a big chip on their collective shoulders, but I worked in Newcastle Polytechnic for 6 years, and I met people there who knew what the Poly's were for, and understood how to represent them. But I remember at the time how surprised the ministers were that all the Polytechnics decided to convert when given the chance.
Older established Universities are academic. They turn out people with a largely theoretical slant on most science and technology subjects. Poly's were set up to be practical skills based. They could take students and equip them to take on high-skill practical work. You could see them as a alternative to business led apprenticeships, leading to BTEC HNC and HND qualifications. Both of these were valuable but different facets of the education system in the UK.
Generally, academically orientated students with the highest 'A' level results (in the days when 'A' levels could be used to differentiate between students) gravitated to Universities, those with adequate results could go to Poly, and still get highly useful qualifications, just not necessarily degrees.
But there was also a difference in teaching methodologies.
'Old' Universities were more likely to drop the student in at the deep end with comparatively little support, and if they sank, threw them out. Those who swam (who were self-motivated and with sufficient discipline to actually get the work handed in and pass the exams despite the distractions), when they graduated, an employer knew that they could resist the temptations of student life, and still get the job done.
Polytechnics, on the other hand, used to offer better support to the students. The staff-to-student ratio was higher, and there was more emphasis on making sure that the students were coping (at least this is was what I saw at Newcastle). This meant that Poly's were a better bet for kids who were still in the 'school' mindset.
In the Computer Studies area, Newcastle Poly. offered HMC and HND courses in Computer Studies, but not a degree, which was catered for by Newcastle University. The one computing degree course offered by the Poly was a business orientated degree, specialising in COBOL as the programming language (we're talking 1980's here), with business oriented methodologies, system analysis and case studies, together with crossover courses from Business Studies so that the students would have an understanding of Data Processing and where it fitted in to a business.
The HNC and HND CS courses turned out people who's skills meant they knew enough about computer systems so they could program effectively, but had a less deep understanding of the fundamentals of a computer than their University contemporaries.
With the generally useless 2-year 'foundation' degrees replacing many of the BTEC qualifications, I really don't know what the split is now, and I think that employers have similar lack of understanding.
60 disks in 4U!
I admit they are special racks (they are nearer 30" wide and goodness knows how deep), but in the IBM P7 775 supercomputer disk enclosures, you can get 384 2.5" disks in 4Us of vertical space.
On more mainstream systems, and having used dual-connected SAS drives for about the last 5 years, I will say that the biggest problem here is the repair of a failed expander card in the disk drawer. The problem is that although they are redundant, so the loss of a SAS expander does not stop the service, the repair action is not normally concurrent. This means that you have to take an outage in order to restore the full resilience, even if you have the connected to dual servers unless you have the data moved or mirrored to disk in another unit. The saving grace is that you can plan the outage, but you have to be careful if you are wanting very high availability.
I learned this the hard way when planning for service work in what had been delivered as a totally redundant system. A bit embarrassing when you end up having to stop all of the workload on a top 500 HPC system just to carry out the work for a single expander card (no, I was not responsible for the design, I only help run it, and it could have been mitigated with a bit more thought)
By the way, this dual connectivity is not a new thing. IBM's SSA disk subsystem also had dual connectivity for both disks and servers back in the mid 1990's. Very popular for HA/CMP configurations, and allowed for 48 disks in 4U of space.
Ingres and 2BSD
Ingres was available for free (or at media and postage costs) to Universities and Colleges who had a UNIX source code license. I believe that it was on any 2 BSD add-on tape (it was certainly on the 2.6 BSD tape I had in 1982).
The University I was at (Durham, UK) was using Ingres to teach relational database in 1978, and I came across it in my second year in 1979.
I must admit that I could not stand it as a subject, because the lecturer was using set theory to try to teach relational algebra, and my maths was beginning to look a little shaky by that time, but when I ended up actually doing real work, I found QUEL quite usable. It took quite an effort to switch to SQL when I had to work on Digital's Rdb, Oracle and DB/2.
I don't count databse as a current skill now, but I still regard the experience I gained as invaluable.
Re: Symantec writeup very poor @Gorbachov
And that is the point of my OP. The writeup is so vague that we're all guessing.
I admit that the client side attack I sketched out requires access as a user on the client system, but that is a lot easier to get than breaking privileged access. All the usual vectors of Java, side-jacking and social manipulation etc could end up with a process owned by the user in question, which would have whatever access the user has on the client system (but no special privilege). This would mean that it could execute a series of shell commands as the user, run an SSH client program itself, read the user's public key and any private keys stored on the client system, and if the keys are passphrase-less keys, use them to gain access to other systems.
Here is a scenario, possibly far-fetched, and I've not worked it all through, but it could set LD_LIBRARY_PATH somewhere like .bashrc so that a local directory appears before some of the system library directories. It then looks at the Linux distro, and fetches a specific hacked SSL or other (including libc, I suppose) library for that distro off of the internet, and puts it in the directory.
Following this, every legitimate program including SSH client sessions that the user starts could be running with malicious code from the bogus library. If it replaced the right routines, you could have a key-logger, and this key-logger will be able to capture the passphrases as they are typed, giving access to all of the user's private SSH keys. It could also capture any passwords that are typed for remote systems.
OK. No breach of privilege required so far. Everything has been done as the user in question.
So, the user is an admin, who foolishly has the private key of a remote account that has some privilege in their keystore. The malicious code then has access to the remote system with privileged to attack that system.
Or, say, the admin has sensibly used a non-privileged account to access the far system, but then uses sudo to issue commands on that system via a compromised SSH session. Compromised client can then capture the password that the user uses with sudo, and again has access to the remote system with privileged to attack that system (unless sudo is really locked down hard).
In both cases, it could inject commands, or even start it's own SSH client session using the captured credentials.
How safe do you feel?
Please note that this attack could be used on almost any OS that allows dynamic binding of libraries at runtime, and provides an over-ride of the default system paths to the libraries. I've sketched it out as a Linux/UNIX attack as that is what I know best, but I seriously suspect that similar attacks are possible on other OSs.
Eternal vigilance is called for, especially for admins, regardless of the platform they are using.
Re: @AC 14:58
I'm not doing the GNU/Linux 'drivel' as you call it. I'm just pointing out that SSH is as much a part of Linux as Audacity or LibreOffice, or a host of other Open Source projects. They're part of most distros, sure, but not a part of Linux itself. I suggest that you just don't understand what a Linux distribution is.
As an analogy, would you claim that Apache or VMWare player or even Skype is part of Windows if a particular machine vendor chooses to pre-install it on the systems that they sell?
It's not even the case that OpenSSH is the only SSH implementation out there. F-Secure have their own completely separate SSH implementation, as have SSH Communications Security, and there are also other free SSH implementations like LSH and Putty (client).
@AC re slipstream SSH datastream
Yes it would be, especially if it could be done from outside the SSH client/server communication stream. But this does not appear to be what has happened. This is hijacking one end or the other, and intercepting/injecting the data at one end of the secure pipe as it were.
Just to point out that SSH is *NOT* part of Linux. It's not in the kernel, nor part of the GNU toolchain, and although it is in the repositories of most distributions, it's also available for most UNIX systems, and also for Windows and probably any other network enabled operating system as well. It's a cross-platform tool. What is important is how and by what vector it was compromised.
So there is a vector (possibly OS specific) that was used to break into SSH, and SSH itself is a vector to compromise whatever OS is being used. Which may be Linux.
Symantec writeup very poor
I know it's difficult to publish information about a vulnerability without providing a means of using it, but the Symantec write-up is pants! I mean, what does "Rather, the backdoor code was injected into the SSH process " actually mean?
Was it added to the binary before it was run, was it added to one of the run-time libraries, was one of the in-core runtime libraries hacked, or was the running instance of the process altered?
It also does not state whether this is a ssh server attack or an attack via the ssh client.
I can think of several ways of compromising the client side of things (each ssh session has it's own instance of the ssh client process), and these can be attacked using well known PATH and LD_LIBRARY_PATH attacks without needing privileged access to the client system, or the on-disk binary or the libraries can be attacked and altered if you have access to a privileged account.
Once into the client process, you will have access to all of the private key information on the current system (although you may already have access to that anyway), but I can see how you could catch and re-use key and password information as it passes through the compromised client process. You would also be able to subvert any and all stream traffic, including fixed passwords, SSH passphrases, sudo passwords etc. for any session that is run through the SSH client (using the client as a keylogger). About the only thing that you would not be able to do would be to compromise one-shot authentication devices.
Injecting arbitrary commands would be a minor trick, although hiding them is more difficult.
And if the SSH key management is lax (same key used for multiple servers and user identities, especially if some of them are privileged), then you have a recipe for system compromise on a massive scale.
But don't blame this on the Linux security model. Any system with some form of trusted remote execution could be compromised in a similar way.
Re: Shills @Bill
You need not be a MS shill, just part of a system where one supplier can control a market, compelling ordinary people like yourself to defend the indefensible. Microsoft want you to not have an alternative.
There is no reason why Linux cannot become as good or a better gaming platform than Windows. It's only market penetration that make gaming companies develop on Windows. It's possible that the Steam effort or Crossover may just change things.
Re: Shills @Ken
But that's the point. Nobody can become famous posting as an AC. They just merge into the crowd.
I'm not saying that the Reg should remove the ability to post AC, hell, I use it myself when commenting on something that may upset someone in my acquaintance. It's just that I'm so pissed off trying to work out who is who when they are making such cowardly accusations.
Why is so much of the muck-slinging, accusing everybody of being shills, being done by AC's.
Really, folks. If you want your comments to be taken seriously, at least post them with an identifiable handle, even it it is not your real name!
In case you forget, it's not possible to differentiate one AC except by content.
In it's true meaning...
...I'm sure that if you really do, you don't need a tool to grok the information!
Never impressed by the 380Z
I always regarded the 380Z as a bit of a lemon, mainly because I did not see one until 1982, after I had my own BBC Model B. I guess that if I had seen it earlier, I might have had a different opinion, although I'm not sure, having first used UNIX in 1978.
It always struck me as slow (especially with the high resolution graphics board), but I did appreciate that it ran CP/M, and thus had a large library of software, provided that you could get it on the slightly unusual disk format (not that there was a standard disk format at the time).
The one I had control of used to be used mainly by one member of staff who wanted to use Wordstar and the QUME Sprint 5 daisy-wheel printer. There was one postgrad who had a strange project to try to connect it up to the Newcastle Connection (aka UNIX United!) as a client machine over RS232 - there being no Cambridge Ring hardware for the 380Z (daft really, as the filesystem API was too different between CP/M and UNIX). He never completed the project, because it turned out that he was a draft dodger from his home in Greece, and he went home to see his family, and was promptly arrested as he stepped off the plane! It did mean that I got to see the UNIX United! source code, as I had to add it to 'my' V7 UNIX PDP11.
I used to be all for DAB when it really was new. Over the years, I've bought two mains powered DAB radios, a car DAB radio which re-transmitted on FM, now no longer made, a pocket DAB radio, and an add-on for an iPod.
Slowly, all of the interesting stations I used to listen to have dropped off DAB, or gone low-bit rate/mono (really - Planet Rock in MONO!).
And to cap it all, there are vast swaths of no DAB reception where I live.
I still keep the DAB radio in the car, but only for Radio 4 Extra. None of the others even get turned on any more. Instead, I normally listen to Radio 4 on FM or occasionally Radio 2 for some of the ex-Radio 1 DJs, and sometimes ClassicFM or Radio 3 when I'm in a classical mood. Other than that, it's music and podcasts stored on my phone.
It's a technology that has failed, and should either be turned off or re-launched in a form common with the rest of the world.
Re: DAB is pointless @Ben
A two word retort to your Internet access in cars comment - usage caps.
Re: 6502/6809's rool btw... @ Jamie
The problem with may of the complex instructions on the Z80 was that they took so many T-states to execute. This meant that on paper, a 4MHz Z80 looked like it should outperform a 2MHz 6502, but as the average Z80 instruction took 3.5 T-states, a 6502 clocked at half the speed, with an average of 1.5 T-states per instruction could run more instructions in the same time.
This meant that with careful programming, it was often possible to get functionally identical code running faster on the 6502 than on a Z80. It was horses for courses, of course, but many of the sorts of things that these processors would be running would be integer, simple data handling or block memory problems that did not need the more powerful instruction set of the Z80 anyway. I've commented on this with a worked example before here
But this comes back to the crux of the article. In order to get the best out of the machines back then, it was necessary to know the instruction set very well. And this is what is missing in today's programmers.
Re: 6502/6809's rool btw... @Steve
What made Page 0 really special on the 6502 was the ability to treat any pair of bytes as a vector, and jump using a 2 byte instruction (one for the op code, the other for the address in page 0) to anywhere in the systems address space very quickly. Because this was used extensively in the BBC Micro OS for almost all OS calls (see the Advanced BBC Micro User Guide), it mean that you could intercept the OS call and do something else instead (it was called re-vectoring).
I used this many times. For example, in Econet 1.2, all file I/O (but not loading programs) across the network was done a byte at a time (very slow, and crippled the network, which only ran at around 200Kb/s anyway). I wrote a piece of intercept code which would re-vector OSREAD and OSWRITE so that they would buffer the file a page (256 bytes) at a time (IIRC I hijacked the cassette file system and serial buffers to hold the code and the buffered page), which sped things up hugely. Could only do one file at a time, but would handle random access files correctly.
When used with the Acorn ISO Pascal ROMs, it sped up compiling a program from disk from a couple of minutes to seconds, and meant that it was possible for a whole class to be working in our 16 seat BBC Micro lab at the same time.
Talking about ISO Pascal, which came on 2 ROMs, I also re-vectored the Switch ROM vector (can't remember it's name) so that I could load the editor and runtime ROM into sideways RAM, edit the Pascal program, issue a compile command (which would switch to the compiler ROM), and have it overwrite the editor/runtme ROM image with the compiler ROM image, compile the code, and then switch back after the compile was finished. Great fun! Infringing on Copyright, of course, but meant that I could work in Pascal on my BEEB that did not have the ROMs installed!
Re: peek? @Simon
OK, I accept that the Atom (and probably System 1 and System 2) had them first,
BTW. My BBC micro was mine. I paid for it, not my parents. I ordered it on the day that they opened the orders process to the public, and it's got an issue 3 (an early) board, has a serial number in the 7000's, came with OS 0.9 in EPROM, and last time I powered it on 18 months ago, still worked.
I had an advantage that I knew C, PL/1 and APL before I got my BEEB.
Re: I'm kinda conflicted...
6502 was an elegant and orthogonal machine code, spoiled by the gaps in the instruction set for instructions that didn't work in the original MosTEK silicon.
By the time the 6510 came along (as well as some of the later 6502B and C chips) many of these missing instructions would work, but nobody used them because of backward compatibility.
6809 was probably a more capable and complete machine code and architecture (it benefited from being a later chip), but I still have a fondness for 6502 (and PDP11).
Oh no. Peek and Poke.
I prefer ? and !
<smug>Guess what machine I had</smug>
Re: Really? @dogged
If the screen is damaged, then there is a good chance that the front glass/touchscreen will also be damaged anyway.
And the rest is really just a plastic moulding, so won't add significantly to the cost.
When I've replaced the screen on a couple of phones, I've always decided to replace the glass as a matter of course. If you're going to take the effort to dismantle a phone, replacing the class seems like a minor extra expense.
Re: VMware Snapshots? For real?
I do not know VMware Snapshots, but I'm assuming that they work like other snapshot systems.
Blockwise filesystem snapshots can have a place in regular backups, but only really if you limit the time you want to go back to the number of snapshots you keep. And this is determined by the amount of change in your systems and the amount of storage (usually disk) that you are prepared to keep back for the snapshots. In addition, they are probably useless for disaster recovery, unless you are maintaining cross-site snapshots (I don't actually know if you can do this, but I would guess that if you had cross-site mirroring, it would also be possible to keep snapshots on your other site).
If your backup requirements are longer term, or require recovery of individual files, then an agent based backup scheme is about the only way you can satisfy the requirements, IMO. This is especially true if you have a heterogeneous environment.
Of course if you are backing up the C: drive of all of your identical virtualised Windows boxes, then there are probably huge benefits in just backing up one copy of a de-duplicated, shared image at the de-dup'd level, rather than agent based backups of each system. But that is a particular system deployment method that does not match all requirements.
It might be to weaken the case that Google is selling in the UK. The 'sales' staff roll up in a barge moored in the Thames, host all their junkets, sell all their advertising, and then sail away,
HMRC and the Parliamentary Select Committee will not be able to express incredulity at Google reporting so little UK business.
- Opportunity selfie: Martian winds have given the spunky ol' rover a spring cleaning
- Spanish village called 'Kill the Jews' mulls rebranding exercise
- NASA finds first Earth-sized planet in a habitable zone around star
- New Facebook phone app allows you to stalk your mates
- Battle of the Linux clouds! Linode DOUBLES RAM to take on Digital Ocean