systemd is pants!

OK, let’s get this out in the open… systemd is *PANTS*

Phew! That’s better.

I’ve just spent most of the day trying to get a Linux system to reliably mount a disk attached via iSCSI over ethernet at boot time and, more importantly, to reliably get it to be unmounted at shutdown before the network rug is pulled from beneath it.

Now, in the days of the init scripts it was pretty easy to stuff a script in-between the networking coming up and nfs-server starting. It was also dead easy to make sure that your script ran at shutdown before the iSCSI and networking were closed down as well.

Now, with systemd the first part is more tricky as the program tries to do everything at once. If you want to have a native systemd service which starts before nfs-server then you have to modify that service description too. You might as well just have an init script which runs last which shuts down nfs-server before mounting the iSCSI disk and then starts it again when it finishes.

Now, it gets worse when the system is shutting down. Oh yes!

You see, systemd always tries to do things in a hurry. It seems that the design philosophy was better to do things quickly rather than correctly, and this is especially true at shutdown.

In a discussion thread on the systemd-devel mailing list titled “[systemd-devel] systemd and LSB Required/Should-Stop” it’s stated by Lennart Poettering:

On Fri, 24.06.11 14:04, Ville Skyttä (ville.skytta at iki.fi) wrote:

> Hello,
> 
> Am I right that systemd does currently have no support for the LSB
> Required-Stop and Should-Stop keywords?  If yes, are there plans to
> support them in the future?

That is true, we do not support this right now. In systemd the shutdown
order is always the inverse startup order of two services. I am not
entirely opposed to make this more flexible but so far I have not heard
of a good usecase where the flexibility to detach the shutdown oder from
the startup order was necessary.

Now, what this means is that any program or script called during the shutdown process is in a race to complete before the services it depends upon disappear. There is no ability to pause the shutdown to wait for vital events to happen, such as the synchronising and unmounting of disks. This is completely broken, by design.

Emulating a Sun.

A week ago, after completing the installation of a Raspberry Pi into an old Sun CDROM drive external enclosure, I posted a picture of the enclosure on Facebook. The response from an old friend was, “Can you run SunView on it?”

20130825-193902.jpg

Of course, a Raspberry Pi is no Sun Workstation so, the answer was, not directly. However, I immediately did a Google search for Sun emulators and was very surprised to find that someone had actually written one. And so my project for the next week was born.

After downloading the source to TME (The Machine Emulator) and installing all the development libraries I tried to build the blighter. Unfortuately this was not as simple as you may have thought as the configure script had a bug which only showed itself if not on a x86 machine. This and the build system’s requirement for an ancient version of the libltdl module loading library for libtool took me nearly three days to work through. Still, I did now have a binary to play with.

It was then time to try it out. Somewhat sillily I followed some instructions I found on the Internet. These instructions included how to create a disk image file and then how to configure it from within the “tape” booted kernel’s format command. Following these instructions caused the emulator to Bus Error and crash when it tried to write to the virtual disk for the first time. I wasted a day trying to debug this… but it turned out that the instructions were wrong!

Having gone back to basics and used the size and parameters from a Seagate ST1480N disk, used in the SPARCstation 2, I was able to format, partition and install the miniroot. I thought this was the end of my problems… until I tried to install the OS.

The “suninstall” program just refused to be able to seek to the correct tape file for the installation files even though mt(1) worked perfectly. Puzzling. Another number of hours trying to find why this wasn’t working and I find that the VM config file needed parameters removing from the tape drive definition as the default only worked with NetBSD as the OS and not SunOS. :-/

Everything seemed good. After a couple of hours I had a working VM on the Pi and it didn’t even seem that slow. So, I logged in and started SunView. Woohoo! Task complete, almost. The mouse emulation didn’t work properly. I thought that maybe it was due to the VM running too slowly to read the input correctly. Still, I could a least post “task complete” as I did have SunView running, which is what Richard had asked.

20130825-194105.jpg

It now took me another day or so of debugging to determine that the Mouse Systems mouse emulation was broken when encoding the movement delta values. The original code seems to require the compiler to pretend that hex values are signed integers when assigning to unsigned 8 bit integers. Well, dodgy!

Having fixed that bug all now works, if rather slowly.

Given the interest on Google+ and Twitter about this emulation I’ve spent today creating an automatic download, configure and build script with an easy to use prototype VM creation system for an emulated Sun3 running SunOS 4.1.1, which you can download here.

On the Origin of Technological Civilisation.

This morning a friend posted an image of a supernova on Facebook and wondered just home many civilisations died as a result. Now, if you take the standard Drake equation and use that as a basis of your estimation of technological life and hence civilisation then you may get the idea that at least one did, given the massive gamma ray bust associated with such an event. However, I don’t believe this at all, and here’s why:

The parameters usually plugged into the Drake Equation assume that eventually that the development of technological lifeforms is almost inevitable once you get life going, as is even the equivalent of multi-cellular and complex lifeforms before this. Given my perspective as an Earth Scientist by training and hence knowledge about how the Earth, the only place we know life exists, I very much dispute these assumptions.

Life is common, complex life probably not so much

If we look at the Earth as an example of how life may develop on a planet we find from the evidence that simple, single celled life appeared pretty darned quickly after the end of the late bombardment where it would have been practically impossible for anything to survive. So, we can assume that this kind of life is probably likely to spring up almost anywhere in the Universe given similar starting conditions. So, life in the Universe is common.

However, after this great “leap” life got lazy. It didn’t really change a great deal for over 2.5 billion years. OK, it had to cope with the rise in oxygen and switch power sources but otherwise it didn’t do a great deal other than maybe become symbiotic and file its DNA away into a special container. Basically, there was no evolutionary massive advantage to change, so it didn’t.

From the fossil record it currently looks as though a global climatic event effectively pushed life to co-operate so as to survive in challenging environments. Without this push, life on Earth would probably still be single celled.

So, just about a billion years ago we got multi-cellular life… Woo-hoo! It took a while before this became complex though and it seems that only when some of these found eating other life to be a convenient method of energy collection did the arms race begin and complex life began.

It was all a big accident!

Climbing the ladder to technology? Maybe not.

So, from this slow start it only took about 500 million years to get to creatures which could potentially have enough brain power to be intelligent enough to wield tools. So, why didn’t we see technological dinosaurs?

Well, technological intelligence requires a couple of things, firstly the abstract, innate intelligence and flexible world modelling capabilities so as to visualise the tool and make the imaginative leap to think them up in the first place. Secondly, you need the twist of evolutionary fate which gives the organism the body parts required to fashion and use technology.

It’s becoming more and more apparent that many creatures from many strands of life are capable of the first part. Not only great apes or primates or even mammals but birds (i.e. dinosaurs) and even molluscs (Cephalopoda, i.e. octopus). However, most of these intelligent beings are handicapped so that technological advancement isn’t practical. They either don’t have the tools, don’t have the time, live in the wrong sort of environment or aren’t social.

Also, in many ways, the pre-requisites for being technological aren’t usually the best for long-term survival in an evolutionary sense. Generalists generally find it hard to compete against specialists, unless there are specific environmental drivers which cause the specialists to fail. Humans almost didn’t make it.

So, humans are an accident?

Basically, yes. We are an aberration. We only made it as a sheer fluke. Given the odds we shouldn’t be here at all and the planet Earth would be no different than it has been since the last great extinction.

So, what does this mean for the Drake Equation?

We have to remember that we’re looking down the wrong end of the telescope at this problem and hence get a very skewed idea. We are here to observe and hence is seems that that must be proof of the inevitability of us appearing. The original parameters of the Drake Equation reflect this and are, in my opinion given the evidence, several tens of orders of magnitude too optimistic.

Well, even given the hugely, mind bogglingly big numbers of potential life harbouring planets out there it’s very probable that only a really, really tiny percentage managed to get beyond single cellular organisms. Even then the combination of factors which would allow a society to develop technology and become a civilisation are so remote and actively discouraged.

Is there anyone else out there then?

Probably not. Sorry.

Given the odds it’s quite possible that we are the very first technological beings to exist within the Universe, given that the Sol system was possibly one of the first to appear after enough building blocks had been created by the previous generations of stars.  Even if we are not, given the number of star systems out there, the time scales involved and the probable life of any species being only a couple of millions of years at best we’ve probably missed the previous ones and others will appear after we’re long gone.

We are but a fleeting island in entropy’s march.

Openindiana: How could the developers go so wrong?

Well, today I’ve been playing with OpenIndiana, the OpenSolaris derivative created after Oracle killed off its ancestor.

Well, to say that I was rather disappointed would be an understatement. It’s rather obvious that the developers of the distribution are not system administrators of integrated networked environments otherwise they would not have made such stupid design decisions.

Anyway, here’s the story of my day:

I downloaded the live DVD desktop version initially as I assumed that this would, when installed, effectively replicate a Solaris desktop environment. Seeing as Solaris in this configuration is capable of being a fully functional server as well I assumed that this would be the case for Openindiana.

So, I created a virtual machine under VirtualBox on the Mac, booted the DVD image and started the install. I was surprised about how little interaction there was during the install process as all it asked about was how to partition the disk and to create a root password and a new user. After the install things went down hill.

Now, it seems that the Openindiana bods are trying to ape Linux. When you boot up you get a GDM login screen, but can’t log in as root. So, you log in as the user you created, not too much of a problem, except that you now can’t start any of the configuration applications, they fail silently after you type the root password. You can’t sudo commands as it says that you don’t have permission…

Finally, I managed to get past this roadblock by trying ‘su –‘ which then asked me to change the root password! Once this was done I could actually run the configuration utilities. Not that it got me very much further, as there seems to be no way to set a static IP address out of the box.

I decided to trash that version and download the server version DVD. Maybe that would be better? Surely it would, it’s designed to be a server…

I booted the DVD image and the text installer started, very similar to the old Solaris installer to begin with, except all it asked about again was the disk partitioning, root password/user creation and networking, giving only the options for no networking or automatic network configuration. There was no manual network configuration! What?!!!! This is a server install!

Also missing from the installer was any way of setting up network authentication services or modifying what was installed. The installer had been lobotomised.

Once the OS had installed and booted up there were some more nasty surprises. Again, you couldn’t set a static IP address and any changes to the networking were silently reverted. It was only with some Googling that I managed to hunt down the culprit, the network/physical:nwam service, which is the default configuration. WHY?!!! This is a SERVER not a laptop!

Once this was fixed I managed to at least get a static IP address set up but it’s far more convoluted than with Solaris 10 or before.

Other strangeness in the design… All the X installation is there, except for the X server. Eh? What’s the point of that?

By default the GUI package manager isn’t installed. Once you do, however, it’s set up by default not to see any not installed packages, which is confusing. If you know where to look you can change this but it’s a stupid default.

Getting NFS client stuff working was a challenge as well. When you manually run mount everything seems to work out of the box. NFS filesystems mount fine and everything looks dandy. So, you put some mounts into /etc/vfstab and ‘mount -a‘ works as expected. Reboot, however, and nothing happens! This is due to the fact that most of the NFS client services are not turned on by default but magically work if you run mount. Turning on the nfs/client:default service doesn’t enable the other services it requires, however, but you don’t see this until a reboot. Stupid! It should work the same way at all times. If it works magically on the command line it should work at boot as well and vice versa. Unpredictability on a server is a killer.

On the bright side, at least the kernel is enterprise grade.

Elite: Dangerous, possibly the ultimate Elite-type game?

In a musing on my old blog site situated at LiveJournal eight years ago I outlined my idea for the best multi-player virtual Universe.

In November David Braben announced a new Kickstarter project to attempt to fund the very long awaited fourth Elite game in the series, “Elite: Dangerous” and this time it’s networked!

Elite: Dangerous

From the description so far it looks as though the procedural Universe creation is very similar to that I thought of in my original LiveJournal posting all those years ago, at least down to the star system level. At least in the first iteration there’s not going to be any chance to exit your ship and wander around the planets etc. But that’s fine.

The only problem I see is with rather large amount of money the Kickstarter project is asking for, £1.25 million! The fund raising is more than half the way through and still the pledges are only just above the half way mark and the rate of increase is slow. Somehow I don’t see it reaching the funding target by the 5th January.

I’ve put my money where my mouth is and pledged as much as I can reasonably afford. If you too would like to see Elite IV come into being then go over to Kickstarter and help with the fund raising.

Update: New video of gameplay:

New gameplay video.