<br><br><div class="gmail_quote">On 12 July 2012 12:06, Richard Ibbotson <span dir="ltr"><<a href="mailto:richard.ibbotson@gmail.com" target="_blank">richard.ibbotson@gmail.com</a>></span> wrote:<br><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
<div class="im">> On 11/07/12 22:19, Gary Stainburn wrote:<br>
> > I've always use RedHat followed by Fedora for my servers (and<br>
> > workstations for that matter). I did try moving to CentOS but was<br>
> > frustrated by the restrictive RPM base. I haven't tried anything<br>
> > else yet, but was wondering what people's thoughts/suggestions<br>
> > were<br>
<br>
</div><div class="im">On Thursday 12 July 2012 11:28:47 Chris Davies wrote:<br>
> The reason I moved to Debian, and similar reasons have been quoted<br>
> by the SysAdmin team, is that the package management "just works",<br>
> and hides much of the dependency hell from the administrator. I<br>
> believe yum now does a fair job on RH-type systems but I for one<br>
> got absolutely fed up with the number of rpms that needed<br>
> downloaded individually each time a new package was required. Once<br>
> bitten, and all that.<br>
<br>
</div>My own 10p's-worth on this is that I started on RedHat and Debian back<br>
in 1995. At that time the RPM package manager was a pile of *******<br>
****. To be nice about it. Present day I find it hard to choose<br>
between apt-get and yum. Sometimes I use smart. TBH I find that<br>
Fedora or RedHat is easy to use for servers. Debian is also good.<br>
<br>
Last week I tried to install a Debian Apache server with a Debian 6.05<br>
netinstall image. It produced a non-bootable hung system. Had to use<br>
System Rescue CD to get it to boot. Probably a good idea to remember<br>
that this is x86 hardware. Which probably isn't seen very much in<br>
commercial circles. Outside of a University. Okay, so Universities<br>
and organisations like NASA use x86 PC clusters but probably not in<br>
commercial organisations. I worked in a BT data centre in Sheffield a<br>
few times. Not much x86 hardware in the servers that are being used.<br>
<span class="HOEnZb"><font color="#888888"><br></font></span></blockquote><div><br></div><div>You say that, but BT had a project to consolidate their internal systems a few years ago and while it started on Sun kit, it ended on HP x86 blades. I won't say it was entirely successful as it wasn't but there's been a trend in enterprises to get away from midrange to x86 clusters or blades, mostly driven by HP and organisations who got stung by Sun's changes in licensing after the Oracle takeover. Again, it's not always successful but if I was building that sort of thing these days I'd be inclined to use CentOS or RedHat on carrier grade x86 hardware for the price.</div>
<div><br></div><div>s/</div><div> </div></div>-- <br>Twitter: @sfgreenwood<div>"TBA are particularly glib"<br></div><br>