Talks (was: Re: [YLUG] A basic question)

Alex Howells alex.howells at 0wn3d.us
Fri Dec 28 21:05:35 GMT 2007


On 28/12/2007, Steve Kemp <steve at steve.org.uk> wrote:
>   I've run into two distinct problems which have the same cause,
>  validating "normal" fields is hard.  There are people who'll just
>  abuse the mail format.
>
>   Specifically I tried to:
>
>    1.  Validate the 'Date' header on mails, if present.

Yeah, this should be an obvious one since spammers often set the date
to something way in the past/future as an attempt to make your mail
client display it at the top - did you have any success filtering mail
if it was +/- a few days out?

>    2.  Attempt to deny "foreign" messages based upon an examination
>       of the character set used.  (I'll save the charset rants for
>       another time...)

I've noticed Google is capable of showing you only search results in a
chosen language, and do think that'd be rather cool for e-mail as
well; provided it doesn't try to intepret an attached binary as
language, of course.

Frankly 100% of the e-mail I care about arrives in English, I don't
speak any other languages barring a passing familiarity with Welsh and
perhaps French.  Being able to say "Not English? GET IN THE
QUARANTINE!" would rock.

> >       Is there an individual quarantine per user/domain, perhaps
> >       accessible through a web interface?
>
>   Ahem.

I know I've seen your system ;)  Was curious about other stuff out there too!

> >   *  Scalable web hosting  -  deploying WWW services behind load balancers
> >       and utilizing technologies like caching on dynamic content.
> >       Would be particularly interested in hearing from someone who's
> >       deployed a Ruby on Rails application in such a manner, given that
> >       it appears such a c*nt to deploy in any sane fashion :-)
>
>   I've no experience of large-scale hosting, but I do highly recommend
>  fragment caching via Danga's memcached or similar software.

Amazing isn't it?  There *must* be some big sites hosted at work but
we rarely see anything "below" the O/S ;)

>   My Debian site uses it extensively and I'd be happy to show you the
>  cache hit/miss rates if you're curious.  In short when you have per-user
>  pages being generated on the fly caching them doesn't work out as well
>  as it could - because too many parts (potentially) change.
>   Caching the computationally intensive sections is a significant
>  win.  (Probably pretty obvious..?)

Is memcached something that'll apply to any dynamic content, so stuff
generated by mod_perl, PHP, Ruby and all that?  What're the drawbacks
to the technology?  Does it require much modification to site code to
define what's cached?

>   The only thing that I'd say here, with a nod to Joel On Software, is
>  that your "product" (be it a commercial program, another piece of
>  software, or a release of a complete distribution) should be released
>  via a single tool.
>
>   If you can type 'make release', '/home/$foo/release-me.sh
>   --version=2.0.1' then you've got it made.
>
>   If you cannot then you're at the mercy of the people who can do
>  the magic.  I've watched large projects switch to this model of
>  allowing easy release of:
>
>     * Final builds
>     * PDF documentation.
>     * Release notes.
>     * Source Control tagging / etc.

That "seems" ridiculous but I can think of at least 15-20 projects
which do it this way. So a large percentage. Would you care to comment
on why they don't adopt a 'single tool' approach, given its clearly
superior?

Sounds like you'd be a good candidate for giving one of these proposed
talks! :-P



More information about the York mailing list