[OWW-Discuss] Why $0.00 Is the Future of Business
Bill F
bill.altmail at gmail.com
Sun Mar 2 16:32:35 EST 2008
I'd be glad to discuss options tomorrow (Monday).
There's a split between the way a wiki can be exported and the way the
database is managed. Since we currently use MySQL 4 with a single server,
there are some limitations on what we can do, yet there are options re:
replication we've not explored. The MediaWiki server also affords more
options we can look at as well.
It turned out that we did have a little bit of luck in that the backup
provided by our hosting company, Rackspace, was only a binary file backup
and not a mysql dump. The backup file images for the database were in good
shape. This can be a problem: this binary backup can result in database
corruption since the size of some of the files is so large that there are
seconds or more of time between the start and end of the file copy within
which transactions being written to the file can result in an inconsistent
backup image. This is an optional service we didn't know Rackspace hadn't
enabled when we moved to our current server.
I immediately deployed a script I've used before, automysqlbackup.sh, as
soon as the database was stable. This script is reliably creating static sql
backups on a nightly basis. We're also enabling the binary log option to
allow for intra-day incremental backup in addition to the nightly full
database backup. This means that we'll be able to reload and be up and
running in minutes the next time. I know this is little solace to people who
lose information; we still want to maximize the amount of data that can be
recovered, hopefully to the point of losing no data.
At the MediaWiki level, we externally back up all of the wiki pages in the
database on a nightly basis. This is a disconnected backup from the SQL
dump. This didn't help us since the backup takes place around midnight,
before the Rackspace backup we reloaded from.
Using the Mediawiki API, we can do incremental exports of all new pages on a
regular basis. This would allow us to recover the MediaWiki application
level pages via a simple import via a special page. The nice thing about the
API is that it can be run on a remote system, thus allowing us to provide
for a constant auto-import of the pages into another server. It's not clear
how we would manage the process of rolling the wiki onto a remote system for
operational purposes. But we would never lose the contents of the database
up until the time it went down. To implement this, we need to write a script
that uses the mediawiki api. I'll be glad to provide details to anyone
interested in working with me to do this. I can provide bindings in Python,
Perl, C#, or PHP for this.
We currently have a full backup of all of the images and attachments
uploaded. None were lost last week. But the database records pointing to
them are only current as of around 4:00 AM on the 28th. Since these files
are in a set of well-known directories, rsync can work the same way the page
collection works. I'll have this set up today on an MIT system. We can sync
on an hourly basis to keep the images up to date.
We have no problem with the performance of Rackspace in this situation. They
responded to our request within 15 minutes. Since we had to reload all of
the databases for the private wikis in addition to the main wiki (the MySQL
INNODB tables use a single large binary file shared by all databases), I
also needed to do full backups of all of the other databases, load in the
4:00 AM database image, then load in the backups to get them back to the
10:00 PM state they were in. We didn't lose data in any applications except
for the main OpenWetWare wiki.
Because we use OpenID for access control on all of the private wiki's, they
were inaccessible until the main wiki came back up.
The blog was completely disconnected from the OpenWetWare database. Because
of this, there was no effect on it at all except for the need to swap in the
new database image and then refresh it with the 10:00 PM info when it the
database server was re-enabled.
Let's see what we can do to improve the system. As I mentioned, I'm open to
suggestions but we're already hard at work hardening the system. I'm not
saying this to appear disinterested in external assistance; we need to do
what we know we can immediately.
Thanks.
Bill Flanagan
On Sun, Mar 2, 2008 at 3:20 PM, Alexander Wait Zaranek <
await at genetics.med.harvard.edu> wrote:
> On Tue, Feb 26, 2008 at 8:15 PM, Alexander Wait Zaranek
> <await at genetics.med.harvard.edu> wrote:
> > On Tue, Feb 26, 2008 at 6:03 PM, Bryan Bishop <kanzure at gmail.com> wrote:
> > > Encourage people to set up servers and backups
> > > of the wiki all over the place, with central aggregation nodes to
> make
> > > sure all of the updates are propagated.
> > >
> > actually, i wanted to offer to do this last steering committee
> > meeting. Anyone else doing it already? We could also run a mysql
> > slave so edits were up to the minute and not just a dump. Setup a
> > dedicated virtual machine on one of our clusters? I'd love to see it
> > happen...
> >
> "Feb. 28, 2008. OpenWetWare.org sustained a database failure on Feb.
> 28, but is back online. We're deeply sorry for any inconvenience this
> may have caused. We'll update the community on what we've done to
> recover and add more reliability to our procedures and infrastructure.
> Documents edited or created in www.OpenWetWare.org on Feb 28 between
> 4:00 AM EST to 7:00 PM EST will need to be updated or re-entered."
>
> So, it's never ideal to work on master-slave replication *after* a
> database failure but there's no time like the present. There's a
> bunch of talented freelance admins around the Church lab that could
> help with this. And we have the bandwidth/infrastructure too.
>
> How can we help?
> Sasha
> _______________________________________________
> OpenWetWare Discussion Mailing List
> discuss at openwetware.org
> http://mailman.mit.edu/mailman/listinfo/oww-discuss
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://mailman.mit.edu/pipermail/oww-discuss/attachments/20080302/8860759a/attachment-0001.htm
More information about the Oww-discuss
mailing list