Difference between revisions of "TODO"
m |
|||
Line 9: | Line 9: | ||
'''Note: the s3 bucket needs to be secure as it contains logins''' |
'''Note: the s3 bucket needs to be secure as it contains logins''' |
||
+ | ==== Alright, that's done. Now I need to probably test the process of loading from the backups to make sure it is working ==== |
||
− | Also, need to create a backup of the LocalSettings.php configuration file. |
+ | Also, need to create a backup of the LocalSettings.php configuration file. And the /images directory once ppl start uploading files here... |
+ | |||
+ | ==== Create repo with full configuration scripts/instructions to restore the wiki from a backup. ==== |
||
+ | probably will more the wiki to a cheaper lightsail server soon, since it's using <800Mb of ram, hopefully 1Gb will be enough. and in a year, probably will handoff the server to someone else |
||
==== Page / data backups ==== |
==== Page / data backups ==== |
Revision as of 00:22, 21 May 2021
Setup automated backups
See: https://www.mediawiki.org/wiki/Manual:Backing_up_a_wiki
Database backups
For example, use automysqlbackup and send the daily/weekly backups to an aws s3 bucket
Maybe follow this tutorialor similar. Also, would be nice to keep daily backups for past month, but then only weekly longer than that. Not sure how to do that, but I'm sure it's been done before.
Note: the s3 bucket needs to be secure as it contains logins
Alright, that's done. Now I need to probably test the process of loading from the backups to make sure it is working
Also, need to create a backup of the LocalSettings.php configuration file. And the /images directory once ppl start uploading files here...
Create repo with full configuration scripts/instructions to restore the wiki from a backup.
probably will more the wiki to a cheaper lightsail server soon, since it's using <800Mb of ram, hopefully 1Gb will be enough. and in a year, probably will handoff the server to someone else
Page / data backups
Another option is to Export or Publish the mediawiki and upload the data to a public git repo.
Ideally, the wiki could also be published to static-pages hosted by GitHub pages
That way, there will always be a read-only copy of all the pages
Looking at using https://www.mediawiki.org/wiki/Parsoid to translate xml dumps to html
Alternatively, https://github.com/openzim/mwoffliner is basically a web-scraper so it's kinda dumb to webscrape your own server but whatever
Make sections for 1st year FAQ
Looks like someone already started: http://gt-student-wiki.org/mediawiki/index.php/First_Year_FAQ
Setup email server
Done; gmail account created
Secure secret keys
https://www.mediawiki.org/wiki/Manual:Securing_database_passwords
Also, configure SSL
Should be fairly easy to do (using AWS route 53 rn)
Configure account creation
Require GT emails to create an account
Something similar to this: https://www.mediawiki.org/wiki/Topic:Rg8scj0rl4iqulmg