A screen shot of the GitHub home page taken on 2015-11-08
|URL||• • •|
|Archiving status||Not saved yet|
|IRC channel||(on EFnet)|
- See also GitHub Downloads
GitHub is a software repository powered by Git. Does not seem to have any site issues, often 24 hours uptime (see site status). Looks pretty sunny at the moment, but when disaster strikes it would be a problem archiving the private repositories.
As of 12th August 2012: 1,963,652 people hosting over 3,460,582 repositories 1,117,147 public repositories are forks, which greatly reduces the amount of data required to archive it. As of 22 November 2015: There are 32,000,000 repositories, with a similar fork ratio. Back-of-the-envelope calculations suggest 120TB of data in git repositories.
Acquisition by Microsoft
It was reported by Bloomberg on June 3rd that Microsoft agreed to buy GitHub for a currently unknown sum of money, and that the deal would be announced the following day (June 4th). With users in the FOSS community fearful of past horror stories of Microsoft's "embrace, extend, extinguish" schemes in the 1990s and 2000s and calls to move to rival GitLab in the wake of the news, a discussion into the feasibility of archiving GitHub has commenced in (on EFnet).
"git clone" is the simplest one. However, it does not get some project data that is not stored in git, including issue reports, comments, pull requests.
github-backup runs in a git repository and chases down that information, committing it to a "github" branch. It also chases down the forks and efficiently downloads them as well.
See also Software Heritage.
GitHub Replacement Engines
If we ever have to archive the data out of GitHub, the data will need to be exportable to a GitHub-style engine.
Currently, the best GitHub-style engine that has a Wiki, issues, Git Repo hosting, and is free and open source to use is GitLab. The engine is used by and paid for by many major organizations, so it is likely to live on in a stable way.
We will need a complete migration system to move a git repository and all related GitHub service information of a repository to GitLab.
Things to Scrape
In case of emergency, these are the items we need to grab.
- Git Repository - Accomplished by github-backup
- Forked Repositories - Accomplished by github-backup
- Notes on Commits/Lines of Code - Not supported by github-backup yet. API support was just added for it.
- Github Gollum Wiki - No tool yet, but just clone the whole thing, and then push it to GitLab.
- Releases - Tags on Github can have binaries attached. These are of high priority to archive.
List of Repositories
A list of repositories from Github API data are maintained by an archive team member at za3k.com. It scrapes continuously. Public downloads are updated once a day. This list does not include gists.
The metadata generated by the GitHub API are archived to Google BigQuery every hour by GithubArchive.
It obviously doesn't grab events dating before 2011, so a targeted repository scrape may still be ideal.
But at least it could be possible to grab all info about a single repository using Google BigQuery's free version, since it would use a low amount of CPU power. However, we need to create such an export script for it when the time comes.
- Issues + Comments - Accomplished by github-backup
- Milestones - Github Backup currently does not archive this yet.
- Labels - Github Backup currently does not archive this yet.
- Hooks - Needs some kind of tool to archive Github Hooks
- • • •