Skip to content
Snippets Groups Projects
gitlab.md 75.82 KiB

GitLab is a web-based DevOps lifecycle tool that provides a Git-repository manager providing wiki, issue-tracking and continuous integration/continuous deployment pipeline features, using an open-source license, developed by GitLab Inc (Wikipedia). Tor uses GitLab mainly for issue tracking, wiki hosting and code review for now, at https://gitlab.torproject.org, after migrating from howto/trac.

Note that continuous integration is documented separately, in the CI page.

Tutorial

How to get an account?

You might already have an account! If you were active on Trac, your account was migrated with the same username and email address as Trac, unless you have an LDAP account, in which case that was used. So head over to the password reset page to get access to your account.

If your account was not migrated, send a mail to gitlab-admin@torproject.org to request a new one.

If you did not have an account in Trac and want a new account, you should request a new one at https://gitlab.onionize.space/.

How to report an issue in Tor software?

You first need to figure out which project the issue resides in. The project list is a good place to get started. Here are a few quick links for popular projects:

If you do not have a GitLab account or can't figure it out for any reason, you can also use the mailing lists. The tor-dev@lists.torproject.org mailing list is the best for now.

How to report an issue in the bugtracker itself?

If you have access to GitLab, you can file a new issue after you have searched the GitLab project for similar bugs.

If you do not have access to GitLab, you can email gitlab-admin@torproject.org.

Note about confidential issues

Note that you can mark issues as "confidentials" which will make them private to the members of the project the issue is reported on (the "developers" group and above, specifically).

Keep in mind, however, that it is still possible issue information gets leaked in cleartext, however. For example, GitLab sends email notifications in cleartext for private issue, an known upstream issue. (We have decided we cannot fix this ourselves in GitLab for now.) Some repositories might also have "web hooks" that notify IRC bots in clear text as well, although at the time of writing all projects are correctly configured.

How to contribute code?

As reporting an issue, you first need to figure out which project you are working on in the GitLab project list. Then, if you are not familiar with merge requests, you should read the merge requests introduction in the GitLab documentation. If you are unfamiliar with merge requests but familiar with GitHub's pull requests, those are similar.

Note that we do not necessarily use merge requests in all teams yet, and Gitolite still has the canonical version of the code. See issue 36 for a followup on this.

Also note that different teams might have different workflows. If a team has a special workflow that diverges from the one here, it should be documented here. Those are the workflows we know about:

If you do not have access to GitLab, please use one of the mailing lists: tor-dev@lists.torproject.org would be best.

How to quote a comment in a reply?

The "Reply" button only creates a new comment without any quoted text by default. It seems the solution to that is currently highlighting the text to quote and then pressing the r-key. See also the other keyboard shortcuts.

Alternatively, you can copy-paste the text in question in the comment form, select the pasted text, and hit the Insert a quote button which look like a styled, curly, and closing quotation mark .

How-to

Continuous Integration (CI)

All CI documentation resides in a different document see service/ci.

Email interactions

You can interact with GitLab by email too.

Creating a new issue

Clicking on the project issues gives a link at the bottom of the page, which says say "Email a new issue to this project".

That link should go into the "To" field of your email. The email subject becomes the title of the issue and the body the description. You can use shortcuts in the body, like /assign @foo, /estimate 1d, etc.

See the upstream docs for more details.

Commenting on an issue

If you just reply to the particular comment notification you received by email, as you would reply to an email in a thread, that comment will show up in the issue.

You need to have email notifications enabled for this to work, naturally.

You can also add a new comment to any issue by copy-pasting the issue-specific email address in the right sidebar (labeled "Issue email", introduced in GitLab 13.8).

This also works with shortcuts like /estimate 1d or /spend -1h. Note: for those you won't get notification emails back, though, while for others like /assign @foo you would.

See the upstream docs for more details.

Quick status updates by email

There are a bunch of quick actions available which are handy to update an issue. As mentioned above they can be sent by email as well, both within a comment (be it as a reply to a previous one or in a new one) or just instead of it. So, for example, if you want to update the amount of time spent on ticket $foo by one hour, find any notification email for that issue and reply to it by replacing any quoted text with /spend 1h.

How to migrate a Git repository from legacy to GitLab?

See the git documentation for this procedure.

How to mirror a Git repository from legacy to GitLab?

See the git documentation for this procedure.

How to mirror a Git repository from GitLab to GitHub

Some repositories are mirrored to THE torproject organization on GitHub. This section explains how that works and how to create a new mirror from GitLab. In this example, we're going to mirror the tor browser manual.

  1. head to the "Mirroring repositories" section of the settings/repository part of the project

  2. as a Git repository URL, enter:

    ssh://git@github.com/torproject/tb-manual.git
  3. click "detect host keys"

  4. choose "SSH" as the "Authentication method"

  5. don't check any of the boxes, click "Mirror repository"

  6. the page will reload and show the mirror in the list of "Mirrored repositories". click the little "paperclip" icon which says "Copy SSH public key"

  7. head over to https://github.com and authenticate with torproject-pusher, password is in external-services-git, in the password manager

  8. go to settings/keys and hit new SSH key

  9. paste the public key in the bigger field, as a title, use the URL of the repository, for example:

    Title: https://gitlab.torproject.org/tpo/web/manual mirror key
    Key: ssh-rsa AAAA[...]
  10. click "Add SSH key"

  11. to speed up the process, you can import the repository in GitHub, otherwise create a new repository. in both cases make sure you change the namespace from the default (torproject-pusher, which is incorrect) to the torproject namespace (which is correct)

  12. then hit the "reload" button in the repositories mirror list

If there is an error, it will show up as a little red "Error" button. Hovering your mouse over the button will show you the error.

How to find the right emoji?

It's possible to add "reaction emojis" to comments and issues and merge requests in GitLab. Just hit the little smiley face and a dialog will pop up. You can then browse through the list and pick the right emoji for how you feel about the comment, but remember to be nice!

It's possible you get lost in the list. You can type the name of the emoji to restrict your search, but be warned that some emojis have particular, non-standard names that might not be immediately obvious. For example, 🎉, U+1F389 PARTY POPPER, is found as tada in the list! See this upstream issue for more details.

Hooking up a project with the bots

By default, new projects do not have notifications setup in #tor-bots like all the others. To do this, you need to configure a "Webhook", in the Settings -> Webhooks section of the project. The URL should be:

https://kgb-bot.torproject.org/webhook/

... and you should select the notifications you wish to see in #tor-bots. You can also enable notifications to other channels by adding more parameters to the URL, like (say) ?channel=tor-foo. Important note: do not try to put the # in the channel name, or if you do, URL-encode it (e.g. like %23tor-foo), otherwise this will silently fail to change the target channel. Other parameters are documented the KGB documentation.

Note that GitLab admins might be able to configure system-wide hooks in the admin section, although it's not entirely clear how does relate to the per-project hooks so those have not been enabled. Furthermore, it is possible for GitLab admins with root access to enable webhooks on all projects, with the webhook rake task. For example, running this on the GitLab server (currently gitlab-02) will enable the above hook on all repositories:

sudo gitlab-rake gitlab:web_hook:add URL='https://kgb-bot.torproject.org/webhook/'

Note that by default, the rake task only enables Push events. You need the following patch to enable others:

modified   lib/tasks/gitlab/web_hook.rake
@@ -10,7 +10,19 @@ namespace :gitlab do
       puts "Adding webhook '#{web_hook_url}' to:"
       projects.find_each(batch_size: 1000) do |project|
         print "- #{project.name} ... "
-        web_hook = project.hooks.new(url: web_hook_url)
+        web_hook = project.hooks.new(
+          url: web_hook_url,
+          push_events: true,
+          issues_events: true,
+          confidential_issues_events: false,
+          merge_requests_events: true,
+          tag_push_events: true,
+          note_events: true,
+          confidential_note_events: false,
+          job_events: true,
+          pipeline_events: true,
+          wiki_page_events: true,
+        )
         if web_hook.save
           puts "added".color(:green)
         else

See also the upstream issue and our GitLab issue 7 for details.

You can also remove a given hook from all repos with:

sudo gitlab-rake gitlab:web_hook:rm URL='https://kgb-bot.torproject.org/webhook/'

And, finally, list all hooks with:

sudo gitlab-rake gitlab:web_hook:list

Setting up two-factor authentication (2FA)

We strongly recommend you enable two-factor authentication on GitLab. This is well documented in the GitLab manual, but basically:

  1. first, pick a 2FA "app" (and optionally a hardware token) if you don't have one already

  2. head to your account settings

  3. register your 2FA app and save the recovery codes somewhere. if you need to enter a URL by hand, you can scan the qrcode with your phone or create one by following this format:

    otpauth://totp/$ACCOUNT?secret=$KEY&issuer=gitlab.torproject.org

    where...

    • $ACCOUNT is the Account field in the 2FA form
    • $KEY is the Key field in the 2FA form, without spaces
  4. register the 2FA hardware token if available

GitLab requires a 2FA "app" even if you intend to use a hardware token. The 2FA "app" must implement the TOTP protocol, for example the Google Authenticator or a free alternative (for example free OTP plus, see also this list from the Nextcloud project). The hardware token must implement the U2F protocol, which is supported by security tokens like the YubiKey, Nitrokey, or similar.

Deleting sensitive attachments

If a user uploaded a secret attachment by mistake, just deleting the issue is not sufficient: it turns out that doesn't remove the attachments from disk!

To fix this, ask a sysadmin to find the file in the /var/opt/gitlab/gitlab-rails/uploads/ directory. Assuming the attachment URL is:

https://gitlab.torproject.org/anarcat/test/uploads/7dca7746b5576f6c6ec34bb62200ba3a/openvpn_5.png

There should be a "hashed" directory and a hashed filename in there, which looks something like:

./@hashed/08/5b/085b2a38876eeddc33e3fbf612912d3d52a45c37cee95cf42cd3099d0a3fd8cb/7dca7746b5576f6c6ec34bb62200ba3a/openvpn_5.png

The second directory (7dca7746b5576f6c6ec34bb62200ba3a above) is the one visible in the attachment URL. The last part is the actual attachment filename, but since those can overlap between issues, it's safer to look for the hash. So to find the above attachement, you should use:

find /var/opt/gitlab/gitlab-rails/uploads/ -name 7dca7746b5576f6c6ec34bb62200ba3a

And delete the file in there. The following should do the trick:

find /var/opt/gitlab/gitlab-rails/uploads/ -name 7dca7746b5576f6c6ec34bb62200ba3a | sed 's/^/rm /' > delete.sh

Verify delete.sh and run it if happy.

Note that GitLab is working on an attachment manager that should allow web operators to delete old files, but it's unclear how or when this will be implemented, if ever.

Publishing GitLab pages

GitLab features a way to publish websites directly from the continuous integration pipelines, called GitLab pages. Complete documentation on how to publish such pages is better served by the official documentation, but creating a .gitlab-ci.yml should get you rolling. For example, this will publish a hugo site:

image: registry.gitlab.com/pages/hugo/hugo_extended:0.65.3
pages:
  script:
    - hugo
  artifacts:
    paths:
      - public
  only:
    - main

If .gitlab-ci.yml already contains a job in the build stage that generates the required artifacts in the public directory, then including the pages-deploy.yml CI template should be sufficient:

include:
  - project: tpo/tpa/ci-templates
    file: pages-deploy.yml

GitLab pages are published under the *.pages.torproject.org wildcard domain. There are two types of projects hosted at the TPO GitLab: sub-group projects, usually under the tpo/ super-group, and user projects, for example anarcat/myproject. You can also publish a page specifically for a user. The URLs will look something like this:

Type of GitLab page Name of the project created in GitLab Website URL
User pages username.pages.torproject.net https://username.pages.torproject.net
User projects user/projectname https://username.pages.torproject.net/projectname
Group projects tpo/group/projectname https://tpo.pages.torproject.net/group/projectname

Accepting merge requests on wikis

Wiki permissions are not great, but there's a workaround: accept merge requests for a git replica of the wiki.

This documentation was moved to the documentation section.

Renaming a branch globally

While git supports renaming branches locally with the git branch --move $to_name command, this doesn't actually rename the remote branch. That process is more involved.

Changing the name of a default branch both locally and on remotes can be partially automated with the use of anarcat's branch rename script. The script basically renames the branch locally, pushes the new branch and deletes the old one, with special handling of GitLab remotes, where it "un-protects" and "re-protects" the branch.

You should run the script with an account that has "Maintainer" or "Owner" access to GitLab, so that it can do the above GitLab API changes. You will then need to provide an access token through the GITLAB_PRIVATE_TOKEN environment variable, which should have the scope api.

So, for example, this will rename the master branch to main on the local and remote repositories:

GITLAB_PRIVATE_TOKEN=REDACTED git-branch-rename-remote

If you want to rename another branch or remote, you can specify those on the commandline as well. For example, this will rename the develop branch to dev on the gitlab remote:

GITLAB_PRIVATE_TOKEN=REDACTED git-branch-rename-remote --remote gitlab --from-branch develop --to-branch dev

The command can also be used to fix other repositories so that they correctly rename their local branch too. In that case, the GitLab repository is already up to date, so there is no need for an access token.

Other users, then can just run this command will rename master to main on the local repository, including remote tracking branches:

git-branch-rename-remote

Obviously, users without any extra data in their local repository can just destroy their local repository and clone a new one to get the correct configuration.

Keep in mind that there may be a few extra steps and considerations to make when changing the name of a heavily used branch, detailed below.

Modifying open Merge Requests

A merge request that is open against the modified branch may be bricked as a result of deleting the old branch name from the Gitlab remote. To avoid this, after creating and pushing the new branch name, edit each merge request to target the new branch name before deleting the old branch.

Updating gitolite

Many GitLab repositories are mirrored or maintained manually on Gitolite (git-rw.torproject.org) and Gitweb. The ssh step for the above automation script will fail for Gitolite and these steps need to be done manually by a sysadmin. Open a TPA ticket with a list of the Gitolite repositories you would like to update and a sysadmin will perform the following magic:

cd /srv/git.torproject.org/repositories/
for repo in $list; do
    git -C "$repo" symbolic-ref HEAD refs/heads/$to_branch
done

This will update Gitolite, but it won't update Gitweb until the repositories have been pushed to. To update Gitweb immediately, ask your friendly sysadmin to run the above command on the Gitweb server as well.

Updating Transifex

If your repository relies on Transifex for translations, make sure to update the Transifex config to pull from the new branch. To do so, open a l10n ticket with the new branch name changes.

Find the project associated with a project ID

Sometimes you'll find a numeric project ID instead of a human-readable one. For example, you can see on the arti project that it says:

Project ID: 647

So you can easily find the project ID of a project right on the project's front page. But what if you only have the ID and need to find what project it represents? You can talk with the API, with a URL like:

https://gitlab.torproject.org/api/v4/projects/<PROJECT_ID>

For example, this is how I found the above arti project from the Project ID 647:

$ curl -s 'https://gitlab.torproject.org/api/v4/projects/647' | jq .web_url
"https://gitlab.torproject.org/tpo/core/arti"

Connect to the PostgreSQL server

The GitLab Omnibus setup is special: it ships its own embedded PostgreSQL server (!), which means the regular sudo -u postgres psql command doesn't work.

To get access to the PostgreSQL server, you need to follow the upstream instructions which are, at the time of writing:

sudo gitlab-psql -d gitlabhq_production

This actually expands to the following command:

sudo -u gitlab-psql /opt/gitlab/embedded/bin/psql -p 5432 -h /var/opt/gitlab/postgresql -d gitlabhq_production -d gitlabhq_production

An emergency dump, therefore, could be taken with:

cd /tmp ; sudo -u gitlab-psql /opt/gitlab/embedded/bin/pg_dump -p 5432 -h /var/opt/gitlab/postgresql -d gitlabhq_production -d gitlabhq_production | pv -s 2G > /srv/gitlab-backup/db/db-$(date '+%Y-%m-%d').sql

Yes, that is silly. See also issue 20.

Pager playbook

TODO: document how to handle common problems in GitLab

Troubleshooting

Upstream recommends running this command to self-test a GitLab instance:

sudo gitlab-rake gitlab:check SANITIZE=true

This command also shows general info about the GitLab instance:

sudo gitlab-rake gitlab:check SANITIZE=true

it is especially useful to find on-disk files and package versions.

GitLab pages not found

If you're looking for a way to track GitLab pages error, know that the webserver logs are in /var/log/nginx/gitlab_pages_access, but that only proxies requests for the GitLab Pages engine, which (JSON!) logs live in /var/log/gitlab/gitlab-pages/current.

If you get a "error":"domain does not exist" problem, make sure the entire pipeline actually succeeds. Typically, the "pages:deploy" job can fail with:

Artifacts for pages are too large

In that case, you need to go into the Admin Area -> Settings -> Preferences -> Pages and bump the size limit. It defaults to 100MB and we bumped it to 1024MB at the time of writing. Note that GitLab CI/CD also have a similar setting which might (or might not?) affect such problems.

PostgreSQL debugging

The PostgreSQL configuration in GitLab is particular. See the connect to the PostgreSQL server section above on how to connect to it.

Disk full on GitLab server

If the main GitLab server is running out of space (as opposed to runners, see Runner disk fills up for that scenario), then it's projects that are taking up space. We've typically had trouble with artifacts taking up space, for example (team#40615 (closed), team#40517 (closed)).

You can see the largest disk users in the GitLab admin area in Overview -> Projects -> Sort by: Largest repository.

Note that, although it's unlikely, it's technically possible that an archived project takes up space, so make sure you check the "Show archived projects" option in the "Sort by" drop down.

In the past, we have worked around that problem by reducing the default artifact retention period from 4 to 2 weeks (team#40516 (closed)) but obviously does not take effect immediately. More recently, we have tried to tweak individual project's retention policies and scheduling strategies (details in team#40615 (closed)).

Please be aware of the known upstream issues that affect those diagnostics as well.

To obtain a list of project sorted by space usage, log on to GitLab using an account with administrative privileges and open the Projects page sorted by Largest repository. The total space consumed by each project is displayed and clicking on a specific project shows a breakdown of how this space is consumed by different components of the project (repository, LFS, CI artifacts, etc.).

If a project is consuming an unexpected amount of space for artifacts, the scripts from the tpo/tpa/gitlab-tools project can by utilized to obtain a breakdown of the space used by job logs and artifacts, per job or per pipeline. These scripts can also be used to manually remove such data, see the gitlab-tools README.

It's also possible to compile some CI artifact usage statistics directly on the GitLab server. To see if expiration policies work (or if "kept" artifacts or old job.log are a problem), use this command (which takes a while to run):

find -mtime +14 -print0 | du --files0-from=- -c -h | tee find-mtime+14-du.log

To limit this to job.log, of course, you can do:

find -name "job.log" -mtime +14 -print0 | du --files0-from=- -c -h | tee find-mtime+14-joblog-du.log

Email routing

Incoming email get routed through either eugeni or the submission service, then end up on the Postfix server on gitlab-02, and from there, to a dovecot mailbox. You can use postfix-trace to confirm the message correctly ended up there.

Normally, GitLab should be picking mails from the mailbox (/srv/mail/git@gitlab.torproject.org/Maildir/) regularly, and deleting them when done. If that is not happening, look at the mailroom logs:

tail -f /var/log/gitlab/mailroom/mail_room_json.log | jq -c

A working run will look something like this:

{"severity":"INFO","time":"2022-08-29T20:15:57.734+00:00","context":{"email":"git@gitlab.torproject.org","name":"inbox"},"action":"Processing started"}
{"severity":"INFO","time":"2022-08-29T20:15:57.734+00:00","context":{"email":"git@gitlab.torproject.org","name":"inbox"},"uid":7788,"action":"asking arbiter to deliver","arbitrator":"MailRoom::Arbitration::Redis"}.734+00:00","context":{"email":"git@gitlab.torproject.org","name":"inbox"},"action":"Getting new messages","unread":{"count":1,"ids":[7788]},"to_be_delivered":{"count":1,"ids":[7788]}}ext":{"email":"git@gitlab.torproject.org","name":"inbox"},"uid":7788,"action":"sending to deliverer","deliverer":"MailRoom::Delivery::Sidekiq","byte_size":4162}","delivery_method":"Sidekiq","action":"message pushed"}
{"severity":"INFO","time":"2022-08-29T20:15:57.744+00:00","context":{"email":"git@gitlab.torproject.org","name":"inbox"},"action":"Processing started"}
{"severity":"INFO","time":"2022-08-29T20:15:57.744+00:00","context":{"email":"git@gitlab.torproject.org","name":"inbox"},"action":"Getting new messages","unread":{"count":0,"ids":[]},"to_be_delivered":{"count":0,"ids":[]}}0","context":{"email":"git@gitlab.torproject.org","name":"inbox"},"action":"Idling"}

Emails should be processed every minute or so.

Disaster recovery

In case the entire GitLab machine is destroyed, a new server should be provisionned in the howto/ganeti cluster (or elsewhere) and backups should be restored using the below procedure.

Running an emergency backup

A full backup can be ran as root with:

/usr/bin/gitlab-rake gitlab:backup:create

Backups are stored as a tar file in /srv/gitlab-backup and do not include secrets, which are backed up separately, for example with:

umask 0077 && tar -C /var/opt/gitlab -czf /srv/gitlab-backup/config_backup$(date +"\%Y\%m\%dT\%H\%M").tar.gz

See /etc/cron.d/gitlab-config-backup, and the gitlab::backup and profile::gitlab::app classes for the actual jobs that runs nightly.

Recovering this wiki from backups

If you need to immediately restore the wiki from backups, you can head to the backup server and restore the directory:

/var/opt/gitlab/git-data/repositories/@hashed/11/f8/11f8e31ccbdbb7d91589ecf40713d3a8a5d17a7ec0cebf641f975af50a1eba8d.git

The hash above is the SHA256 checksum of the wiki-replica project id (695):

$ printf 695 | sha256sum 
11f8e31ccbdbb7d91589ecf40713d3a8a5d17a7ec0cebf641f975af50a1eba8d  -

On the backup server, that would be something like:

bconsole
restore
5
46
cd /var/opt/gitlab/git-data/repositories/@hashed/11/f8
mark 11f8e31ccbdbb7d91589ecf40713d3a8a5d17a7ec0cebf641f975af50a1eba8d.git
done
yes

The files will end up in /var/tmp/bacula-restore on gitlab-02. Note that the number 46, above, will vary according to other servers backed up on the backup server, of course.

This should give you a copy of the git repository, which you can then use, presumably, to read this procedure and restore the rest of GitLab.

(Although then, how did you read this part of the procedure? Anyways, I thought this could save your future self one day. You'll thank me later.)

Restoring from backups

The upstream documentation has a fairly good restore procedure, but because our backup procedure is non-standard -- we exclude repositories and artifacts, for example -- you should follow this procedure instead.

Note that the procedure assumes some familiarity with the general backup and restore procedures, particularly how to restore a bunch of files from the backup server (see the restore files section.

This entire procedure will take many hours to complete. In our tests, it took:

  1. an hour or two to setup a VM
  2. less than an hour to do a basic GitLab install
  3. 20 minutes to restore the basic system (database, tickets are visible at this point)
  4. an hour to restore repositories
  5. another hour to restore artifacts

This gives a time to recovery of about 5 to 6 hours. Most of that time is spent waiting for files to be copied, interspersed with a few manual commands.

So here's the procedure that was followed to deploy a development server, from backups, in tpo/tpa/team#40820 (run everything as root):

  1. install GitLab using Puppet: basically create a server large enough for everything, apply the Puppet role::gitlab

    That includes creating new certificates and DNS records, if not already present (those may be different if you are created a dev server from backups, for example, which was the case for the the above ticket).

    Also note that you need to install the same GitLab version as the one from the backup. If you are unsure of the GitLab version that's in the backup (bad day uh?), try to restore the /var/opt/gitlab/gitlab-rails/VERSION file from the backup server first.

  2. at this point, a blank GitLab installation should be running. verify that you can reach the login page, possibly trying to login with the root account, because a working GitLab installation is a pre-requisite for the rest of the restore procedure.

    (it might be technically possible to restore the entire server from scratch using only the backup server, but that procedure has not been established or tested.)

  3. on the backup server (currently bacula-director-01), restore the latest GitLab backup job from the /srv/gitlab-backup and the secrets from /etc/gitlab:

    # bconsole
    *restore
    To select the JobIds, you have the following choices:
    [...]
     5: Select the most recent backup for a client
    [...]
    Select item:  (1-13): 5
    Defined Clients:
    [...]
        47: gitlab-02.torproject.org-fd
    [...]
    Select the Client (1-98): 47
    Automatically selected FileSet: Standard Set
    [...]
    Building directory tree for JobId(s) 199535,199637,199738,199847,199951 ...  ++++++++++++++++++++++++++++++++
    596,949 files inserted into the tree.
    [...]
    cwd is: /
    $ cd /etc
    cwd is: /etc/
    $ mark gitlab
    84 files marked.
    $ cd /srv
    cwd is: /srv/
    $ mark gitlab-backup
    12 files marked.
    $ done

    This took about 20 minutes in a simulation done in June 2022, including 5 minutes to load the file list.

  4. move the files in place and fix ownership, possibly moving pre-existing backups out of place, if the new server has been running for more than 24 hours:

    mkdir /srv/gitlab-backup.blank
    mv /srv/gitlab-backup/* /srv/gitlab-backup.blank
    cd /var/tmp/bacula-restores/srv/gitlab-backup
    mv *.tar.gz backup_information.yml  db /srv/gitlab-backup/
    cd /srv/gitlab-backup/
    chown git:git *.tar.gz backup_information.yml db/ db/database.sql.gz 
  5. stop GitLab services that talk with the database (those might have changed since the time of writing, review upstream documentation just in case:

    gitlab-ctl stop puma
    gitlab-ctl stop sidekiq
  6. restore the secrets files (note: this wasn't actually tested, but should work):

    chown root:root /var/tmp/bacula-restores/etc/gitlab/*
    mv /var/tmp/bacula-restores/etc/gitlab/{gitlab-secrets.json,gitlab.rb} /etc/gitlab/

    Note that if you're setting up a development environment, you do not want to perform that step, which means that CI/CD variables and 2FA tokens will be lost, which means people will need to reset those and login with their recovery codes. This is what you want for a dev server, because you do not want a possible dev server compromise to escalate to the production server, or the dev server to have access to the prod deployments.

    Also note that this step was not performed on the dev server test and this lead to problems during login: while it was possible to use a recovery code to bypass 2FA, it wasn't possible to reset the 2FA configuration afterwards.

  7. restore the files:

    gitlab-backup restore

    This last step will ask you to confirm the restore, because it actually destroys the existing install. It will also ask you to confirm the rewrite of the authorized_keys file, which you want to accept (unless you specifically want to restore that from backup as well).

  8. restart the services and check everything:

    gitlab-ctl reconfigure
    gitlab-ctl restart
    gitlab-rake gitlab:check SANITIZE=true
    gitlab-rake gitlab:doctor:secrets
    gitlab-rake gitlab:lfs:check
    gitlab-rake gitlab:uploads:check
    gitlab-rake gitlab:artifacts:check

    Note: in the simulation, GitLab was started like this instead, which just worked as well:

    gitlab-ctl start puma
    gitlab-ctl start sidekiq

    We did try the "verification" tasks above, but many of them failed, especially in the gitlab:doctor:secrets job, possibly because we didn't restore the secrets (deliberately).

At this point, basic functionality like logging-in and issues should be working again, but not wikis (because they are not restored yet). Note that it's normal to see a 502 error message ("Whoops, GitLab is taking too much time to respond.") when GitLab restarts: it takes a long time to start (think minutes)... You can follow its progress in /var/log/gitlab/gitlab-rails/*.log.

Be warned that the new server will start sending email notifications, for example for issues with an due date, which might be confusing for users if this is a development server. If this is a production server, that's a good thing. If it's a development server, you may want to disable email altogether in the GitLab server, with this line in hiera/roles/gitlab_dev.yml in the tor-puppet.git repository:

profile::gitlab::app::email_enabled: false

So the above procedure only restores a part of the system, namely what is covered by the nightly backup job. To restore the rest (at the time of writing: artifacts and repositories, which includes wikis!), you also need to specifically restore those files from the backup server.

For example, this procedure will restore the repositories from the backup server:

    $ cd /var/opt/gitlab/git-data
    cwd is: /var/opt/gitlab
    $ mark repositories
    113,766 files marked.
    $ done

The files will then end up in /var/tmp/bacula-restores/var/opt/gitlab/git-data. They will need to be given to the right users and moved into place:

chown -R git:root /var/tmp/bacula-restores/var/opt/gitlab/git-data/repositories
mv /var/opt/gitlab/git-data/repositories /var/opt/gitlab/git-data/repositories.orig
mv /var/tmp/bacula-restores/var/opt/gitlab/git-data/repositories /var/opt/gitlab/git-data/repositories/

During the last simulation, restoring repositories took an hour.

Restoring artifacts is similar:

$ cd /srv/gitlab-shared
cwd is: /srv/gitlab-shared/
$ mark artifacts
434,788 files marked.
$ done

Then the files need to be given and moved as well, notice the git:git instead of git:root:

chown -R git:git /var/tmp/bacula-restores/srv/gitlab-shared/artifacts
mv /var/opt/gitlab/gitlab-rails/shared/artifacts/ /var/opt/gitlab/gitlab-rails/shared/artifacts.orig
mv /var/tmp/bacula-restores/srv/gitlab-shared/artifacts /var/opt/gitlab/gitlab-rails/shared/artifacts/

Restoring the artifacts took another hour of copying.

And that's it! Note that this procedure may vary if the subset of files backed up by the GitLab backup job changes.

Reference

Installation

Main GitLab installation

The current GitLab server was setup in the howto/ganeti cluster in a regular virtual machine. It was configured with howto/puppet with the roles::gitlab. That, in turn, includes a series of profile classes which configure:

  • profile::gitlab::web: nginx vhost and TLS cert, which depends on profile::nginx built for the howto/cache service and relying on the puppet/nginx module from the Forge
  • profile::gitlab::app: the core of the configuration of gitlab itself, uses the puppet/gitlab module from the Forge, with Prometheus, Grafana, and Nginx support disabled, but Redis, PostgreSQL, and other exporters enabled
  • profile::dovecot::private: a simple IMAP server to receive mails destined to GitLab

This installs the GitLab Omnibus distribution which duplicates a lot of resources we would otherwise manage elsewhere in Puppet, including (but possibly not limited to):

This therefore leads to a "particular" situation regarding monitoring and PostgreSQL backups, in particular. See issue 20 for details.

The install takes a long time to complete. It's going to take a few minutes to download, unpack, and configure GitLab. There's no precise timing of this procedure yet, but assume each of those steps takes about 2 to 5 minutes.

After the install, the administrator account details are stored in /etc/gitlab/initial_root_password. After logging in, you most likely want to disable new signups as recommended, or possibly restore from backups.

Note that the first gitlab server (gitlab-01) was setup using the Ansible recipes used by the Debian.org project. That install was not working so well (e.g. 503 errors on merge requests) so we migrated to the omnibus package in March 2020, which seems to work better. There might still be some leftovers of that configuration here and there, but some effort was done during the 2022 hackweek (2022-06-28) to clean that up in Puppet at least. See tpo/tpa/gitlab#127 for some of that cleanup work.

GitLab CI installation

See the CI documentation for documentation specific to GitLab CI.

GitLab pages installation

To setup GitLab pages, we followed the GitLab Pages administration manual. The steps taken were as follows:

  1. add pages.torproject.net to the public suffix list (issue 40121 and upstream PR) (although that takes months or years to propagate everywhere)
  2. add *.pages.torproject.net and pages.torproject.net to DNS (dns/domains.git repository), as A records so that LE DNS-01 challenges still work, along with a CAA record to allow the wildcard on pages.torproject.net
  3. get the wildcard cert from Let's Encrypt (in letsencrypt-domains.git)
  4. deploy the TLS certificate, some GitLab config and a nginx vhost to gitlab-02 with Puppet
  5. run the status-site pipeline to regenerate the pages

The GitLab pages configuration lives in the profile::gitlab::app Puppet class. The following GitLab settings were added:

gitlab_pages             => {
  ssl_certificate     => '/etc/ssl/torproject/certs/pages.torproject.net.crt-chained',
  ssl_certificate_key => '/etc/ssl/private/pages.torproject.net.key',
},
pages_external_url       => 'https://pages.torproject.net',

The virtual host for the pages.torproject.net domain was configured through the profile::gitlab::web class.

SLA

Design

GitLab is a fairly large program with multiple components. The upstream documentation has a good details of the architecture but this section aims at providing a shorter summary. Here's an overview diagram, first:

GitLab's architecture diagram

Note: the above image may be broken as upstream frequently changes the URL. It should be visible in the Simplified component overview of the architecture documentation.

The web frontend is Nginx (which we incidentally also use in our howto/cache system) but GitLab wrote their own reverse proxy called GitLab Workhorse which in turn talks to the underlying GitLab Rails application, served by the Unicorn application server. The Rails app stores its data in a howto/postgresql database (although not our own deployment, for now, should be fixed). GitLab also offloads long-term background tasks to a tool called sidekiq.

Those all server HTTP(S) requests but GitLab is of course also accessible over SSH to push/pull git repositories. This is handled by a separate component called gitlab-shell which acts as a shell for the git user.

Workhorse, Rails, sidekiq and gitlab-shell all talk with Redis to store temporary information, caches and session information. They can also communicate with the Gitaly server which handles all communication with the git repositories themselves.

Continuous integration

GitLab also features Continuous Integration (CI). CI is handled by GitLab runners which can be deployed by anyone and registered in the Rails app to pull CI jobs. This is documented in the service/ci page.

Spam control

TODO: document lobby.

Discuss alternatives, e.g. this hackernews discussion about mediawiki moving to gitlab. Their gitlab migration documentation might give us hints on how to improve the spam situation on our end.

A few ideas on tools:

Scalability

We have not looked a lot into GitLab scalability. Upstream has reference architectures which explain how to scale for various user sizes. We have not yet looked into this, and so far have just thrown hardware at GitLab when performance issues come up.

GitLab pages

GitLab pages is "a simple HTTP server written in Go, made to serve GitLab Pages with CNAMEs and SNI using HTTP/HTTP2". In practice, the way this works is that artifacts from GitLab CI jobs get sent back to the central server.

GitLab pages is designed to scale horizontally: multiple pages servers can be deployed and fetch their content and configuration through NFS. They are rearchitecturing this with Object storage (ie. S3 through minio by default, or external existing providers) which might simplify running this but this actually adds complexity to a previously fairly simple design. Note that they have tried using CephFS instead of NFS but that did not work for some reason.

The new pages architecture also relies on the GitLab rails API for configuration (it was a set of JSON files before), which makes it dependent on the Rails API for availability, although that part of the design has exponential back-off time for unavailability of the rails API, so maybe it would survive a downtime of the rails API.

GitLab pages is not currently in use in our setup, but could be used as an alternative to the static mirroring system. See the discussion there for more information about how that compares with the static mirror system.

Update: some tests of GitLab pages were performed in January 2021, with moderate success. There are still concerns about the reliability and scalability of the service, but the service could be used for small sites at this stage. See the GitLab pages installation instructions for details on how this was setup.

Note that the pages are actually on disk, in /var/opt/gitlab/gitlab-rails/shared/pages/GROUP/.../PROJECT, for example the status site pipeline publishes to:

/var/opt/gitlab/gitlab-rails/shared/pages/tpo/tpa/status-site/

Maybe this could be abused to act as a static source in the static mirror system?

Update: see service/static-shim for the chosen solution to deploy websites built in GitLab CI to the static mirror system.

Issues

File or search for issues in the gitlab project.

Known

See also issues YOU have voted on.

Resolved

Monitoring and testing

Monitoring right now is minimal: normal host-level metrics like disk space, CPU usage, web port and TLS certificates are monitored by Nagios with our normal infrastructure, as a black box.

Prometheus monitoring is built into the GitLab Omnibus package, so it is not configured through our Puppet like other Prometheus servers. It has still been (manually) integrated in our Prometheus setup and Grafana dashboards (see pager playbook) have been deployed.

More work is underway to improve monitoring in this issue (not hardcoding exporters). We could also use the following tools:

There is a development server called gitlab-dev-01 that can be used to test dangerous things if there is a concern a change could break the production server.

Logs and metrics

GitLab keeps an extensive (excessive?) amount of logs, in /var/log/gitlab, which includes PII, including IP addresses.

To see live logs, you can type the handy command:

gitlab-ctl tail

... but that is sort of like drinking from a fire hose. You can inspect the logs of a specific component by passing it as an argument, for example to inspect the mail importer:

gitlab-ctl tail mailroom

Each component is in his own directory, so the equivalent to the above is:

tail -f /var/log/gitlab/mailroom/{current,mail_room_json.log}

Notice how both regular and JSON logs are kept.

Logs seem to be kept for a month.

Backups

There is a backup job ( tpo-gitlab-backup, in the root user crontab) that is a simple wrapper script which calls gitlab-backup to dump some components of the GitLab installation in the backup directory (/srv/gitlab-backup).

The backup system is deployed by Puppet and (at the time of writing!) skips repositories* and artifacts. It contains:

  • GitLab CI build logs (builds.tar.gz)
  • a compressed database dump (db/database.sql.gz)
  • Git Large Files (Git LFS, lfs.tar.gz)
  • packages (packages.tar.gz)
  • GitLab pages (pages.tar.gz)
  • some terraform thing (terraform_state.tar.gz)
  • uploaded files (uploads.tar.gz)

The backup job is ran nightly. GitLab also creates a backup on upgrade. Those are purged after two weeks by the wrapper script.

The backup job does NOT contain those components because they take up a tremendous amount of disk space, and are already backed up by Bacula. Those need to be restored from the regular backup server, separately:

  • Git repositories (found in /var/opt/gitlab/git-data/repositories/)
  • GitLab CI artifacts (normally found in /var/opt/gitlab/gitlab-rails/shared/artifacts/, in our case bind-mounted over /srv/gitlab-shared/artifacts)

It is assumed that the existing howto/backup system will pick up those files, but also the actual backup files in /srv/gitlab-backup and store them for our normal rotation periods.

This implies that the files covered by the gitlab-backup job are also already backed up by Bacula and are therefore duplicated on the backup storage server. See issue 40518 to review that strategy.

Ideally, this rather exotic backup system would be harmonized with our existing backup system, but this would require (for example) using our existing PostgreSQL infrastructure (issue 20). Other ideas (including filesystem snapshots) are also in issue 40518.

Other documentation

Discussion

Meetings

Some meetings about tools discussed GitLab explicitly. Those are the minutes:

Overview

The GitLab project at Tor has been a long time coming. If you look at the history section above, you'll see it has been worked on since at least 2016, at which point an external server was setup for the "network team" to do code review. This server was ultimately retired.

The current server has been worked on since 2019, with the master ticket, issue 29400, created in the footsteps of the 2019 Brussels meeting. The service launched some time in June 2020, with a full migration of Trac tickets.

Goals

Must have

  • replacement of the Trac issue tracking server
  • rough equivalent of Trac features in GitLab

Nice to have

  • identical representation of Trac issues in GitLab, including proper issue numbering

Non-Goals

  • replacement of Gitolite (git hosting)
  • replacement of Gitweb (git hosting)
  • replacement of Jenkins (CI) -- although that was eventually done
  • replacement of the static site hosting system

Those are not part of the first phase of the project, but it is understood that if one of those features gets used more heavily in GitLab, the original service MUST be eventually migrated into GitLab and turned off. We do not want to run multiple similar services at the same time (for example run both gitolite and gitaly on all git repositories, or run Jenkins and GitLab runners).

Approvals required

The GitLab migration was approved at the 2019 Brussels dev meeting.

Proposed Solution

The solution to the "code review" and "project management" problems are to deploy a GitLab instance which does not aim at managing all source code, in the first stage.

Cost

Staff not evaluated.

In terms of hardware, we start with a single virtual machine and agree that, in the worst case, we can throw a full Hetzner PX62-NVMe node at the problem (~70EUR/mth).

Alternatives considered

GitLab is such a broad project that multiple alternatives exist for different components:

  • GitHub
    • Pros:
      • widely used in the open source community
      • Good integration between ticketing system and code
    • Cons
      • It is hosted by a third party (Microsoft!)
      • Closed source
  • GitLab:
  • Pros:
    • Mostly free software
    • Feature-rich
  • Cons:
    • Complex software, high maintenance
    • "Opencore" - some interesting features are closed-source

GitLab command line clients

If you want to do batch operations or integrations with GitLab, you might want to use one of those tools, depending on your environment or prefered programming language:

GitLab upstream has a list of third-party commandline tools that is interesting as well.

Migration tools

ahf implemente the gitlab using his own home-made tools that talk to the GitLab and Trac API. but there's also tracboat which is designed to migrate from trac to GitLab.

We did not use Tracboat because it uses gitlab's DB directly and thus only works with some very specific version. Each time the database schema changes at GitLab, Tracboat needs to port to it. We prefered to use something that talked with the GitLab API.

We also didn't like the output entirely, so we modified it but still used some of its regular expressions and parser.

We also needed to implement the "ticket movement" hack (with the legacy project) which wasn't implemented in Tracboat.

Finally, we didn't want to do complete user migration, but lazily transfer only some users.

Git repository integrity solutions

This section is a summary of the discussion in ticket tpo/tpa/gitlab#81. A broader discussion of the security issues with GitLab vs Gitolite and the choices made during that migration are available in Gitolite: security concerns.

Some developers expressed concerns about using GitLab as a canonical location for source code repositories, mainly because of the much broader attack surface GitLab provides, compared to the legacy, gitolite-based infrastructure, especially considering that the web application basically has write access to everything.

One solution to this problem is to use cryptographic signatures. We already use OpenPGP extensively in the Tor infrastructure, and it's well integrated in git, so it's an obvious candidate. But it's not necessarily obvious how OpenPGP would be used to sign code inside Tor, so this section provides a short review of existing solutions in that space.

Guix: sign all commits

Guix uses OpenPGP to sign commits, using an approach that is basically:

  1. The repository contains a .guix-authorizations file that lists the OpenPGP key fingerprints of authorized committers.
  2. A commit is considered authentic if and only if it is signed by one of the keys listed in the .guix-authorizations file of each of its parents. This is the authorization invariant.

[...] Since .guix-authorizations is a regular file under version control, granting or revoking commit authorization does not require special support.

Note the big caveat:

It has one downside: it prevents pull-request-style workflows. Indeed, merging the branch of a contributor not listed in .guix-authorizations would break the authorization invariant. It’s a good tradeoff for Guix because our workflow relies on patches carved into stone tablets (patch tracker), but it’s not suitable for every project out there.

Also note there's a bootstrapping problem in their design:

Which commit do we pick as the first one where we can start verifying the authorization invariant?

They solve this with an out of band "channel introduction" mechanism which declares a good hash and a signing key.

This also requires a custom client. But it serves as a good example of an extreme approach (validate everything) one could take.

Arista: sign all commits in Gerrit

Arista wrote a blog post called Commit Signing with Git at Enterprise Scale (archive) which takes a radically different approach.

  • all OpenPGP keys are centrally managed (which solves the "web of trust" mess) in a Vault
  • Gerrit is the gatekeeper: for patches to be merged, they must be signed by a trusted key

It is a rather obtuse system: because the final patches are rebased on top of the history, the git signatures are actually lost so they have a system to keep a reference to the Gerrit change id in the git history, which does have a copy of the OpenPGP signature.

Gerwitz: sign all commits or at least merge commits

Mike Gerwitz wrote an article in 2012 (which he warns is out of date) but which already correctly identified the issues with merge and rebase workflows. He argues there is a way to implement the desired workflow by signing merges: because maintainers are the one committing merge requests to the tree, they are in a position to actually sign the code provided by third-parties. Therefore it can be assume that if a merge commit is signed, then the code it imported is also signed.

The article also provides a crude checking script for such a scenario.

Obviously, in the case of GitLab, it would make the "merge" button less useful, as it would break the trust chain. But it's possible to merge "out of band" (in a local checkout) and push the result, which GitLab generally correctly detect as closing the merge request.

Torwalds: signed tags

Linus Torwalds, the original author and maintainer of the Linux kernel, simply signs the release tags. In an article called "what does a pgp signature on a git commit prove?", Konstantin Ryabitsev (the kernel.org sysadmin), provides a good primer on OpenPGP signing in git. It also shows how to validate Linux releases by checking the tag and argues this is sufficient to ensure trust.

Vick: git signatures AKA git notes

The git-signatures project, authored by Lance R. Vick, makes it possible to "attach an arbitrary number of GPG signatures to a given commit or tag.":

Git already supports commit signing. These tools are intended to compliment that support by allowing a code reviewer and/or release engineer attach their signatures as well.

Downside: third-party tool not distributed with git and not packaged in Debian.

The idea of using git-notes was also proposed by Owen Jacobsen.

Walters: extended validation tags

The git-evtag projects from Colin Walters tries to address the perceived vulnerability of the SHA-1 hash by implementing a new signing procedure for tags, based on SHA-512 and OpenPGP.

Ryabitsev: b4 and patch attestations

Konstantin Ryabitsev (the kernel.org sysadmin, again) proposed a new cryptographic scheme to sign patches in Linux, he called "patch attestation". The protocol is designed to survive mailing list transports, rebases and all sorts of mangling. It does not use GnuPG and is based on a Trust On First Use (TOFU) model.

The model is not without critics.

Update, 2021-06-04: there was another iteration of that concept, this time based on DKIM-like headers, with support for OpenPGP signatures but also "native" ed25519.

One key takeaway from this approach, which we could reuse, is the way public keys are stored. In patatt, the git repository itself holds the public keys:

On the other hand, within the context of git repositories, we already have a suitable mechanism for distributing developer public keys, which is the repository itself. Consider this:

  • git is already decentralized and can be mirrored to multiple locations, avoiding any single points of failure
  • all contents are already versioned and key additions/removals can be audited and “git blame’d”
  • git commits themselves can be cryptographically signed, which allows a small subset of developers to act as “trusted introducers” to many other contributors (mimicking the “keysigning” process)

The idea of using git itself for keyring management was originally suggested by the did:git project, though we do not currently implement the proposed standard itself.

<https://github.com/dhuseby/did-git-spec/blob/master/did-git-spec.md>

It's unclear, however, why the latter spec wasn't reused. To be investigated.