• Supported object storage providers
  • Configure a single storage connection for all object types (consolidated form)
  • Configure each object type to define its own storage connection (storage-specific form)
  • Configure the connection settings
  • Full example using the consolidated form and Amazon S3
  • Migrate to object storage
  • Transition to consolidated form
  • Migrate objects to a different object storage provider
  • Alternatives to file system storage
  • Troubleshooting
  • Object storage

    GitLab supports using an object storage service for holding numerous types of data. It’s recommended over NFS and in general it’s better in larger setups as object storage is typically much more performant, reliable, and scalable.

    To configure the object storage, you have two options:

    If you store data locally, see how to migrate to object storage .

    Supported object storage providers

    GitLab is tightly integrated with the Fog library, so you can see which providers can be used with GitLab.

    Specifically, GitLab has been tested by vendors and customers on a number of object storage providers:

    Configure a single storage connection for all object types (consolidated form)

    Introduced in GitLab 13.2.

    Most types of objects, such as CI artifacts, LFS files, and upload attachments can be saved in object storage by specifying a single credential for object storage with multiple buckets.

    Configuring the object storage using the consolidated form has a number of advantages:

    When the consolidated form is used, direct upload is enabled automatically. Thus, only the following providers can be used:

    The consolidated form configuration can’t be used for backups or Mattermost. Backups can be configured with server side encryption separately. See the table for a complete list of supported object storage types.

    Enabling the consolidated form enables object storage for all object types. If not all buckets are specified, you may see an error like:

    Object storage for <object type> must have a bucket specified
    

    If you want to use local storage for specific object types, you can disable object storage for specific features .

    Configure the common parameters

    In the consolidated form, the object_store section defines a common set of parameters.

    Setting Description
    enabled Enable or disable object storage.
    proxy_download Set to true to enable proxying all files served . Option allows to reduce egress traffic as this allows clients to download directly from remote storage instead of proxying all data.
    connection Various connection options described below.
    storage_options Options to use when saving new objects, such as server side encryption . Introduced in GitLab 13.3.
    objects Object-specific configuration .

    For an example, see how to use the consolidated form and Amazon S3 .

    Configure the parameters of each object

    Each object type must at least define the bucket name where it will be stored.

    The following table lists the valid objects that can be used:

    Type Description
    artifacts CI artifacts
    external_diffs Merge request diffs
    uploads User uploads
    lfs Git Large File Storage objects
    packages Project packages (for example, PyPI, Maven, or NuGet)
    dependency_proxy Dependency Proxy
    terraform_state Terraform state files
    pages Pages

    Within each object type, three parameters can be defined:

    Setting Required? Description
    bucket Yes* Bucket name for the object type. Not required if enabled is set to false .
    enabled No Overrides the common parameter .
    proxy_download No Overrides the common parameter .

    For an example, see how to use the consolidated form and Amazon S3 .

    Disable object storage for specific features

    As seen above, object storage can be disabled for specific types by setting the enabled flag to false . For example, to disable object storage for CI artifacts:

    A bucket is not needed if the feature is disabled entirely. For example, no bucket is needed if CI artifacts are disabled with this setting:

    Configure each object type to define its own storage connection (storage-specific form)

    With the storage-specific form, every object defines its own object storage connection and configuration. If you’re using GitLab 13.2 and later, you should transition to the consolidated form .

    The use of encrypted S3 buckets with non-consolidated form is not supported. You may get ETag mismatch errors if you use it.

    note
    For the storage-specific form, direct upload may become the default because it does not require a shared folder.

    For configuring object storage in GitLab 13.1 and earlier, or for storage types not supported by consolidated form, refer to the following guides:

    Container Registry (optional feature) Autoscale runner caching (optional for improved performance) Job artifacts including archived job logs Packages (optional feature) Dependency Proxy (optional feature)
    Object storage type Supported by consolidated form?
    Backups No
    No
    Mattermost No
    No
    Yes
    LFS objects Yes
    Uploads Yes
    Merge request diffs Yes
    Yes
    Yes
    Terraform state files Yes
    Pages content Yes

    Configure the connection settings

    Both consolidated and storage-specific form must configure a connection. The following sections describe parameters that can be used in the connection setting.

    Amazon S3

    The connection settings match those provided by fog-aws :

    Setting Description Default
    provider Always AWS for compatible hosts. AWS
    aws_access_key_id AWS credentials, or compatible.
    aws_secret_access_key AWS credentials, or compatible.
    aws_signature_version AWS signature version to use. 2 or 4 are valid options. Digital Ocean Spaces and other providers may need 2 . 4
    enable_signature_v4_streaming Set to true to enable HTTP chunked transfers with AWS v4 signatures . Oracle Cloud S3 needs this to be false . true
    region AWS region.
    host DEPRECATED: Use endpoint instead. S3 compatible host for when not using AWS. For example, localhost or storage.example.com . HTTPS and port 443 is assumed. s3.amazonaws.com
    endpoint Can be used when configuring an S3 compatible service such as MinIO , by entering a URL such as http://127.0.0.1:9000 . This takes precedence over host . Always use endpoint for consolidated form. (optional)
    path_style Set to true to use host/bucket_name/object style paths instead of bucket_name.host/object . Set to true for using MinIO . Leave as false for AWS S3. false .
    use_iam_profile Set to true to use IAM profile instead of access keys. false
    aws_credentials_refresh_threshold_seconds Sets the automatic refresh threshold in seconds when using temporary credentials in IAM. 15

    Use Amazon instance profiles

    Instead of supplying AWS access and secret keys in object storage configuration, you can configure GitLab to use Amazon Identity Access and Management (IAM) roles to set up an Amazon instance profile . When this is used, GitLab fetches temporary credentials each time an S3 bucket is accessed, so no hard-coded values are needed in the configuration.

    Prerequisites:

    To set up an instance profile:

      Create an IAM role with the necessary permissions. The following example is a role for an S3 bucket named test-bucket :

      "Version" : "2012-10-17" , "Statement" : [ "Sid" : "VisualEditor0" , "Effect" : "Allow" , "Action" : [ "s3:PutObject" , "s3:GetObject" , "s3:DeleteObject" "Resource" : "arn:aws:s3:::test-bucket/*" Attach this role to the EC2 instance hosting your GitLab instance.
    1. Set the use_iam_profile GitLab configuration option to true .

    Encrypted S3 buckets

    Version history Introduced in GitLab 13.1 for instance profiles only and S3 default encryption .
  • Introduced in GitLab 13.2 for static credentials when the consolidated form and S3 default encryption is used.
  • When configured either with an instance profile or with the consolidated form, GitLab Workhorse properly uploads files to S3 buckets that have SSE-S3 or SSE-KMS encryption enabled by default . AWS KMS keys and SSE-C encryption are not supported since this requires sending the encryption keys in every request .

    Server-side encryption headers

    Introduced in GitLab 13.3.

    Setting a default encryption on an S3 bucket is the easiest way to enable encryption, but you may want to set a bucket policy to ensure only encrypted objects are uploaded . To do this, you must configure GitLab to send the proper encryption headers in the storage_options configuration section:

    Setting Description
    server_side_encryption Encryption mode ( AES256 or aws:kms ).
    server_side_encryption_kms_key_id Amazon Resource Name. Only needed when aws:kms is used in server_side_encryption . See the Amazon documentation on using KMS encryption .

    As with the case for default encryption, these options only work when the Workhorse S3 client is enabled. One of the following two conditions must be fulfilled:

      use_iam_profile is true in the connection settings.
    • Consolidated form is in use.

    ETag mismatch errors occur if server side encryption headers are used without enabling the Workhorse S3 client.

    Oracle Cloud S3

    Oracle Cloud S3 must be sure to use the following settings:

    Setting Value
    enable_signature_v4_streaming false
    path_style true

    If enable_signature_v4_streaming is set to true , you may see the following error in production.log :

    Google Cloud Storage (GCS)

    Here are the valid connection parameters for GCS:

    Setting Description Example
    provider Provider name. Google
    google_project GCP project name. gcp-project-12345
    google_json_key_location JSON key path. /path/to/gcp-project-12345-abcde.json
    google_json_key_string JSON key string. { "type": "service_account", "project_id": "example-project-382839", ... }
    google_application_default Set to true to use Google Cloud Application Default Credentials to locate service account credentials.

    GitLab reads the value of google_json_key_location , then google_json_key_string , and finally, google_application_default . It uses the first of these settings that has a value.

    The service account must have permission to access the bucket. For more information, see the Cloud Storage authentication documentation .

    note
    Bucket encryption with the Cloud Key Management Service (KMS) is not supported and results in ETag mismatch errors .

    GCS example

    For Omnibus installations, this is an example of the connection setting in the consolidated form:

    GCS example with ADC

    Introduced in GitLab 13.6.

    Google Cloud Application Default Credentials (ADC) are typically used with GitLab to use the default service account. This eliminates the need to supply credentials for the instance. For example, in the consolidated form:

    gitlab_rails['object_store']['connection'] = {
      'provider' => 'Google',
      'google_project' => '<GOOGLE PROJECT>',
      'google_application_default' => true
    

    If you use ADC, be sure that:

    Azure Blob storage

    Introduced in GitLab 13.4.

    Although Azure uses the word container to denote a collection of blobs, GitLab standardizes on the term bucket . Be sure to configure Azure container names in the bucket settings.

    Azure Blob storage can only be used with the consolidated form because a single set of credentials are used to access multiple containers. The storage-specific form is not supported. For more details, see how to transition to consolidated form .

    The following are the valid connection parameters for Azure. For more information, see the Azure Blob Storage documentation .

    Setting Description Example
    provider Provider name. AzureRM
    azure_storage_account_name Name of the Azure Blob Storage account used to access the storage. azuretest
    azure_storage_access_key Storage account access key used to access the container. This is typically a secret, 512-bit encryption key encoded in base64. czV2OHkvQj9FKEgrTWJRZVRoV21ZcTN0Nnc5eiRDJkYpSkBOY1JmVWpYbjJy\nNHU3eCFBJUQqRy1LYVBkU2dWaw==\n
    azure_storage_domain Domain name used to contact the Azure Blob Storage API (optional). Defaults to blob.core.windows.net . Set this if you are using Azure China, Azure Germany, Azure US Government, or some other custom Azure domain. blob.core.windows.net

      For Omnibus installations, this is an example of the connection setting in the consolidated form:

      gitlab_rails['object_store']['connection'] = {
        'provider' => 'AzureRM',
        'azure_storage_account_name' => '<AZURE STORAGE ACCOUNT NAME>',
        'azure_storage_access_key' => '<AZURE STORAGE ACCESS KEY>',
        'azure_storage_domain' => '<AZURE STORAGE DOMAIN>'
      

      For source installations, Workhorse also needs to be configured with Azure credentials. This isn’t needed in Omnibus installs, because the Workhorse settings are populated from the previous settings.

        Edit /home/git/gitlab-workhorse/config.toml and add or amend the following lines:

        [object_storage]
          provider = "AzureRM"
        [object_storage.azurerm]
          azure_storage_account_name = "<AZURE STORAGE ACCOUNT NAME>"
          azure_storage_access_key = "<AZURE STORAGE ACCESS KEY>"
        

      If you are using a custom Azure storage domain, azure_storage_domain does not have to be set in the Workhorse configuration. This information is exchanged in an API call between GitLab Rails and Workhorse.

    Storj Gateway (SJ) note
    The Storj Gateway does not support multi-threaded copying (see UploadPartCopy in the table). While an implementation is planned , you must disable multi-threaded copying until completion.

    The Storj Network provides an S3-compatible API gateway. Use the following configuration example:

    gitlab_rails['object_store']['connection'] = {
      'provider' => 'AWS',
      'endpoint' => 'https://gateway.storjshare.io',
      'path_style' => true,
      'region' => 'eu1',
      'aws_access_key_id' => 'ACCESS_KEY'
    
    
    
    
        
    ,
      'aws_secret_access_key' => 'SECRET_KEY',
      'aws_signature_version' => 2,
      'enable_signature_v4_streaming' => false
    

    The signature version must be 2 . Using v4 results in a HTTP 411 Length Required error. For more information, see issue #4419 .

    Full example using the consolidated form and Amazon S3

    The following example uses AWS S3 to enable object storage for all supported services:

    Linux package (Omnibus)

      Edit /etc/gitlab/gitlab.rb and add the following lines, substituting the values you want:

      # Consolidated object storage configuration
      gitlab_rails['object_store']['enabled'] = true
      gitlab_rails['object_store']['proxy_download'] = true
      gitlab_rails['object_store']['connection'] = {
        'provider' => 'AWS',
        'region' => 'eu-central-1',
        'aws_access_key_id' => '<AWS_ACCESS_KEY_ID>',
        'aws_secret_access_key' => '<AWS_SECRET_ACCESS_KEY>'
      # OPTIONAL: The following lines are only needed if server side encryption is required
      gitlab_rails['object_store']['storage_options'] = {
        'server_side_encryption' => '<AES256 or aws:kms>',
        'server_side_encryption_kms_key_id' => '<arn:aws:kms:xxx>'
      gitlab_rails['object_store']['objects']['artifacts']['bucket'] = 'gitlab-artifacts'
      gitlab_rails['object_store']['objects']['external_diffs']['bucket'] = 'gitlab-mr-diffs'
      gitlab_rails['object_store']['objects']['lfs']['bucket'] = 'gitlab-lfs'
      gitlab_rails['object_store']['objects']['uploads']['bucket'] = 'gitlab-uploads'
      gitlab_rails['object_store']['objects']['packages']['bucket'] = 'gitlab-packages'
      gitlab_rails['object_store']['objects']['dependency_proxy']['bucket'] = 'gitlab-dependency-proxy'
      gitlab_rails['object_store']['objects']['terraform_state']['bucket'] = 'gitlab-terraform-state'
      gitlab_rails['object_store']['objects']['ci_secure_files']['bucket'] = 'gitlab-ci-secure-files'
      gitlab_rails['object_store']['objects']['pages']['bucket'] = 'gitlab-pages'
      

      If you’re using AWS IAM profiles , omit the AWS access key and secret access key/value pairs. For example:

      gitlab_rails['object_store']['connection'] = {
        'provider' => 'AWS',
        'region' => 'eu-central-1',
        'use_iam_profile' => true
      

      Save the file and reconfigure GitLab:

      sudo gitlab-ctl reconfigure
      
    Helm chart (Kubernetes)

      Put the following content in a file named object_storage.yaml to be used as a Kubernetes Secret :

      provider: AWS
      region: us-east-1
      aws_access_key_id: <AWS_ACCESS_KEY_ID>
      aws_secret_access_key: <AWS_SECRET_ACCESS_KEY>
      

      If you’re using AWS IAM profiles , omit the AWS access key and secret access key/value pairs. For example:

      provider: AWS
      region: us-east-1
      use_iam_profile: true
      

      Create the Kubernetes Secret:

      kubectl create secret generic -n <namespace> gitlab-object-storage --from-file=connection=object_storage.yaml
      

      Export the Helm values:

      helm get values gitlab > gitlab_values.yaml
      

      Edit gitlab_values.yaml:

      global:
        appConfig:
          object_store:
            enabled: false
            proxy_download: true
            storage_options: {}
              # server_side_encryption:
              # server_side_encryption_kms_key_id
            connection:
              secret: gitlab-object-storage
          lfs:
            enabled: true
            proxy_download: true
            bucket: gitlab-lfs
            connection: {}
              # secret:
              # key:
          artifacts:
            enabled: true
            proxy_download: true
            bucket: gitlab-artifacts
            connection: {}
              # secret:
              # key:
          uploads:
            enabled: true
            proxy_download: true
            bucket: gitlab-uploads
            connection: {}
              # secret:
              # key:
          packages:
            enabled: true
            proxy_download: true
            bucket: gitlab-packages
            connection: {}
          externalDiffs:
            enabled: true
            when:
            proxy_download: true
            bucket: gitlab-mr-diffs
            connection: {}
          terraformState:
            enabled: true
            bucket: gitlab-terraform-state
            connection: {}
          ciSecureFiles:
            enabled: true
            bucket: gitlab-ci-secure-files
            connection: {}
          dependencyProxy:
            enabled: true
            proxy_download: true
            bucket: gitlab-dependency-proxy
            connection: {}
      

      Save the file and apply the new values:

      helm upgrade -f gitlab_values.yaml gitlab gitlab/gitlab
      
    Docker

      Edit docker-compose.yml :

      version: "3.6"
      services:
        gitlab:
          environment:
            GITLAB_OMNIBUS_CONFIG: |
              # Consolidated object storage configuration
              gitlab_rails['object_store']['enabled'] = true
              gitlab_rails['object_store']['proxy_download'] = true
              gitlab_rails['object_store']['connection'] = {
                'provider' => 'AWS',
                'region' => 'eu-central-1',
                'aws_access_key_id' => '<AWS_ACCESS_KEY_ID>',
                'aws_secret_access_key' => '<AWS_SECRET_ACCESS_KEY>'
              # OPTIONAL: The following lines are only needed if server side encryption is required
              gitlab_rails['object_store']['storage_options'] = {
                'server_side_encryption' => '<AES256 or aws:kms>',
                'server_side_encryption_kms_key_id' => '<arn:aws:kms:xxx>'
              gitlab_rails['object_store']['objects']['artifacts']['bucket'] = 'gitlab-artifacts'
              gitlab_rails['object_store']['objects']['external_diffs']['bucket'] = 'gitlab-mr-diffs'
              gitlab_rails['object_store']['objects']['lfs']['bucket'] = 'gitlab-lfs'
              gitlab_rails['object_store']['objects']['uploads']['bucket'] = 'gitlab-uploads'
              gitlab_rails['object_store']['objects']['packages']['bucket'] = 'gitlab-packages'
              gitlab_rails['object_store']['objects']['dependency_proxy']['bucket'] = 'gitlab-dependency-proxy'
              gitlab_rails['object_store']['objects']['terraform_state']['bucket'] = 'gitlab-terraform-state'
              gitlab_rails['object_store']['objects']['ci_secure_files']['bucket'] = 'gitlab-ci-secure-files'
              gitlab_rails['object_store']['objects']['pages']['bucket'] = 'gitlab-pages'
      

      If you’re using AWS IAM profiles , omit the AWS access key and secret access key/value pairs. For example:

      gitlab_rails['object_store']['connection'] = {
        'provider' => 'AWS',
        'region' => 'eu-central-1',
        'use_iam_profile' => true
      

      Save the file and restart GitLab:

      docker compose up -d
      
    Self-compiled (source)

      Edit /home/git/gitlab/config/gitlab.yml and add or amend the following lines:

      production: &base
        object_store:
          enabled: true
          proxy_download: true
          connection:
            provider: AWS
            aws_access_key_id: <AWS_ACCESS_KEY_ID>
            aws_secret_access_key: <AWS_SECRET_ACCESS_KEY>
            region: eu-central-1
          storage_options:
            server_side_encryption: <AES256 or aws:kms>
            server_side_encryption_key_kms_id: <arn:aws:kms:xxx>
          objects:
            artifacts:
              bucket: gitlab-artifacts
            external_diffs:
              bucket: gitlab-mr-diffs
            lfs:
              bucket: gitlab-lfs
            uploads:
              bucket: gitlab-uploads
            packages:
              bucket: gitlab-packages
            dependency_proxy:
              bucket: gitlab-dependency-proxy
            terraform_state:
              bucket: gitlab-terraform-state
            ci_secure_files:
              bucket: gitlab-ci-secure-files
            pages:
              bucket: gitlab-pages
      

      If you’re using AWS IAM profiles , omit the AWS access key and secret access key/value pairs. For example:

      connection:
        provider: AWS
        region: eu-central-1
        use_iam_profile: true
      

      Edit /home/git/gitlab-workhorse/config.toml and add or amend the following lines:

      [object_storage]
        provider = "AWS"
      [object_storage.s3]
        aws_access_key_id = "<AWS_ACCESS_KEY_ID>"
        aws_secret_access_key = "<AWS_SECRET_ACCESS_KEY>"
      

      If you’re using AWS IAM profiles, omit the AWS access key and secret access key/value pairs. For example:

      [object_storage.s3]
        use_iam_profile = true
      

      Save the file and restart GitLab:

      # For systems running systemd
      sudo systemctl restart gitlab.target
      # For systems running SysV init
      sudo service gitlab restart
      

    Migrate to object storage

    To migrate existing local data to object storage see the following guides:

    Transition to consolidated form

    Prior to GitLab 13.2:

    For example, an Omnibus GitLab install might have the following configuration:

    Although this provides flexibility in that it makes it possible for GitLab to store objects across different cloud providers, it also creates additional complexity and unnecessary redundancy. Since both GitLab Rails and Workhorse components need access to object storage, the consolidated form avoids excessive duplication of credentials.

    The consolidated form is used only if all lines from the original form is omitted. To move to the consolidated form, remove the original configuration (for example, artifacts_object_store_enabled , or uploads_object_store_connection )

    Migrate objects to a different object storage provider

    You may need to migrate GitLab data in object storage to a different object storage provider. The following steps show you how do this using Rclone .

    The steps assume you are moving the uploads bucket, but the same process works for other buckets.

    Prerequisites:

    • Choose the computer to run Rclone on. Depending on how much data you are migrating, Rclone may have to run for a long time so you should avoid using a laptop or desktop computer that can go into power saving. You can use your GitLab server to run Rclone.
      Install Rclone.
    1. Configure Rclone by running the following:

      rclone config
      

      The configuration process is interactive. Add at least two “remotes”: one for the object storage provider your data is currently on ( old ), and one for the provider you are moving to ( new ).

    2. Verify that you can read the old data. The following example refers to the uploads bucket , but your bucket may have a different name:

      rclone ls old:uploads | head
      

      This should print a partial list of the objects currently stored in your uploads bucket. If you get an error, or if the list is empty, go back and update your Rclone configuration using rclone config .

    3. Perform an initial copy. You do not need to take your GitLab server offline for this step.

      rclone sync -P old:uploads new:uploads
      
    4. After the first sync completes, use the web UI or command-line interface of your new object storage provider to verify that there are objects in the new bucket. If there are none, or if you encounter an error while running rclone sync , check your Rclone configuration and try again.

    After you have done at least one successful Rclone copy from the old location to the new location, schedule maintenance and take your GitLab server offline. During your maintenance window you must do two things:

    1. Perform a final rclone sync run, knowing that your users cannot add new objects so you do not leave any behind in the old bucket.
    2. Update the object storage configuration of your GitLab server to use the new provider for uploads .

    Alternatives to file system storage

    If you’re working to scale out your GitLab implementation, or add fault tolerance and redundancy, you may be looking at removing dependencies on block or network file systems. See the following additional guides:

    1. Make sure the git user home directory is on local disk.
    2. Configure database lookup of SSH keys to eliminate the need for a shared authorized_keys file.
    3. Prevent local disk usage for job logs .
    4. Disable Pages local storage .

    Troubleshooting

    Objects are not included in GitLab backups

    As noted in the backup documentation , objects are not included in GitLab backups. You can enable backups with your object storage provider instead.

    Use separate buckets

    Using separate buckets for each data type is the recommended approach for GitLab. This ensures there are no collisions across the various types of data GitLab stores. There are plans to enable the use of a single bucket in the future.

    With Omnibus and source installations it is possible to split a single real bucket into multiple virtual buckets. If your object storage bucket is called my-gitlab-objects you can configure uploads to go into my-gitlab-objects/uploads , artifacts into my-gitlab-objects/artifacts , etc. The application acts as if these are separate buckets. Use of bucket prefixes may not work correctly with Helm backups .

    Helm-based installs require separate buckets to handle backup restorations .

    S3 API compatibility issues

    Not all S3 providers are fully compatible with the Fog library that GitLab uses. Symptoms include an error in production.log :

    411 Length Required
    

    Proxy Download

    Clients can download files in object storage by receiving a pre-signed, time-limited URL, or by GitLab proxying the data from object storage to the client. Downloading files from object storage directly helps reduce the amount of egress traffic GitLab needs to process.

    When the files are stored on local block storage or NFS, GitLab has to act as a proxy. This is not the default behavior with object storage.

    The proxy_download setting controls this behavior: the default is generally false . Verify this in the documentation for each use case. Set it to true if you want GitLab to proxy the files.

    When not proxying files, GitLab returns an HTTP 302 redirect with a pre-signed, time-limited object storage URL . This can result in some of the following problems:

      If GitLab is using non-secure HTTP to access the object storage, clients may generate https->http downgrade errors and refuse to process the redirect. The solution to this is for GitLab to use HTTPS. LFS, for example, generates this error:

       LFS: lfsapi/client: refusing insecure redirect, https->http
      

      Clients need to trust the certificate authority that issued the object storage certificate, or may return common TLS errors such as:

       x509: certificate signed by unknown authority
      

      Clients need network access to the object storage. Network firewalls could block access. Errors that might result if this access is not in place include:

       Received status code 403 from server: Forbidden
      

      Object storage buckets need to allow Cross-Origin Resource Sharing (CORS) access from the URL of the GitLab instance. Attempting to load a PDF in the repository page may show the following error:

      An error occurred while loading the file. Please try again later.
      

      See the LFS documentation for more details.

    Additionally for a short time period users could share pre-signed, time-limited object storage URLs with others without authentication. Also bandwidth charges may be incurred between the object storage provider and the client.

    ETag mismatch

    Using the default GitLab settings, some object storage back-ends such as MinIO and Alibaba might generate ETag mismatch errors.

    If you are seeing this ETag mismatch error with Amazon Web Services S3, it’s likely this is due to encryption settings on your bucket . To fix this issue, you have two options:

    The first option is recommended for MinIO. Otherwise, the workaround for MinIO is to use the --compat parameter on the server.

    Without the consolidated form or instance profiles enabled, GitLab Workhorse uploads files to S3 using pre-signed URLs that do not have a Content-MD5 HTTP header computed for them. To ensure data is not corrupted, Workhorse checks that the MD5 hash of the data sent equals the ETag header returned from the S3 server. When encryption is enabled, this is not the case, which causes Workhorse to report an ETag mismatch error during an upload.

    When the consolidated form is:

    • Used with an S3-compatible object storage or an istance profile, Workhorse uses its internal S3 client which has S3 credentials so that it can compute the Content-MD5 header. This eliminates the need to compare ETag headers returned from the S3 server.
    • Not used with an S3-compatible object storage, Workhorse falls back to using pre-signed URLs.

    Encrypting buckets with the GCS Cloud Key Management Service (KMS) is not supported and results in ETag mismatch errors.

    Multi-threaded copying

    GitLab uses the S3 Upload Part Copy API to accelerate the copying of files within a bucket. Ceph S3 prior to Kraken 11.0.2 does not support this and returns a 404 error when files are copied during the upload process .

    The feature can be disabled using the :s3_multithreaded_uploads feature flag. To disable the feature, ask a GitLab administrator with Rails console access to run the following command:

    Feature.disable(:s3_multithreaded_uploads)
    

    Docs

    Edit this page to fix an error or add an improvement in a merge request.
    Create an issue to suggest an improvement to this page.

    Product

    Create an issue if there's something you don't like about this feature.
    Propose functionality by submitting a feature request.
    Join First Look to help shape new features.

    Feature availability and product trials

    View pricing to see all GitLab tiers and features, or to upgrade.
    Try GitLab for free with access to all features for 30 days.

    Get Help

    If you didn't find what you were looking for, search the docs.
    If you want help with something specific and could use community support, post on the GitLab forum.
    For problems setting up or using this feature (depending on your GitLab subscription).

    Request support