Skip to content

Advanced Configuration

The following documentation provides guidelines to configure various aspects of OctoPerf On Premise Infra, like:

  • System Settings: setup docker behind a proxy, Elasticsearch system configuration etc.,
  • Mailing: support resetting account password via email,
  • Storage: store resources (recorded requests/responses and more) either on local disk or Amazon S3,
  • High Availability: HA is supported via built-in Hazelcast clustering.

OctoPerf EE is a Spring Boot application which uses YML configuration files.

Docker Proxy Settings

Docker can run from behind a proxy if required. The procedure is detailed in Docker Documentation.


Elasticsearch Database requires the systemctl vm.max_map_count to be set at least to 262144:

sysctl -w vm.max_map_count=262144

To set this value permanently, update the vm.max_map_count setting in /etc/sysctl.conf. To verify after rebooting, run sysctl vm.max_map_count.

See Elasticsearch VM Max Map Count for more information.

Custom Application YML

To define your own configuration settings, you must edit the application.yml provided in config/ folder.

Custom Application YML

  1. Stop OctoPerf EE with the command make clean (data is preserved)
  2. Edit config/application.yml file provided in OctoPerf EE Setup,
  3. Start OctoPerf EE with the command make.

The configuration defined in your own application.yml takes precedence over the default built-in configuration.


Please refer to Spring boot documentation for general settings. (such as file upload limit)


The application.yml file must be indented with 2 spaces for each level.

Environment Variables

When you need to simply override a single setting, defining an environment variable when launching the OctoPerf EE container is the good way to go.

How to define an environment variable

  1. Edit docker-compose.yml provided in OctoPerf EE Setup,
  2. locate environment: section and add relevant environment variables,
  3. Restart the application using docker compose.

Environment variables take precedence over settings defined in YML configuration files.

Custom Settings

Advertised Server

The backend advertises itself using server settings:

  scheme: https
  insecure: false # true = support self-signed SSL Certificates
  port: 8090 # internal server port
    port: 443 # publicly exposed port

The server port is configurable via:

  • server.port: it defines the internal server port running inside the enterprise-edition container,
  • server.public.port: it defines the port advertised to agents and JMeter containers. If not defined, it falls back to server.port setting.


server.public.port setting is useful when the backend is behind a load-balancer. Typically load balancers run on port 80 or 443. For security reasons, the internal server cannot run port <= 1024. That's why we have both settings.

This must be configured to be accessible from all load generators. The JMeter containers launched on agent hosts by OctoPerf EE needs to be able to access to the backend through this hostname. The configuration above shows our Saas server configuration.


Never use localhost because it designates the container itself. This hostname must be accessible from the agents/users perspective.

Server Port

As the backend is behind Nginx by default, changing the default port (which is 80) requires a few steps. Let's say you want to run the server on port 443:

  • In nginx/default.conf, locate listen 80; and change it to 443,
  • In docker-compose.yml, locate 80:80 and replace by 443:443,
  • Set server.public.port environment setting to 443,
  • Stop and destroy containers by running docker-compose down,
  • Restart containers by running docker-compose --build -d.

The Nginx frontend port must always be the same as backend server port because the backend advertises this port to docker agents (to communicate).


After changing the port, you will have to reinstall all agents, otherwise they will fail to connect to the backend.

Internal Proxies

When the backend runs inside a Kubernetes cluster behind an ingress controller, you need to specify server.tomcat.remoteip.internal-proxies with the IPs of both internal ingress controllers (usually private ips like 10.42.xx.xx) and your ingress server public IPs.

Without this configuration, the remote IP of the users and agents performing requests to the backend server are incorrectly detected.

Here is an example configuration:

  shutdown: graceful # Finish Handling current requests before shutting down
  forward-headers-strategy: native # Use X-Forwarded-For HTTP header to resolve remote IPs
      internal-proxies: "10\\.\\d{1,3}\\.\\d{1,3}\\.\\d{1,3}" # Regexp: Ignore internal IPs like 10.xx.xx.xx, resolve remote IP instead

For an exhaustive list of available properties, see Common Application Properties.


General Settings

The backend connects to Elasticsearch database using elasticsearch.hostname setting:

  hostname: elasticsearch
  username: elastic # default username is "elastic"
  password: # empty by default; if none provided, no authentication is used
    buffer-limit: 104857600 # 100Mb by default
    prefix: octoperf_
    size: 256
    keep-alive-min: 15

Elasticsearch can be configured with:

  • hostname: This must be configured to an IP or a hostname accessible from the OctoPerf EE container. Multiple hosts can be defined separated by comas,
  • username: username to authenticate with Elasticsearch if minimal security is enabled. elastic is the default username,
  • password: password to authenticate with Elasticsearch. If empty, no authentication is used,
  • client:

    • buffer-limit: maximum response size the client can handle,
  • indices:

    • prefix: octoperf_ by default, indices are created with octoperf_ prefix. Example: octoperf_user indice contains the user accounts.
    • alias: none by default. Indices can be created with the given alias. This setting is useful to point to use aliases to upgrade a database with no service interruption. Make sure that existing indices already have same aliases configured before editing this setting (otherwise those settings won't be found by the backend).
  • scroll:

    • size: scroll window size, 256 elements by default,
    • keep-alive-min: scroll keep alive timeout between each window, 15 by default. In minutes.


Never use localhost hostname. Use the IP address of the machine hosting your database, even if it's on the same machine.

Snapshots Settings

Snapshots are the equivalent of an incremental backup of all the database indices. The backend can trigger an elasticsearch snapshot creation every night with the configuration below:

    enabled: true 
    repository: nfs-server # name of the repository to backup
    keep-last: 7 # 7 days rolling window

Container Orchestrator

Both docker (by default) and podman are supported:

    orchestrator: docker # or podman

It affects the command-line being generated to start the agent.

SMTP Server

an SMTP mail server can be specified in order to support user account password recovery:

    enabled: true
    port: 587
    password: passw0rd
      mail.smtp.auth: true
      mail.smtp.starttls.enable: true
      mail.smtp.starttls.required: true
      name: MyCompany Support

For a complete list of supported properties, see Javax Mail Documentation.

High Availability


  • You need to have at least 3 Hosts in a cluster with networking enabled between them.

To enable high-availability, the different backend servers need to form a cluster. Each backend server must be able to communicate with others. To enable HA using Hazelcast, define the following configuration:

  driver: hazelcast
  members: enterprise-edition
  quorum: 2

It takes the following configuration:

  • driver: either noop (no clustering), hazelcast (using Hazelcast) or ignite (using Apache Ignite),
  • members: comma separated list of hostnames or IPs of the members of the cluster,
  • quorum: minimum number of nodes necessary to operate. quorum = (n+1) / 2, where n is the total number of nodes in the cluster.

3 Containers Setup

Let's describe what's being configured here:

  • driver: hazelcast indicates to use Hazelcast clustering, Hazelcast is a Distributed Java Framework we use internally to achieve HA,
  • hazelcast.members: hostnames of all the backends separated by comas,
  • hazelcast.quorum: The quorum indicates the minimum number of machines within the cluster to be operational. quorum = (n + 1) / 2, where n is the number of members within the cluster.

hazelcast.members can be either a DNS hostname pointing to multiple IPs or multiple IPs separated by comas.

Resources Storage

Resources like recorded requests, recorded responses or files attached to a project are stored inside the container in folder /home/octoperf/data:

    max-file-count: 1024 # max number of files per project / result
    max-total-size: 1047741824 # max total bytes per project / result
  driver: fs
    folder: data
Folder location is resolved inside the Docker container

When no Docker volume mapping is configured, the data is lost when the container is destroyed. In order to avoid this, you need to setup a volume mapping like for the configuration. Example: /home/ubuntu/docker/enterprise-edition/data:/home/octoperf/data

When running the backend in HA, it's not possible to store data on the local disk. The request being sent to get resources may hit any backend server, but the wanted resource may be stored on another server. In this case, it's better to store resources on a global service shared by all backends like Amazon S3.

Configuring Amazon S3 Storage

  driver: s3
    region: eu-west-1
    bucket: my-bucket

Let's describe what is being configured here:

  • driver: Amazon S3 driver,
  • s3:

    • region: specifies the region where the target S3 bucket is located,
    • bucket: S3 bucket name, usually,
    • access-key: an AWS access key with access granted to the given bucket,
    • secret-key: an AWS secret key associated to the access key.

We suggest to setup an Amazon User with permission granted only to the target S3 bucket. Here is an example Policy:

    "Version": "2012-10-17",
    "Statement": [
            "Sid": "Stmt1427454857000",
            "Effect": "Allow",
            "Action": [
            "Resource": [

User Management

User management can be customised using the following configuration:

  administrators: # coma separated list of usernames
    enabled: true # default: true. Are new users allowed to create an account?
      enabled: false # default: false. Should users receive an email to activate their account?
      block-known-domains: true # default: false.
      custom-domains: # default: none. custom user-defined email domains to block
    enabled: true # default: true. Can users recover their account password by email?
    enabled: true # default: true. Can users login using the regular login form?

Email activation requires a valid email configuration (otherwise users won't receive the activation email).

Data Eviction

Unnecessary data like old results or orphaned files can be automatically deleted using the following settings:

      enabled: true # false by default
      months: 24
      enabled: true # false by default
      months: 24
    orphaned-project-blob: # orphaned project files
      enabled: true # false by default

By default, evicting data is disabled.

Runtime settings

Several runtime settings can be configured:

      timeout-min: 20 # Duration in min before any unresponsive test is killed
      max-iterations: 5 # Maximum number of iterations allowed for a check VU

Single Sign-On (SSO)

OpenID Connect (Oauth2)

OctoPerf EE supports Oauth2 / OpenID Connect authentication through Spring Boot properties. Here is an example application.yml:

            client-id: okta-client-id
            client-name: okta
            client-secret: okta-client-secret
            allowed-domains: "," # (Optional) restrict domains allowed to login. If empty, all are accepted
            username-suffix: "" # (Optional) Example: if the username is not a valid email.
            name: "Okta" # (Optional)
            description: "Login With Okta" # (Optional)
            scope: openid,email # oauth2 scopes
            client-authentication-method: client_secret_basic
            authorization-grant-type: authorization_code
            redirect-uri: "{baseUrl}/{action}/oauth2/code/{registrationId}" # DO NOT MODIFY THIS SETTING
          okta: 1
            user-name-attribute: sub

The Oauth2 provider (here okta) must then be declared in ui configuration config-prod.json to enable Oauth2 login. The user-name-attribute setting must match the username of the user logging in. user-name-attribute must point to an email attribute. If not, please also define username-suffix property.

Only emails are allowed as usernames. In that case, when the user jsmith logs in, it creates an account with username

Microsoft Azure

            provider: azure
            client-id: ${CLIENT_ID} # CLIENT_ID environment variable
            client-secret: ${CLIENT_SECRET} # CLIENT_SECRET environment variable
            scope: openid, email, profile
            issuer-uri:${TENANT_ID}/v2.0 # TENANT_ID environment variable
            user-name-attribute: email

When using Microsoft Azure authentication with OpenID Connect, it's important to use at least v2.0 API version endpoints. As stated in their notes and caveats:

We recommend that all OIDC compliant apps and libraries use the v2.0 endpoint to ensure compatibility.

This is need otherwise you may experience issues retrieving the account email and associating it with an OctoPerf account.


By default, OctoPerf EE creates and manages users inside its own database.

OctoPerf EE supports seamsless integration with a third party authentication server (Single Sign On, aka SSO) based on the LDAP protocol.


It's strongly recommended to be assisted by your System Administrator when configuring OctoPerf to support LDAP authentication.

Configuring LDAP Authentication

  driver: ldap
    # ldap or ldaps
    protocol: ldap
    hostname: localhost
    port: 389
    base: dc=example,dc=com
    principal-suffix: ""

      # anonymous, simple, default-tls, digest-md5 or external-tls
      method: simple
      username: cn=read-only-admin,dc=example,dc=com
      password: password

    # Cache Ldap users
      enabled: true
      durationSec: 300

    # user password encryption
    # OpenLDAP: leave as plain (even if encoded within Ldap)
      type: plain
      # sha, ssha, sha-256, md5 etc.
      algorithm: sha

    # LDAP Connection pooling
    pooled: false

    # User mapping
    object-class: person
      id: uid
      name: cn
      mail: mail

    # Ignorable errors
      partial-result: false
      name-not-found: false
      size-limit-exceeded: true

    # Filters that will be added to the LDAP query as "key=value AND key2=value2"
        memberOf: CN=groupname,OU=ApplicationGroupes,OU=Users,DC=domain,DC=com


The configuration above shows the default configuration used by the OctoPerf server when ldap driver is configured.

The LDAP authentication supports a variety of settings:

  • protocol: either ldap or ldaps,
  • hostname: hostname of the LDAP server,
  • port: 389 by default, LDAP server connection port,
  • base: an LDAP query defines the directory tree starting point,
  • principal-suffix: Empty by default. Example: Suffix which completes the User name to authenticate with the LDAP server,

  • authentication:

    • method: (authentication method to use. Can be anonymous, simple, default-tls, digest-md5 or external-tls. The last one requires additional JVM configuration.
    • username: depending on the authentication method. Must be a valid Ldap Distinguished Name (DN),
    • password: depending on the authentication method. Password associated to the given username,
  • cache:

    • enabled: true by default. Enables User cache to reduce the load on the LDAP server,
    • durationSec: 300 sec by default. Any change on the LDAP server takes up to this duration to be taken into account,
  • password: Defines how the user password should be encrypted,

    • type: plain by default. Set hash and refer to algorithm to set the right hashing algorithm,
    • algorithm: sha by default, required if type is set to hash. Supports: sha, sha-256, sha-384, sha-512, ssha, ssha-256, ssha-384, ssha-512, md5, smd5, crypt, crypt-md5, crypt-sha-256, crpyt-sha-512, crypt-bcrypt and pkcs5s2.
  • pooled: whenever to use a connection pool to connect to the LDAP server. Pooling speeds up the connection,

  • object-class: objectClass defines the class attribute of the users. Users are filtered based on this value,

  • attributes:

    • id: required. Maps the given attribute (uid by default) to the user id,
    • name: optional. Maps the given attribute (cn by default) to the user firstname and lastname,
    • mail: optional. Maps the given attribute (mail by default) to the user mail. If empty, it uses the id concatenated with the principal-suffix.
  • ignore:

    • partial-result: optional. Set to true to ignore PartialResultException. AD servers typically have a problem with referrals. Normally a referral should be followed automatically, but this does not seem to work with AD servers. The problem manifests itself with a PartialResultException being thrown when a referral is encountered by the server. Setting this property to true presents a workaround to this problem,
    • name-not-found: Specify whether NameNotFoundException should be ignored in searches. In previous version, NameNotFoundException caused by the search base not being found was silently ignored,
    • size-limit-exceeded: Specify whether SizeLimitExceededException should be ignored in searches. This is typically what you want if you specify count limit in your search controls.


Choosing between Internal and LDAP authentication must be done as soon as possible. Existing internal users won't match LDAP ones if changed later, even if they share the same username.

When using LDAP authentication, a few features related to user management are disabled or non-functional:

  • changing or resetting a user password,
  • editing user profile information,
  • and registering a new user through the registration form.


You should configure the UI config-prod.json file using LDAP login settings.

Virtual User Validation Storage

Virtual user validation http requests and responses are stored within Elasticsearch by default.

  driver: elasticsearch # or blob, which means on disk
  retention-days: 1 # how much time validation data should be kept

Validation http requests can either be stored in blob storage (with validation.driver: blob; see Resources Storage section), or within the database (with validation.driver: elasticsearch).

Errors storage

The Error table keeps track of only two errors of each kind. It also has some max limits per JMeter container and test result that can be configured:

      max-per-test: 500 # max errors per test result,
      max-per-instance: 100 # max errors per jmeter container / instance,
      max-per-request: 2 # max identical errors/code per request.


Increasing these values can have adverse effect on the storage size required. It can also prevent your reports to load if you try to display too many errors per report.

Nightly Cleanup Batches

Some internal batch tasks (disabled by default) can be run every night to reclaim unnecessary disk space being used. Those batch jobs can be enabled using the following properties:

# Available in version >= 12.0.3
    # scans results to keeps only the first 500 error requests / responses
      enabled: false
    # scans virtual users modified within last 2 days to removed orphaned recorded requests / responses
      enabled: false
    # Removed recorded requests / responses of virtual users with 12 <= lastModified <= 24 months
      enabled: false
      min-months: 12
      max-months: 24
    # scans projects storage folder and removes them if the project doesn't exist anymore
      enabled: false
    # scans results storage folder and removes them if the result doesn't exist anymore
      enabled: false

When upgrading from version 12.0.2 and below, we suggest to temporarly enable those batches to clean previously undeleted files. Those batches run during the night.


Google's reCAPTCHA can be used to prevents bots from creating accounts on your OctoPerf platform.

  enabled: true
  private.key: your-private-key

It takes the following configuration:

  • enabled: true to activate Recaptcha,
  • endpoint: to use Google's recaptcha service,
  • private.key: your reCAPTCHA private key (obtained in the Google's reCAPTCHA admin console).


The public reCAPTCHA key must be configured in the config-prod.json file of the frontend, using the json key "captcha": "you-public-key". Please have a look at the GUI / Frontend configuration section to know how to map the configuration file.

JMeter Plugins

By default, JMeter plugins are automatically downloaded by each JMeter container on each load generator. This feature can be configured as following:

    download-from: JMETER

JMeter plugins are downloaded from:

  • SERVER: OctoPerf server downloads them when importing the JMX. The plugin files are available within project files,
  • JMETER: PluginsManagerCMD is executed just before running the test with install-for-jmx argument,
  • NONE: plugins are not downloaded at all.

File Upload Limit

The default file upload limit is set to 100MB.

    max-request-size: 200MB
    max-file-size: 200MB

This example sets the limit to 200MB.


Tests can be marked as failed when the error percentage is higher than a certain value.

    max-errors-percent: 5

You can configure system-wide notifications to be sent by your OctoPerf server:

    enabled: true
      - channel: user-activity
        token: XXX
          - CLOUD_HOST_ISSUE
          - JMX_IMPORT
          - TEST_ENDED
          - TEST_ERROR
          - TEST_FAILED
          - TEST_PASSED
          - TEST_STATUS
          - TEST_STARTED
          - USER_ACTIVATED
          - USER_LOGIN
          - USER_LOGOUT

Amazon Instance Tags

Aws Instances created by OctoPerf can be tagged with the following settings:

      key: value
      key2: value2

Those tags are applied to all Amazon providers configured on the platform.

UI configuration

The file config-prod.json lets you configure the UI of OctoPerf. An example of this configuration file:

  "version": "octoperf-1",
  "backendApiUrl": "",
  "docUrl": "",
  "adminEmail": "",
  "login": {
    "@type": "form",
    "register": true,
    "activate": true,
    "forgotPassword": true
  "notice": {
    "text": "BETA version: classic UI is available at [](",
    "height": 20,
    "color": "white",
    "backgroundColor": "#171e29"
  "twitter": true,
  "cloud": true,
  "recaptcha": "recaptcha-key"

This file supports a variety of settings:

  • version: the frontend application version used to store data in the browser local storage (changing this version resets all stored local preferences for every user),
  • backendApiUrl: the URL of the backend server, i.e. for OctoPerf SaaS,
  • docUrl: the URL of the documentation, i.e. for OctoPerf's public documentation,
  • adminEmail: the email address the users may contact for Administrator operations such as removing a Workspace,
  • login: see configuration bellow,
  • notice: replaces the application footer with a custom message (height is in pixels and text is markdown formatted),
  • twitter: [true/false] displays a Twitter feed on the right side of the application,
  • cloud: [true/false] activates a tab in the Administration page that lists all currently started Cloud instances,
  • recaptcha: "reCAPTCHA_public_key" your reCAPTCHA public key, see backend's configuration.

You can map a docker volume to this file /usr/share/nginx/html/ui/assets/config-prod.json to customize its content, and for example, update the administrator email address.

This must be done in the docker-compose.yml file. Simply edit the enterprise-ui service to map the file:

    - /path/to/config-prod.json:/usr/share/nginx/html/ui/assets/config-prod.json

UI Login configuration

UI Form login configuration

This is the most basic mode of login configuration used to connect to the backend server. We assume here that the backend manages user on its own.

  "login": {
    "@type": "form",
    "register": true,
    "activate": true,
    "forgotPassword": true
The supported settings are:

  • @type: the type of login configuration, must be set to "form",
  • register: [true/false] register page availability,
  • register: [true/false] activate page availability,
  • forgotPassword: [true/false] forgot password page availability.

UI OAuth login configuration

The page http://FRONTEND_URL/#/access/connect/{clientId} is automatically generated based on the oauth2 settings defined server-side in application.yml.

  "login": {
    "@type": "oauth",
    "id": "clientId"

The supported settings are:

UI LDAP login configuration

If using the server-side #ldap configuration you can set the frontend configuration to:

  "login": {
    "@type": "ldap"

This will disable register, activate, and forgot password pages.

The supported settings is:

  • @type: the type of login configuration, must be set to "ldap".


It is possible to ask the OctoPerf enterprise-edition backend to expose the value of all the properties. This is useful for debugging purpose but make sure to disable it once you're done or you'll be publicly exposing critical security information.

To do this, use the following property:

        include: "*"

After restarting the enterprise edition server, you will be able to access all the environment settings via http://<baseURL>/actuator/env.



This should only be activated for debugging purpose since with this setting critical security information is publicly exposed.