Skip to main content
Version: 0.9.1

Configuration

Configuration of marathon is done using a YAML formatted configuration usually placed in the root of the project and named Marathonfile.

Below is a very simple example of Marathonfile (without the platform configuration):

name: "My awesome tests"
outputDir: "marathon"
debug: false

There are a lot of options in marathon. This can be overwhelming especially when you're just starting out. We will split the options into general options below, complex options that you can find as subsections in the menu on the left and platform-specific options under each platform section.

If you're unsure how to properly format your options in Marathonfile take a look at the samples or take a look at the deserialisation logic in the configuration module of the project. Each option might use a default deserializer from yaml or a custom one. Usually the custom deserializer expects the type option for polymorphic types to understand which specific object we need to instantiate.

Important notes

File-system path handling

When specifying relative host file paths in the configuration they will be resolved relative to the directory of the Marathonfile, e.g. if you have /home/user/app/Marathonfile with baseOutputDir = "./output" then the actual path to the output directory will be /home/user/app/output.

Required

Below you will find a list of currently supported configuration parameters and examples of how to set them up. Keep in mind that not all of the additional parameters are supported by each platform. If you find that something doesn't work - please submit an issue for a platform at fault.

Test run configuration name

This string specifies the name of this test run configuration. It is used mainly in the generated test reports.

name: "My test run for sample app"

Output directory

Directory path to use as the root folder for all the runner output (logs, reports, etc).

For gradle, the output path will automatically be set to a marathon folder in your reports folder unless it's overridden.

outputDir: "build/reports/marathon"

Platform-specific options

Marathon requires you to specify the platform for each run, for example:

vendorConfiguration:
type: "Android"

Refer to platform configuration for additional options inside the vendorConfiguration block.

Optional

Ignore failures

By default, the build fails if some tests failed. If you want the build to succeed even if some tests failed, use true.

ignoreFailures: true

Code coverage

Depending on the vendor implementation, code coverage may not be supported. By default, code coverage is disabled. If this option is enabled, code coverage will be collected and marathon assumes that code coverage generation will be setup by the user (e.g. proper build flags, jacoco jar added to classpath, etc).

isCodeCoverageEnabled: true

Test output timeout

This parameter specifies that the underlying test executor should time out if there is no test output within this duration. By default, this is set to 5 minutes.

testOutputTimeoutMillis: 30000

Test batch timeout

This parameter specifies that the underlying test executor should time out if the batch execution exceeds this duration. By default, this is set to 30 minutes.

testBatchTimeoutMillis: 900000

Device provider init timeout

When the test run starts, the device provider is expected to provide some devices. This should not take more than 3 minutes by default. If your setup requires this to be changed, please override as following:

deviceInitializationTimeoutMillis: 300000

Analytics tracking

To better understand the use-cases that marathon is used for, we're asking you to provide us with anonymised information about your usage. By default, this is enabled. Use false to disable.

analyticsTracking: false
note

analyticsTracking can also be enabled (default value) / disabled directly from the CLI. It is disabled if it's set to be disabled in either the config or the CLI.

BugSnag reporting

To better understand crashes, we report crashes with anonymised info. By default, this is enabled. Use false to disable.

bugsnagReporting: false
note

bugsnagReporting can also be enabled (default value) / disabled directly from the CLI. It is disabled if it's set to be disabled in either the config or the CLI.

Uncompleted test retry quota

By default, tests that don't have any status reported after execution (for example a device disconnected during the execution) retry indefinitely. You can limit the number of total execution for such cases using this option.

uncompletedTestRetryQuota: 100

Execution strategy

When executing tests with retries there are multiple trade-offs to be made. Two execution strategies are supported: any success or all success. By default, ANY_SUCCESS strategy is used with fast execution i.e. if one of the test retries succeeds then the test is considered successfully executed and all non-started retries are removed.

Any success

Test passes if any of its executions are passing. This mode works only if there is no complex sharding strategy applied. This is the default.

executionStrategy:
mode: ANY_SUCCESS
info

Complex sharding with ANY_SUCCESS mode doesn't make sense when user asks for N number of tests to run explicitly, and we pass on the first one.

All success

Test passes if and only if all its executions are passing. This mode works only if there are no retries, i.e. no complex flakiness strategy, no retry strategy.

executionStrategy:
mode: ALL_SUCCESS
info

Adding retries with retry/flakiness strategies means users wants to trade off cost for reliability, i.e. add more retries and pass if one of test retries passes, so retries only make sense for ANY_SUCCESS mode.

When we use ALL_SUCCESS mode it means the user want to verify each test with a number of tries (they are not retries per se) and pass only if all of them succeed. This is the case when fixing a flaky test or adding a new test, and we want to have a signal that the test is fixed/not flaky.

Fast execution mode

When the test reaches a state where a decision about its state can be made, marathon can remove additional in-progress retries. This decision point is different depending on the execution mode used. Let's walk through two examples.

Assume ANY_SUCCESS strategy is used and 100 retries are scheduled for a test A via flakiness strategy. Let's say first 3 failed and the 4th attempt succeeded. At this point the test should already be considered passed since ANY_SUCCESS out of all retries leads to the result by definition of ANY_SUCCESS execution strategy. To save cost one can remove additional non-started retries by using fast execution mode (this is the default behavior for ANY_SUCCESS strategy). On the other hand one could disable fast execution and get much more accurate statistics about this test by executing more retries and calculating the probability of passing as a measure of flakiness for test A.

Assume ALL_SUCCESS strategy is used and 100 retries are scheduled using sharding strategy. Let's say first 3 passed and the 4th attempt failed. At this point the test should already be considered failed since any failure out of all retries leads to the result by definition of ALL_SUCCESS execution strategy. You can save cost by removing additional non-started retries by using fast execution mode (this is the default behaviour for ALL_SUCCESS strategy). On the other hand one could disable fast execution and verify the flakiness rate with a defined precision, in this case there are 100 retries, so you would get precision up to 1% for test A.

executionStrategy:
fast: true

Debug mode

Enables very verbose logging to stdout of all the marathon components. Very useful for debugging.

debug: true

Screen recording policy

By default, screen recordings will only be pulled for tests that failed (ON_FAILURE option). This is to save space and also to reduce the test duration time since we're not pulling additional files. If you need to save screen recordings regardless of the test pass/failure please use the ON_ANY option:

screenRecordingPolicy: "ON_ANY"

Output configuration

Max file path

By default, the max file path for any output file is capped at 255 characters due to limitations of some OSs. This is the reason why some test runs have lots of "File path length cannot exceed" messages in the log. Since there is currently no API to programmatically establish this limit, it's the user's responsibility to set a larger value if the OS supports this and the user desires it.

outputConfiguration:
maxPath: 1024