Skip to content

reportportal/agent-js-testcafe

Repository files navigation

@reportportal/agent-js-testcafe

Agent to integrate TestCafe with ReportPortal.

Installation

Install the agent in your project:

npm install --save-dev testcafe-reporter-agent-js-testcafe@npm:@reportportal/testcafe-reporter-agent-js-testcafe

Note: This package is namespaced. Therefore the following command can be used to install the reporter in a way that TestCafe can detect it. (Related issue in TestCafé repository)

Configuration

1. Create rp.json file with reportportal configuration:

{
  "apiKey": "<API_KEY>",
  "endpoint": "https://your.reportportal.server/api/v1",
  "project": "YourReportPortalProjectName",
  "launch": "YourLauncherName",
  "attributes": [
    {
      "key": "YourKey",
      "value": "YourValue"
    },
    {
      "value": "YourValue"
    }
  ],
  "description": "Your launch description"
}
Option Necessity Default Description
apiKey Required User's reportportal token from which you want to send requests. It can be found on the profile page of this user.
endpoint Required URL of your server. For example 'https://server:8080/api/v1'.
launch Required Name of launch at creation.
project Required The name of the project in which the launches will be created.
attributes Optional [] Launch attributes.
description Optional '' Launch description.
rerun Optional false Enable rerun
rerunOf Optional Not set UUID of launch you want to rerun. If not specified, reportportal will update the latest launch with the same name
mode Optional 'DEFAULT' Results will be submitted to Launches page
'DEBUG' - Results will be submitted to Debug page (values must be upper case).
debug Optional false This flag allows seeing the logs of the client-javascript. Useful for debugging.
restClientConfig Optional Not set axios like http client config. May contain agent property for configure http(s) client, and other client options e.g. proxy, timeout. For debugging and displaying logs the debug: true option can be used.
Visit client-javascript for more details.
headers Optional {} The object with custom headers for internal http client.
launchUuidPrint Optional false Whether to print the current launch UUID.
launchUuidPrintOutput Optional 'STDOUT' Launch UUID printing output. Possible values: 'STDOUT', 'STDERR', 'FILE', 'ENVIRONMENT'. Works only if launchUuidPrint set to true. File format: rp-launch-uuid-${launch_uuid}.tmp. Env variable: RP_LAUNCH_UUID, note that the env variable is only available in the reporter process (it cannot be obtained from tests).
skippedIssue Optional true reportportal provides feature to mark skipped tests as not 'To Investigate'.
Option could be equal boolean values:
true - skipped tests considered as issues and will be marked as 'To Investigate' on reportportal.
false - skipped tests will not be marked as 'To Investigate' on application.
token Deprecated Not set Use apiKey instead.

2.1 Create .testcaferc.json TestCafe configuration file and add agent-js-testcafe to the reporter property

{
  "browsers": "chrome",
  "src": "./tests/**/*.js",
  "screenshots": {
    "path": "./screenshots/"
  },
  "reporter": [
    {
      "name": "list"
    },
    {
      "name": "agent-js-testcafe"
    }
  ],
  "takeScreenshotsOnFails": true
}

Run tests via testcafe command.

2.2 As an alternative if you are using API you can create testcafe.js file and use the reporter with provided config manually:

const createTestCafe = require('testcafe');
const { createReporter } = require('testcafe-reporter-agent-js-testcafe/build/createReporter');
const rpConfig = require('./rp.json');

async function start() {
  const testcafe = await createTestCafe('localhost');
  const runner = testcafe.createRunner();

  await runner.reporter(createReporter(rpConfig)).run(); // or just set 'agent-js-testcafe'

  await testcafe.close();
}

start();

Run tests via node testcafe.js.

Note: TestCafe options from .testcaferc.json can be overwritten programmatically in testcafe.js

Reporting

This reporter provides Reporting API to use it directly in tests to send some additional data to the report.

There are two ways to add additional data to the tests.

Using TestCafe meta method:

1. For fixture:

fixture`Getting Started`.page`http://devexpress.github.io/testcafe/example`
  .meta({
    description: 'Suite description',
    attributes: [{ key: 'page', value: 'testCafeExample' }, { value: 'sample' }],
  });

2. For test:

test('My first test', async (page) => {
  await page
    .typeText('#developer-name', 'John Smith')
    .click('#submit-button')
    .expect(Selector('#article-header').innerText)
    .eql('Thank you, John Smith!');
}).meta({
  description: 'Test form behavior',
  attributes: [{ key: 'test', value: 'form' }],
});

Only attributes and description properties supported.

Using ReportingApi:

To start using the ReportingApi in tests, just import it from '@reportportal/agent-js-testcafe':

const { ReportingApi } = require('testcafe-reporter-agent-js-testcafe/build/reportingApi');

Reporting API methods

The API provide methods for attaching data (logs, attributes, testCaseId, status).

addAttributes

Add attributes(tags) to the current test. Should be called inside of corresponding test or fixture.
ReportingApi.addAttributes(attributes: Array<Attribute>);
required: attributes
Example:

test('should have the correct attributes', async (t) => {
  ReportingApi.addAttributes([
    {
      key: 'testKey',
      value: 'testValue',
    },
    {
      value: 'testValueTwo',
    },
  ]);
  await t.expect(true).eql(true);
});
setTestCaseId

Set test case id to the current test. Should be called inside of corresponding test or fixture.
ReportingApi.setTestCaseId(id: string);
required: id
If testCaseId not specified, it will be generated automatically.
Example:

test('should have the correct testCaseId', async (t) => {
  ReportingApi.setTestCaseId('itemTestCaseId');
  await t.expect(true).eql(true);
});
log

Send logs to report portal for the current test. Should be called inside of corresponding test or fixture.
ReportingApi.log(level: LOG_LEVELS, message: string, file?: Attachment);
required: level, message
where level can be one of the following: TRACE, DEBUG, WARN, INFO, ERROR, FATAL
Example:

test('should contain logs with attachments', async (page) => {
  const fileName = 'test.jpg';
  const fileContent = fs.readFileSync(path.resolve(__dirname, './attachments', fileName));
  const attachment = {
    name: fileName,
    type: 'image/jpg',
    content: fileContent.toString('base64'),
  };
  ReportingApi.log('INFO', 'info log with attachment', attachment);

  await page.expect(true).eql(true);
});
info, debug, warn, error, trace, fatal

Send logs with corresponding level to report portal for the current test or for provided by name. Should be called inside of corresponding test or fixture.
ReportingApi.info(message: string, file?: Attachment);
ReportingApi.debug(message: string, file?: Attachment);
ReportingApi.warn(message: string, file?: Attachment);
ReportingApi.error(message: string, file?: Attachment);
ReportingApi.trace(message: string, file?: Attachment);
ReportingApi.fatal(message: string, file?: Attachment);
required: message
Example:

test('should contain logs with attachments', async (page) => {
    ReportingApi.info('Log message');
    ReportingApi.debug('Log message');
    ReportingApi.warn('Log message');
    ReportingApi.error('Log message');
    ReportingApi.trace('Log message');
    ReportingApi.fatal('Log message');
    
    await page.expect(true).eql(true);
});
launchLog

Send logs to report portal for the current launch. Should be called inside of the any test or fixture.
ReportingApi.launchLog(level: LOG_LEVELS, message: string, file?: Attachment);
required: level, message
where level can be one of the following: TRACE, DEBUG, WARN, INFO, ERROR, FATAL
Example:

test('should contain logs with attachments', async (page) => {
  const fileName = 'test.jpg';
  const fileContent = fs.readFileSync(path.resolve(__dirname, './attachments', fileName));
  const attachment = {
    name: fileName,
    type: 'image/jpg',
    content: fileContent.toString('base64'),
  };
  ReportingApi.launchLog('INFO', 'info log with attachment', attachment);

  await page.expect(true).eql(true);
});
launchInfo, launchDebug, launchWarn, launchError, launchTrace, launchFatal

Send logs with corresponding level to report portal for the current launch. Should be called inside of the any test or fixture.
ReportingApi.launchInfo(message: string, file?: Attachment);
ReportingApi.launchDebug(message: string, file?: Attachment);
ReportingApi.launchWarn(message: string, file?: Attachment);
ReportingApi.launchError(message: string, file?: Attachment);
ReportingApi.launchTrace(message: string, file?: Attachment);
ReportingApi.launchFatal(message: string, file?: Attachment);
required: message
Example:

test('should contain logs with attachments', async (page) => {
    ReportingApi.launchInfo('Log message');
    ReportingApi.launchDebug('Log message');
    ReportingApi.launchWarn('Log message');
    ReportingApi.launchError('Log message');
    ReportingApi.launchTrace('Log message');
    ReportingApi.launchFatal('Log message');
    
    await page.expect(true).eql(true);
});
setStatus

Assign corresponding status to the current test item.
ReportingApi.setStatus(status: string);
required: status
where status must be one of the following: passed, failed, stopped, skipped, interrupted, cancelled, info, warn
Example:

test('should have status FAILED', async (page) => {
    ReportingApi.setStatus('failed');
    
    await page.expect(true).eql(true);
});
setStatusFailed, setStatusPassed, setStatusSkipped, setStatusStopped, setStatusInterrupted, setStatusCancelled, setStatusInfo, setStatusWarn

Assign corresponding status to the current test item.
ReportingApi.setStatusFailed();
ReportingApi.setStatusPassed();
ReportingApi.setStatusSkipped();
ReportingApi.setStatusStopped();
ReportingApi.setStatusInterrupted();
ReportingApi.setStatusCancelled();
ReportingApi.setStatusInfo();
ReportingApi.setStatusWarn();
Example:

test('should call ReportingApi to set statuses', async (page) => {
    ReportingAPI.setStatusFailed();
    ReportingAPI.setStatusPassed();
    ReportingAPI.setStatusSkipped();
    ReportingAPI.setStatusStopped();
    ReportingAPI.setStatusInterrupted();
    ReportingAPI.setStatusCancelled();
    ReportingAPI.setStatusInfo();
    ReportingAPI.setStatusWarn();
});
setLaunchStatus

Assign corresponding status to the current launch.
ReportingApi.setLaunchStatus(status: string);
required: status
where status must be one of the following: passed, failed, stopped, skipped, interrupted, cancelled, info, warn
Example:

test('launch should have status FAILED', async (page) => {
    ReportingApi.setLaunchStatus('failed');
    
    await page.expect(true).eql(true);
});
setLaunchStatusFailed, setLaunchStatusPassed, setLaunchStatusSkipped, setLaunchStatusStopped, setLaunchStatusInterrupted, setLaunchStatusCancelled, setLaunchStatusInfo, setLaunchStatusWarn

Assign corresponding status to the current test item.
ReportingApi.setLaunchStatusFailed();
ReportingApi.setLaunchStatusPassed();
ReportingApi.setLaunchStatusSkipped();
ReportingApi.setLaunchStatusStopped();
ReportingApi.setLaunchStatusInterrupted();
ReportingApi.setLaunchStatusCancelled();
ReportingApi.setLaunchStatusInfo();
ReportingApi.setLaunchStatusWarn();
Example:

test('should call ReportingApi to set launch statuses', async (page) => {
    ReportingAPI.setLaunchStatusFailed();
    ReportingAPI.setLaunchStatusPassed();
    ReportingAPI.setLaunchStatusSkipped();
    ReportingAPI.setLaunchStatusStopped();
    ReportingAPI.setLaunchStatusInterrupted();
    ReportingAPI.setLaunchStatusCancelled();
    ReportingAPI.setLaunchStatusInfo();
    ReportingAPI.setLaunchStatusWarn();
});

Integration with Sauce Labs

To integrate with Sauce Labs just add attributes for the test case:

[{
 "key": "SLID",
 "value": "# of the job in Sauce Labs"
}, {
 "key": "SLDC",
 "value": "EU (your job region in Sauce Labs)"
}]