Agent to integrate Playwright with ReportPortal.
- More about Playwright
- More about ReportPortal
Install the agent in your project:
npm install --save-dev @reportportal/agent-js-playwright
1. Create playwright.config.ts
or *.config.js
file with reportportal configuration:
import { PlaywrightTestConfig } from '@playwright/test';
const RPconfig = {
apiKey: '00000000-0000-0000-0000-000000000000',
endpoint: 'https://your.reportportal.server/api/v1',
project: 'Your reportportal project name',
launch: 'Your launch name',
attributes: [
{
key: 'key',
value: 'value',
},
{
value: 'value',
},
],
description: 'Your launch description',
};
const config: PlaywrightTestConfig = {
reporter: [['@reportportal/agent-js-playwright', RPconfig]],
testDir: './tests',
};
export default config;
The full list of available options presented below.
Option | Necessity | Default | Description |
---|---|---|---|
apiKey | Required | User's reportportal token from which you want to send requests. It can be found on the profile page of this user. | |
endpoint | Required | URL of your server. For example 'https://server:8080/api/v1'. | |
launch | Required | Name of launch at creation. | |
project | Required | The name of the project in which the launches will be created. | |
attributes | Optional | [] | Launch attributes. |
description | Optional | '' | Launch description. |
rerun | Optional | false | Enable rerun |
rerunOf | Optional | Not set | UUID of launch you want to rerun. If not specified, reportportal will update the latest launch with the same name |
mode | Optional | 'DEFAULT' | Results will be submitted to Launches page 'DEBUG' - Results will be submitted to Debug page. |
skippedIssue | Optional | true | reportportal provides feature to mark skipped tests as not 'To Investigate'. Option could be equal boolean values: true - skipped tests considered as issues and will be marked as 'To Investigate' on reportportal. false - skipped tests will not be marked as 'To Investigate' on application. |
debug | Optional | false | This flag allows seeing the logs of the client-javascript. Useful for debugging. |
launchId | Optional | Not set | The ID of an already existing launch. The launch must be in 'IN_PROGRESS' status while the tests are running. Please note that if this ID is provided, the launch will not be finished at the end of the run and must be finished separately. |
restClientConfig | Optional | Not set | The object with agent property for configure http(s) client, may contain other client options eg. timeout . Visit client-javascript for more details. |
launchUuidPrint | Optional | false | Whether to print the current launch UUID. |
launchUuidPrintOutput | Optional | 'STDOUT' | Launch UUID printing output. Possible values: 'STDOUT', 'STDERR'. Works only if launchUuidPrint set to true . |
includeTestSteps | Optional | false | Allows you to see the test steps at the log level. |
includePlaywrightProjectNameToCodeReference | Optional | false | Includes Playwright project name to code reference. See testCaseId and codeRef calculation . It may be useful when you want to see the different history for the same test cases within different playwright projects. |
extendTestDescriptionWithLastError | Optional | true | If set to true the last error log will be attached to the test case description and reported on the error log level |
uploadVideo | Optional | true | Whether to attach the Playwright's video to the test case. |
uploadTrace | Optional | true | Whether to attach the Playwright's trace to the test case. |
token | Deprecated | Not set | Use apiKey instead. |
The following options can be overridden using ENVIRONMENT variables:
Option | ENV variable |
---|---|
launchId | RP_LAUNCH_ID |
2. Add script to package.json
file:
{
"scripts": {
"test": "npx playwright test --config=playwright.config.ts"
}
}
When organizing tests, specify titles for test.describe
blocks, as this is necessary to build the correct structure of reports.
It is also required to specify playwright project names in playwright.config.ts
when running the same tests in different playwright projects.
Attachments can be easily added during test run via testInfo.attach
according to the Playwright docs.
import { test, expect } from '@playwright/test';
test('basic test', async ({ page }, testInfo) => {
await page.goto('https://playwright.dev');
// Capture a screenshot and attach it
const screenshot = await page.screenshot();
await testInfo.attach('screenshot', { body: screenshot, contentType: 'image/png' });
});
Note: attachment path can be provided instead of body.
As an alternative to this approach the ReportingAPI
methods can be used.
Note: ReportingAPI
methods will send attachments to ReportPortal right after their call, unlike attachments provided via testInfo.attach
that will be reported only on the test item finish.
You can use the following console
native methods to report logs to tests:
console.log();
console.info();
console.debug();
console.warn();
console.error();
console's log
, info
,dubug
reports as info log.
console's error
, warn
reports as error log if message contains "error" mention, otherwise as warn log.
As an alternative to this approach the ReportingAPI
methods can be used.
This reporter provides Reporting API to use it directly in tests to send some additional data to the report.
To start using the ReportingApi
in tests, just import it from '@reportportal/agent-js-playwright'
:
import { ReportingApi } from '@reportportal/agent-js-playwright';
The API provide methods for attaching data (logs, attributes, testCaseId, status).
All ReportingApi methods have an optional suite parameter.
If you want to add a data to the suite, you must pass the suite name as the last parameter.
Add attributes (tags) to the current test. Should be called inside of corresponding test.
ReportingApi.addAttributes(attributes: Array<Attribute>, suite?: string);
required: attributes
optional: suite
Example:
test('should have the correct attributes', () => {
ReportingApi.addAttributes([
{
key: 'testKey',
value: 'testValue',
},
{
value: 'testValueTwo',
},
]);
expect(true).toBe(true);
});
Set test case id to the current test (About test case id). Should be called inside of corresponding test.
ReportingApi.setTestCaseId(id: string, suite?: string);
required: id
optional: suite
If testCaseId
not specified, it will be generated automatically based on codeRef.
Example:
test('should have the correct testCaseId', () => {
ReportingApi.setTestCaseId('itemTestCaseId');
expect(true).toBe(true);
});
Send logs to report portal for the current test. Should be called inside of corresponding test.
ReportingApi.log(level: LOG_LEVELS, message: string, file?: Attachment, suite?: string);
required: level
, message
optional: file
, suite
where level
can be one of the following: TRACE, DEBUG, WARN, INFO, ERROR, FATAL
Example:
test('should contain logs with attachments',() => {
const fileName = 'test.jpg';
const fileContent = fs.readFileSync(path.resolve(__dirname, './attachments', fileName));
const attachment = {
name: fileName,
type: 'image/jpg',
content: fileContent.toString('base64'),
};
ReportingApi.log('INFO', 'info log with attachment', attachment);
expect(true).toBe(true);
});
Send logs with corresponding level to report portal for the current test. Should be called inside of corresponding test.
ReportingApi.info(message: string, file?: Attachment, suite?: string);
ReportingApi.debug(message: string, file?: Attachment, suite?: string);
ReportingApi.warn(message: string, file?: Attachment, suite?: string);
ReportingApi.error(message: string, file?: Attachment, suite?: string);
ReportingApi.trace(message: string, file?: Attachment, suite?: string);
ReportingApi.fatal(message: string, file?: Attachment, suite?: string);
required: message
optional: file
, suite
Example:
test('should contain logs with attachments', () => {
ReportingApi.info('Log message');
ReportingApi.debug('Log message');
ReportingApi.warn('Log message');
ReportingApi.error('Log message');
ReportingApi.trace('Log message');
ReportingApi.fatal('Log message');
expect(true).toBe(true);
});
Send logs to report portal for the current launch. Should be called inside of the any test or suite.
ReportingApi.launchLog(level: LOG_LEVELS, message: string, file?: Attachment);
required: level
, message
optional: file
where level
can be one of the following: TRACE, DEBUG, WARN, INFO, ERROR, FATAL
Example:
test('should contain logs with attachments', async () => {
const fileName = 'test.jpg';
const fileContent = fs.readFileSync(path.resolve(__dirname, './attachments', fileName));
const attachment = {
name: fileName,
type: 'image/jpg',
content: fileContent.toString('base64'),
};
ReportingApi.launchLog('INFO', 'info log with attachment', attachment);
await expect(true).toBe(true);
});
Send logs with corresponding level to report portal for the current launch. Should be called inside of the any test or suite.
ReportingApi.launchInfo(message: string, file?: Attachment);
ReportingApi.launchDebug(message: string, file?: Attachment);
ReportingApi.launchWarn(message: string, file?: Attachment);
ReportingApi.launchError(message: string, file?: Attachment);
ReportingApi.launchTrace(message: string, file?: Attachment);
ReportingApi.launchFatal(message: string, file?: Attachment);
required: message
optional: file
Example:
test('should contain logs with attachments', () => {
ReportingApi.launchInfo('Log message');
ReportingApi.launchDebug('Log message');
ReportingApi.launchWarn('Log message');
ReportingApi.launchError('Log message');
ReportingApi.launchTrace('Log message');
ReportingApi.launchFatal('Log message');
expect(true).toBe(true);
});
Assign corresponding status to the current test item. Should be called inside of corresponding test.
ReportingApi.setStatus(status: string, suite?: string);
required: status
optional: suite
where status
must be one of the following: passed, failed, stopped, skipped, interrupted, cancelled
Example:
test('should have status FAILED', () => {
ReportingApi.setStatus('failed');
expect(true).toBe(true);
});
setStatusFailed, setStatusPassed, setStatusSkipped, setStatusStopped, setStatusInterrupted, setStatusCancelled
Assign corresponding status to the current test item. Should be called inside of corresponding test.
ReportingApi.setStatusFailed(suite?: string);
ReportingApi.setStatusPassed(suite?: string);
ReportingApi.setStatusSkipped(suite?: string);
ReportingApi.setStatusStopped(suite?: string);
ReportingApi.setStatusInterrupted(suite?: string);
ReportingApi.setStatusCancelled(suite?: string);
optional: suite
Example:
test('should call ReportingApi to set statuses', () => {
ReportingAPI.setStatusFailed();
ReportingAPI.setStatusPassed();
ReportingAPI.setStatusSkipped();
ReportingAPI.setStatusStopped();
ReportingAPI.setStatusInterrupted();
ReportingAPI.setStatusCancelled();
});
Assign corresponding status to the current launch. Should be called inside of the any test or suite.
ReportingApi.setLaunchStatus(status: string);
required: status
where status
must be one of the following: passed, failed, stopped, skipped, interrupted, cancelled
Example:
test('launch should have status FAILED', () => {
ReportingApi.setLaunchStatus('failed');
expect(true).toBe(true);
});
setLaunchStatusFailed, setLaunchStatusPassed, setLaunchStatusSkipped, setLaunchStatusStopped, setLaunchStatusInterrupted, setLaunchStatusCancelled
Assign corresponding status to the current test item. Should be called inside of the any test or suite.
ReportingApi.setLaunchStatusFailed();
ReportingApi.setLaunchStatusPassed();
ReportingApi.setLaunchStatusSkipped();
ReportingApi.setLaunchStatusStopped();
ReportingApi.setLaunchStatusInterrupted();
ReportingApi.setLaunchStatusCancelled();
Example:
test('should call ReportingApi to set launch statuses', () => {
ReportingAPI.setLaunchStatusFailed();
ReportingAPI.setLaunchStatusPassed();
ReportingAPI.setLaunchStatusSkipped();
ReportingAPI.setLaunchStatusStopped();
ReportingAPI.setLaunchStatusInterrupted();
ReportingAPI.setLaunchStatusCancelled();
});
To integrate with Sauce Labs just add attributes for the test case:
[{
"key": "SLID",
"value": "# of the job in Sauce Labs"
}, {
"key": "SLDC",
"value": "EU (your job region in Sauce Labs)"
}]
There is known issue that in some cases launches not finished as expected in ReportPortal while using static annotations (.skip()
, .fixme()
) that expect the test to be 'SKIPPED'.
This may happen in case of error thrown from before
/beforeAll
hooks, retries enabled and fullyParallel: false
. Associated with #85.
In this case as a workaround we suggest to use .skip()
and .fixme()
annotations inside the test body:
use
test('example fail', async ({}) => {
test.fixme();
expect(1).toBeGreaterThan(2);
});
instead of
test.fixme('example fail', async ({}) => {
expect(1).toBeGreaterThan(2);
});