Thanks for getting here. If you have a good will to improve CodeceptJS we are always glad to help. Ask questions, raise issues, ping in Twitter.
Go over the steps in this guide as first contributions
To start you need:
- Fork and clone the repo.
- Run
npm i --forceto install all required libraries - Do the changes.
- Add/Update Test (if possible)
- Update documentation
- Run
npm run defto generate types - Run
npm run docsif you change the documentation - Commit and Push to your fork
- Make Pull Request
To run codeceptjs from this repo use:
node bin/codecept.js
To run examples:
node bin/codecept.js run -c examples
Depending on a type of change you should do the following.
To see recorder queue in logs enable NodeJS debug output by passing DEBUG=codeceptjs:* env variable:
DEBUG=codeceptjs:* npx codeceptjs run
Please keep in mind that CodeceptJS have unified API for Playwright, WebDriverIO, Appium, Puppeteer, TestCafe. Tests written using those helpers should be compatible at syntax level. However, some helpers may contain unique methods. That happens. If, for instance, WebDriverIO has method XXX and Playwright doesn't, you can implement XXX inside Playwright using the same method signature.
_Whenever a new method or new behavior is added it should be documented in a docblock. Valid JS-example is required! Do not edit docs/helpers/, those files are generated from docblocks in corresponding helpers! _
Working test is highly appreciated. To run the test suite you need:
- selenium server + chromedriver
- PHP installed
To launch PHP demo application run:
php -S 127.0.0.1:8000 -t test/data/appExecute test suite:
mocha test/helper/WebDriver_test.js
mocha test/helper/Puppeteer_test.jsUse --grep to execute tests only for changed parts.
If you need to add new HTML page for a test, please create new .php file in to tests/data/app/view/form:
Adding myexample page:
tests/data/app/view/form/myexample.phpThen is should be accessible at:
http://localhost:8000/form/myexampleWhenever a new method or new behavior is added it should be documented in a docblock. Valid JS-example is required!
Adding a test is highly appreciated.
Start JSON server to run tests:
npm run json-serverEdit a test at test/rest/REST_test.js or test/rest/ApiDataFactory_test.js
_Whenever a new method or new behavior is added it should be documented in a docblock. Valid JS-example is required! Do not edit docs/helpers/, those files are generated from docblocks in corresponding helpers! _
It is recommended to run mobile tests on CI. So do the changes, make pull request, see the CI status. Appium tests are executed at Saucelabs.
Before applying any Core changes please raise an issue to discuss that change with core team. Please try to add corresponding testcase to runner or unit.
Documentation is stored in /docs directory in Markdown format.
Documentation for helpers is a part of a source code.
Whenever you need to update docs for a helper do it inside a .js file.
After you updated docblock in JS file, generate markdown files with next command:
npm run docs
Documentation parts can be shared across helpers. Those parts are located in docs/webapi/*.mustache. Inside a docblock those files can be included like this:
/**
* {{> click }}
*/
click() {
// ...
}Note: Due to the (lib)[https://documentation.js.org/] that we are using to generate docs, the fast and cheap way to fix format issue that text after the mustache template is appended without formatting is moving the texts to above the mustache template.
/**
* // Before
* Click action
* {{> click }}
* Click action
*/
click() {
// ...
} /**
* // After
* Click action
* {{> click }}
*/
click() {
// ...
}Typings are generated in typings/ directory via jsdoc
After you updated docblock in JS file, generate typing files with next command:
npm run def
Whenever you implemented a feature/bugfix
Run unit tests:
mocha test/unitRun general tests:
mocha test/runnerInstead of manually running php, json_server and selenium for before tests you
can use docker-compose to run those automatically.
You can find docker-compose.yml file in test directory and run all commands
from this directory. Currently, we provide following commands to run tests with
respective dependencies:
docker-compose run --rm test-unitdocker-compose run --rm test-helpers
# or pass path to helper test to run specific helper,
# for example to run only WebDriver tests:
docker-compose run --rm test-helpers test/helper/WebDriver_test.js
# Or to run only rest and ApiDataFactory tests
docker-compose run --rm test-helpers test/restTo that we provide three separate services respectively for WebDriver, Nightmare and Puppeteer tests:
docker-compose run --rm test-acceptance.webdriverio
docker-compose run --rm test-acceptance.puppeteerBy default, dockerized tests are run against node 12.10.0, you can run it against
specific version as long as there is Docker container available for such
version. To do that you need to build codecept's Docker image prior to running
tests and pass NODE_VERSION as build argument.
For example to prepare test-helpers containers based on node 9.4.0:
docker-compose build --build-arg NODE_VERSION=9.4.0 test-helpersAnd now every command based on test-helpers service will use node 9.4.0. The
same argument can be passed when building unit and acceptance tests services.
We're currently using a bunch of CI services to build and test codecept in different environments. Here's short summary of what are differences between separate services
Here we use CodeceptJS docker image to build and execute tests inside it. We
start with building Docker container based on Dockerfile present in main project
directory. Then we run (in this order) unit tests, all helpers present in
test/helpers, then we go with test/rest directory and finally if everything
passed so far it executes acceptance tests. For easier maintenance and local
debugging CircleCI uses docker-compose.yml file from test directory.
You can find Circle config in .circleci directory.