In case of questions about the contribution process or for discussion of specific issues please visit the akka/dev gitter chat.
Depending on which version (or sometimes module) you want to work on, you should target a specific branch as explained below:
master
– active development branch of Akka 2.5.xrelease-2.4
– maintenance branch of Akka 2.4.xrelease-2.3
– maintenance branch of Akka 2.3.xartery-dev
– work on the upcoming remoting implementation, codenamed "artery"- similarly
release-2.#
branches contain legacy versions of Akka
Akka uses tags to categorise issues into groups or mark their phase in development.
Most notably many tags start with a t:
prefix (as in topic:
), which categorises issues in terms of which module they relate to. Examples are:
In general all issues are open for anyone working on them, however if you're new to the project and looking for an issue that will be accepted and likely is a nice one to get started you should check out the following tags:
- community - which identifies issues that the core team will likely not have time to work on, or the issue is a nice entry level ticket. If you're not sure how to solve a ticket but would like to work on it feel free to ask in the issue about clarification or tips.
- nice-to-have (low-priority) - are tasks which make sense, however are not very high priority (in face of other very high priority issues). If you see something interesting in this list, a contribution would be really wonderful!
Another group of tickets are those which start from a number. They're used to signal in what phase of development an issue is:
- 0 - new - is assigned when a ticket is unclear on its purpose or if it is valid or not. Sometimes the additional tag
discuss
is used to mark such tickets, if they propose large scale changes and need more discussion before moving into triaged (or being closed as invalid). - 1 - triaged - roughly speaking means "this ticket makes sense". Triaged tickets are safe to pick up for contributing in terms of likeliness of a patch for it being accepted. It is not recommended to start working on a ticket that is not triaged.
- 2 - pick next - used to mark issues which are next up in the queue to be worked on. Sometimes it's also used to mark which PRs are expected to be reviewed/merged for the next release. The tag is non-binding, and mostly used as an organisational helper.
- 3 - in progress - means someone is working on this ticket. If you see a ticket that has the tag, however seems inactive, it could have been an omission with removing the tag, feel free to ping the ticket then if it's still being worked on.
The last group of special tags indicate specific states a ticket is in:
- bug - bugs take priority in being fixed above features. The core team dedicates a number of days to working on bugs each sprint. Bugs which have reproducers are also great for community contributions as they're well isolated. Sometimes we're not as lucky to have reproducers though, then a bugfix should also include a test reproducing the original error along with the fix.
- failed - tickets indicate a Jenkins failure (for example from a nightly build). These tickets usually start with the
FAILED: ...
message, and include a stacktrace + link to the Jenkins failure. The tickets are collected and worked on with priority to keep the build stable and healthy. Often times it may be simple timeout issues (Jenkins boxes are slow), though sometimes real bugs are discovered this way.
Pull request validation states:
validating => [tested | needs-attention]
- signify pull request validation status.
These guidelines apply to all Akka projects, by which we mean both the akka/akka
repository,
as well as any plugins or additional repositories located under the Akka GitHub organisation.
These guidelines are meant to be a living document that should be changed and adapted as needed. We encourage changes that make it easier to achieve our goals in an efficient way.
The steps below describe how to get a patch into a main development branch (e.g. master
).
The steps are exactly the same for everyone involved in the project (be it core team, or first time contributor).
- Make sure an issue exists in the issue tracker for the work you want to contribute.
- If there is no ticket for it, create one first.
- Fork the project on GitHub. You'll need to create a feature-branch for your work on your fork, as this way you'll be able to submit a pull request against the mainline Akka.
- Create a branch on your fork and work on the feature. For example:
git checkout -b wip-custom-headers-akka-http
- Please make sure to follow the general quality guidelines (specified below) when developing your patch.
- Please write additional tests covering your feature and adjust existing ones if needed before submitting your pull request. The
validatePullRequest
sbt task (explained below) may come in handy to verify your changes are correct.
- Once your feature is complete, prepare the commit following our Creating Commits And Writing Commit Messages. For example, a good commit message would be:
Adding compression support for Manifests #22222
(note the reference to the ticket it aimed to resolve). - If it's a new feature, or a change of behaviour, document it on the akka-docs, remember, an undocumented feature is not a feature. If the feature was touching Scala or Java DSL, make sure to document it in both the Java and Scala documentation (usually in a file of the same name, but under
/scala/
instead of/java/
etc). - Now it's finally time to submit the pull request!
- If you have not already done so, you will be asked by our CLA bot to sign the Lightbend CLA online. CLA stands for Contributor License Agreement and is a way of protecting intellectual property disputes from harming the project.
- If you're not already on the contributors white-list, the @akka-ci bot will ask
Can one of the repo owners verify this patch?
, to which a core member will reply by commentingOK TO TEST
. This is just a sanity check to prevent malicious code from being run on the Jenkins cluster. - Now both committers and interested people will review your code. This process is to ensure the code we merge is of the best possible quality, and that no silly mistakes slip though. You're expected to follow-up these comments by adding new commits to the same branch. The commit messages of those commits can be more lose, for example:
Removed debugging using printline
, as they all will be squashed into one commit before merging into the main branch.- The community and team are really nice people, so don't be afraid to ask follow up questions if you didn't understand some comment, or would like clarification on how to continue with a given feature. We're here to help, so feel free to ask and discuss any kind of questions you might have during review!
- After the review you should fix the issues as needed (pushing a new commit for new review etc.), iterating until the reviewers give their thumbs up–which is signalled usually by a comment saying
LGTM
, which means "Looks Good To Me".- In general a PR is expected to get 2 LGTMs from the team before it is merged. If the PR is trivial, or under special circumstances (such as most of the team being on vacation, a PR was very thoroughly reviewed/tested and surely is correct) one LGTM may be fine as well.
- If the code change needs to be applied to other branches as well (for example a bugfix needing to be backported to a previous version), one of the team will either ask you to submit a PR with the same commit to the old branch, or do this for you.
- Backport pull requests such as these are marked using the phrase
for validation
in the title to make the purpose clear in the pull request list. They can be merged once validation passes without additional review (if there are no conflicts).
- Backport pull requests such as these are marked using the phrase
- Once everything is said and done, your pull request gets merged 🎉 Your feature will be available with the next “earliest” release milestone (i.e. if back-ported so that it will be in release x.y.z, find the relevant milestone for that release). And of course you will be given credit for the fix in the release stats during the release's announcement. You've made it!
The TL;DR; of the above very precise workflow version is:
- Fork Akka
- Hack and test on your feature (on a branch)
- Document it
- Submit a PR
- Sign the CLA if necessary
- Keep polishing it until received enough LGTM
- Profit!
Note that the Akka sbt project is large, so sbt
needs to be run with lots of heap (1-2 GB). This can be specified using a command line argument sbt -mem 2048
or in the environment variable SBT_OPTS
but then as a regular JVM memory flag, for example SBT_OPTS=-Xmx2G
, on some platforms you can also edit the global defaults for sbt in /usr/local/etc/sbtopts
.
The Akka build includes a special task called validatePullRequest
which investigates the changes made as well as dirty
(uncommitted changes) in your local working directory and figures out which projects are impacted by those changes,
then running tests only on those projects.
For example changing something in akka-actor
would cause tests to be run in all projects which depend on it
(e.g. akka-actor-tests
, akka-stream
, akka-docs
etc.).
To use the task simply type validatePullRequest
, and the output should include entries like shown below:
> validatePullRequest
[info] Diffing [HEAD] to determine changed modules in PR...
[info] Detected uncomitted changes in directories (including in dependency analysis): [akka-protobuf,project]
[info] Detected changes in directories: [akka-actor-tests, project, akka-stream, akka-docs, akka-persistence]
By default changes are diffed with the master
branch when working locally, if you want to validate against a different
target PR branch you can do so by setting the PR_TARGET_BRANCH environment variable for sbt:
PR_TARGET_BRANCH=origin/example sbt validatePullRequest
Binary compatibility rules and guarantees are described in depth in the Binary Compatibility Rules section of the documentation.
Akka uses MiMa (which is short for Lightbend Migration Manager) to validate binary compatibility of incoming pull requests. If your PR fails due to binary compatibility issues, you may see an error like this:
[info] akka-stream: found 1 potential binary incompatibilities while checking against com.typesafe.akka:akka-stream_2.11:2.4.2 (filtered 222)
[error] * method foldAsync(java.lang.Object,scala.Function2)akka.stream.scaladsl.FlowOps in trait akka.stream.scaladsl.FlowOps is present only in current version
[error] filter with: ProblemFilters.exclude[ReversedMissingMethodProblem]("akka.stream.scaladsl.FlowOps.foldAsync")
In such situations it's good to consult with a core team member if the violation can be safely ignored (by adding the above snippet to project/MiMa.scala
), or if it would indeed break binary compatibility.
Situations when it may be fine to ignore a MiMa issued warning include:
- if it is touching any class marked as
private[akka]
,/** INTERNAL API*/
or similar markers - if it is concerning internal classes (often recognisable by package names like
dungeon
,impl
,internal
etc.) - if it is adding API to classes / traits which are only meant for extension by Akka itself, i.e. should not be extended by end-users
- other tricky situations
For a pull request to be considered at all it has to meet these requirements:
-
Regardless if the code introduces new features or fixes bugs or regressions, it must have comprehensive tests.
-
The code must be well documented in the Lightbend's standard documentation format (see the ‘Documentation’ section below).
-
The commit messages must properly describe the changes, see further below.
-
All Lightbend projects must include Lightbend copyright notices. Each project can choose between one of two approaches:
- All source files in the project must have a Lightbend copyright notice in the file header.
- The Notices file for the project includes the Lightbend copyright notice and no other files contain copyright notices. See http://www.apache.org/legal/src-headers.html for instructions for managing this approach for copyrights.
Akka uses the first choice, having copyright notices in every file header.
Some additional guidelines regarding source code are:
- Files should start with a
Copyright (C) 2017 Lightbend Inc. <http://www.lightbend.com>
copyright header. - Keep the code DRY.
- Apply the Boy Scout Rule whenever you have the chance to.
- Never delete or change existing copyright notices, just add additional info.
- Do not use
@author
tags since it does not encourage Collective Code Ownership.- Contributors , each project should make sure that the contributors gets the credit they deserve—in a text file or page on the project website and in the release notes etc.
If these requirements are not met then the code should not be merged into master, or even reviewed - regardless of how good or important it is. No exceptions.
Whether or not a pull request (or parts of it) shall be back- or forward-ported will be discussed on the pull request discussion page, it shall therefore not be part of the commit messages. If desired the intent can be expressed in the pull request description.
All documentation must abide by the following maxims:
- Example code should be run as part of an automated test suite.
- Version should be programmatically specifiable to the build.
- Generation should be completely automated and available for scripting.
- Artifacts that must be included in the Lightbend stack should be published to a maven “documentation” repository as documentation artifacts.
All documentation is preferred to be in Lightbend's standard documentation format reStructuredText compiled using Lightbend's customized Sphinx based documentation generation system, which among other things allows all code in the documentation to be externalized into compiled files and imported into the documentation.
To learn about how to build the documentation locally see the Reference Docs section about this topic: Build the documentation.
For more info, or for a starting point for new projects, look at the Lightbend Documentation Template project.
For larger projects that have invested a lot of time and resources into their current documentation and samples scheme (like for example Play), it is understandable that it will take some time to migrate to this new model. In these cases someone from the project needs to take the responsibility of manual QA and verifier for the documentation and samples.
Akka generates JavaDoc-style API documentation using the genjavadoc sbt plugin, since the sources are written mostly in Scala.
Generating JavaDoc is not enabled by default, as it's not needed on day-to-day development as it's expected to just work. If you'd like to check if your links and formatting look good in JavaDoc (and not only in ScalaDoc), you can generate it by running:
sbt -Dakka.genjavadoc.enabled=true javaunidoc:doc
Which will generate JavaDoc style docs in ./target/javaunidoc/index.html
.
All the external runtime dependencies for the project, including transitive dependencies, must have an open source license that is equal to, or compatible with, Apache 2.
This must be ensured by manually verifying the license for all the dependencies for the project:
- Whenever a committer to the project changes a version of a dependency (including Scala) in the build file.
- Whenever a committer to the project adds a new dependency.
- Whenever a new release is cut (public or private for a customer).
Which licenses are compatible with Apache 2 are defined in this doc, where you can see that the licenses that are listed under Category A
are automatically compatible with Apache 2, while the ones listed under Category B
need additional action:
Each license in this category requires some degree of reciprocity; therefore, additional action must be taken in order to minimize the chance that a user of an Apache product will create a derivative work of a reciprocally-licensed portion of an Apache product without being aware of the applicable requirements.
Each project must also create and maintain a list of all dependencies and their licenses, including all their transitive dependencies. This can be done either in the documentation or in the build file next to each dependency.
Follow these guidelines when creating public commits and writing commit messages.
-
If your work spans multiple local commits (for example; if you do safe point commits while working in a feature branch or work in a branch for a long time doing merges/rebases etc.) then please do not commit it all but rewrite the history by squashing the commits into a single big commit which you write a good commit message for (like discussed in the following sections). For more info read this article: Git Workflow. Every commit should be able to be used in isolation, cherry picked etc.
-
The first line should be a descriptive sentence what the commit is doing, including the ticket number. It should be possible to fully understand what the commit does—but not necessarily how it does it—by just reading this single line. We follow the “imperative present tense” style for commit messages (more info here).
It is not ok to only list the ticket number, type "minor fix" or similar. If the commit is a small fix, then you are done. If not, go to 3.
-
Following the single line description should be a blank line followed by an enumerated list with the details of the commit.
-
You can request review by a specific team member for your commit (depending on the degree of automation we reach, the list may change over time):
Review by @gituser
- if you want to notify someone on the team. The others can, and are encouraged to participate.
Example:
enable Travis CI #1
* Details 1
* Details 2
* Details 3
Akka uses Jenkins GitHub pull request builder plugin that automatically merges the code, builds it, runs the tests and comments on the pull request in GitHub.
Upon a submission of a pull request the GitHub pull request builder plugin will post a following comment:
Can one of the repo owners verify this patch?
This requires a member from a core team to start the pull request validation process by posting a comment consisting only of OK TO TEST
.
From now on, whenever new commits are pushed to the pull request, a validation job will be automatically started and the results of the validation posted to the pull request.
A pull request validation job can be started manually by posting PLS BUILD
comment on the pull request.
In order to speed up PR validation times, the Akka build contains a special sbt task called validatePullRequest
,
which is smart enough to figure out which projects should be built if a PR only has changes in some parts of the project.
For example, if your PR only touches akka-persistence
, no akka-remote
tests need to be run, however the task
will validate all projects that depend on akka-persistence
(including samples).
Also, tests tagged as PerformanceTest
, TimingTest
, LongRunningTest
and all multi-node tests are excluded from PR validation.
You can exclude the same kind of tests in your local build by starting sbt with:
sbt -Dakka.test.tags.exclude=performance,timing,long-running -Dakka.test.multi-in-test=false
In order to force the validatePullRequest
task to build the entire project, regardless of dependency analysis of a PRs
changes one can use the special PLS BUILD ALL
command (typed in a comment on GitHub, on the pull request), which will cause
the validator to test all projects.
Note, that OK TO TEST
will only be picked up when the user asking for it is considered an admin. Public (!) members of the akka organization are automatically considered admins and users manually declared admin in the Jenkins job (currently no one is explicitly listed). PLS BUILD
and PLS BUILD ALL
can be issued by everyone that is an admin or everyone who was whitelisted in the Jenkins Job (whitelisting != declaring someone an admin).
Akka uses Scalariform to enforce some of the code style rules.
Java code is currently not automatically reformatted by sbt (expecting to have a plugin to do this soon). Thus we ask Java contributions to follow these simple guidelines:
- 2 spaces
{
on same line as method name- in all other aspects, follow the Oracle Java Style Guide
Avoid short test timeouts, since Jenkins server may GC heavily causing spurious test failures. GC pause or other hiccups of 2 seconds are common in our CI environment. Please note that usually giving a larger timeout does not slow down the tests, as in an expectMessage
call for example it usually will complete quickly.
There are a number of ways timeouts can be defined in Akka tests. The following ways to use timeouts are recommended (in order of preference):
remaining
is first choice (requireswithin
block)remainingOrDefault
is second choice3.seconds
is third choice if not using testkit- lower timeouts must come with a very good reason (e.g. Awaiting on a known to be "already completed"
Future
)
Special care should be given to expectNoMsg
calls, which indeed will wait the entire timeout before continuing, therefore a shorter timeout should be used in those, for example 200
or 300.millis
.
You can read up on remaining
and friends in TestKit.scala.
For external contributions of entire features, the normal way is to establish it
as a stand-alone feature first, to show that there is a need for the feature. The
next step would be to add it to Akka as an "may change"-feature (in the
akka-contrib subproject) and marking it's public api with the ApiMayChange
annotation,
then when the feature is hardened, well documented and
tested it becomes an officially supported Akka feature.
List of Akka features marked as may change
Each project should be configured to use a continuous integration (CI) tool (i.e. a build server à la Jenkins).
Lightbend is sponsoring a Jenkins server farm, sometimes referred to as "the Lausanne cluster". The cluster is made out of real bare-metal boxes, and maintained by the Akka team (and other very helpful people at Lightbend).
In addition to PR validation the cluster is also used for nightly and performance test runs.