👍🎉 We're excited you want to contribute! Read on! 🎉👍
Questions • Pull requests • Tackling issues • Reporting bugs • Labels
Before you ask, check our existing questions to see if your question has already been answered. If not, go ahead an open an issue or join us in gitter to ask a question.
Please be sure to use markdown code blocks when posting code on GitHub or Gitter:
```marko
<div>some marko ${code}</div>
```
```js
const some = js.code;
```
Not sure if that typo is worth a pull request? Found a bug and know how to fix it? Do it! We will appreciate it. Any significant improvement should be documented as a GitHub issue before anybody starts working on it.
We are always thrilled to receive pull requests. We do our best to process them quickly. If your pull request is not accepted on the first try, don't get discouraged! We'll work with you to come to an acceptable solution.
Prior to merging your PR, you will need to sign the JS Foundation CLA. It's pretty straight-forward and only takes a minute. You can even sign it now if you're thinking about contributing.
TIP: If you're new to GitHub or open source you can check out this free course on how to contribute to an open source project.
Before submitting your PR, make sure that all new and previous tests pass and that coverage has not decreased:
npm run test-coverage
While developing you can run a single test group and use grep to filter the tests:
npm test -- --grep=lifecycle
Marko makes use of directory based test suites. Take a look at the render
test suite:
test/ ⤷ render/ ⤷ fixtures/ ⤷ attrs/ ⤷ for-tag/ ⤷ expected.html ⤷ template.marko ⤷ test.js ⤷ nested-tags/ ⤷ while-tag/ ⤷ html.test.js
The html.test.js
file will run and read all the directories under render/fixtures
and for each directory (attrs
, for-tag
, etc.) it will run test.js
, render template.marko
and assert that it is equivalent to the content of expected.html
.
To add a new test, you'll find the appropriate test suite, copy a fixture, and modify it to add the new test.
A few of the tests suites use the same fixtures for multiple test scenarios. For example, the component-browser
tests run once rendering the component in a browser environment and a second time rendering in a server environment, then hydrating in the browser.
For some tests, it might be necessary to skip the test in one of these scenarios. This is done by exporting a skip_hydrate
(or similiarly named) property from the fixture. The value of the property should be a string explaining why the test is skipped.
If you've discovered an issue and are able to reproduce it, but don't have a fix, consider submitting a PR with a failing test case. You can mark a fixture as expected to fail by appending exporting a fails
property from the fixture. The value of the fails
property should be a string with the issue number. Upon merging a failing test case, a maintainer will update the corresponding issue to add the has failing test
label.
In the case that a fixture is used in multiple test scenarios, you can mark the test as failing in a specific scenario by exporting a fails_hydrate
(or similarly named) property from the fixture.
Expected failures won't cause Travis CI to report a error, but document that there is an issue and give others a starting point for fixing the problem.
If you need to dig a bit deeper into a failing test, use the --inspect-brk
flag, open Chrome DevTools, and click on the green nodejs icon () to start debugging. Learn more about debugging node from this video.
npm test -- --grep=test-name --inspect-brk
In addition to setting breakpoints, you can also add debugger;
statements in both your JavaScript files and Marko templates:
$ debugger;
<div>Hello ${input.name}!</div>
A number of the test suites make use snapshot comparisons. For example, the render
tests compare the rendered html against a stored snapshot. Similarly, the compiler
tests compare the generated JavaScript module againt a stored snapshot. Any changes compared to the snapshot should be looked at closely, but there are some cases where it is fine that the output has changed and the snapshot needs to be updated.
To update a snapshot, you can copy the contents from the actual
file to the expected
file in the fixture directory. You can also use the UPDATE_EXPECTATIONS
env variable to cause the test runner to update the expected
file for all currently failing tests in a suite:
UPDATE_EXPECTATIONS=1 npm test
Comment on the issue and let us know you'd like to tackle it. If for some reason you aren't going to be able to complete the work, let us know as soon as you can so we can open it up for another developer to work on.
Here's some to get started with:
- good first issue: great for new contributors
- help wanted issues: won't be tackled in the near future by the maintainers... we need your help!
- unassigned issues: open issues that no one has claimed... yet
- has failing test issues: open issues that already have a failing test case in the repo. make it pass!
A great way to contribute to the project is to send a detailed report when you encounter an issue. Even better: submit a PR with a failing test case (see how).
Check that our issue database doesn't already include that problem or suggestion before submitting an issue. If you find a match, you can use the "subscribe" button to get notified on updates. Rather than leaving a "+1" or "I have this too" comment, you can add a reaction to let us know that this is also affecting you without cluttering the conversation. However, if you have ways to reproduce the issue or have additional information that may help resolving the issue, please leave a comment.
We have an ISSUE_TEMPLATE that will populate your textarea when you go to open an issue. Use the relevant section and remove the rest.
Please provide as much detail as possible.
If you discover a security issue, please DO NOT file a public issue, instead send your report privately.
Security reports are greatly appreciated and we will publicly thank you for it. We currently do not offer a paid security bounty program.
Once you post an issue, a maintainer will add one or more labels to it. Below is a guideline for the maintainers and anyone else who is interested in what the various labels mean.
Every issue should be assigned one of these.
- bug: A bug report
- unverified-bug: A bug report that has not been verified
- feature: A feature request
- question: A question about how to do something in Marko
- community: Related to community building, improving the contribution process, etc.
- tech debt: Related to refactoring code, test structure, etc.
- docs: Related to documentation/website
What part of the Marko stack does this issue apply to? In most cases there should only be one of these.
- parser: Relates to
htmljs-parser
- compiler: Relates to the compiler (server only)
- runtime: Relates to the runtime (isomorphic/universal)
- core-taglib: Relates to custom tags that ship with Marko
- components: Relates to components
- tools: Relates to editor plugins, commandline tools, etc.
In many cases, additional actions should be taken when applying one of these.
- backlog: Tasks planned to be worked on
- in progress: This is currently being worked on.
- needs review: This issue needs to be followed up on.
- resolved: The question was answered, the bug was fixed, or the feature was implemented.
- duplicate: Someone has already posted the same or a very similar issue. A comment should be added that references the original issue.
- declined: This feature will not be implemented.
- not a bug: This is not a bug, but either user error or intended behavior.
- inactivity: There was not enough info to reproduce the bug or not enough interest in the feature to hash out an implementation plan and the conversation has stalled.
- no issue: This wasn't so much an issue as a comment
- good first issue: Small tasks that would be good for first time contributors.
- help wanted: Not on the roadmap, but we'd love for someone in the community to tackle it.
- blocked: Cannot be completed until something else happens first. This should be described in a comment with a link to the blocking issue.
- needs more info: The original poster needs to provide more information before action can be taken.
- user land: Something that probably won't be added to core, but could be implemented/proved out as a separate module.