Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You use Azure Pipelines to build and test a React.js application.
You have a pipeline that has a single job.
You discover that installing JavaScript packages from npm takes approximately five minutes each time you run the pipeline.
You need to recommend a solution to reduce the pipeline execution time.
Solution: You recommend defining a container job that uses a custom container that has the JavaScript packages preinstalled.
Does this meet the goal?
Click on the arrows to vote for the correct answer
A. B.B
Instead enable pipeline caching.
Note:
npm-cache is a command line utility that caches dependencies installed via npm, bower, jspm and composer.
It is useful for build processes that run [npm|bower|composer|jspm] install every time as part of their build process. Since dependencies don't change often, this often means slower build times. npm-cache helps alleviate this problem by caching previously installed dependencies on the build machine.
https://www.npmjs.com/package/npm-cacheThe proposed solution of defining a container job that uses a custom container with the preinstalled JavaScript packages would meet the goal of reducing pipeline execution time.
When a pipeline runs, it creates an isolated environment where each task runs in its own process. This means that each time the pipeline runs, it has to download all the required JavaScript packages from npm, which can take a significant amount of time, especially for larger applications.
By defining a container job with a custom container that has the required JavaScript packages preinstalled, the pipeline can reuse the same container for subsequent runs, which can significantly reduce the pipeline's execution time. This approach can also ensure that the pipeline runs in a consistent environment, which can improve the reliability of the pipeline.
It is worth noting that defining a container job may require some additional configuration and setup. For example, you may need to create a Dockerfile that specifies the required packages and dependencies and build the container image before using it in the pipeline. Additionally, the custom container may require regular maintenance and updates to ensure that it stays up to date with the latest packages and security patches.
Overall, the proposed solution of defining a container job with a custom container that has the required JavaScript packages preinstalled is a valid solution to reduce the pipeline execution time.