A SUT (SUT1) is a client-server system based on a thin client. The client is primarily a display and input interface, while the server provides almost all the resources and functionality of the system. Another SUT (SUT2) is a client-server system based on a fat client that relies little on the server and provides most of the resources and functionality of the system. A given TAS is used to implement automated tests on both SUT1 and SUT2. The main objective of the TAS is to cover as many system functionalities as possible through automated tests executed as fast as possible. Which of the following statements about the automation solution is BEST in this scenario?
An API's response to a request made to the corresponding endpoint should return some specific data about a payment transaction in JSON format. In particular, your goal is to write the test automation code, keeping it as short as possible, aimed at determining whether that response includes certain properties (transaction_id, amount, status, timestamp) with the data types and formats expected. Assuming that the TAF provides all the necessary support to validate the specified API response, how would you BEST achieve your goal?
(Which of the following statements about how test automation is applied across different software development lifecycle models is TRUE?)
To improve the maintainability of test automation code, it is recommended to adopt design principles and design patterns that allow the code to be structured into:
As a TAE, you are evaluating a test automation tool to automate some UI tests for a web app. The automated tests will first locate the required HTML elements on the web page using their corresponding identifiers (locators), then perform actions on those elements, and finally check the presence of any expected text for an HTML element. These tests are independent of each other and are organized into a test suite that must be run every night against the most recent build of the web app. There is a high risk that the web app will crash while running some automated tests. Based only on the given information, which of the following is your MOST important concern related to the evaluation of the test automation tool?
A suite of automated test cases was run multiple times on the same release of the SUT in the same test environment. Consider analyzing a test histogram that shows the distribution of test results (pass, fail, etc.) for each test case across these runs. Which of the following potential issues is MOST likely to be identified as a result of such an analysis?
(Which of the following statements refers to a typical advantage of test automation?)
A CI/CD pipeline consists of two phases: build and deployment. The build phase, among other activities, runs automated test cases at the following test levels: Component Testing (CT) and Component Integration Testing (CIT). If the build phase is successful, the deployment phase is started. The deployment phase first provisions the test environment infrastructure needed to deploy the SUT, then deploys the SUT to this environment, and finally triggers another separate pipeline that runs automated test cases at the following test levels: System Testing (ST) and Acceptance Testing (AT). Which of the following statements is TRUE?
Automated tests at the UI level for a web app adopt an asynchronous waiting mechanism that allows them to synchronize test steps with the app, so that they are executed correctly and at the right time, only when the app is ready and has processed the previous step: this is done when there are no timeouts or pending asynchronous requests. In this way, the tests automatically synchronize with the app's web pages. The same initialization tasks to set test preconditions are implemented as test steps for all tests. Regarding the pre-processing (Setup) features defined at the test suite level, the TAS provides both a Suite Setup (which runs exactly once when the suite starts) and a Test Setup (which runs at the start of each test case in the suite). Which of the following recommendations would you provide for improving the TAS (assuming it is possible to perform all of them)?
Consider a TAS implemented to perform automated testing on native mobile apps at the UI level, where the TAF implements a client-server architecture. The client runs on-premise and allows creation of automated test scripts using TAF libraries to recognize and interact with the app’s UI objects. The server runs in the cloud as part of a PaaS service, receiving commands from the client, translating them into actions for the mobile device, and sending the results to the client. The cloud platform hosts several mobile devices dedicated for use by this TAS. The device on which to run test scripts/test suites is specified at run time. You are currently verifying whether the test automation environment and all other TAS/TAF components work correctly. Which of the following activities would you perform to achieve your goal?