Consider a TAS that uses a keyword-driven framework. The SUT is a web application and there is a large set of keywords available for writing the automated tests that relate to highly specific user actions linked directly to the GUI of the SUT. The automated test written with the keywords are statically analyzed by a custom tool which highlight’s repeated instances of identical sequence of keywords. The waiting mechanism implemented by the TAS for a webpage load is based on a synchronous sampling within a given timeout. The TAS allows checking a webpage load every seconds until a timeout value
(In User Acceptance Testing (UAT) for a new SUT, in addition to the manual tests performed by the end-users, automated tests are performed that focus on the execution of repetitive and routine test scenarios. In which of the following environments are all these tests typically performed?)
A new TAS allows the implementation of automated data-driven test scripts. All the tasks planned for the initial deployment of this TAS, aimed at installing and configuring the TAS components and provisioning the infrastructure, will be performed manually by a dedicated, specialized team. This TAS is expected to be deployed in the future in other similar environments. As a TAE, you see a risk that the correct and reproducible deployment of the TAS cannot be guaranteed. Which of the following options is BEST suited for mitigating this risk?
(Which of the following aspects of “design for testability” is MOST directly associated with the need to define precisely which interfaces are available in the SUT for test automation at different test levels?)
A defect in a SUT has been resolved and validated by an automated defect re-test in the current release of the software. This retest has now been added to the automated regression test suite.
Which statement BEST describes a reason why this defect could re-occur in future releases?
Consider a TAS that exclusively uses the APIs of a SUT. To make this work, significant changes have been required to the SUT by adding a set of dedicated test interfaces to the APIs. All the automated tests will use these test interfaces when interacting with the SUT. Assume that you are currently verifying the correctness of the automated test environment and test tool setup.
Which of the following would you expect to be the MOST specific risk associated with this scenario?
As a TAE you are evaluating a functional test automation tool that will be for several projects within your organization. The projects require that tool to work effectively and efficiently with SUT’s in distributed environments. The test automated tool also needs to interface with other existing test tools (test management tool and defect tracking tool.) The existing test tools subject to planned updates and their interface to the test automated tool may not work property after these updates.
Which of the following are the two LEAST important concerns related to the evaluation of the test automation in this scenario?
Is the test automation tool able to launch processors and execute test cases on multiple machines in different environments?
Does the test automation tool support a licensing scheme that allows accessing different sets?
Does the test automation tool have a large feature set, but only part of the features will be sets?
Do the release notes for the planned updates on existing specify the impacts on their interfaces to other tools?
Does the test automation tool need to install specific libraries that could impact the SUT?
Which of the following BEST describes why it is important to separate test definition from test execution in a TAA?
A SUT has an existing automated test suite.
Which of the following statements relating to the introduction of new features in the SUT is TRUE?
A release candidate of a SUT, after being fully integrated with all other necessary systems, has successfully passed all required functional tests (90% were automated tests and 10% were manual tests). Now, it is necessary to perform reliability tests aimed at evaluating whether, under certain conditions, that release will be able to guarantee an MTBF (Mean Time Between Failures) in the production environment higher than a certain threshold (expressed in CPU time). Which of the following test environments is BEST suited to perform these reliability tests?