3 Months Free Update
3 Months Free Update
3 Months Free Update
You have executed an automated test suite for a product that was released into production. Although all the tests passed, there was a major failure in production in an area that was covered well by your automated tests.
You have run the automated tests again and one of the tests is now failing and this is directly related to the production defect that was raised. You decide to run the automated test suite again on the same version of the SUT and the test now passes.
What SHOULD you do now to verify the validity of the automated tests?
Which of the following statements does NOT describe good practice for maintaining the TAS?
You are evaluating several test modelling tools and are wanting to automatically generate test cases within the tool where many different combinations of input data are created.
You are then wanting to export the test cases into a csv file which can then be read by a functional test execution tool using a data-driven or keyword-driven scripting method.
You have investigated several tools and there is only one tool that provides all the necessary features defined by your team with the exception of the export facility. It does not provide an export into either .xls or .csv formats.
What would be the BEST next step regarding the selection of this tool?
Your functional regression test automation suite ran successfully for the first two sprints and no failures were encountered during the runs. The automation suite records the status of each test case as either 'pass' or 'fail' and has excellent recovery capability built in.
For the third sprint, the TAS log reported several test cases with a status of 'fail'. You investigated each test case and found that most failures were due to a defect in one of the keyword scripts, rather than in the SUT. For those where the failure was in the SUT, defect reports were raised but several were returned by the developers asking for more information to enable them to reproduce the problem.
Which additional log items SHOULD you add to the TAS that would BEST improve failure analysis and defect reporting for future sprints?
a) Dynamic measurement information about the SUT.
b) A status of TAS error’, in additional to pass' and 'fail', for each test case.
c) Use of a colour coding scheme so that 'pass’ is in red and fail' is in green.
d) A counter to determine how many times each test case has been executed.
e) System configuration information including software/firmware and operating system versions.
f) A copy of the source code for all Keyword scripts executed.
You are reviewing the testability of your SUT.
Which of the following BEST refers to the characteristic of OBSERVABILITY?
When the SUT provides insight into the behaviour of the system, providing the users the with the status of the various actions performed so that they can check that expected behaviour equals actual behaviour, what is this called?
Assume that you are the TAE responsible for the correct functioning of a TAS, deployed in a test environment that consists of a few machines running the same version of the operating system. The TAS has been working and stable since its deployment, it has been used to run an automated test suite consisting of many similar automated test. The infrastructure team is planning to update the operating system on these machines by installing a new the service pack for security reasons. Since the vendor of the operating system assurance full backward compatibility, the infrastructure team assurance that there will be no impacts on the functioning of the TAS.
What is the BEST approach to confirm the correct functioning of the TAS in this scenario?
Consider A TAS for testing a desktop application via its GUI. All the test cases of the automated test suite contain the same identical sequences of steps at the beginning (to create the necessary objects when doing a preliminary configuration of the test environment and at the end (to remove everything created –specifically for the test itself during the preliminary configuration of the test environment). All automated test cases use the same set of assertion functions from a shared library, for verifying the values in the GUI fields ( e.g text boxes).
What is the BEST recommendation for improving the TAS?
Your TAS has been running successfully on a Windows/GUI based SUT for some years. The SUT has undergone minimal change over the years to maintain business as usual, deploying six-monthly releases for minor enhancements and bug fixes using a waterfall lifecycle.
The TAS has not changed at all during this period. The current project for the SUT will be using the Scrum methodology to deliver a more modern, competitive, user interface. It is in the release planning stage with an agreed release backlog and set of sprints outlined.
The move from lengthy waterfall releases to shorter sprints has led you to conduct a review of the current TAS to make sure it is robust and fully optimised for the timescale challenges of the new project.
What two steps would be BEST to undertake during the review?
a) Ensure that new automation code is using the same naming conventions as existing code.
b) Perform a full regression run in Sprint 1 to identify what improvements could be made to the TAS for future sprints.
c) Ensure that the TAS is using the latest libraries for the operating system.
d) Review the functions that act upon the controls for the GUI for possible consolidation.
e) Involve the test team to see what ease-of-use improvements they would like to see made to the TAS.
Your goal is to verify completeness, consistency and correct behavior of an automated test suite. The TAS has been proven to successfully install in the SUT environment. All the preliminary checks to verify the correct functioning of the automated test environment and test tool configuration, installation and setup have successfully completed.
Which of the following is NOT a relevant check for achieving your goal in this scenario?
A TAS uses a commercial test automation tool and the default logs generated by the inconsistent formats such as different types of messages (pass/fail steps, screenshots, warnings, etc.) To solve this issue some custom logging functions have been created from the test scripts, making it possible to log the different types of messages with the same format. However, this may cause a problem due to excessive size of the logs which can make it difficult to find the required information. Assume that all the default logs will be disabled when running the automated tests and that some tests will not generate excessively sized logs.
Which of the following represents the BEST suggestion for implementing the custom logging functions?
Consider a TAS that uses a keyword-driven framework. The SUT is a web application and there is a large set of keywords available for writing the automated tests that relate to highly specific user actions linked directly to the GUI of the SUT. The automated test written with the keywords are statically analyzed by a custom tool which highlight’s repeated instances of identical sequence of keywords. The waiting mechanism implemented by the TAS for a webpage load is based on a synchronous sampling within a given timeout. The TAS allows checking a webpage load every seconds until a timeout value
Consider a TAS that exclusively uses the APIs of a SUT. To make this work, significant changes have been required to the SUT by adding a set of dedicated test interfaces to the APIs. All the automated tests will use these test interfaces when interacting with the SUT. Assume that you are currently verifying the correctness of the automated test environment and test tool setup.
Which of the following would you expect to be the MOST specific risk associated with this scenario?
You are working on a web-based application called Book Vault that allows people to upload books and order books. This application must be available on all major browsers.
You have been testing the application manually and management have asked you to consider automating some of the tests.
You have investigated a number of commercial and free tools which can automate tests at a web browser level and one tool in particular meets your requirements and you have implemented a trial version.
You have basic programming skills and the main goal is to automate a few functional tests to see if the tool is compatible with the application and can recognise the objects and controls.
Which scripting technique would be MOST suitable in this scenario in order to meet the objectives?
Consider a TAS deployed into production. The SUT is a web application and the test suite consists of a set of automated regression tests developed via GUI. A keyword-driven framework has been adopted for automating the regression tests. The tests are based on identification at low-levels of the web page components (e.g class indexes, tab sequence indexes and coordinates) in the next planned release the SUT will be subject to significant corrective maintenance (bug-fixes) and evolution (new features) Maintenance costs to update the test scripts should be as low as possible and the scripts must be highly reusable.
Which of the following statements is most likely to be TRUE?
A major component of your organisation's Test Automaton Solution (TAS) is a popular open-source third-party capture-replay tool for automated functional testing.
Which two of the following must the Test Automation Engineer (TAE) ensure happens for this TAS?
a) The third party tool is placed under configuration management control.
b) The annual support and maintenance costs are agreed with the tool's vendor.
c) It is Important to obtain information about updates and new versions of the tool so that the third party tool is kept up to date.
d) Ensure that the TAS test scripts are integrated into the tool's framework.
e) Ensure that no changes are made to the tool, because modifications are not allowed for third party products.
Your project is transitioning from manual to automated testing. You have decided to implement a pilot project so that lessons learned can inform future time estimates and schedules.
Which two of the following represent the types of test cases that are MOST suited to a test automation pilot project?
a) High added value test cases that require little effort to automate.
b) Test that are run infrequently as these will be simpler to automate
c) Reliability test cases that can show added value soon
d) Technically challenging test cases to provide the best validation of manual test conversion
e) Tests that are least Important to the business as these are safer to trial
Your company is new to test automation and as TAE. you have designed a TAS which successfully supports the SUT for the current project.
There are other systems currently in operation which have been tested manually and more systems are planned over the coming years. Based on this success, your company requires test automation be rolled out to other current and future SUTs with consistency being a key objective.
Which of the following is the BEST way to achieve that?
You have inherited a TAS that is working well it uses keyword-driven scripting and was well architected. The automation architect who built the system has now moved on to another company. The TAS is working across several projects and has a multiple library of keywords, categorised by project. The individual project teams maintain these keyword scripts.
Based only on the given information, what is the MOST significant risk for the TAS?