test maturity model integrated (tmmi) survey...

20
17ª Dorset Square, London, NW1 6QB T: +44 (0)207 871 2300 E: [email protected] www.experimentus.com Test Maturity Model integrated (TMMi) Survey results How Mature Are CompaniesSoftware Quality Management Processes In Today‟s Market? 2011 update Listen | Challenge | Understand | Interpret | Create

Upload: others

Post on 17-Jul-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

17ª Dorset Square, London, NW1 6QB

T: +44 (0)207 871 2300 E: [email protected] www.experimentus.com

Test Maturity Model integrated (TMMi)

Survey results

How Mature Are Companies‟ Software Quality Management Processes In

Today‟s Market?

2011 update

Listen | Challenge | Understand | Interpret | Create

Page 2: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 2 Copyright © 2011 Experimentus Ltd

Page 3: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 3 Copyright © 2011 Experimentus Ltd

Table of Contents

Executive Summary .......................................................................................................................... 4

Notable Results ............................................................................................................................. 4

Areas for Concern ......................................................................................................................... 4

Conclusion ..................................................................................................................................... 5

Background ....................................................................................................................................... 6

About TMMi ................................................................................................................................... 6

The Survey .................................................................................................................................... 7

The Practice areas of TMMi Level 2 in detail ..................................................................................... 9

Overall rating ................................................................................................................................. 9

Test Policy and Strategy .............................................................................................................. 10

Test Design and Execution .......................................................................................................... 11

Test Planning .............................................................................................................................. 13

Test Environment ........................................................................................................................ 14

Test Monitoring and Control ......................................................................................................... 15

Conclusion ................................................................................................................................... 16

Industry Overview ........................................................................................................................... 17

Appendix 1 ...................................................................................................................................... 18

Background to the TMMi Model and the TMMi Foundation .......................................................... 18

Copyright © 2011 Experimentus Ltd. This document and any information herein are confidential and copyright property of Experimentus Ltd and without infringement neither the whole nor any extract may be disclosed, loaned, copied or used for manufacturing, provision of services or other purposes whatsoever without prior written consent. No liability is accepted for loss or damages from any cause whatsoever from the use of the document. Experimentus Ltd retain the right to alter the document at any time unless a written statement to the contrary has been appended.

Page 4: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 4 Copyright © 2011 Experimentus Ltd

Executive Summary There is very little data that exists on the maturity of testing practices throughout the software

quality life cycle in the Industry and lots of people refer to the maturity with a level of certainty not

backed up by any real data. In 2009 Experimentus established a survey for this very purpose.

Two years on from that first survey, data is still being collected. This document reflects the up to

date view of organisation test maturity based upon the industry standard Test Maturity Model

integrated (TMMi) – see www.tmmifoundation.org.

Of the 100 plus companies, across 9 industry sectors, who responded:

In 2009 72.5% of the respondents were at TMMi Level 1 heading for Level 2,

meaning they are working in a chaotic, hero-based way but starting to build project-

based processes. However, the 2010 survey sees that change in the right direction;

now only 63% are at TMMi Level 1 heading for Level 2, showing a reduction of13%.

In 2009 27.5% were at TMMi Level 2 heading towards Level 3, meaning they have

some established project based process and are moving towards implementing

process at an organisational level. For the 2010 survey we have seen a growth in

the average maturity of test process to 37% now achieving Level 2 and heading for

Level 3, showing an overall improvement of 35%.

As in our 2009 survey, unfortunately none of the 2010 respondents reached Level 3

although many have improved what they do and are therefore moving closer to

achieving this.

0% 20% 40% 60% 80%

TMMi Level 1

TMMi Level 2

72.5%

27.5%

63.0%

37.0%

2010/2009 Maturity accross respondents

2010

2009

Notable Results The results suggest software testers are still consistent at designing and planning testing, but

they continue to be not so good at setting goals, monitoring tests and managing the plans.

Combine this with not being consistent in how testing is estimated and it should come as no

surprise that testing is still seen as a bottleneck to delivery.

Meanwhile, test environments are well planned and reflect „production-like‟ test environments

fairly consistently for the later stages of testing (e.g. acceptance testing). While this will always be

a difficult area to manage, the importance is clearly recognised.

Areas for Concern It is still surprising that metrics continues to be a major weakness in the management of software

quality and testing. The marked drop in the numbers who have metrics in place has now shifted to

organisations now revisiting this (as a result of not getting the right metrics in the first place or the

data was not consistent in its meaning). Thus affecting the measurements and therefore the

For the 2010

survey we have

seen a growth in

the average

maturity of test

process to 37%

now achieving

Level 2 and

heading for Level

3, showing an

overall

improvement of

35%

Page 5: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 5 Copyright © 2011 Experimentus Ltd

decisions which can be made as a result. This leaves many organisations without the means to

manage properly, control and improve what is undoubtedly a large proportion of the project costs.

Conclusion It is not as though testing and software quality management is a new discipline in the software

development lifecycle, where appropriate good practices not being developed to cope with new

processes. Far from it - It is perhaps a damning indictment of the industry that after all these years

we can consistently design and plan testing, but have no thought or regard for effectively

measuring the success and efficiency of this activity (which, combined with the costs of rework,

forms a significant proportion of project costs).

The ambition of testing to be recognised as a profession will never be realised unless there is

measurement in place to manage and report on activities to assist in creating roadmaps for self-

improvement and the delivery of better quality software on time and within budget.

Management also have a considerable amount of hidden costs in the development lifecycle

which, without the right metrics, they cannot hope to understand or control. It does raise the

question as to why any organisation would want to run without consistent, meaningful and

actionable metrics, especially when this can lead to self-improvement.

Overall, the trend towards better management of software quality and testing is improving and

over time will dramatically reduce what are currently “hidden costs”. The questions still to be

answered are:

“How quickly will the industry adopt appropriate good practices?”

and

“With the current state of immaturity, where can you turn to get good appropriate

practices?”

Until recently the TMMi model had covered Maturity Levels 2 and 3 („Managed‟ and „Defined‟) and

with the release of Levels 4 and 5 covering „Management and Measurement‟ and „Optimisation‟,

the TMMi model provides a valuable source for appropriate good practices. The use of the model

has already been established by both service providers and clients to great success, with

dedicated testing practices and even government departments beginning to benefit from its use

and in some cases looking to be certified as a way of differentiating their capability from that of

their competitors.

Overall the trend

towards better

management of

software quality

and testing is

improving and

over time will

dramatically

reduce what are

currently “hidden

costs”

Page 6: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 6 Copyright © 2011 Experimentus Ltd

Background

In December 2010, the TMMi Foundation published the full TMMi model from Level 1 to 5, with

the requirements for organisations and individuals to have their assessment method for the TMMi

Model reviewed and accredited. At the same time they provided guidance on what is required to

attain Accredited Assessor status.

This release was a significant achievement for the TMMi Foundation, marking the beginning of an

industry-wide roadmap for implementing software quality management into the application

development lifecycle. Their roots go back to 2004 when a small group of quality process

improvement enthusiasts from around Europe met for the first time and decided it would make

sense to develop and support a single, non-commercial test improvement model. Since then,

there have been a growing number of supporters who acknowledge the positive difference the

TMMi model makes to the delivery of increased quality and reduced costs.

In September 2008, Experimentus became the first company to have an accredited assessment

method, accredited Assessors and Lead Assessors in the UK. After receiving its accreditation,

Experimentus conducted a survey to understand the maturity of the software quality processes

across the IT industry. Over 100 respondents, from many different industries, completed the

survey. It was decided that as the survey had been well received and as there is little other data

available about the maturity of the software testing industry that the survey would be updated.

This paper details the results of the 2010 survey and reflects the changes that have been taking

place within software quality management and test practices of organisations over the last two

years.

About TMMi In the same way as the CMMI (Capability Maturity Model Integrated) process model is split over 5

Levels, the following diagram depicts the 5 Levels of the TMMi model.

Chaotic process

Dependent on heroes

No understanding of the cost of quality

LEVEL 1: INITIAL

Test policy and strategy

Test planning

Test monitoring and control

Test design and execution

Test environment

LEVEL 2: MANAGED

Test organisation

Test training programme

Test life cycle and integration

Non-functional testing

Peer reviews

LEVEL 3: DEFINED

Test measurement

Software quality evaluation

Advanced peer reviews

LEVEL 4: MANAGEMENT & MEASUREMENT

Defect prevention

Test process optimisation

Quality control

LEVEL 5: OPTIMISATION

Page 7: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 7 Copyright © 2011 Experimentus Ltd

Each maturity level of the model is made up of a series of components (see diagram below).

At the top we have the Maturity Level which indicates an organisation‟s, project‟s or team‟s level

of maturity. The Maturity Level is made up of a series of Process Areas (such as „Test Policy and

Strategy‟). These are the areas in which goals need to be achieved to verify a Maturity Level has

been reached.

Each Process Area contains Generic and Specific Goals and Specific Practices, which in turn

provide the details required to implement the Process Area. So, for example, under the „Test

Policy and Strategy‟ Process Area there are the following supporting Practices:

Specific Goal - Establish a test policy

Specific Practice 1.1 Define test goals

Specific Practice 1.2 Define the test policy

Specific Practice 1.3 Distribute the test policy to stakeholders

Each Process Area is looked at in terms not only of its existence but also its deployment and the

effectiveness of that deployment.

Please refer to Appendix 1 for more background to the TMMi model.

The Survey The survey was designed to align major testing activities with Process Areas analysed in a TMMi

assessment, with the purpose of providing a survey indicating the alignment of current testing

practices to this industry standard. The initial survey took place during the last quarter of 2008

and closed in February 2009 and the second survey was closed in November 2010, with each

respondent asked to review a series of statements and to answer the statement with one of

following responses:

Strongly agree – that the process in question is in place and well established

Page 8: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 8 Copyright © 2011 Experimentus Ltd

Slightly agree – that the process in question existed but is not deployed

successfully

Slightly disagree – that the process was in an embryonic stage

Strongly disagree – no process existed

No opinion/Don‟t know

For the purposes of this report, any person who answered with a „No opinion/Don‟t know‟

response has been removed from the graphs, therefore the total % may not always equal 100%.

We would like to thank all of those who took the time to complete the survey and hope the report

is useful in initially understanding how they compare to their peers.

Page 9: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 9 Copyright © 2011 Experimentus Ltd

The Practice areas of TMMi Level 2 in detail

Overall rating The following chart reflects a 35% increase in Level 2 maturity for all respondents since the

survey results published in March 2009. However no one has yet reached Level 3.

0% 20% 40% 60% 80%

TMMi Level 1

TMMi Level 2

72.5%

27.5%

63.0%

37.0%

2010/2009 Maturity accross respondents

2010

2009

What does being in TMMi Level 1 mean?

TMMi Level 1 reflects that there is little or no standardisation of process. The delivery of good

testing and software quality management depends on people who know what they are doing (they

may all be doing something different and not able to understand each other‟s approach) and the

classic hero culture of the 24 hour test team who are indispensable. The problems arise when a

system matter expert‟s knowledge is not accessible by others or they are no longer available. The

impact to the business can be great, resulting not only in delays, but increased risks and costs,

many examples of which have been well publicised.

What does TMMi Level 2 mean?

TMMi Level 2 reflects that for our 2010 survey 37% of the surveyed respondents have some

processes in place, in all of the five Process Areas on which this level focuses. These are:

1. Test Policy and Strategy

2. Test Design and Execution

3. Test Planning

4. Test Environment

5. Test Monitoring and Control

The small percentage of people with Level 2 is surprising, considering the number of industry

sectors in the survey who have mission-critical applications.

Looking at the results across

the 9 industry sectors who

responded, we saw a swing in

respondents from the 2009

survey to more IT support-

driven respondents, which we

believe is related to the growth

in test outsourcing and the

subsequent need for IT

support services to ensure

they are as mature as possible

to remain efficient and effective

Industry Sectors

TMMi Level 2

TMMi Level 1

Page 10: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 10 Copyright © 2011 Experimentus Ltd

for their clients. For this survey the finance respondents, although at a lower volume than the last

survey, do seem to be showing growing maturity in their test process.

Test Policy and Strategy This Process Area looks at how well established is the organisational view of software testing,

specifically looking at the two key organisational test documents, test policy and test strategy and

how they are deployed and understood.

A test policy has been established and agreed by the stakeholders and is aligned to

business quality policies.

A test policy is the

executive sign-off that

confirms the goals and

objectives of testing in

an organisation and

ensures that the

objectives of the testing

area is aligned with the

needs of the business

and clearly understood

and agreed by all. So if

getting to market quickly

is the business

objective, then testing

would be directed to

enable speed to market. Matching test and business objectives makes sense in order to better

serve the business.

Survey results explained: 2010 sees an increase in those working with an established test

policy from 25% in 2009 to 29%. However, 71% are working without any commitment from

executive management that what they are doing is aligned to the needs of the business.

Experimentus view: It is key that executive sponsorship for testing is obtained if it is to be taken

seriously, hence the need for a test policy. It is good news that more companies do generate test

policies but it is still very disappointing that 43% have a policy that has not been deployed, only

1% less than 2009. Working under a clear policy to deliver better quality outcomes combined with

the resultant efficiency savings is now seen to be gaining some momentum. Typical reasons for

not deploying a test policy include: the value of having a test policy not being clearly understood

by senior management; the test community having not clearly articulated the benefits to

management; that policies are generated but perhaps fall into misuse because they are poorly

written, understood and communicated. It seems pointless to put the effort into obtaining

executive authorisation and then not using that in a more open way.

Effective policies sponsored by the business typically deliver efficiency and productivity savings.

An organisation-wide or programme-wide test strategy has been established and

deployed.

The role of a test strategy is to define the detail of how testing will be implemented, either for an

organisation, programme of work or project. At the organisational level it may be more of a

framework of process and controls from which projects can select, dependent on risk and

development approach.

0% 10% 20% 30% 40% 50%

no process exists

embryonic stage

exists but is not deployed successfully

well established

15%

13%

44%

25%

12%

14%

43%

29%

A test policy has been established

2010

2009

It is key that

executive

sponsorship for

testing is

obtained if it is to

be taken

seriously

Page 11: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 11 Copyright © 2011 Experimentus Ltd

At the programme and

project level it is used to

specify how testing will be

carried out, thus ensuring

that at all levels of the

organisation, a common,

and more importantly, a

reusable process is

deployed. This ensures

that each tester is able to

work in a consistent,

repeatable manner, with

any process improvements

made being reflected

throughout the targeted

area.

Survey Results explained: Disappointingly, the results for fully deployed test strategies have

dropped to 29% for this survey from 33% in 2009. In contrast the amount of organisations that

have a test strategy but have yet to fully deploy it has gone up from 35% to 39% - perhaps

indicating that some existing test strategies are falling into non-use or are in the process of being

revised.

Experimentus view: With less than a third of respondents operating with an effective test

strategy, as we highlighted in 2009 it raises the question, how effective is the testing, and what

value is it adding to the organisation? These organisations are running tests, but are they focused

on the objectives of the project, or merely what the tester believes should be done? Working

without a test strategy is like using a map in the dark; you might get to where you want, but you

will probably get lost on the way, take a lot longer and spend a lot more time and money than you

planned!

These results are surprising because most testers will say they work to a strategy, but this does

not seem to be qualified by the results. For those who do, it is not clear where they get their

focus, if they do not have any test policy.

Test Design and Execution This Process Area looks at the approach to the design of test, how well tests are executed

against the software under test and how well defects are managed and tracked to a satisfactory

solution.

Test scripts are developed, documented and prioritised.

Test scripts describe the

tests that will be run to

prove that the software

under test meets its

quality targets. Prioritising

and documenting these

scripts ensures that the

risks at release are lower,

while the quality of the

delivery is high.

Result: The very positive

news is that the volume of organisations that document and prioritise their test scripts has grown

from 53% to 57%, with no one saying that they do not do any prioritisation or documentation.

With less than a

third of

respondents

operating with an

effective Test

Strategy it raises

the question,

how effective is

the testing, and

what value is it

adding to the

organisation?

We still see the

production of test

scripts as more

important than

having a test

strategy or

policy, so no

goals, just tests!

0% 10% 20% 30% 40%

no process exists

embryonic stage

exists but is not deployed successfully

well established

17%

13%

35%

33%

16%

14%

39%

29%

Organisation/programme-wide test strategy has been established and deployed

2010

2009

Page 12: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 12 Copyright © 2011 Experimentus Ltd

Experimentus view: It is still very surprising that as an industry we still see the production of test

scripts as more important than having a test strategy or policy (so, no goals, just tests). It is not all

negative as the results for this statement are better overall, with no one suggesting that they do

not document or prioritise test scripts being a big step forward. There is no excuse not to

document tests. All development methods expect documented test scripts whether structured,

iterative or agile (albeit some may be produced to record the test that took place rather than as

the test to be executed – exploratory testing for example).

A Test Approach based on identified risks is established and agreed upon.

The test approach sits

below the test strategy in

the test documentation

hierarchy (or can form

part of the test strategy)

and confirms the detail of

how testing will actually

be implemented for a

project or program e.g.

templates as well as

process. This enables a

tactical approach to

dealing with risks.

Result: A small increase from 2009 from 35% to 37% of respondents actively using a risk

process when defining their test approach.

A more positive result is seen if we look at those in the strongly disagree category (who do not

use risk at all) with a large decrease from 13% to 4%.

Experimentus view: With risk assessment forming an important cornerstone to testing and with

the amount of conferences focussed on risk-based testing, it is encouraging that there is an

increased awareness and use of risk-based testing, but surprising that this statement did not get

significantly better results. Our experience reflects these results in that many organisations

struggle to get a repeatable risk process established.

Incidents found during testing are reported using an incident classification scheme and

managed using a documented procedure.

Incident classification

relates to the process

by which the priority

and severity of defects

is defined. The incident

process governs how

defects progress from

identification through to

classification. This

enables strategies to be

developed to reduce

the amount of incidents

and thereby costs and

time of re-work.

Result: This result has

remained constant between surveys at 73%. However, those who have now got a procedure

significantly increase from 10% to 24%.

0% 10% 20% 30% 40%

no process exists

embryonic stage

exists but is not deployed successfully

well established

13%

13%

38%

35%

4%

12%

39%

37%

A test approach, based on identified risks agreed

2010

2009

0% 20% 40% 60% 80%

no process exists

embryonic stage

exists but is not deployed successfully

well established

8%

6%

10%

73%

4%

0%

24%

73%

Incident classification scheme and process implemented

2010

2009

It is encouraging

that there is an

increased

awareness and

use of risk based

testing

Page 13: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 13 Copyright © 2011 Experimentus Ltd

The Experimentus view: Incident management continues to show it is the most intuitive process

within the testing lifecycle. It is the one thing that the test industry can be proud it does well! This

is not surprising but we cannot say whether the incidents are consistently classified, meaningful

and actionable.

Test Planning This Process Area looks at the processes of the test plan development including estimating.

A Test Plan is established and maintained as the basis for managing testing and

communication to stakeholders.

A test plan is the

document that defines

what the day to day

activity on a testing

project is. To be useful it

needs to be updated as

the project/programme

changes. It plays a large

part in ensuring that

testing starts and

completes on time and

that risks and any

delays are identified

early.

Result: Another maturity increase from 54% to 57% who are now documenting their plans and

maintaining the plan once agreed.

The Experimentus view: Although the numbers have improved from 2009 there are still many

test projects working without a plan, this is not good. Not having a plan is like getting lost and

trying to get home with a blank sheet of paper; you do not know where you are or when you will

arrive at your destination!

An approach to test estimation based on a repeatable process and/or historical data is in

place.

A mature organisation

bases its estimation on

the time, people and

environments required

to deliver the test

project on a common

repeatable process.

This ensures that as

actual post release data

becomes known, the

estimating data can be

refined to provide more

and more accurate

estimates in the future.

Result: A significant increase on the last survey with a massive increase from 17% in 2009 to

39% in 2010 of those who are now using a repeatable estimation process.

The Experimentus view: This is the best result of the 2010 survey and indicates a big step

forward for software test planning, and a better understanding of the value of a repeatable

process for estimating. A good, well-understood, repeatable estimation process ensures that

0% 10% 20% 30% 40% 50% 60%

no process exists

embryonic stage

exists but is not deployed successfully

well established

8%

8%

27%

54%

0%

8%

35%

57%

A test plan is established and maintained

2010

2009

0% 10% 20% 30% 40%

no process exists

embryonic stage

exists but is not deployed successfully

well established

13%

33%

33%

17%

10%

22%

27%

39%

Test estimation based on a repeatable process is used

2010

2009

A massive

increase from

17% in 2009, to

33% in 2010 who

are now using a

repeatable

estimation

process

Incident

Management

continues to

show it is the

most intuitive

process in the

testing lifecycle.

It is the one thing

that the test

industry can be

proud it does well

Page 14: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 14 Copyright © 2011 Experimentus Ltd

software testing planning will be better validated and have a significantly better chance of

obtaining the budget it requires.

Test Environment This Process Area looks at how well-planned and implemented are the test environments.

Test environments are specified early and their availability is ensured on time in projects.

Understanding the test

environment requirements

early enough and in

sufficient detail, enables

costs to be built in early to

a project plan, thus

providing early visibility to

the business. The details

of any new environments

should be provided by the

technical architects with

the test team deciding

when they are needed

and how they are to be

delivered (phased or all

together). Over time, this will ensure efficiency and accuracy of test environments with associated

costs which have been planned for.

Result: The number of respondents who specify their test environments early has dropped from

40% in 2009 to 37% in 2010, but the number with a process not yet deployed for this area has

gone up by nearly 10%, which is a positive sign.

The Experimentus view: Test environments are an area often blamed for delays in testing due to

poor construction or generally not getting what is requested. The drop in the number planned

early and available and working on time is perhaps a signal of the fact that as testers we need to

recognise - that sometimes we are not as effective as we should be at defining the environments

we require and getting them booked and implemented in time!

For later test stages (e.g. system or acceptance testing), the test environment is as ‘’real

life’’ as possible.

For the later stages in

testing, to ensure that the

software will work in

production, it is crucial

that the test environment

replicates the production

environment as far as

software/hardware and

configuration is

concerned. It may be

smaller than a production

environment in terms of

data, but not so small that

it will not react in the

same way. Increased

risks in deployment are mitigated by understanding the relationship between this and the live

environment.

0% 10% 20% 30% 40% 50%

no process exists

embryonic stage

exists but is not deployed successfully

well established

8%

17%

33%

40%

6%

16%

41%

37%

Test environments are specified early and their availability is ensured

2010

2009

0% 10% 20% 30% 40% 50% 60%

no process exists

embryonic stage

exists but is not deployed successfully

well established

4%

13%

25%

56%

4%

10%

29%

55%

For later test stages the test environment is as ‘’real life’’ as possible

2010

2009

Test

Environments is

an area often

blamed for

delays in testing

Page 15: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 15 Copyright © 2011 Experimentus Ltd

Result: Very little change from the 2009 results, with 55% of test environments reflecting real life.

The Experimentus view: It continues to be a good sign that the system and user acceptance

test phases do use production like environments. Good environments and data ensure that the

outcome of the test execution, positive or negative, can be believed. This will continue to be a

challenge for organisations with complex systems and is perhaps reflected in the increases seen

in more test approaches which are based on risk.

Test Monitoring and Control This Process Area is focussed on the collection and use of metrics to report progress and identify

possible improvements to the process.

A set of goal oriented test process performance indicators have been established and

deployed.

For testing to measure

whether or not it is effective,

it needs to set itself some

performance indicators such

as achieving a certain level

of Defect Detection

Percentage (DDP). You

cannot manage what you do

not measure.

Result: A significant drop

from 2009‟s results, from

30% to 14%. With a

significant increase in those

that have a set of

performance indicators but

have not yet implemented them fully.

The Experimentus view: In our opinion this is not only the worst result of the entire survey, but

the industry seems to have moved backwards in its use of metrics. The increase in those

organisations that have indicators, but have not implemented them fully, may be a refection of the

goals originally set up not providing the right level of meaningful and actionable measurement.

Organisations may therefore be revisiting their performance indicators, which is seen in the

increase from 25% to 49%. Interestingly there is a decrease of around 50% in organisations that

previously did not have test goals, reflecting an increased take up of performance indicators by

the more immature organisations.

Progress of testing is monitored against the Test Plan.

We saw in „Test

Planning‟ that it is

important to keep the

plan up to date. It is

equally important to

measure actual activity

against planned activity.

Any slippage can then be

identified and managed

before it creates any real

issues.

Result: A significant

increase from 42% in

2009 to 59% in 2010 of

0% 10% 20% 30% 40% 50%

Dont have test goals

Dont think we have test goals

Might have test goals

Have test goals

25%

17%

25%

29%

14%

22%

49%

14%

A set of goal oriented test process performance indicators has been established and deployed

2010

2009

0% 10% 20% 30% 40% 50% 60%

no process exists

embryonic stage

exists but is not deployed successfully

well established

8%

13%

35%

42%

4%

8%

29%

59%

Progress of testing is monitored against the Test Plan

2010

2009

In our opinion,

this is not only

the worst result

of the entire

survey, but as an

industry we seem

to have moved

backwards in our

use of metrics

Page 16: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 16 Copyright © 2011 Experimentus Ltd

those who do track progress against the agreed plan.

The Experimentus view: It is very encouraging to see test organisations getting better at

progress monitoring. It is still a long way from good, but it is improving. A project that does not

monitor its progress has no clue where it is or when it will finish!

Conclusion The survey results show that we are improving as an industry with the overall maturity of testing

increasing over the two year period since the last survey. The signs are that this will continue as

more and more companies realise the benefits that improvements to software quality

management and test processes can bring.

The signs from the survey are encouraging from those organisations that were previously the

least mature, making positive improvements to the way they work, with more mature

organisations looking to and using testing efficiency as a way of meeting current budget restraint.

There seems to be a more focussed, professional attitude which needs data to substantiate

current activities and to measure improvements made.

In the last 2 years a lot of economic and political change has taken place that is driving down the

budget available for IT. This has focus, and will continue to focus, everyone on being more

efficient or effective at what they do. There is of course the alternative where process and

methods just get ignored in the haste to save money. This as will be obvious to anyone reading

this, leads to inflated costs, at less quality delivered than expected and often delivered late - all

those things which the business regularly bring up as areas for improvement.

The improvement in maturity overall is heading in the right direction and it is expected that as

software quality and testing come under increasing pressure in the coming year that the rate of

increase will be greater year on year to reflect both economic and competitive advantage.

Page 17: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 17 Copyright © 2011 Experimentus Ltd

Industry Overview

The challenge we see for the industry is to increase the rate of maturity of software quality

management and test processes, to enable testing to be clearly labelled as a profession within

the application life cycle. The ambition of testing to be recognised as a profession will never be

realised unless there is measurement in place to manage and report on activities which assist in

creating roadmaps for self improvement and the delivery of better quality software on time and in

budget.

Management also have a considerable amount of hidden costs in the development lifecycle

which, without metrics, they cannot hope to understand or control. It does raise the question as to

why any organisation would want to run without consistent, meaningful and actionable metrics.

Particularly now that the TMMi Foundation has completed the development of the full TMMi model

by adding Maturity Levels 4 and 5 covering „Management and Measurement‟ and „Optimisation‟,

there is no reason why the rate of improvement of maturity cannot be increased further. The

spread of appropriate good practices into software quality management and test processes can

only be a good thing, raising the professionalism of testing and management‟s capability to

manage.

We certainly expect the growth in the maturity of testing to continue. We propose that in order to

maximise the benefits of improvements, the industry must focus in the areas of goal setting and

metrics/reporting as an established practice, to demonstrate to the business the value provided

and progress made, as well as delivering input into determining the roadmap for further

improvements to meet the challenges in the coming years. The short-term implication of this will

be to bring into sharp focus the wasted cost and effort which exist hidden in the development

lifecycle.

Experimentus has seen a number of trends in the industry which reflect an increase in the desire

to improve the way we manage software quality and testing:

System Integrators and specialist test companies are under increasing pressure from

their clients and prospects to demonstrate their real commitment to delivering and

improving quality. We are now increasingly being asked to certify their processes

against good appropriate practices in order for them to provide competitive and

demonstrable differentiation

Organisations who are still suffering from the same problems they have had for some

years, now are under increasing pressure from the business to solve those problems so

that they can better compete

The move from “would like to” to “must” regarding increasing efficiency and savings

while maintain or enhancing the quality of deliverables

For International organisations, test and quality management capability assessments

are on the increase to provide frameworks for the better management and delivery of

quality and identify hidden costs

Now that the full TMMi model has been published and with the anticipated increase in the number

of accredited assessors, there is every reason and opportunity for organisations to achieve and

increase their rate of maturity, improve quality and eliminate those hidden costs.

It is also interesting to note that the number of people joining the TMMi Foundation have

increased dramatically with the model gradually being adopted by many organisations.

The spread of

appropriate good

practices into

software quality

management and

test processes

can only be a

good thing

Page 18: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 18 Copyright © 2011 Experimentus Ltd

Appendix 1

Background to the TMMi Model and the TMMi Foundation The Test Maturity Model integrated (TMMi) has been developed to complement the existing CMMI (Capability Maturity

Model Integrated) framework and to specifically deal in detail with software quality.

It provides a structured presentation of maturity levels, allowing for standard TMMi assessments and certification, enabling

a consistent deployment of the standards and the collection of industry metrics.

TMMi has a rapidly growing uptake across Europe, Asia and the USA and owes its popularity to being the only

independent test process measurement method.

The independent TMMi Foundation initiative (www.tmmifoundation.org) has been established with the sole intent of

developing the TMMi standard.

The TMMi Foundation provides:

A standard, staged TMMi model that can be used in isolation or in support of other process improvement models.

TMMi Assessment Method Application Requirements (TAMAR) for TMMi in accordance with ISO15504 and the

process to certify commercial assessment methods against the standard model.

Certification and training/examination process, procedures and standards for formal, public accreditation of

Assessors and Lead Assessors and the on-going management.

An independently managed data repository to support TMMi assessment method accreditation, assessor and

assessment certification/validation and validated assessment data and certificates.

Page 19: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 19 Copyright © 2011 Experimentus Ltd

Page 20: Test Maturity Model integrated (TMMi) Survey resultsdocs.media.bitpipe.com/io_10x/io_102267/item... · The use of the model has already been established by both service providers

TMMi Survey Results - White Paper - 2011 update v1.0 20 Copyright © 2011 Experimentus Ltd

LISTEN

We actively listen to our clients‟ enterprise

computing issues and challenges, giving careful

consideration to their individual requirements.

CHALLENGE

Focussed on consistently improving operational

efficiencies and providing confidence to make

risk-management decisions, we pioneer new

ways to challenge our clients‟ thinking.

UNDERSTAND

Our proven expertise and broad understanding is

based on over 25 years of consultancy, analysis

and testing of large enterprise applications.

INTERPRET

Our unique analytical approach and

commitment ensures that our clients‟

requirements are translated into first class

deliverables with measurable results

and productivity gains.

CREATE

We create superior quality solutions that

deliver not only innovation, but help our

clients to reduce cost, mitigate risk

and enhance revenue.

Experimentus Ltd

17a Dorset Square

London

NW1 6QB

Tel:+44 (0)207 871 2300

Web:[email protected]

www.experimentus.com