Welcome!

Containers Expo Blog Authors: Liz McMillan, Pat Romanski, Yeshim Deniz, Elizabeth White, Zakia Bouachraoui

Related Topics: @DevOpsSummit, Linux Containers, Containers Expo Blog

@DevOpsSummit: Blog Feed Post

Scaled Agile Factory - Going Beyond the Usual Suspects | @DevOpsSummit #DevOps

The question of how to scale Agile development has been around a while

If you address the question of how to scale Agile projects by considering what framework to use, you are only looking at one aspect of the problem. Scaling is all about coordination – managing enterprise considerations and cross program dependencies, and the defacto frameworks (SAFe, LeSS and DAD) focus on the people and process dimensions. However, in combination with a factory approach you may be able to automate many of the compliance and dependency management issues.

The question of how to scale Agile development has been around a while. In January 2015 I commented [1] on a clear trend in which my customers were voicing concerns about loss of consistency; inability to govern; lack of coordination and increasing time to market. Since then I have observed many large organizations adopting SAFe (Scaled Agile Framework) perhaps because it appears to be the only game in town, or perhaps there’s good marketing or there’s strength in numbers. But the criticisms of SAFe [2] haven’t gone away. The central and continuing concern is that SAFe compromises core Agile principles of self-organizing, cross functional teams that have full responsibility for the delivery of potentially shippable increments. And that SAFe looks suspiciously like WaterScrumFall because there’s a high overhead of portfolio, value stream and project management and in all probability intentional architecture. Now I have mixed opinions on SAFe; I see positives in value streams and product management which are important. And for organizations that have large programs, highly organized, structured approaches will be seen as lower risk, and indeed something that takes them some way down the Agile path, without straying too far from conventional management comfort zones. But the outcomes are unlikely to be inherently “agile”.

What SAFe does, is provide a rather conventional approach for a complex problem that most enterprises have – that is to deliver high integrity solutions that can work in an enterprise context that demands cross project dependency management, consistent data and reference architecture, collaboration with the existing portfolio complexities, compliance with enterprise standards and governance etc.

There are alternatives. Craig Larman and Bas Vodde in their books [3] and more recently with their LeSS initiative [4] have pursued a very different path that starts with Scrum and scales by understanding the needs for coordination while adhering to core Scrum principles. In their forthcoming book [5] they say, “LeSS is (1) lightweight, (2) simple to understand, and (3) difficult to master—due to essential complexity”. And this allows us to contrast the different approaches; while SAFe clearly works at some level, it has its roots in conventional large scale project management, whereas LeSS is lightweight, but requires much deeper understanding of the systems dynamics because of the inherent complexity of all the coordination requirements.
So the real question underlying Scaling Agile should be, “can we address some of the coordination requirements in a manner that reduces complexity and eliminates some of the need for additional layers of management or events?”

In my post Service Factory 2.0 [6] I describe the conceptual background of the Software Factory, ideas pioneered by my old friend Keith Short and Jack Greenfield while they were at Microsoft. Today these have evolved and become specialized around a framework of tools, repeatable processes and patterns for creation and assembly of services – manifest as first order components with formal interfaces. If you consider that all core business functionality is increasingly composed of services and their operations, this provides us with a reference architecture that by design implements separation of concerns.  Figure 1 below is a conceptual view of the scope of the service factory in terms of managed objects.

Figure 1

Naturally there will be other patterns involved in any solution including UX and workflow; but for most enterprise class solutions, a high proportion of the functionality should comprise services and business rules. This allows a consistent, highly automated approach to architecture, requirements specification, design and test. The factory effectively implements a proven service reference architecture as a platform which can be customized for individual programs, projects and technologies. While the factory platform is similar to an architecture runway in scope and intent, because it is model based the factory framework enables continuous evolution of the platform capabilities that are automatically inherited into work in progress factory code allowing program wide changes to be incorporated, often with minimal or no change to the service specific code. And the factory is a repository of the life cycle artifacts that can be leveraged for various management tasks including governance of traceability, impact analysis, etc.

Let’s consider the key coordination and management activities typically required in a scaled agile program and how the factory may facilitate that. In Figure 2 below I have used the SAFe layers of portfolio, program and team, albeit without the value stream level. But apart from that, the view is intended to be framework neutral. The diagram illustrates how the Factory provides a full life cycle backplane that establishes the underlying (meta) model that ensures all the participants are in compliance with core concepts from portfolio planning through to delivery and production. The populated model is therefore able to provide essential information particularly about dependencies, but equally complete traceability and governance between business requirements and delivered services and rules.
Figure 2

In the LeSS framework, Larman and Vodde disregard constructs such as ART (Agile Release Train) and 8 to 12 week PIs (Program Increments) and also the Hardening Sprint, preferring to use the standard Scrum duration of 2 to 4 weeks, with continuous integration.  We might assume these management constructs (ARTs and PIs) were invented because of the difficulty of effective inter team communications, and the overhead that creates. Larman and Vodde have a lot of good ideas about how to manage effective inter Teams communications, but their approach is essentially based on feature or story dependencies. In the factory environment inter-team dependencies can be more accurately based on service operation dependencies. But this is just the tip of the iceberg, in terms of how the factory can be leveraged to manage activity. The service specification becomes a pivotal work product; as it becomes available other teams and team members have access to detailed behaviour information that can inform and facilitate working ahead for developing depending services, test case development including automation assistance in case development, plus devops.

Another huge issue in Agile projects is managing technical debt. In the factory environment the standard patterns generate all the infrastructure and shared code, and when changes take place, as discussed above, teams can perform an architecture update, inheriting all the changes to bring their project in line with the latest reference architecture. And because the factory is driven by the specification level, independent of technology it is inherently late binding, allowing change if technology with minimum impact.

By utilizing work products that are part of a defined process based on the full life cycle reference architecture, the factory approach reduces the essential complexity of the scaling task. All moving parts are under management. This doesn’t detract from the inherent purity of the Agile approach. Rather, the defined process provides clarity over the minimum necessary “planning work” (intentional architecture, rules definition and business requirements). And the reference architecture and factory process provide a firm foundation on which emergent architecture can be more effectively carried out by the cross functional delivery team. This is likely to encourage programs to adopt a hybrid approach to their organizing model, embracing elements of LeSS like guidance to optimize multiple concurrent Scrum based sprints with continuous delivery, and to complement them with minimum necessary elements of SAFe, particularly those that are required to communicate to more conventional stakeholders. But, at least for mid-sized projects, maybe up to 10 teams, it seems overkill to adopt the formality of PI Planning and everything that comes with it.

In this post I have focused on the scaled Agile approach for projects, but it’s also useful to understand the factory has other important outcomes. First the highly modular nature of the reference architecture (see Figure 1) leads to “solutions” that are inherently agile. Because solutions are comprised primarily of services and rules with strong formal contracts, implemented using component based implementation patterns, the horizon of change is likely to be very constrained. Further the specification approach in which services, rules, data and processes are defined independent of implementation and technology provides high levels of visibility to the business, and high transparency and traceability of the implemented solution. Finally, we have a way to realize architecture in inherently agile solutions without Big Upfront Design, or indeed Big Upfront Organization, that will of course never become legacy because they can be continuously evolved.


[2] Reasoned criticisms of SAFe

[3] Scaling Lean & Agile, Craig Larman and Bas Vodde, Addison Wesley, 2009



Read the original blog entry...

More Stories By David Sprott

David Sprott is a consultant, researcher and educator specializing in service oriented architecture, application modernization and cloud computing. Since 1997 David founded and led the well known think tank CBDI Forum providing unique research and guidance around loose coupled architecture, technologies and practices to F5000 companies and governments worldwide. As CEO of Everware-CBDI International a UK based corporation, he directs the global research and international consulting operations of the leading independent advisors on Service Oriented Application Modernization.

IoT & Smart Cities Stories
The deluge of IoT sensor data collected from connected devices and the powerful AI required to make that data actionable are giving rise to a hybrid ecosystem in which cloud, on-prem and edge processes become interweaved. Attendees will learn how emerging composable infrastructure solutions deliver the adaptive architecture needed to manage this new data reality. Machine learning algorithms can better anticipate data storms and automate resources to support surges, including fully scalable GPU-c...
Machine learning has taken residence at our cities' cores and now we can finally have "smart cities." Cities are a collection of buildings made to provide the structure and safety necessary for people to function, create and survive. Buildings are a pool of ever-changing performance data from large automated systems such as heating and cooling to the people that live and work within them. Through machine learning, buildings can optimize performance, reduce costs, and improve occupant comfort by ...
The explosion of new web/cloud/IoT-based applications and the data they generate are transforming our world right before our eyes. In this rush to adopt these new technologies, organizations are often ignoring fundamental questions concerning who owns the data and failing to ask for permission to conduct invasive surveillance of their customers. Organizations that are not transparent about how their systems gather data telemetry without offering shared data ownership risk product rejection, regu...
René Bostic is the Technical VP of the IBM Cloud Unit in North America. Enjoying her career with IBM during the modern millennial technological era, she is an expert in cloud computing, DevOps and emerging cloud technologies such as Blockchain. Her strengths and core competencies include a proven record of accomplishments in consensus building at all levels to assess, plan, and implement enterprise and cloud computing solutions. René is a member of the Society of Women Engineers (SWE) and a m...
Poor data quality and analytics drive down business value. In fact, Gartner estimated that the average financial impact of poor data quality on organizations is $9.7 million per year. But bad data is much more than a cost center. By eroding trust in information, analytics and the business decisions based on these, it is a serious impediment to digital transformation.
Digital Transformation: Preparing Cloud & IoT Security for the Age of Artificial Intelligence. As automation and artificial intelligence (AI) power solution development and delivery, many businesses need to build backend cloud capabilities. Well-poised organizations, marketing smart devices with AI and BlockChain capabilities prepare to refine compliance and regulatory capabilities in 2018. Volumes of health, financial, technical and privacy data, along with tightening compliance requirements by...
Predicting the future has never been more challenging - not because of the lack of data but because of the flood of ungoverned and risk laden information. Microsoft states that 2.5 exabytes of data are created every day. Expectations and reliance on data are being pushed to the limits, as demands around hybrid options continue to grow.
Digital Transformation and Disruption, Amazon Style - What You Can Learn. Chris Kocher is a co-founder of Grey Heron, a management and strategic marketing consulting firm. He has 25+ years in both strategic and hands-on operating experience helping executives and investors build revenues and shareholder value. He has consulted with over 130 companies on innovating with new business models, product strategies and monetization. Chris has held management positions at HP and Symantec in addition to ...
Enterprises have taken advantage of IoT to achieve important revenue and cost advantages. What is less apparent is how incumbent enterprises operating at scale have, following success with IoT, built analytic, operations management and software development capabilities - ranging from autonomous vehicles to manageable robotics installations. They have embraced these capabilities as if they were Silicon Valley startups.
As IoT continues to increase momentum, so does the associated risk. Secure Device Lifecycle Management (DLM) is ranked as one of the most important technology areas of IoT. Driving this trend is the realization that secure support for IoT devices provides companies the ability to deliver high-quality, reliable, secure offerings faster, create new revenue streams, and reduce support costs, all while building a competitive advantage in their markets. In this session, we will use customer use cases...