3 Ways DevOps Teams Can Accelerate CI/CD
April 19, 2018

Matt Geddes
Tintri

As development speed has become a competitive advantage, the DevOps team has sought to enable continuous integration and continuous delivery (CI/CD).

CI/CD helps developers overcome enormous strains on the software development and testing process. Over the years, software development has migrated from the enterprise level, where there are vast resources, to smaller development shops that race to keep pace with demand generated from billions of smartphones and other mobile consumer devices.

One of the biggest challenges to a successful software development cycle is the management and integration of disparate information. Think about a group project where everyone divides and conquers on tasks with a goal of integrating their work at a later date. At this point, information can come in different formats, languages and styles.

Before CI/CD, developers worked on specific software modules and merged with others later for integration and testing, which would take weeks or months to address bugs and issue fixes for a final release. We all know this process as the waterfall method. With continuous integration (CI), developers now merge changes into the main branch of the software on a daily basis. Programmers can protect against potential risks by automating testing, which helps prevent integration problems commonly associated with the waterfall method. Continuous delivery (CD) promotes a steady flow of software releases through automation. For some applications, deployment to customers may take place almost immediately after tests are completed.

For the CI/CD process to be successful, it must be fast and efficient. Any potential roadblocks that delay any part of the process increase cycle times and slow down delivery.

Typically, those who manage the IT infrastructure are seen as being responsible for providing the resources needed to support the CI/CD process. But it's also the job of operations to make sure infrastructure facilitates agile development processes that enable CI/CD with minimal bottlenecks in the pipeline.

Operate at VM and Container Level

Development workflows typically run on VMs or in containers, and so automation should be built at the same level of granularity. That requires the right storage, which can greatly accelerate CI/CD workflows by doing away with legacy logical unit numbers (LUN) or volumes — operating in VMs and containers instead.

The key is to automate everything directly at the VM, vDisk, or container level to remove complexity and make it possible to leverage storage-related processes to accelerate your CI/CD pipeline. Snapshots, cloning and replication — common features of all-flash storage — can also be factored in to save time and storage capacity.

Remove Capacity and Performance as a Constraint

It's not uncommon for storage capacity and performance to become bottlenecks in the CI/CD pipeline. If the development team has to sit and wait for processes to complete or for capacity and performance to become available, it can affect software build times and slow down testing.

By applying flash storage at the VM and container level, you can automatically prevent CI/CD workloads from interfering with each other and bogging things down. This allows you to spin up tens, hundreds, or even thousands of test workloads in parallel, and you never have to worry about load balancing or the placement of different types, sizes, or numbers of VMs inside LUNs.

Gains in performance efficiency can also be achieved by offloading operations. For example, if an application includes databases, each person on the development team needs to work with and test against their own database. Given the size of modern databases, if you try to give every developer a physical copy, storage capacity will quickly become an issue, not to mention the time spent waiting for physical copies to complete.

VM-level cloning allows you to have as many environments with full database copies as you need and to create new environments almost instantaneously.

Synchronize VMs to Accelerate Provisioning of Development and Test Environments

It's important to refresh test workloads with the latest data from production as often as possible. One of the best ways to avoid bugs and integration problems is to perform integration tests with production data. However, to accomplish this testing, you have to refresh the data and make multiple copies and rebuild the environment, including all the necessary VM settings. This process can be cumbersome, time-intensive and error-prone.

But by synchronizing VMs, you can accomplish refreshes and streamline the process without rebuilding the entire environment. As a result, there's less disruption and you can refresh regularly.

By factoring the above into your infrastructure design and purchasing decisions, you can help ensure your DevOps teams never run out of resources and spend their time focused on code, not infrastructure.

Matt Geddes is Field CTO at Tintri

The Latest

August 15, 2018

Microservices are a hot topic in IT circles these days. The idea of a modular approach to system building – where you have numerous, smaller software services that talk to each other instead of monolithic components – has many benefits ...

August 13, 2018

Agile is expanding within the enterprise. Agile adoption is growing within organizations, both more broadly and deeply, according to the 12th annual State of Agile report from CollabNet VersionOne. A higher percentage of respondents this year report that "all or almost all" of their teams are agile, and that agile principles and practices are being adopted at higher levels in the organization ...

August 09, 2018

For the past 13 years, the Ponemon Institute has examined the cost associated with data breaches of less than 100,000 records, finding that the costs have steadily risen over the course of the study. The average cost of a data breach was $3.86 million in the 2018 study, compared to $3.50 million in 2014 – representing nearly 10 percent net increase over the past 5 years of the study ...

August 08, 2018

Hidden costs in data breaches – such as lost business, negative impact on reputation and employee time spent on recovery – are difficult and expensive to manage, according to the 2018 Cost of a Data Breach Study, sponsored by IBM Security and conducted by Ponemon Institute. The study found that the average cost of a data breach globally is $3.86 million ...

August 06, 2018

The previous chapter in this WhiteHat Security series discussed dependencies as the second step of the Twelve-Factor App. This next chapter examines the security component of step three of the Twelve-Factor methodology — storing configurations within the environment.

August 02, 2018

Results from new Forrester Consulting research reveal the 20 most important Agile and DevOps quality metrics that separate DevOps/Agile experts from their less advanced peers ...

July 31, 2018

Even organizations that understand the importance of cybersecurity in theory often stumble when it comes to marrying security initiatives with their development and operations processes. Most businesses agree that everyone should be responsible for security, but this principle is not being upheld on a day-to-day basis in many organizations. That’s bad news for everyone. Here are some best practices for implementing SecOps ...

July 30, 2018

While the technologies, processes, and cultural shifts of DevOps have improved the ability of software teams to deliver reliable work rapidly and effectively, security has not been a focal point in the transformation of cloud IT infrastructure. SecOps is a methodology that seeks to address this by operationalizing and hardening security throughout the software lifecycle ...

July 26, 2018

Organizations are shifting away from traditional, monolithic architectures, with three-quarters of survey respondents delivering at least some of their applications and more than one-third delivering most of their applications as microservices, according to the State of DevOps Observability Report from Scalyr ...

July 24, 2018

What top considerations must companies make to ensure – or at least help improve – Agile at scale? The following are key techniques and practices to help accelerate Agile delivery rollouts and scale Agile and DevOps in the Enterprise ...

Share this