Displaying items by tag: database
Using automation to modernize mainframe applications will bring a codebase to today’s common coding standards and architectures. But in many cases, modernization to an application’s functional equivalent isn’t always enough. Organizations can do more to make their modern code more efficient and readable. By building refactoring phases into their modernization projects, organizations can eliminate the Pandora’s box of dead or non-functional code that many developers don’t want to open, especially if it contains elements that just don’t work.
Using TSRI’s automated refactoring engine, remediation was complete in an hour.
What is Refactoring and How is it Used?
Refactoring, by definition, is an iterative process that automatically identifies and remediates pattern-based issues throughout a modernized application’s codebase. For example, unreferenced variables or unnecessary redundant snippets could exist throughout the application. This scan, known as dead/redundant code refactoring, will find repetitions of any of this unusable code to flag, then remove it from the codebase. One of TSRI’s current projects found 25,000 instances of a similar issue that would have required 15 minutes of manual remediation per instance—not including the inevitable introduction of human error that would require further remediation. The number of development hours would take more than a year for a single developer to complete.
Using TSRI’s automated refactoring engine, however, remediation was complete in an hour.
Calling refactoring its own post-modernization phase is, in some ways, misleading. Refactoring typically occurs all the way through an automated mainframe transformation. As an example, in a typical COBOL or PL/1 mainframe modernization, TSRI would refactor the code from a monolithic application to a multi-tier application, with Java or C# handling back-end logic, a relational database layer through a Database Access Object (DAO) layer, and the user interface (screens) modernized in a web-based format. Believe it or not, many legacy applications still run on 3270 green-screens or other terminals, like in the graphic below.
Once the automated modernization of the legacy application is complete, the application has become a functionally equivalent, like-for-like system. However, any deprecated code, functions that may have never worked as planned, or routines that were written but never implemented will still exist. A process written in perhaps 1981—or even 1961—may have taken far more code to execute than a simple microservice could handle today.
Situations like this are where refactoring becomes indispensable.
Where to begin?
Before a formal refactoring process can begin, it’s important to understand your goals and objectives, such as performance, quality, cybersecurity, and maintainability. This will typically mean multiple workshops to define which areas of the modernized codebase need attention and the best candidates for refactoring, based upon the defined goals. These refactorings will either be semi-automated (fully automated with some human input) or custom written (based upon feedback from code scanners or subject-matter experts.)
The refactoring workshops can reveal many different candidates for refactoring:
- Maintainability: By removing or remediating bugs, dead or orphaned code, or any other anomalies the codebase can be reduced by as much as one third while pointing developers in the direction of any bugs in need of remediation.
- Readability: Renaming obscure functions or variables for a modern developer to fit within naming conventions that are both understandable and within the context of the code’s functionality.
- Security: Third-party tools such as Fortify and CAST can be utilized to find vulnerabilities, but once found they need to be remediated through creation of refactoring rules.
- Performance: Adding reusable microservices or RESTful endpoints to connect to other applications in the cloud can greatly improve the efficiency of the application, as can functionality that enables multiple services to run in parallel rather than sequentially.
What are the Challenges?
- Challenge 1: One reason refactoring must be an iterative process is that some functionality can change with each pass. Occasionally, those changes will introduce bugs to the application. However, each automated iteration will go though regression testing, then refactored again to remediate those bugs prior to the application returning to a production environment.
- Challenge 2: The legacy architecture itself may pose challenges. On a mainframe, if a COBOL application needs to access data, it will call on the entire database and cycle through until it finds the records it needs. Within a mainframe architecture this can be done quickly. But if a cloud-based application needs to call a single data record out of millions or billions from halfway across the world (on cloud servers), the round trip of checking each record becomes far less efficient—and, in turn, slower. By refactoring the database, the calls can go directly to the relevant records and ignore everything else that exists in the database.
- Challenge 3: Not every modernization and refactoring exercise meets an organization’s quality requirements. For example, the codebase for a platform that runs military defense systems is not just complex, it’s mission critical. Armed forces will set a minimum quality standard that any transformation must meet. Oftentimes these standards can only be achieved through refactoring. A third-party tool like SonarQube in conjunction with an automated toolset like TSRI’s JANUS Studio® can be utilized to discover and point to solutions for refactoring to reach and exceed the required quality gate.
In conclusion, while an automated modernization will quickly and accurately transform legacy mainframe applications to a modern, functionally equivalent, cloud-based or hybrid architecture, refactoring will make the application durable and reliable into the future.
TSRI is Here for You
As a leading provider of software modernization services, TSRI enables technology readiness for the cloud and other modern architecture environments. We bring software applications into the future quickly, accurately, and efficiently with low risk and minimal business disruption, accomplishing in months what would otherwise take years.
"New Book by Ulrich and Newcomb: Information Systems Transformation:
Architecture-Driven Modernization Case Studies with
|Kirkland, WA. (Feburary 22, 2010) – Book Release
Information Systems Transformation: Architecture-Driven Modernization Case Studies
By William M. Ulrich and Philip H. Newcomb
Published by Morgan Kaufmann
Copyright Feb 2010
$59.95 USD €43.95 EUR £29.99 GBP
|What The Experts Are Saying:
According to Grady Booch, IBM Fellow & Chief Scientist, Software Engineering:
"Ulrich and Newcomb's book offers a comprehensive examination of the challenges of growing software-intensive systems … (Read more...)
According to Ed Yourdon, noted Author and Consultant:
"Modernization is going to be a more and more important part of the overall IT strategy. William Ulrich and Philip Newcomb's important new book ... (Read more...)
According to Richard Soley Ph.D. Chairman/CEO, Object Management Group (OMG):
“Estimates by internationally-known researchers of the worldwide legacy code base is now approaching a half-trillion lines. That only counts so-called "legacy languages" like COBOL--which drive the world. Add in database schemas … (Read more...)
About the Book
Information Systems Transformation: Architecture-Driven Modernization Case Studies, a new book by William Ulrich and Philip Newcomb, provides a practical guide to organizations seeking ways to understand and modernize existing systems as part of their information management strategies. It includes an introduction to ADM disciplines and standards, including alignment with business architecture, as well as a series of scenarios outlining how ADM is applied to various initiatives. Ten chapters, containing in-depth, modernization case studies, distill the theory and delineate principles, processes, and best practices for every industry, ensuring the book's leading position as a reference text for all of those organizations relying on complex software systems to maintain their economic, competitive and operational viability. (Read more...)
|William M. Ulrich is President of Tactical Strategy Group, Inc. (TSGI)
and a management consultant. Mr. Ulrich has been in the modernization field since 1980 and continues to serve as a strategic advisor on business and IT transformation projects for corporations and government agencies. In 2005, Mr. Ulrich was awarded the Keeping America Strong Award for his work in information systems modernization. He is Co-Chair of the OMG Architecture-Driven Modernization Task Force and the OMG Business Architecture Special Interest Group, Editorial Director of the Business Architecture Institute, and author of Legacy Systems: Transformation Strategies.
|Philip H. Newcomb is Founder and CEO of The Software Revolution, Incorporated (TSRI)
and creator of TSRI's acclaimed architecture-driven modernization services and toolset JANUS Studio® . He is coauthor of Reverse Engineering (Kluwer 1996) with Linda Wills, Coeditor of the 2nd Working Conference on Reverse Engineering (IEEE 1995) with Elliot Chikofsky and principal author of the Abstract Syntax Tree Metamodeling Specification (OMG Specification 2009). With more than 35 publications and 70 successfully completed information system modernization projects he is a recognized leader in the application of artificial intelligence, automatic programming and formal methods to industrial-scale software modernization.
|About Morgan Kaufmann:
For more information about TSRI contact:
Vice President of Sales
Phone: (425) 284-2770
Fax: (425) 284-2785