Showing posts with label Paul Evans. Show all posts
Showing posts with label Paul Evans. Show all posts

Wednesday, May 19, 2010

HP Shows Benefits From Successful Application Consolidation With Own Massive Global Supply Chain Project

Transcript of a BriefingsDirect podcast on how HP tackled an internal multi-year effort to streamline supply chain efficiencies and effectiveness through applications consolidation.

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Download the transcript. Sponsor: HP.

Access more information on Application Consolidation.
Read the full-length case study on HP's Application Consolidation.
Learn more about the Application Transformation Experience Workshop.

Dana Gardner: Hi, this is Dana Gardner, principal analyst at Interarbor Solutions, and you’re listening to BriefingsDirect.

Today, we present a sponsored podcast discussion on learning about best practices and execution accelerators for large, often global application-consolidation projects. We'll take a look a telling case-study, a massive multi-year application-consolidation project at HP that involved hundreds of applications and thousands of people around the world.

We'll look first at why the Global Part Supply Chain project at HP was undertaken, but just as importantly, why the project needed to be invigorated after it bogged down by sheer scale and complexity. The project quickly became hugely successful, however, and we'll learn how and why.

These are by no means trivial projects, and often involve every aspect of IT, as well as require a backing of the business leadership and the users to be done well. The goal through these complex undertakings is to radically improve how applications are developed, managed, and governed across their lifecycle to better support dynamic business environments. The stakes, therefore, are potentially huge for both IT and the business.

We're here with an executive from HP to look at proper planning and execution for massive application-consolidation projects by specifically looking at an HP project itself.

Please join me now in welcoming Paul Evans, Worldwide Marketing Lead on Applications Transformation at HP. Welcome to the show, Paul.

Paul Evans: Hi, Dana.

Gardner: Tell me why applications are so numerous in many of these extended business processes, about the whole notion of part supply chains, and why that's such a big deal for HP.

Evans: As you can imagine, HP is an extremely large organization. It makes products, as well as sells services, etc. In terms of product, just imagine your average PC, or your average server, and think of the number of components that are made up inside of that device. It runs into hundreds of thousands, whether it's memory chips, disk drives, screens, keyboards, or whatever.

For a company like HP, in the event that someone needs a spare part for whatever reason, they don't expect to wait a significant period of time for it to turn up. They want it delivered 24 hours later by whatever means that suits them.

So, it's essential for us to have that global supply chain of spare parts tailored toward the ones that we believe we need more -- rather than less -- and that we can supply those parts quickly and easily and, at the same time, cost effectively. That's important for any organization that is dealing in physical components or in the provision of a service. You want to maintain customer satisfaction or increased customer satisfaction.

Customer centric

For us, it was essential that a massive global supply chain organization was extremely customer-centric, but at the same time, very cost-effective. We were doing our utmost to reduce costs, increase the agility of the applications to service the customers, and fuel growth, as our organization and our business grows. The organization has got to respond to that.

So the primary reasoning here was that this is a large organization, dealing with multiple components with pressures on it both from the business and the IT sides.

Gardner: With HP, of course, there have been mergers and acquisitions over the years. Suppliers come and go. Product lines may start small, but then ramp up rapidly. So, we're talking about many different organizations and many different applications that had to be brought into this now hopefully lean process.

Evans: One of the primary reasons we had to do this is that HP has been an amalgam of companies like Hewlett-Packard, originally, Compaq, Tandem, DEC. All of these organizations had their own bills of materials, their own skills, and basically this thing has just grown like Topsy.

What we were trying to do here was to say that we just couldn't continue to treat these systems as un-integrated. We had a lot of legacy environments that were expensive to run, a lot of redundancy, and a lot of overlap.

The whole notion of this coming about through mergers and acquisitions is very common in the marketplace. It's not unique just to HP.



The goal here clearly was to produce one integrated solution that treated the HP customer as an individual, and in the back-end consolidated the applications -- the ones we really needed to move forward. And also, a goal was to retire those applications that were no longer necessary to support the business processes.

The whole notion of this coming about through mergers and acquisitions is very common in the marketplace. It's not unique just to HP. The question of whether you just live with everybody’s apps or you begin to consolidate and rationalize is a major question that customers are asking themselves.

Gardner: If you look at this problem from the perspective of IT, of course, you have multiple platforms, legacy applications, and a mixture of different architectures and hardware. And they have their own set of requirements.

But, if you look at it through the lens of the user, many users are wed to that application, even if it's an older interface. They don't really care about what's underlying in terms of infrastructure. They just want to be able to get their work done without being disrupted.

What we have are perhaps multiple agendas that need to be aligned, and politics and persuasion come into play. Tell me what may have gone awry for a period of time with this project and how some of these other issues about multiple agendas can be managed?

Siloed thinking

Evans: Well, this is a challenge in any situation, and this has been true not only with this particular supply chain project with HP, but for all of us. The rationalization that has taken place inside HP around its IT organization and technology is that because we are human beings, most people think in a very siloed way.

They see their suite of applications supporting their business. They like them. They love them. They’ve grown up with them, and they want to continue using them. Their view is, "Mine is perfect to suit my business requirement. Why would I need anything else?"

That's okay, when you're very close to the coalface. You can always make decisions and always deem to the fact that the applications you use are strategic -- an interesting word that a lot of people use. But, as you zoom out from that environment and begin to get a more holistic view of the silos, you can begin to see that the duplication and replication is grossly inefficient and grossly expensive.

We've seen that in HP. We saw it in this particular supply chain situation. We were looking at three totally different solutions in three different companies: Compaq, HP, and DEC. We were looking over 300 applications. Clearly, that was not the way forward, because it wasn't only just a cost-reduction exercise.

If you're looking into the future and saying you need a much faster, speedier, agile situation to be working with, you can't do that in the whole legacy environment.



If you're looking into the future and saying you need a much faster, speedier, agile situation to be working with, you can't do that in the whole legacy environment. It's just something that's tying you down. That problem is not unique to HP. I definitely understand that.

Gardner: When you decided to look into your parts supply chain activities, I understand there were hundreds of applications involved, multiple sites, geographies, and countries. Was this something that was driven by the business? Was it driven by IT? Both? How did the impetus for this begin?

Evans: Well, from the IT side, there was clearly a view from the top down that said living with 300 applications in the supply-chain world was unacceptable. But also from the business side, the real push was that we had to improve certain metrics. We have this metric called Spend-to-Revenue ratio which is, in fact, what are we spending for parts as opposed to what we are getting in terms of revenue? We were clearly below par in those spaces.

We had some business imperatives that were driving this project that said we needed to save money, we needed to be able to deliver faster, and we needed to be able to do it more reliably. If we tell a customer they're going to get the part within 24 hours, we deliver in 24 hours -- not 36 or 48, because we weren't quite sure where it was. We had to maintain the business acumen.

Complexity kicks in

At the same time, when viewed from a technological angle, we were running old, expensive applications. As always, when you're running far too many applications, the complexity kicks in. How does that all work?

This volume of applications -- or applications bloat, as some people call it -- is a real impediment to agility. You just can't move forward quickly with 300 apps in an environment where you know you're probably looking at a tenth of that. It’s a bit like saying, "How could I run fast if I have 300 feet?" You can't. You can do it with two, but you couldn't do it with 300.

So, our whole goal here was to align business and IT in terms of a technological response to a business driver.

Gardner: From the business side, I suppose they're very concerned about business process, primarily, the applications. They're probably not concerned about some of the more strategic, long-term IT implications -- those being how do we better manage applications as a life-cycle.

At that point, it gives IT the opportunity to come and say, "Let's look at this methodologically. How do we now put in the governance, put in the processes where we can create the applications, manage them, but also sunset them appropriately?"

These two people were the drivers. The buck stopped with these people. They had to make the big decisions.



So how does that factor in, that notion of making this a mature process, and no longer a cherry-picking, complex stew of different styles?

Evans: The area you just addressed is probably one of the primary ones. When we submitted the project, we were basically driving it by committee. Individual business units were saying, "I need applications x, y, z." Another group says, "Actually, we need a, b, c." There was virtually no ability to get to any consensus. The goal here is to go from 300 apps to 30 apps. We’re never going to do it, if you could all self-justify the applications you need.

What we did was discard the committee approach. We took the approach, basically led by one person from the business side, who had supply chain experience, and one from the IT side who had supply chain experience, but both had their specialist areas. These two people were the drivers. The buck stopped with these people. They had to make the big decisions.

To support them, they had a sponsorship committee of senior executives, to which they could always escalate, if there was a problem making a final decision about what was necessary.

Randy Mott, the HP CIO, has the direct support of Mark Hurd, the HP chairman and CEO. In my experience, that's absolutely essential in any project a customer undertakes. They have to have executive sponsorship from the top.

If you don't, any time you get to an impasse, there's no way out. It just distills into argument and bickering. You need somebody who's going to make the decision and says, "We're going this way and we're not going that way."

Getting on track

So for us, setting up this whole governance team of two people to make the hard decisions, and their being supported by a project management team who are there to go off and enact the decisions that were made was the way we really began to move this project forward, get it on track, get it on time, and get it in budget.

Gardner: I see. So the impasse -- the bogging down of this process where it sort of went off the rails in terms of an expected time line -- that's where it was bogged down by committee versus committee. It's when you broke through that, almost at an organizational managerial level, that you were able to accelerate. Is that right?

Evans: Absolutely. In my interaction with customers, I see this time-and-time again. We’ve always said that the experience that HP has gained internally we would share with our customers. We even have a regular customer event, where we share all our best practices and we are not afraid to share the things that go wrong. In this instance, when we started by saying let's have a big committee to help my decisions, it was the wrong approach. We were going nowhere. We had to rationalize and say no.

Access more information on Application Consolidation.
Read the full-length case study on HP's Application Consolidation.
Learn more about the Application Transformation Experience Workshop.

Two respected individuals, one from the IT side and one from the business side, who were totally aligned on what they were doing, shared the same vision in what they were trying to achieve. By virtue of that, we could enforce throughout decisions, sometimes unpopular.

eople sometimes do not understand why a particular application is going to get turned off in place of another. But those were the hard realities we had to take to get the cost down and get the efficiency and the effective result.

Gardner: So, we're talking now about decision-making. We're talking about governance. We're talking about the intersection of IT governance with political governance. This is something you can't buy. You don't necessarily purchase a box that does this. This is a combination of technology, professional services, methodology, standards, experience, and even, I would imagine, a change in management among the leadership.

Tell me how HP internally focused across these multiple disciplines -- not just product, not just technology -- and then how that related to what you do with your supply chain customers now?

Evans: A lot of people would say it's just technological problem. You’ve got 300 apps running on old platforms using old technology and you want to use the latest and greatest, the fastest, the smartest ... whatever. But, as we’ve discussed, at least 50-60 percent of the solution has nothing to do with technology. It had all to do with making decisions, making the right decisions that would lead us to the right outcome for the business.

We knew what we wanted to achieve. We knew that we had to be more agile. We had to get our costs down. We had to optimize this whole spend-to-revenue ratio. As always with the supply chain business, inventory had to go down. Going up is not a good plan, because you're paying for parts that are sitting on the shelf.

Agile and sleek

One of the goals was to get the solution so agile and sleek that we didn't have to use air transportation to get parts from A to B. We could use surface transportation. If we could put the parts in the right place, where they needed to be to get to the right customer, rather having to use air, which of course is very fast, but it’s very expensive compared to surface, we could also get a dramatic reduction in the CO2 emission that we were putting out by virtue of that transportation.

There were lots of things here that had nothing to do with technology. They all had to do with business benefits and outcomes that we wanted to achieve, both internally to HP, like saving money, but also to the customers in terms of delivery of a better service.

Some will call them peripheral, and some may call them fundamental, but things like using different transportation techniques to cut CO2 we felt were pretty important.

What we've done as always with these experiences is translate them into how can we be smarter, better, and more helpful to our customers in learning from these experiences. In the whole HP-IT story, we have outpoured so many best practices and good ideas and bad ideas, which we're quite happy to share with people.

But, of course, hindsight is a wonderful thing.



Similarly, we'd like to think that those organizations that are out there with a supply chain challenge could now look at this and say, "Maybe we could do the same thing." Definitely the alignment between business and IT is probably one of the most paramount of facets. Let me do with which platform, which network, which disk drive, or which operating system. You can have a lot of fun with that. But, in this instance, a lot of the success was driven by setting up the right governance and decision-making structure with the right sponsorship.

Gardner: Now, I'd like to look at some of the paybacks. As I understand it, you turned the corner on this project back in 2006. At the time, you didn't realize that these were "the good old days." The agility and lean aspects of processes in a supply chain are great during growth, but they're also extremely important, when there is a contraction.

So, is there an opportunity to look back and say, "Wow. We didn't know it at the time, but by conducting this application consolidation with the proper governance, we were able to dial down on our delivery of products and services when that was required, and now dial back up." That probably is something of a lesson at the economic level, but can we apply some metrics of success from your project? Any thoughts about what the paybacks were, especially in a topsy-turvy general economy?

Evans: In taking the more holistic view and talking to a lot of customers they would say, "Maybe 18 months ago or two years ago, we knew that we had a legacy app problem. We knew we were spending too much on the underpinning infrastructure, but we could sort of afford it. Was it perfect? No. Was it a bit of a mess? Yes. Should we have really been focusing on the legacy apps issue, thinking maybe the economy is never going to sustain what it was?" But, of course, hindsight is a wonderful thing.

Smarter and better

Now, when I am talking to clients, I mean their comments are, "We need to be smarter. We need to be better. We need to retain our customers, deliver better quality of service to our customers, and we have to do it at a low cost." We've seen a massive change in the approach with the legacy environment, whether that’s applications or infrastructure.

Over the last 12 months what people have realized that it is now time for those organizations that want to remain competitive and innovative. Unfortunately, I still see a lot of companies that believe that doing nothing is the thing to do and will just wait for the economy to rebound. I don't believe it's going to rebound to the same place. It may come back and it may be stronger, but it may end up on a different place.

The organizations that are not waiting for that, but are trying to be innovative, competitive, move away from the competition, and give themselves some breathing space are the ones who are going to sustain themselves.

Within HP, we were not in great shape a few years back in terms of our IT spend. It was way too high, and we openly admit that. We had to take some fairly drastic actions, and it is probably well-known and documented. We went from about 6,000 apps went down to about 1,700 today, and we'll probably plateau at around 1,500. We went from 83 data centers to 6.

We're running a better, faster, cheaper organization that is more agile.



We were not in great shape, but we took action. I don’t believe we took action because we knew the economy was going to change. We don’t think we're that clever. We just had to take it, because economically it was just not the right solution, and nor technically.

We had to focus on driving this both from business and IT. As I said in this small example, we went from 300 apps to 30 apps. We had a 39 percent reduction in our inventory dollars. We reduced our supply chain expenses. We reduced the cost of doing next day delivery. We're heading toward reducing our CO2 emissions by 40 percent on those next day deliveries.

But overall, the global supply chain, this measure of spent revenue, we drove down by 19 percent. We're running a better, faster, cheaper organization that is more agile. As you said, it positions us better to exploit situations as they change and feel that they’ve become more of an opportunity rather than a threat.

Gardner: For those organizations that are in some sort of a multi-year approach, looking at their portfolio of applications, probably shocked by how many there are, what the redundancies are, what the actual landscape looks like, but perhaps also a little bit chagrin by the daunting complexity, where do you suggest they start for resources? Is there a way to start thinking strategically about both the technology, the business, and process issues, as well as those governance, operational, and methodological issues?

First, take stock

Evans: A number of people I have talked to say that their biggest challenge is that they don’t know what they’ve got. So, first and foremost, the advice is always that you need to take stock of what you’ve got, because if you don’t know what you are dealing with, then you’ve got a problem.

I’ll give you an example. I spoke to two large organizations at a recent event we ran here in HP. One organization openly admitted they didn't know what their problems were. They knew they had a massive, complex, and growing applications portfolio that was basically losing touch with the business. That was one side.

Another customer openly admitted that they knew the applications that were causing them problems. They said, "We have these 14 apps that are killing us, and we need to do something about it. We need to streamline those apps. We need to use contemporary technology."

They need to use a new software environment that gives them a much smaller code base, if they are moving from COBOL to something like Java or C#, using new database technologies. Using new testing techniques that don’t mean that we load testing to the end of the session. Then, when time gets tight, what gets cut? The testing gets cut.

Underneath all that is the abilitty to save money, which, of course, is fairly important.



That was a good example we used in the supply chain example. We used an HP product, Quality Center, that gave us this process, this rigor that said, "We're going to test things. We're going to throw out different scenarios, and we're going to test it to death. We're not going to test it to death at the end when it’s too late, but we're going to do that throughout the development cycle so that we can make those adjustments and modifications as we go along."

So, we ended up with a high-quality product at the end. Talking to a lot of customers, the speed by which they can develop, as well as modify, applications, and that is connected directly to customer satisfaction, is paramount.

In the financial services industry, your application is your business. If you are in the telco industry, the level of service you can offer is very much aligned to the application. If you can improve the speed, and the momentum you can create in terms of introducing new editions and you can do those with a very high quality and a high level of integrity, then you're heading towards delivering a much better service to your customer.

There are so many lessons learned here addressing what people have in terms of portfolio and then also delivering new, contemporary, revised types of applications and/or infrastructure.
Underneath all that is the ability to save money, which, of course, is fairly important, isn’t it?

Gardner: We talked a bit about the how and the why of this massive application consolidation projects issue, but this specific Global Part Supply Chain project at HP is now a case-study, which you have written up and which is available for people to get some more detail from. Are there are some other resources, sites, or places where they can go for not only learning more about how you solve this problem, but where they can start on their journey or continue the journey.

Open about experiences

Evans: We have always said that we're going to be very open about our experiences, only because I think people don’t want to begin new things. They don't want to be the first to take a leap, but as I said, pretty much every customer in my mind is doing some form of application transformation, whether small, large, or medium scale.

We have always said that the experiences we gain from our own work we would share openly, and sometimes we’re quite happy to say where we did go wrong. In this instance, we’ve written up a case study to give people an insight in more detail than I have been able to provide today. We're going to post that on our portal. If people want to go there, it’s relatively simple.

It’s http://hp.com/go/applicationtransformation, and they’ll find the case-study. They’ll find videos and other materials of other customers who have embarked on these journeys, whether they’ve been driving that from the top down, from an application’s nature, or whether it’s people who are coming in from the infrastructure, who will say, "I have aging obsolete infrastructure that I need to change, but I know there is a collateral impact on my application. How do I go about that?"

We're trying to cover all the bases in terms of those people that are coming with top-down applications, bottom-up infrastructure, or looking to create a new software environment. If they go to that URL, they can find all the materials, and I hope that they might find useful.

The point is that we get this ability to have an elastic environment.



Gardner: Paul, before we close out, perhaps a look to the future. I've heard so much now about cloud computing and software as a service. This is not necessarily just talking about custom, packaged, and on-premises apps. We now need to think about different sourcing options. How does that relate to this process of application transformation and the rationale around where to go for the best economic bang for the buck?

Evans: Cloud is just a part of the application transformation journey. If you think of history over the last 50 years -- and that’s all technology is; it’s only 50 years old in this space -- we’ve done everything inside. We did everything ourselves. We did everything in big machines, crammed everything in, then we’ve gone more distributed. We’ve gone to PCs and all rest of it. We began to spread the web, before we even knew what the word "Web" meant.

Now, we’ve gotten used to interacting with the Internet, and more importantly, the web. We're beginning to say, you know, maybe there are some services that we don't need inside of firewall, or in a private cloud -- so it’s inside but not inside." The point is that we get this ability to have an elastic environment. We haven’t got dedicated systems to run a service that maybe we only want to use 20 percent of the time.

The notion of using the web or technologies that have been formed from the web development is just like falling off a log. We'll argue probably for at least the next 12 months about what is the cloud and what’s not the cloud, but the use of the World Wide Web is a part of our day-to-day business that's irreversible.

We're never going to go backward now, whether it’s just interaction with consumers who want to get questions answered and order a PC or whatever, or the provision of services that we'll use and our customers will use by utilizing the web. It’s just going to be there.

Also, merging with this whole notion of the cloud is mobility. The mobile, the smartphone, or call it what you want, is going to be the most voluminous device that will attach to the web in the future. People are not just going to want to play games, send SMS, and all the rest of it. They're going to email, they are going to want to do things, and they are going to want to interact in a far richer environment than they do today.

I think these technologies are converging rapidly in terms of a notion that says we’ve got to update and transform what we’ve got and at the same time start the more strategic view of what are we going to incorporate going forward. We may not incorporate them today, but we sure need to leave that socket open that says I may want to plug it in the future.

Gardner: Well great. We've been talking about best practices in execution accelerators around large application consolidation projects. We’ve been joined by Paul Evans, World Wide Marketing Lead on Applications Transformation at HP. Thank you, Paul.

Evans: Thanks, Dana.

Gardner: This is Dana Gardner, Principal Analyst at Interarbor Solutions. You’ve been listening to a sponsored BriefingsDirect podcast. Thanks for listening, and come back next time.

Access more information on Application Consolidation.
Read the full-length case study on HP's Application Consolidation.
Learn more about the Application Transformation Experience Workshop.

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Download the transcript. Sponsor: HP.

Transcript of a BriefingsDirect podcast on how HP tackled an internal multi-year effort to streamline supply chain efficiencies and effectiveness through applications consolidation. Copyright Interarbor Solutions, LLC, 2005-2010. All rights reserved.

You may also be interested in:

Monday, January 18, 2010

Technical and Economic Incentives Mount Around Seeking Alternatives to Mainframe Applications

Transcript of the third in a series of sponsored BriefingsDirect podcasts on the rationale and strategies for application transformation.

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Download the transcript. Learn more. Sponsor: Hewlett-Packard.


Gain more insights into "Application Transformation: Getting to the Bottom Line" via a series of HP virtual conferences. For more on Application Transformation, and to get real time answers to your questions, register to access the virtual conferences for your region:

Access the Asia Pacific event.
Access the EMEA event.
Access the Americas event.


Dana Gardner: Hi, this is Dana Gardner, principal analyst at Interarbor Solutions, and you’re listening to BriefingsDirect.

Today, we present a sponsored podcast discussion on why it's time to exploit alternatives to mainframe computing applications and systems. As enterprises seek to cut their total IT costs, they need to examine alternatives to hard to change and manage legacy systems. There are a growing number of technical and economic incentives for modernizing and transforming applications and the data center infrastructure that support them.

Today, we'll examine some case studies that demonstrate how costs can be cut significantly, while productivity and agility are boosted by replacing aging systems with newer, more efficient standards-based architectures.

This podcast is the third and final episode in a series that examines "Application Transformation: Getting to the Bottom Line." The podcast, incidentally, runs in conjunction with a series of Hewlett-Packard (HP) webinars and virtual conferences on the same subject.

Access the Asia Pacific event. Access the EMEA event. Access the Americas event.

Here with us now to examine alternatives to mainframe computing, is John Pickett, Worldwide Mainframe Modernization Program manager at HP. Hello, John.

John Pickett: Hey, Dana. How are you?

Gardner: Good, thanks. We're also joined by Les Wilson, America's Mainframe Modernization director at HP. Welcome to the show, Les.

Les Wilson: Thank you very much, Dana. Hello.

Gardner: And, we're also joined by Paul Evans, Worldwide Marketing Lead on Applications Transformation at HP. Welcome back, Paul.

Paul Evans: Hello, Dana.

Gardner: Paul, let me start with you if you don't mind. We hear an awful lot about legacy modernization. We usually look at it from a technical perspective. But it appears to me that in many of the discussions I have with organizations is that they are looking for more strategic levels of benefit, to finding business agility and flexibility benefits. The technical and economic considerations, while important in the short-term, pale in comparison to some of the longer term and more strategic benefits.

Pushed to the top

Evans: Where we find ourselves now -- and it has been brought on by the economic situation -- is that it has just pushed to the top an issue that's been out there for a long time. We have seen organizations doing a lot with their infrastructure, consolidating it, virtualizing it, all the right things. At the same time, we know, and a lot of CIOs or IT directors listening to this broadcast will know, that the legacy applications environment has somewhat been ignored.

Now, with the pressure on cost, people are saying, We've got to do something, but what can come out of that and what is coming out of that?" People are looking at this and saying, "We need to accomplish two things. We need a longer term strategy. We need an operational plan that fits into that, supported by our annual budget."

Foremost is this desire to get away from this ridiculous backlog of application changes, to get more agility into the system, and to get these core applications, which are the ones that provide the differentiation and the innovation for organizations, able to communicate with a far more mobile workforce.

At an event last week in America, a customer came up to me and said, "Seventy percent of our applications are batch running on a mainframe. How do I go to a line-of-business manager and say that I can connect that to a guy out there with a smartphone? How do I do that?" Today, it looks like an impossible gap to get across.

What people have to look at is where we're going strategically with our technology and our business alignment. At the same time, how can we have a short-term plan that starts delivering on some of the real benefits that people can get out there?

Gardner: In the first two parts of our series, we looked at several case studies that showed some remarkable return on investment (ROI). So, this is not just a nice to have strategic maturity process, but really pays dividends financially, and then has that longer term strategic roll-out.

Evans: Absolutely. These things have got to pay for themselves. An analyst last week, looked me in the face and said, "People want to get off the mainframe. They understand now that the costs associated with it are just not supportable and are not necessary."

One of the sessions you will hear in the virtual conference will be from Geoffrey Moore, where he talks about this whole difference between core applications and context -- context being applications that are there for productivity reasons, not for innovation or differentiation.

Lowest-cost platform


With a productivity application you want to get delivery on the lowest-cost platform you possibly can. The problem is that 20 or 30 years ago, people put everything on the mainframe. They wrote it all in code. Therefore, the challenge now is, what do you not need in code that can be in a package? What do you not need on the mainframe that could be running on a much more lower cost infrastructure or a completely different means of delivery, such as software as a service (SaaS).

The point is that there are demonstrably much less expensive ways of delivering these things. People have to just lift their heads up and look around, come and talk to us, and listen to the series and they will begin to see people who have done this before, and who have demonstrated that it works, as well as some of the amazing financial rewards that can be generated from this sort of work.

Gardner: John Pickett, let's go to you. We've talked about this, but I think showing it is always more impressive. The case studies that demonstrate the real-world returns tend to be the real education points. Could you share with us some of the case studies that you will be looking at during the upcoming virtual conference and walk us through how the alternative to mainframe process works?

Pickett: Sure, Dana. As Paul indicated, it's not really just about the overall cost, but it's about agility and being able to leverage the existing skills as well.

One of the case studies that I will go over is from the National Agricultural Cooperative Federation (NACF). It's a mouthful, but take a look at the number of banks that the NACF has. It has 5,500 branches and regional offices, so essentially it's one of the largest banks in Korea.

One of the items that they were struggling with was how to overcome some of the technology and performance limitations of the platform that they had. Certainly, in the banking environment, high availability and making sure that the applications and the services are running were absolutely key.

At the same time, they also knew that the path to the future was going to be through the IT systems that they had and they were managing. What they ended up doing was modernizing their overall environment, essentially moving their core banking structure from their current mainframe environment to a system running HP-UX. It included the customer and account information. They were able to integrate that with the sales and support piece, so they had more of a 360 degree view of the customer.

We talk about reducing costs. In this particular example, they were able to save $40 million on an annual basis. That's nice, and certainly saving that much money is significant, but, at the same time, they were able to improve their system response time two- to three-fold. So, it was a better response for the users.

But, from a business perspective, they were able to reduce their time to market. For developing a new product or service, that they were able to decrease that time from one month to five days.

Makes you more agile

If you are a bank and now you can produce a service much faster than your competition, that certainly makes it a lot easier and makes you a lot more agile. So, the agility is not just for the data center, it's for the business as well.

To take this story just a little bit further, they saw that in addition to the savings I just mentioned, they were able to triple the capacity of the systems in their environment. So, it's not only running faster and being able to have more capacity so you are set for the future, but you are also able to roll out business services a whole lot quicker than you were previously.

Gardner: I imagine that with many of these mainframe systems, particularly in a banking environment, they could be 15 or 20 years old. The requirements back then were dramatically different. If the world had not changed in 20 years, these systems might be up to snuff, but the world has changed dramatically. Look at the change we have seen in just the last nine months. Is that what we are facing here? We have a general set of different requirements around these types of applications.

Pickett: There are a couple of things, Dana. It's not only different requirements, but it's also being driven by a couple of different factors. Paul mentioned the cost and being able to be more efficient in today's economy. Any data center manager or CIO is challenged by that today. Given the high cost of legacy and mainframe environment, there's a significant amount of money to be saved.

It's not a one-size-fits-all. It's identifying the right choice for the application, and the right platform for the application as well.



Another example of what we were just talking about is that, if we shift to Europe, Middle East, and Africa region, there is very large insurance company in Spain. It ended up modernizing 14,000 million instructions per second (MIPS). Even though the applications had been developed over a number of years and decades, they were able to make the transition in a relatively short length of time. In a three- to six-month time frame they were able to move that forward.

With that, they saw a 2x increase in their batch performance. It's recognized as one of the largest batch re-hosts that are out there. It's just not an HP thing. They worked with Oracle on that as well to be able to drive Oracle 11g within the environment.

So, it's taking the old, but also integrating with the new. It's not a one-size-fits-all. It's identifying the right choice for the application, and the right platform for the application as well.

Gardner: So, this isn't a matter of swapping out hardware and getting a cheaper fit that way. This is looking at the entire process, the context of the applications, the extended process and architectural requirements in the future, and then looking at how to make the transition, the all important migration aspect.

Pickett: Yes. As we heard last week at a conference that both Paul and I were at, if all you're looking to do is to take your application and put it on to a newer, shinier box, then you are missing something.

Gardner: Let's go now to Les Wilson. Les, tell us a little bit about some studies that have been done and some of the newer insights into the incentives as to why the timing now for moving off of mainframes is so right.

Customer cases

Wilson: Thanks, Dana. I spend virtually every day talking directly to customers and to HP account teams on the subject of modernizing mainframes, and I'll be talking in detail about two particular customer case studies during the webinar.

Before I get into those details though, I want to preface my remarks by giving you some higher level views of what I see happening in the Americas. First of all, the team here is enjoying an unprecedented demand for our services from the customer base. It's up by a factor of 2 over 2008, and I think that some of the concepts that John and Paul have discussed around the reasons for that are very clear.

There's another point about HP's capabilities, as well, that makes us a very attractive partner for mainframe modernization solutions. Following the acquisition of EDS, we are really able to provide a one-stop shop for all of the services that any mainframe customer could require.

That includes anything from optimization of code, refactoring of code on the mainframe itself, all the way through re-hosting, migration, and transformation services. We've positioned ourselves as definitely the alternative to IBM mainframe customers.

In terms of customer situations, we've always had a very active business working with organizations in manufacturing, retail, and communications. One thing that I've perceived in the last year specifically -- it will come as no surprise to you -- is that financial institutions, and some of the largest ones in the world, are now approaching HP with questions about the commitment they have to their mainframe environments.

We're seeing a tremendous amount of interest from some of the largest banks in the United States, insurance companies, and benefits management organizations, in particular.

Second, maybe benefiting from some of the stimulus funds, a large number of government departments are approaching us as well. We've been very excited by customer interest in financial services and public sector. I just wanted to give you that by way of context.

In terms of the detailed case studies, when John Pickett first asked me to participate in the webinar, as well as in this particular recording, I was kind of struck with a plethora of choices. I thought, "Which case study should I choose that best represents some of the business that we are doing today?" So, I've picked two.

The first is a project we recently completed at a wood and paper products company. This is a worldwide concern. In this particular instance we worked with their Americas division on a re-hosting project of applications that are written in the Software AG environment. I hope that many of the listeners will be familiar with the database ADABAS and the language, Natural. These applications were written some years ago, utilizing those Software AG tools.

Demand was lowered

They had divested one of the major divisions within the company, and that meant that the demand for mainframe services was dramatically lowered. So, they chose to take the residual applications, the Software AG applications, representing about 300-350 MIPS, and migrate those in their current state, away from the mainframe, to an HP platform.

Many folks listening to this will understand that the Software AG environment can either be transformed and rewritten to run, say, in an Oracle or a Java environment, or we can maintain the customer's investment in the applications and simply migrate the ADABAS and Natural, almost as they are, from the mainframe to an alternative HP infrastructure. The latter is what we did.

By not needing to touch the mainframe code or the business rules, we were able to complete this project in a period of six months, from beginning to end. They are saving over $1 million today in avoiding the large costs associated with mainframe software, as well as maintenance and depreciation on the mainframe environment.

They're very, very pleased with the work that's being done. Indeed, we're now looking at an additional two applications in other parts of their business with the aim of re-hosting those applications as well.

They are saving over $1 million today in avoiding the large costs associated with mainframe software, as well as maintenance and depreciation on the mainframe environment.



The more monolithic approach to applications development and maintenance on the mainframe is a model that was probably appropriate in the days of the large conglomerates, where we saw a lot of companies trying to centralize all of that processing in large data centers. This consolidation made a lot of sense, when folks were looking for economies of scale in the mainframe world.

Today, we're seeing customers driving for that degree of agility you have just mentioned. In fact, my second case study represents that concept in spades. This is a large multinational manufacturing concern. They never allow their name to be used in these webcasts, so we will just refer to them as "a manufacturing company." They have a large number of businesses in their portfolio.

Our particular customer in this case study is the manufacturer of electronic appliances. One of the driving factors for their mainframe migration was precisely what you just said, Dana, that the ability to divest themselves from the large mainframe corporate environment, where most of the processing had been done for the last 20 years.

They wanted control of their own destiny to a certain extent, and they also wanted to prepare themselves for potential investment, divestment, and acquisition, just to make sure that they were masters of their own future.

Gardner: You mentioned earlier, John, about a two-times increase in the demand since 2008. I wonder if this demand increase is a blip. Is this something that is just temporary, or has the economy -- and some people call it the reset economy, actually changed the game -- and therefore IT needs to respond to that?

In a nutshell the question is whether this is going to be a two-year process, or are we changing the dynamic of IT and how business and IT need to come together in general?

Not a blip

Pickett: First, Dana, it's not a blip at all. We're seeing increased movement from mainframe over to HP systems, whether it's on an HP-UX platform or a Windows Server or SQL platform. Certainly, it's not a blip at all.

As a matter of fact, just within the past week, there was a survey by AFCOM, a group that represents data-center workers. It indicated that, over the next two years, 46 percent of the mainframe users said that they're considering replacing one or more of their mainframes.

Now, let that sink in -- 46 percent say they are going to be replacing high-end systems over the next two years. That's an absurdly high number. So, it certainly points to a trend that we are seeing in that particular environment -- not a blip at all.

Dana, that also points to the skills piece. A lot of times when we talk to people in a mainframe environment the question is, "I've got a mainframe, but what about the mainframe people that I have? They're good people, they know the process, and they have been around for a while." We found that HP, and moving to an HP centralized environment is really a soft landing for these people.

They can use the process skills that they have developed over time. They're absolutely the best at what they do in the process environment, but it doesn’t have to be tied to the legacy platform that they have been working on for the last 10 or 20 years.

We've found that there is a very strong migration for those skills and landing in a place where they can use and develop them for years to come.



We've found that you can take those same processes and apply them to a large HP Integrity Superdome environment, or NonStop environment. We've found that there is a very strong migration for those skills and landing in a place where they can use and develop them for years to come.

Gardner: Les, why do you see this as a longer term trend, and what are the technological changes that we can expect that will make this even more enticing, that is to say, lower cost, more efficient, and higher throughput systems that allow for the agility to take place as well?

Wilson: Good question, Dana, and you have two parts to it. Let me address the first one about the trend. I've been involved in this kind of business on and off since 1992. We have numbers going back to the late 1980s as to the fact that at that time there were over 50,000 mainframes installed worldwide.

When I next got into this business in 2004, the analyst firms confirmed that the number was now around 15,000-16,000. Just this week, we have had information, confirmed by another analyst, that the number of installed mainframes is now at about 10,400. We've seen a 15-20 year trend away from the mainframe, and that will continue, given this unprecedented level of interest we are seeing right now.

You talked about technology trends. Absolutely. Five years ago, it would have been fair to say that there were still mainframe environments and applications that could not be replaced by their open-system equivalents. Today, I don't think that that's true at all.

Airline reservation system

To give you an example, HP, in cooperation with American Airlines, has just announced that we're going to be embarking on a three-year transition of all of the TPF-based airline reservation systems that we HP has been providing as services to customers for 20 years.

That TPF environment will be re-engineered in its entirety over the course of the next three years to provide those same and enhanced airline reservation systems to customers on a Microsoft-HP bladed environment.

That's an unprecedented change in what was always seen as a mainframe centric application, airlines reservations, with the number of throughputs and the amount of transactions that need to be done every second. When these kinds of applications can be transformed to open systems' platforms, it's open season on any mainframe application.

Furthermore, the trend in terms of open-systems price performance improvement continues at 30-35 percent per annum. You just need to look at the latest Intel processors, whether they be x86 or Itanium-based, to see that. That price performance trend is huge in the open systems market.

I've been tracking what's been going on in the IBM System Z arena, and since there are no other competitors in this market, we see nothing more than 15 percent, maybe 18 percent, per annum price performance improvement. As time goes on, HP and industry standard platforms continue, and will continue, to outpace the mainframe technology. So, this trend is bound to happen.

People have to take considered opinions. Investments here are huge. The importance of legacy systems is second to none.



Gardner: Paul, we've heard quite a bit of compelling information. Tell us about the upcoming conference, and perhaps steps that folks can take to get more information or even get started as they consider their alternatives to mainframes?

Evans: Based on what you've heard from John and Les, there is clearly an interest out there in terms of understanding. I don't think this is, as they say in America, a slam dunk. The usual statement is, "How do you eat an elephant? and the answer is, "One bite at a time."

The point here is that this is not going to happen overnight. People have to take considered opinions. Investments here are huge. The importance of legacy systems is second to none. All that means that the things that John and Les are talking about are going to happen strategically over a long time. But, we have people coming to us every day saying, "Please, can you help me understand how do I start, where do I go, where do I go now, or where do I go next week, next year, or next month?

The reason behind the conference was to take a sort of multi-sided view of this. One side is the business requirement, which people like Geoffrey Moore will be talking about -- where the business is going and what does it need.

We'll be looking at a customer case study from the Italian Ministry of Education, looking at how they used multiple modernization strategies to fix their needs. We'll be looking at tools we developed, so that people can understand what the code is doing. We'll be hearing from Les, John, and customers -- Barclays Bank in London -- about what they have been doing and the results they have been getting.

Then, at the very end, we'll be hearing from Dale Vecchio, vice president of Gartner research, about what he believes is really going on.

Efficiency engine

The thing that underpins this is that the business requirement several decades ago drove the introduction of the mainframe. People needed an efficiency engine for doing payroll, human resources, whatever it may be, moving data around. The mainframe was invented and was built perfectly. It was an efficiency engine.

As time has gone on, people look at technology now to become an effectiveness engine. We've seen the blending of technologies between mainframes, PCs, and midrange systems. People now take this whole efficiency thing for granted. No one runs their own payroll, even to the point that people now look to BPOs or those sorts of things.

As we go forward, with people being highly mobile, with mobile devices dramatically exploding all over the place in terms of smartphones, Net PCs, or whatever, people are looking to blend technologies that will deliver both the efficiency and the effectiveness, but also the innovation. Technology is now the strategic asset that people will use going forward. There needs to be a technological response to that.

Over the last year or two, either John or Les referred to the enormous amounts of raw power we can now get from, say, an Intel microprocessor. What we want to do is harness that power and give people the ability to innovate and differentiate, but, at the same time, run those context applications that keep their companies alive.

That's really what we're doing with the conference -- demonstrating, in real terms, how we can get this technology to the bottom-line and how we can exploit it going forward.

Gardner: Well, great. We've been hearing about some case studies that demonstrate how costs can be cut significantly, while productivity and agility are boosted.

I want to thank our guests in today’s discussion. We've been joined by John Pickett, Worldwide Mainframe Modernization Program manager. Thank you, John.

Pickett: Thank you, Dana.

Gardner: We've also been joined by Les Wilson, America’s Mainframe Modernization director. Thank you, Les.

Wilson: Thank you for the opportunity, Dana.

Gardner: And also Paul Evans, worldwide marketing lead on Applications Transformation at HP. Thanks again Paul.

Evans: It's a pleasure.

Gardner: This is Dana Gardner, principal analyst at Interarbor Solutions. You have been listening to a sponsored BriefingsDirect podcast. Thanks for listening, and come back next time.

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Download the transcript. Learn more. Sponsor: Hewlett-Packard.




Gain more insights into "Application Transformation: Getting to the Bottom Line" via a series of HP virtual conferences. For more on Application Transformation, and to get real time answers to your questions, register to access the virtual conferences for your region:

Access the Asia Pacific event.
Access the EMEA event.
Access the Americas event.




Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Download the transcript. Learn more. Sponsor: Hewlett-Packard.


Transcript of the third in a series of sponsored BriefingsDirect podcasts on the rationale and strategies for application transformation. Copyright Interarbor Solutions, LLC, 2005-2009. All rights reserved.

Thursday, October 29, 2009

Separating Core from Context Brings High Returns in Legacy Application Transformation

Transcript of the second in a series of sponsored BriefingsDirect podcasts on the rationale and strategies for application transformation.

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Download the transcript. Learn more. Sponsor: Hewlett-Packard.


Gain more insights into "Application Transformation: Getting to the Bottom Line" via a series of HP virtual conferences Nov. 3-5. For more on Application Transformation, and to get real time answers to your questions, register to the virtual conferences for your region:
Register here to attend the Asia Pacific event on Nov. 3.
Register here to attend the EMEA event on Nov. 4.
Register here to attend the Americas event on Nov. 5.


Dana Gardner: Hi, this is Dana Gardner, principal analyst at Interarbor Solutions, and you’re listening to BriefingsDirect.

Today, we present a sponsored podcast discussion on separating core from context, when it comes to legacy enterprise applications and their modernization processes. As enterprises seek to cut their total IT costs, they need to identify what legacy assets are working for them and carrying their own weight, and which ones are merely hitching a high cost -- but largely unnecessary -- ride.

The widening cost-in-productivity division exists between older, hand-coded software assets, supported by aging systems, and replacement technologies on newer, more efficient standards-based systems. Somewhere in the mix, there are core legacy assets distinct from so-called contextal assets. There are peripheral legacy processes and tools that are costly vestiges of bygone architectures. There is legacy wheat and legacy chaff.

Today we need to identify productivity-enhancing resources and learn how to preserve and modernize them -- while also identifying and replacing the baggage or chaff. The goal is to find the most efficient and low-cost means to support them both, through up-to-date data-center architecture and off-the-shelf components and services.

This podcast is the second in a series of three to examine Application Transformation: Getting to the Bottom Line. We will discuss the rationale and likely returns from assessing the true role and character of legacy applications and their actual costs. The podcast, incidentally, runs in conjunction with some Hewlett-Packard (HP) webinars and virtual conferences on the same subject.

Register here to attend the Asia Pacific event on Nov. 3. Register here to attend the EMEA event on Nov. 4. Register here to attend the Americas event on Nov. 5.

With us to delve deeper into the low cost, high reward transformation of legacy enterprise applications is Steve Woods, distinguished software engineer at HP. Hello, Steve.

Steve Woods: Hello. How are you doing?

Gardner: Good. We are also joined by Paul Evans, worldwide marketing lead on Applications Transformation at HP. Hello, Paul.

Paul Evans: Hello, Dana. Thank you.

Gardner: We talked in the earlier podcast in our series, a case study, about transformation and why that's important through the example of a very large education organization in Italy and what they found. We looked at how this can work very strategically and with great economic benefit, but I think now we are trying to get into a bit more of the how.

Tell us a little bit, Paul, about what the stakes are. Why is it so important to do this now?

Evans: In a way, this podcast is about two types of IT assets. You talked before about core and context. That whole approach to classifying business processes and their associated applications was invented by Geoffrey Moore, who wrote Crossing the Chasm, Inside the Tornado, etc.

He came up with this notion of core and context applications. Core being those that provide the true innovation and differentiation for an organization. Those are the ones that keep your customers. Those are the ones that improve the service levels. Those are the ones that generate your money. They are really important, which is why they're called "core."

Lower cost

The "context" applications were not less important, but they are more for productivity. You should be looking to understand how that could be done in terms of lower cost provisioning. When these applications were invented to provide the core capabilities, it was 5, 10, 15, or 20 years ago. What we have to understand is that what was core 10 years ago may not be core anymore. There are ways of effectively doing it at a much different price point.

As Moore points out, organizations should be looking to build "core," because that is the unique intellectual property of the organization, and to then buy "context." They need to understand, how do I get the lowest-cost provision of something that doesn't make a huge difference to my product or service, but I need it anyway.

An human resources system may not be something that you are going to build your business model on, but you need one. You need to be able to service your employees and all the things they need. But, you need to do that at the lowest-cost provision. As time has gone on, this demarcation between core and context has gotten really confused.

As you said, we're putting together a series of events, and Moore will be the keynote speaker on these events. So, we will elucidate more around core and context.

The other speaker at the event is also an inventor, this time from inside HP, Steve Woods. Steve has taken this notion of core and context and has teamed it with some extremely exciting technology and very innovative thinking to develop some unique tools that we use inside the services from HP, which allow us then really to dive into this. That's going to be one of the sessions that we're also going to be delivering on this series of events.

Gardner: Okay, Steve Woods, we can use a lot of different terms here, "core and context," "wheat and chaff." I thought another metaphor would be "baby and bathwater." What happens is that it's difficult to separate the good from the potentially wasteful in the legacy inventory.

I think this has caused people to resist modernizing. They have resisted tinkering with legacy installations in the past. Why are they willing to do it now? Why the heightened interest at this time?

Woods: A good deal of it has to do with the pain that they're going through. We have had customers who had assessments with us before, as much as a year ago, and now they're coming back and saying they want to get started and actually do something. So, a good deal of the interest is caused by the need to drive down costs.

Also, there's the realization that a lot of these tools -- extract, transform, and load (ETL) tools, enterprise application integration (EAI) tools, reporting, and business process management (BPM) -- are proving themselves now. We can't say that there is a risk in going to these tools. They realize that the strength of these tools is that they bring a lot of agility, solve skill sets issues, and make you much more responsive to the business needs of the organization.

Gardner: This definition of core, as Paul said, is changing over time and also varies greatly from organization to organization. Is there no one size fits all approach to this?

Context not code

Woods: I don't think there really is a one size fits all, but as we use our tools to analyze code, we find sometimes as much as 65 percent or more of an application could really not be core. It could just be context.

As we make these discoveries, we find that in the organization there are political battles to be fought. When you identify these elements that are not core and that could be moved out of handwritten code, you're transferring power from the developers -- say, of COBOL -- to the users of the more modern tools, like the BPM tools.

So there is always an issue. What we try to do, when we present our findings, is to be very objective. You can't argue that we found that 65 percent of the application is not doing core. You can then focus the conversation on something more productive. What do we do with this? The worst thing you could possibly do is take a million lines of COBOL that's generating reports and rewrite that in Java or C# hard-written code.

We take the concept of core versus context not just to a possible off-the-shelf application, but at architectural component level. In many cases, we find that this is helpful for them to identify legacy code that could be moved very incrementally to these new architectures.

Gardner: What's been the holdup? What's difficult? You did mention politics, and we will get into that later, but what's been the roadblock from perspective of these tools? Why has that been decreasing in terms of the ability to automate and manage these large projects?

Woods: A typical COBOL application -- this is true of all legacy code, but particularly mainframe legacy code -- can be as much as 5, 10, or 15 million lines of code. I think the sheer idea of the size of the application is an impediment. There is some sort of inertia there. An object at rest tends to stay at rest, and it's been at rest for years, sometimes 30 years.

So, the biggest impediment is the belief that it's just too big and complex to move and it's even too big and complex to understand. Our approach is a very lightweight process, where we go in and answer to a lot of questions, remove a lot of uncertainty, and give them some very powerful visualizations and understanding of the source code and what their options are.

Gardner: So, as we've progressed in terms of the tools, the automation, and the ability to handle large sets of code, the inertia also involves the nontechnical aspects. What do we mean by politics? Are there fiefdoms? Are there territories? Is this strictly a traditional kind of human nature thing? Perhaps you could help us understanding that a bit better.

Doing things efficiently

Woods: Organizations that we go in have not been living in a vacuum, so many of have been doing greenfield development; when they start out to say they need a system that does primarily reporting, or a system that does primarily data integration. In most organizations those fiefdoms, if you will, have grown pretty robust, and they will continue to grow. The realization is that they actually can do those things quite efficiently.

When you go to the legacy side of the house, you start finding that 65 percent of this application is just doing ETL. It's just parsing files and putting them into databases. Why don't you replace that with a tool? The big resistance there is that, if we replace it with a tool, then the people who are maintaining the application right now are either going to have to learn that tool or they're not going to have a job.

So, there's a lot of resistance in the sense that we don't want to lose anymore ground to the target architecture fiefdom, so we are going to not identify this application as having so many elements of context functionality. Our process, in a very objective way, just says that these are the percentages that we're finding. We'll show you the code, you can agree or disagree that's what it is doing, and then let's make decisions based upon those facts.

If we get the facts on the table, particularly visually, then we find that we get a lot of consensus. It may be partial consensus, but it's consensus nonetheless, and we open up the possibilities and different options, rather than just continuing to move through with hand-written code.

If you look at this whole core-context thing, at the moment, organizations are still in survival mode.



Gardner: Paul, you've mentioned in the past that we've moved from the nice-to-have to the must-have, when it comes to legacy applications transformation and modernization. The economy has changed things in many respects, of course, but it seems as if the lean IT goal is no longer something that's a vision. It's really moved up the pecking order or the hierarchy of priorities.

Is this perhaps something that's going to impact this political logjam? Are other organizations and business and financial outcome folks, who are just going to steamroll these political issues?

Evans: Well, I totally think so, and it's happening already. If you look at this whole core-context thing, at the moment, organizations are still in survival mode. Money is still tight in terms of consumer spending. Money is still tight in terms of company spending. Therefore, you're in this position where keeping your customers or trying to get new customers is absolutely fundamental for staying alive. And, you do that by improving service levels, improving your services, and improving your product.

If you stay still and say, "Well, we'll just glide for the next 6 to 12 months and keep our fingers crossed," you're going to be in deep trouble. A lot of people are trying to understand how to use the newer technologies, whether it's things like Web 2.0 or social networking tools, to maintain that customer outreach.

Those of us who went to the business school, marketing school remember -- it takes $10 to get a customer into your store, but it only takes $1 to keep them coming back. People are now worrying about those dollars. How much do we have to spend to keep our customer base?

Therefore, the line-of-business people are now pushing on technology and saying, "You can't back off. You can't not give us what we want. We have to have this ability to innovate and differentiate, because that way we will keep our customers and we will keep this organization alive."

Public and private sectors

That applies equally to the public and private sectors. The public sector organizations have this mandate of improving service, whether it's in healthcare, insurance, tax, or whatever. So all of these commitments are being made and people have to deliver on them, albeit that the money, the IT budget behind it, is shrinking or has shrunk.

So, the challenge here is, "Last year I ran my IT department on my theoretical $100. I spent $80 on keeping things going, and $20 on improving things." That was never enough for the line-of-business manager. They will say, "I want to make a change. I want it now, or I want it next week. I don't want it in six months time. So explain to me how you are going to do that."

That was tough a year ago, but the problem now is that your $100 IT budget is now $80. Now, it's a bit of a challenge, because now all the money you have got you are going to spend on keeping the old stuff alive. I don't think the line-of-business managers, or whoever they are, are going to sit back and say, "That's okay. That's okay. We don't mind." They're going to come and say that they expect you to innovate more.

This goes back to what Steve was talking about, what we talked about, and what Moore will raise in the event, which is to understand what drives your company. Understand the values, the differentiation, and the innovations that you want and put your money on those and then find a way of dramatically reducing the amount of money you spend on the contextual stuff, which is pure productivity.

The point of the tools is that they allow us to see the code. They allow us to understand what's good and bad and to make very clear, rational, and logical decision.



Steve's tools are probably the best thing out there today that highlight to an organization, "You don't need this in handwritten code. You could put this to a low cost package, running on a low cost environment, as opposed to running it in COBOL on a mainframe." That's how people save money and that's how we've seen people get, as we have talked earlier, a return on investment (ROI) of 18 months or less.

So it is possible, it can be done, and it's definitely not as difficult as people think. The point of the tools is that they allow us to see the code. They allow us to understand what's good and bad and to make very clear, rational, and logical decision.

Gardner: Steve Woods, we spoke earlier about how the core assets are going to be variable from organization to organization, but are there some common themes with the contextual services? We certainly see a lot of very low-cost alternatives now creeping up through software as a service (SaaS), cloud-based, outsourced, mix-sourced, co-located, and lots of different options. Is there some common theme now among what is not core that organizations need to consider?

Woods: Absolutely. One of the things that we do find, when we're brought in to look at legacy applications, is that by virtue of the fact that they are still around, the applications have resisted all the waves of innovation that have preceded. Sometimes, they tend to be of a very definite nature.

A number of them tend to be big data hubs. One of the first things we ask is for the architectural topology diagram, if they have it, or we just draw it on a whiteboard,, they tend to be big spiders. There tends to be a central hub database and you see that they start drawing all these different lines to other different systems within the organization.

The things that have been left behind -- this is the good news -- tend to be the very things that are very amenable for moving to modern architecture in a very incremental way. It's not unusual to find 50-65 percent of an application is just doing ETL functionality.

A good thing

The real benefit to that -- and this is particularly true in a tough economy -- is that if I can identify 65 percent of the application that's just doing data integration, and I create or I have already established the data integration center of excellence within the organization, already have those technologies, or implement those technologies, then I can incrementally start moving that functionality over to the new architecture. When I say incrementally, that's a good thing, because that's beneficial in two ways.

It reduces my risk, because I am doing it a step at a time. It also produces a much better ROI, because the return on the incremental improvement is going to be trickling over time, rather than waiting for 18 months or two years for some big bang type of improvement. Identifying this context code can give you a lot of incremental ROI opportunities, and make you a much more solid IT investment decision picture.

Gardner: So, one of these innovations that's taken place for the past several years is the move towards more distributed data, hosting that data on lower-cost storage architectures, and virtualizing behind the database or the storage itself. That can reduce cost dramatically.

Woods: Absolutely. One of the things that we feel is that decentralizing the architecture improves your efficiency and your redundancy. There is much more opportunity for building a solid, maintainable architecture than there would be if you kept a sort of monolithic approach that's typical on the mainframe.

Gardner: Once we've done this exercise, variable as it may be from organization to organization, separating the core from the non-core, what comes next? What's the next step that typically happens as this transformation and modernization of legacy assets unfolds?

So, if you accept the premise of moving context code to componentized architecture, then the next thing you should be looking for is where is the clone code and how is it arranged?



Woods: That's a very good question. It's really important to understand this leap in logic here. If I accept the notion that a majority of the code in a legacy application can be moved to these model driven architectures, such as BPM and ETL tools, the next premise is, "If I go out and buy these tools, a lot of functionality is provided with these tools right out of the box. It's going to give me my monitoring code, my management code, and in many cases, even some of the testing capabilities are sort of baked into the product."

If that's true, then the next leap of logic is that in my 1.5 million lines of COBOL or my five million lines of COBOL there is a lot of code that's irrelevant, because it's performing management, monitoring, logging, tracing, and testing. If that's true, I need to know where it's at.

The way you find where it's at is identifying the duplicate source code, what we call clone code. Because when you find the clone code, in most cases, it's a superset of that code that's no longer relevant, if you are making this transformation from handwritten code to a model-driven architecture.

What I created at HP is a tool, an algorithm, that can go into any language legacy code and find the duplicate code, and not only find it, but visualize it in very compelling ways. That helps us drill down to identify what I call the unintended design. When we find these unintended designs, they lead us to ask very critical questions that are paramount to understanding how to design the transformation strategy.

So, if you accept the premise of moving context code to componentized architecture, then the next thing you should be looking for is where is the clone code and how is it arranged?

Gardner: Do we have any examples of how this has worked in practice? Are there use cases or an actual organization that you are familiar with? What have been some of the results of going through this process? How long did it take? What did they save? What were the business outcomes?

Viewing the application

Woods: We've often worked with financial services companies and insurance companies, and we have just recently worked with one that gave us an application that was around 1.2 or 1.5 million lines of code. They said, "Here is our application," and they gave us the source code. When we looked into the source code, we found that there were actually four applications, if you looked at just the way the code was structured, which was good news, because it gives us a way of breaking down the functionality.

In this one organization, we found that a high percentage of that code was really just taking files, as I said before, unbundling those files, parsing them, and putting them into databases. So they have kind of let that be the tip of the spear. They said, "That's our start point," because they're often asking themselves where to start.

When you take handwritten code and move it to an ETL tool, there's ample industry evidence that a typical ROI over the course of four years can be between 150 percent and 450 percent improvement in efficiencies. That's just the magic of taking all this difficult-to-maintain spaghetti code and moving it to a very visually oriented tool that gives you much more agility and allows you to respond to changes in the business and the business' needs much more quickly and with skill sets that are readily available.

Gardner: You know, Paul, I've heard a little different story from some of the actual suppliers of legacy systems. A lot of times they say that the last thing you want to do is start monkeying around with the code. What you really want to do is pull it off of an old piece of hardware and put it on a new piece of hardware, perhaps with a virtualization layer involved as well. Why is that not the right way to go?

Evans: Now you've put me in an interesting position. I suppose our view is that there are different strategies. We don't profess one strategy to help people transform or modernize their apps. The first thing they have to do is understand them, and that's what Steve's tools do.

The point is that we don't have a preconceived view of what this thing should run on. That's one thing. We're not wedded to one architectural style.



It is possible to take an approach that says that all we need to do is provide more horsepower. Somebody comes along and says, "Hey, transaction rates are dropping. Users are getting upset because an ATM transaction is taking a minute, when it should take 15 seconds. Surely all we need to do is just give the thing more horsepower and the problem goes away."

I would say the problem goes away -- for 12 months, maybe, or if you're lucky 18 -- but you haven't actually fixed the problem. You've just treated the symptoms.

At HP, we're not wedded to one style of computer architecture as the hub of what we do. We look at the customer requirement. Do we have systems that are equal in performance, if not greater, than a mainframe? Yeah, you bet we do. Our Superdome systems are like that. Are they the same price? No, they are considerably less. Do we have blades, PCs, and normal distributed service? Yeah.

The point is that we don't have a preconceived view of what this thing should run on. That's one thing. We're not wedded to one architectural style. We look at the customer's requirements and then we understand what's necessary in terms of the throughput TP rates or whatever it may be.

So, there is obviously an approach that people can say, "Don't jig around." It's very easy to inject fear into this and just say to put more power underneath it, don't touch the code, and life will be wonderful. We're totally against that approach, but it doesn't mean that one of our strategies is not re-hosting. There are organizations whose applications would benefit from that.

We still believe that can be done on relatively inexpensive hardware. We can re-host an application by keeping the business logic the same, keeping the language the same, but moving it from an expensive system to a less expensive system.

Freeing up cash

People use that strategy to free up cash very quickly. It's one of the fastest ROIs we have, and they are beginning to save instantly. They make the decision that says, "We need to put that money back in the bank, because we need to do that to keep our shareholders happy." Or, they can reinvest that into their next modernization project, and then they're on an upward spiral.

There are approaches to everything, which is why we have seven different strategies for modernization to suit the customer's requirement, but I think the view of just putting more horsepower underneath, closing your eyes, and hoping is not the way forward.

Gardner: Steve, do you have anything more to add to that, treating the symptom rather than the real issues?

Woods: As Paul said, if you treat this as a symptom, we refer to that as a short-term strategy, just to save money to reinvest into the business.

The only thing I would really add to that is that the problem is sometimes not nearly as big as it seems. If you look at the analogy of the clone codes that we find, and all the different areas that we can look at the code and say that it may not be as relevant to a transformation process as you think it is.

The subject matter experts and the stakeholders very slowly start to understand that this is actually possible. It's not as big as we thought.



I do this presentation called "Honey I Shrunk the Mainframe." If you start looking at these different aspects between the clone code and what I call the asymmetrical transformation from handwritten code to model driven architecture, you start looking at these different things. You start really seeing it.

We see this, when we go in to do the workshops. The subject matter experts and the stakeholders very slowly start to understand that this is actually possible. It's not as big as we thought. There are ways to transform it that we didn't realize, and we can do this incrementally. We don't have to do it all at once.

Once we start having those conversations, those who might have been arguing for a re-host suddenly realize that rearchitecting is not as difficult as they think, particularly if you do it asymmetrically. Maybe they should reconsider the re-host and consider going to those context-core concept and start moving the context to these well-proven platforms, such as the ETL tools, the reporting tools, and service-oriented architecture (SOA).

Gardner: Steve, tell us a little bit about how other folks can learn more about this, and then give us a sneak peek or preview into what you are going to be discussing at the upcoming virtual event.

Woods: That's one of the things that we have been talking about -- our tools called the Visual Intelligence Tools. It's a shame you can't see me, because I'm gesturing with my hands as I talk, and If I had the visuals in front of me, I would be pointing to them. This is something to really appreciate -- the images that we give to our customers when we do the analysis. You really have to see it with your own eyes.

We are going to be doing a virtual event on November 3, 4, and 5, and during this you will hear some of the same things I've been talking about, but you will hear them as I'm actually using the tools and showing you what's going to happen with those tools, what those images look like, and why they are meaningful to designing a transformation strategy.

Gardner: Very good. We've been learning more about Application Transformation: Getting to the Bottom Line, and we have been able to separate core from context, and appreciate better how that's an intriguing strategy for approaching this legacy modernization problem and begin to enjoy much greater economic and business benefits as a result.

Helping us weave through this has been Steve Woods, distinguished software engineer at HP. Thanks for your input, Steve.

Woods: Thank you.

Gardner: We've also been joined by Paul Evans, worldwide marketing lead on Applications Transformation at HP. Paul, you are becoming a regular on our show.

Evans: Oh, I'm sorry. I hope I am not getting too repetitive.

Gardner: Not at all. Thanks again for your input.

This is Dana Gardner, principal analyst at Interarbor Solutions. You've been listening to a sponsored BriefingsDirect podcast. Thanks for listening, and come back next time.

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Download the transcript. Learn more. Sponsor: Hewlett-Packard.


Gain more insights into "Application Transformation: Getting to the Bottom Line" via a series of HP virtual conferences Nov. 3-5. For more on Application Transformation, and to get real time answers to your questions, register to the virtual conferences for your region:
Register here to attend the Asia Pacific event on Nov. 3.
Register here to attend the EMEA event on Nov. 4.
Register here to attend the Americas event on Nov. 5.


Transcript of the second in a series of sponsored BriefingsDirect podcasts on the rationale and strategies for application transformation. Copyright Interarbor Solutions, LLC, 2005-2009. All rights reserved.