Thursday, December 05, 2013

Service Virtualization Solves Bottlenecks Amid Complex Billing Process for German Telco

Transcript of a BriefingsDirect podcast on how a large telco in Germany has optimized the testing and development procedure with advanced service virtualization tools from HP.

Listen to the podcast. Find it on iTunes. Download the transcript. Sponsor: HP.

Dana Gardner: Hello, and welcome to the next edition of the HP Discover Podcast Series. I'm Dana Gardner, Principal Analyst at Interarbor Solutions, your host and moderator for this ongoing discussion of IT innovation and applications transformation.

Gardner
Once again, we're focusing on how software testing improvements and advanced service virtualization solutions are enabling IT leaders to deliver better experiences for businesses and end users alike.

Today, we’re here to learn how German telco EWE TEL has solved performance complexity across an extended enterprise billing process by using service virtualization. In doing so, EWE has significantly improved applications performance and quality for their end users, while also gaining predictive insights in the composite application services behavior.

Here to explain, how EWE is leveraging service virtualization technologies and techniques for composite applications, we're joined by Bernd Schindelasch, Leader for Quality Management and Testing at EWE TEL based in Oldenburg, Germany. Bernd will be presenting on this use-case next week at the HP Discover conference in Barcelona. Welcome to BriefingsDirect, Bernd.

Bernd Schindelasch: Hi, Dana. Thank you for having me. [Disclosure: HP is a sponsor of BriefingsDirect podcasts.]

Gardner: First, tell us a little bit about EWE TEL, what it does and what you do there.

Schindelasch
Schindelasch: EWE is a telecommunications company. We operate the network for EWE and we provide a large range of telecommunications services. So we invest a lot of money into infrastructure and we supply the region with high-speed Internet access. EWE TEL was founded in 1996, is a fully owned subsidiary of EWE, and has about 1,400 employees.

Gardner: Your software and IT systems are obviously so important. This is how you interact with your end-users. So these applications must be kept performing.

Schindelasch: Yes, indeed. Our IT systems are very important for us to fulfill our customers’ needs. We have about 40 applications, which are involved in the process of a customer, starting from customer self-service application, to the activation component, and the billing system. It’s a quite complex infrastructure and it’s all based on our IT systems.

Gardner: What have you done over the past several years to put together a team or a process through which you can make sure that your applications are performing and continue to perform time and time again?

Schindelasch: We have a special situation here. Because the telecommunications business is very specialized, we need very customized IT solutions. Often, the effort to customize standard software is so high that we decided to develop a lot of our applications on our own.

Developed in house

Nearly half of our applications are developed in house, for example, the customer self service portal I just mentioned, or our customer care system or Activation Manager.

We had to find a way to test it. So we created a team to test all those systems we developed on our own. We recruited personnel from the operating departments and added IT staff, and we started to certify them all as testers. We created a whole new team with a common foundation, and that made it very easy for us to agree on roles, tasks, processes, and so on, concerning our tests. 

Gardner: Today, we’re interested in hearing about how you adopted service virtualization as a technology and a process. Tell me about the problem that led you to discover service virtualization as a solution?

Schindelasch: When we created this new team, we faced the problem of testing the systems end to end. When you have 40 applications and have to test an end-to-end process over all of those applications, all the contributing applications have to be available and have to have a certain level of quality to be useful.
We created a whole new team with a common foundation, and that made it very easy for us to agree on roles, tasks, processes, and so on, concerning our tests. 

What we encountered was that the order interface of another service provider was often unavailable and responses from that system were faulty. So we hadn’t been able to test our processes end to end.

We once tried to do a load test and, because of the bottleneck of that other interface, we experienced the failure of that other interface and weren’t able to test our own systems. That’s the reason we needed a solution to bypass this problem with the other interface. That was the initial initiative that we had.

Gardner: I think you’re representative of many more companies that are dealing with extended enterprise applications and services, ones they can’t fully control, can’t access, and can't get to, but they have to continue to be responsible for the quality of the end process. It can be a quite difficult problem to solve.

Why weren’t traditional testing or scripting technologies able to help you in this regard?

Schindelasch: We tried it recently. We developed diverse simulations based on traditional mockup scripts. These are very useful for developers to do unit testing, but they weren’t configurable for testers to be used to create the right situations for positive and negative tests.

Additionally, there was a big effort to create these mockups, and sometimes the effort to create the mockup would have been bigger than the real development effort. That was the problem we had.

Complex and costly

Gardner: So any simulations you were approaching were going to be very complex and very costly. It didn't really seem to make sense. So what did you do then?

Schindelasch: We constantly analyzed the market and searched for products that might be able to help us with our problem. In 2012, we found such solutions and finally made a proof of concept (POC) with HP Service Virtualization.

We found that it supported different protocols, all the protocols we needed, and with a rule set to predict the responses. During the POC we found that benefits were both for developers and testers. Even our architects found it to be a good solution. So in the end, we decided to purchase that software this year.

Gardner: Tell us how you’ve implemented HP Service Virtualization and how this pilot project has proceeded.

Schindelasch: We implemented service virtualization in a pilot project and we virtualized even that order interface we talked about. We had to integrate service virtualization as a proxy between our customer care system and the order system. The actual steps you have to take vary by the used protocols, but you have to put it in between them and let the system work as a proxy. Then, you have the ability to let it learn.
That reduced our efforts and cost in development and testing and it’s the basis for further test automation at low testing cost.

It’s in the middle, between your systems, and records all messages and their responses. Afterward, you can just replay this message response or you can improve the rules manually. For example, you can add data tables so you can configure the system to work with the actual test data you are using for you test cases to be able to support positive and negative tests. 

Gardner: For those folks that aren’t familiar with HP Service Virtualization for composite applications, how has this developed in terms of its speed and its cost? What are some of the attributes of it that appeal to you?

Schindelasch: Our main objective was to find a way to do our end-to-end testing to optimize it, but we were able to gain more benefits by using service virtualization. We’ve reduced the effort to create simulations by 80 percent, which is a huge amount, and have been able to virtualize services that were still under development.

So we have been able to uncouple the tests of the self service application from a new technical feasibility check. Therefore, we’ve been able to test earlier in our processes. That reduced our efforts and cost in development and testing and it’s the basis for further test automation at low testing cost.

In the end, we’ve improved quality. It’s even better for our customers, because we’re able to deliver fast and have a better time to market for new products. 

Future attributes

Gardner: Are there other attributes that you’d like to see in future products, perhaps with network-virtualization attributes? I know that you’ve been doing this with certain middleware, messaging, and workflow technology. What would you like to see next?

Schindelasch: One important thing is that development is shifting to agile more and more. Therefore, the people using the software have changed. So we have to have better integration with development tools.

From a virtualization perspective, there will be new protocols, more complex rules to address every situation you can think of without complicated scripting or anything like that. I think that’s what’s coming in the future.

Gardner: And, Bernd, has the use of HP Service Virtualization allowed you to proceed toward more agile development and, as well, to start to benefit from DevOps, more tight association and integration between development and deployment and operations?
Service virtualization has the potential to change the performance model, so you can let your application answer slower or faster.

Schindelasch: We already put it together with our development, I think it’s very crucial to cooperate with development and testing, because there wouldn’t be a real benefit to virtualize the service after development already mocked up in an old-fashioned way.

We brought them together. We had the training for a lot of developers. They started to see the benefits and started to use service virtualization the way the testers already did.

We’re working together more closely and earlier in the process. What’s coming in the future is that the developers will start to use service virtualization for their continuous integration, because service virtualization has the potential to change the performance model, so you can let your application answer slower or faster.

If you put it into fast mode, then you use it in continuous integration. That’s a really big benefit for the developers, because their continuous integration will be faster and therefore they will be able to deploy faster. So for our development, it’s a real benefit.

Gardner: I should think that for an organization like yours, where you’re a services provider, being able to meet your service-level agreements (SLAs) is important to you. This could probably have a very positive impact on that.

Schindelasch: Yes, definitely.

Lessons learned

Gardner: Before we end our discussion, I wonder if you could maybe offer some insights to those who are considering the use of service virtualization with composite applications now that you have been doing it. Are there any lessons learned? Are there any suggestions that you would make for others as they begin to explore new service virtualization in the testing phase?

Schindelasch: One thing I’ve already mentioned is that it’s important to work together with development and testing. To gain maximum benefit from HP Service Virtualization, you have to design your future solutions. What service do you want to virtualize, which protocols will you use, and where are the best places to intercept? Do I want to replace real systems or create the whole environment as virtualized? In which way do I want to use performance model and so on?

It’s very important to really understand what your needs are before you start using the tools and just virtualize everything. It’s easy to virtualize, but there is no real benefit if you virtualize a lot of things you didn’t really want. As always, it’s important to think first, design your future solutions, and then start to do it.
It’s very important to really understand what your needs are before you start using the tools and just virtualize everything.

Gardner: I am afraid we’ll have to leave it there. We’ve been learning how German telco EWE has solved performance complexity across an extended enterprise billing process using HP Service Virtualization.

We have heard how EWE has significantly improved applications performance and quality for their end users, while also gaining predictive insights in the composite applications services behavior, even back into the development phases. Bernd will be presenting on this use case next week at the HP Discover conference in Barcelona.

I’d like to thank our supporter for the series, HP Software, and reminder our audience to carry on the dialogue through the IT Strategy & Performance group on LinkedIn. You can always access this and another episodes in our HP Discover podcast series on iTunes under BriefingsDirect.

And so a big thanks to our guest. We’ve been joined by Bernd Schindelasch, Leader for Quality Management and Testing at EWE TEL based in Oldenburg, Germany. Thank you so much, Bernd.

Schindelasch: Thank you, Dana.

Gardner: This is Dana Gardner, Principal Analyst at Interarbor Solutions, your host and moderator for this ongoing discussion on IT innovation. Thanks again for listening, and come back next time.

Listen to the podcast. Find it on iTunes. Download the transcript. Sponsor: HP.

Transcript of a BriefingsDirect podcast on how a large telco in Germany has optimized the testing and development procedure with advanced service virtualization tools from HP. Copyright Interarbor Solutions, LLC, 2005-2013. All rights reserved.

You may also be interested in:

Wednesday, December 04, 2013

Identity and Access Management as a Service Gets Boost with SailPoint's IdentityNow Cloud Service

Transcript of a BriefingsDirect podcast on the need for and innovation in improved identity and access management.

Listen to the podcast. Find it on iTunes. Download the transcript. Sponsor: SailPoint Technologies.

Dana Gardner: Hi, this is Dana Gardner, Principal Analyst at Interarbor Solutions, and you’re listening to BriefingsDirect.

Gardner
Today, we present a sponsored podcast discussion on the changing needs for, and heightened value around, improved identity and access management (IAM). We'll examine now how business trends are forcing organizations to safely allow access to all kinds of applications and myriad resources anytime, anywhere, and from any device.

According to research firm MarketsandMarkets, the demand for IAM is therefore estimated to grow from more than $5 billion this year to over $10 billion in 2018. What's driving the doubling of the market in five years? Well, as with much of the current IT space, it's about cloud, mobile, bring your own device (BYOD), consumerization of IT, and broader security concerns.

But the explosive growth also factors the move to more pervasive use of identity and access management as a service (IDaaS).

So join us now as we explore how new IDaaS offerings are helping companies far better protect and secure their informational assets. Here to share insights into this future of identity management is Paul Trulove, Vice President of Product Marketing at SailPoint Technologies in Austin, Texas. Welcome, Paul. [Disclosure: SailPoint is a sponsor of BriefingsDirect podcasts.]

Paul Trulove: Thanks, Dana. Glad to be here.

Gardner: The word "control" comes up so often when I talk to people about security and IT management issues, and companies seem to feel that they are losing control, especially with such trends as BYOD. How do companies regain that control, or do we need to think about this differently. Is it no longer an issue of control?

Trulove: The reality in today's market is that a certain level of control will always be required. But as we look at the rapid adoption of new corporate enterprise resources, things like cloud-based applications or mobile devices where you could access corporate information anywhere in the world at any time on any device, the reality is that we have to put a base level of controls in place that allow organizations to protect the most sensitive assets. But you have to also provide ready access to the data, so that the organizations can move at the pace of what the business is demanding today.

Gardner: The expectations of users has changed. When they can go sign up for a software-as-a-service (SaaS) application or access cloud services, they're used to having more of their own freedom. How is that something that we can balance, allow them to get the best of their opportunity and their productivity benefits, but at the same time, allow for the enterprise to be as low risk as possible?

Trulove
Trulove: That's the area that the organization has to find the right balance for their particular business that meets the internal demands, the external regulatory requirements, and really meet the expectations of their customer base. While the productivity aspect can't be ignored, taking a blind approach to allowing an individual end-user to begin to migrate structured data out of something like an SAP or other enterprise resource planning (ERP) systems, up to a personal Box.com account is something most organizations are just not going to allow.

Each organization has to step back, redefine the different types of policies that they're trying to put in place, and then put the right kind of controls that mitigate risk in terms of inappropriate acts, access to critical enterprise resources and data, but also allow the end user to have a little bit more control and little bit more freedom to do things that make them the most productive.

Uptake in SaaS

Gardner: We've seen a significant uptake in SaaS, certainly at the number of apps level, communications, and email, but it seems as if some of the infrastructure services around IAM are lagging. Is there a maturity issue here, or is it just a natural way that markets evolve? What's the case in understanding why the applications have gone fast, but we're now just embarking on IDaaS?

Trulove: We're seeing a common trend in IT if you look back over time, where a lot of the front-end business applications were the first to move to a new paradigm. Things like ERP and service resource management (SRM)-type applications have all migrated fairly quickly.

Over the last decade, we've really seen a lot of the sales management applications, like Salesforce and NetSuite come on as full force. Now, there are things like Workday and even some of the work force management becoming very popular. However, the infrastructure generally lagged for a variety of reasons.

In the IAM space, this is a critical aspect of enterprise security and risk management as it relates to guarding the critical assets of the organization. Security practitioners are going to look at new technology very thoroughly before they begin to move things like IAM out to a new delivery paradigm such as SaaS.

The other thing is that organizations right now are still fundamentally protecting internal applications. So there's less of a need to move your infrastructure out into the cloud until you begin to change the overall delivery paradigm for your internal application.
As customers implement more and more of their software out in the cloud, that's a good time for them to begin to explore IDaaS.

What we're seeing in the market, and definitely from a customer perspective, is that as customers implement more and more of their software out in the cloud, that's a good time for them to begin to explore IDaaS.

Look at some of the statistics being thrown around. In some cases, we've seen that 80 percent of new software purchases are being pushed to a SaaS model. Those kinds of companies are much more likely to embrace moving infrastructure to support that large cloud investment with fewer applications to be managed back in the data center.

Gardner: As you mentioned, SaaS has been around for 10 years, but the notion of mobile-first applications now has picked up in just the last two or three years. I have to imagine that's another accelerant to looking at IAM differently when you get the devices.

We've talked a little bit about SaaS and IDaaS, coming on as a follow up, how does the mobile side of things impact this?

Trulove: Mobile plays a huge part in organizations' looking at IDaaS, and the reason is that you’re moving the device that's interacting with the identity management service outside the bounds of the firewall and the network. So, having a point of presence in the cloud gives you a very easy way to generate all of the content out to the devices that are being operated outside of the traditional bounds of the IT organization, which was generally networked in to the PCs, laptops, etc that are on the network itself.

Moving to IDaaS

Gardner: I'd like to get into what hurdles organizations need to overcome to move in to IDaaS, but let's define this a little better for folks that might not be that familiar with it. How does SailPoint define IDaaS? What are we really talking about?

Trulove: SailPoint looks at IDaaS as a set of capabilities across compliance and governance, access request and provisioning, password management, single sign-on (SSO), and Web access management that allow for an organization to do fundamentally the same types of business processes and activities that they do with an internal IAM systems, but delivered from the cloud.

We also believe that it's critical, when you talk about IDaaS to not only talk about the cloud applications that are being managed by that service, but as importantly, the internal applications behind the firewall that still have to be part of that IAM program.

Gardner: So, this is not just green field. You have to work with what's already in place, and it has to work pretty much right the first time.

Trulove: Yes, it does. We really caution organizations against looking at cloud applications in a siloed manner from all the things that they're traditionally managing in the data center. Bringing up a secondary IAM system to only focus on your cloud apps, while leaving everything that is legacy in place, is a very dangerous situation. You lose visibility, transparency, and that global perspective that most organizations have struggled to get with the current IAM approaches across all of those areas that I talked about.
We see a little bit less of the data export concerns with companies here in the US, but it's a much bigger concern for companies in Europe and Asia in particular.

Gardner: So, we recognize that these large trends are forcing a change, users want their freedom, more mobile devices, more different services from different places, and security being as important if not more than ever. What is holding organizations back from moving towards IDaaS, given that it can help accommodate this very complex set of requirements?

Trulove: It can. The number one area, and it's really made up of several different things, is the data security, data privacy, and data export concerns. Obviously, the level at which each of those interplay with one another, in terms of creating concern within a particular organization, has a lot to do with where the company is physically located. So, we see a little bit less of the data export concerns with companies here in the US, but it's a much bigger concern for companies in Europe and Asia in particular.

Data security and privacy are the two that are very common and are probably at the top of every IT security professional’s list of reasons why they're not looking at IDaaS.

Gardner: It would seem that just three or four years ago, when we were talking about the advent of cloud services, quite a few people thought that cloud was less secure. But I’ve certainly been mindful of increased and improved security as a result of cloud, particularly when the cloud organization is much more comprehensive in how they view security.

They're able to implement patches with regularity. In fact, many of them have just better processes than individual enterprises ever could. So, is that the case here as well? Are we dealing with perceptions? Is there a case to be made for IDaaS being, in fact, a much better solution overall?

IAM as secure

Trulove: Much like organizations have come to recognize the other categories of SaaS as being secure, the same thing is happening within the context of IAM. Even a lot of the cloud storage services, like Box.com, are now signing up large organizations that have significant data security and privacy concerns. But, they're able to do that in a way and provide the service in a way where that assurance is in place that they have control over the environment.

And so, I think the same thing will happen with identity, and it's one of the areas where SailPoint is very focused on delivering capabilities and assurances to the customers that are looking at IDaaS, so that they feel comfortable putting the kinds of information and operating the different types of IAM components, so that they get over that fear of the unknown.

Gardner: Before we get into some of the details about how you’re approaching this, and what your services can provide, I'm curious about what companies can expect to get when they pursue the full cloud and services panoply of possibilities across apps, data, IT management, and other services. What are some of the business drivers? What do you get if you do this right and you make the leap to the services’ strata?

Trulove: One of the biggest benefits of moving from a traditional IAM approach to something that is delivered as IDaaS is the rapid time to value. It's also one of the biggest changes that the organization has to be prepared to make, much like they would have as they move from a Siebel- to a Salesforce-type model back in the day.

IAM delivered as a service needs to be much more about configuration, versus that customized solution where you attempt to map the product and technology directly back to existing business processes.
The benefit that they get out of that is a much lower total cost of ownership (TCO), especially around the deployment aspects of IDaaS.

One of the biggest changes from a business perspective is that the business has to be ready to make investments in business process management, and the changes that go along with that, so that they can accommodate the reality of something that's being delivered as a service, versus completely tailoring a solution to every aspect of their business.

The benefit that they get out of that is a much lower total cost of ownership (TCO), especially around the deployment aspects of IDaaS.

Gardner: It's interesting that you mentioned business process and business process management. It seems to me that by elevating to the cloud for a number of services and then having the access and management controls follow that path, you’re able to get a great deal of flexibility and agility in how you define who it is you’re working with, for how long, for when.

It seems to me that you can use policies and create rules that can be extended far beyond your organization’s boundaries, defining workgroups, defining access to assets, creating and spinning up virtualized companies, and then shutting them down when you need. So, is there a new level of consideration about a boundaryless organization here as well?

Trulove: There is. One of the things that is going to be very interesting is the opportunity to essentially bring up multiple IDaaS environments for different constituents. As an organization, I may have two or three fundamentally distinct user bases for my IAM services.

Separate systems

I may have an internal population that is made up of employees, and contractors that essentially work for the organization that need access to a certain set of systems. So I may bring up a particular environment to manage those employees that have specific policies and workflows and controls. Then, I may bring up a separate system that allows for business partners or individual customers to have access to very different environments within the context of either cloud or on-prem IT resources.

The advantage is that I can deploy these services uniquely across those. I can vary the services that are deployed. Maybe I provide only SSO and basic provisioning services for my external user populations. But for those internal employees, I not only do that, but I add access certifications, and segregation of duties (SOD) policy management. I need to have much better controls over my internal accounts, because they really do guard the keys to the kingdom in terms of data and application access.

Gardner: We began this conversation talking about balance. It certainly seems to me that that level of ability, agility, and defining new types of business benefits far outweighs some of the issues around risk and security that organizations are bound to have to solve one way or the other. So, it strikes me as a very compelling and interesting set of benefits to pursue.

Let's look now, Paul, at your products. You've delivered the SailPoint IdentityNow suite. You've got a series of capabilities, and there are more to come. As you were defining and building out this set of services, what were some of the major requirements that you had, that you needed to check off before you brought this to market?

Trulove: The number one capability that we really talk to a lot of customers about is an integrated set of IAM services that span everything from that compliance and governance to access request provisioning and password management all the way to access management and SSO.
They can get value out of it, not necessarily on day one, but within weeks, as opposed to months.

One of the things that we found as a critical driver for the success of these types of initiatives within organizations is that they don't become siloed, and that as you implement a single service, you get to take advantage of a lot of the work that you've done as you bring on the second, third, or fourth services.

The other big thing is that it needs to be ready immediately. Unlike a traditional IAM solution, where you might have deployment environments to buy and implement software to purchase and deploy and configure, customers really expect IDaaS to be ready for them to start implementing the day that they buy.

It's a quick time-to-value, where the organization deploying it can start immediately. They can get value out of it, not necessarily on day one, but within weeks, as opposed to months. Those things were very critical in deploying the service.

The third thing is that it is ready for enterprise-level requirements. It needs to meet the use cases that a large enterprise would have across those different capabilities, but also as important, that it meets data security, privacy, and export concerns that a large enterprise would have relative to beginning to move infrastructure out to the cloud.

Even as a cloud service, it needs a very secure way to get back into the enterprise and still manage the on-prem resources that aren’t going away anytime soon. n one hand we would talk to customers about managing things like Google Apps, Salesforce and Workday. In the same breath, they also talk about still needing to manage the mainframe and the on-premises enterprise ERP system that they have in place.

So, being able to span both of those environments to provide that secure connectivity from the cloud back into the enterprise apps was really a key design consideration for us as we brought this product to market.

Hybrid model

Gardner: It sounds if it's a hybrid model from the get-go. We hear about public cloud, private cloud, and then hybrid. It sounds as if hybrid is really a starting point and an end point for you right away.

Trulove: It's hybrid only in that it's designed to manage both cloud and on-prem applications. The service itself all runs in the cloud. All of the functionality, the data repositories, all of those things are 100 percent deployed as a service within the cloud. The hybrid nature of it is more around the application that it's designed to manage.

Gardner: You support a hybrid environment, but I see, given what you've just said, that means that all the stock in trade and benefits as a service offering are there, no hardware or software, going from a CAPEX to OPEX model, and probably far lower cost over time were all built in.

Trulove: Exactly. The deployment model is very much that classic SaaS, a multitenant application where we basically run a single version of the service across all of the different customers that are utilizing it.

Obviously, we've put a lot of time, energy, and focus on data protection, so that everybody’s data is protected uniquely for their organization. But we get the benefits of that SaaS deployment model where we can push a single version of the application out for everybody to use when we add a new service or we add new capabilities to existing services. We take care of upright processes and really give the customers that are subscribing to the services the option of when and how they want to turn new things on.
We've put a lot of time, energy, and focus on data protection, so that everybody’s data is protected uniquely for their organization.

Gardner: Let's just take a moment and look at the SailPoint IdentityNow suite. Tell me what it consists of, and how this provides a benefit and on-ramp to a better way of doing IT as a service and business as a service.

Trulove: The IdentityNow suite is made up of multiple individual services that can be deployed distinctly from one another, but all leverage a common back-end governance foundation and common data repository.

The first service is SSO and it very much empowers users to sign on to cloud, mobile, and web applications from a single application platform. It provides central visibility for end users into all the different application environments that they maybe interacting with on a daily basis, both from a launch-pad type of an environment, where I can go to a single dashboard and sign on to any application that I'm authorized to use.

Or I may be using back-end Integrated Windows Authentication, where as soon as I sign into my desktop at work in the morning, I'm automatically signed into all my applications as I used them during the day, and I don’t have to do anything else.

The second service is around password management. This is enabling that end-user self-service capability. When end users need to change their password or, more commonly, reset them because they’ve forgotten them over a long weekend, they don’t have to call the help desk.

Strong authentication

They can go through a process of authenticating through challenge questions or other mechanisms and then gain access to reset that password and even use some strong authentication mechanisms like one-time password tokens that are going to be issued, allow the user to get in and then, change that password to something that they will use on an ongoing basis.

The third service is around access certifications, and this automates that process of allowing organizations to put in place controls through which managers or other users within the organization are reviewing who has access to what on a regular basis. It's a very business-driven process today, where an application owner or business manager is going to go in, look at the series of accounts and entitlements that a user has, and fundamentally make a decision whether that access is correct at a point in time.

One of the key things that we're providing as part of the access certification service is the ability to automatically revoke those application accounts that are no longer required. So there's a direct tie into the provisioning capabilities of being able to say, Paul doesn’t need access to this particular active directory group or this particular capability within the ERP system. I'm going to revoke it. Then, the system will automatically connect to that application and terminate that account or disable that account, so the user no longer has access.

The final two services are around access request and provisioning and advanced policy and analytics. On the access request and provisioning side, this is all about streamlining, how users get access. It can be the automated birth-right provisioning of user accounts based on a new employee or contractor joining new organization, reconciling when a user moves to a new role, what they should or should not have, or terminating access on the back end when a user leaves the organization.
What most customers see, as they begin to deploy IDaaS is the ability to get value very quickly.

All of those capabilities are provided in an automated provisioning model. Then we have that self-service access request, where a user can come in on an ad-hoc basis and say, "I'm starting a new project on Monday and I need some access to support that. I'm going to go in, search for that access. I'm going to request it." Then, it can go through a flexible approval model before it actually gets provisioned out into the infrastructure.

The final service around advanced policy and analytics is a set of deeper capabilities around identifying where risks lie within the organization, where people might have inappropriate access around a segregation of duty violation.

It's putting an extra level of control in place, both of a detective nature, in terms of what the actual environment is and which accounts that may conflict that people already have. More importantly, it's putting preventive controls in place, so that you can attach that to an access request or provisioning event and determine whether a policy violation exists before a provisioning action is actually taken.

Gardner: You've delivered quite a bit in terms of this suite's offering this year. Before we hear some more about some of the roadmap and future capabilities, what are your customers finding now that they are gaining as a result of moving to IDaaS as well, as the opportunity for specific services within the suite? What do you get when you do this right?

Trulove: What most customers see, as they begin to deploy IDaaS is the ability to get value very quickly. Most of our customers are starting with a single service and they are using that as a launching pad into a broader deployment over time.

So you could take SSO as a distinct project. We have customers that are implementing that SSO capability to get rapid time to value that is very distinct and very visible to the business and the end users within their organization.

Password management

Once they have that deployed and up and running, they're leveraging that to go back in and add something like password management or access certification or any combination thereof.

We’re not stipulating how a customer starts. We're giving them a lot of flexibility to start with very small distinct projects, get the system up and running quickly, show demonstrable value to the business, and then continue to build out over time both the breadth of capabilities that they are using but also the depth of functionality within each capability.

Gardner: Do you have any instances, Paul, where folks are saying, "We wanted to go mobile, but we're being held back. Now that we've taken a plunge, this has really opened up a whole new way for us to deliver data and applications to different devices and mobile, whether it’s the campus setting or road warrior setting." Any thoughts about how this is, in particular, aiding and abetting mobile.

Trulove: Mobile is driving a significant increase in why customers are looking at IDaaS. The main reason is that mobile devices operate outside of the corporate network in most cases. If you're on a smartphone and you are on a 3G, 4G, LTE type network, you have to have a very secure way to get back into those enterprise resources to perform particular operations or access certain kinds of data.

One of the benefits that an IDaaS service gives you is a point of presence in cloud that allows the mobile devices to have something that is very accessible from wherever they are. Then, there is a direct and very secure connection back into those on-prem enterprise resources as well as out to the other cloud applications that you are managing.
The other big thing we're seeing in addition to mobile devices is just the adoption of cloud applications.

The reality in a lot of cases is that, as organizations add those BYOD type policies and the number of mobile devices that are trying to access corporate data increase significantly, providing an IAM infrastructure that is delivered from the cloud is a very convenient way to help bring a lot of those mobile devices under control across your compliance, governance, provisioning, and access request type activities.

The other big thing we're seeing in addition to mobile devices is just the adoption of cloud applications. As organizations go out and acquire multiple cloud applications, having a point of presence to manage those in the cloud makes a big difference.

In fact, we've seen several deployment projects of something like Workday actually gated by needing to put in the identity infrastructure before the business was going to allow their end users to begin to use that service. So the combination of both mobile and cloud adoption are driving a renewed focus on IDaaS.

Gardner:  I know you can't actually pre-announce, and I am not asking you to, but as we consider what you can now do with these capabilities, perhaps you can paint a little bit of a vision for us as to where you think your offerings, and therefore the market and the opportunity for improvement in the user organizations, is headed.

Trulove: If you look at the road map that we have for the IdentityNow product, the first three services are available today, and that’s SSO, password management, and access certification. Those are the key services that we're seeing businesses drive into the cloud as early adopters. Behind that, we'll be deploying the access request and provisioning service and the advanced policy and analytic services in the first half of 2014.

Continued maturation

Beyond that, what we're really looking at is continued maturation of the individual services to address a lot of the emerging requirements that we're seeing from customers, not only across the cloud and mobile application environments, but as importantly as they begin to deploy the cloud services and link back to their on-prem identity and access management infrastructure, as well as the applications that they are continuing to run and manage from the data center.

Gardner: So, more inclusive, and therefore more powerful, in terms of the agility, when you can consider all the different aspects of what falls under the umbrella of IAM.

Trulove: We're also looking at new and innovative ways to reduce the deployment timeframes, by building a lot of capabilities that are defined out of the box. These are  things like business processes, where there will be catalog of the best practices that we see a majority of customers implement. That has become a drop-down for an admin to go in and pick, as they are configuring the application.
We're also looking at new and innovative ways to reduce the deployment timeframes, by building a lot of capabilities that are defined out of the box.

We'll be investing very heavily in areas like that, where we can take the learning as we deploy and build that back in as a set of best practices as a default to reduce the time required to set up the application and get it deployed in a particular environment.

Gardner: Well, great. I'm afraid we'll have to leave it there. You've been listening to a sponsored BriefingsDirect podcast discussion on the changing needs for and heightened value around improved IAM, and we have seen how explosive expected growth and change is forcing a move to more a pervasive use of identity and access management as a service or IDaaS.

And, of course, we've learned more about SailPoint Technologies and how they're delivering the means for organizations to safely allow access to all kinds of applications and resources anytime anywhere and from any device.

With that, I'd like to thank our guest, Paul Trulove, Vice President of Product Marketing at SailPoint Technologies. Thanks, Paul.

Trulove: Thank you, Dana. I appreciate the time.

Gardner: This is Dana Gardner, Principal Analyst at Interarbor Solutions. A big thank you also to our audience for joining us, and a reminder to come back and join us again next time.

Listen to the podcast. Find it on iTunes. Download the transcript. Sponsor: SailPoint Technologies.

Transcript of a BriefingsDirect podcast on the need for and innovation in improved identity and access management. Copyright Interarbor Solutions, LLC, 2005-2013. All rights reserved.

You may also be interested in:

Tuesday, December 03, 2013

BI and Big Data Analytics Force an Overdue Reckoning Between IT and Business Interests

Transcript of a Briefings Direct podcast on how advanced analytics aligns business and IT goals to bring further collaboration on innovation within the enterprise.

Listen to the podcast. Find it on iTunes. Download the transcript. Sponsor: Dell Software.

Dana Gardner: Hi, this is Dana Gardner, Principal Analyst at Interarbor Solutions, and you’re listening to BriefingsDirect.

Gardner
Today, we present a sponsored podcast discussion on how big data and business intelligence (BI) trends are hastening a breach between enterprise IT groups and business units. We'll examine how a traditional ebb and flow between IT centralization and decentralization are now swinging in the direction of business groups and even shadow IT. This runs the risk of neglecting essential management security and scalability requirements.

Our discussion now will focus on how big data and analytics should actually force more collaboration and lifecycle-based relationships among and between business and IT groups. For those organizations -- where innovation is being divorced from IT discipline -- we'll explore ways that a comprehensive and virtuous adoption of rigorous and protected data insights can both make the business stronger and make IT more valued.

Here now to share his insights on gaining sustainable, competitive advantage by using enterprise big data strategically, we’re joined by John Whittaker, Senior Director of Marketing for Dell Software's Information Management Solutions Group. Welcome, John. [Disclosure: Dell Software is a sponsor of BriefingsDirect podcasts.]

John Whittaker: Thank you very much, Dana. I’m looking forward to having this conversation and sharing a little bit with your audience about how we at Dell Software look at this particular problem. I’ll offer some suggestions about how we might be able to implement big data and gain lasting sustainable advantage with the solutions that exist today, but doing so in a manner that’s going to deliver long-term success.

Gardner: John, we seem to go back and forth between resources in organizations being tightly controlled and governed by IT, and then resources and control resting largely with the line of business or even, as I mentioned, with a shadow IT group of some sort. So over the past 20 or more years, why has this problem been so difficult to overcome? Why is it persistent? Why do we keep going back and forth?

Whittaker: That’s an interesting question, and I agree. I've been in IT for longer than 20 years and certainly in your study of history you can see that this ebb and flow of centralized management to gain some constraints or some controls in governance and security has been one of the primary motivators of IT. It’s one of the big benefits they provide, but in the backdrop, you have lines of business that want to innovate and want to go in new directions.

Whittaker
Behind all of these things occurring, you have these mega trends that show up, these new technology innovations and these new approaches. There periodically seems to be times when the innovation cycle is really an arc.

We’re entering one of those times right now with big data and the advent of analytics, and it’s driving lines of business to push into these new technologies, and maybe  in ways that IT isn’t ready for just yet.

This, as you mentioned, has been going on for some time. The last iteration where this occurred was back in the ’90s when e-commerce and the Web captured the imagination of business. We saw a lot of similarities to what's occurring today.

Big-data push

It ultimately caused some problems back in the ’90s around e-commerce and leveraging this great new innovation of the Internet, but doing it in a way that was more decentralized. It was a little bit more of the Wild West-based approach and ultimately led to some pretty significant issues that I think we are going to see out of the big data and analytics push that’s occurring right now.

Gardner: I suppose to be fair to each constituency here, it’s the job of IT to be cautious and to try to dot all the i’s and cross the t’s. There were a lot of people in 1996-97 who didn’t necessarily think the Internet was going to be that big of a thing, it seemed to have lots of risk associated with it. So, I suppose due diligence needed to be brought to bear.

http://software.dell.comOn the other hand, if the businesses didn’t recognize that this could be a huge opportunity and we needed to take those risks -- create a website, and enter into a direct dialogue with customers to a new channel -- they would have missed a big opportunity. So these are sort of natural roles, but they can’t be too brittle.

Whittaker: You’re absolutely right. At their core, both groups had, and have, good motivations. IT lives in a world of constraints, of governance, security, and of needing to deliver something that’s going to be stable, that’s going to scale, that’s going to be secure, and that’s not going break governance.
Nobody in either group is trying to harm the business or anything close to it.

Those are laudable goals to have in mind. From the line-of-business perspective, the business wants to innovate and doesn’t want to be outmoded by its competitors. They rightfully see that all these great innovations are coming, and analysts, pundits, and experts are talking about how this is going to make a huge difference for businesses.

So they inevitably want to embrace those, and you have this cognitive dissonance occurring between the IT goals around constraints and the desire to keep things running in a clean and efficient manner. IT is seeing this new technology and saying, “Hold on. We don’t necessarily want to jump into this. This is going to break our model.”

Ultimately, IT gets to a point where maybe they suggest we shouldn’t do it or we should push it off for some time. That’s where the chasm between the two gets started. From the business perspective, the answer “no” is unacceptable, if they feel that’s what they need to do to achieve success in business. They own the profit and loss responsibilities. That’s where these problems come from.

Nobody in either group is trying to harm the business or anything close to it. They just have different motivations and perspectives on how to approach something, and when one gets wildly far apart from the other, that’s where these problems tend to occur. Again, when these big innovation cycles happen, you’re more likely to see a lot of these problems start to occur.

I definitely remember back in 1996-1997. We didn’t call it shadow IT at the time, but you saw IT-like personnel being hired into functional business areas to institute these new technologies, and that ultimately led to a pretty serious hangover at the end of that innovation cycle.

Gardner: What’s the risk of ignoring IT, doing an end-run around them, or downplaying the role? What form does it take?

On their own

Whittaker: Ignoring IT can have some pretty serious problems. It all starts with the fact that, and by and large, businesses can embrace these new technologies without the aid of IT.  Cloud-based implementations have made it possible for lines of business to rapidly deploy some of these new big data technologies, and you have vendors in some cases telling them they don’t need IT’s help. So it’s not all that difficult for lines of business to go out on their own and implement a big data technology.

But they don’t typically have the discipline to apply across-the-board governance capabilities and discipline into their deployment and that leads to potential issues with regulatory requirements. It also leads to security issues, and ultimately can lead to problems where you have seriously bad data management issues.

You have data sunk in silos, and maybe the CEO wants to know how much business we’re doing with x, y, and z. No one can deliver that, because we call x, y, and z, something in one system, a different name in another system, and a different name in the third system. Trying to pull that data together becomes really difficult. When you have lines of business independently operating disparate solutions, those core governance issues tend to break down.

Additionally, although they are great at spotting innovation opportunities, line of business people are not necessarily in the business of building scalable, secure, stable environments. That’s not the core of, say, marketing. They need to understand how the technology can be leveraged, but maintaining and managing it is not core to their charter. It tends to be ignored.

We saw in the early 2000s as the last innovation’s hangover started to occur. We saw people discovering that their systems were not scalable, that they weren’t secure, and that they were unstable because they didn’t have somebody there doing the basic care and feeding that is core to IT. That can lead to some very substantial costs. The cost could be enormous and totally unbound if you start talking about security issues that come from a lack of discipline applied at a governance level.
There are a lot of lessons that can be learned from the concept of working closely together, iterating rapidly, and being open to innovation and the idea that changes occur.

Gardner: John, it strikes me that there are some examples within IT that help understand this potential problem and even grab some remediation, and that’s in software development. We’ve seen the complexity in groups working without a lot of coordination and shared process insights and have run aground.

For many years, we saw a very high failure rate among software development projects, but more recently, we’ve seen improvements -- agile, scrum, opening up the process, small iterative steps that then revert back to an opportunity to take stock and know what everyone is doing, checking in, checking out with centralization -- but without stifling innovation. Is there really a lesson here in what’s happened within software development that could be brought to the whole organization?

Whittaker: Absolutely. In fact, within Dell Software itself we embrace agile and use scrum internally. There are a lot of lessons that can be learned from the concept of working closely together, iterating rapidly, and being open to innovation and the idea that changes occur.

Particularly in these major innovation cycles, it’s important to go with the flow and implement some of these new technologies and new capabilities early, so you can have that brain trust built internally among the broad team. You don’t want IT to hold the reins entirely, and at the same time, you don’t want line of business to do it.

We really need to break that model, that back and forth, centralization-decentralization swing that keeps occurring. We need to get to a point where we really are partnering and have good collaboration, where innovation can be embraced and adopted, and the business can meet its goals. But it has to be done in a way that IT can implement sound governance and implement solutions that can scale, are stable, are reliable, and are going to lead to long-term success.

Back-and-forth

We’ve got to get out of these back-and-forth cycles that occur or we’re going to continue to have these problems. As innovations occur more rapidly, you’re going to have more and more problems like these occurring if you don’t find a way to get IT and line-of-business motivations and interests allied.

Gardner: What’s different this time, John? Are the stakes higher because we’re talking about data analysis? That’s basically intelligence about what’s going on within your markets, your organization, your processes, your supply chain, your ecosystem, all of which could have a huge bearing.

We have the ability now to tackle massive amounts of data very rapidly, but if we don’t bring this together holistically, it seems as if there is a larger risk. I’m thinking about a competitive risk. Others that do this well could enter your market and really disrupt.

Whittaker: You’re absolutely right. There’s great potential benefit that organizations receive or can get out of leveraging big data and analytics, that of being able to determine predictively what is going to occur in their business and what are the most efficient routes to market and what areas of improvements can occur.

The businesses that leverage this are going to outmode, outperform, and ultimately win in the markets currently dominated by organizations who aren’t paying attention and who aren’t implementing solutions today. They’re getting a little bit ahead of this cycle so that they are ready and are able to be successful down the road.
We’re really moving into an era where the context of what’s happening is critically important.

We’re really moving into an era where the context of what’s happening is critically important. A data-driven management model is going to be embraced and it’s ultimately going to lead to more successful organizations. Companies and organizations that embrace this today are going to be the winners tomorrow.

If you’re ignoring this or putting this off, you’re really taking a tremendous risk, because this next iteration of innovation that’s occurring around analytics applies to large data sources. It’s being able to build the correlations and determine that this is a more efficient approach, or conversely, that we have a problem with this outlier that’s going to give us issues down the road.

If you’re not doing that as an organization, you really are running a pretty tremendous risk that somebody else is going to walk in and be able to make smarter decisions, faster.

Gardner: At the same time, your customers are gaining insights into how to procure all the better. And so any rewards that might be out there, if you are in a sales role of any kind, would become much more apparent.

Whittaker: That’s definitely true as well. The construct and the conversation has really shifted. With the advent of social media and the pace at which information is shared and opinions are made, it’s no longer the company that is the primary voice about its products and its capabilities or its positions and point of views.

Customers more empowered

It needs to have those. It needs to get them out. It needs to push them. But in this new world we live in, the customers are so much more empowered than they have ever been before, and it should be a good thing. For companies that are delivering great products and solving real problems for their customers, this should be great news.

If you’re not listening to what your customers are saying in social media and if you’re not paying attention to the ongoing story line and conversation of your firm in the social sphere, you’re really putting yourself at risk. You’re missing out on a tremendous opportunity to engage with your customers in a new, interesting, and very useful way.

That’s a lot of what we built. We have a lot of capabilities here at Dell Software around data management, data integration, and data analysis. On the analysis side, we spend a great deal of time with products like Kitenga and our social networking analytics platforms to do that semantic analysis and look into that form of big data.

But big data is more than just social. It’s also sensor data. The iterative thing is another area where businesses should be innovating and organizations should be pushing to take advantage of it. That’s where line of business should be saying, “We need to get out into this area, or if we don’t, we’re going to be outmoded by our competitors.” And IT should be encouraging it. They should be pushing for more innovation, bringing new ideas, and being a real partner and collaborator at the table within the business and organization. That’s the right way to do this.
IT could use big data analytics to improve its own environment and to answer this crisis of confidence that exists.

Gardner: Otherwise, we run the risk of peeling back an onion only to find other layers unconnected to one another. We don’t get that one view of the customer, that one view of the patient, or that one view of an extended business process. We just create more silos of data. Centralized IT organizations perhaps are better than almost anyone in understanding how to connect those, rather than keep them spinning off on their own.

Whittaker: Absolutely. And IT itself should be applying some of these technologies. In fairness to line of business, there exists a bit of a crisis of confidence in IT, and there’s really no better way to push against that or fight against that then to be able to run analytics on the solutions you’re providing. How well is IT performing? Are you benchmarking against past performance? How do you benchmark against your industry?

That’s another component. Big-data analytics can be utilized by IT not just to deliver capabilities to the organization or push out and help with connecting to the customer. IT could use big data analytics to improve its own environment and to answer this crisis of confidence that exists.

You could turn these tools internally and look at rates of response as compared to your industry, how your network is performing, how your database is performing, or how the code you write is performing. Are your developers efficient in building clean code?

Everybody has been watching the major shift in the healthcare environment in North America. A big component of that probably should have been more benchmark analysis, analytics on code quality, and things of that nature. That’s a great current and topical example of how IT should be utilizing some of these technologies, not just externally, not just bringing it to line of business, but within its own environment, to prove that it’s building systems that are going to be scalable, secure, and stable.

Gardner: What needs to take place in order for this higher level of coordination and collaboration to take place? Are there any key steps that you have in mind for embarking on this?

Four key areas

Whittaker: I think that there are four key areas that need to occur for this collaboration to happen. Number one, senior executives need to be aligned to what the organization is trying to achieve. They need to articulate a common vision that accounts for the shared interest of both IT and line of business and make it clear that they expect collaboration. That should come at the top of the organization.

We need to get out of the smoke-stacked, completely siloed, organizational approaches and get to something that’s far, far more collaborative, and that needs to come from the top. The current approach is not acceptable. These groups need to work together. That’s a key component. If you don’t have buy-in at the top, it makes it really hard for this collaboration to occur.

Number two, IT needs to get its house in order. This means many things, but primarily, it means overcoming the crisis of confidence line of business has in IT by coming to the table with an approach that works for line of business, something that business aligns with such that it feels like it has IT involvement and that they’re buying into the future that the business wants to head towards. IT needs to show that they have a plan that does not compromise the innovations that the business needs.

IT absolutely can no longer just say no. That’s not an acceptable position. Certainly, if you look back, there were IT organizations that were saying, “No, we’re not going to connect to the Internet. It’s not secure. The answer is just going to be no.”

That didn’t work out for them and it’s not going to work out here either. They should be embracing this shift. We shouldn’t perpetuate this cycle by driving more shadow IT and creating ultimately more for IT down the road as inevitable problems start to emerge.
We shouldn’t perpetuate this cycle by driving more shadow IT and creating ultimately more for IT down the road as inevitable problems start to emerge.

Number three, clear the air and put the executive plan in place. Tensions between IT and line of business have gotten to the point where they can’t be ignored any more. Put the stakeholders together in a room, air out the difficulties, and move forward with a clean slate. This is a tremendous opportunity to build a plan that meets both parties’ needs and allows them to start executing on something that’s really going to make a huge impact for the business.

Finally, the fourth point, seek solutions that emphasize collaboration between IT and the business. Many vendors today are encouraging groups to go rogue and operate in silos, and that’s causing a lot of the problem. At Dell, we’re much more about pushing a more collaborative approach. We think IT is terrific, but business has a point. They need innovation and they need IT to step up. And the business needs to embrace IT.

Instead of conflicting with each other and doing your own thing, back up your commitment to collaboration and utilize tools that empower it. That’s where we’re going to win, and that’s how business is going to succeed in the future.

Gardner: Just to be clear John, it sounds as if these aren’t just issues for large enterprises. Mid-size and mid-market organizations, I should think, are in the same issue set or the same ballgame.

Whittaker: You’re absolutely right. If you have an IT department and you have functional business units, you probably have this problem. Certainly, you can benefit from more collaboration, from implementing and instituting an approach to leverage big data and analytics in order to make smarter business decisions is something that everybody is going to need today.

This isn’t something that the G20, the Fortune 500, or Fortune 2000 alone can benefit from. This goes way down in the hierarchy, in the stack, certainly down to the small- and medium-sized business (SMB) level. And maybe even lower. If you’re a data-intensive small business, you probably need to start implementing and taking a look at big data and what analytics based approaches and data-driven decision making opportunities exist within your organization, or you will be outmoded by organizations that do embrace that.

Cloud-based approach

More and more, we’re seeing, particularly in the mid-market, embracing of a cloud-based approach. It's important to point out that that approach is fine and terrific. We love the cloud and we’re big proponents of it, but using a cloud-based solution doesn’t free line of business from the need to collaborate with IT. It will not eliminate this problem.

Yes, you may get a little bit more of stable solution in the cloud, as opposed to managing it yourself, because it's still not core to the functional group. Ultimately, IT is the group that really knows how and where they can apply appropriate constraints, really control those relationships a little better, and ensure that they have solutions in place that allow for you to analyze all the data in the environment.

You need analysis of all the data that exists right now in your cloud implementation, data that exists in all the systems throughout, so you can see all the components and find out where the correlations lie. And if you have siloed data, if you have gone your own way, you’re going to be missing that component.

Gardner: I imagine too that you can’t outsource your alignment of your business strategy with your technology capabilities. In fact, for IT, being able to have more choice with these models for workloads and deployments, frees them up to take more of a role in this alignment.
All businesses want the same thing. They want to find sustainable competitive advantages.

I would think that the role of enterprise architects starts to blend into IT, because they’re no longer looking at the red light-green light issues, keeping the hardware operating. Now, they can really take more of a role and step up to a higher plane on these alignment issues.

Whittaker: Absolutely. We’re seeing terrific IT departments and leadership starting to take a larger role, starting to ultimately become drivers of innovation. That’s really what we want to see. All businesses want the same thing. They want to find sustainable competitive advantages. They want to control spending. They want to reduce risk to the business.

And the most effective and efficient path to achieving all three is getting IT and the business aligned and allowing that collaboration to occur. That’s really at the crux of how businesses are going to gain competitive advantage out of technology in the future.

Gardner: Well, John, I’m afraid we will have to leave it there. We’re about out of time. You’ve been listening to a sponsored BriefingsDirect podcast discussion on how big data and business intelligence trends are hastening a breach or perhaps an alignment between enterprise IT groups and business units. And now is the time to be thinking about which of those directions you will be taking.

You have seen how a more cooperative and innovation fostering relationship, one that can perhaps exploit big data benefits for all is in order and that the stakes are quite high.

And we have learned more about how gaining sustainable competitive advantages using big data strategically among all aspects of enterprises and small businesses should be something made a priority and happen as soon as possible.

Any last words John on recommendations or what you think were some of the more important points that we went over today?

Embrace new technology

Whittaker: Thank you, Dana, this was terrific. And thank you, to the audience, for listening. I would say, again, the big points are, embrace the new technology that’s coming out. The innovation is going to make your business far more successful, and your organization will prosper from these new innovations that will occur.

Number two, do it in a manner that is collaborative between IT and line of business. The CIO, the CMO, the CFO, the CEO, the heads of all of the functional departments, whether you are in sales, marketing, finance, operation, wherever you are, should be aligning with their IT counterparts. It's the combined collaborative approach that’s going to win the day.

And finally, this should really be driven top-down. Senior executives, this is an opportunity to get everybody on the same page to go after and leverage a pretty enormous opportunity before it becomes a huge problem. Let’s get out there right now. We’re still in the early days, but that doesn’t mean there’s not a lot to be gained. And ultimately, in the long-term, we’re going to have more successful organizations able to achieve even greater output through this collaboration and the leveraging of big data analytics.

Gardner: Very good. Thank you to our guest, John Whittaker, Senior Director of Marketing for Dell Software’s Information Management Solutions Group. It was really good talking with you today, John.

Whittaker: Thanks a lot, Dana.

Gardner: And also, a big thank you to our audience for joining this insightful discussion. This is Dana Gardner, Principal Analyst at Interarbor Solutions. Don’t forget to come back next time.

Listen to the podcast. Find it on iTunes. Download the transcript. Sponsor: Dell Software.

Transcript of a Briefings Direct podcast on how advanced analytics aligns business and IT goals to bring further collaboration on innovation within the enterprise. Copyright Interarbor Solutions, LLC, 2005-2013. All rights reserved.

You may also be interested in: