|By AppDynamics Blog||
|November 2, 2012 03:03 PM EDT||
Welcome back to my series on Deploying APM in the Enterprise. In Part 2 we discussed APM maturity from an unusual perspective as maturity models go. In this blog we are going to discuss the process of getting what you want and need in the way of an APM product.
Warning – This is an astronomically long blog post so make sure you have plenty of time to read it through.
Let’s assume for a minute that you actually read Part 2 – APM Maturity As You’ve Probably Never Seen It Before and have an idea where you and your organization currently are on the maturity scale. To advance to a higher level of maturity you can work on process and people until the cows come home but eventually you are going to want a tool that provides insight your people and processes never will. So that is where I am going to focus the rest of this series while touching on the people and process aspect throughout.
If you have decided you need or want to buy a new software tool here is my tried and proven method for making it happen. Just to make things a little more interesting each step in the process is a song title that relates somehow and that I actually like. So feel free to comment on the blog post and my taste in music if you feel like it.
Throughout this blog post I will share some key lessons learned from my own experience dealing with software vendors. This information will be highlighted in bold red text so it is easier to pick out.
It Hurts So Good – John Cougar
Nobody will agree to spend money on a tool unless there is some problem putting the hurt on your business (lost revenue, productivity impact, customer satisfaction, etc…). Find and document a tangible problem. Preferably an issue with a business/mission critical application like your e-commerce platform, online trading, payment gateway, risk calculation, settlement system, etc… Find some application or service that is impacting your business in a meaningful way due to poor performance and/or downtime and document the following:
- Number of issues and severity level
- Mean Time To Repair (MTTR – usually the average amount of time from first impact to problem resolution)
- Quantifiable measure of impact on business (dollars lost per minute, potential customers lost, trades lost per minute, etc…)
- Average number of employees involved in troubleshooting each issue
- Root cause of each incident
You will use this data in your evaluation document and your business justification down the road.
The World I Know – Collective Soul
“So I walk up on high / And I step to the edge / To see my world below” If you haven’t already done it, you need to take inventory of what you already own and document your findings. You will use this information for years to come as long as you keep it up to date.
- What tools exist and what category should they be put in? (Database Monitoring, Network Monitoring, OS Monitoring, Desktop Monitoring, etc…)
- How many licenses do we have and are the current?
- What are they good at?
- What are they not good at?
- What would be classified as an APM tool?
- If I already have an APM tool why is it not being used properly?
- Put labels on your existing tools and understand what they do!
Down in a Hole – Alice In Chains
Now that you have the overall landscape of your monitoring ecosystem laid out you need to see if there are any gaping holes. AppDynamics is an APM company so we suggest you compare your existing tools to the Gartner definition of APM to see what you might be missing. Here it is paraphrased in my own words:
- End User Experience Monitoring: Measuring the response time of your application all the way to the end user. It’s not good enough to just understand how fast your application runs within the confines of the data center(s).
- Application Topology Mapping: Automatic detection and display of all components involved in the delivery of your application. You need to know what application components are in use at any given time, but especially when there is an issue impacting your users.
- Business Transaction Profiling: Detecting and measuring the response time of all application component activity initiated by a single user request. This is not the same as measuring the response time of a web page!!!
- Deep Application Diagnostics: Detecting and measuring the run time code execution within your application containers. If your current or prospective solution does not load into the application container you will NOT have this important capability.
- Analytics: Intelligence applied to data which provides you with actionable information. This is not the same as reporting and analytics can (and should) be a key differentiator between competing solutions.
You can do this same type of gap analysis for other types of monitoring but you will need to figure out the main aspects of each type for yourself (sorry).
So What’cha Want – The Beastie Boys
So lets assume that you need an APM solution to fill that large void in your monitoring capabilities and to solve that pesky problem which has been lingering for the past 6 months. You know which problem I’m referring to, it’s the one that crops up for a few hours every other week, it wreaks havoc on your business, then it magically disappears after half of your IT staff is on a massive conference call trying to figure out what’s going on. It’s the problem that your boss has to answer for in the daily update calls and can never explain. It’s the problem that will eventually get someone fired or “re-organized” and if you can fix it you will be a hero or a rockstar (you get to choose which term you prefer since you fixed the problem).
In order to pick the right APM solution (picking the wrong one can turn that rockstar potential completely upside down) you need to develop a method for comparing different solutions. First you need to narrow down the crowded list of vendors to a group of 2 or 3 at most to do a Proof Of Concept (POC). This initial narrowing of the field is usually done through feature comparisons (vendor websites), references from people you know, phone calls with vendors, phone calls with analysts (Gartner, Forrester, etc…), crystal balls, voodoo rituals, and possibly animal sacrifice.
Use your favorite spreadsheet program (I always wonder how many people don’t actually use Excel at work) and create a matrix with all of the products you might be interested in (AppDynamics should be first on the list, wink, wink) and compare each product to the level of support they have for each of your requirements. Your requirements at this point should be pretty generic as the detailed requirements are best suited for the POC phase. Here are some requirements to get you started:
- Automatic detection, naming, and monitoring of Business Transactions
- Automatic discovery and deep instrumentation of application code
- End user experience monitoring
- Analytics based alerting
- Automatic discovery and display of my application topology
- Support for my application technologies
- Support for my application architecture (cloud, monolithic, distributed, etc…)
- Open-ness of vendor (did they skirt around my questions?, did they talk in circles?)
- Many more requirements you should add…
One you have this matrix built it should be easy to narrow your choices (hopefully AppDynamics made the short list, wink, wink).
One piece of advice before moving along to the POC requirements section… Be very explicit and detailed with your questions as they apply to your environment. If you get a partial answer from a vendor or a response that is not quite what you are looking for make sure you dig deeper right away to figure out if they are blowing smoke or not. If they keep talking in circle you can be pretty sure they can’t really do what you are asking.
Little Lies – Fleetwood Mac
No, this is not the point where I tell you to start making stuff up so you can get your software purchased. This is the part where I warn you that vendors can get very “creative” with their marketing. Some vendors are worse about this than others but it can be very difficulty to wade through the enormous amounts of BS that are splattered across many vendors websites.
Remember this key principle when dealing with all vendors … Don’t trust anything a vendor tells you, make them show you in a live demo!!!
I’m not saying that vendors are evil liars, but they will give you their best answer to your questions and it might not align with the intent of your question. This has happened to me many times before and I learned the hard way that although a vendors answer was technically true, the answer was far from reality when compared with the intent of my question.
And now back to our regularly scheduled broadcast.
3 is the Magic Number – De La Soul
The POC demands it own set of much more detailed requirements. This is the time to really look under the covers at a product and see what it can do versus what you have already been told. Again, using your favorite spreadsheet tool create a new sheet for each tool in the POC. I prefer to use a weighting and grading scale to help differentiate between tools.
- Weighting: Some requirements will be more important to you than others. I use a Low, Medium, and High scale with associated values of 1, 2, and 3. This is your multiplier for the grade you give each requirement
- Grading: Some tools have better support for a requirement than others. I use a grading scale of None, Poor, Adequate, and Excellent with associated values of 0, 1, 2, and 3.
I also like to break up the POC requirements into categories. This helps keep me organized, ensures I don’t miss testing something on my list, and helps me write the analysis document at the end of the POC. Some examples of categories and requirements are shown below.
- Authentication and Authorization
- Supports Microsoft AD
- Granular Role Based Permission System
- User Interface
- Web based user interface compatible with Internet Explorer 8+
- Displays application topology without administrator or user configuration
- Deep Application Diagnostics
- Automatically discovers and instruments custom code
- Automatically traces complete call stack when performance is abnormally poor
- Automated intelligence to ensure instrumentation does not use excessive overhead
Your list should be much longer and more detailed. This requirements list is the basis of the POC and all follow up documentation so make sure it is thorough and not slanted towards a particular vendor. Think of it as your Christmas list when you were a kid, don’t be afraid to ask for things that might seem impossible but that could be really useful.
Another key lesson learned coming your way… Don’t let the vendor control the POC. You define the environment (Dev, Test, and even Prod if it can be done safely), you do the installation, you do the configuration, you do everything related to the POC. You are the one who has to use the tool after you buy it so be sure to personally do everything during the POC.
Be sure you provide the same playing field for each vendor so that your results are really comparable. Having a successful POC using a team of 5 vendor engineers versus another successful POC using just 1 engineer is comparing apples and oranges.
We Are the Champions – Queen
After you wrap up all of your POCs you should have enough data to pick a winner. Hopefully you built your spreadsheets so that they automatically add up the numbers related to all of the requirements. You will have a statistical winner based purely upon you spreadsheet data and that usually aligns with the overall feeling you have after the POC is complete but sometimes it might not. You may have a vendor/product that is statistically the winner but your gut is telling you to choose a different vendor. In this case you need to figure out exactly why you feel this way. You cannot justify vendor selection based upon a gut instinct with no explanation. Maybe one vendor was just a royal pain to deal with, maybe the solution worked well but was agonizingly painful to deploy. The key here is being able to express why your instinct is pushing you a certain direction and quantify the potential impact. Something similar to the following statement can be used as justification of your position:
“Even though Vendor X Product statistically scored highest in the evaluation there is one overriding factor that prohibits selecting them as the overall winner. Deployment and configuration of Vendor X Product is difficult and time consuming. Based upon the observed deployment and configuration time of 2 weeks for 1 application during the POC it would take approximately 19 years to configure monitoring for the 500 applications which are in scope. Deploying Vendor X Product does not make sense in our environment.”
The information you have been building throughout this entire process should be used to create an overall evaluation document. This document should have the following information at a minimum:
- Description of problem
- Description of proposed solution
- Vendors/Products evaluated
- Evaluation Criterial (Requirements)
- Evaluation Results
- Next steps
It can also be helpful to create a short (3–10 pages) presentation to accompany the evaluation document which you can use to brief management on your findings. This presentation should contain only the most important facts since you have all the detail anyone should ever want in the full evaluation document.
Another important document you can create is the business justification. The business justification cuts out all the technical details related to the product you want to purchase and gets right down to the economics of the matter. I am not going to dive deep into how to write a business justification but to help you get started you should make sure you get a ROI (Return On Investment) calculator from each vendor that participates in your POC. Vendors want to help you buy their solution and have a wealth of information available to help you build your business justification so just ask them for help if you need it. Just make sure that everything in the business justification is factually based and relevant to your business.
Lastly it is really helpful to have broad support for your initiative. Seek out people across your organization that will support your recommendation of product or who will validate the problem that you are trying to solve. If you have full support of an Application Owner whose business is being directly impacted this greatly increases your chances of success.
If you read this entire blog post please accept my gratitude! I know it was a long post but it’s a huge topic to cover. Hopefully you picked up some good information and will return for my next post about deploying the product you just worked so hard to get in the door. As always, your comments are welcome. I’d love to hear your tips for getting new solutions approved or any war stories from your vendor evaluations.
Actian Corporation has announced the latest version of the Actian Vector in Hadoop (VectorH) database, generally available at the end of July. VectorH is based on the same query engine that powers Actian Vector, which recently doubled the TPC-H benchmark record for non-clustered systems at the 3000GB scale factor (see tpc.org/3323). The ability to easily ingest information from different data sources and rapidly develop queries to make better business decisions is becoming increasingly importan...
Jul. 29, 2016 06:15 PM EDT Reads: 946
SYS-CON Events announced today that 910Telecom will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Housed in the classic Denver Gas & Electric Building, 910 15th St., 910Telecom is a carrier-neutral telecom hotel located in the heart of Denver. Adjacent to CenturyLink, AT&T, and Denver Main, 910Telecom offers connectivity to all major carriers, Internet service providers, Internet backbones and ...
Jul. 29, 2016 06:00 PM EDT Reads: 910
[guide] Cloud Analytics for Dummies | @CloudExpo @Informatica #API #Cloud #Analytics #BusinessIntelligence
Cloud analytics is dramatically altering business intelligence. Some businesses will capitalize on these promising new technologies and gain key insights that’ll help them gain competitive advantage. And others won’t. Whether you’re a business leader, an IT manager, or an analyst, we want to help you and the people you need to influence with a free copy of “Cloud Analytics for Dummies,” the essential guide to this explosive new space for business intelligence.
Jul. 29, 2016 05:15 PM EDT Reads: 974
You think you know what’s in your data. But do you? Most organizations are now aware of the business intelligence represented by their data. Data science stands to take this to a level you never thought of – literally. The techniques of data science, when used with the capabilities of Big Data technologies, can make connections you had not yet imagined, helping you discover new insights and ask new questions of your data. In his session at @ThingsExpo, Sarbjit Sarkaria, data science team lead ...
Jul. 29, 2016 04:45 PM EDT Reads: 1,239
Qosmos has announced new milestones in the detection of encrypted traffic and in protocol signature coverage. Qosmos latest software can accurately classify traffic encrypted with SSL/TLS (e.g., Google, Facebook, WhatsApp), P2P traffic (e.g., BitTorrent, MuTorrent, Vuze), and Skype, while preserving the privacy of communication content. These new classification techniques mean that traffic optimization, policy enforcement, and user experience are largely unaffected by encryption. In respect wit...
Jul. 29, 2016 04:15 PM EDT Reads: 419
ReadyTalk has expanded the capabilities of the FoxDen collaboration platform announced late last year to include FoxDen Connect, an in-room video collaboration experience that launches with a single touch. With FoxDen Connect, users can now not only engage in HD video conferencing between iOS and Android mobile devices or Chrome browsers, but also set up in-person meeting rooms for video interactions. A host’s mobile device automatically recognizes the presence of a meeting room via beacon tech...
Jul. 29, 2016 04:15 PM EDT Reads: 382
Manufacturers are embracing the Industrial Internet the same way consumers are leveraging Fitbits – to improve overall health and wellness. Both can provide consistent measurement, visibility, and suggest performance improvements customized to help reach goals. Fitbit users can view real-time data and make adjustments to increase their activity. In his session at @ThingsExpo, Mark Bernardo Professional Services Leader, Americas, at GE Digital, discussed how leveraging the Industrial Internet a...
Jul. 29, 2016 03:15 PM EDT Reads: 649
On Dice.com, the number of job postings asking for skill in Amazon Web Services increased 76 percent between June 2015 and June 2016. Salesforce.com saw its own skill mentions increase 37 percent, while DevOps and Cloud rose 35 percent and 28 percent, respectively. Even as they expand their presence in the cloud, companies are also looking for tech professionals who can manage projects, crunch data, and figure out how to make systems run more autonomously. Mentions of ‘data science’ as a skill ...
Jul. 29, 2016 03:00 PM EDT Reads: 440
In his session at Cloud Expo, Robert Cohen, an economist and senior fellow at the Economic Strategy Institute, will provide economic scenarios that describe how the rapid adoption of software-defined everything including cloud services, SDDC and open networking will change GDP, industry growth, productivity and jobs. This session will also include a drill down for several industries such as finance, social media, cloud service providers and pharmaceuticals.
Jul. 29, 2016 02:15 PM EDT Reads: 374
IoT generates lots of temporal data. But how do you unlock its value? You need to discover patterns that are repeatable in vast quantities of data, understand their meaning, and implement scalable monitoring across multiple data streams in order to monetize the discoveries and insights. Motif discovery and deep learning platforms are emerging to visualize sensor data, to search for patterns and to build application that can monitor real time streams efficiently. In his session at @ThingsExpo, ...
Jul. 29, 2016 02:00 PM EDT Reads: 1,203
Security, data privacy, reliability and regulatory compliance are critical factors when evaluating whether to move business applications from in-house client hosted environments to a cloud platform. In her session at 18th Cloud Expo, Vandana Viswanathan, Associate Director at Cognizant, In this session, will provide an orientation to the five stages required to implement a cloud hosted solution validation strategy.
Jul. 29, 2016 01:00 PM EDT Reads: 1,195
"delaPlex is a software development company. We do team-based outsourcing development," explained Mark Rivers, COO and Co-founder of delaPlex Software, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 29, 2016 01:00 PM EDT Reads: 2,113
Ixia (Nasdaq: XXIA) has announced that NoviFlow Inc.has deployed IxNetwork® to validate the company’s designs and accelerate the delivery of its proven, reliable products. Based in Montréal, NoviFlow Inc. supports network carriers, hyperscale data center operators, and enterprises seeking greater network control and flexibility, network scalability, and the capacity to handle extremely large numbers of flows, while maintaining maximum network performance. To meet these requirements, NoviFlow in...
Jul. 29, 2016 01:00 PM EDT Reads: 767
Choosing the right cloud for your workloads is a balancing act that can cost your organization time, money and aggravation - unless you get it right the first time. Economics, speed, performance, accessibility, administrative needs and security all play a vital role in dictating your approach to the cloud. Without knowing the right questions to ask, you could wind up paying for capacity you'll never need or underestimating the resources required to run your applications.
Jul. 29, 2016 01:00 PM EDT Reads: 716
Ovum, a leading technology analyst firm, has published an in-depth report, Ovum Decision Matrix: Selecting a DevOps Release Management Solution, 2016–17. The report focuses on the automation aspects of DevOps, Release Management and compares solutions from the leading vendors.
Jul. 29, 2016 12:00 PM EDT Reads: 1,835
Continuous testing helps bridge the gap between developing quickly and maintaining high quality products. But to implement continuous testing, CTOs must take a strategic approach to building a testing infrastructure and toolset that empowers their team to move fast. Download our guide to laying the groundwork for a scalable continuous testing strategy.
Jul. 29, 2016 12:00 PM EDT Reads: 2,102
Adding public cloud resources to an existing application can be a daunting process. The tools that you currently use to manage the software and hardware outside the cloud aren’t always the best tools to efficiently grow into the cloud. All of the major configuration management tools have cloud orchestration plugins that can be leveraged, but there are also cloud-native tools that can dramatically improve the efficiency of managing your application lifecycle. In his session at 18th Cloud Expo, ...
Jul. 29, 2016 11:30 AM EDT Reads: 1,278
SYS-CON Events announced today that LeaseWeb USA, a cloud Infrastructure-as-a-Service (IaaS) provider, will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. LeaseWeb is one of the world's largest hosting brands. The company helps customers define, develop and deploy IT infrastructure tailored to their exact business needs, by combining various kinds cloud solutions.
Jul. 29, 2016 11:15 AM EDT Reads: 1,344
Amazon has gradually rolled out parts of its IoT offerings in the last year, but these are just the tip of the iceberg. In addition to optimizing their back-end AWS offerings, Amazon is laying the ground work to be a major force in IoT – especially in the connected home and office. Amazon is extending its reach by building on its dominant Cloud IoT platform, its Dash Button strategy, recently announced Replenishment Services, the Echo/Alexa voice recognition control platform, the 6-7 strategic...
Jul. 29, 2016 11:15 AM EDT Reads: 687
[webcast] Mastering DevOps Automation | @DevOpsSummit @IBMDevOps #IBM #Cloud #DevOps #ContinuousDelivery
Deploying applications in hybrid cloud environments is hard work. Your team spends most of the time maintaining your infrastructure, configuring dev/test and production environments, and deploying applications across environments – which can be both time consuming and error prone. But what if you could automate provisioning and deployment to deliver error free environments faster? What could you do with your free time?
Jul. 29, 2016 11:00 AM EDT Reads: 449