Apache Authors: John Mertic, Pat Romanski, Liz McMillan, Elizabeth White, Janakiram MSV

Related Topics: Java IoT, Industrial IoT, Microservices Expo, IoT User Interface, @CloudExpo, Apache

Java IoT: Blog Feed Post

Deploying APM in the Enterprise | Part 4

The Path of the Rockstar

Deploying APM in the Enterprise. In the last installment we covered how you find, test, and justify purchasing an APM solution. This blog will focus on what to do after you’ve made a purchase and started down the path of deploying your coveted APM tool (ahem, ahem, AppDynamics, ahem). Just clearing my throat, let’s jump right in…

APM RockstarWelcome to Part 4 of my series It’s time for a celebration, time to break out the champagne, time to spike the football and do your end zone dance (easy there Michael Jackson, don’t hurt yourself). All of the hours you spent turning data into meaningful information, dealing with software vendors, writing requirements, testing solutions, documenting your findings, writing business justifications, and generally bending over backwards to ensure that no objection would stand in your way has culminated in management approving your purchase of APM software. Now the real work begins…

The 7 Ps
A co-worker of mine shared some words of wisdom with me a long time ago which have served me well over the years. It’s a little saying called the 7 P’s and goes something like this… Piss Poor Planning Promotes Piss Poor Performance. Deploying and using APM software is not a time for spontaneity or just winging it. If you want to make mistakes and derive little value from the investment you just put your reputation behind then by all mean just jump in with little or no planning. If you want to be a rockstar you need a solid plan for deploying, configuring, verifying, operationalizing, using, and evangelizing your APM tool (ahem, ahem, AppDynamics, ahem). Just clearing my throat again, I think there’s a bug going around ;-P

This blog post is a great general outline for planning your implementation. Everything covered in this post should be part of your planning process and should be considered the bare minimum for APM deployment planning within your organization.

Best Practices
The planning stage is a perfect time to ask your APM vendor for documentation on best practices related to deploying their software. Your vendor (AppDynamics, wink) has seen their software deployed in many situations across many industry verticals. They will have important advice for you on how to make the deployment and operation of their product as successful as possible. Use your vendors depth and breadth of information to your advantage, you’re paying them so it’s the least they can do.

Controller: The Brain, Narf!

The first major decision will be an easy one. You probably already covered this during the evaluation, vendor selection, and negotiation phases but we will recap here. You need to decide if you will host your own controller or use the vendors SaaS environment. In case you don’t already know, a controller is the server component that collects, stores, analyzes, etc… the monitoring data from the agents. Basically the controller is the brains behind the operation. There are many factors that you need to consider when deciding to use a SaaS or On-Premise model and we will not cover them in this post. Your vendor of choice (ahem, ahem, AppDynamics, ahem) will help you decide which option is right for your business circumstances.

Easy peasy lemon squeazy! I have just embedded those words in your head for potentially days, weeks or years to come. Sorry about that but it really describes the SaaS option well. You don’t have to get a server racked, VM allocated, disk space configured, solve a Rubiks Cube in 3 minutes or less, or whatever other convoluted deployment process your company may have in order to host your own software. All you really need to do is point your agents at the SaaS controller and you are off and running. Your chosen APM vendor (AppDynamics of course) will handle the server sizing, capacity, maintenance, etc… for you. Nice!


So you’ve decided to host your own controller(s). We have many clients that choose this route for one reason or another and we make every effort to support you just as well as using the SaaS option. In this case we wont be doing all of the work for you so you need to get cracking on your server deployment process. I hope it’s super easy and streamlined and you can have a new host set up and ready to load software in an hour or less. In reality it may take you a few weeks or even months so you need to be familiar with the lead time so that you can appropriately plan the rest of the deployment. You NEED a controller so there is no point in deploying agents without one. Use this lead time to generate the most awesome plan ever!

Agents: Deploying and Configuring
Agents need applications to do anything meaningful so it’s a requirement that you figure out what applications you want (or will be allowed) to monitor. You most likely had at least one problematic yet important application in mind when you started your search for an APM tool. Create a list of the applications that need monitoring and prioritize that list. I personally prefer creating a top 10 list (you could also call it a “next 10” list) that is an equal mix of application I suspect will be difficult to instrument as well as applications I think will be really easy. I do this because you usually don’t work at deploying agents to application components in a serial manner. It’s typically a parallel process where you can jump from one deployment to the next while you are waiting for approvals, personnel, or anything else that gets in your way of doing actual work.

Deploying APM agents should be easy. Add a very small amount of software to the server you want to monitor, reference the agent software in your application configuration and restart your application. It’s basically that easy to deploy an agent. It should also be really easy to configure. In fact, the agent should automatically detect what it needs to monitor and simply just work. This is how AppDynamics works but the same does not hold true for most other APM vendors. Hopefully you saw this when you ran each vendor solution through your POC environment. In the interest of full disclosure I will admit that there are circumstances where NO APM solution can automatically detect your application properly and there is more configuration work to do. This is a problem that every APM vendor has to deal with but thankfully AppDynamics sees this condition with only a very small subset of its customer base. Usually you plug in our agent and we show you what you need to see. It just works!

Awesome, now that we just saved you 80% of the configuration time versus deploying “the other guys” what’s next?

After you deploy agents (whether it be straight to production or advancing though pre-production environments) and you have used the monitored application a bit, you want to look at the user interface to see if the information contained within looks correct.

  • Look at your application flow map to see if you are missing any application components.
  • Check the business transactions to see if the expected transactions are there and reporting metrics.
  • Do you have end user experience metrics showing up?
  • Do you have transaction snapshots showing your custom code executing in the run time?
  • Send out test alerts to see if they make it to their destination. (Alerting is important so we will cover it in another blog post)

If things don’t look right you need to figure out why. It might be that your application really is different than you thought (we see this quite often), or it could be a problem with the monitoring. Resolve any issues you see before declaring deployment and configuration victory.

Production Load Cannot Be Simulated Exactly!!!
To realize the most value from your APM purchase you MUST run it in production. No matter how good your Quality Engineering team is they cannot code all of the crazy things your users will try to do in production. It can also be very difficult to duplicate your application environment in production. Example, you have 5000 JVMs spread across multiple cloud provider data centers. Replicating that environment would be time consuming and really expensive.

Beyond the technology aspects of running in production you also need to consider your existing processes. Your shiny new APM tool will provide incredible insight into application issues as long as you have it integrated into your processes. Here are some points to consider:

  • Are alerts configured so that they are routed to the proper people?
  • Does the operations center know about the new alerts that will be coming from your new APM product?
  • Is there a process that application owners can follow to request monitoring by your new tool?
  • Is there a process to smoothly and efficiently on-board a new application?
  • Is the APM tool integrated with other corporate systems? (LDAP, Events Aggregations, Business Intelligence, etc…)

What I am trying to say is; Give your company every opportunity to use the hell out of your new tool!

Teach Them Well
Educate and evangelize, this will pay dividends ten fold.

Create a short training curriculum for anyone who will need to work with your APM tool. You should have training material for basic usage, advanced concepts (memory leaks, policies, dashboard creation, etc…), and operations (alerts/events) training. You need to make sure the people who will touch the product or consume the data have the information they need to be successful. Their success drives your success.

Tell everyone you can about the success you are having with your new tool. Don’t be annoying to the point where people run the other way when they see you coming but make sure they know what you are working on and how much of an impact it is having on the business.
For every problem you solve with your new APM tool take 30 minutes to put together a 3–5 slide presentation. Include the following information on each presentation you create:

  • Problem Description: Describe the application, problem, and impact level.
  • Resolution: Describe the resolution steps and root cause. Use screenshots from your APM tool.
  • Business Impact: Describe how long it took to resolve the issue, how long it normally takes without APM, and quantify the impact to the business of this outage for both scenarios (with and without APM).

These short presentations will equip you with the information you need to defend your decision to purchase APM, justify a larger investment, and propel yourself to rockstar status within your organization.

There is a lot of work that needs to be done to successfully deploy, configure and use an APM tool in the enterprise but the potential rewards are staggering. Think about how much lost revenue can be avoided by ensuring your revenue generating applications don’t go down at peak times. People notice when the decisions you make and the work you do directly impact the bottom line. Put in the effort and get noticed!

Join me next week for the next installment in this series. It will be a blog post dedicated to alerts, yes they are that important.

Read the original blog entry...

More Stories By AppDynamics Blog

In high-production environments where release cycles are measured in hours or minutes — not days or weeks — there's little room for mistakes and no room for confusion. Everyone has to understand what's happening, in real time, and have the means to do whatever is necessary to keep applications up and running optimally.

DevOps is a high-stakes world, but done well, it delivers the agility and performance to significantly impact business competitiveness.

@ThingsExpo Stories
The Open Connectivity Foundation (OCF), sponsor of the IoTivity open source project, and AllSeen Alliance, which provides the AllJoyn® open source IoT framework, today announced that the two organizations’ boards have approved a merger under the OCF name and bylaws. This merger will advance interoperability between connected devices from both groups, enabling the full operating potential of IoT and representing a significant step towards a connected ecosystem.
November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Penta Security is a leading vendor for data security solutions, including its encryption solution, D’Amo. By using FPE technology, D’Amo allows for the implementation of encryption technology to sensitive data fields without modification to schema in the database environment. With businesses having their data become increasingly more complicated in their mission-critical applications (such as ERP, CRM, HRM), continued ...
SYS-CON Events announced today that Embotics, the cloud automation company, will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Embotics is the cloud automation company for IT organizations and service providers that need to improve provisioning or enable self-service capabilities. With a relentless focus on delivering a premier user experience and unmatched customer support, Embotics is the fas...
Intelligent machines are here. Robots, self-driving cars, drones, bots and many IoT devices are becoming smarter with Machine Learning. In her session at @ThingsExpo, Sudha Jamthe, CEO of IoTDisruptions.com, will discuss the next wave of business disruption at the junction of IoT and AI, impacting many industries and set to change our lives, work and world as we know it.
SYS-CON Events announced today that Cloudbric, a leading website security provider, will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Cloudbric is an elite full service website protection solution specifically designed for IT novices, entrepreneurs, and small and medium businesses. First launched in 2015, Cloudbric is based on the enterprise level Web Application Firewall by Penta Security Sys...
Smart Cities are here to stay, but for their promise to be delivered, the data they produce must not be put in new siloes. In his session at @ThingsExpo, Mathias Herberts, Co-founder and CTO of Cityzen Data, will deep dive into best practices that will ensure a successful smart city journey.
The explosion of new web/cloud/IoT-based applications and the data they generate are transforming our world right before our eyes. In this rush to adopt these new technologies, organizations are often ignoring fundamental questions concerning who owns the data and failing to ask for permission to conduct invasive surveillance of their customers. Organizations that are not transparent about how their systems gather data telemetry without offering shared data ownership risk product rejection, regu...
Data is the fuel that drives the machine learning algorithmic engines and ultimately provides the business value. In his session at Cloud Expo, Ed Featherston, a director and senior enterprise architect at Collaborative Consulting, will discuss the key considerations around quality, volume, timeliness, and pedigree that must be dealt with in order to properly fuel that engine.
SYS-CON Events announced today that MathFreeOn will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. MathFreeOn is Software as a Service (SaaS) used in Engineering and Math education. Write scripts and solve math problems online. MathFreeOn provides online courses for beginners or amateurs who have difficulties in writing scripts. In accordance with various mathematical topics, there are more tha...
Successful digital transformation requires new organizational competencies and capabilities. Research tells us that the biggest impediment to successful transformation is human; consequently, the biggest enabler is a properly skilled and empowered workforce. In the digital age, new individual and collective competencies are required. In his session at 19th Cloud Expo, Bob Newhouse, CEO and founder of Agilitiv, will draw together recent research and lessons learned from emerging and established ...
In his general session at 19th Cloud Expo, Manish Dixit, VP of Product and Engineering at Dice, will discuss how Dice leverages data insights and tools to help both tech professionals and recruiters better understand how skills relate to each other and which skills are in high demand using interactive visualizations and salary indicator tools to maximize earning potential. Manish Dixit is VP of Product and Engineering at Dice. As the leader of the Product, Engineering and Data Sciences team a...
Virgil consists of an open-source encryption library, which implements Cryptographic Message Syntax (CMS) and Elliptic Curve Integrated Encryption Scheme (ECIES) (including RSA schema), a Key Management API, and a cloud-based Key Management Service (Virgil Keys). The Virgil Keys Service consists of a public key service and a private key escrow service. 

The best way to leverage your Cloud Expo presence as a sponsor and exhibitor is to plan your news announcements around our events. The press covering Cloud Expo and @ThingsExpo will have access to these releases and will amplify your news announcements. More than two dozen Cloud companies either set deals at our shows or have announced their mergers and acquisitions at Cloud Expo. Product announcements during our show provide your company with the most reach through our targeted audiences.
@ThingsExpo has been named the Top 5 Most Influential Internet of Things Brand by Onalytica in the ‘The Internet of Things Landscape 2015: Top 100 Individuals and Brands.' Onalytica analyzed Twitter conversations around the #IoT debate to uncover the most influential brands and individuals driving the conversation. Onalytica captured data from 56,224 users. The PageRank based methodology they use to extract influencers on a particular topic (tweets mentioning #InternetofThings or #IoT in this ...
More and more brands have jumped on the IoT bandwagon. We have an excess of wearables – activity trackers, smartwatches, smart glasses and sneakers, and more that track seemingly endless datapoints. However, most consumers have no idea what “IoT” means. Creating more wearables that track data shouldn't be the aim of brands; delivering meaningful, tangible relevance to their users should be. We're in a period in which the IoT pendulum is still swinging. Initially, it swung toward "smart for smar...
Web Real-Time Communication APIs have quickly revolutionized what browsers are capable of. In addition to video and audio streams, we can now bi-directionally send arbitrary data over WebRTC's PeerConnection Data Channels. With the advent of Progressive Web Apps and new hardware APIs such as WebBluetooh and WebUSB, we can finally enable users to stitch together the Internet of Things directly from their browsers while communicating privately and securely in a decentralized way.
In past @ThingsExpo presentations, Joseph di Paolantonio has explored how various Internet of Things (IoT) and data management and analytics (DMA) solution spaces will come together as sensor analytics ecosystems. This year, in his session at @ThingsExpo, Joseph di Paolantonio from DataArchon, will be adding the numerous Transportation areas, from autonomous vehicles to “Uber for containers.” While IoT data in any one area of Transportation will have a huge impact in that area, combining sensor...
With an estimated 50 billion devices connected to the Internet by 2020, several industries will begin to expand their capabilities for retaining end point data at the edge to better utilize the range of data types and sheer volume of M2M data generated by the Internet of Things. In his session at @ThingsExpo, Don DeLoach, CEO and President of Infobright, discussed the infrastructures businesses will need to implement to handle this explosion of data by providing specific use cases for filterin...
What happens when the different parts of a vehicle become smarter than the vehicle itself? As we move toward the era of smart everything, hundreds of entities in a vehicle that communicate with each other, the vehicle and external systems create a need for identity orchestration so that all entities work as a conglomerate. Much like an orchestra without a conductor, without the ability to secure, control, and connect the link between a vehicle’s head unit, devices, and systems and to manage the ...
Ask someone to architect an Internet of Things (IoT) solution and you are guaranteed to see a reference to the cloud. This would lead you to believe that IoT requires the cloud to exist. However, there are many IoT use cases where the cloud is not feasible or desirable. In his session at @ThingsExpo, Dave McCarthy, Director of Products at Bsquare Corporation, will discuss the strategies that exist to extend intelligence directly to IoT devices and sensors, freeing them from the constraints of ...