Welcome!

Microservices Expo Authors: Carmen Gonzalez, Pat Romanski, Elizabeth White, AppNeta Blog, Liz McMillan

Related Topics: @CloudExpo, Microservices Expo, Containers Expo Blog, Cognitive Computing , Machine Learning , Agile Computing, Cloud Security

@CloudExpo: Article

Beyond Intrusion Detection: Eight Best Practices for Cloud SIEM Deployment

Improving visibility across the enterprise

For all the right reasons, your company has been thinking about deploying SIEM…to create an alert system when those with less than good intentions come knocking; to remediate potential network threats; to comply with federal, state or industry regulations; and identify the risks and vulnerabilities throughout the enterprise IT infrastructure and architecture. If you maintain even a modest (SMB -> Fortune 1000) organization that has any online identity, SIEM should be the cornerstone of your asset protection strategy.

First and foremost, SIEM (and to a certain extent log management) is about visibility. Who is doing what and when on your network. It is as much about understanding the holistic landscape of your infrastructure as it is protecting proprietary assets. Without it, it’s akin to coaching the Big Game without any idea who is the opponent; or for that matter if you even have a starting left guard.

But fun metaphors aside, SIEM is a critical enterprise tool. And just like any enterprise solution, it requires forethought, vigilance and most importantly, a good game plan. And when deployed properly it can change your IT department from infrastructure-based, to information-centric. And as such you get to make better decisions, faster.

And with every technology there are best practices and pitfalls. In past articles I have spoken at length regarding the advantages of deploying and managing SIEM in the cloud. Many of these surround the affordability, manageability, control and capability of the solution. For many, security from the cloud is still an emerging concept. But for those who’ve already made the leap, they are reaping the significant benefits. But I want to move beyond the arguments of “going cloud” when deciding on security solutions. Today I want to focus on what happens next. How do you start collecting that ROI once a cloud-based security-as-a-service has been chosen?

The reason most enterprise deployments fail (on premise or cloud) can be typically traced to two causes: (1.) Lack of buy-in from the executive level or employee resistance to change, but more often the culprit is (2.) lack of vision or process. Too many companies jump in and apply a solution because they heard it was important or were sold a Porsche when all they needed was a family SUV. Of course one of the benefits of cloud-based security is the ability to "buy" the SUV and instantly scale up to that Porsche, if and when, the business need requires it (without touching CapEx budgets!)! But with that here are 8 best practices you should implement when moving forward with your cloud-based security initiative:

Best Practice #1: Identify your goals and match your scope to them. There are five questions you need to ask before moving forward with any deployment. 1. WHY do you need SIEM (compliance? user and/or partner expansion? BYOD? Breach detection?) HOW will SIEM be deployed to properly address these issues (what processes, functionality and capabilities are needed; which needs to be outsourced/replaced/improved) WHAT needs to be collected, analyzed and reported? HOW BIG does the deployment need to scale to accurately and cost effectively meet your specific business need? And WHERE is the information situated that should/must be monitored?

Best practice #2: Incremental usage. The quickest route to success is taking baby steps. The idea is to prove the concept and then expand the scope. To some this might be to start with log management and add  SIEM once you understand the requirements, commitment and volume. Now because security-as-a-service is so flexible and can ramp up or down instantly, an easy entry point might be to start with only those elements that fulfill compliance. The project might be overwhelming, but if you take it in bite-sized phases, you will find the victories come easier and the ROI is justified. When dealing with a cloud security deployment, it is easy to turn on the fire hose when only a garden hose is needed. But the beauty of a cloud deployment is the ease and flexibility of scaling. Again, another example of incremental usage would be either to apply SIEM against specific use case scenarios or possibly just migrate a division or a department or a function (as opposed to the entire enterprise).

Best Practice #3: Determine what IS and ISN’T a threat to your network. Returning to the fire hose metaphor, when deploying a SIEM initiative, it is very easy to get lost in a sea of data. It can be like trying to drink from that proverbial fire hose. The trick is to recognize what constitutes a true risk and eliminate false positives. And this requires some internal analysis to create a series of rules that sift out the white noise and differentiate “normal” traffic from suspicious activity. For instance, if there is an attempted access to your partner portal from Russia—is that normal? Do you even have a partner in Minsk? But even a simple filter isn’t quite enough. Risk is three dimensional and it can hide in plain sight. That’s why you continue to filter based on time of day, IP address, server, attempts, network availability and a myriad of other forensic qualifiers before the alert is grave enough to require immediate attention.

Best practice #4: Map response plans. Now that an incident gets your attention, what do you do? Do you launch an account investigation, suspend the user, deactivate a password, apply a  denial-of-service against the IP or a number of remediations based on the severity, vulnerability and identity of the transgressor. This goes back to workflow and process. Who is going to what to whom and how? SIEM is a process-reliant technology. You simply can’t flip a switch and say you’ve put up a magic forcefield around your network. Your response plan is your blueprint to closing the vulnerability gaps and ensuring compliance.

Best practice #5 Correlate data from multiple sources. The practice of situational awareness is what adds the muscle into a SIEM initiative. Like #4, it isn’t enough to plug in a solution and press “go.” Situational awareness takes into account a multitude of different endpoints, servers, data streams, assets and inventories, events and flows, from across the enterprise and puts information into context.  Context is the most important portion of risk assessment.  For example, a shark is a threat. However if that shark is 10 miles away, it is not a direct or immediate threat. Doesn't mean you're not vulnerable if that shark gets hungry. Having an engine that not only creates accurate perspective, but analyzes, understands and acts upon behaviors is key. And to do that a centralized SIEM engine needs the data from more than just a single source or single server.

Best Practice #6: Requires Real time monitoring 7/24/365. For many companies this is a challenge, but hackers don’t sleep. And although a great deal of SIEM and Log Management is automated, it still requires the vigilance of 24 hour monitoring. Trees might be falling in the forest, but if there is no one to see them, breaches occur, networks are compromised. I’ve witnessed plenty of IT departments that don’t have the resources. Again, this is a considerable advantage that security-as-s-service provides and allows you to sleep just a little better at night. Knowing that this one crucial element of your security is professionally addressed without additional staff or budget makes the cloud that much more valuable.

Best Practice #7 Remain calm! One thing we’ve noticed is that soon after the deployment of a SIEM/Log Management it seems there are alerts and issues you never dreamed about. Things are bound to look worse before they get better and it can seem overwhelming; kind of opening a Pandora’s Box of malware and botnets.  For the most part it is because you now know what you didn’t know before.  In some respect it is like looking at your hotel room comforter under black light and a microscope. But once you realize what you’re looking at and that much or the remediation can be automated, soon, (with a bit of fine tuning and normalizing correlation feeds) you will be measure that the anomalous events lessen and the alert prioritizations allow you to make timely and intelligent decisions.

Best practice #8: Evolution. Security is a moving target. You need to revisit you processes and workflows every few months to make sure you are up to date with compliance requirements, new users/access points and expanded or redefined workflows. This is more than recognizing the latest virus threats. New users access your network with regularity. New layers of regulations are added. There are new applications requiring monitoring. All in all, by giving your cloud-based SIEM and log management solutions the new and necessary data, your enterprise will be more secure than it was yesterday.

More Stories By Kevin Nikkhoo

With more than 32 years of experience in information technology, and an extensive and successful entrepreneurial background, Kevin Nikkhoo is the CEO of the dynamic security-as-a-service startup Cloud Access. CloudAccess is at the forefront of the latest evolution of IT asset protection--the cloud.

Kevin holds a Bachelor of Science in Computer Engineering from McGill University, Master of Computer Engineering at California State University, Los Angeles, and an MBA from the University of Southern California with emphasis in entrepreneurial studies.

@MicroservicesExpo Stories
Everyone wants to use containers, but monitoring containers is hard. New ephemeral architecture introduces new challenges in how monitoring tools need to monitor and visualize containers, so your team can make sense of everything. In his session at @DevOpsSummit, David Gildeh, co-founder and CEO of Outlyer, will go through the challenges and show there is light at the end of the tunnel if you use the right tools and understand what you need to be monitoring to successfully use containers in your...
With major technology companies and startups seriously embracing Cloud strategies, now is the perfect time to attend @CloudExpo | @ThingsExpo, June 6-8, 2017, at the Javits Center in New York City, NY and October 31 - November 2, 2017, Santa Clara Convention Center, CA. Learn what is going on, contribute to the discussions, and ensure that your enterprise is on the right path to Digital Transformation.
What if you could build a web application that could support true web-scale traffic without having to ever provision or manage a single server? Sounds magical, and it is! In his session at 20th Cloud Expo, Chris Munns, Senior Developer Advocate for Serverless Applications at Amazon Web Services, will show how to build a serverless website that scales automatically using services like AWS Lambda, Amazon API Gateway, and Amazon S3. We will review several frameworks that can help you build serverle...
The IT industry is undergoing a significant evolution to keep up with cloud application demand. We see this happening as a mindset shift, from traditional IT teams to more well-rounded, cloud-focused job roles. The IT industry has become so cloud-minded that Gartner predicts that by 2020, this cloud shift will impact more than $1 trillion of global IT spending. This shift, however, has left some IT professionals feeling a little anxious about what lies ahead. The good news is that cloud computin...
Lots of cloud technology predictions and analysis are still dealing with future spending and planning, but there are plenty of real-world cloud use cases and implementations happening now. One approach, taken by stalwart GE, is to use SaaS applications for non-differentiated uses. For them, that means moving functions like HR, finance, taxes and scheduling to SaaS, while spending their software development time and resources on the core apps that make GE better, such as inventory, planning and s...
After more than five years of DevOps, definitions are evolving, boundaries are expanding, ‘unicorns’ are no longer rare, enterprises are on board, and pundits are moving on. Can we now look at an evolution of DevOps? Should we? Is the foundation of DevOps ‘done’, or is there still too much left to do? What is mature, and what is still missing? What does the next 5 years of DevOps look like? In this Power Panel at DevOps Summit, moderated by DevOps Summit Conference Chair Andi Mann, panelists l...
Without a clear strategy for cost control and an architecture designed with cloud services in mind, costs and operational performance can quickly get out of control. To avoid multiple architectural redesigns requires extensive thought and planning. Boundary (now part of BMC) launched a new public-facing multi-tenant high resolution monitoring service on Amazon AWS two years ago, facing challenges and learning best practices in the early days of the new service.
The rise of containers and microservices has skyrocketed the rate at which new applications are moved into production environments today. While developers have been deploying containers to speed up the development processes for some time, there still remain challenges with running microservices efficiently. Most existing IT monitoring tools don’t actually maintain visibility into the containers that make up microservices. As those container applications move into production, some IT operations t...
DevOps is often described as a combination of technology and culture. Without both, DevOps isn't complete. However, applying the culture to outdated technology is a recipe for disaster; as response times grow and connections between teams are delayed by technology, the culture will die. A Nutanix Enterprise Cloud has many benefits that provide the needed base for a true DevOps paradigm.
DevOps tends to focus on the relationship between Dev and Ops, putting an emphasis on the ops and application infrastructure. But that’s changing with microservices architectures. In her session at DevOps Summit, Lori MacVittie, Evangelist for F5 Networks, will focus on how microservices are changing the underlying architectures needed to scale, secure and deliver applications based on highly distributed (micro) services and why that means an expansion into “the network” for DevOps.
DevOps is often described as a combination of technology and culture. Without both, DevOps isn't complete. However, applying the culture to outdated technology is a recipe for disaster; as response times grow and connections between teams are delayed by technology, the culture will die. A Nutanix Enterprise Cloud has many benefits that provide the needed base for a true DevOps paradigm. In his Day 3 Keynote at 20th Cloud Expo, Chris Brown, a Solutions Marketing Manager at Nutanix, will explore t...
As Enterprise business moves from Monoliths to Microservices, adoption and successful implementations of Microservices become more evident. The goal of Microservices is to improve software delivery speed and increase system safety as scale increases. Documenting hurdles and problems for the use of Microservices will help consultants, architects and specialists to avoid repeating the same mistakes and learn how and when to use (or not use) Microservices at the enterprise level. The circumstance w...
SYS-CON Events announced today that Auditwerx will exhibit at SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. Auditwerx specializes in SOC 1, SOC 2, and SOC 3 attestation services throughout the U.S. and Canada. As a division of Carr, Riggs & Ingram (CRI), one of the top 20 largest CPA firms nationally, you can expect the resources, skills, and experience of a much larger firm combined with the accessibility and attent...
SYS-CON Events announced today that HTBase will exhibit at SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. HTBase (Gartner 2016 Cool Vendor) delivers a Composable IT infrastructure solution architected for agility and increased efficiency. It turns compute, storage, and fabric into fluid pools of resources that are easily composed and re-composed to meet each application’s needs. With HTBase, companies can quickly prov...
Buzzword alert: Microservices and IoT at a DevOps conference? What could possibly go wrong? In this Power Panel at DevOps Summit, moderated by Jason Bloomberg, the leading expert on architecting agility for the enterprise and president of Intellyx, panelists peeled away the buzz and discuss the important architectural principles behind implementing IoT solutions for the enterprise. As remote IoT devices and sensors become increasingly intelligent, they become part of our distributed cloud enviro...
@DevOpsSummit has been named the ‘Top DevOps Influencer' by iTrend. iTrend processes millions of conversations, tweets, interactions, news articles, press releases, blog posts - and extract meaning form them and analyzes mobile and desktop software platforms used to communicate, various metadata (such as geo location), and automation tools. In overall placement, @DevOpsSummit ranked as the number one ‘DevOps Influencer' followed by @CloudExpo at third, and @MicroservicesE at 24th.
By now, every company in the world is on the lookout for the digital disruption that will threaten their existence. In study after study, executives believe that technology has either already disrupted their industry, is in the process of disrupting it or will disrupt it in the near future. As a result, every organization is taking steps to prepare for or mitigate unforeseen disruptions. Yet in almost every industry, the disruption trend continues unabated.
In his General Session at 16th Cloud Expo, David Shacochis, host of The Hybrid IT Files podcast and Vice President at CenturyLink, investigated three key trends of the “gigabit economy" though the story of a Fortune 500 communications company in transformation. Narrating how multi-modal hybrid IT, service automation, and agile delivery all intersect, he will cover the role of storytelling and empathy in achieving strategic alignment between the enterprise and its information technology.
While DevOps most critically and famously fosters collaboration, communication, and integration through cultural change, culture is more of an output than an input. In order to actively drive cultural evolution, organizations must make substantial organizational and process changes, and adopt new technologies, to encourage a DevOps culture. Moderated by Andi Mann, panelists discussed how to balance these three pillars of DevOps, where to focus attention (and resources), where organizations might...
Microservices are a very exciting architectural approach that many organizations are looking to as a way to accelerate innovation. Microservices promise to allow teams to move away from monolithic "ball of mud" systems, but the reality is that, in the vast majority of organizations, different projects and technologies will continue to be developed at different speeds. How to handle the dependencies between these disparate systems with different iteration cycles? Consider the "canoncial problem" ...