This article covers Event and Log Management (ELM) best practices that can decrease the potential for security breaches. To learn more about our ELM solution, contact our team of experts.Contact us
Has someone made any unauthorized changes to your Active Directory policies or Access Control Lists (ACLs) for a directory on a server containing company Intellectual Property? Has someone gained unauthorized access to data that is regulated by law, such as HIPAA? Is somebody trying to hack into your internal systems? What if your compliance officer asks you for SOX-centric reports?
Every day, computer networks across the globe are generating records of the events that occur. Some are routine. Others are indicators of a decline in network health or attempted security breaches. Log files contain a wealth of information to reduce an organization’s exposure to intruders, malware, damage, loss and legal liabilities. Log data needs to be collected, stored, analyzed and monitored to meet and report on regulatory compliance standards like Sarbanes Oxley, Basel II, HIPAA, GLB, FISMA, PCI DSS, NISPOM. This is a daunting task since log files come from many different sources, in different formats, and in massive volumes, and many organizations don’t have a proper log management strategy in place to monitor and secure their network.
In response this article will discuss common Event and Log Management (ELM) requirements and best practices to decrease the potential for security breaches and reduce the possibility of legal or compliance issues.
Every system in your network generates some type of log file. In fact, a log entry is created for each event or transaction that takes place on any machine or piece of hardware–think of it as acting as your “journal of record”. Microsoft-based systems generate Windows Event Log files, and UNIX-based servers and networking devices use the System Log or Syslog standard. Web Application servers like Apache or IIS, as well as Load Balancers, Firewalls, Proxy Servers, or Content Security appliances generate W3C/IIS log files.
Centralized Log Management should be a key component of your compliance initiatives, because with centralized logs in place, you can monitor, audit, and report on file access, unauthorized activity by users, policy changes, and other critical activities performed against files or folders containing proprietary or regulated personal data such as employee, patient or financial records. A centralized log management strategy should include overseeing Event Logs, Syslog and W3C logs. And this is key because information breaches come equally from internal and external sources. For example, Windows Event Logs will give you visibility into potential harmful activities conducted by disgruntled employees, while Syslog management will give you control over your network perimeter.
Windows-based systems have several different event logs that should be monitored consistently. Of these logs, the most important is the Security Log. It provides key information about who is on logged onto the network and what they are doing. Security logs are important to security personnel to understand if vulnerability exists in the security implementation.
Syslog is a log message format and log transmission protocol defined as a standard by the Internet Engineering Task Force (IETF) in RFC-3164 with draft improvements in RFC-5424. Networking devices, UNIX and Linux systems, and many software and hardware platforms, implement Syslog as a standard logging format and means to transmit and collect those log files in a centralized log management repository. Using Syslog information, you can capture highly detailed information about the status of a device or a number of devices. The information can be sorted and parsed to see atypical behavior through changes in operational or performance patterns. These changes may indicate a single or multiple problems. Storage of Syslog log data can also support compliance efforts by providing audit logs to trace any event that may affect network reliability and protection of data. This is important as it proves control of all information to auditors.
Similarly, W3C logs also provide information on user and server activity. These audit logs should too be monitored as they provide valuable information that you can use to identify any unauthorized attempts to compromise, for example, your Web server. IIS log files are a fixed (meaning that it cannot be customized) ASCII format, which record more information than other log file formats, including basic items, such as the IP address of the user, user name, request date and time, service status code, and number of bytes received. In addition, the IIS log file format includes detailed items, such as the elapsed time, number of bytes sent, action, and target file.
By deploying a centralized log management solution, you can easily manage the frequently overwhelming amount of log information generated by your systems. Real-time access to log data will allow you to filter and locate that one “needle in a haystack” event that could be the cause of a security breach.
Every day your servers and other network devices produce 10’s of thousands of log entries. In fact, the average enterprise will accumulate up to 4GB of log data a day. Over 95% of the data within the log files consists of detailed entries recording every successful event or transaction taking place on the system. For example; a server crash, user logins, start and stop of applications, and file access.
Many administrators are surprised to learn that “simple” log files can result in such a large amount of data that is collected and then stored. It is the perception of this mostly routine data that is at the heart of an organization’s failure to capture, store and analyze the log data being generated constantly. You should never underestimate the importance of the data found within these files.
When manually collecting and reviewing log data, you need to be aware that the more servers you have producing log data, the potential for awareness and locating a security related or compliance issue decreases exponentially over time.
Security is always in the forefront of any size organization’s IT strategy. Usually this strategy focuses on the perimeter of the network to prevent unauthorized access or attacks from malicious parties, that are not associated with the organization.
While external security is essential, what about the internal aspects? A nosy employee who wants to look at confidential company financial data and changes their access permissions? Or a disgruntled employee who has created a back door into a key server and is about to delete terabytes of customer data? While these may be extreme cases, are you prepared to counteract these possible events? The potential for a security breach is just as likely from an internal source as it is from the outside. In fact, it may even be higher. As we discussed in the Executive Summary, the potential for liability is considerable when some unauthorized individual accesses data that is considered protected by legislative act.
Through establishment of a comprehensive ELM strategy for security monitoring of Windows event logs for internal activities and changes that are out of the range of normal business activities, you can locate and prevent small events before they turn into a major catastrophe.
Your organization may or may not face regulatory compliance. If you are a private entity, most likely you do not. However, this should not prevent you from understanding what the regulatory standards have defined as requirements. Leveraging these standards can provide you with a blueprint for your own internal security plans and log management strategy.
If your organization is public, non-compliance with Sarbanes-Oxley, for example, can result in heavy fines and legal liability for the officers. Many of the requirements of the other legislative or industry specific initiatives for security and compliance, as they relate to log management, overlap with those of Sarbanes-Oxley. As a result, all public and many private companies look to that standard for guidance in building a log management strategy.
A quick review of each of the standards below will provide you with a high level overview to understand each of them and how they can affect your log management strategy.
In Sarbanes-Oxley, the phrase “internal controls” in section 404 of the act is central to compliance efforts. Public companies’ annual reports must include:
[…] an internal control report, which shall –
Several misconceptions center on who and what is impacted by these requirements. First, Sarbanes-Oxley applies only to publicly traded organizations in excess of $75 million, or by extension, private firms to be acquired by public companies. Second, it isn’t just the financial data itself or the reporting of that data with which Sarbanes- Oxley is concerned when it refers to “control structure failure;” this has been very broadly interpreted to include just about anything that could even affect the reliability of that data.
Because of the inherent differences in network configurations, business models, markets, and the preferences of auditors,the best practices that are suggested later in this document should be viewed as a starting point. The below sidebar synthesizes some of the ELM requirements that should be included in your audit and compliance strategies.
While these suggestions are focused towards Sarbanes- Oxley, they can also be used in addressing the requirements of Basel II, GLBA, HIPAA, PCI, and other efforts.
In most cases, this will be part of a larger compliance strategy, so be sure to consult with your audit specialists for more detail relevant to compliance in your specific industry.
The term audit policy, in Microsoft Windows lexicon, simply refers to the types of security events you want to be recorded in the security event logs of your servers and workstations. On Microsoft Windows NT® systems, you must set the audit policy by hand on individual servers and workstations, but in Windows 2000® or Windows 2003® Active Directory® domains, with Group Policy enabled, you can associate uniform audit policy settings for groups of servers or the entire domain. For a summary of key logging categories to enable, please refer to the “BASELINE ELM STRATEGY FOR SECURITY, COMPLIANCE AND AUDIT” table.
By default, Windows event logs and Syslog files are decentralized, which each network device or system recording its own event log activity. To obtain a broader picture of trends going on across the network, administrators tasked with security and compliance-centric initiatives must find a way to merge those records into a central data store for complete monitoring, analysis and reporting. Log data collection and storing is critical since some compliance standards mandate data retention for 7 years or more! Automation can really help here because it will save time and ensure the log data reliability. Remember:
In a typical setup, an administrator will configure an ELM tool to gather event log records nightly (or periodically) from servers and workstations throughout their network. This process involves saving and clearing the active event log files from each system, reading log entries out of the log files into a central database (e.g. Microsoft SQL or Oracle), and finally compressing the saved log files and storing them centrally on a secure server.
Keeping your log data in two formats—as database records and as compressed flat files—offers a distinct auditing advantage. Event log data in flat files compresses extremely well, often down to 5% of the original size. Therefore, in terms of storage cost, it costs very little to keep archived log data for many years should an auditor ever need it. However, flat files are a very poor medium for analysis and reporting, so keeping an active working set of data (often 60 to 90 days) in a database allows ad hoc reporting as well as scheduled reporting to be available for recent events. Look for an ELM tool that provides an easy mechanism for rapid re-import of older saved log files back into your database should they ever be needed. It has been our experience that the majority of employee hours when facing an audit are dedicated to simply chasing flat files around and attempting to extract the same types of data from all of them. Having data at the ready in a central database greatly reduces the potential for lost hours when an auditor comes knocking.
Most organizations have a heterogeneous IT environment, with a broad mix of operating systems, devices and systems. Even though your environment may trend towards Windows desktop and server OSs, you may also want the option of choosing more than just Windows event log monitoring. Syslog support is important to have not only for routers, switches, IDS and firewalls, but also for UNIX or LINUX systems.
Most software products require the use of agents to perform real time monitoring of log files. If any factor influences your choice of a solution this should be the one. If you can opt for a no-agents-required implementation of a monitoring solution, do it. This will save a lot of headaches in the initial implementation, as your network grows, and in the ongoing maintenance of your monitoring solution.
When developing a log monitoring plan, every organization has different rules on what sorts of events they must monitor. IT departments will frequently focus on security events as the sole indicator of any issues. While monitoring the security event log is essential, other event logs can also indicate issues with applications, hardware issues or malicious software. At a minimum, all monitored events should be traceable back their origination point. The “BASELINE ELM STRATEGY FOR SECURITY, COMPLIANCE AND AUDIT” sidebar table in the above section provides the kick-off point for your log monitoring implementation.
Depending on your requirements and the flexibility of the ELM solution you deploy, you should define a methodology for continuous monitoring based on how frequently you want to check logs for events of interest in real-time. Each defined event is polled at a regular interval and will generate an alert or notification when an entry of interest is detected.
The number of events configured, number of target systems and polling frequency will dictate the amount of bandwidth consumed during a polling cycle. If you already know the events of interest on certain systems that you want to monitor, then configure away. If you are establishing your event monitoring for the first time, it may better to start by enabling all events and configuring a higher polling frequency. After your familiarity level increases, you can then pare down the number of events and decreasing the polling frequency.
Reporting is one area to which you should pay particular attention. It provides you with significant data on security trends and proves compliance. Reporting can help you substantiate the need to change security policies based on events that could result or have resulted in compromised security.
Any ELM solution that you implement needs to answer the following questions: