#js&&****&……# emacs: -*- mode: python; py-indent-offset: 4; indent-tabs-mode: t -*- # vi: set ft=python sts=4 ts=4 sw=4 noet : # This file is part of Fail2Ban. # # Fail2Ban is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # Fail2Ban is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Fail2Ban; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. __author__ = "Cyril Jaquier and Fail2Ban Contributors" __copyright__ = "Copyright (c) 2004 Cyril Jaquier, 2011-2013 Yaroslav Halchenko" __license__ = "GPL" import codecs import datetime import fcntl import logging import os import re import sys import time from .actions import Actions from .failmanager import FailManagerEmpty, FailManager from .ipdns import DNSUtils, IPAddr from .observer import Observers from .ticket import FailTicket from .jailthread import JailThread from .datedetector import DateDetector, validateTimeZone from .mytime import MyTime from .failregex import FailRegex, Regex, RegexException from .action import CommandAction from .utils import Utils from ..helpers import getLogger, PREFER_ENC # Gets the instance of the logger. logSys = getLogger(__name__) ## # Log reader class. # # This class reads a log file and detects login failures or anything else # that matches a given regular expression. This class is instantiated by # a Jail object. class Filter(JailThread): ## # Constructor. # # Initialize the filter object with default values. # @param jail the jail object def __init__(self, jail, useDns='warn'): JailThread.__init__(self) ## The jail which contains this filter. self.jail = jail ## The failures manager. self.failManager = FailManager() ## Regular expression pre-filtering matching the failures. self.__prefRegex = None ## The regular expression list matching the failures. self.__failRegex = list() ## The regular expression list with expressions to ignore. self.__ignoreRegex = list() ## Use DNS setting self.setUseDns(useDns) ## The amount of time to look back. self.__findTime = 600 ## Ignore own IPs flag: self.__ignoreSelf = True ## The ignore IP list. self.__ignoreIpSet = set() self.__ignoreIpList = [] ## External command self.__ignoreCommand = False ## Cache for ignoreip: self.__ignoreCache = None ## Size of line buffer self.__lineBufferSize = 1 ## Line buffer self.__lineBuffer = [] ## Store last time stamp, applicable for multi-line self.__lastTimeText = "" self.__lastDate = None ## Next service (cleanup) time self.__nextSvcTime = -(1<<63) ## if set, treat log lines without explicit time zone to be in this time zone self.__logtimezone = None ## Default or preferred encoding (to decode bytes from file or journal): self.__encoding = PREFER_ENC ## Cache temporary holds failures info (used by multi-line for wrapping e. g. conn-id to host): self.__mlfidCache = None ## Error counter (protected, so can be used in filter implementations) ## if it reached 100 (at once), run-cycle will go idle self._errors = 0 ## Next time to update log or journal position in database: self._nextUpdateTM = 0 ## Pending updates (must be executed at next update time or during stop): self._pendDBUpdates = {} ## return raw host (host is not dns): self.returnRawHost = False ## check each regex (used for test purposes): self.checkAllRegex = False ## avoid finding of pending failures (without ID/IP, used in fail2ban-regex): self.ignorePending = True ## callback called on ignoreregex match : self.onIgnoreRegex = None ## if true ignores obsolete failures (failure time < now - findTime): self.checkFindTime = True ## shows that filter is in operation mode (processing new messages): self.inOperation = True ## Ticks counter self.ticks = 0 ## Processed lines counter self.procLines = 0 ## Thread name: self.name="f2b/f."+self.jailName self.dateDetector = DateDetector() logSys.debug("Created %s", self) def __repr__(self): return "%s(%r)" % (self.__class__.__name__, self.jail) @property def jailName(self): return (self.jail is not None and self.jail.name or "~jailless~") def clearAllParams(self): """ Clear all lists/dicts parameters (used by reloading) """ self.delFailRegex() self.delIgnoreRegex() self.delIgnoreIP() def reload(self, begin=True): """ Begin or end of reloading resp. refreshing of all parameters """ if begin: self.clearAllParams() if hasattr(self, 'getLogPaths'): self._reload_logs = dict((k, 1) for k in self.getLogPaths()) else: if hasattr(self, '_reload_logs'): # if it was not reloaded - remove obsolete log file: for path in self._reload_logs: self.delLogPath(path) delattr(self, '_reload_logs') @property def mlfidCache(self): if self.__mlfidCache: return self.__mlfidCache self.__mlfidCache = Utils.Cache(maxCount=100, maxTime=5*60) return self.__mlfidCache @property def prefRegex(self): return self.__prefRegex @prefRegex.setter def prefRegex(self, value): if value: self.__prefRegex = Regex(value, useDns=self.__useDns) else: self.__prefRegex = None ## # Add a regular expression which matches the failure. # # The regular expression can also match any other pattern than failures # and thus can be used for many purporse. # @param value the regular expression def addFailRegex(self, value): multiLine = self.__lineBufferSize > 1 try: regex = FailRegex(value, prefRegex=self.__prefRegex, multiline=multiLine, useDns=self.__useDns) self.__failRegex.append(regex) except RegexException as e: logSys.error(e) raise e def delFailRegex(self, index=None): try: # clear all: if index is None: del self.__failRegex[:] return # delete by index: del self.__failRegex[index] except IndexError: logSys.error("Cannot remove regular expression. Index %d is not " "valid", index) ## # Get the regular expressions as list. # # @return the regular expression list def getFailRegex(self): return [regex.getRegex() for regex in self.__failRegex] ## # Add the regular expression which matches the failure. # # The regular expression can also match any other pattern than failures # and thus can be used for many purpose. # @param value the regular expression def addIgnoreRegex(self, value): try: regex = Regex(value, useDns=self.__useDns) self.__ignoreRegex.append(regex) except RegexException as e: logSys.error(e) raise e def delIgnoreRegex(self, index=None): try: # clear all: if index is None: del self.__ignoreRegex[:] return # delete by index: del self.__ignoreRegex[index] except IndexError: logSys.error("Cannot remove regular expression. Index %d is not " "valid", index) ## # Get the regular expression which matches the failure. # # @return the regular expression def getIgnoreRegex(self): ignoreRegex = list() for regex in self.__ignoreRegex: ignoreRegex.append(regex.getRegex()) return ignoreRegex ## # Set the Use DNS mode # @param value the usedns mode def setUseDns(self, value): if isinstance(value, bool): value = {True: 'yes', False: 'no'}[value] value = value.lower() # must be a string by now if not (value in ('yes', 'warn', 'no', 'raw')): logSys.error("Incorrect value %r specified for usedns. " "Using safe 'no'", value) value = 'no' logSys.debug("Setting usedns = %s for %s", value, self) self.__useDns = value ## # Get the usedns mode # @return the usedns mode def getUseDns(self): return self.__useDns ## # Set the time needed to find a failure. # # This value tells the filter how long it has to take failures into # account. # @param value the time def setFindTime(self, value): value = MyTime.str2seconds(value) self.__findTime = value self.failManager.setMaxTime(value) logSys.info(" findtime: %s", value) ## # Get the time needed to find a failure. # # @return the time def getFindTime(self): return self.__findTime ## # Set the date detector pattern, removing Defaults # # @param pattern the date template pattern def setDatePattern(self, pattern): if pattern is None: self.dateDetector = None return else: dd = DateDetector() dd.default_tz = self.__logtimezone if not isinstance(pattern, (list, tuple)): pattern = list(filter(bool, list(map(str.strip, re.split('\n+', pattern))))) for pattern in pattern: dd.appendTemplate(pattern) self.dateDetector = dd ## # Get the date detector pattern, or Default Detectors if not changed # # @return pattern of the date template pattern def getDatePattern(self): if self.dateDetector is not None: templates = self.dateDetector.templates # lazy template init, by first match if not len(templates) or len(templates) > 2: return None, "Default Detectors" elif len(templates): if hasattr(templates[0], "pattern"): pattern = templates[0].pattern else: pattern = None return pattern, templates[0].name return None ## # Set the log default time zone # # @param tz the symbolic timezone (for now fixed offset only: UTC[+-]HHMM) def setLogTimeZone(self, tz): validateTimeZone(tz); # avoid setting of wrong value, but hold original self.__logtimezone = tz if self.dateDetector: self.dateDetector.default_tz = self.__logtimezone ## # Get the log default timezone # # @return symbolic timezone (a string) def getLogTimeZone(self): return self.__logtimezone ## # Set the maximum retry value. # # @param value the retry value def setMaxRetry(self, value): self.failManager.setMaxRetry(value) logSys.info(" maxRetry: %s", value) ## # Get the maximum retry value. # # @return the retry value def getMaxRetry(self): return self.failManager.getMaxRetry() ## # Set the maximum line buffer size. # # @param value the line buffer size def setMaxLines(self, value): if int(value) <= 0: raise ValueError("maxlines must be integer greater than zero") self.__lineBufferSize = int(value) logSys.info(" maxLines: %i", self.__lineBufferSize) ## # Get the maximum line buffer size. # # @return the line buffer size def getMaxLines(self): return self.__lineBufferSize ## # Set the log file encoding # # @param encoding the encoding used with log files def setLogEncoding(self, encoding): if encoding.lower() == "auto": encoding = PREFER_ENC codecs.lookup(encoding) # Raise LookupError if invalid codec self.__encoding = encoding logSys.info(" encoding: %s", encoding) return encoding ## # Get the log file encoding # # @return log encoding value def getLogEncoding(self): return self.__encoding ## # Main loop. # # This function is the main loop of the thread. It checks if the # file has been modified and looks for failures. # @return True when the thread exits nicely def run(self): # pragma: no cover raise Exception("run() is abstract") ## # External command, for ignoredips # @property def ignoreCommand(self): return self.__ignoreCommand @ignoreCommand.setter def ignoreCommand(self, command): self.__ignoreCommand = command ## # Cache parameters for ignoredips # @property def ignoreCache(self): return [self.__ignoreCache[0], self.__ignoreCache[1].maxCount, self.__ignoreCache[1].maxTime] \ if self.__ignoreCache else None @ignoreCache.setter def ignoreCache(self, command): if command: self.__ignoreCache = command['key'], Utils.Cache( maxCount=int(command.get('max-count', 100)), maxTime=MyTime.str2seconds(command.get('max-time', 5*60)) ) else: self.__ignoreCache = None def performBan(self, ip=None): """Performs a ban for IPs (or given ip) that are reached maxretry of the jail.""" while True: try: ticket = self.failManager.toBan(ip) except FailManagerEmpty: break self.jail.putFailTicket(ticket) if ip: break self.performSvc() def performSvc(self, force=False): """Performs a service tasks (clean failure list).""" tm = MyTime.time() # avoid too early clean up: if force or tm >= self.__nextSvcTime: self.__nextSvcTime = tm + 5 # clean up failure list: self.failManager.cleanup(tm) def addAttempt(self, ip, *matches): """Generate a failed attempt for ip""" if not isinstance(ip, IPAddr): ip = IPAddr(ip) matches = list(matches) # tuple to list # Generate the failure attempt for the IP: unixTime = MyTime.time() ticket = FailTicket(ip, unixTime, matches=matches) logSys.info( "[%s] Attempt %s - %s", self.jailName, ip, datetime.datetime.fromtimestamp(unixTime).strftime("%Y-%m-%d %H:%M:%S") ) attempts = self.failManager.addFailure(ticket, len(matches) or 1) # Perform the ban if this attempt is resulted to: if attempts >= self.failManager.getMaxRetry(): self.performBan(ip) return 1 ## # Ignore own IP/DNS. # @property def ignoreSelf(self): return self.__ignoreSelf @ignoreSelf.setter def ignoreSelf(self, value): self.__ignoreSelf = value ## # Add an IP/DNS to the ignore list. # # IP addresses in the ignore list are not taken into account # when finding failures. CIDR mask and DNS are also accepted. # @param ip IP address to ignore def addIgnoreIP(self, ipstr): # An empty string is always false if ipstr == "": return # Create IP address object ip = IPAddr(ipstr) # Avoid exact duplicates if ip in self.__ignoreIpSet or ip in self.__ignoreIpList: logSys.log(logging.MSG, " Ignore duplicate %r (%r), already in ignore list", ip, ipstr) return # log and append to ignore list logSys.debug(" Add %r to ignore list (%r)", ip, ipstr) # if single IP (not DNS or a subnet) add to set, otherwise to list: if ip.isSingle: self.__ignoreIpSet.add(ip) else: self.__ignoreIpList.append(ip) def delIgnoreIP(self, ip=None): # clear all: if ip is None: self.__ignoreIpSet.clear() del self.__ignoreIpList[:] return # delete by ip: logSys.debug(" Remove %r from ignore list", ip) if ip in self.__ignoreIpSet: self.__ignoreIpSet.remove(ip) else: self.__ignoreIpList.remove(ip) def logIgnoreIp(self, ip, log_ignore, ignore_source="unknown source"): if log_ignore: logSys.info("[%s] Ignore %s by %s", self.jailName, ip, ignore_source) def getIgnoreIP(self): return self.__ignoreIpList + list(self.__ignoreIpSet) ## # Check if IP address/DNS is in the ignore list. # # Check if the given IP address matches an IP address/DNS or a CIDR # mask in the ignore list. # @param ip IP address object or ticket # @return True if IP address is in ignore list def inIgnoreIPList(self, ip, log_ignore=True): ticket = None if isinstance(ip, FailTicket): ticket = ip ip = ticket.getID() elif not isinstance(ip, IPAddr): ip = IPAddr(ip) return self._inIgnoreIPList(ip, ticket, log_ignore) def _inIgnoreIPList(self, ip, ticket, log_ignore=True): aInfo = None # cached ? if self.__ignoreCache: key, c = self.__ignoreCache if ticket: aInfo = Actions.ActionInfo(ticket, self.jail) key = CommandAction.replaceDynamicTags(key, aInfo) else: aInfo = { 'ip': ip } key = CommandAction.replaceTag(key, aInfo) v = c.get(key) if v is not None: return v # check own IPs should be ignored and 'ip' is self IP: if self.__ignoreSelf and ip in DNSUtils.getSelfIPs(): self.logIgnoreIp(ip, log_ignore, ignore_source="ignoreself rule") if self.__ignoreCache: c.set(key, True) return True # check if the IP is covered by ignore IP (in set or in subnet/dns): if ip in self.__ignoreIpSet: self.logIgnoreIp(ip, log_ignore, ignore_source="ip") return True for net in self.__ignoreIpList: if ip.isInNet(net): self.logIgnoreIp(ip, log_ignore, ignore_source=("ip" if net.isValid else "dns")) if self.__ignoreCache: c.set(key, True) return True if self.__ignoreCommand: if ticket: if not aInfo: aInfo = Actions.ActionInfo(ticket, self.jail) command = CommandAction.replaceDynamicTags(self.__ignoreCommand, aInfo) else: if not aInfo: aInfo = { 'ip': ip } command = CommandAction.replaceTag(self.__ignoreCommand, aInfo) logSys.debug('ignore command: %s', command) ret, ret_ignore = CommandAction.executeCmd(command, success_codes=(0, 1)) ret_ignore = ret and ret_ignore == 0 self.logIgnoreIp(ip, log_ignore and ret_ignore, ignore_source="command") if self.__ignoreCache: c.set(key, ret_ignore) return ret_ignore if self.__ignoreCache: c.set(key, False) return False def _logWarnOnce(self, nextLTM, *args): """Log some issue as warning once per day, otherwise level 7""" if MyTime.time() < getattr(self, nextLTM, 0): if logSys.getEffectiveLevel() <= 7: logSys.log(7, *(args[0])) else: setattr(self, nextLTM, MyTime.time() + 24*60*60) for args in args: logSys.warning('[%s] ' + args[0], self.jailName, *args[1:]) def processLine(self, line, date=None): """Split the time portion from log msg and return findFailures on them """ logSys.log(7, "Working on line %r", line) noDate = False if date: tupleLine = line line = "".join(line) self.__lastTimeText = tupleLine[1] self.__lastDate = date else: # try to parse date: timeMatch = self.dateDetector.matchTime(line) m = timeMatch[0] if m: s = m.start(1) e = m.end(1) m = line[s:e] tupleLine = (line[:s], m, line[e:]) if m: # found and not empty - retrive date: date = self.dateDetector.getTime(m, timeMatch) if date is not None: # Lets get the time part date = date[0] self.__lastTimeText = m self.__lastDate = date else: logSys.error("findFailure failed to parse timeText: %s", m) # matched empty value - date is optional or not available - set it to last known or now: elif self.__lastDate and self.__lastDate > MyTime.time() - 60: # set it to last known: tupleLine = ("", self.__lastTimeText, line) date = self.__lastDate else: # set it to now: date = MyTime.time() else: tupleLine = ("", "", line) # still no date - try to use last known: if date is None: noDate = True if self.__lastDate and self.__lastDate > MyTime.time() - 60: tupleLine = ("", self.__lastTimeText, line) date = self.__lastDate elif self.checkFindTime and self.inOperation: date = MyTime.time() if self.checkFindTime and date is not None: # if in operation (modifications have been really found): if self.inOperation: # if weird date - we'd simulate now for timeing issue (too large deviation from now): delta = int(date - MyTime.time()) if abs(delta) > 60: # log timing issue as warning once per day: self._logWarnOnce("_next_simByTimeWarn", ("Detected a log entry %s %s the current time in operation mode. " "This looks like a %s problem. Treating such entries as if they just happened.", MyTime.seconds2str(abs(delta)), "before" if delta < 0 else "after", "latency" if -3300 <= delta < 0 else "timezone" ), ("Please check a jail for a timing issue. Line with odd timestamp: %s", line)) # simulate now as date: date = MyTime.time() self.__lastDate = date else: # in initialization (restore) phase, if too old - ignore: if date < MyTime.time() - self.getFindTime(): # log time zone issue as warning once per day: self._logWarnOnce("_next_ignByTimeWarn", ("Ignoring all log entries older than %ss; these are probably" + " messages generated while fail2ban was not running.", self.getFindTime()), ("Please check a jail for a timing issue. Line with odd timestamp: %s", line)) # ignore - too old (obsolete) entry: return [] # save last line (lazy convert of process line tuple to string on demand): self.processedLine = lambda: "".join(tupleLine[::2]) return self.findFailure(tupleLine, date, noDate=noDate) def processLineAndAdd(self, line, date=None): """Processes the line for failures and populates failManager """ try: for (_, ip, unixTime, fail) in self.processLine(line, date): logSys.debug("Processing line with time:%s and ip:%s", unixTime, ip) # ensure the time is not in the future, e. g. by some estimated (assumed) time: if self.checkFindTime and unixTime > MyTime.time(): unixTime = MyTime.time() tick = FailTicket(ip, unixTime, data=fail) if self._inIgnoreIPList(ip, tick): continue logSys.info( "[%s] Found %s - %s", self.jailName, ip, MyTime.time2str(unixTime) ) attempts = self.failManager.addFailure(tick) # avoid RC on busy filter (too many failures) - if attempts for IP/ID reached maxretry, # we can speedup ban, so do it as soon as possible: if attempts >= self.failManager.getMaxRetry(): self.performBan(ip) # report to observer - failure was found, for possibly increasing of it retry counter (asynchronous) if Observers.Main is not None: Observers.Main.add('failureFound', self.jail, tick) self.procLines += 1 # every 100 lines check need to perform service tasks: if self.procLines % 100 == 0: self.performSvc() # reset (halve) error counter (successfully processed line): if self._errors: self._errors //= 2 except Exception as e: logSys.error("Failed to process line: %r, caught exception: %r", line, e, exc_info=logSys.getEffectiveLevel()<=logging.DEBUG) # incr common error counter: self.commonError() def commonError(self, reason="common", exc=None): # incr error counter, stop processing (going idle) after 100th error : self._errors += 1 # sleep a little bit (to get around time-related errors): time.sleep(self.sleeptime) if self._errors >= 100: logSys.error("Too many errors at once (%s), going idle", self._errors) self._errors //= 2 self.idle = True def _ignoreLine(self, buf, orgBuffer, failRegex=None): # if multi-line buffer - use matched only, otherwise (single line) - original buf: if failRegex and self.__lineBufferSize > 1: orgBuffer = failRegex.getMatchedTupleLines() buf = Regex._tupleLinesBuf(orgBuffer) # search ignored: fnd = None for ignoreRegexIndex, ignoreRegex in enumerate(self.__ignoreRegex): ignoreRegex.search(buf, orgBuffer) if ignoreRegex.hasMatched(): fnd = ignoreRegexIndex logSys.log(7, " Matched ignoreregex %d and was ignored", fnd) if self.onIgnoreRegex: self.onIgnoreRegex(fnd, ignoreRegex) # remove ignored match: if not self.checkAllRegex or self.__lineBufferSize > 1: # todo: check ignoreRegex.getUnmatchedTupleLines() would be better (fix testGetFailuresMultiLineIgnoreRegex): if failRegex: self.__lineBuffer = failRegex.getUnmatchedTupleLines() if not self.checkAllRegex: break return fnd def _updateUsers(self, fail, user=()): users = fail.get('users') # only for regex contains user: if user: if not users: fail['users'] = users = set() users.add(user) return users return users def _mergeFailure(self, mlfid, fail, failRegex): mlfidFail = self.mlfidCache.get(mlfid) if self.__mlfidCache else None users = None nfflgs = 0 if fail.get("mlfgained"): nfflgs |= (8|1) if not fail.get('nofail'): fail['nofail'] = fail["mlfgained"] elif fail.get('nofail'): nfflgs |= 1 if fail.pop('mlfforget', None): nfflgs |= 2 # if multi-line failure id (connection id) known: if mlfidFail: mlfidGroups = mlfidFail[1] # update users set (hold all users of connect): users = self._updateUsers(mlfidGroups, fail.get('user')) # be sure we've correct current state ('nofail' and 'mlfgained' only from last failure) if mlfidGroups.pop('nofail', None): nfflgs |= 4 if mlfidGroups.pop('mlfgained', None): nfflgs |= 4 # gained resets all pending failures (retaining users to check it later) if nfflgs & 8: mlfidGroups.pop('mlfpending', None) # if we had no pending failures then clear the matches (they are already provided): if (nfflgs & 4) == 0 and not mlfidGroups.get('mlfpending', 0): mlfidGroups.pop("matches", None) # overwrite multi-line failure with all values, available in fail: mlfidGroups.update(((k,v) for k,v in fail.items() if v is not None)) # new merged failure data: fail = mlfidGroups # if forget (disconnect/reset) - remove cached entry: if nfflgs & 2: self.mlfidCache.unset(mlfid) elif not (nfflgs & 2): # not mlfforget users = self._updateUsers(fail, fail.get('user')) mlfidFail = [self.__lastDate, fail] self.mlfidCache.set(mlfid, mlfidFail) # check users in order to avoid reset failure by multiple logon-attempts: if fail.pop('mlfpending', 0) or users and len(users) > 1: # we've pending failures or new user, reset 'nofail' because of failures or multiple users attempts: fail.pop('nofail', None) fail.pop('mlfgained', None) nfflgs &= ~(8|1) # reset nofail and gained # merge matches: if (nfflgs & 1) == 0: # current nofail state (corresponding users) m = fail.pop("nofail-matches", []) m += fail.get("matches", []) if (nfflgs & 8) == 0: # no gain signaled m += failRegex.getMatchedTupleLines() fail["matches"] = m elif (nfflgs & 3) == 1: # not mlfforget and nofail: fail["nofail-matches"] = fail.get("nofail-matches", []) + failRegex.getMatchedTupleLines() # return merged: return fail ## # Finds the failure in a line given split into time and log parts. # # Uses the failregex pattern to find it and timeregex in order # to find the logging time. # @return a dict with IP and timestamp. def findFailure(self, tupleLine, date, noDate=False): failList = list() ll = logSys.getEffectiveLevel() defcidr = IPAddr.CIDR_UNSPEC if self.__useDns == "raw" or self.returnRawHost: defcidr = IPAddr.CIDR_RAW if self.__lineBufferSize > 1: self.__lineBuffer.append(tupleLine) orgBuffer = self.__lineBuffer = self.__lineBuffer[-self.__lineBufferSize:] else: orgBuffer = self.__lineBuffer = [tupleLine] if ll <= 5: logSys.log(5, "Looking for match of %r", orgBuffer) buf = Regex._tupleLinesBuf(orgBuffer) # Checks if we must ignore this line (only if fewer ignoreregex than failregex). if self.__ignoreRegex and len(self.__ignoreRegex) < len(self.__failRegex) - 2: if self._ignoreLine(buf, orgBuffer) is not None: # The ignoreregex matched. Return. return failList # Pre-filter fail regex (if available): preGroups = {} if self.__prefRegex: if ll <= 5: logSys.log(5, " Looking for prefregex %r", self.__prefRegex.getRegex()) self.__prefRegex.search(buf, orgBuffer) if not self.__prefRegex.hasMatched(): if ll <= 5: logSys.log(5, " Prefregex not matched") return failList preGroups = self.__prefRegex.getGroups() if ll <= 7: logSys.log(7, " Pre-filter matched %s", preGroups) repl = preGroups.pop('content', None) # Content replacement: if repl: self.__lineBuffer, buf = [('', '', repl)], None # Iterates over all the regular expressions. for failRegexIndex, failRegex in enumerate(self.__failRegex): try: # buffer from tuples if changed: if buf is None: buf = Regex._tupleLinesBuf(self.__lineBuffer) if ll <= 5: logSys.log(5, " Looking for failregex %d - %r", failRegexIndex, failRegex.getRegex()) failRegex.search(buf, orgBuffer) if not failRegex.hasMatched(): continue # current failure data (matched group dict): fail = failRegex.getGroups() # The failregex matched. if ll <= 7: logSys.log(7, " Matched failregex %d: %s", failRegexIndex, fail) # Checks if we must ignore this match. if self.__ignoreRegex and self._ignoreLine(buf, orgBuffer, failRegex) is not None: # The ignoreregex matched. Remove ignored match. buf = None if not self.checkAllRegex: break continue if noDate: self._logWarnOnce("_next_noTimeWarn", ("Found a match but no valid date/time found for %r.", tupleLine[1]), ("Match without a timestamp: %s", "\n".join(failRegex.getMatchedLines())), ("Please try setting a custom date pattern (see man page jail.conf(5)).",) ) if date is None and self.checkFindTime: continue # we should check all regex (bypass on multi-line, otherwise too complex): if not self.checkAllRegex or self.__lineBufferSize > 1: self.__lineBuffer, buf = failRegex.getUnmatchedTupleLines(), None # merge data if multi-line failure: cidr = defcidr raw = (defcidr == IPAddr.CIDR_RAW) if preGroups: currFail, fail = fail, preGroups.copy() fail.update(currFail) # first try to check we have mlfid case (caching of connection id by multi-line): mlfid = fail.get('mlfid') if mlfid is not None: fail = self._mergeFailure(mlfid, fail, failRegex) # bypass if no-failure case: if fail.get('nofail'): if ll <= 7: logSys.log(7, "Nofail by mlfid %r in regex %s: %s", mlfid, failRegexIndex, fail.get('mlfforget', "waiting for failure")) if not self.checkAllRegex: return failList else: # matched lines: fail["matches"] = fail.get("matches", []) + failRegex.getMatchedTupleLines() # failure-id: fid = fail.get('fid') # ip-address or host: ip = fail.get('ip4') if ip is not None: cidr = int(fail.get('cidr') or IPAddr.FAM_IPv4) raw = True else: ip = fail.get('ip6') if ip is not None: cidr = int(fail.get('cidr') or IPAddr.FAM_IPv6) raw = True else: ip = fail.get('dns') if ip is None: # first try to check we have mlfid case (cache connection id): if fid is None and mlfid is None: # if no failure-id also (obscure case, wrong regex), throw error inside getFailID: fid = failRegex.getFailID() ip = fid raw = True # if mlfid case (not failure): if ip is None: if ll <= 7: logSys.log(7, "No failure-id by mlfid %r in regex %s: %s", mlfid, failRegexIndex, fail.get('mlfforget', "waiting for identifier")) fail['mlfpending'] = 1; # mark failure is pending if not self.checkAllRegex and self.ignorePending: return failList fids = [None] # if raw - add single ip or failure-id, # otherwise expand host to multiple ips using dns (or ignore it if not valid): elif raw: # check ip/host equal failure-id, if not - failure with complex id: if fid is None or fid == ip: fid = IPAddr(ip, cidr) else: fail['ip'] = IPAddr(ip, cidr) fid = IPAddr(fid, defcidr) fids = [fid] # otherwise, try to use dns conversion: else: fids = DNSUtils.textToIp(ip, self.__useDns) # if checkAllRegex we must make a copy (to be sure next RE doesn't change merged/cached failure): if self.checkAllRegex and mlfid is not None: fail = fail.copy() # append failure with match to the list: for fid in fids: failList.append([failRegexIndex, fid, date, fail]) if not self.checkAllRegex: break except RegexException as e: # pragma: no cover - unsure if reachable logSys.error(e) return failList def status(self, flavor="basic"): """Status of failures detected by filter. """ ret = [("Currently failed", self.failManager.size()), ("Total failed", self.failManager.getFailTotal())] return ret class FileFilter(Filter): def __init__(self, jail, **kwargs): Filter.__init__(self, jail, **kwargs) ## The log file path. self.__logs = dict() self.__autoSeek = dict() ## # Add a log file path # # @param path log file path def addLogPath(self, path, tail=False, autoSeek=True): if path in self.__logs: if hasattr(self, '_reload_logs') and path in self._reload_logs: del self._reload_logs[path] else: logSys.error(path + " already exists") else: log = FileContainer(path, self.getLogEncoding(), tail) db = self.jail.database if db is not None: lastpos = db.addLog(self.jail, log) if lastpos and not tail: log.setPos(lastpos) self.__logs[path] = log logSys.info("Added logfile: %r (pos = %s, hash = %s)" , path, log.getPos(), log.getHash()) if autoSeek and not tail: self.__autoSeek[path] = autoSeek self._addLogPath(path) # backend specific def _addLogPath(self, path): # nothing to do by default # to be overridden by backends pass ## # Delete a log path # # @param path the log file to delete def delLogPath(self, path): try: log = self.__logs.pop(path) except KeyError: return logSys.info("Removed logfile: %r", path) self._delLogPath(path) return def _delLogPath(self, path): # pragma: no cover - overwritten function # nothing to do by default # to be overridden by backends pass ## # Get the log file names # # @return log paths def getLogPaths(self): return list(self.__logs.keys()) ## # Get the log containers # # @return log containers def getLogs(self): return list(self.__logs.values()) ## # Get the count of log containers # # @return count of log containers def getLogCount(self): return len(self.__logs) ## # Check whether path is already monitored. # # @param path The path # @return True if the path is already monitored else False def containsLogPath(self, path): return path in self.__logs ## # Set the log file encoding # # @param encoding the encoding used with log files def setLogEncoding(self, encoding): encoding = super(FileFilter, self).setLogEncoding(encoding) for log in self.__logs.values(): log.setEncoding(encoding) def getLog(self, path): return self.__logs.get(path, None) ## # Gets all the failure in the log file. # # Gets all the failure in the log file which are newer than # MyTime.time()-self.findTime. When a failure is detected, a FailTicket # is created and is added to the FailManager. def getFailures(self, filename, inOperation=None): if self.idle: return False log = self.getLog(filename) if log is None: logSys.error("Unable to get failures in %s", filename) return False # We should always close log (file), otherwise may be locked (log-rotate, etc.) try: # Try to open log file. try: has_content = log.open() # see http://python.org/dev/peps/pep-3151/ except IOError as e: logSys.error("Unable to open %s", filename) if e.errno != 2: # errno.ENOENT logSys.exception(e) return False except OSError as e: # pragma: no cover - requires race condition to trigger this logSys.error("Error opening %s", filename) logSys.exception(e) return False except Exception as e: # pragma: no cover - Requires implementation error in FileContainer to generate logSys.error("Internal error in FileContainer open method - please report as a bug to https://github.com/fail2ban/fail2ban/issues") logSys.exception(e) return False # seek to find time for first usage only (prevent performance decline with polling of big files) if self.__autoSeek: startTime = self.__autoSeek.pop(filename, None) if startTime: # if default, seek to "current time" - "find time": if isinstance(startTime, bool): startTime = MyTime.time() - self.getFindTime() # prevent completely read of big files first time (after start of service), # initial seek to start time using half-interval search algorithm: try: self.seekToTime(log, startTime) except Exception as e: # pragma: no cover logSys.error("Error during seek to start time in \"%s\"", filename) raise logSys.exception(e) return False if has_content: while not self.idle: line = log.readline() if not self.active: break; # jail has been stopped if line is None: # The jail reached the bottom, simply set in operation for this log # (since we are first time at end of file, growing is only possible after modifications): log.inOperation = True break # acquire in operation from log and process: self.inOperation = inOperation if inOperation is not None else log.inOperation self.processLineAndAdd(line) finally: log.close() if self.jail.database is not None: self._pendDBUpdates[log] = 1 if ( self.ticks % 100 == 0 or MyTime.time() >= self._nextUpdateTM or not self.active ): self._updateDBPending() self._nextUpdateTM = MyTime.time() + Utils.DEFAULT_SLEEP_TIME * 5 return True ## # Seeks to line with date (search using half-interval search algorithm), to start polling from it # def seekToTime(self, container, date, accuracy=3): fs = container.getFileSize() if logSys.getEffectiveLevel() <= logging.DEBUG: logSys.debug("Seek to find time %s (%s), file size %s", date, MyTime.time2str(date), fs) if not fs: return minp = container.getPos() maxp = fs tryPos = minp lastPos = -1 foundPos = 0 foundTime = None cntr = 0 unixTime = None movecntr = accuracy while maxp > minp: if tryPos is None: pos = int(minp + (maxp - minp) / 2) else: pos, tryPos = tryPos, None # because container seek will go to start of next line (minus CRLF): pos = max(0, pos-2) seekpos = pos = container.seek(pos) cntr += 1 # within next 5 lines try to find any legal datetime: lncntr = 5; dateTimeMatch = None nextp = None while True: line = container.readline(False) if line is None: break (timeMatch, template) = self.dateDetector.matchTime(line) if timeMatch: dateTimeMatch = self.dateDetector.getTime( line[timeMatch.start():timeMatch.end()], (timeMatch, template)) else: nextp = container.tell() if nextp > maxp: pos = seekpos break pos = nextp if not dateTimeMatch and lncntr: lncntr -= 1 continue break # not found at this step - stop searching if dateTimeMatch: unixTime = dateTimeMatch[0] if unixTime >= date: if foundTime is None or unixTime <= foundTime: foundPos = pos foundTime = unixTime if pos == maxp: pos = seekpos if pos < maxp: maxp = pos else: if foundTime is None or unixTime >= foundTime: foundPos = pos foundTime = unixTime if nextp is None: nextp = container.tell() pos = nextp if pos > minp: minp = pos # if we can't move (position not changed) if pos == lastPos: movecntr -= 1 if movecntr <= 0: break # we have found large area without any date matched # or end of search - try min position (because can be end of previous line): if minp != lastPos: lastPos = tryPos = minp continue break lastPos = pos # always use smallest pos, that could be found: foundPos = container.seek(minp, False) container.setPos(foundPos) if logSys.getEffectiveLevel() <= logging.DEBUG: logSys.debug("Position %s from %s, found time %s (%s) within %s seeks", lastPos, fs, foundTime, (MyTime.time2str(foundTime) if foundTime is not None else ''), cntr) def status(self, flavor="basic"): """Status of Filter plus files being monitored. """ ret = super(FileFilter, self).status(flavor=flavor) path = list(self.__logs.keys()) ret.append(("File list", path)) return ret def _updateDBPending(self): """Apply pending updates (log position) to database. """ db = self.jail.database while True: try: log, args = self._pendDBUpdates.popitem() except KeyError: break db.updateLog(self.jail, log) def onStop(self): """Stop monitoring of log-file(s). Invoked after run method. """ # ensure positions of pending logs are up-to-date: if self._pendDBUpdates and self.jail.database: self._updateDBPending() # stop files monitoring: for path in list(self.__logs.keys()): self.delLogPath(path) def stop(self): """Stop filter """ # normally onStop will be called automatically in thread after its run ends, # but for backwards compatibilities we'll invoke it in caller of stop method. self.onStop() # stop thread: super(Filter, self).stop() ## # FileContainer class. # # This class manages a file handler and takes care of log rotation detection. # In order to detect log rotation, the hash (MD5) of the first line of the file # is computed and compared to the previous hash of this line. try: import hashlib try: md5sum = hashlib.md5 # try to use it (several standards like FIPS forbid it): md5sum(' ').hexdigest() except: # pragma: no cover md5sum = hashlib.sha1 except ImportError: # pragma: no cover # hashlib was introduced in Python 2.5. For compatibility with those # elderly Pythons, import from md5 import md5 md5sum = md5.new class FileContainer: def __init__(self, filename, encoding, tail=False, doOpen=False): self.__filename = filename self.waitForLineEnd = True self.setEncoding(encoding) self.__tail = tail self.__handler = None self.__pos = 0 self.__pos4hash = 0 self.__hash = '' self.__hashNextTime = time.time() + 30 # Try to open the file. Raises an exception if an error occurred. handler = open(filename, 'rb') if doOpen: # fail2ban-regex only (don't need to reopen it and check for rotation) self.__handler = handler return try: stats = os.fstat(handler.fileno()) self.__ino = stats.st_ino if stats.st_size: firstLine = handler.readline() # first line available and contains new-line: if firstLine != firstLine.rstrip(b'\r\n'): # Computes the MD5 of the first line. self.__hash = md5sum(firstLine).hexdigest() # if tail mode scroll to the end of file if tail: handler.seek(0, 2) self.__pos = handler.tell() finally: handler.close() ## shows that log is in operation mode (expecting new messages only from here): self.inOperation = tail def __hash__(self): return hash(self.__filename) def __eq__(self, other): return (id(self) == id(other) or self.__filename == (other.__filename if isinstance(other, FileContainer) else other) ) def __repr__(self): return 'file-log:'+self.__filename def getFileName(self): return self.__filename def getFileSize(self): h = self.__handler if h is not None: stats = os.fstat(h.fileno()) return stats.st_size return os.path.getsize(self.__filename); def setEncoding(self, encoding): codecs.lookup(encoding) # Raises LookupError if invalid self.__encoding = encoding def getEncoding(self): return self.__encoding def getHash(self): return self.__hash def getPos(self): return self.__pos def setPos(self, value): self.__pos = value def open(self, forcePos=None): h = open(self.__filename, 'rb') try: # Set the file descriptor to be FD_CLOEXEC fd = h.fileno() flags = fcntl.fcntl(fd, fcntl.F_GETFD) fcntl.fcntl(fd, fcntl.F_SETFD, flags | fcntl.FD_CLOEXEC) myHash = self.__hash # Stat the file before even attempting to read it stats = os.fstat(h.fileno()) rotflg = stats.st_size < self.__pos or stats.st_ino != self.__ino if rotflg or not len(myHash) or time.time() > self.__hashNextTime: myHash = '' firstLine = h.readline() # Computes the MD5 of the first line (if it is complete) if firstLine != firstLine.rstrip(b'\r\n'): myHash = md5sum(firstLine).hexdigest() self.__hashNextTime = time.time() + 30 elif stats.st_size == self.__pos: myHash = self.__hash # Compare size, hash and inode if rotflg or myHash != self.__hash: if self.__hash != '': logSys.log(logging.MSG, "Log rotation detected for %s, reason: %r", self.__filename, (stats.st_size, self.__pos, stats.st_ino, self.__ino, myHash, self.__hash)) self.__ino = stats.st_ino self.__pos = 0 self.__hash = myHash # if nothing to read from file yet (empty or no new data): if forcePos is not None: self.__pos = forcePos elif stats.st_size <= self.__pos: return False # Sets the file pointer to the last position. h.seek(self.__pos) # leave file open (to read content): self.__handler = h; h = None finally: # close (no content or error only) if h: h.close(); h = None return True def seek(self, offs, endLine=True): h = self.__handler if h is None: self.open(offs) h = self.__handler # seek to given position h.seek(offs, 0) # goto end of next line if offs and endLine: h.readline() # get current real position return h.tell() def tell(self): # get current real position return self.__handler.tell() @staticmethod def decode_line(filename, enc, line): try: return line.decode(enc, 'strict') except (UnicodeDecodeError, UnicodeEncodeError) as e: # avoid warning if got incomplete end of line (e. g. '\n' in "...[0A" followed by "00]..." for utf-16le: if (e.end == len(line) and line[e.start] in b'\r\n'): return line[0:e.start].decode(enc, 'replace') global _decode_line_warn lev = 7 if not _decode_line_warn.get(filename, 0): lev = logging.WARNING _decode_line_warn.set(filename, 1) logSys.log(lev, "Error decoding line from '%s' with '%s'.", filename, enc) if logSys.getEffectiveLevel() <= lev: logSys.log(lev, "Consider setting logencoding to appropriate encoding for this jail. " "Continuing to process line ignoring invalid characters: %r", line) # decode with replacing error chars: line = line.decode(enc, 'replace') return line def readline(self, complete=True): """Read line from file In opposite to pythons readline it doesn't return new-line, so returns either the line if line is complete (and complete=True) or None if line is not complete (and complete=True) or there is no content to read. If line is complete (and complete is True), it also shift current known position to begin of next line. Also it is safe against interim new-line bytes (e. g. part of multi-byte char) in given encoding. """ if self.__handler is None: return "" # read raw bytes up to \n char: b = self.__handler.readline() if not b: return None bl = len(b) # convert to log-encoding (new-line char could disappear if it is part of multi-byte sequence): r = FileContainer.decode_line( self.getFileName(), self.getEncoding(), b) # trim new-line at end and check the line was written complete (contains a new-line): l = r.rstrip('\r\n') if complete: if l == r: # try to fill buffer in order to find line-end in log encoding: fnd = 0 while 1: r = self.__handler.readline() if not r: break b += r bl += len(r) # convert to log-encoding: r = FileContainer.decode_line( self.getFileName(), self.getEncoding(), b) # ensure new-line is not in the middle (buffered 2 strings, e. g. in utf-16le it is "...[0A"+"00]..."): e = r.find('\n') if e >= 0 and e != len(r)-1: l, r = r[0:e], r[0:e+1] # back to bytes and get offset to seek after NL: r = r.encode(self.getEncoding(), 'replace') self.__handler.seek(-bl+len(r), 1) return l # trim new-line at end and check the line was written complete (contains a new-line): l = r.rstrip('\r\n') if l != r: return l if self.waitForLineEnd: # not fulfilled - seek back and return: self.__handler.seek(-bl, 1) return None return l def close(self): if self.__handler is not None: # Saves the last real position. self.__pos = self.__handler.tell() # Closes the file. self.__handler.close() self.__handler = None def __iter__(self): return self def __next__(self): line = self.readline() if line is None: self.close() raise StopIteration return line _decode_line_warn = Utils.Cache(maxCount=1000, maxTime=24*60*60); ## # JournalFilter class. # # Base interface class for systemd journal filters class JournalFilter(Filter): # pragma: systemd no cover def clearAllParams(self): super(JournalFilter, self).clearAllParams() self.delJournalMatch() def addJournalMatch(self, match): # pragma: no cover - Base class, not used pass def delJournalMatch(self, match=None): # pragma: no cover - Base class, not used pass def getJournalMatch(self, match): # pragma: no cover - Base class, not used return []