Overview
Comment: | switched sg_import to python3.2 reworked SysLogHandlerQueue to stick to one socket reworked FReadlineQueue to use io.FileIO for file operation |
---|---|
Downloads: | Tarball | ZIP archive | SQL archive |
Timelines: | family | ancestors | descendants | both | master | trunk |
Files: | files | file ages | folders |
SHA3-256: |
67e8b3309d6e56706b82456eba721093 |
User & Date: | arcade@b1t.name on 2012-07-09 14:25:12.000 |
Other Links: | branch diff | manifest | tags |
Context
2012-07-10
| ||
14:02 | remove leftover fix conversion from multibyte check-in: dbc0727869 user: arcade@b1t.name tags: master, trunk | |
2012-07-09
| ||
14:25 | switched sg_import to python3.2 reworked SysLogHandlerQueue to stick to one socket reworked FReadlineQueue to use io.FileIO for file operation check-in: 67e8b3309d user: arcade@b1t.name tags: master, trunk | |
11:26 | added gevent wrapper for SysLogHandler automatically recode all log lines from utf-8 to str simplify invoking added some comment check-in: 2654b86697 user: arcade@b1t.name tags: master, trunk | |
Changes
Modified sg_import.py
from [f97ad6d498]
to [5ad2e51f1f].
|
| | | 1 2 3 4 5 6 7 8 | #!/usr/bin/env python3.2 # This script converts SquidGuard database into format that can be imported to # squid-tagger. It should be run in SquidGuard database directory and it would # produce csv stream that can be redirected to squid-tagger for imports: # cd /var/db/squidGuard ; path/to/sg_import.py | path/to/squid-tagger.py -l -f |
︙ | ︙ |
Modified squid-tagger.py
from [faa884292f]
to [568859367f].
︙ | ︙ | |||
34 35 36 37 38 39 40 | class Config: __slots__ = frozenset(['_config', '_default', '_section', 'options']) _default = { 'log': { 'silent': 'no', }, 'database': { | < | 34 35 36 37 38 39 40 41 42 43 44 45 46 47 | class Config: __slots__ = frozenset(['_config', '_default', '_section', 'options']) _default = { 'log': { 'silent': 'no', }, 'database': { 'database': 'squidTag', },} # function to read in config file def __init__(self): import ConfigParser, optparse, os |
︙ | ︙ | |||
96 97 98 99 100 101 102 | config = Config() import logging, logging.handlers # wrapper around logging handler to make it queue records and don't stall when sending them class SysLogHandlerQueue(logging.handlers.SysLogHandler): | | < < | | | | < < | > | | | | > | > < | | | | | | 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 | config = Config() import logging, logging.handlers # wrapper around logging handler to make it queue records and don't stall when sending them class SysLogHandlerQueue(logging.handlers.SysLogHandler): __slots__ = frozenset(['_running', '_tail', '_worker']) def __init__(self): logging.handlers.SysLogHandler.__init__(self, '/dev/log') self._tail = gevent.queue.Queue() self._worker = None def emit(self, record): # my syslog is broken and cannot into UTF-8 BOM record.msg = str(record.msg) self._tail.put(record) if self._worker == None: # in case queue is empty we will spawn new worker # all workers are logged so we can kill them on close() self._worker = gevent.spawn(self._writer) def _writer(self): # here we are locking the queue so we can be sure we are the only one print('syslog start') while not self._tail.empty(): logging.handlers.SysLogHandler.emit(self, self._tail.get()) self._worker = None print('syslog end') def close(self): if self._worker != None: gevent.kill(self._worker) logging.handlers.SysLogHandler.close(self) logger = logging.getLogger('squidTag') logger.setLevel(logging.INFO) handler = SysLogHandlerQueue() handler.setFormatter(logging.Formatter(str('squidTag[%(process)s]: %(message)s'))) logger.addHandler(handler) # tiny wrapper around a file to make reads from it geventable # or should i move this somewhere? class FReadlineQueue(gevent.queue.Queue): # storing file descriptor, leftover __slots__ = frozenset(['_io', '_fileno', '_tail']) def __init__(self, fd, closefd = True): import io # initialising class gevent.queue.Queue.__init__(self) # storing file descriptor self._fileno = fd.fileno() self._io = io.FileIO(self._fileno, 'r', closefd) # using empty tail self._tail = '' # setting up event self._install_wait() def _install_wait(self): # putting file to nonblocking mode fcntl.fcntl(self._fileno, fcntl.F_SETFL, fcntl.fcntl(self._fileno, fcntl.F_GETFL) | os.O_NONBLOCK) # installing event handler gevent.core.read_event(self._fileno, self._wait_helper) def _wait_helper(self, ev, evtype): # reading one buffer from stream buf = self._io.read(4096) # splitting stream by line ends rows = buf.decode('l1').split('\n') # adding tail to the first element if there is some tail if len(self._tail) > 0: rows[0] = self._tail + rows[0] # popping out last (incomplete) element self._tail = rows.pop(-1) # dropping all complete elements to the queue for row in rows: self.put_nowait(row) logger.info('< ' + row) if len(buf) > 0: # no EOF, reinstalling event handler gevent.core.read_event(self._fileno, self._wait_helper) else: # EOF found, sending EOF to queue self.put_nowait(None) stdin = FReadlineQueue(sys.stdin, False) # wrapper against file handler that makes possible to queue some writes without stalling class FWritelineQueue(gevent.queue.JoinableQueue): # storing fileno, io interface, leftover __slots__ = frozenset(['_fileno', '_io', '_tail']) |
︙ | ︙ |