1 | #!/usr/bin/env python
|
---|
2 | """
|
---|
3 | # Multiplexer for HTTP streams, used for broadcast of MPJEG and WAV streams
|
---|
4 | #
|
---|
5 | # Licence: BSDLike
|
---|
6 | # URL: http://rickvanderzwet.nl/svn/personal/py-tcpmultiplexer/
|
---|
7 | #
|
---|
8 | # Rick van der Zwet <info@rickvanderzwet.nl>
|
---|
9 | """
|
---|
10 |
|
---|
11 | import SocketServer
|
---|
12 | import argparse
|
---|
13 | import logging
|
---|
14 | import logging.handlers
|
---|
15 | import os
|
---|
16 | import socket
|
---|
17 | import sys
|
---|
18 | import threading
|
---|
19 | import time
|
---|
20 | import urllib2
|
---|
21 |
|
---|
22 | # Some boring defaults
|
---|
23 | DEFAULT_HOST = '0.0.0.0'
|
---|
24 | DEFAULT_PORT = 9999
|
---|
25 | DEFAULT_CONFIG = 'streams.yaml'
|
---|
26 | DEFAULT_DOCUMENTROOT = './htdocs'
|
---|
27 | DEFAULT_TIMEOUT = None
|
---|
28 | DEFAULT_LOGFILE = 'py-tcpmultiplexer.log'
|
---|
29 |
|
---|
30 |
|
---|
31 | # URL : TARGET
|
---|
32 | DEFAULT_STREAMS = {
|
---|
33 | # '/cam1/video' : 'http://172.16.0.67:8080/videofeed',
|
---|
34 | # '/cam1/audio' : 'http://172.16.0.67:8080/audio.wav',
|
---|
35 | }
|
---|
36 |
|
---|
37 |
|
---|
38 | # Global variables used as ring-buffers or shared-storage
|
---|
39 | urlheader = dict()
|
---|
40 | dataheader = dict()
|
---|
41 | urldata = dict()
|
---|
42 | stream_running = dict()
|
---|
43 | running = True
|
---|
44 | recv_threads = []
|
---|
45 | document_root = None
|
---|
46 | recv_timeout = None
|
---|
47 |
|
---|
48 | logger = logging.getLogger(__name__)
|
---|
49 | logger.setLevel(logging.INFO)
|
---|
50 | formatter = logging.Formatter('%(asctime)s - %(threadName)s - %(levelname)s - %(message)s')
|
---|
51 |
|
---|
52 | ch = logging.StreamHandler()
|
---|
53 | ch.setFormatter(formatter)
|
---|
54 |
|
---|
55 | logger.addHandler(ch)
|
---|
56 |
|
---|
57 | class ThreadedTCPServer(SocketServer.ThreadingMixIn, SocketServer.TCPServer):
|
---|
58 | pass
|
---|
59 |
|
---|
60 |
|
---|
61 | class MyTCPHandler(SocketServer.BaseRequestHandler):
|
---|
62 | """
|
---|
63 | The RequestHandler class for our server.
|
---|
64 |
|
---|
65 | It is instantiated once per connection to the server, and must
|
---|
66 | override the handle() method to implement communication to the
|
---|
67 | client.
|
---|
68 | """
|
---|
69 |
|
---|
70 | def handle(self):
|
---|
71 | global running, dataheader, document_root, stream_running, urlheader, urldata
|
---|
72 | # self.request is the TCP socket connected to the client
|
---|
73 | self.data = self.request.recv(1024).strip()
|
---|
74 | urlid = self.data.split('\n')[0].split()[1]
|
---|
75 | req_type = self.data.split('\n')[0].split()[0]
|
---|
76 | # XXX: Check for .. paths
|
---|
77 | document_file = document_root + urlid
|
---|
78 | logging.info("Connection %s from '%s' for '%s'", req_type, self.client_address[0], urlid)
|
---|
79 | try:
|
---|
80 | if urlid == '/':
|
---|
81 | self.request.send("HTTP/1.1 200 OK\n\n<html><head><title>Overview</title></head><body><h3>Streams</h3><ul>")
|
---|
82 | for urlid in filter(lambda x: stream_running[x], stream_running.keys()):
|
---|
83 | self.request.send('<li><a href="%(urlid)s">%(urlid)s</a></li>' % {'urlid': urlid})
|
---|
84 | self.request.send("</ul><h3>Files</h3><ul>")
|
---|
85 | for root, dirs, files in os.walk(document_root):
|
---|
86 | # Please do not show any hidden files or directories
|
---|
87 | [dirs.remove(name) for name in dirs if name.startswith('.')]
|
---|
88 | [files.remove(name) for name in files if name.startswith('.')]
|
---|
89 | for name in sorted(files):
|
---|
90 | self.request.send('<li><a href="%(item)s">%(item)s</li>' % { 'item' : os.path.join(root.replace(document_root,'',1),name) })
|
---|
91 | self.request.send("</ul></body></html>")
|
---|
92 | elif urldata.has_key(urlid):
|
---|
93 | self.request.send('HTTP/1.1 200 OK\n' + urlheader[urlid] + 'X-Proxy-Server: TCPMultiplexer.py\n' + '\n' + dataheader[urlid])
|
---|
94 | if req_type == 'HEAD':
|
---|
95 | return
|
---|
96 | elif req_type == 'GET':
|
---|
97 | urldata[urlid][self] = []
|
---|
98 | while running:
|
---|
99 | if len(urldata[urlid][self]) == 0:
|
---|
100 | time.sleep(0.1)
|
---|
101 | continue
|
---|
102 | self.request.send(urldata[urlid][self].pop(0))
|
---|
103 | elif os.path.isfile(document_file):
|
---|
104 | data = open(document_file,'r').read()
|
---|
105 | self.request.send('HTTP/1.1 200 OK\nContent-Length: %s\n\n%s' % (len(data),data))
|
---|
106 | else:
|
---|
107 | self.request.send("HTTP/1.1 404 NOT FOUND\nContent-Type: text/html\n\n<h1>404</h1>The page '%s' does not exists" % urlid)
|
---|
108 | except IOError:
|
---|
109 | logging.info("Connection closed from '%s'", self.client_address[0])
|
---|
110 | del urldata[urlid][self]
|
---|
111 |
|
---|
112 |
|
---|
113 |
|
---|
114 | def get_data(url, urlid, *args):
|
---|
115 | """ Fetch the DATA from the WAV or MJPEG Stream """
|
---|
116 | global running, dataheader, urlheader, urldata, stream_running, recv_timeout
|
---|
117 | # Fill buffers if needed
|
---|
118 | logger.info("Starting thread '%s' (%s)", url, urlid)
|
---|
119 | while running:
|
---|
120 | stream_running[urlid] = False
|
---|
121 | urlheader[urlid] = None
|
---|
122 | dataheader[urlid] = None
|
---|
123 | urldata[urlid] = None
|
---|
124 | try:
|
---|
125 | f = urllib2.urlopen(url,timeout=recv_timeout)
|
---|
126 | urlheader[urlid] = ''.join(f.info().headers)
|
---|
127 | urldata[urlid] = dict()
|
---|
128 |
|
---|
129 | # Find datatype
|
---|
130 | datatype = None
|
---|
131 | urltype = f.info().gettype()
|
---|
132 | logger.info("url.info().gettype() is '%s'", urltype)
|
---|
133 | if urltype == 'audio/x-wav':
|
---|
134 | datatype = 'wav'
|
---|
135 | elif urltype == 'multipart/x-mixed-replace':
|
---|
136 | datatype = 'mjpeg'
|
---|
137 |
|
---|
138 | # Be verbose to the user
|
---|
139 | if datatype:
|
---|
140 | logger.info("'%s' Identified as %s", url, datatype)
|
---|
141 | else:
|
---|
142 | logger.warning("Cannot find file type of '%s'", url)
|
---|
143 |
|
---|
144 | # Set the correct data header
|
---|
145 | if datatype == 'wav':
|
---|
146 | # WAV header
|
---|
147 | dataheader[urlid] = f.read(44)
|
---|
148 | elif datatype == 'mjpeg':
|
---|
149 | data = f.read(1024)
|
---|
150 |
|
---|
151 | # Get the required headers and acurate datasize
|
---|
152 | headers = []
|
---|
153 | datasize = 1024 * 1024
|
---|
154 | for header in data.splitlines():
|
---|
155 | if not header.strip():
|
---|
156 | # Newlines in the beginning are evil
|
---|
157 | if headers:
|
---|
158 | break
|
---|
159 | else:
|
---|
160 | continue
|
---|
161 | if header.startswith('DataLen:') or header.startswith('Content-length:'):
|
---|
162 | datasize = int(header.split(':')[1])
|
---|
163 | headers.append(header)
|
---|
164 | boundry = headers[0]
|
---|
165 |
|
---|
166 | logger.info("Data Length: %s", datasize)
|
---|
167 | logger.info("Boundry line: %s", boundry)
|
---|
168 | logger.info("Image headers %s", headers)
|
---|
169 |
|
---|
170 | data = data + f.read(datasize * 2)
|
---|
171 | valid_image = boundry + data.split(boundry)[1]
|
---|
172 | dataheader[urlid] = valid_image + '\n'.join(headers) + '\n'
|
---|
173 | else:
|
---|
174 | dataheader[urlid] = ''
|
---|
175 | logger.info("Using dataheader of length %s", len(dataheader[urlid]))
|
---|
176 |
|
---|
177 | # Main data loader
|
---|
178 | logger.info("Stream ready to serve")
|
---|
179 | stream_running[urlid] = True
|
---|
180 | recv_buffer_size = 1024 * 8
|
---|
181 | while running:
|
---|
182 | data = f.read(recv_buffer_size)
|
---|
183 | if not len(data) == recv_buffer_size:
|
---|
184 | raise IOError("Connection corrupted, got '%s' instead of '%s'" % (len(data), recv_buffer_size))
|
---|
185 | logger.debug("Received data chunk with length: %s", len(data))
|
---|
186 | for key in urldata[urlid].keys():
|
---|
187 | urldata[urlid][key].append(data)
|
---|
188 | except (urllib2.URLError, IOError) as e:
|
---|
189 | #Enforce a connection reset
|
---|
190 | logger.warning("URL reset '%s' (%s)", url, e)
|
---|
191 | stream_running[urlid] = False
|
---|
192 | time.sleep(1)
|
---|
193 | pass
|
---|
194 | logger.info("Closing Thread '%s'", url)
|
---|
195 |
|
---|
196 |
|
---|
197 | if __name__ == "__main__":
|
---|
198 | parser = argparse.ArgumentParser(description=__doc__,formatter_class=argparse.RawDescriptionHelpFormatter)
|
---|
199 | parser.add_argument('--host', dest='host', default=DEFAULT_HOST, help='Listen to IP [default: %s]' % DEFAULT_HOST)
|
---|
200 | parser.add_argument('--port', dest='port', default=DEFAULT_PORT, type=int, help='Listen to PORT [default: %s]' % DEFAULT_PORT)
|
---|
201 | parser.add_argument('--logfile', dest='logfile', default=DEFAULT_LOGFILE, help='File to write logfiles to [default: %s]' % DEFAULT_LOGFILE)
|
---|
202 | parser.add_argument('--timeout', dest='timeout', default=DEFAULT_TIMEOUT, type=int, help='Default socket timeout [default: %s]' % DEFAULT_TIMEOUT)
|
---|
203 | parser.add_argument('--stream-cfg', dest='stream_cfg', default=DEFAULT_CONFIG, help='YAML Stream configuration [default: %s]' % DEFAULT_CONFIG)
|
---|
204 | parser.add_argument('--document-root', dest='document_root', default=DEFAULT_DOCUMENTROOT, help='Document Root File Directory [default: %s]' % DEFAULT_DOCUMENTROOT)
|
---|
205 | args = parser.parse_args()
|
---|
206 |
|
---|
207 | # Add file logger
|
---|
208 | ch = logging.handlers.WatchedFileHandler(args.logfile)
|
---|
209 | ch.setFormatter(formatter)
|
---|
210 | logger.addHandler(ch)
|
---|
211 |
|
---|
212 | document_root = os.path.abspath(os.path.join(os.path.dirname(__file__),args.document_root))
|
---|
213 | logger.info("Serving '/htdocs' from document_root '%s'", document_root)
|
---|
214 |
|
---|
215 | # Set the timeout
|
---|
216 | logger.info("Changing socket timeout from '%s' to '%s'", socket.getdefaulttimeout(), args.timeout)
|
---|
217 | recv_timeout = args.timeout
|
---|
218 |
|
---|
219 |
|
---|
220 | # Inport streams
|
---|
221 | streams = DEFAULT_STREAMS
|
---|
222 | try:
|
---|
223 | import yaml
|
---|
224 | streams.update(yaml.load(open(args.stream_cfg)))
|
---|
225 | except (ImportError, IOError) as e:
|
---|
226 | logger.warning("Stream config file '%s' not readable or parsable (%s)", args.stream_cfg, e)
|
---|
227 |
|
---|
228 |
|
---|
229 | # Create the base server
|
---|
230 | try:
|
---|
231 | while True:
|
---|
232 | try:
|
---|
233 | ThreadedTCPServer.allow_reuse_address = True
|
---|
234 | server = ThreadedTCPServer((args.host, args.port), MyTCPHandler)
|
---|
235 | break
|
---|
236 | except IOError, e:
|
---|
237 | logger.warning('For conection %s:%s to become available (%s)', args.host, args.port , e)
|
---|
238 | time.sleep(1)
|
---|
239 | except KeyboardInterrupt:
|
---|
240 | sys.exit(1)
|
---|
241 |
|
---|
242 | for urlid, url in streams.iteritems():
|
---|
243 | recv_threads.append(threading.Thread(name=urlid, target=get_data, args=(url,urlid)))
|
---|
244 | recv_threads[-1].setDaemon(True)
|
---|
245 | recv_threads[-1].start()
|
---|
246 |
|
---|
247 | # Activate the server; this will keep running until you
|
---|
248 | # interrupt the program with Ctrl-C
|
---|
249 | try:
|
---|
250 | logger.info('Serving at %s:%s', args.host, args.port)
|
---|
251 | server.serve_forever()
|
---|
252 | except KeyboardInterrupt, IOError:
|
---|
253 | logger.info('Shutting down, please wait...')
|
---|
254 | running = False
|
---|
255 | server.shutdown()
|
---|
256 | [thread.join() for thread in recv_threads]
|
---|
257 | logger.info('All done, good bye!')
|
---|
258 |
|
---|