2019-09-13 13:17:58 +00:00
|
|
|
try:
|
|
|
|
from BaseHTTPServer import BaseHTTPRequestHandler
|
|
|
|
except ImportError:
|
|
|
|
from http.server import BaseHTTPRequestHandler
|
|
|
|
|
|
|
|
try:
|
|
|
|
from BaseHTTPServer import HTTPServer
|
|
|
|
except ImportError:
|
|
|
|
from http.server import HTTPServer
|
|
|
|
|
|
|
|
try:
|
|
|
|
import urllib.parse as urlparse
|
|
|
|
except ImportError:
|
|
|
|
import urlparse
|
|
|
|
|
|
|
|
import json
|
|
|
|
import logging
|
|
|
|
import os
|
|
|
|
import socket
|
|
|
|
import sys
|
|
|
|
import threading
|
|
|
|
import time
|
|
|
|
|
|
|
|
|
|
|
|
logging.getLogger().setLevel(logging.INFO)
|
|
|
|
file_handler = logging.FileHandler('/var/log/clickhouse-server/test-server.log', 'a', encoding='utf-8')
|
|
|
|
file_handler.setFormatter(logging.Formatter('%(asctime)s %(message)s'))
|
|
|
|
logging.getLogger().addHandler(file_handler)
|
|
|
|
logging.getLogger().addHandler(logging.StreamHandler())
|
|
|
|
|
2019-09-14 07:44:46 +00:00
|
|
|
communication_port = int(sys.argv[1])
|
2019-09-13 13:53:17 +00:00
|
|
|
bucket = sys.argv[2]
|
2019-09-13 13:17:58 +00:00
|
|
|
|
|
|
|
def GetFreeTCPPortsAndIP(n):
|
|
|
|
result = []
|
|
|
|
sockets = []
|
|
|
|
for i in range(n):
|
|
|
|
tcp = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
|
|
|
|
tcp.bind((socket.gethostname(), 0))
|
|
|
|
addr, port = tcp.getsockname()
|
|
|
|
result.append(port)
|
|
|
|
sockets.append(tcp)
|
|
|
|
[ s.close() for s in sockets ]
|
|
|
|
return result, addr
|
|
|
|
|
2019-09-14 07:44:46 +00:00
|
|
|
(redirecting_to_http_port, simple_server_port, preserving_data_port, redirecting_preserving_data_port), localhost = GetFreeTCPPortsAndIP(4)
|
2019-09-13 13:17:58 +00:00
|
|
|
data = {
|
|
|
|
'redirecting_to_http_port': redirecting_to_http_port,
|
|
|
|
'preserving_data_port': preserving_data_port,
|
|
|
|
'redirecting_preserving_data_port': redirecting_preserving_data_port,
|
|
|
|
}
|
|
|
|
redirecting_host = localhost
|
|
|
|
|
|
|
|
|
2019-09-14 07:44:46 +00:00
|
|
|
class SimpleHTTPServerHandler(BaseHTTPRequestHandler):
|
2019-09-13 13:17:58 +00:00
|
|
|
def do_GET(self):
|
2019-09-14 07:44:46 +00:00
|
|
|
logging.info('GET {}'.format(self.path))
|
|
|
|
if self.path == '/milovidov/test.csv':
|
|
|
|
self.send_response(200)
|
|
|
|
self.send_header('Content-type', 'text/plain')
|
|
|
|
self.end_headers()
|
|
|
|
self.wfile.write('42,87,44\n55,33,81\n1,0,9\n')
|
|
|
|
else:
|
|
|
|
self.send_response(404)
|
|
|
|
self.end_headers()
|
2019-09-13 13:17:58 +00:00
|
|
|
self.finish()
|
|
|
|
|
|
|
|
|
2019-09-14 07:44:46 +00:00
|
|
|
class RedirectingToHTTPHandler(BaseHTTPRequestHandler):
|
2019-09-13 13:17:58 +00:00
|
|
|
def do_GET(self):
|
|
|
|
self.send_response(307)
|
|
|
|
self.send_header('Content-type', 'text/xml')
|
2019-09-14 07:44:46 +00:00
|
|
|
self.send_header('Location', 'http://{}:{}/milovidov/test.csv'.format(localhost, simple_server_port))
|
2019-09-13 13:17:58 +00:00
|
|
|
self.end_headers()
|
|
|
|
self.wfile.write(r'''<?xml version="1.0" encoding="UTF-8"?>
|
|
|
|
<Error>
|
|
|
|
<Code>TemporaryRedirect</Code>
|
|
|
|
<Message>Please re-send this request to the specified temporary endpoint.
|
|
|
|
Continue to use the original request endpoint for future requests.</Message>
|
|
|
|
<Endpoint>storage.yandexcloud.net</Endpoint>
|
|
|
|
</Error>'''.encode())
|
|
|
|
self.finish()
|
|
|
|
|
|
|
|
|
|
|
|
class PreservingDataHandler(BaseHTTPRequestHandler):
|
|
|
|
protocol_version = 'HTTP/1.1'
|
|
|
|
|
|
|
|
def parse_request(self):
|
|
|
|
result = BaseHTTPRequestHandler.parse_request(self)
|
|
|
|
# Adaptation to Python 3.
|
|
|
|
if sys.version_info.major == 2 and result == True:
|
|
|
|
expect = self.headers.get('Expect', "")
|
|
|
|
if (expect.lower() == "100-continue" and self.protocol_version >= "HTTP/1.1" and self.request_version >= "HTTP/1.1"):
|
|
|
|
if not self.handle_expect_100():
|
|
|
|
return False
|
|
|
|
return result
|
|
|
|
|
|
|
|
def send_response_only(self, code, message=None):
|
|
|
|
if message is None:
|
|
|
|
if code in self.responses:
|
|
|
|
message = self.responses[code][0]
|
|
|
|
else:
|
|
|
|
message = ''
|
|
|
|
if self.request_version != 'HTTP/0.9':
|
|
|
|
self.wfile.write("%s %d %s\r\n" % (self.protocol_version, code, message))
|
|
|
|
|
|
|
|
def handle_expect_100(self):
|
|
|
|
logging.info('Received Expect-100')
|
|
|
|
self.send_response_only(100)
|
|
|
|
self.end_headers()
|
|
|
|
return True
|
|
|
|
|
|
|
|
def do_POST(self):
|
|
|
|
self.send_response(200)
|
|
|
|
query = urlparse.urlparse(self.path).query
|
|
|
|
logging.info('POST ' + query)
|
|
|
|
if query == 'uploads':
|
|
|
|
post_data = r'''<?xml version="1.0" encoding="UTF-8"?>
|
|
|
|
<hi><UploadId>TEST</UploadId></hi>'''.encode()
|
|
|
|
self.send_header('Content-length', str(len(post_data)))
|
|
|
|
self.send_header('Content-type', 'text/plain')
|
|
|
|
self.end_headers()
|
|
|
|
self.wfile.write(post_data)
|
|
|
|
else:
|
|
|
|
post_data = self.rfile.read(int(self.headers.get('Content-Length')))
|
|
|
|
self.send_header('Content-type', 'text/plain')
|
|
|
|
self.end_headers()
|
|
|
|
data['received_data_completed'] = True
|
|
|
|
data['finalize_data'] = post_data
|
|
|
|
data['finalize_data_query'] = query
|
|
|
|
self.finish()
|
|
|
|
|
|
|
|
def do_PUT(self):
|
|
|
|
self.send_response(200)
|
|
|
|
self.send_header('Content-type', 'text/plain')
|
|
|
|
self.send_header('ETag', 'hello-etag')
|
|
|
|
self.end_headers()
|
|
|
|
query = urlparse.urlparse(self.path).query
|
|
|
|
path = urlparse.urlparse(self.path).path
|
|
|
|
logging.info('Content-Length = ' + self.headers.get('Content-Length'))
|
|
|
|
logging.info('PUT ' + query)
|
|
|
|
assert self.headers.get('Content-Length')
|
|
|
|
assert self.headers['Expect'] == '100-continue'
|
|
|
|
put_data = self.rfile.read()
|
|
|
|
data.setdefault('received_data', []).append(put_data)
|
|
|
|
logging.info('PUT to {}'.format(path))
|
|
|
|
self.server.storage[path] = put_data
|
|
|
|
self.finish()
|
|
|
|
|
|
|
|
def do_GET(self):
|
|
|
|
path = urlparse.urlparse(self.path).path
|
|
|
|
if path in self.server.storage:
|
|
|
|
self.send_response(200)
|
|
|
|
self.send_header('Content-type', 'text/plain')
|
|
|
|
self.send_header('Content-length', str(len(self.server.storage[path])))
|
|
|
|
self.end_headers()
|
|
|
|
self.wfile.write(self.server.storage[path])
|
|
|
|
else:
|
|
|
|
self.send_response(404)
|
|
|
|
self.end_headers()
|
|
|
|
self.finish()
|
|
|
|
|
|
|
|
|
|
|
|
class RedirectingPreservingDataHandler(BaseHTTPRequestHandler):
|
|
|
|
protocol_version = 'HTTP/1.1'
|
|
|
|
|
|
|
|
def parse_request(self):
|
|
|
|
result = BaseHTTPRequestHandler.parse_request(self)
|
|
|
|
# Adaptation to Python 3.
|
|
|
|
if sys.version_info.major == 2 and result == True:
|
|
|
|
expect = self.headers.get('Expect', "")
|
|
|
|
if (expect.lower() == "100-continue" and self.protocol_version >= "HTTP/1.1" and self.request_version >= "HTTP/1.1"):
|
|
|
|
if not self.handle_expect_100():
|
|
|
|
return False
|
|
|
|
return result
|
|
|
|
|
|
|
|
def send_response_only(self, code, message=None):
|
|
|
|
if message is None:
|
|
|
|
if code in self.responses:
|
|
|
|
message = self.responses[code][0]
|
|
|
|
else:
|
|
|
|
message = ''
|
|
|
|
if self.request_version != 'HTTP/0.9':
|
|
|
|
self.wfile.write("%s %d %s\r\n" % (self.protocol_version, code, message))
|
|
|
|
|
|
|
|
def handle_expect_100(self):
|
|
|
|
logging.info('Received Expect-100')
|
|
|
|
return True
|
|
|
|
|
|
|
|
def do_POST(self):
|
|
|
|
query = urlparse.urlparse(self.path).query
|
|
|
|
if query:
|
|
|
|
query = '?{}'.format(query)
|
|
|
|
self.send_response(307)
|
|
|
|
self.send_header('Content-type', 'text/xml')
|
|
|
|
self.send_header('Location', 'http://{host}:{port}/{bucket}/test.csv{query}'.format(host=localhost, port=preserving_data_port, bucket=bucket, query=query))
|
|
|
|
self.end_headers()
|
|
|
|
self.wfile.write(r'''<?xml version="1.0" encoding="UTF-8"?>
|
|
|
|
<Error>
|
|
|
|
<Code>TemporaryRedirect</Code>
|
|
|
|
<Message>Please re-send this request to the specified temporary endpoint.
|
|
|
|
Continue to use the original request endpoint for future requests.</Message>
|
|
|
|
<Endpoint>{host}:{port}</Endpoint>
|
|
|
|
</Error>'''.format(host=localhost, port=preserving_data_port).encode())
|
|
|
|
self.finish()
|
|
|
|
|
|
|
|
def do_PUT(self):
|
|
|
|
query = urlparse.urlparse(self.path).query
|
|
|
|
if query:
|
|
|
|
query = '?{}'.format(query)
|
|
|
|
self.send_response(307)
|
|
|
|
self.send_header('Content-type', 'text/xml')
|
|
|
|
self.send_header('Location', 'http://{host}:{port}/{bucket}/test.csv{query}'.format(host=localhost, port=preserving_data_port, bucket=bucket, query=query))
|
|
|
|
self.end_headers()
|
|
|
|
self.wfile.write(r'''<?xml version="1.0" encoding="UTF-8"?>
|
|
|
|
<Error>
|
|
|
|
<Code>TemporaryRedirect</Code>
|
|
|
|
<Message>Please re-send this request to the specified temporary endpoint.
|
|
|
|
Continue to use the original request endpoint for future requests.</Message>
|
|
|
|
<Endpoint>{host}:{port}</Endpoint>
|
|
|
|
</Error>'''.format(host=localhost, port=preserving_data_port).encode())
|
|
|
|
self.finish()
|
|
|
|
|
|
|
|
|
2019-09-14 07:44:46 +00:00
|
|
|
class CommunicationServerHandler(BaseHTTPRequestHandler):
|
|
|
|
def do_GET(self):
|
|
|
|
self.send_response(200)
|
|
|
|
self.end_headers()
|
|
|
|
self.wfile.write(json.dumps(data))
|
|
|
|
self.finish()
|
|
|
|
|
|
|
|
|
2019-09-13 13:17:58 +00:00
|
|
|
servers = []
|
2019-09-14 07:44:46 +00:00
|
|
|
servers.append(HTTPServer((localhost, communication_port), CommunicationServerHandler))
|
|
|
|
servers.append(HTTPServer((localhost, redirecting_to_http_port), RedirectingToHTTPHandler))
|
|
|
|
servers.append(HTTPServer((localhost, preserving_data_port), PreservingDataHandler))
|
2019-09-13 13:17:58 +00:00
|
|
|
servers[-1].storage = {}
|
2019-09-14 07:44:46 +00:00
|
|
|
servers.append(HTTPServer((localhost, simple_server_port), SimpleHTTPServerHandler))
|
|
|
|
servers.append(HTTPServer((localhost, redirecting_preserving_data_port), RedirectingPreservingDataHandler))
|
2019-09-13 13:17:58 +00:00
|
|
|
jobs = [ threading.Thread(target=server.serve_forever) for server in servers ]
|
|
|
|
[ job.start() for job in jobs ]
|
|
|
|
|
|
|
|
time.sleep(60) # Timeout
|
|
|
|
|
|
|
|
logging.info('Shutting down')
|
|
|
|
[ server.shutdown() for server in servers ]
|
|
|
|
logging.info('Joining threads')
|
|
|
|
[ job.join() for job in jobs ]
|
|
|
|
logging.info('Done')
|