1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584
|
# Copyright (C) 2007 Giampaolo Rodola' <g.rodola@gmail.com>.
# Use of this source code is governed by MIT license that can be
# found in the LICENSE file.
"""
This module contains the main FTPServer class which listens on a
host:port and dispatches the incoming connections to a handler.
The concurrency is handled asynchronously by the main process thread,
meaning the handler cannot block otherwise the whole server will hang.
Other than that we have 2 subclasses changing the asynchronous concurrency
model using multiple threads or processes.
You might be interested in these in case your code contains blocking
parts which cannot be adapted to the base async model or if the
underlying filesystem is particularly slow, see:
https://github.com/giampaolo/pyftpdlib/issues/197
https://github.com/giampaolo/pyftpdlib/issues/212
Two classes are provided:
- ThreadingFTPServer
- MultiprocessFTPServer
...spawning a new thread or process every time a client connects.
The main thread will be async-based and be used only to accept new
connections.
Every time a new connection comes in that will be dispatched to a
separate thread/process which internally will run its own IO loop.
This way the handler handling that connections will be free to block
without hanging the whole FTP server.
"""
import errno
import os
import select
import signal
import sys
import threading
import time
import traceback
from .ioloop import Acceptor
from .log import PREFIX
from .log import PREFIX_MPROC
from .log import config_logging
from .log import debug
from .log import is_logging_configured
from .log import logger
from .prefork import fork_processes
__all__ = ['FTPServer', 'ThreadedFTPServer']
_BSD = 'bsd' in sys.platform
# ===================================================================
# --- base class
# ===================================================================
class FTPServer(Acceptor):
"""Creates a socket listening on <address>, dispatching the requests
to a <handler> (typically FTPHandler class).
Depending on the type of address specified IPv4 or IPv6 connections
(or both, depending from the underlying system) will be accepted.
All relevant session information is stored in class attributes
described below.
- (int) max_cons:
number of maximum simultaneous connections accepted (defaults
to 512). Can be set to 0 for unlimited but it is recommended
to always have a limit to avoid running out of file descriptors
(DoS).
- (int) max_cons_per_ip:
number of maximum connections accepted for the same IP address
(defaults to 0 == unlimited).
"""
max_cons = 512
max_cons_per_ip = 0
def __init__(self, address_or_socket, handler, ioloop=None, backlog=100):
"""Creates a socket listening on 'address' dispatching
connections to a 'handler'.
- (tuple) address_or_socket: the (host, port) pair on which
the command channel will listen for incoming connections or
an existent socket object.
- (instance) handler: the handler class to use.
- (instance) ioloop: a pyftpdlib.ioloop.IOLoop instance
- (int) backlog: the maximum number of queued connections
passed to listen(). If a connection request arrives when
the queue is full the client may raise ECONNRESET.
Defaults to 5.
"""
Acceptor.__init__(self, ioloop=ioloop)
self.handler = handler
self.backlog = backlog
self.ip_map = []
# in case of FTPS class not properly configured we want errors
# to be raised here rather than later, when client connects
if hasattr(handler, 'get_ssl_context'):
handler.get_ssl_context()
if callable(getattr(address_or_socket, 'listen', None)):
sock = address_or_socket
sock.setblocking(0)
self.set_socket(sock)
else:
self.bind_af_unspecified(address_or_socket)
self.listen(backlog)
def __enter__(self):
return self
def __exit__(self, *args):
self.close_all()
@property
def address(self):
"""The address this server is listening on as a (ip, port) tuple."""
return self.socket.getsockname()[:2]
def _map_len(self):
return len(self.ioloop.socket_map)
def _accept_new_cons(self):
"""Return True if the server is willing to accept new connections."""
if not self.max_cons:
return True
else:
return self._map_len() <= self.max_cons
def _log_start(self, prefork=False):
def get_fqname(obj):
try:
return obj.__module__ + "." + obj.__class__.__name__
except AttributeError:
try:
return obj.__module__ + "." + obj.__name__
except AttributeError:
return str(obj)
if not is_logging_configured():
# If we get to this point it means the user hasn't
# configured any logger. We want logging to be on
# by default (stderr).
config_logging(prefix=PREFIX_MPROC if prefork else PREFIX)
if self.handler.passive_ports:
pasv_ports = "%s->%s" % ( # noqa: UP031
self.handler.passive_ports[0],
self.handler.passive_ports[-1],
)
else:
pasv_ports = None
model = 'prefork + ' if prefork else ''
if 'ThreadedFTPServer' in __all__ and issubclass(
self.__class__, ThreadedFTPServer
):
model += 'multi-thread'
elif 'MultiprocessFTPServer' in __all__ and issubclass(
self.__class__, MultiprocessFTPServer
):
model += 'multi-process'
elif issubclass(self.__class__, FTPServer):
model += 'async'
else:
model += 'unknown (custom class)'
logger.info("concurrency model: " + model)
logger.info(
"masquerade (NAT) address: %s", self.handler.masquerade_address
)
logger.info("passive ports: %s", pasv_ports)
logger.debug("poller: %r", get_fqname(self.ioloop))
logger.debug("authorizer: %r", get_fqname(self.handler.authorizer))
if os.name == 'posix':
logger.debug("use sendfile(2): %s", self.handler.use_sendfile)
logger.debug("handler: %r", get_fqname(self.handler))
logger.debug("max connections: %s", self.max_cons or "unlimited")
logger.debug(
"max connections per ip: %s", self.max_cons_per_ip or "unlimited"
)
logger.debug("timeout: %s", self.handler.timeout or "unlimited")
logger.debug("banner: %r", self.handler.banner)
logger.debug("max login attempts: %r", self.handler.max_login_attempts)
if getattr(self.handler, 'certfile', None):
logger.debug("SSL certfile: %r", self.handler.certfile)
if getattr(self.handler, 'keyfile', None):
logger.debug("SSL keyfile: %r", self.handler.keyfile)
def serve_forever(
self, timeout=None, blocking=True, handle_exit=True, worker_processes=1
):
"""Start serving.
- (float) timeout: the timeout passed to the underlying IO
loop expressed in seconds.
- (bool) blocking: if False loop once and then return the
timeout of the next scheduled call next to expire soonest
(if any).
- (bool) handle_exit: when True catches KeyboardInterrupt and
SystemExit exceptions (generally caused by SIGTERM / SIGINT
signals) and gracefully exits after cleaning up resources.
Also, logs server start and stop.
- (int) worker_processes: pre-fork a certain number of child
processes before starting.
Each child process will keep using a 1-thread, async
concurrency model, handling multiple concurrent connections.
If the number is None or <= 0 the number of usable cores
available on this machine is detected and used.
It is a good idea to use this option in case the app risks
blocking for too long on a single function call (e.g.
hard-disk is slow, long DB query on auth etc.).
By splitting the work load over multiple processes the delay
introduced by a blocking function call is amortized and divided
by the number of worker processes.
"""
log = handle_exit and blocking
if worker_processes != 1 and os.name == 'posix':
if not blocking:
raise ValueError(
"'worker_processes' and 'blocking' are mutually exclusive"
)
if log:
self._log_start(prefork=True)
fork_processes(worker_processes)
elif log:
self._log_start()
proto = "FTP+SSL" if hasattr(self.handler, 'ssl_protocol') else "FTP"
logger.info(
">>> starting %s server on %s:%s, pid=%i <<<"
% (proto, self.address[0], self.address[1], os.getpid())
)
if handle_exit:
try:
self.ioloop.loop(timeout, blocking)
except (KeyboardInterrupt, SystemExit):
logger.info("received interrupt signal")
if blocking:
if log:
logger.info(
">>> shutting down FTP server, %s socket(s), pid=%i "
"<<<",
self._map_len(),
os.getpid(),
)
self.close_all()
else:
self.ioloop.loop(timeout, blocking)
def handle_accepted(self, sock, addr):
"""Called when remote client initiates a connection."""
handler = None
ip = None
try:
handler = self.handler(sock, self, ioloop=self.ioloop)
if not handler.connected:
return
ip = addr[0]
self.ip_map.append(ip)
# For performance and security reasons we should always set a
# limit for the number of file descriptors that socket_map
# should contain. When we're running out of such limit we'll
# use the last available channel for sending a 421 response
# to the client before disconnecting it.
if not self._accept_new_cons():
handler.handle_max_cons()
return
# accept only a limited number of connections from the same
# source address.
if self.max_cons_per_ip:
if self.ip_map.count(ip) > self.max_cons_per_ip:
handler.handle_max_cons_per_ip()
return
try:
handler.handle()
except Exception:
handler.handle_error()
else:
return handler
except Exception:
# This is supposed to be an application bug that should
# be fixed. We do not want to tear down the server though
# (DoS). We just log the exception, hoping that someone
# will eventually file a bug. References:
# - https://github.com/giampaolo/pyftpdlib/issues/143
# - https://github.com/giampaolo/pyftpdlib/issues/166
# - https://groups.google.com/forum/#!topic/pyftpdlib/h7pPybzAx14
logger.error(traceback.format_exc())
if handler is not None:
handler.close()
elif ip is not None and ip in self.ip_map:
self.ip_map.remove(ip)
def handle_error(self):
"""Called to handle any uncaught exceptions."""
try:
raise # noqa: PLE0704
except Exception:
logger.error(traceback.format_exc())
self.close()
def close_all(self):
"""Stop serving and also disconnects all currently connected
clients.
"""
return self.ioloop.close()
# ===================================================================
# --- extra implementations
# ===================================================================
class _SpawnerBase(FTPServer):
"""Base class shared by multiple threads/process dispatcher.
Not supposed to be used.
"""
# How many seconds to wait when join()ing parent's threads
# or processes.
join_timeout = 5
# How often thread/process finished tasks should be cleaned up.
refresh_interval = 5
_lock = None
_exit = None
def __init__(self, address_or_socket, handler, ioloop=None, backlog=100):
FTPServer.__init__(
self, address_or_socket, handler, ioloop=ioloop, backlog=backlog
)
self._active_tasks = []
self._active_tasks_idler = self.ioloop.call_every(
self.refresh_interval,
self._refresh_tasks,
_errback=self.handle_error,
)
def _start_task(self, *args, **kwargs):
raise NotImplementedError('must be implemented in subclass')
def _map_len(self):
if len(self._active_tasks) >= self.max_cons:
# Since refresh()ing is a potentially expensive operation
# (O(N)) do it only if we're exceeding max connections
# limit. Other than in here, tasks are refreshed every 10
# seconds anyway.
self._refresh_tasks()
return len(self._active_tasks)
def _refresh_tasks(self):
"""join() terminated tasks and update internal _tasks list.
This gets called every X secs.
"""
if self._active_tasks:
logger.debug(
f"refreshing tasks ({len(self._active_tasks)} join()"
" potentials)"
)
with self._lock:
new = []
for t in self._active_tasks:
if not t.is_alive():
self._join_task(t)
else:
new.append(t)
self._active_tasks = new
def _loop(self, handler):
"""Serve handler's IO loop in a separate thread or process."""
with self.ioloop.factory() as ioloop:
handler.ioloop = ioloop
try:
handler.add_channel()
except OSError as err:
if err.errno == errno.EBADF:
# we might get here in case the other end quickly
# disconnected (see test_quick_connect())
debug(
"call: %s._loop(); add_channel() returned EBADF", self
)
return
else:
raise
# Here we localize variable access to minimize overhead.
poll = ioloop.poll
sched_poll = ioloop.sched.poll
poll_timeout = getattr(self, 'poll_timeout', None)
soonest_timeout = poll_timeout
while (
ioloop.socket_map or ioloop.sched._tasks
) and not self._exit.is_set():
try:
if ioloop.socket_map:
poll(timeout=soonest_timeout)
if ioloop.sched._tasks:
soonest_timeout = sched_poll()
# Handle the case where socket_map is empty but some
# cancelled scheduled calls are still around causing
# this while loop to hog CPU resources.
# In theory this should never happen as all the sched
# functions are supposed to be cancel()ed on close()
# but by using threads we can incur into
# synchronization issues such as this one.
# https://github.com/giampaolo/pyftpdlib/issues/245
if not ioloop.socket_map:
# get rid of cancel()led calls
ioloop.sched.reheapify()
soonest_timeout = sched_poll()
if soonest_timeout:
time.sleep(min(soonest_timeout, 1))
else:
soonest_timeout = None
except (KeyboardInterrupt, SystemExit):
# note: these two exceptions are raised in all sub
# processes
self._exit.set()
except OSError as err:
# on Windows we can get WSAENOTSOCK if the client
# rapidly connect and disconnects
if os.name == 'nt' and err.winerror == 10038:
for fd in list(ioloop.socket_map.keys()):
try:
select.select([fd], [], [], 0)
except OSError:
try:
logger.info(
"discarding broken socket %r",
ioloop.socket_map[fd],
)
del ioloop.socket_map[fd]
except KeyError:
# dict changed during iteration
pass
else:
raise
else:
if poll_timeout:
if (
soonest_timeout is None
or soonest_timeout > poll_timeout
):
soonest_timeout = poll_timeout
def handle_accepted(self, sock, addr):
handler = FTPServer.handle_accepted(self, sock, addr)
if handler is not None:
# unregister the handler from the main IOLoop used by the
# main thread to accept connections
self.ioloop.unregister(handler._fileno)
t = self._start_task(
target=self._loop, args=(handler,), name='ftpd'
)
t.name = repr(addr)
t.start()
# it is a different process so free resources here
if hasattr(t, 'pid'):
handler.close()
with self._lock:
# add the new task
self._active_tasks.append(t)
def _log_start(self):
FTPServer._log_start(self)
def serve_forever(self, timeout=1.0, blocking=True, handle_exit=True):
self._exit.clear()
if handle_exit:
log = handle_exit and blocking
if log:
self._log_start()
try:
self.ioloop.loop(timeout, blocking)
except (KeyboardInterrupt, SystemExit):
pass
if blocking:
if log:
logger.info(
">>> shutting down FTP server (%s active workers) <<<",
self._map_len(),
)
self.close_all()
else:
self.ioloop.loop(timeout, blocking)
def _terminate_task(self, t):
if hasattr(t, 'terminate'):
logger.debug(f"terminate()ing task {t!r}")
try:
if not _BSD:
t.terminate()
else:
# XXX - On FreeBSD using SIGTERM doesn't work
# as the process hangs on kqueue.control() or
# select.select(). Use SIGKILL instead.
os.kill(t.pid, signal.SIGKILL)
except ProcessLookupError:
pass
def _join_task(self, t):
logger.debug(f"join()ing task {t!r}")
t.join(self.join_timeout)
if t.is_alive():
logger.warning(
"task %r remained alive after %r secs", t, self.join_timeout
)
def close_all(self):
self._active_tasks_idler.cancel()
# this must be set after getting active tasks as it causes
# thread objects to get out of the list too soon
self._exit.set()
with self._lock:
for t in self._active_tasks:
self._terminate_task(t)
for t in self._active_tasks:
self._join_task(t)
del self._active_tasks[:]
FTPServer.close_all(self)
class ThreadedFTPServer(_SpawnerBase):
"""A modified version of base FTPServer class which spawns a
thread every time a new connection is established.
"""
# The timeout passed to thread's IOLoop.poll() call on every
# loop. Necessary since threads ignore KeyboardInterrupt.
poll_timeout = 1.0
_lock = threading.Lock()
_exit = threading.Event()
def _start_task(self, *args, **kwargs):
return threading.Thread(*args, **kwargs)
if os.name == 'posix':
try:
import multiprocessing
multiprocessing.Lock()
except Exception: # noqa
# see https://github.com/giampaolo/pyftpdlib/issues/496
pass
else:
__all__ += ['MultiprocessFTPServer']
class MultiprocessFTPServer(_SpawnerBase):
"""A modified version of base FTPServer class which spawns a
process every time a new connection is established.
"""
_lock = multiprocessing.Lock()
_exit = multiprocessing.Event()
def _start_task(self, *args, **kwargs):
return multiprocessing.Process(*args, **kwargs)
|