173 lines
4.7 KiB
Python
173 lines
4.7 KiB
Python
import logging
|
|
import multiprocessing
|
|
import pickle
|
|
import random
|
|
import signal
|
|
import sys
|
|
from asyncio import sleep
|
|
|
|
import pytest
|
|
from sanic_testing.testing import HOST, PORT
|
|
|
|
from sanic import Blueprint, text
|
|
from sanic.compat import use_context
|
|
from sanic.log import logger
|
|
|
|
|
|
@pytest.mark.skipif(
|
|
not hasattr(signal, "SIGALRM"),
|
|
reason="SIGALRM is not implemented for this platform, we have to come "
|
|
"up with another timeout strategy to test these",
|
|
)
|
|
@pytest.mark.skipif(
|
|
sys.platform not in ("linux", "darwin"),
|
|
reason="This test requires fork context",
|
|
)
|
|
def test_multiprocessing(app):
|
|
"""Tests that the number of children we produce is correct"""
|
|
# Selects a number at random so we can spot check
|
|
num_workers = random.choice(range(2, multiprocessing.cpu_count() * 2 + 1))
|
|
process_list = set()
|
|
|
|
@app.after_server_start
|
|
async def shutdown(app):
|
|
await sleep(2.1)
|
|
app.stop()
|
|
|
|
def stop_on_alarm(*args):
|
|
for process in multiprocessing.active_children():
|
|
process_list.add(process.pid)
|
|
|
|
signal.signal(signal.SIGALRM, stop_on_alarm)
|
|
signal.alarm(2)
|
|
with use_context("fork"):
|
|
app.run(HOST, 4120, workers=num_workers, debug=True)
|
|
|
|
assert len(process_list) == num_workers + 1
|
|
|
|
|
|
@pytest.mark.skipif(
|
|
not hasattr(signal, "SIGALRM"),
|
|
reason="SIGALRM is not implemented for this platform",
|
|
)
|
|
@pytest.mark.skipif(
|
|
sys.platform not in ("linux", "darwin"),
|
|
reason="This test requires fork context",
|
|
)
|
|
def test_multiprocessing_with_blueprint(app):
|
|
# Selects a number at random so we can spot check
|
|
num_workers = random.choice(range(2, multiprocessing.cpu_count() * 2 + 1))
|
|
process_list = set()
|
|
|
|
@app.after_server_start
|
|
async def shutdown(app):
|
|
await sleep(2.1)
|
|
app.stop()
|
|
|
|
def stop_on_alarm(*args):
|
|
for process in multiprocessing.active_children():
|
|
process_list.add(process.pid)
|
|
|
|
signal.signal(signal.SIGALRM, stop_on_alarm)
|
|
signal.alarm(2)
|
|
|
|
bp = Blueprint("test_text")
|
|
app.blueprint(bp)
|
|
with use_context("fork"):
|
|
app.run(HOST, 4121, workers=num_workers, debug=True)
|
|
|
|
assert len(process_list) == num_workers + 1
|
|
|
|
|
|
# this function must be outside a test function so that it can be
|
|
# able to be pickled (local functions cannot be pickled).
|
|
def handler(request):
|
|
return text("Hello")
|
|
|
|
|
|
def stop(app):
|
|
app.stop()
|
|
|
|
|
|
# Multiprocessing on Windows requires app to be able to be pickled
|
|
@pytest.mark.parametrize("protocol", [3, 4])
|
|
def test_pickle_app(app, protocol):
|
|
app.route("/")(handler)
|
|
app.after_server_start(stop)
|
|
app.router.reset()
|
|
app.signal_router.reset()
|
|
p_app = pickle.dumps(app, protocol=protocol)
|
|
del app
|
|
up_p_app = pickle.loads(p_app)
|
|
assert up_p_app
|
|
up_p_app.run(single_process=True)
|
|
|
|
|
|
@pytest.mark.parametrize("protocol", [3, 4])
|
|
def test_pickle_app_with_bp(app, protocol):
|
|
bp = Blueprint("test_text")
|
|
bp.route("/")(handler)
|
|
bp.after_server_start(stop)
|
|
app.blueprint(bp)
|
|
app.router.reset()
|
|
app.signal_router.reset()
|
|
p_app = pickle.dumps(app, protocol=protocol)
|
|
del app
|
|
up_p_app = pickle.loads(p_app)
|
|
assert up_p_app
|
|
up_p_app.run(single_process=True)
|
|
|
|
|
|
@pytest.mark.parametrize("protocol", [3, 4])
|
|
def test_pickle_app_with_static(app, protocol):
|
|
app.route("/")(handler)
|
|
app.after_server_start(stop)
|
|
app.static("/static", "/tmp/static")
|
|
app.router.reset()
|
|
app.signal_router.reset()
|
|
p_app = pickle.dumps(app, protocol=protocol)
|
|
del app
|
|
up_p_app = pickle.loads(p_app)
|
|
assert up_p_app
|
|
up_p_app.run(single_process=True)
|
|
|
|
|
|
@pytest.mark.skipif(
|
|
sys.platform not in ("linux", "darwin"),
|
|
reason="This test requires fork context",
|
|
)
|
|
def test_main_process_event(app, caplog):
|
|
# Selects a number at random so we can spot check
|
|
num_workers = random.choice(range(2, multiprocessing.cpu_count() * 2 + 1))
|
|
|
|
app.after_server_start(stop)
|
|
|
|
@app.listener("main_process_start")
|
|
def main_process_start(app, loop):
|
|
logger.info("main_process_start")
|
|
|
|
@app.listener("main_process_stop")
|
|
def main_process_stop(app, loop):
|
|
logger.info("main_process_stop")
|
|
|
|
@app.main_process_start
|
|
def main_process_start2(app, loop):
|
|
logger.info("main_process_start")
|
|
|
|
@app.main_process_stop
|
|
def main_process_stop2(app, loop):
|
|
logger.info("main_process_stop")
|
|
|
|
with use_context("fork"):
|
|
with caplog.at_level(logging.INFO):
|
|
app.run(HOST, PORT, workers=num_workers)
|
|
|
|
assert (
|
|
caplog.record_tuples.count(("sanic.root", 20, "main_process_start"))
|
|
== 2
|
|
)
|
|
assert (
|
|
caplog.record_tuples.count(("sanic.root", 20, "main_process_stop"))
|
|
== 2
|
|
)
|