Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion Lib/test/_test_multiprocessing.py
Original file line number Diff line number Diff line change
Expand Up @@ -318,7 +318,7 @@ def _test_create_grandchild_process(cls, wconn):
def _test_report_parent_status(cls, wconn):
from multiprocessing.process import parent_process
wconn.send("alive" if parent_process().is_alive() else "not alive")
parent_process().join(timeout=5)
parent_process().join(timeout=support.SHORT_TIMEOUT)
wconn.send("alive" if parent_process().is_alive() else "not alive")

def test_process(self):
Expand Down
4 changes: 2 additions & 2 deletions Lib/test/fork_wait.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@

import os, sys, time, unittest
import threading
import test.support as support
from test import support


LONGSLEEP = 2
Expand Down Expand Up @@ -62,7 +62,7 @@ def test_wait(self):
self.threads.append(thread)

# busy-loop to wait for threads
deadline = time.monotonic() + 10.0
deadline = time.monotonic() + support.SHORT_TIMEOUT
while len(self.alive) < NUM_THREADS:
time.sleep(0.1)
if deadline < time.monotonic():
Expand Down
3 changes: 2 additions & 1 deletion Lib/test/signalinterproctester.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
import sys
import time
import unittest
from test import support


class SIGUSR1Exception(Exception):
Expand All @@ -27,7 +28,7 @@ def wait_signal(self, child, signame):
# (if set)
child.wait()

timeout = 10.0
timeout = support.SHORT_TIMEOUT
deadline = time.monotonic() + timeout

while time.monotonic() < deadline:
Expand Down
14 changes: 8 additions & 6 deletions Lib/test/test_asyncio/test_events.py
Original file line number Diff line number Diff line change
Expand Up @@ -1475,12 +1475,12 @@ def reader(data):
return len(data)

test_utils.run_until(self.loop, lambda: reader(data) >= 1,
timeout=10)
timeout=support.SHORT_TIMEOUT)
self.assertEqual(b'1', data)

transport.write(b'2345')
test_utils.run_until(self.loop, lambda: reader(data) >= 5,
timeout=10)
timeout=support.SHORT_TIMEOUT)
self.assertEqual(b'12345', data)
self.assertEqual('CONNECTED', proto.state)

Expand Down Expand Up @@ -1531,27 +1531,29 @@ def reader(data):
return len(data)

write_transport.write(b'1')
test_utils.run_until(self.loop, lambda: reader(data) >= 1, timeout=10)
test_utils.run_until(self.loop, lambda: reader(data) >= 1,
timeout=support.SHORT_TIMEOUT)
self.assertEqual(b'1', data)
self.assertEqual(['INITIAL', 'CONNECTED'], read_proto.state)
self.assertEqual('CONNECTED', write_proto.state)

os.write(master, b'a')
test_utils.run_until(self.loop, lambda: read_proto.nbytes >= 1,
timeout=10)
timeout=support.SHORT_TIMEOUT)
self.assertEqual(['INITIAL', 'CONNECTED'], read_proto.state)
self.assertEqual(1, read_proto.nbytes)
self.assertEqual('CONNECTED', write_proto.state)

write_transport.write(b'2345')
test_utils.run_until(self.loop, lambda: reader(data) >= 5, timeout=10)
test_utils.run_until(self.loop, lambda: reader(data) >= 5,
timeout=support.SHORT_TIMEOUT)
self.assertEqual(b'12345', data)
self.assertEqual(['INITIAL', 'CONNECTED'], read_proto.state)
self.assertEqual('CONNECTED', write_proto.state)

os.write(master, b'bcde')
test_utils.run_until(self.loop, lambda: read_proto.nbytes >= 5,
timeout=10)
timeout=support.SHORT_TIMEOUT)
self.assertEqual(['INITIAL', 'CONNECTED'], read_proto.state)
self.assertEqual(5, read_proto.nbytes)
self.assertEqual('CONNECTED', write_proto.state)
Expand Down
11 changes: 7 additions & 4 deletions Lib/test/test_asyncio/test_sslproto.py
Original file line number Diff line number Diff line change
Expand Up @@ -274,7 +274,8 @@ async def client(addr):

with self.tcp_server(serve, timeout=self.TIMEOUT) as srv:
self.loop.run_until_complete(
asyncio.wait_for(client(srv.addr), timeout=10))
asyncio.wait_for(client(srv.addr),
timeout=support.SHORT_TIMEOUT))

# No garbage is left if SSL is closed uncleanly
client_context = weakref.ref(client_context)
Expand Down Expand Up @@ -335,7 +336,8 @@ async def client(addr):

with self.tcp_server(serve, timeout=self.TIMEOUT) as srv:
self.loop.run_until_complete(
asyncio.wait_for(client(srv.addr), timeout=10))
asyncio.wait_for(client(srv.addr),
timeout=support.SHORT_TIMEOUT))

# No garbage is left for SSL client from loop.create_connection, even
# if user stores the SSLTransport in corresponding protocol instance
Expand Down Expand Up @@ -491,7 +493,8 @@ async def client(addr):

with self.tcp_server(serve, timeout=self.TIMEOUT) as srv:
self.loop.run_until_complete(
asyncio.wait_for(client(srv.addr), timeout=10))
asyncio.wait_for(client(srv.addr),
timeout=support.SHORT_TIMEOUT))

def test_start_tls_server_1(self):
HELLO_MSG = b'1' * self.PAYLOAD_SIZE
Expand Down Expand Up @@ -619,7 +622,7 @@ async def client(addr):
*addr,
ssl=client_sslctx,
server_hostname='',
ssl_handshake_timeout=10.0),
ssl_handshake_timeout=support.SHORT_TIMEOUT),
0.5)

with self.tcp_server(server,
Expand Down
2 changes: 1 addition & 1 deletion Lib/test/test_asyncio/utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -107,7 +107,7 @@ async def once():
gen.close()


def run_until(loop, pred, timeout=30):
def run_until(loop, pred, timeout=support.SHORT_TIMEOUT):
deadline = time.monotonic() + timeout
while not pred():
if timeout is not None:
Expand Down
41 changes: 21 additions & 20 deletions Lib/test/test_concurrent_futures.py
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
import test.support
from test import support

# Skip tests if _multiprocessing wasn't built.
test.support.import_module('_multiprocessing')
support.import_module('_multiprocessing')
# Skip tests if sem_open implementation is broken.
test.support.import_module('multiprocessing.synchronize')
support.import_module('multiprocessing.synchronize')

from test.support.script_helper import assert_python_ok

Expand Down Expand Up @@ -101,11 +101,11 @@ def make_dummy_object(_):

class BaseTestCase(unittest.TestCase):
def setUp(self):
self._thread_key = test.support.threading_setup()
self._thread_key = support.threading_setup()

def tearDown(self):
test.support.reap_children()
test.support.threading_cleanup(*self._thread_key)
support.reap_children()
support.threading_cleanup(*self._thread_key)


class ExecutorMixin:
Expand All @@ -132,7 +132,7 @@ def tearDown(self):
self.executor = None

dt = time.monotonic() - self.t1
if test.support.verbose:
if support.verbose:
print("%.2fs" % dt, end=' ')
self.assertLess(dt, 300, "synchronization issue: test lasted too long")

Expand Down Expand Up @@ -712,7 +712,7 @@ def test_shutdown_race_issue12456(self):
self.executor.map(str, [2] * (self.worker_count + 1))
self.executor.shutdown()

@test.support.cpython_only
@support.cpython_only
def test_no_stale_references(self):
# Issue #16284: check that the executors don't unnecessarily hang onto
# references.
Expand All @@ -724,7 +724,7 @@ def test_no_stale_references(self):
self.executor.submit(my_object.my_method)
del my_object

collected = my_object_collected.wait(timeout=5.0)
collected = my_object_collected.wait(timeout=support.SHORT_TIMEOUT)
self.assertTrue(collected,
"Stale reference not collected within timeout.")

Expand Down Expand Up @@ -836,7 +836,7 @@ def test_traceback(self):
self.assertIs(type(cause), futures.process._RemoteTraceback)
self.assertIn('raise RuntimeError(123) # some comment', cause.tb)

with test.support.captured_stderr() as f1:
with support.captured_stderr() as f1:
try:
raise exc
except RuntimeError:
Expand Down Expand Up @@ -929,7 +929,7 @@ def __reduce__(self):


class ExecutorDeadlockTest:
TIMEOUT = 15
TIMEOUT = support.SHORT_TIMEOUT

@classmethod
def _sleep_id(cls, x, delay):
Expand Down Expand Up @@ -994,7 +994,7 @@ def test_crash(self):
for func, args, error, name in crash_cases:
with self.subTest(name):
# The captured_stderr reduces the noise in the test report
with test.support.captured_stderr():
with support.captured_stderr():
executor = self.executor_type(
max_workers=2, mp_context=get_context(self.ctx))
res = executor.submit(func, *args)
Expand Down Expand Up @@ -1061,7 +1061,7 @@ def fn(callback_future):
self.assertTrue(was_cancelled)

def test_done_callback_raises(self):
with test.support.captured_stderr() as stderr:
with support.captured_stderr() as stderr:
raising_was_called = False
fn_was_called = False

Expand Down Expand Up @@ -1116,7 +1116,7 @@ def fn(callback_future):
self.assertTrue(was_cancelled)

def test_done_callback_raises_already_succeeded(self):
with test.support.captured_stderr() as stderr:
with support.captured_stderr() as stderr:
def raising_fn(callback_future):
raise Exception('doh!')

Expand Down Expand Up @@ -1235,7 +1235,8 @@ def notification():
t = threading.Thread(target=notification)
t.start()

self.assertRaises(futures.CancelledError, f1.result, timeout=5)
self.assertRaises(futures.CancelledError,
f1.result, timeout=support.SHORT_TIMEOUT)
t.join()

def test_exception_with_timeout(self):
Expand Down Expand Up @@ -1264,7 +1265,7 @@ def notification():
t = threading.Thread(target=notification)
t.start()

self.assertTrue(isinstance(f1.exception(timeout=5), OSError))
self.assertTrue(isinstance(f1.exception(timeout=support.SHORT_TIMEOUT), OSError))
t.join()

def test_multiple_set_result(self):
Expand Down Expand Up @@ -1300,12 +1301,12 @@ def test_multiple_set_exception(self):

def setUpModule():
global _threads_key
_threads_key = test.support.threading_setup()
_threads_key = support.threading_setup()


def tearDownModule():
test.support.threading_cleanup(*_threads_key)
test.support.reap_children()
support.threading_cleanup(*_threads_key)
support.reap_children()

# cleanup multiprocessing
multiprocessing.process._cleanup()
Expand All @@ -1315,7 +1316,7 @@ def tearDownModule():
# bpo-37421: Explicitly call _run_finalizers() to remove immediately
# temporary directories created by multiprocessing.util.get_temp_dir().
multiprocessing.util._run_finalizers()
test.support.gc_collect()
support.gc_collect()


if __name__ == "__main__":
Expand Down
12 changes: 6 additions & 6 deletions Lib/test/test_fork1.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,15 +10,15 @@
import unittest

from test.fork_wait import ForkWait
from test.support import reap_children, get_attribute, verbose
from test import support


# Skip test if fork does not exist.
get_attribute(os, 'fork')
support.get_attribute(os, 'fork')

class ForkTest(ForkWait):
def wait_impl(self, cpid):
deadline = time.monotonic() + 10.0
deadline = time.monotonic() + support.SHORT_TIMEOUT
while time.monotonic() <= deadline:
# waitpid() shouldn't hang, but some of the buildbots seem to hang
# in the forking tests. This is an attempt to fix the problem.
Expand Down Expand Up @@ -56,7 +56,7 @@ def importer():
if m == complete_module:
os._exit(0)
else:
if verbose > 1:
if support.verbose > 1:
print("Child encountered partial module")
os._exit(1)
else:
Expand Down Expand Up @@ -90,7 +90,7 @@ def fork_with_import_lock(level):
imp.release_lock()
except RuntimeError:
if in_child:
if verbose > 1:
if support.verbose > 1:
print("RuntimeError in child")
os._exit(1)
raise
Expand All @@ -105,7 +105,7 @@ def fork_with_import_lock(level):


def tearDownModule():
reap_children()
support.reap_children()

if __name__ == "__main__":
unittest.main()
2 changes: 1 addition & 1 deletion Lib/test/test_pydoc.py
Original file line number Diff line number Diff line change
Expand Up @@ -1336,7 +1336,7 @@ def my_url_handler(url, content_type):
self.assertIn('0.0.0.0', serverthread.docserver.address)

starttime = time.monotonic()
timeout = 1 #seconds
timeout = test.support.SHORT_TIMEOUT

while serverthread.serving:
time.sleep(.01)
Expand Down
2 changes: 1 addition & 1 deletion Lib/test/test_sched.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@
from test import support


TIMEOUT = 10
TIMEOUT = support.SHORT_TIMEOUT


class Timer:
Expand Down
6 changes: 3 additions & 3 deletions Lib/test/test_signal.py
Original file line number Diff line number Diff line change
Expand Up @@ -644,7 +644,7 @@ def handler(signum, frame):
# wait until the child process is loaded and has started
first_line = process.stdout.readline()

stdout, stderr = process.communicate(timeout=5.0)
stdout, stderr = process.communicate(timeout=support.SHORT_TIMEOUT)
except subprocess.TimeoutExpired:
process.kill()
return False
Expand Down Expand Up @@ -1192,7 +1192,7 @@ def second_handler(signum=None, frame=None):
self.setsig(signal.SIGALRM, second_handler) # for ITIMER_REAL

expected_sigs = 0
deadline = time.monotonic() + 15.0
deadline = time.monotonic() + support.SHORT_TIMEOUT

while expected_sigs < N:
os.kill(os.getpid(), signal.SIGPROF)
Expand Down Expand Up @@ -1226,7 +1226,7 @@ def handler(signum, frame):
self.setsig(signal.SIGALRM, handler) # for ITIMER_REAL

expected_sigs = 0
deadline = time.monotonic() + 15.0
deadline = time.monotonic() + support.SHORT_TIMEOUT

while expected_sigs < N:
# Hopefully the SIGALRM will be received somewhere during
Expand Down
2 changes: 1 addition & 1 deletion Lib/test/test_socketserver.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ def signal_alarm(n):
# Remember real select() to avoid interferences with mocking
_real_select = select.select

def receive(sock, n, timeout=20):
def receive(sock, n, timeout=test.support.SHORT_TIMEOUT):
r, w, x = _real_select([sock], [], [], timeout)
if sock in r:
return sock.recv(n)
Expand Down
2 changes: 1 addition & 1 deletion Lib/test/test_ssl.py
Original file line number Diff line number Diff line change
Expand Up @@ -2156,7 +2156,7 @@ def test_context_setget(self):
def ssl_io_loop(self, sock, incoming, outgoing, func, *args, **kwargs):
# A simple IO loop. Call func(*args) depending on the error we get
# (WANT_READ or WANT_WRITE) move data between the socket and the BIOs.
timeout = kwargs.get('timeout', 10)
timeout = kwargs.get('timeout', support.SHORT_TIMEOUT)
deadline = time.monotonic() + timeout
count = 0
while True:
Expand Down
Loading