import multiprocessing import os import psycopg import platform import random import asyncio import blacksheep as bs import jinja2 from pathlib import Path from psycopg_pool import AsyncConnectionPool READ_ROW_SQL = 'SELECT "id", "randomnumber" FROM "world" WHERE id = %s' WRITE_ROW_SQL = 'UPDATE "world" SET "randomnumber"=%s WHERE id=%s' ADDITIONAL_ROW = [0, "Additional fortune added at request time."] CORE_COUNT = multiprocessing.cpu_count() MAX_POOL_SIZE = CORE_COUNT * 2 MIN_POOL_SIZE = max(1, MAX_POOL_SIZE // 2) db_pool = None async def setup_db(app): global db_pool conninfo = ( f"postgresql://{os.getenv('PGUSER', 'benchmarkdbuser')}:{os.getenv('PGPASS', 'benchmarkdbpass')}" f"@tfb-database:5432/hello_world" ) db_pool = AsyncConnectionPool( conninfo=conninfo, min_size=MIN_POOL_SIZE, max_size=MAX_POOL_SIZE, open=False, timeout=5.0, max_lifetime=1800, ) await db_pool.open() async def shutdown_db(app): global db_pool if db_pool is not None: await db_pool.close() db_pool = None def load_fortunes_template(): with Path("templates/fortune.html").open("r") as f: return jinja2.Template(f.read()) fortune_template = load_fortunes_template() app = bs.Application() app.on_start += setup_db app.on_stop += shutdown_db def get_num_queries(request): try: value = request.query.get('queries') if value is None: return 1 query_count = int(value[0]) except (KeyError, IndexError, ValueError): return 1 return min(max(query_count, 1), 500) JSON_CONTENT_TYPE = b"application/json" @bs.get('/json') async def json_test(request): return bs.json({'message': 'Hello, world!'}) @bs.get('/db') async def single_db_query_test(request): row_id = random.randint(1, 10000) async with db_pool.connection() as db_conn: async with db_conn.cursor() as cursor: await cursor.execute(READ_ROW_SQL, (row_id,)) number = await cursor.fetchone() return bs.json({'id': row_id, 'randomNumber': number[1]}) @bs.get('/queries') async def multiple_db_queries_test(request): num_queries = get_num_queries(request) row_ids = random.sample(range(1, 10000), num_queries) worlds = [] async with db_pool.connection() as db_conn: async with db_conn.cursor() as cursor: for row_id in row_ids: await cursor.execute(READ_ROW_SQL, (row_id,)) number = await cursor.fetchone() worlds.append({"id": row_id, "randomNumber": number[1]}) return bs.json(worlds) @bs.get('/fortunes') async def fortunes_test(request): async with db_pool.connection() as db_conn: async with db_conn.cursor() as cursor: await cursor.execute("SELECT * FROM Fortune") fortunes = await cursor.fetchall() fortunes.append(ADDITIONAL_ROW) fortunes.sort(key=lambda row: row[1]) data = fortune_template.render(fortunes=fortunes) return bs.html(data) @bs.get('/updates') async def db_updates_test(request): num_queries = get_num_queries(request) updates = sorted(zip( random.sample(range(1, 10000), num_queries), random.sample(range(1, 10000), num_queries) ), key=lambda x: x[1]) worlds = [{"id": row_id, "randomNumber": number} for row_id, number in updates] for _ in range(5): async with db_pool.connection() as db_conn: try: await db_conn.execute("SET TRANSACTION ISOLATION LEVEL READ COMMITTED") async with db_conn.cursor() as cursor: for row_id, number in updates: await cursor.execute(READ_ROW_SQL, (row_id,)) await cursor.fetchone() for _ in range(5): try: await cursor.executemany(WRITE_ROW_SQL, [(number, row_id) for row_id, number in updates]) return bs.json(worlds) except psycopg.errors.DeadlockDetected: await db_conn.rollback() continue # await cursor.executemany(WRITE_ROW_SQL, [(number, row_id) for row_id, number in updates]) except (psycopg.errors.OperationalError, psycopg.errors.PipelineAborted): await db_conn.rollback() continue raise Exception("connect error") @bs.get('/plaintext') async def plaintext_test(request): return bs.Response(200, content=bs.Content(b"text/plain", b'Hello, World!')) if platform.python_implementation() == 'PyPy': import logging from socketify import ASGI workers = int(multiprocessing.cpu_count()) if os.environ.get('TRAVIS') == 'true': workers = 2 def run_app(): ASGI(app).listen(8080, lambda config: logging.info(f"Listening on port http://localhost:{config.port} now\n")).run() def create_fork(): n = os.fork() if not n > 0: run_app() for i in range(1, workers): create_fork() run_app()