test
This commit is contained in:
parent
3653f92784
commit
fab771129d
|
@ -1,3 +1,6 @@
|
||||||
NETADM_CONFIG__RUN__HOST = "0.0.0.0"
|
NETADM_CONFIG__RUN__HOST = "0.0.0.0"
|
||||||
NETADM_CONFIG__RUN__PORT = 8000
|
NETADM_CONFIG__RUN__PORT = 8000
|
||||||
NETADM_CONFIG__RUN__RELOAD = 1
|
NETADM_CONFIG__RUN__RELOAD = 1
|
||||||
|
|
||||||
|
NETADM_CONFIG__DB__URL=postgresql+asyncpg://user:pwd@localhost:5432/app
|
||||||
|
NETADM_CONFIG__DB__ECHO=1
|
|
@ -0,0 +1,115 @@
|
||||||
|
# A generic, single database configuration.
|
||||||
|
|
||||||
|
[alembic]
|
||||||
|
# path to migration scripts.
|
||||||
|
# Use forward slashes (/) also on windows to provide an os agnostic path
|
||||||
|
script_location = alembic
|
||||||
|
|
||||||
|
# template used to generate migration file names; The default value is %%(rev)s_%%(slug)s
|
||||||
|
# Uncomment the line below if you want the files to be prepended with date and time
|
||||||
|
file_template = %%(year)d_%%(month).2d_%%(day).2d_%%(hour).2d%%(minute).2d-%%(rev)s_%%(slug)s
|
||||||
|
|
||||||
|
# sys.path path, will be prepended to sys.path if present.
|
||||||
|
# defaults to the current working directory.
|
||||||
|
prepend_sys_path = .
|
||||||
|
|
||||||
|
# timezone to use when rendering the date within the migration file
|
||||||
|
# as well as the filename.
|
||||||
|
# If specified, requires the python>=3.9 or backports.zoneinfo library.
|
||||||
|
# Any required deps can installed by adding `alembic[tz]` to the pip requirements
|
||||||
|
# string value is passed to ZoneInfo()
|
||||||
|
# leave blank for localtime
|
||||||
|
# timezone =
|
||||||
|
|
||||||
|
# max length of characters to apply to the "slug" field
|
||||||
|
# truncate_slug_length = 40
|
||||||
|
|
||||||
|
# set to 'true' to run the environment during
|
||||||
|
# the 'revision' command, regardless of autogenerate
|
||||||
|
# revision_environment = false
|
||||||
|
|
||||||
|
# set to 'true' to allow .pyc and .pyo files without
|
||||||
|
# a source .py file to be detected as revisions in the
|
||||||
|
# versions/ directory
|
||||||
|
# sourceless = false
|
||||||
|
|
||||||
|
# version location specification; This defaults
|
||||||
|
# to alembic/versions. When using multiple version
|
||||||
|
# directories, initial revisions must be specified with --version-path.
|
||||||
|
# The path separator used here should be the separator specified by "version_path_separator" below.
|
||||||
|
# version_locations = %(here)s/bar:%(here)s/bat:alembic/versions
|
||||||
|
|
||||||
|
# version path separator; As mentioned above, this is the character used to split
|
||||||
|
# version_locations. The default within new alembic.ini files is "os", which uses os.pathsep.
|
||||||
|
# If this key is omitted entirely, it falls back to the legacy behavior of splitting on spaces and/or commas.
|
||||||
|
# Valid values for version_path_separator are:
|
||||||
|
#
|
||||||
|
# version_path_separator = :
|
||||||
|
# version_path_separator = ;
|
||||||
|
# version_path_separator = space
|
||||||
|
# version_path_separator = newline
|
||||||
|
version_path_separator = os # Use os.pathsep. Default configuration used for new projects.
|
||||||
|
|
||||||
|
# set to 'true' to search source files recursively
|
||||||
|
# in each "version_locations" directory
|
||||||
|
# new in Alembic version 1.10
|
||||||
|
# recursive_version_locations = false
|
||||||
|
|
||||||
|
# the output encoding used when revision files
|
||||||
|
# are written from script.py.mako
|
||||||
|
# output_encoding = utf-8
|
||||||
|
|
||||||
|
sqlalchemy.url = driver://user:pass@localhost/dbname
|
||||||
|
|
||||||
|
|
||||||
|
[post_write_hooks]
|
||||||
|
# post_write_hooks defines scripts or Python functions that are run
|
||||||
|
# on newly generated revision scripts. See the documentation for further
|
||||||
|
# detail and examples
|
||||||
|
|
||||||
|
# format using "black" - use the console_scripts runner, against the "black" entrypoint
|
||||||
|
hooks = black
|
||||||
|
black.type = console_scripts
|
||||||
|
black.entrypoint = black
|
||||||
|
black.options = -l 79 REVISION_SCRIPT_FILENAME
|
||||||
|
|
||||||
|
# lint with attempts to fix using "ruff" - use the exec runner, execute a binary
|
||||||
|
# hooks = ruff
|
||||||
|
# ruff.type = exec
|
||||||
|
# ruff.executable = %(here)s/.venv/bin/ruff
|
||||||
|
# ruff.options = --fix REVISION_SCRIPT_FILENAME
|
||||||
|
|
||||||
|
# Logging configuration
|
||||||
|
[loggers]
|
||||||
|
keys = root,sqlalchemy,alembic
|
||||||
|
|
||||||
|
[handlers]
|
||||||
|
keys = console
|
||||||
|
|
||||||
|
[formatters]
|
||||||
|
keys = generic
|
||||||
|
|
||||||
|
[logger_root]
|
||||||
|
level = WARN
|
||||||
|
handlers = console
|
||||||
|
qualname =
|
||||||
|
|
||||||
|
[logger_sqlalchemy]
|
||||||
|
level = WARN
|
||||||
|
handlers =
|
||||||
|
qualname = sqlalchemy.engine
|
||||||
|
|
||||||
|
[logger_alembic]
|
||||||
|
level = INFO
|
||||||
|
handlers =
|
||||||
|
qualname = alembic
|
||||||
|
|
||||||
|
[handler_console]
|
||||||
|
class = StreamHandler
|
||||||
|
args = (sys.stderr,)
|
||||||
|
level = NOTSET
|
||||||
|
formatter = generic
|
||||||
|
|
||||||
|
[formatter_generic]
|
||||||
|
format = %(levelname)-5.5s [%(name)s] %(message)s
|
||||||
|
datefmt = %H:%M:%S
|
|
@ -0,0 +1 @@
|
||||||
|
Generic single-database configuration with an async dbapi.
|
|
@ -0,0 +1,94 @@
|
||||||
|
import asyncio
|
||||||
|
from logging.config import fileConfig
|
||||||
|
|
||||||
|
from sqlalchemy import pool
|
||||||
|
from sqlalchemy.engine import Connection
|
||||||
|
from sqlalchemy.ext.asyncio import async_engine_from_config
|
||||||
|
|
||||||
|
from alembic import context
|
||||||
|
|
||||||
|
# this is the Alembic Config object, which provides
|
||||||
|
# access to the values within the .ini file in use.
|
||||||
|
config = context.config
|
||||||
|
|
||||||
|
# Interpret the config file for Python logging.
|
||||||
|
# This line sets up loggers basically.
|
||||||
|
if config.config_file_name is not None:
|
||||||
|
fileConfig(config.config_file_name)
|
||||||
|
|
||||||
|
# add your model's MetaData object here
|
||||||
|
# for 'autogenerate' support
|
||||||
|
# from myapp import mymodel
|
||||||
|
# target_metadata = mymodel.Base.metadata
|
||||||
|
# target_metadata = None
|
||||||
|
from models import Base
|
||||||
|
|
||||||
|
target_metadata = Base.metadata
|
||||||
|
# other values from the config, defined by the needs of env.py,
|
||||||
|
# can be acquired:
|
||||||
|
# my_important_option = config.get_main_option("my_important_option")
|
||||||
|
# ... etc.
|
||||||
|
from config import settings
|
||||||
|
|
||||||
|
config.set_main_option("sqlalchemy.url", settings.db_url)
|
||||||
|
|
||||||
|
|
||||||
|
def run_migrations_offline() -> None:
|
||||||
|
"""Run migrations in 'offline' mode.
|
||||||
|
|
||||||
|
This configures the context with just a URL
|
||||||
|
and not an Engine, though an Engine is acceptable
|
||||||
|
here as well. By skipping the Engine creation
|
||||||
|
we don't even need a DBAPI to be available.
|
||||||
|
|
||||||
|
Calls to context.execute() here emit the given string to the
|
||||||
|
script output.
|
||||||
|
|
||||||
|
"""
|
||||||
|
url = config.get_main_option("sqlalchemy.url")
|
||||||
|
context.configure(
|
||||||
|
url=url,
|
||||||
|
target_metadata=target_metadata,
|
||||||
|
literal_binds=True,
|
||||||
|
dialect_opts={"paramstyle": "named"},
|
||||||
|
)
|
||||||
|
|
||||||
|
with context.begin_transaction():
|
||||||
|
context.run_migrations()
|
||||||
|
|
||||||
|
|
||||||
|
def do_run_migrations(connection: Connection) -> None:
|
||||||
|
context.configure(connection=connection, target_metadata=target_metadata)
|
||||||
|
|
||||||
|
with context.begin_transaction():
|
||||||
|
context.run_migrations()
|
||||||
|
|
||||||
|
|
||||||
|
async def run_async_migrations() -> None:
|
||||||
|
"""In this scenario we need to create an Engine
|
||||||
|
and associate a connection with the context.
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
connectable = async_engine_from_config(
|
||||||
|
config.get_section(config.config_ini_section, {}),
|
||||||
|
prefix="sqlalchemy.",
|
||||||
|
poolclass=pool.NullPool,
|
||||||
|
)
|
||||||
|
|
||||||
|
async with connectable.connect() as connection:
|
||||||
|
await connection.run_sync(do_run_migrations)
|
||||||
|
|
||||||
|
await connectable.dispose()
|
||||||
|
|
||||||
|
|
||||||
|
def run_migrations_online() -> None:
|
||||||
|
"""Run migrations in 'online' mode."""
|
||||||
|
|
||||||
|
asyncio.run(run_async_migrations())
|
||||||
|
|
||||||
|
|
||||||
|
if context.is_offline_mode():
|
||||||
|
run_migrations_offline()
|
||||||
|
else:
|
||||||
|
run_migrations_online()
|
|
@ -0,0 +1,26 @@
|
||||||
|
"""${message}
|
||||||
|
|
||||||
|
Revision ID: ${up_revision}
|
||||||
|
Revises: ${down_revision | comma,n}
|
||||||
|
Create Date: ${create_date}
|
||||||
|
|
||||||
|
"""
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
from alembic import op
|
||||||
|
import sqlalchemy as sa
|
||||||
|
${imports if imports else ""}
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = ${repr(up_revision)}
|
||||||
|
down_revision: Union[str, None] = ${repr(down_revision)}
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = ${repr(branch_labels)}
|
||||||
|
depends_on: Union[str, Sequence[str], None] = ${repr(depends_on)}
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
${upgrades if upgrades else "pass"}
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
${downgrades if downgrades else "pass"}
|
19
config.py
19
config.py
|
@ -3,7 +3,7 @@ from pydantic_settings import (
|
||||||
BaseSettings,
|
BaseSettings,
|
||||||
SettingsConfigDict,
|
SettingsConfigDict,
|
||||||
)
|
)
|
||||||
|
from pydantic import PostgresDsn
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
@ -29,6 +29,22 @@ class SwaggerConfig(BaseModel):
|
||||||
swagger_favicon_url: str = "/static/swagger/favicon.png"
|
swagger_favicon_url: str = "/static/swagger/favicon.png"
|
||||||
|
|
||||||
|
|
||||||
|
class DatabaseConfig(BaseModel):
|
||||||
|
url: PostgresDsn
|
||||||
|
echo: bool = False
|
||||||
|
echo_pool: bool = False
|
||||||
|
pool_size: int = 50
|
||||||
|
max_overflow: int = 10
|
||||||
|
|
||||||
|
naming_convention: dict[str, str] = {
|
||||||
|
"ix": "ix_%(column_0_label)s",
|
||||||
|
"uq": "uq_%(table_name)s_%(column_0_N_name)s",
|
||||||
|
"ck": "ck_%(table_name)s_%(constraint_name)s",
|
||||||
|
"fk": "fk_%(table_name)s_%(column_0_name)s_%(referred_table_name)s",
|
||||||
|
"pk": "pk_%(table_name)s",
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
class Settings(BaseSettings):
|
class Settings(BaseSettings):
|
||||||
model_config = SettingsConfigDict(
|
model_config = SettingsConfigDict(
|
||||||
env_file=(
|
env_file=(
|
||||||
|
@ -41,6 +57,7 @@ class Settings(BaseSettings):
|
||||||
)
|
)
|
||||||
run: RunConfig
|
run: RunConfig
|
||||||
swagger: SwaggerConfig = SwaggerConfig()
|
swagger: SwaggerConfig = SwaggerConfig()
|
||||||
|
db: DatabaseConfig
|
||||||
|
|
||||||
|
|
||||||
settings = Settings()
|
settings = Settings()
|
||||||
|
|
23
main.py
23
main.py
|
@ -1,4 +1,5 @@
|
||||||
import logging
|
import logging
|
||||||
|
from contextlib import asynccontextmanager
|
||||||
from config import settings, STATIC_DIR
|
from config import settings, STATIC_DIR
|
||||||
import uvicorn
|
import uvicorn
|
||||||
from fastapi import FastAPI
|
from fastapi import FastAPI
|
||||||
|
@ -6,15 +7,33 @@ from fastapi.responses import ORJSONResponse
|
||||||
from starlette.staticfiles import StaticFiles
|
from starlette.staticfiles import StaticFiles
|
||||||
from web.routers import router as web_router
|
from web.routers import router as web_router
|
||||||
|
|
||||||
|
from models import db_helper
|
||||||
|
|
||||||
|
|
||||||
|
@asynccontextmanager
|
||||||
|
async def lifespan(app: FastAPI):
|
||||||
|
# startup
|
||||||
|
yield
|
||||||
|
# shutdown
|
||||||
|
await db_helper.dispose()
|
||||||
|
|
||||||
|
|
||||||
main_app = FastAPI(
|
main_app = FastAPI(
|
||||||
default_response_class=ORJSONResponse,
|
default_response_class=ORJSONResponse,
|
||||||
|
lifespan=lifespan,
|
||||||
docs_url=None,
|
docs_url=None,
|
||||||
)
|
)
|
||||||
|
|
||||||
main_app.include_router(web_router)
|
main_app.include_router(
|
||||||
|
web_router,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
main_app.mount("/static", StaticFiles(directory=STATIC_DIR), name="static")
|
main_app.mount(
|
||||||
|
"/static",
|
||||||
|
StaticFiles(directory=STATIC_DIR),
|
||||||
|
name="static",
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
|
|
@ -0,0 +1,8 @@
|
||||||
|
from .db_helper import db_helper
|
||||||
|
from .base import Base
|
||||||
|
|
||||||
|
|
||||||
|
__all__ = (
|
||||||
|
"db_helper",
|
||||||
|
"Base",
|
||||||
|
)
|
|
@ -0,0 +1,20 @@
|
||||||
|
from sqlalchemy import MetaData
|
||||||
|
from sqlalchemy.orm import DeclarativeBase
|
||||||
|
from sqlalchemy.orm import Mapped
|
||||||
|
from sqlalchemy.orm import mapped_column
|
||||||
|
from sqlalchemy.orm import declared_attr
|
||||||
|
|
||||||
|
from config import settings
|
||||||
|
from utils import camel_case_to_snake_case
|
||||||
|
|
||||||
|
|
||||||
|
class Base(DeclarativeBase):
|
||||||
|
__abstract__ = True
|
||||||
|
|
||||||
|
metadata = MetaData(
|
||||||
|
naming_convention=settings.db.naming_convention,
|
||||||
|
)
|
||||||
|
|
||||||
|
@declared_attr.directive
|
||||||
|
def __tablename__(cls) -> str:
|
||||||
|
return f"{camel_case_to_snake_case(cls.__name__)}s"
|
|
@ -0,0 +1,50 @@
|
||||||
|
from typing import AsyncGenerator
|
||||||
|
|
||||||
|
from sqlalchemy.ext.asyncio import (
|
||||||
|
create_async_engine,
|
||||||
|
AsyncEngine,
|
||||||
|
async_sessionmaker,
|
||||||
|
AsyncSession,
|
||||||
|
)
|
||||||
|
|
||||||
|
from config import settings
|
||||||
|
|
||||||
|
|
||||||
|
class DatabaseHelper:
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
url: str,
|
||||||
|
echo: bool = False,
|
||||||
|
echo_pool: bool = False,
|
||||||
|
pool_size: int = 5,
|
||||||
|
max_overflow: int = 10,
|
||||||
|
) -> None:
|
||||||
|
self.engine: AsyncEngine = create_async_engine(
|
||||||
|
url=url,
|
||||||
|
echo=echo,
|
||||||
|
echo_pool=echo_pool,
|
||||||
|
pool_size=pool_size,
|
||||||
|
max_overflow=max_overflow,
|
||||||
|
)
|
||||||
|
self.session_factory: async_sessionmaker[AsyncSession] = async_sessionmaker(
|
||||||
|
bind=self.engine,
|
||||||
|
autoflush=False,
|
||||||
|
autocommit=False,
|
||||||
|
expire_on_commit=False,
|
||||||
|
)
|
||||||
|
|
||||||
|
async def dispose(self) -> None:
|
||||||
|
await self.engine.dispose()
|
||||||
|
|
||||||
|
async def session_getter(self) -> AsyncGenerator[AsyncSession, None]:
|
||||||
|
async with self.session_factory() as session:
|
||||||
|
yield session
|
||||||
|
|
||||||
|
|
||||||
|
db_helper = DatabaseHelper(
|
||||||
|
url=str(settings.db.url),
|
||||||
|
echo=settings.db.echo,
|
||||||
|
echo_pool=settings.db.echo_pool,
|
||||||
|
pool_size=settings.db.pool_size,
|
||||||
|
max_overflow=settings.db.max_overflow,
|
||||||
|
)
|
|
@ -0,0 +1,6 @@
|
||||||
|
from sqlalchemy.orm import Mapped
|
||||||
|
from sqlalchemy.orm import mapped_column
|
||||||
|
|
||||||
|
|
||||||
|
class IdIntPkMixin:
|
||||||
|
id: Mapped[int] = mapped_column(primary_key=True)
|
|
@ -0,0 +1,5 @@
|
||||||
|
__all__ = (
|
||||||
|
"camel_case_to_snake_case",
|
||||||
|
)
|
||||||
|
|
||||||
|
from .case_converter import camel_case_to_snake_case
|
|
@ -0,0 +1,29 @@
|
||||||
|
"""
|
||||||
|
Taken from
|
||||||
|
https://github.com/mahenzon/ri-sdk-python-wrapper/blob/master/ri_sdk_codegen/utils/case_converter.py
|
||||||
|
"""
|
||||||
|
|
||||||
|
|
||||||
|
def camel_case_to_snake_case(input_str: str) -> str:
|
||||||
|
"""
|
||||||
|
>>> camel_case_to_snake_case("SomeSDK")
|
||||||
|
'some_sdk'
|
||||||
|
>>> camel_case_to_snake_case("RServoDrive")
|
||||||
|
'r_servo_drive'
|
||||||
|
>>> camel_case_to_snake_case("SDKDemo")
|
||||||
|
'sdk_demo'
|
||||||
|
"""
|
||||||
|
chars = []
|
||||||
|
for c_idx, char in enumerate(input_str):
|
||||||
|
if c_idx and char.isupper():
|
||||||
|
nxt_idx = c_idx + 1
|
||||||
|
# idea of the flag is to separate abbreviations
|
||||||
|
# as new words, show them in lower case
|
||||||
|
flag = nxt_idx >= len(input_str) or input_str[nxt_idx].isupper()
|
||||||
|
prev_char = input_str[c_idx - 1]
|
||||||
|
if prev_char.isupper() and flag:
|
||||||
|
pass
|
||||||
|
else:
|
||||||
|
chars.append("_")
|
||||||
|
chars.append(char.lower())
|
||||||
|
return "".join(chars)
|
Loading…
Reference in New Issue