mirror of
https://github.com/creyD/creyPY.git
synced 2026-04-13 11:50:31 +02:00
Compare commits
52 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
50031556f9 | ||
|
|
2940ddbdcd | ||
| 807af12fa1 | |||
|
|
dce897c247 | ||
|
|
89997372ef | ||
| c8c5977978 | |||
| 974bc591d6 | |||
| eb895398ab | |||
| 867abd7054 | |||
| 26e18f6b31 | |||
| 8a3a60dbb0 | |||
| e52a5f421b | |||
| a6ded91185 | |||
| eb64874c47 | |||
| b7200852a4 | |||
| 3d18205205 | |||
| 99c84b676c | |||
| 6806de23b3 | |||
| 6a93ab05a3 | |||
|
|
c5b2ab9932 | ||
| 5a32a5908b | |||
| b7df0bfdcd | |||
| 378d1d60f1 | |||
| e381992f8e | |||
| 6d5411a8ae | |||
| 89351d714b | |||
| c24f8933fb | |||
| 0bed0e0da4 | |||
| 8463eef907 | |||
| 5903de2aad | |||
| 0bf89fe14d | |||
| d54146e05b | |||
| d6f79c3ed8 | |||
| 3f4a0ee00d | |||
| 714178d68f | |||
| c7e205f14b | |||
| 39ae74becb | |||
| 5f39966223 | |||
| c91e684f08 | |||
| f11b8b8864 | |||
| 983553e97a | |||
| 8740eafce2 | |||
| aa44b9ebe9 | |||
| 851573d964 | |||
| cfa1da08d3 | |||
| 4a5a777ef5 | |||
| c9a9b1bc0a | |||
| d9f6e82736 | |||
| 65e93a023b | |||
|
|
6ce0cfbd14 | ||
|
|
da7ec0b28e | ||
| 2727c452b6 |
35
.github/workflows/ci.yml
vendored
35
.github/workflows/ci.yml
vendored
@@ -4,12 +4,15 @@ on:
|
|||||||
push:
|
push:
|
||||||
branches:
|
branches:
|
||||||
- master
|
- master
|
||||||
|
- dev
|
||||||
paths-ignore:
|
paths-ignore:
|
||||||
- "**/.github/**"
|
|
||||||
- "**/.gitignore"
|
- "**/.gitignore"
|
||||||
- "**/.vscode/**"
|
- "**/.vscode/**"
|
||||||
- "**/README.md"
|
- "**/README.md"
|
||||||
- "**/CHANGELOG.md"
|
- "**/CHANGELOG.md"
|
||||||
|
pull_request:
|
||||||
|
branches:
|
||||||
|
- dev
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
@@ -42,6 +45,7 @@ jobs:
|
|||||||
|
|
||||||
tag_and_publish:
|
tag_and_publish:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
if: github.ref_name == 'master' || github.ref_name == 'dev'
|
||||||
needs: test
|
needs: test
|
||||||
permissions:
|
permissions:
|
||||||
id-token: write # IMPORTANT: this permission is mandatory for trusted publishing
|
id-token: write # IMPORTANT: this permission is mandatory for trusted publishing
|
||||||
@@ -51,7 +55,7 @@ jobs:
|
|||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
with:
|
with:
|
||||||
fetch-tags: true
|
fetch-tags: true
|
||||||
ref: ${{ github.ref }}
|
ref: ${{ github.ref_name }}
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
|
|
||||||
- name: setup git
|
- name: setup git
|
||||||
@@ -59,18 +63,29 @@ jobs:
|
|||||||
git config --local user.email "15138480+creyD@users.noreply.github.com"
|
git config --local user.email "15138480+creyD@users.noreply.github.com"
|
||||||
git config --local user.name "creyD"
|
git config --local user.name "creyD"
|
||||||
|
|
||||||
|
- name: set version format
|
||||||
|
id: version_format
|
||||||
|
run: |
|
||||||
|
if [[ ${{ github.ref_name }} == 'master' ]]; then
|
||||||
|
echo "version_format=\${major}.\${minor}.\${patch}" >> $GITHUB_OUTPUT
|
||||||
|
else
|
||||||
|
echo "version_format=\${major}.\${minor}.\${patch}rc\${increment}" >> $GITHUB_OUTPUT
|
||||||
|
fi
|
||||||
|
|
||||||
- name: Git Version
|
- name: Git Version
|
||||||
uses: codacy/git-version@2.8.0
|
uses: PaulHatch/semantic-version@v5.4.0
|
||||||
id: git_version
|
id: git_version
|
||||||
with:
|
with:
|
||||||
minor-identifier: "feat:"
|
tag_prefix: ""
|
||||||
major-identifier: "breaking:"
|
major_pattern: "breaking:"
|
||||||
|
minor_pattern: "feat:"
|
||||||
|
enable_prerelease_mode: false
|
||||||
|
version_format: ${{ steps.version_format.outputs.version_format }}
|
||||||
|
|
||||||
- name: Create Tag
|
- name: Create & Push Tag
|
||||||
run: git tag ${{ steps.git_version.outputs.version }}
|
run: |
|
||||||
|
git tag ${{ steps.git_version.outputs.version }}
|
||||||
- name: Push tag
|
git push origin ${{ steps.git_version.outputs.version }}
|
||||||
run: git push origin ${{ steps.git_version.outputs.version }}
|
|
||||||
|
|
||||||
- name: Set up Python
|
- name: Set up Python
|
||||||
uses: actions/setup-python@v5
|
uses: actions/setup-python@v5
|
||||||
|
|||||||
8
.vscode/settings.json
vendored
8
.vscode/settings.json
vendored
@@ -26,10 +26,16 @@
|
|||||||
"**/db.sqlite3": true,
|
"**/db.sqlite3": true,
|
||||||
"**/.DS_Store": true,
|
"**/.DS_Store": true,
|
||||||
"**/*.pyc": true,
|
"**/*.pyc": true,
|
||||||
"**/__pycache__/": true
|
"**/__pycache__/": true,
|
||||||
|
"**/build": true,
|
||||||
|
"**/dist": true,
|
||||||
|
"**/*.egg-info": true,
|
||||||
},
|
},
|
||||||
"search.exclude": {
|
"search.exclude": {
|
||||||
"**/.git": true,
|
"**/.git": true,
|
||||||
|
"**/build": true,
|
||||||
|
"**/*.egg-info": true,
|
||||||
|
"**/dist": true,
|
||||||
"**/.venv": true,
|
"**/.venv": true,
|
||||||
"**/tmp": true,
|
"**/tmp": true,
|
||||||
"htmlcov/*": true,
|
"htmlcov/*": true,
|
||||||
|
|||||||
19
CHANGELOG.md
19
CHANGELOG.md
@@ -2,6 +2,25 @@
|
|||||||
|
|
||||||
All notable changes to this project will be documented in this file.
|
All notable changes to this project will be documented in this file.
|
||||||
|
|
||||||
|
## 2.0.0
|
||||||
|
|
||||||
|
- Fixed #1 Rename misspelled additonal_data to additional_data on create_obj_from_data
|
||||||
|
- Fixed #3 Inverse partial flag: bool = False because it was wrong on update_obj_from_data
|
||||||
|
|
||||||
|
Notes:
|
||||||
|
|
||||||
|
You will need to change calls to `create_obj_from_data` according to #1 (rename additonal_data to additional_data)
|
||||||
|
|
||||||
|
You will need to change calls to `update_obj_from_data` according to #3 (if you supplied `partial`, you will need to reverse it: `true` -> `false` and `false` -> `true`)
|
||||||
|
|
||||||
|
## 1.3.0
|
||||||
|
|
||||||
|
- Addition of pagination proxy and pagination=off functionality (Thanks to @vikbhas)
|
||||||
|
|
||||||
|
## 1.2.5
|
||||||
|
|
||||||
|
- Bumped dependencies
|
||||||
|
|
||||||
## 1.2.4
|
## 1.2.4
|
||||||
|
|
||||||
- Enabled newer versions for all dependencies
|
- Enabled newer versions for all dependencies
|
||||||
|
|||||||
@@ -23,9 +23,9 @@ def get_object_or_404(
|
|||||||
|
|
||||||
# TODO: Add testing
|
# TODO: Add testing
|
||||||
def create_obj_from_data(
|
def create_obj_from_data(
|
||||||
data: BaseModel, model: Type[T], db: Session, additonal_data={}, exclude={}
|
data: BaseModel, model: Type[T], db: Session, additional_data={}, exclude={}
|
||||||
) -> T:
|
) -> T:
|
||||||
obj = model(**data.model_dump(exclude=exclude) | additonal_data)
|
obj = model(**data.model_dump(exclude=exclude) | additional_data)
|
||||||
db.add(obj)
|
db.add(obj)
|
||||||
db.commit()
|
db.commit()
|
||||||
db.refresh(obj)
|
db.refresh(obj)
|
||||||
@@ -38,13 +38,13 @@ def update_obj_from_data(
|
|||||||
model: Type[T],
|
model: Type[T],
|
||||||
id: UUID | str,
|
id: UUID | str,
|
||||||
db: Session,
|
db: Session,
|
||||||
partial: bool = False, # TODO: inverse, because it is currently the wrong way around
|
partial: bool = True,
|
||||||
ignore_fields=[],
|
ignore_fields=[],
|
||||||
additional_data={},
|
additional_data={},
|
||||||
exclude={},
|
exclude={},
|
||||||
) -> T:
|
) -> T:
|
||||||
obj = get_object_or_404(model, id, db)
|
obj = get_object_or_404(model, id, db)
|
||||||
data_dict = data.model_dump(exclude_unset=not partial, exclude=exclude)
|
data_dict = data.model_dump(exclude_unset=partial, exclude=exclude)
|
||||||
data_dict.update(additional_data) # merge additional_data into data_dict
|
data_dict.update(additional_data) # merge additional_data into data_dict
|
||||||
for field in data_dict:
|
for field in data_dict:
|
||||||
if field not in ignore_fields:
|
if field not in ignore_fields:
|
||||||
|
|||||||
@@ -1 +1,2 @@
|
|||||||
from .session import * # noqa
|
from .session import * # noqa
|
||||||
|
from .async_session import * # noqa
|
||||||
|
|||||||
31
creyPY/fastapi/db/async_session.py
Normal file
31
creyPY/fastapi/db/async_session.py
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
import os
|
||||||
|
from typing import AsyncGenerator
|
||||||
|
from dotenv import load_dotenv
|
||||||
|
from sqlalchemy.ext.asyncio import AsyncSession, create_async_engine
|
||||||
|
from sqlalchemy.orm import sessionmaker
|
||||||
|
|
||||||
|
|
||||||
|
load_dotenv()
|
||||||
|
|
||||||
|
host = os.getenv("POSTGRES_HOST", "localhost")
|
||||||
|
user = os.getenv("POSTGRES_USER", "postgres")
|
||||||
|
password = os.getenv("POSTGRES_PASSWORD", "root")
|
||||||
|
port = os.getenv("POSTGRES_PORT", "5432")
|
||||||
|
name = os.getenv("POSTGRES_DB", "fastapi")
|
||||||
|
|
||||||
|
SQLALCHEMY_DATABASE_URL = f"postgresql+psycopg://{user}:{password}@{host}:{port}/"
|
||||||
|
|
||||||
|
|
||||||
|
async_engine = create_async_engine(SQLALCHEMY_DATABASE_URL + name, pool_pre_ping=True)
|
||||||
|
AsyncSessionLocal = sessionmaker(
|
||||||
|
bind=async_engine,
|
||||||
|
class_=AsyncSession,
|
||||||
|
expire_on_commit=False,
|
||||||
|
autoflush=False,
|
||||||
|
autocommit=False,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
async def get_async_db() -> AsyncGenerator[AsyncSession, None]:
|
||||||
|
async with AsyncSessionLocal() as db:
|
||||||
|
yield db
|
||||||
@@ -1,14 +1,44 @@
|
|||||||
from math import ceil
|
from math import ceil
|
||||||
from typing import Any, Generic, Optional, Self, Sequence, TypeVar
|
from typing import Any, Generic, Optional, Self, Sequence, TypeVar, Union, overload
|
||||||
|
from contextlib import suppress
|
||||||
|
from pydantic import BaseModel
|
||||||
from fastapi_pagination import Params
|
from fastapi_pagination import Params
|
||||||
from fastapi_pagination.bases import AbstractPage, AbstractParams
|
from fastapi_pagination.bases import AbstractPage, AbstractParams
|
||||||
from fastapi_pagination.types import GreaterEqualOne, GreaterEqualZero
|
from fastapi_pagination.types import (
|
||||||
|
GreaterEqualOne,
|
||||||
|
GreaterEqualZero,
|
||||||
|
AdditionalData,
|
||||||
|
SyncItemsTransformer,
|
||||||
|
AsyncItemsTransformer,
|
||||||
|
ItemsTransformer,
|
||||||
|
)
|
||||||
|
from fastapi_pagination.api import create_page, apply_items_transformer
|
||||||
|
from fastapi_pagination.utils import verify_params
|
||||||
|
from fastapi_pagination.ext.sqlalchemy import create_paginate_query
|
||||||
|
from fastapi_pagination.bases import AbstractParams, RawParams
|
||||||
from pydantic.json_schema import SkipJsonSchema
|
from pydantic.json_schema import SkipJsonSchema
|
||||||
|
from sqlalchemy.sql.selectable import Select
|
||||||
|
from sqlalchemy.orm.session import Session
|
||||||
|
from sqlalchemy import select, func
|
||||||
|
from sqlalchemy.ext.asyncio import AsyncSession, async_scoped_session
|
||||||
|
from fastapi import Query
|
||||||
|
from sqlalchemy.util import await_only, greenlet_spawn
|
||||||
|
|
||||||
T = TypeVar("T")
|
T = TypeVar("T")
|
||||||
|
|
||||||
|
|
||||||
|
class PaginationParams(BaseModel, AbstractParams):
|
||||||
|
page: int = Query(1, ge=1, description="Page number")
|
||||||
|
size: int = Query(50, ge=1, le=100, description="Page size")
|
||||||
|
pagination: bool = Query(True, description="Toggle pagination")
|
||||||
|
|
||||||
|
def to_raw_params(self) -> RawParams:
|
||||||
|
if not self.pagination:
|
||||||
|
return RawParams(limit=None, offset=None)
|
||||||
|
|
||||||
|
return RawParams(limit=self.size, offset=(self.page - 1) * self.size)
|
||||||
|
|
||||||
|
|
||||||
# TODO: Add complete fastapi-pagination proxy here
|
# TODO: Add complete fastapi-pagination proxy here
|
||||||
# TODO: Add pagination off functionality
|
# TODO: Add pagination off functionality
|
||||||
# SkipJsonSchema is used to avoid generating invalid JSON schema in FastAPI
|
# SkipJsonSchema is used to avoid generating invalid JSON schema in FastAPI
|
||||||
@@ -21,7 +51,7 @@ class Page(AbstractPage[T], Generic[T]):
|
|||||||
has_next: bool | SkipJsonSchema[None] = None
|
has_next: bool | SkipJsonSchema[None] = None
|
||||||
has_prev: bool | SkipJsonSchema[None] = None
|
has_prev: bool | SkipJsonSchema[None] = None
|
||||||
|
|
||||||
__params_type__ = Params
|
__params_type__ = PaginationParams
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def create(
|
def create(
|
||||||
@@ -70,3 +100,105 @@ def parse_page(response, page: int, size: int) -> Page:
|
|||||||
has_next=response.has_next,
|
has_next=response.has_next,
|
||||||
has_prev=response.has_prev,
|
has_prev=response.has_prev,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def create_count_query(query: Select) -> Select:
|
||||||
|
return select(func.count()).select_from(query.subquery())
|
||||||
|
|
||||||
|
|
||||||
|
def unwrap_scalars(
|
||||||
|
items: Sequence[Sequence[T]],
|
||||||
|
force_unwrap: bool = True,
|
||||||
|
) -> Union[Sequence[T], Sequence[Sequence[T]]]:
|
||||||
|
return [item[0] if force_unwrap else item for item in items]
|
||||||
|
|
||||||
|
|
||||||
|
def _get_sync_conn_from_async(conn: Any) -> Session: # pragma: no cover
|
||||||
|
if isinstance(conn, async_scoped_session):
|
||||||
|
conn = conn()
|
||||||
|
|
||||||
|
with suppress(AttributeError):
|
||||||
|
return conn.sync_session # type: ignore
|
||||||
|
|
||||||
|
with suppress(AttributeError):
|
||||||
|
return conn.sync_connection # type: ignore
|
||||||
|
|
||||||
|
raise TypeError("conn must be an AsyncConnection or AsyncSession")
|
||||||
|
|
||||||
|
|
||||||
|
@overload
|
||||||
|
def paginate(
|
||||||
|
connection: Session,
|
||||||
|
query: Select,
|
||||||
|
params: Optional[AbstractParams] = None,
|
||||||
|
transformer: Optional[SyncItemsTransformer] = None,
|
||||||
|
additional_data: Optional[AdditionalData] = None,
|
||||||
|
) -> Any:
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
@overload
|
||||||
|
async def paginate(
|
||||||
|
connection: AsyncSession,
|
||||||
|
query: Select,
|
||||||
|
params: Optional[AbstractParams] = None,
|
||||||
|
transformer: Optional[AsyncItemsTransformer] = None,
|
||||||
|
additional_data: Optional[AdditionalData] = None,
|
||||||
|
) -> Any:
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
def _paginate(
|
||||||
|
connection: Session,
|
||||||
|
query: Select,
|
||||||
|
params: Optional[AbstractParams] = None,
|
||||||
|
transformer: Optional[ItemsTransformer] = None,
|
||||||
|
additional_data: Optional[AdditionalData] = None,
|
||||||
|
async_: bool = False,
|
||||||
|
):
|
||||||
|
|
||||||
|
if async_:
|
||||||
|
|
||||||
|
def _apply_items_transformer(*args: Any, **kwargs: Any) -> Any:
|
||||||
|
return await_only(apply_items_transformer(*args, **kwargs, async_=True))
|
||||||
|
|
||||||
|
else:
|
||||||
|
_apply_items_transformer = apply_items_transformer
|
||||||
|
|
||||||
|
params, raw_params = verify_params(params, "limit-offset", "cursor")
|
||||||
|
count_query = create_count_query(query)
|
||||||
|
total = connection.scalar(count_query)
|
||||||
|
|
||||||
|
if params.pagination is False and total > 0:
|
||||||
|
params = Params(page=1, size=total)
|
||||||
|
else:
|
||||||
|
params = Params(page=params.page, size=params.size)
|
||||||
|
|
||||||
|
query = create_paginate_query(query, params)
|
||||||
|
items = connection.execute(query).all()
|
||||||
|
|
||||||
|
items = unwrap_scalars(items)
|
||||||
|
t_items = _apply_items_transformer(items, transformer)
|
||||||
|
|
||||||
|
return create_page(
|
||||||
|
t_items,
|
||||||
|
params=params,
|
||||||
|
total=total,
|
||||||
|
**(additional_data or {}),
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def paginate(
|
||||||
|
connection: Session,
|
||||||
|
query: Select,
|
||||||
|
params: Optional[AbstractParams] = None,
|
||||||
|
transformer: Optional[ItemsTransformer] = None,
|
||||||
|
additional_data: Optional[AdditionalData] = None,
|
||||||
|
):
|
||||||
|
if isinstance(connection, AsyncSession):
|
||||||
|
connection = _get_sync_conn_from_async(connection)
|
||||||
|
return greenlet_spawn(
|
||||||
|
_paginate, connection, query, params, transformer, additional_data, async_=True
|
||||||
|
)
|
||||||
|
|
||||||
|
return _paginate(connection, query, params, transformer, additional_data, async_=False)
|
||||||
|
|||||||
@@ -21,3 +21,6 @@ psycopg-pool>=3.2.2 # PostgreSQL
|
|||||||
h11>=0.14.0 # Testing
|
h11>=0.14.0 # Testing
|
||||||
httpcore>=1.0.5 # Testing
|
httpcore>=1.0.5 # Testing
|
||||||
httpx>=0.27.0 # Testing
|
httpx>=0.27.0 # Testing
|
||||||
|
|
||||||
|
asyncpg>=0.30.0 #SQLAlchemy
|
||||||
|
greenlet>=3.1.1 #Async
|
||||||
|
|||||||
Reference in New Issue
Block a user