Server IP : 66.29.132.122 / Your IP : 18.188.118.20 Web Server : LiteSpeed System : Linux business142.web-hosting.com 4.18.0-553.lve.el8.x86_64 #1 SMP Mon May 27 15:27:34 UTC 2024 x86_64 User : admazpex ( 531) PHP Version : 7.2.34 Disable Function : NONE MySQL : OFF | cURL : ON | WGET : ON | Perl : ON | Python : ON | Sudo : OFF | Pkexec : OFF Directory : /proc/self/root/proc/self/root/proc/thread-self/root/proc/thread-self/root/opt/hc_python/lib/python3.8/site-packages/sqlalchemy/orm/ |
Upload File : |
# orm/bulk_persistence.py # Copyright (C) 2005-2024 the SQLAlchemy authors and contributors # <see AUTHORS file> # # This module is part of SQLAlchemy and is released under # the MIT License: https://www.opensource.org/licenses/mit-license.php # mypy: ignore-errors """additional ORM persistence classes related to "bulk" operations, specifically outside of the flush() process. """ from __future__ import annotations from typing import Any from typing import cast from typing import Dict from typing import Iterable from typing import Optional from typing import overload from typing import TYPE_CHECKING from typing import TypeVar from typing import Union from . import attributes from . import context from . import evaluator from . import exc as orm_exc from . import loading from . import persistence from .base import NO_VALUE from .context import AbstractORMCompileState from .context import FromStatement from .context import ORMFromStatementCompileState from .context import QueryContext from .. import exc as sa_exc from .. import util from ..engine import Dialect from ..engine import result as _result from ..sql import coercions from ..sql import dml from ..sql import expression from ..sql import roles from ..sql import select from ..sql import sqltypes from ..sql.base import _entity_namespace_key from ..sql.base import CompileState from ..sql.base import Options from ..sql.dml import DeleteDMLState from ..sql.dml import InsertDMLState from ..sql.dml import UpdateDMLState from ..util import EMPTY_DICT from ..util.typing import Literal if TYPE_CHECKING: from ._typing import DMLStrategyArgument from ._typing import OrmExecuteOptionsParameter from ._typing import SynchronizeSessionArgument from .mapper import Mapper from .session import _BindArguments from .session import ORMExecuteState from .session import Session from .session import SessionTransaction from .state import InstanceState from ..engine import Connection from ..engine import cursor from ..engine.interfaces import _CoreAnyExecuteParams _O = TypeVar("_O", bound=object) @overload def _bulk_insert( mapper: Mapper[_O], mappings: Union[Iterable[InstanceState[_O]], Iterable[Dict[str, Any]]], session_transaction: SessionTransaction, *, isstates: bool, return_defaults: bool, render_nulls: bool, use_orm_insert_stmt: Literal[None] = ..., execution_options: Optional[OrmExecuteOptionsParameter] = ..., ) -> None: ... @overload def _bulk_insert( mapper: Mapper[_O], mappings: Union[Iterable[InstanceState[_O]], Iterable[Dict[str, Any]]], session_transaction: SessionTransaction, *, isstates: bool, return_defaults: bool, render_nulls: bool, use_orm_insert_stmt: Optional[dml.Insert] = ..., execution_options: Optional[OrmExecuteOptionsParameter] = ..., ) -> cursor.CursorResult[Any]: ... def _bulk_insert( mapper: Mapper[_O], mappings: Union[Iterable[InstanceState[_O]], Iterable[Dict[str, Any]]], session_transaction: SessionTransaction, *, isstates: bool, return_defaults: bool, render_nulls: bool, use_orm_insert_stmt: Optional[dml.Insert] = None, execution_options: Optional[OrmExecuteOptionsParameter] = None, ) -> Optional[cursor.CursorResult[Any]]: base_mapper = mapper.base_mapper if session_transaction.session.connection_callable: raise NotImplementedError( "connection_callable / per-instance sharding " "not supported in bulk_insert()" ) if isstates: if TYPE_CHECKING: mappings = cast(Iterable[InstanceState[_O]], mappings) if return_defaults: # list of states allows us to attach .key for return_defaults case states = [(state, state.dict) for state in mappings] mappings = [dict_ for (state, dict_) in states] else: mappings = [state.dict for state in mappings] else: if TYPE_CHECKING: mappings = cast(Iterable[Dict[str, Any]], mappings) if return_defaults: # use dictionaries given, so that newly populated defaults # can be delivered back to the caller (see #11661). This is **not** # compatible with other use cases such as a session-executed # insert() construct, as this will confuse the case of # insert-per-subclass for joined inheritance cases (see # test_bulk_statements.py::BulkDMLReturningJoinedInhTest). # # So in this conditional, we have **only** called # session.bulk_insert_mappings() which does not have this # requirement mappings = list(mappings) else: # for all other cases we need to establish a local dictionary # so that the incoming dictionaries aren't mutated mappings = [dict(m) for m in mappings] _expand_composites(mapper, mappings) connection = session_transaction.connection(base_mapper) return_result: Optional[cursor.CursorResult[Any]] = None mappers_to_run = [ (table, mp) for table, mp in base_mapper._sorted_tables.items() if table in mapper._pks_by_table ] if return_defaults: # not used by new-style bulk inserts, only used for legacy bookkeeping = True elif len(mappers_to_run) > 1: # if we have more than one table, mapper to run where we will be # either horizontally splicing, or copying values between tables, # we need the "bookkeeping" / deterministic returning order bookkeeping = True else: bookkeeping = False for table, super_mapper in mappers_to_run: # find bindparams in the statement. For bulk, we don't really know if # a key in the params applies to a different table since we are # potentially inserting for multiple tables here; looking at the # bindparam() is a lot more direct. in most cases this will # use _generate_cache_key() which is memoized, although in practice # the ultimate statement that's executed is probably not the same # object so that memoization might not matter much. extra_bp_names = ( [ b.key for b in use_orm_insert_stmt._get_embedded_bindparams() if b.key in mappings[0] ] if use_orm_insert_stmt is not None else () ) records = ( ( None, state_dict, params, mapper, connection, value_params, has_all_pks, has_all_defaults, ) for ( state, state_dict, params, mp, conn, value_params, has_all_pks, has_all_defaults, ) in persistence._collect_insert_commands( table, ((None, mapping, mapper, connection) for mapping in mappings), bulk=True, return_defaults=bookkeeping, render_nulls=render_nulls, include_bulk_keys=extra_bp_names, ) ) result = persistence._emit_insert_statements( base_mapper, None, super_mapper, table, records, bookkeeping=bookkeeping, use_orm_insert_stmt=use_orm_insert_stmt, execution_options=execution_options, ) if use_orm_insert_stmt is not None: if not use_orm_insert_stmt._returning or return_result is None: return_result = result elif result.returns_rows: assert bookkeeping return_result = return_result.splice_horizontally(result) if return_defaults and isstates: identity_cls = mapper._identity_class identity_props = [p.key for p in mapper._identity_key_props] for state, dict_ in states: state.key = ( identity_cls, tuple([dict_[key] for key in identity_props]), None, ) if use_orm_insert_stmt is not None: assert return_result is not None return return_result @overload def _bulk_update( mapper: Mapper[Any], mappings: Union[Iterable[InstanceState[_O]], Iterable[Dict[str, Any]]], session_transaction: SessionTransaction, *, isstates: bool, update_changed_only: bool, use_orm_update_stmt: Literal[None] = ..., enable_check_rowcount: bool = True, ) -> None: ... @overload def _bulk_update( mapper: Mapper[Any], mappings: Union[Iterable[InstanceState[_O]], Iterable[Dict[str, Any]]], session_transaction: SessionTransaction, *, isstates: bool, update_changed_only: bool, use_orm_update_stmt: Optional[dml.Update] = ..., enable_check_rowcount: bool = True, ) -> _result.Result[Any]: ... def _bulk_update( mapper: Mapper[Any], mappings: Union[Iterable[InstanceState[_O]], Iterable[Dict[str, Any]]], session_transaction: SessionTransaction, *, isstates: bool, update_changed_only: bool, use_orm_update_stmt: Optional[dml.Update] = None, enable_check_rowcount: bool = True, ) -> Optional[_result.Result[Any]]: base_mapper = mapper.base_mapper search_keys = mapper._primary_key_propkeys if mapper._version_id_prop: search_keys = {mapper._version_id_prop.key}.union(search_keys) def _changed_dict(mapper, state): return { k: v for k, v in state.dict.items() if k in state.committed_state or k in search_keys } if isstates: if update_changed_only: mappings = [_changed_dict(mapper, state) for state in mappings] else: mappings = [state.dict for state in mappings] else: mappings = [dict(m) for m in mappings] _expand_composites(mapper, mappings) if session_transaction.session.connection_callable: raise NotImplementedError( "connection_callable / per-instance sharding " "not supported in bulk_update()" ) connection = session_transaction.connection(base_mapper) # find bindparams in the statement. see _bulk_insert for similar # notes for the insert case extra_bp_names = ( [ b.key for b in use_orm_update_stmt._get_embedded_bindparams() if b.key in mappings[0] ] if use_orm_update_stmt is not None else () ) for table, super_mapper in base_mapper._sorted_tables.items(): if not mapper.isa(super_mapper) or table not in mapper._pks_by_table: continue records = persistence._collect_update_commands( None, table, ( ( None, mapping, mapper, connection, ( mapping[mapper._version_id_prop.key] if mapper._version_id_prop else None ), ) for mapping in mappings ), bulk=True, use_orm_update_stmt=use_orm_update_stmt, include_bulk_keys=extra_bp_names, ) persistence._emit_update_statements( base_mapper, None, super_mapper, table, records, bookkeeping=False, use_orm_update_stmt=use_orm_update_stmt, enable_check_rowcount=enable_check_rowcount, ) if use_orm_update_stmt is not None: return _result.null_result() def _expand_composites(mapper, mappings): composite_attrs = mapper.composites if not composite_attrs: return composite_keys = set(composite_attrs.keys()) populators = { key: composite_attrs[key]._populate_composite_bulk_save_mappings_fn() for key in composite_keys } for mapping in mappings: for key in composite_keys.intersection(mapping): populators[key](mapping) class ORMDMLState(AbstractORMCompileState): is_dml_returning = True from_statement_ctx: Optional[ORMFromStatementCompileState] = None @classmethod def _get_orm_crud_kv_pairs( cls, mapper, statement, kv_iterator, needs_to_be_cacheable ): core_get_crud_kv_pairs = UpdateDMLState._get_crud_kv_pairs for k, v in kv_iterator: k = coercions.expect(roles.DMLColumnRole, k) if isinstance(k, str): desc = _entity_namespace_key(mapper, k, default=NO_VALUE) if desc is NO_VALUE: yield ( coercions.expect(roles.DMLColumnRole, k), ( coercions.expect( roles.ExpressionElementRole, v, type_=sqltypes.NullType(), is_crud=True, ) if needs_to_be_cacheable else v ), ) else: yield from core_get_crud_kv_pairs( statement, desc._bulk_update_tuples(v), needs_to_be_cacheable, ) elif "entity_namespace" in k._annotations: k_anno = k._annotations attr = _entity_namespace_key( k_anno["entity_namespace"], k_anno["proxy_key"] ) yield from core_get_crud_kv_pairs( statement, attr._bulk_update_tuples(v), needs_to_be_cacheable, ) else: yield ( k, ( v if not needs_to_be_cacheable else coercions.expect( roles.ExpressionElementRole, v, type_=sqltypes.NullType(), is_crud=True, ) ), ) @classmethod def _get_multi_crud_kv_pairs(cls, statement, kv_iterator): plugin_subject = statement._propagate_attrs["plugin_subject"] if not plugin_subject or not plugin_subject.mapper: return UpdateDMLState._get_multi_crud_kv_pairs( statement, kv_iterator ) return [ dict( cls._get_orm_crud_kv_pairs( plugin_subject.mapper, statement, value_dict.items(), False ) ) for value_dict in kv_iterator ] @classmethod def _get_crud_kv_pairs(cls, statement, kv_iterator, needs_to_be_cacheable): assert ( needs_to_be_cacheable ), "no test coverage for needs_to_be_cacheable=False" plugin_subject = statement._propagate_attrs["plugin_subject"] if not plugin_subject or not plugin_subject.mapper: return UpdateDMLState._get_crud_kv_pairs( statement, kv_iterator, needs_to_be_cacheable ) return list( cls._get_orm_crud_kv_pairs( plugin_subject.mapper, statement, kv_iterator, needs_to_be_cacheable, ) ) @classmethod def get_entity_description(cls, statement): ext_info = statement.table._annotations["parententity"] mapper = ext_info.mapper if ext_info.is_aliased_class: _label_name = ext_info.name else: _label_name = mapper.class_.__name__ return { "name": _label_name, "type": mapper.class_, "expr": ext_info.entity, "entity": ext_info.entity, "table": mapper.local_table, } @classmethod def get_returning_column_descriptions(cls, statement): def _ent_for_col(c): return c._annotations.get("parententity", None) def _attr_for_col(c, ent): if ent is None: return c proxy_key = c._annotations.get("proxy_key", None) if not proxy_key: return c else: return getattr(ent.entity, proxy_key, c) return [ { "name": c.key, "type": c.type, "expr": _attr_for_col(c, ent), "aliased": ent.is_aliased_class, "entity": ent.entity, } for c, ent in [ (c, _ent_for_col(c)) for c in statement._all_selected_columns ] ] def _setup_orm_returning( self, compiler, orm_level_statement, dml_level_statement, dml_mapper, *, use_supplemental_cols=True, ): """establish ORM column handlers for an INSERT, UPDATE, or DELETE which uses explicit returning(). called within compilation level create_for_statement. The _return_orm_returning() method then receives the Result after the statement was executed, and applies ORM loading to the state that we first established here. """ if orm_level_statement._returning: fs = FromStatement( orm_level_statement._returning, dml_level_statement, _adapt_on_names=False, ) fs = fs.execution_options(**orm_level_statement._execution_options) fs = fs.options(*orm_level_statement._with_options) self.select_statement = fs self.from_statement_ctx = fsc = ( ORMFromStatementCompileState.create_for_statement(fs, compiler) ) fsc.setup_dml_returning_compile_state(dml_mapper) dml_level_statement = dml_level_statement._generate() dml_level_statement._returning = () cols_to_return = [c for c in fsc.primary_columns if c is not None] # since we are splicing result sets together, make sure there # are columns of some kind returned in each result set if not cols_to_return: cols_to_return.extend(dml_mapper.primary_key) if use_supplemental_cols: dml_level_statement = dml_level_statement.return_defaults( # this is a little weird looking, but by passing # primary key as the main list of cols, this tells # return_defaults to omit server-default cols (and # actually all cols, due to some weird thing we should # clean up in crud.py). # Since we have cols_to_return, just return what we asked # for (plus primary key, which ORM persistence needs since # we likely set bookkeeping=True here, which is another # whole thing...). We dont want to clutter the # statement up with lots of other cols the user didn't # ask for. see #9685 *dml_mapper.primary_key, supplemental_cols=cols_to_return, ) else: dml_level_statement = dml_level_statement.returning( *cols_to_return ) return dml_level_statement @classmethod def _return_orm_returning( cls, session, statement, params, execution_options, bind_arguments, result, ): execution_context = result.context compile_state = execution_context.compiled.compile_state if ( compile_state.from_statement_ctx and not compile_state.from_statement_ctx.compile_options._is_star ): load_options = execution_options.get( "_sa_orm_load_options", QueryContext.default_load_options ) querycontext = QueryContext( compile_state.from_statement_ctx, compile_state.select_statement, params, session, load_options, execution_options, bind_arguments, ) return loading.instances(result, querycontext) else: return result class BulkUDCompileState(ORMDMLState): class default_update_options(Options): _dml_strategy: DMLStrategyArgument = "auto" _synchronize_session: SynchronizeSessionArgument = "auto" _can_use_returning: bool = False _is_delete_using: bool = False _is_update_from: bool = False _autoflush: bool = True _subject_mapper: Optional[Mapper[Any]] = None _resolved_values = EMPTY_DICT _eval_condition = None _matched_rows = None _identity_token = None @classmethod def can_use_returning( cls, dialect: Dialect, mapper: Mapper[Any], *, is_multitable: bool = False, is_update_from: bool = False, is_delete_using: bool = False, is_executemany: bool = False, ) -> bool: raise NotImplementedError() @classmethod def orm_pre_session_exec( cls, session, statement, params, execution_options, bind_arguments, is_pre_event, ): ( update_options, execution_options, ) = BulkUDCompileState.default_update_options.from_execution_options( "_sa_orm_update_options", { "synchronize_session", "autoflush", "identity_token", "is_delete_using", "is_update_from", "dml_strategy", }, execution_options, statement._execution_options, ) bind_arguments["clause"] = statement try: plugin_subject = statement._propagate_attrs["plugin_subject"] except KeyError: assert False, "statement had 'orm' plugin but no plugin_subject" else: if plugin_subject: bind_arguments["mapper"] = plugin_subject.mapper update_options += {"_subject_mapper": plugin_subject.mapper} if "parententity" not in statement.table._annotations: update_options += {"_dml_strategy": "core_only"} elif not isinstance(params, list): if update_options._dml_strategy == "auto": update_options += {"_dml_strategy": "orm"} elif update_options._dml_strategy == "bulk": raise sa_exc.InvalidRequestError( 'Can\'t use "bulk" ORM insert strategy without ' "passing separate parameters" ) else: if update_options._dml_strategy == "auto": update_options += {"_dml_strategy": "bulk"} sync = update_options._synchronize_session if sync is not None: if sync not in ("auto", "evaluate", "fetch", False): raise sa_exc.ArgumentError( "Valid strategies for session synchronization " "are 'auto', 'evaluate', 'fetch', False" ) if update_options._dml_strategy == "bulk" and sync == "fetch": raise sa_exc.InvalidRequestError( "The 'fetch' synchronization strategy is not available " "for 'bulk' ORM updates (i.e. multiple parameter sets)" ) if not is_pre_event: if update_options._autoflush: session._autoflush() if update_options._dml_strategy == "orm": if update_options._synchronize_session == "auto": update_options = cls._do_pre_synchronize_auto( session, statement, params, execution_options, bind_arguments, update_options, ) elif update_options._synchronize_session == "evaluate": update_options = cls._do_pre_synchronize_evaluate( session, statement, params, execution_options, bind_arguments, update_options, ) elif update_options._synchronize_session == "fetch": update_options = cls._do_pre_synchronize_fetch( session, statement, params, execution_options, bind_arguments, update_options, ) elif update_options._dml_strategy == "bulk": if update_options._synchronize_session == "auto": update_options += {"_synchronize_session": "evaluate"} # indicators from the "pre exec" step that are then # added to the DML statement, which will also be part of the cache # key. The compile level create_for_statement() method will then # consume these at compiler time. statement = statement._annotate( { "synchronize_session": update_options._synchronize_session, "is_delete_using": update_options._is_delete_using, "is_update_from": update_options._is_update_from, "dml_strategy": update_options._dml_strategy, "can_use_returning": update_options._can_use_returning, } ) return ( statement, util.immutabledict(execution_options).union( {"_sa_orm_update_options": update_options} ), ) @classmethod def orm_setup_cursor_result( cls, session, statement, params, execution_options, bind_arguments, result, ): # this stage of the execution is called after the # do_orm_execute event hook. meaning for an extension like # horizontal sharding, this step happens *within* the horizontal # sharding event handler which calls session.execute() re-entrantly # and will occur for each backend individually. # the sharding extension then returns its own merged result from the # individual ones we return here. update_options = execution_options["_sa_orm_update_options"] if update_options._dml_strategy == "orm": if update_options._synchronize_session == "evaluate": cls._do_post_synchronize_evaluate( session, statement, result, update_options ) elif update_options._synchronize_session == "fetch": cls._do_post_synchronize_fetch( session, statement, result, update_options ) elif update_options._dml_strategy == "bulk": if update_options._synchronize_session == "evaluate": cls._do_post_synchronize_bulk_evaluate( session, params, result, update_options ) return result return cls._return_orm_returning( session, statement, params, execution_options, bind_arguments, result, ) @classmethod def _adjust_for_extra_criteria(cls, global_attributes, ext_info): """Apply extra criteria filtering. For all distinct single-table-inheritance mappers represented in the table being updated or deleted, produce additional WHERE criteria such that only the appropriate subtypes are selected from the total results. Additionally, add WHERE criteria originating from LoaderCriteriaOptions collected from the statement. """ return_crit = () adapter = ext_info._adapter if ext_info.is_aliased_class else None if ( "additional_entity_criteria", ext_info.mapper, ) in global_attributes: return_crit += tuple( ae._resolve_where_criteria(ext_info) for ae in global_attributes[ ("additional_entity_criteria", ext_info.mapper) ] if ae.include_aliases or ae.entity is ext_info ) if ext_info.mapper._single_table_criterion is not None: return_crit += (ext_info.mapper._single_table_criterion,) if adapter: return_crit = tuple(adapter.traverse(crit) for crit in return_crit) return return_crit @classmethod def _interpret_returning_rows(cls, mapper, rows): """translate from local inherited table columns to base mapper primary key columns. Joined inheritance mappers always establish the primary key in terms of the base table. When we UPDATE a sub-table, we can only get RETURNING for the sub-table's columns. Here, we create a lookup from the local sub table's primary key columns to the base table PK columns so that we can get identity key values from RETURNING that's against the joined inheritance sub-table. the complexity here is to support more than one level deep of inheritance, where we have to link columns to each other across the inheritance hierarchy. """ if mapper.local_table is not mapper.base_mapper.local_table: return rows # this starts as a mapping of # local_pk_col: local_pk_col. # we will then iteratively rewrite the "value" of the dict with # each successive superclass column local_pk_to_base_pk = {pk: pk for pk in mapper.local_table.primary_key} for mp in mapper.iterate_to_root(): if mp.inherits is None: break elif mp.local_table is mp.inherits.local_table: continue t_to_e = dict(mp._table_to_equated[mp.inherits.local_table]) col_to_col = {sub_pk: super_pk for super_pk, sub_pk in t_to_e[mp]} for pk, super_ in local_pk_to_base_pk.items(): local_pk_to_base_pk[pk] = col_to_col[super_] lookup = { local_pk_to_base_pk[lpk]: idx for idx, lpk in enumerate(mapper.local_table.primary_key) } primary_key_convert = [ lookup[bpk] for bpk in mapper.base_mapper.primary_key ] return [tuple(row[idx] for idx in primary_key_convert) for row in rows] @classmethod def _get_matched_objects_on_criteria(cls, update_options, states): mapper = update_options._subject_mapper eval_condition = update_options._eval_condition raw_data = [ (state.obj(), state, state.dict) for state in states if state.mapper.isa(mapper) and not state.expired ] identity_token = update_options._identity_token if identity_token is not None: raw_data = [ (obj, state, dict_) for obj, state, dict_ in raw_data if state.identity_token == identity_token ] result = [] for obj, state, dict_ in raw_data: evaled_condition = eval_condition(obj) # caution: don't use "in ()" or == here, _EXPIRE_OBJECT # evaluates as True for all comparisons if ( evaled_condition is True or evaled_condition is evaluator._EXPIRED_OBJECT ): result.append( ( obj, state, dict_, evaled_condition is evaluator._EXPIRED_OBJECT, ) ) return result @classmethod def _eval_condition_from_statement(cls, update_options, statement): mapper = update_options._subject_mapper target_cls = mapper.class_ evaluator_compiler = evaluator._EvaluatorCompiler(target_cls) crit = () if statement._where_criteria: crit += statement._where_criteria global_attributes = {} for opt in statement._with_options: if opt._is_criteria_option: opt.get_global_criteria(global_attributes) if global_attributes: crit += cls._adjust_for_extra_criteria(global_attributes, mapper) if crit: eval_condition = evaluator_compiler.process(*crit) else: # workaround for mypy https://github.com/python/mypy/issues/14027 def _eval_condition(obj): return True eval_condition = _eval_condition return eval_condition @classmethod def _do_pre_synchronize_auto( cls, session, statement, params, execution_options, bind_arguments, update_options, ): """setup auto sync strategy "auto" checks if we can use "evaluate" first, then falls back to "fetch" evaluate is vastly more efficient for the common case where session is empty, only has a few objects, and the UPDATE statement can potentially match thousands/millions of rows. OTOH more complex criteria that fails to work with "evaluate" we would hope usually correlates with fewer net rows. """ try: eval_condition = cls._eval_condition_from_statement( update_options, statement ) except evaluator.UnevaluatableError: pass else: return update_options + { "_eval_condition": eval_condition, "_synchronize_session": "evaluate", } update_options += {"_synchronize_session": "fetch"} return cls._do_pre_synchronize_fetch( session, statement, params, execution_options, bind_arguments, update_options, ) @classmethod def _do_pre_synchronize_evaluate( cls, session, statement, params, execution_options, bind_arguments, update_options, ): try: eval_condition = cls._eval_condition_from_statement( update_options, statement ) except evaluator.UnevaluatableError as err: raise sa_exc.InvalidRequestError( 'Could not evaluate current criteria in Python: "%s". ' "Specify 'fetch' or False for the " "synchronize_session execution option." % err ) from err return update_options + { "_eval_condition": eval_condition, } @classmethod def _get_resolved_values(cls, mapper, statement): if statement._multi_values: return [] elif statement._ordered_values: return list(statement._ordered_values) elif statement._values: return list(statement._values.items()) else: return [] @classmethod def _resolved_keys_as_propnames(cls, mapper, resolved_values): values = [] for k, v in resolved_values: if mapper and isinstance(k, expression.ColumnElement): try: attr = mapper._columntoproperty[k] except orm_exc.UnmappedColumnError: pass else: values.append((attr.key, v)) else: raise sa_exc.InvalidRequestError( "Attribute name not found, can't be " "synchronized back to objects: %r" % k ) return values @classmethod def _do_pre_synchronize_fetch( cls, session, statement, params, execution_options, bind_arguments, update_options, ): mapper = update_options._subject_mapper select_stmt = ( select(*(mapper.primary_key + (mapper.select_identity_token,))) .select_from(mapper) .options(*statement._with_options) ) select_stmt._where_criteria = statement._where_criteria # conditionally run the SELECT statement for pre-fetch, testing the # "bind" for if we can use RETURNING or not using the do_orm_execute # event. If RETURNING is available, the do_orm_execute event # will cancel the SELECT from being actually run. # # The way this is organized seems strange, why don't we just # call can_use_returning() before invoking the statement and get # answer?, why does this go through the whole execute phase using an # event? Answer: because we are integrating with extensions such # as the horizontal sharding extention that "multiplexes" an individual # statement run through multiple engines, and it uses # do_orm_execute() to do that. can_use_returning = None def skip_for_returning(orm_context: ORMExecuteState) -> Any: bind = orm_context.session.get_bind(**orm_context.bind_arguments) nonlocal can_use_returning per_bind_result = cls.can_use_returning( bind.dialect, mapper, is_update_from=update_options._is_update_from, is_delete_using=update_options._is_delete_using, is_executemany=orm_context.is_executemany, ) if can_use_returning is not None: if can_use_returning != per_bind_result: raise sa_exc.InvalidRequestError( "For synchronize_session='fetch', can't mix multiple " "backends where some support RETURNING and others " "don't" ) elif orm_context.is_executemany and not per_bind_result: raise sa_exc.InvalidRequestError( "For synchronize_session='fetch', can't use multiple " "parameter sets in ORM mode, which this backend does not " "support with RETURNING" ) else: can_use_returning = per_bind_result if per_bind_result: return _result.null_result() else: return None result = session.execute( select_stmt, params, execution_options=execution_options, bind_arguments=bind_arguments, _add_event=skip_for_returning, ) matched_rows = result.fetchall() return update_options + { "_matched_rows": matched_rows, "_can_use_returning": can_use_returning, } @CompileState.plugin_for("orm", "insert") class BulkORMInsert(ORMDMLState, InsertDMLState): class default_insert_options(Options): _dml_strategy: DMLStrategyArgument = "auto" _render_nulls: bool = False _return_defaults: bool = False _subject_mapper: Optional[Mapper[Any]] = None _autoflush: bool = True _populate_existing: bool = False select_statement: Optional[FromStatement] = None @classmethod def orm_pre_session_exec( cls, session, statement, params, execution_options, bind_arguments, is_pre_event, ): ( insert_options, execution_options, ) = BulkORMInsert.default_insert_options.from_execution_options( "_sa_orm_insert_options", {"dml_strategy", "autoflush", "populate_existing", "render_nulls"}, execution_options, statement._execution_options, ) bind_arguments["clause"] = statement try: plugin_subject = statement._propagate_attrs["plugin_subject"] except KeyError: assert False, "statement had 'orm' plugin but no plugin_subject" else: if plugin_subject: bind_arguments["mapper"] = plugin_subject.mapper insert_options += {"_subject_mapper": plugin_subject.mapper} if not params: if insert_options._dml_strategy == "auto": insert_options += {"_dml_strategy": "orm"} elif insert_options._dml_strategy == "bulk": raise sa_exc.InvalidRequestError( 'Can\'t use "bulk" ORM insert strategy without ' "passing separate parameters" ) else: if insert_options._dml_strategy == "auto": insert_options += {"_dml_strategy": "bulk"} if insert_options._dml_strategy != "raw": # for ORM object loading, like ORMContext, we have to disable # result set adapt_to_context, because we will be generating a # new statement with specific columns that's cached inside of # an ORMFromStatementCompileState, which we will re-use for # each result. if not execution_options: execution_options = context._orm_load_exec_options else: execution_options = execution_options.union( context._orm_load_exec_options ) if not is_pre_event and insert_options._autoflush: session._autoflush() statement = statement._annotate( {"dml_strategy": insert_options._dml_strategy} ) return ( statement, util.immutabledict(execution_options).union( {"_sa_orm_insert_options": insert_options} ), ) @classmethod def orm_execute_statement( cls, session: Session, statement: dml.Insert, params: _CoreAnyExecuteParams, execution_options: OrmExecuteOptionsParameter, bind_arguments: _BindArguments, conn: Connection, ) -> _result.Result: insert_options = execution_options.get( "_sa_orm_insert_options", cls.default_insert_options ) if insert_options._dml_strategy not in ( "raw", "bulk", "orm", "auto", ): raise sa_exc.ArgumentError( "Valid strategies for ORM insert strategy " "are 'raw', 'orm', 'bulk', 'auto" ) result: _result.Result[Any] if insert_options._dml_strategy == "raw": result = conn.execute( statement, params or {}, execution_options=execution_options ) return result if insert_options._dml_strategy == "bulk": mapper = insert_options._subject_mapper if ( statement._post_values_clause is not None and mapper._multiple_persistence_tables ): raise sa_exc.InvalidRequestError( "bulk INSERT with a 'post values' clause " "(typically upsert) not supported for multi-table " f"mapper {mapper}" ) assert mapper is not None assert session._transaction is not None result = _bulk_insert( mapper, cast( "Iterable[Dict[str, Any]]", [params] if isinstance(params, dict) else params, ), session._transaction, isstates=False, return_defaults=insert_options._return_defaults, render_nulls=insert_options._render_nulls, use_orm_insert_stmt=statement, execution_options=execution_options, ) elif insert_options._dml_strategy == "orm": result = conn.execute( statement, params or {}, execution_options=execution_options ) else: raise AssertionError() if not bool(statement._returning): return result if insert_options._populate_existing: load_options = execution_options.get( "_sa_orm_load_options", QueryContext.default_load_options ) load_options += {"_populate_existing": True} execution_options = execution_options.union( {"_sa_orm_load_options": load_options} ) return cls._return_orm_returning( session, statement, params, execution_options, bind_arguments, result, ) @classmethod def create_for_statement(cls, statement, compiler, **kw) -> BulkORMInsert: self = cast( BulkORMInsert, super().create_for_statement(statement, compiler, **kw), ) if compiler is not None: toplevel = not compiler.stack else: toplevel = True if not toplevel: return self mapper = statement._propagate_attrs["plugin_subject"] dml_strategy = statement._annotations.get("dml_strategy", "raw") if dml_strategy == "bulk": self._setup_for_bulk_insert(compiler) elif dml_strategy == "orm": self._setup_for_orm_insert(compiler, mapper) return self @classmethod def _resolved_keys_as_col_keys(cls, mapper, resolved_value_dict): return { col.key if col is not None else k: v for col, k, v in ( (mapper.c.get(k), k, v) for k, v in resolved_value_dict.items() ) } def _setup_for_orm_insert(self, compiler, mapper): statement = orm_level_statement = cast(dml.Insert, self.statement) statement = self._setup_orm_returning( compiler, orm_level_statement, statement, dml_mapper=mapper, use_supplemental_cols=False, ) self.statement = statement def _setup_for_bulk_insert(self, compiler): """establish an INSERT statement within the context of bulk insert. This method will be within the "conn.execute()" call that is invoked by persistence._emit_insert_statement(). """ statement = orm_level_statement = cast(dml.Insert, self.statement) an = statement._annotations emit_insert_table, emit_insert_mapper = ( an["_emit_insert_table"], an["_emit_insert_mapper"], ) statement = statement._clone() statement.table = emit_insert_table if self._dict_parameters: self._dict_parameters = { col: val for col, val in self._dict_parameters.items() if col.table is emit_insert_table } statement = self._setup_orm_returning( compiler, orm_level_statement, statement, dml_mapper=emit_insert_mapper, use_supplemental_cols=True, ) if ( self.from_statement_ctx is not None and self.from_statement_ctx.compile_options._is_star ): raise sa_exc.CompileError( "Can't use RETURNING * with bulk ORM INSERT. " "Please use a different INSERT form, such as INSERT..VALUES " "or INSERT with a Core Connection" ) self.statement = statement @CompileState.plugin_for("orm", "update") class BulkORMUpdate(BulkUDCompileState, UpdateDMLState): @classmethod def create_for_statement(cls, statement, compiler, **kw): self = cls.__new__(cls) dml_strategy = statement._annotations.get( "dml_strategy", "unspecified" ) toplevel = not compiler.stack if toplevel and dml_strategy == "bulk": self._setup_for_bulk_update(statement, compiler) elif ( dml_strategy == "core_only" or dml_strategy == "unspecified" and "parententity" not in statement.table._annotations ): UpdateDMLState.__init__(self, statement, compiler, **kw) elif not toplevel or dml_strategy in ("orm", "unspecified"): self._setup_for_orm_update(statement, compiler) return self def _setup_for_orm_update(self, statement, compiler, **kw): orm_level_statement = statement toplevel = not compiler.stack ext_info = statement.table._annotations["parententity"] self.mapper = mapper = ext_info.mapper self._resolved_values = self._get_resolved_values(mapper, statement) self._init_global_attributes( statement, compiler, toplevel=toplevel, process_criteria_for_toplevel=toplevel, ) if statement._values: self._resolved_values = dict(self._resolved_values) new_stmt = statement._clone() if new_stmt.table._annotations["parententity"] is mapper: new_stmt.table = mapper.local_table # note if the statement has _multi_values, these # are passed through to the new statement, which will then raise # InvalidRequestError because UPDATE doesn't support multi_values # right now. if statement._ordered_values: new_stmt._ordered_values = self._resolved_values elif statement._values: new_stmt._values = self._resolved_values new_crit = self._adjust_for_extra_criteria( self.global_attributes, mapper ) if new_crit: new_stmt = new_stmt.where(*new_crit) # if we are against a lambda statement we might not be the # topmost object that received per-execute annotations # do this first as we need to determine if there is # UPDATE..FROM UpdateDMLState.__init__(self, new_stmt, compiler, **kw) use_supplemental_cols = False if not toplevel: synchronize_session = None else: synchronize_session = compiler._annotations.get( "synchronize_session", None ) can_use_returning = compiler._annotations.get( "can_use_returning", None ) if can_use_returning is not False: # even though pre_exec has determined basic # can_use_returning for the dialect, if we are to use # RETURNING we need to run can_use_returning() at this level # unconditionally because is_delete_using was not known # at the pre_exec level can_use_returning = ( synchronize_session == "fetch" and self.can_use_returning( compiler.dialect, mapper, is_multitable=self.is_multitable ) ) if synchronize_session == "fetch" and can_use_returning: use_supplemental_cols = True # NOTE: we might want to RETURNING the actual columns to be # synchronized also. however this is complicated and difficult # to align against the behavior of "evaluate". Additionally, # in a large number (if not the majority) of cases, we have the # "evaluate" answer, usually a fixed value, in memory already and # there's no need to re-fetch the same value # over and over again. so perhaps if it could be RETURNING just # the elements that were based on a SQL expression and not # a constant. For now it doesn't quite seem worth it new_stmt = new_stmt.return_defaults(*new_stmt.table.primary_key) if toplevel: new_stmt = self._setup_orm_returning( compiler, orm_level_statement, new_stmt, dml_mapper=mapper, use_supplemental_cols=use_supplemental_cols, ) self.statement = new_stmt def _setup_for_bulk_update(self, statement, compiler, **kw): """establish an UPDATE statement within the context of bulk insert. This method will be within the "conn.execute()" call that is invoked by persistence._emit_update_statement(). """ statement = cast(dml.Update, statement) an = statement._annotations emit_update_table, _ = ( an["_emit_update_table"], an["_emit_update_mapper"], ) statement = statement._clone() statement.table = emit_update_table UpdateDMLState.__init__(self, statement, compiler, **kw) if self._ordered_values: raise sa_exc.InvalidRequestError( "bulk ORM UPDATE does not support ordered_values() for " "custom UPDATE statements with bulk parameter sets. Use a " "non-bulk UPDATE statement or use values()." ) if self._dict_parameters: self._dict_parameters = { col: val for col, val in self._dict_parameters.items() if col.table is emit_update_table } self.statement = statement @classmethod def orm_execute_statement( cls, session: Session, statement: dml.Update, params: _CoreAnyExecuteParams, execution_options: OrmExecuteOptionsParameter, bind_arguments: _BindArguments, conn: Connection, ) -> _result.Result: update_options = execution_options.get( "_sa_orm_update_options", cls.default_update_options ) if update_options._dml_strategy not in ( "orm", "auto", "bulk", "core_only", ): raise sa_exc.ArgumentError( "Valid strategies for ORM UPDATE strategy " "are 'orm', 'auto', 'bulk', 'core_only'" ) result: _result.Result[Any] if update_options._dml_strategy == "bulk": enable_check_rowcount = not statement._where_criteria assert update_options._synchronize_session != "fetch" if ( statement._where_criteria and update_options._synchronize_session == "evaluate" ): raise sa_exc.InvalidRequestError( "bulk synchronize of persistent objects not supported " "when using bulk update with additional WHERE " "criteria right now. add synchronize_session=None " "execution option to bypass synchronize of persistent " "objects." ) mapper = update_options._subject_mapper assert mapper is not None assert session._transaction is not None result = _bulk_update( mapper, cast( "Iterable[Dict[str, Any]]", [params] if isinstance(params, dict) else params, ), session._transaction, isstates=False, update_changed_only=False, use_orm_update_stmt=statement, enable_check_rowcount=enable_check_rowcount, ) return cls.orm_setup_cursor_result( session, statement, params, execution_options, bind_arguments, result, ) else: return super().orm_execute_statement( session, statement, params, execution_options, bind_arguments, conn, ) @classmethod def can_use_returning( cls, dialect: Dialect, mapper: Mapper[Any], *, is_multitable: bool = False, is_update_from: bool = False, is_delete_using: bool = False, is_executemany: bool = False, ) -> bool: # normal answer for "should we use RETURNING" at all. normal_answer = ( dialect.update_returning and mapper.local_table.implicit_returning ) if not normal_answer: return False if is_executemany: return dialect.update_executemany_returning # these workarounds are currently hypothetical for UPDATE, # unlike DELETE where they impact MariaDB if is_update_from: return dialect.update_returning_multifrom elif is_multitable and not dialect.update_returning_multifrom: raise sa_exc.CompileError( f'Dialect "{dialect.name}" does not support RETURNING ' "with UPDATE..FROM; for synchronize_session='fetch', " "please add the additional execution option " "'is_update_from=True' to the statement to indicate that " "a separate SELECT should be used for this backend." ) return True @classmethod def _do_post_synchronize_bulk_evaluate( cls, session, params, result, update_options ): if not params: return mapper = update_options._subject_mapper pk_keys = [prop.key for prop in mapper._identity_key_props] identity_map = session.identity_map for param in params: identity_key = mapper.identity_key_from_primary_key( (param[key] for key in pk_keys), update_options._identity_token, ) state = identity_map.fast_get_state(identity_key) if not state: continue evaluated_keys = set(param).difference(pk_keys) dict_ = state.dict # only evaluate unmodified attributes to_evaluate = state.unmodified.intersection(evaluated_keys) for key in to_evaluate: if key in dict_: dict_[key] = param[key] state.manager.dispatch.refresh(state, None, to_evaluate) state._commit(dict_, list(to_evaluate)) # attributes that were formerly modified instead get expired. # this only gets hit if the session had pending changes # and autoflush were set to False. to_expire = evaluated_keys.intersection(dict_).difference( to_evaluate ) if to_expire: state._expire_attributes(dict_, to_expire) @classmethod def _do_post_synchronize_evaluate( cls, session, statement, result, update_options ): matched_objects = cls._get_matched_objects_on_criteria( update_options, session.identity_map.all_states(), ) cls._apply_update_set_values_to_objects( session, update_options, statement, [(obj, state, dict_) for obj, state, dict_, _ in matched_objects], ) @classmethod def _do_post_synchronize_fetch( cls, session, statement, result, update_options ): target_mapper = update_options._subject_mapper returned_defaults_rows = result.returned_defaults_rows if returned_defaults_rows: pk_rows = cls._interpret_returning_rows( target_mapper, returned_defaults_rows ) matched_rows = [ tuple(row) + (update_options._identity_token,) for row in pk_rows ] else: matched_rows = update_options._matched_rows objs = [ session.identity_map[identity_key] for identity_key in [ target_mapper.identity_key_from_primary_key( list(primary_key), identity_token=identity_token, ) for primary_key, identity_token in [ (row[0:-1], row[-1]) for row in matched_rows ] if update_options._identity_token is None or identity_token == update_options._identity_token ] if identity_key in session.identity_map ] if not objs: return cls._apply_update_set_values_to_objects( session, update_options, statement, [ ( obj, attributes.instance_state(obj), attributes.instance_dict(obj), ) for obj in objs ], ) @classmethod def _apply_update_set_values_to_objects( cls, session, update_options, statement, matched_objects ): """apply values to objects derived from an update statement, e.g. UPDATE..SET <values> """ mapper = update_options._subject_mapper target_cls = mapper.class_ evaluator_compiler = evaluator._EvaluatorCompiler(target_cls) resolved_values = cls._get_resolved_values(mapper, statement) resolved_keys_as_propnames = cls._resolved_keys_as_propnames( mapper, resolved_values ) value_evaluators = {} for key, value in resolved_keys_as_propnames: try: _evaluator = evaluator_compiler.process( coercions.expect(roles.ExpressionElementRole, value) ) except evaluator.UnevaluatableError: pass else: value_evaluators[key] = _evaluator evaluated_keys = list(value_evaluators.keys()) attrib = {k for k, v in resolved_keys_as_propnames} states = set() for obj, state, dict_ in matched_objects: to_evaluate = state.unmodified.intersection(evaluated_keys) for key in to_evaluate: if key in dict_: # only run eval for attributes that are present. dict_[key] = value_evaluators[key](obj) state.manager.dispatch.refresh(state, None, to_evaluate) state._commit(dict_, list(to_evaluate)) # attributes that were formerly modified instead get expired. # this only gets hit if the session had pending changes # and autoflush were set to False. to_expire = attrib.intersection(dict_).difference(to_evaluate) if to_expire: state._expire_attributes(dict_, to_expire) states.add(state) session._register_altered(states) @CompileState.plugin_for("orm", "delete") class BulkORMDelete(BulkUDCompileState, DeleteDMLState): @classmethod def create_for_statement(cls, statement, compiler, **kw): self = cls.__new__(cls) dml_strategy = statement._annotations.get( "dml_strategy", "unspecified" ) if ( dml_strategy == "core_only" or dml_strategy == "unspecified" and "parententity" not in statement.table._annotations ): DeleteDMLState.__init__(self, statement, compiler, **kw) return self toplevel = not compiler.stack orm_level_statement = statement ext_info = statement.table._annotations["parententity"] self.mapper = mapper = ext_info.mapper self._init_global_attributes( statement, compiler, toplevel=toplevel, process_criteria_for_toplevel=toplevel, ) new_stmt = statement._clone() if new_stmt.table._annotations["parententity"] is mapper: new_stmt.table = mapper.local_table new_crit = cls._adjust_for_extra_criteria( self.global_attributes, mapper ) if new_crit: new_stmt = new_stmt.where(*new_crit) # do this first as we need to determine if there is # DELETE..FROM DeleteDMLState.__init__(self, new_stmt, compiler, **kw) use_supplemental_cols = False if not toplevel: synchronize_session = None else: synchronize_session = compiler._annotations.get( "synchronize_session", None ) can_use_returning = compiler._annotations.get( "can_use_returning", None ) if can_use_returning is not False: # even though pre_exec has determined basic # can_use_returning for the dialect, if we are to use # RETURNING we need to run can_use_returning() at this level # unconditionally because is_delete_using was not known # at the pre_exec level can_use_returning = ( synchronize_session == "fetch" and self.can_use_returning( compiler.dialect, mapper, is_multitable=self.is_multitable, is_delete_using=compiler._annotations.get( "is_delete_using", False ), ) ) if can_use_returning: use_supplemental_cols = True new_stmt = new_stmt.return_defaults(*new_stmt.table.primary_key) if toplevel: new_stmt = self._setup_orm_returning( compiler, orm_level_statement, new_stmt, dml_mapper=mapper, use_supplemental_cols=use_supplemental_cols, ) self.statement = new_stmt return self @classmethod def orm_execute_statement( cls, session: Session, statement: dml.Delete, params: _CoreAnyExecuteParams, execution_options: OrmExecuteOptionsParameter, bind_arguments: _BindArguments, conn: Connection, ) -> _result.Result: update_options = execution_options.get( "_sa_orm_update_options", cls.default_update_options ) if update_options._dml_strategy == "bulk": raise sa_exc.InvalidRequestError( "Bulk ORM DELETE not supported right now. " "Statement may be invoked at the " "Core level using " "session.connection().execute(stmt, parameters)" ) if update_options._dml_strategy not in ("orm", "auto", "core_only"): raise sa_exc.ArgumentError( "Valid strategies for ORM DELETE strategy are 'orm', 'auto', " "'core_only'" ) return super().orm_execute_statement( session, statement, params, execution_options, bind_arguments, conn ) @classmethod def can_use_returning( cls, dialect: Dialect, mapper: Mapper[Any], *, is_multitable: bool = False, is_update_from: bool = False, is_delete_using: bool = False, is_executemany: bool = False, ) -> bool: # normal answer for "should we use RETURNING" at all. normal_answer = ( dialect.delete_returning and mapper.local_table.implicit_returning ) if not normal_answer: return False # now get into special workarounds because MariaDB supports # DELETE...RETURNING but not DELETE...USING...RETURNING. if is_delete_using: # is_delete_using hint was passed. use # additional dialect feature (True for PG, False for MariaDB) return dialect.delete_returning_multifrom elif is_multitable and not dialect.delete_returning_multifrom: # is_delete_using hint was not passed, but we determined # at compile time that this is in fact a DELETE..USING. # it's too late to continue since we did not pre-SELECT. # raise that we need that hint up front. raise sa_exc.CompileError( f'Dialect "{dialect.name}" does not support RETURNING ' "with DELETE..USING; for synchronize_session='fetch', " "please add the additional execution option " "'is_delete_using=True' to the statement to indicate that " "a separate SELECT should be used for this backend." ) return True @classmethod def _do_post_synchronize_evaluate( cls, session, statement, result, update_options ): matched_objects = cls._get_matched_objects_on_criteria( update_options, session.identity_map.all_states(), ) to_delete = [] for _, state, dict_, is_partially_expired in matched_objects: if is_partially_expired: state._expire(dict_, session.identity_map._modified) else: to_delete.append(state) if to_delete: session._remove_newly_deleted(to_delete) @classmethod def _do_post_synchronize_fetch( cls, session, statement, result, update_options ): target_mapper = update_options._subject_mapper returned_defaults_rows = result.returned_defaults_rows if returned_defaults_rows: pk_rows = cls._interpret_returning_rows( target_mapper, returned_defaults_rows ) matched_rows = [ tuple(row) + (update_options._identity_token,) for row in pk_rows ] else: matched_rows = update_options._matched_rows for row in matched_rows: primary_key = row[0:-1] identity_token = row[-1] # TODO: inline this and call remove_newly_deleted # once identity_key = target_mapper.identity_key_from_primary_key( list(primary_key), identity_token=identity_token, ) if identity_key in session.identity_map: session._remove_newly_deleted( [ attributes.instance_state( session.identity_map[identity_key] ) ] )