mirror of
https://github.com/sqlalchemy/sqlalchemy.git
synced 2026-05-07 01:10:52 -04:00
08cef20f4a
Improve handling of two phase transaction identifiers for PostgreSQL when the identifier is provided by the user. As part of this change the psycopg dialect was updated to use the DBAPI two phase transaction API instead of executing the SQL directly. Fixes: #13229 Change-Id: If8301a7253b4a0c88e5323c9a052c3a9fa258780
4404 lines
138 KiB
Python
4404 lines
138 KiB
Python
import collections.abc as collections_abc
|
|
from contextlib import contextmanager
|
|
from contextlib import nullcontext
|
|
import copy
|
|
from io import StringIO
|
|
import re
|
|
import threading
|
|
from unittest import mock
|
|
from unittest.mock import call
|
|
from unittest.mock import Mock
|
|
from unittest.mock import patch
|
|
import weakref
|
|
|
|
import sqlalchemy as tsa
|
|
from sqlalchemy import bindparam
|
|
from sqlalchemy import create_engine
|
|
from sqlalchemy import create_mock_engine
|
|
from sqlalchemy import event
|
|
from sqlalchemy import func
|
|
from sqlalchemy import inspect
|
|
from sqlalchemy import INT
|
|
from sqlalchemy import Integer
|
|
from sqlalchemy import LargeBinary
|
|
from sqlalchemy import MetaData
|
|
from sqlalchemy import select
|
|
from sqlalchemy import Sequence
|
|
from sqlalchemy import String
|
|
from sqlalchemy import testing
|
|
from sqlalchemy import text
|
|
from sqlalchemy import TypeDecorator
|
|
from sqlalchemy import util
|
|
from sqlalchemy import VARCHAR
|
|
from sqlalchemy.connectors.asyncio import AsyncAdapt_dbapi_module
|
|
from sqlalchemy.engine import BindTyping
|
|
from sqlalchemy.engine.base import Connection
|
|
from sqlalchemy.engine.base import Engine
|
|
from sqlalchemy.pool import AsyncAdaptedQueuePool
|
|
from sqlalchemy.pool import NullPool
|
|
from sqlalchemy.pool import QueuePool
|
|
from sqlalchemy.sql import column
|
|
from sqlalchemy.sql import literal
|
|
from sqlalchemy.sql.elements import literal_column
|
|
from sqlalchemy.testing import assert_raises
|
|
from sqlalchemy.testing import assert_raises_message
|
|
from sqlalchemy.testing import config
|
|
from sqlalchemy.testing import engines
|
|
from sqlalchemy.testing import eq_
|
|
from sqlalchemy.testing import expect_raises_message
|
|
from sqlalchemy.testing import fixtures
|
|
from sqlalchemy.testing import is_
|
|
from sqlalchemy.testing import is_false
|
|
from sqlalchemy.testing import is_not
|
|
from sqlalchemy.testing import is_true
|
|
from sqlalchemy.testing import ne_
|
|
from sqlalchemy.testing import provision
|
|
from sqlalchemy.testing.assertions import expect_deprecated
|
|
from sqlalchemy.testing.assertsql import CompiledSQL
|
|
from sqlalchemy.testing.provision import normalize_sequence
|
|
from sqlalchemy.testing.schema import Column
|
|
from sqlalchemy.testing.schema import Table
|
|
from sqlalchemy.testing.util import gc_collect
|
|
from sqlalchemy.testing.util import picklers
|
|
|
|
|
|
class SomeException(Exception):
|
|
pass
|
|
|
|
|
|
class Foo:
|
|
def __str__(self):
|
|
return "foo"
|
|
|
|
def __unicode__(self):
|
|
return "fóó"
|
|
|
|
|
|
class ExecuteTest(fixtures.TablesTest):
|
|
|
|
@classmethod
|
|
def define_tables(cls, metadata):
|
|
Table(
|
|
"users",
|
|
metadata,
|
|
Column("user_id", INT, primary_key=True, autoincrement=False),
|
|
Column("user_name", VARCHAR(20)),
|
|
)
|
|
Table(
|
|
"users_autoinc",
|
|
metadata,
|
|
Column(
|
|
"user_id", INT, primary_key=True, test_needs_autoincrement=True
|
|
),
|
|
Column("user_name", VARCHAR(20)),
|
|
)
|
|
|
|
def test_no_strings(self, connection):
|
|
with expect_raises_message(
|
|
tsa.exc.ObjectNotExecutableError,
|
|
"Not an executable object: 'select 1'",
|
|
):
|
|
connection.execute("select 1")
|
|
|
|
def test_raw_positional_invalid(self, connection):
|
|
assert_raises_message(
|
|
tsa.exc.ArgumentError,
|
|
"List argument must consist only of tuples or dictionaries",
|
|
connection.exec_driver_sql,
|
|
"insert into users (user_id, user_name) values (?, ?)",
|
|
[2, "fred"],
|
|
)
|
|
|
|
assert_raises_message(
|
|
tsa.exc.ArgumentError,
|
|
"List argument must consist only of tuples or dictionaries",
|
|
connection.exec_driver_sql,
|
|
"insert into users (user_id, user_name) values (?, ?)",
|
|
[[3, "ed"], [4, "horse"]],
|
|
)
|
|
|
|
def test_raw_named_invalid(self, connection):
|
|
# this is awkward b.c. this is just testing if regular Python
|
|
# is raising TypeError if they happened to send arguments that
|
|
# look like the legacy ones which also happen to conflict with
|
|
# the positional signature for the method. some combinations
|
|
# can get through and fail differently
|
|
assert_raises(
|
|
TypeError,
|
|
connection.exec_driver_sql,
|
|
"insert into users (user_id, user_name) "
|
|
"values (%(id)s, %(name)s)",
|
|
{"id": 2, "name": "ed"},
|
|
{"id": 3, "name": "horse"},
|
|
{"id": 4, "name": "horse"},
|
|
)
|
|
assert_raises(
|
|
TypeError,
|
|
connection.exec_driver_sql,
|
|
"insert into users (user_id, user_name) "
|
|
"values (%(id)s, %(name)s)",
|
|
id=4,
|
|
name="sally",
|
|
)
|
|
|
|
def test_dialect_has_table_assertion(self):
|
|
with expect_raises_message(
|
|
tsa.exc.ArgumentError,
|
|
r"The argument passed to Dialect.has_table\(\) should be a",
|
|
):
|
|
testing.db.dialect.has_table(testing.db, "some_table")
|
|
|
|
def test_not_an_executable(self):
|
|
for obj in (
|
|
Table("foo", MetaData(), Column("x", Integer)),
|
|
Column("x", Integer),
|
|
tsa.and_(True),
|
|
tsa.and_(True).compile(),
|
|
column("foo"),
|
|
column("foo").compile(),
|
|
select(1).cte(),
|
|
# select(1).subquery(),
|
|
MetaData(),
|
|
Integer(),
|
|
tsa.Index(name="foo"),
|
|
tsa.UniqueConstraint("x"),
|
|
):
|
|
with testing.db.connect() as conn:
|
|
assert_raises_message(
|
|
tsa.exc.ObjectNotExecutableError,
|
|
"Not an executable object",
|
|
conn.execute,
|
|
obj,
|
|
)
|
|
|
|
def test_stmt_exception_bytestring_raised(self):
|
|
name = "méil"
|
|
users = self.tables.users
|
|
with testing.db.connect() as conn:
|
|
assert_raises_message(
|
|
tsa.exc.StatementError,
|
|
"A value is required for bind parameter 'uname'\n"
|
|
".*SELECT users.user_name AS .méil.",
|
|
conn.execute,
|
|
select(users.c.user_name.label(name)).where(
|
|
users.c.user_name == bindparam("uname")
|
|
),
|
|
{"uname_incorrect": "foo"},
|
|
)
|
|
|
|
def test_stmt_exception_bytestring_utf8(self):
|
|
# uncommon case for Py3K, bytestring object passed
|
|
# as the error message
|
|
message = "some message méil".encode()
|
|
|
|
err = tsa.exc.SQLAlchemyError(message)
|
|
eq_(str(err), "some message méil")
|
|
|
|
def test_stmt_exception_bytestring_latin1(self):
|
|
# uncommon case for Py3K, bytestring object passed
|
|
# as the error message
|
|
message = "some message méil".encode("latin-1")
|
|
|
|
err = tsa.exc.SQLAlchemyError(message)
|
|
eq_(str(err), "some message m\\xe9il")
|
|
|
|
def test_stmt_exception_unicode_hook_unicode(self):
|
|
# uncommon case for Py2K, Unicode object passed
|
|
# as the error message
|
|
message = "some message méil"
|
|
|
|
err = tsa.exc.SQLAlchemyError(message)
|
|
eq_(str(err), "some message méil")
|
|
|
|
def test_stmt_exception_object_arg(self):
|
|
err = tsa.exc.SQLAlchemyError(Foo())
|
|
eq_(str(err), "foo")
|
|
|
|
def test_stmt_exception_str_multi_args(self):
|
|
err = tsa.exc.SQLAlchemyError("some message", 206)
|
|
eq_(str(err), "('some message', 206)")
|
|
|
|
def test_stmt_exception_str_multi_args_bytestring(self):
|
|
message = "some message méil".encode()
|
|
|
|
err = tsa.exc.SQLAlchemyError(message, 206)
|
|
eq_(str(err), str((message, 206)))
|
|
|
|
def test_stmt_exception_str_multi_args_unicode(self):
|
|
message = "some message méil"
|
|
|
|
err = tsa.exc.SQLAlchemyError(message, 206)
|
|
eq_(str(err), str((message, 206)))
|
|
|
|
def test_generative_engine_event_dispatch_hasevents(self, testing_engine):
|
|
def l1(*arg, **kw):
|
|
pass
|
|
|
|
eng = testing_engine()
|
|
assert not eng._has_events
|
|
event.listen(eng, "before_execute", l1)
|
|
eng2 = eng.execution_options(foo="bar")
|
|
assert eng2._has_events
|
|
|
|
def test_scalar(self, connection):
|
|
conn = connection
|
|
users = self.tables.users
|
|
conn.execute(
|
|
users.insert(),
|
|
[
|
|
{"user_id": 1, "user_name": "sandy"},
|
|
{"user_id": 2, "user_name": "spongebob"},
|
|
],
|
|
)
|
|
res = conn.scalar(select(users.c.user_name).order_by(users.c.user_id))
|
|
eq_(res, "sandy")
|
|
|
|
def test_scalars(self, connection):
|
|
conn = connection
|
|
users = self.tables.users
|
|
conn.execute(
|
|
users.insert(),
|
|
[
|
|
{"user_id": 1, "user_name": "sandy"},
|
|
{"user_id": 2, "user_name": "spongebob"},
|
|
],
|
|
)
|
|
res = conn.scalars(select(users.c.user_name).order_by(users.c.user_id))
|
|
eq_(res.all(), ["sandy", "spongebob"])
|
|
|
|
@testing.combinations(
|
|
({"user_id": 1, "user_name": "name1"},),
|
|
([{"user_id": 1, "user_name": "name1"}],),
|
|
(({"user_id": 1, "user_name": "name1"},),),
|
|
(
|
|
[
|
|
{"user_id": 1, "user_name": "name1"},
|
|
{"user_id": 2, "user_name": "name2"},
|
|
],
|
|
),
|
|
argnames="parameters",
|
|
)
|
|
def test_params_interpretation(self, connection, parameters):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), parameters)
|
|
|
|
|
|
class ConvenienceExecuteTest(fixtures.TablesTest):
|
|
|
|
@classmethod
|
|
def define_tables(cls, metadata):
|
|
cls.table = Table(
|
|
"exec_test",
|
|
metadata,
|
|
Column("a", Integer),
|
|
Column("b", Integer),
|
|
test_needs_acid=True,
|
|
)
|
|
|
|
def _trans_fn(self, is_transaction=False):
|
|
def go(conn, x, value=None):
|
|
if is_transaction:
|
|
conn = conn.connection
|
|
conn.execute(self.table.insert().values(a=x, b=value))
|
|
|
|
return go
|
|
|
|
def _trans_rollback_fn(self, is_transaction=False):
|
|
def go(conn, x, value=None):
|
|
if is_transaction:
|
|
conn = conn.connection
|
|
conn.execute(self.table.insert().values(a=x, b=value))
|
|
raise SomeException("breakage")
|
|
|
|
return go
|
|
|
|
def _assert_no_data(self):
|
|
with testing.db.connect() as conn:
|
|
eq_(
|
|
conn.scalar(select(func.count("*")).select_from(self.table)),
|
|
0,
|
|
)
|
|
|
|
def _assert_fn(self, x, value=None):
|
|
with testing.db.connect() as conn:
|
|
eq_(conn.execute(self.table.select()).fetchall(), [(x, value)])
|
|
|
|
def test_transaction_engine_ctx_commit(self):
|
|
fn = self._trans_fn()
|
|
ctx = testing.db.begin()
|
|
testing.run_as_contextmanager(ctx, fn, 5, value=8)
|
|
self._assert_fn(5, value=8)
|
|
|
|
def test_transaction_engine_ctx_begin_fails_dont_enter_enter(self):
|
|
"""test #7272"""
|
|
engine = engines.testing_engine()
|
|
|
|
mock_connection = Mock(
|
|
return_value=Mock(begin=Mock(side_effect=Exception("boom")))
|
|
)
|
|
with mock.patch.object(engine, "_connection_cls", mock_connection):
|
|
# context manager isn't entered, doesn't actually call
|
|
# connect() or connection.begin()
|
|
engine.begin()
|
|
|
|
eq_(mock_connection.return_value.close.mock_calls, [])
|
|
|
|
def test_transaction_engine_ctx_begin_fails_include_enter(self):
|
|
"""test #7272
|
|
|
|
Note this behavior for 2.0 required that we add a new flag to
|
|
Connection _allow_autobegin=False, so that the first-connect
|
|
initialization sequence in create.py does not actually run begin()
|
|
events. previously, the initialize sequence used a future=False
|
|
connection unconditionally (and I didn't notice this).
|
|
|
|
"""
|
|
engine = engines.testing_engine()
|
|
|
|
close_mock = Mock()
|
|
with (
|
|
mock.patch.object(
|
|
engine._connection_cls,
|
|
"begin",
|
|
Mock(side_effect=Exception("boom")),
|
|
),
|
|
mock.patch.object(engine._connection_cls, "close", close_mock),
|
|
):
|
|
with expect_raises_message(Exception, "boom"):
|
|
with engine.begin():
|
|
pass
|
|
|
|
eq_(close_mock.mock_calls, [call()])
|
|
|
|
def test_transaction_engine_ctx_rollback(self):
|
|
fn = self._trans_rollback_fn()
|
|
ctx = testing.db.begin()
|
|
assert_raises_message(
|
|
Exception,
|
|
"breakage",
|
|
testing.run_as_contextmanager,
|
|
ctx,
|
|
fn,
|
|
5,
|
|
value=8,
|
|
)
|
|
self._assert_no_data()
|
|
|
|
def test_transaction_connection_ctx_commit(self):
|
|
fn = self._trans_fn(True)
|
|
with testing.db.connect() as conn:
|
|
ctx = conn.begin()
|
|
testing.run_as_contextmanager(ctx, fn, 5, value=8)
|
|
self._assert_fn(5, value=8)
|
|
|
|
def test_transaction_connection_ctx_rollback(self):
|
|
fn = self._trans_rollback_fn(True)
|
|
with testing.db.connect() as conn:
|
|
ctx = conn.begin()
|
|
assert_raises_message(
|
|
Exception,
|
|
"breakage",
|
|
testing.run_as_contextmanager,
|
|
ctx,
|
|
fn,
|
|
5,
|
|
value=8,
|
|
)
|
|
self._assert_no_data()
|
|
|
|
def test_connection_as_ctx(self):
|
|
fn = self._trans_fn()
|
|
with testing.db.begin() as conn:
|
|
fn(conn, 5, value=8)
|
|
self._assert_fn(5, value=8)
|
|
|
|
|
|
class ExecuteDriverTest(fixtures.TablesTest):
|
|
__backend__ = True
|
|
|
|
@classmethod
|
|
def define_tables(cls, metadata):
|
|
Table(
|
|
"users",
|
|
metadata,
|
|
Column("user_id", INT, primary_key=True, autoincrement=False),
|
|
Column("user_name", VARCHAR(20)),
|
|
)
|
|
Table(
|
|
"users_autoinc",
|
|
metadata,
|
|
Column(
|
|
"user_id", INT, primary_key=True, test_needs_autoincrement=True
|
|
),
|
|
Column("user_name", VARCHAR(20)),
|
|
)
|
|
|
|
def test_no_params_option(self):
|
|
stmt = (
|
|
"SELECT '%'"
|
|
+ testing.db.dialect.statement_compiler(
|
|
testing.db.dialect, None
|
|
).default_from()
|
|
)
|
|
|
|
with testing.db.connect() as conn:
|
|
result = (
|
|
conn.execution_options(no_parameters=True)
|
|
.exec_driver_sql(stmt)
|
|
.scalar()
|
|
)
|
|
eq_(result, "%")
|
|
|
|
@testing.requires.qmark_paramstyle
|
|
def test_raw_qmark(self, connection):
|
|
conn = connection
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) values (?, ?)",
|
|
(1, "jack"),
|
|
)
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) values (?, ?)",
|
|
(2, "fred"),
|
|
)
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) values (?, ?)",
|
|
[(3, "ed"), (4, "horse")],
|
|
)
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) values (?, ?)",
|
|
[(5, "barney"), (6, "donkey")],
|
|
)
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) values (?, ?)",
|
|
(7, "sally"),
|
|
)
|
|
res = conn.exec_driver_sql("select * from users order by user_id")
|
|
assert res.fetchall() == [
|
|
(1, "jack"),
|
|
(2, "fred"),
|
|
(3, "ed"),
|
|
(4, "horse"),
|
|
(5, "barney"),
|
|
(6, "donkey"),
|
|
(7, "sally"),
|
|
]
|
|
|
|
res = conn.exec_driver_sql(
|
|
"select * from users where user_name=?", ("jack",)
|
|
)
|
|
assert res.fetchall() == [(1, "jack")]
|
|
|
|
@testing.requires.format_paramstyle
|
|
def test_raw_sprintf(self, connection):
|
|
conn = connection
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) values (%s, %s)",
|
|
(1, "jack"),
|
|
)
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) values (%s, %s)",
|
|
[(2, "ed"), (3, "horse")],
|
|
)
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) values (%s, %s)",
|
|
(4, "sally"),
|
|
)
|
|
conn.exec_driver_sql("insert into users (user_id) values (%s)", (5,))
|
|
res = conn.exec_driver_sql("select * from users order by user_id")
|
|
assert res.fetchall() == [
|
|
(1, "jack"),
|
|
(2, "ed"),
|
|
(3, "horse"),
|
|
(4, "sally"),
|
|
(5, None),
|
|
]
|
|
|
|
res = conn.exec_driver_sql(
|
|
"select * from users where user_name=%s", ("jack",)
|
|
)
|
|
assert res.fetchall() == [(1, "jack")]
|
|
|
|
@testing.requires.pyformat_paramstyle
|
|
def test_raw_python(self, connection):
|
|
conn = connection
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) "
|
|
"values (%(id)s, %(name)s)",
|
|
{"id": 1, "name": "jack"},
|
|
)
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) "
|
|
"values (%(id)s, %(name)s)",
|
|
[{"id": 2, "name": "ed"}, {"id": 3, "name": "horse"}],
|
|
)
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) "
|
|
"values (%(id)s, %(name)s)",
|
|
dict(id=4, name="sally"),
|
|
)
|
|
res = conn.exec_driver_sql("select * from users order by user_id")
|
|
assert res.fetchall() == [
|
|
(1, "jack"),
|
|
(2, "ed"),
|
|
(3, "horse"),
|
|
(4, "sally"),
|
|
]
|
|
|
|
@testing.requires.named_paramstyle
|
|
def test_raw_named(self, connection):
|
|
conn = connection
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) values (:id, :name)",
|
|
{"id": 1, "name": "jack"},
|
|
)
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) values (:id, :name)",
|
|
[{"id": 2, "name": "ed"}, {"id": 3, "name": "horse"}],
|
|
)
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) values (:id, :name)",
|
|
{"id": 4, "name": "sally"},
|
|
)
|
|
res = conn.exec_driver_sql("select * from users order by user_id")
|
|
assert res.fetchall() == [
|
|
(1, "jack"),
|
|
(2, "ed"),
|
|
(3, "horse"),
|
|
(4, "sally"),
|
|
]
|
|
|
|
def test_raw_tuple_params(self, connection):
|
|
"""test #7820
|
|
|
|
There was an apparent improvement in the distill params
|
|
methodology used in exec_driver_sql which allows raw tuples to
|
|
pass through. In 1.4 there seems to be a _distill_cursor_params()
|
|
function that says it can handle this kind of parameter, but it isn't
|
|
used and when I tried to substitute it in for exec_driver_sql(),
|
|
things still fail.
|
|
|
|
In any case, add coverage here for the use case of passing
|
|
direct tuple params to exec_driver_sql including as the first
|
|
param, to note that it isn't mis-interpreted the way it is
|
|
in 1.x.
|
|
|
|
"""
|
|
|
|
with patch.object(connection.dialect, "do_execute") as do_exec:
|
|
connection.exec_driver_sql(
|
|
"UPDATE users SET user_name = 'query_one' WHERE "
|
|
"user_id = %s OR user_id IN %s",
|
|
(3, (1, 2)),
|
|
)
|
|
|
|
connection.exec_driver_sql(
|
|
"UPDATE users SET user_name = 'query_two' WHERE "
|
|
"user_id IN %s OR user_id = %s",
|
|
((1, 2), 3),
|
|
)
|
|
|
|
eq_(
|
|
do_exec.mock_calls,
|
|
[
|
|
call(
|
|
mock.ANY,
|
|
"UPDATE users SET user_name = 'query_one' "
|
|
"WHERE user_id = %s OR user_id IN %s",
|
|
connection.dialect.execute_sequence_format((3, (1, 2))),
|
|
mock.ANY,
|
|
),
|
|
call(
|
|
mock.ANY,
|
|
"UPDATE users SET user_name = 'query_two' "
|
|
"WHERE user_id IN %s OR user_id = %s",
|
|
connection.dialect.execute_sequence_format(((1, 2), 3)),
|
|
mock.ANY,
|
|
),
|
|
],
|
|
)
|
|
|
|
def test_non_dict_mapping(self, connection):
|
|
"""ensure arbitrary Mapping works for execute()"""
|
|
|
|
class NotADict(collections_abc.Mapping):
|
|
def __init__(self, _data):
|
|
self._data = _data
|
|
|
|
def __iter__(self):
|
|
return iter(self._data)
|
|
|
|
def __len__(self):
|
|
return len(self._data)
|
|
|
|
def __getitem__(self, key):
|
|
return self._data[key]
|
|
|
|
def keys(self):
|
|
return self._data.keys()
|
|
|
|
nd = NotADict({"a": 10, "b": 15})
|
|
eq_(dict(nd), {"a": 10, "b": 15})
|
|
|
|
result = connection.execute(
|
|
select(
|
|
bindparam("a", type_=Integer), bindparam("b", type_=Integer)
|
|
),
|
|
nd,
|
|
)
|
|
eq_(result.first(), (10, 15))
|
|
|
|
def test_row_works_as_mapping(self, connection):
|
|
"""ensure the RowMapping object works as a parameter dictionary for
|
|
execute."""
|
|
|
|
result = connection.execute(
|
|
select(literal(10).label("a"), literal(15).label("b"))
|
|
)
|
|
row = result.first()
|
|
eq_(row, (10, 15))
|
|
eq_(row._mapping, {"a": 10, "b": 15})
|
|
|
|
result = connection.execute(
|
|
select(
|
|
bindparam("a", type_=Integer).label("a"),
|
|
bindparam("b", type_=Integer).label("b"),
|
|
),
|
|
row._mapping,
|
|
)
|
|
row = result.first()
|
|
eq_(row, (10, 15))
|
|
eq_(row._mapping, {"a": 10, "b": 15})
|
|
|
|
def test_exception_wrapping_dbapi(self):
|
|
with testing.db.connect() as conn:
|
|
assert_raises_message(
|
|
tsa.exc.DBAPIError,
|
|
r"not_a_valid_statement",
|
|
conn.exec_driver_sql,
|
|
"not_a_valid_statement",
|
|
)
|
|
|
|
def test_exception_wrapping_orig_accessors(self):
|
|
de = None
|
|
|
|
with testing.db.connect() as conn:
|
|
try:
|
|
conn.exec_driver_sql("not_a_valid_statement")
|
|
except tsa.exc.DBAPIError as de_caught:
|
|
de = de_caught
|
|
|
|
assert isinstance(de.orig, conn.dialect.dbapi.Error)
|
|
|
|
# get the driver module name, the one which we know will provide
|
|
# for exceptions
|
|
top_level_dbapi_module = conn.dialect.dbapi
|
|
if isinstance(top_level_dbapi_module, AsyncAdapt_dbapi_module):
|
|
driver_module = top_level_dbapi_module.exceptions_module
|
|
else:
|
|
driver_module = top_level_dbapi_module
|
|
top_level_dbapi_module = driver_module.__name__.split(".")[0]
|
|
|
|
# check that it's not us
|
|
ne_(top_level_dbapi_module, "sqlalchemy")
|
|
|
|
# then make sure driver_exception is from that module
|
|
assert type(de.driver_exception).__module__.startswith(
|
|
top_level_dbapi_module
|
|
)
|
|
|
|
@testing.requires.sqlite
|
|
def test_exception_wrapping_non_dbapi_error(self):
|
|
e = create_engine("sqlite://")
|
|
e.dialect.is_disconnect = is_disconnect = Mock()
|
|
|
|
with e.connect() as c:
|
|
c.connection.cursor = Mock(
|
|
return_value=Mock(
|
|
execute=Mock(
|
|
side_effect=TypeError("I'm not a DBAPI error")
|
|
)
|
|
)
|
|
)
|
|
assert_raises_message(
|
|
TypeError,
|
|
"I'm not a DBAPI error",
|
|
c.exec_driver_sql,
|
|
"select ",
|
|
)
|
|
eq_(is_disconnect.call_count, 0)
|
|
|
|
def test_exception_wrapping_non_standard_dbapi_error(self):
|
|
class DBAPIError(Exception):
|
|
pass
|
|
|
|
class OperationalError(DBAPIError):
|
|
pass
|
|
|
|
class NonStandardException(OperationalError):
|
|
pass
|
|
|
|
# TODO: this test is assuming too much of arbitrary dialects and would
|
|
# be better suited tested against a single mock dialect that does not
|
|
# have any special behaviors
|
|
with (
|
|
patch.object(testing.db.dialect, "dbapi", Mock(Error=DBAPIError)),
|
|
patch.object(
|
|
testing.db.dialect, "loaded_dbapi", Mock(Error=DBAPIError)
|
|
),
|
|
patch.object(
|
|
testing.db.dialect, "is_disconnect", lambda *arg: False
|
|
),
|
|
patch.object(
|
|
testing.db.dialect,
|
|
"do_execute",
|
|
Mock(side_effect=NonStandardException),
|
|
),
|
|
patch.object(
|
|
testing.db.dialect.execution_ctx_cls,
|
|
"handle_dbapi_exception",
|
|
Mock(),
|
|
),
|
|
):
|
|
with testing.db.connect() as conn:
|
|
assert_raises(
|
|
tsa.exc.OperationalError, conn.exec_driver_sql, "select 1"
|
|
)
|
|
|
|
def test_exception_wrapping_non_dbapi_statement(self):
|
|
class MyType(TypeDecorator):
|
|
impl = Integer
|
|
cache_ok = True
|
|
|
|
def process_bind_param(self, value, dialect):
|
|
raise SomeException("nope")
|
|
|
|
def _go(conn):
|
|
assert_raises_message(
|
|
tsa.exc.StatementError,
|
|
r"\(.*.SomeException\) " r"nope\n\[SQL\: u?SELECT 1 ",
|
|
conn.execute,
|
|
select(1).where(column("foo") == literal("bar", MyType())),
|
|
)
|
|
|
|
with testing.db.connect() as conn:
|
|
_go(conn)
|
|
|
|
def test_stmt_exception_pickleable_no_dbapi(self):
|
|
self._test_stmt_exception_pickleable(Exception("hello world"))
|
|
|
|
@testing.crashes(
|
|
"postgresql+psycopg2",
|
|
"Older versions don't support cursor pickling, newer ones do",
|
|
)
|
|
@testing.fails_on(
|
|
"+mysqlconnector",
|
|
"Exception doesn't come back exactly the same from pickle",
|
|
)
|
|
@testing.fails_on(
|
|
"+mssqlpython",
|
|
"Exception pickling not working due to additional parameter",
|
|
)
|
|
@testing.fails_on(
|
|
"oracle+cx_oracle",
|
|
"cx_oracle exception seems to be having some issue with pickling",
|
|
)
|
|
@testing.fails_on(
|
|
"oracle+oracledb",
|
|
"oracledb exception seems to be having some issue with pickling",
|
|
)
|
|
def test_stmt_exception_pickleable_plus_dbapi(self):
|
|
raw = testing.db.raw_connection()
|
|
the_orig = None
|
|
try:
|
|
try:
|
|
cursor = raw.cursor()
|
|
cursor.execute("SELECTINCORRECT")
|
|
except testing.db.dialect.dbapi.Error as orig:
|
|
# py3k has "orig" in local scope...
|
|
the_orig = orig
|
|
finally:
|
|
raw.close()
|
|
self._test_stmt_exception_pickleable(the_orig)
|
|
|
|
def _test_stmt_exception_pickleable(self, orig):
|
|
for sa_exc in (
|
|
tsa.exc.StatementError(
|
|
"some error",
|
|
"select * from table",
|
|
{"foo": "bar"},
|
|
orig,
|
|
False,
|
|
),
|
|
tsa.exc.InterfaceError(
|
|
"select * from table", {"foo": "bar"}, orig, True
|
|
),
|
|
tsa.exc.NoReferencedTableError("message", "tname"),
|
|
tsa.exc.NoReferencedColumnError("message", "tname", "cname"),
|
|
tsa.exc.CircularDependencyError(
|
|
"some message", [1, 2, 3], [(1, 2), (3, 4)]
|
|
),
|
|
):
|
|
for loads, dumps in picklers():
|
|
repickled = loads(dumps(sa_exc))
|
|
eq_(repickled.args[0], sa_exc.args[0])
|
|
if isinstance(sa_exc, tsa.exc.StatementError):
|
|
eq_(repickled.params, {"foo": "bar"})
|
|
eq_(repickled.statement, sa_exc.statement)
|
|
if hasattr(sa_exc, "connection_invalidated"):
|
|
eq_(
|
|
repickled.connection_invalidated,
|
|
sa_exc.connection_invalidated,
|
|
)
|
|
eq_(repickled.orig.args[0], orig.args[0])
|
|
|
|
def test_dont_wrap_mixin(self):
|
|
class MyException(Exception, tsa.exc.DontWrapMixin):
|
|
pass
|
|
|
|
class MyType(TypeDecorator):
|
|
impl = Integer
|
|
cache_ok = True
|
|
|
|
def process_bind_param(self, value, dialect):
|
|
raise MyException("nope")
|
|
|
|
def _go(conn):
|
|
assert_raises_message(
|
|
MyException,
|
|
"nope",
|
|
conn.execute,
|
|
select(1).where(column("foo") == literal("bar", MyType())),
|
|
)
|
|
|
|
conn = testing.db.connect()
|
|
try:
|
|
_go(conn)
|
|
finally:
|
|
conn.close()
|
|
|
|
def test_empty_insert(self, connection):
|
|
"""test that execute() interprets [] as a list with no params and
|
|
warns since it has nothing to do with such an executemany.
|
|
"""
|
|
users_autoinc = self.tables.users_autoinc
|
|
|
|
with expect_deprecated(
|
|
r"Empty parameter sequence passed to execute\(\). "
|
|
"This use is deprecated and will raise an exception in a "
|
|
"future SQLAlchemy release"
|
|
):
|
|
connection.execute(
|
|
users_autoinc.insert().values(
|
|
user_name=bindparam("name", None)
|
|
),
|
|
[],
|
|
)
|
|
|
|
eq_(len(connection.execute(users_autoinc.select()).all()), 1)
|
|
|
|
@testing.only_on("sqlite")
|
|
def test_raw_insert_with_empty_list(self, connection):
|
|
"""exec_driver_sql instead does not raise if an empty list is passed.
|
|
Let the driver do that if it wants to.
|
|
"""
|
|
conn = connection
|
|
with expect_raises_message(
|
|
tsa.exc.ProgrammingError, "Incorrect number of bindings supplied"
|
|
):
|
|
conn.exec_driver_sql(
|
|
"insert into users (user_id, user_name) values (?, ?)", []
|
|
)
|
|
|
|
def test_works_after_dispose_testing_engine(self):
|
|
eng = engines.testing_engine()
|
|
for i in range(3):
|
|
with eng.connect() as conn:
|
|
eq_(conn.scalar(select(1)), 1)
|
|
eng.dispose()
|
|
|
|
@testing.requires.insertmanyvalues
|
|
def test_cursor_execute_insertmanyvalues(self, connection, metadata):
|
|
"""test #13018, that before_cursor_execute and after_cursor_execute
|
|
get the inner INSERT statements / params for an insertmanyvalues
|
|
|
|
"""
|
|
canary = Mock()
|
|
|
|
t = Table(
|
|
"t",
|
|
metadata,
|
|
Column("id", Integer, primary_key=True),
|
|
Column("data", String(50)),
|
|
)
|
|
t.create(connection)
|
|
|
|
event.listen(connection, "before_cursor_execute", canary.bce)
|
|
event.listen(connection, "after_cursor_execute", canary.ace)
|
|
|
|
result = connection.execute(
|
|
t.insert().returning(
|
|
t.c.id, t.c.data, sort_by_parameter_order=True
|
|
),
|
|
[{"data": f"d{i}"} for i in range(10)],
|
|
)
|
|
eq_(result.all(), [(i + 1, f"d{i}") for i in range(10)])
|
|
|
|
eq_(
|
|
[(c1.args[2], c1.args[3]) for c1 in canary.bce.mock_calls],
|
|
[(c1.args[2], c1.args[3]) for c1 in canary.ace.mock_calls],
|
|
)
|
|
|
|
|
|
class CompiledCacheTest(fixtures.TestBase):
|
|
__sparse_driver_backend__ = True
|
|
|
|
def test_cache(self, connection, metadata):
|
|
users = Table(
|
|
"users",
|
|
metadata,
|
|
Column(
|
|
"user_id", INT, primary_key=True, test_needs_autoincrement=True
|
|
),
|
|
Column("user_name", VARCHAR(20)),
|
|
Column("extra_data", VARCHAR(20)),
|
|
)
|
|
users.create(connection)
|
|
|
|
conn = connection
|
|
cache = {}
|
|
cached_conn = conn.execution_options(compiled_cache=cache)
|
|
|
|
ins = users.insert()
|
|
with patch.object(
|
|
ins, "_compiler", Mock(side_effect=ins._compiler)
|
|
) as compile_mock:
|
|
cached_conn.execute(ins, {"user_name": "u1"})
|
|
cached_conn.execute(ins, {"user_name": "u2"})
|
|
cached_conn.execute(ins, {"user_name": "u3"})
|
|
eq_(compile_mock.call_count, 1)
|
|
assert len(cache) == 1
|
|
eq_(conn.exec_driver_sql("select count(*) from users").scalar(), 3)
|
|
|
|
@testing.only_on(
|
|
["sqlite", "mysql", "postgresql"],
|
|
"uses blob value that is problematic for some DBAPIs",
|
|
)
|
|
def test_cache_noleak_on_statement_values(self, metadata, connection):
|
|
# This is a non regression test for an object reference leak caused
|
|
# by the compiled_cache.
|
|
|
|
photo = Table(
|
|
"photo",
|
|
metadata,
|
|
Column(
|
|
"id", Integer, primary_key=True, test_needs_autoincrement=True
|
|
),
|
|
Column("photo_blob", LargeBinary()),
|
|
)
|
|
metadata.create_all(connection)
|
|
|
|
cache = {}
|
|
cached_conn = connection.execution_options(compiled_cache=cache)
|
|
|
|
class PhotoBlob(bytearray):
|
|
pass
|
|
|
|
blob = PhotoBlob(100)
|
|
ref_blob = weakref.ref(blob)
|
|
|
|
ins = photo.insert()
|
|
with patch.object(
|
|
ins, "_compiler", Mock(side_effect=ins._compiler)
|
|
) as compile_mock:
|
|
cached_conn.execute(ins, {"photo_blob": blob})
|
|
eq_(compile_mock.call_count, 1)
|
|
eq_(len(cache), 1)
|
|
eq_(
|
|
connection.exec_driver_sql("select count(*) from photo").scalar(),
|
|
1,
|
|
)
|
|
|
|
del blob
|
|
|
|
gc_collect()
|
|
|
|
# The compiled statement cache should not hold any reference to the
|
|
# the statement values (only the keys).
|
|
eq_(ref_blob(), None)
|
|
|
|
def test_keys_independent_of_ordering(self, connection, metadata):
|
|
users = Table(
|
|
"users",
|
|
metadata,
|
|
Column(
|
|
"user_id", INT, primary_key=True, test_needs_autoincrement=True
|
|
),
|
|
Column("user_name", VARCHAR(20)),
|
|
Column("extra_data", VARCHAR(20)),
|
|
)
|
|
users.create(connection)
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
{"user_id": 1, "user_name": "u1", "extra_data": "e1"},
|
|
)
|
|
cache = {}
|
|
cached_conn = connection.execution_options(compiled_cache=cache)
|
|
|
|
upd = users.update().where(users.c.user_id == bindparam("b_user_id"))
|
|
|
|
with patch.object(
|
|
upd, "_compiler", Mock(side_effect=upd._compiler)
|
|
) as compile_mock:
|
|
cached_conn.execute(
|
|
upd,
|
|
util.OrderedDict(
|
|
[
|
|
("b_user_id", 1),
|
|
("user_name", "u2"),
|
|
("extra_data", "e2"),
|
|
]
|
|
),
|
|
)
|
|
cached_conn.execute(
|
|
upd,
|
|
util.OrderedDict(
|
|
[
|
|
("b_user_id", 1),
|
|
("extra_data", "e3"),
|
|
("user_name", "u3"),
|
|
]
|
|
),
|
|
)
|
|
cached_conn.execute(
|
|
upd,
|
|
util.OrderedDict(
|
|
[
|
|
("extra_data", "e4"),
|
|
("user_name", "u4"),
|
|
("b_user_id", 1),
|
|
]
|
|
),
|
|
)
|
|
eq_(compile_mock.call_count, 1)
|
|
eq_(len(cache), 1)
|
|
|
|
@testing.requires.schemas
|
|
def test_schema_translate_in_key(self, metadata, connection):
|
|
Table("x", metadata, Column("q", Integer))
|
|
Table("x", metadata, Column("q", Integer), schema=config.test_schema)
|
|
metadata.create_all(connection)
|
|
|
|
m = MetaData()
|
|
t1 = Table("x", m, Column("q", Integer))
|
|
ins = t1.insert()
|
|
stmt = select(t1.c.q)
|
|
|
|
cache = {}
|
|
|
|
conn = connection.execution_options(compiled_cache=cache)
|
|
conn.execute(ins, {"q": 1})
|
|
eq_(conn.scalar(stmt), 1)
|
|
|
|
conn = connection.execution_options(
|
|
compiled_cache=cache,
|
|
schema_translate_map={None: config.test_schema},
|
|
)
|
|
conn.execute(ins, {"q": 2})
|
|
eq_(conn.scalar(stmt), 2)
|
|
|
|
conn = connection.execution_options(
|
|
compiled_cache=cache,
|
|
schema_translate_map={None: None},
|
|
)
|
|
# should use default schema again even though statement
|
|
# was compiled with test_schema in the map
|
|
eq_(conn.scalar(stmt), 1)
|
|
|
|
conn = connection.execution_options(
|
|
compiled_cache=cache,
|
|
)
|
|
eq_(conn.scalar(stmt), 1)
|
|
|
|
|
|
class MockStrategyTest(fixtures.TestBase):
|
|
def _engine_fixture(self):
|
|
buf = StringIO()
|
|
|
|
def dump(sql, *multiparams, **params):
|
|
buf.write(str(sql.compile(dialect=engine.dialect)))
|
|
|
|
engine = create_mock_engine("postgresql+psycopg2://", executor=dump)
|
|
return engine, buf
|
|
|
|
def test_sequence_not_duped(self):
|
|
engine, buf = self._engine_fixture()
|
|
metadata = MetaData()
|
|
t = Table(
|
|
"testtable",
|
|
metadata,
|
|
Column(
|
|
"pk",
|
|
Integer,
|
|
normalize_sequence(config, Sequence("testtable_pk_seq")),
|
|
primary_key=True,
|
|
),
|
|
)
|
|
|
|
t.create(engine)
|
|
t.drop(engine)
|
|
|
|
eq_(re.findall(r"CREATE (\w+)", buf.getvalue()), ["SEQUENCE", "TABLE"])
|
|
|
|
eq_(re.findall(r"DROP (\w+)", buf.getvalue()), ["TABLE", "SEQUENCE"])
|
|
|
|
|
|
class SchemaTranslateTest(fixtures.TestBase, testing.AssertsExecutionResults):
|
|
__requires__ = ("schemas",)
|
|
__sparse_driver_backend__ = True
|
|
|
|
@testing.fixture
|
|
def plain_tables(self, metadata):
|
|
t1 = Table(
|
|
"t1", metadata, Column("x", Integer), schema=config.test_schema
|
|
)
|
|
t2 = Table(
|
|
"t2", metadata, Column("x", Integer), schema=config.test_schema
|
|
)
|
|
t3 = Table("t3", metadata, Column("x", Integer), schema=None)
|
|
|
|
return t1, t2, t3
|
|
|
|
@testing.fixture
|
|
def same_named_tables(self, metadata, connection):
|
|
ts1 = Table(
|
|
"t1", metadata, Column("x", String(10)), schema=config.test_schema
|
|
)
|
|
tsnone = Table("t1", metadata, Column("x", String(10)), schema=None)
|
|
|
|
metadata.create_all(connection)
|
|
|
|
connection.execute(ts1.insert().values(x="ts1"))
|
|
connection.execute(tsnone.insert().values(x="tsnone"))
|
|
return ts1, tsnone
|
|
|
|
def test_create_table(self, plain_tables, connection):
|
|
map_ = {
|
|
None: config.test_schema,
|
|
"foo": config.test_schema,
|
|
"bar": None,
|
|
}
|
|
|
|
metadata = MetaData()
|
|
t1 = Table("t1", metadata, Column("x", Integer))
|
|
t2 = Table("t2", metadata, Column("x", Integer), schema="foo")
|
|
t3 = Table("t3", metadata, Column("x", Integer), schema="bar")
|
|
|
|
with self.sql_execution_asserter(connection) as asserter:
|
|
conn = connection.execution_options(schema_translate_map=map_)
|
|
|
|
t1.create(conn)
|
|
t2.create(conn)
|
|
t3.create(conn)
|
|
|
|
t3.drop(conn)
|
|
t2.drop(conn)
|
|
t1.drop(conn)
|
|
|
|
asserter.assert_(
|
|
CompiledSQL("CREATE TABLE __[SCHEMA__none].t1 (x INTEGER)"),
|
|
CompiledSQL("CREATE TABLE __[SCHEMA_foo].t2 (x INTEGER)"),
|
|
CompiledSQL("CREATE TABLE __[SCHEMA_bar].t3 (x INTEGER)"),
|
|
CompiledSQL("DROP TABLE __[SCHEMA_bar].t3"),
|
|
CompiledSQL("DROP TABLE __[SCHEMA_foo].t2"),
|
|
CompiledSQL("DROP TABLE __[SCHEMA__none].t1"),
|
|
)
|
|
|
|
def test_ddl_hastable(self, plain_tables, connection):
|
|
map_ = {
|
|
None: config.test_schema,
|
|
"foo": config.test_schema,
|
|
"bar": None,
|
|
}
|
|
|
|
metadata = MetaData()
|
|
Table("t1", metadata, Column("x", Integer))
|
|
Table("t2", metadata, Column("x", Integer), schema="foo")
|
|
Table("t3", metadata, Column("x", Integer), schema="bar")
|
|
|
|
conn = connection.execution_options(schema_translate_map=map_)
|
|
metadata.create_all(conn)
|
|
|
|
insp = inspect(connection)
|
|
is_true(insp.has_table("t1", schema=config.test_schema))
|
|
is_true(insp.has_table("t2", schema=config.test_schema))
|
|
is_true(insp.has_table("t3", schema=None))
|
|
|
|
conn = connection.execution_options(schema_translate_map=map_)
|
|
|
|
# if this test fails, the tables won't get dropped. so need a
|
|
# more robust fixture for this
|
|
metadata.drop_all(conn)
|
|
|
|
insp = inspect(connection)
|
|
is_false(insp.has_table("t1", schema=config.test_schema))
|
|
is_false(insp.has_table("t2", schema=config.test_schema))
|
|
is_false(insp.has_table("t3", schema=None))
|
|
|
|
def test_option_on_execute(self, plain_tables, connection):
|
|
# provided by metadata fixture provided by plain_tables fixture
|
|
self.metadata.create_all(connection)
|
|
|
|
map_ = {
|
|
None: config.test_schema,
|
|
"foo": config.test_schema,
|
|
"bar": None,
|
|
}
|
|
|
|
metadata = MetaData()
|
|
t1 = Table("t1", metadata, Column("x", Integer))
|
|
t2 = Table("t2", metadata, Column("x", Integer), schema="foo")
|
|
t3 = Table("t3", metadata, Column("x", Integer), schema="bar")
|
|
|
|
with self.sql_execution_asserter(connection) as asserter:
|
|
conn = connection
|
|
execution_options = {"schema_translate_map": map_}
|
|
conn.execute(
|
|
t1.insert(), {"x": 1}, execution_options=execution_options
|
|
)
|
|
conn.execute(
|
|
t2.insert(), {"x": 1}, execution_options=execution_options
|
|
)
|
|
conn.execute(
|
|
t3.insert(), {"x": 1}, execution_options=execution_options
|
|
)
|
|
|
|
conn.execute(
|
|
t1.update().values(x=1).where(t1.c.x == 1),
|
|
execution_options=execution_options,
|
|
)
|
|
conn.execute(
|
|
t2.update().values(x=2).where(t2.c.x == 1),
|
|
execution_options=execution_options,
|
|
)
|
|
conn.execute(
|
|
t3.update().values(x=3).where(t3.c.x == 1),
|
|
execution_options=execution_options,
|
|
)
|
|
|
|
eq_(
|
|
conn.execute(
|
|
select(t1.c.x), execution_options=execution_options
|
|
).scalar(),
|
|
1,
|
|
)
|
|
eq_(
|
|
conn.execute(
|
|
select(t2.c.x), execution_options=execution_options
|
|
).scalar(),
|
|
2,
|
|
)
|
|
eq_(
|
|
conn.execute(
|
|
select(t3.c.x), execution_options=execution_options
|
|
).scalar(),
|
|
3,
|
|
)
|
|
|
|
conn.execute(t1.delete(), execution_options=execution_options)
|
|
conn.execute(t2.delete(), execution_options=execution_options)
|
|
conn.execute(t3.delete(), execution_options=execution_options)
|
|
|
|
asserter.assert_(
|
|
CompiledSQL("INSERT INTO __[SCHEMA__none].t1 (x) VALUES (:x)"),
|
|
CompiledSQL("INSERT INTO __[SCHEMA_foo].t2 (x) VALUES (:x)"),
|
|
CompiledSQL("INSERT INTO __[SCHEMA_bar].t3 (x) VALUES (:x)"),
|
|
CompiledSQL(
|
|
"UPDATE __[SCHEMA__none].t1 SET x=:x WHERE "
|
|
"__[SCHEMA__none].t1.x = :x_1"
|
|
),
|
|
CompiledSQL(
|
|
"UPDATE __[SCHEMA_foo].t2 SET x=:x WHERE "
|
|
"__[SCHEMA_foo].t2.x = :x_1"
|
|
),
|
|
CompiledSQL(
|
|
"UPDATE __[SCHEMA_bar].t3 SET x=:x WHERE "
|
|
"__[SCHEMA_bar].t3.x = :x_1"
|
|
),
|
|
CompiledSQL(
|
|
"SELECT __[SCHEMA__none].t1.x FROM __[SCHEMA__none].t1"
|
|
),
|
|
CompiledSQL("SELECT __[SCHEMA_foo].t2.x FROM __[SCHEMA_foo].t2"),
|
|
CompiledSQL("SELECT __[SCHEMA_bar].t3.x FROM __[SCHEMA_bar].t3"),
|
|
CompiledSQL("DELETE FROM __[SCHEMA__none].t1"),
|
|
CompiledSQL("DELETE FROM __[SCHEMA_foo].t2"),
|
|
CompiledSQL("DELETE FROM __[SCHEMA_bar].t3"),
|
|
)
|
|
|
|
def test_schema_translate_map_keys_change_name_added(
|
|
self, same_named_tables, connection
|
|
):
|
|
"""test #10024"""
|
|
|
|
metadata = MetaData()
|
|
|
|
translate_table = Table(
|
|
"t1", metadata, Column("x", String(10)), schema=config.test_schema
|
|
)
|
|
|
|
eq_(
|
|
connection.scalar(
|
|
select(translate_table),
|
|
execution_options={"schema_translate_map": {"foo": "bar"}},
|
|
),
|
|
"ts1",
|
|
)
|
|
|
|
eq_(
|
|
connection.scalar(
|
|
select(translate_table),
|
|
execution_options={
|
|
"schema_translate_map": {
|
|
"foo": "bar",
|
|
config.test_schema: None,
|
|
}
|
|
},
|
|
),
|
|
"tsnone",
|
|
)
|
|
|
|
def test_schema_translate_map_keys_change_name_removed(
|
|
self, same_named_tables, connection
|
|
):
|
|
"""test #10024"""
|
|
|
|
metadata = MetaData()
|
|
|
|
translate_table = Table(
|
|
"t1", metadata, Column("x", String(10)), schema=config.test_schema
|
|
)
|
|
|
|
eq_(
|
|
connection.scalar(
|
|
select(translate_table),
|
|
execution_options={
|
|
"schema_translate_map": {
|
|
"foo": "bar",
|
|
config.test_schema: None,
|
|
}
|
|
},
|
|
),
|
|
"tsnone",
|
|
)
|
|
|
|
eq_(
|
|
connection.scalar(
|
|
select(translate_table),
|
|
execution_options={"schema_translate_map": {"foo": "bar"}},
|
|
),
|
|
"ts1",
|
|
)
|
|
|
|
def test_schema_translate_map_keys_change_none_removed(
|
|
self, same_named_tables, connection
|
|
):
|
|
"""test #10024"""
|
|
|
|
connection.engine.clear_compiled_cache()
|
|
|
|
metadata = MetaData()
|
|
|
|
translate_table = Table("t1", metadata, Column("x", String(10)))
|
|
|
|
eq_(
|
|
connection.scalar(
|
|
select(translate_table),
|
|
execution_options={
|
|
"schema_translate_map": {None: config.test_schema}
|
|
},
|
|
),
|
|
"ts1",
|
|
)
|
|
|
|
with expect_raises_message(
|
|
tsa.exc.StatementError,
|
|
"schema translate map which previously had `None` "
|
|
"present as a key now no longer has it present",
|
|
):
|
|
connection.scalar(
|
|
select(translate_table),
|
|
execution_options={"schema_translate_map": {"foo": "bar"}},
|
|
),
|
|
|
|
def test_schema_translate_map_keys_change_none_added(
|
|
self, same_named_tables, connection
|
|
):
|
|
"""test #10024"""
|
|
|
|
connection.engine.clear_compiled_cache()
|
|
|
|
metadata = MetaData()
|
|
|
|
translate_table = Table("t1", metadata, Column("x", String(10)))
|
|
|
|
eq_(
|
|
connection.scalar(
|
|
select(translate_table),
|
|
execution_options={"schema_translate_map": {"foo": "bar"}},
|
|
),
|
|
"tsnone",
|
|
)
|
|
|
|
with expect_raises_message(
|
|
tsa.exc.StatementError,
|
|
"schema translate map which previously did not have `None` "
|
|
"present as a key now has `None` present; compiled statement may "
|
|
"lack adequate placeholders.",
|
|
):
|
|
connection.scalar(
|
|
select(translate_table),
|
|
execution_options={
|
|
"schema_translate_map": {None: config.test_schema}
|
|
},
|
|
),
|
|
|
|
def test_crud(self, plain_tables, connection):
|
|
# provided by metadata fixture provided by plain_tables fixture
|
|
self.metadata.create_all(connection)
|
|
|
|
map_ = {
|
|
None: config.test_schema,
|
|
"foo": config.test_schema,
|
|
"bar": None,
|
|
}
|
|
|
|
metadata = MetaData()
|
|
t1 = Table("t1", metadata, Column("x", Integer))
|
|
t2 = Table("t2", metadata, Column("x", Integer), schema="foo")
|
|
t3 = Table("t3", metadata, Column("x", Integer), schema="bar")
|
|
|
|
with self.sql_execution_asserter(connection) as asserter:
|
|
conn = connection.execution_options(schema_translate_map=map_)
|
|
|
|
conn.execute(t1.insert(), {"x": 1})
|
|
conn.execute(t2.insert(), {"x": 1})
|
|
conn.execute(t3.insert(), {"x": 1})
|
|
|
|
conn.execute(t1.update().values(x=1).where(t1.c.x == 1))
|
|
conn.execute(t2.update().values(x=2).where(t2.c.x == 1))
|
|
conn.execute(t3.update().values(x=3).where(t3.c.x == 1))
|
|
|
|
eq_(conn.scalar(select(t1.c.x)), 1)
|
|
eq_(conn.scalar(select(t2.c.x)), 2)
|
|
eq_(conn.scalar(select(t3.c.x)), 3)
|
|
|
|
conn.execute(t1.delete())
|
|
conn.execute(t2.delete())
|
|
conn.execute(t3.delete())
|
|
|
|
asserter.assert_(
|
|
CompiledSQL("INSERT INTO __[SCHEMA__none].t1 (x) VALUES (:x)"),
|
|
CompiledSQL("INSERT INTO __[SCHEMA_foo].t2 (x) VALUES (:x)"),
|
|
CompiledSQL("INSERT INTO __[SCHEMA_bar].t3 (x) VALUES (:x)"),
|
|
CompiledSQL(
|
|
"UPDATE __[SCHEMA__none].t1 SET x=:x WHERE "
|
|
"__[SCHEMA__none].t1.x = :x_1"
|
|
),
|
|
CompiledSQL(
|
|
"UPDATE __[SCHEMA_foo].t2 SET x=:x WHERE "
|
|
"__[SCHEMA_foo].t2.x = :x_1"
|
|
),
|
|
CompiledSQL(
|
|
"UPDATE __[SCHEMA_bar].t3 SET x=:x WHERE "
|
|
"__[SCHEMA_bar].t3.x = :x_1"
|
|
),
|
|
CompiledSQL(
|
|
"SELECT __[SCHEMA__none].t1.x FROM __[SCHEMA__none].t1"
|
|
),
|
|
CompiledSQL("SELECT __[SCHEMA_foo].t2.x FROM __[SCHEMA_foo].t2"),
|
|
CompiledSQL("SELECT __[SCHEMA_bar].t3.x FROM __[SCHEMA_bar].t3"),
|
|
CompiledSQL("DELETE FROM __[SCHEMA__none].t1"),
|
|
CompiledSQL("DELETE FROM __[SCHEMA_foo].t2"),
|
|
CompiledSQL("DELETE FROM __[SCHEMA_bar].t3"),
|
|
)
|
|
|
|
def test_via_engine(self, plain_tables, metadata):
|
|
with config.db.begin() as connection:
|
|
metadata.create_all(connection)
|
|
|
|
map_ = {
|
|
None: config.test_schema,
|
|
"foo": config.test_schema,
|
|
"bar": None,
|
|
}
|
|
|
|
metadata = MetaData()
|
|
t2 = Table("t2", metadata, Column("x", Integer), schema="foo")
|
|
|
|
with self.sql_execution_asserter(config.db) as asserter:
|
|
eng = config.db.execution_options(schema_translate_map=map_)
|
|
with eng.connect() as conn:
|
|
conn.execute(select(t2.c.x))
|
|
asserter.assert_(
|
|
CompiledSQL("SELECT __[SCHEMA_foo].t2.x FROM __[SCHEMA_foo].t2")
|
|
)
|
|
|
|
|
|
class ExecutionOptionsTest(fixtures.TestBase):
|
|
def test_engine_level_options(self):
|
|
eng = engines.testing_engine(
|
|
options={"execution_options": {"foo": "bar"}}
|
|
)
|
|
with eng.connect() as conn:
|
|
eq_(conn._execution_options["foo"], "bar")
|
|
eq_(
|
|
conn.execution_options(bat="hoho")._execution_options["foo"],
|
|
"bar",
|
|
)
|
|
eq_(
|
|
conn.execution_options(bat="hoho")._execution_options["bat"],
|
|
"hoho",
|
|
)
|
|
eq_(
|
|
conn.execution_options(foo="hoho")._execution_options["foo"],
|
|
"hoho",
|
|
)
|
|
eng.update_execution_options(foo="hoho")
|
|
conn2 = eng.connect()
|
|
eq_(conn2._execution_options["foo"], "hoho")
|
|
conn2.close()
|
|
|
|
def test_generative_engine_execution_options(self):
|
|
eng = engines.testing_engine(
|
|
options={"execution_options": {"base": "x1"}}
|
|
)
|
|
|
|
is_(eng.engine, eng)
|
|
|
|
eng1 = eng.execution_options(foo="b1")
|
|
is_(eng1.engine, eng1)
|
|
eng2 = eng.execution_options(foo="b2")
|
|
eng1a = eng1.execution_options(bar="a1")
|
|
eng2a = eng2.execution_options(foo="b3", bar="a2")
|
|
is_(eng2a.engine, eng2a)
|
|
|
|
eq_(eng._execution_options, {"base": "x1"})
|
|
eq_(eng1._execution_options, {"base": "x1", "foo": "b1"})
|
|
eq_(eng2._execution_options, {"base": "x1", "foo": "b2"})
|
|
eq_(eng1a._execution_options, {"base": "x1", "foo": "b1", "bar": "a1"})
|
|
eq_(eng2a._execution_options, {"base": "x1", "foo": "b3", "bar": "a2"})
|
|
is_(eng1a.pool, eng.pool)
|
|
|
|
# test pool is shared
|
|
eng2.dispose()
|
|
is_(eng1a.pool, eng2.pool)
|
|
is_(eng.pool, eng2.pool)
|
|
|
|
def test_autocommit_option_preserved_first_connect(self, testing_engine):
|
|
eng = testing_engine()
|
|
eng.update_execution_options(autocommit=True)
|
|
conn = eng.connect()
|
|
eq_(conn._execution_options, {"autocommit": True})
|
|
conn.close()
|
|
|
|
def test_initialize_rollback(self, testing_engine):
|
|
"""test a rollback happens during first connect"""
|
|
eng = testing_engine()
|
|
with patch.object(eng.dialect, "do_rollback") as do_rollback:
|
|
assert do_rollback.call_count == 0
|
|
connection = eng.connect()
|
|
assert do_rollback.call_count == 1
|
|
connection.close()
|
|
|
|
def test_dialect_init_uses_options(self, testing_engine):
|
|
eng = testing_engine()
|
|
|
|
def my_init(connection):
|
|
connection.execution_options(foo="bar").execute(select(1))
|
|
|
|
with patch.object(eng.dialect, "initialize", my_init):
|
|
conn = eng.connect()
|
|
eq_(conn._execution_options, {})
|
|
conn.close()
|
|
|
|
@testing.combinations(
|
|
({}, {}, {}),
|
|
({"a": "b"}, {}, {"a": "b"}),
|
|
({"a": "b", "d": "e"}, {"a": "c"}, {"a": "c", "d": "e"}),
|
|
argnames="conn_opts, exec_opts, expected",
|
|
)
|
|
def test_execution_opts_per_invoke(
|
|
self, connection, conn_opts, exec_opts, expected
|
|
):
|
|
opts = []
|
|
|
|
@event.listens_for(connection, "before_cursor_execute")
|
|
def before_cursor_execute(
|
|
conn, cursor, statement, parameters, context, executemany
|
|
):
|
|
opts.append(context.execution_options)
|
|
|
|
if conn_opts:
|
|
connection = connection.execution_options(**conn_opts)
|
|
|
|
if exec_opts:
|
|
connection.execute(select(1), execution_options=exec_opts)
|
|
else:
|
|
connection.execute(select(1))
|
|
|
|
eq_(opts, [expected])
|
|
|
|
@testing.combinations(
|
|
({}, {}, {}, {}),
|
|
({}, {"a": "b"}, {}, {"a": "b"}),
|
|
({}, {"a": "b", "d": "e"}, {"a": "c"}, {"a": "c", "d": "e"}),
|
|
(
|
|
{"q": "z", "p": "r"},
|
|
{"a": "b", "p": "x", "d": "e"},
|
|
{"a": "c"},
|
|
{"q": "z", "p": "x", "a": "c", "d": "e"},
|
|
),
|
|
argnames="stmt_opts, conn_opts, exec_opts, expected",
|
|
)
|
|
def test_execution_opts_per_invoke_execute_events(
|
|
self, connection, stmt_opts, conn_opts, exec_opts, expected
|
|
):
|
|
opts = []
|
|
|
|
@event.listens_for(connection, "before_execute")
|
|
def before_execute(
|
|
conn, clauseelement, multiparams, params, execution_options
|
|
):
|
|
opts.append(("before", execution_options))
|
|
|
|
@event.listens_for(connection, "after_execute")
|
|
def after_execute(
|
|
conn,
|
|
clauseelement,
|
|
multiparams,
|
|
params,
|
|
execution_options,
|
|
result,
|
|
):
|
|
opts.append(("after", execution_options))
|
|
|
|
stmt = select(1)
|
|
|
|
if stmt_opts:
|
|
stmt = stmt.execution_options(**stmt_opts)
|
|
|
|
if conn_opts:
|
|
connection = connection.execution_options(**conn_opts)
|
|
|
|
if exec_opts:
|
|
connection.execute(stmt, execution_options=exec_opts)
|
|
else:
|
|
connection.execute(stmt)
|
|
|
|
eq_(opts, [("before", expected), ("after", expected)])
|
|
|
|
def test_dialect_conn_options(self, testing_engine):
|
|
engine = testing_engine("sqlite://", options=dict(_initialize=False))
|
|
engine.dialect = Mock()
|
|
with engine.connect() as conn:
|
|
c2 = conn.execution_options(foo="bar")
|
|
eq_(
|
|
engine.dialect.set_connection_execution_options.mock_calls,
|
|
[call(c2, {"foo": "bar"})],
|
|
)
|
|
|
|
def test_dialect_engine_options(self, testing_engine):
|
|
engine = testing_engine("sqlite://")
|
|
engine.dialect = Mock()
|
|
e2 = engine.execution_options(foo="bar")
|
|
eq_(
|
|
engine.dialect.set_engine_execution_options.mock_calls,
|
|
[call(e2, {"foo": "bar"})],
|
|
)
|
|
|
|
def test_dialect_engine_construction_options(self):
|
|
dialect = Mock()
|
|
engine = Engine(
|
|
Mock(), dialect, Mock(), execution_options={"foo": "bar"}
|
|
)
|
|
eq_(
|
|
dialect.set_engine_execution_options.mock_calls,
|
|
[call(engine, {"foo": "bar"})],
|
|
)
|
|
|
|
def test_propagate_engine_to_connection(self, testing_engine):
|
|
engine = testing_engine(
|
|
"sqlite://", options=dict(execution_options={"foo": "bar"})
|
|
)
|
|
with engine.connect() as conn:
|
|
eq_(conn._execution_options, {"foo": "bar"})
|
|
|
|
def test_propagate_option_engine_to_connection(self, testing_engine):
|
|
e1 = testing_engine(
|
|
"sqlite://", options=dict(execution_options={"foo": "bar"})
|
|
)
|
|
e2 = e1.execution_options(bat="hoho")
|
|
c1 = e1.connect()
|
|
c2 = e2.connect()
|
|
eq_(c1._execution_options, {"foo": "bar"})
|
|
eq_(c2._execution_options, {"foo": "bar", "bat": "hoho"})
|
|
|
|
c1.close()
|
|
c2.close()
|
|
|
|
def test_get_engine_execution_options(self, testing_engine):
|
|
engine = testing_engine("sqlite://")
|
|
engine.dialect = Mock()
|
|
e2 = engine.execution_options(foo="bar")
|
|
|
|
eq_(e2.get_execution_options(), {"foo": "bar"})
|
|
|
|
def test_get_connection_execution_options(self, testing_engine):
|
|
engine = testing_engine("sqlite://", options=dict(_initialize=False))
|
|
engine.dialect = Mock()
|
|
with engine.connect() as conn:
|
|
c = conn.execution_options(foo="bar")
|
|
|
|
eq_(c.get_execution_options(), {"foo": "bar"})
|
|
|
|
|
|
class EngineEventsTest(fixtures.TestBase):
|
|
|
|
def teardown_test(self):
|
|
Engine.dispatch._clear()
|
|
Engine._has_events = False
|
|
|
|
def _assert_stmts(self, expected, received):
|
|
list(received)
|
|
|
|
for stmt, params, posn in expected:
|
|
if not received:
|
|
assert False, "Nothing available for stmt: %s" % stmt
|
|
while received:
|
|
teststmt, testparams, testmultiparams = received.pop(0)
|
|
teststmt = (
|
|
re.compile(r"[\n\t ]+", re.M).sub(" ", teststmt).strip()
|
|
)
|
|
if teststmt.startswith(stmt) and (
|
|
testparams == params or testparams == posn
|
|
):
|
|
break
|
|
|
|
def test_engine_connect(self, testing_engine):
|
|
e1 = testing_engine(config.db_url)
|
|
|
|
canary = Mock()
|
|
|
|
# use a real def to trigger legacy signature decorator
|
|
# logic, if present
|
|
def thing(conn):
|
|
canary(conn)
|
|
|
|
event.listen(e1, "engine_connect", thing)
|
|
|
|
c1 = e1.connect()
|
|
c1.close()
|
|
|
|
c2 = e1.connect()
|
|
c2.close()
|
|
|
|
eq_(canary.mock_calls, [mock.call(c1), mock.call(c2)])
|
|
|
|
def test_per_engine_independence(self, testing_engine):
|
|
e1 = testing_engine(config.db_url)
|
|
e2 = testing_engine(config.db_url)
|
|
|
|
canary = Mock()
|
|
event.listen(e1, "before_execute", canary)
|
|
s1 = select(1)
|
|
s2 = select(2)
|
|
|
|
with e1.connect() as conn:
|
|
conn.execute(s1)
|
|
|
|
with e2.connect() as conn:
|
|
conn.execute(s2)
|
|
eq_([arg[1][1] for arg in canary.mock_calls], [s1])
|
|
event.listen(e2, "before_execute", canary)
|
|
|
|
with e1.connect() as conn:
|
|
conn.execute(s1)
|
|
|
|
with e2.connect() as conn:
|
|
conn.execute(s2)
|
|
eq_([arg[1][1] for arg in canary.mock_calls], [s1, s1, s2])
|
|
|
|
def test_per_engine_plus_global(self, testing_engine):
|
|
canary = Mock()
|
|
event.listen(Engine, "before_execute", canary.be1)
|
|
e1 = testing_engine(config.db_url)
|
|
e2 = testing_engine(config.db_url)
|
|
|
|
event.listen(e1, "before_execute", canary.be2)
|
|
|
|
event.listen(Engine, "before_execute", canary.be3)
|
|
|
|
with e1.connect() as conn:
|
|
conn.execute(select(1))
|
|
eq_(canary.be1.call_count, 1)
|
|
eq_(canary.be2.call_count, 1)
|
|
|
|
with e2.connect() as conn:
|
|
conn.execute(select(1))
|
|
|
|
eq_(canary.be1.call_count, 2)
|
|
eq_(canary.be2.call_count, 1)
|
|
eq_(canary.be3.call_count, 2)
|
|
|
|
def test_option_engine_registration_issue_one(self):
|
|
"""test #12289"""
|
|
|
|
e1 = create_engine(testing.db.url)
|
|
e2 = e1.execution_options(foo="bar")
|
|
e3 = e2.execution_options(isolation_level="AUTOCOMMIT")
|
|
|
|
eq_(
|
|
e3._execution_options,
|
|
{"foo": "bar", "isolation_level": "AUTOCOMMIT"},
|
|
)
|
|
|
|
def test_option_engine_registration_issue_two(self):
|
|
"""test #12289"""
|
|
|
|
e1 = create_engine(testing.db.url)
|
|
e2 = e1.execution_options(foo="bar")
|
|
|
|
@event.listens_for(e2, "engine_connect")
|
|
def r1(*arg, **kw):
|
|
pass
|
|
|
|
e3 = e2.execution_options(bat="hoho")
|
|
|
|
@event.listens_for(e3, "engine_connect")
|
|
def r2(*arg, **kw):
|
|
pass
|
|
|
|
eq_(e3._execution_options, {"foo": "bar", "bat": "hoho"})
|
|
|
|
def test_emit_sql_in_autobegin(self, testing_engine):
|
|
e1 = testing_engine(config.db_url)
|
|
|
|
canary = Mock()
|
|
|
|
@event.listens_for(e1, "begin")
|
|
def begin(connection):
|
|
result = connection.execute(select(1)).scalar()
|
|
canary.got_result(result)
|
|
|
|
with e1.connect() as conn:
|
|
conn.execute(select(1)).scalar()
|
|
|
|
assert conn.in_transaction()
|
|
|
|
conn.commit()
|
|
|
|
assert not conn.in_transaction()
|
|
|
|
eq_(canary.mock_calls, [call.got_result(1)])
|
|
|
|
def test_per_connection_plus_engine(self, testing_engine):
|
|
canary = Mock()
|
|
e1 = testing_engine(config.db_url)
|
|
|
|
event.listen(e1, "before_execute", canary.be1)
|
|
|
|
conn = e1.connect()
|
|
event.listen(conn, "before_execute", canary.be2)
|
|
conn.execute(select(1))
|
|
|
|
eq_(canary.be1.call_count, 1)
|
|
eq_(canary.be2.call_count, 1)
|
|
|
|
@testing.combinations(
|
|
(True, False),
|
|
(True, True),
|
|
(False, False),
|
|
argnames="mock_out_on_connect, add_our_own_onconnect",
|
|
)
|
|
def test_insert_connect_is_definitely_first(
|
|
self, mock_out_on_connect, add_our_own_onconnect, testing_engine
|
|
):
|
|
"""test issue #5708.
|
|
|
|
We want to ensure that a single "connect" event may be invoked
|
|
*before* dialect initialize as well as before dialect on_connects.
|
|
|
|
This is also partially reliant on the changes we made as a result of
|
|
#5497, however here we go further with the changes and remove use
|
|
of the pool first_connect() event entirely so that the startup
|
|
for a dialect is fully consistent.
|
|
|
|
"""
|
|
if mock_out_on_connect:
|
|
if add_our_own_onconnect:
|
|
|
|
def our_connect(connection):
|
|
m1.our_connect("our connect event")
|
|
|
|
patcher = mock.patch.object(
|
|
config.db.dialect.__class__,
|
|
"on_connect",
|
|
lambda self: our_connect,
|
|
)
|
|
else:
|
|
patcher = mock.patch.object(
|
|
config.db.dialect.__class__,
|
|
"on_connect",
|
|
lambda self: None,
|
|
)
|
|
else:
|
|
patcher = nullcontext()
|
|
|
|
with patcher:
|
|
e1 = testing_engine(config.db_url)
|
|
|
|
initialize = e1.dialect.initialize
|
|
|
|
def init(connection):
|
|
initialize(connection)
|
|
connection.execute(select(1))
|
|
|
|
# begin mock added as part of migration to future only
|
|
# where we don't want anything related to begin() happening
|
|
# as part of create
|
|
# note we can't use an event to ensure begin() is not called
|
|
# because create also blocks events from happening
|
|
with (
|
|
mock.patch.object(
|
|
e1.dialect, "initialize", side_effect=init
|
|
) as m1,
|
|
mock.patch.object(e1._connection_cls, "begin") as begin_mock,
|
|
):
|
|
|
|
@event.listens_for(e1, "connect", insert=True)
|
|
def go1(dbapi_conn, xyz):
|
|
m1.foo("custom event first")
|
|
|
|
@event.listens_for(e1, "connect")
|
|
def go2(dbapi_conn, xyz):
|
|
m1.foo("custom event last")
|
|
|
|
c1 = e1.connect()
|
|
|
|
m1.bar("ok next connection")
|
|
|
|
c2 = e1.connect()
|
|
|
|
# this happens with sqlite singletonthreadpool.
|
|
# we can almost use testing.requires.independent_connections
|
|
# but sqlite file backend will also have independent
|
|
# connections here.
|
|
its_the_same_connection = (
|
|
c1.connection.dbapi_connection
|
|
is c2.connection.dbapi_connection
|
|
)
|
|
c1.close()
|
|
c2.close()
|
|
|
|
eq_(begin_mock.mock_calls, [])
|
|
|
|
if add_our_own_onconnect:
|
|
calls = [
|
|
mock.call.foo("custom event first"),
|
|
mock.call.our_connect("our connect event"),
|
|
mock.call(mock.ANY),
|
|
mock.call.foo("custom event last"),
|
|
mock.call.bar("ok next connection"),
|
|
]
|
|
else:
|
|
calls = [
|
|
mock.call.foo("custom event first"),
|
|
mock.call(mock.ANY),
|
|
mock.call.foo("custom event last"),
|
|
mock.call.bar("ok next connection"),
|
|
]
|
|
|
|
if not its_the_same_connection:
|
|
if add_our_own_onconnect:
|
|
calls.extend(
|
|
[
|
|
mock.call.foo("custom event first"),
|
|
mock.call.our_connect("our connect event"),
|
|
mock.call.foo("custom event last"),
|
|
]
|
|
)
|
|
else:
|
|
calls.extend(
|
|
[
|
|
mock.call.foo("custom event first"),
|
|
mock.call.foo("custom event last"),
|
|
]
|
|
)
|
|
eq_(m1.mock_calls, calls)
|
|
|
|
def test_new_exec_driver_sql_no_events(self):
|
|
m1 = Mock()
|
|
|
|
with testing.db.connect() as conn:
|
|
event.listen(conn, "before_execute", m1.before_execute)
|
|
event.listen(conn, "after_execute", m1.after_execute)
|
|
conn.exec_driver_sql(str(select(1).compile(testing.db)))
|
|
eq_(m1.mock_calls, [])
|
|
|
|
def test_add_event_after_connect(self, testing_engine):
|
|
# new feature as of #2978
|
|
|
|
canary = Mock()
|
|
e1 = testing_engine(config.db_url)
|
|
assert not e1._has_events
|
|
|
|
conn = e1.connect()
|
|
|
|
event.listen(e1, "before_execute", canary.be1)
|
|
conn.execute(select(1))
|
|
|
|
eq_(canary.be1.call_count, 1)
|
|
|
|
def test_force_conn_events_false(self, testing_engine):
|
|
canary = Mock()
|
|
e1 = testing_engine(config.db_url)
|
|
assert not e1._has_events
|
|
|
|
event.listen(e1, "before_execute", canary.be1)
|
|
|
|
conn = e1._connection_cls(
|
|
e1, connection=e1.raw_connection(), _has_events=False
|
|
)
|
|
|
|
conn.execute(select(1))
|
|
|
|
eq_(canary.be1.call_count, 0)
|
|
|
|
def test_cursor_events_ctx_execute_scalar(self, testing_engine):
|
|
canary = Mock()
|
|
e1 = testing_engine(config.db_url)
|
|
|
|
event.listen(e1, "before_cursor_execute", canary.bce)
|
|
event.listen(e1, "after_cursor_execute", canary.ace)
|
|
|
|
stmt = str(select(1).compile(dialect=e1.dialect))
|
|
|
|
with e1.connect() as conn:
|
|
dialect = conn.dialect
|
|
|
|
ctx = dialect.execution_ctx_cls._init_statement(
|
|
dialect, conn, conn.connection, {}, stmt, {}
|
|
)
|
|
|
|
ctx._execute_scalar(stmt, Integer())
|
|
|
|
eq_(
|
|
canary.bce.mock_calls,
|
|
[call(conn, ctx.cursor, stmt, ctx.parameters[0], ctx, False)],
|
|
)
|
|
eq_(
|
|
canary.ace.mock_calls,
|
|
[call(conn, ctx.cursor, stmt, ctx.parameters[0], ctx, False)],
|
|
)
|
|
|
|
def test_cursor_events_execute(self, testing_engine):
|
|
canary = Mock()
|
|
e1 = testing_engine(config.db_url)
|
|
|
|
event.listen(e1, "before_cursor_execute", canary.bce)
|
|
event.listen(e1, "after_cursor_execute", canary.ace)
|
|
|
|
stmt = str(select(1).compile(dialect=e1.dialect))
|
|
|
|
with e1.connect() as conn:
|
|
result = conn.exec_driver_sql(stmt)
|
|
eq_(result.scalar(), 1)
|
|
|
|
ctx = result.context
|
|
eq_(
|
|
canary.bce.mock_calls,
|
|
[call(conn, ctx.cursor, stmt, ctx.parameters[0], ctx, False)],
|
|
)
|
|
eq_(
|
|
canary.ace.mock_calls,
|
|
[call(conn, ctx.cursor, stmt, ctx.parameters[0], ctx, False)],
|
|
)
|
|
|
|
@testing.combinations(
|
|
(
|
|
([{"x": 5, "y": 10}, {"x": 8, "y": 9}],),
|
|
{},
|
|
[{"x": 5, "y": 10}, {"x": 8, "y": 9}],
|
|
{},
|
|
),
|
|
(({"z": 10},), {}, [], {"z": 10}),
|
|
argnames="multiparams, params, expected_multiparams, expected_params",
|
|
)
|
|
def test_modify_parameters_from_event_one(
|
|
self,
|
|
multiparams,
|
|
params,
|
|
expected_multiparams,
|
|
expected_params,
|
|
testing_engine,
|
|
):
|
|
# this is testing both the normalization added to parameters
|
|
# as of I97cb4d06adfcc6b889f10d01cc7775925cffb116 as well as
|
|
# that the return value from the event is taken as the new set
|
|
# of parameters.
|
|
def before_execute(
|
|
conn, clauseelement, multiparams, params, execution_options
|
|
):
|
|
eq_(multiparams, expected_multiparams)
|
|
eq_(params, expected_params)
|
|
return clauseelement, (), {"q": "15"}
|
|
|
|
def after_execute(
|
|
conn, clauseelement, multiparams, params, result, execution_options
|
|
):
|
|
eq_(multiparams, ())
|
|
eq_(params, {"q": "15"})
|
|
|
|
e1 = testing_engine(config.db_url)
|
|
event.listen(e1, "before_execute", before_execute, retval=True)
|
|
event.listen(e1, "after_execute", after_execute)
|
|
|
|
with e1.connect() as conn:
|
|
result = conn.execute(
|
|
select(bindparam("q", type_=String)), *multiparams, **params
|
|
)
|
|
eq_(result.all(), [("15",)])
|
|
|
|
@testing.provide_metadata
|
|
def test_modify_parameters_from_event_two(self, connection):
|
|
t = Table("t", self.metadata, Column("q", Integer))
|
|
|
|
t.create(connection)
|
|
|
|
def before_execute(
|
|
conn, clauseelement, multiparams, params, execution_options
|
|
):
|
|
return clauseelement, [{"q": 15}, {"q": 19}], {}
|
|
|
|
event.listen(connection, "before_execute", before_execute, retval=True)
|
|
connection.execute(t.insert(), {"q": 12})
|
|
event.remove(connection, "before_execute", before_execute)
|
|
|
|
eq_(
|
|
connection.execute(select(t).order_by(t.c.q)).fetchall(),
|
|
[(15,), (19,)],
|
|
)
|
|
|
|
def test_modify_parameters_from_event_three(
|
|
self, connection, testing_engine
|
|
):
|
|
def before_execute(
|
|
conn, clauseelement, multiparams, params, execution_options
|
|
):
|
|
return clauseelement, [{"q": 15}, {"q": 19}], {"q": 7}
|
|
|
|
e1 = testing_engine(config.db_url)
|
|
event.listen(e1, "before_execute", before_execute, retval=True)
|
|
|
|
with expect_raises_message(
|
|
tsa.exc.InvalidRequestError,
|
|
"Event handler can't return non-empty multiparams "
|
|
"and params at the same time",
|
|
):
|
|
with e1.connect() as conn:
|
|
conn.execute(select(literal("1")))
|
|
|
|
@testing.only_on("sqlite")
|
|
def test_dont_modify_statement_driversql(self, connection):
|
|
m1 = mock.Mock()
|
|
|
|
@event.listens_for(connection, "before_execute", retval=True)
|
|
def _modify(
|
|
conn, clauseelement, multiparams, params, execution_options
|
|
):
|
|
m1.run_event()
|
|
return clauseelement.replace("hi", "there"), multiparams, params
|
|
|
|
# the event does not take effect for the "driver SQL" option
|
|
eq_(connection.exec_driver_sql("select 'hi'").scalar(), "hi")
|
|
|
|
# event is not called at all
|
|
eq_(m1.mock_calls, [])
|
|
|
|
@testing.only_on("sqlite")
|
|
def test_modify_statement_internal_driversql(self, connection):
|
|
m1 = mock.Mock()
|
|
|
|
@event.listens_for(connection, "before_execute", retval=True)
|
|
def _modify(
|
|
conn, clauseelement, multiparams, params, execution_options
|
|
):
|
|
m1.run_event()
|
|
return clauseelement.replace("hi", "there"), multiparams, params
|
|
|
|
eq_(
|
|
connection.exec_driver_sql("select 'hi'").scalar(),
|
|
"hi",
|
|
)
|
|
|
|
eq_(m1.mock_calls, [])
|
|
|
|
def test_modify_statement_clauseelement(self, connection):
|
|
@event.listens_for(connection, "before_execute", retval=True)
|
|
def _modify(
|
|
conn, clauseelement, multiparams, params, execution_options
|
|
):
|
|
return select(literal_column("'there'")), multiparams, params
|
|
|
|
eq_(connection.scalar(select(literal_column("'hi'"))), "there")
|
|
|
|
def test_argument_format_execute(self, testing_engine):
|
|
def before_execute(
|
|
conn, clauseelement, multiparams, params, execution_options
|
|
):
|
|
assert isinstance(multiparams, (list, tuple))
|
|
assert isinstance(params, collections_abc.Mapping)
|
|
|
|
def after_execute(
|
|
conn, clauseelement, multiparams, params, result, execution_options
|
|
):
|
|
assert isinstance(multiparams, (list, tuple))
|
|
assert isinstance(params, collections_abc.Mapping)
|
|
|
|
e1 = testing_engine(config.db_url)
|
|
event.listen(e1, "before_execute", before_execute)
|
|
event.listen(e1, "after_execute", after_execute)
|
|
|
|
with e1.connect() as conn:
|
|
conn.execute(select(1))
|
|
conn.execute(select(1).compile(dialect=e1.dialect).statement)
|
|
|
|
@testing.emits_warning("The garbage collector is trying to clean up")
|
|
def test_execute_events(self):
|
|
stmts = []
|
|
cursor_stmts = []
|
|
|
|
def execute(
|
|
conn, clauseelement, multiparams, params, execution_options
|
|
):
|
|
stmts.append((str(clauseelement), params, multiparams))
|
|
|
|
def cursor_execute(
|
|
conn, cursor, statement, parameters, context, executemany
|
|
):
|
|
cursor_stmts.append((str(statement), parameters, None))
|
|
|
|
# TODO: this test is kind of a mess
|
|
|
|
for engine in [
|
|
engines.testing_engine(),
|
|
engines.testing_engine().connect(),
|
|
]:
|
|
event.listen(engine, "before_execute", execute)
|
|
event.listen(engine, "before_cursor_execute", cursor_execute)
|
|
m = MetaData()
|
|
t1 = Table(
|
|
"t1",
|
|
m,
|
|
Column("c1", Integer, primary_key=True),
|
|
Column(
|
|
"c2",
|
|
String(50),
|
|
default=func.lower("Foo"),
|
|
primary_key=True,
|
|
),
|
|
implicit_returning=False,
|
|
)
|
|
|
|
if isinstance(engine, Connection):
|
|
ctx = None
|
|
conn = engine
|
|
else:
|
|
ctx = conn = engine.connect()
|
|
|
|
trans = conn.begin()
|
|
try:
|
|
m.create_all(conn, checkfirst=False)
|
|
try:
|
|
conn.execute(t1.insert(), dict(c1=5, c2="some data"))
|
|
conn.execute(t1.insert(), dict(c1=6))
|
|
eq_(
|
|
conn.execute(text("select * from t1")).fetchall(),
|
|
[(5, "some data"), (6, "foo")],
|
|
)
|
|
finally:
|
|
m.drop_all(conn)
|
|
trans.commit()
|
|
finally:
|
|
if ctx:
|
|
ctx.close()
|
|
|
|
compiled = [
|
|
("CREATE TABLE t1", {}, None),
|
|
(
|
|
"INSERT INTO t1 (c1, c2)",
|
|
{"c2": "some data", "c1": 5},
|
|
(),
|
|
),
|
|
("INSERT INTO t1 (c1, c2)", {"c1": 6}, ()),
|
|
("select * from t1", {}, None),
|
|
("DROP TABLE t1", {}, None),
|
|
]
|
|
|
|
cursor = [
|
|
("CREATE TABLE t1", {}, ()),
|
|
(
|
|
"INSERT INTO t1 (c1, c2)",
|
|
{"c2": "some data", "c1": 5},
|
|
(5, "some data"),
|
|
),
|
|
("SELECT lower", {"lower_2": "Foo"}, ("Foo",)),
|
|
(
|
|
"INSERT INTO t1 (c1, c2)",
|
|
{"c2": "foo", "c1": 6},
|
|
(6, "foo"),
|
|
),
|
|
("select * from t1", {}, ()),
|
|
("DROP TABLE t1", {}, ()),
|
|
]
|
|
self._assert_stmts(compiled, stmts)
|
|
self._assert_stmts(cursor, cursor_stmts)
|
|
|
|
def test_options(self):
|
|
canary = []
|
|
|
|
def execute(conn, *args, **kw):
|
|
canary.append("execute")
|
|
|
|
def cursor_execute(conn, *args, **kw):
|
|
canary.append("cursor_execute")
|
|
|
|
engine = engines.testing_engine()
|
|
event.listen(engine, "before_execute", execute)
|
|
event.listen(engine, "before_cursor_execute", cursor_execute)
|
|
conn = engine.connect()
|
|
c2 = conn.execution_options(foo="bar")
|
|
eq_(c2._execution_options, {"foo": "bar"})
|
|
c2.execute(select(1))
|
|
c3 = c2.execution_options(bar="bat")
|
|
eq_(c3._execution_options, {"foo": "bar", "bar": "bat"})
|
|
eq_(canary, ["execute", "cursor_execute"])
|
|
|
|
def test_generative_engine_event_dispatch(self):
|
|
canary = []
|
|
|
|
def l1(*arg, **kw):
|
|
canary.append("l1")
|
|
|
|
def l2(*arg, **kw):
|
|
canary.append("l2")
|
|
|
|
def l3(*arg, **kw):
|
|
canary.append("l3")
|
|
|
|
eng = engines.testing_engine(
|
|
options={"execution_options": {"base": "x1"}}
|
|
)
|
|
event.listen(eng, "before_execute", l1)
|
|
|
|
eng1 = eng.execution_options(foo="b1")
|
|
event.listen(eng, "before_execute", l2)
|
|
event.listen(eng1, "before_execute", l3)
|
|
|
|
with eng.connect() as conn:
|
|
conn.execute(select(1))
|
|
|
|
eq_(canary, ["l1", "l2"])
|
|
|
|
with eng1.connect() as conn:
|
|
conn.execute(select(1))
|
|
|
|
eq_(canary, ["l1", "l2", "l3", "l1", "l2"])
|
|
|
|
def test_clslevel_engine_event_options(self):
|
|
canary = []
|
|
|
|
def l1(*arg, **kw):
|
|
canary.append("l1")
|
|
|
|
def l2(*arg, **kw):
|
|
canary.append("l2")
|
|
|
|
def l3(*arg, **kw):
|
|
canary.append("l3")
|
|
|
|
def l4(*arg, **kw):
|
|
canary.append("l4")
|
|
|
|
event.listen(Engine, "before_execute", l1)
|
|
|
|
eng = engines.testing_engine(
|
|
options={"execution_options": {"base": "x1"}}
|
|
)
|
|
event.listen(eng, "before_execute", l2)
|
|
|
|
eng1 = eng.execution_options(foo="b1")
|
|
event.listen(eng, "before_execute", l3)
|
|
event.listen(eng1, "before_execute", l4)
|
|
|
|
with eng.connect() as conn:
|
|
conn.execute(select(1))
|
|
|
|
eq_(canary, ["l1", "l2", "l3"])
|
|
|
|
with eng1.connect() as conn:
|
|
conn.execute(select(1))
|
|
|
|
eq_(canary, ["l1", "l2", "l3", "l4", "l1", "l2", "l3"])
|
|
|
|
canary[:] = []
|
|
|
|
event.remove(Engine, "before_execute", l1)
|
|
event.remove(eng1, "before_execute", l4)
|
|
event.remove(eng, "before_execute", l3)
|
|
|
|
with eng1.connect() as conn:
|
|
conn.execute(select(1))
|
|
eq_(canary, ["l2"])
|
|
|
|
def test_cant_listen_to_option_engine(self):
|
|
from sqlalchemy.engine import base
|
|
|
|
def evt(*arg, **kw):
|
|
pass
|
|
|
|
assert_raises_message(
|
|
tsa.exc.InvalidRequestError,
|
|
r"Can't assign an event directly to the "
|
|
"<class 'sqlalchemy.engine.base.OptionEngine'> class",
|
|
event.listen,
|
|
base.OptionEngine,
|
|
"before_cursor_execute",
|
|
evt,
|
|
)
|
|
|
|
def test_dispose_event(self, testing_engine):
|
|
canary = Mock()
|
|
eng = testing_engine(testing.db.url)
|
|
event.listen(eng, "engine_disposed", canary)
|
|
|
|
conn = eng.connect()
|
|
conn.close()
|
|
eng.dispose()
|
|
|
|
conn = eng.connect()
|
|
conn.close()
|
|
|
|
eq_(canary.mock_calls, [call(eng)])
|
|
|
|
eng.dispose()
|
|
|
|
eq_(canary.mock_calls, [call(eng), call(eng)])
|
|
|
|
@testing.combinations(True, False, argnames="close")
|
|
def test_close_parameter(self, testing_engine, close):
|
|
eng = testing_engine(
|
|
options=dict(
|
|
pool_size=1,
|
|
max_overflow=0,
|
|
poolclass=(
|
|
QueuePool
|
|
if not testing.db.dialect.is_async
|
|
else AsyncAdaptedQueuePool
|
|
),
|
|
)
|
|
)
|
|
|
|
conn = eng.connect()
|
|
dbapi_conn_one = conn.connection.dbapi_connection
|
|
conn.close()
|
|
|
|
eng_copy = copy.copy(eng)
|
|
eng_copy.dispose(close=close)
|
|
|
|
copy_conn = eng_copy.connect()
|
|
dbapi_conn_two = copy_conn.connection.dbapi_connection
|
|
|
|
is_not(dbapi_conn_one, dbapi_conn_two)
|
|
|
|
conn = eng.connect()
|
|
if close:
|
|
is_not(dbapi_conn_one, conn.connection.dbapi_connection)
|
|
else:
|
|
is_(dbapi_conn_one, conn.connection.dbapi_connection)
|
|
|
|
conn.close()
|
|
copy_conn.close()
|
|
|
|
def test_retval_flag(self):
|
|
canary = []
|
|
|
|
def tracker(name):
|
|
def go(conn, *args, **kw):
|
|
canary.append(name)
|
|
|
|
return go
|
|
|
|
def execute(
|
|
conn, clauseelement, multiparams, params, execution_options
|
|
):
|
|
canary.append("execute")
|
|
return clauseelement, multiparams, params
|
|
|
|
def cursor_execute(
|
|
conn, cursor, statement, parameters, context, executemany
|
|
):
|
|
canary.append("cursor_execute")
|
|
return statement, parameters
|
|
|
|
engine = engines.testing_engine()
|
|
|
|
assert_raises(
|
|
tsa.exc.ArgumentError,
|
|
event.listen,
|
|
engine,
|
|
"begin",
|
|
tracker("begin"),
|
|
retval=True,
|
|
)
|
|
|
|
event.listen(engine, "before_execute", execute, retval=True)
|
|
event.listen(
|
|
engine, "before_cursor_execute", cursor_execute, retval=True
|
|
)
|
|
with engine.connect() as conn:
|
|
conn.execute(select(1))
|
|
eq_(canary, ["execute", "cursor_execute"])
|
|
|
|
def test_execution_options(self):
|
|
engine = engines.testing_engine()
|
|
|
|
engine_tracker = Mock()
|
|
conn_tracker = Mock()
|
|
|
|
event.listen(engine, "set_engine_execution_options", engine_tracker)
|
|
event.listen(engine, "set_connection_execution_options", conn_tracker)
|
|
|
|
e2 = engine.execution_options(e1="opt_e1")
|
|
c1 = engine.connect()
|
|
c2 = c1.execution_options(c1="opt_c1")
|
|
c3 = e2.connect()
|
|
c4 = c3.execution_options(c3="opt_c3")
|
|
eq_(engine_tracker.mock_calls, [call(e2, {"e1": "opt_e1"})])
|
|
eq_(
|
|
conn_tracker.mock_calls,
|
|
[call(c2, {"c1": "opt_c1"}), call(c4, {"c3": "opt_c3"})],
|
|
)
|
|
|
|
def test_execution_options_modify_inplace(self):
|
|
engine = engines.testing_engine()
|
|
|
|
@event.listens_for(engine, "set_engine_execution_options")
|
|
def engine_tracker(conn, opt):
|
|
opt["engine_tracked"] = True
|
|
|
|
@event.listens_for(engine, "set_connection_execution_options")
|
|
def conn_tracker(conn, opt):
|
|
opt["conn_tracked"] = True
|
|
|
|
with (
|
|
mock.patch.object(
|
|
engine.dialect, "set_connection_execution_options"
|
|
) as conn_opt,
|
|
mock.patch.object(
|
|
engine.dialect, "set_engine_execution_options"
|
|
) as engine_opt,
|
|
):
|
|
e2 = engine.execution_options(e1="opt_e1")
|
|
c1 = engine.connect()
|
|
c2 = c1.execution_options(c1="opt_c1")
|
|
|
|
is_not(e2, engine)
|
|
is_(c1, c2)
|
|
|
|
eq_(e2._execution_options, {"e1": "opt_e1", "engine_tracked": True})
|
|
eq_(c2._execution_options, {"c1": "opt_c1", "conn_tracked": True})
|
|
eq_(
|
|
engine_opt.mock_calls,
|
|
[mock.call(e2, {"e1": "opt_e1", "engine_tracked": True})],
|
|
)
|
|
eq_(
|
|
conn_opt.mock_calls,
|
|
[mock.call(c1, {"c1": "opt_c1", "conn_tracked": True})],
|
|
)
|
|
|
|
@testing.requires.sequences
|
|
@testing.provide_metadata
|
|
def test_cursor_execute(self):
|
|
canary = []
|
|
|
|
def tracker(name):
|
|
def go(conn, cursor, statement, parameters, context, executemany):
|
|
canary.append((statement, context))
|
|
|
|
return go
|
|
|
|
engine = engines.testing_engine()
|
|
|
|
t = Table(
|
|
"t",
|
|
self.metadata,
|
|
Column(
|
|
"x",
|
|
Integer,
|
|
normalize_sequence(config, Sequence("t_id_seq")),
|
|
primary_key=True,
|
|
),
|
|
implicit_returning=False,
|
|
)
|
|
self.metadata.create_all(engine)
|
|
|
|
with engine.begin() as conn:
|
|
event.listen(
|
|
conn, "before_cursor_execute", tracker("cursor_execute")
|
|
)
|
|
conn.execute(t.insert())
|
|
|
|
# we see the sequence pre-executed in the first call
|
|
assert "t_id_seq" in canary[0][0]
|
|
assert "INSERT" in canary[1][0]
|
|
# same context
|
|
is_(canary[0][1], canary[1][1])
|
|
|
|
def test_transactional(self):
|
|
canary = []
|
|
|
|
def tracker(name):
|
|
def go(conn, *args, **kw):
|
|
canary.append(name)
|
|
|
|
return go
|
|
|
|
engine = engines.testing_engine()
|
|
event.listen(engine, "before_execute", tracker("execute"))
|
|
event.listen(
|
|
engine, "before_cursor_execute", tracker("cursor_execute")
|
|
)
|
|
event.listen(engine, "begin", tracker("begin"))
|
|
event.listen(engine, "commit", tracker("commit"))
|
|
event.listen(engine, "rollback", tracker("rollback"))
|
|
|
|
with engine.connect() as conn:
|
|
trans = conn.begin()
|
|
conn.execute(select(1))
|
|
trans.rollback()
|
|
trans = conn.begin()
|
|
conn.execute(select(1))
|
|
trans.commit()
|
|
|
|
eq_(
|
|
canary,
|
|
[
|
|
"begin",
|
|
"execute",
|
|
"cursor_execute",
|
|
"rollback",
|
|
"begin",
|
|
"execute",
|
|
"cursor_execute",
|
|
"commit",
|
|
],
|
|
)
|
|
|
|
def test_transactional_named(self):
|
|
canary = []
|
|
|
|
def tracker(name):
|
|
def go(*args, **kw):
|
|
canary.append((name, set(kw)))
|
|
|
|
return go
|
|
|
|
engine = engines.testing_engine()
|
|
event.listen(engine, "before_execute", tracker("execute"), named=True)
|
|
event.listen(
|
|
engine,
|
|
"before_cursor_execute",
|
|
tracker("cursor_execute"),
|
|
named=True,
|
|
)
|
|
event.listen(engine, "begin", tracker("begin"), named=True)
|
|
event.listen(engine, "commit", tracker("commit"), named=True)
|
|
event.listen(engine, "rollback", tracker("rollback"), named=True)
|
|
|
|
with engine.connect() as conn:
|
|
trans = conn.begin()
|
|
conn.execute(select(1))
|
|
trans.rollback()
|
|
trans = conn.begin()
|
|
conn.execute(select(1))
|
|
trans.commit()
|
|
|
|
eq_(
|
|
canary,
|
|
[
|
|
("begin", {"conn"}),
|
|
(
|
|
"execute",
|
|
{
|
|
"conn",
|
|
"clauseelement",
|
|
"multiparams",
|
|
"params",
|
|
"execution_options",
|
|
},
|
|
),
|
|
(
|
|
"cursor_execute",
|
|
{
|
|
"conn",
|
|
"cursor",
|
|
"executemany",
|
|
"statement",
|
|
"parameters",
|
|
"context",
|
|
},
|
|
),
|
|
("rollback", {"conn"}),
|
|
("begin", {"conn"}),
|
|
(
|
|
"execute",
|
|
{
|
|
"conn",
|
|
"clauseelement",
|
|
"multiparams",
|
|
"params",
|
|
"execution_options",
|
|
},
|
|
),
|
|
(
|
|
"cursor_execute",
|
|
{
|
|
"conn",
|
|
"cursor",
|
|
"executemany",
|
|
"statement",
|
|
"parameters",
|
|
"context",
|
|
},
|
|
),
|
|
("commit", {"conn"}),
|
|
],
|
|
)
|
|
|
|
@testing.requires.savepoints
|
|
@testing.requires.two_phase_transactions
|
|
def test_transactional_advanced(self):
|
|
canary1 = []
|
|
|
|
def tracker1(name):
|
|
def go(*args, **kw):
|
|
canary1.append(name)
|
|
|
|
return go
|
|
|
|
canary2 = []
|
|
|
|
def tracker2(name):
|
|
def go(*args, **kw):
|
|
canary2.append(name)
|
|
|
|
return go
|
|
|
|
engine = engines.testing_engine()
|
|
for name in [
|
|
"begin",
|
|
"savepoint",
|
|
"rollback_savepoint",
|
|
"release_savepoint",
|
|
"rollback",
|
|
"begin_twophase",
|
|
"prepare_twophase",
|
|
"commit_twophase",
|
|
]:
|
|
event.listen(engine, "%s" % name, tracker1(name))
|
|
|
|
conn = engine.connect()
|
|
for name in [
|
|
"begin",
|
|
"savepoint",
|
|
"rollback_savepoint",
|
|
"release_savepoint",
|
|
"rollback",
|
|
"begin_twophase",
|
|
"prepare_twophase",
|
|
"commit_twophase",
|
|
]:
|
|
event.listen(conn, "%s" % name, tracker2(name))
|
|
|
|
trans = conn.begin()
|
|
trans2 = conn.begin_nested()
|
|
conn.execute(select(1))
|
|
trans2.rollback()
|
|
trans2 = conn.begin_nested()
|
|
conn.execute(select(1))
|
|
trans2.commit()
|
|
trans.rollback()
|
|
|
|
trans = conn.begin_twophase()
|
|
conn.execute(select(1))
|
|
trans.prepare()
|
|
trans.commit()
|
|
|
|
eq_(
|
|
canary1,
|
|
[
|
|
"begin",
|
|
"savepoint",
|
|
"rollback_savepoint",
|
|
"savepoint",
|
|
"release_savepoint",
|
|
"rollback",
|
|
"begin_twophase",
|
|
"prepare_twophase",
|
|
"commit_twophase",
|
|
],
|
|
)
|
|
eq_(
|
|
canary2,
|
|
[
|
|
"begin",
|
|
"savepoint",
|
|
"rollback_savepoint",
|
|
"savepoint",
|
|
"release_savepoint",
|
|
"rollback",
|
|
"begin_twophase",
|
|
"prepare_twophase",
|
|
"commit_twophase",
|
|
],
|
|
)
|
|
conn.close()
|
|
|
|
|
|
class HandleErrorTest(fixtures.TestBase):
|
|
__sparse_driver_backend__ = True
|
|
|
|
def teardown_test(self):
|
|
Engine.dispatch._clear()
|
|
Engine._has_events = False
|
|
|
|
def test_handle_error(self):
|
|
engine = engines.testing_engine()
|
|
canary = Mock(return_value=None)
|
|
|
|
event.listen(engine, "handle_error", canary)
|
|
|
|
with engine.connect() as conn:
|
|
try:
|
|
conn.exec_driver_sql("SELECT FOO FROM I_DONT_EXIST")
|
|
assert False
|
|
except tsa.exc.DBAPIError as e:
|
|
ctx = canary.mock_calls[0][1][0]
|
|
|
|
eq_(ctx.original_exception, e.orig)
|
|
is_(ctx.sqlalchemy_exception, e)
|
|
eq_(ctx.statement, "SELECT FOO FROM I_DONT_EXIST")
|
|
|
|
def test_exception_event_reraise(self):
|
|
engine = engines.testing_engine()
|
|
|
|
class MyException(Exception):
|
|
pass
|
|
|
|
@event.listens_for(engine, "handle_error", retval=True)
|
|
def err(context):
|
|
stmt = context.statement
|
|
exception = context.original_exception
|
|
if "ERROR ONE" in str(stmt):
|
|
return MyException("my exception")
|
|
elif "ERROR TWO" in str(stmt):
|
|
return exception
|
|
else:
|
|
return None
|
|
|
|
conn = engine.connect()
|
|
# case 1: custom exception
|
|
assert_raises_message(
|
|
MyException,
|
|
"my exception",
|
|
conn.exec_driver_sql,
|
|
"SELECT 'ERROR ONE' FROM I_DONT_EXIST",
|
|
)
|
|
# case 2: return the DBAPI exception we're given;
|
|
# no wrapping should occur
|
|
assert_raises(
|
|
conn.dialect.dbapi.Error,
|
|
conn.exec_driver_sql,
|
|
"SELECT 'ERROR TWO' FROM I_DONT_EXIST",
|
|
)
|
|
# case 3: normal wrapping
|
|
assert_raises(
|
|
tsa.exc.DBAPIError,
|
|
conn.exec_driver_sql,
|
|
"SELECT 'ERROR THREE' FROM I_DONT_EXIST",
|
|
)
|
|
|
|
def test_exception_event_reraise_chaining(self):
|
|
engine = engines.testing_engine()
|
|
|
|
class MyException1(Exception):
|
|
pass
|
|
|
|
class MyException2(Exception):
|
|
pass
|
|
|
|
class MyException3(Exception):
|
|
pass
|
|
|
|
@event.listens_for(engine, "handle_error", retval=True)
|
|
def err1(context):
|
|
stmt = context.statement
|
|
|
|
if (
|
|
"ERROR ONE" in str(stmt)
|
|
or "ERROR TWO" in str(stmt)
|
|
or "ERROR THREE" in str(stmt)
|
|
):
|
|
return MyException1("my exception")
|
|
elif "ERROR FOUR" in str(stmt):
|
|
raise MyException3("my exception short circuit")
|
|
|
|
@event.listens_for(engine, "handle_error", retval=True)
|
|
def err2(context):
|
|
stmt = context.statement
|
|
if (
|
|
"ERROR ONE" in str(stmt) or "ERROR FOUR" in str(stmt)
|
|
) and isinstance(context.chained_exception, MyException1):
|
|
raise MyException2("my exception chained")
|
|
elif "ERROR TWO" in str(stmt):
|
|
return context.chained_exception
|
|
else:
|
|
return None
|
|
|
|
conn = engine.connect()
|
|
|
|
with patch.object(
|
|
engine.dialect.execution_ctx_cls, "handle_dbapi_exception"
|
|
) as patched:
|
|
assert_raises_message(
|
|
MyException2,
|
|
"my exception chained",
|
|
conn.exec_driver_sql,
|
|
"SELECT 'ERROR ONE' FROM I_DONT_EXIST",
|
|
)
|
|
eq_(patched.call_count, 1)
|
|
|
|
with patch.object(
|
|
engine.dialect.execution_ctx_cls, "handle_dbapi_exception"
|
|
) as patched:
|
|
assert_raises(
|
|
MyException1,
|
|
conn.exec_driver_sql,
|
|
"SELECT 'ERROR TWO' FROM I_DONT_EXIST",
|
|
)
|
|
eq_(patched.call_count, 1)
|
|
|
|
with patch.object(
|
|
engine.dialect.execution_ctx_cls, "handle_dbapi_exception"
|
|
) as patched:
|
|
# test that non None from err1 isn't cancelled out
|
|
# by err2
|
|
assert_raises(
|
|
MyException1,
|
|
conn.exec_driver_sql,
|
|
"SELECT 'ERROR THREE' FROM I_DONT_EXIST",
|
|
)
|
|
eq_(patched.call_count, 1)
|
|
|
|
with patch.object(
|
|
engine.dialect.execution_ctx_cls, "handle_dbapi_exception"
|
|
) as patched:
|
|
assert_raises(
|
|
tsa.exc.DBAPIError,
|
|
conn.exec_driver_sql,
|
|
"SELECT 'ERROR FIVE' FROM I_DONT_EXIST",
|
|
)
|
|
eq_(patched.call_count, 1)
|
|
|
|
with patch.object(
|
|
engine.dialect.execution_ctx_cls, "handle_dbapi_exception"
|
|
) as patched:
|
|
assert_raises_message(
|
|
MyException3,
|
|
"my exception short circuit",
|
|
conn.exec_driver_sql,
|
|
"SELECT 'ERROR FOUR' FROM I_DONT_EXIST",
|
|
)
|
|
eq_(patched.call_count, 1)
|
|
|
|
@testing.only_on("sqlite", "using specific DB message")
|
|
def test_exception_no_autorollback(self):
|
|
"""with the 2.0 engine, a SQL statement will have run
|
|
"autobegin", so that we are in a transaction. so if an error
|
|
occurs, we report the error but stay in the transaction.
|
|
|
|
previously, we'd see the rollback failing due to autorollback
|
|
when transaction isn't started.
|
|
"""
|
|
engine = engines.testing_engine()
|
|
conn = engine.connect()
|
|
|
|
def boom(connection):
|
|
raise provision.dbapi_error(
|
|
config,
|
|
engine.dialect.dbapi.OperationalError,
|
|
"rollback failed",
|
|
)
|
|
|
|
with patch.object(conn.dialect, "do_rollback", boom):
|
|
assert_raises_message(
|
|
tsa.exc.OperationalError,
|
|
"no such table: i_dont_exist",
|
|
conn.exec_driver_sql,
|
|
"insert into i_dont_exist (x) values ('y')",
|
|
)
|
|
|
|
# we're still in a transaction
|
|
assert conn._transaction
|
|
|
|
# only fails when we actually call rollback
|
|
assert_raises_message(
|
|
tsa.exc.OperationalError,
|
|
"rollback failed",
|
|
conn.rollback,
|
|
)
|
|
|
|
def test_actual_autorollback(self):
|
|
"""manufacture an autorollback scenario that works in 2.x."""
|
|
|
|
engine = engines.testing_engine()
|
|
conn = engine.connect()
|
|
|
|
def boom(connection):
|
|
raise provision.dbapi_error(
|
|
config,
|
|
engine.dialect.dbapi.OperationalError,
|
|
"rollback failed",
|
|
)
|
|
|
|
@event.listens_for(conn, "begin")
|
|
def _do_begin(conn):
|
|
# run a breaking statement before begin actually happens
|
|
conn.exec_driver_sql("insert into i_dont_exist (x) values ('y')")
|
|
|
|
with patch.object(conn.dialect, "do_rollback", boom):
|
|
assert_raises_message(
|
|
tsa.exc.OperationalError,
|
|
"rollback failed",
|
|
conn.begin,
|
|
)
|
|
|
|
def test_exception_event_ad_hoc_context(self):
|
|
"""test that handle_error is called with a context in
|
|
cases where _handle_dbapi_error() is normally called without
|
|
any context.
|
|
|
|
"""
|
|
|
|
engine = engines.testing_engine()
|
|
|
|
listener = Mock(return_value=None)
|
|
event.listen(engine, "handle_error", listener)
|
|
|
|
nope = SomeException("nope")
|
|
|
|
class MyType(TypeDecorator):
|
|
impl = Integer
|
|
cache_ok = True
|
|
|
|
def process_bind_param(self, value, dialect):
|
|
raise nope
|
|
|
|
with engine.connect() as conn:
|
|
assert_raises_message(
|
|
tsa.exc.StatementError,
|
|
r"\(.*.SomeException\) " r"nope\n\[SQL\: u?SELECT 1 ",
|
|
conn.execute,
|
|
select(1).where(column("foo") == literal("bar", MyType())),
|
|
)
|
|
|
|
ctx = listener.mock_calls[0][1][0]
|
|
assert ctx.statement.startswith("SELECT 1 ")
|
|
is_(ctx.is_disconnect, False)
|
|
is_(ctx.original_exception, nope)
|
|
|
|
def test_exception_event_non_dbapi_error(self):
|
|
"""test that handle_error is called with a context in
|
|
cases where DBAPI raises an exception that is not a DBAPI
|
|
exception, e.g. internal errors or encoding problems.
|
|
|
|
"""
|
|
engine = engines.testing_engine()
|
|
|
|
listener = Mock(return_value=None)
|
|
event.listen(engine, "handle_error", listener)
|
|
|
|
nope = TypeError("I'm not a DBAPI error")
|
|
with engine.connect() as c:
|
|
c.connection.cursor = Mock(
|
|
return_value=Mock(execute=Mock(side_effect=nope))
|
|
)
|
|
|
|
assert_raises_message(
|
|
TypeError,
|
|
"I'm not a DBAPI error",
|
|
c.exec_driver_sql,
|
|
"select ",
|
|
)
|
|
ctx = listener.mock_calls[0][1][0]
|
|
eq_(ctx.statement, "select ")
|
|
is_(ctx.is_disconnect, False)
|
|
is_(ctx.original_exception, nope)
|
|
|
|
def test_exception_event_disable_handlers(self):
|
|
engine = engines.testing_engine()
|
|
|
|
class MyException1(Exception):
|
|
pass
|
|
|
|
@event.listens_for(engine, "handle_error")
|
|
def err1(context):
|
|
stmt = context.statement
|
|
|
|
if "ERROR_ONE" in str(stmt):
|
|
raise MyException1("my exception short circuit")
|
|
|
|
with engine.connect() as conn:
|
|
assert_raises(
|
|
tsa.exc.DBAPIError,
|
|
conn.execution_options(
|
|
skip_user_error_events=True
|
|
).exec_driver_sql,
|
|
"SELECT ERROR_ONE FROM I_DONT_EXIST",
|
|
)
|
|
|
|
assert_raises(
|
|
MyException1,
|
|
conn.execution_options(
|
|
skip_user_error_events=False
|
|
).exec_driver_sql,
|
|
"SELECT ERROR_ONE FROM I_DONT_EXIST",
|
|
)
|
|
|
|
def _test_alter_disconnect(self, orig_error, evt_value):
|
|
engine = engines.testing_engine()
|
|
|
|
@event.listens_for(engine, "handle_error")
|
|
def evt(ctx):
|
|
ctx.is_disconnect = evt_value
|
|
|
|
with patch.object(
|
|
engine.dialect, "is_disconnect", Mock(return_value=orig_error)
|
|
):
|
|
with engine.connect() as c:
|
|
try:
|
|
c.exec_driver_sql("SELECT x FROM nonexistent")
|
|
assert False
|
|
except tsa.exc.StatementError as st:
|
|
eq_(st.connection_invalidated, evt_value)
|
|
|
|
def test_alter_disconnect_to_true(self):
|
|
self._test_alter_disconnect(False, True)
|
|
self._test_alter_disconnect(True, True)
|
|
|
|
def test_alter_disconnect_to_false(self):
|
|
self._test_alter_disconnect(True, False)
|
|
self._test_alter_disconnect(False, False)
|
|
|
|
@testing.requires.independent_connections
|
|
def _test_alter_invalidate_pool_to_false(self, set_to_false):
|
|
orig_error = True
|
|
|
|
engine = engines.testing_engine()
|
|
|
|
@event.listens_for(engine, "handle_error")
|
|
def evt(ctx):
|
|
if set_to_false:
|
|
ctx.invalidate_pool_on_disconnect = False
|
|
|
|
c1, c2, c3 = (
|
|
engine.pool.connect(),
|
|
engine.pool.connect(),
|
|
engine.pool.connect(),
|
|
)
|
|
crecs = [conn._connection_record for conn in (c1, c2, c3)]
|
|
c1.close()
|
|
c2.close()
|
|
c3.close()
|
|
|
|
with patch.object(
|
|
engine.dialect, "is_disconnect", Mock(return_value=orig_error)
|
|
):
|
|
with engine.connect() as c:
|
|
target_crec = c.connection._connection_record
|
|
try:
|
|
c.exec_driver_sql("SELECT x FROM nonexistent")
|
|
assert False
|
|
except tsa.exc.StatementError as st:
|
|
eq_(st.connection_invalidated, True)
|
|
|
|
for crec in crecs:
|
|
if crec is target_crec or not set_to_false:
|
|
is_not(crec.dbapi_connection, crec.get_connection())
|
|
else:
|
|
is_(crec.dbapi_connection, crec.get_connection())
|
|
|
|
def test_alter_invalidate_pool_to_false(self):
|
|
self._test_alter_invalidate_pool_to_false(True)
|
|
|
|
def test_alter_invalidate_pool_stays_true(self):
|
|
self._test_alter_invalidate_pool_to_false(False)
|
|
|
|
def test_handle_error_event_connect_isolation_level(self):
|
|
engine = engines.testing_engine()
|
|
|
|
class MySpecialException(Exception):
|
|
pass
|
|
|
|
@event.listens_for(engine, "handle_error")
|
|
def handle_error(ctx):
|
|
raise MySpecialException("failed operation")
|
|
|
|
ProgrammingError = engine.dialect.dbapi.ProgrammingError
|
|
with engine.connect() as conn:
|
|
with patch.object(
|
|
conn.dialect,
|
|
"get_isolation_level",
|
|
Mock(
|
|
side_effect=provision.dbapi_error(
|
|
config, ProgrammingError, "random error"
|
|
)
|
|
),
|
|
):
|
|
assert_raises(MySpecialException, conn.get_isolation_level)
|
|
|
|
def test_handle_error_not_on_connection(self, connection):
|
|
with expect_raises_message(
|
|
tsa.exc.InvalidRequestError,
|
|
r"The handle_error\(\) event hook as of SQLAlchemy 2.0 is "
|
|
r"established "
|
|
r"on the Dialect, and may only be applied to the Engine as a "
|
|
r"whole or to a specific Dialect as a whole, not on a "
|
|
r"per-Connection basis.",
|
|
):
|
|
|
|
@event.listens_for(connection, "handle_error")
|
|
def handle_error(ctx):
|
|
pass
|
|
|
|
@testing.only_on("sqlite+pysqlite")
|
|
def test_cursor_close_resultset_failed_connectionless(self):
|
|
engine = engines.testing_engine()
|
|
|
|
the_conn = []
|
|
the_cursor = []
|
|
|
|
@event.listens_for(engine, "after_cursor_execute")
|
|
def go(
|
|
connection, cursor, statement, parameters, context, executemany
|
|
):
|
|
the_cursor.append(cursor)
|
|
the_conn.append(connection)
|
|
|
|
with mock.patch(
|
|
"sqlalchemy.engine.cursor.CursorResult.__init__",
|
|
Mock(side_effect=tsa.exc.InvalidRequestError("duplicate col")),
|
|
):
|
|
with engine.connect() as conn:
|
|
assert_raises(
|
|
tsa.exc.InvalidRequestError,
|
|
conn.execute,
|
|
text("select 1"),
|
|
)
|
|
|
|
# cursor is closed
|
|
assert_raises_message(
|
|
engine.dialect.dbapi.ProgrammingError,
|
|
"Cannot operate on a closed cursor",
|
|
the_cursor[0].execute,
|
|
"select 1",
|
|
)
|
|
|
|
# connection is closed
|
|
assert the_conn[0].closed
|
|
|
|
@testing.only_on("sqlite+pysqlite")
|
|
def test_cursor_close_resultset_failed_explicit(self):
|
|
engine = engines.testing_engine()
|
|
|
|
the_cursor = []
|
|
|
|
@event.listens_for(engine, "after_cursor_execute")
|
|
def go(
|
|
connection, cursor, statement, parameters, context, executemany
|
|
):
|
|
the_cursor.append(cursor)
|
|
|
|
conn = engine.connect()
|
|
|
|
with mock.patch(
|
|
"sqlalchemy.engine.cursor.CursorResult.__init__",
|
|
Mock(side_effect=tsa.exc.InvalidRequestError("duplicate col")),
|
|
):
|
|
assert_raises(
|
|
tsa.exc.InvalidRequestError,
|
|
conn.execute,
|
|
text("select 1"),
|
|
)
|
|
|
|
# cursor is closed
|
|
assert_raises_message(
|
|
engine.dialect.dbapi.ProgrammingError,
|
|
"Cannot operate on a closed cursor",
|
|
the_cursor[0].execute,
|
|
"select 1",
|
|
)
|
|
|
|
# connection not closed
|
|
assert not conn.closed
|
|
|
|
conn.close()
|
|
|
|
|
|
class OnConnectTest(fixtures.TestBase):
|
|
__requires__ = ("sqlite",)
|
|
|
|
def setup_test(self):
|
|
e = create_engine("sqlite://")
|
|
|
|
connection = Mock(get_server_version_info=Mock(return_value="5.0"))
|
|
|
|
def connect(*args, **kwargs):
|
|
return connection
|
|
|
|
dbapi = Mock(
|
|
sqlite_version_info=(99, 9, 9),
|
|
version_info=(99, 9, 9),
|
|
sqlite_version="99.9.9",
|
|
paramstyle="named",
|
|
connect=Mock(side_effect=connect),
|
|
)
|
|
|
|
sqlite3 = e.dialect.dbapi
|
|
dbapi.Error = (sqlite3.Error,)
|
|
dbapi.ProgrammingError = sqlite3.ProgrammingError
|
|
|
|
self.dbapi = dbapi
|
|
self.ProgrammingError = sqlite3.ProgrammingError
|
|
|
|
def test_wraps_connect_in_dbapi(self):
|
|
dbapi = self.dbapi
|
|
dbapi.connect = Mock(side_effect=self.ProgrammingError("random error"))
|
|
try:
|
|
create_engine("sqlite://", module=dbapi).connect()
|
|
assert False
|
|
except tsa.exc.DBAPIError as de:
|
|
assert not de.connection_invalidated
|
|
|
|
def test_handle_error_event_connect(self):
|
|
dbapi = self.dbapi
|
|
dbapi.connect = Mock(side_effect=self.ProgrammingError("random error"))
|
|
|
|
class MySpecialException(Exception):
|
|
pass
|
|
|
|
eng = create_engine("sqlite://", module=dbapi)
|
|
|
|
@event.listens_for(eng, "handle_error")
|
|
def handle_error(ctx):
|
|
assert ctx.engine is eng
|
|
assert ctx.connection is None
|
|
raise MySpecialException("failed operation")
|
|
|
|
assert_raises(MySpecialException, eng.connect)
|
|
|
|
def test_handle_error_event_revalidate(self):
|
|
dbapi = self.dbapi
|
|
|
|
class MySpecialException(Exception):
|
|
pass
|
|
|
|
eng = create_engine("sqlite://", module=dbapi, _initialize=False)
|
|
|
|
@event.listens_for(eng, "handle_error")
|
|
def handle_error(ctx):
|
|
assert ctx.engine is eng
|
|
assert ctx.connection is conn
|
|
assert isinstance(
|
|
ctx.sqlalchemy_exception, tsa.exc.ProgrammingError
|
|
)
|
|
raise MySpecialException("failed operation")
|
|
|
|
conn = eng.connect()
|
|
conn.invalidate()
|
|
|
|
dbapi.connect = Mock(side_effect=self.ProgrammingError("random error"))
|
|
|
|
assert_raises(MySpecialException, getattr, conn, "connection")
|
|
|
|
def test_handle_error_event_implicit_revalidate(self):
|
|
dbapi = self.dbapi
|
|
|
|
class MySpecialException(Exception):
|
|
pass
|
|
|
|
eng = create_engine("sqlite://", module=dbapi, _initialize=False)
|
|
|
|
@event.listens_for(eng, "handle_error")
|
|
def handle_error(ctx):
|
|
assert ctx.engine is eng
|
|
assert ctx.connection is conn
|
|
assert isinstance(
|
|
ctx.sqlalchemy_exception, tsa.exc.ProgrammingError
|
|
)
|
|
raise MySpecialException("failed operation")
|
|
|
|
conn = eng.connect()
|
|
conn.invalidate()
|
|
|
|
dbapi.connect = Mock(side_effect=self.ProgrammingError("random error"))
|
|
|
|
assert_raises(MySpecialException, conn.execute, select(1))
|
|
|
|
def test_handle_error_custom_connect(self):
|
|
dbapi = self.dbapi
|
|
|
|
class MySpecialException(Exception):
|
|
pass
|
|
|
|
def custom_connect():
|
|
raise self.ProgrammingError("random error")
|
|
|
|
eng = create_engine("sqlite://", module=dbapi, creator=custom_connect)
|
|
|
|
@event.listens_for(eng, "handle_error")
|
|
def handle_error(ctx):
|
|
assert ctx.engine is eng
|
|
assert ctx.connection is None
|
|
raise MySpecialException("failed operation")
|
|
|
|
assert_raises(MySpecialException, eng.connect)
|
|
|
|
def test_handle_error_event_connect_invalidate_flag(self):
|
|
dbapi = self.dbapi
|
|
dbapi.connect = Mock(
|
|
side_effect=self.ProgrammingError(
|
|
"Cannot operate on a closed database."
|
|
)
|
|
)
|
|
|
|
class MySpecialException(Exception):
|
|
pass
|
|
|
|
eng = create_engine("sqlite://", module=dbapi)
|
|
|
|
@event.listens_for(eng, "handle_error")
|
|
def handle_error(ctx):
|
|
assert ctx.is_disconnect
|
|
ctx.is_disconnect = False
|
|
|
|
try:
|
|
eng.connect()
|
|
assert False
|
|
except tsa.exc.DBAPIError as de:
|
|
assert not de.connection_invalidated
|
|
|
|
def test_cant_connect_stay_invalidated(self):
|
|
class MySpecialException(Exception):
|
|
pass
|
|
|
|
eng = create_engine("sqlite://")
|
|
|
|
@event.listens_for(eng, "handle_error")
|
|
def handle_error(ctx):
|
|
assert ctx.is_disconnect
|
|
|
|
conn = eng.connect()
|
|
|
|
conn.invalidate()
|
|
|
|
eng.pool._creator = Mock(
|
|
side_effect=self.ProgrammingError(
|
|
"Cannot operate on a closed database."
|
|
)
|
|
)
|
|
|
|
try:
|
|
conn.connection
|
|
assert False
|
|
except tsa.exc.DBAPIError:
|
|
assert conn.invalidated
|
|
|
|
def test_dont_touch_non_dbapi_exception_on_connect(self):
|
|
dbapi = self.dbapi
|
|
dbapi.connect = Mock(side_effect=TypeError("I'm not a DBAPI error"))
|
|
|
|
e = create_engine("sqlite://", module=dbapi)
|
|
e.dialect.is_disconnect = is_disconnect = Mock()
|
|
assert_raises_message(TypeError, "I'm not a DBAPI error", e.connect)
|
|
eq_(is_disconnect.call_count, 0)
|
|
|
|
def test_ensure_dialect_does_is_disconnect_no_conn(self):
|
|
"""test that is_disconnect() doesn't choke if no connection,
|
|
cursor given."""
|
|
dialect = testing.db.dialect
|
|
dbapi = dialect.dbapi
|
|
assert not dialect.is_disconnect(
|
|
provision.dbapi_error(config, dbapi.OperationalError, "test"),
|
|
None,
|
|
None,
|
|
)
|
|
|
|
def test_dont_create_transaction_on_initialize(self):
|
|
"""test that engine init doesn't invoke autobegin.
|
|
|
|
this happened implicitly in 1.4 due to use of a non-future
|
|
connection for initialize.
|
|
|
|
to fix for 2.0 we added a new flag _allow_autobegin=False
|
|
for init purposes only.
|
|
|
|
"""
|
|
e = create_engine("sqlite://")
|
|
|
|
init_connection = None
|
|
|
|
def mock_initialize(connection):
|
|
# definitely trigger what would normally be an autobegin
|
|
connection.execute(select(1))
|
|
nonlocal init_connection
|
|
init_connection = connection
|
|
|
|
with (
|
|
mock.patch.object(e._connection_cls, "begin") as mock_begin,
|
|
mock.patch.object(
|
|
e.dialect, "initialize", Mock(side_effect=mock_initialize)
|
|
) as mock_init,
|
|
):
|
|
conn = e.connect()
|
|
|
|
eq_(mock_begin.mock_calls, [])
|
|
is_not(init_connection, None)
|
|
is_not(conn, init_connection)
|
|
is_false(init_connection._allow_autobegin)
|
|
eq_(mock_init.mock_calls, [mock.call(init_connection)])
|
|
|
|
# assert the mock works too
|
|
conn.begin()
|
|
eq_(mock_begin.mock_calls, [mock.call()])
|
|
|
|
conn.close()
|
|
|
|
def test_invalidate_on_connect(self):
|
|
"""test that is_disconnect() is called during connect.
|
|
|
|
interpretation of connection failures are not supported by
|
|
every backend.
|
|
|
|
"""
|
|
dbapi = self.dbapi
|
|
dbapi.connect = Mock(
|
|
side_effect=self.ProgrammingError(
|
|
"Cannot operate on a closed database."
|
|
)
|
|
)
|
|
e = create_engine("sqlite://", module=dbapi)
|
|
try:
|
|
e.connect()
|
|
assert False
|
|
except tsa.exc.DBAPIError as de:
|
|
assert de.connection_invalidated
|
|
|
|
@testing.only_on("sqlite+pysqlite")
|
|
def test_initialize_connect_calls(self):
|
|
"""test for :ticket:`5497`, on_connect not called twice"""
|
|
|
|
m1 = Mock()
|
|
cls_ = testing.db.dialect.__class__
|
|
|
|
class SomeDialect(cls_):
|
|
def initialize(self, connection):
|
|
super().initialize(connection)
|
|
m1.initialize(connection)
|
|
|
|
def on_connect(self):
|
|
oc = super().on_connect()
|
|
|
|
def my_on_connect(conn):
|
|
if oc:
|
|
oc(conn)
|
|
m1.on_connect(conn)
|
|
|
|
return my_on_connect
|
|
|
|
u1 = Mock(
|
|
username=None,
|
|
password=None,
|
|
host=None,
|
|
port=None,
|
|
query={},
|
|
database=None,
|
|
_instantiate_plugins=lambda kw: (u1, [], kw),
|
|
_get_entrypoint=Mock(
|
|
return_value=Mock(get_dialect_cls=lambda u: SomeDialect)
|
|
),
|
|
)
|
|
eng = create_engine(u1, poolclass=QueuePool)
|
|
# make sure other dialects aren't getting pulled in here
|
|
eq_(eng.name, "sqlite")
|
|
c = eng.connect()
|
|
dbapi_conn_one = c.connection.dbapi_connection
|
|
c.close()
|
|
|
|
eq_(
|
|
m1.mock_calls,
|
|
[call.on_connect(dbapi_conn_one), call.initialize(mock.ANY)],
|
|
)
|
|
|
|
c = eng.connect()
|
|
|
|
eq_(
|
|
m1.mock_calls,
|
|
[call.on_connect(dbapi_conn_one), call.initialize(mock.ANY)],
|
|
)
|
|
|
|
c2 = eng.connect()
|
|
dbapi_conn_two = c2.connection.dbapi_connection
|
|
|
|
is_not(dbapi_conn_one, dbapi_conn_two)
|
|
|
|
eq_(
|
|
m1.mock_calls,
|
|
[
|
|
call.on_connect(dbapi_conn_one),
|
|
call.initialize(mock.ANY),
|
|
call.on_connect(dbapi_conn_two),
|
|
],
|
|
)
|
|
|
|
c.close()
|
|
c2.close()
|
|
|
|
@testing.only_on("sqlite+pysqlite")
|
|
def test_initialize_connect_race(self):
|
|
"""test for :ticket:`6337` fixing the regression in :ticket:`5497`,
|
|
dialect init is mutexed"""
|
|
|
|
m1 = []
|
|
cls_ = testing.db.dialect.__class__
|
|
|
|
class SomeDialect(cls_):
|
|
supports_statement_cache = True
|
|
|
|
def initialize(self, connection):
|
|
super().initialize(connection)
|
|
m1.append("initialize")
|
|
|
|
def on_connect(self):
|
|
oc = super().on_connect()
|
|
|
|
def my_on_connect(conn):
|
|
if oc:
|
|
oc(conn)
|
|
m1.append("on_connect")
|
|
|
|
return my_on_connect
|
|
|
|
u1 = Mock(
|
|
username=None,
|
|
password=None,
|
|
host=None,
|
|
port=None,
|
|
query={},
|
|
database=None,
|
|
_instantiate_plugins=lambda kw: (u1, [], kw),
|
|
_get_entrypoint=Mock(
|
|
return_value=Mock(get_dialect_cls=lambda u: SomeDialect)
|
|
),
|
|
)
|
|
|
|
for j in range(5):
|
|
m1[:] = []
|
|
eng = create_engine(
|
|
u1,
|
|
poolclass=NullPool,
|
|
connect_args={"check_same_thread": False},
|
|
)
|
|
|
|
def go():
|
|
c = eng.connect()
|
|
c.execute(text("select 1"))
|
|
c.close()
|
|
|
|
threads = [threading.Thread(target=go) for i in range(10)]
|
|
for t in threads:
|
|
t.start()
|
|
for t in threads:
|
|
t.join()
|
|
|
|
eq_(m1, ["on_connect", "initialize"] + ["on_connect"] * 9)
|
|
|
|
|
|
class DialectEventTest(fixtures.TestBase):
|
|
@contextmanager
|
|
def _run_test(self, retval):
|
|
m1 = Mock()
|
|
|
|
m1.do_execute.return_value = retval
|
|
m1.do_executemany.return_value = retval
|
|
m1.do_execute_no_params.return_value = retval
|
|
e = engines.testing_engine(options={"_initialize": False})
|
|
|
|
event.listen(e, "do_execute", m1.do_execute)
|
|
event.listen(e, "do_executemany", m1.do_executemany)
|
|
event.listen(e, "do_execute_no_params", m1.do_execute_no_params)
|
|
|
|
e.dialect.do_execute = m1.real_do_execute
|
|
e.dialect.do_executemany = m1.real_do_executemany
|
|
e.dialect.do_execute_no_params = m1.real_do_execute_no_params
|
|
|
|
def mock_the_cursor(cursor, *arg):
|
|
arg[-1].get_result_proxy = Mock(return_value=Mock(context=arg[-1]))
|
|
return retval
|
|
|
|
m1.real_do_execute.side_effect = m1.do_execute.side_effect = (
|
|
mock_the_cursor
|
|
)
|
|
m1.real_do_executemany.side_effect = m1.do_executemany.side_effect = (
|
|
mock_the_cursor
|
|
)
|
|
m1.real_do_execute_no_params.side_effect = (
|
|
m1.do_execute_no_params.side_effect
|
|
) = mock_the_cursor
|
|
|
|
with e.begin() as conn:
|
|
yield conn, m1
|
|
|
|
def _assert(self, retval, m1, m2, mock_calls):
|
|
eq_(m1.mock_calls, mock_calls)
|
|
if retval:
|
|
eq_(m2.mock_calls, [])
|
|
else:
|
|
eq_(m2.mock_calls, mock_calls)
|
|
|
|
def _test_do_execute(self, retval):
|
|
with self._run_test(retval) as (conn, m1):
|
|
result = conn.exec_driver_sql(
|
|
"insert into table foo", {"foo": "bar"}
|
|
)
|
|
self._assert(
|
|
retval,
|
|
m1.do_execute,
|
|
m1.real_do_execute,
|
|
[
|
|
call(
|
|
result.context.cursor,
|
|
"insert into table foo",
|
|
{"foo": "bar"},
|
|
result.context,
|
|
)
|
|
],
|
|
)
|
|
|
|
def _test_do_executemany(self, retval):
|
|
with self._run_test(retval) as (conn, m1):
|
|
result = conn.exec_driver_sql(
|
|
"insert into table foo", [{"foo": "bar"}, {"foo": "bar"}]
|
|
)
|
|
self._assert(
|
|
retval,
|
|
m1.do_executemany,
|
|
m1.real_do_executemany,
|
|
[
|
|
call(
|
|
result.context.cursor,
|
|
"insert into table foo",
|
|
[{"foo": "bar"}, {"foo": "bar"}],
|
|
result.context,
|
|
)
|
|
],
|
|
)
|
|
|
|
def _test_do_execute_no_params(self, retval):
|
|
with self._run_test(retval) as (conn, m1):
|
|
result = conn.execution_options(
|
|
no_parameters=True
|
|
).exec_driver_sql("insert into table foo")
|
|
self._assert(
|
|
retval,
|
|
m1.do_execute_no_params,
|
|
m1.real_do_execute_no_params,
|
|
[
|
|
call(
|
|
result.context.cursor,
|
|
"insert into table foo",
|
|
result.context,
|
|
)
|
|
],
|
|
)
|
|
|
|
def _test_cursor_execute(self, retval):
|
|
with self._run_test(retval) as (conn, m1):
|
|
dialect = conn.dialect
|
|
|
|
stmt = "insert into table foo"
|
|
params = {"foo": "bar"}
|
|
ctx = dialect.execution_ctx_cls._init_statement(
|
|
dialect,
|
|
conn,
|
|
conn.connection,
|
|
{},
|
|
stmt,
|
|
[params],
|
|
)
|
|
|
|
conn._cursor_execute(ctx.cursor, stmt, params, ctx)
|
|
|
|
self._assert(
|
|
retval,
|
|
m1.do_execute,
|
|
m1.real_do_execute,
|
|
[call(ctx.cursor, "insert into table foo", {"foo": "bar"}, ctx)],
|
|
)
|
|
|
|
def test_do_execute_w_replace(self):
|
|
self._test_do_execute(True)
|
|
|
|
def test_do_execute_wo_replace(self):
|
|
self._test_do_execute(False)
|
|
|
|
def test_do_executemany_w_replace(self):
|
|
self._test_do_executemany(True)
|
|
|
|
def test_do_executemany_wo_replace(self):
|
|
self._test_do_executemany(False)
|
|
|
|
def test_do_execute_no_params_w_replace(self):
|
|
self._test_do_execute_no_params(True)
|
|
|
|
def test_do_execute_no_params_wo_replace(self):
|
|
self._test_do_execute_no_params(False)
|
|
|
|
def test_cursor_execute_w_replace(self):
|
|
self._test_cursor_execute(True)
|
|
|
|
def test_cursor_execute_wo_replace(self):
|
|
self._test_cursor_execute(False)
|
|
|
|
def test_connect_replace_params(self):
|
|
e = engines.testing_engine(options={"_initialize": False})
|
|
|
|
@event.listens_for(e, "do_connect")
|
|
def evt(dialect, conn_rec, cargs, cparams):
|
|
cargs[:] = ["foo", "hoho"]
|
|
cparams.clear()
|
|
cparams["bar"] = "bat"
|
|
conn_rec.info["boom"] = "bap"
|
|
|
|
m1 = Mock()
|
|
e.dialect.connect = m1.real_connect
|
|
|
|
with e.connect() as conn:
|
|
eq_(m1.mock_calls, [call.real_connect("foo", "hoho", bar="bat")])
|
|
eq_(conn.info["boom"], "bap")
|
|
|
|
def test_connect_do_connect(self):
|
|
e = engines.testing_engine(options={"_initialize": False})
|
|
|
|
m1 = Mock()
|
|
|
|
@event.listens_for(e, "do_connect")
|
|
def evt1(dialect, conn_rec, cargs, cparams):
|
|
cargs[:] = ["foo", "hoho"]
|
|
cparams.clear()
|
|
cparams["bar"] = "bat"
|
|
conn_rec.info["boom"] = "one"
|
|
|
|
@event.listens_for(e, "do_connect")
|
|
def evt2(dialect, conn_rec, cargs, cparams):
|
|
conn_rec.info["bap"] = "two"
|
|
return m1.our_connect(cargs, cparams)
|
|
|
|
with e.connect() as conn:
|
|
# called with args
|
|
eq_(
|
|
m1.mock_calls,
|
|
[call.our_connect(["foo", "hoho"], {"bar": "bat"})],
|
|
)
|
|
|
|
eq_(conn.info["boom"], "one")
|
|
eq_(conn.info["bap"], "two")
|
|
|
|
# returned our mock connection
|
|
is_(conn.connection.dbapi_connection, m1.our_connect())
|
|
|
|
def test_connect_do_connect_info_there_after_recycle(self):
|
|
# test that info is maintained after the do_connect()
|
|
# event for a soft invalidation.
|
|
|
|
e = engines.testing_engine(options={"_initialize": False})
|
|
|
|
@event.listens_for(e, "do_connect")
|
|
def evt1(dialect, conn_rec, cargs, cparams):
|
|
conn_rec.info["boom"] = "one"
|
|
|
|
conn = e.connect()
|
|
eq_(conn.info["boom"], "one")
|
|
|
|
conn.connection.invalidate(soft=True)
|
|
conn.close()
|
|
with e.connect() as conn:
|
|
eq_(conn.info["boom"], "one")
|
|
|
|
def test_connect_do_connect_info_there_after_invalidate(self):
|
|
# test that info is maintained after the do_connect()
|
|
# event for a hard invalidation.
|
|
|
|
e = engines.testing_engine(options={"_initialize": False})
|
|
|
|
@event.listens_for(e, "do_connect")
|
|
def evt1(dialect, conn_rec, cargs, cparams):
|
|
assert not conn_rec.info
|
|
conn_rec.info["boom"] = "one"
|
|
|
|
conn = e.connect()
|
|
eq_(conn.info["boom"], "one")
|
|
|
|
conn.connection.invalidate()
|
|
|
|
with e.connect() as conn:
|
|
eq_(conn.info["boom"], "one")
|
|
|
|
def test_connect_do_connect_no_cargs_cparams_leak_nullpool(self):
|
|
"""test #13144, cargs/cparams in do_connect are not shared"""
|
|
|
|
e = engines.testing_engine(
|
|
options={"_initialize": False, "poolclass": NullPool}
|
|
)
|
|
|
|
m1 = Mock()
|
|
e.dialect.connect = m1.mock_connect
|
|
|
|
cargs_list = []
|
|
cparams_list = []
|
|
call_count = 0
|
|
|
|
@event.listens_for(e, "do_connect")
|
|
def evt(dialect, conn_rec, cargs, cparams):
|
|
nonlocal call_count
|
|
cargs_list.append(cargs)
|
|
cparams_list.append(cparams)
|
|
|
|
cargs.append(f"extra_arg_{call_count}")
|
|
cparams[f"extra_param_{call_count}"] = "value"
|
|
call_count += 1
|
|
|
|
for _ in range(3):
|
|
with e.connect():
|
|
pass
|
|
|
|
all_list_keys = ["extra_arg_0", "extra_arg_1", "extra_arg_2"]
|
|
|
|
for carg, expected in zip(cargs_list, all_list_keys):
|
|
eq_(set(carg).intersection(all_list_keys), {expected})
|
|
|
|
all_param_keys = ["extra_param_0", "extra_param_1", "extra_param_2"]
|
|
for cparam, expected in zip(cparams_list, all_param_keys):
|
|
eq_(set(cparam).intersection(all_param_keys), {expected})
|
|
|
|
eq_(
|
|
len(set(id(carg) for carg in cargs_list)),
|
|
3,
|
|
"cargs should be different objects",
|
|
)
|
|
eq_(
|
|
len(set(id(cparam) for cparam in cparams_list)),
|
|
3,
|
|
"cparams should be different objects",
|
|
)
|
|
|
|
|
|
class SetInputSizesTest(fixtures.TablesTest):
|
|
__backend__ = True
|
|
|
|
__requires__ = ("independent_connections", "insert_returning")
|
|
|
|
@classmethod
|
|
def define_tables(cls, metadata):
|
|
Table(
|
|
"users",
|
|
metadata,
|
|
Column("user_id", INT, primary_key=True, autoincrement=False),
|
|
Column("user_name", VARCHAR(20)),
|
|
)
|
|
|
|
@testing.fixture
|
|
def input_sizes_fixture(self, testing_engine):
|
|
canary = mock.Mock()
|
|
|
|
def do_set_input_sizes(cursor, list_of_tuples, context):
|
|
canary.do_set_input_sizes(cursor, list_of_tuples, context)
|
|
|
|
def pre_exec(self):
|
|
self.translate_set_input_sizes = None
|
|
self.include_set_input_sizes = None
|
|
self.exclude_set_input_sizes = None
|
|
|
|
engine = testing_engine()
|
|
engine.connect().close()
|
|
|
|
# the idea of this test is we fully replace the dialect
|
|
# do_set_input_sizes with a mock, and we can then intercept
|
|
# the setting passed to the dialect. the test table uses very
|
|
# "safe" datatypes so that the DBAPI does not actually need
|
|
# setinputsizes() called in order to work.
|
|
|
|
with (
|
|
mock.patch.object(
|
|
engine.dialect, "bind_typing", BindTyping.SETINPUTSIZES
|
|
),
|
|
mock.patch.object(
|
|
engine.dialect, "do_set_input_sizes", do_set_input_sizes
|
|
),
|
|
mock.patch.object(
|
|
engine.dialect.execution_ctx_cls, "pre_exec", pre_exec
|
|
),
|
|
):
|
|
yield engine, canary
|
|
|
|
@testing.requires.insertmanyvalues
|
|
def test_set_input_sizes_insertmanyvalues_no_event(
|
|
self, input_sizes_fixture
|
|
):
|
|
engine, canary = input_sizes_fixture
|
|
|
|
with engine.begin() as conn:
|
|
conn.execute(
|
|
self.tables.users.insert().returning(
|
|
self.tables.users.c.user_id
|
|
),
|
|
[
|
|
{"user_id": 1, "user_name": "n1"},
|
|
{"user_id": 2, "user_name": "n2"},
|
|
{"user_id": 3, "user_name": "n3"},
|
|
],
|
|
)
|
|
|
|
eq_(
|
|
canary.mock_calls,
|
|
[
|
|
call.do_set_input_sizes(
|
|
mock.ANY,
|
|
[
|
|
(
|
|
"user_id_0",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(Integer),
|
|
),
|
|
(
|
|
"user_name_0",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(String),
|
|
),
|
|
(
|
|
"user_id_1",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(Integer),
|
|
),
|
|
(
|
|
"user_name_1",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(String),
|
|
),
|
|
(
|
|
"user_id_2",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(Integer),
|
|
),
|
|
(
|
|
"user_name_2",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(String),
|
|
),
|
|
],
|
|
mock.ANY,
|
|
)
|
|
],
|
|
)
|
|
|
|
def test_set_input_sizes_no_event(self, input_sizes_fixture):
|
|
engine, canary = input_sizes_fixture
|
|
|
|
with engine.begin() as conn:
|
|
conn.execute(
|
|
self.tables.users.update()
|
|
.where(self.tables.users.c.user_id == 15)
|
|
.values(user_id=15, user_name="n1"),
|
|
)
|
|
|
|
eq_(
|
|
canary.mock_calls,
|
|
[
|
|
call.do_set_input_sizes(
|
|
mock.ANY,
|
|
[
|
|
(
|
|
"user_id",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(Integer),
|
|
),
|
|
(
|
|
"user_name",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(String),
|
|
),
|
|
(
|
|
"user_id_1",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(Integer),
|
|
),
|
|
],
|
|
mock.ANY,
|
|
)
|
|
],
|
|
)
|
|
|
|
def test_set_input_sizes_expanding_param(self, input_sizes_fixture):
|
|
engine, canary = input_sizes_fixture
|
|
|
|
with engine.connect() as conn:
|
|
conn.execute(
|
|
select(self.tables.users).where(
|
|
self.tables.users.c.user_name.in_(["x", "y", "z"])
|
|
)
|
|
)
|
|
|
|
eq_(
|
|
canary.mock_calls,
|
|
[
|
|
call.do_set_input_sizes(
|
|
mock.ANY,
|
|
[
|
|
(
|
|
"user_name_1_1",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(String),
|
|
),
|
|
(
|
|
"user_name_1_2",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(String),
|
|
),
|
|
(
|
|
"user_name_1_3",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(String),
|
|
),
|
|
],
|
|
mock.ANY,
|
|
)
|
|
],
|
|
)
|
|
|
|
@testing.requires.tuple_in
|
|
def test_set_input_sizes_expanding_tuple_param(self, input_sizes_fixture):
|
|
engine, canary = input_sizes_fixture
|
|
|
|
from sqlalchemy import tuple_
|
|
|
|
with engine.connect() as conn:
|
|
conn.execute(
|
|
select(self.tables.users).where(
|
|
tuple_(
|
|
self.tables.users.c.user_id,
|
|
self.tables.users.c.user_name,
|
|
).in_([(1, "x"), (2, "y")])
|
|
)
|
|
)
|
|
|
|
eq_(
|
|
canary.mock_calls,
|
|
[
|
|
call.do_set_input_sizes(
|
|
mock.ANY,
|
|
[
|
|
(
|
|
"param_1_1_1",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(Integer),
|
|
),
|
|
(
|
|
"param_1_1_2",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(String),
|
|
),
|
|
(
|
|
"param_1_2_1",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(Integer),
|
|
),
|
|
(
|
|
"param_1_2_2",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(String),
|
|
),
|
|
],
|
|
mock.ANY,
|
|
)
|
|
],
|
|
)
|
|
|
|
def test_set_input_sizes_event(self, input_sizes_fixture):
|
|
engine, canary = input_sizes_fixture
|
|
|
|
SPECIAL_STRING = mock.Mock()
|
|
|
|
@event.listens_for(engine, "do_setinputsizes")
|
|
def do_setinputsizes(
|
|
inputsizes, cursor, statement, parameters, context
|
|
):
|
|
for k in inputsizes:
|
|
if k.type._type_affinity is String:
|
|
inputsizes[k] = (
|
|
SPECIAL_STRING,
|
|
None,
|
|
0,
|
|
)
|
|
|
|
with engine.begin() as conn:
|
|
conn.execute(
|
|
self.tables.users.update()
|
|
.where(self.tables.users.c.user_id == 15)
|
|
.values(user_id=15, user_name="n1"),
|
|
)
|
|
|
|
eq_(
|
|
canary.mock_calls,
|
|
[
|
|
call.do_set_input_sizes(
|
|
mock.ANY,
|
|
[
|
|
(
|
|
"user_id",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(Integer),
|
|
),
|
|
(
|
|
"user_name",
|
|
(SPECIAL_STRING, None, 0),
|
|
testing.eq_type_affinity(String),
|
|
),
|
|
(
|
|
"user_id_1",
|
|
mock.ANY,
|
|
testing.eq_type_affinity(Integer),
|
|
),
|
|
],
|
|
mock.ANY,
|
|
)
|
|
],
|
|
)
|
|
|
|
|
|
class DialectDoesntSupportCachingTest(fixtures.TestBase):
|
|
"""test the opt-in caching flag added in :ticket:`6184`."""
|
|
|
|
__only_on__ = "sqlite+pysqlite"
|
|
|
|
__requires__ = ("sqlite_memory",)
|
|
|
|
@testing.fixture()
|
|
def sqlite_no_cache_dialect(self, testing_engine):
|
|
from sqlalchemy.dialects.sqlite.pysqlite import SQLiteDialect_pysqlite
|
|
from sqlalchemy.dialects.sqlite.base import SQLiteCompiler
|
|
from sqlalchemy.sql import visitors
|
|
|
|
class MyCompiler(SQLiteCompiler):
|
|
def translate_select_structure(self, select_stmt, **kwargs):
|
|
select = select_stmt
|
|
|
|
if not getattr(select, "_mydialect_visit", None):
|
|
select = visitors.cloned_traverse(select_stmt, {}, {})
|
|
if select._limit_clause is not None:
|
|
# create a bindparam with a fixed name and hardcode
|
|
# it to the given limit. this breaks caching.
|
|
select._limit_clause = bindparam(
|
|
"limit", value=select._limit, literal_execute=True
|
|
)
|
|
|
|
select._mydialect_visit = True
|
|
|
|
return select
|
|
|
|
class MyDialect(SQLiteDialect_pysqlite):
|
|
statement_compiler = MyCompiler
|
|
supports_statement_cache = False
|
|
|
|
from sqlalchemy.dialects import registry
|
|
|
|
def go(name):
|
|
return MyDialect
|
|
|
|
with mock.patch.object(registry, "load", go):
|
|
eng = testing_engine()
|
|
yield eng
|
|
|
|
@testing.fixture
|
|
def data_fixture(self, sqlite_no_cache_dialect):
|
|
m = MetaData()
|
|
t = Table("t1", m, Column("x", Integer))
|
|
with sqlite_no_cache_dialect.begin() as conn:
|
|
t.create(conn)
|
|
conn.execute(t.insert(), [{"x": 1}, {"x": 2}, {"x": 3}, {"x": 4}])
|
|
|
|
return t
|
|
|
|
def test_no_cache(self, sqlite_no_cache_dialect, data_fixture):
|
|
eng = sqlite_no_cache_dialect
|
|
|
|
def go(lim):
|
|
with eng.connect() as conn:
|
|
result = conn.execute(
|
|
select(data_fixture).order_by(data_fixture.c.x).limit(lim)
|
|
)
|
|
return result
|
|
|
|
r1 = go(2)
|
|
r2 = go(3)
|
|
|
|
eq_(r1.all(), [(1,), (2,)])
|
|
eq_(r2.all(), [(1,), (2,), (3,)])
|
|
|
|
def test_it_caches(self, sqlite_no_cache_dialect, data_fixture):
|
|
eng = sqlite_no_cache_dialect
|
|
eng.dialect.__class__.supports_statement_cache = True
|
|
del eng.dialect.__dict__["_supports_statement_cache"]
|
|
|
|
def go(lim):
|
|
with eng.connect() as conn:
|
|
result = conn.execute(
|
|
select(data_fixture).order_by(data_fixture.c.x).limit(lim)
|
|
)
|
|
return result
|
|
|
|
r1 = go(2)
|
|
r2 = go(3)
|
|
|
|
eq_(r1.all(), [(1,), (2,)])
|
|
|
|
# wrong answer
|
|
eq_(
|
|
r2.all(),
|
|
[
|
|
(1,),
|
|
(2,),
|
|
],
|
|
)
|