mirror of
https://github.com/sqlalchemy/sqlalchemy.git
synced 2026-05-13 04:07:20 -04:00
a8029f5a7e
* ORM Insert now includes "bulk" mode that will run essentially the same process as session.bulk_insert_mappings; interprets the given list of values as ORM attributes for key names * ORM UPDATE has a similar feature, without RETURNING support, for session.bulk_update_mappings * Added support for upserts to do RETURNING ORM objects as well * ORM UPDATE/DELETE with list of parameters + WHERE criteria is a not implemented; use connection * ORM UPDATE/DELETE defaults to "auto" synchronize_session; use fetch if RETURNING is present, evaluate if not, as "fetch" is much more efficient (no expired object SELECT problem) and less error prone if RETURNING is available UPDATE: howver this is inefficient! please continue to use evaluate for simple cases, auto can move to fetch if criteria not evaluable * "Evaluate" criteria will now not preemptively unexpire and SELECT attributes that were individually expired. Instead, if evaluation of the criteria indicates that the necessary attrs were expired, we expire the object completely (delete) or expire the SET attrs unconditionally (update). This keeps the object in the same unloaded state where it will refresh those attrs on the next pass, for this generally unusual case. (originally #5664) * Core change! update/delete rowcount comes from len(rows) if RETURNING was used. SQLite at least otherwise did not support this. adjusted test_rowcount accordingly * ORM DELETE with a list of parameters at all is also a not implemented as this would imply "bulk", and there is no bulk_delete_mappings (could be, but we dont have that) * ORM insert().values() with single or multi-values translates key names based on ORM attribute names * ORM returning() implemented for insert, update, delete; explcit returning clauses now interpret rows in an ORM context, with support for qualifying loader options as well * session.bulk_insert_mappings() assigns polymorphic identity if not set. * explicit RETURNING + synchronize_session='fetch' is now supported with UPDATE and DELETE. * expanded return_defaults() to work with DELETE also. * added support for composite attributes to be present in the dictionaries used by bulk_insert_mappings and bulk_update_mappings, which is also the new ORM bulk insert/update feature, that will expand the composite values into their individual mapped attributes the way they'd be on a mapped instance. * bulk UPDATE supports "synchronize_session=evaluate", is the default. this does not apply to session.bulk_update_mappings, just the new version * both bulk UPDATE and bulk INSERT, the latter with or without RETURNING, support *heterogenous* parameter sets. session.bulk_insert/update_mappings did this, so this feature is maintained. now cursor result can be both horizontally and vertically spliced :) This is now a long story with a lot of options, which in itself is a problem to be able to document all of this in some way that makes sense. raising exceptions for use cases we haven't supported is pretty important here too, the tradition of letting unsupported things just not work is likely not a good idea at this point, though there are still many cases that aren't easily avoidable Fixes: #8360 Fixes: #7864 Fixes: #7865 Change-Id: Idf28379f8705e403a3c6a937f6a798a042ef2540
3528 lines
112 KiB
Python
3528 lines
112 KiB
Python
import collections
|
|
import collections.abc as collections_abc
|
|
from contextlib import contextmanager
|
|
import csv
|
|
from io import StringIO
|
|
import operator
|
|
import pickle
|
|
from unittest.mock import Mock
|
|
from unittest.mock import patch
|
|
|
|
from sqlalchemy import CHAR
|
|
from sqlalchemy import column
|
|
from sqlalchemy import exc
|
|
from sqlalchemy import exc as sa_exc
|
|
from sqlalchemy import ForeignKey
|
|
from sqlalchemy import func
|
|
from sqlalchemy import INT
|
|
from sqlalchemy import Integer
|
|
from sqlalchemy import literal
|
|
from sqlalchemy import literal_column
|
|
from sqlalchemy import MetaData
|
|
from sqlalchemy import select
|
|
from sqlalchemy import sql
|
|
from sqlalchemy import String
|
|
from sqlalchemy import table
|
|
from sqlalchemy import testing
|
|
from sqlalchemy import text
|
|
from sqlalchemy import true
|
|
from sqlalchemy import tuple_
|
|
from sqlalchemy import type_coerce
|
|
from sqlalchemy import TypeDecorator
|
|
from sqlalchemy import VARCHAR
|
|
from sqlalchemy.engine import cursor as _cursor
|
|
from sqlalchemy.engine import default
|
|
from sqlalchemy.engine import Row
|
|
from sqlalchemy.engine.result import SimpleResultMetaData
|
|
from sqlalchemy.engine.row import KEY_INTEGER_ONLY
|
|
from sqlalchemy.ext.compiler import compiles
|
|
from sqlalchemy.sql import ColumnElement
|
|
from sqlalchemy.sql import expression
|
|
from sqlalchemy.sql import LABEL_STYLE_TABLENAME_PLUS_COL
|
|
from sqlalchemy.sql.selectable import LABEL_STYLE_NONE
|
|
from sqlalchemy.sql.selectable import TextualSelect
|
|
from sqlalchemy.sql.sqltypes import NULLTYPE
|
|
from sqlalchemy.sql.util import ClauseAdapter
|
|
from sqlalchemy.testing import assert_raises
|
|
from sqlalchemy.testing import assert_raises_message
|
|
from sqlalchemy.testing import assertions
|
|
from sqlalchemy.testing import engines
|
|
from sqlalchemy.testing import eq_
|
|
from sqlalchemy.testing import expect_raises
|
|
from sqlalchemy.testing import expect_raises_message
|
|
from sqlalchemy.testing import fixtures
|
|
from sqlalchemy.testing import in_
|
|
from sqlalchemy.testing import is_
|
|
from sqlalchemy.testing import is_true
|
|
from sqlalchemy.testing import le_
|
|
from sqlalchemy.testing import mock
|
|
from sqlalchemy.testing import ne_
|
|
from sqlalchemy.testing import not_in
|
|
from sqlalchemy.testing.schema import Column
|
|
from sqlalchemy.testing.schema import Table
|
|
|
|
|
|
class CursorResultTest(fixtures.TablesTest):
|
|
__backend__ = True
|
|
|
|
@classmethod
|
|
def define_tables(cls, metadata):
|
|
Table(
|
|
"users",
|
|
metadata,
|
|
Column(
|
|
"user_id", INT, primary_key=True, test_needs_autoincrement=True
|
|
),
|
|
Column("user_name", VARCHAR(20)),
|
|
test_needs_acid=True,
|
|
)
|
|
Table(
|
|
"addresses",
|
|
metadata,
|
|
Column(
|
|
"address_id",
|
|
Integer,
|
|
primary_key=True,
|
|
test_needs_autoincrement=True,
|
|
),
|
|
Column("user_id", Integer, ForeignKey("users.user_id")),
|
|
Column("address", String(30)),
|
|
test_needs_acid=True,
|
|
)
|
|
|
|
Table(
|
|
"users2",
|
|
metadata,
|
|
Column("user_id", INT, primary_key=True),
|
|
Column("user_name", VARCHAR(20)),
|
|
test_needs_acid=True,
|
|
)
|
|
Table(
|
|
"test",
|
|
metadata,
|
|
Column(
|
|
"x", Integer, primary_key=True, test_needs_autoincrement=False
|
|
),
|
|
Column("y", String(50)),
|
|
)
|
|
|
|
@testing.requires.insert_returning
|
|
def test_splice_horizontally(self, connection):
|
|
users = self.tables.users
|
|
addresses = self.tables.addresses
|
|
|
|
r1 = connection.execute(
|
|
users.insert().returning(users.c.user_name, users.c.user_id),
|
|
[
|
|
dict(user_id=1, user_name="john"),
|
|
dict(user_id=2, user_name="jack"),
|
|
],
|
|
)
|
|
|
|
r2 = connection.execute(
|
|
addresses.insert().returning(
|
|
addresses.c.address_id,
|
|
addresses.c.address,
|
|
addresses.c.user_id,
|
|
),
|
|
[
|
|
dict(address_id=1, user_id=1, address="foo@bar.com"),
|
|
dict(address_id=2, user_id=2, address="bar@bat.com"),
|
|
],
|
|
)
|
|
|
|
rows = r1.splice_horizontally(r2).all()
|
|
eq_(
|
|
rows,
|
|
[
|
|
("john", 1, 1, "foo@bar.com", 1),
|
|
("jack", 2, 2, "bar@bat.com", 2),
|
|
],
|
|
)
|
|
|
|
eq_(rows[0]._mapping[users.c.user_id], 1)
|
|
eq_(rows[0]._mapping[addresses.c.user_id], 1)
|
|
eq_(rows[1].address, "bar@bat.com")
|
|
|
|
with expect_raises_message(
|
|
exc.InvalidRequestError, "Ambiguous column name 'user_id'"
|
|
):
|
|
rows[0].user_id
|
|
|
|
def test_keys_no_rows(self, connection):
|
|
|
|
for i in range(2):
|
|
r = connection.execute(
|
|
text("update users set user_name='new' where user_id=10")
|
|
)
|
|
|
|
with expect_raises_message(
|
|
exc.ResourceClosedError,
|
|
"This result object does not return rows",
|
|
):
|
|
r.keys()
|
|
|
|
def test_row_keys_removed(self, connection):
|
|
r = connection.execute(
|
|
text("select * from users where user_id=2")
|
|
).first()
|
|
|
|
with expect_raises(AttributeError):
|
|
r.keys()
|
|
|
|
def test_row_contains_key_no_strings(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
dict(user_id=1, user_name="john"),
|
|
dict(user_id=2, user_name="jack"),
|
|
],
|
|
)
|
|
|
|
r = connection.execute(
|
|
text("select * from users where user_id=2")
|
|
).first()
|
|
|
|
not_in("user_name", r)
|
|
in_("user_name", r._mapping)
|
|
not_in("foobar", r)
|
|
not_in("foobar", r._mapping)
|
|
|
|
def test_row_iteration(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
{"user_id": 7, "user_name": "jack"},
|
|
{"user_id": 8, "user_name": "ed"},
|
|
{"user_id": 9, "user_name": "fred"},
|
|
],
|
|
)
|
|
r = connection.execute(users.select())
|
|
rows = []
|
|
for row in r:
|
|
rows.append(row)
|
|
eq_(len(rows), 3)
|
|
|
|
def test_scalars(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
{"user_id": 7, "user_name": "jack"},
|
|
{"user_id": 8, "user_name": "ed"},
|
|
{"user_id": 9, "user_name": "fred"},
|
|
],
|
|
)
|
|
r = connection.scalars(users.select().order_by(users.c.user_id))
|
|
eq_(r.all(), [7, 8, 9])
|
|
|
|
def test_result_tuples(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
{"user_id": 7, "user_name": "jack"},
|
|
{"user_id": 8, "user_name": "ed"},
|
|
{"user_id": 9, "user_name": "fred"},
|
|
],
|
|
)
|
|
r = connection.execute(
|
|
users.select().order_by(users.c.user_id)
|
|
).tuples()
|
|
eq_(r.all(), [(7, "jack"), (8, "ed"), (9, "fred")])
|
|
|
|
def test_row_tuple(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
{"user_id": 7, "user_name": "jack"},
|
|
{"user_id": 8, "user_name": "ed"},
|
|
{"user_id": 9, "user_name": "fred"},
|
|
],
|
|
)
|
|
r = connection.execute(users.select().order_by(users.c.user_id))
|
|
eq_([row.t for row in r], [(7, "jack"), (8, "ed"), (9, "fred")])
|
|
|
|
def test_row_next(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
{"user_id": 7, "user_name": "jack"},
|
|
{"user_id": 8, "user_name": "ed"},
|
|
{"user_id": 9, "user_name": "fred"},
|
|
],
|
|
)
|
|
r = connection.execute(users.select())
|
|
rows = []
|
|
while True:
|
|
row = next(r, "foo")
|
|
if row == "foo":
|
|
break
|
|
rows.append(row)
|
|
eq_(len(rows), 3)
|
|
|
|
@testing.requires.subqueries
|
|
def test_anonymous_rows(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
{"user_id": 7, "user_name": "jack"},
|
|
{"user_id": 8, "user_name": "ed"},
|
|
{"user_id": 9, "user_name": "fred"},
|
|
],
|
|
)
|
|
|
|
sel = (
|
|
select(users.c.user_id)
|
|
.where(users.c.user_name == "jack")
|
|
.scalar_subquery()
|
|
)
|
|
for row in connection.execute(select(sel + 1, sel + 3)):
|
|
eq_(row._mapping["anon_1"], 8)
|
|
eq_(row._mapping["anon_2"], 10)
|
|
|
|
def test_row_comparison(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=7, user_name="jack"))
|
|
rp = connection.execute(users.select()).first()
|
|
|
|
eq_(rp, rp)
|
|
is_(not (rp != rp), True)
|
|
|
|
equal = (7, "jack")
|
|
|
|
eq_(rp, equal)
|
|
eq_(equal, rp)
|
|
is_((not (rp != equal)), True)
|
|
is_(not (equal != equal), True)
|
|
|
|
def endless():
|
|
while True:
|
|
yield 1
|
|
|
|
ne_(rp, endless())
|
|
ne_(endless(), rp)
|
|
|
|
# test that everything compares the same
|
|
# as it would against a tuple
|
|
for compare in [False, 8, endless(), "xyz", (7, "jack")]:
|
|
for op in [
|
|
operator.eq,
|
|
operator.ne,
|
|
operator.gt,
|
|
operator.lt,
|
|
operator.ge,
|
|
operator.le,
|
|
]:
|
|
|
|
try:
|
|
control = op(equal, compare)
|
|
except TypeError:
|
|
# Py3K raises TypeError for some invalid comparisons
|
|
assert_raises(TypeError, op, rp, compare)
|
|
else:
|
|
eq_(control, op(rp, compare))
|
|
|
|
try:
|
|
control = op(compare, equal)
|
|
except TypeError:
|
|
# Py3K raises TypeError for some invalid comparisons
|
|
assert_raises(TypeError, op, compare, rp)
|
|
else:
|
|
eq_(control, op(compare, rp))
|
|
|
|
@testing.provide_metadata
|
|
def test_column_label_overlap_fallback(self, connection):
|
|
content = Table("content", self.metadata, Column("type", String(30)))
|
|
bar = Table("bar", self.metadata, Column("content_type", String(30)))
|
|
self.metadata.create_all(connection)
|
|
connection.execute(content.insert().values(type="t1"))
|
|
|
|
row = connection.execute(
|
|
content.select().set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
).first()
|
|
in_(content.c.type, row._mapping)
|
|
not_in(bar.c.content_type, row._mapping)
|
|
|
|
# in 1.x, would warn for string match, but return a result
|
|
not_in(sql.column("content_type"), row)
|
|
|
|
not_in(bar.c.content_type, row._mapping)
|
|
|
|
row = connection.execute(
|
|
select(func.now().label("content_type"))
|
|
).first()
|
|
|
|
not_in(content.c.type, row._mapping)
|
|
|
|
not_in(bar.c.content_type, row._mapping)
|
|
|
|
# in 1.x, would warn for string match, but return a result
|
|
not_in(sql.column("content_type"), row._mapping)
|
|
|
|
def test_pickled_rows(self, connection):
|
|
users = self.tables.users
|
|
addresses = self.tables.addresses
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
{"user_id": 7, "user_name": "jack"},
|
|
{"user_id": 8, "user_name": "ed"},
|
|
{"user_id": 9, "user_name": "fred"},
|
|
],
|
|
)
|
|
|
|
for use_pickle in False, True:
|
|
for use_labels in False, True:
|
|
result = connection.execute(
|
|
users.select()
|
|
.order_by(users.c.user_id)
|
|
.set_label_style(
|
|
LABEL_STYLE_TABLENAME_PLUS_COL
|
|
if use_labels
|
|
else LABEL_STYLE_NONE
|
|
)
|
|
).fetchall()
|
|
|
|
if use_pickle:
|
|
result = pickle.loads(pickle.dumps(result))
|
|
|
|
eq_(result, [(7, "jack"), (8, "ed"), (9, "fred")])
|
|
if use_labels:
|
|
eq_(result[0]._mapping["users_user_id"], 7)
|
|
eq_(
|
|
list(result[0]._fields),
|
|
["users_user_id", "users_user_name"],
|
|
)
|
|
else:
|
|
eq_(result[0]._mapping["user_id"], 7)
|
|
eq_(list(result[0]._fields), ["user_id", "user_name"])
|
|
|
|
eq_(result[0][0], 7)
|
|
|
|
assert_raises(
|
|
exc.NoSuchColumnError,
|
|
lambda: result[0]._mapping["fake key"],
|
|
)
|
|
|
|
# previously would warn
|
|
|
|
if use_pickle:
|
|
with expect_raises_message(
|
|
exc.NoSuchColumnError,
|
|
"Row was unpickled; lookup by ColumnElement is "
|
|
"unsupported",
|
|
):
|
|
result[0]._mapping[users.c.user_id]
|
|
else:
|
|
eq_(result[0]._mapping[users.c.user_id], 7)
|
|
|
|
if use_pickle:
|
|
with expect_raises_message(
|
|
exc.NoSuchColumnError,
|
|
"Row was unpickled; lookup by ColumnElement is "
|
|
"unsupported",
|
|
):
|
|
result[0]._mapping[users.c.user_name]
|
|
else:
|
|
eq_(result[0]._mapping[users.c.user_name], "jack")
|
|
|
|
assert_raises(
|
|
exc.NoSuchColumnError,
|
|
lambda: result[0]._mapping[addresses.c.user_id],
|
|
)
|
|
|
|
assert_raises(
|
|
exc.NoSuchColumnError,
|
|
lambda: result[0]._mapping[addresses.c.address_id],
|
|
)
|
|
|
|
def test_column_error_printing(self, connection):
|
|
result = connection.execute(select(1))
|
|
row = result.first()
|
|
|
|
class unprintable:
|
|
def __str__(self):
|
|
raise ValueError("nope")
|
|
|
|
msg = r"Could not locate column in row for column '%s'"
|
|
|
|
for accessor, repl in [
|
|
("x", "x"),
|
|
(Column("q", Integer), "q"),
|
|
(Column("q", Integer) + 12, r"q \+ :q_1"),
|
|
(unprintable(), "unprintable element.*"),
|
|
]:
|
|
assert_raises_message(
|
|
exc.NoSuchColumnError, msg % repl, result._getter, accessor
|
|
)
|
|
|
|
is_(result._getter(accessor, False), None)
|
|
|
|
assert_raises_message(
|
|
exc.NoSuchColumnError,
|
|
msg % repl,
|
|
lambda: row._mapping[accessor],
|
|
)
|
|
|
|
def test_fetchmany(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[{"user_id": i, "user_name": "n%d" % i} for i in range(7, 15)],
|
|
)
|
|
r = connection.execute(users.select())
|
|
rows = []
|
|
for row in r.fetchmany(size=2):
|
|
rows.append(row)
|
|
eq_(len(rows), 2)
|
|
|
|
def test_fetchmany_arraysize_default(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[{"user_id": i, "user_name": "n%d" % i} for i in range(1, 150)],
|
|
)
|
|
r = connection.execute(users.select())
|
|
arraysize = r.cursor.arraysize
|
|
rows = list(r.fetchmany())
|
|
|
|
eq_(len(rows), min(arraysize, 150))
|
|
|
|
def test_fetchmany_arraysize_set(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[{"user_id": i, "user_name": "n%d" % i} for i in range(7, 15)],
|
|
)
|
|
r = connection.execute(users.select())
|
|
r.cursor.arraysize = 4
|
|
rows = list(r.fetchmany())
|
|
eq_(len(rows), 4)
|
|
|
|
def test_column_slices(self, connection):
|
|
users = self.tables.users
|
|
addresses = self.tables.addresses
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
connection.execute(users.insert(), dict(user_id=2, user_name="jack"))
|
|
connection.execute(
|
|
addresses.insert(),
|
|
dict(address_id=1, user_id=2, address="foo@bar.com"),
|
|
)
|
|
|
|
r = connection.execute(text("select * from addresses")).first()
|
|
eq_(r[0:1], (1,))
|
|
eq_(r[1:], (2, "foo@bar.com"))
|
|
eq_(r[:-1], (1, 2))
|
|
|
|
def test_mappings(self, connection):
|
|
users = self.tables.users
|
|
addresses = self.tables.addresses
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
connection.execute(users.insert(), dict(user_id=2, user_name="jack"))
|
|
connection.execute(
|
|
addresses.insert(),
|
|
dict(address_id=1, user_id=2, address="foo@bar.com"),
|
|
)
|
|
|
|
r = connection.execute(text("select * from addresses"))
|
|
eq_(
|
|
r.mappings().all(),
|
|
[{"address_id": 1, "user_id": 2, "address": "foo@bar.com"}],
|
|
)
|
|
|
|
def test_column_accessor_basic_compiled_mapping(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
dict(user_id=1, user_name="john"),
|
|
dict(user_id=2, user_name="jack"),
|
|
],
|
|
)
|
|
|
|
r = connection.execute(
|
|
users.select().where(users.c.user_id == 2)
|
|
).first()
|
|
eq_(r.user_id, 2)
|
|
eq_(r._mapping["user_id"], 2)
|
|
eq_(r._mapping[users.c.user_id], 2)
|
|
|
|
eq_(r.user_name, "jack")
|
|
eq_(r._mapping["user_name"], "jack")
|
|
eq_(r._mapping[users.c.user_name], "jack")
|
|
|
|
def test_column_accessor_basic_compiled_traditional(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
dict(user_id=1, user_name="john"),
|
|
dict(user_id=2, user_name="jack"),
|
|
],
|
|
)
|
|
|
|
r = connection.execute(
|
|
users.select().where(users.c.user_id == 2)
|
|
).first()
|
|
|
|
eq_(r.user_id, 2)
|
|
eq_(r._mapping["user_id"], 2)
|
|
eq_(r._mapping[users.c.user_id], 2)
|
|
|
|
eq_(r.user_name, "jack")
|
|
eq_(r._mapping["user_name"], "jack")
|
|
eq_(r._mapping[users.c.user_name], "jack")
|
|
|
|
@testing.combinations(
|
|
(select(literal_column("1").label("col1")), ("col1",)),
|
|
(
|
|
select(
|
|
literal_column("1").label("col1"),
|
|
literal_column("2").label("col2"),
|
|
),
|
|
("col1", "col2"),
|
|
),
|
|
argnames="sql,cols",
|
|
)
|
|
def test_compiled_star_doesnt_interfere_w_description(
|
|
self, connection, sql, cols
|
|
):
|
|
"""test #6665"""
|
|
|
|
row = connection.execute(
|
|
select("*").select_from(sql.subquery())
|
|
).first()
|
|
eq_(row._fields, cols)
|
|
eq_(row._mapping["col1"], 1)
|
|
|
|
def test_row_getitem_string(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
dict(user_id=1, user_name="john"),
|
|
dict(user_id=2, user_name="jack"),
|
|
],
|
|
)
|
|
|
|
r = connection.execute(
|
|
text("select * from users where user_id=2")
|
|
).first()
|
|
|
|
with expect_raises_message(TypeError, "tuple indices must be"):
|
|
r["foo"]
|
|
|
|
eq_(r._mapping["user_name"], "jack")
|
|
|
|
def test_row_getitem_column(self, connection):
|
|
col = literal_column("1").label("foo")
|
|
|
|
row = connection.execute(select(col)).first()
|
|
|
|
with expect_raises_message(TypeError, "tuple indices must be"):
|
|
row[col]
|
|
|
|
eq_(row._mapping[col], 1)
|
|
|
|
def test_column_accessor_basic_text(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
dict(user_id=1, user_name="john"),
|
|
dict(user_id=2, user_name="jack"),
|
|
],
|
|
)
|
|
r = connection.execute(
|
|
text("select * from users where user_id=2")
|
|
).first()
|
|
|
|
eq_(r.user_id, 2)
|
|
|
|
eq_(r.user_name, "jack")
|
|
|
|
eq_(r._mapping["user_id"], 2)
|
|
|
|
eq_(r.user_name, "jack")
|
|
eq_(r._mapping["user_name"], "jack")
|
|
|
|
# cases which used to succeed w warning
|
|
with expect_raises_message(
|
|
exc.NoSuchColumnError, "Could not locate column in row"
|
|
):
|
|
r._mapping[users.c.user_id]
|
|
with expect_raises_message(
|
|
exc.NoSuchColumnError, "Could not locate column in row"
|
|
):
|
|
r._mapping[users.c.user_name]
|
|
|
|
def test_column_accessor_text_colexplicit(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
dict(user_id=1, user_name="john"),
|
|
dict(user_id=2, user_name="jack"),
|
|
],
|
|
)
|
|
r = connection.execute(
|
|
text("select * from users where user_id=2").columns(
|
|
users.c.user_id, users.c.user_name
|
|
)
|
|
).first()
|
|
|
|
eq_(r.user_id, 2)
|
|
eq_(r._mapping["user_id"], 2)
|
|
eq_(r._mapping[users.c.user_id], 2)
|
|
|
|
eq_(r.user_name, "jack")
|
|
eq_(r._mapping["user_name"], "jack")
|
|
eq_(r._mapping[users.c.user_name], "jack")
|
|
|
|
def test_column_accessor_textual_select(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
dict(user_id=1, user_name="john"),
|
|
dict(user_id=2, user_name="jack"),
|
|
],
|
|
)
|
|
# this will create column() objects inside
|
|
# the select(), these need to match on name anyway
|
|
r = connection.execute(
|
|
select(column("user_id"), column("user_name"))
|
|
.select_from(table("users"))
|
|
.where(text("user_id=2"))
|
|
).first()
|
|
|
|
# keyed access works in many ways
|
|
eq_(r.user_id, 2)
|
|
eq_(r.user_name, "jack")
|
|
eq_(r._mapping["user_id"], 2)
|
|
eq_(r.user_name, "jack")
|
|
eq_(r._mapping["user_name"], "jack")
|
|
|
|
# error cases that previously would warn
|
|
with expect_raises_message(
|
|
exc.NoSuchColumnError, "Could not locate column in row"
|
|
):
|
|
r._mapping[users.c.user_id]
|
|
with expect_raises_message(
|
|
exc.NoSuchColumnError, "Could not locate column in row"
|
|
):
|
|
r._mapping[users.c.user_name]
|
|
|
|
def test_column_accessor_dotted_union(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
|
|
# test a little sqlite < 3.10.0 weirdness - with the UNION,
|
|
# cols come back as "users.user_id" in cursor.description
|
|
r = connection.execute(
|
|
text(
|
|
"select users.user_id, users.user_name "
|
|
"from users "
|
|
"UNION select users.user_id, "
|
|
"users.user_name from users"
|
|
)
|
|
).first()
|
|
eq_(r._mapping["user_id"], 1)
|
|
eq_(r._mapping["user_name"], "john")
|
|
eq_(list(r._fields), ["user_id", "user_name"])
|
|
|
|
def test_column_accessor_sqlite_raw(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
|
|
r = connection.execute(
|
|
text(
|
|
"select users.user_id, users.user_name "
|
|
"from users "
|
|
"UNION select users.user_id, "
|
|
"users.user_name from users",
|
|
).execution_options(sqlite_raw_colnames=True)
|
|
).first()
|
|
|
|
if testing.against("sqlite < 3.10.0"):
|
|
not_in("user_id", r)
|
|
not_in("user_name", r)
|
|
eq_(r["users.user_id"], 1)
|
|
eq_(r["users.user_name"], "john")
|
|
|
|
eq_(list(r._fields), ["users.user_id", "users.user_name"])
|
|
else:
|
|
not_in("users.user_id", r._mapping)
|
|
not_in("users.user_name", r._mapping)
|
|
eq_(r._mapping["user_id"], 1)
|
|
eq_(r._mapping["user_name"], "john")
|
|
|
|
eq_(list(r._fields), ["user_id", "user_name"])
|
|
|
|
def test_column_accessor_sqlite_translated(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
|
|
r = connection.execute(
|
|
text(
|
|
"select users.user_id, users.user_name "
|
|
"from users "
|
|
"UNION select users.user_id, "
|
|
"users.user_name from users",
|
|
)
|
|
).first()
|
|
eq_(r._mapping["user_id"], 1)
|
|
eq_(r._mapping["user_name"], "john")
|
|
|
|
if testing.against("sqlite < 3.10.0"):
|
|
eq_(r._mapping["users.user_id"], 1)
|
|
eq_(r._mapping["users.user_name"], "john")
|
|
else:
|
|
not_in("users.user_id", r._mapping)
|
|
not_in("users.user_name", r._mapping)
|
|
|
|
eq_(list(r._fields), ["user_id", "user_name"])
|
|
|
|
def test_column_accessor_labels_w_dots(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
# test using literal tablename.colname
|
|
r = connection.execute(
|
|
text(
|
|
'select users.user_id AS "users.user_id", '
|
|
'users.user_name AS "users.user_name" '
|
|
"from users",
|
|
).execution_options(sqlite_raw_colnames=True)
|
|
).first()
|
|
eq_(r._mapping["users.user_id"], 1)
|
|
eq_(r._mapping["users.user_name"], "john")
|
|
not_in("user_name", r._mapping)
|
|
eq_(list(r._fields), ["users.user_id", "users.user_name"])
|
|
|
|
def test_column_accessor_unary(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
|
|
# unary expressions
|
|
r = connection.execute(
|
|
select(users.c.user_name.distinct()).order_by(users.c.user_name)
|
|
).first()
|
|
eq_(r._mapping[users.c.user_name], "john")
|
|
eq_(r.user_name, "john")
|
|
|
|
@testing.fixture
|
|
def _ab_row_fixture(self, connection):
|
|
r = connection.execute(
|
|
select(literal(1).label("a"), literal(2).label("b"))
|
|
).first()
|
|
return r
|
|
|
|
def test_named_tuple_access(self, _ab_row_fixture):
|
|
r = _ab_row_fixture
|
|
eq_(r.a, 1)
|
|
eq_(r.b, 2)
|
|
|
|
def test_named_tuple_missing_attr(self, _ab_row_fixture):
|
|
r = _ab_row_fixture
|
|
with expect_raises_message(
|
|
AttributeError, "Could not locate column in row for column 'c'"
|
|
):
|
|
r.c
|
|
|
|
def test_named_tuple_no_delete_present(self, _ab_row_fixture):
|
|
r = _ab_row_fixture
|
|
with expect_raises_message(AttributeError, "can't delete attribute"):
|
|
del r.a
|
|
|
|
def test_named_tuple_no_delete_missing(self, _ab_row_fixture):
|
|
r = _ab_row_fixture
|
|
# including for non-existent attributes
|
|
with expect_raises_message(AttributeError, "can't delete attribute"):
|
|
del r.c
|
|
|
|
def test_named_tuple_no_assign_present(self, _ab_row_fixture):
|
|
r = _ab_row_fixture
|
|
with expect_raises_message(AttributeError, "can't set attribute"):
|
|
r.a = 5
|
|
|
|
with expect_raises_message(AttributeError, "can't set attribute"):
|
|
r.a += 5
|
|
|
|
def test_named_tuple_no_assign_missing(self, _ab_row_fixture):
|
|
r = _ab_row_fixture
|
|
# including for non-existent attributes
|
|
with expect_raises_message(AttributeError, "can't set attribute"):
|
|
r.c = 5
|
|
|
|
def test_named_tuple_no_self_assign_missing(self, _ab_row_fixture):
|
|
r = _ab_row_fixture
|
|
with expect_raises_message(
|
|
AttributeError, "Could not locate column in row for column 'c'"
|
|
):
|
|
r.c += 5
|
|
|
|
def test_mapping_tuple_readonly_errors(self, connection):
|
|
r = connection.execute(
|
|
select(literal(1).label("a"), literal(2).label("b"))
|
|
).first()
|
|
r = r._mapping
|
|
eq_(r["a"], 1)
|
|
eq_(r["b"], 2)
|
|
|
|
with expect_raises_message(
|
|
KeyError, "Could not locate column in row for column 'c'"
|
|
):
|
|
r["c"]
|
|
|
|
with expect_raises_message(
|
|
TypeError, "'RowMapping' object does not support item assignment"
|
|
):
|
|
r["a"] = 5
|
|
|
|
with expect_raises_message(
|
|
TypeError, "'RowMapping' object does not support item assignment"
|
|
):
|
|
r["a"] += 5
|
|
|
|
def test_column_accessor_err(self, connection):
|
|
r = connection.execute(select(1)).first()
|
|
assert_raises_message(
|
|
AttributeError,
|
|
"Could not locate column in row for column 'foo'",
|
|
getattr,
|
|
r,
|
|
"foo",
|
|
)
|
|
assert_raises_message(
|
|
KeyError,
|
|
"Could not locate column in row for column 'foo'",
|
|
lambda: r._mapping["foo"],
|
|
)
|
|
|
|
def test_graceful_fetch_on_non_rows(self):
|
|
"""test that calling fetchone() etc. on a result that doesn't
|
|
return rows fails gracefully.
|
|
|
|
"""
|
|
|
|
# these proxies don't work with no cursor.description present.
|
|
# so they don't apply to this test at the moment.
|
|
# result.FullyBufferedCursorResult,
|
|
# result.BufferedRowCursorResult,
|
|
# result.BufferedColumnCursorResult
|
|
|
|
users = self.tables.users
|
|
|
|
with testing.db.connect() as conn:
|
|
keys_lambda = lambda r: r.keys() # noqa: E731
|
|
|
|
for meth in [
|
|
lambda r: r.fetchone(),
|
|
lambda r: r.fetchall(),
|
|
lambda r: r.first(),
|
|
lambda r: r.scalar(),
|
|
lambda r: r.fetchmany(),
|
|
lambda r: r._getter("user"),
|
|
keys_lambda,
|
|
lambda r: r.columns("user"),
|
|
lambda r: r.cursor_strategy.fetchone(r, r.cursor),
|
|
]:
|
|
trans = conn.begin()
|
|
result = conn.execute(users.insert(), dict(user_id=1))
|
|
|
|
assert_raises_message(
|
|
exc.ResourceClosedError,
|
|
"This result object does not return rows. "
|
|
"It has been closed automatically.",
|
|
meth,
|
|
result,
|
|
)
|
|
trans.rollback()
|
|
|
|
def test_fetchone_til_end(self, connection):
|
|
result = connection.exec_driver_sql("select * from users")
|
|
eq_(result.fetchone(), None)
|
|
eq_(result.fetchone(), None)
|
|
eq_(result.fetchone(), None)
|
|
result.close()
|
|
assert_raises_message(
|
|
exc.ResourceClosedError,
|
|
"This result object is closed.",
|
|
result.fetchone,
|
|
)
|
|
|
|
def test_row_case_sensitive(self, connection):
|
|
row = connection.execute(
|
|
select(
|
|
literal_column("1").label("case_insensitive"),
|
|
literal_column("2").label("CaseSensitive"),
|
|
)
|
|
).first()
|
|
|
|
eq_(list(row._fields), ["case_insensitive", "CaseSensitive"])
|
|
|
|
in_("case_insensitive", row._keymap)
|
|
in_("CaseSensitive", row._keymap)
|
|
not_in("casesensitive", row._keymap)
|
|
|
|
eq_(row._mapping["case_insensitive"], 1)
|
|
eq_(row._mapping["CaseSensitive"], 2)
|
|
|
|
assert_raises(KeyError, lambda: row._mapping["Case_insensitive"])
|
|
assert_raises(KeyError, lambda: row._mapping["casesensitive"])
|
|
|
|
def test_row_case_sensitive_unoptimized(self, testing_engine):
|
|
with testing_engine().connect() as ins_conn:
|
|
row = ins_conn.execute(
|
|
select(
|
|
literal_column("1").label("case_insensitive"),
|
|
literal_column("2").label("CaseSensitive"),
|
|
text("3 AS screw_up_the_cols"),
|
|
)
|
|
).first()
|
|
|
|
eq_(
|
|
list(row._fields),
|
|
["case_insensitive", "CaseSensitive", "screw_up_the_cols"],
|
|
)
|
|
|
|
in_("case_insensitive", row._keymap)
|
|
in_("CaseSensitive", row._keymap)
|
|
not_in("casesensitive", row._keymap)
|
|
|
|
eq_(row._mapping["case_insensitive"], 1)
|
|
eq_(row._mapping["CaseSensitive"], 2)
|
|
eq_(row._mapping["screw_up_the_cols"], 3)
|
|
|
|
assert_raises(KeyError, lambda: row._mapping["Case_insensitive"])
|
|
assert_raises(KeyError, lambda: row._mapping["casesensitive"])
|
|
assert_raises(KeyError, lambda: row._mapping["screw_UP_the_cols"])
|
|
|
|
def test_row_as_args(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
r = connection.execute(
|
|
users.select().where(users.c.user_id == 1)
|
|
).first()
|
|
connection.execute(users.delete())
|
|
connection.execute(users.insert(), r._mapping)
|
|
eq_(connection.execute(users.select()).fetchall(), [(1, "john")])
|
|
|
|
@testing.requires.tuple_in
|
|
def test_row_tuple_interpretation(self, connection):
|
|
"""test #7292"""
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
dict(user_id=1, user_name="u1"),
|
|
dict(user_id=2, user_name="u2"),
|
|
dict(user_id=3, user_name="u3"),
|
|
],
|
|
)
|
|
rows = connection.execute(
|
|
select(users.c.user_id, users.c.user_name)
|
|
).all()
|
|
|
|
# was previously needed
|
|
# rows = [(x, y) for x, y in rows]
|
|
|
|
new_stmt = (
|
|
select(users)
|
|
.where(tuple_(users.c.user_id, users.c.user_name).in_(rows))
|
|
.order_by(users.c.user_id)
|
|
)
|
|
|
|
eq_(
|
|
connection.execute(new_stmt).all(),
|
|
[(1, "u1"), (2, "u2"), (3, "u3")],
|
|
)
|
|
|
|
def test_result_as_args(self, connection):
|
|
users = self.tables.users
|
|
users2 = self.tables.users2
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
dict(user_id=1, user_name="john"),
|
|
dict(user_id=2, user_name="ed"),
|
|
],
|
|
)
|
|
r = connection.execute(users.select())
|
|
connection.execute(users2.insert(), [row._mapping for row in r])
|
|
eq_(
|
|
connection.execute(
|
|
users2.select().order_by(users2.c.user_id)
|
|
).fetchall(),
|
|
[(1, "john"), (2, "ed")],
|
|
)
|
|
|
|
connection.execute(users2.delete())
|
|
r = connection.execute(users.select())
|
|
connection.execute(users2.insert(), [row._mapping for row in r])
|
|
eq_(
|
|
connection.execute(
|
|
users2.select().order_by(users2.c.user_id)
|
|
).fetchall(),
|
|
[(1, "john"), (2, "ed")],
|
|
)
|
|
|
|
@testing.requires.duplicate_names_in_cursor_description
|
|
def test_ambiguous_column(self, connection):
|
|
users = self.tables.users
|
|
addresses = self.tables.addresses
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
result = connection.execute(
|
|
users.outerjoin(addresses)
|
|
.select()
|
|
.set_label_style(LABEL_STYLE_NONE)
|
|
)
|
|
r = result.first()
|
|
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Ambiguous column name",
|
|
lambda: r._mapping["user_id"],
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Ambiguous column name",
|
|
result._getter,
|
|
"user_id",
|
|
)
|
|
|
|
# pure positional targeting; users.c.user_id
|
|
# and addresses.c.user_id are known!
|
|
# works as of 1.1 issue #3501
|
|
eq_(r._mapping[users.c.user_id], 1)
|
|
eq_(r._mapping[addresses.c.user_id], None)
|
|
|
|
# try to trick it - fake_table isn't in the result!
|
|
# we get the correct error
|
|
fake_table = Table("fake", MetaData(), Column("user_id", Integer))
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Could not locate column in row for column 'fake.user_id'",
|
|
lambda: r._mapping[fake_table.c.user_id],
|
|
)
|
|
|
|
r = pickle.loads(pickle.dumps(r))
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Ambiguous column name",
|
|
lambda: r._mapping["user_id"],
|
|
)
|
|
|
|
@testing.requires.duplicate_names_in_cursor_description
|
|
def test_ambiguous_column_by_col(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
ua = users.alias()
|
|
u2 = users.alias()
|
|
result = connection.execute(
|
|
select(users.c.user_id, ua.c.user_id)
|
|
.select_from(users.join(ua, true()))
|
|
.set_label_style(LABEL_STYLE_NONE)
|
|
)
|
|
row = result.first()
|
|
|
|
# as of 1.1 issue #3501, we use pure positional
|
|
# targeting for the column objects here
|
|
eq_(row._mapping[users.c.user_id], 1)
|
|
|
|
eq_(row._mapping[ua.c.user_id], 1)
|
|
|
|
# this now works as of 1.1 issue #3501;
|
|
# previously this was stuck on "ambiguous column name"
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Could not locate column in row",
|
|
lambda: row._mapping[u2.c.user_id],
|
|
)
|
|
|
|
@testing.requires.duplicate_names_in_cursor_description
|
|
def test_ambiguous_column_contains(self, connection):
|
|
users = self.tables.users
|
|
addresses = self.tables.addresses
|
|
|
|
# ticket 2702. in 0.7 we'd get True, False.
|
|
# in 0.8, both columns are present so it's True;
|
|
# but when they're fetched you'll get the ambiguous error.
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
result = connection.execute(
|
|
select(users.c.user_id, addresses.c.user_id).select_from(
|
|
users.outerjoin(addresses)
|
|
)
|
|
)
|
|
row = result.first()
|
|
|
|
eq_(
|
|
set(
|
|
[
|
|
users.c.user_id in row._mapping,
|
|
addresses.c.user_id in row._mapping,
|
|
]
|
|
),
|
|
set([True]),
|
|
)
|
|
|
|
@testing.combinations(
|
|
(("name_label", "*"), False),
|
|
(("*", "name_label"), False),
|
|
(("user_id", "name_label", "user_name"), False),
|
|
(("user_id", "name_label", "*", "user_name"), True),
|
|
argnames="cols,other_cols_are_ambiguous",
|
|
)
|
|
@testing.requires.select_star_mixed
|
|
def test_label_against_star(
|
|
self, connection, cols, other_cols_are_ambiguous
|
|
):
|
|
"""test #8536"""
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
|
|
stmt = select(
|
|
*[
|
|
text("*")
|
|
if colname == "*"
|
|
else users.c.user_name.label("name_label")
|
|
if colname == "name_label"
|
|
else users.c[colname]
|
|
for colname in cols
|
|
]
|
|
)
|
|
|
|
row = connection.execute(stmt).first()
|
|
|
|
eq_(row._mapping["name_label"], "john")
|
|
|
|
if other_cols_are_ambiguous:
|
|
with expect_raises_message(
|
|
exc.InvalidRequestError, "Ambiguous column name"
|
|
):
|
|
row._mapping["user_id"]
|
|
with expect_raises_message(
|
|
exc.InvalidRequestError, "Ambiguous column name"
|
|
):
|
|
row._mapping["user_name"]
|
|
else:
|
|
eq_(row._mapping["user_id"], 1)
|
|
eq_(row._mapping["user_name"], "john")
|
|
|
|
def test_loose_matching_one(self, connection):
|
|
users = self.tables.users
|
|
addresses = self.tables.addresses
|
|
|
|
connection.execute(users.insert(), {"user_id": 1, "user_name": "john"})
|
|
connection.execute(
|
|
addresses.insert(),
|
|
{"address_id": 1, "user_id": 1, "address": "email"},
|
|
)
|
|
|
|
# use some column labels in the SELECT
|
|
result = connection.execute(
|
|
TextualSelect(
|
|
text(
|
|
"select users.user_name AS users_user_name, "
|
|
"users.user_id AS user_id, "
|
|
"addresses.address_id AS address_id "
|
|
"FROM users JOIN addresses "
|
|
"ON users.user_id = addresses.user_id "
|
|
"WHERE users.user_id=1 "
|
|
),
|
|
[users.c.user_id, users.c.user_name, addresses.c.address_id],
|
|
positional=False,
|
|
)
|
|
)
|
|
row = result.first()
|
|
eq_(row._mapping[users.c.user_id], 1)
|
|
eq_(row._mapping[users.c.user_name], "john")
|
|
|
|
def test_loose_matching_two(self, connection):
|
|
users = self.tables.users
|
|
addresses = self.tables.addresses
|
|
|
|
connection.execute(users.insert(), {"user_id": 1, "user_name": "john"})
|
|
connection.execute(
|
|
addresses.insert(),
|
|
{"address_id": 1, "user_id": 1, "address": "email"},
|
|
)
|
|
|
|
# use some column labels in the SELECT
|
|
result = connection.execute(
|
|
TextualSelect(
|
|
text(
|
|
"select users.user_name AS users_user_name, "
|
|
"users.user_id AS user_id, "
|
|
"addresses.user_id "
|
|
"FROM users JOIN addresses "
|
|
"ON users.user_id = addresses.user_id "
|
|
"WHERE users.user_id=1 "
|
|
),
|
|
[users.c.user_id, users.c.user_name, addresses.c.user_id],
|
|
positional=False,
|
|
)
|
|
)
|
|
row = result.first()
|
|
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Ambiguous column name",
|
|
lambda: row._mapping[users.c.user_id],
|
|
)
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Ambiguous column name",
|
|
lambda: row._mapping[addresses.c.user_id],
|
|
)
|
|
eq_(row._mapping[users.c.user_name], "john")
|
|
|
|
def test_ambiguous_column_by_col_plus_label(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="john"))
|
|
result = connection.execute(
|
|
select(
|
|
users.c.user_id,
|
|
type_coerce(users.c.user_id, Integer).label("foo"),
|
|
)
|
|
)
|
|
row = result.first()
|
|
eq_(row._mapping[users.c.user_id], 1)
|
|
eq_(row[1], 1)
|
|
|
|
def test_fetch_partial_result_map(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=7, user_name="ed"))
|
|
|
|
t = text("select * from users").columns(user_name=String())
|
|
eq_(connection.execute(t).fetchall(), [(7, "ed")])
|
|
|
|
def test_fetch_unordered_result_map(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=7, user_name="ed"))
|
|
|
|
class Goofy1(TypeDecorator):
|
|
impl = String
|
|
cache_ok = True
|
|
|
|
def process_result_value(self, value, dialect):
|
|
return value + "a"
|
|
|
|
class Goofy2(TypeDecorator):
|
|
impl = String
|
|
cache_ok = True
|
|
|
|
def process_result_value(self, value, dialect):
|
|
return value + "b"
|
|
|
|
class Goofy3(TypeDecorator):
|
|
impl = String
|
|
cache_ok = True
|
|
|
|
def process_result_value(self, value, dialect):
|
|
return value + "c"
|
|
|
|
t = text(
|
|
"select user_name as a, user_name as b, "
|
|
"user_name as c from users"
|
|
).columns(a=Goofy1(), b=Goofy2(), c=Goofy3())
|
|
eq_(connection.execute(t).fetchall(), [("eda", "edb", "edc")])
|
|
|
|
@testing.requires.subqueries
|
|
def test_column_label_targeting(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=7, user_name="ed"))
|
|
|
|
for s in (
|
|
users.select().alias("foo"),
|
|
users.select().alias(users.name),
|
|
):
|
|
row = connection.execute(
|
|
s.select().set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
).first()
|
|
eq_(row._mapping[s.c.user_id], 7)
|
|
eq_(row._mapping[s.c.user_name], "ed")
|
|
|
|
def test_ro_mapping_py3k(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="foo"))
|
|
result = connection.execute(users.select())
|
|
|
|
row = result.first()
|
|
dict_row = row._asdict()
|
|
|
|
# dictionaries aren't ordered in Python 3 until 3.7
|
|
odict_row = collections.OrderedDict(
|
|
[("user_id", 1), ("user_name", "foo")]
|
|
)
|
|
eq_(dict_row, odict_row)
|
|
|
|
mapping_row = row._mapping
|
|
|
|
eq_(list(mapping_row), list(mapping_row.keys()))
|
|
eq_(odict_row.keys(), mapping_row.keys())
|
|
eq_(odict_row.values(), mapping_row.values())
|
|
eq_(odict_row.items(), mapping_row.items())
|
|
|
|
@testing.combinations(
|
|
(lambda result: result),
|
|
(lambda result: result.first(),),
|
|
(lambda result: result.first()._mapping),
|
|
argnames="get_object",
|
|
)
|
|
def test_keys(self, connection, get_object):
|
|
users = self.tables.users
|
|
addresses = self.tables.addresses
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="foo"))
|
|
result = connection.execute(users.select())
|
|
|
|
obj = get_object(result)
|
|
|
|
if isinstance(obj, Row):
|
|
keys = obj._mapping.keys()
|
|
else:
|
|
keys = obj.keys()
|
|
|
|
# in 1.4, keys() is now a view that includes support for testing
|
|
# of columns and other objects
|
|
eq_(len(keys), 2)
|
|
eq_(list(keys), ["user_id", "user_name"])
|
|
eq_(keys, ["user_id", "user_name"])
|
|
ne_(keys, ["user_name", "user_id"])
|
|
in_("user_id", keys)
|
|
not_in("foo", keys)
|
|
in_(users.c.user_id, keys)
|
|
not_in(0, keys)
|
|
not_in(addresses.c.user_id, keys)
|
|
not_in(addresses.c.address, keys)
|
|
|
|
if isinstance(obj, Row):
|
|
eq_(obj._fields, ("user_id", "user_name"))
|
|
|
|
def test_row_mapping_keys(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="foo"))
|
|
result = connection.execute(users.select())
|
|
eq_(result.keys(), ["user_id", "user_name"])
|
|
row = result.first()
|
|
eq_(list(row._mapping.keys()), ["user_id", "user_name"])
|
|
eq_(row._fields, ("user_id", "user_name"))
|
|
|
|
in_("user_id", row._fields)
|
|
not_in("foo", row._fields)
|
|
in_(users.c.user_id, row._mapping.keys())
|
|
|
|
def test_row_keys_legacy_dont_warn(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="foo"))
|
|
result = connection.execute(users.select())
|
|
row = result.first()
|
|
|
|
eq_(dict(row._mapping), {"user_id": 1, "user_name": "foo"})
|
|
|
|
eq_(row._fields, ("user_id", "user_name"))
|
|
|
|
def test_row_namedtuple_legacy_ok(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="foo"))
|
|
result = connection.execute(users.select())
|
|
row = result.first()
|
|
eq_(row.user_id, 1)
|
|
eq_(row.user_name, "foo")
|
|
|
|
def test_keys_anon_labels(self, connection):
|
|
"""test [ticket:3483]"""
|
|
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="foo"))
|
|
result = connection.execute(
|
|
select(
|
|
users.c.user_id,
|
|
users.c.user_name.label(None),
|
|
func.count(literal_column("1")),
|
|
).group_by(users.c.user_id, users.c.user_name)
|
|
)
|
|
|
|
eq_(result.keys(), ["user_id", "user_name_1", "count_1"])
|
|
row = result.first()
|
|
eq_(row._fields, ("user_id", "user_name_1", "count_1"))
|
|
eq_(list(row._mapping.keys()), ["user_id", "user_name_1", "count_1"])
|
|
|
|
def test_items(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="foo"))
|
|
r = connection.execute(users.select()).first()
|
|
eq_(
|
|
[(x[0].lower(), x[1]) for x in list(r._mapping.items())],
|
|
[("user_id", 1), ("user_name", "foo")],
|
|
)
|
|
|
|
def test_len(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="foo"))
|
|
r = connection.execute(users.select()).first()
|
|
eq_(len(r), 2)
|
|
|
|
r = connection.exec_driver_sql(
|
|
"select user_name, user_id from users"
|
|
).first()
|
|
eq_(len(r), 2)
|
|
r = connection.exec_driver_sql("select user_name from users").first()
|
|
eq_(len(r), 1)
|
|
|
|
def test_sorting_in_python(self, connection):
|
|
users = self.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
dict(user_id=1, user_name="foo"),
|
|
dict(user_id=2, user_name="bar"),
|
|
dict(user_id=3, user_name="def"),
|
|
],
|
|
)
|
|
|
|
rows = connection.execute(
|
|
users.select().order_by(users.c.user_name)
|
|
).fetchall()
|
|
|
|
eq_(rows, [(2, "bar"), (3, "def"), (1, "foo")])
|
|
|
|
eq_(sorted(rows), [(1, "foo"), (2, "bar"), (3, "def")])
|
|
|
|
def test_column_order_with_simple_query(self, connection):
|
|
# should return values in column definition order
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="foo"))
|
|
r = connection.execute(
|
|
users.select().where(users.c.user_id == 1)
|
|
).first()
|
|
eq_(r[0], 1)
|
|
eq_(r[1], "foo")
|
|
eq_([x.lower() for x in r._fields], ["user_id", "user_name"])
|
|
eq_(list(r._mapping.values()), [1, "foo"])
|
|
|
|
def test_column_order_with_text_query(self, connection):
|
|
# should return values in query order
|
|
users = self.tables.users
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="foo"))
|
|
|
|
r = connection.exec_driver_sql(
|
|
"select user_name, user_id from users"
|
|
).first()
|
|
eq_(r[0], "foo")
|
|
eq_(r[1], 1)
|
|
eq_([x.lower() for x in r._fields], ["user_name", "user_id"])
|
|
eq_(list(r._mapping.values()), ["foo", 1])
|
|
|
|
@testing.crashes("oracle", "FIXME: unknown, verify not fails_on()")
|
|
@testing.provide_metadata
|
|
def test_column_accessor_shadow(self, connection):
|
|
shadowed = Table(
|
|
"test_shadowed",
|
|
self.metadata,
|
|
Column("shadow_id", INT, primary_key=True),
|
|
Column("shadow_name", VARCHAR(20)),
|
|
Column("parent", VARCHAR(20)),
|
|
Column("row", VARCHAR(40)),
|
|
Column("_parent", VARCHAR(20)),
|
|
Column("_row", VARCHAR(20)),
|
|
)
|
|
self.metadata.create_all(connection)
|
|
connection.execute(
|
|
shadowed.insert(),
|
|
dict(
|
|
shadow_id=1,
|
|
shadow_name="The Shadow",
|
|
parent="The Light",
|
|
row="Without light there is no shadow",
|
|
_parent="Hidden parent",
|
|
_row="Hidden row",
|
|
),
|
|
)
|
|
r = connection.execute(
|
|
shadowed.select().where(shadowed.c.shadow_id == 1)
|
|
).first()
|
|
|
|
eq_(r.shadow_id, 1)
|
|
eq_(r._mapping["shadow_id"], 1)
|
|
eq_(r._mapping[shadowed.c.shadow_id], 1)
|
|
|
|
eq_(r.shadow_name, "The Shadow")
|
|
eq_(r._mapping["shadow_name"], "The Shadow")
|
|
eq_(r._mapping[shadowed.c.shadow_name], "The Shadow")
|
|
|
|
eq_(r.parent, "The Light")
|
|
eq_(r._mapping["parent"], "The Light")
|
|
eq_(r._mapping[shadowed.c.parent], "The Light")
|
|
|
|
eq_(r.row, "Without light there is no shadow")
|
|
eq_(r._mapping["row"], "Without light there is no shadow")
|
|
eq_(r._mapping[shadowed.c.row], "Without light there is no shadow")
|
|
|
|
eq_(r._mapping["_parent"], "Hidden parent")
|
|
eq_(r._mapping["_row"], "Hidden row")
|
|
|
|
def test_nontuple_row(self):
|
|
"""ensure the C version of BaseRow handles
|
|
duck-type-dependent rows.
|
|
|
|
|
|
As of 1.4 they are converted internally to tuples in any case.
|
|
|
|
"""
|
|
|
|
class MyList:
|
|
def __init__(self, data):
|
|
self.internal_list = data
|
|
|
|
def __len__(self):
|
|
return len(self.internal_list)
|
|
|
|
def __getitem__(self, i):
|
|
return list.__getitem__(self.internal_list, i)
|
|
|
|
proxy = Row(
|
|
object(),
|
|
[None],
|
|
{"key": (0, None, "key"), 0: (0, None, "key")},
|
|
Row._default_key_style,
|
|
MyList(["value"]),
|
|
)
|
|
eq_(list(proxy), ["value"])
|
|
eq_(proxy[0], "value")
|
|
eq_(proxy._mapping["key"], "value")
|
|
|
|
def test_no_rowcount_on_selects_inserts(self, metadata, testing_engine):
|
|
"""assert that rowcount is only called on deletes and updates.
|
|
|
|
This because cursor.rowcount may can be expensive on some dialects
|
|
such as Firebird, however many dialects require it be called
|
|
before the cursor is closed.
|
|
|
|
"""
|
|
|
|
engine = testing_engine()
|
|
|
|
t = Table("t1", metadata, Column("data", String(10)))
|
|
metadata.create_all(engine)
|
|
|
|
with patch.object(
|
|
engine.dialect.execution_ctx_cls, "rowcount"
|
|
) as mock_rowcount:
|
|
with engine.begin() as conn:
|
|
mock_rowcount.__get__ = Mock()
|
|
conn.execute(
|
|
t.insert(),
|
|
[{"data": "d1"}, {"data": "d2"}, {"data": "d3"}],
|
|
)
|
|
|
|
eq_(len(mock_rowcount.__get__.mock_calls), 0)
|
|
|
|
eq_(
|
|
conn.execute(t.select()).fetchall(),
|
|
[("d1",), ("d2",), ("d3",)],
|
|
)
|
|
eq_(len(mock_rowcount.__get__.mock_calls), 0)
|
|
|
|
conn.execute(t.update(), {"data": "d4"})
|
|
|
|
eq_(len(mock_rowcount.__get__.mock_calls), 1)
|
|
|
|
conn.execute(t.delete())
|
|
eq_(len(mock_rowcount.__get__.mock_calls), 2)
|
|
|
|
def test_row_is_sequence(self):
|
|
|
|
row = Row(
|
|
object(),
|
|
[None],
|
|
{"key": (None, 0), 0: (None, 0)},
|
|
Row._default_key_style,
|
|
["value"],
|
|
)
|
|
is_true(isinstance(row, collections_abc.Sequence))
|
|
|
|
def test_row_special_names(self):
|
|
metadata = SimpleResultMetaData(["key", "count", "index", "foo"])
|
|
row = Row(
|
|
metadata,
|
|
[None, None, None, None],
|
|
metadata._keymap,
|
|
Row._default_key_style,
|
|
["kv", "cv", "iv", "f"],
|
|
)
|
|
is_true(isinstance(row, collections_abc.Sequence))
|
|
|
|
eq_(row.key, "kv")
|
|
eq_(row.count, "cv")
|
|
eq_(row.index, "iv")
|
|
|
|
eq_(row._mapping["foo"], "f")
|
|
eq_(row._mapping["count"], "cv")
|
|
eq_(row._mapping["index"], "iv")
|
|
|
|
metadata = SimpleResultMetaData(["key", "q", "p"])
|
|
|
|
row = Row(
|
|
metadata,
|
|
[None, None, None],
|
|
metadata._keymap,
|
|
Row._default_key_style,
|
|
["kv", "cv", "iv"],
|
|
)
|
|
is_true(isinstance(row, collections_abc.Sequence))
|
|
|
|
eq_(row.key, "kv")
|
|
eq_(row.q, "cv")
|
|
eq_(row.p, "iv")
|
|
eq_(row.index("cv"), 1)
|
|
eq_(row.count("cv"), 1)
|
|
eq_(row.count("x"), 0)
|
|
|
|
def test_new_row_no_dict_behaviors(self):
|
|
"""This mode is not used currently but will be once we are in 2.0."""
|
|
metadata = SimpleResultMetaData(
|
|
[
|
|
"a",
|
|
"b",
|
|
"count",
|
|
]
|
|
)
|
|
row = Row(
|
|
metadata,
|
|
[None, None, None],
|
|
metadata._keymap,
|
|
KEY_INTEGER_ONLY,
|
|
["av", "bv", "cv"],
|
|
)
|
|
|
|
eq_(dict(row._mapping), {"a": "av", "b": "bv", "count": "cv"})
|
|
|
|
with assertions.expect_raises(
|
|
TypeError,
|
|
"tuple indices must be integers or slices, not str",
|
|
):
|
|
eq_(row["a"], "av")
|
|
|
|
with assertions.expect_raises_message(
|
|
TypeError,
|
|
"tuple indices must be integers or slices, not str",
|
|
):
|
|
eq_(row["count"], "cv")
|
|
|
|
eq_(list(row._mapping), ["a", "b", "count"])
|
|
|
|
def test_row_is_hashable(self):
|
|
|
|
row = Row(
|
|
object(),
|
|
[None, None, None],
|
|
{"key": (None, 0), 0: (None, 0)},
|
|
Row._default_key_style,
|
|
(1, "value", "foo"),
|
|
)
|
|
eq_(hash(row), hash((1, "value", "foo")))
|
|
|
|
@testing.provide_metadata
|
|
def test_row_getitem_indexes_compiled(self, connection):
|
|
values = Table(
|
|
"rp",
|
|
self.metadata,
|
|
Column("key", String(10), primary_key=True),
|
|
Column("value", String(10)),
|
|
)
|
|
values.create(connection)
|
|
|
|
connection.execute(values.insert(), dict(key="One", value="Uno"))
|
|
row = connection.execute(values.select()).first()
|
|
eq_(row._mapping["key"], "One")
|
|
eq_(row._mapping["value"], "Uno")
|
|
eq_(row[0], "One")
|
|
eq_(row[1], "Uno")
|
|
eq_(row[-2], "One")
|
|
eq_(row[-1], "Uno")
|
|
eq_(row[1:0:-1], ("Uno",))
|
|
|
|
@testing.only_on("sqlite")
|
|
def test_row_getitem_indexes_raw(self, connection):
|
|
row = connection.exec_driver_sql(
|
|
"select 'One' as key, 'Uno' as value"
|
|
).first()
|
|
eq_(row._mapping["key"], "One")
|
|
eq_(row._mapping["value"], "Uno")
|
|
eq_(row[0], "One")
|
|
eq_(row[1], "Uno")
|
|
eq_(row[-2], "One")
|
|
eq_(row[-1], "Uno")
|
|
eq_(row[1:0:-1], ("Uno",))
|
|
|
|
@testing.requires.cextensions
|
|
@testing.provide_metadata
|
|
def test_row_c_sequence_check(self, connection):
|
|
users = self.tables.users2
|
|
|
|
connection.execute(users.insert(), dict(user_id=1, user_name="Test"))
|
|
row = connection.execute(
|
|
users.select().where(users.c.user_id == 1)
|
|
).fetchone()
|
|
|
|
s = StringIO()
|
|
writer = csv.writer(s)
|
|
# csv performs PySequenceCheck call
|
|
writer.writerow(row)
|
|
assert s.getvalue().strip() == "1,Test"
|
|
|
|
@testing.requires.selectone
|
|
def test_empty_accessors(self, connection):
|
|
statements = [
|
|
(
|
|
"select 1",
|
|
[
|
|
lambda r: r.last_inserted_params(),
|
|
lambda r: r.last_updated_params(),
|
|
lambda r: r.prefetch_cols(),
|
|
lambda r: r.postfetch_cols(),
|
|
lambda r: r.inserted_primary_key,
|
|
],
|
|
"Statement is not a compiled expression construct.",
|
|
),
|
|
(
|
|
select(1),
|
|
[
|
|
lambda r: r.last_inserted_params(),
|
|
lambda r: r.inserted_primary_key,
|
|
],
|
|
r"Statement is not an insert\(\) expression construct.",
|
|
),
|
|
(
|
|
select(1),
|
|
[lambda r: r.last_updated_params()],
|
|
r"Statement is not an update\(\) expression construct.",
|
|
),
|
|
(
|
|
select(1),
|
|
[lambda r: r.prefetch_cols(), lambda r: r.postfetch_cols()],
|
|
r"Statement is not an insert\(\) "
|
|
r"or update\(\) expression construct.",
|
|
),
|
|
]
|
|
|
|
for stmt, meths, msg in statements:
|
|
if isinstance(stmt, str):
|
|
r = connection.exec_driver_sql(stmt)
|
|
else:
|
|
r = connection.execute(stmt)
|
|
try:
|
|
for meth in meths:
|
|
assert_raises_message(
|
|
sa_exc.InvalidRequestError, msg, meth, r
|
|
)
|
|
|
|
finally:
|
|
r.close()
|
|
|
|
@testing.requires.dbapi_lastrowid
|
|
def test_lastrowid(self, connection):
|
|
users = self.tables.users
|
|
|
|
r = connection.execute(
|
|
users.insert(), dict(user_id=1, user_name="Test")
|
|
)
|
|
eq_(r.lastrowid, r.context.get_lastrowid())
|
|
|
|
def test_raise_errors(self, connection):
|
|
users = self.tables.users
|
|
|
|
class Wrapper:
|
|
def __init__(self, context):
|
|
self.context = context
|
|
|
|
def __getattr__(self, name):
|
|
if name in ("rowcount", "get_lastrowid"):
|
|
raise Exception("canary")
|
|
return getattr(self.context, name)
|
|
|
|
r = connection.execute(
|
|
users.insert(), dict(user_id=1, user_name="Test")
|
|
)
|
|
r.context = Wrapper(r.context)
|
|
with expect_raises_message(Exception, "canary"):
|
|
r.rowcount
|
|
with expect_raises_message(Exception, "canary"):
|
|
r.lastrowid
|
|
|
|
@testing.combinations("plain", "mapping", "scalar", argnames="result_type")
|
|
@testing.combinations(
|
|
"stream_results", "yield_per", "yield_per_meth", argnames="optname"
|
|
)
|
|
@testing.combinations(10, 50, argnames="value")
|
|
@testing.combinations(
|
|
"meth", "passed_in", "stmt", argnames="send_opts_how"
|
|
)
|
|
def test_stream_options(
|
|
self,
|
|
connection,
|
|
optname,
|
|
value,
|
|
send_opts_how,
|
|
result_type,
|
|
close_result_when_finished,
|
|
):
|
|
table = self.tables.test
|
|
|
|
connection.execute(
|
|
table.insert(),
|
|
[{"x": i, "y": "t_%d" % i} for i in range(15, 3000)],
|
|
)
|
|
|
|
if optname == "stream_results":
|
|
opts = {"stream_results": True, "max_row_buffer": value}
|
|
elif optname == "yield_per":
|
|
opts = {"yield_per": value}
|
|
elif optname == "yield_per_meth":
|
|
opts = {"stream_results": True}
|
|
else:
|
|
assert False
|
|
|
|
if send_opts_how == "meth":
|
|
result = connection.execution_options(**opts).execute(
|
|
table.select()
|
|
)
|
|
elif send_opts_how == "passed_in":
|
|
result = connection.execute(table.select(), execution_options=opts)
|
|
elif send_opts_how == "stmt":
|
|
result = connection.execute(
|
|
table.select().execution_options(**opts)
|
|
)
|
|
else:
|
|
assert False
|
|
|
|
if result_type == "mapping":
|
|
result = result.mappings()
|
|
real_result = result._real_result
|
|
elif result_type == "scalar":
|
|
result = result.scalars()
|
|
real_result = result._real_result
|
|
else:
|
|
real_result = result
|
|
|
|
if optname == "yield_per_meth":
|
|
result = result.yield_per(value)
|
|
|
|
if result_type == "mapping" or result_type == "scalar":
|
|
real_result = result._real_result
|
|
else:
|
|
real_result = result
|
|
|
|
close_result_when_finished(result, consume=True)
|
|
|
|
if optname == "yield_per" and value is not None:
|
|
expected_opt = {
|
|
"stream_results": True,
|
|
"max_row_buffer": value,
|
|
"yield_per": value,
|
|
}
|
|
elif optname == "stream_results" and value is not None:
|
|
expected_opt = {
|
|
"stream_results": True,
|
|
"max_row_buffer": value,
|
|
}
|
|
else:
|
|
expected_opt = None
|
|
|
|
if expected_opt is not None:
|
|
eq_(real_result.context.execution_options, expected_opt)
|
|
|
|
if value is None:
|
|
assert isinstance(
|
|
real_result.cursor_strategy, _cursor.CursorFetchStrategy
|
|
)
|
|
return
|
|
|
|
assert isinstance(
|
|
real_result.cursor_strategy, _cursor.BufferedRowCursorFetchStrategy
|
|
)
|
|
eq_(real_result.cursor_strategy._max_row_buffer, value)
|
|
|
|
if optname == "yield_per" or optname == "yield_per_meth":
|
|
eq_(real_result.cursor_strategy._bufsize, value)
|
|
else:
|
|
eq_(real_result.cursor_strategy._bufsize, min(value, 5))
|
|
eq_(len(real_result.cursor_strategy._rowbuffer), 1)
|
|
|
|
next(result)
|
|
next(result)
|
|
|
|
if optname == "yield_per" or optname == "yield_per_meth":
|
|
eq_(len(real_result.cursor_strategy._rowbuffer), value - 1)
|
|
else:
|
|
# based on default growth of 5
|
|
eq_(len(real_result.cursor_strategy._rowbuffer), 4)
|
|
|
|
for i, row in enumerate(result):
|
|
if i == 186:
|
|
break
|
|
|
|
if optname == "yield_per" or optname == "yield_per_meth":
|
|
eq_(
|
|
len(real_result.cursor_strategy._rowbuffer),
|
|
value - (188 % value),
|
|
)
|
|
else:
|
|
# based on default growth of 5
|
|
eq_(
|
|
len(real_result.cursor_strategy._rowbuffer),
|
|
7 if value == 10 else 42,
|
|
)
|
|
|
|
if optname == "yield_per" or optname == "yield_per_meth":
|
|
# ensure partition is set up to same size
|
|
partition = next(result.partitions())
|
|
eq_(len(partition), value)
|
|
|
|
|
|
class KeyTargetingTest(fixtures.TablesTest):
|
|
run_inserts = "once"
|
|
run_deletes = None
|
|
__backend__ = True
|
|
|
|
@classmethod
|
|
def define_tables(cls, metadata):
|
|
Table(
|
|
"keyed1",
|
|
metadata,
|
|
Column("a", CHAR(2), key="b"),
|
|
Column("c", CHAR(2), key="q"),
|
|
)
|
|
Table("keyed2", metadata, Column("a", CHAR(2)), Column("b", CHAR(2)))
|
|
Table("keyed3", metadata, Column("a", CHAR(2)), Column("d", CHAR(2)))
|
|
Table("keyed4", metadata, Column("b", CHAR(2)), Column("q", CHAR(2)))
|
|
Table("content", metadata, Column("t", String(30), key="type"))
|
|
Table("bar", metadata, Column("ctype", String(30), key="content_type"))
|
|
|
|
if testing.requires.schemas.enabled:
|
|
Table(
|
|
"wschema",
|
|
metadata,
|
|
Column("a", CHAR(2), key="b"),
|
|
Column("c", CHAR(2), key="q"),
|
|
schema=testing.config.test_schema,
|
|
)
|
|
|
|
Table(
|
|
"users",
|
|
metadata,
|
|
Column("id", Integer, primary_key=True),
|
|
Column("team_id", metadata, ForeignKey("teams.id")),
|
|
)
|
|
Table(
|
|
"teams",
|
|
metadata,
|
|
Column("id", Integer, primary_key=True),
|
|
)
|
|
|
|
@classmethod
|
|
def insert_data(cls, connection):
|
|
conn = connection
|
|
conn.execute(cls.tables.keyed1.insert(), dict(b="a1", q="c1"))
|
|
conn.execute(cls.tables.keyed2.insert(), dict(a="a2", b="b2"))
|
|
conn.execute(cls.tables.keyed3.insert(), dict(a="a3", d="d3"))
|
|
conn.execute(cls.tables.keyed4.insert(), dict(b="b4", q="q4"))
|
|
conn.execute(cls.tables.content.insert(), dict(type="t1"))
|
|
|
|
conn.execute(cls.tables.teams.insert(), dict(id=1))
|
|
conn.execute(cls.tables.users.insert(), dict(id=1, team_id=1))
|
|
|
|
if testing.requires.schemas.enabled:
|
|
conn.execute(
|
|
cls.tables["%s.wschema" % testing.config.test_schema].insert(),
|
|
dict(b="a1", q="c1"),
|
|
)
|
|
|
|
@testing.requires.schemas
|
|
def test_keyed_accessor_wschema(self, connection):
|
|
keyed1 = self.tables["%s.wschema" % testing.config.test_schema]
|
|
row = connection.execute(keyed1.select()).first()
|
|
|
|
eq_(row.b, "a1")
|
|
eq_(row.q, "c1")
|
|
eq_(row.a, "a1")
|
|
eq_(row.c, "c1")
|
|
|
|
def test_keyed_accessor_single(self, connection):
|
|
keyed1 = self.tables.keyed1
|
|
row = connection.execute(keyed1.select()).first()
|
|
|
|
eq_(row.b, "a1")
|
|
eq_(row.q, "c1")
|
|
eq_(row.a, "a1")
|
|
eq_(row.c, "c1")
|
|
|
|
def test_keyed_accessor_single_labeled(self, connection):
|
|
keyed1 = self.tables.keyed1
|
|
row = connection.execute(
|
|
keyed1.select().set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
).first()
|
|
|
|
eq_(row.keyed1_b, "a1")
|
|
eq_(row.keyed1_q, "c1")
|
|
eq_(row.keyed1_a, "a1")
|
|
eq_(row.keyed1_c, "c1")
|
|
|
|
def _test_keyed_targeting_no_label_at_all(self, expression, conn):
|
|
lt = literal_column("2")
|
|
stmt = select(literal_column("1"), expression, lt).select_from(
|
|
self.tables.keyed1
|
|
)
|
|
row = conn.execute(stmt).first()
|
|
|
|
eq_(row._mapping[expression], "a1")
|
|
eq_(row._mapping[lt], 2)
|
|
|
|
# Postgresql for example has the key as "?column?", which dupes
|
|
# easily. we get around that because we know that "2" is unique
|
|
eq_(row._mapping["2"], 2)
|
|
|
|
def test_keyed_targeting_no_label_at_all_one(self, connection):
|
|
class not_named_max(expression.ColumnElement):
|
|
name = "not_named_max"
|
|
inherit_cache = True
|
|
|
|
@compiles(not_named_max)
|
|
def visit_max(element, compiler, **kw):
|
|
# explicit add
|
|
if "add_to_result_map" in kw:
|
|
kw["add_to_result_map"](None, None, (element,), NULLTYPE)
|
|
return "max(a)"
|
|
|
|
# assert that there is no "AS max_" or any label of any kind.
|
|
eq_(str(select(not_named_max())), "SELECT max(a)")
|
|
|
|
nnm = not_named_max()
|
|
self._test_keyed_targeting_no_label_at_all(nnm, connection)
|
|
|
|
def test_keyed_targeting_no_label_at_all_two(self, connection):
|
|
class not_named_max(expression.ColumnElement):
|
|
name = "not_named_max"
|
|
inherit_cache = True
|
|
|
|
@compiles(not_named_max)
|
|
def visit_max(element, compiler, **kw):
|
|
# we don't add to keymap here; compiler should be doing it
|
|
return "max(a)"
|
|
|
|
# assert that there is no "AS max_" or any label of any kind.
|
|
eq_(str(select(not_named_max())), "SELECT max(a)")
|
|
|
|
nnm = not_named_max()
|
|
self._test_keyed_targeting_no_label_at_all(nnm, connection)
|
|
|
|
def test_keyed_targeting_no_label_at_all_text(self, connection):
|
|
t1 = text("max(a)")
|
|
t2 = text("min(a)")
|
|
|
|
stmt = select(t1, t2).select_from(self.tables.keyed1)
|
|
row = connection.execute(stmt).first()
|
|
|
|
eq_(row._mapping[t1], "a1")
|
|
eq_(row._mapping[t2], "a1")
|
|
|
|
@testing.requires.duplicate_names_in_cursor_description
|
|
def test_keyed_accessor_composite_conflict_2(self, connection):
|
|
keyed1 = self.tables.keyed1
|
|
keyed2 = self.tables.keyed2
|
|
|
|
row = connection.execute(
|
|
select(keyed1, keyed2)
|
|
.select_from(keyed1.join(keyed2, true()))
|
|
.set_label_style(LABEL_STYLE_NONE)
|
|
).first()
|
|
|
|
# column access is unambiguous
|
|
eq_(row._mapping[self.tables.keyed2.c.b], "b2")
|
|
|
|
# row.a is ambiguous
|
|
assert_raises_message(
|
|
exc.InvalidRequestError, "Ambig", getattr, row, "a"
|
|
)
|
|
|
|
# for "b" we have kind of a choice. the name "b" is not ambiguous in
|
|
# cursor.description in this case. It is however ambiguous as far as
|
|
# the objects we have queried against, because keyed1.c.a has key="b"
|
|
# and keyed1.c.b is "b". historically this was allowed as
|
|
# non-ambiguous, however the column it targets changes based on
|
|
# whether or not the dupe is present so it's ambiguous
|
|
# eq_(row.b, "b2")
|
|
assert_raises_message(
|
|
exc.InvalidRequestError, "Ambig", getattr, row, "b"
|
|
)
|
|
|
|
# illustrate why row.b above is ambiguous, and not "b2"; because
|
|
# if we didn't have keyed2, now it matches row.a. a new column
|
|
# shouldn't be able to grab the value from a previous column.
|
|
row = connection.execute(select(keyed1)).first()
|
|
eq_(row.b, "a1")
|
|
|
|
def test_keyed_accessor_composite_conflict_2_fix_w_uselabels(
|
|
self, connection
|
|
):
|
|
keyed1 = self.tables.keyed1
|
|
keyed2 = self.tables.keyed2
|
|
|
|
row = connection.execute(
|
|
select(keyed1, keyed2)
|
|
.select_from(keyed1.join(keyed2, true()))
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
).first()
|
|
|
|
# column access is unambiguous
|
|
eq_(row._mapping[self.tables.keyed2.c.b], "b2")
|
|
|
|
eq_(row._mapping["keyed2_b"], "b2")
|
|
eq_(row._mapping["keyed1_a"], "a1")
|
|
|
|
def test_keyed_accessor_composite_names_precedent(self, connection):
|
|
keyed1 = self.tables.keyed1
|
|
keyed4 = self.tables.keyed4
|
|
|
|
row = connection.execute(
|
|
select(keyed1, keyed4).select_from(keyed1.join(keyed4, true()))
|
|
).first()
|
|
eq_(row.b, "b4")
|
|
eq_(row.q, "q4")
|
|
eq_(row.a, "a1")
|
|
eq_(row.c, "c1")
|
|
|
|
@testing.requires.duplicate_names_in_cursor_description
|
|
def test_keyed_accessor_composite_keys_precedent(self, connection):
|
|
keyed1 = self.tables.keyed1
|
|
keyed3 = self.tables.keyed3
|
|
|
|
row = connection.execute(
|
|
select(keyed1, keyed3)
|
|
.select_from(keyed1.join(keyed3, true()))
|
|
.set_label_style(LABEL_STYLE_NONE)
|
|
).first()
|
|
eq_(row.q, "c1")
|
|
|
|
# prior to 1.4 #4887, this raised an "ambiguous column name 'a'""
|
|
# message, because "b" is linked to "a" which is a dupe. but we know
|
|
# where "b" is in the row by position.
|
|
eq_(row.b, "a1")
|
|
|
|
# "a" is of course ambiguous
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Ambiguous column name 'a'",
|
|
getattr,
|
|
row,
|
|
"a",
|
|
)
|
|
eq_(row.d, "d3")
|
|
|
|
def test_keyed_accessor_composite_labeled(self, connection):
|
|
keyed1 = self.tables.keyed1
|
|
keyed2 = self.tables.keyed2
|
|
|
|
row = connection.execute(
|
|
select(keyed1, keyed2)
|
|
.select_from(keyed1.join(keyed2, true()))
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
).first()
|
|
eq_(row.keyed1_b, "a1")
|
|
eq_(row.keyed1_a, "a1")
|
|
eq_(row.keyed1_q, "c1")
|
|
eq_(row.keyed1_c, "c1")
|
|
eq_(row.keyed2_a, "a2")
|
|
eq_(row.keyed2_b, "b2")
|
|
|
|
assert_raises(KeyError, lambda: row._mapping["keyed2_c"])
|
|
assert_raises(KeyError, lambda: row._mapping["keyed2_q"])
|
|
|
|
def test_keyed_accessor_column_is_repeated_multiple_times(
|
|
self, connection
|
|
):
|
|
# test new logic added as a result of the combination of #4892 and
|
|
# #4887. We allow duplicate columns, but we also have special logic
|
|
# to disambiguate for the same column repeated, and as #4887 adds
|
|
# stricter ambiguous result column logic, the compiler has to know to
|
|
# not add these dupe columns to the result map, else they register as
|
|
# ambiguous.
|
|
|
|
keyed2 = self.tables.keyed2
|
|
keyed3 = self.tables.keyed3
|
|
|
|
stmt = (
|
|
select(
|
|
keyed2.c.a,
|
|
keyed3.c.a,
|
|
keyed2.c.a,
|
|
keyed2.c.a,
|
|
keyed3.c.a,
|
|
keyed3.c.a,
|
|
keyed3.c.d,
|
|
keyed3.c.d,
|
|
)
|
|
.select_from(keyed2.join(keyed3, true()))
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
)
|
|
|
|
result = connection.execute(stmt)
|
|
|
|
# ensure the result map is the same number of cols so we can
|
|
# use positional targeting
|
|
eq_(
|
|
[rec[0] for rec in result.context.compiled._result_columns],
|
|
[
|
|
"keyed2_a",
|
|
"keyed3_a",
|
|
"keyed2_a__1",
|
|
"keyed2_a__2",
|
|
"keyed3_a__1",
|
|
"keyed3_a__2",
|
|
"keyed3_d",
|
|
"keyed3_d__1",
|
|
],
|
|
)
|
|
row = result.first()
|
|
|
|
# keyed access will ignore the dupe cols
|
|
eq_(row._mapping[keyed2.c.a], "a2")
|
|
eq_(row._mapping[keyed3.c.a], "a3")
|
|
eq_(result._getter(keyed3.c.a)(row), "a3")
|
|
eq_(row._mapping[keyed3.c.d], "d3")
|
|
|
|
# however we can get everything positionally
|
|
eq_(row, ("a2", "a3", "a2", "a2", "a3", "a3", "d3", "d3"))
|
|
eq_(row[0], "a2")
|
|
eq_(row[1], "a3")
|
|
eq_(row[2], "a2")
|
|
eq_(row[3], "a2")
|
|
eq_(row[4], "a3")
|
|
eq_(row[5], "a3")
|
|
eq_(row[6], "d3")
|
|
eq_(row[7], "d3")
|
|
|
|
def test_columnclause_schema_column_one(self, connection):
|
|
# originally addressed by [ticket:2932], however liberalized
|
|
# Column-targeting rules are deprecated
|
|
a, b = sql.column("a"), sql.column("b")
|
|
stmt = select(a, b).select_from(table("keyed2"))
|
|
row = connection.execute(stmt).first()
|
|
|
|
in_(a, row._mapping)
|
|
in_(b, row._mapping)
|
|
|
|
keyed2 = self.tables.keyed2
|
|
|
|
not_in(keyed2.c.a, row._mapping)
|
|
not_in(keyed2.c.b, row._mapping)
|
|
|
|
def test_columnclause_schema_column_two(self, connection):
|
|
keyed2 = self.tables.keyed2
|
|
|
|
stmt = select(keyed2.c.a, keyed2.c.b)
|
|
row = connection.execute(stmt).first()
|
|
|
|
in_(keyed2.c.a, row._mapping)
|
|
in_(keyed2.c.b, row._mapping)
|
|
|
|
# in 1.x, would warn for string match, but return a result
|
|
a, b = sql.column("a"), sql.column("b")
|
|
not_in(a, row._mapping)
|
|
not_in(b, row._mapping)
|
|
|
|
def test_columnclause_schema_column_three(self, connection):
|
|
# this is also addressed by [ticket:2932]
|
|
stmt = text("select a, b from keyed2").columns(a=CHAR, b=CHAR)
|
|
row = connection.execute(stmt).first()
|
|
|
|
in_(stmt.selected_columns.a, row._mapping)
|
|
in_(stmt.selected_columns.b, row._mapping)
|
|
|
|
keyed2 = self.tables.keyed2
|
|
a, b = sql.column("a"), sql.column("b")
|
|
|
|
# in 1.x, would warn for string match, but return a result
|
|
not_in(keyed2.c.a, row._mapping)
|
|
not_in(keyed2.c.b, row._mapping)
|
|
not_in(a, row._mapping)
|
|
not_in(b, row._mapping)
|
|
not_in(stmt.subquery().c.a, row._mapping)
|
|
not_in(stmt.subquery().c.b, row._mapping)
|
|
|
|
def test_columnclause_schema_column_four(self, connection):
|
|
# originally addressed by [ticket:2932], however liberalized
|
|
# Column-targeting rules are deprecated
|
|
|
|
a, b = sql.column("keyed2_a"), sql.column("keyed2_b")
|
|
stmt = text("select a AS keyed2_a, b AS keyed2_b from keyed2").columns(
|
|
a, b
|
|
)
|
|
row = connection.execute(stmt).first()
|
|
|
|
in_(a, row._mapping)
|
|
in_(b, row._mapping)
|
|
|
|
in_(stmt.selected_columns.keyed2_a, row._mapping)
|
|
in_(stmt.selected_columns.keyed2_b, row._mapping)
|
|
|
|
keyed2 = self.tables.keyed2
|
|
|
|
# in 1.x, would warn for string match, but return a result
|
|
not_in(keyed2.c.a, row._mapping)
|
|
not_in(keyed2.c.b, row._mapping)
|
|
not_in(stmt.subquery().c.keyed2_a, row._mapping)
|
|
not_in(stmt.subquery().c.keyed2_b, row._mapping)
|
|
|
|
def test_columnclause_schema_column_five(self, connection):
|
|
# this is also addressed by [ticket:2932]
|
|
|
|
stmt = text("select a AS keyed2_a, b AS keyed2_b from keyed2").columns(
|
|
keyed2_a=CHAR, keyed2_b=CHAR
|
|
)
|
|
row = connection.execute(stmt).first()
|
|
|
|
in_(stmt.selected_columns.keyed2_a, row._mapping)
|
|
in_(stmt.selected_columns.keyed2_b, row._mapping)
|
|
|
|
keyed2 = self.tables.keyed2
|
|
|
|
# in 1.x, would warn for string match, but return a result
|
|
not_in(keyed2.c.a, row._mapping)
|
|
not_in(keyed2.c.b, row._mapping)
|
|
not_in(stmt.subquery().c.keyed2_a, row._mapping)
|
|
not_in(stmt.subquery().c.keyed2_b, row._mapping)
|
|
|
|
def _adapt_result_columns_fixture_one(self):
|
|
keyed1 = self.tables.keyed1
|
|
stmt = (
|
|
select(keyed1.c.b, keyed1.c.q.label("foo"))
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
.subquery()
|
|
)
|
|
|
|
return select(stmt.c.keyed1_b, stmt.c.foo)
|
|
|
|
def _adapt_result_columns_fixture_two(self):
|
|
return text("select a AS keyed2_a, b AS keyed2_b from keyed2").columns(
|
|
column("keyed2_a", CHAR), column("keyed2_b", CHAR)
|
|
)
|
|
|
|
def _adapt_result_columns_fixture_three(self):
|
|
keyed1 = self.tables.keyed1
|
|
stmt = select(keyed1.c.b, keyed1.c.q.label("foo")).subquery()
|
|
|
|
return select(stmt.c.b, stmt.c.foo)
|
|
|
|
def _adapt_result_columns_fixture_four(self):
|
|
keyed1 = self.tables.keyed1
|
|
|
|
stmt1 = select(keyed1).set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
|
|
a1 = keyed1.alias()
|
|
stmt2 = ClauseAdapter(a1).traverse(stmt1)
|
|
|
|
return stmt2
|
|
|
|
def _adapt_result_columns_fixture_five(self):
|
|
users, teams = self.tables("users", "teams")
|
|
return select(users.c.id, teams.c.id).select_from(
|
|
users.outerjoin(teams)
|
|
)
|
|
|
|
def _adapt_result_columns_fixture_six(self):
|
|
# this has _result_columns structure that is not ordered
|
|
# the same as the cursor.description.
|
|
return text("select a AS keyed2_a, b AS keyed2_b from keyed2").columns(
|
|
keyed2_b=CHAR,
|
|
keyed2_a=CHAR,
|
|
)
|
|
|
|
def _adapt_result_columns_fixture_seven(self):
|
|
# this has _result_columns structure that is not ordered
|
|
# the same as the cursor.description.
|
|
return text("select a AS keyed2_a, b AS keyed2_b from keyed2").columns(
|
|
keyed2_b=CHAR, bogus_col=CHAR
|
|
)
|
|
|
|
@testing.combinations(
|
|
_adapt_result_columns_fixture_one,
|
|
_adapt_result_columns_fixture_two,
|
|
_adapt_result_columns_fixture_three,
|
|
_adapt_result_columns_fixture_four,
|
|
_adapt_result_columns_fixture_five,
|
|
_adapt_result_columns_fixture_six,
|
|
_adapt_result_columns_fixture_seven,
|
|
argnames="stmt_fn",
|
|
)
|
|
def test_adapt_result_columns(self, connection, stmt_fn):
|
|
"""test adaptation of a CursorResultMetadata to another one.
|
|
|
|
|
|
This copies the _keymap from one to the other in terms of the
|
|
selected columns of a target selectable.
|
|
|
|
This is used by the statement caching process to re-use the
|
|
CursorResultMetadata from the cached statement against the same
|
|
statement sent separately.
|
|
|
|
"""
|
|
|
|
stmt1 = stmt_fn(self)
|
|
stmt2 = stmt_fn(self)
|
|
|
|
eq_(stmt1._generate_cache_key(), stmt2._generate_cache_key())
|
|
|
|
column_linkage = dict(
|
|
zip(stmt1.selected_columns, stmt2.selected_columns)
|
|
)
|
|
|
|
for i in range(2):
|
|
try:
|
|
result = connection.execute(stmt1)
|
|
|
|
mock_context = Mock(
|
|
compiled=result.context.compiled, invoked_statement=stmt2
|
|
)
|
|
existing_metadata = result._metadata
|
|
adapted_metadata = existing_metadata._adapt_to_context(
|
|
mock_context
|
|
)
|
|
|
|
eq_(existing_metadata.keys, adapted_metadata.keys)
|
|
|
|
for k in existing_metadata._keymap:
|
|
if isinstance(k, ColumnElement) and k in column_linkage:
|
|
other_k = column_linkage[k]
|
|
else:
|
|
other_k = k
|
|
|
|
is_(
|
|
existing_metadata._keymap[k],
|
|
adapted_metadata._keymap[other_k],
|
|
)
|
|
finally:
|
|
result.close()
|
|
|
|
@testing.combinations(
|
|
_adapt_result_columns_fixture_one,
|
|
_adapt_result_columns_fixture_two,
|
|
_adapt_result_columns_fixture_three,
|
|
_adapt_result_columns_fixture_four,
|
|
_adapt_result_columns_fixture_five,
|
|
_adapt_result_columns_fixture_six,
|
|
_adapt_result_columns_fixture_seven,
|
|
argnames="stmt_fn",
|
|
)
|
|
def test_adapt_result_columns_from_cache(self, connection, stmt_fn):
|
|
stmt1 = stmt_fn(self)
|
|
stmt2 = stmt_fn(self)
|
|
|
|
cache = {}
|
|
result = connection.execute(
|
|
stmt1,
|
|
execution_options={"compiled_cache": cache},
|
|
)
|
|
result.close()
|
|
assert cache
|
|
|
|
result = connection.execute(
|
|
stmt2,
|
|
execution_options={"compiled_cache": cache},
|
|
)
|
|
|
|
row = result.first()
|
|
for col in stmt2.selected_columns:
|
|
if "bogus" in col.name:
|
|
assert col not in row._mapping
|
|
else:
|
|
assert col in row._mapping
|
|
|
|
|
|
class PositionalTextTest(fixtures.TablesTest):
|
|
run_inserts = "once"
|
|
run_deletes = None
|
|
__backend__ = True
|
|
|
|
@classmethod
|
|
def define_tables(cls, metadata):
|
|
Table(
|
|
"text1",
|
|
metadata,
|
|
Column("a", CHAR(2)),
|
|
Column("b", CHAR(2)),
|
|
Column("c", CHAR(2)),
|
|
Column("d", CHAR(2)),
|
|
)
|
|
|
|
@classmethod
|
|
def insert_data(cls, connection):
|
|
connection.execute(
|
|
cls.tables.text1.insert(),
|
|
[dict(a="a1", b="b1", c="c1", d="d1")],
|
|
)
|
|
|
|
def test_via_column(self, connection):
|
|
c1, c2, c3, c4 = column("q"), column("p"), column("r"), column("d")
|
|
stmt = text("select a, b, c, d from text1").columns(c1, c2, c3, c4)
|
|
|
|
result = connection.execute(stmt)
|
|
row = result.first()
|
|
|
|
eq_(row._mapping[c2], "b1")
|
|
eq_(row._mapping[c4], "d1")
|
|
eq_(row[1], "b1")
|
|
eq_(row._mapping["b"], "b1")
|
|
eq_(list(row._mapping.keys()), ["a", "b", "c", "d"])
|
|
eq_(row._fields, ("a", "b", "c", "d"))
|
|
eq_(row._mapping["r"], "c1")
|
|
eq_(row._mapping["d"], "d1")
|
|
|
|
def test_fewer_cols_than_sql_positional(self, connection):
|
|
c1, c2 = column("q"), column("p")
|
|
stmt = text("select a, b, c, d from text1").columns(c1, c2)
|
|
|
|
# no warning as this can be similar for non-positional
|
|
result = connection.execute(stmt)
|
|
row = result.first()
|
|
|
|
eq_(row._mapping[c1], "a1")
|
|
eq_(row._mapping["c"], "c1")
|
|
|
|
def test_fewer_cols_than_sql_non_positional(self, connection):
|
|
c1, c2 = column("a"), column("p")
|
|
stmt = text("select a, b, c, d from text1").columns(c2, c1, d=CHAR)
|
|
|
|
# no warning as this can be similar for non-positional
|
|
result = connection.execute(stmt)
|
|
row = result.first()
|
|
|
|
# c1 name matches, locates
|
|
eq_(row._mapping[c1], "a1")
|
|
eq_(row._mapping["c"], "c1")
|
|
|
|
# c2 name does not match, doesn't locate
|
|
assert_raises_message(
|
|
exc.NoSuchColumnError,
|
|
"in row for column 'p'",
|
|
lambda: row._mapping[c2],
|
|
)
|
|
|
|
def test_more_cols_than_sql_positional(self, connection):
|
|
c1, c2, c3, c4 = column("q"), column("p"), column("r"), column("d")
|
|
stmt = text("select a, b from text1").columns(c1, c2, c3, c4)
|
|
|
|
with assertions.expect_warnings(
|
|
r"Number of columns in textual SQL \(4\) is "
|
|
r"smaller than number of columns requested \(2\)"
|
|
):
|
|
result = connection.execute(stmt)
|
|
|
|
row = result.first()
|
|
eq_(row._mapping[c2], "b1")
|
|
|
|
assert_raises_message(
|
|
exc.NoSuchColumnError,
|
|
"in row for column 'r'",
|
|
lambda: row._mapping[c3],
|
|
)
|
|
|
|
def test_more_cols_than_sql_nonpositional(self, connection):
|
|
c1, c2, c3, c4 = column("b"), column("a"), column("r"), column("d")
|
|
stmt = TextualSelect(
|
|
text("select a, b from text1"), [c1, c2, c3, c4], positional=False
|
|
)
|
|
|
|
# no warning for non-positional
|
|
result = connection.execute(stmt)
|
|
|
|
row = result.first()
|
|
eq_(row._mapping[c1], "b1")
|
|
eq_(row._mapping[c2], "a1")
|
|
|
|
assert_raises_message(
|
|
exc.NoSuchColumnError,
|
|
"in row for column 'r'",
|
|
lambda: row._mapping[c3],
|
|
)
|
|
|
|
def test_more_cols_than_sql_nonpositional_labeled_cols(self, connection):
|
|
text1 = self.tables.text1
|
|
c1, c2, c3, c4 = text1.c.b, text1.c.a, column("r"), column("d")
|
|
|
|
# the compiler will enable loose matching for this statement
|
|
# so that column._label is taken into account
|
|
stmt = TextualSelect(
|
|
text("select a, b AS text1_b from text1"),
|
|
[c1, c2, c3, c4],
|
|
positional=False,
|
|
)
|
|
|
|
# no warning for non-positional
|
|
result = connection.execute(stmt)
|
|
|
|
row = result.first()
|
|
eq_(row._mapping[c1], "b1")
|
|
eq_(row._mapping[c2], "a1")
|
|
|
|
assert_raises_message(
|
|
exc.NoSuchColumnError,
|
|
"in row for column 'r'",
|
|
lambda: row._mapping[c3],
|
|
)
|
|
|
|
def test_dupe_col_obj(self, connection):
|
|
c1, c2, c3 = column("q"), column("p"), column("r")
|
|
stmt = text("select a, b, c, d from text1").columns(c1, c2, c3, c2)
|
|
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Duplicate column expression requested in "
|
|
"textual SQL: <.*.ColumnClause.*; p>",
|
|
connection.execute,
|
|
stmt,
|
|
)
|
|
|
|
def test_anon_aliased_unique(self, connection):
|
|
text1 = self.tables.text1
|
|
|
|
c1 = text1.c.a.label(None)
|
|
c2 = text1.alias().c.c
|
|
c3 = text1.alias().c.b
|
|
c4 = text1.alias().c.d.label(None)
|
|
|
|
stmt = text("select a, b, c, d from text1").columns(c1, c2, c3, c4)
|
|
result = connection.execute(stmt)
|
|
row = result.first()
|
|
|
|
eq_(row._mapping[c1], "a1")
|
|
eq_(row._mapping[c2], "b1")
|
|
eq_(row._mapping[c3], "c1")
|
|
eq_(row._mapping[c4], "d1")
|
|
|
|
# in 1.x, would warn for string match, but return a result
|
|
assert_raises_message(
|
|
exc.NoSuchColumnError,
|
|
"Could not locate column in row for column 'text1.a'",
|
|
lambda: row._mapping[text1.c.a],
|
|
)
|
|
|
|
# in 1.x, would warn for string match, but return a result
|
|
assert_raises_message(
|
|
exc.NoSuchColumnError,
|
|
"Could not locate column in row for column 'text1.d'",
|
|
lambda: row._mapping[text1.c.d],
|
|
)
|
|
|
|
# text1.c.b goes nowhere....because we hit key fallback
|
|
# but the text1.c.b doesn't derive from text1.c.c
|
|
assert_raises_message(
|
|
exc.NoSuchColumnError,
|
|
"Could not locate column in row for column 'text1.b'",
|
|
lambda: row._mapping[text1.c.b],
|
|
)
|
|
|
|
def test_anon_aliased_overlapping(self, connection):
|
|
text1 = self.tables.text1
|
|
|
|
c1 = text1.c.a.label(None)
|
|
c2 = text1.alias().c.a
|
|
c3 = text1.alias().c.a.label(None)
|
|
c4 = text1.c.a.label(None)
|
|
|
|
stmt = text("select a, b, c, d from text1").columns(c1, c2, c3, c4)
|
|
result = connection.execute(stmt)
|
|
row = result.first()
|
|
|
|
eq_(row._mapping[c1], "a1")
|
|
eq_(row._mapping[c2], "b1")
|
|
eq_(row._mapping[c3], "c1")
|
|
eq_(row._mapping[c4], "d1")
|
|
|
|
# in 1.x, would warn for string match, but return a result
|
|
assert_raises_message(
|
|
exc.NoSuchColumnError,
|
|
"Could not locate column in row for column 'text1.a'",
|
|
lambda: row._mapping[text1.c.a],
|
|
)
|
|
|
|
def test_anon_aliased_name_conflict(self, connection):
|
|
text1 = self.tables.text1
|
|
|
|
c1 = text1.c.a.label("a")
|
|
c2 = text1.alias().c.a
|
|
c3 = text1.alias().c.a.label("a")
|
|
c4 = text1.c.a.label("a")
|
|
|
|
# all cols are named "a". if we are positional, we don't care.
|
|
# this is new logic in 1.1
|
|
stmt = text("select a, b as a, c as a, d as a from text1").columns(
|
|
c1, c2, c3, c4
|
|
)
|
|
result = connection.execute(stmt)
|
|
row = result.first()
|
|
|
|
eq_(row._mapping[c1], "a1")
|
|
eq_(row._mapping[c2], "b1")
|
|
eq_(row._mapping[c3], "c1")
|
|
eq_(row._mapping[c4], "d1")
|
|
|
|
# fails, because we hit key fallback and find conflicts
|
|
# in columns that are presnet
|
|
assert_raises_message(
|
|
exc.NoSuchColumnError,
|
|
"Could not locate column in row for column 'text1.a'",
|
|
lambda: row._mapping[text1.c.a],
|
|
)
|
|
|
|
|
|
class AlternateCursorResultTest(fixtures.TablesTest):
|
|
__requires__ = ("sqlite",)
|
|
|
|
@classmethod
|
|
def setup_bind(cls):
|
|
cls.engine = engine = engines.testing_engine(
|
|
"sqlite://", options={"scope": "class"}
|
|
)
|
|
return engine
|
|
|
|
@classmethod
|
|
def define_tables(cls, metadata):
|
|
Table(
|
|
"test",
|
|
metadata,
|
|
Column("x", Integer, primary_key=True),
|
|
Column("y", String(50)),
|
|
)
|
|
|
|
@classmethod
|
|
def insert_data(cls, connection):
|
|
connection.execute(
|
|
cls.tables.test.insert(),
|
|
[{"x": i, "y": "t_%d" % i} for i in range(1, 12)],
|
|
)
|
|
|
|
@contextmanager
|
|
def _proxy_fixture(self, cls):
|
|
self.table = self.tables.test
|
|
|
|
class ExcCtx(default.DefaultExecutionContext):
|
|
def post_exec(self):
|
|
if cls is _cursor.CursorFetchStrategy:
|
|
pass
|
|
elif cls is _cursor.BufferedRowCursorFetchStrategy:
|
|
self.cursor_fetch_strategy = cls(
|
|
self.cursor, self.execution_options
|
|
)
|
|
elif cls is _cursor.FullyBufferedCursorFetchStrategy:
|
|
self.cursor_fetch_strategy = cls(
|
|
self.cursor,
|
|
self.cursor.description,
|
|
self.cursor.fetchall(),
|
|
)
|
|
else:
|
|
assert False
|
|
|
|
self.patcher = patch.object(
|
|
self.engine.dialect, "execution_ctx_cls", ExcCtx
|
|
)
|
|
|
|
with self.patcher:
|
|
yield
|
|
|
|
def _test_proxy(self, cls):
|
|
with self._proxy_fixture(cls):
|
|
rows = []
|
|
with self.engine.connect() as conn:
|
|
r = conn.execute(select(self.table))
|
|
assert isinstance(r.cursor_strategy, cls)
|
|
for i in range(5):
|
|
rows.append(r.fetchone())
|
|
eq_(rows, [(i, "t_%d" % i) for i in range(1, 6)])
|
|
|
|
rows = r.fetchmany(3)
|
|
eq_(rows, [(i, "t_%d" % i) for i in range(6, 9)])
|
|
|
|
rows = r.fetchall()
|
|
eq_(rows, [(i, "t_%d" % i) for i in range(9, 12)])
|
|
|
|
r = conn.execute(select(self.table))
|
|
rows = r.fetchmany(None)
|
|
eq_(rows[0], (1, "t_1"))
|
|
# number of rows here could be one, or the whole thing
|
|
assert len(rows) == 1 or len(rows) == 11
|
|
|
|
r = conn.execute(select(self.table).limit(1))
|
|
r.fetchone()
|
|
eq_(r.fetchone(), None)
|
|
|
|
r = conn.execute(select(self.table).limit(5))
|
|
rows = r.fetchmany(6)
|
|
eq_(rows, [(i, "t_%d" % i) for i in range(1, 6)])
|
|
|
|
# result keeps going just fine with blank results...
|
|
eq_(r.fetchmany(2), [])
|
|
|
|
eq_(r.fetchmany(2), [])
|
|
|
|
eq_(r.fetchall(), [])
|
|
|
|
eq_(r.fetchone(), None)
|
|
|
|
# until we close
|
|
r.close()
|
|
|
|
self._assert_result_closed(r)
|
|
|
|
r = conn.execute(select(self.table).limit(5))
|
|
eq_(r.first(), (1, "t_1"))
|
|
self._assert_result_closed(r)
|
|
|
|
r = conn.execute(select(self.table).limit(5))
|
|
eq_(r.scalar(), 1)
|
|
self._assert_result_closed(r)
|
|
|
|
def _assert_result_closed(self, r):
|
|
assert_raises_message(
|
|
sa_exc.ResourceClosedError, "object is closed", r.fetchone
|
|
)
|
|
|
|
assert_raises_message(
|
|
sa_exc.ResourceClosedError, "object is closed", r.fetchmany, 2
|
|
)
|
|
|
|
assert_raises_message(
|
|
sa_exc.ResourceClosedError, "object is closed", r.fetchall
|
|
)
|
|
|
|
def test_basic_plain(self):
|
|
self._test_proxy(_cursor.CursorFetchStrategy)
|
|
|
|
def test_basic_buffered_row_result_proxy(self):
|
|
self._test_proxy(_cursor.BufferedRowCursorFetchStrategy)
|
|
|
|
def test_basic_fully_buffered_result_proxy(self):
|
|
self._test_proxy(_cursor.FullyBufferedCursorFetchStrategy)
|
|
|
|
def test_basic_buffered_column_result_proxy(self):
|
|
self._test_proxy(_cursor.CursorFetchStrategy)
|
|
|
|
def test_resultprocessor_plain(self):
|
|
self._test_result_processor(_cursor.CursorFetchStrategy, False)
|
|
|
|
def test_resultprocessor_plain_cached(self):
|
|
self._test_result_processor(_cursor.CursorFetchStrategy, True)
|
|
|
|
def test_resultprocessor_buffered_row(self):
|
|
self._test_result_processor(
|
|
_cursor.BufferedRowCursorFetchStrategy, False
|
|
)
|
|
|
|
def test_resultprocessor_buffered_row_cached(self):
|
|
self._test_result_processor(
|
|
_cursor.BufferedRowCursorFetchStrategy, True
|
|
)
|
|
|
|
def test_resultprocessor_fully_buffered(self):
|
|
self._test_result_processor(
|
|
_cursor.FullyBufferedCursorFetchStrategy, False
|
|
)
|
|
|
|
def test_resultprocessor_fully_buffered_cached(self):
|
|
self._test_result_processor(
|
|
_cursor.FullyBufferedCursorFetchStrategy, True
|
|
)
|
|
|
|
def _test_result_processor(self, cls, use_cache):
|
|
class MyType(TypeDecorator):
|
|
impl = String()
|
|
cache_ok = True
|
|
|
|
def process_result_value(self, value, dialect):
|
|
return "HI " + value
|
|
|
|
with self._proxy_fixture(cls):
|
|
with self.engine.connect() as conn:
|
|
if use_cache:
|
|
cache = {}
|
|
conn = conn.execution_options(compiled_cache=cache)
|
|
|
|
stmt = select(literal("THERE", type_=MyType()))
|
|
for i in range(2):
|
|
r = conn.execute(stmt)
|
|
eq_(r.scalar(), "HI THERE")
|
|
|
|
@testing.fixture
|
|
def row_growth_fixture(self):
|
|
with self._proxy_fixture(_cursor.BufferedRowCursorFetchStrategy):
|
|
with self.engine.begin() as conn:
|
|
conn.execute(
|
|
self.table.insert(),
|
|
[{"x": i, "y": "t_%d" % i} for i in range(15, 3000)],
|
|
)
|
|
yield conn
|
|
|
|
@testing.combinations(
|
|
("no option", None, {0: 5, 1: 25, 9: 125, 135: 625, 274: 1000}),
|
|
("lt 1000", 27, {0: 5, 16: 27, 70: 27, 150: 27, 250: 27}),
|
|
(
|
|
"gt 1000",
|
|
1500,
|
|
{0: 5, 1: 25, 9: 125, 135: 625, 274: 1500, 1351: 1500},
|
|
),
|
|
(
|
|
"gt 1500",
|
|
2000,
|
|
{0: 5, 1: 25, 9: 125, 135: 625, 274: 2000, 1351: 2000},
|
|
),
|
|
id_="iaa",
|
|
argnames="max_row_buffer,checks",
|
|
)
|
|
def test_buffered_row_growth(
|
|
self, row_growth_fixture, max_row_buffer, checks
|
|
):
|
|
if max_row_buffer:
|
|
result = row_growth_fixture.execution_options(
|
|
max_row_buffer=max_row_buffer
|
|
).execute(self.table.select())
|
|
else:
|
|
result = row_growth_fixture.execute(self.table.select())
|
|
|
|
assertion = {}
|
|
max_size = max(checks.values())
|
|
for idx, row in enumerate(result, 0):
|
|
if idx in checks:
|
|
assertion[idx] = result.cursor_strategy._bufsize
|
|
le_(len(result.cursor_strategy._rowbuffer), max_size)
|
|
|
|
def test_buffered_fetchmany_fixed(self, row_growth_fixture):
|
|
"""The BufferedRow cursor strategy will defer to the fetchmany
|
|
size passed when given rather than using the buffer growth
|
|
heuristic.
|
|
|
|
"""
|
|
result = row_growth_fixture.execute(self.table.select())
|
|
eq_(len(result.cursor_strategy._rowbuffer), 1)
|
|
|
|
rows = result.fetchmany(300)
|
|
eq_(len(rows), 300)
|
|
eq_(len(result.cursor_strategy._rowbuffer), 0)
|
|
|
|
rows = result.fetchmany(300)
|
|
eq_(len(rows), 300)
|
|
eq_(len(result.cursor_strategy._rowbuffer), 0)
|
|
|
|
bufsize = result.cursor_strategy._bufsize
|
|
result.fetchone()
|
|
|
|
# the fetchone() caused it to buffer a full set of rows
|
|
eq_(len(result.cursor_strategy._rowbuffer), bufsize - 1)
|
|
|
|
# assert partitions uses fetchmany(), therefore controlling
|
|
# how the buffer is used
|
|
lens = []
|
|
for partition in result.partitions(180):
|
|
lens.append(len(partition))
|
|
eq_(len(result.cursor_strategy._rowbuffer), 0)
|
|
|
|
for lp in lens[0:-1]:
|
|
eq_(lp, 180)
|
|
|
|
def test_buffered_fetchmany_yield_per(self, connection):
|
|
table = self.tables.test
|
|
|
|
connection.execute(
|
|
table.insert(),
|
|
[{"x": i, "y": "t_%d" % i} for i in range(15, 3000)],
|
|
)
|
|
|
|
result = connection.execute(table.select())
|
|
assert isinstance(result.cursor_strategy, _cursor.CursorFetchStrategy)
|
|
|
|
result.fetchmany(5)
|
|
|
|
result = result.yield_per(100)
|
|
assert isinstance(
|
|
result.cursor_strategy, _cursor.BufferedRowCursorFetchStrategy
|
|
)
|
|
eq_(result.cursor_strategy._bufsize, 100)
|
|
eq_(result.cursor_strategy._growth_factor, 0)
|
|
eq_(len(result.cursor_strategy._rowbuffer), 0)
|
|
|
|
result.fetchone()
|
|
eq_(len(result.cursor_strategy._rowbuffer), 99)
|
|
|
|
for i, row in enumerate(result):
|
|
if i == 188:
|
|
break
|
|
|
|
# buffer of 98, plus buffer of 99 - 89, 10 rows
|
|
eq_(len(result.cursor_strategy._rowbuffer), 10)
|
|
|
|
@testing.combinations(True, False, argnames="close_on_init")
|
|
@testing.combinations(
|
|
"fetchone", "fetchmany", "fetchall", argnames="fetch_style"
|
|
)
|
|
def test_buffered_fetch_auto_soft_close(
|
|
self, connection, close_on_init, fetch_style
|
|
):
|
|
"""test #7274"""
|
|
|
|
table = self.tables.test
|
|
|
|
connection.execute(
|
|
table.insert(),
|
|
[{"x": i, "y": "t_%d" % i} for i in range(15, 30)],
|
|
)
|
|
|
|
result = connection.execute(table.select().limit(15))
|
|
assert isinstance(result.cursor_strategy, _cursor.CursorFetchStrategy)
|
|
|
|
if close_on_init:
|
|
# close_on_init - the initial buffering will exhaust the cursor,
|
|
# should soft close immediately
|
|
result = result.yield_per(30)
|
|
else:
|
|
# not close_on_init - soft close will occur after fetching an
|
|
# empty buffer
|
|
result = result.yield_per(5)
|
|
assert isinstance(
|
|
result.cursor_strategy, _cursor.BufferedRowCursorFetchStrategy
|
|
)
|
|
|
|
with mock.patch.object(result, "_soft_close") as soft_close:
|
|
if fetch_style == "fetchone":
|
|
while True:
|
|
row = result.fetchone()
|
|
|
|
if row:
|
|
eq_(soft_close.mock_calls, [])
|
|
else:
|
|
# fetchone() is also used by first(), scalar()
|
|
# and one() which want to embed a hard close in one
|
|
# step
|
|
eq_(soft_close.mock_calls, [mock.call(hard=False)])
|
|
break
|
|
elif fetch_style == "fetchmany":
|
|
while True:
|
|
rows = result.fetchmany(5)
|
|
|
|
if rows:
|
|
eq_(soft_close.mock_calls, [])
|
|
else:
|
|
eq_(soft_close.mock_calls, [mock.call()])
|
|
break
|
|
elif fetch_style == "fetchall":
|
|
rows = result.fetchall()
|
|
|
|
eq_(soft_close.mock_calls, [mock.call()])
|
|
else:
|
|
assert False
|
|
|
|
result.close()
|
|
|
|
def test_buffered_fetchmany_yield_per_all(self, connection):
|
|
table = self.tables.test
|
|
|
|
connection.execute(
|
|
table.insert(),
|
|
[{"x": i, "y": "t_%d" % i} for i in range(15, 500)],
|
|
)
|
|
|
|
result = connection.execute(table.select())
|
|
assert isinstance(result.cursor_strategy, _cursor.CursorFetchStrategy)
|
|
|
|
result.fetchmany(5)
|
|
|
|
result = result.yield_per(0)
|
|
assert isinstance(
|
|
result.cursor_strategy, _cursor.BufferedRowCursorFetchStrategy
|
|
)
|
|
eq_(result.cursor_strategy._bufsize, 0)
|
|
eq_(result.cursor_strategy._growth_factor, 0)
|
|
eq_(len(result.cursor_strategy._rowbuffer), 0)
|
|
|
|
result.fetchone()
|
|
eq_(len(result.cursor_strategy._rowbuffer), 490)
|
|
|
|
for i, row in enumerate(result):
|
|
if i == 188:
|
|
break
|
|
|
|
eq_(len(result.cursor_strategy._rowbuffer), 301)
|
|
|
|
# already buffered, so this doesn't change things
|
|
result.yield_per(10)
|
|
|
|
result.fetchmany(5)
|
|
eq_(len(result.cursor_strategy._rowbuffer), 296)
|
|
|
|
self._test_result_processor(
|
|
_cursor.BufferedRowCursorFetchStrategy, False
|
|
)
|
|
|
|
@testing.combinations(
|
|
_cursor.CursorFetchStrategy,
|
|
_cursor.BufferedRowCursorFetchStrategy,
|
|
# does not handle error in fetch
|
|
# _cursor.FullyBufferedCursorFetchStrategy,
|
|
argnames="strategy_cls",
|
|
)
|
|
@testing.combinations(
|
|
"fetchone",
|
|
"fetchmany",
|
|
"fetchmany_w_num",
|
|
"fetchall",
|
|
argnames="method_name",
|
|
)
|
|
def test_handle_error_in_fetch(self, strategy_cls, method_name):
|
|
class cursor:
|
|
def raise_(self):
|
|
raise IOError("random non-DBAPI error during cursor operation")
|
|
|
|
def fetchone(self):
|
|
self.raise_()
|
|
|
|
def fetchmany(self, num=None):
|
|
self.raise_()
|
|
|
|
def fetchall(self):
|
|
self.raise_()
|
|
|
|
def close(self):
|
|
self.raise_()
|
|
|
|
with self._proxy_fixture(strategy_cls):
|
|
with self.engine.connect() as conn:
|
|
r = conn.execute(select(self.table))
|
|
assert isinstance(r.cursor_strategy, strategy_cls)
|
|
with mock.patch.object(r, "cursor", cursor()):
|
|
|
|
with testing.expect_raises_message(
|
|
IOError, "random non-DBAPI"
|
|
):
|
|
if method_name == "fetchmany_w_num":
|
|
r.fetchmany(10)
|
|
else:
|
|
getattr(r, method_name)()
|
|
getattr(r, method_name)()
|
|
|
|
r.close()
|
|
|
|
def test_buffered_row_close_error_during_fetchone(self):
|
|
def raise_(**kw):
|
|
raise IOError("random non-DBAPI error during cursor operation")
|
|
|
|
with self._proxy_fixture(_cursor.BufferedRowCursorFetchStrategy):
|
|
with self.engine.connect() as conn:
|
|
r = conn.execute(select(self.table).limit(1))
|
|
|
|
r.fetchone()
|
|
with mock.patch.object(
|
|
r, "_soft_close", raise_
|
|
), testing.expect_raises_message(IOError, "random non-DBAPI"):
|
|
r.first()
|
|
r.close()
|
|
|
|
|
|
class MergeCursorResultTest(fixtures.TablesTest):
|
|
__backend__ = True
|
|
|
|
__requires__ = ("independent_cursors",)
|
|
|
|
@classmethod
|
|
def define_tables(cls, metadata):
|
|
Table(
|
|
"users",
|
|
metadata,
|
|
Column("user_id", INT, primary_key=True, autoincrement=False),
|
|
Column("user_name", VARCHAR(20)),
|
|
test_needs_acid=True,
|
|
)
|
|
|
|
@classmethod
|
|
def insert_data(cls, connection):
|
|
users = cls.tables.users
|
|
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
{"user_id": 7, "user_name": "u1"},
|
|
{"user_id": 8, "user_name": "u2"},
|
|
{"user_id": 9, "user_name": "u3"},
|
|
{"user_id": 10, "user_name": "u4"},
|
|
{"user_id": 11, "user_name": "u5"},
|
|
{"user_id": 12, "user_name": "u6"},
|
|
],
|
|
)
|
|
|
|
@testing.fixture
|
|
def merge_fixture(self):
|
|
users = self.tables.users
|
|
|
|
def results(connection):
|
|
|
|
r1 = connection.execute(
|
|
users.select()
|
|
.where(users.c.user_id.in_([7, 8]))
|
|
.order_by(users.c.user_id)
|
|
)
|
|
r2 = connection.execute(
|
|
users.select()
|
|
.where(users.c.user_id.in_([9]))
|
|
.order_by(users.c.user_id)
|
|
)
|
|
r3 = connection.execute(
|
|
users.select()
|
|
.where(users.c.user_id.in_([10, 11]))
|
|
.order_by(users.c.user_id)
|
|
)
|
|
r4 = connection.execute(
|
|
users.select()
|
|
.where(users.c.user_id.in_([12]))
|
|
.order_by(users.c.user_id)
|
|
)
|
|
return r1, r2, r3, r4
|
|
|
|
return results
|
|
|
|
def test_merge_results(self, connection, merge_fixture):
|
|
r1, r2, r3, r4 = merge_fixture(connection)
|
|
|
|
result = r1.merge(r2, r3, r4)
|
|
|
|
eq_(result.keys(), ["user_id", "user_name"])
|
|
row = result.fetchone()
|
|
eq_(row, (7, "u1"))
|
|
result.close()
|
|
|
|
def test_close(self, connection, merge_fixture):
|
|
r1, r2, r3, r4 = merge_fixture(connection)
|
|
|
|
result = r1.merge(r2, r3, r4)
|
|
|
|
for r in [result, r1, r2, r3, r4]:
|
|
assert not r.closed
|
|
|
|
result.close()
|
|
for r in [result, r1, r2, r3, r4]:
|
|
assert r.closed
|
|
|
|
def test_fetchall(self, connection, merge_fixture):
|
|
r1, r2, r3, r4 = merge_fixture(connection)
|
|
|
|
result = r1.merge(r2, r3, r4)
|
|
eq_(
|
|
result.fetchall(),
|
|
[
|
|
(7, "u1"),
|
|
(8, "u2"),
|
|
(9, "u3"),
|
|
(10, "u4"),
|
|
(11, "u5"),
|
|
(12, "u6"),
|
|
],
|
|
)
|
|
for r in [r1, r2, r3, r4]:
|
|
assert r._soft_closed
|
|
|
|
def test_first(self, connection, merge_fixture):
|
|
r1, r2, r3, r4 = merge_fixture(connection)
|
|
|
|
result = r1.merge(r2, r3, r4)
|
|
eq_(
|
|
result.first(),
|
|
(7, "u1"),
|
|
)
|
|
for r in [r1, r2, r3, r4]:
|
|
assert r.closed
|
|
|
|
def test_columns(self, connection, merge_fixture):
|
|
r1, r2, r3, r4 = merge_fixture(connection)
|
|
|
|
result = r1.merge(r2, r3, r4)
|
|
eq_(
|
|
result.columns("user_name").fetchmany(4),
|
|
[("u1",), ("u2",), ("u3",), ("u4",)],
|
|
)
|
|
result.close()
|
|
|
|
|
|
class GenerativeResultTest(fixtures.TablesTest):
|
|
__backend__ = True
|
|
|
|
@classmethod
|
|
def define_tables(cls, metadata):
|
|
Table(
|
|
"users",
|
|
metadata,
|
|
Column("user_id", INT, primary_key=True, autoincrement=False),
|
|
Column("user_name", VARCHAR(20)),
|
|
Column("x", Integer),
|
|
Column("y", Integer),
|
|
test_needs_acid=True,
|
|
)
|
|
Table(
|
|
"users_autoinc",
|
|
metadata,
|
|
Column(
|
|
"user_id", INT, primary_key=True, test_needs_autoincrement=True
|
|
),
|
|
Column("user_name", VARCHAR(20)),
|
|
test_needs_acid=True,
|
|
)
|
|
|
|
def test_fetchall(self, connection):
|
|
users = self.tables.users
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
{"user_id": 7, "user_name": "jack", "x": 1, "y": 2},
|
|
{"user_id": 8, "user_name": "ed", "x": 2, "y": 3},
|
|
{"user_id": 9, "user_name": "fred", "x": 15, "y": 20},
|
|
],
|
|
)
|
|
|
|
result = connection.execute(select(users).order_by(users.c.user_id))
|
|
eq_(
|
|
result.all(),
|
|
[(7, "jack", 1, 2), (8, "ed", 2, 3), (9, "fred", 15, 20)],
|
|
)
|
|
|
|
@testing.combinations(
|
|
((1, 0), [("jack", 7), ("ed", 8), ("fred", 9)]),
|
|
((3,), [(2,), (3,), (20,)]),
|
|
((-2, -1), [(1, 2), (2, 3), (15, 20)]),
|
|
argnames="columns, expected",
|
|
)
|
|
def test_columns(self, connection, columns, expected):
|
|
users = self.tables.users
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
{"user_id": 7, "user_name": "jack", "x": 1, "y": 2},
|
|
{"user_id": 8, "user_name": "ed", "x": 2, "y": 3},
|
|
{"user_id": 9, "user_name": "fred", "x": 15, "y": 20},
|
|
],
|
|
)
|
|
|
|
result = connection.execute(select(users).order_by(users.c.user_id))
|
|
|
|
all_ = result.columns(*columns).all()
|
|
eq_(all_, expected)
|
|
|
|
assert type(all_[0]) is Row
|
|
|
|
def test_columns_twice(self, connection):
|
|
users = self.tables.users
|
|
connection.execute(
|
|
users.insert(),
|
|
[{"user_id": 7, "user_name": "jack", "x": 1, "y": 2}],
|
|
)
|
|
|
|
result = connection.execute(select(users).order_by(users.c.user_id))
|
|
|
|
all_ = (
|
|
result.columns("x", "y", "user_name", "user_id")
|
|
.columns("user_name", "x")
|
|
.all()
|
|
)
|
|
eq_(all_, [("jack", 1)])
|
|
|
|
assert type(all_[0]) is Row
|
|
|
|
def test_columns_plus_getter(self, connection):
|
|
users = self.tables.users
|
|
connection.execute(
|
|
users.insert(),
|
|
[{"user_id": 7, "user_name": "jack", "x": 1, "y": 2}],
|
|
)
|
|
|
|
result = connection.execute(select(users).order_by(users.c.user_id))
|
|
|
|
result = result.columns("x", "y", "user_name")
|
|
getter = result._metadata._getter("y")
|
|
|
|
eq_(getter(result.first()), 2)
|
|
|
|
def test_partitions(self, connection):
|
|
users = self.tables.users
|
|
connection.execute(
|
|
users.insert(),
|
|
[
|
|
{
|
|
"user_id": i,
|
|
"user_name": "user %s" % i,
|
|
"x": i * 5,
|
|
"y": i * 20,
|
|
}
|
|
for i in range(500)
|
|
],
|
|
)
|
|
|
|
result = connection.execute(select(users).order_by(users.c.user_id))
|
|
|
|
start = 0
|
|
for partition in result.columns(0, 1).partitions(20):
|
|
eq_(
|
|
partition,
|
|
[(i, "user %s" % i) for i in range(start, start + 20)],
|
|
)
|
|
start += 20
|
|
|
|
assert result._soft_closed
|