mirror of
https://github.com/sqlalchemy/sqlalchemy.git
synced 2026-05-06 17:01:07 -04:00
459ebc668a
Added new generalized aggregate function ordering to functions via the :func:`_functions.FunctionElement.aggregate_order_by` method, which receives an expression and generates the appropriate embedded "ORDER BY" or "WITHIN GROUP (ORDER BY)" phrase depending on backend database. This new function supersedes the use of the PostgreSQL :func:`_postgresql.aggregate_order_by` function, which remains present for backward compatibility. To complement the new parameter, the :paramref:`_functions.aggregate_strings.order_by` which adds ORDER BY capability to the :class:`_functions.aggregate_strings` dialect-agnostic function which works for all included backends. Thanks much to Reuven Starodubski with help on this patch. Co-authored-by: Mike Bayer <mike_mp@zzzcomputing.com> Fixes: #12853 Closes: #12856 Pull-request: https://github.com/sqlalchemy/sqlalchemy/pull/12856 Pull-request-sha: d93fb591751227eb1f96052ea3ad449f511f70b3 Change-Id: I8eb41ff2d57695963a358b5f0017ca9372f15f70
8248 lines
282 KiB
Python
8248 lines
282 KiB
Python
"""
|
|
compiler tests.
|
|
|
|
These tests are among the very first that were written when SQLAlchemy
|
|
began in 2005. As a result the testing style here is very dense;
|
|
it's an ongoing job to break these into much smaller tests with correct pep8
|
|
styling and coherent test organization.
|
|
|
|
"""
|
|
|
|
from __future__ import annotations
|
|
|
|
import datetime
|
|
import decimal
|
|
import re
|
|
from typing import TYPE_CHECKING
|
|
|
|
from sqlalchemy import alias
|
|
from sqlalchemy import and_
|
|
from sqlalchemy import asc
|
|
from sqlalchemy import bindparam
|
|
from sqlalchemy import Boolean
|
|
from sqlalchemy import case
|
|
from sqlalchemy import cast
|
|
from sqlalchemy import CheckConstraint
|
|
from sqlalchemy import Column
|
|
from sqlalchemy import Date
|
|
from sqlalchemy import desc
|
|
from sqlalchemy import distinct
|
|
from sqlalchemy import Enum
|
|
from sqlalchemy import exc
|
|
from sqlalchemy import except_
|
|
from sqlalchemy import exists
|
|
from sqlalchemy import Float
|
|
from sqlalchemy import ForeignKey
|
|
from sqlalchemy import func
|
|
from sqlalchemy import Index
|
|
from sqlalchemy import insert
|
|
from sqlalchemy import insert_sentinel
|
|
from sqlalchemy import Integer
|
|
from sqlalchemy import intersect
|
|
from sqlalchemy import join
|
|
from sqlalchemy import literal
|
|
from sqlalchemy import literal_column
|
|
from sqlalchemy import MetaData
|
|
from sqlalchemy import not_
|
|
from sqlalchemy import null
|
|
from sqlalchemy import nulls_first
|
|
from sqlalchemy import nulls_last
|
|
from sqlalchemy import nullsfirst
|
|
from sqlalchemy import nullslast
|
|
from sqlalchemy import Numeric
|
|
from sqlalchemy import or_
|
|
from sqlalchemy import OrderByList
|
|
from sqlalchemy import outerjoin
|
|
from sqlalchemy import over
|
|
from sqlalchemy import schema
|
|
from sqlalchemy import select
|
|
from sqlalchemy import Sequence
|
|
from sqlalchemy import sql
|
|
from sqlalchemy import String
|
|
from sqlalchemy import Table
|
|
from sqlalchemy import testing
|
|
from sqlalchemy import Text
|
|
from sqlalchemy import text
|
|
from sqlalchemy import TIMESTAMP
|
|
from sqlalchemy import true
|
|
from sqlalchemy import try_cast
|
|
from sqlalchemy import tuple_
|
|
from sqlalchemy import type_coerce
|
|
from sqlalchemy import types
|
|
from sqlalchemy import union
|
|
from sqlalchemy import union_all
|
|
from sqlalchemy import update
|
|
from sqlalchemy import util
|
|
from sqlalchemy.dialects import mssql
|
|
from sqlalchemy.dialects import mysql
|
|
from sqlalchemy.dialects import oracle
|
|
from sqlalchemy.dialects import postgresql
|
|
from sqlalchemy.dialects import sqlite
|
|
from sqlalchemy.dialects.postgresql.base import PGCompiler
|
|
from sqlalchemy.dialects.postgresql.base import PGDialect
|
|
from sqlalchemy.engine import default
|
|
from sqlalchemy.ext.compiler import compiles
|
|
from sqlalchemy.sql import column
|
|
from sqlalchemy.sql import compiler
|
|
from sqlalchemy.sql import elements
|
|
from sqlalchemy.sql import label
|
|
from sqlalchemy.sql import operators
|
|
from sqlalchemy.sql import table
|
|
from sqlalchemy.sql import util as sql_util
|
|
from sqlalchemy.sql.elements import BooleanClauseList
|
|
from sqlalchemy.sql.elements import ColumnElement
|
|
from sqlalchemy.sql.elements import CompilerColumnElement
|
|
from sqlalchemy.sql.elements import Grouping
|
|
from sqlalchemy.sql.expression import ClauseElement
|
|
from sqlalchemy.sql.expression import ClauseList
|
|
from sqlalchemy.sql.expression import ColumnClause
|
|
from sqlalchemy.sql.expression import TableClause
|
|
from sqlalchemy.sql.selectable import LABEL_STYLE_NONE
|
|
from sqlalchemy.sql.selectable import LABEL_STYLE_TABLENAME_PLUS_COL
|
|
from sqlalchemy.testing import assert_raises
|
|
from sqlalchemy.testing import assert_raises_message
|
|
from sqlalchemy.testing import AssertsCompiledSQL
|
|
from sqlalchemy.testing import eq_
|
|
from sqlalchemy.testing import eq_ignore_whitespace
|
|
from sqlalchemy.testing import expect_raises
|
|
from sqlalchemy.testing import expect_raises_message
|
|
from sqlalchemy.testing import fixtures
|
|
from sqlalchemy.testing import is_
|
|
from sqlalchemy.testing import is_none
|
|
from sqlalchemy.testing import is_true
|
|
from sqlalchemy.testing import mock
|
|
from sqlalchemy.testing import ne_
|
|
from sqlalchemy.testing import Variation
|
|
from sqlalchemy.testing.schema import pep435_enum
|
|
from sqlalchemy.types import UserDefinedType
|
|
|
|
|
|
if TYPE_CHECKING:
|
|
from sqlalchemy import Select
|
|
|
|
|
|
table1 = table(
|
|
"mytable",
|
|
column("myid", Integer),
|
|
column("name", String),
|
|
column("description", String),
|
|
)
|
|
|
|
table2 = table(
|
|
"myothertable", column("otherid", Integer), column("othername", String)
|
|
)
|
|
|
|
table3 = table(
|
|
"thirdtable", column("userid", Integer), column("otherstuff", String)
|
|
)
|
|
|
|
metadata = MetaData()
|
|
|
|
# table with a schema
|
|
table4 = Table(
|
|
"remotetable",
|
|
metadata,
|
|
Column("rem_id", Integer, primary_key=True),
|
|
Column("datatype_id", Integer),
|
|
Column("value", String(20)),
|
|
schema="remote_owner",
|
|
)
|
|
|
|
# table with a 'multipart' schema
|
|
table5 = Table(
|
|
"remotetable",
|
|
metadata,
|
|
Column("rem_id", Integer, primary_key=True),
|
|
Column("datatype_id", Integer),
|
|
Column("value", String(20)),
|
|
schema="dbo.remote_owner",
|
|
)
|
|
|
|
parent = Table("parent", metadata, Column("id", Integer, primary_key=True))
|
|
child = Table(
|
|
"child",
|
|
metadata,
|
|
Column("id", Integer, primary_key=True),
|
|
Column("parent_id", ForeignKey("parent.id")),
|
|
)
|
|
users = table(
|
|
"users", column("user_id"), column("user_name"), column("password")
|
|
)
|
|
|
|
addresses = table(
|
|
"addresses",
|
|
column("address_id"),
|
|
column("user_id"),
|
|
column("street"),
|
|
column("city"),
|
|
column("state"),
|
|
column("zip"),
|
|
)
|
|
|
|
keyed = Table(
|
|
"keyed",
|
|
metadata,
|
|
Column("x", Integer, key="colx"),
|
|
Column("y", Integer, key="coly"),
|
|
Column("z", Integer),
|
|
)
|
|
|
|
|
|
class TestCompilerFixture(fixtures.TestBase, AssertsCompiledSQL):
|
|
def test_dont_access_statement(self):
|
|
def visit_foobar(self, element, **kw):
|
|
self.statement.table
|
|
|
|
class Foobar(ClauseElement):
|
|
__visit_name__ = "foobar"
|
|
|
|
with mock.patch.object(
|
|
testing.db.dialect.statement_compiler,
|
|
"visit_foobar",
|
|
visit_foobar,
|
|
create=True,
|
|
):
|
|
assert_raises_message(
|
|
NotImplementedError,
|
|
"compiler accessed .statement; use "
|
|
"compiler.current_executable",
|
|
self.assert_compile,
|
|
Foobar(),
|
|
"",
|
|
)
|
|
|
|
def test_no_stack(self):
|
|
def visit_foobar(self, element, **kw):
|
|
self.current_executable.table
|
|
|
|
class Foobar(ClauseElement):
|
|
__visit_name__ = "foobar"
|
|
|
|
with mock.patch.object(
|
|
testing.db.dialect.statement_compiler,
|
|
"visit_foobar",
|
|
visit_foobar,
|
|
create=True,
|
|
):
|
|
compiler = testing.db.dialect.statement_compiler(
|
|
testing.db.dialect, None
|
|
)
|
|
assert_raises_message(
|
|
IndexError,
|
|
"Compiler does not have a stack entry",
|
|
compiler.process,
|
|
Foobar(),
|
|
)
|
|
|
|
|
|
class SelectTest(fixtures.TestBase, AssertsCompiledSQL):
|
|
__dialect__ = "default"
|
|
|
|
def test_compiler_column_element_is_slots(self):
|
|
class SomeColThing(CompilerColumnElement):
|
|
__slots__ = ("name",)
|
|
__visit_name__ = "some_col_thing"
|
|
|
|
def __init__(self, name):
|
|
self.name = name
|
|
|
|
c1 = SomeColThing("some name")
|
|
eq_(c1.name, "some name")
|
|
assert not hasattr(c1, "__dict__")
|
|
|
|
def test_compile_label_is_slots(self):
|
|
c1 = compiler._CompileLabel(column("q"), "somename")
|
|
|
|
eq_(c1.name, "somename")
|
|
assert not hasattr(c1, "__dict__")
|
|
|
|
def test_attribute_sanity(self):
|
|
assert hasattr(table1, "c")
|
|
assert hasattr(table1.select().subquery(), "c")
|
|
assert not hasattr(table1.c.myid.self_group(), "columns")
|
|
assert not hasattr(table1.c.myid, "columns")
|
|
assert not hasattr(table1.c.myid, "c")
|
|
assert not hasattr(table1.select().subquery().c.myid, "c")
|
|
assert not hasattr(table1.select().subquery().c.myid, "columns")
|
|
assert not hasattr(table1.alias().c.myid, "columns")
|
|
assert not hasattr(table1.alias().c.myid, "c")
|
|
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Scalar Select expression has no "
|
|
"columns; use this object directly within a "
|
|
"column-level expression.",
|
|
getattr,
|
|
select(table1.c.myid).scalar_subquery().self_group(),
|
|
"columns",
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Scalar Select expression has no "
|
|
"columns; use this object directly within a "
|
|
"column-level expression.",
|
|
getattr,
|
|
select(table1.c.myid).scalar_subquery(),
|
|
"columns",
|
|
)
|
|
|
|
def test_table_select(self):
|
|
self.assert_compile(
|
|
table1.select(),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM mytable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(table1, table2),
|
|
"SELECT mytable.myid, mytable.name, mytable.description, "
|
|
"myothertable.otherid, myothertable.othername FROM mytable, "
|
|
"myothertable",
|
|
)
|
|
|
|
def test_int_limit_offset_coercion(self):
|
|
for given, exp in [
|
|
("5", 5),
|
|
(5, 5),
|
|
(5.2, 5),
|
|
(decimal.Decimal("5"), 5),
|
|
(None, None),
|
|
]:
|
|
eq_(select().limit(given)._limit, exp)
|
|
eq_(select().offset(given)._offset, exp)
|
|
|
|
assert_raises(ValueError, select().limit, "foo")
|
|
assert_raises(ValueError, select().offset, "foo")
|
|
|
|
def test_limit_offset_no_int_coercion_one(self):
|
|
exp1 = literal_column("Q")
|
|
exp2 = literal_column("Y")
|
|
self.assert_compile(
|
|
select(1).limit(exp1).offset(exp2), "SELECT 1 LIMIT Q OFFSET Y"
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(1).limit(bindparam("x")).offset(bindparam("y")),
|
|
"SELECT 1 LIMIT :x OFFSET :y",
|
|
)
|
|
|
|
def test_limit_offset_no_int_coercion_two(self):
|
|
exp1 = literal_column("Q")
|
|
exp2 = literal_column("Y")
|
|
sel = select(1).limit(exp1).offset(exp2)
|
|
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
"This SELECT structure does not use a simple integer "
|
|
"value for limit",
|
|
getattr,
|
|
sel,
|
|
"_limit",
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
"This SELECT structure does not use a simple integer "
|
|
"value for offset",
|
|
getattr,
|
|
sel,
|
|
"_offset",
|
|
)
|
|
|
|
def test_limit_offset_no_int_coercion_three(self):
|
|
exp1 = bindparam("Q")
|
|
exp2 = bindparam("Y")
|
|
sel = select(1).limit(exp1).offset(exp2)
|
|
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
"This SELECT structure does not use a simple integer "
|
|
"value for limit",
|
|
getattr,
|
|
sel,
|
|
"_limit",
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
"This SELECT structure does not use a simple integer "
|
|
"value for offset",
|
|
getattr,
|
|
sel,
|
|
"_offset",
|
|
)
|
|
|
|
@testing.combinations(
|
|
(
|
|
5,
|
|
10,
|
|
"LIMIT :param_1 OFFSET :param_2",
|
|
{"param_1": 5, "param_2": 10},
|
|
),
|
|
(None, 10, "LIMIT -1 OFFSET :param_1", {"param_1": 10}),
|
|
(5, None, "LIMIT :param_1", {"param_1": 5}),
|
|
(
|
|
0,
|
|
0,
|
|
"LIMIT :param_1 OFFSET :param_2",
|
|
{"param_1": 0, "param_2": 0},
|
|
),
|
|
(
|
|
literal_column("Q"),
|
|
literal_column("Y"),
|
|
"LIMIT Q OFFSET Y",
|
|
{},
|
|
),
|
|
(
|
|
column("Q"),
|
|
column("Y"),
|
|
'LIMIT "Q" OFFSET "Y"',
|
|
{},
|
|
),
|
|
)
|
|
def test_limit_offset(self, lim, offset, exp, params):
|
|
self.assert_compile(
|
|
select(1).limit(lim).offset(offset),
|
|
"SELECT 1 " + exp,
|
|
checkparams=params,
|
|
)
|
|
|
|
@testing.combinations(
|
|
(
|
|
5,
|
|
10,
|
|
{},
|
|
"OFFSET :param_1 ROWS FETCH FIRST :param_2 ROWS ONLY",
|
|
{"param_1": 10, "param_2": 5},
|
|
),
|
|
(None, 10, {}, "LIMIT -1 OFFSET :param_1", {"param_1": 10}),
|
|
(
|
|
5,
|
|
None,
|
|
{},
|
|
"FETCH FIRST :param_1 ROWS ONLY",
|
|
{"param_1": 5},
|
|
),
|
|
(
|
|
0,
|
|
0,
|
|
{},
|
|
"OFFSET :param_1 ROWS FETCH FIRST :param_2 ROWS ONLY",
|
|
{"param_1": 0, "param_2": 0},
|
|
),
|
|
(
|
|
5,
|
|
10,
|
|
{"percent": True},
|
|
"OFFSET :param_1 ROWS FETCH FIRST :param_2 PERCENT ROWS ONLY",
|
|
{"param_1": 10, "param_2": 5},
|
|
),
|
|
(
|
|
5,
|
|
10,
|
|
{"percent": True, "with_ties": True},
|
|
"OFFSET :param_1 ROWS FETCH FIRST :param_2 PERCENT ROWS WITH TIES",
|
|
{"param_1": 10, "param_2": 5},
|
|
),
|
|
(
|
|
5,
|
|
10,
|
|
{"with_ties": True},
|
|
"OFFSET :param_1 ROWS FETCH FIRST :param_2 ROWS WITH TIES",
|
|
{"param_1": 10, "param_2": 5},
|
|
),
|
|
(
|
|
literal_column("Q"),
|
|
literal_column("Y"),
|
|
{},
|
|
"OFFSET Y ROWS FETCH FIRST Q ROWS ONLY",
|
|
{},
|
|
),
|
|
(
|
|
column("Q"),
|
|
column("Y"),
|
|
{},
|
|
'OFFSET "Y" ROWS FETCH FIRST "Q" ROWS ONLY',
|
|
{},
|
|
),
|
|
(
|
|
bindparam("Q", 3),
|
|
bindparam("Y", 7),
|
|
{},
|
|
"OFFSET :Y ROWS FETCH FIRST :Q ROWS ONLY",
|
|
{"Q": 3, "Y": 7},
|
|
),
|
|
(
|
|
literal_column("Q") + literal_column("Z"),
|
|
literal_column("Y") + literal_column("W"),
|
|
{},
|
|
"OFFSET Y + W ROWS FETCH FIRST Q + Z ROWS ONLY",
|
|
{},
|
|
),
|
|
)
|
|
def test_fetch(self, fetch, offset, fetch_kw, exp, params):
|
|
self.assert_compile(
|
|
select(1).fetch(fetch, **fetch_kw).offset(offset),
|
|
"SELECT 1 " + exp,
|
|
checkparams=params,
|
|
)
|
|
|
|
def test_fetch_limit_offset_self_group(self):
|
|
self.assert_compile(
|
|
select(1).limit(1).self_group(),
|
|
"(SELECT 1 LIMIT :param_1)",
|
|
checkparams={"param_1": 1},
|
|
)
|
|
self.assert_compile(
|
|
select(1).offset(1).self_group(),
|
|
"(SELECT 1 LIMIT -1 OFFSET :param_1)",
|
|
checkparams={"param_1": 1},
|
|
)
|
|
self.assert_compile(
|
|
select(1).fetch(1).self_group(),
|
|
"(SELECT 1 FETCH FIRST :param_1 ROWS ONLY)",
|
|
checkparams={"param_1": 1},
|
|
)
|
|
|
|
def test_limit_fetch_interaction(self):
|
|
self.assert_compile(
|
|
select(1).limit(42).fetch(1),
|
|
"SELECT 1 FETCH FIRST :param_1 ROWS ONLY",
|
|
checkparams={"param_1": 1},
|
|
)
|
|
self.assert_compile(
|
|
select(1).fetch(42).limit(1),
|
|
"SELECT 1 LIMIT :param_1",
|
|
checkparams={"param_1": 1},
|
|
)
|
|
self.assert_compile(
|
|
select(1).limit(42).offset(7).fetch(1),
|
|
"SELECT 1 OFFSET :param_1 ROWS FETCH FIRST :param_2 ROWS ONLY",
|
|
checkparams={"param_1": 7, "param_2": 1},
|
|
)
|
|
self.assert_compile(
|
|
select(1).fetch(1).slice(2, 5),
|
|
"SELECT 1 LIMIT :param_1 OFFSET :param_2",
|
|
checkparams={"param_1": 3, "param_2": 2},
|
|
)
|
|
self.assert_compile(
|
|
select(1).slice(2, 5).fetch(1),
|
|
"SELECT 1 OFFSET :param_1 ROWS FETCH FIRST :param_2 ROWS ONLY",
|
|
checkparams={"param_1": 2, "param_2": 1},
|
|
)
|
|
|
|
def test_select_precol_compile_ordering(self):
|
|
s1 = (
|
|
select(column("x"))
|
|
.select_from(text("a"))
|
|
.limit(5)
|
|
.scalar_subquery()
|
|
)
|
|
s2 = select(s1).limit(10)
|
|
|
|
class MyCompiler(compiler.SQLCompiler):
|
|
def get_select_precolumns(self, select, **kw):
|
|
result = ""
|
|
if select._limit:
|
|
result += "FIRST %s " % self.process(
|
|
literal(select._limit), **kw
|
|
)
|
|
if select._offset:
|
|
result += "SKIP %s " % self.process(
|
|
literal(select._offset), **kw
|
|
)
|
|
return result
|
|
|
|
def limit_clause(self, select, **kw):
|
|
return ""
|
|
|
|
dialect = default.DefaultDialect()
|
|
dialect.statement_compiler = MyCompiler
|
|
dialect.paramstyle = "qmark"
|
|
dialect.positional = True
|
|
self.assert_compile(
|
|
s2,
|
|
"SELECT FIRST ? (SELECT FIRST ? x FROM a) AS anon_1",
|
|
checkpositional=(10, 5),
|
|
dialect=dialect,
|
|
)
|
|
|
|
@testing.combinations(
|
|
(
|
|
select(table1.c.name)
|
|
.select_from(table1, table2)
|
|
.where(table1.c.myid == table2.c.otherid),
|
|
"SELECT mytable.name FROM mytable, myothertable "
|
|
"WHERE mytable.myid = myothertable.otherid",
|
|
),
|
|
(
|
|
select(table1.c.name)
|
|
.select_from(table2, table1)
|
|
.where(table1.c.myid == table2.c.otherid),
|
|
"SELECT mytable.name FROM myothertable, mytable "
|
|
"WHERE mytable.myid = myothertable.otherid",
|
|
),
|
|
(
|
|
select(table1.c.name)
|
|
.where(table1.c.myid == table2.c.otherid)
|
|
.select_from(table2, table1),
|
|
"SELECT mytable.name FROM myothertable, mytable "
|
|
"WHERE mytable.myid = myothertable.otherid",
|
|
),
|
|
(
|
|
select(table1.c.name)
|
|
.where(table1.c.myid == table2.c.otherid)
|
|
.select_from(table1, table2),
|
|
"SELECT mytable.name FROM mytable, myothertable "
|
|
"WHERE mytable.myid = myothertable.otherid",
|
|
),
|
|
(
|
|
select(table3.c.userid, table1.c.name)
|
|
.where(table1.c.myid == table2.c.otherid)
|
|
.select_from(table1, table3, table2),
|
|
"SELECT thirdtable.userid, mytable.name "
|
|
"FROM mytable, thirdtable, myothertable "
|
|
"WHERE mytable.myid = myothertable.otherid",
|
|
),
|
|
(
|
|
select(table3.c.userid, table1.c.name)
|
|
.where(table1.c.myid == table2.c.otherid)
|
|
.select_from(table3, table1, table2),
|
|
"SELECT thirdtable.userid, mytable.name "
|
|
"FROM thirdtable, mytable, myothertable "
|
|
"WHERE mytable.myid = myothertable.otherid",
|
|
),
|
|
(
|
|
select(table3.c.userid, table1.c.name)
|
|
.where(table1.c.myid == table2.c.otherid)
|
|
.select_from(table1, table2),
|
|
"SELECT thirdtable.userid, mytable.name "
|
|
"FROM mytable, myothertable, thirdtable "
|
|
"WHERE mytable.myid = myothertable.otherid",
|
|
),
|
|
(
|
|
select(table3.c.userid, table1.c.name)
|
|
.where(table1.c.myid == table2.c.otherid)
|
|
.select_from(table3, table2),
|
|
"SELECT thirdtable.userid, mytable.name "
|
|
"FROM thirdtable, myothertable, mytable "
|
|
"WHERE mytable.myid = myothertable.otherid",
|
|
),
|
|
(
|
|
select(table3.c.userid, table1.c.name)
|
|
.where(table1.c.myid == table2.c.otherid)
|
|
.select_from(table3, table2)
|
|
.join_from(table3, table1, table3.c.userid == table1.c.myid),
|
|
"SELECT thirdtable.userid, mytable.name "
|
|
"FROM thirdtable "
|
|
"JOIN mytable ON thirdtable.userid = mytable.myid, "
|
|
"myothertable WHERE mytable.myid = myothertable.otherid",
|
|
),
|
|
(
|
|
select(table3.c.userid, table1.c.name)
|
|
.where(table1.c.myid == table2.c.otherid)
|
|
.select_from(table2, table3)
|
|
.join_from(table3, table1, table3.c.userid == table1.c.myid),
|
|
"SELECT thirdtable.userid, mytable.name "
|
|
"FROM myothertable, thirdtable "
|
|
"JOIN mytable ON thirdtable.userid = mytable.myid "
|
|
"WHERE mytable.myid = myothertable.otherid",
|
|
),
|
|
)
|
|
def test_select_from_ordering(self, stmt, expected):
|
|
self.assert_compile(stmt, expected)
|
|
|
|
def test_from_subquery(self):
|
|
"""tests placing select statements in the column clause of
|
|
another select, for the
|
|
purposes of selecting from the exported columns of that select."""
|
|
|
|
s = select(table1).where(table1.c.name == "jack").subquery()
|
|
self.assert_compile(
|
|
select(s).where(s.c.myid == 7),
|
|
"SELECT anon_1.myid, anon_1.name, anon_1.description FROM "
|
|
"(SELECT mytable.myid AS myid, "
|
|
"mytable.name AS name, mytable.description AS description "
|
|
"FROM mytable "
|
|
"WHERE mytable.name = :name_1) AS anon_1 WHERE "
|
|
"anon_1.myid = :myid_1",
|
|
)
|
|
|
|
sq = select(table1)
|
|
self.assert_compile(
|
|
sq.subquery().select(),
|
|
"SELECT anon_1.myid, anon_1.name, anon_1.description FROM "
|
|
"(SELECT mytable.myid AS myid, "
|
|
"mytable.name AS name, mytable.description "
|
|
"AS description FROM mytable) AS anon_1",
|
|
)
|
|
|
|
sq = select(table1).alias("sq")
|
|
|
|
self.assert_compile(
|
|
sq.select().where(sq.c.myid == 7),
|
|
"SELECT sq.myid, sq.name, sq.description FROM "
|
|
"(SELECT mytable.myid AS myid, mytable.name AS name, "
|
|
"mytable.description AS description FROM mytable) AS sq "
|
|
"WHERE sq.myid = :myid_1",
|
|
)
|
|
|
|
sq = (
|
|
select(table1, table2)
|
|
.where(and_(table1.c.myid == 7, table2.c.otherid == table1.c.myid))
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
.alias("sq")
|
|
)
|
|
|
|
sqstring = (
|
|
"SELECT mytable.myid AS mytable_myid, mytable.name AS "
|
|
"mytable_name, mytable.description AS mytable_description, "
|
|
"myothertable.otherid AS myothertable_otherid, "
|
|
"myothertable.othername AS myothertable_othername FROM "
|
|
"mytable, myothertable WHERE mytable.myid = :myid_1 AND "
|
|
"myothertable.otherid = mytable.myid"
|
|
)
|
|
|
|
self.assert_compile(
|
|
sq.select(),
|
|
"SELECT sq.mytable_myid, sq.mytable_name, "
|
|
"sq.mytable_description, sq.myothertable_otherid, "
|
|
"sq.myothertable_othername FROM (%s) AS sq" % sqstring,
|
|
)
|
|
|
|
sq2 = (
|
|
select(sq)
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
.alias("sq2")
|
|
)
|
|
|
|
self.assert_compile(
|
|
sq2.select(),
|
|
"SELECT sq2.sq_mytable_myid, sq2.sq_mytable_name, "
|
|
"sq2.sq_mytable_description, sq2.sq_myothertable_otherid, "
|
|
"sq2.sq_myothertable_othername FROM "
|
|
"(SELECT sq.mytable_myid AS "
|
|
"sq_mytable_myid, sq.mytable_name AS sq_mytable_name, "
|
|
"sq.mytable_description AS sq_mytable_description, "
|
|
"sq.myothertable_otherid AS sq_myothertable_otherid, "
|
|
"sq.myothertable_othername AS sq_myothertable_othername "
|
|
"FROM (%s) AS sq) AS sq2" % sqstring,
|
|
)
|
|
|
|
def test_select_from_clauselist(self):
|
|
self.assert_compile(
|
|
select(ClauseList(column("a"), column("b"))).select_from(
|
|
text("sometable")
|
|
),
|
|
"SELECT a, b FROM sometable",
|
|
)
|
|
|
|
def test_use_labels(self):
|
|
self.assert_compile(
|
|
select(table1.c.myid == 5).set_label_style(
|
|
LABEL_STYLE_TABLENAME_PLUS_COL
|
|
),
|
|
"SELECT mytable.myid = :myid_1 AS anon_1 FROM mytable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(func.foo()).set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL),
|
|
"SELECT foo() AS foo_1",
|
|
)
|
|
|
|
# this is native_boolean=False for default dialect
|
|
self.assert_compile(
|
|
select(not_(True)).set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL),
|
|
"SELECT :param_1 = 0 AS anon_1",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(cast("data", Integer)).set_label_style(
|
|
LABEL_STYLE_TABLENAME_PLUS_COL
|
|
),
|
|
"SELECT CAST(:param_1 AS INTEGER) AS anon_1",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(
|
|
func.sum(func.lala(table1.c.myid).label("foo")).label("bar")
|
|
),
|
|
"SELECT sum(lala(mytable.myid)) AS bar FROM mytable",
|
|
)
|
|
|
|
def test_use_labels_keyed(self):
|
|
self.assert_compile(
|
|
select(keyed), "SELECT keyed.x, keyed.y, keyed.z FROM keyed"
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(keyed).set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL),
|
|
"SELECT keyed.x AS keyed_x, keyed.y AS "
|
|
"keyed_y, keyed.z AS keyed_z FROM keyed",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(
|
|
select(keyed)
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
.subquery()
|
|
).set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL),
|
|
"SELECT anon_1.keyed_x AS anon_1_keyed_x, "
|
|
"anon_1.keyed_y AS anon_1_keyed_y, "
|
|
"anon_1.keyed_z AS anon_1_keyed_z "
|
|
"FROM (SELECT keyed.x AS keyed_x, keyed.y AS keyed_y, "
|
|
"keyed.z AS keyed_z FROM keyed) AS anon_1",
|
|
)
|
|
|
|
def test_paramstyles(self):
|
|
stmt = text("select :foo, :bar, :bat from sometable")
|
|
|
|
self.assert_compile(
|
|
stmt,
|
|
"select ?, ?, ? from sometable",
|
|
dialect=default.DefaultDialect(paramstyle="qmark"),
|
|
)
|
|
self.assert_compile(
|
|
stmt,
|
|
"select :foo, :bar, :bat from sometable",
|
|
dialect=default.DefaultDialect(paramstyle="named"),
|
|
)
|
|
self.assert_compile(
|
|
stmt,
|
|
"select %s, %s, %s from sometable",
|
|
dialect=default.DefaultDialect(paramstyle="format"),
|
|
)
|
|
self.assert_compile(
|
|
stmt,
|
|
"select :1, :2, :3 from sometable",
|
|
dialect=default.DefaultDialect(paramstyle="numeric"),
|
|
)
|
|
self.assert_compile(
|
|
stmt,
|
|
"select %(foo)s, %(bar)s, %(bat)s from sometable",
|
|
dialect=default.DefaultDialect(paramstyle="pyformat"),
|
|
)
|
|
|
|
def test_anon_param_name_on_keys(self):
|
|
self.assert_compile(
|
|
keyed.insert(),
|
|
"INSERT INTO keyed (x, y, z) VALUES (%(colx)s, %(coly)s, %(z)s)",
|
|
dialect=default.DefaultDialect(paramstyle="pyformat"),
|
|
)
|
|
self.assert_compile(
|
|
keyed.c.coly == 5,
|
|
"keyed.y = %(coly_1)s",
|
|
checkparams={"coly_1": 5},
|
|
dialect=default.DefaultDialect(paramstyle="pyformat"),
|
|
)
|
|
|
|
def test_dupe_columns(self):
|
|
"""as of 1.4, there's no deduping."""
|
|
|
|
self.assert_compile(
|
|
select(column("a"), column("a"), column("a")),
|
|
"SELECT a, a, a",
|
|
dialect=default.DefaultDialect(),
|
|
)
|
|
|
|
c = column("a")
|
|
self.assert_compile(
|
|
select(c, c, c),
|
|
"SELECT a, a, a",
|
|
dialect=default.DefaultDialect(),
|
|
)
|
|
|
|
a, b = column("a"), column("b")
|
|
self.assert_compile(
|
|
select(a, b, b, b, a, a),
|
|
"SELECT a, b, b, b, a, a",
|
|
dialect=default.DefaultDialect(),
|
|
)
|
|
|
|
# using alternate keys.
|
|
a, b, c = (
|
|
Column("a", Integer, key="b"),
|
|
Column("b", Integer),
|
|
Column("c", Integer, key="a"),
|
|
)
|
|
self.assert_compile(
|
|
select(a, b, c, a, b, c),
|
|
"SELECT a, b, c, a, b, c",
|
|
dialect=default.DefaultDialect(),
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(bindparam("a"), bindparam("b"), bindparam("c")),
|
|
"SELECT :a AS anon_1, :b AS anon_2, :c AS anon_3",
|
|
dialect=default.DefaultDialect(paramstyle="named"),
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(bindparam("a"), bindparam("b"), bindparam("c")),
|
|
"SELECT ? AS anon_1, ? AS anon_2, ? AS anon_3",
|
|
dialect=default.DefaultDialect(paramstyle="qmark"),
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(column("a"), column("a"), column("a")), "SELECT a, a, a"
|
|
)
|
|
|
|
s = select(bindparam("a"), bindparam("b"), bindparam("c"))
|
|
s = s.compile(dialect=default.DefaultDialect(paramstyle="qmark"))
|
|
eq_(s.positiontup, ["a", "b", "c"])
|
|
|
|
def test_overlapping_labels_use_labels(self):
|
|
foo = table("foo", column("id"), column("bar_id"))
|
|
foo_bar = table("foo_bar", column("id"))
|
|
|
|
stmt = select(foo, foo_bar).set_label_style(
|
|
LABEL_STYLE_TABLENAME_PLUS_COL
|
|
)
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT foo.id AS foo_id, foo.bar_id AS foo_bar_id, "
|
|
"foo_bar.id AS foo_bar_id_1 "
|
|
"FROM foo, foo_bar",
|
|
)
|
|
|
|
def test_overlapping_labels_plus_dupes_use_labels(self):
|
|
foo = table("foo", column("id"), column("bar_id"))
|
|
foo_bar = table("foo_bar", column("id"))
|
|
|
|
# current approach is:
|
|
# 1. positional nature of columns is always maintained in all cases
|
|
# 2. two different columns that have the same label, second one
|
|
# is disambiguated
|
|
# 3. if the same column is repeated, it gets deduped using a special
|
|
# 'dedupe' label that will show two underscores
|
|
# 4. The disambiguating label generated in #2 also has to be deduped.
|
|
# 5. The derived columns, e.g. subquery().c etc. do not export the
|
|
# "dedupe" columns, at all. they are unreachable (because they
|
|
# are unreachable anyway in SQL unless you use "SELECT *")
|
|
#
|
|
# this is all new logic necessitated by #4753 since we allow columns
|
|
# to be repeated. We would still like the targeting of this column,
|
|
# both in a result set as well as in a derived selectable, to be
|
|
# unambiguous (DBs like postgresql won't let us reference an ambiguous
|
|
# label in a derived selectable even if its the same column repeated).
|
|
#
|
|
# this kind of thing happens of course because the ORM is in some
|
|
# more exotic cases writing in joins where columns may be duped.
|
|
# it might be nice to fix it on that side also, however SQLAlchemy
|
|
# has deduped columns in SELECT statements for 13 years so having a
|
|
# robust behavior when dupes are present is still very useful.
|
|
|
|
stmt = select(
|
|
foo.c.id,
|
|
foo.c.bar_id,
|
|
foo_bar.c.id,
|
|
foo.c.bar_id,
|
|
foo.c.id,
|
|
foo.c.bar_id,
|
|
foo_bar.c.id,
|
|
foo_bar.c.id,
|
|
).set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT foo.id AS foo_id, "
|
|
"foo.bar_id AS foo_bar_id, " # 1. 1st foo.bar_id, as is
|
|
"foo_bar.id AS foo_bar_id_1, " # 2. 1st foo_bar.id, disamb from 1
|
|
"foo.bar_id AS foo_bar_id__1, " # 3. 2nd foo.bar_id, dedupe from 1
|
|
"foo.id AS foo_id__1, "
|
|
"foo.bar_id AS foo_bar_id__2, " # 4. 3rd foo.bar_id, dedupe again
|
|
"foo_bar.id AS foo_bar_id__3, " # 5. 2nd foo_bar.id
|
|
"foo_bar.id AS foo_bar_id__4 " # 6. 3rd foo_bar.id, dedupe again
|
|
"FROM foo, foo_bar",
|
|
)
|
|
eq_(
|
|
stmt.selected_columns.keys(),
|
|
[
|
|
"foo_id",
|
|
"foo_bar_id",
|
|
"foo_bar_id_1",
|
|
"foo_bar_id_2",
|
|
"foo_id_1",
|
|
"foo_bar_id_2",
|
|
"foo_bar_id_1",
|
|
"foo_bar_id_1",
|
|
],
|
|
)
|
|
|
|
# for the subquery, the labels created for repeated occurrences
|
|
# of the same column are not used. only the label applied to the
|
|
# first occurrence of each column is used
|
|
self.assert_compile(
|
|
select(stmt.subquery()).set_label_style(LABEL_STYLE_NONE),
|
|
"SELECT "
|
|
"anon_1.foo_id, " # from 1st foo.id in derived (line 1)
|
|
"anon_1.foo_bar_id, " # from 1st foo.bar_id in derived (line 2)
|
|
"anon_1.foo_bar_id_1, " # from 1st foo_bar.id in derived (line 3)
|
|
"anon_1.foo_bar_id, " # from 1st foo.bar_id in derived (line 2)
|
|
"anon_1.foo_id, " # from 1st foo.id in derived (line 1)
|
|
"anon_1.foo_bar_id, " # from 1st foo.bar_id in derived (line 2)
|
|
"anon_1.foo_bar_id_1, " # from 1st foo_bar.id in derived (line 3)
|
|
"anon_1.foo_bar_id_1 " # from 1st foo_bar.id in derived (line 3)
|
|
"FROM ("
|
|
"SELECT foo.id AS foo_id, "
|
|
"foo.bar_id AS foo_bar_id, " # 1. 1st foo.bar_id, as is
|
|
"foo_bar.id AS foo_bar_id_1, " # 2. 1st foo_bar.id, disamb from 1
|
|
"foo.bar_id AS foo_bar_id__1, " # 3. 2nd foo.bar_id, dedupe from 1
|
|
"foo.id AS foo_id__1, "
|
|
"foo.bar_id AS foo_bar_id__2, " # 4. 3rd foo.bar_id, dedupe again
|
|
"foo_bar.id AS foo_bar_id__3, " # 5. 2nd foo_bar.id
|
|
"foo_bar.id AS foo_bar_id__4 " # 6. 3rd foo_bar.id, dedupe again
|
|
"FROM foo, foo_bar"
|
|
") AS anon_1",
|
|
)
|
|
|
|
def test_overlapping_labels_plus_dupes_separate_keys_use_labels(self):
|
|
"""test a condition related to #6710.
|
|
|
|
prior to this issue CTE uses selected_columns to render the
|
|
"WITH RECURSIVE (colnames)" part. This test shows that this isn't
|
|
correct when keys are present.
|
|
|
|
See also test_cte ->
|
|
test_wrecur_ovlp_lbls_plus_dupes_separate_keys_use_labels
|
|
|
|
"""
|
|
m = MetaData()
|
|
foo = Table(
|
|
"foo",
|
|
m,
|
|
Column("id", Integer),
|
|
Column("bar_id", Integer, key="bb"),
|
|
)
|
|
foo_bar = Table("foo_bar", m, Column("id", Integer, key="bb"))
|
|
|
|
stmt = select(
|
|
foo.c.id,
|
|
foo.c.bb,
|
|
foo_bar.c.bb,
|
|
foo.c.bb,
|
|
foo.c.id,
|
|
foo.c.bb,
|
|
foo_bar.c.bb,
|
|
foo_bar.c.bb,
|
|
).set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
|
|
# note these keys are not what renders in the SQL. These keys
|
|
# will be addressable in the result set but can't be used in
|
|
# rendering, such as for a CTE
|
|
eq_(
|
|
stmt.selected_columns.keys(),
|
|
[
|
|
"foo_id",
|
|
"foo_bb",
|
|
"foo_bar_bb",
|
|
"foo_bb_1",
|
|
"foo_id_1",
|
|
"foo_bb_1",
|
|
"foo_bar_bb_1",
|
|
"foo_bar_bb_1",
|
|
],
|
|
)
|
|
eq_(
|
|
stmt.subquery().c.keys(),
|
|
[
|
|
"foo_id",
|
|
"foo_bb",
|
|
"foo_bar_bb",
|
|
"foo_bb_1",
|
|
"foo_id_1",
|
|
"foo_bb_1",
|
|
"foo_bar_bb_1",
|
|
"foo_bar_bb_1",
|
|
],
|
|
)
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT foo.id AS foo_id, "
|
|
"foo.bar_id AS foo_bar_id, " # 1. 1st foo.bar_id, as is
|
|
"foo_bar.id AS foo_bar_id_1, " # 2. 1st foo_bar.id, disamb from 1
|
|
"foo.bar_id AS foo_bar_id__1, " # 3. 2nd foo.bar_id, dedupe from 1
|
|
"foo.id AS foo_id__1, "
|
|
"foo.bar_id AS foo_bar_id__2, " # 4. 3rd foo.bar_id, dedupe again
|
|
"foo_bar.id AS foo_bar_id__3, " # 5. 2nd foo_bar.id
|
|
"foo_bar.id AS foo_bar_id__4 " # 6. 3rd foo_bar.id, dedupe again
|
|
"FROM foo, foo_bar",
|
|
)
|
|
|
|
def test_dupe_columns_use_labels(self):
|
|
t = table("t", column("a"), column("b"))
|
|
self.assert_compile(
|
|
select(t.c.a, t.c.a, t.c.b, t.c.a).set_label_style(
|
|
LABEL_STYLE_TABLENAME_PLUS_COL
|
|
),
|
|
"SELECT t.a AS t_a, t.a AS t_a__1, t.b AS t_b, "
|
|
"t.a AS t_a__2 FROM t",
|
|
)
|
|
|
|
def test_dupe_columns_use_labels_derived_selectable(self):
|
|
t = table("t", column("a"), column("b"))
|
|
stmt = (
|
|
select(t.c.a, t.c.a, t.c.b, t.c.a)
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
.subquery()
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(stmt).set_label_style(LABEL_STYLE_NONE),
|
|
"SELECT anon_1.t_a, anon_1.t_a, anon_1.t_b, anon_1.t_a FROM "
|
|
"(SELECT t.a AS t_a, t.a AS t_a__1, t.b AS t_b, t.a AS t_a__2 "
|
|
"FROM t) AS anon_1",
|
|
)
|
|
|
|
def test_dupe_columns_use_labels_mix_annotations(self):
|
|
t = table("t", column("a"), column("b"))
|
|
a, b, a_a = t.c.a, t.c.b, t.c.a._annotate({"some_orm_thing": True})
|
|
|
|
self.assert_compile(
|
|
select(a, a_a, b, a_a).set_label_style(
|
|
LABEL_STYLE_TABLENAME_PLUS_COL
|
|
),
|
|
"SELECT t.a AS t_a, t.a AS t_a__1, t.b AS t_b, "
|
|
"t.a AS t_a__2 FROM t",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(a_a, a, b, a_a).set_label_style(
|
|
LABEL_STYLE_TABLENAME_PLUS_COL
|
|
),
|
|
"SELECT t.a AS t_a, t.a AS t_a__1, t.b AS t_b, "
|
|
"t.a AS t_a__2 FROM t",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(a_a, a_a, b, a).set_label_style(
|
|
LABEL_STYLE_TABLENAME_PLUS_COL
|
|
),
|
|
"SELECT t.a AS t_a, t.a AS t_a__1, t.b AS t_b, "
|
|
"t.a AS t_a__2 FROM t",
|
|
)
|
|
|
|
def test_dupe_columns_use_labels_derived_selectable_mix_annotations(self):
|
|
t = table("t", column("a"), column("b"))
|
|
a, b, a_a = t.c.a, t.c.b, t.c.a._annotate({"some_orm_thing": True})
|
|
stmt = (
|
|
select(a, a_a, b, a_a)
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
.subquery()
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(stmt).set_label_style(LABEL_STYLE_NONE),
|
|
"SELECT anon_1.t_a, anon_1.t_a, anon_1.t_b, anon_1.t_a FROM "
|
|
"(SELECT t.a AS t_a, t.a AS t_a__1, t.b AS t_b, t.a AS t_a__2 "
|
|
"FROM t) AS anon_1",
|
|
)
|
|
|
|
def test_overlapping_labels_plus_dupes_use_labels_mix_annotations(self):
|
|
foo = table("foo", column("id"), column("bar_id"))
|
|
foo_bar = table("foo_bar", column("id"))
|
|
|
|
foo_bar__id = foo_bar.c.id._annotate({"some_orm_thing": True})
|
|
|
|
stmt = select(
|
|
foo.c.bar_id,
|
|
foo_bar.c.id,
|
|
foo_bar.c.id,
|
|
foo_bar__id,
|
|
foo_bar__id,
|
|
).set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT foo.bar_id AS foo_bar_id, foo_bar.id AS foo_bar_id_1, "
|
|
"foo_bar.id AS foo_bar_id__1, foo_bar.id AS foo_bar_id__2, "
|
|
"foo_bar.id AS foo_bar_id__3 FROM foo, foo_bar",
|
|
)
|
|
|
|
def test_dupe_columns_use_labels_from_anon(self):
|
|
t = table("t", column("a"), column("b"))
|
|
a = t.alias()
|
|
|
|
# second and third occurrences of a.c.a are labeled, but are
|
|
# dupes of each other.
|
|
self.assert_compile(
|
|
select(a.c.a, a.c.a, a.c.b, a.c.a).set_label_style(
|
|
LABEL_STYLE_TABLENAME_PLUS_COL
|
|
),
|
|
"SELECT t_1.a AS t_1_a, t_1.a AS t_1_a__1, t_1.b AS t_1_b, "
|
|
"t_1.a AS t_1_a__2 "
|
|
"FROM t AS t_1",
|
|
)
|
|
|
|
def test_nested_label_targeting(self):
|
|
"""test nested anonymous label generation."""
|
|
s1 = table1.select()
|
|
s2 = s1.alias()
|
|
s3 = select(s2).set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
s4 = s3.alias()
|
|
s5 = select(s4).set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
self.assert_compile(
|
|
s5,
|
|
"SELECT anon_1.anon_2_myid AS "
|
|
"anon_1_anon_2_myid, anon_1.anon_2_name AS "
|
|
"anon_1_anon_2_name, anon_1.anon_2_descript"
|
|
"ion AS anon_1_anon_2_description FROM "
|
|
"(SELECT anon_2.myid AS anon_2_myid, "
|
|
"anon_2.name AS anon_2_name, "
|
|
"anon_2.description AS anon_2_description "
|
|
"FROM (SELECT mytable.myid AS myid, "
|
|
"mytable.name AS name, mytable.description "
|
|
"AS description FROM mytable) AS anon_2) "
|
|
"AS anon_1",
|
|
)
|
|
|
|
def test_nested_label_targeting_keyed(self):
|
|
s1 = keyed.select()
|
|
s2 = s1.alias()
|
|
s3 = select(s2).set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
self.assert_compile(
|
|
s3,
|
|
"SELECT anon_1.x AS anon_1_x, "
|
|
"anon_1.y AS anon_1_y, "
|
|
"anon_1.z AS anon_1_z FROM "
|
|
"(SELECT keyed.x AS x, keyed.y "
|
|
"AS y, keyed.z AS z FROM keyed) AS anon_1",
|
|
)
|
|
|
|
s4 = s3.alias()
|
|
s5 = select(s4).set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
self.assert_compile(
|
|
s5,
|
|
"SELECT anon_1.anon_2_x AS anon_1_anon_2_x, "
|
|
"anon_1.anon_2_y AS anon_1_anon_2_y, "
|
|
"anon_1.anon_2_z AS anon_1_anon_2_z "
|
|
"FROM (SELECT anon_2.x AS anon_2_x, "
|
|
"anon_2.y AS anon_2_y, "
|
|
"anon_2.z AS anon_2_z FROM "
|
|
"(SELECT keyed.x AS x, keyed.y AS y, keyed.z "
|
|
"AS z FROM keyed) AS anon_2) AS anon_1",
|
|
)
|
|
|
|
@testing.combinations("per cent", "per % cent", "%percent")
|
|
def test_percent_names_collide_with_anonymizing(self, name):
|
|
table1 = table("t1", column(name))
|
|
|
|
jj = select(table1.c[name]).subquery()
|
|
jjj = join(table1, jj, table1.c[name] == jj.c[name])
|
|
|
|
j2 = (
|
|
jjj.select()
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
.subquery("foo")
|
|
)
|
|
|
|
self.assert_compile(
|
|
j2.select(),
|
|
'SELECT foo."t1_%(name)s", foo."anon_1_%(name)s" FROM '
|
|
'(SELECT t1."%(name)s" AS "t1_%(name)s", anon_1."%(name)s" '
|
|
'AS "anon_1_%(name)s" FROM t1 JOIN (SELECT t1."%(name)s" AS '
|
|
'"%(name)s" FROM t1) AS anon_1 ON t1."%(name)s" = '
|
|
'anon_1."%(name)s") AS foo' % {"name": name},
|
|
)
|
|
|
|
def test_exists(self):
|
|
s = select(table1.c.myid).where(table1.c.myid == 5)
|
|
|
|
self.assert_compile(
|
|
exists(s),
|
|
"EXISTS (SELECT mytable.myid FROM mytable "
|
|
"WHERE mytable.myid = :myid_1)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
exists(s.scalar_subquery()),
|
|
"EXISTS (SELECT mytable.myid FROM mytable "
|
|
"WHERE mytable.myid = :myid_1)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
exists(table1.c.myid).where(table1.c.myid == 5).select(),
|
|
"SELECT EXISTS (SELECT mytable.myid FROM "
|
|
"mytable WHERE mytable.myid = :myid_1) AS anon_1",
|
|
params={"mytable_myid": 5},
|
|
)
|
|
self.assert_compile(
|
|
select(table1, exists(1).select_from(table2)),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description, EXISTS (SELECT 1 "
|
|
"FROM myothertable) AS anon_1 FROM mytable",
|
|
params={},
|
|
)
|
|
self.assert_compile(
|
|
select(table1, exists(1).select_from(table2).label("foo")),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description, EXISTS (SELECT 1 "
|
|
"FROM myothertable) AS foo FROM mytable",
|
|
params={},
|
|
)
|
|
|
|
self.assert_compile(
|
|
table1.select().where(
|
|
exists()
|
|
.where(table2.c.otherid == table1.c.myid)
|
|
.correlate(table1)
|
|
),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM mytable WHERE "
|
|
"EXISTS (SELECT * FROM myothertable WHERE "
|
|
"myothertable.otherid = mytable.myid)",
|
|
)
|
|
self.assert_compile(
|
|
table1.select().where(
|
|
exists()
|
|
.where(table2.c.otherid == table1.c.myid)
|
|
.correlate(table1)
|
|
),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM mytable WHERE "
|
|
"EXISTS (SELECT * FROM myothertable WHERE "
|
|
"myothertable.otherid = mytable.myid)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(
|
|
or_(
|
|
exists().where(table2.c.otherid == "foo"),
|
|
exists().where(table2.c.otherid == "bar"),
|
|
)
|
|
),
|
|
"SELECT (EXISTS (SELECT * FROM myothertable "
|
|
"WHERE myothertable.otherid = :otherid_1)) "
|
|
"OR (EXISTS (SELECT * FROM myothertable WHERE "
|
|
"myothertable.otherid = :otherid_2)) AS anon_1",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(exists(1)), "SELECT EXISTS (SELECT 1) AS anon_1"
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(~exists(1)), "SELECT NOT (EXISTS (SELECT 1)) AS anon_1"
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(~(~exists(1))),
|
|
"SELECT NOT (NOT (EXISTS (SELECT 1))) AS anon_1",
|
|
)
|
|
|
|
self.assert_compile(
|
|
exists(42)
|
|
.select_from(table1)
|
|
.where(table1.c.name == "foo", table1.c.description == "bar"),
|
|
"EXISTS (SELECT 42 FROM mytable WHERE mytable.name = :name_1 "
|
|
"AND mytable.description = :description_1)",
|
|
)
|
|
|
|
def test_exists_method(self):
|
|
subq = (
|
|
select(func.count(table2.c.otherid))
|
|
.where(table2.c.otherid == table1.c.myid)
|
|
.correlate(table1)
|
|
.group_by(table2.c.otherid)
|
|
.having(func.count(table2.c.otherid) > 1)
|
|
.exists()
|
|
)
|
|
|
|
self.assert_compile(
|
|
table1.select().where(subq),
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable WHERE EXISTS (SELECT count(myothertable.otherid) "
|
|
"AS count_1 FROM myothertable WHERE myothertable.otherid = "
|
|
"mytable.myid GROUP BY myothertable.otherid "
|
|
"HAVING count(myothertable.otherid) > :count_2)",
|
|
)
|
|
|
|
def test_where_subquery(self):
|
|
s = (
|
|
select(addresses.c.street)
|
|
.where(addresses.c.user_id == users.c.user_id)
|
|
.alias("s")
|
|
)
|
|
|
|
# don't correlate in a FROM list
|
|
self.assert_compile(
|
|
select(users, s.c.street).select_from(s),
|
|
"SELECT users.user_id, users.user_name, users.password, s.street "
|
|
"FROM (SELECT addresses.street AS street FROM addresses, users "
|
|
"WHERE addresses.user_id = users.user_id) AS s, users",
|
|
)
|
|
self.assert_compile(
|
|
table1.select().where(
|
|
table1.c.myid
|
|
== select(table1.c.myid)
|
|
.where(table1.c.name == "jack")
|
|
.scalar_subquery()
|
|
),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM mytable WHERE "
|
|
"mytable.myid = (SELECT mytable.myid FROM "
|
|
"mytable WHERE mytable.name = :name_1)",
|
|
)
|
|
self.assert_compile(
|
|
table1.select().where(
|
|
table1.c.myid
|
|
== select(table2.c.otherid)
|
|
.where(table1.c.name == table2.c.othername)
|
|
.scalar_subquery()
|
|
),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM mytable WHERE "
|
|
"mytable.myid = (SELECT "
|
|
"myothertable.otherid FROM myothertable "
|
|
"WHERE mytable.name = myothertable.othernam"
|
|
"e)",
|
|
)
|
|
self.assert_compile(
|
|
table1.select().where(
|
|
exists(1).where(table2.c.otherid == table1.c.myid)
|
|
),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM mytable WHERE "
|
|
"EXISTS (SELECT 1 FROM myothertable WHERE "
|
|
"myothertable.otherid = mytable.myid)",
|
|
)
|
|
talias = table1.alias("ta")
|
|
s = (
|
|
select(talias)
|
|
.where(exists(1).where(table2.c.otherid == talias.c.myid))
|
|
.subquery("sq2")
|
|
)
|
|
self.assert_compile(
|
|
select(s, table1),
|
|
"SELECT sq2.myid, sq2.name, "
|
|
"sq2.description, mytable.myid AS myid_1, "
|
|
"mytable.name AS name_1, "
|
|
"mytable.description AS description_1 FROM "
|
|
"(SELECT ta.myid AS myid, ta.name AS name, "
|
|
"ta.description AS description FROM "
|
|
"mytable AS ta WHERE EXISTS (SELECT 1 FROM "
|
|
"myothertable WHERE myothertable.otherid = "
|
|
"ta.myid)) AS sq2, mytable",
|
|
)
|
|
|
|
# test constructing the outer query via append_column(), which
|
|
# occurs in the ORM's Query object
|
|
|
|
s = (
|
|
select()
|
|
.where(exists(1).where(table2.c.otherid == table1.c.myid))
|
|
.select_from(table1)
|
|
)
|
|
s.add_columns.non_generative(s, table1)
|
|
self.assert_compile(
|
|
s,
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM mytable WHERE "
|
|
"EXISTS (SELECT 1 FROM myothertable WHERE "
|
|
"myothertable.otherid = mytable.myid)",
|
|
)
|
|
|
|
def test_orderby_subquery(self):
|
|
self.assert_compile(
|
|
table1.select().order_by(
|
|
select(table2.c.otherid)
|
|
.where(table1.c.myid == table2.c.otherid)
|
|
.scalar_subquery()
|
|
),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM mytable ORDER BY "
|
|
"(SELECT myothertable.otherid FROM "
|
|
"myothertable WHERE mytable.myid = "
|
|
"myothertable.otherid)",
|
|
)
|
|
self.assert_compile(
|
|
table1.select().order_by(
|
|
desc(
|
|
select(table2.c.otherid)
|
|
.where(table1.c.myid == table2.c.otherid)
|
|
.scalar_subquery()
|
|
)
|
|
),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM mytable ORDER BY "
|
|
"(SELECT myothertable.otherid FROM "
|
|
"myothertable WHERE mytable.myid = "
|
|
"myothertable.otherid) DESC",
|
|
)
|
|
|
|
def test_scalar_select(self):
|
|
s = select(table1.c.myid).correlate(None).scalar_subquery()
|
|
self.assert_compile(
|
|
select(table1, s),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description, (SELECT mytable.myid "
|
|
"FROM mytable) AS anon_1 FROM mytable",
|
|
)
|
|
s = select(table1.c.myid).scalar_subquery()
|
|
self.assert_compile(
|
|
select(table2, s),
|
|
"SELECT myothertable.otherid, "
|
|
"myothertable.othername, (SELECT "
|
|
"mytable.myid FROM mytable) AS anon_1 FROM "
|
|
"myothertable",
|
|
)
|
|
s = select(table1.c.myid).correlate(None).scalar_subquery()
|
|
self.assert_compile(
|
|
select(table1, s),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description, (SELECT mytable.myid "
|
|
"FROM mytable) AS anon_1 FROM mytable",
|
|
)
|
|
|
|
s = select(table1.c.myid).scalar_subquery()
|
|
s2 = s.where(table1.c.myid == 5)
|
|
self.assert_compile(
|
|
s2,
|
|
"(SELECT mytable.myid FROM mytable WHERE mytable.myid = :myid_1)",
|
|
)
|
|
self.assert_compile(s, "(SELECT mytable.myid FROM mytable)")
|
|
# test that aliases use scalar_subquery() when used in an explicitly
|
|
# scalar context
|
|
|
|
s = select(table1.c.myid).scalar_subquery()
|
|
self.assert_compile(
|
|
select(table1.c.myid).where(table1.c.myid == s),
|
|
"SELECT mytable.myid FROM mytable WHERE "
|
|
"mytable.myid = (SELECT mytable.myid FROM "
|
|
"mytable)",
|
|
)
|
|
self.assert_compile(
|
|
select(table1.c.myid).where(table1.c.myid < s),
|
|
"SELECT mytable.myid FROM mytable WHERE "
|
|
"mytable.myid < (SELECT mytable.myid FROM "
|
|
"mytable)",
|
|
)
|
|
s = select(table1.c.myid).scalar_subquery()
|
|
self.assert_compile(
|
|
select(table2, s),
|
|
"SELECT myothertable.otherid, "
|
|
"myothertable.othername, (SELECT "
|
|
"mytable.myid FROM mytable) AS anon_1 FROM "
|
|
"myothertable",
|
|
)
|
|
|
|
# test expressions against scalar selects
|
|
|
|
self.assert_compile(
|
|
select(s - literal(8)),
|
|
"SELECT (SELECT mytable.myid FROM mytable) "
|
|
"- :param_1 AS anon_1",
|
|
)
|
|
self.assert_compile(
|
|
select(select(table1.c.name).scalar_subquery() + literal("x")),
|
|
"SELECT (SELECT mytable.name FROM mytable) "
|
|
"|| :param_1 AS anon_1",
|
|
)
|
|
self.assert_compile(
|
|
select(s > literal(8)),
|
|
"SELECT (SELECT mytable.myid FROM mytable) "
|
|
"> :param_1 AS anon_1",
|
|
)
|
|
self.assert_compile(
|
|
select(select(table1.c.name).label("foo")),
|
|
"SELECT (SELECT mytable.name FROM mytable) AS foo",
|
|
)
|
|
|
|
# scalar selects should not have any attributes on their 'c' or
|
|
# 'columns' attribute
|
|
|
|
s = select(table1.c.myid).scalar_subquery()
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Scalar Select expression has no columns; use this "
|
|
"object directly within a column-level expression.",
|
|
lambda: s.c.foo,
|
|
)
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Scalar Select expression has no columns; use this "
|
|
"object directly within a column-level expression.",
|
|
lambda: s.columns.foo,
|
|
)
|
|
|
|
zips = table(
|
|
"zips", column("zipcode"), column("latitude"), column("longitude")
|
|
)
|
|
places = table("places", column("id"), column("nm"))
|
|
zipcode = "12345"
|
|
qlat = (
|
|
select(zips.c.latitude)
|
|
.where(zips.c.zipcode == zipcode)
|
|
.correlate(None)
|
|
.scalar_subquery()
|
|
)
|
|
qlng = (
|
|
select(zips.c.longitude)
|
|
.where(zips.c.zipcode == zipcode)
|
|
.correlate(None)
|
|
.scalar_subquery()
|
|
)
|
|
|
|
q = (
|
|
select(
|
|
places.c.id,
|
|
places.c.nm,
|
|
zips.c.zipcode,
|
|
func.latlondist(qlat, qlng).label("dist"),
|
|
)
|
|
.where(zips.c.zipcode == zipcode)
|
|
.order_by("dist", places.c.nm)
|
|
)
|
|
|
|
self.assert_compile(
|
|
q,
|
|
"SELECT places.id, places.nm, "
|
|
"zips.zipcode, latlondist((SELECT "
|
|
"zips.latitude FROM zips WHERE "
|
|
"zips.zipcode = :zipcode_1), (SELECT "
|
|
"zips.longitude FROM zips WHERE "
|
|
"zips.zipcode = :zipcode_2)) AS dist FROM "
|
|
"places, zips WHERE zips.zipcode = "
|
|
":zipcode_3 ORDER BY dist, places.nm",
|
|
)
|
|
|
|
zalias = zips.alias("main_zip")
|
|
qlat = (
|
|
select(zips.c.latitude)
|
|
.where(zips.c.zipcode == zalias.c.zipcode)
|
|
.scalar_subquery()
|
|
)
|
|
qlng = (
|
|
select(zips.c.longitude)
|
|
.where(zips.c.zipcode == zalias.c.zipcode)
|
|
.scalar_subquery()
|
|
)
|
|
q = select(
|
|
places.c.id,
|
|
places.c.nm,
|
|
zalias.c.zipcode,
|
|
func.latlondist(qlat, qlng).label("dist"),
|
|
).order_by("dist", places.c.nm)
|
|
self.assert_compile(
|
|
q,
|
|
"SELECT places.id, places.nm, "
|
|
"main_zip.zipcode, latlondist((SELECT "
|
|
"zips.latitude FROM zips WHERE "
|
|
"zips.zipcode = main_zip.zipcode), (SELECT "
|
|
"zips.longitude FROM zips WHERE "
|
|
"zips.zipcode = main_zip.zipcode)) AS dist "
|
|
"FROM places, zips AS main_zip ORDER BY "
|
|
"dist, places.nm",
|
|
)
|
|
|
|
a1 = table2.alias("t2alias")
|
|
s1 = (
|
|
select(a1.c.otherid)
|
|
.where(table1.c.myid == a1.c.otherid)
|
|
.scalar_subquery()
|
|
)
|
|
j1 = table1.join(table2, table1.c.myid == table2.c.otherid)
|
|
s2 = select(table1, s1).select_from(j1)
|
|
self.assert_compile(
|
|
s2,
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description, (SELECT "
|
|
"t2alias.otherid FROM myothertable AS "
|
|
"t2alias WHERE mytable.myid = "
|
|
"t2alias.otherid) AS anon_1 FROM mytable "
|
|
"JOIN myothertable ON mytable.myid = "
|
|
"myothertable.otherid",
|
|
)
|
|
|
|
def test_label_comparison_one(self):
|
|
x = func.lala(table1.c.myid).label("foo")
|
|
self.assert_compile(
|
|
select(x).where(x == 5),
|
|
"SELECT lala(mytable.myid) AS foo FROM "
|
|
"mytable WHERE lala(mytable.myid) = "
|
|
":param_1",
|
|
)
|
|
|
|
def test_label_comparison_two(self):
|
|
self.assert_compile(
|
|
label("bar", column("foo", type_=String)) + "foo",
|
|
"foo || :param_1",
|
|
)
|
|
|
|
def test_order_by_labels_enabled_negative_cases(self):
|
|
"""test order_by_labels enabled but the cases where we expect
|
|
ORDER BY the expression without the label name"""
|
|
|
|
lab1 = (table1.c.myid + 12).label("foo")
|
|
lab2 = func.somefunc(table1.c.name).label("bar")
|
|
dialect = default.DefaultDialect()
|
|
|
|
# binary expressions render as the expression without labels
|
|
self.assert_compile(
|
|
select(lab1, lab2).order_by(lab1 + "test"),
|
|
"SELECT mytable.myid + :myid_1 AS foo, "
|
|
"somefunc(mytable.name) AS bar FROM mytable "
|
|
"ORDER BY mytable.myid + :myid_1 + :param_1",
|
|
dialect=dialect,
|
|
)
|
|
|
|
# labels within functions in the columns clause render
|
|
# with the expression
|
|
self.assert_compile(
|
|
select(lab1, func.foo(lab1)).order_by(lab1, func.foo(lab1)),
|
|
"SELECT mytable.myid + :myid_1 AS foo, "
|
|
"foo(mytable.myid + :myid_1) AS foo_1 FROM mytable "
|
|
"ORDER BY foo, foo(mytable.myid + :myid_1)",
|
|
dialect=dialect,
|
|
)
|
|
|
|
# here, 'name' is implicitly available, but w/ #3882 we don't
|
|
# want to render a name that isn't specifically a Label elsewhere
|
|
# in the query
|
|
self.assert_compile(
|
|
select(table1.c.myid).order_by(table1.c.name.label("name")),
|
|
"SELECT mytable.myid FROM mytable ORDER BY mytable.name",
|
|
)
|
|
|
|
# as well as if it doesn't match
|
|
self.assert_compile(
|
|
select(table1.c.myid).order_by(
|
|
func.lower(table1.c.name).label("name")
|
|
),
|
|
"SELECT mytable.myid FROM mytable ORDER BY lower(mytable.name)",
|
|
)
|
|
|
|
@testing.combinations(
|
|
(desc, "DESC"),
|
|
(asc, "ASC"),
|
|
(nulls_first, "NULLS FIRST"),
|
|
(nulls_last, "NULLS LAST"),
|
|
(nullsfirst, "NULLS FIRST"),
|
|
(nullslast, "NULLS LAST"),
|
|
(lambda c: c.desc().nulls_last(), "DESC NULLS LAST"),
|
|
(lambda c: c.desc().nullslast(), "DESC NULLS LAST"),
|
|
(lambda c: c.nulls_first().asc(), "NULLS FIRST ASC"),
|
|
)
|
|
def test_order_by_labels_enabled(self, operator, expected):
|
|
"""test positive cases with order_by_labels enabled. this is
|
|
multipled out to all the ORDER BY modifier operators
|
|
(see #11592)
|
|
|
|
|
|
"""
|
|
lab1 = (table1.c.myid + 12).label("foo")
|
|
lab2 = func.somefunc(table1.c.name).label("bar")
|
|
dialect = default.DefaultDialect()
|
|
|
|
self.assert_compile(
|
|
select(lab1, lab2).order_by(lab1, operator(lab2)),
|
|
"SELECT mytable.myid + :myid_1 AS foo, "
|
|
"somefunc(mytable.name) AS bar FROM mytable "
|
|
f"ORDER BY foo, bar {expected}",
|
|
dialect=dialect,
|
|
)
|
|
|
|
# the function embedded label renders as the function
|
|
self.assert_compile(
|
|
select(lab1, lab2).order_by(func.hoho(lab1), operator(lab2)),
|
|
"SELECT mytable.myid + :myid_1 AS foo, "
|
|
"somefunc(mytable.name) AS bar FROM mytable "
|
|
f"ORDER BY hoho(mytable.myid + :myid_1), bar {expected}",
|
|
dialect=dialect,
|
|
)
|
|
|
|
lx = (table1.c.myid + table1.c.myid).label("lx")
|
|
ly = (func.lower(table1.c.name) + table1.c.description).label("ly")
|
|
|
|
self.assert_compile(
|
|
select(lx, ly).order_by(lx, operator(ly)),
|
|
"SELECT mytable.myid + mytable.myid AS lx, "
|
|
"lower(mytable.name) || mytable.description AS ly "
|
|
f"FROM mytable ORDER BY lx, ly {expected}",
|
|
dialect=dialect,
|
|
)
|
|
|
|
# expression isn't actually the same thing (even though label is)
|
|
self.assert_compile(
|
|
select(lab1, lab2).order_by(
|
|
table1.c.myid.label("foo"),
|
|
operator(table1.c.name.label("bar")),
|
|
),
|
|
"SELECT mytable.myid + :myid_1 AS foo, "
|
|
"somefunc(mytable.name) AS bar FROM mytable "
|
|
f"ORDER BY mytable.myid, mytable.name {expected}",
|
|
dialect=dialect,
|
|
)
|
|
|
|
# it's also an exact match, not aliased etc.
|
|
self.assert_compile(
|
|
select(lab1, lab2).order_by(
|
|
operator(table1.alias().c.name.label("bar"))
|
|
),
|
|
"SELECT mytable.myid + :myid_1 AS foo, "
|
|
"somefunc(mytable.name) AS bar FROM mytable "
|
|
f"ORDER BY mytable_1.name {expected}",
|
|
dialect=dialect,
|
|
)
|
|
|
|
# but! it's based on lineage
|
|
lab2_lineage = lab2.element._clone()
|
|
self.assert_compile(
|
|
select(lab1, lab2).order_by(operator(lab2_lineage.label("bar"))),
|
|
"SELECT mytable.myid + :myid_1 AS foo, "
|
|
"somefunc(mytable.name) AS bar FROM mytable "
|
|
f"ORDER BY bar {expected}",
|
|
dialect=dialect,
|
|
)
|
|
|
|
def test_order_by_labels_disabled(self):
|
|
"""test when the order_by_labels feature is disabled entirely"""
|
|
|
|
lab1 = (table1.c.myid + 12).label("foo")
|
|
lab2 = func.somefunc(table1.c.name).label("bar")
|
|
dialect = default.DefaultDialect()
|
|
dialect.supports_simple_order_by_label = False
|
|
self.assert_compile(
|
|
select(lab1, lab2).order_by(lab1, desc(lab2)),
|
|
"SELECT mytable.myid + :myid_1 AS foo, "
|
|
"somefunc(mytable.name) AS bar FROM mytable "
|
|
"ORDER BY mytable.myid + :myid_1, somefunc(mytable.name) DESC",
|
|
dialect=dialect,
|
|
)
|
|
self.assert_compile(
|
|
select(lab1, lab2).order_by(func.hoho(lab1), desc(lab2)),
|
|
"SELECT mytable.myid + :myid_1 AS foo, "
|
|
"somefunc(mytable.name) AS bar FROM mytable "
|
|
"ORDER BY hoho(mytable.myid + :myid_1), "
|
|
"somefunc(mytable.name) DESC",
|
|
dialect=dialect,
|
|
)
|
|
|
|
def test_no_group_by_labels(self):
|
|
lab1 = (table1.c.myid + 12).label("foo")
|
|
lab2 = func.somefunc(table1.c.name).label("bar")
|
|
dialect = default.DefaultDialect()
|
|
|
|
self.assert_compile(
|
|
select(lab1, lab2).group_by(lab1, lab2),
|
|
"SELECT mytable.myid + :myid_1 AS foo, somefunc(mytable.name) "
|
|
"AS bar FROM mytable GROUP BY mytable.myid + :myid_1, "
|
|
"somefunc(mytable.name)",
|
|
dialect=dialect,
|
|
)
|
|
|
|
def test_conjunctions(self):
|
|
a, b, c = text("a"), text("b"), text("c")
|
|
x = and_(a, b, c)
|
|
assert isinstance(x.type, Boolean)
|
|
assert str(x) == "a AND b AND c"
|
|
self.assert_compile(
|
|
select(x.label("foo")), "SELECT a AND b AND c AS foo"
|
|
)
|
|
|
|
self.assert_compile(
|
|
and_(
|
|
table1.c.myid == 12,
|
|
table1.c.name == "asdf",
|
|
table2.c.othername == "foo",
|
|
text("sysdate() = today()"),
|
|
),
|
|
"mytable.myid = :myid_1 AND mytable.name = :name_1 "
|
|
"AND myothertable.othername = "
|
|
":othername_1 AND sysdate() = today()",
|
|
)
|
|
|
|
self.assert_compile(
|
|
and_(
|
|
table1.c.myid == 12,
|
|
or_(
|
|
table2.c.othername == "asdf",
|
|
table2.c.othername == "foo",
|
|
table2.c.otherid == 9,
|
|
),
|
|
text("sysdate() = today()"),
|
|
),
|
|
"mytable.myid = :myid_1 AND (myothertable.othername = "
|
|
":othername_1 OR myothertable.othername = :othername_2 OR "
|
|
"myothertable.otherid = :otherid_1) AND sysdate() = "
|
|
"today()",
|
|
checkparams={
|
|
"othername_1": "asdf",
|
|
"othername_2": "foo",
|
|
"otherid_1": 9,
|
|
"myid_1": 12,
|
|
},
|
|
)
|
|
|
|
# test a generator
|
|
self.assert_compile(
|
|
and_(
|
|
conj for conj in [table1.c.myid == 12, table1.c.name == "asdf"]
|
|
),
|
|
"mytable.myid = :myid_1 AND mytable.name = :name_1",
|
|
)
|
|
|
|
def test_nested_conjunctions_short_circuit(self):
|
|
"""test that empty or_(), and_() conjunctions are collapsed by
|
|
an enclosing conjunction."""
|
|
|
|
t = table("t", column("x"))
|
|
|
|
self.assert_compile(
|
|
select(t).where(and_(t.c.x == 5, or_(and_(or_(t.c.x == 7))))),
|
|
"SELECT t.x FROM t WHERE t.x = :x_1 AND t.x = :x_2",
|
|
)
|
|
self.assert_compile(
|
|
select(t).where(and_(or_(t.c.x == 12, and_(or_(t.c.x == 8))))),
|
|
"SELECT t.x FROM t WHERE t.x = :x_1 OR t.x = :x_2",
|
|
)
|
|
self.assert_compile(
|
|
select(t).where(
|
|
and_(or_(or_(t.c.x == 12), and_(or_(and_(t.c.x == 8)))))
|
|
),
|
|
"SELECT t.x FROM t WHERE t.x = :x_1 OR t.x = :x_2",
|
|
)
|
|
self.assert_compile(
|
|
select(t).where(
|
|
and_(
|
|
or_(
|
|
or_(t.c.x == 12),
|
|
and_(
|
|
BooleanClauseList._construct_raw(operators.or_),
|
|
or_(and_(t.c.x == 8)),
|
|
BooleanClauseList._construct_raw(operators.and_),
|
|
),
|
|
)
|
|
)
|
|
),
|
|
"SELECT t.x FROM t WHERE t.x = :x_1 OR t.x = :x_2",
|
|
)
|
|
|
|
def test_true_short_circuit(self):
|
|
t = table("t", column("x"))
|
|
|
|
self.assert_compile(
|
|
select(t).where(true()),
|
|
"SELECT t.x FROM t WHERE 1 = 1",
|
|
dialect=default.DefaultDialect(supports_native_boolean=False),
|
|
)
|
|
self.assert_compile(
|
|
select(t).where(true()),
|
|
"SELECT t.x FROM t WHERE true",
|
|
dialect=default.DefaultDialect(supports_native_boolean=True),
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(t),
|
|
"SELECT t.x FROM t",
|
|
dialect=default.DefaultDialect(supports_native_boolean=True),
|
|
)
|
|
|
|
def test_distinct(self):
|
|
self.assert_compile(
|
|
select(table1.c.myid.distinct()),
|
|
"SELECT DISTINCT mytable.myid FROM mytable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(distinct(table1.c.myid)),
|
|
"SELECT DISTINCT mytable.myid FROM mytable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(distinct(table1.c.myid)).set_label_style(
|
|
LABEL_STYLE_TABLENAME_PLUS_COL
|
|
),
|
|
"SELECT DISTINCT mytable.myid FROM mytable",
|
|
)
|
|
|
|
# the bug fixed here as part of #6008 is the same bug that's
|
|
# in 1.3 as well, producing
|
|
# "SELECT anon_2.anon_1 FROM (SELECT distinct mytable.myid
|
|
# FROM mytable) AS anon_2"
|
|
self.assert_compile(
|
|
select(select(distinct(table1.c.myid)).subquery()),
|
|
"SELECT anon_2.anon_1 FROM (SELECT "
|
|
"DISTINCT mytable.myid AS anon_1 FROM mytable) AS anon_2",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(table1.c.myid).distinct(),
|
|
"SELECT DISTINCT mytable.myid FROM mytable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(func.count(table1.c.myid.distinct())),
|
|
"SELECT count(DISTINCT mytable.myid) AS count_1 FROM mytable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(func.count(distinct(table1.c.myid))),
|
|
"SELECT count(DISTINCT mytable.myid) AS count_1 FROM mytable",
|
|
)
|
|
|
|
def test_distinct_on(self):
|
|
with testing.expect_deprecated(
|
|
"Passing expression to",
|
|
"DISTINCT ON is currently supported only by the PostgreSQL "
|
|
"dialect",
|
|
):
|
|
select("*").distinct(table1.c.myid).compile()
|
|
|
|
def test_where_empty(self):
|
|
self.assert_compile(
|
|
select(table1.c.myid).where(
|
|
BooleanClauseList._construct_raw(operators.and_)
|
|
),
|
|
"SELECT mytable.myid FROM mytable",
|
|
)
|
|
self.assert_compile(
|
|
select(table1.c.myid).where(
|
|
BooleanClauseList._construct_raw(operators.or_)
|
|
),
|
|
"SELECT mytable.myid FROM mytable",
|
|
)
|
|
|
|
def test_where_multiple(self):
|
|
self.assert_compile(
|
|
select(table1.c.myid).where(
|
|
table1.c.myid == 12, table1.c.name == "foobar"
|
|
),
|
|
"SELECT mytable.myid FROM mytable WHERE mytable.myid = :myid_1 "
|
|
"AND mytable.name = :name_1",
|
|
)
|
|
|
|
def test_order_by_nulls(self):
|
|
self.assert_compile(
|
|
table2.select().order_by(
|
|
table2.c.otherid,
|
|
table2.c.othername.desc().nulls_first(),
|
|
),
|
|
"SELECT myothertable.otherid, myothertable.othername FROM "
|
|
"myothertable ORDER BY myothertable.otherid, "
|
|
"myothertable.othername DESC NULLS FIRST",
|
|
)
|
|
|
|
self.assert_compile(
|
|
table2.select().order_by(
|
|
table2.c.otherid,
|
|
table2.c.othername.desc().nulls_last(),
|
|
),
|
|
"SELECT myothertable.otherid, myothertable.othername FROM "
|
|
"myothertable ORDER BY myothertable.otherid, "
|
|
"myothertable.othername DESC NULLS LAST",
|
|
)
|
|
|
|
self.assert_compile(
|
|
table2.select().order_by(
|
|
table2.c.otherid.nulls_last(),
|
|
table2.c.othername.desc().nulls_first(),
|
|
),
|
|
"SELECT myothertable.otherid, myothertable.othername FROM "
|
|
"myothertable ORDER BY myothertable.otherid NULLS LAST, "
|
|
"myothertable.othername DESC NULLS FIRST",
|
|
)
|
|
|
|
self.assert_compile(
|
|
table2.select().order_by(
|
|
table2.c.otherid.nulls_first(),
|
|
table2.c.othername.desc(),
|
|
),
|
|
"SELECT myothertable.otherid, myothertable.othername FROM "
|
|
"myothertable ORDER BY myothertable.otherid NULLS FIRST, "
|
|
"myothertable.othername DESC",
|
|
)
|
|
|
|
self.assert_compile(
|
|
table2.select().order_by(
|
|
table2.c.otherid.nulls_first(),
|
|
table2.c.othername.desc().nulls_last(),
|
|
),
|
|
"SELECT myothertable.otherid, myothertable.othername FROM "
|
|
"myothertable ORDER BY myothertable.otherid NULLS FIRST, "
|
|
"myothertable.othername DESC NULLS LAST",
|
|
)
|
|
|
|
def test_orderby_groupby(self):
|
|
self.assert_compile(
|
|
table2.select().order_by(
|
|
table2.c.otherid, asc(table2.c.othername)
|
|
),
|
|
"SELECT myothertable.otherid, myothertable.othername FROM "
|
|
"myothertable ORDER BY myothertable.otherid, "
|
|
"myothertable.othername ASC",
|
|
)
|
|
|
|
self.assert_compile(
|
|
table2.select().order_by(
|
|
table2.c.otherid, table2.c.othername.desc()
|
|
),
|
|
"SELECT myothertable.otherid, myothertable.othername FROM "
|
|
"myothertable ORDER BY myothertable.otherid, "
|
|
"myothertable.othername DESC",
|
|
)
|
|
|
|
# generative order_by
|
|
self.assert_compile(
|
|
table2.select()
|
|
.order_by(table2.c.otherid)
|
|
.order_by(table2.c.othername.desc()),
|
|
"SELECT myothertable.otherid, myothertable.othername FROM "
|
|
"myothertable ORDER BY myothertable.otherid, "
|
|
"myothertable.othername DESC",
|
|
)
|
|
|
|
self.assert_compile(
|
|
table2.select()
|
|
.order_by(table2.c.otherid)
|
|
.order_by(table2.c.othername.desc())
|
|
.order_by(None),
|
|
"SELECT myothertable.otherid, myothertable.othername "
|
|
"FROM myothertable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(table2.c.othername, func.count(table2.c.otherid)).group_by(
|
|
table2.c.othername
|
|
),
|
|
"SELECT myothertable.othername, "
|
|
"count(myothertable.otherid) AS count_1 "
|
|
"FROM myothertable GROUP BY myothertable.othername",
|
|
)
|
|
|
|
# generative group by
|
|
self.assert_compile(
|
|
select(table2.c.othername, func.count(table2.c.otherid)).group_by(
|
|
table2.c.othername
|
|
),
|
|
"SELECT myothertable.othername, "
|
|
"count(myothertable.otherid) AS count_1 "
|
|
"FROM myothertable GROUP BY myothertable.othername",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(table2.c.othername, func.count(table2.c.otherid))
|
|
.group_by(table2.c.othername)
|
|
.group_by(None),
|
|
"SELECT myothertable.othername, "
|
|
"count(myothertable.otherid) AS count_1 "
|
|
"FROM myothertable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(table2.c.othername, func.count(table2.c.otherid))
|
|
.group_by(table2.c.othername)
|
|
.order_by(table2.c.othername),
|
|
"SELECT myothertable.othername, "
|
|
"count(myothertable.otherid) AS count_1 "
|
|
"FROM myothertable "
|
|
"GROUP BY myothertable.othername ORDER BY myothertable.othername",
|
|
)
|
|
|
|
def test_custom_order_by_clause(self):
|
|
class CustomCompiler(PGCompiler):
|
|
def order_by_clause(self, select, **kw):
|
|
return super().order_by_clause(select, **kw) + " CUSTOMIZED"
|
|
|
|
class CustomDialect(PGDialect):
|
|
name = "custom"
|
|
statement_compiler = CustomCompiler
|
|
|
|
stmt = select(table1.c.myid).order_by(table1.c.myid)
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid FROM mytable ORDER BY "
|
|
"mytable.myid CUSTOMIZED",
|
|
dialect=CustomDialect(),
|
|
)
|
|
|
|
def test_custom_group_by_clause(self):
|
|
class CustomCompiler(PGCompiler):
|
|
def group_by_clause(self, select, **kw):
|
|
return super().group_by_clause(select, **kw) + " CUSTOMIZED"
|
|
|
|
class CustomDialect(PGDialect):
|
|
name = "custom"
|
|
statement_compiler = CustomCompiler
|
|
|
|
stmt = select(table1.c.myid).group_by(table1.c.myid)
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid FROM mytable GROUP BY "
|
|
"mytable.myid CUSTOMIZED",
|
|
dialect=CustomDialect(),
|
|
)
|
|
|
|
def test_for_update(self):
|
|
self.assert_compile(
|
|
table1.select().where(table1.c.myid == 7).with_for_update(),
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable WHERE mytable.myid = :myid_1 FOR UPDATE",
|
|
)
|
|
|
|
# not supported by dialect, should just use update
|
|
self.assert_compile(
|
|
table1.select()
|
|
.where(table1.c.myid == 7)
|
|
.with_for_update(nowait=True),
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable WHERE mytable.myid = :myid_1 FOR UPDATE",
|
|
)
|
|
|
|
def test_alias(self):
|
|
# test the alias for a table1. column names stay the same,
|
|
# table name "changes" to "foo".
|
|
self.assert_compile(
|
|
select(table1.alias("foo")),
|
|
"SELECT foo.myid, foo.name, foo.description FROM mytable AS foo",
|
|
)
|
|
|
|
for dialect in (oracle.dialect(),):
|
|
self.assert_compile(
|
|
select(table1.alias("foo")),
|
|
"SELECT foo.myid, foo.name, foo.description FROM mytable foo",
|
|
dialect=dialect,
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(table1.alias()),
|
|
"SELECT mytable_1.myid, mytable_1.name, mytable_1.description "
|
|
"FROM mytable AS mytable_1",
|
|
)
|
|
|
|
# create a select for a join of two tables. use_labels
|
|
# means the column names will have labels tablename_columnname,
|
|
# which become the column keys accessible off the Selectable object.
|
|
# also, only use one column from the second table and all columns
|
|
# from the first table1.
|
|
q = (
|
|
select(table1, table2.c.otherid)
|
|
.where(table1.c.myid == table2.c.otherid)
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
)
|
|
|
|
# make an alias of the "selectable". column names
|
|
# stay the same (i.e. the labels), table name "changes" to "t2view".
|
|
a = q.alias("t2view")
|
|
|
|
# select from that alias, also using labels. two levels of labels
|
|
# should produce two underscores.
|
|
# also, reference the column "mytable_myid" off of the t2view alias.
|
|
self.assert_compile(
|
|
a.select()
|
|
.where(a.c.mytable_myid == 9)
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL),
|
|
"SELECT t2view.mytable_myid AS t2view_mytable_myid, "
|
|
"t2view.mytable_name "
|
|
"AS t2view_mytable_name, "
|
|
"t2view.mytable_description AS t2view_mytable_description, "
|
|
"t2view.myothertable_otherid AS t2view_myothertable_otherid FROM "
|
|
"(SELECT mytable.myid AS mytable_myid, "
|
|
"mytable.name AS mytable_name, "
|
|
"mytable.description AS mytable_description, "
|
|
"myothertable.otherid AS "
|
|
"myothertable_otherid FROM mytable, myothertable "
|
|
"WHERE mytable.myid = "
|
|
"myothertable.otherid) AS t2view "
|
|
"WHERE t2view.mytable_myid = :mytable_myid_1",
|
|
)
|
|
|
|
def test_alias_nesting_table(self):
|
|
self.assert_compile(
|
|
select(table1.alias("foo").alias("bar").alias("bat")),
|
|
"SELECT bat.myid, bat.name, bat.description FROM mytable AS bat",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(table1.alias(None).alias("bar").alias("bat")),
|
|
"SELECT bat.myid, bat.name, bat.description FROM mytable AS bat",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(table1.alias("foo").alias(None).alias("bat")),
|
|
"SELECT bat.myid, bat.name, bat.description FROM mytable AS bat",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(table1.alias("foo").alias("bar").alias(None)),
|
|
"SELECT bar_1.myid, bar_1.name, bar_1.description "
|
|
"FROM mytable AS bar_1",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(table1.alias("foo").alias(None).alias(None)),
|
|
"SELECT anon_1.myid, anon_1.name, anon_1.description "
|
|
"FROM mytable AS anon_1",
|
|
)
|
|
|
|
def test_alias_nesting_subquery(self):
|
|
stmt = select(table1).subquery()
|
|
self.assert_compile(
|
|
select(stmt.alias("foo").alias("bar").alias("bat")),
|
|
"SELECT bat.myid, bat.name, bat.description FROM "
|
|
"(SELECT mytable.myid AS myid, mytable.name AS name, "
|
|
"mytable.description AS description FROM mytable) AS bat",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(stmt.alias("foo").alias(None).alias(None)),
|
|
"SELECT anon_1.myid, anon_1.name, anon_1.description FROM "
|
|
"(SELECT mytable.myid AS myid, mytable.name AS name, "
|
|
"mytable.description AS description FROM mytable) AS anon_1",
|
|
)
|
|
|
|
def test_prefix(self):
|
|
self.assert_compile(
|
|
table1.select()
|
|
.prefix_with("SQL_CALC_FOUND_ROWS")
|
|
.prefix_with("SQL_SOME_WEIRD_MYSQL_THING"),
|
|
"SELECT SQL_CALC_FOUND_ROWS SQL_SOME_WEIRD_MYSQL_THING "
|
|
"mytable.myid, mytable.name, mytable.description FROM mytable",
|
|
)
|
|
|
|
def test_prefix_dialect_specific(self):
|
|
self.assert_compile(
|
|
table1.select()
|
|
.prefix_with("SQL_CALC_FOUND_ROWS", dialect="sqlite")
|
|
.prefix_with("SQL_SOME_WEIRD_MYSQL_THING", dialect="mysql"),
|
|
"SELECT SQL_SOME_WEIRD_MYSQL_THING "
|
|
"mytable.myid, mytable.name, mytable.description FROM mytable",
|
|
dialect=mysql.dialect(),
|
|
)
|
|
|
|
def test_collate(self):
|
|
# columns clause
|
|
self.assert_compile(
|
|
select(column("x").collate("bar")),
|
|
"SELECT x COLLATE bar AS anon_1",
|
|
)
|
|
|
|
# WHERE clause
|
|
self.assert_compile(
|
|
select(column("x")).where(column("x").collate("bar") == "foo"),
|
|
"SELECT x WHERE (x COLLATE bar) = :param_1",
|
|
)
|
|
|
|
# ORDER BY clause
|
|
self.assert_compile(
|
|
select(column("x")).order_by(column("x").collate("bar")),
|
|
"SELECT x ORDER BY x COLLATE bar",
|
|
)
|
|
|
|
def test_literal(self):
|
|
self.assert_compile(
|
|
select(literal("foo")), "SELECT :param_1 AS anon_1"
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(literal("foo") + literal("bar")).select_from(table1),
|
|
"SELECT :param_1 || :param_2 AS anon_1 FROM mytable",
|
|
)
|
|
|
|
def test_calculated_columns(self):
|
|
value_tbl = table(
|
|
"values",
|
|
column("id", Integer),
|
|
column("val1", Float),
|
|
column("val2", Float),
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(
|
|
value_tbl.c.id,
|
|
(value_tbl.c.val2 - value_tbl.c.val1) / value_tbl.c.val1,
|
|
),
|
|
"SELECT values.id, (values.val2 - values.val1) "
|
|
"/ CAST(values.val1 AS FLOAT) AS anon_1 FROM values",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(value_tbl.c.id).where(
|
|
(value_tbl.c.val2 - value_tbl.c.val1) / value_tbl.c.val1 > 2.0,
|
|
),
|
|
"SELECT values.id FROM values WHERE "
|
|
"(values.val2 - values.val1) / "
|
|
"CAST(values.val1 AS FLOAT) > :param_1",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(value_tbl.c.id).where(
|
|
value_tbl.c.val1
|
|
/ (value_tbl.c.val2 - value_tbl.c.val1)
|
|
/ value_tbl.c.val1
|
|
> 2.0,
|
|
),
|
|
"SELECT values.id FROM values WHERE "
|
|
"(values.val1 / CAST((values.val2 - values.val1) AS FLOAT)) "
|
|
"/ CAST(values.val1 AS FLOAT) > :param_1",
|
|
)
|
|
|
|
def test_percent_chars(self):
|
|
t = table(
|
|
"table%name",
|
|
column("percent%"),
|
|
column("%(oneofthese)s"),
|
|
column("spaces % more spaces"),
|
|
)
|
|
self.assert_compile(
|
|
t.select().set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL),
|
|
"""SELECT "table%name"."percent%" AS "table%name_percent%", """
|
|
""""table%name"."%(oneofthese)s" AS """
|
|
""""table%name_%(oneofthese)s", """
|
|
""""table%name"."spaces % more spaces" AS """
|
|
""""table%name_spaces % """
|
|
'''more spaces" FROM "table%name"''',
|
|
)
|
|
|
|
def test_joins(self):
|
|
self.assert_compile(
|
|
join(table2, table1, table1.c.myid == table2.c.otherid).select(),
|
|
"SELECT myothertable.otherid, myothertable.othername, "
|
|
"mytable.myid, mytable.name, mytable.description FROM "
|
|
"myothertable JOIN mytable ON mytable.myid = myothertable.otherid",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(table1).select_from(
|
|
join(table1, table2, table1.c.myid == table2.c.otherid)
|
|
),
|
|
"SELECT mytable.myid, mytable.name, mytable.description FROM "
|
|
"mytable JOIN myothertable ON mytable.myid = myothertable.otherid",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(
|
|
join(
|
|
join(table1, table2, table1.c.myid == table2.c.otherid),
|
|
table3,
|
|
table1.c.myid == table3.c.userid,
|
|
)
|
|
),
|
|
"SELECT mytable.myid, mytable.name, mytable.description, "
|
|
"myothertable.otherid, myothertable.othername, "
|
|
"thirdtable.userid, "
|
|
"thirdtable.otherstuff FROM mytable JOIN myothertable "
|
|
"ON mytable.myid ="
|
|
" myothertable.otherid JOIN thirdtable ON "
|
|
"mytable.myid = thirdtable.userid",
|
|
)
|
|
|
|
self.assert_compile(
|
|
join(
|
|
users, addresses, users.c.user_id == addresses.c.user_id
|
|
).select(),
|
|
"SELECT users.user_id, users.user_name, users.password, "
|
|
"addresses.address_id, addresses.user_id AS user_id_1, "
|
|
"addresses.street, "
|
|
"addresses.city, addresses.state, addresses.zip "
|
|
"FROM users JOIN addresses "
|
|
"ON users.user_id = addresses.user_id",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(table1, table2, table3).select_from(
|
|
join(
|
|
table1, table2, table1.c.myid == table2.c.otherid
|
|
).outerjoin(table3, table1.c.myid == table3.c.userid)
|
|
),
|
|
"SELECT mytable.myid, mytable.name, mytable.description, "
|
|
"myothertable.otherid, myothertable.othername, "
|
|
"thirdtable.userid,"
|
|
" thirdtable.otherstuff FROM mytable "
|
|
"JOIN myothertable ON mytable.myid "
|
|
"= myothertable.otherid LEFT OUTER JOIN thirdtable "
|
|
"ON mytable.myid ="
|
|
" thirdtable.userid",
|
|
)
|
|
self.assert_compile(
|
|
select(table1, table2, table3).select_from(
|
|
outerjoin(
|
|
table1,
|
|
join(table2, table3, table2.c.otherid == table3.c.userid),
|
|
table1.c.myid == table2.c.otherid,
|
|
)
|
|
),
|
|
"SELECT mytable.myid, mytable.name, mytable.description, "
|
|
"myothertable.otherid, myothertable.othername, "
|
|
"thirdtable.userid,"
|
|
" thirdtable.otherstuff FROM mytable LEFT OUTER JOIN "
|
|
"(myothertable "
|
|
"JOIN thirdtable ON myothertable.otherid = "
|
|
"thirdtable.userid) ON "
|
|
"mytable.myid = myothertable.otherid",
|
|
)
|
|
|
|
query = (
|
|
select(table1, table2)
|
|
.where(
|
|
or_(
|
|
table1.c.name == "fred",
|
|
table1.c.myid == 10,
|
|
table2.c.othername != "jack",
|
|
text("EXISTS (select yay from foo where boo = lar)"),
|
|
)
|
|
)
|
|
.select_from(
|
|
outerjoin(table1, table2, table1.c.myid == table2.c.otherid)
|
|
)
|
|
)
|
|
self.assert_compile(
|
|
query,
|
|
"SELECT mytable.myid, mytable.name, mytable.description, "
|
|
"myothertable.otherid, myothertable.othername "
|
|
"FROM mytable LEFT OUTER JOIN myothertable ON mytable.myid = "
|
|
"myothertable.otherid WHERE mytable.name = :name_1 OR "
|
|
"mytable.myid = :myid_1 OR myothertable.othername != :othername_1 "
|
|
"OR EXISTS (select yay from foo where boo = lar)",
|
|
)
|
|
|
|
def test_full_outer_join(self):
|
|
for spec in [
|
|
join(table1, table2, table1.c.myid == table2.c.otherid, full=True),
|
|
outerjoin(
|
|
table1, table2, table1.c.myid == table2.c.otherid, full=True
|
|
),
|
|
table1.join(table2, table1.c.myid == table2.c.otherid, full=True),
|
|
table1.outerjoin(
|
|
table2, table1.c.myid == table2.c.otherid, full=True
|
|
),
|
|
]:
|
|
stmt = select(table1).select_from(spec)
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid, mytable.name, mytable.description FROM "
|
|
"mytable FULL OUTER JOIN myothertable "
|
|
"ON mytable.myid = myothertable.otherid",
|
|
)
|
|
|
|
def test_compound_selects(self):
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
"All selectables passed to CompoundSelect "
|
|
"must have identical numbers of columns; "
|
|
"select #1 has 2 columns, select #2 has 3",
|
|
union(table3.select(), table1.select()).compile,
|
|
)
|
|
|
|
x = union(
|
|
select(table1).where(table1.c.myid == 5),
|
|
select(table1).where(table1.c.myid == 12),
|
|
).order_by(table1.c.myid)
|
|
|
|
self.assert_compile(
|
|
x,
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description "
|
|
"FROM mytable WHERE "
|
|
"mytable.myid = :myid_1 UNION "
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable WHERE mytable.myid = :myid_2 "
|
|
"ORDER BY myid",
|
|
)
|
|
|
|
x = union(select(table1), select(table1))
|
|
x = union(x, select(table1))
|
|
self.assert_compile(
|
|
x,
|
|
"(SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable UNION SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM mytable) UNION SELECT mytable.myid,"
|
|
" mytable.name, mytable.description FROM mytable",
|
|
)
|
|
|
|
u1 = union(
|
|
select(table1.c.myid, table1.c.name),
|
|
select(table2),
|
|
select(table3),
|
|
).order_by("name")
|
|
self.assert_compile(
|
|
u1,
|
|
"SELECT mytable.myid, mytable.name "
|
|
"FROM mytable UNION SELECT myothertable.otherid, "
|
|
"myothertable.othername FROM myothertable "
|
|
"UNION SELECT thirdtable.userid, thirdtable.otherstuff "
|
|
"FROM thirdtable ORDER BY name",
|
|
)
|
|
|
|
u1s = u1.subquery()
|
|
assert u1s.corresponding_column(table2.c.otherid) is u1s.c.myid
|
|
|
|
self.assert_compile(
|
|
union(select(table1.c.myid, table1.c.name), select(table2))
|
|
.order_by("myid")
|
|
.offset(10)
|
|
.limit(5),
|
|
# note table name is omitted here. The CompoundSelect, inside of
|
|
# _label_resolve_dict(), creates a subquery of itself and then
|
|
# turns "named_with_column" off, so that we can order by the
|
|
# "myid" name as relative to the CompoundSelect itself without it
|
|
# having a name.
|
|
"SELECT mytable.myid, mytable.name "
|
|
"FROM mytable UNION SELECT myothertable.otherid, "
|
|
"myothertable.othername "
|
|
"FROM myothertable ORDER BY myid "
|
|
"LIMIT :param_1 OFFSET :param_2",
|
|
{"param_1": 5, "param_2": 10},
|
|
)
|
|
|
|
# these tests are mostly in test_text, however adding one here
|
|
# to check the special thing CompoundSelect does with labels
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
"Can't resolve label reference for ORDER BY / GROUP BY / "
|
|
"DISTINCT etc. Textual "
|
|
"SQL expression 'noname'",
|
|
union(
|
|
select(table1.c.myid, table1.c.name),
|
|
select(table2),
|
|
)
|
|
.order_by("noname")
|
|
.compile,
|
|
)
|
|
|
|
self.assert_compile(
|
|
union(
|
|
select(
|
|
table1.c.myid,
|
|
table1.c.name,
|
|
func.max(table1.c.description),
|
|
)
|
|
.where(table1.c.name == "name2")
|
|
.group_by(table1.c.myid, table1.c.name),
|
|
table1.select().where(table1.c.name == "name1"),
|
|
),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"max(mytable.description) AS max_1 "
|
|
"FROM mytable WHERE mytable.name = :name_1 "
|
|
"GROUP BY mytable.myid, "
|
|
"mytable.name UNION SELECT mytable.myid, mytable.name, "
|
|
"mytable.description "
|
|
"FROM mytable WHERE mytable.name = :name_2",
|
|
)
|
|
|
|
self.assert_compile(
|
|
union(
|
|
select(literal(100).label("value")),
|
|
select(literal(200).label("value")),
|
|
),
|
|
"SELECT :param_1 AS value UNION SELECT :param_2 AS value",
|
|
)
|
|
|
|
self.assert_compile(
|
|
union_all(
|
|
select(table1.c.myid),
|
|
union(select(table2.c.otherid), select(table3.c.userid)),
|
|
),
|
|
"SELECT mytable.myid FROM mytable UNION ALL "
|
|
"(SELECT myothertable.otherid FROM myothertable UNION "
|
|
"SELECT thirdtable.userid FROM thirdtable)",
|
|
)
|
|
|
|
s = select(column("foo"), column("bar"))
|
|
|
|
self.assert_compile(
|
|
union(s.order_by("foo"), s.order_by("bar")),
|
|
"(SELECT foo, bar ORDER BY foo) UNION "
|
|
"(SELECT foo, bar ORDER BY bar)",
|
|
)
|
|
self.assert_compile(
|
|
union(
|
|
s.order_by("foo").self_group(),
|
|
s.order_by("bar").limit(10).self_group(),
|
|
),
|
|
"(SELECT foo, bar ORDER BY foo) UNION (SELECT foo, "
|
|
"bar ORDER BY bar LIMIT :param_1)",
|
|
{"param_1": 10},
|
|
)
|
|
|
|
def test_dupe_cols_hey_we_can_union(self):
|
|
"""test the original inspiration for [ticket:4753]."""
|
|
|
|
s1 = select(table1, table1.c.myid).where(table1.c.myid == 5)
|
|
s2 = select(table1, table2.c.otherid).where(
|
|
table1.c.myid == table2.c.otherid
|
|
)
|
|
|
|
# note myid__1 is a dedupe of same column, same table. see
|
|
# test/sql/test_labels.py for the double underscore thing
|
|
self.assert_compile(
|
|
union(s1, s2).order_by(s1.selected_columns.myid),
|
|
"SELECT mytable.myid, mytable.name, mytable.description, "
|
|
"mytable.myid AS myid__1 FROM mytable "
|
|
"WHERE mytable.myid = :myid_1 "
|
|
"UNION SELECT mytable.myid, mytable.name, mytable.description, "
|
|
"myothertable.otherid FROM mytable, myothertable "
|
|
"WHERE mytable.myid = myothertable.otherid ORDER BY myid",
|
|
)
|
|
|
|
def test_deduping_hash_algo(self):
|
|
"""related to #7153.
|
|
|
|
testing new feature "add_hash" of _anon_label which adds an additional
|
|
integer value as part of what the anon label is deduplicated upon.
|
|
|
|
"""
|
|
|
|
class Thing(ColumnElement):
|
|
def __init__(self, hash_):
|
|
self._hash = hash_
|
|
|
|
def __hash__(self):
|
|
return self._hash
|
|
|
|
t1 = Thing(10)
|
|
t2 = Thing(11)
|
|
|
|
# this is the collision case. therefore we assert that this
|
|
# add_hash has to be below 16 bits.
|
|
# eq_(
|
|
# t1._anon_label('hi', add_hash=65537),
|
|
# t2._anon_label('hi', add_hash=1)
|
|
# )
|
|
with expect_raises(AssertionError):
|
|
t1._anon_label("hi", add_hash=65536)
|
|
|
|
for i in range(50):
|
|
ne_(
|
|
t1._anon_label("hi", add_hash=i),
|
|
t2._anon_label("hi", add_hash=1),
|
|
)
|
|
|
|
def test_deduping_unique_across_selects(self):
|
|
"""related to #7153
|
|
|
|
looking to see that dedupe anon labels use a unique hash not only
|
|
within each statement but across multiple statements.
|
|
|
|
"""
|
|
|
|
s1 = select(null(), null())
|
|
|
|
s2 = select(true(), true())
|
|
|
|
s3 = union(s1, s2)
|
|
|
|
self.assert_compile(
|
|
s3,
|
|
"SELECT NULL AS anon_1, NULL AS anon__1 UNION "
|
|
# without the feature tested in test_deduping_hash_algo we'd get
|
|
# "SELECT true AS anon_2, true AS anon__1",
|
|
"SELECT true AS anon_2, true AS anon__2",
|
|
dialect="default_enhanced",
|
|
)
|
|
|
|
def test_compound_grouping(self):
|
|
s = select(column("foo"), column("bar")).select_from(text("bat"))
|
|
|
|
self.assert_compile(
|
|
union(union(union(s, s), s), s),
|
|
"((SELECT foo, bar FROM bat UNION SELECT foo, bar FROM bat) "
|
|
"UNION SELECT foo, bar FROM bat) UNION SELECT foo, bar FROM bat",
|
|
)
|
|
|
|
self.assert_compile(
|
|
union(s, s, s, s),
|
|
"SELECT foo, bar FROM bat UNION SELECT foo, bar "
|
|
"FROM bat UNION SELECT foo, bar FROM bat "
|
|
"UNION SELECT foo, bar FROM bat",
|
|
)
|
|
|
|
self.assert_compile(
|
|
union(s, union(s, union(s, s))),
|
|
"SELECT foo, bar FROM bat UNION (SELECT foo, bar FROM bat "
|
|
"UNION (SELECT foo, bar FROM bat "
|
|
"UNION SELECT foo, bar FROM bat))",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(s.alias()),
|
|
"SELECT anon_1.foo, anon_1.bar FROM "
|
|
"(SELECT foo, bar FROM bat) AS anon_1",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(union(s, s).alias()),
|
|
"SELECT anon_1.foo, anon_1.bar FROM "
|
|
"(SELECT foo, bar FROM bat UNION "
|
|
"SELECT foo, bar FROM bat) AS anon_1",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(except_(s, s).alias()),
|
|
"SELECT anon_1.foo, anon_1.bar FROM "
|
|
"(SELECT foo, bar FROM bat EXCEPT "
|
|
"SELECT foo, bar FROM bat) AS anon_1",
|
|
)
|
|
|
|
# this query sqlite specifically chokes on
|
|
self.assert_compile(
|
|
union(except_(s, s), s),
|
|
"(SELECT foo, bar FROM bat EXCEPT SELECT foo, bar FROM bat) "
|
|
"UNION SELECT foo, bar FROM bat",
|
|
)
|
|
|
|
self.assert_compile(
|
|
union(s, except_(s, s)),
|
|
"SELECT foo, bar FROM bat "
|
|
"UNION (SELECT foo, bar FROM bat EXCEPT SELECT foo, bar FROM bat)",
|
|
)
|
|
|
|
# this solves it
|
|
self.assert_compile(
|
|
union(except_(s, s).alias().select(), s),
|
|
"SELECT anon_1.foo, anon_1.bar FROM "
|
|
"(SELECT foo, bar FROM bat EXCEPT "
|
|
"SELECT foo, bar FROM bat) AS anon_1 "
|
|
"UNION SELECT foo, bar FROM bat",
|
|
)
|
|
|
|
self.assert_compile(
|
|
except_(union(s, s), union(s, s)),
|
|
"(SELECT foo, bar FROM bat UNION SELECT foo, bar FROM bat) "
|
|
"EXCEPT (SELECT foo, bar FROM bat UNION SELECT foo, bar FROM bat)",
|
|
)
|
|
s2 = union(s, s)
|
|
s3 = union(s2, s2)
|
|
self.assert_compile(
|
|
s3,
|
|
"(SELECT foo, bar FROM bat "
|
|
"UNION SELECT foo, bar FROM bat) "
|
|
"UNION (SELECT foo, bar FROM bat "
|
|
"UNION SELECT foo, bar FROM bat)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
union(intersect(s, s), intersect(s, s)),
|
|
"(SELECT foo, bar FROM bat INTERSECT SELECT foo, bar FROM bat) "
|
|
"UNION (SELECT foo, bar FROM bat INTERSECT "
|
|
"SELECT foo, bar FROM bat)",
|
|
)
|
|
|
|
# tests for [ticket:2528]
|
|
# sqlite hates all of these.
|
|
self.assert_compile(
|
|
union(s.limit(1), s.offset(2)),
|
|
"(SELECT foo, bar FROM bat LIMIT :param_1) "
|
|
"UNION (SELECT foo, bar FROM bat LIMIT -1 OFFSET :param_2)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
union(s.order_by(column("bar")), s.offset(2)),
|
|
"(SELECT foo, bar FROM bat ORDER BY bar) "
|
|
"UNION (SELECT foo, bar FROM bat LIMIT -1 OFFSET :param_1)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
union(
|
|
s.limit(1).alias("a").element, s.limit(2).alias("b").element
|
|
),
|
|
"(SELECT foo, bar FROM bat LIMIT :param_1) "
|
|
"UNION (SELECT foo, bar FROM bat LIMIT :param_2)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
union(s.limit(1).self_group(), s.limit(2).self_group()),
|
|
"(SELECT foo, bar FROM bat LIMIT :param_1) "
|
|
"UNION (SELECT foo, bar FROM bat LIMIT :param_2)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
union(s.limit(1), s.limit(2).offset(3)).alias().select(),
|
|
"SELECT anon_1.foo, anon_1.bar FROM "
|
|
"((SELECT foo, bar FROM bat LIMIT :param_1) "
|
|
"UNION (SELECT foo, bar FROM bat LIMIT :param_2 OFFSET :param_3)) "
|
|
"AS anon_1",
|
|
)
|
|
|
|
# this version works for SQLite
|
|
self.assert_compile(
|
|
union(s.limit(1).alias().select(), s.offset(2).alias().select()),
|
|
"SELECT anon_1.foo, anon_1.bar "
|
|
"FROM (SELECT foo, bar FROM bat"
|
|
" LIMIT :param_1) AS anon_1 "
|
|
"UNION SELECT anon_2.foo, anon_2.bar "
|
|
"FROM (SELECT foo, bar "
|
|
"FROM bat"
|
|
" LIMIT -1 OFFSET :param_2) AS anon_2",
|
|
)
|
|
|
|
def test_cast(self):
|
|
tbl = table(
|
|
"casttest",
|
|
column("id", Integer),
|
|
column("v1", Float),
|
|
column("v2", Float),
|
|
column("ts", TIMESTAMP),
|
|
)
|
|
|
|
def check_results(dialect, expected_results, literal):
|
|
eq_(
|
|
len(expected_results),
|
|
5,
|
|
"Incorrect number of expected results",
|
|
)
|
|
eq_(
|
|
str(cast(tbl.c.v1, Numeric).compile(dialect=dialect)),
|
|
"CAST(casttest.v1 AS %s)" % expected_results[0],
|
|
)
|
|
eq_(
|
|
str(tbl.c.v1.cast(Numeric).compile(dialect=dialect)),
|
|
"CAST(casttest.v1 AS %s)" % expected_results[0],
|
|
)
|
|
eq_(
|
|
str(cast(tbl.c.v1, Numeric(12, 9)).compile(dialect=dialect)),
|
|
"CAST(casttest.v1 AS %s)" % expected_results[1],
|
|
)
|
|
eq_(
|
|
str(cast(tbl.c.ts, Date).compile(dialect=dialect)),
|
|
"CAST(casttest.ts AS %s)" % expected_results[2],
|
|
)
|
|
eq_(
|
|
str(cast(1234, Text).compile(dialect=dialect)),
|
|
"CAST(%s AS %s)" % (literal, expected_results[3]),
|
|
)
|
|
eq_(
|
|
str(cast("test", String(20)).compile(dialect=dialect)),
|
|
"CAST(%s AS %s)" % (literal, expected_results[4]),
|
|
)
|
|
|
|
# fixme: shoving all of this dialect-specific stuff in one test
|
|
# is now officially completely ridiculous AND non-obviously omits
|
|
# coverage on other dialects.
|
|
sel = select(tbl, cast(tbl.c.v1, Numeric)).compile(dialect=dialect)
|
|
|
|
# TODO: another unusual result from disambiguate only:
|
|
# v1__1 vs v1_1 are due to the special meaning
|
|
# WrapsColumnExpression gives to the "_anon_name_label" attribute,
|
|
# where it tries to default to a label name that matches that of
|
|
# the column within.
|
|
|
|
if isinstance(dialect, type(mysql.dialect())):
|
|
eq_(
|
|
str(sel),
|
|
"SELECT casttest.id, casttest.v1, casttest.v2, "
|
|
"casttest.ts, "
|
|
"CAST(casttest.v1 AS DECIMAL) AS v1__1 \n"
|
|
"FROM casttest",
|
|
)
|
|
else:
|
|
eq_(
|
|
str(sel),
|
|
"SELECT casttest.id, casttest.v1, casttest.v2, "
|
|
"casttest.ts, CAST(casttest.v1 AS NUMERIC) AS "
|
|
"v1__1 \nFROM casttest",
|
|
)
|
|
|
|
sel = (
|
|
select(tbl, cast(tbl.c.v1, Numeric))
|
|
.set_label_style(LABEL_STYLE_NONE)
|
|
.compile(dialect=dialect)
|
|
)
|
|
# label style none - dupes v1
|
|
if isinstance(dialect, type(mysql.dialect())):
|
|
eq_(
|
|
str(sel),
|
|
"SELECT casttest.id, casttest.v1, casttest.v2, "
|
|
"casttest.ts, "
|
|
"CAST(casttest.v1 AS DECIMAL) AS v1 \n"
|
|
"FROM casttest",
|
|
)
|
|
else:
|
|
eq_(
|
|
str(sel),
|
|
"SELECT casttest.id, casttest.v1, casttest.v2, "
|
|
"casttest.ts, CAST(casttest.v1 AS NUMERIC) AS "
|
|
"v1 \nFROM casttest",
|
|
)
|
|
|
|
# first test with PostgreSQL engine
|
|
check_results(
|
|
postgresql.dialect(),
|
|
["NUMERIC", "NUMERIC(12, 9)", "DATE", "TEXT", "VARCHAR(20)"],
|
|
"%(param_1)s",
|
|
)
|
|
|
|
# then the Oracle engine
|
|
check_results(
|
|
oracle.dialect(),
|
|
["NUMERIC", "NUMERIC(12, 9)", "DATE", "CLOB", "VARCHAR2(20 CHAR)"],
|
|
":param_1",
|
|
)
|
|
|
|
# then the sqlite engine
|
|
check_results(
|
|
sqlite.dialect(),
|
|
["NUMERIC", "NUMERIC(12, 9)", "DATE", "TEXT", "VARCHAR(20)"],
|
|
"?",
|
|
)
|
|
|
|
# then the MySQL engine
|
|
check_results(
|
|
mysql.dialect(),
|
|
["DECIMAL", "DECIMAL(12, 9)", "DATE", "CHAR", "CHAR(20)"],
|
|
"%s",
|
|
)
|
|
|
|
self.assert_compile(
|
|
cast(text("NULL"), Integer),
|
|
"CAST(NULL AS INTEGER)",
|
|
dialect=sqlite.dialect(),
|
|
)
|
|
self.assert_compile(
|
|
cast(null(), Integer),
|
|
"CAST(NULL AS INTEGER)",
|
|
dialect=sqlite.dialect(),
|
|
)
|
|
self.assert_compile(
|
|
cast(literal_column("NULL"), Integer),
|
|
"CAST(NULL AS INTEGER)",
|
|
dialect=sqlite.dialect(),
|
|
)
|
|
|
|
@testing.combinations(
|
|
(
|
|
"default",
|
|
None,
|
|
"SELECT CAST(t1.txt AS VARCHAR(10)) AS txt FROM t1",
|
|
None,
|
|
),
|
|
(
|
|
"explicit_mssql",
|
|
"Latin1_General_CI_AS",
|
|
"SELECT CAST(t1.txt AS VARCHAR(10)) COLLATE Latin1_General_CI_AS AS txt FROM t1", # noqa
|
|
mssql.dialect(),
|
|
),
|
|
(
|
|
"explicit_mysql",
|
|
"utf8mb4_unicode_ci",
|
|
"SELECT CAST(t1.txt AS CHAR(10)) AS txt FROM t1",
|
|
mysql.dialect(),
|
|
),
|
|
(
|
|
"explicit_postgresql",
|
|
"en_US",
|
|
'SELECT CAST(t1.txt AS VARCHAR(10)) COLLATE "en_US" AS txt FROM t1', # noqa
|
|
postgresql.dialect(),
|
|
),
|
|
(
|
|
"explicit_sqlite",
|
|
"NOCASE",
|
|
'SELECT CAST(t1.txt AS VARCHAR(10)) COLLATE "NOCASE" AS txt FROM t1', # noqa
|
|
sqlite.dialect(),
|
|
),
|
|
id_="iaaa",
|
|
)
|
|
def test_cast_with_collate(self, collation_name, expected_sql, dialect):
|
|
t1 = Table(
|
|
"t1",
|
|
MetaData(),
|
|
Column("txt", String(10, collation=collation_name)),
|
|
)
|
|
stmt = select(func.cast(t1.c.txt, t1.c.txt.type))
|
|
self.assert_compile(stmt, expected_sql, dialect=dialect)
|
|
|
|
def test_over(self):
|
|
self.assert_compile(func.row_number().over(), "row_number() OVER ()")
|
|
self.assert_compile(
|
|
func.row_number().over(
|
|
order_by=[table1.c.name, table1.c.description]
|
|
),
|
|
"row_number() OVER (ORDER BY mytable.name, mytable.description)",
|
|
)
|
|
self.assert_compile(
|
|
func.row_number().over(
|
|
partition_by=[table1.c.name, table1.c.description]
|
|
),
|
|
"row_number() OVER (PARTITION BY mytable.name, "
|
|
"mytable.description)",
|
|
)
|
|
self.assert_compile(
|
|
func.row_number().over(
|
|
partition_by=[table1.c.name], order_by=[table1.c.description]
|
|
),
|
|
"row_number() OVER (PARTITION BY mytable.name "
|
|
"ORDER BY mytable.description)",
|
|
)
|
|
self.assert_compile(
|
|
func.row_number().over(
|
|
partition_by=table1.c.name, order_by=table1.c.description
|
|
),
|
|
"row_number() OVER (PARTITION BY mytable.name "
|
|
"ORDER BY mytable.description)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
func.row_number().over(
|
|
partition_by=table1.c.name,
|
|
order_by=[table1.c.name, table1.c.description],
|
|
),
|
|
"row_number() OVER (PARTITION BY mytable.name "
|
|
"ORDER BY mytable.name, mytable.description)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
func.row_number().over(
|
|
partition_by=[], order_by=[table1.c.name, table1.c.description]
|
|
),
|
|
"row_number() OVER (ORDER BY mytable.name, mytable.description)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
func.row_number().over(
|
|
partition_by=[table1.c.name, table1.c.description], order_by=[]
|
|
),
|
|
"row_number() OVER (PARTITION BY mytable.name, "
|
|
"mytable.description)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
func.row_number().over(partition_by=[], order_by=[]),
|
|
"row_number() OVER ()",
|
|
)
|
|
self.assert_compile(
|
|
select(
|
|
func.row_number()
|
|
.over(order_by=table1.c.description)
|
|
.label("foo")
|
|
),
|
|
"SELECT row_number() OVER (ORDER BY mytable.description) "
|
|
"AS foo FROM mytable",
|
|
)
|
|
|
|
# test from_obj generation.
|
|
# from func:
|
|
self.assert_compile(
|
|
select(func.max(table1.c.name).over(partition_by=["description"])),
|
|
"SELECT max(mytable.name) OVER (PARTITION BY mytable.description) "
|
|
"AS anon_1 FROM mytable",
|
|
)
|
|
# from partition_by
|
|
self.assert_compile(
|
|
select(func.row_number().over(partition_by=[table1.c.name])),
|
|
"SELECT row_number() OVER (PARTITION BY mytable.name) "
|
|
"AS anon_1 FROM mytable",
|
|
)
|
|
# from order_by
|
|
self.assert_compile(
|
|
select(func.row_number().over(order_by=table1.c.name)),
|
|
"SELECT row_number() OVER (ORDER BY mytable.name) "
|
|
"AS anon_1 FROM mytable",
|
|
)
|
|
|
|
# this tests that _from_objects
|
|
# concantenates OK
|
|
self.assert_compile(
|
|
select(column("x") + over(func.foo())),
|
|
"SELECT x + foo() OVER () AS anon_1",
|
|
)
|
|
|
|
# test a reference to a label that in the referecned selectable;
|
|
# this resolves
|
|
expr = (table1.c.myid + 5).label("sum")
|
|
stmt = select(expr).alias()
|
|
self.assert_compile(
|
|
select(stmt.c.sum, func.row_number().over(order_by=stmt.c.sum)),
|
|
"SELECT anon_1.sum, row_number() OVER (ORDER BY anon_1.sum) "
|
|
"AS anon_2 FROM (SELECT mytable.myid + :myid_1 AS sum "
|
|
"FROM mytable) AS anon_1",
|
|
)
|
|
|
|
# test a reference to a label that's at the same level as the OVER
|
|
# in the columns clause; doesn't resolve
|
|
expr = (table1.c.myid + 5).label("sum")
|
|
self.assert_compile(
|
|
select(expr, func.row_number().over(order_by=expr)),
|
|
"SELECT mytable.myid + :myid_1 AS sum, "
|
|
"row_number() OVER "
|
|
"(ORDER BY mytable.myid + :myid_1) AS anon_1 FROM mytable",
|
|
)
|
|
|
|
def test_over_framespec(self):
|
|
expr = table1.c.myid
|
|
self.assert_compile(
|
|
select(func.row_number().over(order_by=expr, rows=(0, None))),
|
|
"SELECT row_number() OVER "
|
|
"(ORDER BY mytable.myid ROWS BETWEEN CURRENT "
|
|
"ROW AND UNBOUNDED FOLLOWING)"
|
|
" AS anon_1 FROM mytable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(func.row_number().over(order_by=expr, rows=(None, None))),
|
|
"SELECT row_number() OVER "
|
|
"(ORDER BY mytable.myid ROWS BETWEEN UNBOUNDED "
|
|
"PRECEDING AND UNBOUNDED FOLLOWING)"
|
|
" AS anon_1 FROM mytable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(func.row_number().over(order_by=expr, range_=(None, 0))),
|
|
"SELECT row_number() OVER "
|
|
"(ORDER BY mytable.myid RANGE BETWEEN "
|
|
"UNBOUNDED PRECEDING AND CURRENT ROW)"
|
|
" AS anon_1 FROM mytable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(func.row_number().over(order_by=expr, range_=(-5, 10))),
|
|
"SELECT row_number() OVER "
|
|
"(ORDER BY mytable.myid RANGE BETWEEN "
|
|
":param_1 PRECEDING AND :param_2 FOLLOWING)"
|
|
" AS anon_1 FROM mytable",
|
|
checkparams={"param_1": 5, "param_2": 10},
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(func.row_number().over(order_by=expr, range_=(1, 10))),
|
|
"SELECT row_number() OVER "
|
|
"(ORDER BY mytable.myid RANGE BETWEEN "
|
|
":param_1 FOLLOWING AND :param_2 FOLLOWING)"
|
|
" AS anon_1 FROM mytable",
|
|
checkparams={"param_1": 1, "param_2": 10},
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(func.row_number().over(order_by=expr, range_=(-10, -1))),
|
|
"SELECT row_number() OVER "
|
|
"(ORDER BY mytable.myid RANGE BETWEEN "
|
|
":param_1 PRECEDING AND :param_2 PRECEDING)"
|
|
" AS anon_1 FROM mytable",
|
|
checkparams={"param_1": 10, "param_2": 1},
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(func.row_number().over(order_by=expr, groups=(None, 0))),
|
|
"SELECT row_number() OVER "
|
|
"(ORDER BY mytable.myid GROUPS BETWEEN "
|
|
"UNBOUNDED PRECEDING AND CURRENT ROW)"
|
|
" AS anon_1 FROM mytable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(func.row_number().over(order_by=expr, groups=(-5, 10))),
|
|
"SELECT row_number() OVER "
|
|
"(ORDER BY mytable.myid GROUPS BETWEEN "
|
|
":param_1 PRECEDING AND :param_2 FOLLOWING)"
|
|
" AS anon_1 FROM mytable",
|
|
checkparams={"param_1": 5, "param_2": 10},
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(func.row_number().over(order_by=expr, groups=(1, 10))),
|
|
"SELECT row_number() OVER "
|
|
"(ORDER BY mytable.myid GROUPS BETWEEN "
|
|
":param_1 FOLLOWING AND :param_2 FOLLOWING)"
|
|
" AS anon_1 FROM mytable",
|
|
checkparams={"param_1": 1, "param_2": 10},
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(func.row_number().over(order_by=expr, groups=(-10, -1))),
|
|
"SELECT row_number() OVER "
|
|
"(ORDER BY mytable.myid GROUPS BETWEEN "
|
|
":param_1 PRECEDING AND :param_2 PRECEDING)"
|
|
" AS anon_1 FROM mytable",
|
|
checkparams={"param_1": 10, "param_2": 1},
|
|
)
|
|
|
|
def test_over_invalid_framespecs(self):
|
|
assert_raises_message(
|
|
exc.ArgumentError,
|
|
"Integer or None expected for range value",
|
|
func.row_number().over,
|
|
range_=("foo", 8),
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.ArgumentError,
|
|
"Integer or None expected for range value",
|
|
func.row_number().over,
|
|
range_=(-5, "foo"),
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.ArgumentError,
|
|
"only one of 'rows', 'range_', or 'groups' may be provided",
|
|
func.row_number().over,
|
|
range_=(-5, 8),
|
|
rows=(-2, 5),
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.ArgumentError,
|
|
"only one of 'rows', 'range_', or 'groups' may be provided",
|
|
func.row_number().over,
|
|
range_=(-5, 8),
|
|
groups=(None, None),
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.ArgumentError,
|
|
"only one of 'rows', 'range_', or 'groups' may be provided",
|
|
func.row_number().over,
|
|
rows=(-2, 5),
|
|
groups=(None, None),
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.ArgumentError,
|
|
"only one of 'rows', 'range_', or 'groups' may be provided",
|
|
func.row_number().over,
|
|
range_=(-5, 8),
|
|
rows=(-2, 5),
|
|
groups=(None, None),
|
|
)
|
|
|
|
def test_over_within_group(self):
|
|
from sqlalchemy import within_group
|
|
|
|
stmt = select(
|
|
table1.c.myid,
|
|
within_group(func.percentile_cont(0.5), table1.c.name.desc()).over(
|
|
range_=(1, 2),
|
|
partition_by=table1.c.name,
|
|
order_by=table1.c.myid,
|
|
),
|
|
)
|
|
eq_ignore_whitespace(
|
|
str(stmt),
|
|
"SELECT mytable.myid, percentile_cont(:percentile_cont_1) "
|
|
"WITHIN GROUP (ORDER BY mytable.name DESC) "
|
|
"OVER (PARTITION BY mytable.name ORDER BY mytable.myid "
|
|
"RANGE BETWEEN :param_1 FOLLOWING AND :param_2 FOLLOWING) "
|
|
"AS anon_1 FROM mytable",
|
|
)
|
|
|
|
stmt = select(
|
|
table1.c.myid,
|
|
within_group(func.percentile_cont(0.5), table1.c.name.desc()).over(
|
|
rows=(1, 2),
|
|
partition_by=table1.c.name,
|
|
order_by=table1.c.myid,
|
|
),
|
|
)
|
|
eq_ignore_whitespace(
|
|
str(stmt),
|
|
"SELECT mytable.myid, percentile_cont(:percentile_cont_1) "
|
|
"WITHIN GROUP (ORDER BY mytable.name DESC) "
|
|
"OVER (PARTITION BY mytable.name ORDER BY mytable.myid "
|
|
"ROWS BETWEEN :param_1 FOLLOWING AND :param_2 FOLLOWING) "
|
|
"AS anon_1 FROM mytable",
|
|
)
|
|
|
|
def test_date_between(self):
|
|
import datetime
|
|
|
|
table = Table("dt", metadata, Column("date", Date))
|
|
self.assert_compile(
|
|
table.select().where(
|
|
table.c.date.between(
|
|
datetime.date(2006, 6, 1), datetime.date(2006, 6, 5)
|
|
)
|
|
),
|
|
"SELECT dt.date FROM dt WHERE dt.date BETWEEN :date_1 AND :date_2",
|
|
checkparams={
|
|
"date_1": datetime.date(2006, 6, 1),
|
|
"date_2": datetime.date(2006, 6, 5),
|
|
},
|
|
)
|
|
|
|
self.assert_compile(
|
|
table.select().where(
|
|
sql.between(
|
|
table.c.date,
|
|
datetime.date(2006, 6, 1),
|
|
datetime.date(2006, 6, 5),
|
|
)
|
|
),
|
|
"SELECT dt.date FROM dt WHERE dt.date BETWEEN :date_1 AND :date_2",
|
|
checkparams={
|
|
"date_1": datetime.date(2006, 6, 1),
|
|
"date_2": datetime.date(2006, 6, 5),
|
|
},
|
|
)
|
|
|
|
def test_delayed_col_naming(self):
|
|
my_str = Column(String)
|
|
|
|
sel1 = select(my_str)
|
|
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"Cannot initialize a sub-selectable with this Column",
|
|
lambda: sel1.subquery().c,
|
|
)
|
|
|
|
# calling label or scalar_subquery doesn't compile
|
|
# anything.
|
|
sel2 = select(func.substr(my_str, 2, 3)).label("my_substr")
|
|
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
"Cannot compile Column object until its 'name' is assigned.",
|
|
sel2.compile,
|
|
dialect=default.DefaultDialect(),
|
|
)
|
|
|
|
sel3 = select(my_str).scalar_subquery()
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
"Cannot compile Column object until its 'name' is assigned.",
|
|
sel3.compile,
|
|
dialect=default.DefaultDialect(),
|
|
)
|
|
|
|
my_str.name = "foo"
|
|
|
|
self.assert_compile(sel1, "SELECT foo")
|
|
self.assert_compile(
|
|
sel2, "(SELECT substr(foo, :substr_2, :substr_3) AS substr_1)"
|
|
)
|
|
|
|
self.assert_compile(sel3, "(SELECT foo)")
|
|
|
|
def test_naming(self):
|
|
# TODO: the part where we check c.keys() are not "compile" tests, they
|
|
# belong probably in test_selectable, or some broken up
|
|
# version of that suite
|
|
|
|
f1 = func.hoho(table1.c.name)
|
|
s1 = select(
|
|
table1.c.myid,
|
|
table1.c.myid.label("foobar"),
|
|
f1,
|
|
func.lala(table1.c.name).label("gg"),
|
|
)
|
|
|
|
eq_(list(s1.subquery().c.keys()), ["myid", "foobar", "hoho", "gg"])
|
|
|
|
meta = MetaData()
|
|
t1 = Table("mytable", meta, Column("col1", Integer))
|
|
|
|
exprs = (
|
|
table1.c.myid == 12,
|
|
func.hoho(table1.c.myid),
|
|
cast(table1.c.name, Numeric),
|
|
literal("x"),
|
|
)
|
|
for col, key, expr, lbl in (
|
|
(table1.c.name, "name", "mytable.name", None),
|
|
(
|
|
exprs[0],
|
|
"_no_label",
|
|
"mytable.myid = :myid_1",
|
|
"anon_1",
|
|
),
|
|
(exprs[1], "hoho", "hoho(mytable.myid)", "hoho_1"),
|
|
(
|
|
exprs[2],
|
|
"name",
|
|
"CAST(mytable.name AS NUMERIC)",
|
|
"name", # due to [ticket:4449]
|
|
),
|
|
(t1.c.col1, "col1", "mytable.col1", None),
|
|
(
|
|
column("some wacky thing"),
|
|
"some wacky thing",
|
|
'"some wacky thing"',
|
|
"",
|
|
),
|
|
(
|
|
exprs[3],
|
|
"_no_label",
|
|
":param_1",
|
|
"anon_1",
|
|
),
|
|
):
|
|
if getattr(col, "table", None) is not None:
|
|
t = col.table
|
|
else:
|
|
t = table1
|
|
|
|
s1 = select(col).select_from(t)
|
|
eq_(col._proxy_key, key if key != "_no_label" else None)
|
|
eq_(list(s1.subquery().c.keys()), [key])
|
|
|
|
if lbl:
|
|
self.assert_compile(
|
|
s1, "SELECT %s AS %s FROM mytable" % (expr, lbl)
|
|
)
|
|
else:
|
|
self.assert_compile(s1, "SELECT %s FROM mytable" % (expr,))
|
|
|
|
s1 = select(s1.subquery())
|
|
if lbl:
|
|
alias_ = "anon_2" if lbl == "anon_1" else "anon_1"
|
|
self.assert_compile(
|
|
s1,
|
|
"SELECT %s.%s FROM (SELECT %s AS %s FROM mytable) AS %s"
|
|
% (alias_, lbl, expr, lbl, alias_),
|
|
)
|
|
elif col.table is not None:
|
|
# sqlite rule labels subquery columns
|
|
self.assert_compile(
|
|
s1,
|
|
"SELECT anon_1.%s FROM (SELECT %s AS %s FROM mytable) "
|
|
"AS anon_1" % (key, expr, key),
|
|
)
|
|
else:
|
|
self.assert_compile(
|
|
s1,
|
|
"SELECT anon_1.%s FROM (SELECT %s FROM mytable) AS anon_1"
|
|
% (expr, expr),
|
|
)
|
|
|
|
def test_hints(self):
|
|
s = select(table1.c.myid).with_hint(table1, "test hint %(name)s")
|
|
|
|
s2 = (
|
|
select(table1.c.myid)
|
|
.with_hint(table1, "index(%(name)s idx)", "oracle")
|
|
.with_hint(table1, "WITH HINT INDEX idx", "mssql")
|
|
)
|
|
|
|
a1 = table1.alias()
|
|
s3 = select(a1.c.myid).with_hint(a1, "index(%(name)s hint)")
|
|
|
|
subs4 = (
|
|
select(table1, table2)
|
|
.select_from(
|
|
table1.join(table2, table1.c.myid == table2.c.otherid)
|
|
)
|
|
.with_hint(table1, "hint1")
|
|
).subquery()
|
|
|
|
s4 = (
|
|
select(table3)
|
|
.select_from(
|
|
table3.join(subs4, subs4.c.othername == table3.c.otherstuff)
|
|
)
|
|
.with_hint(table3, "hint3")
|
|
)
|
|
|
|
t1 = table("QuotedName", column("col1"))
|
|
s6 = (
|
|
select(t1.c.col1)
|
|
.where(t1.c.col1 > 10)
|
|
.with_hint(t1, "%(name)s idx1")
|
|
)
|
|
a2 = t1.alias("SomeName")
|
|
s7 = (
|
|
select(a2.c.col1)
|
|
.where(a2.c.col1 > 10)
|
|
.with_hint(a2, "%(name)s idx1")
|
|
)
|
|
|
|
mysql_d, oracle_d, mssql_d = (
|
|
mysql.dialect(),
|
|
oracle.dialect(),
|
|
mssql.dialect(),
|
|
)
|
|
|
|
for stmt, dialect, expected in [
|
|
(s, mysql_d, "SELECT mytable.myid FROM mytable test hint mytable"),
|
|
(
|
|
s,
|
|
oracle_d,
|
|
"SELECT /*+ test hint mytable */ mytable.myid FROM mytable",
|
|
),
|
|
(
|
|
s,
|
|
mssql_d,
|
|
"SELECT mytable.myid FROM mytable test hint mytable",
|
|
),
|
|
(s2, mysql_d, "SELECT mytable.myid FROM mytable"),
|
|
(
|
|
s2,
|
|
oracle_d,
|
|
"SELECT /*+ index(mytable idx) */ mytable.myid FROM mytable",
|
|
),
|
|
(
|
|
s2,
|
|
mssql_d,
|
|
"SELECT mytable.myid FROM mytable WITH HINT INDEX idx",
|
|
),
|
|
(
|
|
s3,
|
|
mysql_d,
|
|
"SELECT mytable_1.myid FROM mytable AS mytable_1 "
|
|
"index(mytable_1 hint)",
|
|
),
|
|
(
|
|
s3,
|
|
oracle_d,
|
|
"SELECT /*+ index(mytable_1 hint) */ mytable_1.myid FROM "
|
|
"mytable mytable_1",
|
|
),
|
|
(
|
|
s3,
|
|
mssql_d,
|
|
"SELECT mytable_1.myid FROM mytable AS mytable_1 "
|
|
"index(mytable_1 hint)",
|
|
),
|
|
(
|
|
s4,
|
|
mysql_d,
|
|
"SELECT thirdtable.userid, thirdtable.otherstuff "
|
|
"FROM thirdtable "
|
|
"hint3 INNER JOIN (SELECT mytable.myid AS myid, "
|
|
"mytable.name AS name, "
|
|
"mytable.description AS description, "
|
|
"myothertable.otherid AS otherid, "
|
|
"myothertable.othername AS othername FROM mytable hint1 INNER "
|
|
"JOIN myothertable ON "
|
|
"mytable.myid = myothertable.otherid) AS anon_1 "
|
|
"ON anon_1.othername = thirdtable.otherstuff",
|
|
),
|
|
(
|
|
s4,
|
|
mssql_d,
|
|
"SELECT thirdtable.userid, thirdtable.otherstuff "
|
|
"FROM thirdtable "
|
|
"hint3 JOIN (SELECT mytable.myid AS myid, "
|
|
"mytable.name AS name, "
|
|
"mytable.description AS description, "
|
|
"myothertable.otherid AS otherid, "
|
|
"myothertable.othername AS othername FROM mytable hint1 "
|
|
"JOIN myothertable ON "
|
|
"mytable.myid = myothertable.otherid) AS anon_1 "
|
|
"ON anon_1.othername = thirdtable.otherstuff",
|
|
),
|
|
(
|
|
s4,
|
|
oracle_d,
|
|
"SELECT /*+ hint3 */ thirdtable.userid, thirdtable.otherstuff "
|
|
"FROM thirdtable JOIN (SELECT /*+ hint1 */ "
|
|
"mytable.myid AS myid,"
|
|
" mytable.name AS name, mytable.description AS description, "
|
|
"myothertable.otherid AS otherid,"
|
|
" myothertable.othername AS othername "
|
|
"FROM mytable JOIN myothertable ON "
|
|
"mytable.myid = myothertable.otherid) anon_1 ON "
|
|
"anon_1.othername = thirdtable.otherstuff",
|
|
),
|
|
(
|
|
s6,
|
|
oracle_d,
|
|
"""SELECT /*+ "QuotedName" idx1 */ "QuotedName".col1 """
|
|
"""FROM "QuotedName" WHERE "QuotedName".col1 > :col1_1""",
|
|
),
|
|
(
|
|
s7,
|
|
oracle_d,
|
|
"""SELECT /*+ "SomeName" idx1 */ "SomeName".col1 FROM """
|
|
""""QuotedName" "SomeName" WHERE "SomeName".col1 > :col1_1""",
|
|
),
|
|
]:
|
|
self.assert_compile(stmt, expected, dialect=dialect)
|
|
|
|
def test_statement_hints(self):
|
|
stmt = (
|
|
select(table1.c.myid)
|
|
.with_statement_hint("test hint one")
|
|
.with_statement_hint("test hint two", "mysql")
|
|
)
|
|
|
|
self.assert_compile(
|
|
stmt, "SELECT mytable.myid FROM mytable test hint one"
|
|
)
|
|
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid FROM mytable test hint one test hint two",
|
|
dialect="mysql",
|
|
)
|
|
|
|
def test_literal_as_text_fromstring(self):
|
|
self.assert_compile(and_(text("a"), text("b")), "a AND b")
|
|
|
|
def test_literal_as_text_nonstring_raise(self):
|
|
assert_raises(exc.ArgumentError, and_, ("a",), ("b",))
|
|
|
|
|
|
class BindParameterTest(AssertsCompiledSQL, fixtures.TestBase):
|
|
__dialect__ = "default"
|
|
|
|
def test_binds(self):
|
|
for (
|
|
stmt,
|
|
expected_named_stmt,
|
|
expected_positional_stmt,
|
|
expected_default_params_dict,
|
|
expected_default_params_list,
|
|
test_param_dict,
|
|
expected_test_params_dict,
|
|
expected_test_params_list,
|
|
) in [
|
|
(
|
|
select(table1, table2).where(
|
|
and_(
|
|
table1.c.myid == table2.c.otherid,
|
|
table1.c.name == bindparam("mytablename"),
|
|
),
|
|
),
|
|
"SELECT mytable.myid, mytable.name, mytable.description, "
|
|
"myothertable.otherid, myothertable.othername FROM mytable, "
|
|
"myothertable WHERE mytable.myid = myothertable.otherid "
|
|
"AND mytable.name = :mytablename",
|
|
"SELECT mytable.myid, mytable.name, mytable.description, "
|
|
"myothertable.otherid, myothertable.othername FROM mytable, "
|
|
"myothertable WHERE mytable.myid = myothertable.otherid AND "
|
|
"mytable.name = ?",
|
|
{"mytablename": None},
|
|
[None],
|
|
{"mytablename": 5},
|
|
{"mytablename": 5},
|
|
[5],
|
|
),
|
|
(
|
|
select(table1).where(
|
|
or_(
|
|
table1.c.myid == bindparam("myid"),
|
|
table2.c.otherid == bindparam("myid"),
|
|
),
|
|
),
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable, myothertable WHERE mytable.myid = :myid "
|
|
"OR myothertable.otherid = :myid",
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable, myothertable WHERE mytable.myid = ? "
|
|
"OR myothertable.otherid = ?",
|
|
{"myid": None},
|
|
[None, None],
|
|
{"myid": 5},
|
|
{"myid": 5},
|
|
[5, 5],
|
|
),
|
|
(
|
|
text(
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM "
|
|
"mytable, myothertable WHERE mytable.myid = :myid OR "
|
|
"myothertable.otherid = :myid"
|
|
),
|
|
"SELECT mytable.myid, mytable.name, mytable.description FROM "
|
|
"mytable, myothertable WHERE mytable.myid = :myid OR "
|
|
"myothertable.otherid = :myid",
|
|
"SELECT mytable.myid, mytable.name, mytable.description FROM "
|
|
"mytable, myothertable WHERE mytable.myid = ? OR "
|
|
"myothertable.otherid = ?",
|
|
{"myid": None},
|
|
[None, None],
|
|
{"myid": 5},
|
|
{"myid": 5},
|
|
[5, 5],
|
|
),
|
|
(
|
|
select(table1).where(
|
|
or_(
|
|
table1.c.myid == bindparam("myid", unique=True),
|
|
table2.c.otherid == bindparam("myid", unique=True),
|
|
),
|
|
),
|
|
"SELECT mytable.myid, mytable.name, mytable.description FROM "
|
|
"mytable, myothertable WHERE mytable.myid = "
|
|
":myid_1 OR myothertable.otherid = :myid_2",
|
|
"SELECT mytable.myid, mytable.name, mytable.description FROM "
|
|
"mytable, myothertable WHERE mytable.myid = ? "
|
|
"OR myothertable.otherid = ?",
|
|
{"myid_1": None, "myid_2": None},
|
|
[None, None],
|
|
{"myid_1": 5, "myid_2": 6},
|
|
{"myid_1": 5, "myid_2": 6},
|
|
[5, 6],
|
|
),
|
|
(
|
|
bindparam("test", type_=String, required=False) + text("'hi'"),
|
|
":test || 'hi'",
|
|
"? || 'hi'",
|
|
{"test": None},
|
|
[None],
|
|
{},
|
|
{"test": None},
|
|
[None],
|
|
),
|
|
(
|
|
# testing select.params() here - bindparam() objects
|
|
# must get required flag set to False
|
|
select(table1)
|
|
.where(
|
|
or_(
|
|
table1.c.myid == bindparam("myid"),
|
|
table2.c.otherid == bindparam("myotherid"),
|
|
),
|
|
)
|
|
.params({"myid": 8, "myotherid": 7}),
|
|
"SELECT mytable.myid, mytable.name, mytable.description FROM "
|
|
"mytable, myothertable WHERE mytable.myid = "
|
|
":myid OR myothertable.otherid = :myotherid",
|
|
"SELECT mytable.myid, mytable.name, mytable.description FROM "
|
|
"mytable, myothertable WHERE mytable.myid = "
|
|
"? OR myothertable.otherid = ?",
|
|
{"myid": 8, "myotherid": 7},
|
|
[8, 7],
|
|
{"myid": 5},
|
|
{"myid": 5, "myotherid": 7},
|
|
[5, 7],
|
|
),
|
|
(
|
|
select(table1).where(
|
|
or_(
|
|
table1.c.myid
|
|
== bindparam("myid", value=7, unique=True),
|
|
table2.c.otherid
|
|
== bindparam("myid", value=8, unique=True),
|
|
),
|
|
),
|
|
"SELECT mytable.myid, mytable.name, mytable.description FROM "
|
|
"mytable, myothertable WHERE mytable.myid = "
|
|
":myid_1 OR myothertable.otherid = :myid_2",
|
|
"SELECT mytable.myid, mytable.name, mytable.description FROM "
|
|
"mytable, myothertable WHERE mytable.myid = "
|
|
"? OR myothertable.otherid = ?",
|
|
{"myid_1": 7, "myid_2": 8},
|
|
[7, 8],
|
|
{"myid_1": 5, "myid_2": 6},
|
|
{"myid_1": 5, "myid_2": 6},
|
|
[5, 6],
|
|
),
|
|
]:
|
|
self.assert_compile(
|
|
stmt, expected_named_stmt, params=expected_default_params_dict
|
|
)
|
|
self.assert_compile(
|
|
stmt, expected_positional_stmt, dialect=sqlite.dialect()
|
|
)
|
|
nonpositional = stmt.compile()
|
|
positional = stmt.compile(dialect=sqlite.dialect())
|
|
pp = positional.params
|
|
eq_(
|
|
[pp[k] for k in positional.positiontup],
|
|
expected_default_params_list,
|
|
)
|
|
|
|
eq_(
|
|
nonpositional.construct_params(test_param_dict),
|
|
expected_test_params_dict,
|
|
)
|
|
pp = positional.construct_params(test_param_dict)
|
|
eq_(
|
|
[pp[k] for k in positional.positiontup],
|
|
expected_test_params_list,
|
|
)
|
|
|
|
# check that params() doesn't modify original statement
|
|
s = select(table1).where(
|
|
or_(
|
|
table1.c.myid == bindparam("myid"),
|
|
table2.c.otherid == bindparam("myotherid"),
|
|
),
|
|
)
|
|
s2 = s.params({"myid": 8, "myotherid": 7})
|
|
s3 = s2.params({"myid": 9})
|
|
assert s.compile().params == {"myid": None, "myotherid": None}
|
|
assert s2.compile().params == {"myid": 8, "myotherid": 7}
|
|
assert s3.compile().params == {"myid": 9, "myotherid": 7}
|
|
|
|
# test using same 'unique' param object twice in one compile
|
|
s = select(table1.c.myid).where(table1.c.myid == 12).scalar_subquery()
|
|
s2 = select(table1, s).where(table1.c.myid == s)
|
|
self.assert_compile(
|
|
s2,
|
|
"SELECT mytable.myid, mytable.name, mytable.description, "
|
|
"(SELECT mytable.myid FROM mytable WHERE mytable.myid = "
|
|
":myid_1) AS anon_1 FROM mytable WHERE mytable.myid = "
|
|
"(SELECT mytable.myid FROM mytable WHERE mytable.myid = :myid_1)",
|
|
)
|
|
positional = s2.compile(dialect=sqlite.dialect())
|
|
|
|
pp = positional.params
|
|
assert [pp[k] for k in positional.positiontup] == [12, 12]
|
|
|
|
# check that conflicts with "unique" params are caught
|
|
s = select(table1).where(
|
|
or_(table1.c.myid == 7, table1.c.myid == bindparam("myid_1")),
|
|
)
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
"conflicts with unique bind parameter of the same name",
|
|
str,
|
|
s,
|
|
)
|
|
|
|
s = select(table1).where(
|
|
or_(
|
|
table1.c.myid == 7,
|
|
table1.c.myid == 8,
|
|
table1.c.myid == bindparam("myid_1"),
|
|
),
|
|
)
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
"conflicts with unique bind parameter of the same name",
|
|
str,
|
|
s,
|
|
)
|
|
|
|
def test_bind_param_escaping(self):
|
|
"""general bind param escape unit tests added as a result of
|
|
#8053.
|
|
|
|
The final application of an escaped param name
|
|
was moved out of compiler and into DefaultExecutionContext in
|
|
related issue #8056.
|
|
|
|
However in #8113 we made this conditional to suit usage recipes
|
|
posted in the FAQ.
|
|
|
|
|
|
"""
|
|
|
|
SomeEnum = pep435_enum("SomeEnum")
|
|
one = SomeEnum("one", 1)
|
|
SomeEnum("two", 2)
|
|
|
|
t = Table(
|
|
"t",
|
|
MetaData(),
|
|
Column("_id", Integer, primary_key=True),
|
|
Column("_data", Enum(SomeEnum)),
|
|
)
|
|
|
|
class MyCompiler(compiler.SQLCompiler):
|
|
def bindparam_string(self, name, **kw):
|
|
kw["escaped_from"] = name
|
|
return super().bindparam_string('"%s"' % name, **kw)
|
|
|
|
dialect = default.DefaultDialect()
|
|
dialect.statement_compiler = MyCompiler
|
|
|
|
self.assert_compile(
|
|
t.insert(),
|
|
'INSERT INTO t (_id, _data) VALUES (:"_id", :"_data")',
|
|
dialect=dialect,
|
|
)
|
|
|
|
compiled = t.insert().compile(
|
|
dialect=dialect, compile_kwargs=dict(compile_keys=("_id", "_data"))
|
|
)
|
|
|
|
# not escaped
|
|
params = compiled.construct_params(
|
|
{"_id": 1, "_data": one}, escape_names=False
|
|
)
|
|
eq_(params, {"_id": 1, "_data": one})
|
|
|
|
# escaped by default
|
|
params = compiled.construct_params({"_id": 1, "_data": one})
|
|
eq_(params, {'"_id"': 1, '"_data"': one})
|
|
|
|
# escaped here as well
|
|
eq_(compiled.params, {'"_data"': None, '"_id"': None})
|
|
|
|
# bind processors aren't part of this
|
|
eq_(compiled._bind_processors, {"_data": mock.ANY})
|
|
|
|
dialect.paramstyle = "pyformat"
|
|
compiled = t.insert().compile(
|
|
dialect=dialect, compile_kwargs=dict(compile_keys=("_id", "_data"))
|
|
)
|
|
|
|
# FAQ recipe works
|
|
eq_(
|
|
compiled.string % compiled.params,
|
|
"INSERT INTO t (_id, _data) VALUES (None, None)",
|
|
)
|
|
|
|
def test_expanding_non_expanding_conflict(self):
|
|
"""test #8018"""
|
|
|
|
s = select(
|
|
literal("x").in_(bindparam("q")),
|
|
bindparam("q"),
|
|
)
|
|
|
|
with expect_raises_message(
|
|
exc.CompileError,
|
|
r"Can't reuse bound parameter name 'q' in both 'expanding' "
|
|
r"\(e.g. within an IN expression\) and non-expanding contexts. "
|
|
"If this parameter is to "
|
|
"receive a list/array value, set 'expanding=True' on "
|
|
"it for expressions that aren't IN, otherwise use "
|
|
"a different parameter name.",
|
|
):
|
|
str(s)
|
|
|
|
def test_unique_binds_no_clone_collision(self):
|
|
"""test #6824"""
|
|
bp = bindparam("foo", unique=True)
|
|
|
|
bpc1 = bp._clone(maintain_key=True)
|
|
bpc2 = bp._clone(maintain_key=True)
|
|
|
|
stmt1 = select(bp, bpc1, bpc2)
|
|
|
|
# OK, still strange that the double-dedupe logic is still *duping*
|
|
# the label name, but that's a different issue
|
|
self.assert_compile(
|
|
stmt1,
|
|
"SELECT :foo_1 AS anon_1, :foo_1 AS anon__1, :foo_1 AS anon__2",
|
|
)
|
|
|
|
def _test_binds_no_hash_collision(self):
|
|
"""test that construct_params doesn't corrupt dict
|
|
due to hash collisions"""
|
|
|
|
total_params = 100000
|
|
|
|
in_clause = [":in%d" % i for i in range(total_params)]
|
|
params = {"in%d" % i: i for i in range(total_params)}
|
|
t = text("text clause %s" % ", ".join(in_clause))
|
|
eq_(len(t.bindparams), total_params)
|
|
c = t.compile()
|
|
pp = c.construct_params(params)
|
|
eq_(len(set(pp)), total_params, "%s %s" % (len(set(pp)), len(pp)))
|
|
eq_(len(set(pp.values())), total_params)
|
|
|
|
def test_bind_anon_name_no_special_chars(self):
|
|
for paramstyle in "named", "pyformat":
|
|
dialect = default.DefaultDialect()
|
|
dialect.paramstyle = paramstyle
|
|
|
|
for name, named, pyformat in [
|
|
("%(my name)s", ":my_name_s_1", "%(my_name_s_1)s"),
|
|
("myname(foo)", ":myname_foo_1", "%(myname_foo_1)s"),
|
|
(
|
|
"this is a name",
|
|
":this_is_a_name_1",
|
|
"%(this_is_a_name_1)s",
|
|
),
|
|
("_leading_one", ":leading_one_1", "%(leading_one_1)s"),
|
|
("3leading_two", ":3leading_two_1", "%(3leading_two_1)s"),
|
|
("$leading_three", ":leading_three_1", "%(leading_three_1)s"),
|
|
("%(tricky", ":tricky_1", "%(tricky_1)s"),
|
|
("5(tricky", ":5_tricky_1", "%(5_tricky_1)s"),
|
|
]:
|
|
t = table("t", column(name, String))
|
|
expr = t.c[name] == "foo"
|
|
|
|
self.assert_compile(
|
|
expr,
|
|
"t.%s = %s"
|
|
% (
|
|
dialect.identifier_preparer.quote(name),
|
|
named if paramstyle == "named" else pyformat,
|
|
),
|
|
dialect=dialect,
|
|
checkparams={named[1:]: "foo"},
|
|
)
|
|
|
|
def test_bind_anon_name_special_chars_uniqueify_one(self):
|
|
# test that the chars are escaped before doing the counter,
|
|
# otherwise these become the same name and bind params will conflict
|
|
t = table("t", column("_3foo"), column("4%foo"))
|
|
|
|
self.assert_compile(
|
|
(t.c["_3foo"] == "foo") & (t.c["4%foo"] == "bar"),
|
|
't._3foo = :3foo_1 AND t."4%foo" = :4_foo_1',
|
|
checkparams={"3foo_1": "foo", "4_foo_1": "bar"},
|
|
)
|
|
|
|
def test_bind_anon_name_special_chars_uniqueify_two(self):
|
|
t = table("t", column("_3foo"), column("4(foo"))
|
|
|
|
self.assert_compile(
|
|
(t.c["_3foo"] == "foo") & (t.c["4(foo"] == "bar"),
|
|
't._3foo = :3foo_1 AND t."4(foo" = :4_foo_1',
|
|
checkparams={"3foo_1": "foo", "4_foo_1": "bar"},
|
|
)
|
|
|
|
def test_bind_given_anon_name_dont_double(self):
|
|
c = column("id")
|
|
l = c.label(None)
|
|
|
|
# new case as of Id810f485c5f7ed971529489b84694e02a3356d6d
|
|
subq = select(l).subquery()
|
|
|
|
# this creates a ColumnClause as a proxy to the Label() that has
|
|
# an anonymous name, so the column has one too.
|
|
anon_col = subq.c[0]
|
|
assert isinstance(anon_col.name, elements._anonymous_label)
|
|
|
|
# then when BindParameter is created, it checks the label
|
|
# and doesn't double up on the anonymous name which is uncachable
|
|
expr = anon_col > 5
|
|
|
|
self.assert_compile(
|
|
expr, "anon_1.id_1 > :param_1", checkparams={"param_1": 5}
|
|
)
|
|
|
|
# see also test_compare.py -> _statements_w_anonymous_col_names
|
|
# fixture for cache key
|
|
|
|
def test_bind_as_col(self):
|
|
t = table("foo", column("id"))
|
|
|
|
s = select(t, literal("lala").label("hoho"))
|
|
self.assert_compile(s, "SELECT foo.id, :param_1 AS hoho FROM foo")
|
|
|
|
assert [str(c) for c in s.subquery().c] == ["anon_1.id", "anon_1.hoho"]
|
|
|
|
def test_bind_callable(self):
|
|
expr = column("x") == bindparam("key", callable_=lambda: 12)
|
|
self.assert_compile(expr, "x = :key", {"x": 12})
|
|
|
|
def test_bind_params_missing(self):
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
r"A value is required for bind parameter 'x'",
|
|
select(table1)
|
|
.where(
|
|
and_(
|
|
table1.c.myid == bindparam("x", required=True),
|
|
table1.c.name == bindparam("y", required=True),
|
|
)
|
|
)
|
|
.compile()
|
|
.construct_params,
|
|
params=dict(y=5),
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
r"A value is required for bind parameter 'x'",
|
|
select(table1)
|
|
.where(table1.c.myid == bindparam("x", required=True))
|
|
.compile()
|
|
.construct_params,
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
r"A value is required for bind parameter 'x', "
|
|
"in parameter group 2",
|
|
select(table1)
|
|
.where(
|
|
and_(
|
|
table1.c.myid == bindparam("x", required=True),
|
|
table1.c.name == bindparam("y", required=True),
|
|
)
|
|
)
|
|
.compile()
|
|
.construct_params,
|
|
params=dict(y=5),
|
|
_group_number=2,
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
r"A value is required for bind parameter 'x', "
|
|
"in parameter group 2",
|
|
select(table1)
|
|
.where(table1.c.myid == bindparam("x", required=True))
|
|
.compile()
|
|
.construct_params,
|
|
_group_number=2,
|
|
)
|
|
|
|
@testing.combinations(
|
|
(
|
|
select(table1).where(table1.c.myid == 5),
|
|
select(table1).where(table1.c.myid == 10),
|
|
{"myid_1": 5},
|
|
{"myid_1": 10},
|
|
None,
|
|
None,
|
|
),
|
|
(
|
|
select(table1).where(
|
|
table1.c.myid
|
|
== bindparam(None, unique=True, callable_=lambda: 5)
|
|
),
|
|
select(table1).where(
|
|
table1.c.myid
|
|
== bindparam(None, unique=True, callable_=lambda: 10)
|
|
),
|
|
{"param_1": 5},
|
|
{"param_1": 10},
|
|
None,
|
|
None,
|
|
),
|
|
(
|
|
table1.update()
|
|
.where(table1.c.myid == 5)
|
|
.values(name="n1", description="d1"),
|
|
table1.update()
|
|
.where(table1.c.myid == 10)
|
|
.values(name="n2", description="d2"),
|
|
{"description": "d1", "myid_1": 5, "name": "n1"},
|
|
{"description": "d2", "myid_1": 10, "name": "n2"},
|
|
None,
|
|
None,
|
|
),
|
|
(
|
|
table1.update().where(table1.c.myid == 5),
|
|
table1.update().where(table1.c.myid == 10),
|
|
{"description": "d1", "myid_1": 5, "name": "n1"},
|
|
{"description": "d2", "myid_1": 10, "name": "n2"},
|
|
{"description": "d1", "name": "n1"},
|
|
{"description": "d2", "name": "n2"},
|
|
),
|
|
(
|
|
table1.update().where(
|
|
table1.c.myid
|
|
== bindparam(None, unique=True, callable_=lambda: 5)
|
|
),
|
|
table1.update().where(
|
|
table1.c.myid
|
|
== bindparam(None, unique=True, callable_=lambda: 10)
|
|
),
|
|
{"description": "d1", "param_1": 5, "name": "n1"},
|
|
{"description": "d2", "param_1": 10, "name": "n2"},
|
|
{"description": "d1", "name": "n1"},
|
|
{"description": "d2", "name": "n2"},
|
|
),
|
|
(
|
|
union(
|
|
select(table1).where(table1.c.myid == 5),
|
|
select(table1).where(table1.c.myid == 12),
|
|
),
|
|
union(
|
|
select(table1).where(table1.c.myid == 5),
|
|
select(table1).where(table1.c.myid == 15),
|
|
),
|
|
{"myid_1": 5, "myid_2": 12},
|
|
{"myid_1": 5, "myid_2": 15},
|
|
None,
|
|
None,
|
|
),
|
|
)
|
|
def test_construct_params_combine_extracted(
|
|
self, stmt1, stmt2, param1, param2, extparam1, extparam2
|
|
):
|
|
if extparam1:
|
|
keys = list(extparam1)
|
|
else:
|
|
keys = []
|
|
|
|
s1_cache_key = stmt1._generate_cache_key()
|
|
s1_compiled = stmt1.compile(cache_key=s1_cache_key, column_keys=keys)
|
|
|
|
s2_cache_key = stmt2._generate_cache_key()
|
|
|
|
eq_(s1_compiled.construct_params(params=extparam1), param1)
|
|
eq_(
|
|
s1_compiled.construct_params(
|
|
params=extparam1, extracted_parameters=s1_cache_key[1]
|
|
),
|
|
param1,
|
|
)
|
|
|
|
eq_(
|
|
s1_compiled.construct_params(
|
|
params=extparam2, extracted_parameters=s2_cache_key[1]
|
|
),
|
|
param2,
|
|
)
|
|
|
|
s1_compiled_no_cache_key = stmt1.compile()
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
"This compiled object has no original cache key; can't pass "
|
|
"extracted_parameters to construct_params",
|
|
s1_compiled_no_cache_key.construct_params,
|
|
extracted_parameters=s1_cache_key[1],
|
|
)
|
|
|
|
@testing.combinations(True, False, argnames="adapt_before_key")
|
|
def test_construct_params_w_bind_clones_post(self, adapt_before_key):
|
|
"""test that a BindParameter that has been cloned after the cache
|
|
key was generated still matches up when construct_params()
|
|
is called with an extracted parameter collection.
|
|
|
|
This case occurs now with the ORM as the ORM construction will
|
|
frequently run clause adaptation on elements of the statement within
|
|
compilation, after the cache key has been generated. this adaptation
|
|
hits BindParameter objects which will change their key as they
|
|
will usually have unqique=True. So the construct_params() process
|
|
when it links its internal bind_names to the cache key binds,
|
|
must do this badsed on bindparam._identifying_key, which does not
|
|
change across clones, rather than .key which usually will.
|
|
|
|
"""
|
|
|
|
stmt = select(table1.c.myid).where(table1.c.myid == 5)
|
|
|
|
# get the original bindparam.
|
|
original_bind = stmt._where_criteria[0].right
|
|
|
|
# it's anonymous so unique=True
|
|
is_true(original_bind.unique)
|
|
|
|
# test #7903 - adapt the statement *before* we make the cache
|
|
# key also
|
|
if adapt_before_key:
|
|
stmt = sql_util.ClauseAdapter(table1).traverse(stmt)
|
|
|
|
# cache key against the original param
|
|
cache_key = stmt._generate_cache_key()
|
|
|
|
# now adapt the statement
|
|
stmt_adapted = sql_util.ClauseAdapter(table1).traverse(stmt)
|
|
|
|
# new bind parameter has a different key but same
|
|
# identifying key
|
|
new_bind = stmt_adapted._where_criteria[0].right
|
|
eq_(original_bind._identifying_key, new_bind._identifying_key)
|
|
ne_(original_bind.key, new_bind.key)
|
|
|
|
# compile the adapted statement but set the cache key to the one
|
|
# generated from the unadapted statement. this will look like
|
|
# when the ORM runs clause adaption inside of visit_select, after
|
|
# the cache key is generated but before the compiler is given the
|
|
# core select statement to actually render.
|
|
compiled = stmt_adapted.compile(cache_key=cache_key)
|
|
|
|
# params set up as 5
|
|
eq_(
|
|
compiled.construct_params(
|
|
params={},
|
|
),
|
|
{"myid_1": 5},
|
|
)
|
|
|
|
# also works w the original cache key
|
|
eq_(
|
|
compiled.construct_params(
|
|
params={}, extracted_parameters=cache_key[1]
|
|
),
|
|
{"myid_1": 5},
|
|
)
|
|
|
|
# now make a totally new statement with the same cache key
|
|
new_stmt = select(table1.c.myid).where(table1.c.myid == 10)
|
|
new_cache_key = new_stmt._generate_cache_key()
|
|
|
|
# cache keys match
|
|
eq_(cache_key.key, new_cache_key.key)
|
|
|
|
# ensure we get "10" from construct params. if it matched
|
|
# based on .key and not ._identifying_key, it would not see that
|
|
# the bind parameter is part of the cache key.
|
|
eq_(
|
|
compiled.construct_params(
|
|
params={}, extracted_parameters=new_cache_key[1]
|
|
),
|
|
{"myid_1": 10},
|
|
)
|
|
|
|
@testing.combinations(True, False, argnames="adapt_before_key")
|
|
def test_construct_duped_params_w_bind_clones_post(self, adapt_before_key):
|
|
"""same as previous test_construct_params_w_bind_clones_post but
|
|
where the binds have been used
|
|
repeatedly, and the adaption occurs on a per-subquery basis.
|
|
test for #6391
|
|
|
|
"""
|
|
|
|
inner_stmt = select(table1.c.myid).where(table1.c.myid == 5)
|
|
|
|
stmt = union(inner_stmt, inner_stmt, inner_stmt)
|
|
|
|
# get the original bindparam.
|
|
original_bind = inner_stmt._where_criteria[0].right
|
|
|
|
# same bind three times
|
|
is_(stmt.selects[0]._where_criteria[0].right, original_bind)
|
|
is_(stmt.selects[1]._where_criteria[0].right, original_bind)
|
|
is_(stmt.selects[2]._where_criteria[0].right, original_bind)
|
|
|
|
# it's anonymous so unique=True
|
|
is_true(original_bind.unique)
|
|
|
|
# variant that exercises #7903
|
|
if adapt_before_key:
|
|
stmt = sql_util.ClauseAdapter(table1).traverse(stmt)
|
|
|
|
# cache key against the original param
|
|
cache_key = stmt._generate_cache_key()
|
|
|
|
# now adapt the statement and separately adapt the inner
|
|
# SELECTs, since if these subqueries are also ORM then they get adapted
|
|
# separately.
|
|
stmt_adapted = sql_util.ClauseAdapter(table1).traverse(stmt)
|
|
stmt_adapted.selects[0] = sql_util.ClauseAdapter(table1).traverse(
|
|
stmt_adapted.selects[0]
|
|
)
|
|
stmt_adapted.selects[1] = sql_util.ClauseAdapter(table1).traverse(
|
|
stmt_adapted.selects[1]
|
|
)
|
|
stmt_adapted.selects[2] = sql_util.ClauseAdapter(table1).traverse(
|
|
stmt_adapted.selects[2]
|
|
)
|
|
|
|
# new bind parameter has a different key but same
|
|
# identifying key
|
|
|
|
new_bind_one = stmt_adapted.selects[0]._where_criteria[0].right
|
|
new_bind_two = stmt_adapted.selects[1]._where_criteria[0].right
|
|
new_bind_three = stmt_adapted.selects[2]._where_criteria[0].right
|
|
|
|
for new_bind in (new_bind_one, new_bind_two, new_bind_three):
|
|
eq_(original_bind._identifying_key, new_bind._identifying_key)
|
|
ne_(original_bind.key, new_bind.key)
|
|
|
|
# compile the adapted statement but set the cache key to the one
|
|
# generated from the unadapted statement. this will look like
|
|
# when the ORM runs clause adaption inside of visit_select, after
|
|
# the cache key is generated but before the compiler is given the
|
|
# core select statement to actually render.
|
|
compiled = stmt_adapted.compile(cache_key=cache_key)
|
|
|
|
# the same parameter was split into three distinct ones, due to
|
|
# the separate adaption on a per-subquery basis. but they still
|
|
# refer to the original in their _cloned_set and this is what
|
|
# has to match up to what's in the cache key.
|
|
# params set up as 5
|
|
eq_(
|
|
compiled.construct_params(
|
|
params={},
|
|
),
|
|
{"myid_1": 5, "myid_2": 5, "myid_3": 5},
|
|
)
|
|
|
|
# also works w the original cache key
|
|
eq_(
|
|
compiled.construct_params(
|
|
params={}, extracted_parameters=cache_key[1]
|
|
),
|
|
{"myid_1": 5, "myid_2": 5, "myid_3": 5},
|
|
)
|
|
|
|
# now make a totally new statement with the same cache key
|
|
new_inner_stmt = select(table1.c.myid).where(table1.c.myid == 10)
|
|
new_stmt = union(new_inner_stmt, new_inner_stmt, new_inner_stmt)
|
|
|
|
new_cache_key = new_stmt._generate_cache_key()
|
|
|
|
# cache keys match
|
|
eq_(cache_key.key, new_cache_key.key)
|
|
|
|
# ensure we get "10" from construct params. if it matched
|
|
# based on .key and not ._identifying_key, it would not see that
|
|
# the bind parameter is part of the cache key.
|
|
# before #6391 was fixed you would see 5, 5, 10
|
|
eq_(
|
|
compiled.construct_params(
|
|
params={}, extracted_parameters=new_cache_key[1]
|
|
),
|
|
{"myid_1": 10, "myid_2": 10, "myid_3": 10},
|
|
)
|
|
|
|
def test_construct_params_w_bind_clones_pre(self):
|
|
"""test that a BindParameter that has been cloned before the cache
|
|
key was generated, and was doubled up just to make sure it has to
|
|
be unique, still matches up when construct_params()
|
|
is called with an extracted parameter collection.
|
|
|
|
other ORM features like optimized_compare() end up doing something
|
|
like this, such as if there are multiple "has()" or "any()" which would
|
|
have cloned the join condition and changed the values of bound
|
|
parameters.
|
|
|
|
"""
|
|
|
|
stmt = select(table1.c.myid).where(table1.c.myid == 5)
|
|
|
|
original_bind = stmt._where_criteria[0].right
|
|
# it's anonymous so unique=True
|
|
is_true(original_bind.unique)
|
|
|
|
b1 = original_bind._clone()
|
|
b1.value = 10
|
|
b2 = original_bind._clone()
|
|
b2.value = 12
|
|
|
|
# make a new statement that uses the clones as distinct
|
|
# parameters
|
|
modified_stmt = select(table1.c.myid).where(
|
|
or_(table1.c.myid == b1, table1.c.myid == b2)
|
|
)
|
|
|
|
cache_key = modified_stmt._generate_cache_key()
|
|
compiled = modified_stmt.compile(cache_key=cache_key)
|
|
|
|
eq_(
|
|
compiled.construct_params(params={}),
|
|
{"myid_1": 10, "myid_2": 12},
|
|
)
|
|
|
|
# make a new statement doing the same thing and make sure
|
|
# the binds match up correctly
|
|
new_stmt = select(table1.c.myid).where(table1.c.myid == 8)
|
|
|
|
new_original_bind = new_stmt._where_criteria[0].right
|
|
new_b1 = new_original_bind._clone()
|
|
new_b1.value = 20
|
|
new_b2 = new_original_bind._clone()
|
|
new_b2.value = 18
|
|
modified_new_stmt = select(table1.c.myid).where(
|
|
or_(table1.c.myid == new_b1, table1.c.myid == new_b2)
|
|
)
|
|
|
|
new_cache_key = modified_new_stmt._generate_cache_key()
|
|
|
|
# cache keys match
|
|
eq_(cache_key.key, new_cache_key.key)
|
|
|
|
# ensure we get both values
|
|
eq_(
|
|
compiled.construct_params(
|
|
params={}, extracted_parameters=new_cache_key[1]
|
|
),
|
|
{"myid_1": 20, "myid_2": 18},
|
|
)
|
|
|
|
@testing.combinations("default", "default_qmark", argnames="dialect")
|
|
def test_literal_execute_combinations(self, dialect):
|
|
"""test #10142"""
|
|
|
|
a = bindparam("a", value="abc", literal_execute=True)
|
|
b = bindparam("b", value="def", literal_execute=True)
|
|
c = bindparam("c", value="ghi", literal_execute=True)
|
|
self.assert_compile(
|
|
select(a, b, a, c),
|
|
"SELECT 'abc' AS anon_1, 'def' AS anon_2, 'abc' AS anon__1, "
|
|
"'ghi' AS anon_3",
|
|
render_postcompile=True,
|
|
dialect=dialect,
|
|
)
|
|
|
|
def test_tuple_expanding_in_no_values(self):
|
|
expr = tuple_(table1.c.myid, table1.c.name).in_(
|
|
[(1, "foo"), (5, "bar")]
|
|
)
|
|
self.assert_compile(
|
|
expr,
|
|
"(mytable.myid, mytable.name) IN (__[POSTCOMPILE_param_1])",
|
|
checkparams={"param_1": [(1, "foo"), (5, "bar")]},
|
|
check_post_param={"param_1": [(1, "foo"), (5, "bar")]},
|
|
check_literal_execute={},
|
|
)
|
|
|
|
compiled = expr.compile()
|
|
(
|
|
to_update,
|
|
replacement_expr,
|
|
) = compiled._literal_execute_expanding_parameter(
|
|
"param_1", expr.right, [(1, "foo"), (5, "bar")]
|
|
)
|
|
eq_(
|
|
to_update,
|
|
[
|
|
("param_1_1_1", 1),
|
|
("param_1_1_2", "foo"),
|
|
("param_1_2_1", 5),
|
|
("param_1_2_2", "bar"),
|
|
],
|
|
)
|
|
eq_(
|
|
replacement_expr,
|
|
"(:param_1_1_1, :param_1_1_2), (:param_1_2_1, :param_1_2_2)",
|
|
)
|
|
|
|
def test_tuple_expanding_in_values(self):
|
|
expr = tuple_(table1.c.myid, table1.c.name).in_(
|
|
[(1, "foo"), (5, "bar")]
|
|
)
|
|
dialect = default.DefaultDialect()
|
|
dialect.tuple_in_values = True
|
|
self.assert_compile(
|
|
tuple_(table1.c.myid, table1.c.name).in_([(1, "foo"), (5, "bar")]),
|
|
"(mytable.myid, mytable.name) IN (__[POSTCOMPILE_param_1])",
|
|
dialect=dialect,
|
|
checkparams={"param_1": [(1, "foo"), (5, "bar")]},
|
|
check_post_param={"param_1": [(1, "foo"), (5, "bar")]},
|
|
check_literal_execute={},
|
|
)
|
|
|
|
compiled = expr.compile(dialect=dialect)
|
|
(
|
|
to_update,
|
|
replacement_expr,
|
|
) = compiled._literal_execute_expanding_parameter(
|
|
"param_1", expr.right, [(1, "foo"), (5, "bar")]
|
|
)
|
|
eq_(
|
|
to_update,
|
|
[
|
|
("param_1_1_1", 1),
|
|
("param_1_1_2", "foo"),
|
|
("param_1_2_1", 5),
|
|
("param_1_2_2", "bar"),
|
|
],
|
|
)
|
|
eq_(
|
|
replacement_expr,
|
|
"VALUES (:param_1_1_1, :param_1_1_2), "
|
|
"(:param_1_2_1, :param_1_2_2)",
|
|
)
|
|
|
|
def test_construct_params_repeated_postcompile_params_one(self):
|
|
"""test for :ticket:`6202` one - name repeated in positiontup
|
|
(e.g. SQL Server using TOP)
|
|
|
|
"""
|
|
|
|
t = table("t", column("x"))
|
|
stmt = (
|
|
select(1)
|
|
.where(t.c.x == bindparam(None, value="10", literal_execute=True))
|
|
.scalar_subquery()
|
|
)
|
|
|
|
u = union(select(stmt), select(stmt)).subquery().select()
|
|
|
|
compiled = u.compile(
|
|
dialect=default.DefaultDialect(paramstyle="format"),
|
|
compile_kwargs={"render_postcompile": True},
|
|
)
|
|
eq_ignore_whitespace(
|
|
compiled.string,
|
|
"SELECT anon_2.anon_1 FROM (SELECT (SELECT 1 FROM t "
|
|
"WHERE t.x = '10') AS anon_1 UNION SELECT "
|
|
"(SELECT 1 FROM t WHERE t.x = '10') AS anon_1) AS anon_2",
|
|
)
|
|
eq_(compiled.construct_params(_no_postcompile=True), {"param_1": "10"})
|
|
|
|
def test_construct_params_repeated_postcompile_params_two(self):
|
|
"""test for :ticket:`6202` two - same param name used twice
|
|
(e.g. Oracle LIMIT)
|
|
|
|
"""
|
|
t = table("t", column("x"))
|
|
|
|
bp = bindparam(None, value="10")
|
|
stmt = (
|
|
select(1)
|
|
.where(t.c.x == bp.render_literal_execute())
|
|
.scalar_subquery()
|
|
)
|
|
stmt2 = (
|
|
select(1)
|
|
.where(t.c.x == bp.render_literal_execute())
|
|
.scalar_subquery()
|
|
)
|
|
|
|
u = union(select(stmt), select(stmt2)).subquery().select()
|
|
|
|
compiled = u.compile(
|
|
dialect=default.DefaultDialect(paramstyle="named"),
|
|
compile_kwargs={"render_postcompile": True},
|
|
)
|
|
eq_ignore_whitespace(
|
|
compiled.string,
|
|
"SELECT anon_2.anon_1 FROM (SELECT (SELECT 1 "
|
|
"FROM t WHERE t.x = '10') AS anon_1 UNION SELECT "
|
|
"(SELECT 1 FROM t WHERE t.x = '10') AS anon_3) AS anon_2",
|
|
)
|
|
eq_(compiled.construct_params(_no_postcompile=True), {"param_1": "10"})
|
|
|
|
def test_construct_params_positional_plain_repeated(self):
|
|
t = table("t", column("x"))
|
|
stmt = (
|
|
select(1)
|
|
.where(t.c.x == bindparam(None, value="10"))
|
|
.where(t.c.x == bindparam(None, value="12", literal_execute=True))
|
|
.scalar_subquery()
|
|
)
|
|
|
|
u = union(select(stmt), select(stmt)).subquery().select()
|
|
|
|
compiled = u.compile(
|
|
dialect=default.DefaultDialect(paramstyle="format"),
|
|
compile_kwargs={"render_postcompile": True},
|
|
)
|
|
eq_ignore_whitespace(
|
|
compiled.string,
|
|
"SELECT anon_2.anon_1 FROM (SELECT (SELECT 1 FROM t "
|
|
"WHERE t.x = %s AND t.x = '12') AS anon_1 "
|
|
"UNION SELECT (SELECT 1 FROM t WHERE t.x = %s AND t.x = '12') "
|
|
"AS anon_1) AS anon_2",
|
|
)
|
|
eq_(
|
|
compiled.construct_params(_no_postcompile=True),
|
|
{"param_1": "10", "param_2": "12"},
|
|
)
|
|
eq_(compiled.positiontup, ["param_1", "param_1"])
|
|
|
|
def test_tuple_clauselist_in(self):
|
|
self.assert_compile(
|
|
tuple_(table1.c.myid, table1.c.name).in_(
|
|
[tuple_(table2.c.otherid, table2.c.othername)]
|
|
),
|
|
"(mytable.myid, mytable.name) IN "
|
|
"((myothertable.otherid, myothertable.othername))",
|
|
)
|
|
|
|
@testing.variation("scalar_subquery", [True, False])
|
|
def test_select_in(self, scalar_subquery):
|
|
stmt = select(table2.c.otherid, table2.c.othername)
|
|
|
|
if scalar_subquery:
|
|
stmt = stmt.scalar_subquery()
|
|
|
|
self.assert_compile(
|
|
tuple_(table1.c.myid, table1.c.name).in_(stmt),
|
|
"(mytable.myid, mytable.name) IN (SELECT "
|
|
"myothertable.otherid, myothertable.othername FROM myothertable)",
|
|
)
|
|
|
|
def test_expanding_parameter(self):
|
|
self.assert_compile(
|
|
tuple_(table1.c.myid, table1.c.name).in_(
|
|
bindparam("foo", expanding=True)
|
|
),
|
|
"(mytable.myid, mytable.name) IN (__[POSTCOMPILE_foo])",
|
|
)
|
|
|
|
dialect = default.DefaultDialect()
|
|
dialect.tuple_in_values = True
|
|
self.assert_compile(
|
|
tuple_(table1.c.myid, table1.c.name).in_(
|
|
bindparam("foo", expanding=True)
|
|
),
|
|
"(mytable.myid, mytable.name) IN (__[POSTCOMPILE_foo])",
|
|
dialect=dialect,
|
|
)
|
|
|
|
self.assert_compile(
|
|
table1.c.myid.in_(bindparam("foo", expanding=True)),
|
|
"mytable.myid IN (__[POSTCOMPILE_foo])",
|
|
)
|
|
|
|
def test_limit_offset_select_literal_binds(self):
|
|
stmt = select(1).limit(5).offset(6)
|
|
self.assert_compile(
|
|
stmt, "SELECT 1 LIMIT 5 OFFSET 6", literal_binds=True
|
|
)
|
|
|
|
def test_limit_offset_compound_select_literal_binds(self):
|
|
stmt = select(1).union(select(2)).limit(5).offset(6)
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT 1 UNION SELECT 2 LIMIT 5 OFFSET 6",
|
|
literal_binds=True,
|
|
)
|
|
|
|
def test_fetch_offset_select_literal_binds(self):
|
|
stmt = select(1).fetch(5).offset(6)
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT 1 OFFSET 6 ROWS FETCH FIRST 5 ROWS ONLY",
|
|
literal_binds=True,
|
|
)
|
|
|
|
def test_fetch_offset_compound_select_literal_binds(self):
|
|
stmt = select(1).union(select(2)).fetch(5).offset(6)
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT 1 UNION SELECT 2 OFFSET 6 ROWS FETCH FIRST 5 ROWS ONLY",
|
|
literal_binds=True,
|
|
)
|
|
|
|
def test_multiple_col_binds(self):
|
|
self.assert_compile(
|
|
select(literal_column("*")).where(
|
|
or_(
|
|
table1.c.myid == 12,
|
|
table1.c.myid == "asdf",
|
|
table1.c.myid == "foo",
|
|
),
|
|
),
|
|
"SELECT * FROM mytable WHERE mytable.myid = :myid_1 "
|
|
"OR mytable.myid = :myid_2 OR mytable.myid = :myid_3",
|
|
)
|
|
|
|
@testing.combinations("plain", "expanding", argnames="exprtype")
|
|
def test_literal_bind_typeerror(self, exprtype):
|
|
"""test #8800"""
|
|
|
|
if exprtype == "expanding":
|
|
stmt = select(table1).where(
|
|
table1.c.myid.in_([("tuple",), ("tuple",)])
|
|
)
|
|
elif exprtype == "plain":
|
|
stmt = select(table1).where(table1.c.myid == ("tuple",))
|
|
else:
|
|
assert False
|
|
|
|
with expect_raises_message(
|
|
exc.CompileError,
|
|
r"Could not render literal value \"\(\'tuple\',\)\" "
|
|
r"with datatype INTEGER; see parent "
|
|
r"stack trace for more detail.",
|
|
):
|
|
stmt.compile(compile_kwargs={"literal_binds": True})
|
|
|
|
@testing.combinations("plain", "expanding", argnames="exprtype")
|
|
def test_literal_bind_dont_know_how_to_quote(self, exprtype):
|
|
"""test #8800"""
|
|
|
|
class MyType(UserDefinedType):
|
|
def get_col_spec(self, **kw):
|
|
return "MYTYPE"
|
|
|
|
col = column("x", MyType())
|
|
|
|
if exprtype == "expanding":
|
|
stmt = select(table1).where(col.in_([("tuple",), ("tuple",)]))
|
|
elif exprtype == "plain":
|
|
stmt = select(table1).where(col == ("tuple",))
|
|
else:
|
|
assert False
|
|
|
|
with expect_raises_message(
|
|
exc.CompileError,
|
|
r"No literal value renderer is available for literal "
|
|
r"value \"\('tuple',\)\" with datatype MYTYPE",
|
|
):
|
|
stmt.compile(compile_kwargs={"literal_binds": True})
|
|
|
|
@testing.fixture
|
|
def ansi_compiler_fixture(self):
|
|
dialect = default.DefaultDialect()
|
|
|
|
class Compiler(compiler.StrSQLCompiler):
|
|
ansi_bind_rules = True
|
|
|
|
dialect.statement_compiler = Compiler
|
|
|
|
return dialect
|
|
|
|
@testing.combinations(
|
|
(
|
|
"one",
|
|
select(literal("someliteral")),
|
|
"SELECT __[POSTCOMPILE_param_1] AS anon_1",
|
|
dict(
|
|
check_literal_execute={"param_1": "someliteral"},
|
|
check_post_param={},
|
|
),
|
|
),
|
|
(
|
|
"two",
|
|
select(table1.c.myid + 3),
|
|
"SELECT mytable.myid + __[POSTCOMPILE_myid_1] "
|
|
"AS anon_1 FROM mytable",
|
|
dict(check_literal_execute={"myid_1": 3}, check_post_param={}),
|
|
),
|
|
(
|
|
"three",
|
|
select(table1.c.myid.in_([4, 5, 6])),
|
|
"SELECT mytable.myid IN (__[POSTCOMPILE_myid_1]) "
|
|
"AS anon_1 FROM mytable",
|
|
dict(
|
|
check_literal_execute={"myid_1": [4, 5, 6]},
|
|
check_post_param={},
|
|
),
|
|
),
|
|
(
|
|
"four",
|
|
select(func.mod(table1.c.myid, 5)),
|
|
"SELECT mod(mytable.myid, __[POSTCOMPILE_mod_2]) "
|
|
"AS mod_1 FROM mytable",
|
|
dict(check_literal_execute={"mod_2": 5}, check_post_param={}),
|
|
),
|
|
(
|
|
"five",
|
|
select(literal("foo").in_([])),
|
|
"SELECT __[POSTCOMPILE_param_1] IN (__[POSTCOMPILE_param_2]) "
|
|
"AS anon_1",
|
|
dict(
|
|
check_literal_execute={"param_1": "foo", "param_2": []},
|
|
check_post_param={},
|
|
),
|
|
),
|
|
(
|
|
"six",
|
|
select(literal(util.b("foo"))),
|
|
"SELECT __[POSTCOMPILE_param_1] AS anon_1",
|
|
dict(
|
|
check_literal_execute={"param_1": util.b("foo")},
|
|
check_post_param={},
|
|
),
|
|
),
|
|
(
|
|
"seven",
|
|
select(table1.c.myid == bindparam("foo", callable_=lambda: 5)),
|
|
"SELECT mytable.myid = __[POSTCOMPILE_foo] AS anon_1 FROM mytable",
|
|
dict(check_literal_execute={"foo": 5}, check_post_param={}),
|
|
),
|
|
argnames="stmt, expected, kw",
|
|
id_="iaaa",
|
|
)
|
|
def test_render_binds_as_literal(
|
|
self, ansi_compiler_fixture, stmt, expected, kw
|
|
):
|
|
"""test a compiler that renders binds inline into
|
|
SQL in the columns clause."""
|
|
|
|
self.assert_compile(
|
|
stmt, expected, dialect=ansi_compiler_fixture, **kw
|
|
)
|
|
|
|
def test_render_literal_execute_parameter(self):
|
|
self.assert_compile(
|
|
select(table1.c.myid).where(
|
|
table1.c.myid == bindparam("foo", 5, literal_execute=True)
|
|
),
|
|
"SELECT mytable.myid FROM mytable "
|
|
"WHERE mytable.myid = __[POSTCOMPILE_foo]",
|
|
)
|
|
|
|
def test_render_literal_execute_parameter_literal_binds(self):
|
|
self.assert_compile(
|
|
select(table1.c.myid).where(
|
|
table1.c.myid == bindparam("foo", 5, literal_execute=True)
|
|
),
|
|
"SELECT mytable.myid FROM mytable WHERE mytable.myid = 5",
|
|
literal_binds=True,
|
|
)
|
|
|
|
def test_render_literal_execute_sent_parameter_literal_binds(self):
|
|
"""test #6863"""
|
|
|
|
stmt = select(table1.c.myid).where(
|
|
table1.c.myid == bindparam("foo", 5, literal_execute=True)
|
|
)
|
|
eq_ignore_whitespace(
|
|
str(
|
|
stmt.compile(
|
|
compile_kwargs={
|
|
"literal_binds": True,
|
|
"literal_execute": True,
|
|
}
|
|
)
|
|
),
|
|
"SELECT mytable.myid FROM mytable WHERE mytable.myid = 5",
|
|
)
|
|
|
|
def test_render_literal_execute_parameter_render_postcompile(self):
|
|
self.assert_compile(
|
|
select(table1.c.myid).where(
|
|
table1.c.myid == bindparam("foo", 5, literal_execute=True)
|
|
),
|
|
"SELECT mytable.myid FROM mytable WHERE mytable.myid = 5",
|
|
render_postcompile=True,
|
|
)
|
|
|
|
def test_render_expanding_parameter(self):
|
|
self.assert_compile(
|
|
select(table1.c.myid).where(
|
|
table1.c.myid.in_(bindparam("foo", expanding=True))
|
|
),
|
|
"SELECT mytable.myid FROM mytable "
|
|
"WHERE mytable.myid IN (__[POSTCOMPILE_foo])",
|
|
)
|
|
|
|
def test_render_expanding_parameter_literal_binds(self):
|
|
self.assert_compile(
|
|
select(table1.c.myid).where(
|
|
table1.c.myid.in_(bindparam("foo", [1, 2, 3], expanding=True))
|
|
),
|
|
"SELECT mytable.myid FROM mytable "
|
|
"WHERE mytable.myid IN (1, 2, 3)",
|
|
literal_binds=True,
|
|
)
|
|
|
|
def test_render_expanding_parameter_render_postcompile(self):
|
|
# renders the IN the old way, essentially, but creates the bound
|
|
# parameters on the fly.
|
|
|
|
self.assert_compile(
|
|
select(table1.c.myid).where(
|
|
table1.c.myid.in_(bindparam("foo", [1, 2, 3], expanding=True))
|
|
),
|
|
"SELECT mytable.myid FROM mytable "
|
|
"WHERE mytable.myid IN (:foo_1, :foo_2, :foo_3)",
|
|
render_postcompile=True,
|
|
checkparams={"foo_1": 1, "foo_2": 2, "foo_3": 3},
|
|
)
|
|
|
|
@testing.combinations(
|
|
(
|
|
select(table1.c.myid).where(
|
|
table1.c.myid == bindparam("x", value=None)
|
|
),
|
|
"SELECT mytable.myid FROM mytable WHERE mytable.myid = NULL",
|
|
True,
|
|
None,
|
|
),
|
|
(
|
|
select(table1.c.myid).where(table1.c.myid == None),
|
|
"SELECT mytable.myid FROM mytable WHERE mytable.myid IS NULL",
|
|
False,
|
|
None,
|
|
),
|
|
(
|
|
select(table1.c.myid, None),
|
|
"SELECT mytable.myid, NULL AS anon_1 FROM mytable",
|
|
False,
|
|
None,
|
|
),
|
|
(
|
|
select(table1.c.myid).where(
|
|
table1.c.myid.is_(bindparam("x", value=None))
|
|
),
|
|
"SELECT mytable.myid FROM mytable WHERE mytable.myid IS NULL",
|
|
False,
|
|
None,
|
|
),
|
|
(
|
|
# as of SQLAlchemy 1.4, values like these are considered to be
|
|
# SQL expressions up front, so it is coerced to null()
|
|
# immediately and no bindparam() is created
|
|
table1.insert().values({"myid": None}),
|
|
"INSERT INTO mytable (myid) VALUES (NULL)",
|
|
False,
|
|
None,
|
|
),
|
|
(table1.insert(), "INSERT INTO mytable DEFAULT VALUES", False, {}),
|
|
(
|
|
table1.update().values({"myid": None}),
|
|
"UPDATE mytable SET myid=NULL",
|
|
False,
|
|
None,
|
|
),
|
|
(
|
|
table1.update()
|
|
.where(table1.c.myid == bindparam("x"))
|
|
.values({"myid": None}),
|
|
"UPDATE mytable SET myid=NULL WHERE mytable.myid = NULL",
|
|
True,
|
|
None,
|
|
),
|
|
)
|
|
def test_render_nulls_literal_binds(self, stmt, expected, warns, params):
|
|
if warns:
|
|
with testing.expect_warnings(
|
|
r"Bound parameter '.*?' rendering literal "
|
|
"NULL in a SQL expression"
|
|
):
|
|
self.assert_compile(
|
|
stmt, expected, literal_binds=True, params=params
|
|
)
|
|
else:
|
|
self.assert_compile(
|
|
stmt, expected, literal_binds=True, params=params
|
|
)
|
|
|
|
standalone_escape = testing.combinations(
|
|
("normalname", "normalname"),
|
|
("_name", "_name"),
|
|
("[BracketsAndCase]", "_BracketsAndCase_"),
|
|
("has spaces", "has_spaces"),
|
|
argnames="paramname, expected",
|
|
)
|
|
|
|
@standalone_escape
|
|
@testing.variation("use_positional", [True, False])
|
|
def test_standalone_bindparam_escape(
|
|
self, paramname, expected, use_positional
|
|
):
|
|
stmt = select(table1.c.myid).where(
|
|
table1.c.name == bindparam(paramname, value="x")
|
|
)
|
|
|
|
if use_positional:
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid FROM mytable WHERE mytable.name = ?",
|
|
params={paramname: "y"},
|
|
checkpositional=("y",),
|
|
dialect="sqlite",
|
|
)
|
|
else:
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid FROM mytable WHERE mytable.name = :%s"
|
|
% (expected,),
|
|
params={paramname: "y"},
|
|
checkparams={expected: "y"},
|
|
dialect="default",
|
|
)
|
|
|
|
@testing.variation("use_positional", [True, False])
|
|
def test_standalone_bindparam_escape_collision(self, use_positional):
|
|
"""this case is currently not supported
|
|
|
|
it's kinda bad since positional takes the unescaped param
|
|
while non positional takes the escaped one.
|
|
"""
|
|
stmt = select(table1.c.myid).where(
|
|
table1.c.name == bindparam("[brackets]", value="x"),
|
|
table1.c.description == bindparam("_brackets_", value="y"),
|
|
)
|
|
|
|
if use_positional:
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid FROM mytable WHERE mytable.name = ? "
|
|
"AND mytable.description = ?",
|
|
params={"[brackets]": "a", "_brackets_": "b"},
|
|
checkpositional=("a", "a"),
|
|
dialect="sqlite",
|
|
)
|
|
else:
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid FROM mytable WHERE mytable.name = "
|
|
":_brackets_ AND mytable.description = :_brackets_",
|
|
params={"[brackets]": "a", "_brackets_": "b"},
|
|
checkparams={"_brackets_": "b"},
|
|
dialect="default",
|
|
)
|
|
|
|
paramstyle = testing.variation("paramstyle", ["named", "qmark", "numeric"])
|
|
|
|
@standalone_escape
|
|
@paramstyle
|
|
def test_standalone_bindparam_escape_expanding_compile(
|
|
self, paramname, expected, paramstyle
|
|
):
|
|
stmt = select(table1.c.myid).where(
|
|
table1.c.name.in_(bindparam(paramname, value=["a", "b"]))
|
|
)
|
|
|
|
if paramstyle.qmark:
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid FROM mytable "
|
|
"WHERE mytable.name IN (?, ?)",
|
|
params={paramname: ["y", "z", "q"]},
|
|
checkpositional=("y", "z", "q"),
|
|
dialect="sqlite",
|
|
render_postcompile=True,
|
|
)
|
|
elif paramstyle.numeric:
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid FROM mytable "
|
|
"WHERE mytable.name IN (:1, :2)",
|
|
params={paramname: ["y", "z", "q"]},
|
|
checkpositional=("y", "z", "q"),
|
|
dialect=sqlite.dialect(paramstyle="numeric"),
|
|
render_postcompile=True,
|
|
)
|
|
elif paramstyle.named:
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid FROM mytable WHERE mytable.name IN "
|
|
"(:%s_1, :%s_2)" % (expected, expected),
|
|
params={paramname: ["y", "z"]},
|
|
checkparams={
|
|
"%s_1" % expected: "y",
|
|
"%s_2" % expected: "z",
|
|
},
|
|
dialect="default",
|
|
render_postcompile=True,
|
|
)
|
|
else:
|
|
paramstyle.fail()
|
|
|
|
@standalone_escape
|
|
@paramstyle
|
|
def test_standalone_bindparam_escape_expanding(
|
|
self, paramname, expected, paramstyle
|
|
):
|
|
stmt = select(table1.c.myid).where(
|
|
table1.c.name.in_(bindparam(paramname, value=["a", "b"]))
|
|
)
|
|
# this is what DefaultDialect actually does.
|
|
# this should be matched to DefaultDialect._init_compiled()
|
|
if paramstyle.qmark:
|
|
dialect = default.DefaultDialect(paramstyle="qmark")
|
|
elif paramstyle.numeric:
|
|
dialect = default.DefaultDialect(paramstyle="numeric")
|
|
else:
|
|
dialect = default.DefaultDialect()
|
|
|
|
compiled = stmt.compile(dialect=dialect)
|
|
checkparams = compiled.construct_params(
|
|
{paramname: ["y", "z"]}, escape_names=False
|
|
)
|
|
|
|
# nothing actually happened. if the compiler had
|
|
# render_postcompile set, the
|
|
# above weird param thing happens
|
|
eq_(checkparams, {paramname: ["y", "z"]})
|
|
|
|
expanded_state = compiled._process_parameters_for_postcompile(
|
|
checkparams
|
|
)
|
|
eq_(
|
|
expanded_state.additional_parameters,
|
|
{f"{expected}_1": "y", f"{expected}_2": "z"},
|
|
)
|
|
|
|
if paramstyle.qmark or paramstyle.numeric:
|
|
eq_(
|
|
expanded_state.positiontup,
|
|
[f"{expected}_1", f"{expected}_2"],
|
|
)
|
|
|
|
@paramstyle
|
|
def test_expanding_in_repeated(self, paramstyle):
|
|
stmt = (
|
|
select(table1)
|
|
.where(
|
|
table1.c.name.in_(
|
|
bindparam("uname", value=["h", "e"], expanding=True)
|
|
)
|
|
| table1.c.name.in_(
|
|
bindparam("uname2", value=["y"], expanding=True)
|
|
)
|
|
)
|
|
.where(table1.c.myid == 8)
|
|
)
|
|
stmt = stmt.union(
|
|
select(table1)
|
|
.where(
|
|
table1.c.name.in_(
|
|
bindparam("uname", value=["h", "e"], expanding=True)
|
|
)
|
|
| table1.c.name.in_(
|
|
bindparam("uname2", value=["y"], expanding=True)
|
|
)
|
|
)
|
|
.where(table1.c.myid == 9)
|
|
).order_by("myid")
|
|
|
|
if paramstyle.qmark:
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable "
|
|
"WHERE (mytable.name IN (?, ?) OR "
|
|
"mytable.name IN (?)) "
|
|
"AND mytable.myid = ? "
|
|
"UNION SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable "
|
|
"WHERE (mytable.name IN (?, ?) OR "
|
|
"mytable.name IN (?)) "
|
|
"AND mytable.myid = ? ORDER BY myid",
|
|
params={"uname": ["y", "z"], "uname2": ["a"]},
|
|
checkpositional=("y", "z", "a", 8, "y", "z", "a", 9),
|
|
dialect="sqlite",
|
|
render_postcompile=True,
|
|
)
|
|
elif paramstyle.numeric:
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable "
|
|
"WHERE (mytable.name IN (:3, :4) OR "
|
|
"mytable.name IN (:5)) "
|
|
"AND mytable.myid = :1 "
|
|
"UNION SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable "
|
|
"WHERE (mytable.name IN (:3, :4) OR "
|
|
"mytable.name IN (:5)) "
|
|
"AND mytable.myid = :2 ORDER BY myid",
|
|
params={"uname": ["y", "z"], "uname2": ["a"]},
|
|
checkpositional=(8, 9, "y", "z", "a"),
|
|
dialect=sqlite.dialect(paramstyle="numeric"),
|
|
render_postcompile=True,
|
|
)
|
|
elif paramstyle.named:
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable "
|
|
"WHERE (mytable.name IN (:uname_1, :uname_2) OR "
|
|
"mytable.name IN (:uname2_1)) "
|
|
"AND mytable.myid = :myid_1 "
|
|
"UNION SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable "
|
|
"WHERE (mytable.name IN (:uname_1, :uname_2) OR "
|
|
"mytable.name IN (:uname2_1)) "
|
|
"AND mytable.myid = :myid_2 ORDER BY myid",
|
|
params={"uname": ["y", "z"], "uname2": ["a"]},
|
|
checkparams={
|
|
"myid_1": 8,
|
|
"myid_2": 9,
|
|
"uname_1": "y",
|
|
"uname_2": "z",
|
|
"uname2_1": "a",
|
|
},
|
|
dialect="default",
|
|
render_postcompile=True,
|
|
)
|
|
else:
|
|
paramstyle.fail()
|
|
|
|
def test_numeric_dollar_bindparam(self):
|
|
stmt = table1.select().where(
|
|
table1.c.name == "a", table1.c.myid.in_([1, 2])
|
|
)
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable "
|
|
"WHERE mytable.name = $1 "
|
|
"AND mytable.myid IN ($2, $3)",
|
|
checkpositional=("a", 1, 2),
|
|
dialect=default.DefaultDialect(paramstyle="numeric_dollar"),
|
|
render_postcompile=True,
|
|
)
|
|
|
|
def test_bind_escape_extensibility(self):
|
|
"""test #8994, extensibility of the bind escape character lookup.
|
|
|
|
The main test for actual known characters passing through for bound
|
|
params is in
|
|
sqlalchemy.testing.suite.test_dialect.DifficultParametersTest.
|
|
|
|
"""
|
|
dialect = default.DefaultDialect()
|
|
|
|
class Compiler(compiler.StrSQLCompiler):
|
|
bindname_escape_characters = {
|
|
"%": "P",
|
|
# chars that need regex escaping
|
|
"(": "A",
|
|
")": "Z",
|
|
"*": "S",
|
|
"+": "L",
|
|
# completely random "normie" character
|
|
"8": "E",
|
|
":": "C",
|
|
# left bracket is not escaped, right bracket is
|
|
"]": "_",
|
|
" ": "_",
|
|
}
|
|
|
|
dialect.statement_compiler = Compiler
|
|
|
|
self.assert_compile(
|
|
select(
|
|
bindparam("number8ight"),
|
|
bindparam("plus+sign"),
|
|
bindparam("par(en)s and [brackets]"),
|
|
),
|
|
"SELECT :numberEight AS anon_1, :plusLsign AS anon_2, "
|
|
":parAenZs_and_[brackets_ AS anon_3",
|
|
dialect=dialect,
|
|
)
|
|
|
|
|
|
class CrudParamOverlapTest(AssertsCompiledSQL, fixtures.TestBase):
|
|
"""tests for #9075.
|
|
|
|
we apparently allow same-column-named bindparams in values(), even though
|
|
we do *not* allow same-column-named bindparams in other parts of the
|
|
statement, but only if the bindparam is associated with that column in the
|
|
VALUES / SET clause. If you use a name that matches that of a column in
|
|
values() but associate it with a different column, you also get the error.
|
|
|
|
This is supported, see
|
|
test_insert.py::InsertTest::test_binds_that_match_columns and
|
|
test_update.py::UpdateTest::test_binds_that_match_columns. The use
|
|
case makes sense because the "overlapping binds" issue is that using
|
|
a column name in bindparam() will conflict with the bindparam()
|
|
that crud.py is going to make for that column in VALUES / SET; but if we
|
|
are replacing the actual expression that would be in VALUES / SET, then
|
|
it's fine, there is no conflict.
|
|
|
|
The test suite is extended in
|
|
test/orm/test_core_compilation.py with ORM mappings that caused
|
|
the failure that was fixed by #9075.
|
|
|
|
|
|
"""
|
|
|
|
__dialect__ = "default"
|
|
|
|
@testing.fixture(
|
|
params=Variation.generate_cases("type_", ["lowercase", "uppercase"]),
|
|
ids=["lowercase", "uppercase"],
|
|
)
|
|
def crud_table_fixture(self, request):
|
|
type_ = request.param
|
|
|
|
if type_.lowercase:
|
|
table1 = table(
|
|
"mytable",
|
|
column("myid", Integer),
|
|
column("name", String),
|
|
column("description", String),
|
|
)
|
|
elif type_.uppercase:
|
|
table1 = Table(
|
|
"mytable",
|
|
MetaData(),
|
|
Column("myid", Integer),
|
|
Column("name", String),
|
|
Column("description", String),
|
|
)
|
|
else:
|
|
type_.fail()
|
|
|
|
yield table1
|
|
|
|
def test_same_named_binds_insert_values(self, crud_table_fixture):
|
|
table1 = crud_table_fixture
|
|
stmt = insert(table1).values(
|
|
myid=bindparam("myid"),
|
|
description=func.coalesce(bindparam("description"), "default"),
|
|
)
|
|
self.assert_compile(
|
|
stmt,
|
|
"INSERT INTO mytable (myid, description) VALUES "
|
|
"(:myid, coalesce(:description, :coalesce_1))",
|
|
)
|
|
|
|
self.assert_compile(
|
|
stmt,
|
|
"INSERT INTO mytable (myid, description) VALUES "
|
|
"(:myid, coalesce(:description, :coalesce_1))",
|
|
params={"myid": 5, "description": "foo"},
|
|
checkparams={
|
|
"coalesce_1": "default",
|
|
"description": "foo",
|
|
"myid": 5,
|
|
},
|
|
)
|
|
|
|
self.assert_compile(
|
|
stmt,
|
|
"INSERT INTO mytable (myid, name, description) VALUES "
|
|
"(:myid, :name, coalesce(:description, :coalesce_1))",
|
|
params={"myid": 5, "description": "foo", "name": "bar"},
|
|
checkparams={
|
|
"coalesce_1": "default",
|
|
"description": "foo",
|
|
"myid": 5,
|
|
"name": "bar",
|
|
},
|
|
)
|
|
|
|
def test_same_named_binds_update_values(self, crud_table_fixture):
|
|
table1 = crud_table_fixture
|
|
stmt = update(table1).values(
|
|
myid=bindparam("myid"),
|
|
description=func.coalesce(bindparam("description"), "default"),
|
|
)
|
|
self.assert_compile(
|
|
stmt,
|
|
"UPDATE mytable SET myid=:myid, "
|
|
"description=coalesce(:description, :coalesce_1)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
stmt,
|
|
"UPDATE mytable SET myid=:myid, "
|
|
"description=coalesce(:description, :coalesce_1)",
|
|
params={"myid": 5, "description": "foo"},
|
|
checkparams={
|
|
"coalesce_1": "default",
|
|
"description": "foo",
|
|
"myid": 5,
|
|
},
|
|
)
|
|
|
|
self.assert_compile(
|
|
stmt,
|
|
"UPDATE mytable SET myid=:myid, name=:name, "
|
|
"description=coalesce(:description, :coalesce_1)",
|
|
params={"myid": 5, "description": "foo", "name": "bar"},
|
|
checkparams={
|
|
"coalesce_1": "default",
|
|
"description": "foo",
|
|
"myid": 5,
|
|
"name": "bar",
|
|
},
|
|
)
|
|
|
|
def test_different_named_binds_insert_values(self, crud_table_fixture):
|
|
table1 = crud_table_fixture
|
|
stmt = insert(table1).values(
|
|
myid=bindparam("myid"),
|
|
name=func.coalesce(bindparam("description"), "default"),
|
|
)
|
|
self.assert_compile(
|
|
stmt,
|
|
"INSERT INTO mytable (myid, name) VALUES "
|
|
"(:myid, coalesce(:description, :coalesce_1))",
|
|
)
|
|
|
|
with expect_raises_message(
|
|
exc.CompileError, r"bindparam\(\) name 'description' is reserved "
|
|
):
|
|
stmt.compile(column_keys=["myid", "description"])
|
|
|
|
with expect_raises_message(
|
|
exc.CompileError, r"bindparam\(\) name 'description' is reserved "
|
|
):
|
|
stmt.compile(column_keys=["myid", "description", "name"])
|
|
|
|
def test_different_named_binds_update_values(self, crud_table_fixture):
|
|
table1 = crud_table_fixture
|
|
stmt = update(table1).values(
|
|
myid=bindparam("myid"),
|
|
name=func.coalesce(bindparam("description"), "default"),
|
|
)
|
|
self.assert_compile(
|
|
stmt,
|
|
"UPDATE mytable SET myid=:myid, "
|
|
"name=coalesce(:description, :coalesce_1)",
|
|
)
|
|
|
|
with expect_raises_message(
|
|
exc.CompileError, r"bindparam\(\) name 'description' is reserved "
|
|
):
|
|
stmt.compile(column_keys=["myid", "description"])
|
|
|
|
with expect_raises_message(
|
|
exc.CompileError, r"bindparam\(\) name 'description' is reserved "
|
|
):
|
|
stmt.compile(column_keys=["myid", "description", "name"])
|
|
|
|
|
|
class CompileUXTest(fixtures.TestBase):
|
|
"""tests focused on calling stmt.compile() directly, user cases"""
|
|
|
|
@testing.fixture
|
|
def render_postcompile_fixture(self):
|
|
return (
|
|
select(func.count(1))
|
|
.where(column("q") == "x")
|
|
.where(column("z").in_([1, 2, 3]))
|
|
.where(column("z_tuple").in_([(1, "a"), (2, "b"), (3, "c")]))
|
|
.where(
|
|
column("y").op("foobar")(
|
|
bindparam(
|
|
"key", value=[("1", "2"), ("3", "4")], expanding=True
|
|
)
|
|
)
|
|
)
|
|
)
|
|
|
|
def test_render_postcompile_default_stmt(self, render_postcompile_fixture):
|
|
stmt = render_postcompile_fixture
|
|
|
|
compiled = stmt.compile(compile_kwargs={"render_postcompile": True})
|
|
eq_ignore_whitespace(
|
|
compiled.string,
|
|
"SELECT count(:count_2) AS count_1 WHERE q = :q_1 AND z "
|
|
"IN (:z_1_1, :z_1_2, :z_1_3) AND z_tuple IN "
|
|
"((:z_tuple_1_1_1, :z_tuple_1_1_2), "
|
|
"(:z_tuple_1_2_1, :z_tuple_1_2_2), "
|
|
"(:z_tuple_1_3_1, :z_tuple_1_3_2)) "
|
|
"AND (y foobar ((:key_1_1, :key_1_2), (:key_2_1, :key_2_2)))",
|
|
)
|
|
|
|
def test_render_postcompile_named_parameters(
|
|
self, render_postcompile_fixture
|
|
):
|
|
stmt = render_postcompile_fixture
|
|
|
|
compiled = stmt.compile(compile_kwargs={"render_postcompile": True})
|
|
is_none(compiled.positiontup)
|
|
eq_(
|
|
compiled.construct_params(),
|
|
{
|
|
"count_2": 1,
|
|
"q_1": "x",
|
|
"z_1_1": 1,
|
|
"z_1_2": 2,
|
|
"z_1_3": 3,
|
|
"z_tuple_1_1_1": 1,
|
|
"z_tuple_1_1_2": "a",
|
|
"z_tuple_1_2_1": 2,
|
|
"z_tuple_1_2_2": "b",
|
|
"z_tuple_1_3_1": 3,
|
|
"z_tuple_1_3_2": "c",
|
|
"key_1_1": "1",
|
|
"key_1_2": "2",
|
|
"key_2_1": "3",
|
|
"key_2_2": "4",
|
|
},
|
|
)
|
|
|
|
def test_render_postcompile_no_new_params(
|
|
self, render_postcompile_fixture
|
|
):
|
|
stmt = render_postcompile_fixture
|
|
|
|
compiled = stmt.compile(compile_kwargs={"render_postcompile": True})
|
|
params = {"q_1": "g"}
|
|
with expect_raises_message(
|
|
exc.InvalidRequestError,
|
|
"can't construct new parameters when render_postcompile is used; "
|
|
"the statement is hard-linked to the original parameters.",
|
|
):
|
|
compiled.construct_params(params)
|
|
|
|
@testing.variation("render_postcompile", [True, False])
|
|
def test_new_expanded_state_no_params(
|
|
self, render_postcompile_fixture: Select, render_postcompile
|
|
):
|
|
stmt = render_postcompile_fixture
|
|
|
|
compiled = stmt.compile(
|
|
compile_kwargs={"render_postcompile": render_postcompile}
|
|
)
|
|
is_none(compiled.positiontup)
|
|
|
|
es = compiled.construct_expanded_state()
|
|
|
|
is_none(compiled.positiontup)
|
|
|
|
eq_ignore_whitespace(
|
|
es.statement,
|
|
"SELECT count(:count_2) AS count_1 WHERE q = :q_1 AND z "
|
|
"IN (:z_1_1, :z_1_2, :z_1_3) AND z_tuple IN "
|
|
"((:z_tuple_1_1_1, :z_tuple_1_1_2), "
|
|
"(:z_tuple_1_2_1, :z_tuple_1_2_2), "
|
|
"(:z_tuple_1_3_1, :z_tuple_1_3_2)) "
|
|
"AND (y foobar ((:key_1_1, :key_1_2), (:key_2_1, :key_2_2)))",
|
|
)
|
|
|
|
eq_(
|
|
es.parameters,
|
|
{
|
|
"count_2": 1,
|
|
"q_1": "x",
|
|
"z_1_1": 1,
|
|
"z_1_2": 2,
|
|
"z_1_3": 3,
|
|
"z_tuple_1_1_1": 1,
|
|
"z_tuple_1_1_2": "a",
|
|
"z_tuple_1_2_1": 2,
|
|
"z_tuple_1_2_2": "b",
|
|
"z_tuple_1_3_1": 3,
|
|
"z_tuple_1_3_2": "c",
|
|
"key_1_1": "1",
|
|
"key_1_2": "2",
|
|
"key_2_1": "3",
|
|
"key_2_2": "4",
|
|
},
|
|
)
|
|
|
|
@testing.variation("render_postcompile", [True, False])
|
|
@testing.variation("positional", [True, False])
|
|
def test_accessor_no_params(self, render_postcompile, positional):
|
|
stmt = select(column("q"))
|
|
|
|
positional_dialect = default.DefaultDialect(
|
|
paramstyle="qmark" if positional else "pyformat"
|
|
)
|
|
compiled = stmt.compile(
|
|
dialect=positional_dialect,
|
|
compile_kwargs={"render_postcompile": render_postcompile},
|
|
)
|
|
if positional:
|
|
eq_(compiled.positiontup, [])
|
|
else:
|
|
is_none(compiled.positiontup)
|
|
eq_(compiled.params, {})
|
|
eq_(compiled.construct_params(), {})
|
|
|
|
es = compiled.construct_expanded_state()
|
|
if positional:
|
|
eq_(es.positiontup, [])
|
|
eq_(es.positional_parameters, ())
|
|
else:
|
|
is_none(es.positiontup)
|
|
with expect_raises_message(
|
|
exc.InvalidRequestError,
|
|
"statement does not use a positional paramstyle",
|
|
):
|
|
es.positional_parameters
|
|
eq_(es.parameters, {})
|
|
|
|
eq_ignore_whitespace(
|
|
es.statement,
|
|
"SELECT q",
|
|
)
|
|
|
|
@testing.variation("render_postcompile", [True, False])
|
|
def test_new_expanded_state_new_params(
|
|
self, render_postcompile_fixture: Select, render_postcompile
|
|
):
|
|
stmt = render_postcompile_fixture
|
|
|
|
compiled = stmt.compile(
|
|
compile_kwargs={"render_postcompile": render_postcompile}
|
|
)
|
|
is_none(compiled.positiontup)
|
|
|
|
es = compiled.construct_expanded_state(
|
|
{
|
|
"z_tuple_1": [("q", "z", "p"), ("g", "h", "i")],
|
|
"key": ["a", "b"],
|
|
}
|
|
)
|
|
is_none(compiled.positiontup)
|
|
|
|
eq_ignore_whitespace(
|
|
es.statement,
|
|
"SELECT count(:count_2) AS count_1 WHERE q = :q_1 AND z IN "
|
|
"(:z_1_1, :z_1_2, :z_1_3) AND z_tuple IN "
|
|
"((:z_tuple_1_1_1, :z_tuple_1_1_2, :z_tuple_1_1_3), "
|
|
"(:z_tuple_1_2_1, :z_tuple_1_2_2, :z_tuple_1_2_3)) AND "
|
|
"(y foobar (:key_1, :key_2))",
|
|
)
|
|
|
|
eq_(
|
|
es.parameters,
|
|
{
|
|
"count_2": 1,
|
|
"q_1": "x",
|
|
"z_1_1": 1,
|
|
"z_1_2": 2,
|
|
"z_1_3": 3,
|
|
"z_tuple_1_1_1": "q",
|
|
"z_tuple_1_1_2": "z",
|
|
"z_tuple_1_1_3": "p",
|
|
"z_tuple_1_2_1": "g",
|
|
"z_tuple_1_2_2": "h",
|
|
"z_tuple_1_2_3": "i",
|
|
"key_1": "a",
|
|
"key_2": "b",
|
|
},
|
|
)
|
|
|
|
@testing.variation("render_postcompile", [True, False])
|
|
@testing.variation("paramstyle", ["qmark", "numeric"])
|
|
def test_new_expanded_state_new_positional_params(
|
|
self,
|
|
render_postcompile_fixture: Select,
|
|
render_postcompile,
|
|
paramstyle,
|
|
):
|
|
stmt = render_postcompile_fixture
|
|
positional_dialect = default.DefaultDialect(paramstyle=paramstyle.name)
|
|
|
|
compiled = stmt.compile(
|
|
dialect=positional_dialect,
|
|
compile_kwargs={"render_postcompile": render_postcompile},
|
|
)
|
|
|
|
if render_postcompile:
|
|
eq_(
|
|
compiled.positiontup,
|
|
[
|
|
"count_2",
|
|
"q_1",
|
|
"z_1_1",
|
|
"z_1_2",
|
|
"z_1_3",
|
|
"z_tuple_1_1_1",
|
|
"z_tuple_1_1_2",
|
|
"z_tuple_1_2_1",
|
|
"z_tuple_1_2_2",
|
|
"z_tuple_1_3_1",
|
|
"z_tuple_1_3_2",
|
|
"key_1_1",
|
|
"key_1_2",
|
|
"key_2_1",
|
|
"key_2_2",
|
|
],
|
|
)
|
|
else:
|
|
eq_(
|
|
compiled.positiontup,
|
|
["count_2", "q_1", "z_1", "z_tuple_1", "key"],
|
|
)
|
|
es = compiled.construct_expanded_state(
|
|
{
|
|
"z_tuple_1": [("q", "z", "p"), ("g", "h", "i")],
|
|
"key": ["a", "b"],
|
|
}
|
|
)
|
|
if paramstyle.qmark:
|
|
eq_ignore_whitespace(
|
|
es.statement,
|
|
"SELECT count(?) AS count_1 WHERE q = ? "
|
|
"AND z IN (?, ?, ?) AND "
|
|
"z_tuple IN ((?, ?, ?), (?, ?, ?)) AND (y foobar (?, ?))",
|
|
)
|
|
elif paramstyle.numeric:
|
|
eq_ignore_whitespace(
|
|
es.statement,
|
|
"SELECT count(:1) AS count_1 WHERE q = :2 AND z IN "
|
|
"(:3, :4, :5) AND z_tuple "
|
|
"IN ((:6, :7, :8), (:9, :10, :11)) AND (y foobar (:12, :13))",
|
|
)
|
|
else:
|
|
paramstyle.fail()
|
|
|
|
eq_(
|
|
es.parameters,
|
|
{
|
|
"count_2": 1,
|
|
"q_1": "x",
|
|
"z_1_1": 1,
|
|
"z_1_2": 2,
|
|
"z_1_3": 3,
|
|
"z_tuple_1_1_1": "q",
|
|
"z_tuple_1_1_2": "z",
|
|
"z_tuple_1_1_3": "p",
|
|
"z_tuple_1_2_1": "g",
|
|
"z_tuple_1_2_2": "h",
|
|
"z_tuple_1_2_3": "i",
|
|
"key_1": "a",
|
|
"key_2": "b",
|
|
},
|
|
)
|
|
eq_(
|
|
es.positiontup,
|
|
[
|
|
"count_2",
|
|
"q_1",
|
|
"z_1_1",
|
|
"z_1_2",
|
|
"z_1_3",
|
|
"z_tuple_1_1_1",
|
|
"z_tuple_1_1_2",
|
|
"z_tuple_1_1_3",
|
|
"z_tuple_1_2_1",
|
|
"z_tuple_1_2_2",
|
|
"z_tuple_1_2_3",
|
|
"key_1",
|
|
"key_2",
|
|
],
|
|
)
|
|
eq_(
|
|
es.positional_parameters,
|
|
(1, "x", 1, 2, 3, "q", "z", "p", "g", "h", "i", "a", "b"),
|
|
)
|
|
|
|
def test_render_postcompile_positional_parameters(
|
|
self, render_postcompile_fixture
|
|
):
|
|
stmt = render_postcompile_fixture
|
|
|
|
positional_dialect = default.DefaultDialect(paramstyle="qmark")
|
|
compiled = stmt.compile(
|
|
dialect=positional_dialect,
|
|
compile_kwargs={"render_postcompile": True},
|
|
)
|
|
eq_(
|
|
compiled.construct_params(),
|
|
{
|
|
"count_2": 1,
|
|
"q_1": "x",
|
|
"z_1_1": 1,
|
|
"z_1_2": 2,
|
|
"z_1_3": 3,
|
|
"z_tuple_1_1_1": 1,
|
|
"z_tuple_1_1_2": "a",
|
|
"z_tuple_1_2_1": 2,
|
|
"z_tuple_1_2_2": "b",
|
|
"z_tuple_1_3_1": 3,
|
|
"z_tuple_1_3_2": "c",
|
|
"key_1_1": "1",
|
|
"key_1_2": "2",
|
|
"key_2_1": "3",
|
|
"key_2_2": "4",
|
|
},
|
|
)
|
|
eq_(
|
|
compiled.positiontup,
|
|
[
|
|
"count_2",
|
|
"q_1",
|
|
"z_1_1",
|
|
"z_1_2",
|
|
"z_1_3",
|
|
"z_tuple_1_1_1",
|
|
"z_tuple_1_1_2",
|
|
"z_tuple_1_2_1",
|
|
"z_tuple_1_2_2",
|
|
"z_tuple_1_3_1",
|
|
"z_tuple_1_3_2",
|
|
"key_1_1",
|
|
"key_1_2",
|
|
"key_2_1",
|
|
"key_2_2",
|
|
],
|
|
)
|
|
|
|
|
|
class UnsupportedTest(fixtures.TestBase):
|
|
def test_unsupported_element_str_visit_name(self):
|
|
from sqlalchemy.sql.expression import ClauseElement
|
|
|
|
class SomeElement(ClauseElement):
|
|
__visit_name__ = "some_element"
|
|
|
|
assert_raises_message(
|
|
exc.UnsupportedCompilationError,
|
|
r"Compiler <sqlalchemy.sql.compiler.StrSQLCompiler .*"
|
|
r"can't render element of type <class '.*SomeElement'>",
|
|
SomeElement().compile,
|
|
)
|
|
|
|
def test_unsupported_element_meth_visit_name(self):
|
|
from sqlalchemy.sql.expression import ClauseElement
|
|
|
|
def go():
|
|
class SomeElement(ClauseElement):
|
|
@classmethod
|
|
def __visit_name__(cls):
|
|
return "some_element"
|
|
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
r"__visit_name__ on class SomeElement must be a string at "
|
|
r"the class level",
|
|
go,
|
|
)
|
|
|
|
def test_unsupported_type(self):
|
|
class MyType(types.TypeEngine):
|
|
__visit_name__ = "mytype"
|
|
|
|
t = Table("t", MetaData(), Column("q", MyType()))
|
|
|
|
with expect_raises_message(
|
|
exc.CompileError,
|
|
r"\(in table 't', column 'q'\): Compiler .*SQLiteTypeCompiler.* "
|
|
r"can't render element of type MyType\(\)",
|
|
):
|
|
schema.CreateTable(t).compile(dialect=sqlite.dialect())
|
|
|
|
def test_unsupported_operator(self):
|
|
from sqlalchemy.sql.expression import BinaryExpression
|
|
|
|
def myop(x, y):
|
|
pass
|
|
|
|
binary = BinaryExpression(column("foo"), column("bar"), myop)
|
|
assert_raises_message(
|
|
exc.UnsupportedCompilationError,
|
|
r"Compiler <sqlalchemy.sql.compiler.StrSQLCompiler .*"
|
|
r"can't render element of type <function.*",
|
|
binary.compile,
|
|
)
|
|
|
|
|
|
class StringifySpecialTest(fixtures.TestBase):
|
|
def test_basic(self):
|
|
stmt = select(table1).where(table1.c.myid == 10)
|
|
eq_ignore_whitespace(
|
|
str(stmt),
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable WHERE mytable.myid = :myid_1",
|
|
)
|
|
|
|
def test_unnamed_column(self):
|
|
stmt = Column(Integer) == 5
|
|
eq_ignore_whitespace(str(stmt), '"<name unknown>" = :param_1')
|
|
|
|
def test_empty_insert(self):
|
|
stmt = table1.insert().values()
|
|
eq_ignore_whitespace(str(stmt), "INSERT INTO mytable () VALUES ()")
|
|
|
|
def test_insert_return_defaults(self):
|
|
t = Table(
|
|
"t",
|
|
MetaData(),
|
|
Column("myid", Integer, primary_key=True),
|
|
Column("name", String, server_default="some str"),
|
|
Column("description", String, default=func.lower("hi")),
|
|
)
|
|
stmt = t.insert().values().return_defaults()
|
|
eq_ignore_whitespace(
|
|
str(stmt),
|
|
"INSERT INTO t (description) VALUES (lower(:lower_1)) "
|
|
"RETURNING t.myid, t.name, t.description",
|
|
)
|
|
|
|
def test_multirow_insert(self):
|
|
stmt = table1.insert().values([{"myid": 1}, {"myid": 2}])
|
|
eq_ignore_whitespace(
|
|
str(stmt),
|
|
"INSERT INTO mytable (myid) VALUES (:myid_m0), (:myid_m1)",
|
|
)
|
|
|
|
def test_multirow_insert_positional(self):
|
|
stmt = table1.insert().values([{"myid": 1}, {"myid": 2}])
|
|
eq_ignore_whitespace(
|
|
stmt.compile(dialect=sqlite.dialect()).string,
|
|
"INSERT INTO mytable (myid) VALUES (?), (?)",
|
|
)
|
|
|
|
def test_multirow_insert_numeric(self):
|
|
stmt = table1.insert().values([{"myid": 1}, {"myid": 2}])
|
|
eq_ignore_whitespace(
|
|
stmt.compile(dialect=sqlite.dialect(paramstyle="numeric")).string,
|
|
"INSERT INTO mytable (myid) VALUES (:1), (:2)",
|
|
)
|
|
|
|
def test_insert_noparams_numeric(self):
|
|
ii = table1.insert().returning(table1.c.myid)
|
|
eq_ignore_whitespace(
|
|
ii.compile(dialect=sqlite.dialect(paramstyle="numeric")).string,
|
|
"INSERT INTO mytable (myid, name, description) VALUES "
|
|
"(:1, :2, :3) RETURNING myid",
|
|
)
|
|
|
|
def test_cte(self):
|
|
# stringify of these was supported anyway by defaultdialect.
|
|
stmt = select(table1.c.myid).cte()
|
|
stmt = select(stmt)
|
|
eq_ignore_whitespace(
|
|
str(stmt),
|
|
"WITH anon_1 AS (SELECT mytable.myid AS myid FROM mytable) "
|
|
"SELECT anon_1.myid FROM anon_1",
|
|
)
|
|
|
|
@testing.combinations(("cte",), ("alias",), ("subquery",))
|
|
def test_grouped_selectables_print_alone(self, modifier):
|
|
stmt = select(table1).where(table1.c.myid == 10)
|
|
|
|
grouped = getattr(stmt, modifier)()
|
|
eq_ignore_whitespace(
|
|
str(grouped),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM mytable WHERE mytable.myid = :myid_1",
|
|
)
|
|
|
|
def test_next_sequence_value(self):
|
|
# using descriptive text that is intentionally not compatible
|
|
# with any particular backend, since all backends have different
|
|
# syntax
|
|
|
|
seq = Sequence("my_sequence")
|
|
|
|
eq_ignore_whitespace(
|
|
str(seq.next_value()), "<next sequence value: my_sequence>"
|
|
)
|
|
|
|
def test_returning(self):
|
|
stmt = table1.insert().returning(table1.c.myid)
|
|
|
|
eq_ignore_whitespace(
|
|
str(stmt),
|
|
"INSERT INTO mytable (myid, name, description) "
|
|
"VALUES (:myid, :name, :description) RETURNING mytable.myid",
|
|
)
|
|
|
|
def test_array_index(self):
|
|
stmt = select(column("foo", types.ARRAY(Integer))[5])
|
|
|
|
eq_ignore_whitespace(str(stmt), "SELECT foo[:foo_1] AS anon_1")
|
|
|
|
def test_unknown_type(self):
|
|
class MyType(types.TypeEngine):
|
|
__visit_name__ = "mytype"
|
|
|
|
stmt = select(cast(table1.c.myid, MyType))
|
|
|
|
eq_ignore_whitespace(
|
|
str(stmt),
|
|
"SELECT CAST(mytable.myid AS MyType()) AS myid FROM mytable",
|
|
)
|
|
|
|
def test_dialect_sub_compile(self):
|
|
class Widget(ClauseElement):
|
|
__visit_name__ = "widget"
|
|
stringify_dialect = "sqlite"
|
|
|
|
def visit_widget(self, element, **kw):
|
|
return "widget"
|
|
|
|
with mock.patch(
|
|
"sqlalchemy.dialects.sqlite.base.SQLiteCompiler.visit_widget",
|
|
visit_widget,
|
|
create=True,
|
|
):
|
|
eq_(str(Grouping(Widget())), "(widget)")
|
|
|
|
def test_dialect_sub_compile_has_stack(self):
|
|
"""test #10753"""
|
|
|
|
class Widget(ColumnElement):
|
|
__visit_name__ = "widget"
|
|
stringify_dialect = "sqlite"
|
|
|
|
def visit_widget(self, element, **kw):
|
|
assert self.stack
|
|
return "widget"
|
|
|
|
with mock.patch(
|
|
"sqlalchemy.dialects.sqlite.base.SQLiteCompiler.visit_widget",
|
|
visit_widget,
|
|
create=True,
|
|
):
|
|
eq_(str(select(Widget())), "SELECT widget AS anon_1")
|
|
|
|
def test_dialect_sub_compile_has_stack_pg_specific(self):
|
|
"""test #10753"""
|
|
my_table = table(
|
|
"my_table", column("id"), column("data"), column("user_email")
|
|
)
|
|
|
|
from sqlalchemy.dialects.postgresql import insert
|
|
|
|
insert_stmt = insert(my_table).values(
|
|
id="some_existing_id", data="inserted value"
|
|
)
|
|
|
|
do_update_stmt = insert_stmt.on_conflict_do_update(
|
|
index_elements=["id"], set_=dict(data="updated value")
|
|
)
|
|
|
|
# note! two different bound parameter formats. It's weird yes,
|
|
# but this is what I want. They are stringifying without using the
|
|
# correct dialect. We could use the PG compiler at the point of
|
|
# the insert() but that still would not accommodate params in other
|
|
# parts of the statement.
|
|
eq_ignore_whitespace(
|
|
str(select(do_update_stmt.cte())),
|
|
"WITH anon_1 AS (INSERT INTO my_table (id, data) "
|
|
"VALUES (:param_1, :param_2) "
|
|
"ON CONFLICT (id) "
|
|
"DO UPDATE SET data = %(param_3)s) SELECT FROM anon_1",
|
|
)
|
|
|
|
def test_dialect_sub_compile_w_binds(self):
|
|
"""test sub-compile into a new compiler where
|
|
state != CompilerState.COMPILING, but we have to render a bindparam
|
|
string. has to render the correct template.
|
|
|
|
"""
|
|
|
|
class Widget(ClauseElement):
|
|
__visit_name__ = "widget"
|
|
stringify_dialect = "sqlite"
|
|
|
|
def visit_widget(self, element, **kw):
|
|
return f"widget {self.process(bindparam('q'), **kw)}"
|
|
|
|
with mock.patch(
|
|
"sqlalchemy.dialects.sqlite.base.SQLiteCompiler.visit_widget",
|
|
visit_widget,
|
|
create=True,
|
|
):
|
|
eq_(str(Grouping(Widget())), "(widget ?)")
|
|
|
|
def test_within_group(self):
|
|
# stringify of these was supported anyway by defaultdialect.
|
|
from sqlalchemy import within_group
|
|
|
|
stmt = select(
|
|
table1.c.myid,
|
|
within_group(func.percentile_cont(0.5), table1.c.name.desc()),
|
|
)
|
|
eq_ignore_whitespace(
|
|
str(stmt),
|
|
"SELECT mytable.myid, percentile_cont(:percentile_cont_2) "
|
|
"WITHIN GROUP (ORDER BY mytable.name DESC) AS percentile_cont_1 "
|
|
"FROM mytable",
|
|
)
|
|
|
|
@testing.combinations(
|
|
("datetime", datetime.datetime.now()),
|
|
("date", datetime.date.today()),
|
|
("time", datetime.time()),
|
|
argnames="value",
|
|
id_="ia",
|
|
)
|
|
def test_render_datetime(self, value):
|
|
lit = literal(value)
|
|
|
|
eq_ignore_whitespace(
|
|
str(lit.compile(compile_kwargs={"literal_binds": True})),
|
|
"'%s'" % value,
|
|
)
|
|
|
|
def test_with_hint_table(self):
|
|
stmt = (
|
|
select(table1)
|
|
.select_from(
|
|
table1.join(table2, table1.c.myid == table2.c.otherid)
|
|
)
|
|
.with_hint(table1, "use some_hint")
|
|
)
|
|
|
|
# note that some dialects instead use the "with_select_hint"
|
|
# hook to put the 'hint' up front
|
|
eq_ignore_whitespace(
|
|
str(stmt),
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable [use some_hint] "
|
|
"JOIN myothertable ON mytable.myid = myothertable.otherid",
|
|
)
|
|
|
|
def test_with_hint_statement(self):
|
|
stmt = (
|
|
select(table1)
|
|
.select_from(
|
|
table1.join(table2, table1.c.myid == table2.c.otherid)
|
|
)
|
|
.with_statement_hint("use some_hint")
|
|
)
|
|
|
|
eq_ignore_whitespace(
|
|
str(stmt),
|
|
"SELECT mytable.myid, mytable.name, mytable.description "
|
|
"FROM mytable "
|
|
"JOIN myothertable ON mytable.myid = myothertable.otherid "
|
|
"use some_hint",
|
|
)
|
|
|
|
def test_dialect_specific_sql(self):
|
|
my_table = table(
|
|
"my_table", column("id"), column("data"), column("user_email")
|
|
)
|
|
|
|
from sqlalchemy.dialects.postgresql import insert
|
|
|
|
insert_stmt = insert(my_table).values(
|
|
id="some_existing_id", data="inserted value"
|
|
)
|
|
|
|
do_update_stmt = insert_stmt.on_conflict_do_update(
|
|
index_elements=["id"], set_=dict(data="updated value")
|
|
)
|
|
eq_ignore_whitespace(
|
|
str(do_update_stmt),
|
|
"INSERT INTO my_table (id, data) VALUES (%(id)s, %(data)s) "
|
|
"ON CONFLICT (id) DO UPDATE SET data = %(param_1)s",
|
|
)
|
|
|
|
def test_dialect_specific_ddl(self):
|
|
from sqlalchemy.dialects.postgresql import ExcludeConstraint
|
|
|
|
m = MetaData()
|
|
tbl = Table("testtbl", m, Column("room", Integer, primary_key=True))
|
|
cons = ExcludeConstraint(("room", "="))
|
|
tbl.append_constraint(cons)
|
|
|
|
eq_ignore_whitespace(
|
|
str(schema.AddConstraint(cons)),
|
|
"ALTER TABLE testtbl ADD EXCLUDE USING gist (room WITH =)",
|
|
)
|
|
|
|
def test_try_cast(self):
|
|
t1 = Table("t1", MetaData(), Column("id", Integer, primary_key=True))
|
|
expr = select(try_cast(t1.c.id, Integer))
|
|
|
|
eq_ignore_whitespace(
|
|
str(expr),
|
|
"SELECT TRY_CAST(t1.id AS INTEGER) AS id FROM t1",
|
|
)
|
|
|
|
|
|
class KwargPropagationTest(fixtures.TestBase):
|
|
@classmethod
|
|
def setup_test_class(cls):
|
|
class CatchCol(ColumnClause):
|
|
pass
|
|
|
|
class CatchTable(TableClause):
|
|
pass
|
|
|
|
cls.column = CatchCol("x")
|
|
cls.table = CatchTable("y")
|
|
cls.criterion = cls.column == CatchCol("y")
|
|
|
|
@compiles(CatchCol)
|
|
def compile_col(element, compiler, **kw):
|
|
assert "canary" in kw
|
|
return compiler.visit_column(element)
|
|
|
|
@compiles(CatchTable)
|
|
def compile_table(element, compiler, **kw):
|
|
assert "canary" in kw
|
|
return compiler.visit_table(element)
|
|
|
|
def _do_test(self, element):
|
|
d = default.DefaultDialect()
|
|
d.statement_compiler(d, element, compile_kwargs={"canary": True})
|
|
|
|
def test_binary(self):
|
|
self._do_test(self.column == 5)
|
|
|
|
def test_select(self):
|
|
s = (
|
|
select(self.column)
|
|
.select_from(self.table)
|
|
.where(self.column == self.criterion)
|
|
.order_by(self.column)
|
|
)
|
|
self._do_test(s)
|
|
|
|
def test_case(self):
|
|
c = case((self.criterion, self.column), else_=self.column)
|
|
self._do_test(c)
|
|
|
|
def test_cast(self):
|
|
c = cast(self.column, Integer)
|
|
self._do_test(c)
|
|
|
|
|
|
class ExecutionOptionsTest(fixtures.TestBase):
|
|
def test_non_dml(self):
|
|
stmt = table1.select().execution_options(foo="bar")
|
|
compiled = stmt.compile()
|
|
|
|
eq_(compiled.execution_options, {"foo": "bar"})
|
|
|
|
def test_dml(self):
|
|
stmt = table1.insert().execution_options(foo="bar")
|
|
compiled = stmt.compile()
|
|
|
|
eq_(compiled.execution_options, {"foo": "bar"})
|
|
|
|
def test_embedded_element_true_to_none(self):
|
|
stmt = table1.insert().execution_options(foo="bar")
|
|
eq_(stmt._execution_options, {"foo": "bar"})
|
|
s2 = select(table1).select_from(stmt.cte())
|
|
eq_(s2._execution_options, {})
|
|
|
|
compiled = s2.compile()
|
|
eq_(compiled.execution_options, {})
|
|
|
|
def test_embedded_element_true_to_false(self):
|
|
stmt = table1.insert().execution_options(foo="bar")
|
|
eq_(stmt._execution_options, {"foo": "bar"})
|
|
s2 = (
|
|
select(table1).select_from(stmt.cte()).execution_options(foo="bat")
|
|
)
|
|
eq_(s2._execution_options, {"foo": "bat"})
|
|
|
|
compiled = s2.compile()
|
|
eq_(compiled.execution_options, {"foo": "bat"})
|
|
|
|
|
|
class DDLTest(fixtures.TestBase, AssertsCompiledSQL):
|
|
__dialect__ = "default"
|
|
|
|
def _illegal_type_fixture(self):
|
|
class MyType(types.TypeEngine):
|
|
pass
|
|
|
|
@compiles(MyType)
|
|
def compile_(element, compiler, **kw):
|
|
raise exc.CompileError("Couldn't compile type")
|
|
|
|
return MyType
|
|
|
|
def test_reraise_of_column_spec_issue(self):
|
|
MyType = self._illegal_type_fixture()
|
|
t1 = Table("t", MetaData(), Column("x", MyType()))
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
r"\(in table 't', column 'x'\): Couldn't compile type",
|
|
schema.CreateTable(t1).compile,
|
|
)
|
|
|
|
def test_reraise_of_column_spec_issue_unicode(self):
|
|
MyType = self._illegal_type_fixture()
|
|
t1 = Table("t", MetaData(), Column("méil", MyType()))
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
r"\(in table 't', column 'méil'\): Couldn't compile type",
|
|
schema.CreateTable(t1).compile,
|
|
)
|
|
|
|
def test_system_flag(self):
|
|
m = MetaData()
|
|
t = Table(
|
|
"t",
|
|
m,
|
|
Column("x", Integer),
|
|
Column("y", Integer, system=True),
|
|
Column("z", Integer),
|
|
)
|
|
self.assert_compile(
|
|
schema.CreateTable(t), "CREATE TABLE t (x INTEGER, z INTEGER)"
|
|
)
|
|
m2 = MetaData()
|
|
t2 = t.to_metadata(m2)
|
|
self.assert_compile(
|
|
schema.CreateTable(t2), "CREATE TABLE t (x INTEGER, z INTEGER)"
|
|
)
|
|
|
|
def test_composite_pk_constraint_autoinc_first_implicit(self):
|
|
m = MetaData()
|
|
t = Table(
|
|
"t",
|
|
m,
|
|
Column("a", Integer, primary_key=True),
|
|
Column("b", Integer, primary_key=True, autoincrement=True),
|
|
)
|
|
self.assert_compile(
|
|
schema.CreateTable(t),
|
|
"CREATE TABLE t ("
|
|
"a INTEGER NOT NULL, "
|
|
"b INTEGER NOT NULL, "
|
|
"PRIMARY KEY (b, a))",
|
|
)
|
|
|
|
def test_composite_pk_constraint_maintains_order_explicit(self):
|
|
m = MetaData()
|
|
t = Table(
|
|
"t",
|
|
m,
|
|
Column("a", Integer),
|
|
Column("b", Integer, autoincrement=True),
|
|
schema.PrimaryKeyConstraint("a", "b"),
|
|
)
|
|
self.assert_compile(
|
|
schema.CreateTable(t),
|
|
"CREATE TABLE t ("
|
|
"a INTEGER NOT NULL, "
|
|
"b INTEGER NOT NULL, "
|
|
"PRIMARY KEY (a, b))",
|
|
)
|
|
|
|
def test_create_table_exists(self):
|
|
m = MetaData()
|
|
t1 = Table("t1", m, Column("q", Integer))
|
|
self.assert_compile(
|
|
schema.CreateTable(t1, if_not_exists=True),
|
|
"CREATE TABLE IF NOT EXISTS t1 (q INTEGER)",
|
|
)
|
|
|
|
def test_drop_table_exists(self):
|
|
m = MetaData()
|
|
t1 = Table("t1", m, Column("q", Integer))
|
|
self.assert_compile(
|
|
schema.DropTable(t1, if_exists=True),
|
|
"DROP TABLE IF EXISTS t1",
|
|
)
|
|
|
|
def test_create_index_exists(self):
|
|
m = MetaData()
|
|
t1 = Table("t1", m, Column("q", Integer))
|
|
idx = Index("my_idx", t1.c.q)
|
|
self.assert_compile(
|
|
schema.CreateIndex(idx, if_not_exists=True),
|
|
"CREATE INDEX IF NOT EXISTS my_idx ON t1 (q)",
|
|
)
|
|
|
|
def test_drop_index_exists(self):
|
|
m = MetaData()
|
|
t1 = Table("t1", m, Column("q", Integer))
|
|
idx = Index("my_idx", t1.c.q)
|
|
self.assert_compile(
|
|
schema.DropIndex(idx, if_exists=True),
|
|
"DROP INDEX IF EXISTS my_idx",
|
|
)
|
|
|
|
def test_create_table_suffix(self):
|
|
class MyDialect(default.DefaultDialect):
|
|
class MyCompiler(compiler.DDLCompiler):
|
|
def create_table_suffix(self, table):
|
|
return "SOME SUFFIX"
|
|
|
|
ddl_compiler = MyCompiler
|
|
|
|
m = MetaData()
|
|
t1 = Table("t1", m, Column("q", Integer))
|
|
self.assert_compile(
|
|
schema.CreateTable(t1),
|
|
"CREATE TABLE t1 SOME SUFFIX (q INTEGER)",
|
|
dialect=MyDialect(),
|
|
)
|
|
|
|
def test_table_no_cols(self):
|
|
m = MetaData()
|
|
t1 = Table("t1", m)
|
|
self.assert_compile(schema.CreateTable(t1), "CREATE TABLE t1 ()")
|
|
|
|
def test_table_no_cols_w_constraint(self):
|
|
m = MetaData()
|
|
t1 = Table("t1", m, CheckConstraint("a = 1"))
|
|
self.assert_compile(
|
|
schema.CreateTable(t1), "CREATE TABLE t1 (CHECK (a = 1))"
|
|
)
|
|
|
|
def test_table_one_col_w_constraint(self):
|
|
m = MetaData()
|
|
t1 = Table("t1", m, Column("q", Integer), CheckConstraint("a = 1"))
|
|
self.assert_compile(
|
|
schema.CreateTable(t1),
|
|
"CREATE TABLE t1 (q INTEGER, CHECK (a = 1))",
|
|
)
|
|
|
|
def test_schema_translate_map_table(self):
|
|
m = MetaData()
|
|
t1 = Table("t1", m, Column("q", Integer))
|
|
t2 = Table("t2", m, Column("q", Integer), schema="foo")
|
|
t3 = Table("t3", m, Column("q", Integer), schema="bar")
|
|
|
|
schema_translate_map = {None: "z", "bar": None, "foo": "bat"}
|
|
|
|
self.assert_compile(
|
|
schema.CreateTable(t1),
|
|
"CREATE TABLE __[SCHEMA__none].t1 (q INTEGER)",
|
|
schema_translate_map=schema_translate_map,
|
|
)
|
|
self.assert_compile(
|
|
schema.CreateTable(t1),
|
|
"CREATE TABLE z.t1 (q INTEGER)",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
self.assert_compile(
|
|
schema.CreateTable(t2),
|
|
"CREATE TABLE __[SCHEMA_foo].t2 (q INTEGER)",
|
|
schema_translate_map=schema_translate_map,
|
|
)
|
|
self.assert_compile(
|
|
schema.CreateTable(t2),
|
|
"CREATE TABLE bat.t2 (q INTEGER)",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
self.assert_compile(
|
|
schema.CreateTable(t3),
|
|
"CREATE TABLE __[SCHEMA_bar].t3 (q INTEGER)",
|
|
schema_translate_map=schema_translate_map,
|
|
)
|
|
self.assert_compile(
|
|
schema.CreateTable(t3),
|
|
"CREATE TABLE main.t3 (q INTEGER)",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
default_schema_name="main",
|
|
)
|
|
|
|
def test_schema_translate_map_special_chars(self):
|
|
m = MetaData()
|
|
t1 = Table("t1", m, Column("q", Integer))
|
|
t2 = Table("t2", m, Column("q", Integer), schema="foo % ^ #")
|
|
t3 = Table("t3", m, Column("q", Integer), schema="bar {}")
|
|
|
|
schema_translate_map = {None: "z", "bar {}": None, "foo % ^ #": "bat"}
|
|
|
|
self.assert_compile(
|
|
schema.CreateTable(t1),
|
|
"CREATE TABLE __[SCHEMA__none].t1 (q INTEGER)",
|
|
schema_translate_map=schema_translate_map,
|
|
)
|
|
self.assert_compile(
|
|
schema.CreateTable(t1),
|
|
"CREATE TABLE z.t1 (q INTEGER)",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
self.assert_compile(
|
|
schema.CreateTable(t2),
|
|
"CREATE TABLE __[SCHEMA_foo % ^ #].t2 (q INTEGER)",
|
|
schema_translate_map=schema_translate_map,
|
|
)
|
|
self.assert_compile(
|
|
schema.CreateTable(t2),
|
|
"CREATE TABLE bat.t2 (q INTEGER)",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
self.assert_compile(
|
|
schema.CreateTable(t3),
|
|
"CREATE TABLE __[SCHEMA_bar {}].t3 (q INTEGER)",
|
|
schema_translate_map=schema_translate_map,
|
|
)
|
|
self.assert_compile(
|
|
schema.CreateTable(t3),
|
|
"CREATE TABLE main.t3 (q INTEGER)",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
default_schema_name="main",
|
|
)
|
|
|
|
def test_schema_translate_map_no_square_brackets(self):
|
|
m = MetaData()
|
|
t2 = Table("t2", m, Column("q", Integer), schema="bar [")
|
|
t3 = Table("t3", m, Column("q", Integer), schema="foo ]")
|
|
|
|
schema_translate_map = {None: "z", "bar [": None, "foo ]": "bat"}
|
|
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
r"Square bracket characters .* not supported "
|
|
r"in schema translate name 'bar \['",
|
|
schema.CreateTable(t2).compile,
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
r"Square bracket characters .* not supported "
|
|
r"in schema translate name 'foo \]'",
|
|
schema.CreateTable(t3).compile,
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
def test_schema_translate_map_sequence(self):
|
|
s1 = schema.Sequence("s1")
|
|
s2 = schema.Sequence("s2", schema="foo")
|
|
s3 = schema.Sequence("s3", schema="bar")
|
|
|
|
schema_translate_map = {None: "z", "bar": None, "foo": "bat"}
|
|
|
|
self.assert_compile(
|
|
schema.CreateSequence(s1),
|
|
"CREATE SEQUENCE __[SCHEMA__none].s1",
|
|
schema_translate_map=schema_translate_map,
|
|
)
|
|
|
|
self.assert_compile(
|
|
s1.next_value(),
|
|
"<next sequence value: __[SCHEMA__none].s1>",
|
|
schema_translate_map=schema_translate_map,
|
|
dialect="default_enhanced",
|
|
)
|
|
|
|
self.assert_compile(
|
|
schema.CreateSequence(s2),
|
|
"CREATE SEQUENCE __[SCHEMA_foo].s2",
|
|
schema_translate_map=schema_translate_map,
|
|
)
|
|
|
|
self.assert_compile(
|
|
s2.next_value(),
|
|
"<next sequence value: __[SCHEMA_foo].s2>",
|
|
schema_translate_map=schema_translate_map,
|
|
dialect="default_enhanced",
|
|
)
|
|
|
|
self.assert_compile(
|
|
schema.CreateSequence(s3),
|
|
"CREATE SEQUENCE __[SCHEMA_bar].s3",
|
|
schema_translate_map=schema_translate_map,
|
|
)
|
|
|
|
self.assert_compile(
|
|
s3.next_value(),
|
|
"<next sequence value: __[SCHEMA_bar].s3>",
|
|
schema_translate_map=schema_translate_map,
|
|
dialect="default_enhanced",
|
|
)
|
|
|
|
def test_schema_translate_map_sequence_server_default(self):
|
|
s1 = schema.Sequence("s1")
|
|
s2 = schema.Sequence("s2", schema="foo")
|
|
s3 = schema.Sequence("s3", schema="bar")
|
|
|
|
schema_translate_map = {None: "z", "bar": None, "foo": "bat"}
|
|
|
|
m = MetaData()
|
|
|
|
t1 = Table(
|
|
"t1",
|
|
m,
|
|
Column(
|
|
"id", Integer, server_default=s1.next_value(), primary_key=True
|
|
),
|
|
)
|
|
t2 = Table(
|
|
"t2",
|
|
m,
|
|
Column(
|
|
"id", Integer, server_default=s2.next_value(), primary_key=True
|
|
),
|
|
)
|
|
t3 = Table(
|
|
"t3",
|
|
m,
|
|
Column(
|
|
"id", Integer, server_default=s3.next_value(), primary_key=True
|
|
),
|
|
)
|
|
|
|
self.assert_compile(
|
|
schema.CreateTable(t1),
|
|
"CREATE TABLE __[SCHEMA__none].t1 "
|
|
"(id INTEGER DEFAULT <next sequence value: __[SCHEMA__none].s1> "
|
|
"NOT NULL, PRIMARY KEY (id))",
|
|
schema_translate_map=schema_translate_map,
|
|
dialect="default_enhanced",
|
|
)
|
|
self.assert_compile(
|
|
schema.CreateTable(t2),
|
|
"CREATE TABLE __[SCHEMA__none].t2 "
|
|
"(id INTEGER DEFAULT <next sequence value: __[SCHEMA_foo].s2> "
|
|
"NOT NULL, PRIMARY KEY (id))",
|
|
schema_translate_map=schema_translate_map,
|
|
dialect="default_enhanced",
|
|
)
|
|
self.assert_compile(
|
|
schema.CreateTable(t3),
|
|
"CREATE TABLE __[SCHEMA__none].t3 "
|
|
"(id INTEGER DEFAULT <next sequence value: __[SCHEMA_bar].s3> "
|
|
"NOT NULL, PRIMARY KEY (id))",
|
|
schema_translate_map=schema_translate_map,
|
|
dialect="default_enhanced",
|
|
)
|
|
|
|
def test_fk_render(self):
|
|
a = Table("a", MetaData(), Column("q", Integer))
|
|
b = Table("b", MetaData(), Column("p", Integer))
|
|
|
|
self.assert_compile(
|
|
schema.AddConstraint(
|
|
schema.ForeignKeyConstraint([a.c.q], [b.c.p])
|
|
),
|
|
"ALTER TABLE a ADD FOREIGN KEY(q) REFERENCES b (p)",
|
|
)
|
|
|
|
self.assert_compile(
|
|
schema.AddConstraint(
|
|
schema.ForeignKeyConstraint(
|
|
[a.c.q], [b.c.p], onupdate="SET NULL", ondelete="CASCADE"
|
|
)
|
|
),
|
|
"ALTER TABLE a ADD FOREIGN KEY(q) REFERENCES b (p) "
|
|
"ON DELETE CASCADE ON UPDATE SET NULL",
|
|
)
|
|
|
|
self.assert_compile(
|
|
schema.AddConstraint(
|
|
schema.ForeignKeyConstraint(
|
|
[a.c.q], [b.c.p], initially="DEFERRED"
|
|
)
|
|
),
|
|
"ALTER TABLE a ADD FOREIGN KEY(q) REFERENCES b (p) "
|
|
"INITIALLY DEFERRED",
|
|
)
|
|
|
|
def test_fk_illegal_sql_phrases(self):
|
|
a = Table("a", MetaData(), Column("q", Integer))
|
|
b = Table("b", MetaData(), Column("p", Integer))
|
|
|
|
for kw in ("onupdate", "ondelete", "initially"):
|
|
for phrase in (
|
|
"NOT SQL",
|
|
"INITIALLY NOT SQL",
|
|
"FOO RESTRICT",
|
|
"CASCADE WRONG",
|
|
"SET NULL",
|
|
# test that PostgreSQL's syntax added in #11595 is not
|
|
# accepted by base compiler
|
|
"SET NULL(postgresql_db.some_column)",
|
|
):
|
|
const = schema.AddConstraint(
|
|
schema.ForeignKeyConstraint(
|
|
[a.c.q], [b.c.p], **{kw: phrase}
|
|
)
|
|
)
|
|
assert_raises_message(
|
|
exc.CompileError,
|
|
rf"Unexpected SQL phrase: '{re.escape(phrase)}'",
|
|
const.compile,
|
|
)
|
|
|
|
|
|
class SchemaTest(fixtures.TestBase, AssertsCompiledSQL):
|
|
__dialect__ = "default"
|
|
|
|
def test_select(self):
|
|
self.assert_compile(
|
|
table4.select(),
|
|
"SELECT remote_owner.remotetable.rem_id, "
|
|
"remote_owner.remotetable.datatype_id,"
|
|
" remote_owner.remotetable.value "
|
|
"FROM remote_owner.remotetable",
|
|
)
|
|
|
|
self.assert_compile(
|
|
table4.select().where(
|
|
and_(table4.c.datatype_id == 7, table4.c.value == "hi")
|
|
),
|
|
"SELECT remote_owner.remotetable.rem_id, "
|
|
"remote_owner.remotetable.datatype_id,"
|
|
" remote_owner.remotetable.value "
|
|
"FROM remote_owner.remotetable WHERE "
|
|
"remote_owner.remotetable.datatype_id = :datatype_id_1 AND"
|
|
" remote_owner.remotetable.value = :value_1",
|
|
)
|
|
|
|
s = (
|
|
table4.select()
|
|
.where(and_(table4.c.datatype_id == 7, table4.c.value == "hi"))
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
)
|
|
self.assert_compile(
|
|
s,
|
|
"SELECT remote_owner.remotetable.rem_id AS"
|
|
" remote_owner_remotetable_rem_id, "
|
|
"remote_owner.remotetable.datatype_id AS"
|
|
" remote_owner_remotetable_datatype_id, "
|
|
"remote_owner.remotetable.value "
|
|
"AS remote_owner_remotetable_value FROM "
|
|
"remote_owner.remotetable WHERE "
|
|
"remote_owner.remotetable.datatype_id = :datatype_id_1 AND "
|
|
"remote_owner.remotetable.value = :value_1",
|
|
)
|
|
|
|
# multi-part schema name
|
|
self.assert_compile(
|
|
table5.select(),
|
|
'SELECT "dbo.remote_owner".remotetable.rem_id, '
|
|
'"dbo.remote_owner".remotetable.datatype_id, '
|
|
'"dbo.remote_owner".remotetable.value '
|
|
'FROM "dbo.remote_owner".remotetable',
|
|
)
|
|
|
|
# multi-part schema name labels - convert '.' to '_'
|
|
self.assert_compile(
|
|
table5.select().set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL),
|
|
'SELECT "dbo.remote_owner".remotetable.rem_id AS'
|
|
" dbo_remote_owner_remotetable_rem_id, "
|
|
'"dbo.remote_owner".remotetable.datatype_id'
|
|
" AS dbo_remote_owner_remotetable_datatype_id,"
|
|
' "dbo.remote_owner".remotetable.value AS '
|
|
"dbo_remote_owner_remotetable_value FROM"
|
|
' "dbo.remote_owner".remotetable',
|
|
)
|
|
|
|
def test_schema_translate_select(self):
|
|
m = MetaData()
|
|
table1 = Table(
|
|
"mytable",
|
|
m,
|
|
Column("myid", Integer),
|
|
Column("name", String),
|
|
Column("description", String),
|
|
)
|
|
schema_translate_map = {"remote_owner": "foob", None: "bar"}
|
|
|
|
self.assert_compile(
|
|
table1.select().where(table1.c.name == "hi"),
|
|
"SELECT bar.mytable.myid, bar.mytable.name, "
|
|
"bar.mytable.description FROM bar.mytable "
|
|
"WHERE bar.mytable.name = :name_1",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
self.assert_compile(
|
|
table4.select().where(table4.c.value == "hi"),
|
|
"SELECT foob.remotetable.rem_id, foob.remotetable.datatype_id, "
|
|
"foob.remotetable.value FROM foob.remotetable "
|
|
"WHERE foob.remotetable.value = :value_1",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
schema_translate_map = {"remote_owner": "foob"}
|
|
self.assert_compile(
|
|
select(table1, table4).select_from(
|
|
join(table1, table4, table1.c.myid == table4.c.rem_id)
|
|
),
|
|
"SELECT mytable.myid, mytable.name, mytable.description, "
|
|
"foob.remotetable.rem_id, foob.remotetable.datatype_id, "
|
|
"foob.remotetable.value FROM mytable JOIN foob.remotetable "
|
|
"ON mytable.myid = foob.remotetable.rem_id",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
def test_schema_translate_aliases(self):
|
|
schema_translate_map = {None: "bar"}
|
|
|
|
m = MetaData()
|
|
table1 = Table(
|
|
"mytable",
|
|
m,
|
|
Column("myid", Integer),
|
|
Column("name", String),
|
|
Column("description", String),
|
|
)
|
|
table2 = Table(
|
|
"myothertable",
|
|
m,
|
|
Column("otherid", Integer),
|
|
Column("othername", String),
|
|
)
|
|
|
|
alias = table1.alias()
|
|
|
|
stmt = (
|
|
select(table2, alias)
|
|
.select_from(table2.join(alias, table2.c.otherid == alias.c.myid))
|
|
.where(alias.c.name == "foo")
|
|
)
|
|
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT __[SCHEMA__none].myothertable.otherid, "
|
|
"__[SCHEMA__none].myothertable.othername, "
|
|
"mytable_1.myid, mytable_1.name, mytable_1.description "
|
|
"FROM __[SCHEMA__none].myothertable JOIN "
|
|
"__[SCHEMA__none].mytable AS mytable_1 "
|
|
"ON __[SCHEMA__none].myothertable.otherid = mytable_1.myid "
|
|
"WHERE mytable_1.name = :name_1",
|
|
schema_translate_map=schema_translate_map,
|
|
)
|
|
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT bar.myothertable.otherid, bar.myothertable.othername, "
|
|
"mytable_1.myid, mytable_1.name, mytable_1.description "
|
|
"FROM bar.myothertable JOIN bar.mytable AS mytable_1 "
|
|
"ON bar.myothertable.otherid = mytable_1.myid "
|
|
"WHERE mytable_1.name = :name_1",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
def test_schema_translate_crud(self):
|
|
schema_translate_map = {"remote_owner": "foob", None: "bar"}
|
|
|
|
m = MetaData()
|
|
table1 = Table(
|
|
"mytable",
|
|
m,
|
|
Column("myid", Integer),
|
|
Column("name", String),
|
|
Column("description", String),
|
|
)
|
|
|
|
self.assert_compile(
|
|
table1.insert().values(description="foo"),
|
|
"INSERT INTO bar.mytable (description) VALUES (:description)",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
self.assert_compile(
|
|
table1.update()
|
|
.where(table1.c.name == "hi")
|
|
.values(description="foo"),
|
|
"UPDATE bar.mytable SET description=:description "
|
|
"WHERE bar.mytable.name = :name_1",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
self.assert_compile(
|
|
table1.delete().where(table1.c.name == "hi"),
|
|
"DELETE FROM bar.mytable WHERE bar.mytable.name = :name_1",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
self.assert_compile(
|
|
table4.insert().values(value="there"),
|
|
"INSERT INTO foob.remotetable (value) VALUES (:value)",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
self.assert_compile(
|
|
table4.update()
|
|
.where(table4.c.value == "hi")
|
|
.values(value="there"),
|
|
"UPDATE foob.remotetable SET value=:value "
|
|
"WHERE foob.remotetable.value = :value_1",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
self.assert_compile(
|
|
table4.delete().where(table4.c.value == "hi"),
|
|
"DELETE FROM foob.remotetable WHERE "
|
|
"foob.remotetable.value = :value_1",
|
|
schema_translate_map=schema_translate_map,
|
|
render_schema_translate=True,
|
|
)
|
|
|
|
@testing.combinations(
|
|
(
|
|
lambda t1, t2: select(t1, t2),
|
|
"SELECT some_table_1.id, some_table_1.q, "
|
|
"foo.some_table.id AS id_1, foo.some_table.p "
|
|
"FROM some_table AS some_table_1, foo.some_table",
|
|
),
|
|
(
|
|
lambda t1, t2: select(t1, t2).set_label_style(
|
|
LABEL_STYLE_TABLENAME_PLUS_COL
|
|
),
|
|
# the original "tablename_colname" label is preserved despite
|
|
# the alias of some_table
|
|
"SELECT some_table_1.id AS some_table_id, some_table_1.q AS "
|
|
"some_table_q, foo.some_table.id AS foo_some_table_id, "
|
|
"foo.some_table.p AS foo_some_table_p "
|
|
"FROM some_table AS some_table_1, foo.some_table",
|
|
),
|
|
(
|
|
lambda t1, t2: select(t1, t2).join_from(
|
|
t1, t2, t1.c.id == t2.c.id
|
|
),
|
|
"SELECT some_table_1.id, some_table_1.q, "
|
|
"foo.some_table.id AS id_1, foo.some_table.p "
|
|
"FROM some_table AS some_table_1 "
|
|
"JOIN foo.some_table ON some_table_1.id = foo.some_table.id",
|
|
),
|
|
(
|
|
lambda t1, t2: select(t1, t2).where(t1.c.id == t2.c.id),
|
|
"SELECT some_table_1.id, some_table_1.q, "
|
|
"foo.some_table.id AS id_1, foo.some_table.p "
|
|
"FROM some_table AS some_table_1, foo.some_table "
|
|
"WHERE some_table_1.id = foo.some_table.id",
|
|
),
|
|
(
|
|
lambda t1, t2: select(t1).where(t1.c.id == t2.c.id),
|
|
"SELECT some_table_1.id, some_table_1.q "
|
|
"FROM some_table AS some_table_1, foo.some_table "
|
|
"WHERE some_table_1.id = foo.some_table.id",
|
|
),
|
|
(
|
|
lambda t2, subq: select(t2)
|
|
.select_from(t2)
|
|
.join(subq, t2.c.id == subq.c.id),
|
|
"SELECT foo.some_table.id, foo.some_table.p "
|
|
"FROM foo.some_table JOIN "
|
|
"(SELECT some_table_1.id AS id, some_table_1.q AS q "
|
|
"FROM some_table AS some_table_1, foo.some_table "
|
|
"WHERE some_table_1.id = foo.some_table.id) AS anon_1 "
|
|
"ON foo.some_table.id = anon_1.id",
|
|
),
|
|
(
|
|
lambda t1, subq: select(t1, subq.c.id)
|
|
.select_from(t1)
|
|
.join(subq, t1.c.id == subq.c.id),
|
|
# some_table is only aliased inside the subquery. this is not
|
|
# any challenge for the compiler, just checking as this is a new
|
|
# source of aliasing.
|
|
"SELECT some_table.id, some_table.q, anon_1.id AS id_1 "
|
|
"FROM some_table "
|
|
"JOIN (SELECT some_table_1.id AS id, some_table_1.q AS q "
|
|
"FROM some_table AS some_table_1, foo.some_table "
|
|
"WHERE some_table_1.id = foo.some_table.id) AS anon_1 "
|
|
"ON some_table.id = anon_1.id",
|
|
),
|
|
(
|
|
# issue #12451
|
|
lambda t1alias, t2: select(t2, t1alias),
|
|
"SELECT foo.some_table.id, foo.some_table.p, "
|
|
"some_table_1.id AS id_1, some_table_1.q FROM foo.some_table, "
|
|
"some_table AS some_table_1",
|
|
),
|
|
(
|
|
# issue #12451
|
|
lambda t1alias, t2: select(t2).join(
|
|
t1alias, t1alias.c.q == t2.c.p
|
|
),
|
|
"SELECT foo.some_table.id, foo.some_table.p FROM foo.some_table "
|
|
"JOIN some_table AS some_table_1 "
|
|
"ON some_table_1.q = foo.some_table.p",
|
|
),
|
|
(
|
|
# issue #12451
|
|
lambda t1alias, t2: select(t1alias).join(
|
|
t2, t1alias.c.q == t2.c.p
|
|
),
|
|
"SELECT some_table_1.id, some_table_1.q "
|
|
"FROM some_table AS some_table_1 "
|
|
"JOIN foo.some_table ON some_table_1.q = foo.some_table.p",
|
|
),
|
|
(
|
|
# issue #12451
|
|
lambda t1alias, t2alias: select(t1alias, t2alias).join(
|
|
t2alias, t1alias.c.q == t2alias.c.p
|
|
),
|
|
"SELECT some_table_1.id, some_table_1.q, "
|
|
"some_table_2.id AS id_1, some_table_2.p "
|
|
"FROM some_table AS some_table_1 "
|
|
"JOIN foo.some_table AS some_table_2 "
|
|
"ON some_table_1.q = some_table_2.p",
|
|
),
|
|
)
|
|
def test_schema_non_schema_disambiguation(self, stmt, expected):
|
|
"""test #7471, and its regression #12451"""
|
|
|
|
t1 = table("some_table", column("id"), column("q"))
|
|
t2 = table("some_table", column("id"), column("p"), schema="foo")
|
|
t1alias = t1.alias()
|
|
t2alias = t2.alias()
|
|
subq = select(t1).where(t1.c.id == t2.c.id).subquery()
|
|
|
|
stmt = testing.resolve_lambda(
|
|
stmt, t1=t1, t2=t2, subq=subq, t1alias=t1alias, t2alias=t2alias
|
|
)
|
|
|
|
self.assert_compile(stmt, expected)
|
|
|
|
def test_alias(self):
|
|
a = alias(table4, "remtable")
|
|
self.assert_compile(
|
|
a.select().where(a.c.datatype_id == 7),
|
|
"SELECT remtable.rem_id, remtable.datatype_id, "
|
|
"remtable.value FROM"
|
|
" remote_owner.remotetable AS remtable "
|
|
"WHERE remtable.datatype_id = :datatype_id_1",
|
|
)
|
|
|
|
def test_update(self):
|
|
self.assert_compile(
|
|
table4.update()
|
|
.where(table4.c.value == "test")
|
|
.values({table4.c.datatype_id: 12}),
|
|
"UPDATE remote_owner.remotetable SET datatype_id=:datatype_id "
|
|
"WHERE remote_owner.remotetable.value = :value_1",
|
|
)
|
|
|
|
def test_insert(self):
|
|
self.assert_compile(
|
|
table4.insert().values((2, 5, "test")),
|
|
"INSERT INTO remote_owner.remotetable "
|
|
"(rem_id, datatype_id, value) VALUES "
|
|
"(:rem_id, :datatype_id, :value)",
|
|
)
|
|
|
|
def test_schema_lowercase_select(self):
|
|
# test that "schema" works correctly when passed to table
|
|
t1 = table("foo", column("a"), column("b"), schema="bar")
|
|
self.assert_compile(
|
|
select(t1).select_from(t1),
|
|
"SELECT bar.foo.a, bar.foo.b FROM bar.foo",
|
|
)
|
|
|
|
def test_schema_lowercase_select_alias(self):
|
|
# test alias behavior
|
|
t1 = table("foo", schema="bar")
|
|
self.assert_compile(
|
|
select("*").select_from(t1.alias("t")),
|
|
"SELECT * FROM bar.foo AS t",
|
|
)
|
|
|
|
def test_schema_lowercase_select_labels(self):
|
|
# test "schema" with extended_labels
|
|
t1 = table(
|
|
"baz",
|
|
column("id", Integer),
|
|
column("name", String),
|
|
column("meta", String),
|
|
schema="here",
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(t1)
|
|
.select_from(t1)
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL),
|
|
"SELECT here.baz.id AS here_baz_id, here.baz.name AS "
|
|
"here_baz_name, here.baz.meta AS here_baz_meta FROM here.baz",
|
|
)
|
|
|
|
def test_schema_lowercase_select_subquery(self):
|
|
# test schema plays well with subqueries
|
|
t1 = table(
|
|
"yetagain",
|
|
column("anotherid", Integer),
|
|
column("anothername", String),
|
|
schema="here",
|
|
)
|
|
s = (
|
|
text("select id, name from user")
|
|
.columns(id=Integer, name=String)
|
|
.subquery()
|
|
)
|
|
stmt = select(t1.c.anotherid).select_from(
|
|
t1.join(s, t1.c.anotherid == s.c.id)
|
|
)
|
|
compiled = stmt.compile()
|
|
eq_(
|
|
compiled._create_result_map(),
|
|
{
|
|
"anotherid": (
|
|
"anotherid",
|
|
(
|
|
t1.c.anotherid,
|
|
"anotherid",
|
|
"anotherid",
|
|
"here_yetagain_anotherid",
|
|
),
|
|
t1.c.anotherid.type,
|
|
0,
|
|
)
|
|
},
|
|
)
|
|
|
|
def test_schema_lowercase_invalid(self):
|
|
assert_raises_message(
|
|
exc.ArgumentError,
|
|
r"Unsupported argument\(s\): \['not_a_schema'\]",
|
|
table,
|
|
"foo",
|
|
not_a_schema="bar",
|
|
)
|
|
|
|
|
|
class CorrelateTest(fixtures.TestBase, AssertsCompiledSQL):
|
|
__dialect__ = "default"
|
|
|
|
def test_dont_overcorrelate(self):
|
|
self.assert_compile(
|
|
select(table1)
|
|
.select_from(table1)
|
|
.select_from(table1.select().subquery()),
|
|
"SELECT mytable.myid, mytable.name, "
|
|
"mytable.description FROM mytable, (SELECT "
|
|
"mytable.myid AS myid, mytable.name AS "
|
|
"name, mytable.description AS description "
|
|
"FROM mytable) AS anon_1",
|
|
)
|
|
|
|
def _fixture(self):
|
|
t1 = table("t1", column("a"))
|
|
t2 = table("t2", column("a"))
|
|
return t1, t2, select(t1).where(t1.c.a == t2.c.a)
|
|
|
|
def _assert_where_correlated(self, stmt):
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT t2.a FROM t2 WHERE t2.a = "
|
|
"(SELECT t1.a FROM t1 WHERE t1.a = t2.a)",
|
|
)
|
|
|
|
def _assert_where_all_correlated(self, stmt):
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT t1.a, t2.a AS a_1 FROM t1, t2 WHERE t2.a = "
|
|
"(SELECT t1.a WHERE t1.a = t2.a)",
|
|
)
|
|
|
|
# note there's no more "backwards" correlation after
|
|
# we've done #2746
|
|
# def _assert_where_backwards_correlated(self, stmt):
|
|
# self.assert_compile(
|
|
# stmt,
|
|
# "SELECT t2.a FROM t2 WHERE t2.a = "
|
|
# "(SELECT t1.a FROM t2 WHERE t1.a = t2.a)")
|
|
|
|
# def _assert_column_backwards_correlated(self, stmt):
|
|
# self.assert_compile(stmt,
|
|
# "SELECT t2.a, (SELECT t1.a FROM t2 WHERE t1.a = t2.a) "
|
|
# "AS anon_1 FROM t2")
|
|
|
|
def _assert_column_correlated(self, stmt):
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT t2.a, (SELECT t1.a FROM t1 WHERE t1.a = t2.a) "
|
|
"AS anon_1 FROM t2",
|
|
)
|
|
|
|
def _assert_column_all_correlated(self, stmt):
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT t1.a, t2.a AS a_1, "
|
|
"(SELECT t1.a WHERE t1.a = t2.a) AS anon_1 FROM t1, t2",
|
|
)
|
|
|
|
def _assert_having_correlated(self, stmt):
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT t2.a FROM t2 HAVING t2.a = "
|
|
"(SELECT t1.a FROM t1 WHERE t1.a = t2.a)",
|
|
)
|
|
|
|
def _assert_from_uncorrelated(self, stmt):
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT t2.a, anon_1.a AS a_1 FROM t2, "
|
|
"(SELECT t1.a AS a FROM t1, t2 WHERE t1.a = t2.a) AS anon_1",
|
|
)
|
|
|
|
def _assert_from_all_uncorrelated(self, stmt):
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT t1.a, t2.a AS a_1, anon_1.a AS a_2 FROM t1, t2, "
|
|
"(SELECT t1.a AS a FROM t1, t2 WHERE t1.a = t2.a) AS anon_1",
|
|
)
|
|
|
|
def _assert_where_uncorrelated(self, stmt):
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT t2.a FROM t2 WHERE t2.a = "
|
|
"(SELECT t1.a FROM t1, t2 WHERE t1.a = t2.a)",
|
|
)
|
|
|
|
def _assert_column_uncorrelated(self, stmt):
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT t2.a, (SELECT t1.a FROM t1, t2 "
|
|
"WHERE t1.a = t2.a) AS anon_1 FROM t2",
|
|
)
|
|
|
|
def _assert_having_uncorrelated(self, stmt):
|
|
self.assert_compile(
|
|
stmt,
|
|
"SELECT t2.a FROM t2 HAVING t2.a = "
|
|
"(SELECT t1.a FROM t1, t2 WHERE t1.a = t2.a)",
|
|
)
|
|
|
|
def _assert_where_single_full_correlated(self, stmt):
|
|
self.assert_compile(
|
|
stmt, "SELECT t1.a FROM t1 WHERE t1.a = (SELECT t1.a)"
|
|
)
|
|
|
|
def test_correlate_semiauto_where(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_where_correlated(
|
|
select(t2).where(t2.c.a == s1.correlate(t2).scalar_subquery())
|
|
)
|
|
|
|
def test_correlate_semiauto_column(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_column_correlated(
|
|
select(t2, s1.correlate(t2).scalar_subquery())
|
|
)
|
|
|
|
def test_correlate_semiauto_column_correlate_from_subq(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_column_correlated(
|
|
select(t2, s1.scalar_subquery().correlate(t2))
|
|
)
|
|
|
|
def test_correlate_semiauto_from(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_from_uncorrelated(select(t2, s1.correlate(t2).alias()))
|
|
|
|
def test_correlate_semiauto_having(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_having_correlated(
|
|
select(t2).having(t2.c.a == s1.correlate(t2).scalar_subquery())
|
|
)
|
|
|
|
def test_correlate_semiauto_having_from_subq(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_having_correlated(
|
|
select(t2).having(t2.c.a == s1.scalar_subquery().correlate(t2))
|
|
)
|
|
|
|
def test_correlate_except_inclusion_where(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_where_correlated(
|
|
select(t2).where(
|
|
t2.c.a == s1.correlate_except(t1).scalar_subquery()
|
|
)
|
|
)
|
|
|
|
def test_correlate_except_exclusion_where(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_where_uncorrelated(
|
|
select(t2).where(
|
|
t2.c.a == s1.correlate_except(t2).scalar_subquery()
|
|
)
|
|
)
|
|
|
|
def test_correlate_except_inclusion_column(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_column_correlated(
|
|
select(t2, s1.correlate_except(t1).scalar_subquery())
|
|
)
|
|
|
|
def test_correlate_except_exclusion_column(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_column_uncorrelated(
|
|
select(t2, s1.correlate_except(t2).scalar_subquery())
|
|
)
|
|
|
|
def test_correlate_except_inclusion_from(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_from_uncorrelated(
|
|
select(t2, s1.correlate_except(t1).alias())
|
|
)
|
|
|
|
def test_correlate_except_exclusion_from(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_from_uncorrelated(
|
|
select(t2, s1.correlate_except(t2).alias())
|
|
)
|
|
|
|
@testing.combinations(False, None)
|
|
def test_correlate_except_none(self, value):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_where_all_correlated(
|
|
select(t1, t2).where(
|
|
t2.c.a == s1.correlate_except(value).scalar_subquery()
|
|
)
|
|
)
|
|
|
|
def test_correlate_except_empty(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_where_all_correlated(
|
|
select(t1, t2).where(
|
|
t2.c.a == s1.correlate_except().scalar_subquery()
|
|
)
|
|
)
|
|
|
|
def test_correlate_except_having(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_having_correlated(
|
|
select(t2).having(
|
|
t2.c.a == s1.correlate_except(t1).scalar_subquery()
|
|
)
|
|
)
|
|
|
|
def test_correlate_auto_where(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_where_correlated(
|
|
select(t2).where(t2.c.a == s1.scalar_subquery())
|
|
)
|
|
|
|
def test_correlate_auto_column(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_column_correlated(select(t2, s1.scalar_subquery()))
|
|
|
|
def test_correlate_auto_from(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_from_uncorrelated(select(t2, s1.alias()))
|
|
|
|
def test_correlate_auto_having(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_having_correlated(
|
|
select(t2).having(t2.c.a == s1.scalar_subquery())
|
|
)
|
|
|
|
@testing.combinations(False, None)
|
|
def test_correlate_disabled_where(self, value):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_where_uncorrelated(
|
|
select(t2).where(t2.c.a == s1.correlate(value).scalar_subquery())
|
|
)
|
|
|
|
def test_correlate_disabled_column(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_column_uncorrelated(
|
|
select(t2, s1.correlate(None).scalar_subquery())
|
|
)
|
|
|
|
def test_correlate_disabled_from(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_from_uncorrelated(select(t2, s1.correlate(None).alias()))
|
|
|
|
def test_correlate_disabled_having(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_having_uncorrelated(
|
|
select(t2).having(t2.c.a == s1.correlate(None).scalar_subquery())
|
|
)
|
|
|
|
def test_correlate_all_where(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_where_all_correlated(
|
|
select(t1, t2).where(
|
|
t2.c.a == s1.correlate(t1, t2).scalar_subquery()
|
|
)
|
|
)
|
|
|
|
def test_correlate_all_column(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_column_all_correlated(
|
|
select(t1, t2, s1.correlate(t1, t2).scalar_subquery())
|
|
)
|
|
|
|
def test_correlate_all_from(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self._assert_from_all_uncorrelated(
|
|
select(t1, t2, s1.correlate(t1, t2).alias())
|
|
)
|
|
|
|
def test_correlate_where_all_unintentional(self):
|
|
t1, t2, s1 = self._fixture()
|
|
assert_raises_message(
|
|
exc.InvalidRequestError,
|
|
"returned no FROM clauses due to auto-correlation",
|
|
select(t1, t2).where(t2.c.a == s1.scalar_subquery()).compile,
|
|
)
|
|
|
|
def test_correlate_from_all_ok(self):
|
|
t1, t2, s1 = self._fixture()
|
|
self.assert_compile(
|
|
select(t1, t2, s1.subquery()),
|
|
"SELECT t1.a, t2.a AS a_1, anon_1.a AS a_2 FROM t1, t2, "
|
|
"(SELECT t1.a AS a FROM t1, t2 WHERE t1.a = t2.a) AS anon_1",
|
|
)
|
|
|
|
def test_correlate_auto_where_singlefrom(self):
|
|
t1, t2, s1 = self._fixture()
|
|
s = select(t1.c.a)
|
|
s2 = select(t1).where(t1.c.a == s.scalar_subquery())
|
|
self.assert_compile(
|
|
s2, "SELECT t1.a FROM t1 WHERE t1.a = (SELECT t1.a FROM t1)"
|
|
)
|
|
|
|
def test_correlate_semiauto_where_singlefrom(self):
|
|
t1, t2, s1 = self._fixture()
|
|
|
|
s = select(t1.c.a)
|
|
|
|
s2 = select(t1).where(t1.c.a == s.correlate(t1).scalar_subquery())
|
|
self._assert_where_single_full_correlated(s2)
|
|
|
|
def test_correlate_except_semiauto_where_singlefrom(self):
|
|
t1, t2, s1 = self._fixture()
|
|
|
|
s = select(t1.c.a)
|
|
|
|
s2 = select(t1).where(
|
|
t1.c.a == s.correlate_except(t2).scalar_subquery()
|
|
)
|
|
self._assert_where_single_full_correlated(s2)
|
|
|
|
def test_correlate_alone_noeffect(self):
|
|
# new as of #2668
|
|
t1, t2, s1 = self._fixture()
|
|
self.assert_compile(
|
|
s1.correlate(t1, t2), "SELECT t1.a FROM t1, t2 WHERE t1.a = t2.a"
|
|
)
|
|
|
|
def test_correlate_except_froms(self):
|
|
# new as of #2748
|
|
t1 = table("t1", column("a"))
|
|
t2 = table("t2", column("a"), column("b"))
|
|
s = select(t2.c.b).where(t1.c.a == t2.c.a)
|
|
s = s.correlate_except(t2).alias("s")
|
|
|
|
s2 = select(func.foo(s.c.b)).scalar_subquery()
|
|
s3 = select(t1).order_by(s2)
|
|
|
|
self.assert_compile(
|
|
s3,
|
|
"SELECT t1.a FROM t1 ORDER BY "
|
|
"(SELECT foo(s.b) AS foo_1 FROM "
|
|
"(SELECT t2.b AS b FROM t2 WHERE t1.a = t2.a) AS s)",
|
|
)
|
|
|
|
def test_multilevel_froms_correlation(self):
|
|
# new as of #2748
|
|
p = table("parent", column("id"))
|
|
c = table("child", column("id"), column("parent_id"), column("pos"))
|
|
|
|
s = (
|
|
c.select()
|
|
.where(c.c.parent_id == p.c.id)
|
|
.order_by(c.c.pos)
|
|
.limit(1)
|
|
)
|
|
s = s.correlate(p).subquery()
|
|
|
|
s = exists().select_from(s).where(s.c.id == 1)
|
|
s = select(p).where(s)
|
|
self.assert_compile(
|
|
s,
|
|
"SELECT parent.id FROM parent WHERE EXISTS (SELECT * "
|
|
"FROM (SELECT child.id AS id, child.parent_id AS parent_id, "
|
|
"child.pos AS pos FROM child WHERE child.parent_id = parent.id "
|
|
"ORDER BY child.pos LIMIT :param_1) AS anon_1 "
|
|
"WHERE anon_1.id = :id_1)",
|
|
)
|
|
|
|
def test_no_contextless_correlate_except(self):
|
|
# new as of #2748
|
|
|
|
t1 = table("t1", column("x"))
|
|
t2 = table("t2", column("y"))
|
|
t3 = table("t3", column("z"))
|
|
|
|
s = (
|
|
select(t1)
|
|
.where(t1.c.x == t2.c.y)
|
|
.where(t2.c.y == t3.c.z)
|
|
.correlate_except(t1)
|
|
)
|
|
self.assert_compile(
|
|
s, "SELECT t1.x FROM t1, t2, t3 WHERE t1.x = t2.y AND t2.y = t3.z"
|
|
)
|
|
|
|
def test_multilevel_implicit_correlation_disabled(self):
|
|
# test that implicit correlation with multilevel WHERE correlation
|
|
# behaves like 0.8.1, 0.7 (i.e. doesn't happen)
|
|
t1 = table("t1", column("x"))
|
|
t2 = table("t2", column("y"))
|
|
t3 = table("t3", column("z"))
|
|
|
|
s = select(t1.c.x).where(t1.c.x == t2.c.y)
|
|
s2 = select(t3.c.z).where(t3.c.z == s.scalar_subquery())
|
|
s3 = select(t1).where(t1.c.x == s2.scalar_subquery())
|
|
|
|
self.assert_compile(
|
|
s3,
|
|
"SELECT t1.x FROM t1 "
|
|
"WHERE t1.x = (SELECT t3.z "
|
|
"FROM t3 "
|
|
"WHERE t3.z = (SELECT t1.x "
|
|
"FROM t1, t2 "
|
|
"WHERE t1.x = t2.y))",
|
|
)
|
|
|
|
def test_from_implicit_correlation_disabled(self):
|
|
# test that implicit correlation with immediate and
|
|
# multilevel FROM clauses behaves like 0.8.1 (i.e. doesn't happen)
|
|
t1 = table("t1", column("x"))
|
|
t2 = table("t2", column("y"))
|
|
|
|
s = select(t1.c.x).where(t1.c.x == t2.c.y)
|
|
s2 = select(t2, s.subquery())
|
|
s3 = select(t1, s2.subquery())
|
|
|
|
self.assert_compile(
|
|
s3,
|
|
"SELECT t1.x, anon_1.y, anon_1.x AS x_1 FROM t1, "
|
|
"(SELECT t2.y AS y, anon_2.x AS x FROM t2, "
|
|
"(SELECT t1.x AS x FROM t1, t2 WHERE t1.x = t2.y) "
|
|
"AS anon_2) AS anon_1",
|
|
)
|
|
|
|
|
|
class CoercionTest(fixtures.TestBase, AssertsCompiledSQL):
|
|
__dialect__ = default.DefaultDialect(supports_native_boolean=True)
|
|
|
|
def _fixture(self):
|
|
m = MetaData()
|
|
return Table("foo", m, Column("id", Integer))
|
|
|
|
bool_table = table("t", column("x", Boolean))
|
|
|
|
def test_coerce_bool_where(self):
|
|
self.assert_compile(
|
|
select(self.bool_table).where(self.bool_table.c.x),
|
|
"SELECT t.x FROM t WHERE t.x",
|
|
)
|
|
|
|
def test_coerce_bool_where_non_native(self):
|
|
self.assert_compile(
|
|
select(self.bool_table).where(self.bool_table.c.x),
|
|
"SELECT t.x FROM t WHERE t.x = 1",
|
|
dialect=default.DefaultDialect(supports_native_boolean=False),
|
|
)
|
|
|
|
self.assert_compile(
|
|
select(self.bool_table).where(~self.bool_table.c.x),
|
|
"SELECT t.x FROM t WHERE t.x = 0",
|
|
dialect=default.DefaultDialect(supports_native_boolean=False),
|
|
)
|
|
|
|
def test_val_and_false(self):
|
|
t = self._fixture()
|
|
self.assert_compile(and_(t.c.id == 1, False), "false")
|
|
|
|
def test_val_and_true_coerced(self):
|
|
t = self._fixture()
|
|
self.assert_compile(and_(t.c.id == 1, True), "foo.id = :id_1")
|
|
|
|
def test_val_is_null_coerced(self):
|
|
t = self._fixture()
|
|
self.assert_compile(and_(t.c.id == None), "foo.id IS NULL") # noqa
|
|
|
|
def test_val_and_None(self):
|
|
t = self._fixture()
|
|
self.assert_compile(and_(t.c.id == 1, None), "foo.id = :id_1 AND NULL")
|
|
|
|
def test_None_and_val(self):
|
|
t = self._fixture()
|
|
self.assert_compile(and_(None, t.c.id == 1), "NULL AND foo.id = :id_1")
|
|
|
|
def test_None_and_nothing(self):
|
|
# current convention is None in and_()
|
|
# returns None May want
|
|
# to revise this at some point.
|
|
self.assert_compile(and_(None), "NULL")
|
|
|
|
def test_val_and_null(self):
|
|
t = self._fixture()
|
|
self.assert_compile(
|
|
and_(t.c.id == 1, null()), "foo.id = :id_1 AND NULL"
|
|
)
|
|
|
|
|
|
class ResultMapTest(fixtures.TestBase):
|
|
"""test the behavior of the 'entry stack' and the determination
|
|
when the result_map needs to be populated.
|
|
|
|
"""
|
|
|
|
def test_compound_populates(self):
|
|
t = Table("t", MetaData(), Column("a", Integer), Column("b", Integer))
|
|
stmt = select(t).union(select(t))
|
|
comp = stmt.compile()
|
|
eq_(
|
|
comp._create_result_map(),
|
|
{
|
|
"a": ("a", (t.c.a, "a", "a", "t_a"), t.c.a.type, 0),
|
|
"b": ("b", (t.c.b, "b", "b", "t_b"), t.c.b.type, 1),
|
|
},
|
|
)
|
|
|
|
def test_compound_not_toplevel_doesnt_populate(self):
|
|
t = Table("t", MetaData(), Column("a", Integer), Column("b", Integer))
|
|
subq = select(t).union(select(t)).subquery()
|
|
stmt = select(t.c.a).select_from(t.join(subq, t.c.a == subq.c.a))
|
|
comp = stmt.compile()
|
|
eq_(
|
|
comp._create_result_map(),
|
|
{"a": ("a", (t.c.a, "a", "a", "t_a"), t.c.a.type, 0)},
|
|
)
|
|
|
|
def test_compound_only_top_populates(self):
|
|
t = Table("t", MetaData(), Column("a", Integer), Column("b", Integer))
|
|
stmt = select(t.c.a).union(select(t.c.b))
|
|
comp = stmt.compile()
|
|
eq_(
|
|
comp._create_result_map(),
|
|
{"a": ("a", (t.c.a, "a", "a", "t_a"), t.c.a.type, 0)},
|
|
)
|
|
|
|
def test_label_plus_element(self):
|
|
t = Table("t", MetaData(), Column("a", Integer))
|
|
l1 = t.c.a.label("bar")
|
|
tc = type_coerce(t.c.a + "str", String)
|
|
stmt = select(t.c.a, l1, tc)
|
|
comp = stmt.compile()
|
|
tc_anon_label = comp._create_result_map()["anon_1"][1][0]
|
|
eq_(
|
|
comp._create_result_map(),
|
|
{
|
|
"a": ("a", (t.c.a, "a", "a", "t_a"), t.c.a.type, 0),
|
|
"bar": ("bar", (l1, "bar"), l1.type, 1),
|
|
"anon_1": (
|
|
tc._anon_name_label,
|
|
(tc_anon_label, "anon_1", tc, "_no_label"),
|
|
tc.type,
|
|
2,
|
|
),
|
|
},
|
|
)
|
|
|
|
def test_label_conflict_union(self):
|
|
t1 = Table(
|
|
"t1", MetaData(), Column("a", Integer), Column("b", Integer)
|
|
)
|
|
t2 = Table("t2", MetaData(), Column("t1_a", Integer))
|
|
union = select(t2).union(select(t2)).alias()
|
|
|
|
t1_alias = t1.alias()
|
|
stmt = (
|
|
select(t1, t1_alias)
|
|
.select_from(t1.join(union, t1.c.a == union.c.t1_a))
|
|
.set_label_style(LABEL_STYLE_TABLENAME_PLUS_COL)
|
|
)
|
|
comp = stmt.compile()
|
|
eq_(
|
|
set(comp._create_result_map()),
|
|
{"t1_1_b", "t1_1_a", "t1_a", "t1_b"},
|
|
)
|
|
is_(comp._create_result_map()["t1_a"][1][2], t1.c.a)
|
|
|
|
def test_insert_with_select_values(self):
|
|
astring = Column("a", String)
|
|
aint = Column("a", Integer)
|
|
m = MetaData()
|
|
Table("t1", m, astring)
|
|
t2 = Table("t2", m, aint)
|
|
|
|
stmt = (
|
|
t2.insert()
|
|
.values(a=select(astring).scalar_subquery())
|
|
.returning(aint)
|
|
)
|
|
comp = stmt.compile(dialect=postgresql.dialect())
|
|
eq_(
|
|
comp._create_result_map(),
|
|
{"a": ("a", (aint, "a", "a", "t2_a"), aint.type, 0)},
|
|
)
|
|
|
|
def test_insert_from_select(self):
|
|
astring = Column("a", String)
|
|
aint = Column("a", Integer)
|
|
m = MetaData()
|
|
Table("t1", m, astring)
|
|
t2 = Table("t2", m, aint)
|
|
|
|
stmt = t2.insert().from_select(["a"], select(astring)).returning(aint)
|
|
comp = stmt.compile(dialect=postgresql.dialect())
|
|
eq_(
|
|
comp._create_result_map(),
|
|
{"a": ("a", (aint, "a", "a", "t2_a"), aint.type, 0)},
|
|
)
|
|
|
|
def test_nested_api(self):
|
|
from sqlalchemy.engine.cursor import CursorResultMetaData
|
|
|
|
stmt2 = select(table2).subquery()
|
|
|
|
stmt1 = select(table1).select_from(stmt2)
|
|
|
|
contexts = {}
|
|
|
|
int_ = Integer()
|
|
|
|
class MyCompiler(compiler.SQLCompiler):
|
|
def visit_select(self, stmt, *arg, **kw):
|
|
if stmt is stmt2.element:
|
|
with self._nested_result() as nested:
|
|
contexts[stmt2.element] = nested
|
|
text = super().visit_select(
|
|
stmt2.element,
|
|
)
|
|
self._add_to_result_map("k1", "k1", (1, 2, 3), int_)
|
|
else:
|
|
text = super().visit_select(stmt, *arg, **kw)
|
|
self._add_to_result_map("k2", "k2", (3, 4, 5), int_)
|
|
return text
|
|
|
|
comp = MyCompiler(default.DefaultDialect(), stmt1)
|
|
eq_(
|
|
CursorResultMetaData._create_description_match_map(
|
|
contexts[stmt2.element][0]
|
|
),
|
|
{
|
|
"otherid": (
|
|
"otherid",
|
|
(
|
|
table2.c.otherid,
|
|
"otherid",
|
|
"otherid",
|
|
"myothertable_otherid",
|
|
),
|
|
table2.c.otherid.type,
|
|
0,
|
|
),
|
|
"othername": (
|
|
"othername",
|
|
(
|
|
table2.c.othername,
|
|
"othername",
|
|
"othername",
|
|
"myothertable_othername",
|
|
),
|
|
table2.c.othername.type,
|
|
1,
|
|
),
|
|
"k1": ("k1", (1, 2, 3), int_, 2),
|
|
},
|
|
)
|
|
eq_(
|
|
comp._create_result_map(),
|
|
{
|
|
"myid": (
|
|
"myid",
|
|
(table1.c.myid, "myid", "myid", "mytable_myid"),
|
|
table1.c.myid.type,
|
|
0,
|
|
),
|
|
"k2": ("k2", (3, 4, 5), int_, 3),
|
|
"name": (
|
|
"name",
|
|
(table1.c.name, "name", "name", "mytable_name"),
|
|
table1.c.name.type,
|
|
1,
|
|
),
|
|
"description": (
|
|
"description",
|
|
(
|
|
table1.c.description,
|
|
"description",
|
|
"description",
|
|
"mytable_description",
|
|
),
|
|
table1.c.description.type,
|
|
2,
|
|
),
|
|
},
|
|
)
|
|
|
|
def test_select_wraps_for_translate_ambiguity(self):
|
|
# test for issue #3657
|
|
t = table("a", column("x"), column("y"), column("z"))
|
|
|
|
l1, l2, l3 = t.c.z.label("a"), t.c.x.label("b"), t.c.x.label("c")
|
|
orig = [t.c.x, t.c.y, l1, l2, l3]
|
|
stmt = select(*orig)
|
|
wrapped = stmt._generate()
|
|
wrapped = wrapped.add_columns(
|
|
func.ROW_NUMBER().over(order_by=t.c.z)
|
|
).alias()
|
|
|
|
wrapped_again = select(*[c for c in wrapped.c])
|
|
|
|
dialect = default.DefaultDialect()
|
|
|
|
with mock.patch.object(
|
|
dialect.statement_compiler,
|
|
"translate_select_structure",
|
|
lambda self, to_translate, **kw: (
|
|
wrapped_again if to_translate is stmt else to_translate
|
|
),
|
|
):
|
|
compiled = stmt.compile(dialect=dialect)
|
|
|
|
proxied = [obj[0] for (k, n, obj, type_) in compiled._result_columns]
|
|
for orig_obj, proxied_obj in zip(orig, proxied):
|
|
is_(orig_obj, proxied_obj)
|
|
|
|
def test_select_wraps_for_translate_ambiguity_dupe_cols(self):
|
|
# test for issue #3657
|
|
t = table("a", column("x"), column("y"), column("z"))
|
|
|
|
l1, l2, l3 = t.c.z.label("a"), t.c.x.label("b"), t.c.x.label("c")
|
|
|
|
orig = [t.c.x, t.c.y, l1, t.c.y, l2, t.c.x, l3]
|
|
|
|
# create the statement with some duplicate columns. right now
|
|
# the behavior is that these redundant columns are deduped.
|
|
stmt = select(t.c.x, t.c.y, l1, t.c.y, l2, t.c.x, l3).set_label_style(
|
|
LABEL_STYLE_NONE
|
|
)
|
|
|
|
# so the statement has 7 inner columns...
|
|
eq_(len(list(stmt.selected_columns)), 7)
|
|
|
|
# 7 are exposed as of 1.4, no more deduping
|
|
eq_(len(stmt.subquery().c), 7)
|
|
|
|
# will render 7 as well
|
|
eq_(
|
|
len(
|
|
stmt._compile_state_factory(
|
|
stmt, stmt.compile()
|
|
).columns_plus_names
|
|
),
|
|
7,
|
|
)
|
|
|
|
wrapped = stmt._generate()
|
|
wrapped = wrapped.add_columns(
|
|
func.ROW_NUMBER().over(order_by=t.c.z)
|
|
).alias()
|
|
|
|
# so when we wrap here we're going to have only 5 columns
|
|
wrapped_again = select(*[c for c in wrapped.c]).set_label_style(
|
|
LABEL_STYLE_NONE
|
|
)
|
|
|
|
# so the compiler logic that matches up the "wrapper" to the
|
|
# "select_wraps_for" can't use inner_columns to match because
|
|
# these collections are not the same
|
|
|
|
dialect = default.DefaultDialect()
|
|
|
|
with mock.patch.object(
|
|
dialect.statement_compiler,
|
|
"translate_select_structure",
|
|
lambda self, to_translate, **kw: (
|
|
wrapped_again if to_translate is stmt else to_translate
|
|
),
|
|
):
|
|
compiled = stmt.compile(dialect=dialect)
|
|
|
|
proxied = [obj[0] for (k, n, obj, type_) in compiled._result_columns]
|
|
for orig_obj, proxied_obj in zip(orig, proxied):
|
|
is_(orig_obj, proxied_obj)
|
|
|
|
|
|
class OmitFromStatementsTest(fixtures.TestBase, AssertsCompiledSQL):
|
|
"""test the _omit_from_statements parameter.
|
|
|
|
this somewhat awkward parameter was added to suit the case of
|
|
"insert_sentinel" columns that would try very hard not to be noticed
|
|
when not needed, by being omitted from any SQL statement that does not
|
|
refer to them explicitly. If they are referred to explicitly or
|
|
are in a context where their client side default has to be fired off,
|
|
then they are present.
|
|
|
|
If marked public, the feature could be used as a general "I don't want to
|
|
see this column unless I asked it to" use case.
|
|
|
|
"""
|
|
|
|
__dialect__ = "default_enhanced"
|
|
|
|
@testing.fixture
|
|
def t1(self):
|
|
m1 = MetaData()
|
|
|
|
t1 = Table(
|
|
"t1",
|
|
m1,
|
|
Column("id", Integer, primary_key=True),
|
|
Column("a", Integer),
|
|
Column(
|
|
"b", Integer, _omit_from_statements=True, insert_sentinel=True
|
|
),
|
|
Column("c", Integer),
|
|
Column("d", Integer, _omit_from_statements=True),
|
|
Column("e", Integer),
|
|
)
|
|
return t1
|
|
|
|
@testing.fixture
|
|
def t2(self):
|
|
m1 = MetaData()
|
|
|
|
t2 = Table(
|
|
"t2",
|
|
m1,
|
|
Column("id", Integer, primary_key=True),
|
|
Column("a", Integer),
|
|
Column(
|
|
"b",
|
|
Integer,
|
|
_omit_from_statements=True,
|
|
insert_sentinel=True,
|
|
default="10",
|
|
onupdate="20",
|
|
),
|
|
Column("c", Integer, default="14", onupdate="19"),
|
|
Column(
|
|
"d",
|
|
Integer,
|
|
_omit_from_statements=True,
|
|
default="5",
|
|
onupdate="15",
|
|
),
|
|
Column("e", Integer),
|
|
)
|
|
return t2
|
|
|
|
@testing.fixture
|
|
def t3(self):
|
|
m1 = MetaData()
|
|
|
|
t3 = Table(
|
|
"t3",
|
|
m1,
|
|
Column("id", Integer, primary_key=True),
|
|
Column("a", Integer),
|
|
insert_sentinel("b"),
|
|
Column("c", Integer, default="14", onupdate="19"),
|
|
)
|
|
return t3
|
|
|
|
def test_select_omitted(self, t1):
|
|
self.assert_compile(
|
|
select(t1), "SELECT t1.id, t1.a, t1.c, t1.e FROM t1"
|
|
)
|
|
|
|
def test_select_from_subquery_includes_hidden(self, t1):
|
|
s1 = select(t1.c.a, t1.c.b, t1.c.c, t1.c.d, t1.c.e).subquery()
|
|
eq_(s1.c.keys(), ["a", "b", "c", "d", "e"])
|
|
|
|
self.assert_compile(
|
|
select(s1),
|
|
"SELECT anon_1.a, anon_1.b, anon_1.c, anon_1.d, anon_1.e "
|
|
"FROM (SELECT t1.a AS a, t1.b AS b, t1.c AS c, t1.d AS d, "
|
|
"t1.e AS e FROM t1) AS anon_1",
|
|
)
|
|
|
|
def test_select_from_subquery_omitted(self, t1):
|
|
s1 = select(t1).subquery()
|
|
|
|
eq_(s1.c.keys(), ["id", "a", "c", "e"])
|
|
self.assert_compile(
|
|
select(s1),
|
|
"SELECT anon_1.id, anon_1.a, anon_1.c, anon_1.e FROM "
|
|
"(SELECT t1.id AS id, t1.a AS a, t1.c AS c, t1.e AS e FROM t1) "
|
|
"AS anon_1",
|
|
)
|
|
|
|
def test_insert_omitted(self, t1):
|
|
self.assert_compile(
|
|
insert(t1), "INSERT INTO t1 (id, a, c, e) VALUES (:id, :a, :c, :e)"
|
|
)
|
|
|
|
def test_insert_from_select_omitted(self, t1):
|
|
self.assert_compile(
|
|
insert(t1).from_select(["a", "c", "e"], select(t1)),
|
|
"INSERT INTO t1 (a, c, e) SELECT t1.id, t1.a, t1.c, t1.e FROM t1",
|
|
)
|
|
|
|
def test_insert_from_select_included(self, t1):
|
|
self.assert_compile(
|
|
insert(t1).from_select(["a", "b", "c", "d", "e"], select(t1)),
|
|
"INSERT INTO t1 (a, b, c, d, e) SELECT t1.id, t1.a, t1.c, t1.e "
|
|
"FROM t1",
|
|
)
|
|
|
|
def test_insert_from_select_defaults_included(self, t2):
|
|
self.assert_compile(
|
|
insert(t2).from_select(["a", "c", "e"], select(t2)),
|
|
"INSERT INTO t2 (a, c, e, b, d) SELECT t2.id, t2.a, t2.c, t2.e, "
|
|
":b AS anon_1, :d AS anon_2 FROM t2",
|
|
# TODO: do we have a test in test_defaults for this, that the
|
|
# default values get set up as expected?
|
|
)
|
|
|
|
def test_insert_from_select_sentinel_defaults_omitted(self, t3):
|
|
self.assert_compile(
|
|
# a pure SentinelDefault not included here, so there is no 'b'
|
|
insert(t3).from_select(["a", "c"], select(t3)),
|
|
"INSERT INTO t3 (a, c) SELECT t3.id, t3.a, t3.c FROM t3",
|
|
)
|
|
|
|
def test_insert_omitted_return_col_nonspecified(self, t1):
|
|
self.assert_compile(
|
|
insert(t1).returning(t1),
|
|
"INSERT INTO t1 (id, a, c, e) VALUES (:id, :a, :c, :e) "
|
|
"RETURNING t1.id, t1.a, t1.c, t1.e",
|
|
)
|
|
|
|
def test_insert_omitted_return_col_specified(self, t1):
|
|
self.assert_compile(
|
|
insert(t1).returning(t1.c.a, t1.c.b, t1.c.c, t1.c.d, t1.c.e),
|
|
"INSERT INTO t1 (id, a, c, e) VALUES (:id, :a, :c, :e) "
|
|
"RETURNING t1.a, t1.b, t1.c, t1.d, t1.e",
|
|
)
|
|
|
|
def test_insert_omitted_no_params(self, t1):
|
|
self.assert_compile(
|
|
insert(t1), "INSERT INTO t1 () VALUES ()", params={}
|
|
)
|
|
|
|
def test_insert_omitted_no_params_defaults(self, t2):
|
|
# omit columns that nonetheless have client-side defaults
|
|
# are included
|
|
self.assert_compile(
|
|
insert(t2),
|
|
"INSERT INTO t2 (b, c, d) VALUES (:b, :c, :d)",
|
|
params={},
|
|
)
|
|
|
|
def test_insert_omitted_no_params_defaults_no_sentinel(self, t3):
|
|
# omit columns that nonetheless have client-side defaults
|
|
# are included
|
|
self.assert_compile(
|
|
insert(t3),
|
|
"INSERT INTO t3 (c) VALUES (:c)",
|
|
params={},
|
|
)
|
|
|
|
def test_insert_omitted_defaults(self, t2):
|
|
self.assert_compile(
|
|
insert(t2), "INSERT INTO t2 (id, a, c, e) VALUES (:id, :a, :c, :e)"
|
|
)
|
|
|
|
def test_update_omitted(self, t1):
|
|
self.assert_compile(
|
|
update(t1), "UPDATE t1 SET id=:id, a=:a, c=:c, e=:e"
|
|
)
|
|
|
|
def test_update_omitted_defaults(self, t2):
|
|
self.assert_compile(
|
|
update(t2), "UPDATE t2 SET id=:id, a=:a, c=:c, e=:e"
|
|
)
|
|
|
|
def test_update_omitted_no_params_defaults(self, t2):
|
|
# omit columns that nonetheless have client-side defaults
|
|
# are included
|
|
self.assert_compile(
|
|
update(t2), "UPDATE t2 SET b=:b, c=:c, d=:d", params={}
|
|
)
|
|
|
|
def test_select_include_col(self, t1):
|
|
self.assert_compile(
|
|
select(t1, t1.c.b, t1.c.d),
|
|
"SELECT t1.id, t1.a, t1.c, t1.e, t1.b, t1.d FROM t1",
|
|
)
|
|
|
|
def test_update_include_col(self, t1):
|
|
self.assert_compile(
|
|
update(t1).values(a=5, b=10, c=15, d=20, e=25),
|
|
"UPDATE t1 SET a=:a, b=:b, c=:c, d=:d, e=:e",
|
|
checkparams={"a": 5, "b": 10, "c": 15, "d": 20, "e": 25},
|
|
)
|
|
|
|
def test_insert_include_col(self, t1):
|
|
self.assert_compile(
|
|
insert(t1).values(a=5, b=10, c=15, d=20, e=25),
|
|
"INSERT INTO t1 (a, b, c, d, e) VALUES (:a, :b, :c, :d, :e)",
|
|
checkparams={"a": 5, "b": 10, "c": 15, "d": 20, "e": 25},
|
|
)
|
|
|
|
def test_insert_include_col_via_keys(self, t1):
|
|
self.assert_compile(
|
|
insert(t1),
|
|
"INSERT INTO t1 (a, b, c, d, e) VALUES (:a, :b, :c, :d, :e)",
|
|
params={"a": 5, "b": 10, "c": 15, "d": 20, "e": 25},
|
|
checkparams={"a": 5, "b": 10, "c": 15, "d": 20, "e": 25},
|
|
)
|
|
|
|
def test_select_omitted_incl_whereclause(self, t1):
|
|
self.assert_compile(
|
|
select(t1).where(t1.c.d == 5),
|
|
"SELECT t1.id, t1.a, t1.c, t1.e FROM t1 WHERE t1.d = :d_1",
|
|
checkparams={"d_1": 5},
|
|
)
|
|
|
|
def test_select_omitted_incl_order_by(self, t1):
|
|
self.assert_compile(
|
|
select(t1).order_by(t1.c.d),
|
|
"SELECT t1.id, t1.a, t1.c, t1.e FROM t1 ORDER BY t1.d",
|
|
)
|
|
|
|
|
|
class OrderByListTest(fixtures.TestBase, AssertsCompiledSQL):
|
|
__dialect__ = "default"
|
|
|
|
def test_order_by_list(self):
|
|
"""Test standalone OrderByList with various operators"""
|
|
col1 = Column("x", Integer)
|
|
col2 = Column("y", Integer)
|
|
|
|
# Test basic OrderByList creation
|
|
order_list = OrderByList([col1, col2])
|
|
self.assert_compile(
|
|
select(literal(1)).order_by(order_list),
|
|
"SELECT :param_1 AS anon_1 ORDER BY x, y",
|
|
)
|
|
|
|
# Test OrderByList with desc
|
|
order_list_desc = order_list.desc()
|
|
self.assert_compile(
|
|
select(literal(1)).order_by(order_list_desc),
|
|
"SELECT :param_1 AS anon_1 ORDER BY x DESC, y DESC",
|
|
)
|
|
|
|
# Test OrderByList with asc
|
|
order_list_asc = order_list.asc()
|
|
self.assert_compile(
|
|
select(literal(1)).order_by(order_list_asc),
|
|
"SELECT :param_1 AS anon_1 ORDER BY x ASC, y ASC",
|
|
)
|
|
|
|
# Test OrderByList with nulls_first
|
|
order_list_nf = order_list.nulls_first()
|
|
self.assert_compile(
|
|
select(literal(1)).order_by(order_list_nf),
|
|
"SELECT :param_1 AS anon_1 ORDER BY x NULLS FIRST, y NULLS FIRST",
|
|
)
|
|
|
|
# Test OrderByList with nulls_last
|
|
order_list_nl = order_list.nulls_last()
|
|
self.assert_compile(
|
|
select(literal(1)).order_by(order_list_nl),
|
|
"SELECT :param_1 AS anon_1 ORDER BY x NULLS LAST, y NULLS LAST",
|
|
)
|
|
|
|
def test_order_by_list_chained_ops(self):
|
|
"""Test chained operations on OrderByList"""
|
|
col1 = Column("x", Integer)
|
|
col2 = Column("y", Integer)
|
|
|
|
order_list = OrderByList([col1, col2])
|
|
|
|
# Test chained desc().nulls_first()
|
|
chained = order_list.desc().nulls_first()
|
|
self.assert_compile(
|
|
select(literal(1)).order_by(chained),
|
|
"SELECT :param_1 AS anon_1 ORDER BY x DESC NULLS FIRST, "
|
|
"y DESC NULLS FIRST",
|
|
)
|