@@ -92,7 +92,7 @@ def _format_airflow_moved_table_name(source_table, version, category):
92
92
@provide_session
93
93
def merge_conn (conn , session : Session = NEW_SESSION ):
94
94
"""Add new Connection."""
95
- if not session .scalar ( select ( conn .__class__ ).filter_by (conn_id = conn .conn_id ).limit ( 1 ) ):
95
+ if not session .query ( conn .__class__ ).filter_by (conn_id = conn .conn_id ).first ( ):
96
96
session .add (conn )
97
97
session .commit ()
98
98
@@ -959,9 +959,7 @@ def check_conn_id_duplicates(session: Session) -> Iterable[str]:
959
959
960
960
dups = []
961
961
try :
962
- dups = session .execute (
963
- select (Connection .conn_id ).group_by (Connection .conn_id ).having (func .count () > 1 )
964
- ).all ()
962
+ dups = session .query (Connection .conn_id ).group_by (Connection .conn_id ).having (func .count () > 1 ).all ()
965
963
except (exc .OperationalError , exc .ProgrammingError ):
966
964
# fallback if tables hasn't been created yet
967
965
session .rollback ()
@@ -986,11 +984,12 @@ def check_username_duplicates(session: Session) -> Iterable[str]:
986
984
for model in [User , RegisterUser ]:
987
985
dups = []
988
986
try :
989
- dups = session . execute (
990
- select (model .username ) # type: ignore[attr-defined]
987
+ dups = (
988
+ session . query (model .username ) # type: ignore[attr-defined]
991
989
.group_by (model .username ) # type: ignore[attr-defined]
992
990
.having (func .count () > 1 )
993
- ).all ()
991
+ .all ()
992
+ )
994
993
except (exc .OperationalError , exc .ProgrammingError ):
995
994
# fallback if tables hasn't been created yet
996
995
session .rollback ()
@@ -1059,13 +1058,13 @@ def check_task_fail_for_duplicates(session):
1059
1058
"""
1060
1059
minimal_table_obj = table (table_name , * [column (x ) for x in uniqueness ])
1061
1060
try :
1062
- subquery = session . execute (
1063
- select (minimal_table_obj , func .count ().label ("dupe_count" ))
1061
+ subquery = (
1062
+ session . query (minimal_table_obj , func .count ().label ("dupe_count" ))
1064
1063
.group_by (* [text (x ) for x in uniqueness ])
1065
1064
.having (func .count () > text ("1" ))
1066
1065
.subquery ()
1067
1066
)
1068
- dupe_count = session .scalar ( select ( func .sum (subquery .c .dupe_count )))
1067
+ dupe_count = session .query ( func .sum (subquery .c .dupe_count )). scalar ( )
1069
1068
if not dupe_count :
1070
1069
# there are no duplicates; nothing to do.
1071
1070
return
@@ -1102,7 +1101,7 @@ def check_conn_type_null(session: Session) -> Iterable[str]:
1102
1101
1103
1102
n_nulls = []
1104
1103
try :
1105
- n_nulls = session .scalars ( select ( Connection .conn_id ).where (Connection .conn_type .is_ (None ) )).all ()
1104
+ n_nulls = session .query ( Connection .conn_id ).filter (Connection .conn_type .is_ (None )).all ()
1106
1105
except (exc .OperationalError , exc .ProgrammingError , exc .InternalError ):
1107
1106
# fallback if tables hasn't been created yet
1108
1107
session .rollback ()
@@ -1144,7 +1143,7 @@ def check_run_id_null(session: Session) -> Iterable[str]:
1144
1143
dagrun_table .c .run_id .is_ (None ),
1145
1144
dagrun_table .c .execution_date .is_ (None ),
1146
1145
)
1147
- invalid_dagrun_count = session .scalar ( select ( func .count (dagrun_table .c .id )).where (invalid_dagrun_filter ))
1146
+ invalid_dagrun_count = session .query ( func .count (dagrun_table .c .id )).filter (invalid_dagrun_filter ). scalar ( )
1148
1147
if invalid_dagrun_count > 0 :
1149
1148
dagrun_dangling_table_name = _format_airflow_moved_table_name (dagrun_table .name , "2.2" , "dangling" )
1150
1149
if dagrun_dangling_table_name in inspect (session .get_bind ()).get_table_names ():
@@ -1241,7 +1240,7 @@ def _move_dangling_data_to_new_table(
1241
1240
pk_cols = source_table .primary_key .columns
1242
1241
1243
1242
delete = source_table .delete ().where (
1244
- tuple_ (* pk_cols ).in_ (session .select (* target_table .primary_key .columns ).subquery ())
1243
+ tuple_ (* pk_cols ).in_ (session .query (* target_table .primary_key .columns ).subquery ())
1245
1244
)
1246
1245
else :
1247
1246
delete = source_table .delete ().where (
@@ -1263,11 +1262,10 @@ def _dangling_against_dag_run(session, source_table, dag_run):
1263
1262
source_table .c .dag_id == dag_run .c .dag_id ,
1264
1263
source_table .c .execution_date == dag_run .c .execution_date ,
1265
1264
)
1266
-
1267
1265
return (
1268
- select (* [c .label (c .name ) for c in source_table .c ])
1266
+ session . query (* [c .label (c .name ) for c in source_table .c ])
1269
1267
.join (dag_run , source_to_dag_run_join_cond , isouter = True )
1270
- .where (dag_run .c .dag_id .is_ (None ))
1268
+ .filter (dag_run .c .dag_id .is_ (None ))
1271
1269
)
1272
1270
1273
1271
@@ -1306,10 +1304,10 @@ def _dangling_against_task_instance(session, source_table, dag_run, task_instanc
1306
1304
)
1307
1305
1308
1306
return (
1309
- select (* [c .label (c .name ) for c in source_table .c ])
1307
+ session . query (* [c .label (c .name ) for c in source_table .c ])
1310
1308
.join (dag_run , dr_join_cond , isouter = True )
1311
1309
.join (task_instance , ti_join_cond , isouter = True )
1312
- .where (or_ (task_instance .c .dag_id .is_ (None ), dag_run .c .dag_id .is_ (None )))
1310
+ .filter (or_ (task_instance .c .dag_id .is_ (None ), dag_run .c .dag_id .is_ (None )))
1313
1311
)
1314
1312
1315
1313
@@ -1333,9 +1331,9 @@ def _move_duplicate_data_to_new_table(
1333
1331
"""
1334
1332
bind = session .get_bind ()
1335
1333
dialect_name = bind .dialect .name
1336
-
1337
1334
query = (
1338
- select (* [getattr (source_table .c , x .name ).label (str (x .name )) for x in source_table .columns ])
1335
+ session .query (source_table )
1336
+ .with_entities (* [getattr (source_table .c , x .name ).label (str (x .name )) for x in source_table .columns ])
1339
1337
.select_from (source_table )
1340
1338
.join (subquery , and_ (* [getattr (source_table .c , x ) == getattr (subquery .c , x ) for x in uniqueness ]))
1341
1339
)
0 commit comments