{"diffoscope-json-version": 1, "source1": "/srv/reproducible-results/rbuild-debian/r-b-build.ODddrMmb/b1/sqlalchemy_1.4.50+ds1-1_i386.changes", "source2": "/srv/reproducible-results/rbuild-debian/r-b-build.ODddrMmb/b2/sqlalchemy_1.4.50+ds1-1_i386.changes", "unified_diff": null, "details": [{"source1": "Files", "source2": "Files", "unified_diff": "@@ -1,5 +1,5 @@\n \n- e2b00e528839e5301f19c335c12aad90 3719988 doc optional python-sqlalchemy-doc_1.4.50+ds1-1_all.deb\n+ 034143d3b673d02ed73a1ba5a1b7f1fc 3720236 doc optional python-sqlalchemy-doc_1.4.50+ds1-1_all.deb\n a35a05381daf16b69a8c4e74b15c6441 62104 debug optional python3-sqlalchemy-ext-dbgsym_1.4.50+ds1-1_i386.deb\n af1e87ed481dbbc93c449748c4af29ce 21488 python optional python3-sqlalchemy-ext_1.4.50+ds1-1_i386.deb\n 7b9b7746123a45060be2c12bd6e80ed4 1009400 python optional python3-sqlalchemy_1.4.50+ds1-1_all.deb\n"}, {"source1": "python-sqlalchemy-doc_1.4.50+ds1-1_all.deb", "source2": "python-sqlalchemy-doc_1.4.50+ds1-1_all.deb", "unified_diff": null, "details": [{"source1": "file list", "source2": "file list", "unified_diff": "@@ -1,3 +1,3 @@\n -rw-r--r-- 0 0 0 4 2024-01-05 13:47:47.000000 debian-binary\n--rw-r--r-- 0 0 0 13368 2024-01-05 13:47:47.000000 control.tar.xz\n--rw-r--r-- 0 0 0 3706428 2024-01-05 13:47:47.000000 data.tar.xz\n+-rw-r--r-- 0 0 0 13380 2024-01-05 13:47:47.000000 control.tar.xz\n+-rw-r--r-- 0 0 0 3706664 2024-01-05 13:47:47.000000 data.tar.xz\n"}, {"source1": "control.tar.xz", "source2": "control.tar.xz", "unified_diff": null, "details": [{"source1": "control.tar", "source2": "control.tar", "unified_diff": null, "details": [{"source1": "./md5sums", "source2": "./md5sums", "unified_diff": null, "details": [{"source1": "./md5sums", "source2": "./md5sums", "comments": ["Files differ"], "unified_diff": null}]}]}]}, {"source1": "data.tar.xz", "source2": "data.tar.xz", "unified_diff": null, "details": [{"source1": "data.tar", "source2": "data.tar", "unified_diff": null, "details": [{"source1": "./usr/share/doc/python-sqlalchemy-doc/html/changelog/changelog_10.html", "source2": "./usr/share/doc/python-sqlalchemy-doc/html/changelog/changelog_10.html", "unified_diff": "@@ -590,15 +590,15 @@\n
\n

1.0 Changelog\u00b6

\n
\n

1.0.19\u00b6

\n Released: August 3, 2017
\n

oracle\u00b6

\n
    \n-
  • [oracle] [performance] [bug] [py2k] \u00b6

    Fixed performance regression caused by the fix for #3937 where\n+

  • [oracle] [bug] [performance] [py2k] \u00b6

    Fixed performance regression caused by the fix for #3937 where\n cx_Oracle as of version 5.3 dropped the .UNICODE symbol from its\n namespace, which was interpreted as cx_Oracle\u2019s \u201cWITH_UNICODE\u201d mode being\n turned on unconditionally, which invokes functions on the SQLAlchemy\n side which convert all strings to unicode unconditionally and causing\n a performance impact. In fact, per cx_Oracle\u2019s author the\n \u201cWITH_UNICODE\u201d mode has been removed entirely as of 5.1, so the expensive unicode\n conversion functions are no longer necessary and are disabled if\n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -333,15 +333,15 @@\n # oracle\n # tests\n # misc\n ****** 1.0 Changelog\u00b6 ******\n ***** 1.0.19\u00b6 *****\n Released: August 3, 2017\n **** oracle\u00b6 ****\n- * [oracle] [performance] [bug] [py2k] \u00b6\n+ * [oracle] [bug] [performance] [py2k] \u00b6\n Fixed performance regression caused by the fix for #3937 where cx_Oracle\n as of version 5.3 dropped the .UNICODE symbol from its namespace, which\n was interpreted as cx_Oracle\u2019s \u201cWITH_UNICODE\u201d mode being turned on\n unconditionally, which invokes functions on the SQLAlchemy side which\n convert all strings to unicode unconditionally and causing a performance\n impact. In fact, per cx_Oracle\u2019s author the \u201cWITH_UNICODE\u201d mode has been\n removed entirely as of 5.1, so the expensive unicode conversion functions\n"}]}, {"source1": "./usr/share/doc/python-sqlalchemy-doc/html/changelog/changelog_11.html", "source2": "./usr/share/doc/python-sqlalchemy-doc/html/changelog/changelog_11.html", "unified_diff": "@@ -875,15 +875,15 @@\n

\n
\n
\n

1.1.13\u00b6

\n Released: August 3, 2017
\n

oracle\u00b6

\n
    \n-
  • [oracle] [performance] [bug] [py2k] \u00b6

    Fixed performance regression caused by the fix for #3937 where\n+

  • [oracle] [bug] [performance] [py2k] \u00b6

    Fixed performance regression caused by the fix for #3937 where\n cx_Oracle as of version 5.3 dropped the .UNICODE symbol from its\n namespace, which was interpreted as cx_Oracle\u2019s \u201cWITH_UNICODE\u201d mode being\n turned on unconditionally, which invokes functions on the SQLAlchemy\n side which convert all strings to unicode unconditionally and causing\n a performance impact. In fact, per cx_Oracle\u2019s author the\n \u201cWITH_UNICODE\u201d mode has been removed entirely as of 5.1, so the expensive unicode\n conversion functions are no longer necessary and are disabled if\n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -529,15 +529,15 @@\n left side of the range to be positive and for the right to be negative,\n e.g. (1, 3) is \u201c1 FOLLOWING AND 3 FOLLOWING\u201d.\n References: #4053\n \n ***** 1.1.13\u00b6 *****\n Released: August 3, 2017\n **** oracle\u00b6 ****\n- * [oracle] [performance] [bug] [py2k] \u00b6\n+ * [oracle] [bug] [performance] [py2k] \u00b6\n Fixed performance regression caused by the fix for #3937 where cx_Oracle\n as of version 5.3 dropped the .UNICODE symbol from its namespace, which\n was interpreted as cx_Oracle\u2019s \u201cWITH_UNICODE\u201d mode being turned on\n unconditionally, which invokes functions on the SQLAlchemy side which\n convert all strings to unicode unconditionally and causing a performance\n impact. In fact, per cx_Oracle\u2019s author the \u201cWITH_UNICODE\u201d mode has been\n removed entirely as of 5.1, so the expensive unicode conversion functions\n"}]}, {"source1": "./usr/share/doc/python-sqlalchemy-doc/html/changelog/changelog_12.html", "source2": "./usr/share/doc/python-sqlalchemy-doc/html/changelog/changelog_12.html", "unified_diff": "@@ -2979,15 +2979,15 @@\n

    \n
  • \n
\n
\n
\n

oracle\u00b6

\n
    \n-
  • [oracle] [performance] [bug] [py2k] \u00b6

    Fixed performance regression caused by the fix for #3937 where\n+

  • [oracle] [bug] [performance] [py2k] \u00b6

    Fixed performance regression caused by the fix for #3937 where\n cx_Oracle as of version 5.3 dropped the .UNICODE symbol from its\n namespace, which was interpreted as cx_Oracle\u2019s \u201cWITH_UNICODE\u201d mode being\n turned on unconditionally, which invokes functions on the SQLAlchemy\n side which convert all strings to unicode unconditionally and causing\n a performance impact. In fact, per cx_Oracle\u2019s author the\n \u201cWITH_UNICODE\u201d mode has been removed entirely as of 5.1, so the expensive unicode\n conversion functions are no longer necessary and are disabled if\n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -2003,15 +2003,15 @@\n [mssql] [bug] \u00b6\n Added a rule to SQL Server index reflection to ignore the so-called \u201cheap\u201d\n index that is implicitly present on a table that does not specify a clustered\n index.\n References: #4059\n \n **** oracle\u00b6 ****\n- * [oracle] [performance] [bug] [py2k] \u00b6\n+ * [oracle] [bug] [performance] [py2k] \u00b6\n Fixed performance regression caused by the fix for #3937 where cx_Oracle\n as of version 5.3 dropped the .UNICODE symbol from its namespace, which\n was interpreted as cx_Oracle\u2019s \u201cWITH_UNICODE\u201d mode being turned on\n unconditionally, which invokes functions on the SQLAlchemy side which\n convert all strings to unicode unconditionally and causing a performance\n impact. In fact, per cx_Oracle\u2019s author the \u201cWITH_UNICODE\u201d mode has been\n removed entirely as of 5.1, so the expensive unicode conversion functions\n"}]}, {"source1": "./usr/share/doc/python-sqlalchemy-doc/html/changelog/changelog_13.html", "source2": "./usr/share/doc/python-sqlalchemy-doc/html/changelog/changelog_13.html", "unified_diff": "@@ -1803,30 +1803,30 @@\n

    \n
  • \n
\n
\n
\n

oracle\u00b6

\n
    \n-
  • [oracle] [performance] [bug] \u00b6

    Changed the implementation of fetching CLOB and BLOB objects to use\n+

  • [oracle] [bug] \u00b6

    Some modifications to how the cx_oracle dialect sets up per-column\n+outputtype handlers for LOB and numeric datatypes to adjust for potential\n+changes coming in cx_Oracle 8.

    \n+

    References: #5246

    \n+

    \n+
  • \n+
  • [oracle] [bug] [performance] \u00b6

    Changed the implementation of fetching CLOB and BLOB objects to use\n cx_Oracle\u2019s native implementation which fetches CLOB/BLOB objects inline\n with other result columns, rather than performing a separate fetch. As\n always, this can be disabled by setting auto_convert_lobs to False.

    \n

    As part of this change, the behavior of a CLOB that was given a blank\n string on INSERT now returns None on SELECT, which is now consistent with\n that of VARCHAR on Oracle.

    \n

    References: #5314

    \n

    \n
  • \n-
  • [oracle] [bug] \u00b6

    Some modifications to how the cx_oracle dialect sets up per-column\n-outputtype handlers for LOB and numeric datatypes to adjust for potential\n-changes coming in cx_Oracle 8.

    \n-

    References: #5246

    \n-

    \n-
  • \n
\n
\n
\n

firebird\u00b6

\n
    \n
  • [firebird] [change] \u00b6

    Adjusted dialect loading for firebird:// URIs so the external\n sqlalchemy-firebird dialect will be used if it has been installed,\n@@ -2204,15 +2204,15 @@\n

    misc\u00b6

    \n
      \n
    • [usecase] [ext] \u00b6

      Added keyword arguments to the MutableList.sort() function so that a\n key function as well as the \u201creverse\u201d keyword argument can be provided.

      \n

      References: #5114

      \n

      \n
    • \n-
    • [performance] [bug] \u00b6

      Revised an internal change to the test system added as a result of\n+

    • [bug] [performance] \u00b6

      Revised an internal change to the test system added as a result of\n #5085 where a testing-related module per dialect would be loaded\n unconditionally upon making use of that dialect, pulling in SQLAlchemy\u2019s\n testing framework as well as the ORM into the module import space. This\n would only impact initial startup time and memory to a modest extent,\n however it\u2019s best that these additional modules aren\u2019t reverse-dependent on\n straight Core usage.

      \n

      References: #5180

      \n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -1214,28 +1214,28 @@\n [mssql] [bug] [reflection] \u00b6\n Fix a regression introduced by the reflection of computed column in MSSQL when\n using SQL server versions before 2012, which does not support the concat\n function.\n References: #5271\n \n **** oracle\u00b6 ****\n- * [oracle] [performance] [bug] \u00b6\n- Changed the implementation of fetching CLOB and BLOB objects to use\n- cx_Oracle\u2019s native implementation which fetches CLOB/BLOB objects inline\n- with other result columns, rather than performing a separate fetch. As\n- always, this can be disabled by setting auto_convert_lobs to False.\n- As part of this change, the behavior of a CLOB that was given a blank\n- string on INSERT now returns None on SELECT, which is now consistent with\n- that of VARCHAR on Oracle.\n- References: #5314\n-[oracle] [bug] \u00b6\n-Some modifications to how the cx_oracle dialect sets up per-column outputtype\n-handlers for LOB and numeric datatypes to adjust for potential changes coming\n-in cx_Oracle 8.\n-References: #5246\n+ * [oracle] [bug] \u00b6\n+ Some modifications to how the cx_oracle dialect sets up per-column\n+ outputtype handlers for LOB and numeric datatypes to adjust for potential\n+ changes coming in cx_Oracle 8.\n+ References: #5246\n+[oracle] [bug] [performance] \u00b6\n+Changed the implementation of fetching CLOB and BLOB objects to use cx_Oracle\u2019s\n+native implementation which fetches CLOB/BLOB objects inline with other result\n+columns, rather than performing a separate fetch. As always, this can be\n+disabled by setting auto_convert_lobs to False.\n+As part of this change, the behavior of a CLOB that was given a blank string on\n+INSERT now returns None on SELECT, which is now consistent with that of VARCHAR\n+on Oracle.\n+References: #5314\n \n **** firebird\u00b6 ****\n * [firebird] [change] \u00b6\n Adjusted dialect loading for firebird:// URIs so the external sqlalchemy-\n firebird dialect will be used if it has been installed, otherwise fall\n back to the (now deprecated) internal Firebird dialect.\n References: #5278\n@@ -1501,15 +1501,15 @@\n References: #5146\n \n **** misc\u00b6 ****\n * [usecase] [ext] \u00b6\n Added keyword arguments to the MutableList.sort() function so that a key\n function as well as the \u201creverse\u201d keyword argument can be provided.\n References: #5114\n-[performance] [bug] \u00b6\n+[bug] [performance] \u00b6\n Revised an internal change to the test system added as a result of #5085 where\n a testing-related module per dialect would be loaded unconditionally upon\n making use of that dialect, pulling in SQLAlchemy\u2019s testing framework as well\n as the ORM into the module import space. This would only impact initial startup\n time and memory to a modest extent, however it\u2019s best that these additional\n modules aren\u2019t reverse-dependent on straight Core usage.\n References: #5180\n"}]}, {"source1": "./usr/share/doc/python-sqlalchemy-doc/html/changelog/changelog_14.html", "source2": "./usr/share/doc/python-sqlalchemy-doc/html/changelog/changelog_14.html", "unified_diff": "@@ -2857,36 +2857,36 @@\n attributes and entities that are installed as part of an Insert,\n Update, or Delete construct. The\n Select.column_descriptions accessor is also now implemented for\n Core-only selectables.

      \n

      References: #7861

      \n

      \n
    • \n-
    • [orm] [performance] [bug] \u00b6

      Improvements in memory usage by the ORM, removing a significant set of\n-intermediary expression objects that are typically stored when a copy of an\n-expression object is created. These clones have been greatly reduced,\n-reducing the number of total expression objects stored in memory by\n-ORM mappings by about 30%.

      \n-

      References: #7823

      \n-

      \n-
    • \n-
    • [orm] [bug] [regression] \u00b6

      Fixed regression in \u201cdynamic\u201d loader strategy where the\n+

    • [orm] [bug] [regression] \u00b6

      Fixed regression in \u201cdynamic\u201d loader strategy where the\n Query.filter_by() method would not be given an appropriate\n entity to filter from, in the case where a \u201csecondary\u201d table were present\n in the relationship being queried and the mapping were against something\n complex such as a \u201cwith polymorphic\u201d.

      \n

      References: #7868

      \n

      \n
    • \n-
    • [orm] [bug] \u00b6

      Fixed bug where composite() attributes would not work in\n+

    • [orm] [bug] \u00b6

      Fixed bug where composite() attributes would not work in\n conjunction with the selectin_polymorphic() loader strategy for\n joined table inheritance.

      \n

      References: #7801

      \n

      \n
    • \n+
    • [orm] [bug] [performance] \u00b6

      Improvements in memory usage by the ORM, removing a significant set of\n+intermediary expression objects that are typically stored when a copy of an\n+expression object is created. These clones have been greatly reduced,\n+reducing the number of total expression objects stored in memory by\n+ORM mappings by about 30%.

      \n+

      References: #7823

      \n+

      \n+
    • \n
    • [orm] [bug] \u00b6

      Fixed issue where the selectin_polymorphic() loader option would\n not work with joined inheritance mappers that don\u2019t have a fixed\n \u201cpolymorphic_on\u201d column. Additionally added test support for a wider\n variety of usage patterns with this construct.

      \n

      References: #7799

      \n

      \n
    • \n@@ -4612,15 +4612,15 @@\n

      \n \n
    \n
\n
\n

oracle\u00b6

\n
    \n-
  • [oracle] [performance] [bug] \u00b6

    Added a CAST(VARCHAR2(128)) to the \u201ctable name\u201d, \u201cowner\u201d, and other\n+

  • [oracle] [bug] [performance] \u00b6

    Added a CAST(VARCHAR2(128)) to the \u201ctable name\u201d, \u201cowner\u201d, and other\n DDL-name parameters as used in reflection queries against Oracle system\n views such as ALL_TABLES, ALL_TAB_CONSTRAINTS, etc to better enable\n indexing to take place against these columns, as they previously would be\n implicitly handled as NVARCHAR2 due to Python\u2019s use of Unicode for strings;\n these columns are documented in all Oracle versions as being VARCHAR2 with\n lengths varying from 30 to 128 characters depending on server version.\n Additionally, test support has been enabled for Unicode-named DDL\n@@ -5334,24 +5334,15 @@\n

\n
\n
\n

1.4.18\u00b6

\n Released: June 10, 2021
\n

orm\u00b6

\n
    \n-
  • [orm] [performance] [bug] [regression] \u00b6

    Fixed regression involving how the ORM would resolve a given mapped column\n-to a result row, where under cases such as joined eager loading, a slightly\n-more expensive \u201cfallback\u201d could take place to set up this resolution due to\n-some logic that was removed since 1.3. The issue could also cause\n-deprecation warnings involving column resolution to be emitted when using a\n-1.4 style query with joined eager loading.

    \n-

    References: #6596

    \n-

    \n-
  • \n-
  • [orm] [bug] \u00b6

    Clarified the current purpose of the\n+

  • [orm] [bug] \u00b6

    Clarified the current purpose of the\n relationship.bake_queries flag, which in 1.4 is to enable\n or disable \u201clambda caching\u201d of statements within the \u201clazyload\u201d and\n \u201cselectinload\u201d loader strategies; this is separate from the more\n foundational SQL query cache that is used for most statements.\n Additionally, the lazy loader no longer uses its own cache for many-to-one\n SQL queries, which was an implementation quirk that doesn\u2019t exist for any\n other loader scenario. Finally, the \u201clru cache\u201d warning that the lazyloader\n@@ -5361,29 +5352,38 @@\n setting bake_queries=False for such a relationship will remove this\n cache from being used, there\u2019s no particular performance gain in this case\n as using no caching vs. using a cache that needs to refresh often likely\n still wins out on the caching being used side.

    \n

    References: #6072, #6487

    \n

    \n
  • \n-
  • [orm] [bug] [regression] \u00b6

    Adjusted the means by which classes such as scoped_session\n+

  • [orm] [bug] [regression] \u00b6

    Adjusted the means by which classes such as scoped_session\n and AsyncSession are generated from the base\n Session class, such that custom Session\n subclasses such as that used by Flask-SQLAlchemy don\u2019t need to implement\n positional arguments when they call into the superclass method, and can\n continue using the same argument styles as in previous releases.

    \n

    References: #6285

    \n

    \n
  • \n-
  • [orm] [bug] [regression] \u00b6

    Fixed issue where query production for joinedload against a complex left\n+

  • [orm] [bug] [regression] \u00b6

    Fixed issue where query production for joinedload against a complex left\n hand side involving joined-table inheritance could fail to produce a\n correct query, due to a clause adaption issue.

    \n

    References: #6595

    \n

    \n
  • \n+
  • [orm] [bug] [performance] [regression] \u00b6

    Fixed regression involving how the ORM would resolve a given mapped column\n+to a result row, where under cases such as joined eager loading, a slightly\n+more expensive \u201cfallback\u201d could take place to set up this resolution due to\n+some logic that was removed since 1.3. The issue could also cause\n+deprecation warnings involving column resolution to be emitted when using a\n+1.4 style query with joined eager loading.

    \n+

    References: #6596

    \n+

    \n+
  • \n
  • [orm] [bug] \u00b6

    Fixed issue in experimental \u201cselect ORM objects from INSERT/UPDATE\u201d use\n case where an error was raised if the statement were against a\n single-table-inheritance subclass.

    \n

    References: #6591

    \n

    \n
  • \n
  • [orm] [bug] \u00b6

    The warning that\u2019s emitted for relationship() when multiple\n@@ -6227,15 +6227,15 @@\n synonyms can be established linking to these constructs which work\n fully. This is a behavior that was semi-explicitly disallowed previously,\n however since it did not fail in every scenario, explicit support\n for assoc proxy and hybrids has been added.

    \n

    References: #6267

    \n

    \n
  • \n-
  • [orm] [performance] [bug] [regression] [sql] \u00b6

    Fixed a critical performance issue where the traversal of a\n+

  • [orm] [bug] [performance] [regression] [sql] \u00b6

    Fixed a critical performance issue where the traversal of a\n select() construct would traverse a repetitive product of the\n represented FROM clauses as they were each referred towards by columns in\n the columns clause; for a series of nested subqueries with lots of columns\n this could cause a large delay and significant memory growth. This\n traversal is used by a wide variety of SQL and ORM functions, including by\n the ORM Session when it\u2019s configured to have\n \u201ctable-per-bind\u201d, which while this is not a common use case, it seems to be\n@@ -9028,22 +9028,15 @@\n

    \n

    References: #4710

    \n

    \n
  • \n-
  • [engine] [performance] \u00b6

    The pool \u201cpre-ping\u201d feature has been refined to not invoke for a DBAPI\n-connection that was just opened in the same checkout operation. pre ping\n-only applies to a DBAPI connection that\u2019s been checked into the pool\n-and is being checked out again.

    \n-

    References: #4524

    \n-

    \n-
  • \n-
  • [engine] [performance] [change] [py3k] \u00b6

    Disabled the \u201cunicode returns\u201d check that runs on dialect startup when\n+

  • [engine] [change] [performance] [py3k] \u00b6

    Disabled the \u201cunicode returns\u201d check that runs on dialect startup when\n running under Python 3, which for many years has occurred in order to test\n the current DBAPI\u2019s behavior for whether or not it returns Python Unicode\n or Py2K strings for the VARCHAR and NVARCHAR datatypes. The check still\n occurs by default under Python 2, however the mechanism to test the\n behavior will be removed in SQLAlchemy 2.0 when Python 2 support is also\n removed.

    \n

    This logic was very effective when it was needed, however now that Python 3\n@@ -9054,14 +9047,21 @@\n dialect flags by setting the dialect level flag returns_unicode_strings\n to one of String.RETURNS_CONDITIONAL or\n String.RETURNS_BYTES, both of which will enable Unicode conversion\n even under Python 3.

    \n

    References: #5315

    \n

    \n
  • \n+
  • [engine] [performance] \u00b6

    The pool \u201cpre-ping\u201d feature has been refined to not invoke for a DBAPI\n+connection that was just opened in the same checkout operation. pre ping\n+only applies to a DBAPI connection that\u2019s been checked into the pool\n+and is being checked out again.

    \n+

    References: #4524

    \n+

    \n+
  • \n
  • [engine] [bug] \u00b6

    Revised the Connection.execution_options.schema_translate_map\n feature such that the processing of the SQL statement to receive a specific\n schema name occurs within the execution phase of the statement, rather than\n at the compile phase. This is to support the statement being efficiently\n cached. Previously, the current schema being rendered into the statement\n for a particular run would be considered as part of the cache key itself,\n meaning that for a run against hundreds of schemas, there would be hundreds\n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -1998,31 +1998,31 @@\n [orm] [usecase] \u00b6\n Added new attributes UpdateBase.returning_column_descriptions and\n UpdateBase.entity_description to allow for inspection of ORM attributes and\n entities that are installed as part of an Insert, Update, or Delete construct.\n The Select.column_descriptions accessor is also now implemented for Core-only\n selectables.\n References: #7861\n-[orm] [performance] [bug] \u00b6\n-Improvements in memory usage by the ORM, removing a significant set of\n-intermediary expression objects that are typically stored when a copy of an\n-expression object is created. These clones have been greatly reduced, reducing\n-the number of total expression objects stored in memory by ORM mappings by\n-about 30%.\n-References: #7823\n [orm] [bug] [regression] \u00b6\n Fixed regression in \u201cdynamic\u201d loader strategy where the Query.filter_by()\n method would not be given an appropriate entity to filter from, in the case\n where a \u201csecondary\u201d table were present in the relationship being queried and\n the mapping were against something complex such as a \u201cwith polymorphic\u201d.\n References: #7868\n [orm] [bug] \u00b6\n Fixed bug where composite() attributes would not work in conjunction with the\n selectin_polymorphic() loader strategy for joined table inheritance.\n References: #7801\n+[orm] [bug] [performance] \u00b6\n+Improvements in memory usage by the ORM, removing a significant set of\n+intermediary expression objects that are typically stored when a copy of an\n+expression object is created. These clones have been greatly reduced, reducing\n+the number of total expression objects stored in memory by ORM mappings by\n+about 30%.\n+References: #7823\n [orm] [bug] \u00b6\n Fixed issue where the selectin_polymorphic() loader option would not work with\n joined inheritance mappers that don\u2019t have a fixed \u201cpolymorphic_on\u201d column.\n Additionally added test support for a wider variety of usage patterns with this\n construct.\n References: #7799\n [orm] [bug] \u00b6\n@@ -3285,15 +3285,15 @@\n Fixed an issue where sqlalchemy.engine.reflection.has_table() returned\n True for local temporary tables that actually belonged to a different SQL\n Server session (connection). An extra check is now performed to ensure\n that the temp table detected is in fact owned by the current session.\n References: #6910\n \n **** oracle\u00b6 ****\n- * [oracle] [performance] [bug] \u00b6\n+ * [oracle] [bug] [performance] \u00b6\n Added a CAST(VARCHAR2(128)) to the \u201ctable name\u201d, \u201cowner\u201d, and other DDL-\n name parameters as used in reflection queries against Oracle system views\n such as ALL_TABLES, ALL_TAB_CONSTRAINTS, etc to better enable indexing to\n take place against these columns, as they previously would be implicitly\n handled as NVARCHAR2 due to Python\u2019s use of Unicode for strings; these\n columns are documented in all Oracle versions as being VARCHAR2 with\n lengths varying from 30 to 128 characters depending on server version.\n@@ -3822,50 +3822,51 @@\n \u201con\u201d; it\u2019s in this directive where the schema translate map would fail to be\n honored.\n References: #6658\n \n ***** 1.4.18\u00b6 *****\n Released: June 10, 2021\n **** orm\u00b6 ****\n- * [orm] [performance] [bug] [regression] \u00b6\n- Fixed regression involving how the ORM would resolve a given mapped\n- column to a result row, where under cases such as joined eager loading, a\n- slightly more expensive \u201cfallback\u201d could take place to set up this\n- resolution due to some logic that was removed since 1.3. The issue could\n- also cause deprecation warnings involving column resolution to be emitted\n- when using a 1.4 style query with joined eager loading.\n- References: #6596\n-[orm] [bug] \u00b6\n-Clarified the current purpose of the relationship.bake_queries flag, which in\n-1.4 is to enable or disable \u201clambda caching\u201d of statements within the\n-\u201clazyload\u201d and \u201cselectinload\u201d loader strategies; this is separate from the more\n-foundational SQL query cache that is used for most statements. Additionally,\n-the lazy loader no longer uses its own cache for many-to-one SQL queries, which\n-was an implementation quirk that doesn\u2019t exist for any other loader scenario.\n-Finally, the \u201clru cache\u201d warning that the lazyloader and selectinloader\n-strategies could emit when handling a wide array of class/relationship\n-combinations has been removed; based on analysis of some end-user cases, this\n-warning doesn\u2019t suggest any significant issue. While setting bake_queries=False\n-for such a relationship will remove this cache from being used, there\u2019s no\n-particular performance gain in this case as using no caching vs. using a cache\n-that needs to refresh often likely still wins out on the caching being used\n-side.\n-References: #6072, #6487\n+ * [orm] [bug] \u00b6\n+ Clarified the current purpose of the relationship.bake_queries flag,\n+ which in 1.4 is to enable or disable \u201clambda caching\u201d of statements\n+ within the \u201clazyload\u201d and \u201cselectinload\u201d loader strategies; this is\n+ separate from the more foundational SQL query cache that is used for most\n+ statements. Additionally, the lazy loader no longer uses its own cache\n+ for many-to-one SQL queries, which was an implementation quirk that\n+ doesn\u2019t exist for any other loader scenario. Finally, the \u201clru cache\u201d\n+ warning that the lazyloader and selectinloader strategies could emit when\n+ handling a wide array of class/relationship combinations has been\n+ removed; based on analysis of some end-user cases, this warning doesn\u2019t\n+ suggest any significant issue. While setting bake_queries=False for such\n+ a relationship will remove this cache from being used, there\u2019s no\n+ particular performance gain in this case as using no caching vs. using a\n+ cache that needs to refresh often likely still wins out on the caching\n+ being used side.\n+ References: #6072, #6487\n [orm] [bug] [regression] \u00b6\n Adjusted the means by which classes such as scoped_session and AsyncSession are\n generated from the base Session class, such that custom Session subclasses such\n as that used by Flask-SQLAlchemy don\u2019t need to implement positional arguments\n when they call into the superclass method, and can continue using the same\n argument styles as in previous releases.\n References: #6285\n [orm] [bug] [regression] \u00b6\n Fixed issue where query production for joinedload against a complex left hand\n side involving joined-table inheritance could fail to produce a correct query,\n due to a clause adaption issue.\n References: #6595\n+[orm] [bug] [performance] [regression] \u00b6\n+Fixed regression involving how the ORM would resolve a given mapped column to a\n+result row, where under cases such as joined eager loading, a slightly more\n+expensive \u201cfallback\u201d could take place to set up this resolution due to some\n+logic that was removed since 1.3. The issue could also cause deprecation\n+warnings involving column resolution to be emitted when using a 1.4 style query\n+with joined eager loading.\n+References: #6596\n [orm] [bug] \u00b6\n Fixed issue in experimental \u201cselect ORM objects from INSERT/UPDATE\u201d use case\n where an error was raised if the statement were against a single-table-\n inheritance subclass.\n References: #6591\n [orm] [bug] \u00b6\n The warning that\u2019s emitted for relationship() when multiple relationships would\n@@ -4472,15 +4473,15 @@\n Established support for synoynm() in conjunction with hybrid property,\n assocaitionproxy is set up completely, including that synonyms can be\n established linking to these constructs which work fully. This is a\n behavior that was semi-explicitly disallowed previously, however since it\n did not fail in every scenario, explicit support for assoc proxy and\n hybrids has been added.\n References: #6267\n-[orm] [performance] [bug] [regression] [sql] \u00b6\n+[orm] [bug] [performance] [regression] [sql] \u00b6\n Fixed a critical performance issue where the traversal of a select() construct\n would traverse a repetitive product of the represented FROM clauses as they\n were each referred towards by columns in the columns clause; for a series of\n nested subqueries with lots of columns this could cause a large delay and\n significant memory growth. This traversal is used by a wide variety of SQL and\n ORM functions, including by the ORM Session when it\u2019s configured to have\n \u201ctable-per-bind\u201d, which while this is not a common use case, it seems to be\n@@ -6529,21 +6530,15 @@\n returned by the ResultProxy is now the LegacyRow subclass, which maintains\n mapping/tuple hybrid behavior, however the base Row class now behaves more\n fully like a named tuple.\n See also\n RowProxy_is_no_longer_a_\u201cproxy\u201d;_is_now_called_Row_and_behaves_like_an_enhanced\n named_tuple\n References: #4710\n-[engine] [performance] \u00b6\n-The pool \u201cpre-ping\u201d feature has been refined to not invoke for a DBAPI\n-connection that was just opened in the same checkout operation. pre ping only\n-applies to a DBAPI connection that\u2019s been checked into the pool and is being\n-checked out again.\n-References: #4524\n-[engine] [performance] [change] [py3k] \u00b6\n+[engine] [change] [performance] [py3k] \u00b6\n Disabled the \u201cunicode returns\u201d check that runs on dialect startup when running\n under Python 3, which for many years has occurred in order to test the current\n DBAPI\u2019s behavior for whether or not it returns Python Unicode or Py2K strings\n for the VARCHAR and NVARCHAR datatypes. The check still occurs by default under\n Python 2, however the mechanism to test the behavior will be removed in\n SQLAlchemy 2.0 when Python 2 support is also removed.\n This logic was very effective when it was needed, however now that Python 3 is\n@@ -6551,14 +6546,20 @@\n datatypes. In the unlikely case that a third party DBAPI does not support this,\n the conversion logic within String is still available and the third party\n dialect may specify this in its upfront dialect flags by setting the dialect\n level flag returns_unicode_strings to one of String.RETURNS_CONDITIONAL or\n String.RETURNS_BYTES, both of which will enable Unicode conversion even under\n Python 3.\n References: #5315\n+[engine] [performance] \u00b6\n+The pool \u201cpre-ping\u201d feature has been refined to not invoke for a DBAPI\n+connection that was just opened in the same checkout operation. pre ping only\n+applies to a DBAPI connection that\u2019s been checked into the pool and is being\n+checked out again.\n+References: #4524\n [engine] [bug] \u00b6\n Revised the Connection.execution_options.schema_translate_map feature such that\n the processing of the SQL statement to receive a specific schema name occurs\n within the execution phase of the statement, rather than at the compile phase.\n This is to support the statement being efficiently cached. Previously, the\n current schema being rendered into the statement for a particular run would be\n considered as part of the cache key itself, meaning that for a run against\n"}]}, {"source1": "./usr/share/doc/python-sqlalchemy-doc/html/orm/examples.html", "source2": "./usr/share/doc/python-sqlalchemy-doc/html/orm/examples.html", "comments": ["Ordering differences only"], "unified_diff": "@@ -308,46 +308,46 @@\n

\n
\n

Associations\u00b6

\n

Examples illustrating the usage of the \u201cassociation object\u201d pattern,\n where an intermediary class mediates the relationship between two\n classes that are associated in a many-to-many pattern.

\n

Listing of files:

    \n-
  • dict_of_sets_with_default.py - An advanced association proxy example which\n-illustrates nesting of association proxies to produce multi-level Python\n-collections, in this case a dictionary with string keys and sets of integers\n-as values, which conceal the underlying mapped classes.

    \n-

  • \n
  • basic_association.py - Illustrate a many-to-many relationship between an\n \u201cOrder\u201d and a collection of \u201cItem\u201d objects, associating a purchase price\n with each via an association object called \u201cOrderItem\u201d

    \n

  • \n
  • proxied_association.py - Same example as basic_association, adding in\n usage of sqlalchemy.ext.associationproxy to make explicit references\n to OrderItem optional.

    \n

  • \n+
  • dict_of_sets_with_default.py - An advanced association proxy example which\n+illustrates nesting of association proxies to produce multi-level Python\n+collections, in this case a dictionary with string keys and sets of integers\n+as values, which conceal the underlying mapped classes.

    \n+

  • \n
\n

\n
\n
\n

Asyncio Integration\u00b6

\n

Examples illustrating the asyncio engine feature of SQLAlchemy.

\n

Listing of files:

    \n-
  • greenlet_orm.py - Illustrates use of the sqlalchemy.ext.asyncio.AsyncSession object\n-for asynchronous ORM use, including the optional run_sync() method.

    \n-

  • \n
  • async_orm.py - Illustrates use of the sqlalchemy.ext.asyncio.AsyncSession object\n for asynchronous ORM use.

    \n

  • \n-
  • basic.py - Illustrates the asyncio engine / connection interface.

    \n-

  • \n
  • gather_orm_statements.py - Illustrates how to run many statements concurrently using asyncio.gather()\n along many asyncio database connections, merging ORM results into a single\n AsyncSession.

    \n

  • \n+
  • basic.py - Illustrates the asyncio engine / connection interface.

    \n+

  • \n+
  • greenlet_orm.py - Illustrates use of the sqlalchemy.ext.asyncio.AsyncSession object\n+for asynchronous ORM use, including the optional run_sync() method.

    \n+

  • \n
\n

\n
\n
\n

Directed Graphs\u00b6

\n

An example of persistence for a directed graph structure. The\n graph is stored as a collection of edges, each referencing both a\n@@ -391,26 +391,26 @@\n

  • discriminator_on_association.py - Illustrates a mixin which provides a generic association\n using a single target table and a single association table,\n referred to by all parent tables. The association table\n contains a \u201cdiscriminator\u201d column which determines what type of\n parent object associates to each particular row in the association\n table.

    \n

  • \n-
  • table_per_association.py - Illustrates a mixin which provides a generic association\n-via a individually generated association tables for each parent class.\n-The associated objects themselves are persisted in a single table\n-shared among all parents.

    \n-

  • \n
  • generic_fk.py - Illustrates a so-called \u201cgeneric foreign key\u201d, in a similar fashion\n to that of popular frameworks such as Django, ROR, etc. This\n approach bypasses standard referential integrity\n practices, in that the \u201cforeign key\u201d column is not actually\n constrained to refer to any particular table; instead,\n in-application logic is used to determine which table is referenced.

    \n

  • \n+
  • table_per_association.py - Illustrates a mixin which provides a generic association\n+via a individually generated association tables for each parent class.\n+The associated objects themselves are persisted in a single table\n+shared among all parents.

    \n+

  • \n
  • table_per_related.py - Illustrates a generic association which persists association\n objects within individual tables, each one generated to persist\n those objects on behalf of a particular parent class.

    \n

  • \n \n

    \n
    \n@@ -500,33 +500,33 @@\n \n
    \n

    File Listing\u00b6

    \n

    Listing of files:

      \n-
    • __main__.py - Allows the examples/performance package to be run as a script.

      \n+
    • bulk_inserts.py - This series of tests illustrates different ways to INSERT a large number\n+of rows in bulk.

      \n

    • \n
    • short_selects.py - This series of tests illustrates different ways to SELECT a single\n record by primary key

      \n

    • \n-
    • single_inserts.py - In this series of tests, we\u2019re looking at a method that inserts a row\n-within a distinct transaction, and afterwards returns to essentially a\n-\u201cclosed\u201d state. This would be analogous to an API call that starts up\n-a database connection, inserts the row, commits and closes.

      \n-

    • \n-
    • bulk_inserts.py - This series of tests illustrates different ways to INSERT a large number\n-of rows in bulk.

      \n-

    • \n
    • large_resultsets.py - In this series of tests, we are looking at time to load a large number\n of very small and simple rows.

      \n

    • \n
    • bulk_updates.py - This series of tests will illustrate different ways to UPDATE a large number\n of rows in bulk (under construction! there\u2019s just one test at the moment)

      \n

    • \n+
    • single_inserts.py - In this series of tests, we\u2019re looking at a method that inserts a row\n+within a distinct transaction, and afterwards returns to essentially a\n+\u201cclosed\u201d state. This would be analogous to an API call that starts up\n+a database connection, inserts the row, commits and closes.

      \n+

    • \n+
    • __main__.py - Allows the examples/performance package to be run as a script.

      \n+

    • \n
    \n

    \n
    \n
    \n

    Running all tests with time\u00b6

    \n

    This is the default form of run:

    \n
    $ python -m examples.performance single_inserts\n@@ -668,22 +668,22 @@\n 
    \n
    \n

    Relationship Join Conditions\u00b6

    \n

    Examples of various relationship() configurations,\n which make use of the primaryjoin argument to compose special types\n of join conditions.

    \n

    Listing of files:

      \n-
    • cast.py - Illustrate a relationship() that joins two columns where those\n-columns are not of the same type, and a CAST must be used on the SQL\n-side in order to match them.

      \n-

    • \n
    • threeway.py - Illustrate a \u201cthree way join\u201d - where a primary table joins to a remote\n table via an association table, but then the primary table also needs\n to refer to some columns in the remote table directly.

      \n

    • \n+
    • cast.py - Illustrate a relationship() that joins two columns where those\n+columns are not of the same type, and a CAST must be used on the SQL\n+side in order to match them.

      \n+

    • \n
    \n

    \n
    \n
    \n

    Space Invaders\u00b6

    \n

    A Space Invaders game using SQLite as the state machine.

    \n

    Originally developed in 2012. Adapted to work in Python 3.

    \n@@ -813,19 +813,19 @@\n assert type(other) is SomeClass and other.id == self.id\n \n

    Above, if two instance of SomeClass with the same version identifier\n are updated and sent to the database for UPDATE concurrently, if the database\n isolation level allows the two UPDATE statements to proceed, one will fail\n because it no longer is against the last known version identifier.

    \n

    Listing of files:

    \n

    \n
    \n
    \n

    Versioning using Temporal Rows\u00b6

    \n

    Several examples that illustrate the technique of intercepting changes\n that would be first interpreted as an UPDATE on a row, and instead turning\n@@ -879,56 +879,56 @@\n q = (session.query(Animal).\n filter(Animal.facts.any(\n and_(AnimalFact.key == u'weasel-like',\n AnimalFact.value == True))))\n print('weasel-like animals', q.all())\n \n

    Listing of files:

    \n

    \n
    \n
    \n
    \n

    Inheritance Mapping Recipes\u00b6

    \n
    \n

    Basic Inheritance Mappings\u00b6

    \n

    Working examples of single-table, joined-table, and concrete-table\n inheritance as described in Mapping Class Inheritance Hierarchies.

    \n

    Listing of files:

      \n-
    • concrete.py - Concrete-table (table-per-class) inheritance example.

      \n+
    • joined.py - Joined-table (table-per-subclass) inheritance example.

      \n

    • \n
    • single.py - Single-table (table-per-hierarchy) inheritance example.

      \n

    • \n-
    • joined.py - Joined-table (table-per-subclass) inheritance example.

      \n+
    • concrete.py - Concrete-table (table-per-class) inheritance example.

      \n

    • \n
    \n

    \n
    \n
    \n
    \n

    Special APIs\u00b6

    \n
    \n

    Attribute Instrumentation\u00b6

    \n

    Examples illustrating modifications to SQLAlchemy\u2019s attribute management\n system.

    \n

    Listing of files:

    \n

    \n
    \n
    \n

    Horizontal Sharding\u00b6

    \n

    A basic example of using the SQLAlchemy Sharding API.\n@@ -960,20 +960,20 @@\n more plain-spoken alternative, the \u201cdistinct entity\u201d approach\n is a simple method of assigning objects to different tables (and potentially\n database nodes) in an explicit way - described on the wiki at\n EntityName.

    \n

    Listing of files:

      \n
    • separate_databases.py - Illustrates sharding using distinct SQLite databases.

      \n

    • \n-
    • separate_tables.py - Illustrates sharding using a single SQLite database, that will however\n-have multiple tables using a naming convention.

      \n-

    • \n
    • separate_schema_translates.py - Illustrates sharding using a single database with multiple schemas,\n where a different \u201cschema_translates_map\u201d can be used for each shard.

      \n

    • \n+
    • separate_tables.py - Illustrates sharding using a single SQLite database, that will however\n+have multiple tables using a naming convention.

      \n+

    • \n
    \n

    \n
    \n
    \n
    \n

    Extending the ORM\u00b6

    \n
    \n@@ -984,19 +984,19 @@\n object.

    \n

    Examples include demonstrations of the with_loader_criteria()\n option as well as the SessionEvents.do_orm_execute() hook.

    \n

    As of SQLAlchemy 1.4, the Query construct is unified\n with the Select construct, so that these two objects\n are mostly the same.

    \n

    Listing of files:

      \n-
    • filter_public.py - Illustrates a global criteria applied to entities of a particular type.

      \n-

    • \n
    • temporal_range.py - Illustrates a custom per-query criteria that will be applied\n to selected entities.

      \n

    • \n+
    • filter_public.py - Illustrates a global criteria applied to entities of a particular type.

      \n+

    • \n
    \n

    \n
    \n
    \n

    Dogpile Caching\u00b6

    \n

    Illustrates how to embed\n dogpile.cache\n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -123,37 +123,36 @@\n * adjacency_list.py\n \n **** Associations\u00b6 ****\n Examples illustrating the usage of the \u201cassociation object\u201d pattern, where an\n intermediary class mediates the relationship between two classes that are\n associated in a many-to-many pattern.\n Listing of files:\n- * dict_of_sets_with_default.py - An advanced association proxy example\n- which illustrates nesting of association proxies to produce multi-level\n- Python collections, in this case a dictionary with string keys and sets\n- of integers as values, which conceal the underlying mapped classes.\n-basic_association.py - Illustrate a many-to-many relationship between an\n-\u201cOrder\u201d and a collection of \u201cItem\u201d objects, associating a purchase price with\n-each via an association object called \u201cOrderItem\u201d\n+ * basic_association.py - Illustrate a many-to-many relationship between an\n+ \u201cOrder\u201d and a collection of \u201cItem\u201d objects, associating a purchase price\n+ with each via an association object called \u201cOrderItem\u201d\n proxied_association.py - Same example as basic_association, adding in usage of\n sqlalchemy.ext.associationproxy to make explicit references to OrderItem\n optional.\n+dict_of_sets_with_default.py - An advanced association proxy example which\n+illustrates nesting of association proxies to produce multi-level Python\n+collections, in this case a dictionary with string keys and sets of integers as\n+values, which conceal the underlying mapped classes.\n \n **** Asyncio Integration\u00b6 ****\n Examples illustrating the asyncio engine feature of SQLAlchemy.\n Listing of files:\n- * greenlet_orm.py - Illustrates use of the\n- sqlalchemy.ext.asyncio.AsyncSession object for asynchronous ORM use,\n- including the optional run_sync() method.\n-async_orm.py - Illustrates use of the sqlalchemy.ext.asyncio.AsyncSession\n-object for asynchronous ORM use.\n-basic.py - Illustrates the asyncio engine / connection interface.\n+ * async_orm.py - Illustrates use of the sqlalchemy.ext.asyncio.AsyncSession\n+ object for asynchronous ORM use.\n gather_orm_statements.py - Illustrates how to run many statements concurrently\n using asyncio.gather() along many asyncio database connections, merging ORM\n results into a single AsyncSession.\n+basic.py - Illustrates the asyncio engine / connection interface.\n+greenlet_orm.py - Illustrates use of the sqlalchemy.ext.asyncio.AsyncSession\n+object for asynchronous ORM use, including the optional run_sync() method.\n \n **** Directed Graphs\u00b6 ****\n An example of persistence for a directed graph structure. The graph is stored\n as a collection of edges, each referencing both a \u201clower\u201d and an \u201cupper\u201d node\n in a table of nodes. Basic persistence and querying for lower- and upper-\n neighbors are illustrated:\n n2 = Node(2)\n@@ -183,23 +182,23 @@\n with_SQLAlchemy.\n Listing of files:\n * discriminator_on_association.py - Illustrates a mixin which provides a\n generic association using a single target table and a single association\n table, referred to by all parent tables. The association table contains a\n \u201cdiscriminator\u201d column which determines what type of parent object\n associates to each particular row in the association table.\n-table_per_association.py - Illustrates a mixin which provides a generic\n-association via a individually generated association tables for each parent\n-class. The associated objects themselves are persisted in a single table shared\n-among all parents.\n generic_fk.py - Illustrates a so-called \u201cgeneric foreign key\u201d, in a similar\n fashion to that of popular frameworks such as Django, ROR, etc. This approach\n bypasses standard referential integrity practices, in that the \u201cforeign key\u201d\n column is not actually constrained to refer to any particular table; instead,\n in-application logic is used to determine which table is referenced.\n+table_per_association.py - Illustrates a mixin which provides a generic\n+association via a individually generated association tables for each parent\n+class. The associated objects themselves are persisted in a single table shared\n+among all parents.\n table_per_related.py - Illustrates a generic association which persists\n association objects within individual tables, each one generated to persist\n those objects on behalf of a particular parent class.\n \n **** Large Collections\u00b6 ****\n Large collection example.\n Illustrates the options to use with relationship() when the list of related\n@@ -263,29 +262,28 @@\n $ python -m examples.performance bulk_inserts \\\n --dburl mysql+mysqldb://scott:tiger@localhost/test \\\n --profile --num 1000\n See also\n How_can_I_profile_a_SQLAlchemy_powered_application?\n *** File Listing\u00b6 ***\n Listing of files:\n- * __main__.py - Allows the examples/performance package to be run as a\n- script.\n+ * bulk_inserts.py - This series of tests illustrates different ways to\n+ INSERT a large number of rows in bulk.\n short_selects.py - This series of tests illustrates different ways to SELECT a\n single record by primary key\n-single_inserts.py - In this series of tests, we\u2019re looking at a method that\n-inserts a row within a distinct transaction, and afterwards returns to\n-essentially a \u201cclosed\u201d state. This would be analogous to an API call that\n-starts up a database connection, inserts the row, commits and closes.\n-bulk_inserts.py - This series of tests illustrates different ways to INSERT a\n-large number of rows in bulk.\n large_resultsets.py - In this series of tests, we are looking at time to load a\n large number of very small and simple rows.\n bulk_updates.py - This series of tests will illustrate different ways to UPDATE\n a large number of rows in bulk (under construction! there\u2019s just one test at\n the moment)\n+single_inserts.py - In this series of tests, we\u2019re looking at a method that\n+inserts a row within a distinct transaction, and afterwards returns to\n+essentially a \u201cclosed\u201d state. This would be analogous to an API call that\n+starts up a database connection, inserts the row, commits and closes.\n+__main__.py - Allows the examples/performance package to be run as a script.\n \n *** Running all tests with time\u00b6 ***\n This is the default form of run:\n $ python -m examples.performance single_inserts\n Tests to run: test_orm_commit, test_bulk_save,\n test_bulk_insert_dictionaries, test_core,\n test_core_query_caching, test_dbapi_raw_w_connect,\n@@ -426,20 +424,20 @@\n test_subqueryload : load everything, subquery eager loading. (1000 iterations);\n total time 2.977696 sec\n \n **** Relationship Join Conditions\u00b6 ****\n Examples of various relationship() configurations, which make use of the\n primaryjoin argument to compose special types of join conditions.\n Listing of files:\n- * cast.py - Illustrate a relationship() that joins two columns where those\n- columns are not of the same type, and a CAST must be used on the SQL side\n- in order to match them.\n-threeway.py - Illustrate a \u201cthree way join\u201d - where a primary table joins to a\n-remote table via an association table, but then the primary table also needs to\n-refer to some columns in the remote table directly.\n+ * threeway.py - Illustrate a \u201cthree way join\u201d - where a primary table joins\n+ to a remote table via an association table, but then the primary table\n+ also needs to refer to some columns in the remote table directly.\n+cast.py - Illustrate a relationship() that joins two columns where those\n+columns are not of the same type, and a CAST must be used on the SQL side in\n+order to match them.\n \n **** Space Invaders\u00b6 ****\n A Space Invaders game using SQLite as the state machine.\n Originally developed in 2012. Adapted to work in Python 3.\n Runs in a textual console using ASCII art.\n [../_images/space_invaders.jpg]\n To run:\n@@ -545,17 +543,17 @@\n def __eq__(self, other):\n assert type(other) is SomeClass and other.id == self.id\n Above, if two instance of SomeClass with the same version identifier are\n updated and sent to the database for UPDATE concurrently, if the database\n isolation level allows the two UPDATE statements to proceed, one will fail\n because it no longer is against the last known version identifier.\n Listing of files:\n- * test_versioning.py - Unit tests illustrating usage of the history_meta.py\n- module functions.\n-history_meta.py - Versioned mixin class and other utilities.\n+ * history_meta.py - Versioned mixin class and other utilities.\n+test_versioning.py - Unit tests illustrating usage of the history_meta.py\n+module functions.\n \n *** Versioning using Temporal Rows\u00b6 ***\n Several examples that illustrate the technique of intercepting changes that\n would be first interpreted as an UPDATE on a row, and instead turning it into\n an INSERT of a new row, leaving the previous row intact as a historical\n version.\n Compare to the Versioning_with_a_History_Table example which writes a history\n@@ -597,40 +595,39 @@\n \n q = (session.query(Animal).\n filter(Animal.facts.any(\n and_(AnimalFact.key == u'weasel-like',\n AnimalFact.value == True))))\n print('weasel-like animals', q.all())\n Listing of files:\n- * dictlike-polymorphic.py - Mapping a polymorphic-valued vertical table as\n- a dictionary.\n-dictlike.py - Mapping a vertical table as a dictionary.\n+ * dictlike.py - Mapping a vertical table as a dictionary.\n+dictlike-polymorphic.py - Mapping a polymorphic-valued vertical table as a\n+dictionary.\n \n ***** Inheritance Mapping Recipes\u00b6 *****\n **** Basic Inheritance Mappings\u00b6 ****\n Working examples of single-table, joined-table, and concrete-table inheritance\n as described in Mapping_Class_Inheritance_Hierarchies.\n Listing of files:\n- * concrete.py - Concrete-table (table-per-class) inheritance example.\n+ * joined.py - Joined-table (table-per-subclass) inheritance example.\n single.py - Single-table (table-per-hierarchy) inheritance example.\n-joined.py - Joined-table (table-per-subclass) inheritance example.\n+concrete.py - Concrete-table (table-per-class) inheritance example.\n \n ***** Special APIs\u00b6 *****\n **** Attribute Instrumentation\u00b6 ****\n Examples illustrating modifications to SQLAlchemy\u2019s attribute management\n system.\n Listing of files:\n- * active_column_defaults.py - Illustrates use of the\n- AttributeEvents.init_scalar() event, in conjunction with Core column\n- defaults to provide ORM objects that automatically produce the default\n- value when an un-set attribute is accessed.\n+ * custom_management.py - Illustrates customized class instrumentation,\n+ using the sqlalchemy.ext.instrumentation extension package.\n listen_for_events.py - Illustrates how to attach events to all instrumented\n attributes and listen for change events.\n-custom_management.py - Illustrates customized class instrumentation, using the\n-sqlalchemy.ext.instrumentation extension package.\n+active_column_defaults.py - Illustrates use of the AttributeEvents.init_scalar\n+() event, in conjunction with Core column defaults to provide ORM objects that\n+automatically produce the default value when an un-set attribute is accessed.\n \n **** Horizontal Sharding\u00b6 ****\n A basic example of using the SQLAlchemy Sharding API. Sharding refers to\n horizontally scaling data across multiple databases.\n The basic components of a \u201csharded\u201d mapping are:\n * multiple Engine instances, each assigned a \u201cshard id\u201d. These Engine\n instances may refer to different databases, or different schemas /\n@@ -654,34 +651,34 @@\n issue of organizing instances among multiple databases. For a more plain-spoken\n alternative, the \u201cdistinct entity\u201d approach is a simple method of assigning\n objects to different tables (and potentially database nodes) in an explicit way\n - described on the wiki at EntityName.\n Listing of files:\n * separate_databases.py - Illustrates sharding using distinct SQLite\n databases.\n-separate_tables.py - Illustrates sharding using a single SQLite database, that\n-will however have multiple tables using a naming convention.\n separate_schema_translates.py - Illustrates sharding using a single database\n with multiple schemas, where a different \u201cschema_translates_map\u201d can be used\n for each shard.\n+separate_tables.py - Illustrates sharding using a single SQLite database, that\n+will however have multiple tables using a naming convention.\n \n ***** Extending the ORM\u00b6 *****\n **** ORM Query Events\u00b6 ****\n Recipes which illustrate augmentation of ORM SELECT behavior as used by\n Session.execute() with 2.0_style use of select(), as well as the 1.x_style\n Query object.\n Examples include demonstrations of the with_loader_criteria() option as well as\n the SessionEvents.do_orm_execute() hook.\n As of SQLAlchemy 1.4, the Query construct is unified with the Select construct,\n so that these two objects are mostly the same.\n Listing of files:\n- * filter_public.py - Illustrates a global criteria applied to entities of a\n- particular type.\n-temporal_range.py - Illustrates a custom per-query criteria that will be\n-applied to selected entities.\n+ * temporal_range.py - Illustrates a custom per-query criteria that will be\n+ applied to selected entities.\n+filter_public.py - Illustrates a global criteria applied to entities of a\n+particular type.\n \n **** Dogpile Caching\u00b6 ****\n Illustrates how to embed dogpile.cache functionality with ORM queries, allowing\n full cache control as well as the ability to pull \u201clazy loaded\u201d attributes from\n long term cache.\n In this demo, the following techniques are illustrated:\n * Using the SessionEvents.do_orm_execute() event hook\n"}]}, {"source1": "./usr/share/doc/python-sqlalchemy-doc/html/searchindex.js", "source2": "./usr/share/doc/python-sqlalchemy-doc/html/searchindex.js", "unified_diff": null, "details": [{"source1": "js-beautify {}", "source2": "js-beautify {}", "unified_diff": "@@ -7086,17 +7086,17 @@\n \"4138\": 12,\n \"5265\": 12,\n \"5266\": 12,\n \"td\": [12, 63],\n \"protocol\": [12, 17, 48, 51, 55, 59, 65, 66, 80, 102, 113],\n \"5255\": 12,\n \"5271\": 12,\n- \"5314\": 12,\n \"outputtyp\": 12,\n \"5246\": 12,\n+ \"5314\": 12,\n \"5278\": 12,\n \"planner\": [12, 23, 67],\n \"5162\": 12,\n \"slot\": [12, 56, 87, 91],\n \"5228\": 12,\n \"5210\": 12,\n \"eval\": [12, 75, 84, 124, 140],\n@@ -8097,23 +8097,23 @@\n \"6583\": 13,\n \"6652\": 13,\n \"majordalla\": 13,\n \"6649\": 13,\n \"6621\": 13,\n \"6132\": 13,\n \"6658\": 13,\n- \"6596\": 13,\n \"bake_queri\": [13, 23, 123, 140],\n \"foundat\": [13, 25, 27, 42, 46, 160, 161],\n \"analysi\": [13, 76, 95],\n \"win\": 13,\n \"6072\": 13,\n \"6487\": 13,\n \"flask\": [13, 90],\n \"6285\": 13,\n+ \"6596\": 13,\n \"6591\": 13,\n \"6400\": 13,\n \"async_object_sess\": [13, 102],\n \"async_sess\": [13, 102, 123],\n \"in_nested_transact\": [13, 31, 102, 146],\n \"get_transact\": [13, 31, 71, 102, 146, 147],\n \"get_nested_transact\": [13, 31, 102, 146],\n@@ -8521,20 +8521,20 @@\n \"3414\": [13, 25],\n \"alchemy2\": 13,\n \"4644\": 13,\n \"5649\": 13,\n \"get_sequence_nam\": [13, 48, 52],\n \"2056\": 13,\n \"4755\": 13,\n- \"4524\": 13,\n \"upfront\": 13,\n \"returns_unicode_str\": [13, 48],\n \"returns_condit\": [13, 59],\n \"returns_byt\": [13, 59],\n \"5315\": 13,\n+ \"4524\": 13,\n \"hundr\": [13, 21, 24, 25, 31, 76, 136, 137, 155],\n \"4645\": [13, 25],\n \"4808\": [13, 25],\n \"5004\": [13, 25],\n \"har\": [13, 25],\n \"4712\": 13,\n \"5526\": [13, 25],\n@@ -12471,16 +12471,16 @@\n \"n5\": 98,\n \"add_neighbor\": 98,\n \"higher_neighbor\": 98,\n \"directed_graph\": 98,\n \"supplier\": 98,\n \"hasaddress\": 98,\n \"generic_fk\": 98,\n- \"table_per_associ\": 98,\n \"ror\": 98,\n+ \"table_per_associ\": 98,\n \"table_per_rel\": 98,\n \"materialized_path\": 98,\n \"nested_set\": 98,\n \"single_insert\": 98,\n \"bulk_upd\": 98,\n \"test_orm_commit\": 98,\n \"test_bulk_insert_dictionari\": 98,\n@@ -12551,19 +12551,19 @@\n \"poison\": 98,\n \"animalfact\": 98,\n \"custom_manag\": 98,\n \"weather\": 98,\n \"contin\": 98,\n \"spoken\": 98,\n \"separate_databas\": 98,\n- \"separate_t\": 98,\n \"separate_schema_transl\": 98,\n \"schema_translates_map\": 98,\n- \"filter_publ\": 98,\n+ \"separate_t\": 98,\n \"temporal_rang\": 98,\n+ \"filter_publ\": 98,\n \"demo\": 98,\n \"datafil\": 98,\n \"helloworld\": 98,\n \"local_session_cach\": 98,\n \"datamodel\": 98,\n \"postalcod\": 98,\n \"citi\": [98, 124, 133],\n"}]}]}]}]}]}