jpodivin/pep_summarization
Text2Text Generation
•
Updated
•
6
text
stringlengths 330
67k
| status
stringclasses 9
values | title
stringlengths 18
80
| type
stringclasses 3
values | abstract
stringlengths 4
917
|
---|---|---|---|---|
PEP 5 – Guidelines for Language Evolution
Author:
Paul Prescod <paul at prescod.net>
Status:
Superseded
Type:
Process
Created:
26-Oct-2000
Post-History:
Superseded-By:
387
Table of Contents
Abstract
Implementation Details
Scope
Steps For Introducing Backwards-Incompatible Features
Abstract
In the natural evolution of programming languages it is sometimes
necessary to make changes that modify the behavior of older programs.
This PEP proposes a policy for implementing these changes in a manner
respectful of the installed base of Python users.
Implementation Details
Implementation of this PEP requires the addition of a formal warning
and deprecation facility that will be described in another proposal.
Scope
These guidelines apply to future versions of Python that introduce
backward-incompatible behavior. Backward incompatible behavior is a
major deviation in Python interpretation from an earlier behavior
described in the standard Python documentation. Removal of a feature
also constitutes a change of behavior.
This PEP does not replace or preclude other compatibility strategies
such as dynamic loading of backwards-compatible parsers. On the other
hand, if execution of “old code” requires a special switch or pragma
then that is indeed a change of behavior from the point of view of the
user and that change should be implemented according to these
guidelines.
In general, common sense must prevail in the implementation of these
guidelines. For instance changing “sys.copyright” does not constitute
a backwards-incompatible change of behavior!
Steps For Introducing Backwards-Incompatible Features
Propose backwards-incompatible behavior in a PEP. The PEP must
include a section on backwards compatibility that describes in
detail a plan to complete the remainder of these steps.
Once the PEP is accepted as a productive direction, implement an
alternate way to accomplish the task previously provided by the
feature that is being removed or changed. For instance if the
addition operator were scheduled for removal, a new version of
Python could implement an “add()” built-in function.
Formally deprecate the obsolete construct in the Python
documentation.
Add an optional warning mode to the parser that will inform users
when the deprecated construct is used. In other words, all
programs that will behave differently in the future must trigger
warnings in this mode. Compile-time warnings are preferable to
runtime warnings. The warning messages should steer people from
the deprecated construct to the alternative construct.
There must be at least a one-year transition period between the
release of the transitional version of Python and the release of
the backwards incompatible version. Users will have at least a
year to test their programs and migrate them from use of the
deprecated construct to the alternative one.
| Superseded | PEP 5 – Guidelines for Language Evolution | Process | In the natural evolution of programming languages it is sometimes
necessary to make changes that modify the behavior of older programs.
This PEP proposes a policy for implementing these changes in a manner
respectful of the installed base of Python users. |
PEP 6 – Bug Fix Releases
Author:
Aahz <aahz at pythoncraft.com>, Anthony Baxter <anthony at interlink.com.au>
Status:
Superseded
Type:
Process
Created:
15-Mar-2001
Post-History:
15-Mar-2001, 18-Apr-2001, 19-Aug-2004
Table of Contents
Abstract
Motivation
Prohibitions
Not-Quite-Prohibitions
Applicability of Prohibitions
Helping the Bug Fix Releases Happen
Version Numbers
Procedure
Patch Czar History
History
References
Copyright
Note
This PEP is obsolete.
The current release policy is documented in the devguide.
See also PEP 101 for mechanics of the release process.
Abstract
Python has historically had only a single fork of development, with
releases having the combined purpose of adding new features and
delivering bug fixes (these kinds of releases will be referred to as
“major releases”). This PEP describes how to fork off maintenance, or
bug fix, releases of old versions for the primary purpose of fixing
bugs.
This PEP is not, repeat NOT, a guarantee of the existence of bug fix
releases; it only specifies a procedure to be followed if bug fix
releases are desired by enough of the Python community willing to do
the work.
Motivation
With the move to SourceForge, Python development has accelerated.
There is a sentiment among part of the community that there was too
much acceleration, and many people are uncomfortable with upgrading to
new versions to get bug fixes when so many features have been added,
sometimes late in the development cycle.
One solution for this issue is to maintain the previous major release,
providing bug fixes until the next major release. This should make
Python more attractive for enterprise development, where Python may
need to be installed on hundreds or thousands of machines.
Prohibitions
Bug fix releases are required to adhere to the following restrictions:
There must be zero syntax changes. All .pyc and .pyo files must
work (no regeneration needed) with all bugfix releases forked off
from a major release.
There must be zero pickle changes.
There must be no incompatible C API changes. All extensions must
continue to work without recompiling in all bugfix releases in the
same fork as a major release.
Breaking any of these prohibitions requires a BDFL proclamation (and a
prominent warning in the release notes).
Not-Quite-Prohibitions
Where possible, bug fix releases should also:
Have no new features. The purpose of a bug fix release is to fix
bugs, not add the latest and greatest whizzo feature from the HEAD
of the CVS root.
Be a painless upgrade. Users should feel confident that an upgrade
from 2.x.y to 2.x.(y+1) will not break their running systems. This
means that, unless it is necessary to fix a bug, the standard
library should not change behavior, or worse yet, APIs.
Applicability of Prohibitions
The above prohibitions and not-quite-prohibitions apply both for a
final release to a bugfix release (for instance, 2.4 to 2.4.1) and for
one bugfix release to the next in a series (for instance 2.4.1 to
2.4.2).
Following the prohibitions listed in this PEP should help keep the
community happy that a bug fix release is a painless and safe upgrade.
Helping the Bug Fix Releases Happen
Here’s a few pointers on helping the bug fix release process along.
Backport bug fixes. If you fix a bug, and it seems appropriate,
port it to the CVS branch for the current bug fix release. If
you’re unwilling or unable to backport it yourself, make a note in
the commit message, with words like ‘Bugfix candidate’ or
‘Backport candidate’.
If you’re not sure, ask. Ask the person managing the current bug
fix releases if they think a particular fix is appropriate.
If there’s a particular bug you’d particularly like fixed in a bug
fix release, jump up and down and try to get it done. Do not wait
until 48 hours before a bug fix release is due, and then start
asking for bug fixes to be included.
Version Numbers
Starting with Python 2.0, all major releases are required to have a
version number of the form X.Y; bugfix releases will always be of the
form X.Y.Z.
The current major release under development is referred to as release
N; the just-released major version is referred to as N-1.
In CVS, the bug fix releases happen on a branch. For release 2.x, the
branch is named ‘release2x-maint’. For example, the branch for the 2.3
maintenance releases is release23-maint
Procedure
The process for managing bugfix releases is modeled in part on the Tcl
system [1].
The Patch Czar is the counterpart to the BDFL for bugfix releases.
However, the BDFL and designated appointees retain veto power over
individual patches. A Patch Czar might only be looking after a single
branch of development - it’s quite possible that a different person
might be maintaining the 2.3.x and the 2.4.x releases.
As individual patches get contributed to the current trunk of CVS,
each patch committer is requested to consider whether the patch is a
bug fix suitable for inclusion in a bugfix release. If the patch is
considered suitable, the committer can either commit the release to
the maintenance branch, or else mark the patch in the commit message.
In addition, anyone from the Python community is free to suggest
patches for inclusion. Patches may be submitted specifically for
bugfix releases; they should follow the guidelines in PEP 3. In
general, though, it’s probably better that a bug in a specific release
also be fixed on the HEAD as well as the branch.
The Patch Czar decides when there are a sufficient number of patches
to warrant a release. The release gets packaged up, including a
Windows installer, and made public. If any new bugs are found, they
must be fixed immediately and a new bugfix release publicized (with an
incremented version number). For the 2.3.x cycle, the Patch Czar
(Anthony) has been trying for a release approximately every six
months, but this should not be considered binding in any way on any
future releases.
Bug fix releases are expected to occur at an interval of roughly six
months. This is only a guideline, however - obviously, if a major bug
is found, a bugfix release may be appropriate sooner. In general, only
the N-1 release will be under active maintenance at any time. That is,
during Python 2.4’s development, Python 2.3 gets bugfix releases. If,
however, someone qualified wishes to continue the work to maintain an
older release, they should be encouraged.
Patch Czar History
Anthony Baxter is the Patch Czar for 2.3.1 through 2.3.4.
Barry Warsaw is the Patch Czar for 2.2.3.
Guido van Rossum is the Patch Czar for 2.2.2.
Michael Hudson is the Patch Czar for 2.2.1.
Anthony Baxter is the Patch Czar for 2.1.2 and 2.1.3.
Thomas Wouters is the Patch Czar for 2.1.1.
Moshe Zadka is the Patch Czar for 2.0.1.
History
This PEP started life as a proposal on comp.lang.python. The original
version suggested a single patch for the N-1 release to be released
concurrently with the N release. The original version also argued for
sticking with a strict bug fix policy.
Following feedback from the BDFL and others, the draft PEP was written
containing an expanded bugfix release cycle that permitted any
previous major release to obtain patches and also relaxed the strict
bug fix requirement (mainly due to the example of PEP 235, which
could be argued as either a bug fix or a feature).
Discussion then mostly moved to python-dev, where BDFL finally issued
a proclamation basing the Python bugfix release process on Tcl’s,
which essentially returned to the original proposal in terms of being
only the N-1 release and only bug fixes, but allowing multiple bugfix
releases until release N is published.
Anthony Baxter then took this PEP and revised it, based on lessons
from the 2.3 release cycle.
References
[1]
http://www.tcl.tk/cgi-bin/tct/tip/28.html
Copyright
This document has been placed in the public domain.
| Superseded | PEP 6 – Bug Fix Releases | Process | Python has historically had only a single fork of development, with
releases having the combined purpose of adding new features and
delivering bug fixes (these kinds of releases will be referred to as
“major releases”). This PEP describes how to fork off maintenance, or
bug fix, releases of old versions for the primary purpose of fixing
bugs. |
PEP 10 – Voting Guidelines
Author:
Barry Warsaw <barry at python.org>
Status:
Active
Type:
Process
Created:
07-Mar-2002
Post-History:
07-Mar-2002
Table of Contents
Abstract
Rationale
Voting Scores
References
Copyright
Abstract
This PEP outlines the python-dev voting guidelines. These guidelines
serve to provide feedback or gauge the “wind direction” on a
particular proposal, idea, or feature. They don’t have a binding
force.
Rationale
When a new idea, feature, patch, etc. is floated in the Python
community, either through a PEP or on the mailing lists (most likely
on python-dev [1]), it is sometimes helpful to gauge the community’s
general sentiment. Sometimes people just want to register their
opinion of an idea. Sometimes the BDFL wants to take a straw poll.
Whatever the reason, these guidelines have been adopted so as to
provide a common language for developers.
While opinions are (sometimes) useful, but they are never binding.
Opinions that are accompanied by rationales are always valued higher
than bare scores (this is especially true with -1 votes).
Voting Scores
The scoring guidelines are loosely derived from the Apache voting
procedure [2], with of course our own spin on things. There are 4
possible vote scores:
+1 I like it
+0 I don’t care, but go ahead
-0 I don’t care, so why bother?
-1 I hate it
You may occasionally see wild flashes of enthusiasm (either for or
against) with vote scores like +2, +1000, or -1000. These aren’t
really valued much beyond the above scores, but it’s nice to see
people get excited about such geeky stuff.
References
[1]
Python Developer’s Guide,
(http://www.python.org/dev/)
[2]
Apache Project Guidelines and Voting Rules
(http://httpd.apache.org/dev/guidelines.html)
Copyright
This document has been placed in the public domain.
| Active | PEP 10 – Voting Guidelines | Process | This PEP outlines the python-dev voting guidelines. These guidelines
serve to provide feedback or gauge the “wind direction” on a
particular proposal, idea, or feature. They don’t have a binding
force. |
PEP 11 – CPython platform support
Author:
Martin von Löwis <martin at v.loewis.de>,
Brett Cannon <brett at python.org>
Status:
Active
Type:
Process
Created:
07-Jul-2002
Post-History:
18-Aug-2007,
14-May-2014,
20-Feb-2015,
10-Mar-2022
Table of Contents
Abstract
Rationale
Support tiers
Tier 1
Tier 2
Tier 3
All other platforms
Notes
Microsoft Windows
Legacy C Locale
Unsupporting platforms
No-longer-supported platforms
Discussions
Copyright
Abstract
This PEP documents how an operating system (platform) becomes
supported in CPython, what platforms are currently supported, and
documents past support.
Rationale
Over time, the CPython source code has collected various pieces of
platform-specific code, which, at some point in time, was
considered necessary to use CPython on a specific platform.
Without access to this platform, it is not possible to determine
whether this code is still needed. As a result, this code may
either break during CPython’s evolution, or it may become
unnecessary as the platforms evolve as well.
Allowing these fragments to grow poses the risk of
unmaintainability: without having experts for a large number of
platforms, it is not possible to determine whether a certain
change to the CPython source code will work on all supported
platforms.
To reduce this risk, this PEP specifies what is required for a
platform to be considered supported by CPython as well as providing a
procedure to remove code for platforms with few or no CPython
users.
This PEP also lists what platforms are supported by the CPython
interpreter. This lets people know what platforms are directly
supported by the CPython development team.
Support tiers
Platform support is broken down into tiers. Each tier comes with
different requirements which lead to different promises being made
about support.
To be promoted to a tier, steering council support is required and is
expected to be driven by team consensus. Demotion to a lower tier
occurs when the requirements of the current tier are no longer met for
a platform for an extended period of time based on the judgment of
the release manager or steering council. For platforms which no longer
meet the requirements of any tier by b1 of a new feature release, an
announcement will be made to warn the community of the pending removal
of support for the platform (e.g. in the b1 announcement). If the
platform is not brought into line for at least one of the tiers by the
first release candidate, it will be listed as unsupported in this PEP.
Tier 1
STATUS
CI failures block releases.
Changes which would break the main branch are not allowed to be merged;
any breakage should be fixed or reverted immediately.
All core developers are responsible to keep main, and thus these
platforms, working.
Failures on these platforms block a release.
Target Triple
Notes
i686-pc-windows-msvc
x86_64-pc-windows-msvc
x86_64-apple-darwin
BSD libc, clang
x86_64-unknown-linux-gnu
glibc, gcc
Tier 2
STATUS
Must have a reliable buildbot.
At least two core developers are signed up to support the platform.
Changes which break any of these platforms are to be fixed or
reverted within 24 hours.
Failures on these platforms block a release.
Target Triple
Notes
Contacts
aarch64-apple-darwin
clang
Ned Deily, Ronald Oussoren, Dong-hee Na
aarch64-unknown-linux-gnu
glibc, gccglibc, clang
Petr Viktorin, Victor StinnerVictor Stinner, Gregory P. Smith
wasm32-unknown-wasi
WASI SDK, Wasmtime
Brett Cannon, Eric Snow
x86_64-unknown-linux-gnu
glibc, clang
Victor Stinner, Gregory P. Smith
Tier 3
STATUS
Must have a reliable buildbot.
At least one core developer is signed up to support the platform.
No response SLA to failures.
Failures on these platforms do not block a release.
Target Triple
Notes
Contacts
aarch64-pc-windows-msvc
Steve Dower
armv7l-unknown-linux-gnueabihf
Raspberry Pi OS, glibc, gcc
Gregory P. Smith
powerpc64le-unknown-linux-gnu
glibc, clangglibc, gcc
Victor StinnerVictor Stinner
s390x-unknown-linux-gnu
glibc, gcc
Victor Stinner
x86_64-unknown-freebsd
BSD libc, clang
Victor Stinner
All other platforms
Support for a platform may be partial within the code base, such as
from active development around platform support or accidentally.
Code changes to platforms not listed in the above tiers may be rejected
or removed from the code base without a deprecation process if they
cause a maintenance burden or obstruct general improvements.
Platforms not listed here may be supported by the wider Python
community in some way. If your desired platform is not listed above,
please perform a search online to see if someone is already providing
support in some form.
Notes
Microsoft Windows
Windows versions prior to Windows 10 follow Microsoft’s Fixed Lifecycle Policy,
with a mainstream support phase for 5 years after release,
where the product is generally commercially available,
and an additional 5 year extended support phase,
where paid support is still available and certain bug fixes are released.
Extended Security Updates (ESU)
is a paid program available to high-volume enterprise customers
as a “last resort” option to receive certain security updates after extended support ends.
ESU is considered a distinct phase that follows the expiration of extended support.
Windows 10 and later follow Microsoft’s Modern Lifecycle Policy,
which varies per-product, per-version, per-edition and per-channel.
Generally, feature updates (1709, 22H2) occur every 6-12 months
and are supported for 18-36 months;
Server and IoT editions, and LTSC channel releases are supported for 5-10 years,
and the latest feature release of a major version (Windows 10, Windows 11)
generally receives new updates for at least 10 years following release.
Microsoft’s Windows Lifecycle FAQ
has more specific and up-to-date guidance.
CPython’s Windows support currently follows Microsoft’s lifecycles.
A new feature release X.Y.0 will support all Windows versions
whose extended support phase has not yet expired.
Subsequent bug fix releases will support the same Windows versions
as the original feature release, even if no longer supported by Microsoft.
New versions of Windows released while CPython is in maintenance mode
may be supported at the discretion of the core team and release manager.
As of 2024, our current interpretation of Microsoft’s lifecycles is that
Windows for IoT and embedded systems is out of scope for new CPython releases,
as the intent of those is to avoid feature updates. Windows Server will usually
be the oldest version still receiving free security fixes, and that will
determine the earliest supported client release with equivalent API version
(which will usually be past its end-of-life).
Each feature release is built by a specific version of Microsoft
Visual Studio. That version should have mainstream support when the
release is made. Developers of extension modules will generally need
to use the same Visual Studio release; they are concerned both with
the availability of the versions they need to use, and with keeping
the zoo of versions small. The CPython source tree will keep
unmaintained build files for older Visual Studio releases, for which
patches will be accepted. Such build files will be removed from the
source tree 3 years after the extended support for the compiler has
ended (but continue to remain available in revision control).
Legacy C Locale
Starting with CPython 3.7.0, *nix platforms are expected to provide
at least one of C.UTF-8 (full locale), C.utf8 (full locale) or
UTF-8 (LC_CTYPE-only locale) as an alternative to the legacy C
locale.
Any Unicode-related integration problems that occur only in the legacy C
locale and cannot be reproduced in an appropriately configured non-ASCII
locale will be closed as “won’t fix”.
Unsupporting platforms
If a platform drops out of tiered support, a note must be posted
in this PEP that the platform is no longer actively supported. This
note must include:
The name of the system,
The first release number that does not support this platform
anymore, and
The first release where the historical support code is actively
removed.
In some cases, it is not possible to identify the specific list of
systems for which some code is used (e.g. when autoconf tests for
absence of some feature which is considered present on all
supported systems). In this case, the name will give the precise
condition (usually a preprocessor symbol) that will become
unsupported.
At the same time, the CPython build must be changed to produce a
warning if somebody tries to install CPython on this platform. On
platforms using autoconf, configure should also be made emit a warning
about the unsupported platform.
This gives potential users of the platform a chance to step forward
and offer maintenance. We do not treat a platform that loses Tier 3
support any worse than a platform that was never supported.
No-longer-supported platforms
Name: MS-DOS, MS-Windows 3.x
Unsupported in: Python 2.0
Code removed in: Python 2.1
Name: SunOS 4
Unsupported in: Python 2.3
Code removed in: Python 2.4
Name: DYNIX
Unsupported in: Python 2.3
Code removed in: Python 2.4
Name: dgux
Unsupported in: Python 2.3
Code removed in: Python 2.4
Name: Minix
Unsupported in: Python 2.3
Code removed in: Python 2.4
Name: Irix 4 and –with-sgi-dl
Unsupported in: Python 2.3
Code removed in: Python 2.4
Name: Linux 1
Unsupported in: Python 2.3
Code removed in: Python 2.4
Name: Systems defining __d6_pthread_create (configure.in)
Unsupported in: Python 2.3
Code removed in: Python 2.4
Name: Systems defining PY_PTHREAD_D4, PY_PTHREAD_D6,
or PY_PTHREAD_D7 in thread_pthread.h
Unsupported in: Python 2.3
Code removed in: Python 2.4
Name: Systems using –with-dl-dld
Unsupported in: Python 2.3
Code removed in: Python 2.4
Name: Systems using –without-universal-newlines,
Unsupported in: Python 2.3
Code removed in: Python 2.4
Name: MacOS 9
Unsupported in: Python 2.4
Code removed in: Python 2.4
Name: Systems using –with-wctype-functions
Unsupported in: Python 2.6
Code removed in: Python 2.6
Name: Win9x, WinME, NT4
Unsupported in: Python 2.6 (warning in 2.5 installer)
Code removed in: Python 2.6
Name: AtheOS
Unsupported in: Python 2.6 (with “AtheOS” changed to “Syllable”)
Build broken in: Python 2.7 (edit configure to re-enable)
Code removed in: Python 3.0
Details: http://www.syllable.org/discussion.php?id=2320
Name: BeOS
Unsupported in: Python 2.6 (warning in configure)
Build broken in: Python 2.7 (edit configure to re-enable)
Code removed in: Python 3.0
Name: Systems using Mach C Threads
Unsupported in: Python 3.2
Code removed in: Python 3.3
Name: SunOS lightweight processes (LWP)
Unsupported in: Python 3.2
Code removed in: Python 3.3
Name: Systems using –with-pth (GNU pth threads)
Unsupported in: Python 3.2
Code removed in: Python 3.3
Name: Systems using Irix threads
Unsupported in: Python 3.2
Code removed in: Python 3.3
Name: OSF* systems (issue 8606)
Unsupported in: Python 3.2
Code removed in: Python 3.3
Name: OS/2 (issue 16135)
Unsupported in: Python 3.3
Code removed in: Python 3.4
Name: VMS (issue 16136)
Unsupported in: Python 3.3
Code removed in: Python 3.4
Name: Windows 2000
Unsupported in: Python 3.3
Code removed in: Python 3.4
Name: Windows systems where COMSPEC points to command.com
Unsupported in: Python 3.3
Code removed in: Python 3.4
Name: RISC OS
Unsupported in: Python 3.0 (some code actually removed)
Code removed in: Python 3.4
Name: IRIX
Unsupported in: Python 3.7
Code removed in: Python 3.7
Name: Systems without multithreading support
Unsupported in: Python 3.7
Code removed in: Python 3.7
Name: wasm32-unknown-emscripten
Unsupported in: Python 3.13
Code removed in: Unknown
Discussions
April 2022: Consider adding a Tier 3 to tiered platform support
(Victor Stinner)
March 2022: Proposed tiered platform support
(Brett Cannon)
February 2015: Update to PEP 11 to clarify garnering platform support
(Brett Cannon)
May 2014: Where is our official policy of what platforms we do support?
(Brett Cannon)
August 2007: PEP 11 update - Call for port maintainers to step forward
(Skip Montanaro)
Copyright
This document is placed in the public domain or under the
CC0-1.0-Universal license, whichever is more permissive.
| Active | PEP 11 – CPython platform support | Process | This PEP documents how an operating system (platform) becomes
supported in CPython, what platforms are currently supported, and
documents past support. |
PEP 12 – Sample reStructuredText PEP Template
Author:
David Goodger <goodger at python.org>,
Barry Warsaw <barry at python.org>,
Brett Cannon <brett at python.org>
Status:
Active
Type:
Process
Created:
05-Aug-2002
Post-History:
30-Aug-2002
Table of Contents
Abstract
Rationale
How to Use This Template
ReStructuredText PEP Formatting Requirements
General
Section Headings
Paragraphs
Inline Markup
Block Quotes
Literal Blocks
Lists
Tables
Hyperlinks
Internal and PEP/RFC Links
Footnotes
Images
Comments
Escaping Mechanism
Canonical Documentation and Intersphinx
Habits to Avoid
Suggested Sections
Resources
Copyright
Note
For those who have written a PEP before, there is a template
(which is included as a file in the PEPs repository).
Abstract
This PEP provides a boilerplate or sample template for creating your
own reStructuredText PEPs. In conjunction with the content guidelines
in PEP 1, this should make it easy for you to conform your own
PEPs to the format outlined below.
Note: if you are reading this PEP via the web, you should first grab
the text (reStructuredText) source of this PEP in order to complete
the steps below. DO NOT USE THE HTML FILE AS YOUR TEMPLATE!
The source for this (or any) PEP can be found in the
PEPs repository,
as well as via a link at the bottom of each PEP.
Rationale
If you intend to submit a PEP, you MUST use this template, in
conjunction with the format guidelines below, to ensure that your PEP
submission won’t get automatically rejected because of form.
ReStructuredText provides PEP authors with useful functionality and
expressivity, while maintaining easy readability in the source text.
The processed HTML form makes the functionality accessible to readers:
live hyperlinks, styled text, tables, images, and automatic tables of
contents, among other advantages.
How to Use This Template
To use this template you must first decide whether your PEP is going
to be an Informational or Standards Track PEP. Most PEPs are
Standards Track because they propose a new feature for the Python
language or standard library. When in doubt, read PEP 1 for details,
or open a tracker issue on the PEPs repo to ask for assistance.
Once you’ve decided which type of PEP yours is going to be, follow the
directions below.
Make a copy of this file (the .rst file, not the HTML!) and
perform the following edits. Name the new file pep-NNNN.rst, using
the next available number (not used by a published or in-PR PEP).
Replace the “PEP: 12” header with “PEP: NNNN”,
matching the file name. Note that the file name should be padded with
zeros (eg pep-0012.rst), but the header should not (PEP: 12).
Change the Title header to the title of your PEP.
Change the Author header to include your name, and optionally your
email address. Be sure to follow the format carefully: your name
must appear first, and it must not be contained in parentheses.
Your email address may appear second (or it can be omitted) and if
it appears, it must appear in angle brackets. It is okay to
obfuscate your email address.
If none of the authors are Python core developers, include a Sponsor
header with the name of the core developer sponsoring your PEP.
Add the direct URL of the PEP’s canonical discussion thread
(on e.g. Python-Dev, Discourse, etc) under the Discussions-To header.
If the thread will be created after the PEP is submitted as an official
draft, it is okay to just list the venue name initially, but remember to
update the PEP with the URL as soon as the PEP is successfully merged
to the PEPs repository and you create the corresponding discussion thread.
See PEP 1 for more details.
Change the Status header to “Draft”.
For Standards Track PEPs, change the Type header to “Standards
Track”.
For Informational PEPs, change the Type header to “Informational”.
For Standards Track PEPs, if your feature depends on the acceptance
of some other currently in-development PEP, add a Requires header
right after the Type header. The value should be the PEP number of
the PEP yours depends on. Don’t add this header if your dependent
feature is described in a Final PEP.
Change the Created header to today’s date. Be sure to follow the
format carefully: it must be in dd-mmm-yyyy format, where the
mmm is the 3 English letter month abbreviation, i.e. one of Jan,
Feb, Mar, Apr, May, Jun, Jul, Aug, Sep, Oct, Nov, Dec.
For Standards Track PEPs, after the Created header, add a
Python-Version header and set the value to the next planned version
of Python, i.e. the one your new feature will hopefully make its
first appearance in. Do not use an alpha or beta release
designation here. Thus, if the last version of Python was 2.2 alpha
1 and you’re hoping to get your new feature into Python 2.2, set the
header to:Python-Version: 2.2
Add a Topic header if the PEP belongs under one shown at the Topic Index.
Most PEPs don’t.
Leave Post-History alone for now; you’ll add dates and corresponding links
to this header each time you post your PEP to the designated discussion forum
(and update the Discussions-To header with said link, as above).
For each thread, use the date (in the dd-mmm-yyy format) as the
linked text, and insert the URLs inline as anonymous reST hyperlinks,
with commas in between each posting.If you posted threads for your PEP on August 14, 2001 and September 3, 2001,
the Post-History header would look like, e.g.:
Post-History: `14-Aug-2001 <https://www.example.com/thread_1>`__,
`03-Sept-2001 <https://www.example.com/thread_2>`__
You should add the new dates/links here as soon as you post a
new discussion thread.
Add a Replaces header if your PEP obsoletes an earlier PEP. The
value of this header is the number of the PEP that your new PEP is
replacing. Only add this header if the older PEP is in “final”
form, i.e. is either Accepted, Final, or Rejected. You aren’t
replacing an older open PEP if you’re submitting a competing idea.
Now write your Abstract, Rationale, and other content for your PEP,
replacing all this gobbledygook with your own text. Be sure to
adhere to the format guidelines below, specifically on the
prohibition of tab characters and the indentation requirements.
See “Suggested Sections” below for a template of sections to include.
Update your Footnotes section, listing any footnotes and
non-inline link targets referenced by the text.
Run ./build.py to ensure the PEP is rendered without errors,
and check that the output in build/pep-NNNN.html looks as you intend.
Create a pull request against the PEPs repository.
For reference, here are all of the possible header fields (everything
in brackets should either be replaced or have the field removed if
it has a leading * marking it as optional and it does not apply to
your PEP):
PEP: [NNN]
Title: [...]
Author: [Full Name <email at example.com>]
Sponsor: *[Full Name <email at example.com>]
PEP-Delegate:
Discussions-To: [URL]
Status: Draft
Type: [Standards Track | Informational | Process]
Topic: *[Governance | Packaging | Release | Typing]
Requires: *[NNN]
Created: [DD-MMM-YYYY]
Python-Version: *[M.N]
Post-History: [`DD-MMM-YYYY <URL>`__]
Replaces: *[NNN]
Superseded-By: *[NNN]
Resolution:
ReStructuredText PEP Formatting Requirements
The following is a PEP-specific summary of reStructuredText syntax.
For the sake of simplicity and brevity, much detail is omitted. For
more detail, see Resources below. Literal blocks (in which no
markup processing is done) are used for examples throughout, to
illustrate the plaintext markup.
General
Lines should usually not extend past column 79,
excepting URLs and similar circumstances.
Tab characters must never appear in the document at all.
Section Headings
PEP headings must begin in column zero and the initial letter of each
word must be capitalized as in book titles. Acronyms should be in all
capitals. Section titles must be adorned with an underline, a single
repeated punctuation character, which begins in column zero and must
extend at least as far as the right edge of the title text (4
characters minimum). First-level section titles are underlined with
“=” (equals signs), second-level section titles with “-” (hyphens),
and third-level section titles with “’” (single quotes or
apostrophes). For example:
First-Level Title
=================
Second-Level Title
------------------
Third-Level Title
'''''''''''''''''
If there are more than three levels of sections in your PEP, you may
insert overline/underline-adorned titles for the first and second
levels as follows:
============================
First-Level Title (optional)
============================
-----------------------------
Second-Level Title (optional)
-----------------------------
Third-Level Title
=================
Fourth-Level Title
------------------
Fifth-Level Title
'''''''''''''''''
You shouldn’t have more than five levels of sections in your PEP. If
you do, you should consider rewriting it.
You must use two blank lines between the last line of a section’s body
and the next section heading. If a subsection heading immediately
follows a section heading, a single blank line in-between is
sufficient.
The body of each section is not normally indented, although some
constructs do use indentation, as described below. Blank lines are
used to separate constructs.
Paragraphs
Paragraphs are left-aligned text blocks separated by blank lines.
Paragraphs are not indented unless they are part of an indented
construct (such as a block quote or a list item).
Inline Markup
Portions of text within paragraphs and other text blocks may be
styled. For example:
Text may be marked as *emphasized* (single asterisk markup,
typically shown in italics) or **strongly emphasized** (double
asterisks, typically boldface). ``Inline literals`` (using double
backquotes) are typically rendered in a monospaced typeface. No
further markup recognition is done within the double backquotes,
so they're safe for any kind of code snippets.
Block Quotes
Block quotes consist of indented body elements. For example:
This is a paragraph.
This is a block quote.
A block quote may contain many paragraphs.
Block quotes are used to quote extended passages from other sources.
Block quotes may be nested inside other body elements. Use 4 spaces
per indent level.
Literal Blocks
Literal blocks are used for code samples and other preformatted text.
To indicate a literal block, preface the indented text block with
“::” (two colons), or use the .. code-block:: directive.
Indent the text block by 4 spaces; the literal block continues until the end
of the indentation. For example:
This is a typical paragraph. A literal block follows.
::
for a in [5, 4, 3, 2, 1]: # this is program code, shown as-is
print(a)
print("it's...")
“::” is also recognized at the end of any paragraph; if not immediately
preceded by whitespace, one colon will remain visible in the final output:
This is an example::
Literal block
By default, literal blocks will be syntax-highlighted as Python code.
For specific blocks that contain code or data in other languages/formats,
use the .. code-block:: language directive, substituting the “short name”
of the appropriate Pygments lexer
(or text to disable highlighting) for language. For example:
.. code-block:: rst
An example of the ``rst`` lexer (i.e. *reStructuredText*).
For PEPs that predominantly contain literal blocks of a specific language,
use the .. highlight:: language directive with the appropriate language
at the top of the PEP body (below the headers and above the Abstract).
All literal blocks will then be treated as that language,
unless specified otherwise in the specific .. code-block. For example:
.. highlight:: c
Abstract
========
Here's some C code::
printf("Hello, World!\n");
Lists
Bullet list items begin with one of “-”, “*”, or “+” (hyphen,
asterisk, or plus sign), followed by whitespace and the list item
body. List item bodies must be left-aligned and indented relative to
the bullet; the text immediately after the bullet determines the
indentation. For example:
This paragraph is followed by a list.
* This is the first bullet list item. The blank line above the
first list item is required; blank lines between list items
(such as below this paragraph) are optional.
* This is the first paragraph in the second item in the list.
This is the second paragraph in the second item in the list.
The blank line above this paragraph is required. The left edge
of this paragraph lines up with the paragraph above, both
indented relative to the bullet.
- This is a sublist. The bullet lines up with the left edge of
the text blocks above. A sublist is a new list so requires a
blank line above and below.
* This is the third item of the main list.
This paragraph is not part of the list.
Enumerated (numbered) list items are similar, but use an enumerator
instead of a bullet. Enumerators are numbers (1, 2, 3, …), letters
(A, B, C, …; uppercase or lowercase), or Roman numerals (i, ii, iii,
iv, …; uppercase or lowercase), formatted with a period suffix
(“1.”, “2.”), parentheses (“(1)”, “(2)”), or a right-parenthesis
suffix (“1)”, “2)”). For example:
1. As with bullet list items, the left edge of paragraphs must
align.
2. Each list item may contain multiple paragraphs, sublists, etc.
This is the second paragraph of the second list item.
a) Enumerated lists may be nested.
b) Blank lines may be omitted between list items.
Definition lists are written like this:
what
Definition lists associate a term with a definition.
how
The term is a one-line phrase, and the definition is one
or more paragraphs or body elements, indented relative to
the term.
Tables
Simple tables are easy and compact:
===== ===== =======
A B A and B
===== ===== =======
False False False
True False False
False True False
True True True
===== ===== =======
There must be at least two columns in a table (to differentiate from
section titles). Column spans use underlines of hyphens (“Inputs”
spans the first two columns):
===== ===== ======
Inputs Output
------------ ------
A B A or B
===== ===== ======
False False False
True False True
False True True
True True True
===== ===== ======
Text in a first-column cell starts a new row. No text in the first
column indicates a continuation line; the rest of the cells may
consist of multiple lines. For example:
===== =========================
col 1 col 2
===== =========================
1 Second column of row 1.
2 Second column of row 2.
Second line of paragraph.
3 - Second column of row 3.
- Second item in bullet
list (row 3, column 2).
===== =========================
Hyperlinks
When referencing an external web page in the body of a PEP, you should
include the title of the page or a suitable description in the text, with
either an inline hyperlink or a separate explicit target with the URL.
Do not include bare URLs in the body text of the PEP, and use HTTPS
links wherever available.
Hyperlink references use backquotes and a trailing underscore to mark
up the reference text; backquotes are optional if the reference text
is a single word. For example, to reference a hyperlink target named
Python website, you would write:
In this paragraph, we refer to the `Python website`_.
If you intend to only reference a link once, and want to define it inline
with the text, insert the link into angle brackets (<>) after the text
you want to link, but before the closing backtick, with a space between the
text and the opening backtick. You should also use a double-underscore after
the closing backtick instead of a single one, which makes it an anonymous
reference to avoid conflicting with other target names. For example:
Visit the `website <https://www.python.org/>`__ for more.
If you want to use one link multiple places with different linked text,
or want to ensure you don’t have to update your link target names when
changing the linked text, include the target name within angle brackets
following the text to link, with an underscore after the target name
but before the closing angle bracket (or the link will not work).
For example:
For further examples, see the `documentation <pydocs_>`_.
An explicit target provides the URL. Put targets in the Footnotes section
at the end of the PEP, or immediately after the paragraph with the reference.
Hyperlink targets begin with two periods and a space (the “explicit
markup start”), followed by a leading underscore, the reference text,
a colon, and the URL.
.. _Python web site: https://www.python.org/
.. _pydocs: https://docs.python.org/
The reference text and the target text must match (although the match
is case-insensitive and ignores differences in whitespace). Note that
the underscore trails the reference text but precedes the target text.
If you think of the underscore as a right-pointing arrow, it points
away from the reference and toward the target.
Internal and PEP/RFC Links
The same mechanism as hyperlinks can be used for internal references.
Every unique section title implicitly defines an internal hyperlink target.
We can make a link to the Abstract section like this:
Here is a hyperlink reference to the `Abstract`_ section. The
backquotes are optional since the reference text is a single word;
we can also just write: Abstract_.
To refer to PEPs or RFCs, always use the :pep: and :rfc: roles,
never hardcoded URLs.
For example:
See :pep:`1` for more information on how to write a PEP,
and :pep:`the Hyperlink section of PEP 12 <12#hyperlinks>` for how to link.
This renders as:
See PEP 1 for more information on how to write a PEP,
and the Hyperlink section of PEP 12 for how to link.
PEP numbers in the text are never padded, and there is a space (not a dash)
between “PEP” or “RFC” and the number; the above roles will take care of
that for you.
Footnotes
Footnote references consist of a left square bracket, a label, a
right square bracket, and a trailing underscore.
Instead of a number, use a label of the
form “#word”, where “word” is a mnemonic consisting of alphanumerics
plus internal hyphens, underscores, and periods (no whitespace or
other characters are allowed).
For example:
Refer to The TeXbook [#TeXbook]_ for more information.
which renders as
Refer to The TeXbook [1] for more information.
Whitespace must precede the footnote reference. Leave a space between
the footnote reference and the preceding word.
Use footnotes for additional notes, explanations and caveats, as well as
for references to books and other sources not readily available online.
Native reST hyperlink targets or inline hyperlinks in the text should be
used in preference to footnotes for including URLs to online resources.
Footnotes begin with “.. “ (the explicit
markup start), followed by the footnote marker (no underscores),
followed by the footnote body. For example:
.. [#TeXbook] Donald Knuth's *The TeXbook*, pages 195 and 196.
which renders as
[1]
Donald Knuth’s The TeXbook, pages 195 and 196.
Footnotes and footnote references will be numbered automatically, and
the numbers will always match.
Images
If your PEP contains a diagram or other graphic, you may include it in the
processed output using the image directive:
.. image:: diagram.png
Any browser-friendly graphics format is possible; PNG should be
preferred for graphics, JPEG for photos and GIF for animations.
Currently, SVG must be avoided due to compatibility issues with the
PEP build system.
For accessibility and readers of the source text, you should include
a description of the image and any key information contained within
using the :alt: option to the image directive:
.. image:: dataflow.png
:alt: Data flows from the input module, through the "black box"
module, and finally into (and through) the output module.
Comments
A comment is an indented block of arbitrary text immediately
following an explicit markup start: two periods and whitespace. Leave
the “..” on a line by itself to ensure that the comment is not
misinterpreted as another explicit markup construct. Comments are not
visible in the processed document. For example:
..
This section should be updated in the final PEP.
Ensure the date is accurate.
Escaping Mechanism
reStructuredText uses backslashes (”\”) to override the special
meaning given to markup characters and get the literal characters
themselves. To get a literal backslash, use an escaped backslash
(”\\”). There are two contexts in which backslashes have no
special meaning: literal blocks and inline literals (see Inline
Markup above). In these contexts, no markup recognition is done,
and a single backslash represents a literal backslash, without having
to double up.
If you find that you need to use a backslash in your text, consider
using inline literals or a literal block instead.
Canonical Documentation and Intersphinx
As PEP 1 describes,
PEPs are considered historical documents once marked Final,
and their canonical documentation/specification should be moved elsewhere.
To indicate this, use the canonical-doc directive
or an appropriate subclass:
canonical-pypa-spec for packaging standards
canonical-typing-spec for typing standards
Furthermore, you can use
Intersphinx references
to other Sphinx sites,
currently the Python documentation
and packaging.python.org,
to easily cross-reference pages, sections and Python/C objects.
This works with both the “canonical” directives and anywhere in your PEP.
Add the directive between the headers and the first section of the PEP
(typically the Abstract)
and pass as an argument an Intersphinx reference of the canonical doc/spec
(or if the target is not on a Sphinx site, a reST hyperlink).
For example,
to create a banner pointing to the sqlite3 docs,
you would write the following:
.. canonical-doc:: :mod:`python:sqlite3`
which would generate the banner:
Important
This PEP is a historical document. The up-to-date, canonical documentation can now be found at sqlite3.
×
See PEP 1 for how to propose changes.
Or for a PyPA spec,
such as the Core metadata specifications,
you would use:
.. canonical-pypa-spec:: :ref:`packaging:core-metadata`
which renders as:
Attention
This PEP is a historical document. The up-to-date, canonical spec, Core metadata specifications, is maintained on the PyPA specs page.
×
See the PyPA specification update process for how to propose changes.
The argument accepts arbitrary reST,
so you can include multiple linked docs/specs and name them whatever you like,
and you can also include directive content that will be inserted into the text.
The following advanced example:
.. canonical-doc:: the :ref:`python:sqlite3-connection-objects` and :exc:`python:~sqlite3.DataError` docs
Also, see the :ref:`Data Persistence docs <persistence>` for other examples.
would render as:
Important
This PEP is a historical document. The up-to-date, canonical documentation can now be found at the Connection objects and sqlite3.DataError docs.
×
Also, see the Data Persistence docs for other examples.
See PEP 1 for how to propose changes.
Habits to Avoid
Many programmers who are familiar with TeX often write quotation marks
like this:
`single-quoted' or ``double-quoted''
Backquotes are significant in reStructuredText, so this practice
should be avoided. For ordinary text, use ordinary ‘single-quotes’ or
“double-quotes”. For inline literal text (see Inline Markup
above), use double-backquotes:
``literal text: in here, anything goes!``
Suggested Sections
Various sections are found to be common across PEPs and are outlined in
PEP 1. Those sections are provided here for convenience.
PEP: <REQUIRED: pep number>
Title: <REQUIRED: pep title>
Author: <REQUIRED: list of authors' real names and optionally, email addrs>
Sponsor: <real name of sponsor>
PEP-Delegate: <PEP delegate's real name>
Discussions-To: <REQUIRED: URL of current canonical discussion thread>
Status: <REQUIRED: Draft | Active | Accepted | Provisional | Deferred | Rejected | Withdrawn | Final | Superseded>
Type: <REQUIRED: Standards Track | Informational | Process>
Topic: <Governance | Packaging | Release | Typing>
Requires: <pep numbers>
Created: <date created on, in dd-mmm-yyyy format>
Python-Version: <version number>
Post-History: <REQUIRED: dates, in dd-mmm-yyyy format, and corresponding links to PEP discussion threads>
Replaces: <pep number>
Superseded-By: <pep number>
Resolution: <url>
Abstract
========
[A short (~200 word) description of the technical issue being addressed.]
Motivation
==========
[Clearly explain why the existing language specification is inadequate to address the problem that the PEP solves.]
Rationale
=========
[Describe why particular design decisions were made.]
Specification
=============
[Describe the syntax and semantics of any new language feature.]
Backwards Compatibility
=======================
[Describe potential impact and severity on pre-existing code.]
Security Implications
=====================
[How could a malicious user take advantage of this new feature?]
How to Teach This
=================
[How to teach users, new and experienced, how to apply the PEP to their work.]
Reference Implementation
========================
[Link to any existing implementation and details about its state, e.g. proof-of-concept.]
Rejected Ideas
==============
[Why certain ideas that were brought while discussing this PEP were not ultimately pursued.]
Open Issues
===========
[Any points that are still being decided/discussed.]
Footnotes
=========
[A collection of footnotes cited in the PEP, and a place to list non-inline hyperlink targets.]
Copyright
=========
This document is placed in the public domain or under the
CC0-1.0-Universal license, whichever is more permissive.
Resources
Many other constructs and variations are possible,
both those supported by basic Docutils
and the extensions added by Sphinx.
A number of resources are available to learn more about them:
Sphinx ReStructuredText Primer,
a gentle but fairly detailed introduction.
reStructuredText Markup Specification,
the authoritative, comprehensive documentation of the basic reST syntax,
directives, roles and more.
Sphinx Roles
and Sphinx Directives,
the extended constructs added by the Sphinx documentation system used to
render the PEPs to HTML.
If you have questions or require assistance with writing a PEP that the above
resources don’t address, ping @python/pep-editors on GitHub, open an
issue on the PEPs repository
or reach out to a PEP editor directly.
Copyright
This document is placed in the public domain or under the
CC0-1.0-Universal license, whichever is more permissive.
| Active | PEP 12 – Sample reStructuredText PEP Template | Process | This PEP provides a boilerplate or sample template for creating your
own reStructuredText PEPs. In conjunction with the content guidelines
in PEP 1, this should make it easy for you to conform your own
PEPs to the format outlined below. |
PEP 13 – Python Language Governance
Author:
The Python core team and community
Status:
Active
Type:
Process
Topic:
Governance
Created:
16-Dec-2018
Table of Contents
Abstract
Current steering council
Specification
The steering council
Composition
Mandate
Powers
Electing the council
Term
Vacancies
Conflicts of interest
Ejecting core team members
Vote of no confidence
The core team
Role
Prerogatives
Membership
Changing this document
History
Creation of this document
History of council elections
History of amendments
Acknowledgements
Copyright
Abstract
This PEP defines the formal governance process for Python, and records
how this has changed over time. Currently, governance is based around
a steering council. The council has broad authority, which they seek
to exercise as rarely as possible.
Current steering council
The 2024 term steering council consists of:
Barry Warsaw
Emily Morehouse
Gregory P. Smith
Pablo Galindo Salgado
Thomas Wouters
Per the results of the vote tracked in PEP 8105.
The core team consists of those listed in the private
https://github.com/python/voters/ repository which is publicly
shared via https://devguide.python.org/developers/.
Specification
The steering council
Composition
The steering council is a 5-person committee.
Mandate
The steering council shall work to:
Maintain the quality and stability of the Python language and
CPython interpreter,
Make contributing as accessible, inclusive, and sustainable as
possible,
Formalize and maintain the relationship between the core team and
the PSF,
Establish appropriate decision-making processes for PEPs,
Seek consensus among contributors and the core team before acting in
a formal capacity,
Act as a “court of final appeal” for decisions where all other
methods have failed.
Powers
The council has broad authority to make decisions about the project.
For example, they can:
Accept or reject PEPs
Enforce or update the project’s code of conduct
Work with the PSF to manage any project assets
Delegate parts of their authority to other subcommittees or
processes
However, they cannot modify this PEP, or affect the membership of the
core team, except via the mechanisms specified in this PEP.
The council should look for ways to use these powers as little as
possible. Instead of voting, it’s better to seek consensus. Instead of
ruling on individual PEPs, it’s better to define a standard process
for PEP decision making (for example, by accepting one of the other
801x series of PEPs). It’s better to establish a Code of Conduct
committee than to rule on individual cases. And so on.
To use its powers, the council votes. Every council member must either
vote or explicitly abstain. Members with conflicts of interest on a
particular vote must abstain. Passing requires a strict majority of
non-abstaining council members.
Whenever possible, the council’s deliberations and votes shall be held
in public.
Electing the council
A council election consists of two phases:
Phase 1: Candidates advertise their interest in serving. Candidates
must be nominated by a core team member. Self-nominations are
allowed.
Phase 2: Each core team member can vote for zero or more of the
candidates. Voting is performed anonymously. Candidates are ranked
by the total number of votes they receive. If a tie occurs, it may
be resolved by mutual agreement among the candidates, or else the
winner will be chosen at random.
Each phase lasts one to two weeks, at the outgoing council’s discretion.
For the initial election, both phases will last two weeks.
The election process is managed by a returns officer nominated by the
outgoing steering council. For the initial election, the returns
officer will be nominated by the PSF Executive Director.
The council should ideally reflect the diversity of Python
contributors and users, and core team members are encouraged to vote
accordingly.
Term
A new council is elected after each feature release. Each council’s
term runs from when their election results are finalized until the
next council’s term starts. There are no term limits.
Vacancies
Council members may resign their position at any time.
Whenever there is a vacancy during the regular council term, the
council may vote to appoint a replacement to serve out the rest of the
term.
If a council member drops out of touch and cannot be contacted for a
month or longer, then the rest of the council may vote to replace
them.
Conflicts of interest
While we trust council members to act in the best interests of Python
rather than themselves or their employers, the mere appearance of any
one company dominating Python development could itself be harmful and
erode trust. In order to avoid any appearance of conflict of interest,
at most 2 members of the council can work for any single employer.
In a council election, if 3 of the top 5 vote-getters work for the
same employer, then whichever of them ranked lowest is disqualified
and the 6th-ranking candidate moves up into 5th place; this is
repeated until a valid council is formed.
During a council term, if changing circumstances cause this rule to be
broken (for instance, due to a council member changing employment),
then one or more council members must resign to remedy the issue, and
the resulting vacancies can then be filled as normal.
Ejecting core team members
In exceptional circumstances, it may be necessary to remove someone
from the core team against their will. (For example: egregious and
ongoing code of conduct violations.) This can be accomplished by a
steering council vote, but unlike other steering council votes, this
requires at least a two-thirds majority. With 5 members voting, this
means that a 3:2 vote is insufficient; 4:1 in favor is the minimum
required for such a vote to succeed. In addition, this is the one
power of the steering council which cannot be delegated, and this
power cannot be used while a vote of no confidence is in process.
If the ejected core team member is also on the steering council, then
they are removed from the steering council as well.
Vote of no confidence
In exceptional circumstances, the core team may remove a sitting
council member, or the entire council, via a vote of no confidence.
A no-confidence vote is triggered when a core team member calls for
one publicly on an appropriate project communication channel, and
another core team member seconds the proposal.
The vote lasts for two weeks. Core team members vote for or against.
If at least two thirds of voters express a lack of confidence, then
the vote succeeds.
There are two forms of no-confidence votes: those targeting a single
member, and those targeting the council as a whole. The initial call
for a no-confidence vote must specify which type is intended. If a
single-member vote succeeds, then that member is removed from the
council and the resulting vacancy can be handled in the usual way. If
a whole-council vote succeeds, the council is dissolved and a new
council election is triggered immediately.
The core team
Role
The core team is the group of trusted volunteers who manage Python.
They assume many roles required to achieve the project’s goals,
especially those that require a high level of trust. They make the
decisions that shape the future of the project.
Core team members are expected to act as role models for the community
and custodians of the project, on behalf of the community and all
those who rely on Python.
They will intervene, where necessary, in online discussions or at
official Python events on the rare occasions that a situation arises
that requires intervention.
They have authority over the Python Project infrastructure, including
the Python Project website itself, the Python GitHub organization and
repositories, the bug tracker, the mailing lists, IRC channels, etc.
Prerogatives
Core team members may participate in formal votes, typically to nominate new
team members and to elect the steering council.
Membership
Python core team members demonstrate:
a good grasp of the philosophy of the Python Project
a solid track record of being constructive and helpful
significant contributions to the project’s goals, in any form
willingness to dedicate some time to improving Python
As the project matures, contributions go beyond code. Here’s an
incomplete list of areas where contributions may be considered for
joining the core team, in no particular order:
Working on community management and outreach
Providing support on the mailing lists and on IRC
Triaging tickets
Writing patches (code, docs, or tests)
Reviewing patches (code, docs, or tests)
Participating in design decisions
Providing expertise in a particular domain (security, i18n, etc.)
Managing the continuous integration infrastructure
Managing the servers (website, tracker, documentation, etc.)
Maintaining related projects (alternative interpreters, core
infrastructure like packaging, etc.)
Creating visual designs
Core team membership acknowledges sustained and valuable efforts that
align well with the philosophy and the goals of the Python project.
It is granted by receiving at least two-thirds positive votes in a
core team vote that is open for one week and is not vetoed by the
steering council.
Core team members are always looking for promising contributors,
teaching them how the project is managed, and submitting their names
to the core team’s vote when they’re ready.
There’s no time limit on core team membership. However, in order to
provide the general public with a reasonable idea of how many people
maintain Python, core team members who have stopped contributing are
encouraged to declare themselves as “inactive”. Those who haven’t made
any non-trivial contribution in two years may be asked to move
themselves to this category, and moved there if they don’t respond. To
record and honor their contributions, inactive team members will
continue to be listed alongside active core team members; and, if they
later resume contributing, they can switch back to active status at
will. While someone is in inactive status, though, they lose their
active privileges like voting or nominating for the steering council,
and commit access.
The initial active core team members will consist of everyone
currently listed in the “Python core” team on GitHub (access
granted for core members only), and the
initial inactive members will consist of everyone else who has been a
committer in the past.
Changing this document
Changes to this document require at least a two-thirds majority of
votes cast in a core team vote which should be open for two weeks.
History
Creation of this document
The Python project was started by Guido van Rossum, who served as its
Benevolent Dictator for Life (BDFL) from inception until July 2018,
when he stepped down.
After discussion, a number of proposals were put forward for a new
governance model, and the core devs voted to choose between them. The
overall process is described in PEP 8000 and PEP 8001, a review of
other projects was performed in PEP 8002, and the proposals themselves
were written up as the 801x series of PEPs. Eventually the proposal in
PEP 8016 was selected
as the new governance model, and was used to create the initial
version of this PEP. The 8000-series PEPs are preserved for historical
reference (and in particular, PEP 8016 contains additional rationale
and links to contemporary discussions), but this PEP is now the
official reference, and will evolve following the rules described
herein.
History of council elections
January 2019: PEP 8100
December 2019: PEP 8101
December 2020: PEP 8102
December 2021: PEP 8103
December 2022: PEP 8104
December 2023: PEP 8105
History of amendments
2019-04-17: Added the vote length for core devs and changes to this document.
Acknowledgements
This PEP began as PEP 8016, which was written by Nathaniel J. Smith
and Donald Stufft, based on a Django governance document written by
Aymeric Augustin, and incorporated feedback and assistance from
numerous others.
Copyright
This document has been placed in the public domain.
| Active | PEP 13 – Python Language Governance | Process | This PEP defines the formal governance process for Python, and records
how this has changed over time. Currently, governance is based around
a steering council. The council has broad authority, which they seek
to exercise as rarely as possible. |
PEP 20 – The Zen of Python
Author:
Tim Peters <tim.peters at gmail.com>
Status:
Active
Type:
Informational
Created:
19-Aug-2004
Post-History:
22-Aug-2004
Table of Contents
Abstract
The Zen of Python
Easter Egg
References
Copyright
Abstract
Long time Pythoneer Tim Peters succinctly channels the BDFL’s guiding
principles for Python’s design into 20 aphorisms, only 19 of which
have been written down.
The Zen of Python
Beautiful is better than ugly.
Explicit is better than implicit.
Simple is better than complex.
Complex is better than complicated.
Flat is better than nested.
Sparse is better than dense.
Readability counts.
Special cases aren't special enough to break the rules.
Although practicality beats purity.
Errors should never pass silently.
Unless explicitly silenced.
In the face of ambiguity, refuse the temptation to guess.
There should be one-- and preferably only one --obvious way to do it.
Although that way may not be obvious at first unless you're Dutch.
Now is better than never.
Although never is often better than *right* now.
If the implementation is hard to explain, it's a bad idea.
If the implementation is easy to explain, it may be a good idea.
Namespaces are one honking great idea -- let's do more of those!
Easter Egg
>>> import this
References
Originally posted to comp.lang.python/[email protected] under a
thread called “The Way of Python”
Copyright
This document has been placed in the public domain.
| Active | PEP 20 – The Zen of Python | Informational | Long time Pythoneer Tim Peters succinctly channels the BDFL’s guiding
principles for Python’s design into 20 aphorisms, only 19 of which
have been written down. |
PEP 101 – Doing Python Releases 101
Author:
Barry Warsaw <barry at python.org>, Guido van Rossum <guido at python.org>
Status:
Active
Type:
Informational
Created:
22-Aug-2001
Post-History:
Replaces:
102
Table of Contents
Abstract
Things You’ll Need
Types of Releases
How To Make A Release
What Next?
Moving to End-of-life
Windows Notes
Copyright
Abstract
Making a Python release is a thrilling and crazy process. You’ve heard
the expression “herding cats”? Imagine trying to also saddle those
purring little creatures up, and ride them into town, with some of their
buddies firmly attached to your bare back, anchored by newly sharpened
claws. At least they’re cute, you remind yourself.
Actually, no, that’s a slight exaggeration 😉 The Python release
process has steadily improved over the years and now, with the help of our
amazing community, is really not too difficult. This PEP attempts to
collect, in one place, all the steps needed to make a Python release.
Most of the steps are now automated or guided by automation, so manually
following this list is no longer necessary.
Things You’ll Need
As a release manager there are a lot of resources you’ll need to access.
Here’s a hopefully-complete list.
A GPG key.Python releases are digitally signed with GPG; you’ll need a key,
which hopefully will be on the “web of trust” with at least one of
the other release managers.
A bunch of software:
A checkout of the python/release-tools repo.
It contains a requirements.txt file that you need to install
dependencies from first. Afterwards, you can fire up scripts in the
repo, covered later in this PEP.
blurb, the
Misc/NEWS
management tool. You can pip install it.
A fairly complete installation of a recent TeX distribution,
such as texlive. You need that for building the PDF docs.
Access to servers where you will upload files:
downloads.nyc1.psf.io, the server that hosts download files; and
docs.nyc1.psf.io, the server that hosts the documentation.
Administrator access to https://github.com/python/cpython.
An administrator account on www.python.org, including an “API key”.
Write access to the PEP repository.If you’re reading this, you probably already have this–the first
task of any release manager is to draft the release schedule. But
in case you just signed up… sucker! I mean, uh, congratulations!
Posting access to http://blog.python.org, a Blogger-hosted weblog.
The RSS feed from this blog is used for the ‘Python News’ section
on www.python.org.
A subscription to the super secret release manager mailing list, which may
or may not be called python-cabal. Bug Barry about this.
A @python.org email address that you will use to sign your releases
with. Ask postmaster@ for an address; you can either get a full
account, or a redirecting alias + SMTP credentials to send email from
this address that looks legit to major email providers.
Types of Releases
There are several types of releases you will need to make. These include:
alpha
begin beta, also known as beta 1, also known as new branch
beta 2+
release candidate 1
release candidate 2+
final
new branch
begin bugfix mode
begin security-only mode
end-of-life
Some of these release types actually involve more than
one release branch. In particular, a new branch is that point in the
release cycle when a new feature release cycle begins. Under the current
organization of the cpython git repository, the main branch is always
the target for new features. At some point in the release cycle of the
next feature release, a new branch release is made which creates a
new separate branch for stabilization and later maintenance of the
current in-progress feature release (3.n.0) and the main branch is modified
to build a new version (which will eventually be released as 3.n+1.0).
While the new branch release step could occur at one of several points
in the release cycle, current practice is for it to occur at feature code
cutoff for the release which is scheduled for the first beta release.
In the descriptions that follow, steps specific to release types are
labeled accordingly, for now, new branch and final.
How To Make A Release
Here are the steps taken to make a Python release. Some steps are more
fuzzy than others because there’s little that can be automated (e.g.
writing the NEWS entries). Where a step is usually performed by An
Expert, the role of that expert is given. Otherwise, assume the step is
done by the Release Manager (RM), the designated person performing the
release. The roles and their current experts are:
RM = Release Manager
Thomas Wouters <[email protected]> (NL)
Pablo Galindo Salgado <[email protected]> (UK)
Łukasz Langa <[email protected]> (PL)
WE = Windows - Steve Dower <[email protected]>
ME = Mac - Ned Deily <[email protected]> (US)
DE = Docs - Julien Palard <[email protected]> (Central Europe)
Note
It is highly recommended that the RM contact the Experts the day
before the release. Because the world is round and everyone lives
in different timezones, the RM must ensure that the release tag is
created in enough time for the Experts to cut binary releases.
You should not make the release public (by updating the website and
sending announcements) before all experts have updated their bits.
In rare cases where the expert for Windows or Mac is MIA, you may add
a message “(Platform) binaries will be provided shortly” and proceed.
As much as possible, the release steps are automated and guided by the
release script, which is available in a separate repository:
https://github.com/python/release-tools
We use the following conventions in the examples below. Where a release
number is given, it is of the form 3.X.YaN, e.g. 3.13.0a3 for Python 3.13.0
alpha 3, where “a” == alpha, “b” == beta, “rc” == release candidate.
Release tags are named v3.X.YaN. The branch name for minor release
maintenance branches is 3.X.
This helps by performing several automatic editing steps, and guides you
to perform some manual editing steps.
Log into Discord and join the Python Core Devs server. Ask Thomas
or Łukasz for an invite.You probably need to coordinate with other people around the world.
This communication channel is where we’ve arranged to meet.
Check to see if there are any showstopper bugs.Go to https://github.com/python/cpython/issues and look for any open
bugs that can block this release. You’re looking at two relevant labels:
release-blockerStops the release dead in its tracks. You may not
make any release with any open release blocker bugs.
deferred-blockerDoesn’t block this release, but it will block a
future release. You may not make a final or
candidate release with any open deferred blocker
bugs.
Review the release blockers and either resolve them, bump them down to
deferred, or stop the release and ask for community assistance. If
you’re making a final or candidate release, do the same with any open
deferred.
Check the stable buildbots.Go to https://buildbot.python.org/all/#/release_status
Look at the buildbots for the release
you’re making. Ignore any that are offline (or inform the community so
they can be restarted). If what remains are (mostly) green buildbots,
you’re good to go. If you have non-offline red buildbots, you may want
to hold up the release until they are fixed. Review the problems and
use your judgement, taking into account whether you are making an alpha,
beta, or final release.
Make a release clone.On a fork of the cpython repository on GitHub, create a release branch
within it (called the “release clone” from now on). You can use the same
GitHub fork you use for cpython development. Using the standard setup
recommended in the Python Developer’s Guide, your fork would be referred
to as origin and the standard cpython repo as upstream. You will
use the branch on your fork to do the release engineering work, including
tagging the release, and you will use it to share with the other experts
for making the binaries.
For a final or release candidate 2+ release, if you are going
to cherry-pick a subset of changes for the next rc or final from all those
merged since the last rc, you should create a release
engineering branch starting from the most recent release candidate tag,
i.e. v3.8.0rc1. You will then cherry-pick changes from the standard
release branch as necessary into the release engineering branch and
then proceed as usual. If you are going to take all of the changes
since the previous rc, you can proceed as normal.
Make sure the current branch of your release clone is the branch you
want to release from. (git status)
Run blurb release <version> specifying the version number
(e.g. blurb release 3.4.7rc1). This merges all the recent news
blurbs into a single file marked with this release’s version number.
Regenerate Lib/pydoc-topics.py.While still in the Doc directory, run make pydoc-topics. Then copy
build/pydoc-topics/topics.py to ../Lib/pydoc_data/topics.py.
Commit your changes to pydoc_topics.py
(and any fixes you made in the docs).
Consider running autoconf using the currently accepted standard version
in case configure or other autoconf-generated files were last
committed with a newer or older version and may contain spurious or
harmful differences. Currently, autoconf 2.71 is our de facto standard.
if there are differences, commit them.
Make sure the SOURCE_URI in Doc/tools/extensions/pyspecific.py
points to the right branch in the git repository (main or 3.X).
For a new branch release, change the branch in the file from main
to the new release branch you are about to create (3.X).
Bump version numbers via the release script:$ .../release-tools/release.py --bump 3.X.YaN
Reminder: X, Y, and N should be integers.
a should be one of “a”, “b”, or “rc” (e.g. “3.4.3rc1”).
For final releases omit the aN (“3.4.3”). For the first
release of a new version Y should be 0 (“3.6.0”).
This automates updating various release numbers, but you will have to
modify a few files manually. If your $EDITOR environment variable is
set up correctly, release.py will pop up editor windows with the files
you need to edit.
Review the blurb-generated Misc/NEWS file and edit as necessary.
Make sure all changes have been committed. (release.py --bump
doesn’t check in its changes for you.)
Check the years on the copyright notice. If the last release
was some time last year, add the current year to the copyright
notice in several places:
README
LICENSE (make sure to change on trunk and the branch)
Python/getcopyright.c
Doc/copyright.rst
Doc/license.rst
PC/python_ver_rc.h sets up the DLL version resource for Windows
(displayed when you right-click on the DLL and select
Properties). This isn’t a C include file, it’s a Windows
“resource file” include file.
For a final major release, edit the first paragraph of
Doc/whatsnew/3.X.rst to include the actual release date; e.g. “Python
2.5 was released on August 1, 2003.” There’s no need to edit this for
alpha or beta releases.
Do a “git status” in this directory.You should not see any files. I.e. you better not have any uncommitted
changes in your working directory.
Tag the release for 3.X.YaN:$ .../release-tools/release.py --tag 3.X.YaN
This executes a git tag command with the -s option so that the
release tag in the repo is signed with your gpg key. When prompted
choose the private key you use for signing release tarballs etc.
For begin security-only mode and end-of-life releases, review the
two files and update the versions accordingly in all active branches.
Time to build the source tarball. Use the release script to create
the source gzip and xz tarballs,
documentation tar and zip files, and gpg signature files:$ .../release-tools/release.py --export 3.X.YaN
This can take a while for final releases, and it will leave all the
tarballs and signatures in a subdirectory called 3.X.YaN/src, and the
built docs in 3.X.YaN/docs (for final releases).
Note that the script will sign your release with Sigstore. Please use
your @python.org email address for this. See here for more information:
https://www.python.org/download/sigstore/.
Now you want to perform the very important step of checking the
tarball you just created, to make sure a completely clean,
virgin build passes the regression test. Here are the best
steps to take:$ cd /tmp
$ tar xvf /path/to/your/release/clone/<version>//Python-3.2rc2.tgz
$ cd Python-3.2rc2
$ ls
(Do things look reasonable?)
$ ls Lib
(Are there stray .pyc files?)
$ ./configure
(Loads of configure output)
$ make test
(Do all the expected tests pass?)
If you’re feeling lucky and have some time to kill, or if you are making
a release candidate or final release, run the full test suite:
$ make testall
If the tests pass, then you can feel good that the tarball is
fine. If some of the tests fail, or anything else about the
freshly unpacked directory looks weird, you better stop now and
figure out what the problem is.
Push your commits to the remote release branch in your GitHub fork.:# Do a dry run first.
$ git push --dry-run --tags origin
# Make sure you are pushing to your GitHub fork, *not* to the main
# python/cpython repo!
$ git push --tags origin
Notify the experts that they can start building binaries.
Warning
STOP: at this point you must receive the “green light” from other experts
in order to create the release. There are things you can do while you wait
though, so keep reading until you hit the next STOP.
The WE generates and publishes the Windows files using the Azure
Pipelines build scripts in .azure-pipelines/windows-release/,
currently set up at https://dev.azure.com/Python/cpython/_build?definitionId=21.The build process runs in multiple stages, with each stage’s output being
available as a downloadable artifact. The stages are:
Compile all variants of binaries (32-bit, 64-bit, debug/release),
including running profile-guided optimization.
Compile the HTML Help file containing the Python documentation
Codesign all the binaries with the PSF’s certificate
Create packages for python.org, nuget.org, the embeddable distro and
the Windows Store
Perform basic verification of the installers
Upload packages to python.org and nuget.org, purge download caches and
run a test download.
After the uploads are complete, the WE copies the generated hashes from
the build logs and emails them to the RM. The Windows Store packages are
uploaded manually to https://partner.microsoft.com/dashboard/home by the
WE.
The ME builds Mac installer packages and uploads them to
downloads.nyc1.psf.io together with gpg signature files.
scp or rsync all the files built by release.py --export
to your home directory on downloads.nyc1.psf.io.While you’re waiting for the files to finish uploading, you can continue
on with the remaining tasks. You can also ask folks on #python-dev
and/or python-committers to download the files as they finish uploading
so that they can test them on their platforms as well.
Now you need to go to downloads.nyc1.psf.io and move all the files in place
over there. Our policy is that every Python version gets its own
directory, but each directory contains all releases of that version.
On downloads.nyc1.psf.io, cd /srv/www.python.org/ftp/python/3.X.Y
creating it if necessary. Make sure it is owned by group ‘downloads’
and group-writable.
Move the release .tgz, and .tar.xz files into place, as well as the
.asc GPG signature files. The Win/Mac binaries are usually put there
by the experts themselves.Make sure they are world readable. They should also be group
writable, and group-owned by downloads.
Use gpg --verify to make sure they got uploaded intact.
If this is a final or rc release: Move the doc zips and tarballs to
/srv/www.python.org/ftp/python/doc/3.X.Y[rcA], creating the directory
if necessary, and adapt the “current” symlink in .../doc to point to
that directory. Note though that if you’re releasing a maintenance
release for an older version, don’t change the current link.
If this is a final or rc release (even a maintenance release), also
unpack the HTML docs to /srv/docs.python.org/release/3.X.Y[rcA] on
docs.nyc1.psf.io. Make sure the files are in group docs and are
group-writeable.
Let the DE check if the docs are built and work all right.
Note both the documentation and downloads are behind a caching CDN. If
you change archives after downloading them through the website, you’ll
need to purge the stale data in the CDN like this:$ curl -X PURGE https://www.python.org/ftp/python/3.12.0/Python-3.12.0.tar.xz
You should always purge the cache of the directory listing as people
use that to browse the release files:
$ curl -X PURGE https://www.python.org/ftp/python/3.12.0/
For the extra paranoid, do a completely clean test of the release.
This includes downloading the tarball from www.python.org.Make sure the md5 checksums match. Then unpack the tarball,
and do a clean make test.:
$ make distclean
$ ./configure
$ make test
To ensure that the regression test suite passes. If not, you
screwed up somewhere!
Warning
STOP and confirm:
Have you gotten the green light from the WE?
Have you gotten the green light from the ME?
Have you gotten the green light from the DE?
If green, it’s time to merge the release engineering branch back into
the main repo.
In order to push your changes to GitHub, you’ll have to temporarily
disable branch protection for administrators. Go to the
Settings | Branches page:https://github.com/python/cpython/settings/branches/
“Edit” the settings for the branch you’re releasing on.
This will load the settings page for that branch.
Uncheck the “Include administrators” box and press the
“Save changes” button at the bottom.
Merge your release clone into the main development repo:# Pristine copy of the upstream repo branch
$ git clone [email protected]:python/cpython.git merge
$ cd merge
# Checkout the correct branch:
# 1. For feature pre-releases up to and including a
# **new branch** release, i.e. alphas and first beta
# do a checkout of the main branch
$ git checkout main
# 2. Else, for all other releases, checkout the
# appropriate release branch.
$ git checkout 3.X
# Fetch the newly created and signed tag from your clone repo
$ git fetch --tags [email protected]:your-github-id/cpython.git v3.X.YaN
# Merge the temporary release engineering branch back into
$ git merge --no-squash v3.X.YaN
$ git commit -m 'Merge release engineering branch'
If this is a new branch release, i.e. first beta,
now create the new release branch:$ git checkout -b 3.X
Do any steps needed to setup the new release branch, including:
In README.rst, change all references from main to
the new branch, in particular, GitHub repo URLs.
For all releases, do the guided post-release steps with the
release script.:$ .../release-tools/release.py --done 3.X.YaN
For a final or release candidate 2+ release, you may need to
do some post-merge cleanup. Check the top-level README.rst
and include/patchlevel.h files to ensure they now reflect
the desired post-release values for on-going development.
The patchlevel should be the release tag with a +.
Also, if you cherry-picked changes from the standard release
branch into the release engineering branch for this release,
you will now need to manual remove each blurb entry from
the Misc/NEWS.d/next directory that was cherry-picked
into the release you are working on since that blurb entry
is now captured in the merged x.y.z.rst file for the new
release. Otherwise, the blurb entry will appear twice in
the changelog.html file, once under Python next and again
under x.y.z.
Review and commit these changes:$ git commit -m 'Post release updates'
If this is a new branch release (e.g. the first beta),
update the main branch to start development for the
following feature release. When finished, the main
branch will now build Python X.Y+1.
First, set main up to be the next release, i.e.X.Y+1.a0:$ git checkout main
$ .../release-tools/release.py --bump 3.9.0a0
Edit all version references in README.rst
Move any historical “what’s new” entries from Misc/NEWS to
Misc/HISTORY.
Edit Doc/tutorial/interpreter.rst (2 references to ‘[Pp]ython3x’,
one to ‘Python 3.x’, also make the date in the banner consistent).
Edit Doc/tutorial/stdlib.rst and Doc/tutorial/stdlib2.rst, which
have each one reference to ‘[Pp]ython3x’.
Add a new whatsnew/3.x.rst file (with the comment near the top
and the toplevel sections copied from the previous file) and
add it to the toctree in whatsnew/index.rst. But beware that
the initial whatsnew/3.x.rst checkin from previous releases
may be incorrect due to the initial midstream change to blurb
that propagates from release to release! Help break the cycle: if
necessary make the following change:- For full details, see the :source:`Misc/NEWS` file.
+ For full details, see the :ref:`changelog <changelog>`.
Update the version number in configure.ac and re-run autoconf.
Make sure the SOURCE_URI in Doc/tools/extensions/pyspecific.py
points to main.
Update the version numbers for the Windows builds in PC/ and
PCbuild/, which have references to python38.
NOTE, check with Steve Dower about this step, it is probably obsolete.:$ find PC/ PCbuild/ -type f | xargs sed -i 's/python38/python39/g'
$ git mv -f PC/os2emx/python38.def PC/os2emx/python39.def
$ git mv -f PC/python38stub.def PC/python39stub.def
$ git mv -f PC/python38gen.py PC/python39gen.py
Commit these changes to the main branch:$ git status
$ git add ...
$ git commit -m 'Bump to 3.9.0a0'
Do another git status in this directory.You should not see any files. I.e. you better not have any uncommitted
changes in your working directory.
Commit and push to the main repo.:# Do a dry run first.
# For feature pre-releases prior to a **new branch** release,
# i.e. a feature alpha release:
$ git push --dry-run --tags [email protected]:python/cpython.git main
# If it looks OK, take the plunge. There's no going back!
$ git push --tags [email protected]:python/cpython.git main
# For a **new branch** release, i.e. first beta:
$ git push --dry-run --tags [email protected]:python/cpython.git 3.X
$ git push --dry-run --tags [email protected]:python/cpython.git main
# If it looks OK, take the plunge. There's no going back!
$ git push --tags [email protected]:python/cpython.git 3.X
$ git push --tags [email protected]:python/cpython.git main
# For all other releases:
$ git push --dry-run --tags [email protected]:python/cpython.git 3.X
# If it looks OK, take the plunge. There's no going back!
$ git push --tags [email protected]:python/cpython.git 3.X
If this is a new branch release, add a Branch protection rule
for the newly created branch (3.X). Look at the values for the previous
release branch (3.X-1) and use them as a template.
https://github.com/python/cpython/settings/branches/Also, add a needs backport to 3.X label to the GitHub repo.
https://github.com/python/cpython/labels
You can now re-enable enforcement of branch settings against administrators
on GitHub. Go back to the Settings | Branch page:https://github.com/python/cpython/settings/branches/
“Edit” the settings for the branch you’re releasing on.
Re-check the “Include administrators” box and press the
“Save changes” button at the bottom.
Now it’s time to twiddle the web site. Almost none of this is automated, sorry.
To do these steps, you must have the permission to edit the website. If you
don’t have that, ask someone on [email protected] for the proper
permissions. (Or ask Ewa, who coordinated the effort for the new website
with RevSys.)
Log in to https://www.python.org/admin .
Create a new “release” for the release. Currently “Releases” are
sorted under “Downloads”.The easiest thing is probably to copy fields from an existing
Python release “page”, editing as you go.
You can use Markdown or
ReStructured Text
to describe your release. The former is less verbose, while the latter has nifty
integration for things like referencing PEPs.
Leave the “Release page” field on the form empty.
“Save” the release.
Populate the release with the downloadable files.Your friend and mine, Georg Brandl, made a lovely tool
called “add-to-pydotorg.py”. You can find it in the
“release” tree (next to “release.py”). You run the
tool on downloads.nyc1.psf.io, like this:
$ AUTH_INFO=<username>:<python.org-api-key> python add-to-pydotorg.py <version>
This walks the correct download directory for <version>,
looks for files marked with <version>, and populates
the “Release Files” for the correct “release” on the web
site with these files. Note that clears the “Release Files”
for the relevant version each time it’s run. You may run
it from any directory you like, and you can run it as
many times as you like if the files happen to change.
Keep a copy in your home directory on dl-files and
keep it fresh.
If new types of files are added to the release, someone will need to
update add-to-pydotorg.py so it recognizes these new files.
(It’s best to update add-to-pydotorg.py when file types
are removed, too.)
The script will also sign any remaining files that were not
signed with Sigstore until this point. Again, if this happens,
do use your @python.org address for this process. More info:
https://www.python.org/download/sigstore/
In case the CDN already cached a version of the Downloads page
without the files present, you can invalidate the cache using:$ curl -X PURGE https://www.python.org/downloads/release/python-XXX/
If this is a final release:
Add the new version to the Python Documentation by Version
page https://www.python.org/doc/versions/ and
remove the current version from any ‘in development’ section.
For 3.X.Y, edit all the previous X.Y releases’ page(s) to
point to the new release. This includes the content field of the
Downloads -> Releases entry for the release:Note: Python 3.x.(y-1) has been superseded by
`Python 3.x.y </downloads/release/python-3xy/>`_.
And, for those releases having separate release page entries
(phasing these out?), update those pages as well,
e.g. download/releases/3.x.y:
Note: Python 3.x.(y-1) has been superseded by
`Python 3.x.y </download/releases/3.x.y/>`_.
Update the “Current Pre-release Testing Versions web page”.There’s a page that lists all the currently-in-testing versions
of Python:
https://www.python.org/download/pre-releases/
Every time you make a release, one way or another you’ll
have to update this page:
If you’re releasing a version before 3.x.0,
you should add it to this page, removing the previous pre-release
of version 3.x as needed.
If you’re releasing 3.x.0 final, you need to remove the pre-release
version from this page.
This is in the “Pages” category on the Django-based website, and finding
it through that UI is kind of a chore. However! If you’re already logged
in to the admin interface (which, at this point, you should be), Django
will helpfully add a convenient “Edit this page” link to the top of the
page itself. So you can simply follow the link above, click on the
“Edit this page” link, and make your changes as needed. How convenient!
If appropriate, update the “Python Documentation by Version” page:
https://www.python.org/doc/versions/
This lists all releases of Python by version number and links to their
static (not built daily) online documentation. There’s a list at the
bottom of in-development versions, which is where all alphas/betas/RCs
should go. And yes you should be able to click on the link above then
press the shiny, exciting “Edit this page” button.
Write the announcement on https://discuss.python.org/. This is the
fuzzy bit because not much can be automated. You can use an earlier
announcement as a template, but edit it for content!
Once the announcement is up on Discourse, send an equivalent to the
following mailing lists:[email protected]
[email protected]
[email protected]
Also post the announcement to
The Python Insider blog.
To add a new entry, go to
your Blogger home page, here.
Update any release PEPs (e.g. 719) with the release dates.
Update the labels on https://github.com/python/cpython/issues:
Flip all the deferred-blocker issues back to release-blocker
for the next release.
Add version 3.X+1 as when version 3.X enters alpha.
Change non-doc feature requests to version 3.X+1 when version 3.X
enters beta.
Update issues from versions that your release makes
unsupported to the next supported version.
Review open issues, as this might find lurking showstopper bugs,
besides reminding people to fix the easy ones they forgot about.
You can delete the remote release clone branch from your repo clone.
If this is a new branch release, you will need to ensure various
pieces of the development infrastructure are updated for the new branch.
These include:
Update the issue tracker for the new branch: add the new version to
the versions list.
Update the devguide to reflect the new branches and versions.
Create a PR to update the supported releases table on the
downloads page.
(See https://github.com/python/pythondotorg/issues/1302)
Ensure buildbots are defined for the new branch (contact Łukasz
or Zach Ware).
Ensure the various GitHub bots are updated, as needed, for the
new branch, in particular, make sure backporting to the new
branch works (contact core-workflow team)
https://github.com/python/core-workflow/issues
Review the most recent commit history for the main and new release
branches to identify and backport any merges that might have been made
to the main branch during the release engineering phase and that
should be in the release branch.
Verify that CI is working for new PRs for the main and new release
branches and that the release branch is properly protected (no direct
pushes, etc).
Verify that the on-line docs are building properly (this may take up to
24 hours for a complete build on the web site).
What Next?
Verify! Pretend you’re a user: download the files from python.org, and
make Python from it. This step is too easy to overlook, and on several
occasions we’ve had useless release files. Once a general server problem
caused mysterious corruption of all files; once the source tarball got
built incorrectly; more than once the file upload process on SF truncated
files; and so on.
Rejoice. Drink. Be Merry. Write a PEP like this one. Or be
like unto Guido and take A Vacation.
You’ve just made a Python release!
Moving to End-of-life
Under current policy, a release branch normally reaches end-of-life status
5 years after its initial release. The policy is discussed in more detail
in the Python Developer’s Guide.
When end-of-life is reached, there are a number of tasks that need to be
performed either directly by you as release manager or by ensuring someone
else does them. Some of those tasks include:
Optionally making a final release to publish any remaining unreleased
changes.
Freeze the state of the release branch by creating a tag of its current HEAD
and then deleting the branch from the cpython repo. The current HEAD should
be at or beyond the final security release for the branch:git fetch upstream
git tag --sign -m 'Final head of the former 3.3 branch' 3.3 upstream/3.3
git push upstream refs/tags/3.3
If all looks good, delete the branch. This may require the assistance of
someone with repo administrator privileges:git push upstream --delete 3.3 # or perform from GitHub Settings page
Remove the release from the list of “Active Python Releases” on the Downloads
page. To do this, log in to the admin page for python.org, navigate to Boxes,
and edit the downloads-active-releases entry. Simply strip out the relevant
paragraph of HTML for your release. (You’ll probably have to do the curl -X PURGE
trick to purge the cache if you want to confirm you made the change correctly.)
Add retired notice to each release page on python.org for the retired branch.
For example:
https://www.python.org/downloads/release/python-337/https://www.python.org/downloads/release/python-336/
In the developer’s guide, add the branch to the recent end-of-life branches
list (https://devguide.python.org/devcycle/#end-of-life-branches) and update
or remove references to the branch elsewhere in the devguide.
Retire the release from the issue tracker. Tasks include:
remove version label from list of versions
remove the “needs backport to” label for the retired version
review and dispose of open issues marked for this branch
Announce the branch retirement in the usual places:
discuss.python.org
mailing lists (python-dev, python-list, python-announcements)
Python Dev blog
Enjoy your retirement and bask in the glow of a job well done!
Windows Notes
Windows has a MSI installer, various flavors of Windows have
“special limitations”, and the Windows installer also packs
precompiled “foreign” binaries (Tcl/Tk, expat, etc).
The installer is tested as part of the Azure Pipeline. In the past,
those steps were performed manually. We’re keeping this for posterity.
Concurrent with uploading the installer, the WE installs Python
from it twice: once into the default directory suggested by the
installer, and later into a directory with embedded spaces in its
name. For each installation, the WE runs the full regression suite
from a DOS box, and both with and without -0. For maintenance
release, the WE also tests whether upgrade installations succeed.
The WE also tries every shortcut created under Start -> Menu -> the
Python group. When trying IDLE this way, you need to verify that
Help -> Python Documentation works. When trying pydoc this way
(the “Module Docs” Start menu entry), make sure the “Start
Browser” button works, and make sure you can search for a random
module (like “random” <wink>) and then that the “go to selected”
button works.
It’s amazing how much can go wrong here – and even more amazing
how often last-second checkins break one of these things. If
you’re “the Windows geek”, keep in mind that you’re likely the
only person routinely testing on Windows, and that Windows is
simply a mess.
Repeat the testing for each target architecture. Try both an
Admin and a plain User (not Power User) account.
Copyright
This document has been placed in the public domain.
| Active | PEP 101 – Doing Python Releases 101 | Informational | Making a Python release is a thrilling and crazy process. You’ve heard
the expression “herding cats”? Imagine trying to also saddle those
purring little creatures up, and ride them into town, with some of their
buddies firmly attached to your bare back, anchored by newly sharpened
claws. At least they’re cute, you remind yourself. |
PEP 102 – Doing Python Micro Releases
Author:
Anthony Baxter <anthony at interlink.com.au>,
Barry Warsaw <barry at python.org>,
Guido van Rossum <guido at python.org>
Status:
Superseded
Type:
Informational
Created:
09-Jan-2002
Post-History:
Superseded-By:
101
Table of Contents
Replacement Note
Abstract
How to Make A Release
What Next?
Final Release Notes
Windows Notes
Copyright
Replacement Note
Although the size of the to-do list in this PEP is much less scary
than that in PEP 101, it turns out not to be enough justification
for the duplication of information, and with it, the danger of one
of the copies to become out of date. Therefore, this PEP is not
maintained anymore, and micro releases are fully covered by PEP 101.
Abstract
Making a Python release is an arduous process that takes a
minimum of half a day’s work even for an experienced releaser.
Until recently, most – if not all – of that burden was borne by
Guido himself. But several recent releases have been performed by
other folks, so this PEP attempts to collect, in one place, all
the steps needed to make a Python bugfix release.
The major Python release process is covered in PEP 101 - this PEP
is just PEP 101, trimmed down to only include the bits that are
relevant for micro releases, a.k.a. patch, or bug fix releases.
It is organized as a recipe and you can actually print this out and
check items off as you complete them.
How to Make A Release
Here are the steps taken to make a Python release. Some steps are
more fuzzy than others because there’s little that can be
automated (e.g. writing the NEWS entries). Where a step is
usually performed by An Expert, the name of that expert is given.
Otherwise, assume the step is done by the Release Manager (RM),
the designated person performing the release. Almost every place
the RM is mentioned below, this step can also be done by the BDFL
of course!
XXX: We should include a dependency graph to illustrate the steps
that can be taken in parallel, or those that depend on other
steps.
We use the following conventions in the examples below. Where a
release number is given, it is of the form X.Y.MaA, e.g. 2.1.2c1
for Python 2.1.2 release candidate 1, where “a” == alpha, “b” ==
beta, “c” == release candidate. Final releases are tagged with
“releaseXYZ” in CVS. The micro releases are made from the
maintenance branch of the major release, e.g. Python 2.1.2 is made
from the release21-maint branch.
Send an email to [email protected] indicating the release is
about to start.
Put a freeze on check ins into the maintenance branch. At this
point, nobody except the RM should make any commits to the branch
(or his duly assigned agents, i.e. Guido the BDFL, Fred Drake for
documentation, or Thomas Heller for Windows). If the RM screwed up
and some desperate last minute change to the branch is
necessary, it can mean extra work for Fred and Thomas. So try to
avoid this!
On the branch, change Include/patchlevel.h in two places, to
reflect the new version number you’ve just created. You’ll want
to change the PY_VERSION macro, and one or several of the
version subpart macros just above PY_VERSION, as appropriate.
Change the “%define version” line of Misc/RPM/python-2.3.spec to the
same string as PY_VERSION was changed to above. E.g:%define version 2.3.1
You also probably want to reset the %define release line
to ‘1pydotorg’ if it’s not already that.
If you’re changing the version number for Python (e.g. from
Python 2.1.1 to Python 2.1.2), you also need to update the
README file, which has a big banner at the top proclaiming its
identity. Don’t do this if you’re just releasing a new alpha or
beta release, but /do/ do this if you’re release a new micro,
minor or major release.
The LICENSE file also needs to be changed, due to several
references to the release number. As for the README file, changing
these are necessary for a new micro, minor or major release.The LICENSE file contains a table that describes the legal
heritage of Python; you should add an entry for the X.Y.Z
release you are now making. You should update this table in the
LICENSE file on the CVS trunk too.
When the year changes, copyright legends need to be updated in
many places, including the README and LICENSE files.
For the Windows build, additional files have to be updated.PCbuild/BUILDno.txt contains the Windows build number, see the
instructions in this file how to change it. Saving the project
file PCbuild/pythoncore.dsp results in a change to
PCbuild/pythoncore.dsp as well.
PCbuild/python20.wse sets up the Windows installer version
resource (displayed when you right-click on the installer .exe
and select Properties), and also contains the Python version
number.
(Before version 2.3.2, it was required to manually edit
PC/python_nt.rc, this step is now automated by the build
process.)
After starting the process, the most important thing to do next
is to update the Misc/NEWS file. Thomas will need this in order to
do the Windows release and he likes to stay up late. This step
can be pretty tedious, so it’s best to get to it immediately
after making the branch, or even before you’ve made the branch.
The sooner the better (but again, watch for new checkins up
until the release is made!)Add high level items new to this release. E.g. if we’re
releasing 2.2a3, there must be a section at the top of the file
explaining “What’s new in Python 2.2a3”. It will be followed by
a section entitled “What’s new in Python 2.2a2”.
Note that you /hope/ that as developers add new features to the
trunk, they’ve updated the NEWS file accordingly. You can’t be
positive, so double check. If you’re a Unix weenie, it helps to
verify with Thomas about changes on Windows, and Jack Jansen
about changes on the Mac.
This command should help you (but substitute the correct -r tag!):
% cvs log -rr22a1: | python Tools/scripts/logmerge.py > /tmp/news.txt
IOW, you’re printing out all the cvs log entries from the
previous release until now. You can then troll through the
news.txt file looking for interesting things to add to NEWS.
Check your NEWS changes into the maintenance branch. It’s easy
to forget to update the release date in this file!
Check in any changes to IDLE’s NEWS.txt. Update the header in
Lib/idlelib/NEWS.txt to reflect its release version and date.
Update the IDLE version in Lib/idlelib/idlever.py to match.
Once the release process has started, the documentation needs to
be built and posted on python.org according to the instructions
in PEP 101.Note that Fred is responsible both for merging doc changes from
the trunk to the branch AND for merging any branch changes from
the branch to the trunk during the cleaning up phase.
Basically, if it’s in Doc/ Fred will take care of it.
Thomas compiles everything with MSVC 6.0 SP5, and moves the
python23.chm file into the src/chm directory. The installer
executable is then generated with Wise Installation System.The installer includes the MSVC 6.0 runtime in the files
MSVCRT.DLL and MSVCIRT.DLL. It leads to disaster if these files
are taken from the system directory of the machine where the
installer is built, instead it must be absolutely made sure that
these files come from the VCREDIST.EXE redistributable package
contained in the MSVC SP5 CD. VCREDIST.EXE must be unpacked
with winzip, and the Wise Installation System prompts for the
directory.
After building the installer, it should be opened with winzip,
and the MS dlls extracted again and check for the same version
number as those unpacked from VCREDIST.EXE.
Thomas uploads this file to the starship. He then sends the RM
a notice which includes the location and MD5 checksum of the
Windows executable.
Note that Thomas’s creation of the Windows executable may generate
a few more commits on the branch. Thomas will be responsible for
merging Windows-specific changes from trunk to branch, and from
branch to trunk.
Sean performs his Red Hat magic, generating a set of RPMs. He
uploads these files to python.org. He then sends the RM a notice
which includes the location and MD5 checksum of the RPMs.
It’s Build Time!Now, you’re ready to build the source tarball. First cd to your
working directory for the branch. E.g.
% cd …/python-22a3
Do a “cvs update” in this directory. Do NOT include the -A flag!You should not see any “M” files, but you may see several “P”
and/or “U” files. I.e. you better not have any uncommitted
changes in your working directory, but you may pick up some of
Fred’s or Thomas’s last minute changes.
Now tag the branch using a symbolic name like “rXYMaZ”,
e.g. r212% cvs tag r212
Be sure to tag only the python/dist/src subdirectory of the
Python CVS tree!
Change to a neutral directory, i.e. one in which you can do a
fresh, virgin, cvs export of the branch. You will be creating a
new directory at this location, to be named “Python-X.Y.M”. Do
a CVS export of the tagged branch.% cd ~
% cvs -d cvs.sf.net:/cvsroot/python export -rr212 \
-d Python-2.1.2 python/dist/src
Generate the tarball. Note that we’re not using the ‘z’ option
on the tar command because 1) that’s only supported by GNU tar
as far as we know, and 2) we’re going to max out the compression
level, which isn’t a supported option. We generate both tar.gz
tar.bz2 formats, as the latter is about 1/6th smaller.% tar -cf - Python-2.1.2 | gzip -9 > Python-2.1.2.tgz
% tar -cf - Python-2.1.2 | bzip2 -9 > Python-2.1.2.tar.bz2
Calculate the MD5 checksum of the tgz and tar.bz2 files you
just created% md5sum Python-2.1.2.tgz
Note that if you don’t have the md5sum program, there is a
Python replacement in the Tools/scripts/md5sum.py file.
Create GPG keys for each of the files.% gpg -ba Python-2.1.2.tgz
% gpg -ba Python-2.1.2.tar.bz2
% gpg -ba Python-2.1.2.exe
Now you want to perform the very important step of checking the
tarball you just created, to make sure a completely clean,
virgin build passes the regression test. Here are the best
steps to take:% cd /tmp
% tar zxvf ~/Python-2.1.2.tgz
% cd Python-2.1.2
% ls
(Do things look reasonable?)
% ./configure
(Loads of configure output)
% make test
(Do all the expected tests pass?)
If the tests pass, then you can feel good that the tarball is
fine. If some of the tests fail, or anything else about the
freshly unpacked directory looks weird, you better stop now and
figure out what the problem is.
You need to upload the tgz and the exe file to creosote.python.org.
This step can take a long time depending on your network
bandwidth. scp both files from your own machine to creosote.
While you’re waiting, you can start twiddling the web pages to
include the announcement.
In the top of the python.org web site CVS tree, create a
subdirectory for the X.Y.Z release. You can actually copy an
earlier patch release’s subdirectory, but be sure to delete
the X.Y.Z/CVS directory and “cvs add X.Y.Z”, for example:% cd .../pydotorg
% cp -r 2.2.2 2.2.3
% rm -rf 2.2.3/CVS
% cvs add 2.2.3
% cd 2.2.3
Edit the files for content: usually you can globally replace
X.Ya(Z-1) with X.YaZ. However, you’ll need to think about the
“What’s New?” section.
Copy the Misc/NEWS file to NEWS.txt in the X.Y.Z directory for
python.org; this contains the “full scoop” of changes to
Python since the previous release for this version of Python.
Copy the .asc GPG signatures you created earlier here as well.
Also, update the MD5 checksums.
Preview the web page by doing a “make” or “make install” (as
long as you’ve created a new directory for this release!)
Similarly, edit the ../index.ht file, i.e. the python.org home
page. In the Big Blue Announcement Block, move the paragraph
for the new version up to the top and boldify the phrase
“Python X.YaZ is out”. Edit for content, and preview locally,
but do NOT do a “make install” yet!
Now we’re waiting for the scp to creosote to finish. Da de da,
da de dum, hmm, hmm, dum de dum.
Once that’s done you need to go to creosote.python.org and move
all the files in place over there. Our policy is that every
Python version gets its own directory, but each directory may
contain several releases. We keep all old releases, moving them
into a “prev” subdirectory when we have a new release.So, there’s a directory called “2.2” which contains
Python-2.2a2.exe and Python-2.2a2.tgz, along with a “prev”
subdirectory containing Python-2.2a1.exe and Python-2.2a1.tgz.
So…
On creosote, cd to ~ftp/pub/python/X.Y creating it if
necessary.
Move the previous release files to a directory called “prev”
creating the directory if necessary (make sure the directory
has g+ws bits on). If this is the first alpha release of a
new Python version, skip this step.
Move the .tgz file and the .exe file to this directory. Make
sure they are world readable. They should also be group
writable, and group-owned by webmaster.
md5sum the files and make sure they got uploaded intact.
the X.Y/bugs.ht file if necessary. It is best to get
BDFL input for this step.
Go up to the parent directory (i.e. the root of the web page
hierarchy) and do a “make install” there. You’re release is now
live!
Now it’s time to write the announcement for the mailing lists.
This is the fuzzy bit because not much can be automated. You
can use one of Guido’s earlier announcements as a template, but
please edit it for content!Once the announcement is ready, send it to the following
addresses:
[email protected]
[email protected]
[email protected]
Send a SourceForge News Item about the release. From the
project’s “menu bar”, select the “News” link; once in News,
select the “Submit” link. Type a suitable subject (e.g. “Python
2.2c1 released” :-) in the Subject box, add some text to the
Details box (at the very least including the release URL at
www.python.org and the fact that you’re happy with the release)
and click the SUBMIT button.Feel free to remove any old news items.
Now it’s time to do some cleanup. These steps are very important!
Edit the file Include/patchlevel.h so that the PY_VERSION
string says something like “X.YaZ+”. Note the trailing ‘+’
indicating that the trunk is going to be moving forward with
development. E.g. the line should look like:#define PY_VERSION "2.1.2+"
Make sure that the other PY_ version macros contain the
correct values. Commit this change.
For the extra paranoid, do a completely clean test of the
release. This includes downloading the tarball from
www.python.org.
Make sure the md5 checksums match. Then unpack the tarball,
and do a clean make test.% make distclean
% ./configure
% make test
To ensure that the regression test suite passes. If not, you
screwed up somewhere!
Step 5 …
Verify! This can be interleaved with Step 4. Pretend you’re a
user: download the files from python.org, and make Python from
it. This step is too easy to overlook, and on several occasions
we’ve had useless release files. Once a general server problem
caused mysterious corruption of all files; once the source tarball
got built incorrectly; more than once the file upload process on
SF truncated files; and so on.
What Next?
Rejoice. Drink. Be Merry. Write a PEP like this one. Or be
like unto Guido and take A Vacation.
You’ve just made a Python release!
Actually, there is one more step. You should turn over ownership
of the branch to Jack Jansen. All this means is that now he will
be responsible for making commits to the branch. He’s going to
use this to build the MacOS versions. He may send you information
about the Mac release that should be merged into the informational
pages on www.python.org. When he’s done, he’ll tag the branch
something like “rX.YaZ-mac”. He’ll also be responsible for
merging any Mac-related changes back into the trunk.
Final Release Notes
The Final release of any major release, e.g. Python 2.2 final, has
special requirements, specifically because it will be one of the
longest lived releases (i.e. betas don’t last more than a couple
of weeks, but final releases can last for years!).
For this reason we want to have a higher coordination between the
three major releases: Windows, Mac, and source. The Windows and
source releases benefit from the close proximity of the respective
release-bots. But the Mac-bot, Jack Jansen, is 6 hours away. So
we add this extra step to the release process for a final
release:
Hold up the final release until Jack approves, or until we
lose patience <wink>.
The python.org site also needs some tweaking when a new bugfix release
is issued.
The documentation should be installed at doc/<version>/.
Add a link from doc/<previous-minor-release>/index.ht to the
documentation for the new version.
All older doc/<old-release>/index.ht files should be updated to
point to the documentation for the new version.
/robots.txt should be modified to prevent the old version’s
documentation from being crawled by search engines.
Windows Notes
Windows has a GUI installer, various flavors of Windows have
“special limitations”, and the Windows installer also packs
precompiled “foreign” binaries (Tcl/Tk, expat, etc). So Windows
testing is tiresome but very necessary.
Concurrent with uploading the installer, Thomas installs Python
from it twice: once into the default directory suggested by the
installer, and later into a directory with embedded spaces in its
name. For each installation, he runs the full regression suite
from a DOS box, and both with and without -0.
He also tries every shortcut created under Start -> Menu -> the
Python group. When trying IDLE this way, you need to verify that
Help -> Python Documentation works. When trying pydoc this way
(the “Module Docs” Start menu entry), make sure the “Start
Browser” button works, and make sure you can search for a random
module (Thomas uses “random” <wink>) and then that the “go to
selected” button works.
It’s amazing how much can go wrong here – and even more amazing
how often last-second checkins break one of these things. If
you’re “the Windows geek”, keep in mind that you’re likely the
only person routinely testing on Windows, and that Windows is
simply a mess.
Repeat all of the above on at least one flavor of Win9x, and one
of NT/2000/XP. On NT/2000/XP, try both an Admin and a plain User
(not Power User) account.
WRT Step 5 above (verify the release media), since by the time
release files are ready to download Thomas has generally run many
Windows tests on the installer he uploaded, he usually doesn’t do
anything for Step 5 except a full byte-comparison (“fc /b” if
using a Windows shell) of the downloaded file against the file he
uploaded.
Copyright
This document has been placed in the public domain.
| Superseded | PEP 102 – Doing Python Micro Releases | Informational | Making a Python release is an arduous process that takes a
minimum of half a day’s work even for an experienced releaser.
Until recently, most – if not all – of that burden was borne by
Guido himself. But several recent releases have been performed by
other folks, so this PEP attempts to collect, in one place, all
the steps needed to make a Python bugfix release. |
PEP 103 – Collecting information about git
Author:
Oleg Broytman <phd at phdru.name>
Status:
Withdrawn
Type:
Informational
Created:
01-Jun-2015
Post-History:
12-Sep-2015
Table of Contents
Withdrawal
Abstract
Documentation
Documentation for starters
Advanced documentation
Offline documentation
Quick start
Download and installation
Initial configuration
Examples in this PEP
Branches and branches
Remote repositories and remote branches
Updating local and remote-tracking branches
Fetch and pull
Push
Tags
Private information
Commit editing and caveats
Undo
git checkout: restore file’s content
git reset: remove (non-pushed) commits
Unstaging
git reflog: reference log
git revert: revert a commit
One thing that cannot be undone
Merge or rebase?
Null-merges
Branching models
Advanced configuration
Line endings
Useful assets
Advanced topics
Staging area
Root
ReReRe
Database maintenance
Tips and tricks
Command-line options and arguments
bash/zsh completion
bash/zsh prompt
SSH connection sharing
git on server
From Mercurial to git
Git and GitHub
Copyright
Withdrawal
This PEP was withdrawn as it’s too generic and doesn’t really deals
with Python development. It is no longer updated.
The content was moved to Python Wiki. Make further updates in the
wiki.
Abstract
This Informational PEP collects information about git. There is, of
course, a lot of documentation for git, so the PEP concentrates on
more complex (and more related to Python development) issues,
scenarios and examples.
The plan is to extend the PEP in the future collecting information
about equivalence of Mercurial and git scenarios to help migrating
Python development from Mercurial to git.
The author of the PEP doesn’t currently plan to write a Process PEP on
migration Python development from Mercurial to git.
Documentation
Git is accompanied with a lot of documentation, both online and
offline.
Documentation for starters
Git Tutorial: part 1,
part 2.
Git User’s manual.
Everyday GIT With 20 Commands Or So.
Git workflows.
Advanced documentation
Git Magic,
with a number of translations.
Pro Git. The Book about git. Buy it at
Amazon or download in PDF, mobi, or ePub form. It has translations to
many different languages. Download Russian translation from GArik.
Git Wiki.
Git Buch (German).
Offline documentation
Git has builtin help: run git help $TOPIC. For example, run
git help git or git help help.
Quick start
Download and installation
Unix users: download and install using your package manager.
Microsoft Windows: download git-for-windows.
MacOS X: use git installed with XCode or download from MacPorts or
git-osx-installer or
install git with Homebrew: brew install git.
git-cola (repository) is a Git GUI written in
Python and GPL licensed. Linux, Windows, MacOS X.
TortoiseGit is a Windows Shell Interface
to Git based on TortoiseSVN; open source.
Initial configuration
This simple code is often appears in documentation, but it is
important so let repeat it here. Git stores author and committer
names/emails in every commit, so configure your real name and
preferred email:
$ git config --global user.name "User Name"
$ git config --global user.email [email protected]
Examples in this PEP
Examples of git commands in this PEP use the following approach. It is
supposed that you, the user, works with a local repository named
python that has an upstream remote repo named origin. Your
local repo has two branches v1 and master. For most examples
the currently checked out branch is master. That is, it’s assumed
you have done something like that:
$ git clone https://git.python.org/python.git
$ cd python
$ git branch v1 origin/v1
The first command clones remote repository into local directory
python, creates a new local branch master, sets
remotes/origin/master as its upstream remote-tracking branch and
checks it out into the working directory.
The last command creates a new local branch v1 and sets
remotes/origin/v1 as its upstream remote-tracking branch.
The same result can be achieved with commands:
$ git clone -b v1 https://git.python.org/python.git
$ cd python
$ git checkout --track origin/master
The last command creates a new local branch master, sets
remotes/origin/master as its upstream remote-tracking branch and
checks it out into the working directory.
Branches and branches
Git terminology can be a bit misleading. Take, for example, the term
“branch”. In git it has two meanings. A branch is a directed line of
commits (possibly with merges). And a branch is a label or a pointer
assigned to a line of commits. It is important to distinguish when you
talk about commits and when about their labels. Lines of commits are
by itself unnamed and are usually only lengthening and merging.
Labels, on the other hand, can be created, moved, renamed and deleted
freely.
Remote repositories and remote branches
Remote-tracking branches are branches (pointers to commits) in your
local repository. They are there for git (and for you) to remember
what branches and commits have been pulled from and pushed to what
remote repos (you can pull from and push to many remotes).
Remote-tracking branches live under remotes/$REMOTE namespaces,
e.g. remotes/origin/master.
To see the status of remote-tracking branches run:
$ git branch -rv
To see local and remote-tracking branches (and tags) pointing to
commits:
$ git log --decorate
You never do your own development on remote-tracking branches. You
create a local branch that has a remote branch as upstream and do
development on that local branch. On push git pushes commits to the
remote repo and updates remote-tracking branches, on pull git fetches
commits from the remote repo, updates remote-tracking branches and
fast-forwards, merges or rebases local branches.
When you do an initial clone like this:
$ git clone -b v1 https://git.python.org/python.git
git clones remote repository https://git.python.org/python.git to
directory python, creates a remote named origin, creates
remote-tracking branches, creates a local branch v1, configure it
to track upstream remotes/origin/v1 branch and checks out v1 into
the working directory.
Some commands, like git status --branch and git branch --verbose,
report the difference between local and remote branches.
Please remember they only do comparison with remote-tracking branches
in your local repository, and the state of those remote-tracking
branches can be outdated. To update remote-tracking branches you
either fetch and merge (or rebase) commits from the remote repository
or update remote-tracking branches without updating local branches.
Updating local and remote-tracking branches
To update remote-tracking branches without updating local branches run
git remote update [$REMOTE...]. For example:
$ git remote update
$ git remote update origin
Fetch and pull
There is a major difference between
$ git fetch $REMOTE $BRANCH
and
$ git fetch $REMOTE $BRANCH:$BRANCH
The first command fetches commits from the named $BRANCH in the
$REMOTE repository that are not in your repository, updates
remote-tracking branch and leaves the id (the hash) of the head commit
in file .git/FETCH_HEAD.
The second command fetches commits from the named $BRANCH in the
$REMOTE repository that are not in your repository and updates both
the local branch $BRANCH and its upstream remote-tracking branch. But
it refuses to update branches in case of non-fast-forward. And it
refuses to update the current branch (currently checked out branch,
where HEAD is pointing to).
The first command is used internally by git pull.
$ git pull $REMOTE $BRANCH
is equivalent to
$ git fetch $REMOTE $BRANCH
$ git merge FETCH_HEAD
Certainly, $BRANCH in that case should be your current branch. If you
want to merge a different branch into your current branch first update
that non-current branch and then merge:
$ git fetch origin v1:v1 # Update v1
$ git pull --rebase origin master # Update the current branch master
# using rebase instead of merge
$ git merge v1
If you have not yet pushed commits on v1, though, the scenario has
to become a bit more complex. Git refuses to update
non-fast-forwardable branch, and you don’t want to do force-pull
because that would remove your non-pushed commits and you would need
to recover. So you want to rebase v1 but you cannot rebase
non-current branch. Hence, checkout v1 and rebase it before
merging:
$ git checkout v1
$ git pull --rebase origin v1
$ git checkout master
$ git pull --rebase origin master
$ git merge v1
It is possible to configure git to make it fetch/pull a few branches
or all branches at once, so you can simply run
$ git pull origin
or even
$ git pull
Default remote repository for fetching/pulling is origin. Default
set of references to fetch is calculated using matching algorithm: git
fetches all branches having the same name on both ends.
Push
Pushing is a bit simpler. There is only one command push. When you
run
$ git push origin v1 master
git pushes local v1 to remote v1 and local master to remote master.
The same as:
$ git push origin v1:v1 master:master
Git pushes commits to the remote repo and updates remote-tracking
branches. Git refuses to push commits that aren’t fast-forwardable.
You can force-push anyway, but please remember - you can force-push to
your own repositories but don’t force-push to public or shared repos.
If you find git refuses to push commits that aren’t fast-forwardable,
better fetch and merge commits from the remote repo (or rebase your
commits on top of the fetched commits), then push. Only force-push if
you know what you do and why you do it. See the section Commit
editing and caveats below.
It is possible to configure git to make it push a few branches or all
branches at once, so you can simply run
$ git push origin
or even
$ git push
Default remote repository for pushing is origin. Default set of
references to push in git before 2.0 is calculated using matching
algorithm: git pushes all branches having the same name on both ends.
Default set of references to push in git 2.0+ is calculated using
simple algorithm: git pushes the current branch back to its
@{upstream}.
To configure git before 2.0 to the new behaviour run:
$ git config push.default simple
To configure git 2.0+ to the old behaviour run:
$ git config push.default matching
Git doesn’t allow to push a branch if it’s the current branch in the
remote non-bare repository: git refuses to update remote working
directory. You really should push only to bare repositories. For
non-bare repositories git prefers pull-based workflow.
When you want to deploy code on a remote host and can only use push
(because your workstation is behind a firewall and you cannot pull
from it) you do that in two steps using two repositories: you push
from the workstation to a bare repo on the remote host, ssh to the
remote host and pull from the bare repo to a non-bare deployment repo.
That changed in git 2.3, but see the blog post
for caveats; in 2.4 the push-to-deploy feature was further improved.
Tags
Git automatically fetches tags that point to commits being fetched
during fetch/pull. To fetch all tags (and commits they point to) run
git fetch --tags origin. To fetch some specific tags fetch them
explicitly:
$ git fetch origin tag $TAG1 tag $TAG2...
For example:
$ git fetch origin tag 1.4.2
$ git fetch origin v1:v1 tag 2.1.7
Git doesn’t automatically pushes tags. That allows you to have private
tags. To push tags list them explicitly:
$ git push origin tag 1.4.2
$ git push origin v1 master tag 2.1.7
Or push all tags at once:
$ git push --tags origin
Don’t move tags with git tag -f or remove tags with git tag -d
after they have been published.
Private information
When cloning/fetching/pulling/pushing git copies only database objects
(commits, trees, files and tags) and symbolic references (branches and
lightweight tags). Everything else is private to the repository and
never cloned, updated or pushed. It’s your config, your hooks, your
private exclude file.
If you want to distribute hooks, copy them to the working tree, add,
commit, push and instruct the team to update and install the hooks
manually.
Commit editing and caveats
A warning not to edit published (pushed) commits also appears in
documentation but it’s repeated here anyway as it’s very important.
It is possible to recover from a forced push but it’s PITA for the
entire team. Please avoid it.
To see what commits have not been published yet compare the head of the
branch with its upstream remote-tracking branch:
$ git log origin/master.. # from origin/master to HEAD (of master)
$ git log origin/v1..v1 # from origin/v1 to the head of v1
For every branch that has an upstream remote-tracking branch git
maintains an alias @{upstream} (short version @{u}), so the commands
above can be given as:
$ git log @{u}..
$ git log v1@{u}..v1
To see the status of all branches:
$ git branch -avv
To compare the status of local branches with a remote repo:
$ git remote show origin
Read how to recover from upstream rebase.
It is in git help rebase.
On the other hand, don’t be too afraid about commit editing. You can
safely edit, reorder, remove, combine and split commits that haven’t
been pushed yet. You can even push commits to your own (backup) repo,
edit them later and force-push edited commits to replace what have
already been pushed. Not a problem until commits are in a public
or shared repository.
Undo
Whatever you do, don’t panic. Almost anything in git can be undone.
git checkout: restore file’s content
git checkout, for example, can be used to restore the content of
file(s) to that one of a commit. Like this:
git checkout HEAD~ README
The commands restores the contents of README file to the last but one
commit in the current branch. By default the commit ID is simply HEAD;
i.e. git checkout README restores README to the latest commit.
(Do not use git checkout to view a content of a file in a commit,
use git cat-file -p; e.g. git cat-file -p HEAD~:path/to/README).
git reset: remove (non-pushed) commits
git reset moves the head of the current branch. The head can be
moved to point to any commit but it’s often used to remove a commit or
a few (preferably, non-pushed ones) from the top of the branch - that
is, to move the branch backward in order to undo a few (non-pushed)
commits.
git reset has three modes of operation - soft, hard and mixed.
Default is mixed. ProGit explains the
difference very clearly. Bare repositories don’t have indices or
working trees so in a bare repo only soft reset is possible.
Unstaging
Mixed mode reset with a path or paths can be used to unstage changes -
that is, to remove from index changes added with git add for
committing. See The Book for details
about unstaging and other undo tricks.
git reflog: reference log
Removing commits with git reset or moving the head of a branch
sounds dangerous and it is. But there is a way to undo: another
reset back to the original commit. Git doesn’t remove commits
immediately; unreferenced commits (in git terminology they are called
“dangling commits”) stay in the database for some time (default is two
weeks) so you can reset back to it or create a new branch pointing to
the original commit.
For every move of a branch’s head - with git commit, git
checkout, git fetch, git pull, git rebase, git reset
and so on - git stores a reference log (reflog for short). For every
move git stores where the head was. Command git reflog can be used
to view (and manipulate) the log.
In addition to the moves of the head of every branch git stores the
moves of the HEAD - a symbolic reference that (usually) names the
current branch. HEAD is changed with git checkout $BRANCH.
By default git reflog shows the moves of the HEAD, i.e. the
command is equivalent to git reflog HEAD. To show the moves of the
head of a branch use the command git reflog $BRANCH.
So to undo a git reset lookup the original commit in git
reflog, verify it with git show or git log and run git
reset $COMMIT_ID. Git stores the move of the branch’s head in
reflog, so you can undo that undo later again.
In a more complex situation you’d want to move some commits along with
resetting the head of the branch. Cherry-pick them to the new branch.
For example, if you want to reset the branch master back to the
original commit but preserve two commits created in the current branch
do something like:
$ git branch save-master # create a new branch saving master
$ git reflog # find the original place of master
$ git reset $COMMIT_ID
$ git cherry-pick save-master~ save-master
$ git branch -D save-master # remove temporary branch
git revert: revert a commit
git revert reverts a commit or commits, that is, it creates a new
commit or commits that revert(s) the effects of the given commits.
It’s the only way to undo published commits (git commit --amend,
git rebase and git reset change the branch in
non-fast-forwardable ways so they should only be used for non-pushed
commits.)
There is a problem with reverting a merge commit. git revert can
undo the code created by the merge commit but it cannot undo the fact
of merge. See the discussion How to revert a faulty merge.
One thing that cannot be undone
Whatever you undo, there is one thing that cannot be undone -
overwritten uncommitted changes. Uncommitted changes don’t belong to
git so git cannot help preserving them.
Most of the time git warns you when you’re going to execute a command
that overwrites uncommitted changes. Git doesn’t allow you to switch
branches with git checkout. It stops you when you’re going to
rebase with non-clean working tree. It refuses to pull new commits
over non-committed files.
But there are commands that do exactly that - overwrite files in the
working tree. Commands like git checkout $PATHs or git reset
--hard silently overwrite files including your uncommitted changes.
With that in mind you can understand the stance “commit early, commit
often”. Commit as often as possible. Commit on every save in your
editor or IDE. You can edit your commits before pushing - edit commit
messages, change commits, reorder, combine, split, remove. But save
your changes in git database, either commit changes or at least stash
them with git stash.
Merge or rebase?
Internet is full of heated discussions on the topic: “merge or
rebase?” Most of them are meaningless. When a DVCS is being used in a
big team with a big and complex project with many branches there is
simply no way to avoid merges. So the question’s diminished to
“whether to use rebase, and if yes - when to use rebase?” Considering
that it is very much recommended not to rebase published commits the
question’s diminished even further: “whether to use rebase on
non-pushed commits?”
That small question is for the team to decide. To preserve the beauty
of linear history it’s recommended to use rebase when pulling, i.e. do
git pull --rebase or even configure automatic setup of rebase for
every new branch:
$ git config branch.autosetuprebase always
and configure rebase for existing branches:
$ git config branch.$NAME.rebase true
For example:
$ git config branch.v1.rebase true
$ git config branch.master.rebase true
After that git pull origin master becomes equivalent to git pull
--rebase origin master.
It is recommended to create new commits in a separate feature or topic
branch while using rebase to update the mainline branch. When the
topic branch is ready merge it into mainline. To avoid a tedious task
of resolving large number of conflicts at once you can merge the topic
branch to the mainline from time to time and switch back to the topic
branch to continue working on it. The entire workflow would be
something like:
$ git checkout -b issue-42 # create a new issue branch and switch to it
...edit/test/commit...
$ git checkout master
$ git pull --rebase origin master # update master from the upstream
$ git merge issue-42
$ git branch -d issue-42 # delete the topic branch
$ git push origin master
When the topic branch is deleted only the label is removed, commits
are stayed in the database, they are now merged into master:
o--o--o--o--o--M--< master - the mainline branch
\ /
--*--*--* - the topic branch, now unnamed
The topic branch is deleted to avoid cluttering branch namespace with
small topic branches. Information on what issue was fixed or what
feature was implemented should be in the commit messages.
But even that small amount of rebasing could be too big in case of
long-lived merged branches. Imagine you’re doing work in both v1
and master branches, regularly merging v1 into master.
After some time you will have a lot of merge and non-merge commits in
master. Then you want to push your finished work to a shared
repository and find someone has pushed a few commits to v1. Now
you have a choice of two equally bad alternatives: either you fetch
and rebase v1 and then have to recreate all you work in master
(reset master to the origin, merge v1 and cherry-pick all
non-merge commits from the old master); or merge the new v1 and
loose the beauty of linear history.
Null-merges
Git has a builtin merge strategy for what Python core developers call
“null-merge”:
$ git merge -s ours v1 # null-merge v1 into master
Branching models
Git doesn’t assume any particular development model regarding
branching and merging. Some projects prefer to graduate patches from
the oldest branch to the newest, some prefer to cherry-pick commits
backwards, some use squashing (combining a number of commits into
one). Anything is possible.
There are a few examples to start with. git help workflows
describes how the very git authors develop git.
ProGit book has a few chapters devoted to branch management in
different projects: Git Branching - Branching Workflows and
Distributed Git - Contributing to a Project.
There is also a well-known article A successful Git branching model by Vincent
Driessen. It recommends a set of very detailed rules on creating and
managing mainline, topic and bugfix branches. To support the model the
author implemented git flow
extension.
Advanced configuration
Line endings
Git has builtin mechanisms to handle line endings between platforms
with different end-of-line styles. To allow git to do CRLF conversion
assign text attribute to files using .gitattributes.
For files that have to have specific line endings assign eol
attribute. For binary files the attribute is, naturally, binary.
For example:
$ cat .gitattributes
*.py text
*.txt text
*.png binary
/readme.txt eol=CRLF
To check what attributes git uses for files use git check-attr
command. For example:
$ git check-attr -a -- \*.py
Useful assets
GitAlias (repository) is a big collection of
aliases. A careful selection of aliases for frequently used commands
could save you a lot of keystrokes!
GitIgnore and
https://github.com/github/gitignore are collections of .gitignore
files for all kinds of IDEs and programming languages. Python
included!
pre-commit (repositories) is a framework for managing and
maintaining multi-language pre-commit hooks. The framework is written
in Python and has a lot of plugins for many programming languages.
Advanced topics
Staging area
Staging area aka index aka cache is a distinguishing feature of git.
Staging area is where git collects patches before committing them.
Separation between collecting patches and commit phases provides a
very useful feature of git: you can review collected patches before
commit and even edit them - remove some hunks, add new hunks and
review again.
To add files to the index use git add. Collecting patches before
committing means you need to do that for every change, not only to add
new (untracked) files. To simplify committing in case you just want to
commit everything without reviewing run git commit --all (or just
-a) - the command adds every changed tracked file to the index and
then commit. To commit a file or files regardless of patches collected
in the index run git commit [--only|-o] -- $FILE....
To add hunks of patches to the index use git add --patch (or just
-p). To remove collected files from the index use git reset HEAD
-- $FILE... To add/inspect/remove collected hunks use git add
--interactive (-i).
To see the diff between the index and the last commit (i.e., collected
patches) use git diff --cached. To see the diff between the
working tree and the index (i.e., uncollected patches) use just git
diff. To see the diff between the working tree and the last commit
(i.e., both collected and uncollected patches) run git diff HEAD.
See WhatIsTheIndex and
IndexCommandQuickref in Git
Wiki.
Root
Git switches to the root (top-level directory of the project where
.git subdirectory exists) before running any command. Git
remembers though the directory that was current before the switch.
Some programs take into account the current directory. E.g., git
status shows file paths of changed and unknown files relative to the
current directory; git grep searches below the current directory;
git apply applies only those hunks from the patch that touch files
below the current directory.
But most commands run from the root and ignore the current directory.
Imagine, for example, that you have two work trees, one for the branch
v1 and the other for master. If you want to merge v1 from
a subdirectory inside the second work tree you must write commands as
if you’re in the top-level dir. Let take two work trees,
project-v1 and project, for example:
$ cd project/subdirectory
$ git fetch ../project-v1 v1:v1
$ git merge v1
Please note the path in git fetch ../project-v1 v1:v1 is
../project-v1 and not ../../project-v1 despite the fact that
we run the commands from a subdirectory, not from the root.
ReReRe
Rerere is a mechanism that helps to resolve repeated merge conflicts.
The most frequent source of recurring merge conflicts are topic
branches that are merged into mainline and then the merge commits are
removed; that’s often performed to test the topic branches and train
rerere; merge commits are removed to have clean linear history and
finish the topic branch with only one last merge commit.
Rerere works by remembering the states of tree before and after a
successful commit. That way rerere can automatically resolve conflicts
if they appear in the same files.
Rerere can be used manually with git rerere command but most often
it’s used automatically. Enable rerere with these commands in a
working tree:
$ git config rerere.enabled true
$ git config rerere.autoupdate true
You don’t need to turn rerere on globally - you don’t want rerere in
bare repositories or single-branch repositories; you only need rerere
in repos where you often perform merges and resolve merge conflicts.
See Rerere in The
Book.
Database maintenance
Git object database and other files/directories under .git require
periodic maintenance and cleanup. For example, commit editing left
unreferenced objects (dangling objects, in git terminology) and these
objects should be pruned to avoid collecting cruft in the DB. The
command git gc is used for maintenance. Git automatically runs
git gc --auto as a part of some commands to do quick maintenance.
Users are recommended to run git gc --aggressive from time to
time; git help gc recommends to run it every few hundred
changesets; for more intensive projects it should be something like
once a week and less frequently (biweekly or monthly) for lesser
active projects.
git gc --aggressive not only removes dangling objects, it also
repacks object database into indexed and better optimized pack(s); it
also packs symbolic references (branches and tags). Another way to do
it is to run git repack.
There is a well-known message from Linus
Torvalds regarding “stupidity” of git gc --aggressive. The message
can safely be ignored now. It is old and outdated, git gc
--aggressive became much better since that time.
For those who still prefer git repack over git gc --aggressive
the recommended parameters are git repack -a -d -f --depth=20
--window=250. See this detailed experiment
for explanation of the effects of these parameters.
From time to time run git fsck [--strict] to verify integrity of
the database. git fsck may produce a list of dangling objects;
that’s not an error, just a reminder to perform regular maintenance.
Tips and tricks
Command-line options and arguments
git help cli
recommends not to combine short options/flags. Most of the times
combining works: git commit -av works perfectly, but there are
situations when it doesn’t. E.g., git log -p -5 cannot be combined
as git log -p5.
Some options have arguments, some even have default arguments. In that
case the argument for such option must be spelled in a sticky way:
-Oarg, never -O arg because for an option that has a default
argument the latter means “use default value for option -O and
pass arg further to the option parser”. For example, git grep
has an option -O that passes a list of names of the found files to
a program; default program for -O is a pager (usually less),
but you can use your editor:
$ git grep -Ovim # but not -O vim
BTW, if git is instructed to use less as the pager (i.e., if pager
is not configured in git at all it uses less by default, or if it
gets less from GIT_PAGER or PAGER environment variables, or if it
was configured with git config [--global] core.pager less, or
less is used in the command git grep -Oless) git grep
passes +/$pattern option to less which is quite convenient.
Unfortunately, git grep doesn’t pass the pattern if the pager is
not exactly less, even if it’s less with parameters (something
like git config [--global] core.pager less -FRSXgimq); fortunately,
git grep -Oless always passes the pattern.
bash/zsh completion
It’s a bit hard to type git rebase --interactive --preserve-merges
HEAD~5 manually even for those who are happy to use command-line,
and this is where shell completion is of great help. Bash/zsh come
with programmable completion, often automatically installed and
enabled, so if you have bash/zsh and git installed, chances are you
are already done - just go and use it at the command-line.
If you don’t have necessary bits installed, install and enable
bash_completion package. If you want to upgrade your git completion to
the latest and greatest download necessary file from git contrib.
Git-for-windows comes with git-bash for which bash completion is
installed and enabled.
bash/zsh prompt
For command-line lovers shell prompt can carry a lot of useful
information. To include git information in the prompt use
git-prompt.sh.
Read the detailed instructions in the file.
Search the Net for “git prompt” to find other prompt variants.
SSH connection sharing
SSH connection sharing is a feature of OpenSSH and perhaps derivatives
like PuTTY. SSH connection sharing is a way to decrease ssh client
startup time by establishing one connection and reusing it for all
subsequent clients connecting to the same server. SSH connection
sharing can be used to speedup a lot of short ssh sessions like scp,
sftp, rsync and of course git over ssh. If you regularly
fetch/pull/push from/to remote repositories accessible over ssh then
using ssh connection sharing is recommended.
To turn on ssh connection sharing add something like this to your
~/.ssh/config:
Host *
ControlMaster auto
ControlPath ~/.ssh/mux-%r@%h:%p
ControlPersist 600
See OpenSSH wikibook and
search for
more information.
SSH connection sharing can be used at GitHub, GitLab and SourceForge
repositories, but please be advised that BitBucket doesn’t allow it
and forcibly closes master connection after a short inactivity period
so you will see errors like this from ssh: “Connection to bitbucket.org
closed by remote host.”
git on server
The simplest way to publish a repository or a group of repositories is
git daemon. The daemon provides anonymous access, by default it is
read-only. The repositories are accessible by git protocol (git://
URLs). Write access can be enabled but the protocol lacks any
authentication means, so it should be enabled only within a trusted
LAN. See git help daemon for details.
Git over ssh provides authentication and repo-level authorisation as
repositories can be made user- or group-writeable (see parameter
core.sharedRepository in git help config). If that’s too
permissive or too restrictive for some project’s needs there is a
wrapper gitolite that can
be configured to allow access with great granularity; gitolite is
written in Perl and has a lot of documentation.
Web interface to browse repositories can be created using gitweb or cgit. Both are CGI scripts (written in
Perl and C). In addition to web interface both provide read-only dumb
http access for git (http(s):// URLs). Klaus is a small and simple WSGI web
server that implements both web interface and git smart HTTP
transport; supports Python 2 and Python 3, performs syntax
highlighting.
There are also more advanced web-based development environments that
include ability to manage users, groups and projects; private,
group-accessible and public repositories; they often include issue
trackers, wiki pages, pull requests and other tools for development
and communication. Among these environments are Kallithea and pagure,
both are written in Python; pagure was written by Fedora developers
and is being used to develop some Fedora projects. GitPrep is yet another GitHub clone,
written in Perl. Gogs is written in Go.
GitBucket is
written in Scala.
And last but not least, GitLab. It’s
perhaps the most advanced web-based development environment for git.
Written in Ruby, community edition is free and open source (MIT
license).
From Mercurial to git
There are many tools to convert Mercurial repositories to git. The
most famous are, probably, hg-git and
fast-export (many years ago
it was known under the name hg2git).
But a better tool, perhaps the best, is git-remote-hg. It provides transparent
bidirectional (pull and push) access to Mercurial repositories from
git. Its author wrote a comparison of alternatives
that seems to be mostly objective.
To use git-remote-hg, install or clone it, add to your PATH (or copy
script git-remote-hg to a directory that’s already in PATH) and
prepend hg:: to Mercurial URLs. For example:
$ git clone https://github.com/felipec/git-remote-hg.git
$ PATH=$PATH:"`pwd`"/git-remote-hg
$ git clone hg::https://hg.python.org/peps/ PEPs
To work with the repository just use regular git commands including
git fetch/pull/push.
To start converting your Mercurial habits to git see the page
Mercurial for Git users at Mercurial wiki.
At the second half of the page there is a table that lists
corresponding Mercurial and git commands. Should work perfectly in
both directions.
Python Developer’s Guide also has a chapter Mercurial for git
developers that
documents a few differences between git and hg.
Git and GitHub
gitsome - Git/GitHub
command line interface (CLI). Written in Python, work on MacOS, Unix,
Windows. Git/GitHub CLI with autocomplete, includes many GitHub
integrated commands that work with all shells, builtin xonsh with
Python REPL to run Python commands alongside shell commands, command
history, customizable highlighting, thoroughly documented.
Copyright
This document has been placed in the public domain.
| Withdrawn | PEP 103 – Collecting information about git | Informational | This Informational PEP collects information about git. There is, of
course, a lot of documentation for git, so the PEP concentrates on
more complex (and more related to Python development) issues,
scenarios and examples. |
PEP 207 – Rich Comparisons
Author:
Guido van Rossum <guido at python.org>, David Ascher <DavidA at ActiveState.com>
Status:
Final
Type:
Standards Track
Created:
25-Jul-2000
Python-Version:
2.1
Post-History:
Table of Contents
Abstract
Motivation
Previous Work
Concerns
Proposed Resolutions
Implementation Proposal
C API
Changes to the interpreter
Classes
Copyright
Appendix
Abstract
Motivation
Current State of Affairs
Proposed Mechanism
Chained Comparisons
Problem
Solution
Abstract
This PEP proposes several new features for comparisons:
Allow separately overloading of <, >, <=, >=, ==, !=, both in
classes and in C extensions.
Allow any of those overloaded operators to return something else
besides a Boolean result.
Motivation
The main motivation comes from NumPy, whose users agree that A<B
should return an array of elementwise comparison outcomes; they
currently have to spell this as less(A,B) because A<B can only
return a Boolean result or raise an exception.
An additional motivation is that frequently, types don’t have a
natural ordering, but still need to be compared for equality.
Currently such a type must implement comparison and thus define
an arbitrary ordering, just so that equality can be tested.
Also, for some object types an equality test can be implemented
much more efficiently than an ordering test; for example, lists
and dictionaries that differ in length are unequal, but the
ordering requires inspecting some (potentially all) items.
Previous Work
Rich Comparisons have been proposed before; in particular by David
Ascher, after experience with Numerical Python:
http://starship.python.net/crew/da/proposals/richcmp.html
It is also included below as an Appendix. Most of the material in
this PEP is derived from David’s proposal.
Concerns
Backwards compatibility, both at the Python level (classes using
__cmp__ need not be changed) and at the C level (extensions
defining tp_comparea need not be changed, code using
PyObject_Compare() must work even if the compared objects use
the new rich comparison scheme).
When A<B returns a matrix of elementwise comparisons, an easy
mistake to make is to use this expression in a Boolean context.
Without special precautions, it would always be true. This use
should raise an exception instead.
If a class overrides x==y but nothing else, should x!=y be
computed as not(x==y), or fail? What about the similar
relationship between < and >=, or between > and <=?
Similarly, should we allow x<y to be calculated from y>x? And
x<=y from not(x>y)? And x==y from y==x, or x!=y from y!=x?
When comparison operators return elementwise comparisons, what
to do about shortcut operators like A<B<C, A<B and C<D,
A<B or C<D?
What to do about min() and max(), the ‘in’ and ‘not in’
operators, list.sort(), dictionary key comparison, and other
uses of comparisons by built-in operations?
Proposed Resolutions
Full backwards compatibility can be achieved as follows. When
an object defines tp_compare() but not tp_richcompare(), and a
rich comparison is requested, the outcome of tp_compare() is
used in the obvious way. E.g. if “<” is requested, an exception if
tp_compare() raises an exception, the outcome is 1 if
tp_compare() is negative, and 0 if it is zero or positive. Etc.Full forward compatibility can be achieved as follows. When a
classic comparison is requested on an object that implements
tp_richcompare(), up to three comparisons are used: first == is
tried, and if it returns true, 0 is returned; next, < is tried
and if it returns true, -1 is returned; next, > is tried and if
it returns true, +1 is returned. If any operator tried returns
a non-Boolean value (see below), the exception raised by
conversion to Boolean is passed through. If none of the
operators tried returns true, the classic comparison fallbacks
are tried next.
(I thought long and hard about the order in which the three
comparisons should be tried. At one point I had a convincing
argument for doing it in this order, based on the behavior of
comparisons for cyclical data structures. But since that code
has changed again, I’m not so sure that it makes a difference
any more.)
Any type that returns a collection of Booleans instead of a
single boolean should define nb_nonzero() to raise an exception.
Such a type is considered a non-Boolean.
The == and != operators are not assumed to be each other’s
complement (e.g. IEEE 754 floating point numbers do not satisfy
this). It is up to the type to implement this if desired.
Similar for < and >=, or > and <=; there are lots of examples
where these assumptions aren’t true (e.g. tabnanny).
The reflexivity rules are assumed by Python. Thus, the
interpreter may swap y>x with x<y, y>=x with x<=y, and may swap
the arguments of x==y and x!=y. (Note: Python currently assumes
that x==x is always true and x!=x is never true; this should not
be assumed.)
In the current proposal, when A<B returns an array of
elementwise comparisons, this outcome is considered non-Boolean,
and its interpretation as Boolean by the shortcut operators
raises an exception. David Ascher’s proposal tries to deal
with this; I don’t think this is worth the additional complexity
in the code generator. Instead of A<B<C, you can write
(A<B)&(B<C).
The min() and list.sort() operations will only use the
< operator; max() will only use the > operator. The ‘in’ and
‘not in’ operators and dictionary lookup will only use the ==
operator.
Implementation Proposal
This closely follows David Ascher’s proposal.
C API
New functions:PyObject *PyObject_RichCompare(PyObject *, PyObject *, int)
This performs the requested rich comparison, returning a Python
object or raising an exception. The 3rd argument must be one of
Py_LT, Py_LE, Py_EQ, Py_NE, Py_GT or Py_GE.
int PyObject_RichCompareBool(PyObject *, PyObject *, int)
This performs the requested rich comparison, returning a
Boolean: -1 for exception, 0 for false, 1 for true. The 3rd
argument must be one of Py_LT, Py_LE, Py_EQ, Py_NE, Py_GT or
Py_GE. Note that when PyObject_RichCompare() returns a
non-Boolean object, PyObject_RichCompareBool() will raise an
exception.
New typedef:typedef PyObject *(*richcmpfunc) (PyObject *, PyObject *, int);
New slot in type object, replacing spare tp_xxx7:richcmpfunc tp_richcompare;
This should be a function with the same signature as
PyObject_RichCompare(), and performing the same comparison.
At least one of the arguments is of the type whose
tp_richcompare slot is being used, but the other may have a
different type. If the function cannot compare the particular
combination of objects, it should return a new reference to
Py_NotImplemented.
PyObject_Compare() is changed to try rich comparisons if they
are defined (but only if classic comparisons aren’t defined).
Changes to the interpreter
Whenever PyObject_Compare() is called with the intent of getting
the outcome of a particular comparison (e.g. in list.sort(), and
of course for the comparison operators in ceval.c), the code is
changed to call PyObject_RichCompare() or
PyObject_RichCompareBool() instead; if the C code needs to know
the outcome of the comparison, PyObject_IsTrue() is called on
the result (which may raise an exception).
Most built-in types that currently define a comparison will be
modified to define a rich comparison instead. (This is
optional; I’ve converted lists, tuples, complex numbers, and
arrays so far, and am not sure whether I will convert others.)
Classes
Classes can define new special methods __lt__, __le__, __eq__,
__ne__, __gt__, __ge__ to override the corresponding operators.
(I.e., <, <=, ==, !=, >, >=. You gotta love the Fortran
heritage.) If a class defines __cmp__ as well, it is only used
when __lt__ etc. have been tried and return NotImplemented.
Copyright
This document has been placed in the public domain.
Appendix
Here is most of David Ascher’s original proposal (version 0.2.1,
dated Wed Jul 22 16:49:28 1998; I’ve left the Contents, History
and Patches sections out). It addresses almost all concerns
above.
Abstract
A new mechanism allowing comparisons of Python objects to return
values other than -1, 0, or 1 (or raise exceptions) is
proposed. This mechanism is entirely backwards compatible, and can
be controlled at the level of the C PyObject type or of the Python
class definition. There are three cooperating parts to the
proposed mechanism:
the use of the last slot in the type object structure to store a
pointer to a rich comparison function
the addition of special methods for classes
the addition of an optional argument to the builtin cmp()
function.
Motivation
The current comparison protocol for Python objects assumes that
any two Python objects can be compared (as of Python 1.5, object
comparisons can raise exceptions), and that the return value for
any comparison should be -1, 0 or 1. -1 indicates that the first
argument to the comparison function is less than the right one, +1
indicating the contrapositive, and 0 indicating that the two
objects are equal. While this mechanism allows the establishment
of an order relationship (e.g. for use by the sort() method of list
objects), it has proven to be limited in the context of Numeric
Python (NumPy).
Specifically, NumPy allows the creation of multidimensional
arrays, which support most of the numeric operators. Thus:
x = array((1,2,3,4)) y = array((2,2,4,4))
are two NumPy arrays. While they can be added elementwise,:
z = x + y # z == array((3,4,7,8))
they cannot be compared in the current framework - the released
version of NumPy compares the pointers, (thus yielding junk
information) which was the only solution before the recent
addition of the ability (in 1.5) to raise exceptions in comparison
functions.
Even with the ability to raise exceptions, the current protocol
makes array comparisons useless. To deal with this fact, NumPy
includes several functions which perform the comparisons: less(),
less_equal(), greater(), greater_equal(), equal(),
not_equal(). These functions return arrays with the same shape as
their arguments (modulo broadcasting), filled with 0’s and 1’s
depending on whether the comparison is true or not for each
element pair. Thus, for example, using the arrays x and y defined
above:
less(x,y)
would be an array containing the numbers (1,0,0,0).
The current proposal is to modify the Python object interface to
allow the NumPy package to make it so that x < y returns the same
thing as less(x,y). The exact return value is up to the NumPy
package – what this proposal really asks for is changing the
Python core so that extension objects have the ability to return
something other than -1, 0, 1, should their authors choose to do
so.
Current State of Affairs
The current protocol is, at the C level, that each object type
defines a tp_compare slot, which is a pointer to a function which
takes two PyObject* references and returns -1, 0, or 1. This
function is called by the PyObject_Compare() function defined in
the C API. PyObject_Compare() is also called by the builtin
function cmp() which takes two arguments.
Proposed Mechanism
Changes to the C structure for type objectsThe last available slot in the PyTypeObject, reserved up to now
for future expansion, is used to optionally store a pointer to a
new comparison function, of type richcmpfunc defined by:
typedef PyObject *(*richcmpfunc)
Py_PROTO((PyObject *, PyObject *, int));
This function takes three arguments. The first two are the objects
to be compared, and the third is an integer corresponding to an
opcode (one of LT, LE, EQ, NE, GT, GE). If this slot is left NULL,
then rich comparison for that object type is not supported (except
for class instances whose class provide the special methods
described below).
The above opcodes need to be added to the published Python/C API
(probably under the names Py_LT, Py_LE, etc.)
Additions of special methods for classesClasses wishing to support the rich comparison mechanisms must add
one or more of the following new special methods:
def __lt__(self, other):
...
def __le__(self, other):
...
def __gt__(self, other):
...
def __ge__(self, other):
...
def __eq__(self, other):
...
def __ne__(self, other):
...
Each of these is called when the class instance is the on the
left-hand-side of the corresponding operators (<, <=, >, >=, ==,
and != or <>). The argument other is set to the object on the
right side of the operator. The return value of these methods is
up to the class implementor (after all, that’s the entire point of
the proposal).
If the object on the left side of the operator does not define an
appropriate rich comparison operator (either at the C level or
with one of the special methods, then the comparison is reversed,
and the right hand operator is called with the opposite operator,
and the two objects are swapped. This assumes that a < b and b > a
are equivalent, as are a <= b and b >= a, and that == and != are
commutative (e.g. a == b if and only if b == a).
For example, if obj1 is an object which supports the rich
comparison protocol and x and y are objects which do not support
the rich comparison protocol, then obj1 < x will call the __lt__
method of obj1 with x as the second argument. x < obj1 will call
obj1’s __gt__ method with x as a second argument, and x < y will
just use the existing (non-rich) comparison mechanism.
The above mechanism is such that classes can get away with not
implementing either __lt__ and __le__ or __gt__ and
__ge__. Further smarts could have been added to the comparison
mechanism, but this limited set of allowed “swaps” was chosen
because it doesn’t require the infrastructure to do any processing
(negation) of return values. The choice of six special methods was
made over a single (e.g. __richcmp__) method to allow the
dispatching on the opcode to be performed at the level of the C
implementation rather than the user-defined method.
Addition of an optional argument to the builtin cmp()The builtin cmp() is still used for simple comparisons. For rich
comparisons, it is called with a third argument, one of “<”, “<=”,
“>”, “>=”, “==”, “!=”, “<>” (the last two have the same
meaning). When called with one of these strings as the third
argument, cmp() can return any Python object. Otherwise, it can
only return -1, 0 or 1 as before.
Chained Comparisons
Problem
It would be nice to allow objects for which the comparison returns
something other than -1, 0, or 1 to be used in chained
comparisons, such as:
x < y < z
Currently, this is interpreted by Python as:
temp1 = x < y
if temp1:
return y < z
else:
return temp1
Note that this requires testing the truth value of the result of
comparisons, with potential “shortcutting” of the right-side
comparison testings. In other words, the truth-value of the result
of the result of the comparison determines the result of a chained
operation. This is problematic in the case of arrays, since if x,
y and z are three arrays, then the user expects:
x < y < z
to be an array of 0’s and 1’s where 1’s are in the locations
corresponding to the elements of y which are between the
corresponding elements in x and z. In other words, the right-hand
side must be evaluated regardless of the result of x < y, which is
incompatible with the mechanism currently in use by the parser.
Solution
Guido mentioned that one possible way out would be to change the
code generated by chained comparisons to allow arrays to be
chained-compared intelligently. What follows is a mixture of his
idea and my suggestions. The code generated for x < y < z would be
equivalent to:
temp1 = x < y
if temp1:
temp2 = y < z
return boolean_combine(temp1, temp2)
else:
return temp1
where boolean_combine is a new function which does something like
the following:
def boolean_combine(a, b):
if hasattr(a, '__boolean_and__') or \
hasattr(b, '__boolean_and__'):
try:
return a.__boolean_and__(b)
except:
return b.__boolean_and__(a)
else: # standard behavior
if a:
return b
else:
return 0
where the __boolean_and__ special method is implemented for
C-level types by another value of the third argument to the
richcmp function. This method would perform a boolean comparison
of the arrays (currently implemented in the umath module as the
logical_and ufunc).
Thus, objects returned by rich comparisons should always test
true, but should define another special method which creates
boolean combinations of them and their argument.
This solution has the advantage of allowing chained comparisons to
work for arrays, but the disadvantage that it requires comparison
arrays to always return true (in an ideal world, I’d have them
always raise an exception on truth testing, since the meaning of
testing “if a>b:” is massively ambiguous.
The inlining already present which deals with integer comparisons
would still apply, resulting in no performance cost for the most
common cases.
| Final | PEP 207 – Rich Comparisons | Standards Track | This PEP proposes several new features for comparisons: |
PEP 208 – Reworking the Coercion Model
Author:
Neil Schemenauer <nas at arctrix.com>, Marc-André Lemburg <mal at lemburg.com>
Status:
Final
Type:
Standards Track
Created:
04-Dec-2000
Python-Version:
2.1
Post-History:
Table of Contents
Abstract
Rationale
Specification
Reference Implementation
Credits
Copyright
References
Abstract
Many Python types implement numeric operations. When the arguments of
a numeric operation are of different types, the interpreter tries to
coerce the arguments into a common type. The numeric operation is
then performed using this common type. This PEP proposes a new type
flag to indicate that arguments to a type’s numeric operations should
not be coerced. Operations that do not support the supplied types
indicate it by returning a new singleton object. Types which do not
set the type flag are handled in a backwards compatible manner.
Allowing operations handle different types is often simpler, more
flexible, and faster than having the interpreter do coercion.
Rationale
When implementing numeric or other related operations, it is often
desirable to provide not only operations between operands of one type
only, e.g. integer + integer, but to generalize the idea behind the
operation to other type combinations as well, e.g. integer + float.
A common approach to this mixed type situation is to provide a method
of “lifting” the operands to a common type (coercion) and then use
that type’s operand method as execution mechanism. Yet, this strategy
has a few drawbacks:
the “lifting” process creates at least one new (temporary)
operand object,
since the coercion method is not being told about the operation
that is to follow, it is not possible to implement operation
specific coercion of types,
there is no elegant way to solve situations were a common type
is not at hand, and
the coercion method will always have to be called prior to the
operation’s method itself.
A fix for this situation is obviously needed, since these drawbacks
make implementations of types needing these features very cumbersome,
if not impossible. As an example, have a look at the DateTime and
DateTimeDelta [1] types, the first being absolute, the second
relative. You can always add a relative value to an absolute one,
giving a new absolute value. Yet, there is no common type which the
existing coercion mechanism could use to implement that operation.
Currently, PyInstance types are treated specially by the interpreter
in that their numeric methods are passed arguments of different types.
Removing this special case simplifies the interpreter and allows other
types to implement numeric methods that behave like instance types.
This is especially useful for extension types like ExtensionClass.
Specification
Instead of using a central coercion method, the process of handling
different operand types is simply left to the operation. If the
operation finds that it cannot handle the given operand type
combination, it may return a special singleton as indicator.
Note that “numbers” (anything that implements the number protocol, or
part of it) written in Python already use the first part of this
strategy - it is the C level API that we focus on here.
To maintain nearly 100% backward compatibility we have to be very
careful to make numbers that don’t know anything about the new
strategy (old style numbers) work just as well as those that expect
the new scheme (new style numbers). Furthermore, binary compatibility
is a must, meaning that the interpreter may only access and use new
style operations if the number indicates the availability of these.
A new style number is considered by the interpreter as such if and
only if it sets the type flag Py_TPFLAGS_CHECKTYPES. The main
difference between an old style number and a new style one is that the
numeric slot functions can no longer assume to be passed arguments of
identical type. New style slots must check all arguments for proper
type and implement the necessary conversions themselves. This may seem
to cause more work on the behalf of the type implementor, but is in
fact no more difficult than writing the same kind of routines for an
old style coercion slot.
If a new style slot finds that it cannot handle the passed argument
type combination, it may return a new reference of the special
singleton Py_NotImplemented to the caller. This will cause the caller
to try the other operands operation slots until it finds a slot that
does implement the operation for the specific type combination. If
none of the possible slots succeed, it raises a TypeError.
To make the implementation easy to understand (the whole topic is
esoteric enough), a new layer in the handling of numeric operations is
introduced. This layer takes care of all the different cases that need
to be taken into account when dealing with all the possible
combinations of old and new style numbers. It is implemented by the
two static functions binary_op() and ternary_op(), which are both
internal functions that only the functions in Objects/abstract.c
have access to. The numeric API (PyNumber_*) is easy to adapt to
this new layer.
As a side-effect all numeric slots can be NULL-checked (this has to be
done anyway, so the added feature comes at no extra cost).
The scheme used by the layer to execute a binary operation is as
follows:
v
w
Action taken
new
new
v.op(v,w), w.op(v,w)
new
old
v.op(v,w), coerce(v,w), v.op(v,w)
old
new
w.op(v,w), coerce(v,w), v.op(v,w)
old
old
coerce(v,w), v.op(v,w)
The indicated action sequence is executed from left to right until
either the operation succeeds and a valid result (!=
Py_NotImplemented) is returned or an exception is raised. Exceptions
are returned to the calling function as-is. If a slot returns
Py_NotImplemented, the next item in the sequence is executed.
Note that coerce(v,w) will use the old style nb_coerce slot methods
via a call to PyNumber_Coerce().
Ternary operations have a few more cases to handle:
v
w
z
Action taken
new
new
new
v.op(v,w,z), w.op(v,w,z), z.op(v,w,z)
new
old
new
v.op(v,w,z), z.op(v,w,z), coerce(v,w,z), v.op(v,w,z)
old
new
new
w.op(v,w,z), z.op(v,w,z), coerce(v,w,z), v.op(v,w,z)
old
old
new
z.op(v,w,z), coerce(v,w,z), v.op(v,w,z)
new
new
old
v.op(v,w,z), w.op(v,w,z), coerce(v,w,z), v.op(v,w,z)
new
old
old
v.op(v,w,z), coerce(v,w,z), v.op(v,w,z)
old
new
old
w.op(v,w,z), coerce(v,w,z), v.op(v,w,z)
old
old
old
coerce(v,w,z), v.op(v,w,z)
The same notes as above, except that coerce(v,w,z) actually does:
if z != Py_None:
coerce(v,w), coerce(v,z), coerce(w,z)
else:
# treat z as absent variable
coerce(v,w)
The current implementation uses this scheme already (there’s only one
ternary slot: nb_pow(a,b,c)).
Note that the numeric protocol is also used for some other related
tasks, e.g. sequence concatenation. These can also benefit from the
new mechanism by implementing right-hand operations for type
combinations that would otherwise fail to work. As an example, take
string concatenation: currently you can only do string + string. With
the new mechanism, a new string-like type could implement new_type +
string and string + new_type, even though strings don’t know anything
about new_type.
Since comparisons also rely on coercion (every time you compare an
integer to a float, the integer is first converted to float and then
compared…), a new slot to handle numeric comparisons is needed:
PyObject *nb_cmp(PyObject *v, PyObject *w)
This slot should compare the two objects and return an integer object
stating the result. Currently, this result integer may only be -1, 0, 1.
If the slot cannot handle the type combination, it may return a
reference to Py_NotImplemented. [XXX Note that this slot is still
in flux since it should take into account rich comparisons
(i.e. PEP 207).]
Numeric comparisons are handled by a new numeric protocol API:
PyObject *PyNumber_Compare(PyObject *v, PyObject *w)
This function compare the two objects as “numbers” and return an
integer object stating the result. Currently, this result integer may
only be -1, 0, 1. In case the operation cannot be handled by the given
objects, a TypeError is raised.
The PyObject_Compare() API needs to adjusted accordingly to make use
of this new API.
Other changes include adapting some of the built-in functions (e.g.
cmp()) to use this API as well. Also, PyNumber_CoerceEx() will need to
check for new style numbers before calling the nb_coerce slot. New
style numbers don’t provide a coercion slot and thus cannot be
explicitly coerced.
Reference Implementation
A preliminary patch for the CVS version of Python is available through
the Source Forge patch manager [2].
Credits
This PEP and the patch are heavily based on work done by Marc-André
Lemburg [3].
Copyright
This document has been placed in the public domain.
References
[1]
http://www.lemburg.com/files/python/mxDateTime.html
[2]
http://sourceforge.net/patch/?func=detailpatch&patch_id=102652&group_id=5470
[3]
http://www.lemburg.com/files/python/CoercionProposal.html
| Final | PEP 208 – Reworking the Coercion Model | Standards Track | Many Python types implement numeric operations. When the arguments of
a numeric operation are of different types, the interpreter tries to
coerce the arguments into a common type. The numeric operation is
then performed using this common type. This PEP proposes a new type
flag to indicate that arguments to a type’s numeric operations should
not be coerced. Operations that do not support the supplied types
indicate it by returning a new singleton object. Types which do not
set the type flag are handled in a backwards compatible manner.
Allowing operations handle different types is often simpler, more
flexible, and faster than having the interpreter do coercion. |