This module defines a standard interface to break Uniform Resource Locator (URL)
strings up in components (addressing scheme, network location, path etc.), to
combine the components back into a URL string, and to convert a „relative URL”
to an absolute URL given a „base URL.”
The module has been designed to match the internet RFC on Relative Uniform
Resource Locators. It supports the following URL schemes: file, ftp,
gopher, hdl, http, https, imap, itms-services, mailto, mms,
news, nntp, prospero, rsync, rtsp, rtsps, rtspu,
sftp, shttp, sip, sips, snews, svn, svn+ssh,
telnet, wais, ws, wss.
Szczegół implementacyjny CPythona: The inclusion of the itms-services URL scheme can prevent an app from
passing Apple’s App Store review process for the macOS and iOS App Stores.
Handling for the itms-services scheme is always removed on iOS; on
macOS, it may be removed if CPython has been built with the
--with-app-store-compliance option.
The urllib.parse module defines functions that fall into two broad
categories: URL parsing and URL quoting. These are covered in detail in
the following sections.
This module’s functions use the deprecated term netloc (or net_loc),
which was introduced in RFC 1808. However, this term has been obsoleted by
RFC 3986, which introduced the term authority as its replacement.
The use of netloc is continued for backward compatibility.
Parse a URL into six components, returning a 6-item named tuple. This
corresponds to the general structure of a URL:
scheme://netloc/path;parameters?query#fragment.
Each tuple item is a string, possibly empty, or None if
missing_as_none is true.
Not defined component are represented an empty string (by default) or
None if missing_as_none is true.
The components are not broken up
into smaller parts (for example, the network location is a single string), and %
escapes are not expanded. The delimiters as shown above are not part of the
result, except for a leading slash in the path component, which is retained if
present. For example:
Following the syntax specifications in RFC 1808, urlparse recognizes
a netloc only if it is properly introduced by «//». Otherwise the
input is presumed to be a relative URL and thus to start with
a path component.
The scheme argument gives the default addressing scheme, to be
used only if the URL does not specify one. It should be the same type
(text or bytes) as urlstring or None, except that the '' is
always allowed, and is automatically converted to b'' if appropriate.
If the allow_fragments argument is false, fragment identifiers are not
recognized. Instead, they are parsed as part of the path, parameters
or query component, and fragment is set to None or the empty
string (depending on the value of missing_as_none) in the return value.
The return value is a named tuple, which means that its items can
be accessed by index or as named attributes, which are:
Reading the port attribute will raise a ValueError if
an invalid port is specified in the URL. See section
Structured Parse Results for more information on the result object.
Unmatched square brackets in the netloc attribute will raise a
ValueError.
Characters in the netloc attribute that decompose under NFKC
normalization (as used by the IDNA encoding) into any of /, ?,
#, @, or : will raise a ValueError. If the URL is
decomposed before parsing, no error will be raised.
As is the case with all named tuples, the subclass has a few additional methods
and attributes that are particularly useful. One such method is _replace().
The _replace() method will return a new ParseResult object replacing specified
fields with new values.
Zmienione w wersji 3.2: Added IPv6 URL parsing capabilities.
Zmienione w wersji 3.3: The fragment is now parsed for all URL schemes (unless allow_fragments is
false), in accordance with RFC 3986. Previously, an allowlist of
schemes that support fragments existed.
Zmienione w wersji 3.6: Out-of-range port numbers now raise ValueError, instead of
returning None.
Zmienione w wersji 3.8: Characters that affect netloc parsing under NFKC normalization will
now raise ValueError.
Zmienione w wersji 3.15.0a5 (unreleased): Added the missing_as_none parameter.
Parse a query string given as a string argument (data of type
application/x-www-form-urlencoded). Data are returned as a
dictionary. The dictionary keys are the unique query variable names and the
values are lists of values for each name.
The optional argument keep_blank_values is a flag indicating whether blank
values in percent-encoded queries should be treated as blank strings. A true value
indicates that blanks should be retained as blank strings. The default false
value indicates that blank values are to be ignored and treated as if they were
not included.
The optional argument strict_parsing is a flag indicating what to do with
parsing errors. If false (the default), errors are silently ignored. If true,
errors raise a ValueError exception.
The optional encoding and errors parameters specify how to decode
percent-encoded sequences into Unicode characters, as accepted by the
bytes.decode() method.
The optional argument max_num_fields is the maximum number of fields to
read. If set, then throws a ValueError if there are more than
max_num_fields fields read.
The optional argument separator is the symbol to use for separating the
query arguments. It defaults to &.
Use the urllib.parse.urlencode() function (with the doseq
parameter set to True) to convert such dictionaries into query
strings.
Zmienione w wersji 3.2: Add encoding and errors parameters.
Zmienione w wersji 3.8: Added max_num_fields parameter.
Zmienione w wersji 3.10: Added separator parameter with the default value of &. Python
versions earlier than Python 3.10 allowed using both ; and & as
query parameter separator. This has been changed to allow only a single
separator key, with & as the default separator.
Niezalecane od wersji 3.14: Accepting objects with false values (like 0 and []) except empty
strings and byte-like objects and None is now deprecated.
Parse a query string given as a string argument (data of type
application/x-www-form-urlencoded). Data are returned as a list of
name, value pairs.
The optional argument keep_blank_values is a flag indicating whether blank
values in percent-encoded queries should be treated as blank strings. A true value
indicates that blanks should be retained as blank strings. The default false
value indicates that blank values are to be ignored and treated as if they were
not included.
The optional argument strict_parsing is a flag indicating what to do with
parsing errors. If false (the default), errors are silently ignored. If true,
errors raise a ValueError exception.
The optional encoding and errors parameters specify how to decode
percent-encoded sequences into Unicode characters, as accepted by the
bytes.decode() method.
The optional argument max_num_fields is the maximum number of fields to
read. If set, then throws a ValueError if there are more than
max_num_fields fields read.
The optional argument separator is the symbol to use for separating the
query arguments. It defaults to &.
Zmienione w wersji 3.2: Add encoding and errors parameters.
Zmienione w wersji 3.8: Added max_num_fields parameter.
Zmienione w wersji 3.10: Added separator parameter with the default value of &. Python
versions earlier than Python 3.10 allowed using both ; and & as
query parameter separator. This has been changed to allow only a single
separator key, with & as the default separator.
Construct a URL from a tuple as returned by urlparse(). The parts
argument can be any six-item iterable.
This may result in a slightly different, but equivalent URL, if the
URL that was parsed originally had unnecessary delimiters (for example,
a ? with an empty query; the RFC states that these are equivalent).
If keep_empty is true, empty strings are kept in the result (for example,
a ? for an empty query), only None components are omitted.
This allows rebuilding a URL that was parsed with option
missing_as_none=True.
By default, keep_empty is true if parts is the result of the
urlparse() call with missing_as_none=True.
Zmienione w wersji 3.15.0a5 (unreleased): Added the keep_empty parameter.
This is similar to urlparse(), but does not split the params from the URL.
This should generally be used instead of urlparse() if the more recent URL
syntax allowing parameters to be applied to each segment of the path portion
of the URL (see RFC 2396) is wanted. A separate function is needed to
separate the path segments and parameters. This function returns a 5-item
named tuple:
Reading the port attribute will raise a ValueError if
an invalid port is specified in the URL. See section
Structured Parse Results for more information on the result object.
Unmatched square brackets in the netloc attribute will raise a
ValueError.
Characters in the netloc attribute that decompose under NFKC
normalization (as used by the IDNA encoding) into any of /, ?,
#, @, or : will raise a ValueError. If the URL is
decomposed before parsing, no error will be raised.
Following some of the WHATWG spec that updates RFC 3986, leading C0
control and space characters are stripped from the URL. \n,
\r and tab \t characters are removed from the URL at any position.
Combine the elements of a tuple as returned by urlsplit() into a
complete URL as a string. The parts argument can be any five-item
iterable.
This may result in a slightly different, but equivalent URL, if the
URL that was parsed originally had unnecessary delimiters (for example,
a ? with an empty query; the RFC states that these are equivalent).
If keep_empty is true, empty strings are kept in the result (for example,
a ? for an empty query), only None components are omitted.
This allows rebuilding a URL that was parsed with option
missing_as_none=True.
By default, keep_empty is true if parts is the result of the
urlsplit() call with missing_as_none=True.
Zmienione w wersji 3.15.0a5 (unreleased): Added the keep_empty parameter.
Construct a full („absolute”) URL by combining a „base URL” (base) with
another URL (url). Informally, this uses components of the base URL, in
particular the addressing scheme, the network location and (part of) the
path, to provide missing components in the relative URL. For example:
If you do not want that behavior, preprocess the url with urlsplit() and
urlunsplit(), removing possible scheme and netloc parts.
Ostrzeżenie
Because an absolute URL may be passed as the url parameter, it is
generally not secure to use urljoin with an attacker-controlled
url. For example in,
urljoin("https://website.com/users/",username), if username can
contain an absolute URL, the result of urljoin will be the absolute
URL.
Zmienione w wersji 3.5: Behavior updated to match the semantics defined in RFC 3986.
If url contains a fragment identifier, return a modified version of url
with no fragment identifier, and the fragment identifier as a separate
string. If there is no fragment identifier in url, return url unmodified
and an empty string (by default) or None if missing_as_none is true.
The return value is a named tuple, its items can be accessed by index
or as named attributes:
Extract the url from a wrapped URL (that is, a string formatted as
<URL:scheme://host/path>, <scheme://host/path>, URL:scheme://host/path
or scheme://host/path). If url is not a wrapped URL, it is returned
without changes.
The urlsplit() and urlparse() APIs do not perform validation of
inputs. They may not raise errors on inputs that other applications consider
invalid. They may also succeed on some inputs that might not be considered
URLs elsewhere. Their purpose is for practical functionality rather than
purity.
Instead of raising an exception on unusual input, they may instead return some
component parts as empty strings or None (depending on the value of the
missing_as_none argument).
Or components may contain more than perhaps they should.
We recommend that users of these APIs where the values may be used anywhere
with security implications code defensively. Do some verification within your
code before trusting a returned component part. Does that scheme make
sense? Is that a sensible path? Is there anything strange about that
hostname? etc.
What constitutes a URL is not universally well defined. Different applications
have different needs and desired constraints. For instance the living WHATWG
spec describes what user facing web clients such as a web browser require.
While RFC 3986 is more general. These functions incorporate some aspects of
both, but cannot be claimed compliant with either. The APIs and existing user
code with expectations on specific behaviors predate both standards leading us
to be very cautious about making API behavior changes.
The URL parsing functions were originally designed to operate on character
strings only. In practice, it is useful to be able to manipulate properly
quoted and encoded URLs as sequences of ASCII bytes. Accordingly, the
URL parsing functions in this module all operate on bytes and
bytearray objects in addition to str objects.
If str data is passed in, the result will also contain only
str data. If bytes or bytearray data is
passed in, the result will contain only bytes data.
Attempting to mix str data with bytes or
bytearray in a single function call will result in a
TypeError being raised, while attempting to pass in non-ASCII
byte values will trigger UnicodeDecodeError.
To support easier conversion of result objects between str and
bytes, all return values from URL parsing functions provide
either an encode() method (when the result contains str
data) or a decode() method (when the result contains bytes
data). The signatures of these methods match those of the corresponding
str and bytes methods (except that the default encoding
is 'ascii' rather than 'utf-8'). Each produces a value of a
corresponding type that contains either bytes data (for
encode() methods) or str data (for
decode() methods).
Applications that need to operate on potentially improperly quoted URLs
that may contain non-ASCII data will need to do their own decoding from
bytes to characters before invoking the URL parsing methods.
The behaviour described in this section applies only to the URL parsing
functions. The URL quoting functions use their own rules when producing
or consuming byte sequences as detailed in the documentation of the
individual URL quoting functions.
Zmienione w wersji 3.2: URL parsing functions now accept ASCII encoded byte sequences
The result objects from the urlparse(), urlsplit() and
urldefrag() functions are subclasses of the tuple type.
These subclasses add the attributes listed in the documentation for
those functions, the encoding and decoding support described in the
previous section, as well as an additional method:
Return the re-combined version of the original URL as a string. This may
differ from the original URL in that the scheme may be normalized to lower
case and empty components may be dropped. Specifically, empty parameters,
queries, and fragment identifiers will be removed unless the URL was parsed
with missing_as_none=True.
For urldefrag() results, only empty fragment identifiers will be removed.
For urlsplit() and urlparse() results, all noted changes will be
made to the URL returned by this method.
The result of this method remains unchanged if passed back through the original
parsing function:
The URL quoting functions focus on taking program data and making it safe
for use as URL components by quoting special characters and appropriately
encoding non-ASCII text. They also support reversing these operations to
recreate the original data from the contents of a URL component if that
task isn’t already covered by the URL parsing functions above.
Replace special characters in string using the %xx escape. Letters,
digits, and the characters '_.-~' are never quoted. By default, this
function is intended for quoting the path section of a URL. The optional
safe parameter specifies additional ASCII characters that should not be
quoted — its default value is '/'.
Zmienione w wersji 3.7: Moved from RFC 2396 to RFC 3986 for quoting URL strings. „~” is now
included in the set of unreserved characters.
The optional encoding and errors parameters specify how to deal with
non-ASCII characters, as accepted by the str.encode() method.
encoding defaults to 'utf-8'.
errors defaults to 'strict', meaning unsupported characters raise a
UnicodeEncodeError.
encoding and errors must not be supplied if string is a
bytes, or a TypeError is raised.
Note that quote(string,safe,encoding,errors) is equivalent to
quote_from_bytes(string.encode(encoding,errors),safe).
Like quote(), but also replace spaces with plus signs, as required for
quoting HTML form values when building up a query string to go into a URL.
Plus signs in the original string are escaped unless they are included in
safe. It also does not have safe default to '/'.
Replace %xx escapes with their single-character equivalent.
The optional encoding and errors parameters specify how to decode
percent-encoded sequences into Unicode characters, as accepted by the
bytes.decode() method.
Convert a mapping object or a sequence of two-element tuples, which may
contain str or bytes objects, to a percent-encoded ASCII
text string. If the resultant string is to be used as a data for POST
operation with the urlopen() function, then
it should be encoded to bytes, otherwise it would result in a
TypeError.
The resulting string is a series of key=value pairs separated by '&'
characters, where both key and value are quoted using the quote_via
function. By default, quote_plus() is used to quote the values, which
means spaces are quoted as a '+' character and «/» characters are
encoded as %2F, which follows the standard for GET requests
(application/x-www-form-urlencoded). An alternate function that can be
passed as quote_via is quote(), which will encode spaces as %20
and not encode «/» characters. For maximum control of what is quoted, use
quote and specify a value for safe.
When a sequence of two-element tuples is used as the query
argument, the first element of each tuple is a key and the second is a
value. The value element in itself can be a sequence and in that case, if
the optional parameter doseq evaluates to True, individual
key=value pairs separated by '&' are generated for each element of
the value sequence for the key. The order of parameters in the encoded
string will match the order of parameter tuples in the sequence.
The safe, encoding, and errors parameters are passed down to
quote_via (the encoding and errors parameters are only passed
when a query element is a str).
To reverse this encoding process, parse_qs() and parse_qsl() are
provided in this module to parse query strings into Python data structures.
Refer to urllib examples to find out how the
urllib.parse.urlencode() method can be used for generating the query
string of a URL or data for a POST request.
Zmienione w wersji 3.2: query supports bytes and string objects.
Zmienione w wersji 3.5: Added the quote_via parameter.
Niezalecane od wersji 3.14: Accepting objects with false values (like 0 and []) except empty
strings and byte-like objects and None is now deprecated.
This is the current standard (STD66). Any changes to urllib.parse module
should conform to this. Certain deviations could be observed, which are
mostly for backward compatibility purposes and for certain de-facto
parsing requirements as commonly observed in major browsers.
RFC 2732 - Format for Literal IPv6 Addresses in URL’s.
This specifies the parsing requirements of IPv6 URLs.
This Request For Comments includes the rules for joining an absolute and a
relative URL, including a fair number of „Abnormal Examples” which govern the
treatment of border cases.