mirror of
https://github.com/nlohmann/json.git
synced 2024-11-27 16:49:02 +08:00
⬆️ update cpplint (#3225)
This commit is contained in:
parent
29cd970b94
commit
e6bf789f7d
8
third_party/cpplint/README.rst
vendored
8
third_party/cpplint/README.rst
vendored
@ -26,9 +26,10 @@ cpplint - static code checker for C++
|
||||
:target: https://pypi.python.org/pypi/cpplint
|
||||
|
||||
Cpplint is a command-line tool to check C/C++ files for style issues following `Google's C++ style guide <http://google.github.io/styleguide/cppguide.html>`_.
|
||||
Cpplint is developed and maintained by Google Inc. at `google/styleguide <https://github.com/google/styleguide>`_, also see see the `wikipedia entry <http://en.wikipedia.org/wiki/Cpplint>`_
|
||||
Cpplint is developed and maintained by Google Inc. at `google/styleguide <https://github.com/google/styleguide>`_, also see the `wikipedia entry <http://en.wikipedia.org/wiki/Cpplint>`_
|
||||
|
||||
While Google maintains cpplint, Google is not (very) responsive to issues and pull requests, this fork aims to be (somewhat) more open to add fixes to cpplint to enable fixes, when those fixes make cpplint usable in wider contexts.
|
||||
Also see discussion here https://github.com/google/styleguide/pull/528.
|
||||
|
||||
|
||||
Installation
|
||||
@ -56,10 +57,10 @@ For full usage instructions, run:
|
||||
Changes
|
||||
-------
|
||||
|
||||
The modifications in this fork are minor fixes and cosmetic changes:
|
||||
The modifications in this fork are minor fixes and cosmetic changes, such as:
|
||||
|
||||
* more default extensions
|
||||
* python 3 compatibility
|
||||
* more default file extensions
|
||||
* customizable file extensions with the --extensions argument
|
||||
* continuous integration on travis
|
||||
* support for recursive file discovery via the --recursive argument
|
||||
@ -67,6 +68,7 @@ The modifications in this fork are minor fixes and cosmetic changes:
|
||||
* JUnit XML output format
|
||||
* Overriding repository root auto-detection via --repository
|
||||
* Support ``#pragma once`` as an alternative to header include guards
|
||||
* ... and a few more (most of which are open PRs on upstream)
|
||||
|
||||
|
||||
Acknowledgements
|
||||
|
516
third_party/cpplint/cpplint.py
vendored
516
third_party/cpplint/cpplint.py
vendored
@ -59,7 +59,7 @@ import xml.etree.ElementTree
|
||||
# if empty, use defaults
|
||||
_valid_extensions = set([])
|
||||
|
||||
__VERSION__ = '1.4.4'
|
||||
__VERSION__ = '1.5.5'
|
||||
|
||||
try:
|
||||
xrange # Python 2
|
||||
@ -69,7 +69,7 @@ except NameError:
|
||||
|
||||
|
||||
_USAGE = """
|
||||
Syntax: cpplint.py [--verbose=#] [--output=emacs|eclipse|vs7|junit]
|
||||
Syntax: cpplint.py [--verbose=#] [--output=emacs|eclipse|vs7|junit|sed|gsed]
|
||||
[--filter=-x,+y,...]
|
||||
[--counting=total|toplevel|detailed] [--root=subdir]
|
||||
[--repository=path]
|
||||
@ -77,6 +77,7 @@ Syntax: cpplint.py [--verbose=#] [--output=emacs|eclipse|vs7|junit]
|
||||
[--recursive]
|
||||
[--exclude=path]
|
||||
[--extensions=hpp,cpp,...]
|
||||
[--includeorder=default|standardcfirst]
|
||||
[--quiet]
|
||||
[--version]
|
||||
<file> [file] ...
|
||||
@ -102,11 +103,16 @@ Syntax: cpplint.py [--verbose=#] [--output=emacs|eclipse|vs7|junit]
|
||||
|
||||
Flags:
|
||||
|
||||
output=emacs|eclipse|vs7|junit
|
||||
output=emacs|eclipse|vs7|junit|sed|gsed
|
||||
By default, the output is formatted to ease emacs parsing. Visual Studio
|
||||
compatible output (vs7) may also be used. Further support exists for
|
||||
eclipse (eclipse), and JUnit (junit). XML parsers such as those used
|
||||
in Jenkins and Bamboo may also be used. Other formats are unsupported.
|
||||
in Jenkins and Bamboo may also be used.
|
||||
The sed format outputs sed commands that should fix some of the errors.
|
||||
Note that this requires gnu sed. If that is installed as gsed on your
|
||||
system (common e.g. on macOS with homebrew) you can use the gsed output
|
||||
format. Sed commands are written to stdout, not stderr, so you should be
|
||||
able to pipe output straight to a shell to run the fixes.
|
||||
|
||||
verbose=#
|
||||
Specify a number 0-5 to restrict errors to certain verbosity levels.
|
||||
@ -121,11 +127,11 @@ Syntax: cpplint.py [--verbose=#] [--output=emacs|eclipse|vs7|junit]
|
||||
error messages whose category names pass the filters will be printed.
|
||||
(Category names are printed with the message and look like
|
||||
"[whitespace/indent]".) Filters are evaluated left to right.
|
||||
"-FOO" and "FOO" means "do not print categories that start with FOO".
|
||||
"-FOO" means "do not print categories that start with FOO".
|
||||
"+FOO" means "do print categories that start with FOO".
|
||||
|
||||
Examples: --filter=-whitespace,+whitespace/braces
|
||||
--filter=whitespace,runtime/printf,+runtime/printf_format
|
||||
--filter=-whitespace,-runtime/printf,+runtime/printf_format
|
||||
--filter=-,+build/include_what_you_use
|
||||
|
||||
To see a list of all the categories used in cpplint, pass no arg:
|
||||
@ -209,6 +215,15 @@ Syntax: cpplint.py [--verbose=#] [--output=emacs|eclipse|vs7|junit]
|
||||
Examples:
|
||||
--extensions=%s
|
||||
|
||||
includeorder=default|standardcfirst
|
||||
For the build/include_order rule, the default is to blindly assume angle
|
||||
bracket includes with file extension are c-system-headers (default),
|
||||
even knowing this will have false classifications.
|
||||
The default is established at google.
|
||||
standardcfirst means to instead use an allow-list of known c headers and
|
||||
treat all others as separate group of "other system headers". The C headers
|
||||
included are those of the C-standard lib and closely related ones.
|
||||
|
||||
headers=x,y,...
|
||||
The header extensions that cpplint will treat as .h in checks. Values are
|
||||
automatically added to --extensions list.
|
||||
@ -282,6 +297,7 @@ _ERROR_CATEGORIES = [
|
||||
'build/include_alpha',
|
||||
'build/include_order',
|
||||
'build/include_what_you_use',
|
||||
'build/namespaces_headers',
|
||||
'build/namespaces_literals',
|
||||
'build/namespaces',
|
||||
'build/printf_format',
|
||||
@ -338,6 +354,13 @@ _ERROR_CATEGORIES = [
|
||||
'whitespace/todo',
|
||||
]
|
||||
|
||||
# keywords to use with --outputs which generate stdout for machine processing
|
||||
_MACHINE_OUTPUTS = [
|
||||
'junit',
|
||||
'sed',
|
||||
'gsed'
|
||||
]
|
||||
|
||||
# These error categories are no longer enforced by cpplint, but for backwards-
|
||||
# compatibility they may still appear in NOLINT comments.
|
||||
_LEGACY_ERROR_CATEGORIES = [
|
||||
@ -514,6 +537,186 @@ _CPP_HEADERS = frozenset([
|
||||
'cwctype',
|
||||
])
|
||||
|
||||
# C headers
|
||||
_C_HEADERS = frozenset([
|
||||
# System C headers
|
||||
'assert.h',
|
||||
'complex.h',
|
||||
'ctype.h',
|
||||
'errno.h',
|
||||
'fenv.h',
|
||||
'float.h',
|
||||
'inttypes.h',
|
||||
'iso646.h',
|
||||
'limits.h',
|
||||
'locale.h',
|
||||
'math.h',
|
||||
'setjmp.h',
|
||||
'signal.h',
|
||||
'stdalign.h',
|
||||
'stdarg.h',
|
||||
'stdatomic.h',
|
||||
'stdbool.h',
|
||||
'stddef.h',
|
||||
'stdint.h',
|
||||
'stdio.h',
|
||||
'stdlib.h',
|
||||
'stdnoreturn.h',
|
||||
'string.h',
|
||||
'tgmath.h',
|
||||
'threads.h',
|
||||
'time.h',
|
||||
'uchar.h',
|
||||
'wchar.h',
|
||||
'wctype.h',
|
||||
# additional POSIX C headers
|
||||
'aio.h',
|
||||
'arpa/inet.h',
|
||||
'cpio.h',
|
||||
'dirent.h',
|
||||
'dlfcn.h',
|
||||
'fcntl.h',
|
||||
'fmtmsg.h',
|
||||
'fnmatch.h',
|
||||
'ftw.h',
|
||||
'glob.h',
|
||||
'grp.h',
|
||||
'iconv.h',
|
||||
'langinfo.h',
|
||||
'libgen.h',
|
||||
'monetary.h',
|
||||
'mqueue.h',
|
||||
'ndbm.h',
|
||||
'net/if.h',
|
||||
'netdb.h',
|
||||
'netinet/in.h',
|
||||
'netinet/tcp.h',
|
||||
'nl_types.h',
|
||||
'poll.h',
|
||||
'pthread.h',
|
||||
'pwd.h',
|
||||
'regex.h',
|
||||
'sched.h',
|
||||
'search.h',
|
||||
'semaphore.h',
|
||||
'setjmp.h',
|
||||
'signal.h',
|
||||
'spawn.h',
|
||||
'strings.h',
|
||||
'stropts.h',
|
||||
'syslog.h',
|
||||
'tar.h',
|
||||
'termios.h',
|
||||
'trace.h',
|
||||
'ulimit.h',
|
||||
'unistd.h',
|
||||
'utime.h',
|
||||
'utmpx.h',
|
||||
'wordexp.h',
|
||||
# additional GNUlib headers
|
||||
'a.out.h',
|
||||
'aliases.h',
|
||||
'alloca.h',
|
||||
'ar.h',
|
||||
'argp.h',
|
||||
'argz.h',
|
||||
'byteswap.h',
|
||||
'crypt.h',
|
||||
'endian.h',
|
||||
'envz.h',
|
||||
'err.h',
|
||||
'error.h',
|
||||
'execinfo.h',
|
||||
'fpu_control.h',
|
||||
'fstab.h',
|
||||
'fts.h',
|
||||
'getopt.h',
|
||||
'gshadow.h',
|
||||
'ieee754.h',
|
||||
'ifaddrs.h',
|
||||
'libintl.h',
|
||||
'mcheck.h',
|
||||
'mntent.h',
|
||||
'obstack.h',
|
||||
'paths.h',
|
||||
'printf.h',
|
||||
'pty.h',
|
||||
'resolv.h',
|
||||
'shadow.h',
|
||||
'sysexits.h',
|
||||
'ttyent.h',
|
||||
# Additional linux glibc headers
|
||||
'dlfcn.h',
|
||||
'elf.h',
|
||||
'features.h',
|
||||
'gconv.h',
|
||||
'gnu-versions.h',
|
||||
'lastlog.h',
|
||||
'libio.h',
|
||||
'link.h',
|
||||
'malloc.h',
|
||||
'memory.h',
|
||||
'netash/ash.h',
|
||||
'netatalk/at.h',
|
||||
'netax25/ax25.h',
|
||||
'neteconet/ec.h',
|
||||
'netipx/ipx.h',
|
||||
'netiucv/iucv.h',
|
||||
'netpacket/packet.h',
|
||||
'netrom/netrom.h',
|
||||
'netrose/rose.h',
|
||||
'nfs/nfs.h',
|
||||
'nl_types.h',
|
||||
'nss.h',
|
||||
're_comp.h',
|
||||
'regexp.h',
|
||||
'sched.h',
|
||||
'sgtty.h',
|
||||
'stab.h',
|
||||
'stdc-predef.h',
|
||||
'stdio_ext.h',
|
||||
'syscall.h',
|
||||
'termio.h',
|
||||
'thread_db.h',
|
||||
'ucontext.h',
|
||||
'ustat.h',
|
||||
'utmp.h',
|
||||
'values.h',
|
||||
'wait.h',
|
||||
'xlocale.h',
|
||||
# Hardware specific headers
|
||||
'arm_neon.h',
|
||||
'emmintrin.h',
|
||||
'xmmintin.h',
|
||||
])
|
||||
|
||||
# Folders of C libraries so commonly used in C++,
|
||||
# that they have parity with standard C libraries.
|
||||
C_STANDARD_HEADER_FOLDERS = frozenset([
|
||||
# standard C library
|
||||
"sys",
|
||||
# glibc for linux
|
||||
"arpa",
|
||||
"asm-generic",
|
||||
"bits",
|
||||
"gnu",
|
||||
"net",
|
||||
"netinet",
|
||||
"protocols",
|
||||
"rpc",
|
||||
"rpcsvc",
|
||||
"scsi",
|
||||
# linux kernel header
|
||||
"drm",
|
||||
"linux",
|
||||
"misc",
|
||||
"mtd",
|
||||
"rdma",
|
||||
"sound",
|
||||
"video",
|
||||
"xen",
|
||||
])
|
||||
|
||||
# Type names
|
||||
_TYPES = re.compile(
|
||||
r'^(?:'
|
||||
@ -600,9 +803,10 @@ _ALT_TOKEN_REPLACEMENT_PATTERN = re.compile(
|
||||
# _IncludeState.CheckNextIncludeOrder().
|
||||
_C_SYS_HEADER = 1
|
||||
_CPP_SYS_HEADER = 2
|
||||
_LIKELY_MY_HEADER = 3
|
||||
_POSSIBLE_MY_HEADER = 4
|
||||
_OTHER_HEADER = 5
|
||||
_OTHER_SYS_HEADER = 3
|
||||
_LIKELY_MY_HEADER = 4
|
||||
_POSSIBLE_MY_HEADER = 5
|
||||
_OTHER_HEADER = 6
|
||||
|
||||
# These constants define the current inline assembly state
|
||||
_NO_ASM = 0 # Outside of inline assembly block
|
||||
@ -622,6 +826,22 @@ _SEARCH_C_FILE = re.compile(r'\b(?:LINT_C_FILE|'
|
||||
# Match string that indicates we're working on a Linux Kernel file.
|
||||
_SEARCH_KERNEL_FILE = re.compile(r'\b(?:LINT_KERNEL_FILE)')
|
||||
|
||||
# Commands for sed to fix the problem
|
||||
_SED_FIXUPS = {
|
||||
'Remove spaces around =': r's/ = /=/',
|
||||
'Remove spaces around !=': r's/ != /!=/',
|
||||
'Remove space before ( in if (': r's/if (/if(/',
|
||||
'Remove space before ( in for (': r's/for (/for(/',
|
||||
'Remove space before ( in while (': r's/while (/while(/',
|
||||
'Remove space before ( in switch (': r's/switch (/switch(/',
|
||||
'Should have a space between // and comment': r's/\/\//\/\/ /',
|
||||
'Missing space before {': r's/\([^ ]\){/\1 {/',
|
||||
'Tab found, replace by spaces': r's/\t/ /g',
|
||||
'Line ends in whitespace. Consider deleting these extra spaces.': r's/\s*$//',
|
||||
'You don\'t need a ; after a }': r's/};/}/',
|
||||
'Missing space after ,': r's/,\([^ ]\)/, \1/g',
|
||||
}
|
||||
|
||||
_regexp_compile_cache = {}
|
||||
|
||||
# {str, set(int)}: a map from error categories to sets of linenumbers
|
||||
@ -641,13 +861,16 @@ _repository = None
|
||||
# Files to exclude from linting. This is set by the --exclude flag.
|
||||
_excludes = None
|
||||
|
||||
# Whether to supress PrintInfo messages
|
||||
# Whether to supress all PrintInfo messages, UNRELATED to --quiet flag
|
||||
_quiet = False
|
||||
|
||||
# The allowed line length of files.
|
||||
# This is set by --linelength flag.
|
||||
_line_length = 80
|
||||
|
||||
# This allows to use different include order rule than default
|
||||
_include_order = "default"
|
||||
|
||||
try:
|
||||
unicode
|
||||
except NameError:
|
||||
@ -678,7 +901,7 @@ def unicode_escape_decode(x):
|
||||
|
||||
# Treat all headers starting with 'h' equally: .h, .hpp, .hxx etc.
|
||||
# This is set by --headers flag.
|
||||
_hpp_headers = set(['h', 'hh', 'hpp', 'hxx', 'h++', 'cuh'])
|
||||
_hpp_headers = set([])
|
||||
|
||||
# {str, bool}: a map from error categories to booleans which indicate if the
|
||||
# category should be suppressed for every line.
|
||||
@ -687,30 +910,48 @@ _global_error_suppressions = {}
|
||||
def ProcessHppHeadersOption(val):
|
||||
global _hpp_headers
|
||||
try:
|
||||
_hpp_headers = set(val.split(','))
|
||||
# Automatically append to extensions list so it does not have to be set 2 times
|
||||
_valid_extensions.update(_hpp_headers)
|
||||
_hpp_headers = {ext.strip() for ext in val.split(',')}
|
||||
except ValueError:
|
||||
PrintUsage('Header extensions must be comma separated list.')
|
||||
|
||||
def ProcessIncludeOrderOption(val):
|
||||
if val is None or val == "default":
|
||||
pass
|
||||
elif val == "standardcfirst":
|
||||
global _include_order
|
||||
_include_order = val
|
||||
else:
|
||||
PrintUsage('Invalid includeorder value %s. Expected default|standardcfirst')
|
||||
|
||||
def IsHeaderExtension(file_extension):
|
||||
return file_extension in _hpp_headers
|
||||
return file_extension in GetHeaderExtensions()
|
||||
|
||||
def GetHeaderExtensions():
|
||||
return _hpp_headers or ['h']
|
||||
if _hpp_headers:
|
||||
return _hpp_headers
|
||||
if _valid_extensions:
|
||||
return {h for h in _valid_extensions if 'h' in h}
|
||||
return set(['h', 'hh', 'hpp', 'hxx', 'h++', 'cuh'])
|
||||
|
||||
# The allowed extensions for file names
|
||||
# This is set by --extensions flag
|
||||
def GetAllExtensions():
|
||||
if not _valid_extensions:
|
||||
return GetHeaderExtensions().union(set(['c', 'cc', 'cpp', 'cxx', 'c++', 'cu']))
|
||||
return _valid_extensions
|
||||
return GetHeaderExtensions().union(_valid_extensions or set(
|
||||
['c', 'cc', 'cpp', 'cxx', 'c++', 'cu']))
|
||||
|
||||
def ProcessExtensionsOption(val):
|
||||
global _valid_extensions
|
||||
try:
|
||||
extensions = [ext.strip() for ext in val.split(',')]
|
||||
_valid_extensions = set(extensions)
|
||||
except ValueError:
|
||||
PrintUsage('Extensions should be a comma-separated list of values;'
|
||||
'for example: extensions=hpp,cpp\n'
|
||||
'This could not be parsed: "%s"' % (val,))
|
||||
|
||||
def GetNonHeaderExtensions():
|
||||
return GetAllExtensions().difference(GetHeaderExtensions())
|
||||
|
||||
|
||||
|
||||
def ParseNolintSuppressions(filename, raw_line, linenum, error):
|
||||
"""Updates the global list of line error-suppressions.
|
||||
|
||||
@ -843,11 +1084,13 @@ class _IncludeState(object):
|
||||
_MY_H_SECTION = 1
|
||||
_C_SECTION = 2
|
||||
_CPP_SECTION = 3
|
||||
_OTHER_H_SECTION = 4
|
||||
_OTHER_SYS_SECTION = 4
|
||||
_OTHER_H_SECTION = 5
|
||||
|
||||
_TYPE_NAMES = {
|
||||
_C_SYS_HEADER: 'C system header',
|
||||
_CPP_SYS_HEADER: 'C++ system header',
|
||||
_OTHER_SYS_HEADER: 'other system header',
|
||||
_LIKELY_MY_HEADER: 'header this file implements',
|
||||
_POSSIBLE_MY_HEADER: 'header this file may implement',
|
||||
_OTHER_HEADER: 'other header',
|
||||
@ -857,6 +1100,7 @@ class _IncludeState(object):
|
||||
_MY_H_SECTION: 'a header this file implements',
|
||||
_C_SECTION: 'C system header',
|
||||
_CPP_SECTION: 'C++ system header',
|
||||
_OTHER_SYS_SECTION: 'other system header',
|
||||
_OTHER_H_SECTION: 'other header',
|
||||
}
|
||||
|
||||
@ -970,6 +1214,12 @@ class _IncludeState(object):
|
||||
else:
|
||||
self._last_header = ''
|
||||
return error_message
|
||||
elif header_type == _OTHER_SYS_HEADER:
|
||||
if self._section <= self._OTHER_SYS_SECTION:
|
||||
self._section = self._OTHER_SYS_SECTION
|
||||
else:
|
||||
self._last_header = ''
|
||||
return error_message
|
||||
elif header_type == _LIKELY_MY_HEADER:
|
||||
if self._section <= self._MY_H_SECTION:
|
||||
self._section = self._MY_H_SECTION
|
||||
@ -1011,6 +1261,8 @@ class _CppLintState(object):
|
||||
# "eclipse" - format that eclipse can parse
|
||||
# "vs7" - format that Microsoft Visual Studio 7 can parse
|
||||
# "junit" - format that Jenkins, Bamboo, etc can parse
|
||||
# "sed" - returns a gnu sed command to fix the problem
|
||||
# "gsed" - like sed, but names the command gsed, e.g. for macOS homebrew users
|
||||
self.output_format = 'emacs'
|
||||
|
||||
# For JUnit output, save errors and failures until the end so that they
|
||||
@ -1099,7 +1351,9 @@ class _CppLintState(object):
|
||||
self.PrintInfo('Total errors found: %d\n' % self.error_count)
|
||||
|
||||
def PrintInfo(self, message):
|
||||
if not _quiet and self.output_format != 'junit':
|
||||
# _quiet does not represent --quiet flag.
|
||||
# Hide infos from stdout to keep stdout pure for machine consumption
|
||||
if not _quiet and self.output_format not in _MACHINE_OUTPUTS:
|
||||
sys.stdout.write(message)
|
||||
|
||||
def PrintError(self, message):
|
||||
@ -1117,9 +1371,9 @@ class _CppLintState(object):
|
||||
num_failures = len(self._junit_failures)
|
||||
|
||||
testsuite = xml.etree.ElementTree.Element('testsuite')
|
||||
testsuite.attrib['name'] = 'cpplint'
|
||||
testsuite.attrib['errors'] = str(num_errors)
|
||||
testsuite.attrib['failures'] = str(num_failures)
|
||||
testsuite.attrib['name'] = 'cpplint'
|
||||
|
||||
if num_errors == 0 and num_failures == 0:
|
||||
testsuite.attrib['tests'] = str(1)
|
||||
@ -1313,7 +1567,7 @@ class FileInfo(object):
|
||||
If we have a real absolute path name here we can try to do something smart:
|
||||
detecting the root of the checkout and truncating /path/to/checkout from
|
||||
the name so that we get header guards that don't include things like
|
||||
"C:\Documents and Settings\..." or "/home/username/..." in them and thus
|
||||
"C:\\Documents and Settings\\..." or "/home/username/..." in them and thus
|
||||
people on different computers who have checked the source out to different
|
||||
locations won't see bogus errors.
|
||||
"""
|
||||
@ -1459,6 +1713,13 @@ def Error(filename, linenum, category, confidence, message):
|
||||
elif _cpplint_state.output_format == 'junit':
|
||||
_cpplint_state.AddJUnitFailure(filename, linenum, message, category,
|
||||
confidence)
|
||||
elif _cpplint_state.output_format in ['sed', 'gsed']:
|
||||
if message in _SED_FIXUPS:
|
||||
sys.stdout.write(_cpplint_state.output_format + " -i '%s%s' %s # %s [%s] [%d]\n" % (
|
||||
linenum, _SED_FIXUPS[message], filename, message, category, confidence))
|
||||
else:
|
||||
sys.stderr.write('# %s:%s: "%s" [%s] [%d]\n' % (
|
||||
filename, linenum, message, category, confidence))
|
||||
else:
|
||||
final_message = '%s:%s: %s [%s] [%d]\n' % (
|
||||
filename, linenum, message, category, confidence)
|
||||
@ -1599,7 +1860,7 @@ def FindNextMultiLineCommentEnd(lines, lineix):
|
||||
|
||||
def RemoveMultiLineCommentsFromRange(lines, begin, end):
|
||||
"""Clears a range of lines for multi-line comments."""
|
||||
# Having // dummy comments makes the lines non-empty, so we will not get
|
||||
# Having // <empty> comments makes the lines non-empty, so we will not get
|
||||
# unnecessary blank line warnings later in the code.
|
||||
for i in range(begin, end):
|
||||
lines[i] = '/**/'
|
||||
@ -1973,7 +2234,7 @@ def CheckForCopyright(filename, lines, error):
|
||||
"""Logs an error if no Copyright message appears at the top of the file."""
|
||||
|
||||
# We'll say it should occur by line 10. Don't forget there's a
|
||||
# dummy line at the front.
|
||||
# placeholder line at the front.
|
||||
for line in xrange(1, min(len(lines), 11)):
|
||||
if re.search(r'Copyright', lines[line], re.I): break
|
||||
else: # means no copyright line was found
|
||||
@ -2076,7 +2337,8 @@ def GetHeaderGuardCPPVariable(filename):
|
||||
|
||||
# --root=.. , will prepend the outer directory to the header guard
|
||||
full_path = fileinfo.FullName()
|
||||
root_abspath = os.path.abspath(_root)
|
||||
# adapt slashes for windows
|
||||
root_abspath = os.path.abspath(_root).replace('\\', '/')
|
||||
|
||||
maybe_path = StripListPrefix(PathSplitToList(full_path),
|
||||
PathSplitToList(root_abspath))
|
||||
@ -2216,16 +2478,22 @@ def CheckHeaderFileIncluded(filename, include_state, error):
|
||||
continue
|
||||
headername = FileInfo(headerfile).RepositoryName()
|
||||
first_include = None
|
||||
include_uses_unix_dir_aliases = False
|
||||
for section_list in include_state.include_list:
|
||||
for f in section_list:
|
||||
if headername in f[0] or f[0] in headername:
|
||||
include_text = f[0]
|
||||
if "./" in include_text:
|
||||
include_uses_unix_dir_aliases = True
|
||||
if headername in include_text or include_text in headername:
|
||||
return
|
||||
if not first_include:
|
||||
first_include = f[1]
|
||||
|
||||
error(filename, first_include, 'build/include', 5,
|
||||
'%s should include its header file %s' % (fileinfo.RepositoryName(),
|
||||
headername))
|
||||
message = '%s should include its header file %s' % (fileinfo.RepositoryName(), headername)
|
||||
if include_uses_unix_dir_aliases:
|
||||
message += ". Relative paths like . and .. are not allowed."
|
||||
|
||||
error(filename, first_include, 'build/include', 5, message)
|
||||
|
||||
|
||||
def CheckForBadCharacters(filename, lines, error):
|
||||
@ -2876,7 +3144,7 @@ class NestingState(object):
|
||||
# };
|
||||
class_decl_match = Match(
|
||||
r'^(\s*(?:template\s*<[\w\s<>,:=]*>\s*)?'
|
||||
r'(class|struct)\s+(?:[A-Z_]+\s+)*(\w+(?:::\w+)*))'
|
||||
r'(class|struct)\s+(?:[a-zA-Z0-9_]+\s+)*(\w+(?:::\w+)*))'
|
||||
r'(.*)$', line)
|
||||
if (class_decl_match and
|
||||
(not self.stack or self.stack[-1].open_parentheses == 0)):
|
||||
@ -3144,7 +3412,8 @@ def CheckForNonStandardConstructs(filename, clean_lines, linenum,
|
||||
Search(r'\bstd\s*::\s*initializer_list\b', constructor_args[0]))
|
||||
copy_constructor = bool(
|
||||
onearg_constructor and
|
||||
Match(r'(const\s+)?%s(\s*<[^>]*>)?(\s+const)?\s*(?:<\w+>\s*)?&'
|
||||
Match(r'((const\s+(volatile\s+)?)?|(volatile\s+(const\s+)?))?'
|
||||
r'%s(\s*<[^>]*>)?(\s+const)?\s*(?:<\w+>\s*)?&'
|
||||
% re.escape(base_classname), constructor_args[0].strip()))
|
||||
|
||||
if (not is_marked_explicit and
|
||||
@ -3203,7 +3472,7 @@ def CheckSpacingForFunctionCall(filename, clean_lines, linenum, error):
|
||||
# Note that we assume the contents of [] to be short enough that
|
||||
# they'll never need to wrap.
|
||||
if ( # Ignore control structures.
|
||||
not Search(r'\b(if|for|while|switch|return|new|delete|catch|sizeof)\b',
|
||||
not Search(r'\b(if|elif|for|while|switch|return|new|delete|catch|sizeof)\b',
|
||||
fncall) and
|
||||
# Ignore pointers/references to functions.
|
||||
not Search(r' \([^)]+\)\([^)]*(\)|,$)', fncall) and
|
||||
@ -3316,7 +3585,7 @@ def CheckForFunctionLengths(filename, clean_lines, linenum,
|
||||
if Search(r'(;|})', start_line): # Declarations and trivial functions
|
||||
body_found = True
|
||||
break # ... ignore
|
||||
elif Search(r'{', start_line):
|
||||
if Search(r'{', start_line):
|
||||
body_found = True
|
||||
function = Search(r'((\w|:)*)\(', line).group(1)
|
||||
if Match(r'TEST', function): # Handle TEST... macros
|
||||
@ -3509,9 +3778,10 @@ def CheckSpacing(filename, clean_lines, linenum, nesting_state, error):
|
||||
# get rid of comments and strings
|
||||
line = clean_lines.elided[linenum]
|
||||
|
||||
# You shouldn't have spaces before your brackets, except maybe after
|
||||
# 'delete []' or 'return []() {};'
|
||||
if Search(r'\w\s+\[', line) and not Search(r'(?:delete|return)\s+\[', line):
|
||||
# You shouldn't have spaces before your brackets, except for C++11 attributes
|
||||
# or maybe after 'delete []', 'return []() {};', or 'auto [abc, ...] = ...;'.
|
||||
if (Search(r'\w\s+\[(?!\[)', line) and
|
||||
not Search(r'(?:auto&?|delete|return)\s+\[', line)):
|
||||
error(filename, linenum, 'whitespace/braces', 5,
|
||||
'Extra space before [')
|
||||
|
||||
@ -4029,11 +4299,11 @@ def CheckBraces(filename, clean_lines, linenum, error):
|
||||
# its line, and the line after that should have an indent level equal to or
|
||||
# lower than the if. We also check for ambiguous if/else nesting without
|
||||
# braces.
|
||||
if_else_match = Search(r'\b(if\s*\(|else\b)', line)
|
||||
if_else_match = Search(r'\b(if\s*(|constexpr)\s*\(|else\b)', line)
|
||||
if if_else_match and not Match(r'\s*#', line):
|
||||
if_indent = GetIndentLevel(line)
|
||||
endline, endlinenum, endpos = line, linenum, if_else_match.end()
|
||||
if_match = Search(r'\bif\s*\(', line)
|
||||
if_match = Search(r'\bif\s*(|constexpr)\s*\(', line)
|
||||
if if_match:
|
||||
# This could be a multiline if condition, so find the end first.
|
||||
pos = if_match.end() - 1
|
||||
@ -4092,9 +4362,9 @@ def CheckTrailingSemicolon(filename, clean_lines, linenum, error):
|
||||
|
||||
# Block bodies should not be followed by a semicolon. Due to C++11
|
||||
# brace initialization, there are more places where semicolons are
|
||||
# required than not, so we use a whitelist approach to check these
|
||||
# rather than a blacklist. These are the places where "};" should
|
||||
# be replaced by just "}":
|
||||
# required than not, so we explicitly list the allowed rules rather
|
||||
# than listing the disallowed ones. These are the places where "};"
|
||||
# should be replaced by just "}":
|
||||
# 1. Some flavor of block following closing parenthesis:
|
||||
# for (;;) {};
|
||||
# while (...) {};
|
||||
@ -4150,11 +4420,11 @@ def CheckTrailingSemicolon(filename, clean_lines, linenum, error):
|
||||
# - INTERFACE_DEF
|
||||
# - EXCLUSIVE_LOCKS_REQUIRED, SHARED_LOCKS_REQUIRED, LOCKS_EXCLUDED:
|
||||
#
|
||||
# We implement a whitelist of safe macros instead of a blacklist of
|
||||
# We implement a list of safe macros instead of a list of
|
||||
# unsafe macros, even though the latter appears less frequently in
|
||||
# google code and would have been easier to implement. This is because
|
||||
# the downside for getting the whitelist wrong means some extra
|
||||
# semicolons, while the downside for getting the blacklist wrong
|
||||
# the downside for getting the allowed checks wrong means some extra
|
||||
# semicolons, while the downside for getting disallowed checks wrong
|
||||
# would result in compile errors.
|
||||
#
|
||||
# In addition to macros, we also don't want to warn on
|
||||
@ -4575,7 +4845,7 @@ def CheckStyle(filename, clean_lines, linenum, file_extension, nesting_state,
|
||||
# if(match($0, " <<")) complain = 0;
|
||||
# if(match(prev, " +for \\(")) complain = 0;
|
||||
# if(prevodd && match(prevprev, " +for \\(")) complain = 0;
|
||||
scope_or_label_pattern = r'\s*\w+\s*:\s*\\?$'
|
||||
scope_or_label_pattern = r'\s*(?:public|private|protected|signals)(?:\s+(?:slots\s*)?)?:\s*\\?$'
|
||||
classinfo = nesting_state.InnermostClass()
|
||||
initial_spaces = 0
|
||||
cleansed_line = clean_lines.elided[linenum]
|
||||
@ -4699,13 +4969,14 @@ def _DropCommonSuffixes(filename):
|
||||
return os.path.splitext(filename)[0]
|
||||
|
||||
|
||||
def _ClassifyInclude(fileinfo, include, is_system):
|
||||
def _ClassifyInclude(fileinfo, include, used_angle_brackets, include_order="default"):
|
||||
"""Figures out what kind of header 'include' is.
|
||||
|
||||
Args:
|
||||
fileinfo: The current file cpplint is running over. A FileInfo instance.
|
||||
include: The path to a #included file.
|
||||
is_system: True if the #include used <> rather than "".
|
||||
used_angle_brackets: True if the #include used <> rather than "".
|
||||
include_order: "default" or other value allowed in program arguments
|
||||
|
||||
Returns:
|
||||
One of the _XXX_HEADER constants.
|
||||
@ -4715,6 +4986,8 @@ def _ClassifyInclude(fileinfo, include, is_system):
|
||||
_C_SYS_HEADER
|
||||
>>> _ClassifyInclude(FileInfo('foo/foo.cc'), 'string', True)
|
||||
_CPP_SYS_HEADER
|
||||
>>> _ClassifyInclude(FileInfo('foo/foo.cc'), 'foo/foo.h', True, "standardcfirst")
|
||||
_OTHER_SYS_HEADER
|
||||
>>> _ClassifyInclude(FileInfo('foo/foo.cc'), 'foo/foo.h', False)
|
||||
_LIKELY_MY_HEADER
|
||||
>>> _ClassifyInclude(FileInfo('foo/foo_unknown_extension.cc'),
|
||||
@ -4725,17 +4998,23 @@ def _ClassifyInclude(fileinfo, include, is_system):
|
||||
"""
|
||||
# This is a list of all standard c++ header files, except
|
||||
# those already checked for above.
|
||||
is_cpp_h = include in _CPP_HEADERS
|
||||
is_cpp_header = include in _CPP_HEADERS
|
||||
|
||||
# Mark include as C header if in list or in a known folder for standard-ish C headers.
|
||||
is_std_c_header = (include_order == "default") or (include in _C_HEADERS
|
||||
# additional linux glibc header folders
|
||||
or Search(r'(?:%s)\/.*\.h' % "|".join(C_STANDARD_HEADER_FOLDERS), include))
|
||||
|
||||
# Headers with C++ extensions shouldn't be considered C system headers
|
||||
if is_system and os.path.splitext(include)[1] in ['.hpp', '.hxx', '.h++']:
|
||||
is_system = False
|
||||
is_system = used_angle_brackets and not os.path.splitext(include)[1] in ['.hpp', '.hxx', '.h++']
|
||||
|
||||
if is_system:
|
||||
if is_cpp_h:
|
||||
if is_cpp_header:
|
||||
return _CPP_SYS_HEADER
|
||||
else:
|
||||
if is_std_c_header:
|
||||
return _C_SYS_HEADER
|
||||
else:
|
||||
return _OTHER_SYS_HEADER
|
||||
|
||||
# If the target file and the include we're checking share a
|
||||
# basename when we drop common extensions, and the include
|
||||
@ -4800,7 +5079,7 @@ def CheckIncludeLine(filename, clean_lines, linenum, include_state, error):
|
||||
match = _RE_PATTERN_INCLUDE.search(line)
|
||||
if match:
|
||||
include = match.group(2)
|
||||
is_system = (match.group(1) == '<')
|
||||
used_angle_brackets = (match.group(1) == '<')
|
||||
duplicate_line = include_state.FindHeader(include)
|
||||
if duplicate_line >= 0:
|
||||
error(filename, linenum, 'build/include', 4,
|
||||
@ -4815,7 +5094,19 @@ def CheckIncludeLine(filename, clean_lines, linenum, include_state, error):
|
||||
'Do not include .' + extension + ' files from other packages')
|
||||
return
|
||||
|
||||
if not _THIRD_PARTY_HEADERS_PATTERN.match(include):
|
||||
# We DO want to include a 3rd party looking header if it matches the
|
||||
# filename. Otherwise we get an erroneous error "...should include its
|
||||
# header" error later.
|
||||
third_src_header = False
|
||||
for ext in GetHeaderExtensions():
|
||||
basefilename = filename[0:len(filename) - len(fileinfo.Extension())]
|
||||
headerfile = basefilename + '.' + ext
|
||||
headername = FileInfo(headerfile).RepositoryName()
|
||||
if headername in include or include in headername:
|
||||
third_src_header = True
|
||||
break
|
||||
|
||||
if third_src_header or not _THIRD_PARTY_HEADERS_PATTERN.match(include):
|
||||
include_state.include_list[-1].append((include, linenum))
|
||||
|
||||
# We want to ensure that headers appear in the right order:
|
||||
@ -4830,7 +5121,7 @@ def CheckIncludeLine(filename, clean_lines, linenum, include_state, error):
|
||||
# track of the highest type seen, and complains if we see a
|
||||
# lower type after that.
|
||||
error_message = include_state.CheckNextIncludeOrder(
|
||||
_ClassifyInclude(fileinfo, include, is_system))
|
||||
_ClassifyInclude(fileinfo, include, used_angle_brackets, _include_order))
|
||||
if error_message:
|
||||
error(filename, linenum, 'build/include_order', 4,
|
||||
'%s. Should be: %s.h, c system, c++ system, other.' %
|
||||
@ -5084,7 +5375,7 @@ def CheckLanguage(filename, clean_lines, linenum, file_extension,
|
||||
if (IsHeaderExtension(file_extension)
|
||||
and Search(r'\bnamespace\s*{', line)
|
||||
and line[-1] != '\\'):
|
||||
error(filename, linenum, 'build/namespaces', 4,
|
||||
error(filename, linenum, 'build/namespaces_headers', 4,
|
||||
'Do not use unnamed namespaces in header files. See '
|
||||
'https://google-styleguide.googlecode.com/svn/trunk/cppguide.xml#Namespaces'
|
||||
' for more information.')
|
||||
@ -5374,19 +5665,19 @@ def CheckForNonConstReference(filename, clean_lines, linenum,
|
||||
#
|
||||
# We also accept & in static_assert, which looks like a function but
|
||||
# it's actually a declaration expression.
|
||||
whitelisted_functions = (r'(?:[sS]wap(?:<\w:+>)?|'
|
||||
allowed_functions = (r'(?:[sS]wap(?:<\w:+>)?|'
|
||||
r'operator\s*[<>][<>]|'
|
||||
r'static_assert|COMPILE_ASSERT'
|
||||
r')\s*\(')
|
||||
if Search(whitelisted_functions, line):
|
||||
if Search(allowed_functions, line):
|
||||
return
|
||||
elif not Search(r'\S+\([^)]*$', line):
|
||||
# Don't see a whitelisted function on this line. Actually we
|
||||
# Don't see an allowed function on this line. Actually we
|
||||
# didn't see any function name on this line, so this is likely a
|
||||
# multi-line parameter list. Try a bit harder to catch this case.
|
||||
for i in xrange(2):
|
||||
if (linenum > i and
|
||||
Search(whitelisted_functions, clean_lines.elided[linenum - i - 1])):
|
||||
Search(allowed_functions, clean_lines.elided[linenum - i - 1])):
|
||||
return
|
||||
|
||||
decls = ReplaceAll(r'{[^}]*}', ' ', line) # exclude function body
|
||||
@ -5461,7 +5752,7 @@ def CheckCasts(filename, clean_lines, linenum, error):
|
||||
|
||||
if not expecting_function:
|
||||
CheckCStyleCast(filename, clean_lines, linenum, 'static_cast',
|
||||
r'\((int|float|double|bool|char|u?int(16|32|64))\)', error)
|
||||
r'\((int|float|double|bool|char|u?int(16|32|64)|size_t)\)', error)
|
||||
|
||||
# This doesn't catch all cases. Consider (const char * const)"hello".
|
||||
#
|
||||
@ -5614,11 +5905,11 @@ _HEADERS_CONTAINING_TEMPLATES = (
|
||||
)),
|
||||
('<limits>', ('numeric_limits',)),
|
||||
('<list>', ('list',)),
|
||||
('<map>', ('map', 'multimap',)),
|
||||
('<map>', ('multimap',)),
|
||||
('<memory>', ('allocator', 'make_shared', 'make_unique', 'shared_ptr',
|
||||
'unique_ptr', 'weak_ptr')),
|
||||
('<queue>', ('queue', 'priority_queue',)),
|
||||
('<set>', ('set', 'multiset',)),
|
||||
('<set>', ('multiset',)),
|
||||
('<stack>', ('stack',)),
|
||||
('<string>', ('char_traits', 'basic_string',)),
|
||||
('<tuple>', ('tuple',)),
|
||||
@ -5652,6 +5943,16 @@ for _header, _templates in _HEADERS_MAYBE_TEMPLATES:
|
||||
(re.compile(r'[^>.]\b' + _template + r'(<.*?>)?\([^\)]'),
|
||||
_template,
|
||||
_header))
|
||||
# Match set<type>, but not foo->set<type>, foo.set<type>
|
||||
_re_pattern_headers_maybe_templates.append(
|
||||
(re.compile(r'[^>.]\bset\s*\<'),
|
||||
'set<>',
|
||||
'<set>'))
|
||||
# Match 'map<type> var' and 'std::map<type>(...)', but not 'map<type>(...)''
|
||||
_re_pattern_headers_maybe_templates.append(
|
||||
(re.compile(r'(std\b::\bmap\s*\<)|(^(std\b::\b)map\b\(\s*\<)'),
|
||||
'map<>',
|
||||
'<map>'))
|
||||
|
||||
# Other scripts may reach in and modify this pattern.
|
||||
_re_pattern_templates = []
|
||||
@ -5734,18 +6035,19 @@ def UpdateIncludeState(filename, include_dict, io=codecs):
|
||||
"""
|
||||
headerfile = None
|
||||
try:
|
||||
headerfile = io.open(filename, 'r', 'utf8', 'replace')
|
||||
with io.open(filename, 'r', 'utf8', 'replace') as headerfile:
|
||||
linenum = 0
|
||||
for line in headerfile:
|
||||
linenum += 1
|
||||
clean_line = CleanseComments(line)
|
||||
match = _RE_PATTERN_INCLUDE.search(clean_line)
|
||||
if match:
|
||||
include = match.group(2)
|
||||
include_dict.setdefault(include, linenum)
|
||||
return True
|
||||
except IOError:
|
||||
return False
|
||||
linenum = 0
|
||||
for line in headerfile:
|
||||
linenum += 1
|
||||
clean_line = CleanseComments(line)
|
||||
match = _RE_PATTERN_INCLUDE.search(clean_line)
|
||||
if match:
|
||||
include = match.group(2)
|
||||
include_dict.setdefault(include, linenum)
|
||||
return True
|
||||
|
||||
|
||||
|
||||
def CheckForIncludeWhatYouUse(filename, clean_lines, include_state, error,
|
||||
@ -6259,20 +6561,15 @@ def ProcessConfigOverrides(filename):
|
||||
except ValueError:
|
||||
_cpplint_state.PrintError('Line length must be numeric.')
|
||||
elif name == 'extensions':
|
||||
global _valid_extensions
|
||||
try:
|
||||
extensions = [ext.strip() for ext in val.split(',')]
|
||||
_valid_extensions = set(extensions)
|
||||
except ValueError:
|
||||
sys.stderr.write('Extensions should be a comma-separated list of values;'
|
||||
'for example: extensions=hpp,cpp\n'
|
||||
'This could not be parsed: "%s"' % (val,))
|
||||
ProcessExtensionsOption(val)
|
||||
elif name == 'root':
|
||||
global _root
|
||||
# root directories are specified relative to CPPLINT.cfg dir.
|
||||
_root = os.path.join(os.path.dirname(cfg_file), val)
|
||||
elif name == 'headers':
|
||||
ProcessHppHeadersOption(val)
|
||||
elif name == 'includeorder':
|
||||
ProcessIncludeOrderOption(val)
|
||||
else:
|
||||
_cpplint_state.PrintError(
|
||||
'Invalid configuration option (%s) in file %s\n' %
|
||||
@ -6329,7 +6626,8 @@ def ProcessFile(filename, vlevel, extra_check_functions=None):
|
||||
codecs.getwriter('utf8'),
|
||||
'replace').read().split('\n')
|
||||
else:
|
||||
lines = codecs.open(filename, 'r', 'utf8', 'replace').read().split('\n')
|
||||
with codecs.open(filename, 'r', 'utf8', 'replace') as target_file:
|
||||
lines = target_file.read().split('\n')
|
||||
|
||||
# Remove trailing '\r'.
|
||||
# The -1 accounts for the extra trailing blank line we get from split()
|
||||
@ -6389,10 +6687,10 @@ def PrintUsage(message):
|
||||
Args:
|
||||
message: The optional error message.
|
||||
"""
|
||||
sys.stderr.write(_USAGE % (list(GetAllExtensions()),
|
||||
','.join(list(GetAllExtensions())),
|
||||
GetHeaderExtensions(),
|
||||
','.join(GetHeaderExtensions())))
|
||||
sys.stderr.write(_USAGE % (sorted(list(GetAllExtensions())),
|
||||
','.join(sorted(list(GetAllExtensions()))),
|
||||
sorted(GetHeaderExtensions()),
|
||||
','.join(sorted(GetHeaderExtensions()))))
|
||||
|
||||
if message:
|
||||
sys.exit('\nFATAL ERROR: ' + message)
|
||||
@ -6438,6 +6736,7 @@ def ParseArguments(args):
|
||||
'exclude=',
|
||||
'recursive',
|
||||
'headers=',
|
||||
'includeorder=',
|
||||
'quiet'])
|
||||
except getopt.GetoptError:
|
||||
PrintUsage('Invalid arguments.')
|
||||
@ -6455,9 +6754,9 @@ def ParseArguments(args):
|
||||
if opt == '--version':
|
||||
PrintVersion()
|
||||
elif opt == '--output':
|
||||
if val not in ('emacs', 'vs7', 'eclipse', 'junit'):
|
||||
if val not in ('emacs', 'vs7', 'eclipse', 'junit', 'sed', 'gsed'):
|
||||
PrintUsage('The only allowed output formats are emacs, vs7, eclipse '
|
||||
'and junit.')
|
||||
'sed, gsed and junit.')
|
||||
output_format = val
|
||||
elif opt == '--quiet':
|
||||
quiet = True
|
||||
@ -6489,15 +6788,13 @@ def ParseArguments(args):
|
||||
_excludes = set()
|
||||
_excludes.update(glob.glob(val))
|
||||
elif opt == '--extensions':
|
||||
global _valid_extensions
|
||||
try:
|
||||
_valid_extensions = set(val.split(','))
|
||||
except ValueError:
|
||||
PrintUsage('Extensions must be comma seperated list.')
|
||||
ProcessExtensionsOption(val)
|
||||
elif opt == '--headers':
|
||||
ProcessHppHeadersOption(val)
|
||||
elif opt == '--recursive':
|
||||
recursive = True
|
||||
elif opt == '--includeorder':
|
||||
ProcessIncludeOrderOption(val)
|
||||
|
||||
if not filenames:
|
||||
PrintUsage('No files were specified.')
|
||||
@ -6514,6 +6811,7 @@ def ParseArguments(args):
|
||||
_SetFilters(filters)
|
||||
_SetCountingStyle(counting_style)
|
||||
|
||||
filenames.sort()
|
||||
return filenames
|
||||
|
||||
def _ExpandDirectories(filenames):
|
||||
@ -6545,15 +6843,35 @@ def _ExpandDirectories(filenames):
|
||||
for filename in expanded:
|
||||
if os.path.splitext(filename)[1][1:] in GetAllExtensions():
|
||||
filtered.append(filename)
|
||||
|
||||
return filtered
|
||||
|
||||
def _FilterExcludedFiles(filenames):
|
||||
def _FilterExcludedFiles(fnames):
|
||||
"""Filters out files listed in the --exclude command line switch. File paths
|
||||
in the switch are evaluated relative to the current working directory
|
||||
"""
|
||||
exclude_paths = [os.path.abspath(f) for f in _excludes]
|
||||
return [f for f in filenames if os.path.abspath(f) not in exclude_paths]
|
||||
# because globbing does not work recursively, exclude all subpath of all excluded entries
|
||||
return [f for f in fnames
|
||||
if not any(e for e in exclude_paths
|
||||
if _IsParentOrSame(e, os.path.abspath(f)))]
|
||||
|
||||
def _IsParentOrSame(parent, child):
|
||||
"""Return true if child is subdirectory of parent.
|
||||
Assumes both paths are absolute and don't contain symlinks.
|
||||
"""
|
||||
parent = os.path.normpath(parent)
|
||||
child = os.path.normpath(child)
|
||||
if parent == child:
|
||||
return True
|
||||
|
||||
prefix = os.path.commonprefix([parent, child])
|
||||
if prefix != parent:
|
||||
return False
|
||||
# Note: os.path.commonprefix operates on character basis, so
|
||||
# take extra care of situations like '/foo/ba' and '/foo/bar/baz'
|
||||
child_suffix = child[len(prefix):]
|
||||
child_suffix = child_suffix.lstrip(os.sep)
|
||||
return child == os.path.join(prefix, child_suffix)
|
||||
|
||||
def main():
|
||||
filenames = ParseArguments(sys.argv[1:])
|
||||
|
Loading…
Reference in New Issue
Block a user