Update cpplint.py

This brings our copy of cpplint up to date as of 2022-03-20, copied from
https://github.com/cpplint/cpplint, revision 7d9920c459dcdd55b043a2450c6ee3519102d106

Signed-off-by: Michael Richters <gedankenexperimenter@gmail.com>
pull/1156/head
Michael Richters 2 years ago
parent 3c3ff1efc8
commit 5d1d7ae003
No known key found for this signature in database
GPG Key ID: 1288FD13E4EEF0C0

521
bin/cpplint.py vendored

@ -41,6 +41,11 @@ We do a small hack, which is to ignore //'s with "'s after them on the
same line, but it is far from perfect (in either direction). same line, but it is far from perfect (in either direction).
""" """
# cpplint predates fstrings
# pylint: disable=consider-using-f-string
# pylint: disable=invalid-name
import codecs import codecs
import copy import copy
import getopt import getopt
@ -59,7 +64,7 @@ import xml.etree.ElementTree
# if empty, use defaults # if empty, use defaults
_valid_extensions = set([]) _valid_extensions = set([])
__VERSION__ = '1.4.4' __VERSION__ = '1.6.0'
try: try:
xrange # Python 2 xrange # Python 2
@ -69,7 +74,7 @@ except NameError:
_USAGE = """ _USAGE = """
Syntax: cpplint.py [--verbose=#] [--output=emacs|eclipse|vs7|junit] Syntax: cpplint.py [--verbose=#] [--output=emacs|eclipse|vs7|junit|sed|gsed]
[--filter=-x,+y,...] [--filter=-x,+y,...]
[--counting=total|toplevel|detailed] [--root=subdir] [--counting=total|toplevel|detailed] [--root=subdir]
[--repository=path] [--repository=path]
@ -77,6 +82,7 @@ Syntax: cpplint.py [--verbose=#] [--output=emacs|eclipse|vs7|junit]
[--recursive] [--recursive]
[--exclude=path] [--exclude=path]
[--extensions=hpp,cpp,...] [--extensions=hpp,cpp,...]
[--includeorder=default|standardcfirst]
[--quiet] [--quiet]
[--version] [--version]
<file> [file] ... <file> [file] ...
@ -102,11 +108,16 @@ Syntax: cpplint.py [--verbose=#] [--output=emacs|eclipse|vs7|junit]
Flags: Flags:
output=emacs|eclipse|vs7|junit output=emacs|eclipse|vs7|junit|sed|gsed
By default, the output is formatted to ease emacs parsing. Visual Studio By default, the output is formatted to ease emacs parsing. Visual Studio
compatible output (vs7) may also be used. Further support exists for compatible output (vs7) may also be used. Further support exists for
eclipse (eclipse), and JUnit (junit). XML parsers such as those used eclipse (eclipse), and JUnit (junit). XML parsers such as those used
in Jenkins and Bamboo may also be used. Other formats are unsupported. in Jenkins and Bamboo may also be used.
The sed format outputs sed commands that should fix some of the errors.
Note that this requires gnu sed. If that is installed as gsed on your
system (common e.g. on macOS with homebrew) you can use the gsed output
format. Sed commands are written to stdout, not stderr, so you should be
able to pipe output straight to a shell to run the fixes.
verbose=# verbose=#
Specify a number 0-5 to restrict errors to certain verbosity levels. Specify a number 0-5 to restrict errors to certain verbosity levels.
@ -121,11 +132,11 @@ Syntax: cpplint.py [--verbose=#] [--output=emacs|eclipse|vs7|junit]
error messages whose category names pass the filters will be printed. error messages whose category names pass the filters will be printed.
(Category names are printed with the message and look like (Category names are printed with the message and look like
"[whitespace/indent]".) Filters are evaluated left to right. "[whitespace/indent]".) Filters are evaluated left to right.
"-FOO" and "FOO" means "do not print categories that start with FOO". "-FOO" means "do not print categories that start with FOO".
"+FOO" means "do print categories that start with FOO". "+FOO" means "do print categories that start with FOO".
Examples: --filter=-whitespace,+whitespace/braces Examples: --filter=-whitespace,+whitespace/braces
--filter=whitespace,runtime/printf,+runtime/printf_format --filter=-whitespace,-runtime/printf,+runtime/printf_format
--filter=-,+build/include_what_you_use --filter=-,+build/include_what_you_use
To see a list of all the categories used in cpplint, pass no arg: To see a list of all the categories used in cpplint, pass no arg:
@ -209,6 +220,15 @@ Syntax: cpplint.py [--verbose=#] [--output=emacs|eclipse|vs7|junit]
Examples: Examples:
--extensions=%s --extensions=%s
includeorder=default|standardcfirst
For the build/include_order rule, the default is to blindly assume angle
bracket includes with file extension are c-system-headers (default),
even knowing this will have false classifications.
The default is established at google.
standardcfirst means to instead use an allow-list of known c headers and
treat all others as separate group of "other system headers". The C headers
included are those of the C-standard lib and closely related ones.
headers=x,y,... headers=x,y,...
The header extensions that cpplint will treat as .h in checks. Values are The header extensions that cpplint will treat as .h in checks. Values are
automatically added to --extensions list. automatically added to --extensions list.
@ -282,6 +302,7 @@ _ERROR_CATEGORIES = [
'build/include_alpha', 'build/include_alpha',
'build/include_order', 'build/include_order',
'build/include_what_you_use', 'build/include_what_you_use',
'build/namespaces_headers',
'build/namespaces_literals', 'build/namespaces_literals',
'build/namespaces', 'build/namespaces',
'build/printf_format', 'build/printf_format',
@ -338,6 +359,13 @@ _ERROR_CATEGORIES = [
'whitespace/todo', 'whitespace/todo',
] ]
# keywords to use with --outputs which generate stdout for machine processing
_MACHINE_OUTPUTS = [
'junit',
'sed',
'gsed'
]
# These error categories are no longer enforced by cpplint, but for backwards- # These error categories are no longer enforced by cpplint, but for backwards-
# compatibility they may still appear in NOLINT comments. # compatibility they may still appear in NOLINT comments.
_LEGACY_ERROR_CATEGORIES = [ _LEGACY_ERROR_CATEGORIES = [
@ -514,6 +542,186 @@ _CPP_HEADERS = frozenset([
'cwctype', 'cwctype',
]) ])
# C headers
_C_HEADERS = frozenset([
# System C headers
'assert.h',
'complex.h',
'ctype.h',
'errno.h',
'fenv.h',
'float.h',
'inttypes.h',
'iso646.h',
'limits.h',
'locale.h',
'math.h',
'setjmp.h',
'signal.h',
'stdalign.h',
'stdarg.h',
'stdatomic.h',
'stdbool.h',
'stddef.h',
'stdint.h',
'stdio.h',
'stdlib.h',
'stdnoreturn.h',
'string.h',
'tgmath.h',
'threads.h',
'time.h',
'uchar.h',
'wchar.h',
'wctype.h',
# additional POSIX C headers
'aio.h',
'arpa/inet.h',
'cpio.h',
'dirent.h',
'dlfcn.h',
'fcntl.h',
'fmtmsg.h',
'fnmatch.h',
'ftw.h',
'glob.h',
'grp.h',
'iconv.h',
'langinfo.h',
'libgen.h',
'monetary.h',
'mqueue.h',
'ndbm.h',
'net/if.h',
'netdb.h',
'netinet/in.h',
'netinet/tcp.h',
'nl_types.h',
'poll.h',
'pthread.h',
'pwd.h',
'regex.h',
'sched.h',
'search.h',
'semaphore.h',
'setjmp.h',
'signal.h',
'spawn.h',
'strings.h',
'stropts.h',
'syslog.h',
'tar.h',
'termios.h',
'trace.h',
'ulimit.h',
'unistd.h',
'utime.h',
'utmpx.h',
'wordexp.h',
# additional GNUlib headers
'a.out.h',
'aliases.h',
'alloca.h',
'ar.h',
'argp.h',
'argz.h',
'byteswap.h',
'crypt.h',
'endian.h',
'envz.h',
'err.h',
'error.h',
'execinfo.h',
'fpu_control.h',
'fstab.h',
'fts.h',
'getopt.h',
'gshadow.h',
'ieee754.h',
'ifaddrs.h',
'libintl.h',
'mcheck.h',
'mntent.h',
'obstack.h',
'paths.h',
'printf.h',
'pty.h',
'resolv.h',
'shadow.h',
'sysexits.h',
'ttyent.h',
# Additional linux glibc headers
'dlfcn.h',
'elf.h',
'features.h',
'gconv.h',
'gnu-versions.h',
'lastlog.h',
'libio.h',
'link.h',
'malloc.h',
'memory.h',
'netash/ash.h',
'netatalk/at.h',
'netax25/ax25.h',
'neteconet/ec.h',
'netipx/ipx.h',
'netiucv/iucv.h',
'netpacket/packet.h',
'netrom/netrom.h',
'netrose/rose.h',
'nfs/nfs.h',
'nl_types.h',
'nss.h',
're_comp.h',
'regexp.h',
'sched.h',
'sgtty.h',
'stab.h',
'stdc-predef.h',
'stdio_ext.h',
'syscall.h',
'termio.h',
'thread_db.h',
'ucontext.h',
'ustat.h',
'utmp.h',
'values.h',
'wait.h',
'xlocale.h',
# Hardware specific headers
'arm_neon.h',
'emmintrin.h',
'xmmintin.h',
])
# Folders of C libraries so commonly used in C++,
# that they have parity with standard C libraries.
C_STANDARD_HEADER_FOLDERS = frozenset([
# standard C library
"sys",
# glibc for linux
"arpa",
"asm-generic",
"bits",
"gnu",
"net",
"netinet",
"protocols",
"rpc",
"rpcsvc",
"scsi",
# linux kernel header
"drm",
"linux",
"misc",
"mtd",
"rdma",
"sound",
"video",
"xen",
])
# Type names # Type names
_TYPES = re.compile( _TYPES = re.compile(
r'^(?:' r'^(?:'
@ -600,9 +808,10 @@ _ALT_TOKEN_REPLACEMENT_PATTERN = re.compile(
# _IncludeState.CheckNextIncludeOrder(). # _IncludeState.CheckNextIncludeOrder().
_C_SYS_HEADER = 1 _C_SYS_HEADER = 1
_CPP_SYS_HEADER = 2 _CPP_SYS_HEADER = 2
_LIKELY_MY_HEADER = 3 _OTHER_SYS_HEADER = 3
_POSSIBLE_MY_HEADER = 4 _LIKELY_MY_HEADER = 4
_OTHER_HEADER = 5 _POSSIBLE_MY_HEADER = 5
_OTHER_HEADER = 6
# These constants define the current inline assembly state # These constants define the current inline assembly state
_NO_ASM = 0 # Outside of inline assembly block _NO_ASM = 0 # Outside of inline assembly block
@ -622,6 +831,22 @@ _SEARCH_C_FILE = re.compile(r'\b(?:LINT_C_FILE|'
# Match string that indicates we're working on a Linux Kernel file. # Match string that indicates we're working on a Linux Kernel file.
_SEARCH_KERNEL_FILE = re.compile(r'\b(?:LINT_KERNEL_FILE)') _SEARCH_KERNEL_FILE = re.compile(r'\b(?:LINT_KERNEL_FILE)')
# Commands for sed to fix the problem
_SED_FIXUPS = {
'Remove spaces around =': r's/ = /=/',
'Remove spaces around !=': r's/ != /!=/',
'Remove space before ( in if (': r's/if (/if(/',
'Remove space before ( in for (': r's/for (/for(/',
'Remove space before ( in while (': r's/while (/while(/',
'Remove space before ( in switch (': r's/switch (/switch(/',
'Should have a space between // and comment': r's/\/\//\/\/ /',
'Missing space before {': r's/\([^ ]\){/\1 {/',
'Tab found, replace by spaces': r's/\t/ /g',
'Line ends in whitespace. Consider deleting these extra spaces.': r's/\s*$//',
'You don\'t need a ; after a }': r's/};/}/',
'Missing space after ,': r's/,\([^ ]\)/, \1/g',
}
_regexp_compile_cache = {} _regexp_compile_cache = {}
# {str, set(int)}: a map from error categories to sets of linenumbers # {str, set(int)}: a map from error categories to sets of linenumbers
@ -641,13 +866,16 @@ _repository = None
# Files to exclude from linting. This is set by the --exclude flag. # Files to exclude from linting. This is set by the --exclude flag.
_excludes = None _excludes = None
# Whether to supress PrintInfo messages # Whether to supress all PrintInfo messages, UNRELATED to --quiet flag
_quiet = False _quiet = False
# The allowed line length of files. # The allowed line length of files.
# This is set by --linelength flag. # This is set by --linelength flag.
_line_length = 80 _line_length = 80
# This allows to use different include order rule than default
_include_order = "default"
try: try:
unicode unicode
except NameError: except NameError:
@ -678,7 +906,7 @@ def unicode_escape_decode(x):
# Treat all headers starting with 'h' equally: .h, .hpp, .hxx etc. # Treat all headers starting with 'h' equally: .h, .hpp, .hxx etc.
# This is set by --headers flag. # This is set by --headers flag.
_hpp_headers = set(['h', 'hh', 'hpp', 'hxx', 'h++', 'cuh']) _hpp_headers = set([])
# {str, bool}: a map from error categories to booleans which indicate if the # {str, bool}: a map from error categories to booleans which indicate if the
# category should be suppressed for every line. # category should be suppressed for every line.
@ -687,30 +915,48 @@ _global_error_suppressions = {}
def ProcessHppHeadersOption(val): def ProcessHppHeadersOption(val):
global _hpp_headers global _hpp_headers
try: try:
_hpp_headers = set(val.split(',')) _hpp_headers = {ext.strip() for ext in val.split(',')}
# Automatically append to extensions list so it does not have to be set 2 times
_valid_extensions.update(_hpp_headers)
except ValueError: except ValueError:
PrintUsage('Header extensions must be comma separated list.') PrintUsage('Header extensions must be comma separated list.')
def ProcessIncludeOrderOption(val):
if val is None or val == "default":
pass
elif val == "standardcfirst":
global _include_order
_include_order = val
else:
PrintUsage('Invalid includeorder value %s. Expected default|standardcfirst')
def IsHeaderExtension(file_extension): def IsHeaderExtension(file_extension):
return file_extension in _hpp_headers return file_extension in GetHeaderExtensions()
def GetHeaderExtensions(): def GetHeaderExtensions():
return _hpp_headers or ['h'] if _hpp_headers:
return _hpp_headers
if _valid_extensions:
return {h for h in _valid_extensions if 'h' in h}
return set(['h', 'hh', 'hpp', 'hxx', 'h++', 'cuh'])
# The allowed extensions for file names # The allowed extensions for file names
# This is set by --extensions flag # This is set by --extensions flag
def GetAllExtensions(): def GetAllExtensions():
if not _valid_extensions: return GetHeaderExtensions().union(_valid_extensions or set(
return GetHeaderExtensions().union(set(['c', 'cc', 'cpp', 'cxx', 'c++', 'cu'])) ['c', 'cc', 'cpp', 'cxx', 'c++', 'cu']))
return _valid_extensions
def ProcessExtensionsOption(val):
global _valid_extensions
try:
extensions = [ext.strip() for ext in val.split(',')]
_valid_extensions = set(extensions)
except ValueError:
PrintUsage('Extensions should be a comma-separated list of values;'
'for example: extensions=hpp,cpp\n'
'This could not be parsed: "%s"' % (val,))
def GetNonHeaderExtensions(): def GetNonHeaderExtensions():
return GetAllExtensions().difference(GetHeaderExtensions()) return GetAllExtensions().difference(GetHeaderExtensions())
def ParseNolintSuppressions(filename, raw_line, linenum, error): def ParseNolintSuppressions(filename, raw_line, linenum, error):
"""Updates the global list of line error-suppressions. """Updates the global list of line error-suppressions.
@ -843,11 +1089,13 @@ class _IncludeState(object):
_MY_H_SECTION = 1 _MY_H_SECTION = 1
_C_SECTION = 2 _C_SECTION = 2
_CPP_SECTION = 3 _CPP_SECTION = 3
_OTHER_H_SECTION = 4 _OTHER_SYS_SECTION = 4
_OTHER_H_SECTION = 5
_TYPE_NAMES = { _TYPE_NAMES = {
_C_SYS_HEADER: 'C system header', _C_SYS_HEADER: 'C system header',
_CPP_SYS_HEADER: 'C++ system header', _CPP_SYS_HEADER: 'C++ system header',
_OTHER_SYS_HEADER: 'other system header',
_LIKELY_MY_HEADER: 'header this file implements', _LIKELY_MY_HEADER: 'header this file implements',
_POSSIBLE_MY_HEADER: 'header this file may implement', _POSSIBLE_MY_HEADER: 'header this file may implement',
_OTHER_HEADER: 'other header', _OTHER_HEADER: 'other header',
@ -857,6 +1105,7 @@ class _IncludeState(object):
_MY_H_SECTION: 'a header this file implements', _MY_H_SECTION: 'a header this file implements',
_C_SECTION: 'C system header', _C_SECTION: 'C system header',
_CPP_SECTION: 'C++ system header', _CPP_SECTION: 'C++ system header',
_OTHER_SYS_SECTION: 'other system header',
_OTHER_H_SECTION: 'other header', _OTHER_H_SECTION: 'other header',
} }
@ -970,6 +1219,12 @@ class _IncludeState(object):
else: else:
self._last_header = '' self._last_header = ''
return error_message return error_message
elif header_type == _OTHER_SYS_HEADER:
if self._section <= self._OTHER_SYS_SECTION:
self._section = self._OTHER_SYS_SECTION
else:
self._last_header = ''
return error_message
elif header_type == _LIKELY_MY_HEADER: elif header_type == _LIKELY_MY_HEADER:
if self._section <= self._MY_H_SECTION: if self._section <= self._MY_H_SECTION:
self._section = self._MY_H_SECTION self._section = self._MY_H_SECTION
@ -1011,6 +1266,8 @@ class _CppLintState(object):
# "eclipse" - format that eclipse can parse # "eclipse" - format that eclipse can parse
# "vs7" - format that Microsoft Visual Studio 7 can parse # "vs7" - format that Microsoft Visual Studio 7 can parse
# "junit" - format that Jenkins, Bamboo, etc can parse # "junit" - format that Jenkins, Bamboo, etc can parse
# "sed" - returns a gnu sed command to fix the problem
# "gsed" - like sed, but names the command gsed, e.g. for macOS homebrew users
self.output_format = 'emacs' self.output_format = 'emacs'
# For JUnit output, save errors and failures until the end so that they # For JUnit output, save errors and failures until the end so that they
@ -1099,7 +1356,9 @@ class _CppLintState(object):
self.PrintInfo('Total errors found: %d\n' % self.error_count) self.PrintInfo('Total errors found: %d\n' % self.error_count)
def PrintInfo(self, message): def PrintInfo(self, message):
if not _quiet and self.output_format != 'junit': # _quiet does not represent --quiet flag.
# Hide infos from stdout to keep stdout pure for machine consumption
if not _quiet and self.output_format not in _MACHINE_OUTPUTS:
sys.stdout.write(message) sys.stdout.write(message)
def PrintError(self, message): def PrintError(self, message):
@ -1117,9 +1376,9 @@ class _CppLintState(object):
num_failures = len(self._junit_failures) num_failures = len(self._junit_failures)
testsuite = xml.etree.ElementTree.Element('testsuite') testsuite = xml.etree.ElementTree.Element('testsuite')
testsuite.attrib['name'] = 'cpplint'
testsuite.attrib['errors'] = str(num_errors) testsuite.attrib['errors'] = str(num_errors)
testsuite.attrib['failures'] = str(num_failures) testsuite.attrib['failures'] = str(num_failures)
testsuite.attrib['name'] = 'cpplint'
if num_errors == 0 and num_failures == 0: if num_errors == 0 and num_failures == 0:
testsuite.attrib['tests'] = str(1) testsuite.attrib['tests'] = str(1)
@ -1313,7 +1572,7 @@ class FileInfo(object):
If we have a real absolute path name here we can try to do something smart: If we have a real absolute path name here we can try to do something smart:
detecting the root of the checkout and truncating /path/to/checkout from detecting the root of the checkout and truncating /path/to/checkout from
the name so that we get header guards that don't include things like the name so that we get header guards that don't include things like
"C:\Documents and Settings\..." or "/home/username/..." in them and thus "C:\\Documents and Settings\\..." or "/home/username/..." in them and thus
people on different computers who have checked the source out to different people on different computers who have checked the source out to different
locations won't see bogus errors. locations won't see bogus errors.
""" """
@ -1459,6 +1718,13 @@ def Error(filename, linenum, category, confidence, message):
elif _cpplint_state.output_format == 'junit': elif _cpplint_state.output_format == 'junit':
_cpplint_state.AddJUnitFailure(filename, linenum, message, category, _cpplint_state.AddJUnitFailure(filename, linenum, message, category,
confidence) confidence)
elif _cpplint_state.output_format in ['sed', 'gsed']:
if message in _SED_FIXUPS:
sys.stdout.write(_cpplint_state.output_format + " -i '%s%s' %s # %s [%s] [%d]\n" % (
linenum, _SED_FIXUPS[message], filename, message, category, confidence))
else:
sys.stderr.write('# %s:%s: "%s" [%s] [%d]\n' % (
filename, linenum, message, category, confidence))
else: else:
final_message = '%s:%s: %s [%s] [%d]\n' % ( final_message = '%s:%s: %s [%s] [%d]\n' % (
filename, linenum, message, category, confidence) filename, linenum, message, category, confidence)
@ -1599,7 +1865,7 @@ def FindNextMultiLineCommentEnd(lines, lineix):
def RemoveMultiLineCommentsFromRange(lines, begin, end): def RemoveMultiLineCommentsFromRange(lines, begin, end):
"""Clears a range of lines for multi-line comments.""" """Clears a range of lines for multi-line comments."""
# Having // dummy comments makes the lines non-empty, so we will not get # Having // <empty> comments makes the lines non-empty, so we will not get
# unnecessary blank line warnings later in the code. # unnecessary blank line warnings later in the code.
for i in range(begin, end): for i in range(begin, end):
lines[i] = '/**/' lines[i] = '/**/'
@ -1654,6 +1920,7 @@ class CleansedLines(object):
self.raw_lines = lines self.raw_lines = lines
self.num_lines = len(lines) self.num_lines = len(lines)
self.lines_without_raw_strings = CleanseRawStrings(lines) self.lines_without_raw_strings = CleanseRawStrings(lines)
# # pylint: disable=consider-using-enumerate
for linenum in range(len(self.lines_without_raw_strings)): for linenum in range(len(self.lines_without_raw_strings)):
self.lines.append(CleanseComments( self.lines.append(CleanseComments(
self.lines_without_raw_strings[linenum])) self.lines_without_raw_strings[linenum]))
@ -1973,7 +2240,7 @@ def CheckForCopyright(filename, lines, error):
"""Logs an error if no Copyright message appears at the top of the file.""" """Logs an error if no Copyright message appears at the top of the file."""
# We'll say it should occur by line 10. Don't forget there's a # We'll say it should occur by line 10. Don't forget there's a
# dummy line at the front. # placeholder line at the front.
for line in xrange(1, min(len(lines), 11)): for line in xrange(1, min(len(lines), 11)):
if re.search(r'Copyright', lines[line], re.I): break if re.search(r'Copyright', lines[line], re.I): break
else: # means no copyright line was found else: # means no copyright line was found
@ -2076,7 +2343,8 @@ def GetHeaderGuardCPPVariable(filename):
# --root=.. , will prepend the outer directory to the header guard # --root=.. , will prepend the outer directory to the header guard
full_path = fileinfo.FullName() full_path = fileinfo.FullName()
root_abspath = os.path.abspath(_root) # adapt slashes for windows
root_abspath = os.path.abspath(_root).replace('\\', '/')
maybe_path = StripListPrefix(PathSplitToList(full_path), maybe_path = StripListPrefix(PathSplitToList(full_path),
PathSplitToList(root_abspath)) PathSplitToList(root_abspath))
@ -2216,16 +2484,22 @@ def CheckHeaderFileIncluded(filename, include_state, error):
continue continue
headername = FileInfo(headerfile).RepositoryName() headername = FileInfo(headerfile).RepositoryName()
first_include = None first_include = None
include_uses_unix_dir_aliases = False
for section_list in include_state.include_list: for section_list in include_state.include_list:
for f in section_list: for f in section_list:
if headername in f[0] or f[0] in headername: include_text = f[0]
if "./" in include_text:
include_uses_unix_dir_aliases = True
if headername in include_text or include_text in headername:
return return
if not first_include: if not first_include:
first_include = f[1] first_include = f[1]
error(filename, first_include, 'build/include', 5, message = '%s should include its header file %s' % (fileinfo.RepositoryName(), headername)
'%s should include its header file %s' % (fileinfo.RepositoryName(), if include_uses_unix_dir_aliases:
headername)) message += ". Relative paths like . and .. are not allowed."
error(filename, first_include, 'build/include', 5, message)
def CheckForBadCharacters(filename, lines, error): def CheckForBadCharacters(filename, lines, error):
@ -2876,7 +3150,7 @@ class NestingState(object):
# }; # };
class_decl_match = Match( class_decl_match = Match(
r'^(\s*(?:template\s*<[\w\s<>,:=]*>\s*)?' r'^(\s*(?:template\s*<[\w\s<>,:=]*>\s*)?'
r'(class|struct)\s+(?:[A-Z_]+\s+)*(\w+(?:::\w+)*))' r'(class|struct)\s+(?:[a-zA-Z0-9_]+\s+)*(\w+(?:::\w+)*))'
r'(.*)$', line) r'(.*)$', line)
if (class_decl_match and if (class_decl_match and
(not self.stack or self.stack[-1].open_parentheses == 0)): (not self.stack or self.stack[-1].open_parentheses == 0)):
@ -3144,7 +3418,8 @@ def CheckForNonStandardConstructs(filename, clean_lines, linenum,
Search(r'\bstd\s*::\s*initializer_list\b', constructor_args[0])) Search(r'\bstd\s*::\s*initializer_list\b', constructor_args[0]))
copy_constructor = bool( copy_constructor = bool(
onearg_constructor and onearg_constructor and
Match(r'(const\s+)?%s(\s*<[^>]*>)?(\s+const)?\s*(?:<\w+>\s*)?&' Match(r'((const\s+(volatile\s+)?)?|(volatile\s+(const\s+)?))?'
r'%s(\s*<[^>]*>)?(\s+const)?\s*(?:<\w+>\s*)?&'
% re.escape(base_classname), constructor_args[0].strip())) % re.escape(base_classname), constructor_args[0].strip()))
if (not is_marked_explicit and if (not is_marked_explicit and
@ -3203,7 +3478,7 @@ def CheckSpacingForFunctionCall(filename, clean_lines, linenum, error):
# Note that we assume the contents of [] to be short enough that # Note that we assume the contents of [] to be short enough that
# they'll never need to wrap. # they'll never need to wrap.
if ( # Ignore control structures. if ( # Ignore control structures.
not Search(r'\b(if|for|while|switch|return|new|delete|catch|sizeof)\b', not Search(r'\b(if|elif|for|while|switch|return|new|delete|catch|sizeof)\b',
fncall) and fncall) and
# Ignore pointers/references to functions. # Ignore pointers/references to functions.
not Search(r' \([^)]+\)\([^)]*(\)|,$)', fncall) and not Search(r' \([^)]+\)\([^)]*(\)|,$)', fncall) and
@ -3316,7 +3591,7 @@ def CheckForFunctionLengths(filename, clean_lines, linenum,
if Search(r'(;|})', start_line): # Declarations and trivial functions if Search(r'(;|})', start_line): # Declarations and trivial functions
body_found = True body_found = True
break # ... ignore break # ... ignore
elif Search(r'{', start_line): if Search(r'{', start_line):
body_found = True body_found = True
function = Search(r'((\w|:)*)\(', line).group(1) function = Search(r'((\w|:)*)\(', line).group(1)
if Match(r'TEST', function): # Handle TEST... macros if Match(r'TEST', function): # Handle TEST... macros
@ -3509,9 +3784,10 @@ def CheckSpacing(filename, clean_lines, linenum, nesting_state, error):
# get rid of comments and strings # get rid of comments and strings
line = clean_lines.elided[linenum] line = clean_lines.elided[linenum]
# You shouldn't have spaces before your brackets, except maybe after # You shouldn't have spaces before your brackets, except for C++11 attributes
# 'delete []' or 'return []() {};' # or maybe after 'delete []', 'return []() {};', or 'auto [abc, ...] = ...;'.
if Search(r'\w\s+\[', line) and not Search(r'(?:delete|return)\s+\[', line): if (Search(r'\w\s+\[(?!\[)', line) and
not Search(r'(?:auto&?|delete|return)\s+\[', line)):
error(filename, linenum, 'whitespace/braces', 5, error(filename, linenum, 'whitespace/braces', 5,
'Extra space before [') 'Extra space before [')
@ -4029,11 +4305,11 @@ def CheckBraces(filename, clean_lines, linenum, error):
# its line, and the line after that should have an indent level equal to or # its line, and the line after that should have an indent level equal to or
# lower than the if. We also check for ambiguous if/else nesting without # lower than the if. We also check for ambiguous if/else nesting without
# braces. # braces.
if_else_match = Search(r'\b(if\s*\(|else\b)', line) if_else_match = Search(r'\b(if\s*(|constexpr)\s*\(|else\b)', line)
if if_else_match and not Match(r'\s*#', line): if if_else_match and not Match(r'\s*#', line):
if_indent = GetIndentLevel(line) if_indent = GetIndentLevel(line)
endline, endlinenum, endpos = line, linenum, if_else_match.end() endline, endlinenum, endpos = line, linenum, if_else_match.end()
if_match = Search(r'\bif\s*\(', line) if_match = Search(r'\bif\s*(|constexpr)\s*\(', line)
if if_match: if if_match:
# This could be a multiline if condition, so find the end first. # This could be a multiline if condition, so find the end first.
pos = if_match.end() - 1 pos = if_match.end() - 1
@ -4092,9 +4368,9 @@ def CheckTrailingSemicolon(filename, clean_lines, linenum, error):
# Block bodies should not be followed by a semicolon. Due to C++11 # Block bodies should not be followed by a semicolon. Due to C++11
# brace initialization, there are more places where semicolons are # brace initialization, there are more places where semicolons are
# required than not, so we use a whitelist approach to check these # required than not, so we explicitly list the allowed rules rather
# rather than a blacklist. These are the places where "};" should # than listing the disallowed ones. These are the places where "};"
# be replaced by just "}": # should be replaced by just "}":
# 1. Some flavor of block following closing parenthesis: # 1. Some flavor of block following closing parenthesis:
# for (;;) {}; # for (;;) {};
# while (...) {}; # while (...) {};
@ -4150,11 +4426,11 @@ def CheckTrailingSemicolon(filename, clean_lines, linenum, error):
# - INTERFACE_DEF # - INTERFACE_DEF
# - EXCLUSIVE_LOCKS_REQUIRED, SHARED_LOCKS_REQUIRED, LOCKS_EXCLUDED: # - EXCLUSIVE_LOCKS_REQUIRED, SHARED_LOCKS_REQUIRED, LOCKS_EXCLUDED:
# #
# We implement a whitelist of safe macros instead of a blacklist of # We implement a list of safe macros instead of a list of
# unsafe macros, even though the latter appears less frequently in # unsafe macros, even though the latter appears less frequently in
# google code and would have been easier to implement. This is because # google code and would have been easier to implement. This is because
# the downside for getting the whitelist wrong means some extra # the downside for getting the allowed checks wrong means some extra
# semicolons, while the downside for getting the blacklist wrong # semicolons, while the downside for getting disallowed checks wrong
# would result in compile errors. # would result in compile errors.
# #
# In addition to macros, we also don't want to warn on # In addition to macros, we also don't want to warn on
@ -4575,7 +4851,7 @@ def CheckStyle(filename, clean_lines, linenum, file_extension, nesting_state,
# if(match($0, " <<")) complain = 0; # if(match($0, " <<")) complain = 0;
# if(match(prev, " +for \\(")) complain = 0; # if(match(prev, " +for \\(")) complain = 0;
# if(prevodd && match(prevprev, " +for \\(")) complain = 0; # if(prevodd && match(prevprev, " +for \\(")) complain = 0;
scope_or_label_pattern = r'\s*\w+\s*:\s*\\?$' scope_or_label_pattern = r'\s*(?:public|private|protected|signals)(?:\s+(?:slots\s*)?)?:\s*\\?$'
classinfo = nesting_state.InnermostClass() classinfo = nesting_state.InnermostClass()
initial_spaces = 0 initial_spaces = 0
cleansed_line = clean_lines.elided[linenum] cleansed_line = clean_lines.elided[linenum]
@ -4699,13 +4975,14 @@ def _DropCommonSuffixes(filename):
return os.path.splitext(filename)[0] return os.path.splitext(filename)[0]
def _ClassifyInclude(fileinfo, include, is_system): def _ClassifyInclude(fileinfo, include, used_angle_brackets, include_order="default"):
"""Figures out what kind of header 'include' is. """Figures out what kind of header 'include' is.
Args: Args:
fileinfo: The current file cpplint is running over. A FileInfo instance. fileinfo: The current file cpplint is running over. A FileInfo instance.
include: The path to a #included file. include: The path to a #included file.
is_system: True if the #include used <> rather than "". used_angle_brackets: True if the #include used <> rather than "".
include_order: "default" or other value allowed in program arguments
Returns: Returns:
One of the _XXX_HEADER constants. One of the _XXX_HEADER constants.
@ -4715,6 +4992,8 @@ def _ClassifyInclude(fileinfo, include, is_system):
_C_SYS_HEADER _C_SYS_HEADER
>>> _ClassifyInclude(FileInfo('foo/foo.cc'), 'string', True) >>> _ClassifyInclude(FileInfo('foo/foo.cc'), 'string', True)
_CPP_SYS_HEADER _CPP_SYS_HEADER
>>> _ClassifyInclude(FileInfo('foo/foo.cc'), 'foo/foo.h', True, "standardcfirst")
_OTHER_SYS_HEADER
>>> _ClassifyInclude(FileInfo('foo/foo.cc'), 'foo/foo.h', False) >>> _ClassifyInclude(FileInfo('foo/foo.cc'), 'foo/foo.h', False)
_LIKELY_MY_HEADER _LIKELY_MY_HEADER
>>> _ClassifyInclude(FileInfo('foo/foo_unknown_extension.cc'), >>> _ClassifyInclude(FileInfo('foo/foo_unknown_extension.cc'),
@ -4725,17 +5004,23 @@ def _ClassifyInclude(fileinfo, include, is_system):
""" """
# This is a list of all standard c++ header files, except # This is a list of all standard c++ header files, except
# those already checked for above. # those already checked for above.
is_cpp_h = include in _CPP_HEADERS is_cpp_header = include in _CPP_HEADERS
# Mark include as C header if in list or in a known folder for standard-ish C headers.
is_std_c_header = (include_order == "default") or (include in _C_HEADERS
# additional linux glibc header folders
or Search(r'(?:%s)\/.*\.h' % "|".join(C_STANDARD_HEADER_FOLDERS), include))
# Headers with C++ extensions shouldn't be considered C system headers # Headers with C++ extensions shouldn't be considered C system headers
if is_system and os.path.splitext(include)[1] in ['.hpp', '.hxx', '.h++']: is_system = used_angle_brackets and not os.path.splitext(include)[1] in ['.hpp', '.hxx', '.h++']
is_system = False
if is_system: if is_system:
if is_cpp_h: if is_cpp_header:
return _CPP_SYS_HEADER return _CPP_SYS_HEADER
else: if is_std_c_header:
return _C_SYS_HEADER return _C_SYS_HEADER
else:
return _OTHER_SYS_HEADER
# If the target file and the include we're checking share a # If the target file and the include we're checking share a
# basename when we drop common extensions, and the include # basename when we drop common extensions, and the include
@ -4789,10 +5074,12 @@ def CheckIncludeLine(filename, clean_lines, linenum, include_state, error):
# #
# We also make an exception for Lua headers, which follow google # We also make an exception for Lua headers, which follow google
# naming convention but not the include convention. # naming convention but not the include convention.
match = Match(r'#include\s*"([^/]+\.h)"', line) match = Match(r'#include\s*"([^/]+\.(.*))"', line)
if match and not _THIRD_PARTY_HEADERS_PATTERN.match(match.group(1)): if match:
if (IsHeaderExtension(match.group(2)) and
not _THIRD_PARTY_HEADERS_PATTERN.match(match.group(1))):
error(filename, linenum, 'build/include_subdir', 4, error(filename, linenum, 'build/include_subdir', 4,
'Include the directory when naming .h files') 'Include the directory when naming header files')
# we shouldn't include a file more than once. actually, there are a # we shouldn't include a file more than once. actually, there are a
# handful of instances where doing so is okay, but in general it's # handful of instances where doing so is okay, but in general it's
@ -4800,7 +5087,7 @@ def CheckIncludeLine(filename, clean_lines, linenum, include_state, error):
match = _RE_PATTERN_INCLUDE.search(line) match = _RE_PATTERN_INCLUDE.search(line)
if match: if match:
include = match.group(2) include = match.group(2)
is_system = (match.group(1) == '<') used_angle_brackets = (match.group(1) == '<')
duplicate_line = include_state.FindHeader(include) duplicate_line = include_state.FindHeader(include)
if duplicate_line >= 0: if duplicate_line >= 0:
error(filename, linenum, 'build/include', 4, error(filename, linenum, 'build/include', 4,
@ -4815,7 +5102,19 @@ def CheckIncludeLine(filename, clean_lines, linenum, include_state, error):
'Do not include .' + extension + ' files from other packages') 'Do not include .' + extension + ' files from other packages')
return return
if not _THIRD_PARTY_HEADERS_PATTERN.match(include): # We DO want to include a 3rd party looking header if it matches the
# filename. Otherwise we get an erroneous error "...should include its
# header" error later.
third_src_header = False
for ext in GetHeaderExtensions():
basefilename = filename[0:len(filename) - len(fileinfo.Extension())]
headerfile = basefilename + '.' + ext
headername = FileInfo(headerfile).RepositoryName()
if headername in include or include in headername:
third_src_header = True
break
if third_src_header or not _THIRD_PARTY_HEADERS_PATTERN.match(include):
include_state.include_list[-1].append((include, linenum)) include_state.include_list[-1].append((include, linenum))
# We want to ensure that headers appear in the right order: # We want to ensure that headers appear in the right order:
@ -4830,7 +5129,7 @@ def CheckIncludeLine(filename, clean_lines, linenum, include_state, error):
# track of the highest type seen, and complains if we see a # track of the highest type seen, and complains if we see a
# lower type after that. # lower type after that.
error_message = include_state.CheckNextIncludeOrder( error_message = include_state.CheckNextIncludeOrder(
_ClassifyInclude(fileinfo, include, is_system)) _ClassifyInclude(fileinfo, include, used_angle_brackets, _include_order))
if error_message: if error_message:
error(filename, linenum, 'build/include_order', 4, error(filename, linenum, 'build/include_order', 4,
'%s. Should be: %s.h, c system, c++ system, other.' % '%s. Should be: %s.h, c system, c++ system, other.' %
@ -5084,7 +5383,7 @@ def CheckLanguage(filename, clean_lines, linenum, file_extension,
if (IsHeaderExtension(file_extension) if (IsHeaderExtension(file_extension)
and Search(r'\bnamespace\s*{', line) and Search(r'\bnamespace\s*{', line)
and line[-1] != '\\'): and line[-1] != '\\'):
error(filename, linenum, 'build/namespaces', 4, error(filename, linenum, 'build/namespaces_headers', 4,
'Do not use unnamed namespaces in header files. See ' 'Do not use unnamed namespaces in header files. See '
'https://google-styleguide.googlecode.com/svn/trunk/cppguide.xml#Namespaces' 'https://google-styleguide.googlecode.com/svn/trunk/cppguide.xml#Namespaces'
' for more information.') ' for more information.')
@ -5374,19 +5673,19 @@ def CheckForNonConstReference(filename, clean_lines, linenum,
# #
# We also accept & in static_assert, which looks like a function but # We also accept & in static_assert, which looks like a function but
# it's actually a declaration expression. # it's actually a declaration expression.
whitelisted_functions = (r'(?:[sS]wap(?:<\w:+>)?|' allowed_functions = (r'(?:[sS]wap(?:<\w:+>)?|'
r'operator\s*[<>][<>]|' r'operator\s*[<>][<>]|'
r'static_assert|COMPILE_ASSERT' r'static_assert|COMPILE_ASSERT'
r')\s*\(') r')\s*\(')
if Search(whitelisted_functions, line): if Search(allowed_functions, line):
return return
elif not Search(r'\S+\([^)]*$', line): elif not Search(r'\S+\([^)]*$', line):
# Don't see a whitelisted function on this line. Actually we # Don't see an allowed function on this line. Actually we
# didn't see any function name on this line, so this is likely a # didn't see any function name on this line, so this is likely a
# multi-line parameter list. Try a bit harder to catch this case. # multi-line parameter list. Try a bit harder to catch this case.
for i in xrange(2): for i in xrange(2):
if (linenum > i and if (linenum > i and
Search(whitelisted_functions, clean_lines.elided[linenum - i - 1])): Search(allowed_functions, clean_lines.elided[linenum - i - 1])):
return return
decls = ReplaceAll(r'{[^}]*}', ' ', line) # exclude function body decls = ReplaceAll(r'{[^}]*}', ' ', line) # exclude function body
@ -5461,7 +5760,7 @@ def CheckCasts(filename, clean_lines, linenum, error):
if not expecting_function: if not expecting_function:
CheckCStyleCast(filename, clean_lines, linenum, 'static_cast', CheckCStyleCast(filename, clean_lines, linenum, 'static_cast',
r'\((int|float|double|bool|char|u?int(16|32|64))\)', error) r'\((int|float|double|bool|char|u?int(16|32|64)|size_t)\)', error)
# This doesn't catch all cases. Consider (const char * const)"hello". # This doesn't catch all cases. Consider (const char * const)"hello".
# #
@ -5553,7 +5852,8 @@ def CheckCStyleCast(filename, clean_lines, linenum, cast_type, pattern, error):
return False return False
# operator++(int) and operator--(int) # operator++(int) and operator--(int)
if context.endswith(' operator++') or context.endswith(' operator--'): if (context.endswith(' operator++') or context.endswith(' operator--') or
context.endswith('::operator++') or context.endswith('::operator--')):
return False return False
# A single unnamed argument for a function tends to look like old style cast. # A single unnamed argument for a function tends to look like old style cast.
@ -5614,11 +5914,11 @@ _HEADERS_CONTAINING_TEMPLATES = (
)), )),
('<limits>', ('numeric_limits',)), ('<limits>', ('numeric_limits',)),
('<list>', ('list',)), ('<list>', ('list',)),
('<map>', ('map', 'multimap',)), ('<map>', ('multimap',)),
('<memory>', ('allocator', 'make_shared', 'make_unique', 'shared_ptr', ('<memory>', ('allocator', 'make_shared', 'make_unique', 'shared_ptr',
'unique_ptr', 'weak_ptr')), 'unique_ptr', 'weak_ptr')),
('<queue>', ('queue', 'priority_queue',)), ('<queue>', ('queue', 'priority_queue',)),
('<set>', ('set', 'multiset',)), ('<set>', ('multiset',)),
('<stack>', ('stack',)), ('<stack>', ('stack',)),
('<string>', ('char_traits', 'basic_string',)), ('<string>', ('char_traits', 'basic_string',)),
('<tuple>', ('tuple',)), ('<tuple>', ('tuple',)),
@ -5652,6 +5952,16 @@ for _header, _templates in _HEADERS_MAYBE_TEMPLATES:
(re.compile(r'[^>.]\b' + _template + r'(<.*?>)?\([^\)]'), (re.compile(r'[^>.]\b' + _template + r'(<.*?>)?\([^\)]'),
_template, _template,
_header)) _header))
# Match set<type>, but not foo->set<type>, foo.set<type>
_re_pattern_headers_maybe_templates.append(
(re.compile(r'[^>.]\bset\s*\<'),
'set<>',
'<set>'))
# Match 'map<type> var' and 'std::map<type>(...)', but not 'map<type>(...)''
_re_pattern_headers_maybe_templates.append(
(re.compile(r'(std\b::\bmap\s*\<)|(^(std\b::\b)map\b\(\s*\<)'),
'map<>',
'<map>'))
# Other scripts may reach in and modify this pattern. # Other scripts may reach in and modify this pattern.
_re_pattern_templates = [] _re_pattern_templates = []
@ -5734,9 +6044,7 @@ def UpdateIncludeState(filename, include_dict, io=codecs):
""" """
headerfile = None headerfile = None
try: try:
headerfile = io.open(filename, 'r', 'utf8', 'replace') with io.open(filename, 'r', 'utf8', 'replace') as headerfile:
except IOError:
return False
linenum = 0 linenum = 0
for line in headerfile: for line in headerfile:
linenum += 1 linenum += 1
@ -5746,6 +6054,9 @@ def UpdateIncludeState(filename, include_dict, io=codecs):
include = match.group(2) include = match.group(2)
include_dict.setdefault(include, linenum) include_dict.setdefault(include, linenum)
return True return True
except IOError:
return False
def CheckForIncludeWhatYouUse(filename, clean_lines, include_state, error, def CheckForIncludeWhatYouUse(filename, clean_lines, include_state, error,
@ -6221,7 +6532,7 @@ def ProcessConfigOverrides(filename):
continue continue
try: try:
with open(cfg_file) as file_handle: with open(cfg_file, encoding='utf-8') as file_handle:
for line in file_handle: for line in file_handle:
line, _, _ = line.partition('#') # Remove comments. line, _, _ = line.partition('#') # Remove comments.
if not line.strip(): if not line.strip():
@ -6259,20 +6570,15 @@ def ProcessConfigOverrides(filename):
except ValueError: except ValueError:
_cpplint_state.PrintError('Line length must be numeric.') _cpplint_state.PrintError('Line length must be numeric.')
elif name == 'extensions': elif name == 'extensions':
global _valid_extensions ProcessExtensionsOption(val)
try:
extensions = [ext.strip() for ext in val.split(',')]
_valid_extensions = set(extensions)
except ValueError:
sys.stderr.write('Extensions should be a comma-separated list of values;'
'for example: extensions=hpp,cpp\n'
'This could not be parsed: "%s"' % (val,))
elif name == 'root': elif name == 'root':
global _root global _root
# root directories are specified relative to CPPLINT.cfg dir. # root directories are specified relative to CPPLINT.cfg dir.
_root = os.path.join(os.path.dirname(cfg_file), val) _root = os.path.join(os.path.dirname(cfg_file), val)
elif name == 'headers': elif name == 'headers':
ProcessHppHeadersOption(val) ProcessHppHeadersOption(val)
elif name == 'includeorder':
ProcessIncludeOrderOption(val)
else: else:
_cpplint_state.PrintError( _cpplint_state.PrintError(
'Invalid configuration option (%s) in file %s\n' % 'Invalid configuration option (%s) in file %s\n' %
@ -6329,7 +6635,8 @@ def ProcessFile(filename, vlevel, extra_check_functions=None):
codecs.getwriter('utf8'), codecs.getwriter('utf8'),
'replace').read().split('\n') 'replace').read().split('\n')
else: else:
lines = codecs.open(filename, 'r', 'utf8', 'replace').read().split('\n') with codecs.open(filename, 'r', 'utf8', 'replace') as target_file:
lines = target_file.read().split('\n')
# Remove trailing '\r'. # Remove trailing '\r'.
# The -1 accounts for the extra trailing blank line we get from split() # The -1 accounts for the extra trailing blank line we get from split()
@ -6389,10 +6696,10 @@ def PrintUsage(message):
Args: Args:
message: The optional error message. message: The optional error message.
""" """
sys.stderr.write(_USAGE % (list(GetAllExtensions()), sys.stderr.write(_USAGE % (sorted(list(GetAllExtensions())),
','.join(list(GetAllExtensions())), ','.join(sorted(list(GetAllExtensions()))),
GetHeaderExtensions(), sorted(GetHeaderExtensions()),
','.join(GetHeaderExtensions()))) ','.join(sorted(GetHeaderExtensions()))))
if message: if message:
sys.exit('\nFATAL ERROR: ' + message) sys.exit('\nFATAL ERROR: ' + message)
@ -6438,6 +6745,7 @@ def ParseArguments(args):
'exclude=', 'exclude=',
'recursive', 'recursive',
'headers=', 'headers=',
'includeorder=',
'quiet']) 'quiet'])
except getopt.GetoptError: except getopt.GetoptError:
PrintUsage('Invalid arguments.') PrintUsage('Invalid arguments.')
@ -6455,9 +6763,9 @@ def ParseArguments(args):
if opt == '--version': if opt == '--version':
PrintVersion() PrintVersion()
elif opt == '--output': elif opt == '--output':
if val not in ('emacs', 'vs7', 'eclipse', 'junit'): if val not in ('emacs', 'vs7', 'eclipse', 'junit', 'sed', 'gsed'):
PrintUsage('The only allowed output formats are emacs, vs7, eclipse ' PrintUsage('The only allowed output formats are emacs, vs7, eclipse '
'and junit.') 'sed, gsed and junit.')
output_format = val output_format = val
elif opt == '--quiet': elif opt == '--quiet':
quiet = True quiet = True
@ -6489,15 +6797,13 @@ def ParseArguments(args):
_excludes = set() _excludes = set()
_excludes.update(glob.glob(val)) _excludes.update(glob.glob(val))
elif opt == '--extensions': elif opt == '--extensions':
global _valid_extensions ProcessExtensionsOption(val)
try:
_valid_extensions = set(val.split(','))
except ValueError:
PrintUsage('Extensions must be comma seperated list.')
elif opt == '--headers': elif opt == '--headers':
ProcessHppHeadersOption(val) ProcessHppHeadersOption(val)
elif opt == '--recursive': elif opt == '--recursive':
recursive = True recursive = True
elif opt == '--includeorder':
ProcessIncludeOrderOption(val)
if not filenames: if not filenames:
PrintUsage('No files were specified.') PrintUsage('No files were specified.')
@ -6514,6 +6820,7 @@ def ParseArguments(args):
_SetFilters(filters) _SetFilters(filters)
_SetCountingStyle(counting_style) _SetCountingStyle(counting_style)
filenames.sort()
return filenames return filenames
def _ExpandDirectories(filenames): def _ExpandDirectories(filenames):
@ -6545,15 +6852,35 @@ def _ExpandDirectories(filenames):
for filename in expanded: for filename in expanded:
if os.path.splitext(filename)[1][1:] in GetAllExtensions(): if os.path.splitext(filename)[1][1:] in GetAllExtensions():
filtered.append(filename) filtered.append(filename)
return filtered return filtered
def _FilterExcludedFiles(filenames): def _FilterExcludedFiles(fnames):
"""Filters out files listed in the --exclude command line switch. File paths """Filters out files listed in the --exclude command line switch. File paths
in the switch are evaluated relative to the current working directory in the switch are evaluated relative to the current working directory
""" """
exclude_paths = [os.path.abspath(f) for f in _excludes] exclude_paths = [os.path.abspath(f) for f in _excludes]
return [f for f in filenames if os.path.abspath(f) not in exclude_paths] # because globbing does not work recursively, exclude all subpath of all excluded entries
return [f for f in fnames
if not any(e for e in exclude_paths
if _IsParentOrSame(e, os.path.abspath(f)))]
def _IsParentOrSame(parent, child):
"""Return true if child is subdirectory of parent.
Assumes both paths are absolute and don't contain symlinks.
"""
parent = os.path.normpath(parent)
child = os.path.normpath(child)
if parent == child:
return True
prefix = os.path.commonprefix([parent, child])
if prefix != parent:
return False
# Note: os.path.commonprefix operates on character basis, so
# take extra care of situations like '/foo/ba' and '/foo/bar/baz'
child_suffix = child[len(prefix):]
child_suffix = child_suffix.lstrip(os.sep)
return child == os.path.join(prefix, child_suffix)
def main(): def main():
filenames = ParseArguments(sys.argv[1:]) filenames = ParseArguments(sys.argv[1:])

Loading…
Cancel
Save