diff options
author | Zeno Albisser <zeno.albisser@digia.com> | 2013-08-15 21:46:11 +0200 |
---|---|---|
committer | Zeno Albisser <zeno.albisser@digia.com> | 2013-08-15 21:46:11 +0200 |
commit | 679147eead574d186ebf3069647b4c23e8ccace6 (patch) | |
tree | fc247a0ac8ff119f7c8550879ebb6d3dd8d1ff69 /chromium/PRESUBMIT.py | |
download | qtwebengine-chromium-679147eead574d186ebf3069647b4c23e8ccace6.tar.gz |
Initial import.
Diffstat (limited to 'chromium/PRESUBMIT.py')
-rw-r--r-- | chromium/PRESUBMIT.py | 1061 |
1 files changed, 1061 insertions, 0 deletions
diff --git a/chromium/PRESUBMIT.py b/chromium/PRESUBMIT.py new file mode 100644 index 00000000000..0b1f6c7f0b2 --- /dev/null +++ b/chromium/PRESUBMIT.py @@ -0,0 +1,1061 @@ +# Copyright (c) 2012 The Chromium Authors. All rights reserved. +# Use of this source code is governed by a BSD-style license that can be +# found in the LICENSE file. + +"""Top-level presubmit script for Chromium. + +See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts +for more details about the presubmit API built into gcl. +""" + + +import re +import subprocess +import sys + + +_EXCLUDED_PATHS = ( + r"^breakpad[\\\/].*", + r"^native_client_sdk[\\\/]src[\\\/]build_tools[\\\/]make_rules.py", + r"^native_client_sdk[\\\/]src[\\\/]build_tools[\\\/]make_simple.py", + r"^native_client_sdk[\\\/]src[\\\/]tools[\\\/].*.mk", + r"^net[\\\/]tools[\\\/]spdyshark[\\\/].*", + r"^skia[\\\/].*", + r"^v8[\\\/].*", + r".*MakeFile$", + r".+_autogen\.h$", + r".+[\\\/]pnacl_shim\.c$", +) + +# Fragment of a regular expression that matches C++ and Objective-C++ +# implementation files. +_IMPLEMENTATION_EXTENSIONS = r'\.(cc|cpp|cxx|mm)$' + +# Regular expression that matches code only used for test binaries +# (best effort). +_TEST_CODE_EXCLUDED_PATHS = ( + r'.*[/\\](fake_|test_|mock_).+%s' % _IMPLEMENTATION_EXTENSIONS, + r'.+_test_(base|support|util)%s' % _IMPLEMENTATION_EXTENSIONS, + r'.+_(api|browser|perf|pixel|unit|ui)?test(_[a-z]+)?%s' % + _IMPLEMENTATION_EXTENSIONS, + r'.+profile_sync_service_harness%s' % _IMPLEMENTATION_EXTENSIONS, + r'.*[/\\](test|tool(s)?)[/\\].*', + # content_shell is used for running layout tests. + r'content[/\\]shell[/\\].*', + # At request of folks maintaining this folder. + r'chrome[/\\]browser[/\\]automation[/\\].*', +) + +_TEST_ONLY_WARNING = ( + 'You might be calling functions intended only for testing from\n' + 'production code. It is OK to ignore this warning if you know what\n' + 'you are doing, as the heuristics used to detect the situation are\n' + 'not perfect. The commit queue will not block on this warning.\n' + 'Email joi@chromium.org if you have questions.') + + +_INCLUDE_ORDER_WARNING = ( + 'Your #include order seems to be broken. Send mail to\n' + 'marja@chromium.org if this is not the case.') + + +_BANNED_OBJC_FUNCTIONS = ( + ( + 'addTrackingRect:', + ( + 'The use of -[NSView addTrackingRect:owner:userData:assumeInside:] is' + 'prohibited. Please use CrTrackingArea instead.', + 'http://dev.chromium.org/developers/coding-style/cocoa-dos-and-donts', + ), + False, + ), + ( + 'NSTrackingArea', + ( + 'The use of NSTrackingAreas is prohibited. Please use CrTrackingArea', + 'instead.', + 'http://dev.chromium.org/developers/coding-style/cocoa-dos-and-donts', + ), + False, + ), + ( + 'convertPointFromBase:', + ( + 'The use of -[NSView convertPointFromBase:] is almost certainly wrong.', + 'Please use |convertPoint:(point) fromView:nil| instead.', + 'http://dev.chromium.org/developers/coding-style/cocoa-dos-and-donts', + ), + True, + ), + ( + 'convertPointToBase:', + ( + 'The use of -[NSView convertPointToBase:] is almost certainly wrong.', + 'Please use |convertPoint:(point) toView:nil| instead.', + 'http://dev.chromium.org/developers/coding-style/cocoa-dos-and-donts', + ), + True, + ), + ( + 'convertRectFromBase:', + ( + 'The use of -[NSView convertRectFromBase:] is almost certainly wrong.', + 'Please use |convertRect:(point) fromView:nil| instead.', + 'http://dev.chromium.org/developers/coding-style/cocoa-dos-and-donts', + ), + True, + ), + ( + 'convertRectToBase:', + ( + 'The use of -[NSView convertRectToBase:] is almost certainly wrong.', + 'Please use |convertRect:(point) toView:nil| instead.', + 'http://dev.chromium.org/developers/coding-style/cocoa-dos-and-donts', + ), + True, + ), + ( + 'convertSizeFromBase:', + ( + 'The use of -[NSView convertSizeFromBase:] is almost certainly wrong.', + 'Please use |convertSize:(point) fromView:nil| instead.', + 'http://dev.chromium.org/developers/coding-style/cocoa-dos-and-donts', + ), + True, + ), + ( + 'convertSizeToBase:', + ( + 'The use of -[NSView convertSizeToBase:] is almost certainly wrong.', + 'Please use |convertSize:(point) toView:nil| instead.', + 'http://dev.chromium.org/developers/coding-style/cocoa-dos-and-donts', + ), + True, + ), +) + + +_BANNED_CPP_FUNCTIONS = ( + # Make sure that gtest's FRIEND_TEST() macro is not used; the + # FRIEND_TEST_ALL_PREFIXES() macro from base/gtest_prod_util.h should be + # used instead since that allows for FLAKY_ and DISABLED_ prefixes. + ( + 'FRIEND_TEST(', + ( + 'Chromium code should not use gtest\'s FRIEND_TEST() macro. Include', + 'base/gtest_prod_util.h and use FRIEND_TEST_ALL_PREFIXES() instead.', + ), + False, + (), + ), + ( + 'ScopedAllowIO', + ( + 'New code should not use ScopedAllowIO. Post a task to the blocking', + 'pool or the FILE thread instead.', + ), + True, + ( + r"^content[\\\/]shell[\\\/]shell_browser_main\.cc$", + r"^net[\\\/]disk_cache[\\\/]cache_util\.cc$", + ), + ), + ( + 'SkRefPtr', + ( + 'The use of SkRefPtr is prohibited. ', + 'Please use skia::RefPtr instead.' + ), + True, + (), + ), + ( + 'SkAutoRef', + ( + 'The indirect use of SkRefPtr via SkAutoRef is prohibited. ', + 'Please use skia::RefPtr instead.' + ), + True, + (), + ), + ( + 'SkAutoTUnref', + ( + 'The use of SkAutoTUnref is dangerous because it implicitly ', + 'converts to a raw pointer. Please use skia::RefPtr instead.' + ), + True, + (), + ), + ( + 'SkAutoUnref', + ( + 'The indirect use of SkAutoTUnref through SkAutoUnref is dangerous ', + 'because it implicitly converts to a raw pointer. ', + 'Please use skia::RefPtr instead.' + ), + True, + (), + ), +) + + +_VALID_OS_MACROS = ( + # Please keep sorted. + 'OS_ANDROID', + 'OS_BSD', + 'OS_CAT', # For testing. + 'OS_CHROMEOS', + 'OS_FREEBSD', + 'OS_IOS', + 'OS_LINUX', + 'OS_MACOSX', + 'OS_NACL', + 'OS_OPENBSD', + 'OS_POSIX', + 'OS_SOLARIS', + 'OS_WIN', +) + + +def _CheckNoProductionCodeUsingTestOnlyFunctions(input_api, output_api): + """Attempts to prevent use of functions intended only for testing in + non-testing code. For now this is just a best-effort implementation + that ignores header files and may have some false positives. A + better implementation would probably need a proper C++ parser. + """ + # We only scan .cc files and the like, as the declaration of + # for-testing functions in header files are hard to distinguish from + # calls to such functions without a proper C++ parser. + file_inclusion_pattern = r'.+%s' % _IMPLEMENTATION_EXTENSIONS + + base_function_pattern = r'ForTest(ing)?|for_test(ing)?' + inclusion_pattern = input_api.re.compile(r'(%s)\s*\(' % base_function_pattern) + comment_pattern = input_api.re.compile(r'//.*%s' % base_function_pattern) + exclusion_pattern = input_api.re.compile( + r'::[A-Za-z0-9_]+(%s)|(%s)[^;]+\{' % ( + base_function_pattern, base_function_pattern)) + + def FilterFile(affected_file): + black_list = (_EXCLUDED_PATHS + + _TEST_CODE_EXCLUDED_PATHS + + input_api.DEFAULT_BLACK_LIST) + return input_api.FilterSourceFile( + affected_file, + white_list=(file_inclusion_pattern, ), + black_list=black_list) + + problems = [] + for f in input_api.AffectedSourceFiles(FilterFile): + local_path = f.LocalPath() + lines = input_api.ReadFile(f).splitlines() + line_number = 0 + for line in lines: + if (inclusion_pattern.search(line) and + not comment_pattern.search(line) and + not exclusion_pattern.search(line)): + problems.append( + '%s:%d\n %s' % (local_path, line_number, line.strip())) + line_number += 1 + + if problems: + return [output_api.PresubmitPromptOrNotify(_TEST_ONLY_WARNING, problems)] + else: + return [] + + +def _CheckNoIOStreamInHeaders(input_api, output_api): + """Checks to make sure no .h files include <iostream>.""" + files = [] + pattern = input_api.re.compile(r'^#include\s*<iostream>', + input_api.re.MULTILINE) + for f in input_api.AffectedSourceFiles(input_api.FilterSourceFile): + if not f.LocalPath().endswith('.h'): + continue + contents = input_api.ReadFile(f) + if pattern.search(contents): + files.append(f) + + if len(files): + return [ output_api.PresubmitError( + 'Do not #include <iostream> in header files, since it inserts static ' + 'initialization into every file including the header. Instead, ' + '#include <ostream>. See http://crbug.com/94794', + files) ] + return [] + + +def _CheckNoUNIT_TESTInSourceFiles(input_api, output_api): + """Checks to make sure no source files use UNIT_TEST""" + problems = [] + for f in input_api.AffectedFiles(): + if (not f.LocalPath().endswith(('.cc', '.mm'))): + continue + + for line_num, line in f.ChangedContents(): + if 'UNIT_TEST' in line: + problems.append(' %s:%d' % (f.LocalPath(), line_num)) + + if not problems: + return [] + return [output_api.PresubmitPromptWarning('UNIT_TEST is only for headers.\n' + + '\n'.join(problems))] + + +def _CheckNoNewWStrings(input_api, output_api): + """Checks to make sure we don't introduce use of wstrings.""" + problems = [] + for f in input_api.AffectedFiles(): + if (not f.LocalPath().endswith(('.cc', '.h')) or + f.LocalPath().endswith('test.cc')): + continue + + allowWString = False + for line_num, line in f.ChangedContents(): + if 'presubmit: allow wstring' in line: + allowWString = True + elif not allowWString and 'wstring' in line: + problems.append(' %s:%d' % (f.LocalPath(), line_num)) + allowWString = False + else: + allowWString = False + + if not problems: + return [] + return [output_api.PresubmitPromptWarning('New code should not use wstrings.' + ' If you are calling a cross-platform API that accepts a wstring, ' + 'fix the API.\n' + + '\n'.join(problems))] + + +def _CheckNoDEPSGIT(input_api, output_api): + """Make sure .DEPS.git is never modified manually.""" + if any(f.LocalPath().endswith('.DEPS.git') for f in + input_api.AffectedFiles()): + return [output_api.PresubmitError( + 'Never commit changes to .DEPS.git. This file is maintained by an\n' + 'automated system based on what\'s in DEPS and your changes will be\n' + 'overwritten.\n' + 'See http://code.google.com/p/chromium/wiki/UsingNewGit#Rolling_DEPS\n' + 'for more information')] + return [] + + +def _CheckNoBannedFunctions(input_api, output_api): + """Make sure that banned functions are not used.""" + warnings = [] + errors = [] + + file_filter = lambda f: f.LocalPath().endswith(('.mm', '.m', '.h')) + for f in input_api.AffectedFiles(file_filter=file_filter): + for line_num, line in f.ChangedContents(): + for func_name, message, error in _BANNED_OBJC_FUNCTIONS: + if func_name in line: + problems = warnings; + if error: + problems = errors; + problems.append(' %s:%d:' % (f.LocalPath(), line_num)) + for message_line in message: + problems.append(' %s' % message_line) + + file_filter = lambda f: f.LocalPath().endswith(('.cc', '.mm', '.h')) + for f in input_api.AffectedFiles(file_filter=file_filter): + for line_num, line in f.ChangedContents(): + for func_name, message, error, excluded_paths in _BANNED_CPP_FUNCTIONS: + def IsBlacklisted(affected_file, blacklist): + local_path = affected_file.LocalPath() + for item in blacklist: + if input_api.re.match(item, local_path): + return True + return False + if IsBlacklisted(f, excluded_paths): + continue + if func_name in line: + problems = warnings; + if error: + problems = errors; + problems.append(' %s:%d:' % (f.LocalPath(), line_num)) + for message_line in message: + problems.append(' %s' % message_line) + + result = [] + if (warnings): + result.append(output_api.PresubmitPromptWarning( + 'Banned functions were used.\n' + '\n'.join(warnings))) + if (errors): + result.append(output_api.PresubmitError( + 'Banned functions were used.\n' + '\n'.join(errors))) + return result + + +def _CheckNoPragmaOnce(input_api, output_api): + """Make sure that banned functions are not used.""" + files = [] + pattern = input_api.re.compile(r'^#pragma\s+once', + input_api.re.MULTILINE) + for f in input_api.AffectedSourceFiles(input_api.FilterSourceFile): + if not f.LocalPath().endswith('.h'): + continue + contents = input_api.ReadFile(f) + if pattern.search(contents): + files.append(f) + + if files: + return [output_api.PresubmitError( + 'Do not use #pragma once in header files.\n' + 'See http://www.chromium.org/developers/coding-style#TOC-File-headers', + files)] + return [] + + +def _CheckNoTrinaryTrueFalse(input_api, output_api): + """Checks to make sure we don't introduce use of foo ? true : false.""" + problems = [] + pattern = input_api.re.compile(r'\?\s*(true|false)\s*:\s*(true|false)') + for f in input_api.AffectedFiles(): + if not f.LocalPath().endswith(('.cc', '.h', '.inl', '.m', '.mm')): + continue + + for line_num, line in f.ChangedContents(): + if pattern.match(line): + problems.append(' %s:%d' % (f.LocalPath(), line_num)) + + if not problems: + return [] + return [output_api.PresubmitPromptWarning( + 'Please consider avoiding the "? true : false" pattern if possible.\n' + + '\n'.join(problems))] + + +def _CheckUnwantedDependencies(input_api, output_api): + """Runs checkdeps on #include statements added in this + change. Breaking - rules is an error, breaking ! rules is a + warning. + """ + # We need to wait until we have an input_api object and use this + # roundabout construct to import checkdeps because this file is + # eval-ed and thus doesn't have __file__. + original_sys_path = sys.path + try: + sys.path = sys.path + [input_api.os_path.join( + input_api.PresubmitLocalPath(), 'tools', 'checkdeps')] + import checkdeps + from cpp_checker import CppChecker + from rules import Rule + finally: + # Restore sys.path to what it was before. + sys.path = original_sys_path + + added_includes = [] + for f in input_api.AffectedFiles(): + if not CppChecker.IsCppFile(f.LocalPath()): + continue + + changed_lines = [line for line_num, line in f.ChangedContents()] + added_includes.append([f.LocalPath(), changed_lines]) + + deps_checker = checkdeps.DepsChecker(input_api.PresubmitLocalPath()) + + error_descriptions = [] + warning_descriptions = [] + for path, rule_type, rule_description in deps_checker.CheckAddedCppIncludes( + added_includes): + description_with_path = '%s\n %s' % (path, rule_description) + if rule_type == Rule.DISALLOW: + error_descriptions.append(description_with_path) + else: + warning_descriptions.append(description_with_path) + + results = [] + if error_descriptions: + results.append(output_api.PresubmitError( + 'You added one or more #includes that violate checkdeps rules.', + error_descriptions)) + if warning_descriptions: + results.append(output_api.PresubmitPromptOrNotify( + 'You added one or more #includes of files that are temporarily\n' + 'allowed but being removed. Can you avoid introducing the\n' + '#include? See relevant DEPS file(s) for details and contacts.', + warning_descriptions)) + return results + + +def _CheckFilePermissions(input_api, output_api): + """Check that all files have their permissions properly set.""" + args = [sys.executable, 'tools/checkperms/checkperms.py', '--root', + input_api.change.RepositoryRoot()] + for f in input_api.AffectedFiles(): + args += ['--file', f.LocalPath()] + errors = [] + (errors, stderrdata) = subprocess.Popen(args).communicate() + + results = [] + if errors: + results.append(output_api.PresubmitError('checkperms.py failed.', + errors)) + return results + + +def _CheckNoAuraWindowPropertyHInHeaders(input_api, output_api): + """Makes sure we don't include ui/aura/window_property.h + in header files. + """ + pattern = input_api.re.compile(r'^#include\s*"ui/aura/window_property.h"') + errors = [] + for f in input_api.AffectedFiles(): + if not f.LocalPath().endswith('.h'): + continue + for line_num, line in f.ChangedContents(): + if pattern.match(line): + errors.append(' %s:%d' % (f.LocalPath(), line_num)) + + results = [] + if errors: + results.append(output_api.PresubmitError( + 'Header files should not include ui/aura/window_property.h', errors)) + return results + + +def _CheckIncludeOrderForScope(scope, input_api, file_path, changed_linenums): + """Checks that the lines in scope occur in the right order. + + 1. C system files in alphabetical order + 2. C++ system files in alphabetical order + 3. Project's .h files + """ + + c_system_include_pattern = input_api.re.compile(r'\s*#include <.*\.h>') + cpp_system_include_pattern = input_api.re.compile(r'\s*#include <.*>') + custom_include_pattern = input_api.re.compile(r'\s*#include ".*') + + C_SYSTEM_INCLUDES, CPP_SYSTEM_INCLUDES, CUSTOM_INCLUDES = range(3) + + state = C_SYSTEM_INCLUDES + + previous_line = '' + previous_line_num = 0 + problem_linenums = [] + for line_num, line in scope: + if c_system_include_pattern.match(line): + if state != C_SYSTEM_INCLUDES: + problem_linenums.append((line_num, previous_line_num)) + elif previous_line and previous_line > line: + problem_linenums.append((line_num, previous_line_num)) + elif cpp_system_include_pattern.match(line): + if state == C_SYSTEM_INCLUDES: + state = CPP_SYSTEM_INCLUDES + elif state == CUSTOM_INCLUDES: + problem_linenums.append((line_num, previous_line_num)) + elif previous_line and previous_line > line: + problem_linenums.append((line_num, previous_line_num)) + elif custom_include_pattern.match(line): + if state != CUSTOM_INCLUDES: + state = CUSTOM_INCLUDES + elif previous_line and previous_line > line: + problem_linenums.append((line_num, previous_line_num)) + else: + problem_linenums.append(line_num) + previous_line = line + previous_line_num = line_num + + warnings = [] + for (line_num, previous_line_num) in problem_linenums: + if line_num in changed_linenums or previous_line_num in changed_linenums: + warnings.append(' %s:%d' % (file_path, line_num)) + return warnings + + +def _CheckIncludeOrderInFile(input_api, f, changed_linenums): + """Checks the #include order for the given file f.""" + + system_include_pattern = input_api.re.compile(r'\s*#include \<.*') + # Exclude #include <.../...> includes from the check; e.g., <sys/...> includes + # often need to appear in a specific order. + excluded_include_pattern = input_api.re.compile(r'\s*#include \<.*/.*') + custom_include_pattern = input_api.re.compile(r'\s*#include "(?P<FILE>.*)"') + if_pattern = input_api.re.compile( + r'\s*#\s*(if|elif|else|endif|define|undef).*') + # Some files need specialized order of includes; exclude such files from this + # check. + uncheckable_includes_pattern = input_api.re.compile( + r'\s*#include ' + '("ipc/.*macros\.h"|<windows\.h>|".*gl.*autogen.h")\s*') + + contents = f.NewContents() + warnings = [] + line_num = 0 + + # Handle the special first include. If the first include file is + # some/path/file.h, the corresponding including file can be some/path/file.cc, + # some/other/path/file.cc, some/path/file_platform.cc, some/path/file-suffix.h + # etc. It's also possible that no special first include exists. + for line in contents: + line_num += 1 + if system_include_pattern.match(line): + # No special first include -> process the line again along with normal + # includes. + line_num -= 1 + break + match = custom_include_pattern.match(line) + if match: + match_dict = match.groupdict() + header_basename = input_api.os_path.basename( + match_dict['FILE']).replace('.h', '') + if header_basename not in input_api.os_path.basename(f.LocalPath()): + # No special first include -> process the line again along with normal + # includes. + line_num -= 1 + break + + # Split into scopes: Each region between #if and #endif is its own scope. + scopes = [] + current_scope = [] + for line in contents[line_num:]: + line_num += 1 + if uncheckable_includes_pattern.match(line): + return [] + if if_pattern.match(line): + scopes.append(current_scope) + current_scope = [] + elif ((system_include_pattern.match(line) or + custom_include_pattern.match(line)) and + not excluded_include_pattern.match(line)): + current_scope.append((line_num, line)) + scopes.append(current_scope) + + for scope in scopes: + warnings.extend(_CheckIncludeOrderForScope(scope, input_api, f.LocalPath(), + changed_linenums)) + return warnings + + +def _CheckIncludeOrder(input_api, output_api): + """Checks that the #include order is correct. + + 1. The corresponding header for source files. + 2. C system files in alphabetical order + 3. C++ system files in alphabetical order + 4. Project's .h files in alphabetical order + + Each region separated by #if, #elif, #else, #endif, #define and #undef follows + these rules separately. + """ + + warnings = [] + for f in input_api.AffectedFiles(): + if f.LocalPath().endswith(('.cc', '.h')): + changed_linenums = set(line_num for line_num, _ in f.ChangedContents()) + warnings.extend(_CheckIncludeOrderInFile(input_api, f, changed_linenums)) + + results = [] + if warnings: + results.append(output_api.PresubmitPromptOrNotify(_INCLUDE_ORDER_WARNING, + warnings)) + return results + + +def _CheckForVersionControlConflictsInFile(input_api, f): + pattern = input_api.re.compile('^(?:<<<<<<<|>>>>>>>) |^=======$') + errors = [] + for line_num, line in f.ChangedContents(): + if pattern.match(line): + errors.append(' %s:%d %s' % (f.LocalPath(), line_num, line)) + return errors + + +def _CheckForVersionControlConflicts(input_api, output_api): + """Usually this is not intentional and will cause a compile failure.""" + errors = [] + for f in input_api.AffectedFiles(): + errors.extend(_CheckForVersionControlConflictsInFile(input_api, f)) + + results = [] + if errors: + results.append(output_api.PresubmitError( + 'Version control conflict markers found, please resolve.', errors)) + return results + + +def _CheckHardcodedGoogleHostsInLowerLayers(input_api, output_api): + def FilterFile(affected_file): + """Filter function for use with input_api.AffectedSourceFiles, + below. This filters out everything except non-test files from + top-level directories that generally speaking should not hard-code + service URLs (e.g. src/android_webview/, src/content/ and others). + """ + return input_api.FilterSourceFile( + affected_file, + white_list=(r'^(android_webview|base|content|net)[\\\/].*', ), + black_list=(_EXCLUDED_PATHS + + _TEST_CODE_EXCLUDED_PATHS + + input_api.DEFAULT_BLACK_LIST)) + + base_pattern = '"[^"]*google\.com[^"]*"' + comment_pattern = input_api.re.compile('//.*%s' % base_pattern) + pattern = input_api.re.compile(base_pattern) + problems = [] # items are (filename, line_number, line) + for f in input_api.AffectedSourceFiles(FilterFile): + for line_num, line in f.ChangedContents(): + if not comment_pattern.search(line) and pattern.search(line): + problems.append((f.LocalPath(), line_num, line)) + + if problems: + return [output_api.PresubmitPromptOrNotify( + 'Most layers below src/chrome/ should not hardcode service URLs.\n' + 'Are you sure this is correct? (Contact: joi@chromium.org)', + [' %s:%d: %s' % ( + problem[0], problem[1], problem[2]) for problem in problems])] + else: + return [] + + +def _CheckNoAbbreviationInPngFileName(input_api, output_api): + """Makes sure there are no abbreviations in the name of PNG files. + """ + pattern = input_api.re.compile(r'.*_[a-z]_.*\.png$|.*_[a-z]\.png$') + errors = [] + for f in input_api.AffectedFiles(include_deletes=False): + if pattern.match(f.LocalPath()): + errors.append(' %s' % f.LocalPath()) + + results = [] + if errors: + results.append(output_api.PresubmitError( + 'The name of PNG files should not have abbreviations. \n' + 'Use _hover.png, _center.png, instead of _h.png, _c.png.\n' + 'Contact oshima@chromium.org if you have questions.', errors)) + return results + + +def _DepsFilesToCheck(re, changed_lines): + """Helper method for _CheckAddedDepsHaveTargetApprovals. Returns + a set of DEPS entries that we should look up.""" + results = set() + + # This pattern grabs the path without basename in the first + # parentheses, and the basename (if present) in the second. It + # relies on the simple heuristic that if there is a basename it will + # be a header file ending in ".h". + pattern = re.compile( + r"""['"]\+([^'"]+?)(/[a-zA-Z0-9_]+\.h)?['"].*""") + for changed_line in changed_lines: + m = pattern.match(changed_line) + if m: + path = m.group(1) + if not (path.startswith('grit/') or path == 'grit'): + results.add('%s/DEPS' % m.group(1)) + return results + + +def _CheckAddedDepsHaveTargetApprovals(input_api, output_api): + """When a dependency prefixed with + is added to a DEPS file, we + want to make sure that the change is reviewed by an OWNER of the + target file or directory, to avoid layering violations from being + introduced. This check verifies that this happens. + """ + changed_lines = set() + for f in input_api.AffectedFiles(): + filename = input_api.os_path.basename(f.LocalPath()) + if filename == 'DEPS': + changed_lines |= set(line.strip() + for line_num, line + in f.ChangedContents()) + if not changed_lines: + return [] + + virtual_depended_on_files = _DepsFilesToCheck(input_api.re, changed_lines) + if not virtual_depended_on_files: + return [] + + if input_api.is_committing: + if input_api.tbr: + return [output_api.PresubmitNotifyResult( + '--tbr was specified, skipping OWNERS check for DEPS additions')] + if not input_api.change.issue: + return [output_api.PresubmitError( + "DEPS approval by OWNERS check failed: this change has " + "no Rietveld issue number, so we can't check it for approvals.")] + output = output_api.PresubmitError + else: + output = output_api.PresubmitNotifyResult + + owners_db = input_api.owners_db + owner_email, reviewers = input_api.canned_checks._RietveldOwnerAndReviewers( + input_api, + owners_db.email_regexp, + approval_needed=input_api.is_committing) + + owner_email = owner_email or input_api.change.author_email + + reviewers_plus_owner = set(reviewers) + if owner_email: + reviewers_plus_owner.add(owner_email) + missing_files = owners_db.files_not_covered_by(virtual_depended_on_files, + reviewers_plus_owner) + unapproved_dependencies = ["'+%s'," % path[:-len('/DEPS')] + for path in missing_files] + + if unapproved_dependencies: + output_list = [ + output('Missing LGTM from OWNERS of directories added to DEPS:\n %s' % + '\n '.join(sorted(unapproved_dependencies)))] + if not input_api.is_committing: + suggested_owners = owners_db.reviewers_for(missing_files, owner_email) + output_list.append(output( + 'Suggested missing target path OWNERS:\n %s' % + '\n '.join(suggested_owners or []))) + return output_list + + return [] + + +def _CommonChecks(input_api, output_api): + """Checks common to both upload and commit.""" + results = [] + results.extend(input_api.canned_checks.PanProjectChecks( + input_api, output_api, excluded_paths=_EXCLUDED_PATHS)) + results.extend(_CheckAuthorizedAuthor(input_api, output_api)) + results.extend( + _CheckNoProductionCodeUsingTestOnlyFunctions(input_api, output_api)) + results.extend(_CheckNoIOStreamInHeaders(input_api, output_api)) + results.extend(_CheckNoUNIT_TESTInSourceFiles(input_api, output_api)) + results.extend(_CheckNoNewWStrings(input_api, output_api)) + results.extend(_CheckNoDEPSGIT(input_api, output_api)) + results.extend(_CheckNoBannedFunctions(input_api, output_api)) + results.extend(_CheckNoPragmaOnce(input_api, output_api)) + results.extend(_CheckNoTrinaryTrueFalse(input_api, output_api)) + results.extend(_CheckUnwantedDependencies(input_api, output_api)) + results.extend(_CheckFilePermissions(input_api, output_api)) + results.extend(_CheckNoAuraWindowPropertyHInHeaders(input_api, output_api)) + results.extend(_CheckIncludeOrder(input_api, output_api)) + results.extend(_CheckForVersionControlConflicts(input_api, output_api)) + results.extend(_CheckPatchFiles(input_api, output_api)) + results.extend(_CheckHardcodedGoogleHostsInLowerLayers(input_api, output_api)) + results.extend(_CheckNoAbbreviationInPngFileName(input_api, output_api)) + results.extend(_CheckForInvalidOSMacros(input_api, output_api)) + results.extend(_CheckAddedDepsHaveTargetApprovals(input_api, output_api)) + results.extend( + input_api.canned_checks.CheckChangeHasNoTabs( + input_api, + output_api, + source_file_filter=lambda x: x.LocalPath().endswith('.grd'))) + + if any('PRESUBMIT.py' == f.LocalPath() for f in input_api.AffectedFiles()): + results.extend(input_api.canned_checks.RunUnitTestsInDirectory( + input_api, output_api, + input_api.PresubmitLocalPath(), + whitelist=[r'^PRESUBMIT_test\.py$'])) + return results + + +def _CheckSubversionConfig(input_api, output_api): + """Verifies the subversion config file is correctly setup. + + Checks that autoprops are enabled, returns an error otherwise. + """ + join = input_api.os_path.join + if input_api.platform == 'win32': + appdata = input_api.environ.get('APPDATA', '') + if not appdata: + return [output_api.PresubmitError('%APPDATA% is not configured.')] + path = join(appdata, 'Subversion', 'config') + else: + home = input_api.environ.get('HOME', '') + if not home: + return [output_api.PresubmitError('$HOME is not configured.')] + path = join(home, '.subversion', 'config') + + error_msg = ( + 'Please look at http://dev.chromium.org/developers/coding-style to\n' + 'configure your subversion configuration file. This enables automatic\n' + 'properties to simplify the project maintenance.\n' + 'Pro-tip: just download and install\n' + 'http://src.chromium.org/viewvc/chrome/trunk/tools/build/slave/config\n') + + try: + lines = open(path, 'r').read().splitlines() + # Make sure auto-props is enabled and check for 2 Chromium standard + # auto-prop. + if (not '*.cc = svn:eol-style=LF' in lines or + not '*.pdf = svn:mime-type=application/pdf' in lines or + not 'enable-auto-props = yes' in lines): + return [ + output_api.PresubmitNotifyResult( + 'It looks like you have not configured your subversion config ' + 'file or it is not up-to-date.\n' + error_msg) + ] + except (OSError, IOError): + return [ + output_api.PresubmitNotifyResult( + 'Can\'t find your subversion config file.\n' + error_msg) + ] + return [] + + +def _CheckAuthorizedAuthor(input_api, output_api): + """For non-googler/chromites committers, verify the author's email address is + in AUTHORS. + """ + # TODO(maruel): Add it to input_api? + import fnmatch + + author = input_api.change.author_email + if not author: + input_api.logging.info('No author, skipping AUTHOR check') + return [] + authors_path = input_api.os_path.join( + input_api.PresubmitLocalPath(), 'AUTHORS') + valid_authors = ( + input_api.re.match(r'[^#]+\s+\<(.+?)\>\s*$', line) + for line in open(authors_path)) + valid_authors = [item.group(1).lower() for item in valid_authors if item] + if not any(fnmatch.fnmatch(author.lower(), valid) for valid in valid_authors): + input_api.logging.info('Valid authors are %s', ', '.join(valid_authors)) + return [output_api.PresubmitPromptWarning( + ('%s is not in AUTHORS file. If you are a new contributor, please visit' + '\n' + 'http://www.chromium.org/developers/contributing-code and read the ' + '"Legal" section\n' + 'If you are a chromite, verify the contributor signed the CLA.') % + author)] + return [] + + +def _CheckPatchFiles(input_api, output_api): + problems = [f.LocalPath() for f in input_api.AffectedFiles() + if f.LocalPath().endswith(('.orig', '.rej'))] + if problems: + return [output_api.PresubmitError( + "Don't commit .rej and .orig files.", problems)] + else: + return [] + + +def _DidYouMeanOSMacro(bad_macro): + try: + return {'A': 'OS_ANDROID', + 'B': 'OS_BSD', + 'C': 'OS_CHROMEOS', + 'F': 'OS_FREEBSD', + 'L': 'OS_LINUX', + 'M': 'OS_MACOSX', + 'N': 'OS_NACL', + 'O': 'OS_OPENBSD', + 'P': 'OS_POSIX', + 'S': 'OS_SOLARIS', + 'W': 'OS_WIN'}[bad_macro[3].upper()] + except KeyError: + return '' + + +def _CheckForInvalidOSMacrosInFile(input_api, f): + """Check for sensible looking, totally invalid OS macros.""" + preprocessor_statement = input_api.re.compile(r'^\s*#') + os_macro = input_api.re.compile(r'defined\((OS_[^)]+)\)') + results = [] + for lnum, line in f.ChangedContents(): + if preprocessor_statement.search(line): + for match in os_macro.finditer(line): + if not match.group(1) in _VALID_OS_MACROS: + good = _DidYouMeanOSMacro(match.group(1)) + did_you_mean = ' (did you mean %s?)' % good if good else '' + results.append(' %s:%d %s%s' % (f.LocalPath(), + lnum, + match.group(1), + did_you_mean)) + return results + + +def _CheckForInvalidOSMacros(input_api, output_api): + """Check all affected files for invalid OS macros.""" + bad_macros = [] + for f in input_api.AffectedFiles(): + if not f.LocalPath().endswith(('.py', '.js', '.html', '.css')): + bad_macros.extend(_CheckForInvalidOSMacrosInFile(input_api, f)) + + if not bad_macros: + return [] + + return [output_api.PresubmitError( + 'Possibly invalid OS macro[s] found. Please fix your code\n' + 'or add your macro to src/PRESUBMIT.py.', bad_macros)] + + +def CheckChangeOnUpload(input_api, output_api): + results = [] + results.extend(_CommonChecks(input_api, output_api)) + return results + + +def CheckChangeOnCommit(input_api, output_api): + results = [] + results.extend(_CommonChecks(input_api, output_api)) + # TODO(thestig) temporarily disabled, doesn't work in third_party/ + #results.extend(input_api.canned_checks.CheckSvnModifiedDirectories( + # input_api, output_api, sources)) + # Make sure the tree is 'open'. + results.extend(input_api.canned_checks.CheckTreeIsOpen( + input_api, + output_api, + json_url='http://chromium-status.appspot.com/current?format=json')) + results.extend(input_api.canned_checks.CheckRietveldTryJobExecution(input_api, + output_api, 'http://codereview.chromium.org', + ('win_rel', 'linux_rel', 'mac_rel, win:compile'), + 'tryserver@chromium.org')) + + results.extend(input_api.canned_checks.CheckChangeHasBugField( + input_api, output_api)) + results.extend(input_api.canned_checks.CheckChangeHasDescription( + input_api, output_api)) + results.extend(_CheckSubversionConfig(input_api, output_api)) + return results + + +def GetPreferredTrySlaves(project, change): + files = change.LocalPaths() + + if not files or all(re.search(r'[\\/]OWNERS$', f) for f in files): + return [] + + if all(re.search('\.(m|mm)$|(^|[/_])mac[/_.]', f) for f in files): + return ['mac_rel', 'mac:compile'] + if all(re.search('(^|[/_])win[/_.]', f) for f in files): + return ['win_rel', 'win7_aura', 'win:compile'] + if all(re.search('(^|[/_])android[/_.]', f) for f in files): + return ['android_aosp', 'android_dbg', 'android_clang_dbg'] + if all(re.search('^native_client_sdk', f) for f in files): + return ['linux_nacl_sdk', 'win_nacl_sdk', 'mac_nacl_sdk'] + if all(re.search('[/_]ios[/_.]', f) for f in files): + return ['ios_rel_device', 'ios_dbg_simulator'] + + trybots = [ + 'android_clang_dbg', + 'android_dbg', + 'ios_dbg_simulator', + 'ios_rel_device', + 'linux_asan', + 'linux_aura', + 'linux_chromeos', + 'linux_clang:compile', + 'linux_rel', + 'mac_rel', + 'mac:compile', + 'win7_aura', + 'win_rel', + 'win:compile', + 'win_x64_rel:compile', + ] + + # Match things like path/aura/file.cc and path/file_aura.cc. + # Same for chromeos. + if any(re.search('[/_](aura|chromeos)', f) for f in files): + trybots += ['linux_chromeos_clang:compile', 'linux_chromeos_asan'] + + # The AOSP bot doesn't build the chrome/ layer, so ignore any changes to it + # unless they're .gyp(i) files as changes to those files can break the gyp + # step on that bot. + if (not all(re.search('^chrome', f) for f in files) or + any(re.search('\.gypi?$', f) for f in files)): + trybots += ['android_aosp'] + + return trybots |