#!/usr/bin/env python2.7 # Copyright (c) 2015 - present Facebook, Inc. # All rights reserved. # # This source code is licensed under the BSD style license found in the # LICENSE file in the root directory of this source tree. An additional grant # of patent rights can be found in the PATENTS file in the same directory. # example usage: # # run all the tests # ./build_integration_test.py # # run only the ant and gradle tests # ./build_integration_test.py -- ant gradle # # run no test # ./build_integration_test.py -- # # run only the buck tests and record the output # INFER_RECORD_INTEGRATION_TESTS=1 ./build_integration_test.py -- buck from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import json import os import shutil import subprocess import sys import tempfile import unittest SCRIPT_DIR = os.path.dirname(os.path.realpath(__file__)) sys.path.insert(0, os.path.join(SCRIPT_DIR, os.pardir, os.pardir, 'lib', 'python')) from inferlib import issues, utils ROOT_DIR = os.path.join(SCRIPT_DIR, os.pardir, os.pardir, os.pardir) REPORT_JSON = 'report.json' INFER_EXECUTABLE = 'infer' RECORD_ENV = 'INFER_RECORD_INTEGRATION_TESTS' REPORT_FIELDS = [ issues.JSON_INDEX_FILENAME, issues.JSON_INDEX_PROCEDURE, issues.JSON_INDEX_TYPE, ] CODETOANALYZE_DIR = os.path.join(SCRIPT_DIR, 'codetoanalyze') EXPECTED_OUTPUTS_DIR = os.path.join(SCRIPT_DIR, 'expected_outputs') ALL_TESTS = ['ant', 'buck', 'gradle', 'make', 'locale', 'waf'] to_test = ALL_TESTS def should_record_tests(): return RECORD_ENV in os.environ and os.environ[RECORD_ENV] == '1' def quote(s): return '\"%s\"' % s def string_of_error(e): line = '' if issues.JSON_INDEX_LINE in e: line = ' on line %s ' % e[issues.JSON_INDEX_LINE] msg = '%s in file %s, procedure %s%s' % ( e[issues.JSON_INDEX_TYPE], quote(e[issues.JSON_INDEX_FILENAME]), quote(e[issues.JSON_INDEX_PROCEDURE]), line, ) return msg def save_report(reports, filename): # sorting to avoid spurious differences between two lists of reports reports.sort() def filter_report(report): return dict((k, v) for (k, v) in report.items() if k in REPORT_FIELDS) def should_report(report): return len(report) > 0 filtered = filter(should_report, map(filter_report, reports)) utils.dump_json_to_path(filtered, filename, separators=(',', ': '), sort_keys=True) def run_analysis(root, clean_cmd, build_cmd, analyzer, env=None, n=1): os.chdir(root) subprocess.check_call(clean_cmd, env=env) temp_out_dir = tempfile.mkdtemp(suffix='_out', prefix='infer_') infer_cmd = ['infer', '-a', analyzer, '-o', temp_out_dir, '--'] + build_cmd with tempfile.TemporaryFile( mode='w', suffix='.out', prefix='analysis_') as analysis_output: for i in xrange(n): subprocess.check_call(infer_cmd, stdout=analysis_output, env=env) json_path = os.path.join(temp_out_dir, REPORT_JSON) found_errors = utils.load_json_from_path(json_path) shutil.rmtree(temp_out_dir) os.chdir(SCRIPT_DIR) return found_errors def match_pattern(f, p): for key in p.keys(): if f[key] != p[key]: return False return True def is_expected(e, patterns): for p in patterns: if match_pattern(e, p): return True return False def is_missing(p, errors): for e in errors: if match_pattern(e, p): return False return True def unexpected_errors(errors, patterns): return [e for e in errors if not is_expected(e, patterns)] def missing_errors(errors, patterns): return [p for p in patterns if is_missing(p, errors)] def check_results(errors, patterns): unexpected = unexpected_errors(errors, patterns) if unexpected != []: utils.stderr('\nInfer found the following unexpected errors:') for e in unexpected: utils.stderr('\t{}\n'.format(string_of_error(e))) missing = missing_errors(errors, patterns) if missing != []: utils.stderr('\nInfer did not find the following errors:') for p in missing: utils.stderr('\t{}\n'.format(string_of_error(p))) assert unexpected == [] assert missing == [] def is_tool_available(cmd): try: subprocess.call(cmd) except OSError as e: if e.errno == os.errno.ENOENT: return False else: raise return True def do_test(errors, expected_errors_filename): if should_record_tests(): save_report(errors, expected_errors_filename) return else: patterns = utils.load_json_from_path(expected_errors_filename) check_results(errors, patterns) class BuildIntegrationTest(unittest.TestCase): def test_ant_integration(self): if not ('ant' in to_test and is_tool_available(['ant', '-version'])): print('\nSkipping Ant integration test') return print('\nRunning Ant integration test') root = os.path.join(SCRIPT_DIR, os.pardir) errors = run_analysis( root, ['ant', 'clean'], ['ant', 'compile'], INFER_EXECUTABLE) original = os.path.join(EXPECTED_OUTPUTS_DIR, 'ant_report.json') do_test(errors, original) def test_gradle_integration(self): if 'gradle' not in to_test: print('\nSkipping Gradle integration test') return print('\nRunning Gradle integration test using mock gradle') root = os.path.join(ROOT_DIR, 'examples', 'java_hello') env = os.environ env['PATH'] = '{}:{}'.format( os.path.join(SCRIPT_DIR, 'mock'), os.getenv('PATH'), ) errors = run_analysis( root, ['true'], ['gradle', 'build'], INFER_EXECUTABLE, env=env) original = os.path.join(EXPECTED_OUTPUTS_DIR, 'gradle_report.json') do_test(errors, original) def test_buck_integration(self): if not ('buck' in to_test and is_tool_available(['buck', '--version'])): print('\nSkipping Buck integration test') return print('\nRunning Buck integration test') errors = run_analysis( ROOT_DIR, ['buck', 'clean'], ['buck', 'build', 'infer'], INFER_EXECUTABLE) original = os.path.join(EXPECTED_OUTPUTS_DIR, 'buck_report.json') do_test(errors, original) def test_make_integration(self): if 'make' not in to_test: print('\nSkipping make integration test') return print('\nRunning make integration test') root = os.path.join(CODETOANALYZE_DIR, 'make') errors = run_analysis( root, ['make', 'clean'], ['make', 'all'], INFER_EXECUTABLE) original = os.path.join(EXPECTED_OUTPUTS_DIR, 'make_report.json') do_test(errors, original) def test_wonky_locale_integration(self): if 'locale' not in to_test: print('\nSkipping wonky locale integration test') return print('\nRunning wonky locale integration test') root = os.path.join(CODETOANALYZE_DIR, 'make') env = os.environ env['LC_ALL'] = 'C' # check that we are able to remove the previous results by # running the analysis twice errors = run_analysis( root, ['true'], ['clang', '-c', 'utf8_in_function_names.c'], INFER_EXECUTABLE, env=env, n=2) original = os.path.join(EXPECTED_OUTPUTS_DIR, 'locale_report.json') do_test(errors, original) def test_waf_integration(self): if 'waf' not in to_test: print('\nSkipping waf integration test') return print('\nRunning waf integration test') root = os.path.join(CODETOANALYZE_DIR, 'make') errors = run_analysis( root, ['make', 'clean'], ['./waf', 'build'], INFER_EXECUTABLE) original = os.path.join(EXPECTED_OUTPUTS_DIR, 'waf_report.json') do_test(errors, original) if __name__ == '__main__': # hackish capturing of the arguments after '--' try: i = sys.argv.index('--') to_test = sys.argv[i + 1:] sys.argv = sys.argv[:i] except ValueError: pass unittest.main() # run all the tests