Convert Inferprint and Inferanalyze to Reason

Reviewed By: jberdine

Differential Revision: D3750835

fbshipit-source-id: 5c89353
master
Cristiano Calcagno 8 years ago committed by Facebook Github Bot 9
parent 26a6594b90
commit ed8d112e87

@ -48,7 +48,7 @@ endif
#### Backend declarations #### #### Backend declarations ####
INFER_MAIN = backend/infer INFER_MAIN = backend/infer
INFERANALYZE_MAIN = backend/inferanalyze INFERANALYZE_MAIN = backend/InferAnalyze
#### Checkers declarations #### #### Checkers declarations ####
@ -62,7 +62,7 @@ INFERPRINT_ATDGEN_STUB_BASE = backend/jsonbug
INFERPRINT_ATDGEN_STUB_ATD = $(INFERPRINT_ATDGEN_STUB_BASE).atd INFERPRINT_ATDGEN_STUB_ATD = $(INFERPRINT_ATDGEN_STUB_BASE).atd
INFERPRINT_ATDGEN_STUBS = $(addprefix $(INFERPRINT_ATDGEN_STUB_BASE), $(ATDGEN_SUFFIXES)) INFERPRINT_ATDGEN_STUBS = $(addprefix $(INFERPRINT_ATDGEN_STUB_BASE), $(ATDGEN_SUFFIXES))
INFERPRINT_MAIN = backend/inferprint INFERPRINT_MAIN = backend/InferPrint
### InferUnit declarations ### ### InferUnit declarations ###

@ -0,0 +1,157 @@
/*
* Copyright (c) 2009 - 2013 Monoidics ltd.
* Copyright (c) 2013 - present Facebook, Inc.
* All rights reserved.
*
* This source code is licensed under the BSD style license found in the
* LICENSE file in the root directory of this source tree. An additional grant
* of patent rights can be found in the PATENTS file in the same directory.
*/
open! Utils;
/** Main module for the analysis after the capture phase */
let module L = Logging;
let module F = Format;
let () = {
if Config.print_builtins {
Builtin.print_and_exit ()
};
switch Config.modified_targets {
| Some file => MergeCapture.modified_file file
| None => ()
};
if (not (Sys.file_exists Config.results_dir)) {
L.err "ERROR: results directory %s does not exist@.@." Config.results_dir;
Config.print_usage_exit ()
}
};
let analyze_exe_env exe_env => {
let init_time = Unix.gettimeofday ();
L.log_progress_file ();
Specs.clear_spec_tbl ();
Random.self_init ();
let line_reader = Printer.LineReader.create ();
if Config.checkers {
/* run the checkers only */
let call_graph = Exe_env.get_cg exe_env;
Callbacks.iterate_callbacks Checkers.ST.store_summary call_graph exe_env
} else {
/* run the full analysis */
Interproc.do_analysis exe_env;
Printer.write_all_html_files line_reader exe_env;
Interproc.print_stats exe_env;
let elapsed = Unix.gettimeofday () -. init_time;
L.out "Interprocedural footprint analysis terminated in %f sec@." elapsed
}
};
/** Create an exe_env from a cluster. */
let exe_env_from_cluster cluster => {
let _exe_env = Exe_env.create ();
let source_dirs = [cluster];
let sorted_dirs = IList.sort DB.source_dir_compare source_dirs;
IList.iter (fun src_dir => ignore (Exe_env.add_cg _exe_env src_dir)) sorted_dirs;
let exe_env = Exe_env.freeze _exe_env;
exe_env
};
/** Analyze a cluster of files */
let analyze_cluster cluster_num (cluster: Cluster.t) => {
let exe_env = exe_env_from_cluster cluster;
let defined_procs = Cg.get_defined_nodes (Exe_env.get_cg exe_env);
let num_procs = IList.length defined_procs;
L.err "@.Processing cluster #%d with %d procedures@." (cluster_num + 1) num_procs;
analyze_exe_env exe_env
};
let output_json_makefile_stats clusters => {
let clusters_to_analyze = IList.filter ClusterMakefile.cluster_should_be_analyzed clusters;
let num_files = IList.length clusters_to_analyze;
let num_procs = 0;
/* can't compute it at this stage */
let num_lines = 0;
let file_stats =
`Assoc [("files", `Int num_files), ("procedures", `Int num_procs), ("lines", `Int num_lines)];
/* write stats file to disk, intentionally overwriting old file if it already exists */
let f = open_out (Filename.concat Config.results_dir Config.proc_stats_filename);
Yojson.Basic.pretty_to_channel f file_stats
};
let print_prolog () =>
switch Config.cluster_cmdline {
| None =>
L.stdout "Starting analysis (Infer version %s)@\n" Version.versionString;
L.stdout "@\n";
L.stdout "legend:@\n";
L.stdout " \"%s\" analyzing a file@\n" Config.log_analysis_file;
L.stdout " \"%s\" analyzing a procedure@\n" Config.log_analysis_procedure;
if Config.stats_mode {
L.stdout " \"%s\" analyzer crashed@\n" Config.log_analysis_crash;
L.stdout
" \"%s\" timeout: procedure analysis took too much time@\n"
Config.log_analysis_wallclock_timeout;
L.stdout
" \"%s\" timeout: procedure analysis took too many symbolic execution steps@\n"
Config.log_analysis_symops_timeout;
L.stdout
" \"%s\" timeout: procedure analysis took too many recursive iterations@\n"
Config.log_analysis_recursion_timeout
};
L.stdout "@\n@?"
| Some clname => L.stdout "Cluster %s@." clname
};
let process_cluster_cmdline fname =>
switch (Cluster.load_from_file (DB.filename_from_string fname)) {
| None => L.err "Cannot find cluster file %s@." fname
| Some (nr, cluster) => analyze_cluster (nr - 1) cluster
};
let register_perf_stats_report () => {
let stats_dir = Filename.concat Config.results_dir Config.backend_stats_dir_name;
let cluster =
switch Config.cluster_cmdline {
| Some cl => "_" ^ cl
| None => ""
};
let stats_base = Config.perf_stats_prefix ^ "_" ^ Filename.basename cluster ^ ".json";
let stats_file = Filename.concat stats_dir stats_base;
DB.create_dir Config.results_dir;
DB.create_dir stats_dir;
PerfStats.register_report_at_exit stats_file
};
let () = {
register_perf_stats_report ();
if Config.developer_mode {
Printexc.record_backtrace true
};
print_prolog ();
RegisterCheckers.register ();
if (Config.allow_specs_cleanup == true && Config.cluster_cmdline == None) {
DB.Results_dir.clean_specs_dir ()
};
switch Config.cluster_cmdline {
| Some fname => process_cluster_cmdline fname
| None =>
if Config.merge {
MergeCapture.merge_captured_targets ()
};
let clusters = DB.find_source_dirs ();
L.err "Found %d source files in %s@." (IList.length clusters) Config.results_dir;
if (Config.makefile_cmdline != "") {
ClusterMakefile.create_cluster_makefile clusters Config.makefile_cmdline
} else {
IList.iteri (fun i cluster => analyze_cluster i cluster) clusters;
L.stdout "Analysis finished in %as@." pp_elapsed_time ()
};
output_json_makefile_stats clusters
}
};

@ -1,4 +1,4 @@
(* /*
* Copyright (c) 2009 - 2013 Monoidics ltd. * Copyright (c) 2009 - 2013 Monoidics ltd.
* Copyright (c) 2013 - present Facebook, Inc. * Copyright (c) 2013 - present Facebook, Inc.
* All rights reserved. * All rights reserved.
@ -6,8 +6,8 @@
* This source code is licensed under the BSD style license found in the * This source code is licensed under the BSD style license found in the
* LICENSE file in the root directory of this source tree. An additional grant * LICENSE file in the root directory of this source tree. An additional grant
* of patent rights can be found in the PATENTS file in the same directory. * of patent rights can be found in the PATENTS file in the same directory.
*) */
open! Utils open! Utils;
(** Main module for the analysis after the capture phase *) /** Main module for the analysis after the capture phase */

File diff suppressed because it is too large Load Diff

@ -2,7 +2,7 @@
The back end is responsible for the analysis of a project starting from an intermediate representation stored in the results directory, typically `infer-out`. The back end is responsible for the analysis of a project starting from an intermediate representation stored in the results directory, typically `infer-out`.
The main entry point is module [Inferanalyze](inferanalyze.ml) which produces the back-end executable `InferAnalyze`. The main entry point is module [InferAnalyze](InferAnalyze.re) which produces the back-end executable `InferAnalyze`.
Module [Inferprint](inferprint.ml) produces the executable `InferPrint`, which is used to export analysis results. Module [InferPrint](InferPrint.re) produces the executable `InferPrint`, which is used to export analysis results.

@ -1,147 +0,0 @@
(*
* Copyright (c) 2009 - 2013 Monoidics ltd.
* Copyright (c) 2013 - present Facebook, Inc.
* All rights reserved.
*
* This source code is licensed under the BSD style license found in the
* LICENSE file in the root directory of this source tree. An additional grant
* of patent rights can be found in the PATENTS file in the same directory.
*)
open! Utils
(** Main module for the analysis after the capture phase *)
module L = Logging
module F = Format
let () =
if Config.print_builtins then Builtin.print_and_exit () ;
(match Config.modified_targets with Some file -> MergeCapture.modified_file file | None -> ()) ;
if not (Sys.file_exists Config.results_dir) then
begin
L.err "ERROR: results directory %s does not exist@.@." Config.results_dir;
Config.print_usage_exit ()
end
let analyze_exe_env exe_env =
let init_time = Unix.gettimeofday () in
L.log_progress_file ();
Specs.clear_spec_tbl ();
Random.self_init ();
let line_reader = Printer.LineReader.create () in
if Config.checkers then
begin
(* run the checkers only *)
let call_graph = Exe_env.get_cg exe_env in
Callbacks.iterate_callbacks Checkers.ST.store_summary call_graph exe_env
end
else
begin
(* run the full analysis *)
Interproc.do_analysis exe_env;
Printer.write_all_html_files line_reader exe_env;
Interproc.print_stats exe_env;
let elapsed = Unix.gettimeofday () -. init_time in
L.out "Interprocedural footprint analysis terminated in %f sec@." elapsed
end
(** Create an exe_env from a cluster. *)
let exe_env_from_cluster cluster =
let _exe_env = Exe_env.create () in
let source_dirs = [cluster] in
let sorted_dirs = IList.sort DB.source_dir_compare source_dirs in
IList.iter (fun src_dir -> ignore (Exe_env.add_cg _exe_env src_dir)) sorted_dirs;
let exe_env = Exe_env.freeze _exe_env in
exe_env
(** Analyze a cluster of files *)
let analyze_cluster cluster_num (cluster : Cluster.t) =
let exe_env = exe_env_from_cluster cluster in
let defined_procs = Cg.get_defined_nodes (Exe_env.get_cg exe_env) in
let num_procs = IList.length defined_procs in
L.err "@.Processing cluster #%d with %d procedures@."
(cluster_num + 1) num_procs;
analyze_exe_env exe_env
let output_json_makefile_stats clusters =
let clusters_to_analyze =
IList.filter ClusterMakefile.cluster_should_be_analyzed clusters in
let num_files = IList.length clusters_to_analyze in
let num_procs = 0 (* can't compute it at this stage *) in
let num_lines = 0 in
let file_stats =
`Assoc [ ("files", `Int num_files);
("procedures", `Int num_procs);
("lines", `Int num_lines) ] in
(* write stats file to disk, intentionally overwriting old file if it already exists *)
let f = open_out (Filename.concat Config.results_dir Config.proc_stats_filename) in
Yojson.Basic.pretty_to_channel f file_stats
let print_prolog () =
match Config.cluster_cmdline with
| None ->
L.stdout "Starting analysis (Infer version %s)@\n" Version.versionString;
L.stdout "@\n";
L.stdout "legend:@\n";
L.stdout " \"%s\" analyzing a file@\n" Config.log_analysis_file;
L.stdout " \"%s\" analyzing a procedure@\n" Config.log_analysis_procedure;
if Config.stats_mode then (
L.stdout " \"%s\" analyzer crashed@\n" Config.log_analysis_crash;
L.stdout " \"%s\" timeout: procedure analysis took too much time@\n"
Config.log_analysis_wallclock_timeout;
L.stdout " \"%s\" timeout: procedure analysis took too many symbolic execution steps@\n"
Config.log_analysis_symops_timeout;
L.stdout " \"%s\" timeout: procedure analysis took too many recursive iterations@\n"
Config.log_analysis_recursion_timeout;
);
L.stdout "@\n@?";
| Some clname -> L.stdout "Cluster %s@." clname
let process_cluster_cmdline fname =
match Cluster.load_from_file (DB.filename_from_string fname) with
| None ->
L.err "Cannot find cluster file %s@." fname
| Some (nr, cluster) ->
analyze_cluster (nr - 1) cluster
let register_perf_stats_report () =
let stats_dir = Filename.concat Config.results_dir Config.backend_stats_dir_name in
let cluster = match Config.cluster_cmdline with Some cl -> "_" ^ cl | None -> "" in
let stats_base = Config.perf_stats_prefix ^ "_" ^ (Filename.basename cluster) ^ ".json" in
let stats_file = Filename.concat stats_dir stats_base in
DB.create_dir Config.results_dir ;
DB.create_dir stats_dir ;
PerfStats.register_report_at_exit stats_file
let () =
register_perf_stats_report () ;
if Config.developer_mode then
Printexc.record_backtrace true;
print_prolog ();
RegisterCheckers.register ();
if Config.allow_specs_cleanup = true && Config.cluster_cmdline = None then
DB.Results_dir.clean_specs_dir ();
match Config.cluster_cmdline with
| Some fname ->
process_cluster_cmdline fname
| None ->
if Config.merge then MergeCapture.merge_captured_targets ();
let clusters = DB.find_source_dirs () in
L.err "Found %d source files in %s@."
(IList.length clusters) Config.results_dir;
if Config.makefile_cmdline <> ""
then
ClusterMakefile.create_cluster_makefile clusters Config.makefile_cmdline
else
begin
IList.iteri
(fun i cluster -> analyze_cluster i cluster)
clusters;
L.stdout "Analysis finished in %as@." pp_elapsed_time ()
end;
output_json_makefile_stats clusters

File diff suppressed because it is too large Load Diff
Loading…
Cancel
Save