//! Code to save/load the dep-graph from files. use rustc_data_structures::fx::FxHashMap; use rustc::dep_graph::{PreviousDepGraph, SerializedDepGraph, WorkProduct, WorkProductId}; use rustc::session::Session; use rustc::ty::TyCtxt; use rustc::ty::query::OnDiskCache; use rustc::util::common::time_ext; use rustc_serialize::Decodable as RustcDecodable; use rustc_serialize::opaque::Decoder; use std::path::Path; use super::data::*; use super::fs::*; use super::file_format; use super::work_product; pub fn dep_graph_tcx_init<'a, 'tcx>(tcx: TyCtxt<'a, 'tcx, 'tcx>) { if !tcx.dep_graph.is_fully_enabled() { return } tcx.allocate_metadata_dep_nodes(); } type WorkProductMap = FxHashMap; pub enum LoadResult { Ok { data: T }, DataOutOfDate, Error { message: String }, } impl LoadResult<(PreviousDepGraph, WorkProductMap)> { pub fn open(self, sess: &Session) -> (PreviousDepGraph, WorkProductMap) { match self { LoadResult::Error { message } => { sess.warn(&message); Default::default() }, LoadResult::DataOutOfDate => { if let Err(err) = delete_all_session_dir_contents(sess) { sess.err(&format!("Failed to delete invalidated or incompatible \ incremental compilation session directory contents `{}`: {}.", dep_graph_path(sess).display(), err)); } Default::default() } LoadResult::Ok { data } => data } } } fn load_data(report_incremental_info: bool, path: &Path) -> LoadResult<(Vec, usize)> { match file_format::read_file(report_incremental_info, path) { Ok(Some(data_and_pos)) => LoadResult::Ok { data: data_and_pos }, Ok(None) => { // The file either didn't exist or was produced by an incompatible // compiler version. Neither is an error. LoadResult::DataOutOfDate } Err(err) => { LoadResult::Error { message: format!("could not load dep-graph from `{}`: {}", path.display(), err) } } } } fn delete_dirty_work_product(sess: &Session, swp: SerializedWorkProduct) { debug!("delete_dirty_work_product({:?})", swp); work_product::delete_workproduct_files(sess, &swp.work_product); } /// Either a result that has already be computed or a /// handle that will let us wait until it is computed /// by a background thread. pub enum MaybeAsync { Sync(T), Async(std::thread::JoinHandle) } impl MaybeAsync { pub fn open(self) -> std::thread::Result { match self { MaybeAsync::Sync(result) => Ok(result), MaybeAsync::Async(handle) => handle.join() } } } pub type DepGraphFuture = MaybeAsync>; /// Launch a thread and load the dependency graph in the background. pub fn load_dep_graph(sess: &Session) -> DepGraphFuture { // Since `sess` isn't `Sync`, we perform all accesses to `sess` // before we fire the background thread. let time_passes = sess.time_passes(); if sess.opts.incremental.is_none() { // No incremental compilation. return MaybeAsync::Sync(LoadResult::Ok { data: Default::default(), }); } // Calling `sess.incr_comp_session_dir()` will panic if `sess.opts.incremental.is_none()`. // Fortunately, we just checked that this isn't the case. let path = dep_graph_path_from(&sess.incr_comp_session_dir()); let report_incremental_info = sess.opts.debugging_opts.incremental_info; let expected_hash = sess.opts.dep_tracking_hash(); let mut prev_work_products = FxHashMap::default(); // If we are only building with -Zquery-dep-graph but without an actual // incr. comp. session directory, we skip this. Otherwise we'd fail // when trying to load work products. if sess.incr_comp_session_dir_opt().is_some() { let work_products_path = work_products_path(sess); let load_result = load_data(report_incremental_info, &work_products_path); if let LoadResult::Ok { data: (work_products_data, start_pos) } = load_result { // Decode the list of work_products let mut work_product_decoder = Decoder::new(&work_products_data[..], start_pos); let work_products: Vec = RustcDecodable::decode(&mut work_product_decoder).unwrap_or_else(|e| { let msg = format!("Error decoding `work-products` from incremental \ compilation session directory: {}", e); sess.fatal(&msg[..]) }); for swp in work_products { let mut all_files_exist = true; for &(_, ref file_name) in swp.work_product.saved_files.iter() { let path = in_incr_comp_dir_sess(sess, file_name); if !path.exists() { all_files_exist = false; if sess.opts.debugging_opts.incremental_info { eprintln!("incremental: could not find file for work \ product: {}", path.display()); } } } if all_files_exist { debug!("reconcile_work_products: all files for {:?} exist", swp); prev_work_products.insert(swp.id, swp.work_product); } else { debug!("reconcile_work_products: some file for {:?} does not exist", swp); delete_dirty_work_product(sess, swp); } } } } MaybeAsync::Async(std::thread::spawn(move || { time_ext(time_passes, None, "background load prev dep-graph", move || { match load_data(report_incremental_info, &path) { LoadResult::DataOutOfDate => LoadResult::DataOutOfDate, LoadResult::Error { message } => LoadResult::Error { message }, LoadResult::Ok { data: (bytes, start_pos) } => { let mut decoder = Decoder::new(&bytes, start_pos); let prev_commandline_args_hash = u64::decode(&mut decoder) .expect("Error reading commandline arg hash from cached dep-graph"); if prev_commandline_args_hash != expected_hash { if report_incremental_info { println!("[incremental] completely ignoring cache because of \ differing commandline arguments"); } // We can't reuse the cache, purge it. debug!("load_dep_graph_new: differing commandline arg hashes"); // No need to do any further work return LoadResult::DataOutOfDate; } let dep_graph = SerializedDepGraph::decode(&mut decoder) .expect("Error reading cached dep-graph"); LoadResult::Ok { data: (PreviousDepGraph::new(dep_graph), prev_work_products) } } } }) })) } pub fn load_query_result_cache<'sess>(sess: &'sess Session) -> OnDiskCache<'sess> { if sess.opts.incremental.is_none() || !sess.opts.debugging_opts.incremental_queries { return OnDiskCache::new_empty(sess.source_map()); } match load_data(sess.opts.debugging_opts.incremental_info, &query_cache_path(sess)) { LoadResult::Ok{ data: (bytes, start_pos) } => OnDiskCache::new(sess, bytes, start_pos), _ => OnDiskCache::new_empty(sess.source_map()) } }