Merge branch 'master' into feature/yul_compilation
This commit is contained in:
commit
b0ca6e1de9
|
@ -48,6 +48,8 @@
|
|||
versions to a vector of name + contract struct tuples
|
||||
[#908](https://github.com/gakonst/ethers-rs/pull/908)
|
||||
- Add Yul compilation [994](https://github.com/gakonst/ethers-rs/pull/994)
|
||||
- Enforce commutativity of ENS reverse resolution
|
||||
[#996](https://github.com/gakonst/ethers-rs/pull/996)
|
||||
|
||||
## ethers-contract-abigen
|
||||
|
||||
|
|
|
@ -2882,9 +2882,9 @@ dependencies = [
|
|||
|
||||
[[package]]
|
||||
name = "regex"
|
||||
version = "1.5.4"
|
||||
version = "1.5.5"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "d07a8629359eb56f1e2fb1652bb04212c072a87ba68546a04065d525673ac461"
|
||||
checksum = "1a11647b6b25ff05a515cb92c365cec08801e83423a235b51e231e1808747286"
|
||||
dependencies = [
|
||||
"aho-corasick",
|
||||
"memchr",
|
||||
|
@ -3574,8 +3574,8 @@ checksum = "6bdef32e8150c2a081110b42772ffe7d7c9032b606bc226c8260fd97e0976601"
|
|||
|
||||
[[package]]
|
||||
name = "svm-rs"
|
||||
version = "0.2.8"
|
||||
source = "git+https://github.com/roynalnaruto/svm-rs#9b3627cfde2763fdc35afa9cbcea8ebc2926938d"
|
||||
version = "0.2.9"
|
||||
source = "git+https://github.com/roynalnaruto/svm-rs#8e33f55fa2a2afb937749e31b2ffa42600bfe216"
|
||||
dependencies = [
|
||||
"anyhow",
|
||||
"cfg-if 1.0.0",
|
||||
|
@ -3876,9 +3876,9 @@ checksum = "360dfd1d6d30e05fda32ace2c8c70e9c0a9da713275777f5a4dbb8a1893930c6"
|
|||
|
||||
[[package]]
|
||||
name = "tracing"
|
||||
version = "0.1.31"
|
||||
version = "0.1.32"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "f6c650a8ef0cd2dd93736f033d21cbd1224c5a967aa0c258d00fcf7dafef9b9f"
|
||||
checksum = "4a1bdf54a7c28a2bbf701e1d2233f6c77f473486b94bee4f9678da5a148dca7f"
|
||||
dependencies = [
|
||||
"cfg-if 1.0.0",
|
||||
"pin-project-lite",
|
||||
|
@ -3888,9 +3888,9 @@ dependencies = [
|
|||
|
||||
[[package]]
|
||||
name = "tracing-attributes"
|
||||
version = "0.1.19"
|
||||
version = "0.1.20"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "8276d9a4a3a558d7b7ad5303ad50b53d58264641b82914b7ada36bd762e7a716"
|
||||
checksum = "2e65ce065b4b5c53e73bb28912318cb8c9e9ad3921f1d669eb0e68b4c8143a2b"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
|
|
|
@ -24,7 +24,7 @@ async-trait = { version = "0.1.50", default-features = false }
|
|||
serde = { version = "1.0.124", default-features = false, features = ["derive"] }
|
||||
thiserror = { version = "1.0.30", default-features = false }
|
||||
futures-util = { version = "^0.3" }
|
||||
tracing = { version = "0.1.31", default-features = false }
|
||||
tracing = { version = "0.1.32", default-features = false }
|
||||
tracing-futures = { version = "0.2.5", default-features = false }
|
||||
|
||||
# for gas oracles
|
||||
|
|
|
@ -35,7 +35,7 @@ futures-channel = { version = "0.3.16", default-features = false }
|
|||
pin-project = { version = "1.0.7", default-features = false }
|
||||
|
||||
# tracing
|
||||
tracing = { version = "0.1.31", default-features = false }
|
||||
tracing = { version = "0.1.32", default-features = false }
|
||||
tracing-futures = { version = "0.2.5", default-features = false, features = ["std-future"] }
|
||||
|
||||
bytes = { version = "1.1.0", default-features = false, optional = true }
|
||||
|
|
|
@ -112,6 +112,10 @@ pub enum ProviderError {
|
|||
#[error("ens name not found: {0}")]
|
||||
EnsError(String),
|
||||
|
||||
/// Invalid reverse ENS name
|
||||
#[error("reverse ens name not pointing to itself: {0}")]
|
||||
EnsNotOwned(String),
|
||||
|
||||
#[error(transparent)]
|
||||
SerdeJson(#[from] serde_json::Error),
|
||||
|
||||
|
@ -794,7 +798,14 @@ impl<P: JsonRpcClient> Middleware for Provider<P> {
|
|||
/// a string. This should theoretically never happen.
|
||||
async fn lookup_address(&self, address: Address) -> Result<String, ProviderError> {
|
||||
let ens_name = ens::reverse_address(address);
|
||||
self.query_resolver(ParamType::String, &ens_name, ens::NAME_SELECTOR).await
|
||||
let domain: String =
|
||||
self.query_resolver(ParamType::String, &ens_name, ens::NAME_SELECTOR).await?;
|
||||
let reverse_address = self.resolve_name(&domain).await?;
|
||||
if address != reverse_address {
|
||||
Err(ProviderError::EnsNotOwned(domain))
|
||||
} else {
|
||||
Ok(domain)
|
||||
}
|
||||
}
|
||||
|
||||
/// Returns the avatar HTTP link of the avatar that the `ens_name` resolves to (or None
|
||||
|
|
|
@ -33,7 +33,7 @@ trezor-client = { version = "0.0.5", optional = true, default-features = false,
|
|||
# aws
|
||||
rusoto_core = { version = "0.47.0", optional = true }
|
||||
rusoto_kms = { version = "0.47.0", optional = true }
|
||||
tracing = { version = "0.1.31", optional = true }
|
||||
tracing = { version = "0.1.32", optional = true }
|
||||
tracing-futures = { version = "0.2.5", optional = true }
|
||||
spki = { version = "0.5.4", optional = true }
|
||||
|
||||
|
|
|
@ -22,13 +22,13 @@ walkdir = "2.3.2"
|
|||
tokio = { version = "1.15.0", default-features = false, features = ["process", "io-util", "fs", "time"], optional = true }
|
||||
futures-util = { version = "^0.3", optional = true }
|
||||
once_cell = "1.10.0"
|
||||
regex = "1.5.4"
|
||||
regex = "1.5.5"
|
||||
md-5 = "0.10.1"
|
||||
thiserror = "1.0.30"
|
||||
hex = "0.4.3"
|
||||
colored = "2.0.0"
|
||||
glob = "0.3.0"
|
||||
tracing = "0.1.31"
|
||||
tracing = "0.1.32"
|
||||
num_cpus = "1.13.1"
|
||||
tiny-keccak = { version = "2.0.2", default-features = false }
|
||||
tempfile = { version = "3.3.0", optional = true }
|
||||
|
|
|
@ -581,7 +581,7 @@ impl<'a, T: ArtifactOutput> ArtifactsCacheInner<'a, T> {
|
|||
}
|
||||
}
|
||||
|
||||
/// inserts the filtered source with the fiven version
|
||||
/// inserts the filtered source with the given version
|
||||
fn insert_filtered_source(&mut self, file: PathBuf, source: Source, version: Version) {
|
||||
match self.filtered.entry(file) {
|
||||
hash_map::Entry::Occupied(mut entry) => {
|
||||
|
@ -593,35 +593,62 @@ impl<'a, T: ArtifactOutput> ArtifactsCacheInner<'a, T> {
|
|||
}
|
||||
}
|
||||
|
||||
/// Returns only those sources that
|
||||
/// Returns only dirty sources that:
|
||||
/// - are new
|
||||
/// - were changed
|
||||
/// - their imports were changed
|
||||
/// - their artifact is missing
|
||||
/// This also includes their respective imports
|
||||
fn filter(&mut self, sources: Sources, version: &Version) -> Sources {
|
||||
self.fill_hashes(&sources);
|
||||
sources
|
||||
|
||||
let mut imports_of_dirty = HashSet::new();
|
||||
// separates all source files that fit the criteria (dirty) from those that don't (clean)
|
||||
let (mut dirty_sources, clean_sources) = sources
|
||||
.into_iter()
|
||||
.filter_map(|(file, source)| self.requires_solc(file, source, version))
|
||||
.collect()
|
||||
.map(|(file, source)| self.filter_source(file, source, version))
|
||||
.fold(
|
||||
(Sources::default(), Vec::new()),
|
||||
|(mut dirty_sources, mut clean_sources), source| {
|
||||
if source.dirty {
|
||||
// mark all files that are imported by a dirty file
|
||||
imports_of_dirty.extend(self.edges.all_imported_nodes(source.idx));
|
||||
dirty_sources.insert(source.file, source.source);
|
||||
} else {
|
||||
clean_sources.push(source);
|
||||
}
|
||||
|
||||
(dirty_sources, clean_sources)
|
||||
},
|
||||
);
|
||||
|
||||
for clean_source in clean_sources {
|
||||
let FilteredSource { file, source, idx, .. } = clean_source;
|
||||
if imports_of_dirty.contains(&idx) {
|
||||
// file is imported by a dirty file
|
||||
dirty_sources.insert(file, source);
|
||||
} else {
|
||||
self.insert_filtered_source(file, source, version.clone());
|
||||
}
|
||||
}
|
||||
|
||||
// track dirty sources internally
|
||||
for (file, source) in dirty_sources.iter() {
|
||||
self.insert_new_cache_entry(file, source, version.clone());
|
||||
}
|
||||
|
||||
dirty_sources
|
||||
}
|
||||
|
||||
/// Returns `Some` if the file _needs_ to be compiled and `None` if the artifact can be reu-used
|
||||
fn requires_solc(
|
||||
&mut self,
|
||||
file: PathBuf,
|
||||
source: Source,
|
||||
version: &Version,
|
||||
) -> Option<(PathBuf, Source)> {
|
||||
/// Returns the state of the given source file.
|
||||
fn filter_source(&self, file: PathBuf, source: Source, version: &Version) -> FilteredSource {
|
||||
let idx = self.edges.node_id(&file);
|
||||
if !self.is_dirty(&file, version) &&
|
||||
self.edges.imports(&file).iter().all(|file| !self.is_dirty(file, version))
|
||||
{
|
||||
self.insert_filtered_source(file, source, version.clone());
|
||||
None
|
||||
FilteredSource { file, source, idx, dirty: false }
|
||||
} else {
|
||||
self.insert_new_cache_entry(&file, &source, version.clone());
|
||||
|
||||
Some((file, source))
|
||||
FilteredSource { file, source, idx, dirty: true }
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -674,6 +701,14 @@ impl<'a, T: ArtifactOutput> ArtifactsCacheInner<'a, T> {
|
|||
}
|
||||
}
|
||||
|
||||
/// Helper type to represent the state of a source file
|
||||
struct FilteredSource {
|
||||
file: PathBuf,
|
||||
source: Source,
|
||||
idx: usize,
|
||||
dirty: bool,
|
||||
}
|
||||
|
||||
/// Abstraction over configured caching which can be either non-existent or an already loaded cache
|
||||
#[allow(clippy::large_enum_variant)]
|
||||
#[derive(Debug)]
|
||||
|
|
|
@ -179,7 +179,7 @@ impl<'a, T: ArtifactOutput> ProjectCompiler<'a, T> {
|
|||
let Self { edges, project, mut sources } = self;
|
||||
|
||||
let mut cache = ArtifactsCache::new(project, edges)?;
|
||||
// retain and compile only dirty sources
|
||||
// retain and compile only dirty sources and all their imports
|
||||
sources = sources.filtered(&mut cache);
|
||||
|
||||
Ok(PreprocessedState { sources, cache })
|
||||
|
@ -216,18 +216,22 @@ struct CompiledState<'a, T: ArtifactOutput> {
|
|||
impl<'a, T: ArtifactOutput> CompiledState<'a, T> {
|
||||
/// advance to the next state by handling all artifacts
|
||||
///
|
||||
/// Writes all output contracts to disk if enabled in the `Project`
|
||||
/// Writes all output contracts to disk if enabled in the `Project` and if the build was
|
||||
/// successful
|
||||
fn write_artifacts(self) -> Result<ArtifactsState<'a, T>> {
|
||||
let CompiledState { output, cache } = self;
|
||||
|
||||
// write all artifacts
|
||||
let compiled_artifacts = if !cache.project().no_artifacts {
|
||||
// write all artifacts via the handler but only if the build succeeded
|
||||
let compiled_artifacts = if cache.project().no_artifacts {
|
||||
cache.project().artifacts_handler().output_to_artifacts(&output.contracts)
|
||||
} else if output.has_error() {
|
||||
tracing::trace!("skip writing cache file due to solc errors: {:?}", output.errors);
|
||||
cache.project().artifacts_handler().output_to_artifacts(&output.contracts)
|
||||
} else {
|
||||
cache
|
||||
.project()
|
||||
.artifacts_handler()
|
||||
.on_output(&output.contracts, &cache.project().paths)?
|
||||
} else {
|
||||
cache.project().artifacts_handler().output_to_artifacts(&output.contracts)
|
||||
};
|
||||
|
||||
Ok(ArtifactsState { output, cache, compiled_artifacts })
|
||||
|
|
|
@ -10,6 +10,7 @@ use crate::{
|
|||
use crate::artifacts::output_selection::ContractOutputSelection;
|
||||
use serde::{Deserialize, Serialize};
|
||||
use std::{
|
||||
collections::HashSet,
|
||||
fmt::{self, Formatter},
|
||||
fs,
|
||||
path::{Component, Path, PathBuf},
|
||||
|
@ -229,7 +230,7 @@ impl ProjectPathsConfig {
|
|||
pub fn flatten(&self, target: &Path) -> Result<String> {
|
||||
tracing::trace!("flattening file");
|
||||
let graph = Graph::resolve(self)?;
|
||||
self.flatten_node(target, &graph, &mut vec![], false, false)
|
||||
self.flatten_node(target, &graph, &mut Default::default(), false, false)
|
||||
}
|
||||
|
||||
/// Flattens a single node from the dependency graph
|
||||
|
@ -237,7 +238,7 @@ impl ProjectPathsConfig {
|
|||
&self,
|
||||
target: &Path,
|
||||
graph: &Graph,
|
||||
imported: &mut Vec<usize>,
|
||||
imported: &mut HashSet<usize>,
|
||||
strip_version_pragma: bool,
|
||||
strip_license: bool,
|
||||
) -> Result<String> {
|
||||
|
@ -248,9 +249,11 @@ impl ProjectPathsConfig {
|
|||
SolcError::msg(format!("cannot resolve file at \"{:?}\"", target.display()))
|
||||
})?;
|
||||
|
||||
if imported.iter().any(|&idx| idx == *target_index) {
|
||||
if imported.contains(target_index) {
|
||||
// short circuit nodes that were already imported, if both A.sol and B.sol import C.sol
|
||||
return Ok(String::new())
|
||||
}
|
||||
imported.insert(*target_index);
|
||||
|
||||
let target_node = graph.node(*target_index);
|
||||
|
||||
|
@ -278,21 +281,17 @@ impl ProjectPathsConfig {
|
|||
|
||||
for import in imports.iter() {
|
||||
let import_path = self.resolve_import(target_dir, import.data())?;
|
||||
let import_content = self
|
||||
.flatten_node(&import_path, graph, imported, true, true)?
|
||||
.trim()
|
||||
.as_bytes()
|
||||
.to_owned();
|
||||
let s = self.flatten_node(&import_path, graph, imported, true, true)?;
|
||||
let import_content = s.trim().as_bytes();
|
||||
let import_content_len = import_content.len() as isize;
|
||||
let (start, end) = import.loc_by_offset(offset);
|
||||
content.splice(start..end, import_content);
|
||||
content.splice(start..end, import_content.iter().copied());
|
||||
offset += import_content_len - ((end - start) as isize);
|
||||
}
|
||||
|
||||
let result = String::from_utf8(content).map_err(|err| {
|
||||
SolcError::msg(format!("failed to convert extended bytes to string: {}", err))
|
||||
})?;
|
||||
imported.push(*target_index);
|
||||
|
||||
Ok(result)
|
||||
}
|
||||
|
|
|
@ -125,9 +125,11 @@ impl<T: ArtifactOutput> Project<T> {
|
|||
&self.artifacts
|
||||
}
|
||||
|
||||
/// Applies the configured settings to the given `Solc`
|
||||
/// Applies the configured arguments to the given `Solc`
|
||||
///
|
||||
/// This will set the `--allow-paths` to the paths configured for the `Project`, if any.
|
||||
fn configure_solc(&self, mut solc: Solc) -> Solc {
|
||||
if self.allowed_lib_paths.0.is_empty() {
|
||||
if solc.args.is_empty() && !self.allowed_lib_paths.0.is_empty() {
|
||||
solc = solc.arg("--allow-paths").arg(self.allowed_lib_paths.to_string());
|
||||
}
|
||||
solc
|
||||
|
|
|
@ -49,6 +49,13 @@ pub struct Remapping {
|
|||
pub path: String,
|
||||
}
|
||||
|
||||
impl Remapping {
|
||||
/// Convenience function for [`RelativeRemapping::new`]
|
||||
pub fn into_relative(self, root: impl AsRef<Path>) -> RelativeRemapping {
|
||||
RelativeRemapping::new(self, root)
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(thiserror::Error, Debug, PartialEq, PartialOrd)]
|
||||
pub enum RemappingError {
|
||||
#[error("no prefix found")]
|
||||
|
@ -222,6 +229,12 @@ impl RelativeRemapping {
|
|||
self.path.parent = Some(root);
|
||||
self.into()
|
||||
}
|
||||
|
||||
/// Converts this relative remapping into [`Remapping`] without the root path
|
||||
pub fn to_relative_remapping(mut self) -> Remapping {
|
||||
self.path.parent.take();
|
||||
self.into()
|
||||
}
|
||||
}
|
||||
|
||||
// Remappings are printed as `prefix=target`
|
||||
|
@ -263,8 +276,8 @@ impl From<Remapping> for RelativeRemapping {
|
|||
/// resolve as a `weird-erc20/=/var/lib/weird-erc20/src/` remapping.
|
||||
#[derive(Clone, Debug, PartialEq, PartialOrd, Eq, Ord)]
|
||||
pub struct RelativeRemappingPathBuf {
|
||||
parent: Option<PathBuf>,
|
||||
path: PathBuf,
|
||||
pub parent: Option<PathBuf>,
|
||||
pub path: PathBuf,
|
||||
}
|
||||
|
||||
impl RelativeRemappingPathBuf {
|
||||
|
@ -647,6 +660,9 @@ mod tests {
|
|||
assert_eq!(relative.path.relative(), Path::new(&remapping.path));
|
||||
assert_eq!(relative.path.original(), Path::new(&remapping.path));
|
||||
assert!(relative.path.parent.is_none());
|
||||
|
||||
let relative = RelativeRemapping::new(remapping.clone(), "/a/b");
|
||||
assert_eq!(relative.to_relative_remapping(), Remapping::from_str("oz/=c/d/").unwrap());
|
||||
}
|
||||
|
||||
#[test]
|
||||
|
|
|
@ -1,17 +1,52 @@
|
|||
//! Subscribe to events in the compiler pipeline
|
||||
//!
|
||||
//! The _reporter_ is the component of the [`Project::compile()`] pipeline which is responsible
|
||||
//! for reporting on specific steps in the process.
|
||||
//!
|
||||
//! By default, the current reporter is a noop that does
|
||||
//! nothing.
|
||||
//!
|
||||
//! To use another report implementation, it must be set as the current reporter.
|
||||
//! There are two methods for doing so: [`with_scoped`] and
|
||||
//! [`set_global`]. `with_scoped` sets the reporter for the
|
||||
//! duration of a scope, while `set_global` sets a global default report
|
||||
//! for the entire process.
|
||||
|
||||
// https://github.com/tokio-rs/tracing/blob/master/tracing-core/src/dispatch.rs
|
||||
|
||||
use crate::{CompilerInput, CompilerOutput, Solc};
|
||||
use semver::Version;
|
||||
use std::{
|
||||
any::{Any, TypeId},
|
||||
cell::RefCell,
|
||||
error::Error,
|
||||
fmt,
|
||||
path::Path,
|
||||
ptr::NonNull,
|
||||
sync::{
|
||||
atomic::{AtomicUsize, Ordering},
|
||||
atomic::{AtomicBool, AtomicUsize, Ordering},
|
||||
Arc,
|
||||
},
|
||||
};
|
||||
|
||||
thread_local! {
|
||||
static CURRENT_STATE: State = State {
|
||||
scoped: RefCell::new(Report::none()),
|
||||
};
|
||||
}
|
||||
|
||||
static EXISTS: AtomicBool = AtomicBool::new(false);
|
||||
static SCOPED_COUNT: AtomicUsize = AtomicUsize::new(0);
|
||||
|
||||
// tracks the state of `GLOBAL_REPORTER`
|
||||
static GLOBAL_REPORTER_STATE: AtomicUsize = AtomicUsize::new(UN_SET);
|
||||
|
||||
const UN_SET: usize = 0;
|
||||
const SETTING: usize = 1;
|
||||
const SET: usize = 2;
|
||||
|
||||
static mut GLOBAL_REPORTER: Option<Report> = None;
|
||||
|
||||
/// Install this `Reporter` as the global default if one is
|
||||
/// not already set.
|
||||
///
|
||||
|
@ -69,28 +104,70 @@ pub trait Reporter: 'static {
|
|||
|
||||
/// Invoked if the import couldn't be resolved
|
||||
fn on_unresolved_import(&self, _import: &Path) {}
|
||||
|
||||
/// If `self` is the same type as the provided `TypeId`, returns an untyped
|
||||
/// [`NonNull`] pointer to that type. Otherwise, returns `None`.
|
||||
///
|
||||
/// If you wish to downcast a `Reporter`, it is strongly advised to use
|
||||
/// the safe API provided by [`downcast_ref`] instead.
|
||||
///
|
||||
/// This API is required for `downcast_raw` to be a trait method; a method
|
||||
/// signature like [`downcast_ref`] (with a generic type parameter) is not
|
||||
/// object-safe, and thus cannot be a trait method for `Reporter`. This
|
||||
/// means that if we only exposed `downcast_ref`, `Reporter`
|
||||
/// implementations could not override the downcasting behavior
|
||||
///
|
||||
/// # Safety
|
||||
///
|
||||
/// The [`downcast_ref`] method expects that the pointer returned by
|
||||
/// `downcast_raw` points to a valid instance of the type
|
||||
/// with the provided `TypeId`. Failure to ensure this will result in
|
||||
/// undefined behaviour, so implementing `downcast_raw` is unsafe.
|
||||
unsafe fn downcast_raw(&self, id: TypeId) -> Option<NonNull<()>> {
|
||||
if id == TypeId::of::<Self>() {
|
||||
Some(NonNull::from(self).cast())
|
||||
} else {
|
||||
None
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl dyn Reporter {
|
||||
/// Returns `true` if this `Reporter` is the same type as `T`.
|
||||
pub fn is<T: Any>(&self) -> bool {
|
||||
self.downcast_ref::<T>().is_some()
|
||||
}
|
||||
|
||||
/// Returns some reference to this `Reporter` value if it is of type `T`,
|
||||
/// or `None` if it isn't.
|
||||
pub fn downcast_ref<T: Any>(&self) -> Option<&T> {
|
||||
unsafe {
|
||||
let raw = self.downcast_raw(TypeId::of::<T>())?;
|
||||
Some(&*(raw.cast().as_ptr()))
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) fn solc_spawn(solc: &Solc, version: &Version, input: &CompilerInput) {
|
||||
with_global(|r| r.reporter.on_solc_spawn(solc, version, input));
|
||||
get_default(|r| r.reporter.on_solc_spawn(solc, version, input));
|
||||
}
|
||||
|
||||
pub(crate) fn solc_success(solc: &Solc, version: &Version, output: &CompilerOutput) {
|
||||
with_global(|r| r.reporter.on_solc_success(solc, version, output));
|
||||
get_default(|r| r.reporter.on_solc_success(solc, version, output));
|
||||
}
|
||||
|
||||
#[allow(unused)]
|
||||
pub(crate) fn solc_installation_start(version: &Version) {
|
||||
with_global(|r| r.reporter.on_solc_installation_start(version));
|
||||
get_default(|r| r.reporter.on_solc_installation_start(version));
|
||||
}
|
||||
|
||||
#[allow(unused)]
|
||||
pub(crate) fn solc_installation_success(version: &Version) {
|
||||
with_global(|r| r.reporter.on_solc_installation_success(version));
|
||||
get_default(|r| r.reporter.on_solc_installation_success(version));
|
||||
}
|
||||
|
||||
pub(crate) fn unresolved_import(import: &Path) {
|
||||
with_global(|r| r.reporter.on_unresolved_import(import));
|
||||
get_default(|r| r.reporter.on_unresolved_import(import));
|
||||
}
|
||||
|
||||
fn get_global() -> Option<&'static Report> {
|
||||
|
@ -106,10 +183,97 @@ fn get_global() -> Option<&'static Report> {
|
|||
}
|
||||
}
|
||||
|
||||
/// Executes a closure with a reference to this thread's current [reporter].
|
||||
#[inline(always)]
|
||||
pub fn get_default<T, F>(mut f: F) -> T
|
||||
where
|
||||
F: FnMut(&Report) -> T,
|
||||
{
|
||||
if SCOPED_COUNT.load(Ordering::Acquire) == 0 {
|
||||
// fast path if no scoped reporter has been set; use the global
|
||||
// default.
|
||||
return if let Some(glob) = get_global() { f(glob) } else { f(&Report::none()) }
|
||||
}
|
||||
|
||||
get_default_scoped(f)
|
||||
}
|
||||
|
||||
#[inline(never)]
|
||||
fn get_default_scoped<T, F>(mut f: F) -> T
|
||||
where
|
||||
F: FnMut(&Report) -> T,
|
||||
{
|
||||
CURRENT_STATE
|
||||
.try_with(|state| {
|
||||
let scoped = state.scoped.borrow_mut();
|
||||
f(&*scoped)
|
||||
})
|
||||
.unwrap_or_else(|_| f(&Report::none()))
|
||||
}
|
||||
|
||||
/// Executes a closure with a reference to the `Reporter`.
|
||||
pub fn with_global<T>(f: impl FnOnce(&Report) -> T) -> Option<T> {
|
||||
let dispatch = get_global()?;
|
||||
Some(f(dispatch))
|
||||
let report = get_global()?;
|
||||
Some(f(report))
|
||||
}
|
||||
|
||||
/// Sets this reporter as the scoped reporter for the duration of a closure.
|
||||
pub fn with_scoped<T>(report: &Report, f: impl FnOnce() -> T) -> T {
|
||||
// When this guard is dropped, the scoped reporter will be reset to the
|
||||
// prior reporter. Using this (rather than simply resetting after calling
|
||||
// `f`) ensures that we always reset to the prior reporter even if `f`
|
||||
// panics.
|
||||
let _guard = set_scoped(report);
|
||||
f()
|
||||
}
|
||||
|
||||
/// The report state of a thread.
|
||||
struct State {
|
||||
/// This thread's current scoped reporter.
|
||||
scoped: RefCell<Report>,
|
||||
}
|
||||
|
||||
impl State {
|
||||
/// Replaces the current scoped reporter on this thread with the provided
|
||||
/// reporter.
|
||||
///
|
||||
/// Dropping the returned `ResetGuard` will reset the scoped reporter to
|
||||
/// the previous value.
|
||||
#[inline]
|
||||
fn set_scoped(new_report: Report) -> ScopeGuard {
|
||||
let prior = CURRENT_STATE.try_with(|state| state.scoped.replace(new_report)).ok();
|
||||
EXISTS.store(true, Ordering::Release);
|
||||
SCOPED_COUNT.fetch_add(1, Ordering::Release);
|
||||
ScopeGuard(prior)
|
||||
}
|
||||
}
|
||||
|
||||
/// A guard that resets the current scoped reporter to the prior
|
||||
/// scoped reporter when dropped.
|
||||
#[derive(Debug)]
|
||||
pub struct ScopeGuard(Option<Report>);
|
||||
|
||||
impl Drop for ScopeGuard {
|
||||
#[inline]
|
||||
fn drop(&mut self) {
|
||||
SCOPED_COUNT.fetch_sub(1, Ordering::Release);
|
||||
if let Some(report) = self.0.take() {
|
||||
// Replace the reporter and then drop the old one outside
|
||||
// of the thread-local context.
|
||||
let prev = CURRENT_STATE.try_with(|state| state.scoped.replace(report));
|
||||
drop(prev)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Sets the reporter as the scoped reporter for the duration of the lifetime
|
||||
/// of the returned DefaultGuard
|
||||
#[must_use = "Dropping the guard unregisters the reporter."]
|
||||
pub fn set_scoped(reporter: &Report) -> ScopeGuard {
|
||||
// When this guard is dropped, the scoped reporter will be reset to the
|
||||
// prior default. Using this ensures that we always reset to the prior
|
||||
// reporter even if the thread calling this function panics.
|
||||
State::set_scoped(reporter.clone())
|
||||
}
|
||||
|
||||
/// A no-op [`Reporter`] that does nothing.
|
||||
|
@ -165,6 +329,7 @@ impl fmt::Display for SetGlobalReporterError {
|
|||
impl Error for SetGlobalReporterError {}
|
||||
|
||||
/// `Report` trace data to a [`Reporter`].
|
||||
#[derive(Clone)]
|
||||
pub struct Report {
|
||||
reporter: Arc<dyn Reporter + Send + Sync>,
|
||||
}
|
||||
|
@ -184,16 +349,20 @@ impl Report {
|
|||
{
|
||||
Self { reporter: Arc::new(reporter) }
|
||||
}
|
||||
|
||||
/// Returns `true` if this `Report` forwards to a reporter of type
|
||||
/// `T`.
|
||||
#[inline]
|
||||
pub fn is<T: Any>(&self) -> bool {
|
||||
<dyn Reporter>::is::<T>(&*self.reporter)
|
||||
}
|
||||
}
|
||||
|
||||
// tracks the state of `GLOBAL_REPORTER`
|
||||
static GLOBAL_REPORTER_STATE: AtomicUsize = AtomicUsize::new(UN_SET);
|
||||
|
||||
const UN_SET: usize = 0;
|
||||
const SETTING: usize = 1;
|
||||
const SET: usize = 2;
|
||||
|
||||
static mut GLOBAL_REPORTER: Option<Report> = None;
|
||||
impl fmt::Debug for Report {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.pad("Report(...)")
|
||||
}
|
||||
}
|
||||
|
||||
/// Sets this report as the global default for the duration of the entire program.
|
||||
///
|
||||
|
@ -216,3 +385,35 @@ fn set_global_reporter(report: Report) -> Result<(), SetGlobalReporterError> {
|
|||
Err(SetGlobalReporterError { _priv: () })
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn scoped_reporter_works() {
|
||||
struct TestReporter;
|
||||
impl Reporter for TestReporter {}
|
||||
|
||||
with_scoped(&Report::new(TestReporter), || {
|
||||
get_default(|reporter| assert!(reporter.is::<TestReporter>()))
|
||||
});
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn global_and_scoped_reporter_works() {
|
||||
get_default(|reporter| {
|
||||
assert!(reporter.is::<NoReporter>());
|
||||
});
|
||||
|
||||
set_global_reporter(Report::new(BasicStdoutReporter::default())).unwrap();
|
||||
struct TestReporter;
|
||||
impl Reporter for TestReporter {}
|
||||
|
||||
with_scoped(&Report::new(TestReporter), || {
|
||||
get_default(|reporter| assert!(reporter.is::<TestReporter>()))
|
||||
});
|
||||
|
||||
get_default(|reporter| assert!(reporter.is::<BasicStdoutReporter>()))
|
||||
}
|
||||
}
|
||||
|
|
|
@ -91,6 +91,11 @@ impl GraphEdges {
|
|||
&self.edges[from]
|
||||
}
|
||||
|
||||
/// Returns an iterator that yields all imports of a node and all their imports
|
||||
pub fn all_imported_nodes(&self, from: usize) -> impl Iterator<Item = usize> + '_ {
|
||||
NodesIter::new(from, self).skip(1)
|
||||
}
|
||||
|
||||
/// Returns all files imported by the given file
|
||||
pub fn imports(&self, file: impl AsRef<Path>) -> HashSet<&PathBuf> {
|
||||
if let Some(start) = self.indices.get(file.as_ref()).copied() {
|
||||
|
@ -100,6 +105,11 @@ impl GraphEdges {
|
|||
}
|
||||
}
|
||||
|
||||
/// Returns the id of the given file
|
||||
pub fn node_id(&self, file: impl AsRef<Path>) -> usize {
|
||||
self.indices[file.as_ref()]
|
||||
}
|
||||
|
||||
/// Returns true if the `file` was originally included when the graph was first created and not
|
||||
/// added when all `imports` were resolved
|
||||
pub fn is_input_file(&self, file: impl AsRef<Path>) -> bool {
|
||||
|
@ -631,11 +641,18 @@ impl VersionedSources {
|
|||
SolcError::msg(format!("solc \"{}\" should have been installed", version))
|
||||
})?;
|
||||
|
||||
tracing::trace!("verifying solc checksum for {}", solc.solc.display());
|
||||
if solc.verify_checksum().is_err() {
|
||||
tracing::trace!("corrupted solc version, redownloading \"{}\"", version);
|
||||
Solc::blocking_install(version.as_ref())?;
|
||||
tracing::trace!("reinstalled solc: \"{}\"", version);
|
||||
if self.offline {
|
||||
tracing::trace!(
|
||||
"skip verifying solc checksum for {} in offline mode",
|
||||
solc.solc.display()
|
||||
);
|
||||
} else {
|
||||
tracing::trace!("verifying solc checksum for {}", solc.solc.display());
|
||||
if solc.verify_checksum().is_err() {
|
||||
tracing::trace!("corrupted solc version, redownloading \"{}\"", version);
|
||||
Solc::blocking_install(version.as_ref())?;
|
||||
tracing::trace!("reinstalled solc: \"{}\"", version);
|
||||
}
|
||||
}
|
||||
let solc = solc.arg("--allow-paths").arg(allowed_lib_paths.to_string());
|
||||
let version = solc.version()?;
|
||||
|
|
|
@ -157,11 +157,6 @@ fn print_imports(
|
|||
return Ok(())
|
||||
}
|
||||
|
||||
for continues in &**levels_continue {
|
||||
let c = if *continues { symbols.down } else { " " };
|
||||
write!(out, "{} ", c)?;
|
||||
}
|
||||
|
||||
let mut iter = imports.iter().peekable();
|
||||
|
||||
while let Some(import) = iter.next() {
|
||||
|
|
|
@ -447,6 +447,57 @@ fn can_flatten_file_in_dapp_sample() {
|
|||
assert!(result.contains("contract DappTest"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn can_flatten_unique() {
|
||||
let project = TempProject::dapptools().unwrap();
|
||||
|
||||
let f = project
|
||||
.add_source(
|
||||
"A",
|
||||
r#"
|
||||
pragma solidity ^0.8.10;
|
||||
import "./C.sol";
|
||||
import "./B.sol";
|
||||
contract A { }
|
||||
"#,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
project
|
||||
.add_source(
|
||||
"B",
|
||||
r#"
|
||||
pragma solidity ^0.8.10;
|
||||
import "./C.sol";
|
||||
contract B { }
|
||||
"#,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
project
|
||||
.add_source(
|
||||
"C",
|
||||
r#"
|
||||
pragma solidity ^0.8.10;
|
||||
import "./A.sol";
|
||||
contract C { }
|
||||
"#,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
let result = project.flatten(&f).unwrap();
|
||||
|
||||
assert_eq!(
|
||||
result,
|
||||
r#"
|
||||
pragma solidity ^0.8.10;
|
||||
contract C { }
|
||||
contract B { }
|
||||
contract A { }
|
||||
"#
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn can_flatten_file_with_duplicates() {
|
||||
let root = PathBuf::from(env!("CARGO_MANIFEST_DIR")).join("test-data/test-flatten-duplicates");
|
||||
|
@ -592,3 +643,43 @@ contract LinkTest {
|
|||
let s = serde_json::to_string(&bytecode).unwrap();
|
||||
assert_eq!(bytecode.clone(), serde_json::from_str(&s).unwrap());
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn can_recompile_with_changes() {
|
||||
let mut tmp = TempProject::dapptools().unwrap();
|
||||
tmp.project_mut().allowed_lib_paths = vec![tmp.root().join("modules")].into();
|
||||
|
||||
let content = r#"
|
||||
pragma solidity ^0.8.10;
|
||||
import "../modules/B.sol";
|
||||
contract A {}
|
||||
"#;
|
||||
tmp.add_source("A", content).unwrap();
|
||||
|
||||
tmp.add_contract(
|
||||
"modules/B",
|
||||
r#"
|
||||
pragma solidity ^0.8.10;
|
||||
contract B {}
|
||||
"#,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
let compiled = tmp.compile().unwrap();
|
||||
assert!(!compiled.has_compiler_errors());
|
||||
assert!(compiled.find("A").is_some());
|
||||
assert!(compiled.find("B").is_some());
|
||||
|
||||
let compiled = tmp.compile().unwrap();
|
||||
assert!(compiled.find("A").is_some());
|
||||
assert!(compiled.find("B").is_some());
|
||||
assert!(compiled.is_unchanged());
|
||||
|
||||
// modify A.sol
|
||||
tmp.add_source("A", format!("{}\n", content)).unwrap();
|
||||
let compiled = tmp.compile().unwrap();
|
||||
assert!(!compiled.has_compiler_errors());
|
||||
assert!(!compiled.is_unchanged());
|
||||
assert!(compiled.find("A").is_some());
|
||||
assert!(compiled.find("B").is_some());
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue