705 lines
23 KiB
Rust
705 lines
23 KiB
Rust
#![doc = include_str!("../README.md")]
|
|
|
|
pub mod artifacts;
|
|
|
|
pub use artifacts::{CompilerInput, CompilerOutput, EvmVersion};
|
|
use std::collections::btree_map::Entry;
|
|
|
|
pub mod cache;
|
|
|
|
mod compile;
|
|
pub use compile::*;
|
|
|
|
mod config;
|
|
pub use config::{
|
|
AllowedLibPaths, Artifact, ArtifactOutput, MinimalCombinedArtifacts, ProjectPathsConfig,
|
|
SolcConfig,
|
|
};
|
|
|
|
pub mod remappings;
|
|
|
|
use crate::{artifacts::Source, cache::SolFilesCache};
|
|
|
|
pub mod error;
|
|
pub mod utils;
|
|
use crate::artifacts::Sources;
|
|
use error::Result;
|
|
use std::{
|
|
borrow::Cow,
|
|
collections::{BTreeMap, HashMap},
|
|
convert::TryInto,
|
|
fmt, fs, io,
|
|
marker::PhantomData,
|
|
path::PathBuf,
|
|
};
|
|
|
|
/// Handles contract compiling
|
|
#[derive(Debug)]
|
|
pub struct Project<Artifacts: ArtifactOutput = MinimalCombinedArtifacts> {
|
|
/// The layout of the
|
|
pub paths: ProjectPathsConfig,
|
|
/// Where to find solc
|
|
pub solc: Solc,
|
|
/// How solc invocation should be configured.
|
|
pub solc_config: SolcConfig,
|
|
/// Whether caching is enabled
|
|
pub cached: bool,
|
|
/// Whether writing artifacts to disk is enabled
|
|
pub no_artifacts: bool,
|
|
/// Whether writing artifacts to disk is enabled
|
|
pub auto_detect: bool,
|
|
/// How to handle compiler output
|
|
pub artifacts: PhantomData<Artifacts>,
|
|
/// Errors/Warnings which match these error codes are not going to be logged
|
|
pub ignored_error_codes: Vec<u64>,
|
|
/// The paths which will be allowed for library inclusion
|
|
pub allowed_lib_paths: AllowedLibPaths,
|
|
}
|
|
|
|
impl Project {
|
|
/// Convenience function to call `ProjectBuilder::default()`
|
|
///
|
|
/// # Example
|
|
///
|
|
/// Configure with `MinimalCombinedArtifacts` artifacts output
|
|
///
|
|
/// ```rust
|
|
/// use ethers_solc::Project;
|
|
/// let config = Project::builder().build().unwrap();
|
|
/// ```
|
|
///
|
|
/// To configure any a project with any `ArtifactOutput` use either
|
|
///
|
|
/// ```rust
|
|
/// use ethers_solc::Project;
|
|
/// let config = Project::builder().build().unwrap();
|
|
/// ```
|
|
///
|
|
/// or use the builder directly
|
|
///
|
|
/// ```rust
|
|
/// use ethers_solc::{MinimalCombinedArtifacts, ProjectBuilder};
|
|
/// let config = ProjectBuilder::<MinimalCombinedArtifacts>::default().build().unwrap();
|
|
/// ```
|
|
pub fn builder() -> ProjectBuilder {
|
|
ProjectBuilder::default()
|
|
}
|
|
}
|
|
|
|
impl<Artifacts: ArtifactOutput> Project<Artifacts> {
|
|
fn write_cache_file(
|
|
&self,
|
|
sources: Sources,
|
|
artifacts: Vec<(PathBuf, Vec<String>)>,
|
|
) -> Result<()> {
|
|
let mut cache = SolFilesCache::builder()
|
|
.root(&self.paths.root)
|
|
.solc_config(self.solc_config.clone())
|
|
.insert_files(sources)?;
|
|
|
|
// add the artifacts for each file to the cache entry
|
|
for (file, artifacts) in artifacts {
|
|
if let Some(entry) = cache.files.get_mut(&file) {
|
|
entry.artifacts = artifacts;
|
|
}
|
|
}
|
|
|
|
if let Some(cache_dir) = self.paths.cache.parent() {
|
|
fs::create_dir_all(cache_dir)?
|
|
}
|
|
cache.write(&self.paths.cache)
|
|
}
|
|
|
|
/// Returns all sources found under the project's sources path
|
|
pub fn sources(&self) -> io::Result<Sources> {
|
|
Source::read_all_from(self.paths.sources.as_path())
|
|
}
|
|
|
|
/// This emits the cargo [`rerun-if-changed`](https://doc.rust-lang.org/cargo/reference/build-scripts.html#cargorerun-if-changedpath) instruction.
|
|
/// Which tells Cargo to re-run the build script if a file inside the project's sources
|
|
/// directory has changed.
|
|
///
|
|
/// Use this if you compile a project in a `build.rs` file.
|
|
///
|
|
/// # Example `build.rs` file
|
|
///
|
|
///
|
|
/// ```no_run
|
|
/// use ethers_solc::{Project, ProjectPathsConfig};
|
|
/// // configure the project with all its paths, solc, cache etc.
|
|
/// let project = Project::builder()
|
|
/// .paths(ProjectPathsConfig::hardhat(env!("CARGO_MANIFEST_DIR")).unwrap())
|
|
/// .build()
|
|
/// .unwrap();
|
|
/// let output = project.compile().unwrap();
|
|
/// // Tell Cargo that if a source file changes, to rerun this build script.
|
|
/// project.rerun_if_sources_changed();
|
|
/// ```
|
|
pub fn rerun_if_sources_changed(&self) {
|
|
println!("cargo:rerun-if-changed={}", self.paths.sources.display())
|
|
}
|
|
|
|
/// Attempts to read all unique libraries that are used as imports like "hardhat/console.sol"
|
|
fn resolved_libraries(
|
|
&self,
|
|
sources: &Sources,
|
|
) -> io::Result<BTreeMap<PathBuf, (Source, PathBuf)>> {
|
|
let mut libs = BTreeMap::default();
|
|
for source in sources.values() {
|
|
for import in source.parse_imports() {
|
|
if let Some(lib) = utils::resolve_library(&self.paths.libraries, import) {
|
|
if let Entry::Vacant(entry) = libs.entry(import.into()) {
|
|
entry.insert((Source::read(&lib)?, lib));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
Ok(libs)
|
|
}
|
|
|
|
/// Attempts to compile the contracts found at the configured location.
|
|
///
|
|
/// NOTE: this does not check if the contracts were successfully compiled, see
|
|
/// `CompilerOutput::has_error` instead.
|
|
|
|
/// NB: If the `svm` feature is enabled, this function will automatically detect
|
|
/// solc versions across files.
|
|
pub fn compile(&self) -> Result<ProjectCompileOutput<Artifacts>> {
|
|
let sources = self.sources()?;
|
|
|
|
#[cfg(all(feature = "svm", feature = "async"))]
|
|
if self.auto_detect {
|
|
return self.svm_compile(sources)
|
|
}
|
|
|
|
let mut solc = self.solc.clone();
|
|
if !self.allowed_lib_paths.0.is_empty() {
|
|
solc = solc.arg("--allow-paths").arg(self.allowed_lib_paths.to_string());
|
|
}
|
|
self.compile_with_version(&solc, sources)
|
|
}
|
|
|
|
#[cfg(all(feature = "svm", feature = "async"))]
|
|
fn svm_compile(&self, sources: Sources) -> Result<ProjectCompileOutput<Artifacts>> {
|
|
// split them by version
|
|
let mut sources_by_version = BTreeMap::new();
|
|
for (path, source) in sources.into_iter() {
|
|
// will detect and install the solc version
|
|
let version = Solc::detect_version(&source)?;
|
|
// gets the solc binary for that version, it is expected tha this will succeed
|
|
// AND find the solc since it was installed right above
|
|
let mut solc = Solc::find_svm_installed_version(version.to_string())?
|
|
.expect("solc should have been installed");
|
|
|
|
if !self.allowed_lib_paths.0.is_empty() {
|
|
solc = solc.arg("--allow-paths").arg(self.allowed_lib_paths.to_string());
|
|
}
|
|
let entry = sources_by_version.entry(solc).or_insert_with(BTreeMap::new);
|
|
entry.insert(path, source);
|
|
}
|
|
|
|
let mut compiled =
|
|
ProjectCompileOutput::with_ignored_errors(self.ignored_error_codes.clone());
|
|
|
|
// run the compilation step for each version
|
|
for (solc, sources) in sources_by_version {
|
|
compiled.extend(self.compile_with_version(&solc, sources)?);
|
|
}
|
|
if !compiled.has_compiled_contracts() &&
|
|
!compiled.has_compiler_errors() &&
|
|
self.cached &&
|
|
self.paths.cache.exists()
|
|
{
|
|
let cache = SolFilesCache::read(&self.paths.cache)?;
|
|
let artifacts = cache.read_artifacts::<Artifacts>(&self.paths.artifacts)?;
|
|
compiled.artifacts.extend(artifacts);
|
|
}
|
|
Ok(compiled)
|
|
}
|
|
|
|
pub fn compile_with_version(
|
|
&self,
|
|
solc: &Solc,
|
|
mut sources: Sources,
|
|
) -> Result<ProjectCompileOutput<Artifacts>> {
|
|
// add all libraries to the source set while keeping track of their actual disk path
|
|
// (`contracts/contract.sol` -> `/Users/.../contracts.sol`)
|
|
let mut source_name_to_path = HashMap::new();
|
|
// inverse of `source_name_to_path` : (`/Users/.../contracts.sol` ->
|
|
// `contracts/contract.sol`)
|
|
let mut path_to_source_name = HashMap::new();
|
|
|
|
for (import, (source, path)) in self.resolved_libraries(&sources)? {
|
|
// inserting with absolute path here and keep track of the source name <-> path mappings
|
|
sources.insert(path.clone(), source);
|
|
path_to_source_name.insert(path.clone(), import.clone());
|
|
source_name_to_path.insert(import, path);
|
|
}
|
|
|
|
// If there's a cache set, filter to only re-compile the files which were changed
|
|
let sources = if self.cached && self.paths.cache.exists() {
|
|
let cache = SolFilesCache::read(&self.paths.cache)?;
|
|
let changed_files = cache.get_changed_or_missing_artifacts_files::<Artifacts>(
|
|
sources,
|
|
Some(&self.solc_config),
|
|
&self.paths.artifacts,
|
|
);
|
|
|
|
// if nothing changed and all artifacts still exist
|
|
if changed_files.is_empty() {
|
|
let artifacts = cache.read_artifacts::<Artifacts>(&self.paths.artifacts)?;
|
|
return Ok(ProjectCompileOutput::from_unchanged(artifacts))
|
|
}
|
|
changed_files
|
|
} else {
|
|
sources
|
|
};
|
|
|
|
// replace absolute path with source name to make solc happy
|
|
let sources = apply_mappings(sources, path_to_source_name);
|
|
|
|
let input = CompilerInput::with_sources(sources)
|
|
.normalize_evm_version(&solc.version()?)
|
|
.with_remappings(self.paths.remappings.clone());
|
|
let output = solc.compile(&input)?;
|
|
if output.has_error() {
|
|
return Ok(ProjectCompileOutput::from_compiler_output(
|
|
output,
|
|
self.ignored_error_codes.clone(),
|
|
))
|
|
}
|
|
|
|
if self.cached {
|
|
// get all contract names of the files and map them to the disk file
|
|
let artifacts = output
|
|
.contracts
|
|
.iter()
|
|
.map(|(path, contracts)| {
|
|
let path = PathBuf::from(path);
|
|
let file = source_name_to_path.get(&path).cloned().unwrap_or(path);
|
|
(file, contracts.keys().cloned().collect::<Vec<_>>())
|
|
})
|
|
.collect::<Vec<_>>();
|
|
|
|
// reapply to disk paths
|
|
let sources = apply_mappings(input.sources, source_name_to_path);
|
|
|
|
// create cache file
|
|
self.write_cache_file(sources, artifacts)?;
|
|
}
|
|
|
|
// TODO: There seems to be some type redundancy here, c.f. discussion with @mattsse
|
|
if !self.no_artifacts {
|
|
Artifacts::on_output(&output, &self.paths)?;
|
|
}
|
|
Ok(ProjectCompileOutput::from_compiler_output(output, self.ignored_error_codes.clone()))
|
|
}
|
|
}
|
|
|
|
fn apply_mappings(sources: Sources, mut mappings: HashMap<PathBuf, PathBuf>) -> Sources {
|
|
sources
|
|
.into_iter()
|
|
.map(|(import, source)| {
|
|
if let Some(path) = mappings.remove(&import) {
|
|
(path, source)
|
|
} else {
|
|
(import, source)
|
|
}
|
|
})
|
|
.collect()
|
|
}
|
|
|
|
pub struct ProjectBuilder<Artifacts: ArtifactOutput = MinimalCombinedArtifacts> {
|
|
/// The layout of the
|
|
paths: Option<ProjectPathsConfig>,
|
|
/// Where to find solc
|
|
solc: Option<Solc>,
|
|
/// How solc invocation should be configured.
|
|
solc_config: Option<SolcConfig>,
|
|
/// Whether caching is enabled, default is true.
|
|
cached: bool,
|
|
/// Whether writing artifacts to disk is enabled, default is true.
|
|
no_artifacts: bool,
|
|
/// Whether automatic solc version detection is enabled
|
|
auto_detect: bool,
|
|
artifacts: PhantomData<Artifacts>,
|
|
/// Which error codes to ignore
|
|
pub ignored_error_codes: Vec<u64>,
|
|
/// All allowed paths
|
|
pub allowed_paths: Vec<PathBuf>,
|
|
}
|
|
|
|
impl<Artifacts: ArtifactOutput> ProjectBuilder<Artifacts> {
|
|
pub fn paths(mut self, paths: ProjectPathsConfig) -> Self {
|
|
self.paths = Some(paths);
|
|
self
|
|
}
|
|
|
|
pub fn solc(mut self, solc: impl Into<Solc>) -> Self {
|
|
self.solc = Some(solc.into());
|
|
self
|
|
}
|
|
|
|
pub fn solc_config(mut self, solc_config: SolcConfig) -> Self {
|
|
self.solc_config = Some(solc_config);
|
|
self
|
|
}
|
|
|
|
pub fn ignore_error_code(mut self, code: u64) -> Self {
|
|
self.ignored_error_codes.push(code);
|
|
self
|
|
}
|
|
|
|
/// Disables cached builds
|
|
pub fn ephemeral(mut self) -> Self {
|
|
self.cached = false;
|
|
self
|
|
}
|
|
|
|
/// Disables writing artifacts to disk
|
|
pub fn no_artifacts(mut self) -> Self {
|
|
self.no_artifacts = true;
|
|
self
|
|
}
|
|
|
|
/// Disables automatic solc version detection
|
|
pub fn no_auto_detect(mut self) -> Self {
|
|
self.auto_detect = false;
|
|
self
|
|
}
|
|
|
|
/// Set arbitrary `ArtifactOutputHandler`
|
|
pub fn artifacts<A: ArtifactOutput>(self) -> ProjectBuilder<A> {
|
|
let ProjectBuilder {
|
|
paths,
|
|
solc,
|
|
solc_config,
|
|
cached,
|
|
no_artifacts,
|
|
auto_detect,
|
|
ignored_error_codes,
|
|
allowed_paths,
|
|
..
|
|
} = self;
|
|
ProjectBuilder {
|
|
paths,
|
|
solc,
|
|
solc_config,
|
|
cached,
|
|
no_artifacts,
|
|
auto_detect,
|
|
artifacts: PhantomData::default(),
|
|
ignored_error_codes,
|
|
allowed_paths,
|
|
}
|
|
}
|
|
|
|
/// Adds an allowed-path to the solc executable
|
|
pub fn allowed_path<T: Into<PathBuf>>(mut self, path: T) -> Self {
|
|
self.allowed_paths.push(path.into());
|
|
self
|
|
}
|
|
|
|
/// Adds multiple allowed-path to the solc executable
|
|
pub fn allowed_paths<I, S>(mut self, args: I) -> Self
|
|
where
|
|
I: IntoIterator<Item = S>,
|
|
S: Into<PathBuf>,
|
|
{
|
|
for arg in args {
|
|
self = self.allowed_path(arg);
|
|
}
|
|
self
|
|
}
|
|
|
|
pub fn build(self) -> Result<Project<Artifacts>> {
|
|
let Self {
|
|
paths,
|
|
solc,
|
|
solc_config,
|
|
cached,
|
|
no_artifacts,
|
|
auto_detect,
|
|
artifacts,
|
|
ignored_error_codes,
|
|
mut allowed_paths,
|
|
} = self;
|
|
|
|
let solc = solc.unwrap_or_default();
|
|
let solc_config = solc_config.map(Ok).unwrap_or_else(|| SolcConfig::builder().build())?;
|
|
|
|
let paths = paths.map(Ok).unwrap_or_else(ProjectPathsConfig::current_hardhat)?;
|
|
|
|
if allowed_paths.is_empty() {
|
|
// allow every contract under root by default
|
|
allowed_paths.push(paths.root.clone())
|
|
}
|
|
|
|
Ok(Project {
|
|
paths,
|
|
solc,
|
|
solc_config,
|
|
cached,
|
|
no_artifacts,
|
|
auto_detect,
|
|
artifacts,
|
|
ignored_error_codes,
|
|
allowed_lib_paths: allowed_paths.try_into()?,
|
|
})
|
|
}
|
|
}
|
|
|
|
impl<Artifacts: ArtifactOutput> Default for ProjectBuilder<Artifacts> {
|
|
fn default() -> Self {
|
|
Self {
|
|
paths: None,
|
|
solc: None,
|
|
solc_config: None,
|
|
cached: true,
|
|
no_artifacts: false,
|
|
auto_detect: true,
|
|
artifacts: PhantomData::default(),
|
|
ignored_error_codes: Vec::new(),
|
|
allowed_paths: vec![],
|
|
}
|
|
}
|
|
}
|
|
|
|
/// The outcome of `Project::compile`
|
|
#[derive(Debug, Clone, PartialEq, Default)]
|
|
pub struct ProjectCompileOutput<T: ArtifactOutput> {
|
|
/// If solc was invoked multiple times in `Project::compile` then this contains a merged
|
|
/// version of all `CompilerOutput`s. If solc was called only once then `compiler_output`
|
|
/// holds the `CompilerOutput` of that call.
|
|
compiler_output: Option<CompilerOutput>,
|
|
/// All artifacts that were read from cache
|
|
artifacts: BTreeMap<PathBuf, T::Artifact>,
|
|
ignored_error_codes: Vec<u64>,
|
|
}
|
|
|
|
impl<T: ArtifactOutput> ProjectCompileOutput<T> {
|
|
pub fn with_ignored_errors(ignored_errors: Vec<u64>) -> Self {
|
|
Self {
|
|
compiler_output: None,
|
|
artifacts: Default::default(),
|
|
ignored_error_codes: ignored_errors,
|
|
}
|
|
}
|
|
|
|
pub fn from_unchanged(artifacts: BTreeMap<PathBuf, T::Artifact>) -> Self {
|
|
Self { compiler_output: None, artifacts, ignored_error_codes: vec![] }
|
|
}
|
|
|
|
pub fn from_compiler_output(
|
|
compiler_output: CompilerOutput,
|
|
ignored_error_codes: Vec<u64>,
|
|
) -> Self {
|
|
Self {
|
|
compiler_output: Some(compiler_output),
|
|
artifacts: Default::default(),
|
|
ignored_error_codes,
|
|
}
|
|
}
|
|
|
|
/// Get the (merged) solc compiler output
|
|
/// ```no_run
|
|
/// use std::collections::BTreeMap;
|
|
/// use ethers_solc::artifacts::Contract;
|
|
/// use ethers_solc::Project;
|
|
///
|
|
/// let project = Project::builder().build().unwrap();
|
|
/// let contracts: BTreeMap<String, Contract> =
|
|
/// project.compile().unwrap().output().contracts_into_iter().collect();
|
|
/// ```
|
|
pub fn output(self) -> CompilerOutput {
|
|
self.compiler_output.unwrap_or_default()
|
|
}
|
|
|
|
/// Combine two outputs
|
|
pub fn extend(&mut self, compiled: ProjectCompileOutput<T>) {
|
|
let ProjectCompileOutput { compiler_output, artifacts, .. } = compiled;
|
|
self.artifacts.extend(artifacts);
|
|
if let Some(compiled) = compiler_output {
|
|
if let Some(output) = self.compiler_output.as_mut() {
|
|
output.errors.extend(compiled.errors);
|
|
output.sources.extend(compiled.sources);
|
|
output.contracts.extend(compiled.contracts);
|
|
} else {
|
|
self.compiler_output = Some(compiled);
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Whether this type does not contain compiled contracts
|
|
pub fn is_unchanged(&self) -> bool {
|
|
!self.has_compiled_contracts()
|
|
}
|
|
|
|
/// Whether this type has a compiler output
|
|
pub fn has_compiled_contracts(&self) -> bool {
|
|
if let Some(output) = self.compiler_output.as_ref() {
|
|
!output.contracts.is_empty()
|
|
} else {
|
|
false
|
|
}
|
|
}
|
|
|
|
/// Whether there were errors
|
|
pub fn has_compiler_errors(&self) -> bool {
|
|
if let Some(output) = self.compiler_output.as_ref() {
|
|
output.has_error()
|
|
} else {
|
|
false
|
|
}
|
|
}
|
|
|
|
/// Finds the first contract with the given name and removes it from the set
|
|
pub fn remove(&mut self, contract: impl AsRef<str>) -> Option<T::Artifact> {
|
|
let contract = contract.as_ref();
|
|
if let Some(output) = self.compiler_output.as_mut() {
|
|
if let contract @ Some(_) = output
|
|
.contracts
|
|
.values_mut()
|
|
.find_map(|c| c.remove(contract).map(T::contract_to_artifact))
|
|
{
|
|
return contract
|
|
}
|
|
}
|
|
let key = self
|
|
.artifacts
|
|
.iter()
|
|
.find_map(|(path, _)| {
|
|
T::contract_name(path).filter(|name| name == contract).map(|_| path)
|
|
})?
|
|
.clone();
|
|
self.artifacts.remove(&key)
|
|
}
|
|
}
|
|
|
|
impl<T: ArtifactOutput> ProjectCompileOutput<T>
|
|
where
|
|
T::Artifact: Clone,
|
|
{
|
|
/// Finds the first contract with the given name
|
|
pub fn find(&self, contract: impl AsRef<str>) -> Option<Cow<T::Artifact>> {
|
|
let contract = contract.as_ref();
|
|
if let Some(output) = self.compiler_output.as_ref() {
|
|
if let contract @ Some(_) = output.contracts.values().find_map(|c| {
|
|
c.get(contract).map(|c| T::contract_to_artifact(c.clone())).map(Cow::Owned)
|
|
}) {
|
|
return contract
|
|
}
|
|
}
|
|
self.artifacts.iter().find_map(|(path, art)| {
|
|
T::contract_name(path).filter(|name| name == contract).map(|_| Cow::Borrowed(art))
|
|
})
|
|
}
|
|
}
|
|
|
|
impl<T: ArtifactOutput + 'static> ProjectCompileOutput<T> {
|
|
/// All artifacts together with their contract name
|
|
///
|
|
/// # Example
|
|
///
|
|
/// ```no_run
|
|
/// use std::collections::BTreeMap;
|
|
/// use ethers_solc::artifacts::CompactContract;
|
|
/// use ethers_solc::Project;
|
|
///
|
|
/// let project = Project::builder().build().unwrap();
|
|
/// let contracts: BTreeMap<String, CompactContract> = project.compile().unwrap().into_artifacts().collect();
|
|
/// ```
|
|
pub fn into_artifacts(mut self) -> Box<dyn Iterator<Item = (String, T::Artifact)>> {
|
|
let artifacts = self.artifacts.into_iter().filter_map(|(path, art)| {
|
|
T::contract_name(&path)
|
|
.map(|name| (format!("{:?}:{}", path.file_name().unwrap(), name), art))
|
|
});
|
|
|
|
let artifacts: Box<dyn Iterator<Item = (String, T::Artifact)>> =
|
|
if let Some(output) = self.compiler_output.take() {
|
|
Box::new(artifacts.chain(T::output_to_artifacts(output).into_values().flatten()))
|
|
} else {
|
|
Box::new(artifacts)
|
|
};
|
|
artifacts
|
|
}
|
|
}
|
|
|
|
impl<T: ArtifactOutput> fmt::Display for ProjectCompileOutput<T> {
|
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
|
if let Some(output) = self.compiler_output.as_ref() {
|
|
output.diagnostics(&self.ignored_error_codes).fmt(f)
|
|
} else {
|
|
f.write_str("Nothing to compile")
|
|
}
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
|
|
#[test]
|
|
#[cfg(all(feature = "svm", feature = "async"))]
|
|
fn test_build_all_versions() {
|
|
use super::*;
|
|
|
|
let paths = ProjectPathsConfig::builder()
|
|
.root("./test-data/test-contract-versions")
|
|
.sources("./test-data/test-contract-versions")
|
|
.build()
|
|
.unwrap();
|
|
let project = Project::builder().paths(paths).no_artifacts().ephemeral().build().unwrap();
|
|
let compiled = project.compile().unwrap();
|
|
assert!(!compiled.has_compiler_errors());
|
|
let contracts = compiled.output().contracts;
|
|
// Contracts A to F
|
|
assert_eq!(contracts.keys().count(), 5);
|
|
}
|
|
|
|
#[test]
|
|
#[cfg(all(feature = "svm", feature = "async"))]
|
|
fn test_build_many_libs() {
|
|
use super::*;
|
|
|
|
let root = std::fs::canonicalize("./test-data/test-contract-libs").unwrap();
|
|
|
|
let paths = ProjectPathsConfig::builder()
|
|
.root(&root)
|
|
.sources(root.join("src"))
|
|
.lib(root.join("lib1"))
|
|
.lib(root.join("lib2"))
|
|
.build()
|
|
.unwrap();
|
|
let project = Project::builder()
|
|
.paths(paths)
|
|
.no_artifacts()
|
|
.ephemeral()
|
|
.no_artifacts()
|
|
.build()
|
|
.unwrap();
|
|
let compiled = project.compile().unwrap();
|
|
assert!(!compiled.has_compiler_errors());
|
|
let contracts = compiled.output().contracts;
|
|
assert_eq!(contracts.keys().count(), 3);
|
|
}
|
|
|
|
#[test]
|
|
#[cfg(all(feature = "svm", feature = "async"))]
|
|
fn test_build_remappings() {
|
|
use super::*;
|
|
|
|
let root = std::fs::canonicalize("./test-data/test-contract-remappings").unwrap();
|
|
let paths = ProjectPathsConfig::builder()
|
|
.root(&root)
|
|
.sources(root.join("src"))
|
|
.lib(root.join("lib"))
|
|
.build()
|
|
.unwrap();
|
|
let project = Project::builder().no_artifacts().paths(paths).ephemeral().build().unwrap();
|
|
let compiled = project.compile().unwrap();
|
|
assert!(!compiled.has_compiler_errors());
|
|
let contracts = compiled.output().contracts;
|
|
assert_eq!(contracts.keys().count(), 2);
|
|
}
|
|
}
|