mirror of
https://github.com/Start9Labs/start-os.git
synced 2026-03-26 18:31:52 +00:00
1220 lines
40 KiB
Rust
1220 lines
40 KiB
Rust
use std::collections::BTreeMap;
|
|
use std::path::{Path, PathBuf};
|
|
use std::pin::Pin;
|
|
use std::sync::Arc;
|
|
use std::time::SystemTime;
|
|
|
|
use deno_core::anyhow::{anyhow, bail};
|
|
use deno_core::error::AnyError;
|
|
use deno_core::{
|
|
resolve_import, Extension, FastString, JsRuntime, ModuleLoader, ModuleSource,
|
|
ModuleSourceFuture, ModuleSpecifier, ModuleType, OpDecl, ResolutionKind, RuntimeOptions,
|
|
Snapshot,
|
|
};
|
|
use helpers::{script_dir, spawn_local, Rsync};
|
|
use models::{PackageId, ProcedureName, Version, VolumeId};
|
|
use serde::{Deserialize, Serialize};
|
|
use serde_json::Value;
|
|
use tokio::io::AsyncReadExt;
|
|
use tokio::sync::Mutex;
|
|
|
|
lazy_static::lazy_static! {
|
|
static ref DENO_GLOBAL_JS: ModuleSpecifier = "file:///deno_global.js".parse().unwrap();
|
|
static ref LOAD_MODULE_JS: ModuleSpecifier = "file:///loadModule.js".parse().unwrap();
|
|
static ref EMBASSY_JS: ModuleSpecifier = "file:///embassy.js".parse().unwrap();
|
|
}
|
|
|
|
pub trait PathForVolumeId: Send + Sync {
|
|
fn path_for(
|
|
&self,
|
|
data_dir: &Path,
|
|
package_id: &PackageId,
|
|
version: &Version,
|
|
volume_id: &VolumeId,
|
|
) -> Option<PathBuf>;
|
|
fn readonly(&self, volume_id: &VolumeId) -> bool;
|
|
}
|
|
|
|
#[derive(Serialize, Deserialize, Debug, Clone)]
|
|
pub struct JsCode(Arc<str>);
|
|
|
|
#[derive(Debug, Clone, Copy)]
|
|
pub enum JsError {
|
|
Unknown,
|
|
Javascript,
|
|
Engine,
|
|
BoundryLayerSerDe,
|
|
Tokio,
|
|
FileSystem,
|
|
Code(i32),
|
|
Timeout,
|
|
NotValidProcedureName,
|
|
}
|
|
|
|
impl JsError {
|
|
pub fn as_code_num(&self) -> i32 {
|
|
match self {
|
|
JsError::Unknown => 1,
|
|
JsError::Javascript => 2,
|
|
JsError::Engine => 3,
|
|
JsError::BoundryLayerSerDe => 4,
|
|
JsError::Tokio => 5,
|
|
JsError::FileSystem => 6,
|
|
JsError::NotValidProcedureName => 7,
|
|
JsError::Code(code) => *code,
|
|
JsError::Timeout => 143,
|
|
}
|
|
}
|
|
}
|
|
|
|
#[derive(Debug, Clone, Serialize, Deserialize)]
|
|
#[serde(rename_all = "camelCase")]
|
|
pub struct MetadataJs {
|
|
file_type: String,
|
|
is_dir: bool,
|
|
is_file: bool,
|
|
is_symlink: bool,
|
|
len: u64,
|
|
modified: Option<u64>,
|
|
accessed: Option<u64>,
|
|
created: Option<u64>,
|
|
readonly: bool,
|
|
gid: u32,
|
|
mode: u32,
|
|
uid: u32,
|
|
}
|
|
|
|
#[cfg(target_arch = "x86_64")]
|
|
const SNAPSHOT_BYTES: &[u8] = include_bytes!("./artifacts/JS_SNAPSHOT.x86_64.bin");
|
|
|
|
#[cfg(target_arch = "aarch64")]
|
|
const SNAPSHOT_BYTES: &[u8] = include_bytes!("./artifacts/JS_SNAPSHOT.aarch64.bin");
|
|
|
|
#[derive(Clone)]
|
|
struct JsContext {
|
|
sandboxed: bool,
|
|
datadir: PathBuf,
|
|
run_function: String,
|
|
version: Version,
|
|
package_id: PackageId,
|
|
volumes: Arc<dyn PathForVolumeId>,
|
|
input: Value,
|
|
variable_args: Vec<serde_json::Value>,
|
|
rsyncs: Arc<Mutex<(usize, BTreeMap<usize, Rsync>)>>,
|
|
}
|
|
#[derive(Debug, Clone, serde::Deserialize, serde::Serialize)]
|
|
#[serde(rename_all = "kebab-case")]
|
|
enum ResultType {
|
|
Error(String),
|
|
ErrorCode(i32, String),
|
|
Result(serde_json::Value),
|
|
}
|
|
#[derive(Clone, Default)]
|
|
struct AnswerState(std::sync::Arc<deno_core::parking_lot::Mutex<Value>>);
|
|
|
|
#[derive(Clone, Debug)]
|
|
struct ModsLoader {
|
|
code: JsCode,
|
|
}
|
|
|
|
impl ModuleLoader for ModsLoader {
|
|
fn resolve(
|
|
&self,
|
|
specifier: &str,
|
|
referrer: &str,
|
|
_is_main: ResolutionKind,
|
|
) -> Result<ModuleSpecifier, AnyError> {
|
|
if referrer.contains("embassy") {
|
|
bail!("Embassy.js cannot import anything else");
|
|
}
|
|
let s = resolve_import(specifier, referrer).unwrap();
|
|
Ok(s)
|
|
}
|
|
|
|
fn load(
|
|
&self,
|
|
module_specifier: &ModuleSpecifier,
|
|
maybe_referrer: Option<&ModuleSpecifier>,
|
|
is_dyn_import: bool,
|
|
) -> Pin<Box<ModuleSourceFuture>> {
|
|
let module_specifier = module_specifier.as_str().to_owned();
|
|
let module = match &*module_specifier {
|
|
"file:///deno_global.js" => Ok(ModuleSource::new(
|
|
ModuleType::JavaScript,
|
|
FastString::Static("const old_deno = Deno; Deno = null; export default old_deno"),
|
|
&DENO_GLOBAL_JS,
|
|
)),
|
|
"file:///loadModule.js" => Ok(ModuleSource::new(
|
|
ModuleType::JavaScript,
|
|
FastString::Static(include_str!("./artifacts/loadModule.js")),
|
|
&LOAD_MODULE_JS,
|
|
)),
|
|
"file:///embassy.js" => Ok(ModuleSource::new(
|
|
ModuleType::JavaScript,
|
|
self.code.0.clone().into(),
|
|
&EMBASSY_JS,
|
|
)),
|
|
|
|
x => Err(anyhow!("Not allowed to import: {}", x)),
|
|
};
|
|
let module = module.and_then(|m| {
|
|
if is_dyn_import {
|
|
bail!("Will not import dynamic");
|
|
}
|
|
match &maybe_referrer {
|
|
Some(x) if x.as_str() == "file:///embassy.js" => {
|
|
bail!("StartJS is not allowed to import")
|
|
}
|
|
_ => (),
|
|
}
|
|
Ok(m)
|
|
});
|
|
Box::pin(async move { module })
|
|
}
|
|
}
|
|
|
|
pub struct JsExecutionEnvironment {
|
|
sandboxed: bool,
|
|
base_directory: PathBuf,
|
|
module_loader: ModsLoader,
|
|
package_id: PackageId,
|
|
version: Version,
|
|
volumes: Arc<dyn PathForVolumeId>,
|
|
}
|
|
|
|
impl JsExecutionEnvironment {
|
|
pub async fn load_from_package(
|
|
data_directory: impl AsRef<std::path::Path>,
|
|
package_id: &PackageId,
|
|
version: &Version,
|
|
volumes: Box<dyn PathForVolumeId>,
|
|
) -> Result<JsExecutionEnvironment, (JsError, String)> {
|
|
let data_dir = data_directory.as_ref();
|
|
let base_directory = data_dir;
|
|
let js_code = JsCode({
|
|
let file_path = script_dir(data_dir, package_id, version).join("embassy.js");
|
|
let mut file = match tokio::fs::File::open(file_path.clone()).await {
|
|
Ok(x) => x,
|
|
Err(e) => {
|
|
tracing::debug!("path: {:?}", file_path);
|
|
tracing::debug!("{:?}", e);
|
|
return Err((
|
|
JsError::FileSystem,
|
|
format!("The file opening '{:?}' created error: {}", file_path, e),
|
|
));
|
|
}
|
|
};
|
|
let mut buffer = Default::default();
|
|
if let Err(err) = file.read_to_string(&mut buffer).await {
|
|
tracing::debug!("{:?}", err);
|
|
return Err((
|
|
JsError::FileSystem,
|
|
format!("The file reading created error: {}", err),
|
|
));
|
|
};
|
|
buffer.into()
|
|
});
|
|
Ok(JsExecutionEnvironment {
|
|
base_directory: base_directory.to_owned(),
|
|
module_loader: ModsLoader { code: js_code },
|
|
package_id: package_id.clone(),
|
|
version: version.clone(),
|
|
volumes: volumes.into(),
|
|
sandboxed: false,
|
|
})
|
|
}
|
|
pub fn read_only_effects(mut self) -> Self {
|
|
self.sandboxed = true;
|
|
self
|
|
}
|
|
|
|
pub async fn run_action<I: Serialize, O: for<'de> Deserialize<'de>>(
|
|
self,
|
|
procedure_name: ProcedureName,
|
|
input: Option<I>,
|
|
variable_args: Vec<serde_json::Value>,
|
|
) -> Result<O, (JsError, String)> {
|
|
let input = match serde_json::to_value(input) {
|
|
Ok(a) => a,
|
|
Err(err) => {
|
|
tracing::error!("{}", err);
|
|
tracing::debug!("{:?}", err);
|
|
return Err((
|
|
JsError::BoundryLayerSerDe,
|
|
"Couldn't convert input".to_string(),
|
|
));
|
|
}
|
|
};
|
|
let safer_handle = spawn_local(|| self.execute(procedure_name, input, variable_args)).await;
|
|
let output = safer_handle.await.unwrap()?;
|
|
match serde_json::from_value(output.clone()) {
|
|
Ok(x) => Ok(x),
|
|
Err(err) => {
|
|
tracing::error!("{}", err);
|
|
tracing::debug!("{:?}", err);
|
|
Err((
|
|
JsError::BoundryLayerSerDe,
|
|
format!(
|
|
"Couldn't convert output = {:#?} to the correct type",
|
|
serde_json::to_string_pretty(&output).unwrap_or_default()
|
|
),
|
|
))
|
|
}
|
|
}
|
|
}
|
|
fn declarations() -> Vec<OpDecl> {
|
|
vec![
|
|
fns::chown::decl(),
|
|
fns::chmod::decl(),
|
|
fns::fetch::decl(),
|
|
fns::read_file::decl(),
|
|
fns::metadata::decl(),
|
|
fns::write_file::decl(),
|
|
fns::rename::decl(),
|
|
fns::remove_file::decl(),
|
|
fns::create_dir::decl(),
|
|
fns::remove_dir::decl(),
|
|
fns::read_dir::decl(),
|
|
fns::disk_usage::decl(),
|
|
fns::current_function::decl(),
|
|
fns::log_trace::decl(),
|
|
fns::log_warn::decl(),
|
|
fns::log_error::decl(),
|
|
fns::log_debug::decl(),
|
|
fns::log_info::decl(),
|
|
fns::get_input::decl(),
|
|
fns::get_variable_args::decl(),
|
|
fns::set_value::decl(),
|
|
fns::is_sandboxed::decl(),
|
|
fns::sleep::decl(),
|
|
fns::rsync::decl(),
|
|
fns::rsync_wait::decl(),
|
|
fns::rsync_progress::decl(),
|
|
]
|
|
}
|
|
|
|
async fn execute(
|
|
self,
|
|
procedure_name: ProcedureName,
|
|
input: Value,
|
|
variable_args: Vec<serde_json::Value>,
|
|
) -> Result<Value, (JsError, String)> {
|
|
let base_directory = self.base_directory.clone();
|
|
let answer_state = AnswerState::default();
|
|
let ext_answer_state = answer_state.clone();
|
|
let js_ctx = JsContext {
|
|
datadir: base_directory,
|
|
run_function: procedure_name
|
|
.js_function_name()
|
|
.map(Ok)
|
|
.unwrap_or_else(|| {
|
|
Err((
|
|
JsError::NotValidProcedureName,
|
|
format!("procedure is not value: {:?}", procedure_name),
|
|
))
|
|
})?,
|
|
package_id: self.package_id.clone(),
|
|
volumes: self.volumes.clone(),
|
|
version: self.version.clone(),
|
|
sandboxed: self.sandboxed,
|
|
input,
|
|
variable_args,
|
|
rsyncs: Default::default(),
|
|
};
|
|
let ext = Extension::builder("embassy")
|
|
.ops(Self::declarations())
|
|
.state(move |state| {
|
|
state.put(ext_answer_state.clone());
|
|
state.put(js_ctx);
|
|
})
|
|
.build();
|
|
|
|
let loader = std::rc::Rc::new(self.module_loader.clone());
|
|
let runtime_options = RuntimeOptions {
|
|
module_loader: Some(loader),
|
|
extensions: vec![ext],
|
|
startup_snapshot: Some(Snapshot::Static(SNAPSHOT_BYTES)),
|
|
..Default::default()
|
|
};
|
|
let mut runtime = JsRuntime::new(runtime_options);
|
|
|
|
let future = async move {
|
|
let mod_id = runtime
|
|
.load_main_module(&"file:///loadModule.js".parse().unwrap(), None)
|
|
.await?;
|
|
let evaluated = runtime.mod_evaluate(mod_id);
|
|
let res = runtime.run_event_loop(false).await;
|
|
res?;
|
|
evaluated.await??;
|
|
Ok::<_, AnyError>(())
|
|
};
|
|
|
|
future.await.map_err(|e| {
|
|
tracing::debug!("{:?}", e);
|
|
(JsError::Javascript, format!("{}", e))
|
|
})?;
|
|
|
|
let answer = answer_state.0.lock().clone();
|
|
Ok(answer)
|
|
}
|
|
}
|
|
|
|
/// Note: Make sure that we have the assumption that all these methods are callable at any time, and all call restrictions should be in rust
|
|
mod fns {
|
|
use std::cell::RefCell;
|
|
use std::collections::BTreeMap;
|
|
use std::convert::TryFrom;
|
|
use std::fs::Permissions;
|
|
use std::os::unix::fs::MetadataExt;
|
|
use std::os::unix::prelude::PermissionsExt;
|
|
use std::path::{Path, PathBuf};
|
|
use std::rc::Rc;
|
|
use std::time::Duration;
|
|
|
|
use container_init::ProcessId;
|
|
use deno_core::anyhow::{anyhow, bail};
|
|
use deno_core::error::AnyError;
|
|
use deno_core::*;
|
|
use helpers::{to_tmp_path, AtomicFile, Rsync, RsyncOptions};
|
|
use itertools::Itertools;
|
|
use models::VolumeId;
|
|
use serde::{Deserialize, Serialize};
|
|
use serde_json::Value;
|
|
use tokio::io::AsyncWriteExt;
|
|
use tokio::process::Command;
|
|
|
|
use super::{AnswerState, JsContext};
|
|
use crate::{system_time_as_unix_ms, MetadataJs};
|
|
|
|
#[derive(serde::Serialize, serde::Deserialize, Debug, Clone, Default)]
|
|
struct FetchOptions {
|
|
method: Option<String>,
|
|
headers: Option<BTreeMap<String, String>>,
|
|
body: Option<String>,
|
|
}
|
|
#[derive(serde::Serialize, serde::Deserialize, Debug, Clone, Default)]
|
|
struct FetchResponse {
|
|
method: String,
|
|
ok: bool,
|
|
status: u32,
|
|
headers: BTreeMap<String, String>,
|
|
body: Option<String>,
|
|
}
|
|
#[op]
|
|
async fn fetch(
|
|
state: Rc<RefCell<OpState>>,
|
|
url: url::Url,
|
|
options: Option<FetchOptions>,
|
|
) -> Result<FetchResponse, AnyError> {
|
|
let sandboxed = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
ctx.sandboxed
|
|
};
|
|
|
|
if sandboxed {
|
|
bail!("Will not run fetch in sandboxed mode");
|
|
}
|
|
|
|
let client = reqwest::Client::new();
|
|
let options = options.unwrap_or_default();
|
|
let method = options
|
|
.method
|
|
.unwrap_or_else(|| "GET".to_string())
|
|
.to_uppercase();
|
|
let mut request_builder = match &*method {
|
|
"GET" => client.get(url),
|
|
"POST" => client.post(url),
|
|
"PUT" => client.put(url),
|
|
"DELETE" => client.delete(url),
|
|
"HEAD" => client.head(url),
|
|
"PATCH" => client.patch(url),
|
|
x => bail!("Unsupported method: {}", x),
|
|
};
|
|
if let Some(headers) = options.headers {
|
|
for (key, value) in headers {
|
|
request_builder = request_builder.header(key, value);
|
|
}
|
|
}
|
|
if let Some(body) = options.body {
|
|
request_builder = request_builder.body(body);
|
|
}
|
|
let response = request_builder.send().await?;
|
|
|
|
let fetch_response = FetchResponse {
|
|
method,
|
|
ok: response.status().is_success(),
|
|
status: response.status().as_u16() as u32,
|
|
headers: response
|
|
.headers()
|
|
.iter()
|
|
.filter_map(|(head, value)| {
|
|
Some((format!("{}", head), value.to_str().ok()?.to_string()))
|
|
})
|
|
.collect(),
|
|
body: response.text().await.ok(),
|
|
};
|
|
|
|
Ok(fetch_response)
|
|
}
|
|
|
|
#[op]
|
|
async fn read_file(
|
|
state: Rc<RefCell<OpState>>,
|
|
volume_id: VolumeId,
|
|
path_in: PathBuf,
|
|
) -> Result<String, AnyError> {
|
|
let volume_path = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
ctx.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &volume_id)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", volume_id))?
|
|
};
|
|
//get_path_for in volume.rs
|
|
let path_in = path_in.strip_prefix("/").unwrap_or(&path_in);
|
|
let new_file = volume_path.join(path_in);
|
|
if !is_subset(&volume_path, &new_file).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
new_file.to_string_lossy(),
|
|
volume_path.to_string_lossy(),
|
|
);
|
|
}
|
|
let answer = tokio::fs::read_to_string(new_file).await?;
|
|
Ok(answer)
|
|
}
|
|
#[op]
|
|
async fn metadata(
|
|
state: Rc<RefCell<OpState>>,
|
|
volume_id: VolumeId,
|
|
path_in: PathBuf,
|
|
) -> Result<MetadataJs, AnyError> {
|
|
let volume_path = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
ctx.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &volume_id)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", volume_id))?
|
|
};
|
|
//get_path_for in volume.rs
|
|
let path_in = path_in.strip_prefix("/").unwrap_or(&path_in);
|
|
let new_file = volume_path.join(path_in);
|
|
if !is_subset(&volume_path, &new_file).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
new_file.to_string_lossy(),
|
|
volume_path.to_string_lossy(),
|
|
);
|
|
}
|
|
let answer = tokio::fs::metadata(new_file).await?;
|
|
let metadata_js = MetadataJs {
|
|
file_type: format!("{:?}", answer.file_type()),
|
|
is_dir: answer.is_dir(),
|
|
is_file: answer.is_file(),
|
|
is_symlink: answer.is_symlink(),
|
|
len: answer.len(),
|
|
modified: answer
|
|
.modified()
|
|
.ok()
|
|
.as_ref()
|
|
.and_then(system_time_as_unix_ms),
|
|
accessed: answer
|
|
.accessed()
|
|
.ok()
|
|
.as_ref()
|
|
.and_then(system_time_as_unix_ms),
|
|
created: answer
|
|
.created()
|
|
.ok()
|
|
.as_ref()
|
|
.and_then(system_time_as_unix_ms),
|
|
readonly: answer.permissions().readonly(),
|
|
gid: answer.gid(),
|
|
mode: answer.mode(),
|
|
uid: answer.uid(),
|
|
};
|
|
|
|
Ok(metadata_js)
|
|
}
|
|
#[op]
|
|
async fn write_file(
|
|
state: Rc<RefCell<OpState>>,
|
|
volume_id: VolumeId,
|
|
path_in: PathBuf,
|
|
write: String,
|
|
) -> Result<(), AnyError> {
|
|
let (volumes, volume_path) = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
let volume_path = ctx
|
|
.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &volume_id)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", volume_id))?;
|
|
(ctx.volumes.clone(), volume_path)
|
|
};
|
|
if volumes.readonly(&volume_id) {
|
|
bail!("Volume {} is readonly", volume_id);
|
|
}
|
|
|
|
let path_in = path_in.strip_prefix("/").unwrap_or(&path_in);
|
|
let new_file = volume_path.join(path_in);
|
|
let parent_new_file = new_file
|
|
.parent()
|
|
.ok_or_else(|| anyhow!("Expecting that file is not root"))?;
|
|
// With the volume check
|
|
if !is_subset(&volume_path, &parent_new_file).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
new_file.to_string_lossy(),
|
|
volume_path.to_string_lossy(),
|
|
);
|
|
}
|
|
let new_volume_tmp = to_tmp_path(&volume_path).map_err(|e| anyhow!("{}", e))?;
|
|
let hashed_name = {
|
|
use std::os::unix::ffi::OsStrExt;
|
|
|
|
use sha2::{Digest, Sha256};
|
|
let mut hasher = Sha256::new();
|
|
|
|
hasher.update(path_in.as_os_str().as_bytes());
|
|
let result = hasher.finalize();
|
|
format!("{:X}", result)
|
|
};
|
|
let temp_file = new_volume_tmp.join(&hashed_name);
|
|
let mut file = AtomicFile::new(&new_file, Some(&temp_file))
|
|
.await
|
|
.map_err(|e| anyhow!("{}", e))?;
|
|
file.write_all(write.as_bytes()).await?;
|
|
file.save().await.map_err(|e| anyhow!("{}", e))?;
|
|
Ok(())
|
|
}
|
|
#[op]
|
|
async fn rename(
|
|
state: Rc<RefCell<OpState>>,
|
|
src_volume: VolumeId,
|
|
src_path: PathBuf,
|
|
dst_volume: VolumeId,
|
|
dst_path: PathBuf,
|
|
) -> Result<(), AnyError> {
|
|
let (volumes, volume_path, volume_path_out) = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
let volume_path = ctx
|
|
.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &src_volume)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", src_volume))?;
|
|
let volume_path_out = ctx
|
|
.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &dst_volume)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", dst_volume))?;
|
|
(ctx.volumes.clone(), volume_path, volume_path_out)
|
|
};
|
|
if volumes.readonly(&dst_volume) {
|
|
bail!("Volume {} is readonly", dst_volume);
|
|
}
|
|
|
|
let src_path = src_path.strip_prefix("/").unwrap_or(&src_path);
|
|
let old_file = volume_path.join(src_path);
|
|
let parent_old_file = old_file
|
|
.parent()
|
|
.ok_or_else(|| anyhow!("Expecting that file is not root"))?;
|
|
// With the volume check
|
|
if !is_subset(&volume_path, &parent_old_file).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
old_file.to_string_lossy(),
|
|
volume_path.to_string_lossy(),
|
|
);
|
|
}
|
|
|
|
let dst_path = dst_path.strip_prefix("/").unwrap_or(&dst_path);
|
|
let new_file = volume_path_out.join(dst_path);
|
|
let parent_new_file = new_file
|
|
.parent()
|
|
.ok_or_else(|| anyhow!("Expecting that file is not root"))?;
|
|
// With the volume check
|
|
if !is_subset(&volume_path_out, &parent_new_file).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
new_file.to_string_lossy(),
|
|
volume_path_out.to_string_lossy(),
|
|
);
|
|
}
|
|
tokio::fs::rename(old_file, new_file).await?;
|
|
Ok(())
|
|
}
|
|
|
|
#[op]
|
|
async fn rsync(
|
|
state: Rc<RefCell<OpState>>,
|
|
src_volume: VolumeId,
|
|
src_path: PathBuf,
|
|
dst_volume: VolumeId,
|
|
dst_path: PathBuf,
|
|
options: RsyncOptions,
|
|
) -> Result<usize, AnyError> {
|
|
let (volumes, volume_path, volume_path_out, rsyncs) = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
let volume_path = ctx
|
|
.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &src_volume)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", src_volume))?;
|
|
let volume_path_out = ctx
|
|
.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &dst_volume)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", dst_volume))?;
|
|
(
|
|
ctx.volumes.clone(),
|
|
volume_path,
|
|
volume_path_out,
|
|
ctx.rsyncs.clone(),
|
|
)
|
|
};
|
|
if volumes.readonly(&dst_volume) {
|
|
bail!("Volume {} is readonly", dst_volume);
|
|
}
|
|
|
|
let src_path = src_path.strip_prefix("/").unwrap_or(&src_path);
|
|
let src = volume_path.join(src_path);
|
|
// With the volume check
|
|
if !is_subset(&volume_path, &src).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
src.to_string_lossy(),
|
|
volume_path.to_string_lossy(),
|
|
);
|
|
}
|
|
if tokio::fs::metadata(&src).await.is_err() {
|
|
bail!("Source at {} does not exists", src.to_string_lossy());
|
|
}
|
|
|
|
let dst_path = src_path.strip_prefix("/").unwrap_or(&dst_path);
|
|
let dst = volume_path_out.join(dst_path);
|
|
// With the volume check
|
|
if !is_subset(&volume_path_out, &dst).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
dst.to_string_lossy(),
|
|
volume_path_out.to_string_lossy(),
|
|
);
|
|
}
|
|
|
|
let running_rsync = Rsync::new(src, dst, options)
|
|
.await
|
|
.map_err(|e| anyhow::anyhow!("{:?}", e.source))?;
|
|
let insert_id = {
|
|
let mut rsyncs = rsyncs.lock().await;
|
|
let next = rsyncs.0 + 1;
|
|
rsyncs.0 = next;
|
|
rsyncs.1.insert(next, running_rsync);
|
|
next
|
|
};
|
|
Ok(insert_id)
|
|
}
|
|
|
|
#[op]
|
|
async fn rsync_wait(state: Rc<RefCell<OpState>>, id: usize) -> Result<(), AnyError> {
|
|
let rsyncs = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
ctx.rsyncs.clone()
|
|
};
|
|
let running_rsync = match rsyncs.lock().await.1.remove(&id) {
|
|
Some(a) => a,
|
|
None => bail!("Couldn't find rsync at id {id}"),
|
|
};
|
|
running_rsync
|
|
.wait()
|
|
.await
|
|
.map_err(|x| anyhow::anyhow!("{}", x.source))?;
|
|
Ok(())
|
|
}
|
|
#[op]
|
|
async fn rsync_progress(state: Rc<RefCell<OpState>>, id: usize) -> Result<f64, AnyError> {
|
|
use futures::StreamExt;
|
|
let rsyncs = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
ctx.rsyncs.clone()
|
|
};
|
|
let mut running_rsync = match rsyncs.lock().await.1.remove(&id) {
|
|
Some(a) => a,
|
|
None => bail!("Couldn't find rsync at id {id}"),
|
|
};
|
|
let progress = running_rsync.progress.next().await.unwrap_or_default();
|
|
rsyncs.lock().await.1.insert(id, running_rsync);
|
|
Ok(progress)
|
|
}
|
|
#[op]
|
|
async fn remove_file(
|
|
state: Rc<RefCell<OpState>>,
|
|
volume_id: VolumeId,
|
|
path_in: PathBuf,
|
|
) -> Result<(), AnyError> {
|
|
let (volumes, volume_path) = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
let volume_path = ctx
|
|
.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &volume_id)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", volume_id))?;
|
|
(ctx.volumes.clone(), volume_path)
|
|
};
|
|
if volumes.readonly(&volume_id) {
|
|
bail!("Volume {} is readonly", volume_id);
|
|
}
|
|
let path_in = path_in.strip_prefix("/").unwrap_or(&path_in);
|
|
let new_file = volume_path.join(path_in);
|
|
// With the volume check
|
|
if !is_subset(&volume_path, &new_file).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
new_file.to_string_lossy(),
|
|
volume_path.to_string_lossy(),
|
|
);
|
|
}
|
|
tokio::fs::remove_file(new_file).await?;
|
|
Ok(())
|
|
}
|
|
#[op]
|
|
async fn remove_dir(
|
|
state: Rc<RefCell<OpState>>,
|
|
volume_id: VolumeId,
|
|
path_in: PathBuf,
|
|
) -> Result<(), AnyError> {
|
|
let (volumes, volume_path) = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
let volume_path = ctx
|
|
.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &volume_id)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", volume_id))?;
|
|
(ctx.volumes.clone(), volume_path)
|
|
};
|
|
if volumes.readonly(&volume_id) {
|
|
bail!("Volume {} is readonly", volume_id);
|
|
}
|
|
let path_in = path_in.strip_prefix("/").unwrap_or(&path_in);
|
|
let new_file = volume_path.join(path_in);
|
|
// With the volume check
|
|
if !is_subset(&volume_path, &new_file).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
new_file.to_string_lossy(),
|
|
volume_path.to_string_lossy(),
|
|
);
|
|
}
|
|
tokio::fs::remove_dir_all(new_file).await?;
|
|
Ok(())
|
|
}
|
|
#[op]
|
|
async fn create_dir(
|
|
state: Rc<RefCell<OpState>>,
|
|
volume_id: VolumeId,
|
|
path_in: PathBuf,
|
|
) -> Result<(), AnyError> {
|
|
let (volumes, volume_path) = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
let volume_path = ctx
|
|
.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &volume_id)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", volume_id))?;
|
|
(ctx.volumes.clone(), volume_path)
|
|
};
|
|
if volumes.readonly(&volume_id) {
|
|
bail!("Volume {} is readonly", volume_id);
|
|
}
|
|
let path_in = path_in.strip_prefix("/").unwrap_or(&path_in);
|
|
let new_file = volume_path.join(path_in);
|
|
|
|
// With the volume check
|
|
if !is_subset(&volume_path, &new_file).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
new_file.to_string_lossy(),
|
|
volume_path.to_string_lossy(),
|
|
);
|
|
}
|
|
tokio::fs::create_dir_all(new_file).await?;
|
|
Ok(())
|
|
}
|
|
#[op]
|
|
async fn read_dir(
|
|
state: Rc<RefCell<OpState>>,
|
|
volume_id: VolumeId,
|
|
path_in: PathBuf,
|
|
) -> Result<Vec<String>, AnyError> {
|
|
let volume_path = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
ctx.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &volume_id)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", volume_id))?
|
|
};
|
|
let path_in = path_in.strip_prefix("/").unwrap_or(&path_in);
|
|
let new_file = volume_path.join(path_in);
|
|
|
|
// With the volume check
|
|
if !is_subset(&volume_path, &new_file).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
new_file.to_string_lossy(),
|
|
volume_path.to_string_lossy(),
|
|
);
|
|
}
|
|
let mut reader = tokio::fs::read_dir(&new_file).await?;
|
|
let mut paths: Vec<String> = Vec::new();
|
|
let origin_path = format!("{}/", new_file.to_str().unwrap_or_default());
|
|
let remove_new_file = |other_path: String| other_path.replacen(&origin_path, "", 1);
|
|
let has_origin_path = |other_path: &String| other_path.starts_with(&origin_path);
|
|
while let Some(entry) = reader.next_entry().await? {
|
|
entry
|
|
.path()
|
|
.to_str()
|
|
.into_iter()
|
|
.map(ToString::to_string)
|
|
.filter(&has_origin_path)
|
|
.map(&remove_new_file)
|
|
.for_each(|x| paths.push(x));
|
|
}
|
|
paths.sort();
|
|
Ok(paths)
|
|
}
|
|
|
|
#[op]
|
|
async fn disk_usage(
|
|
state: Rc<RefCell<OpState>>,
|
|
volume_id: Option<VolumeId>,
|
|
path_in: Option<PathBuf>,
|
|
) -> Result<(u64, u64), AnyError> {
|
|
let (base_path, volume_path) = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
let volume_path = if let Some(volume_id) = volume_id {
|
|
Some(
|
|
ctx.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &volume_id)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", volume_id))?,
|
|
)
|
|
} else {
|
|
None
|
|
};
|
|
(ctx.datadir.join("package-data"), volume_path)
|
|
};
|
|
let path = if let (Some(volume_path), Some(path_in)) = (volume_path, path_in) {
|
|
let path_in = path_in.strip_prefix("/").unwrap_or(&path_in);
|
|
Some(volume_path.join(path_in))
|
|
} else {
|
|
None
|
|
};
|
|
|
|
if let Some(path) = path {
|
|
let size = String::from_utf8(
|
|
Command::new("df")
|
|
.arg("--output=size")
|
|
.arg("--block-size=1")
|
|
.arg(&base_path)
|
|
.stdout(std::process::Stdio::piped())
|
|
.output()
|
|
.await?
|
|
.stdout,
|
|
)?
|
|
.lines()
|
|
.nth(1)
|
|
.unwrap_or_default()
|
|
.parse()?;
|
|
let used = String::from_utf8(
|
|
Command::new("du")
|
|
.arg("-s")
|
|
.arg("--block-size=1")
|
|
.arg(path)
|
|
.stdout(std::process::Stdio::piped())
|
|
.output()
|
|
.await?
|
|
.stdout,
|
|
)?
|
|
.split_ascii_whitespace()
|
|
.next()
|
|
.unwrap_or_default()
|
|
.parse()?;
|
|
Ok((used, size))
|
|
} else {
|
|
String::from_utf8(
|
|
Command::new("df")
|
|
.arg("--output=used,size")
|
|
.arg("--block-size=1")
|
|
.arg(&base_path)
|
|
.stdout(std::process::Stdio::piped())
|
|
.output()
|
|
.await?
|
|
.stdout,
|
|
)?
|
|
.lines()
|
|
.nth(1)
|
|
.unwrap_or_default()
|
|
.split_ascii_whitespace()
|
|
.next_tuple()
|
|
.and_then(|(used, size)| Some((used.parse().ok()?, size.parse().ok()?)))
|
|
.ok_or_else(|| anyhow!("invalid output from df"))
|
|
}
|
|
}
|
|
|
|
#[op]
|
|
fn current_function(state: &mut OpState) -> Result<String, AnyError> {
|
|
let ctx = state.borrow::<JsContext>();
|
|
Ok(ctx.run_function.clone())
|
|
}
|
|
|
|
#[op]
|
|
async fn log_trace(state: Rc<RefCell<OpState>>, input: String) -> Result<(), AnyError> {
|
|
let ctx = {
|
|
let state = state.borrow();
|
|
state.borrow::<JsContext>().clone()
|
|
};
|
|
tracing::trace!(
|
|
package_id = tracing::field::display(&ctx.package_id),
|
|
run_function = tracing::field::display(&ctx.run_function),
|
|
"{}",
|
|
input
|
|
);
|
|
Ok(())
|
|
}
|
|
#[op]
|
|
async fn log_warn(state: Rc<RefCell<OpState>>, input: String) -> Result<(), AnyError> {
|
|
let ctx = {
|
|
let state = state.borrow();
|
|
state.borrow::<JsContext>().clone()
|
|
};
|
|
tracing::warn!(
|
|
package_id = tracing::field::display(&ctx.package_id),
|
|
run_function = tracing::field::display(&ctx.run_function),
|
|
"{}",
|
|
input
|
|
);
|
|
Ok(())
|
|
}
|
|
#[op]
|
|
async fn log_error(state: Rc<RefCell<OpState>>, input: String) -> Result<(), AnyError> {
|
|
let ctx = {
|
|
let state = state.borrow();
|
|
state.borrow::<JsContext>().clone()
|
|
};
|
|
tracing::error!(
|
|
package_id = tracing::field::display(&ctx.package_id),
|
|
run_function = tracing::field::display(&ctx.run_function),
|
|
"{}",
|
|
input
|
|
);
|
|
Ok(())
|
|
}
|
|
#[op]
|
|
async fn log_debug(state: Rc<RefCell<OpState>>, input: String) -> Result<(), AnyError> {
|
|
let ctx = {
|
|
let state = state.borrow();
|
|
state.borrow::<JsContext>().clone()
|
|
};
|
|
tracing::debug!(
|
|
package_id = tracing::field::display(&ctx.package_id),
|
|
run_function = tracing::field::display(&ctx.run_function),
|
|
"{}",
|
|
input
|
|
);
|
|
Ok(())
|
|
}
|
|
#[op]
|
|
async fn log_info(state: Rc<RefCell<OpState>>, input: String) -> Result<(), AnyError> {
|
|
let (package_id, run_function) = {
|
|
let state = state.borrow();
|
|
let ctx: JsContext = state.borrow::<JsContext>().clone();
|
|
(ctx.package_id, ctx.run_function)
|
|
};
|
|
tracing::info!(
|
|
package_id = tracing::field::display(&package_id),
|
|
run_function = tracing::field::display(&run_function),
|
|
"{}",
|
|
input
|
|
);
|
|
Ok(())
|
|
}
|
|
|
|
#[op]
|
|
fn get_input(state: &mut OpState) -> Result<Value, AnyError> {
|
|
let ctx = state.borrow::<JsContext>();
|
|
Ok(ctx.input.clone())
|
|
}
|
|
#[op]
|
|
fn get_variable_args(state: &mut OpState) -> Result<Vec<Value>, AnyError> {
|
|
let ctx = state.borrow::<JsContext>();
|
|
Ok(ctx.variable_args.clone())
|
|
}
|
|
#[op]
|
|
fn set_value(state: &mut OpState, value: Value) -> Result<(), AnyError> {
|
|
let mut answer = state.borrow::<AnswerState>().0.lock();
|
|
*answer = value;
|
|
Ok(())
|
|
}
|
|
#[op]
|
|
fn is_sandboxed(state: &mut OpState) -> Result<bool, AnyError> {
|
|
let ctx = state.borrow::<JsContext>();
|
|
Ok(ctx.sandboxed)
|
|
}
|
|
|
|
#[derive(Debug, Clone, Serialize, Deserialize)]
|
|
#[serde(rename_all = "camelCase")]
|
|
pub struct StartCommand {
|
|
process_id: ProcessId,
|
|
}
|
|
|
|
#[op]
|
|
async fn sleep(time_ms: u64) -> Result<(), AnyError> {
|
|
tokio::time::sleep(Duration::from_millis(time_ms)).await;
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[op]
|
|
async fn chown(
|
|
state: Rc<RefCell<OpState>>,
|
|
volume_id: VolumeId,
|
|
path_in: PathBuf,
|
|
ownership: u32,
|
|
) -> Result<(), AnyError> {
|
|
let sandboxed = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
ctx.sandboxed
|
|
};
|
|
|
|
if sandboxed {
|
|
bail!("Will not run chown in sandboxed mode");
|
|
}
|
|
|
|
let (volumes, volume_path) = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
let volume_path = ctx
|
|
.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &volume_id)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", volume_id))?;
|
|
(ctx.volumes.clone(), volume_path)
|
|
};
|
|
if volumes.readonly(&volume_id) {
|
|
bail!("Volume {} is readonly", volume_id);
|
|
}
|
|
let path_in = path_in.strip_prefix("/").unwrap_or(&path_in);
|
|
let new_file = volume_path.join(path_in);
|
|
// With the volume check
|
|
if !is_subset(&volume_path, &new_file).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
new_file.to_string_lossy(),
|
|
volume_path.to_string_lossy(),
|
|
);
|
|
}
|
|
let output = tokio::process::Command::new("chown")
|
|
.arg("--recursive")
|
|
.arg(format!("{ownership}"))
|
|
.arg(new_file.as_os_str())
|
|
.output()
|
|
.await?;
|
|
if !output.status.success() {
|
|
return Err(anyhow!("Chown Error"));
|
|
}
|
|
Ok(())
|
|
}
|
|
#[op]
|
|
async fn chmod(
|
|
state: Rc<RefCell<OpState>>,
|
|
volume_id: VolumeId,
|
|
path_in: PathBuf,
|
|
mode: u32,
|
|
) -> Result<(), AnyError> {
|
|
let sandboxed = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
ctx.sandboxed
|
|
};
|
|
|
|
if sandboxed {
|
|
bail!("Will not run chmod in sandboxed mode");
|
|
}
|
|
|
|
let (volumes, volume_path) = {
|
|
let state = state.borrow();
|
|
let ctx: &JsContext = state.borrow();
|
|
let volume_path = ctx
|
|
.volumes
|
|
.path_for(&ctx.datadir, &ctx.package_id, &ctx.version, &volume_id)
|
|
.ok_or_else(|| anyhow!("There is no {} in volumes", volume_id))?;
|
|
(ctx.volumes.clone(), volume_path)
|
|
};
|
|
if volumes.readonly(&volume_id) {
|
|
bail!("Volume {} is readonly", volume_id);
|
|
}
|
|
let path_in = path_in.strip_prefix("/").unwrap_or(&path_in);
|
|
let new_file = volume_path.join(path_in);
|
|
// With the volume check
|
|
if !is_subset(&volume_path, &new_file).await? {
|
|
bail!(
|
|
"Path '{}' has broken away from parent '{}'",
|
|
new_file.to_string_lossy(),
|
|
volume_path.to_string_lossy(),
|
|
);
|
|
}
|
|
tokio::fs::set_permissions(new_file, Permissions::from_mode(mode)).await?;
|
|
Ok(())
|
|
}
|
|
/// We need to make sure that during the file accessing, we don't reach beyond our scope of control
|
|
async fn is_subset(
|
|
parent: impl AsRef<Path>,
|
|
child: impl AsRef<Path>,
|
|
) -> Result<bool, AnyError> {
|
|
let child = {
|
|
let mut child_count = 0;
|
|
let mut child = child.as_ref();
|
|
loop {
|
|
if child.ends_with("..") {
|
|
child_count += 1;
|
|
} else if child_count > 0 {
|
|
child_count -= 1;
|
|
} else {
|
|
let meta = tokio::fs::metadata(child).await;
|
|
if meta.is_ok() {
|
|
break;
|
|
}
|
|
}
|
|
child = match child.parent() {
|
|
Some(child) => child,
|
|
None => {
|
|
return Ok(false);
|
|
}
|
|
};
|
|
}
|
|
tokio::fs::canonicalize(child).await?
|
|
};
|
|
let parent = tokio::fs::canonicalize(parent).await?;
|
|
Ok(child.starts_with(parent))
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn test_is_subset() {
|
|
let home = std::env::var("HOME").unwrap();
|
|
let home = Path::new(&home);
|
|
assert!(!is_subset(home, &home.join("code/fakedir/../../.."))
|
|
.await
|
|
.unwrap())
|
|
}
|
|
}
|
|
|
|
fn system_time_as_unix_ms(system_time: &SystemTime) -> Option<u64> {
|
|
system_time
|
|
.duration_since(SystemTime::UNIX_EPOCH)
|
|
.ok()?
|
|
.as_millis()
|
|
.try_into()
|
|
.ok()
|
|
}
|