Added support for optionally deploying to subdir (for docs later)
parent
a3cf021fc6
commit
b4c8216ebc
|
@ -1,2 +1,3 @@
|
||||||
/target
|
/target
|
||||||
/data/
|
/data/
|
||||||
|
*.tar.gz
|
||||||
|
|
|
@ -791,6 +791,20 @@ name = "serde"
|
||||||
version = "1.0.136"
|
version = "1.0.136"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "ce31e24b01e1e524df96f1c2fdd054405f8d7376249a5110886fb4b658484789"
|
checksum = "ce31e24b01e1e524df96f1c2fdd054405f8d7376249a5110886fb4b658484789"
|
||||||
|
dependencies = [
|
||||||
|
"serde_derive",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "serde_derive"
|
||||||
|
version = "1.0.136"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "08597e7152fcd306f41838ed3e37be9eaeed2b61c42e2117266a554fab4662f9"
|
||||||
|
dependencies = [
|
||||||
|
"proc-macro2",
|
||||||
|
"quote",
|
||||||
|
"syn",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "serde_json"
|
name = "serde_json"
|
||||||
|
@ -843,6 +857,7 @@ dependencies = [
|
||||||
"hyper",
|
"hyper",
|
||||||
"metrics",
|
"metrics",
|
||||||
"metrics-exporter-prometheus",
|
"metrics-exporter-prometheus",
|
||||||
|
"serde",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
"tar",
|
"tar",
|
||||||
"tokio",
|
"tokio",
|
||||||
|
|
|
@ -25,3 +25,4 @@ uuid = { version = "1.0.0-alpha.1", features = ["v4"] }
|
||||||
serde_json = "1.0.79"
|
serde_json = "1.0.79"
|
||||||
metrics = "0.18.1"
|
metrics = "0.18.1"
|
||||||
metrics-exporter-prometheus = "0.9.0"
|
metrics-exporter-prometheus = "0.9.0"
|
||||||
|
serde = { version = "1.0", features = ["derive"] }
|
||||||
|
|
|
@ -0,0 +1,8 @@
|
||||||
|
#!/usr/bin/env sh
|
||||||
|
|
||||||
|
curl \
|
||||||
|
-XPOST \
|
||||||
|
-T test.tar.gz \
|
||||||
|
-H 'Authorization: Bearer test' \
|
||||||
|
-v \
|
||||||
|
http://localhost:3000/api/deploy?dir=docs
|
|
@ -1,19 +1,27 @@
|
||||||
use std::{collections::HashSet, io::ErrorKind, path::Path};
|
use std::{collections::HashSet, io::ErrorKind, path::Path};
|
||||||
|
|
||||||
use axum::{
|
use axum::{
|
||||||
extract::{BodyStream, Extension},
|
extract::{BodyStream, Extension, Query},
|
||||||
http::StatusCode,
|
http::StatusCode,
|
||||||
response::IntoResponse,
|
response::IntoResponse,
|
||||||
};
|
};
|
||||||
use flate2::read::GzDecoder;
|
use flate2::read::GzDecoder;
|
||||||
use futures_util::TryStreamExt;
|
use futures_util::TryStreamExt;
|
||||||
|
use serde::Deserialize;
|
||||||
use tar::Archive;
|
use tar::Archive;
|
||||||
use tokio_util::io::StreamReader;
|
use tokio_util::io::StreamReader;
|
||||||
|
|
||||||
use crate::STATIC_DIR_NAME;
|
use crate::DEFAULT_STATIC_DIR_NAME;
|
||||||
|
|
||||||
|
#[derive(Deserialize)]
|
||||||
|
pub struct StaticDirParams
|
||||||
|
{
|
||||||
|
dir: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
pub async fn post_deploy(
|
pub async fn post_deploy(
|
||||||
Extension(data_dir): Extension<String>,
|
Extension(data_dir): Extension<String>,
|
||||||
|
Query(params): Query<StaticDirParams>,
|
||||||
res: BodyStream,
|
res: BodyStream,
|
||||||
) -> impl IntoResponse
|
) -> impl IntoResponse
|
||||||
{
|
{
|
||||||
|
@ -26,46 +34,53 @@ pub async fn post_deploy(
|
||||||
let mut file = tokio::fs::File::create(&file_path).await.unwrap();
|
let mut file = tokio::fs::File::create(&file_path).await.unwrap();
|
||||||
tokio::io::copy(&mut read, &mut file).await;
|
tokio::io::copy(&mut read, &mut file).await;
|
||||||
|
|
||||||
|
let mut static_path = Path::new(&data_dir).join(DEFAULT_STATIC_DIR_NAME);
|
||||||
|
|
||||||
|
if params.dir.is_some() {
|
||||||
|
static_path = static_path.join(params.dir.unwrap());
|
||||||
|
}
|
||||||
|
|
||||||
|
// Make sure the static directory exists
|
||||||
|
tokio::fs::create_dir_all(&static_path).await;
|
||||||
|
|
||||||
|
let fp_clone = file_path.clone();
|
||||||
// Extract the contents of the tarball synchronously
|
// Extract the contents of the tarball synchronously
|
||||||
match tokio::task::spawn_blocking(move || {
|
let res =
|
||||||
let file = match std::fs::File::open(file_path) {
|
match tokio::task::spawn_blocking(move || process_archive(&fp_clone, &static_path)).await {
|
||||||
Ok(v) => v,
|
Ok(_) => StatusCode::OK,
|
||||||
Err(_) => return StatusCode::INTERNAL_SERVER_ERROR,
|
Err(_) => StatusCode::INTERNAL_SERVER_ERROR,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
// Remove archive file after use
|
||||||
|
tokio::fs::remove_file(&file_path).await;
|
||||||
|
|
||||||
|
res
|
||||||
|
}
|
||||||
|
|
||||||
|
fn process_archive(archive_path: &Path, static_dir: &Path) -> Result<(), ()>
|
||||||
|
{
|
||||||
|
let file = std::fs::File::open(archive_path).map_err(|_| ())?;
|
||||||
let tar = GzDecoder::new(file);
|
let tar = GzDecoder::new(file);
|
||||||
let mut archive = Archive::new(tar);
|
let mut archive = Archive::new(tar);
|
||||||
|
|
||||||
let mut paths = HashSet::new();
|
let mut paths = HashSet::new();
|
||||||
|
|
||||||
let entries = match archive.entries() {
|
let entries = archive.entries().map_err(|_| ())?;
|
||||||
Ok(e) => e,
|
|
||||||
Err(_) => return StatusCode::INTERNAL_SERVER_ERROR,
|
|
||||||
};
|
|
||||||
|
|
||||||
// Extract each entry into the output directory
|
// Extract each entry into the output directory
|
||||||
let static_dir = Path::new(&data_dir).join(STATIC_DIR_NAME);
|
|
||||||
for entry_res in entries {
|
for entry_res in entries {
|
||||||
if let Ok(mut entry) = entry_res {
|
let mut entry = entry_res.map_err(|_| ())?;
|
||||||
if let Err(_) = entry.unpack_in(&static_dir) {
|
entry.unpack_in(static_dir).map_err(|_| ())?;
|
||||||
return StatusCode::INTERNAL_SERVER_ERROR;
|
|
||||||
}
|
|
||||||
|
|
||||||
if let Ok(path) = entry.path() {
|
if let Ok(path) = entry.path() {
|
||||||
paths.insert(path.into_owned());
|
paths.insert(path.into_owned());
|
||||||
}
|
}
|
||||||
} else {
|
|
||||||
return StatusCode::INTERNAL_SERVER_ERROR;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Remove any old files that weren't present in new archive
|
// Remove any old files that weren't present in new archive
|
||||||
let mut items = vec![];
|
let mut items = vec![];
|
||||||
|
|
||||||
// Start by populating the vec with the initial files
|
// Start by populating the vec with the initial files
|
||||||
let iter = match static_dir.read_dir() {
|
let iter = static_dir.read_dir().map_err(|_| ())?;
|
||||||
Ok(v) => v,
|
|
||||||
Err(_) => return StatusCode::INTERNAL_SERVER_ERROR,
|
|
||||||
};
|
|
||||||
iter.filter_map(|r| r.ok())
|
iter.filter_map(|r| r.ok())
|
||||||
.for_each(|e| items.push(e.path()));
|
.for_each(|e| items.push(e.path()));
|
||||||
|
|
||||||
|
@ -86,11 +101,5 @@ pub async fn post_deploy(
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
StatusCode::OK
|
Ok(())
|
||||||
})
|
|
||||||
.await
|
|
||||||
{
|
|
||||||
Ok(s) => s,
|
|
||||||
Err(_) => StatusCode::INTERNAL_SERVER_ERROR,
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -14,7 +14,7 @@ mod api;
|
||||||
mod matrix;
|
mod matrix;
|
||||||
mod metrics;
|
mod metrics;
|
||||||
|
|
||||||
const STATIC_DIR_NAME: &str = "static";
|
const DEFAULT_STATIC_DIR_NAME: &str = "static";
|
||||||
|
|
||||||
#[tokio::main]
|
#[tokio::main]
|
||||||
async fn main()
|
async fn main()
|
||||||
|
@ -30,7 +30,7 @@ async fn main()
|
||||||
// Get required variables from env vars
|
// Get required variables from env vars
|
||||||
let api_key = std::env::var("API_KEY").expect("No API_KEY was provided.");
|
let api_key = std::env::var("API_KEY").expect("No API_KEY was provided.");
|
||||||
let data_dir = std::env::var("DATA_DIR").expect("No DATA_DIR was provided.");
|
let data_dir = std::env::var("DATA_DIR").expect("No DATA_DIR was provided.");
|
||||||
let static_dir = format!("{}/{}", data_dir, STATIC_DIR_NAME);
|
let static_dir = format!("{}/{}", data_dir, DEFAULT_STATIC_DIR_NAME);
|
||||||
|
|
||||||
std::fs::create_dir_all(&static_dir);
|
std::fs::create_dir_all(&static_dir);
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue