Compare commits

..

No commits in common. "97e42588ed6f912175b8dadfcd32034ef6df6eb9" and "27afb3496d6c1c1df4165426bfe2e8fe999b9a66" have entirely different histories.

8 changed files with 204 additions and 236 deletions

View File

@ -1,17 +1,11 @@
api_key = "test" api_key = "test"
pkg_workers = 2 port = 8000
log_level = "rieterd=debug" log_level = "tower_http=debug,rieterd=debug"
[fs] [fs]
type = "local" type = "locl"
data_dir = "./data" data_dir = "./data"
[db] [db]
type = "sqlite" type = "sqlite"
db_dir = "./data" db_dir = "./data"
# [db]
# type = "postgres"
# host = "localhost"
# db = "rieter"
# user = "rieter"
# password = "rieter"

View File

@ -1,4 +1,4 @@
use crate::{distro::MetaDistroMgr, Config, FsConfig, Global}; use crate::{distro::MetaDistroMgr, Config, Global};
use std::{io, path::PathBuf, sync::Arc}; use std::{io, path::PathBuf, sync::Arc};
@ -12,6 +12,13 @@ use tracing_subscriber::{layer::SubscriberExt, util::SubscriberInitExt};
#[derive(Parser)] #[derive(Parser)]
#[command(author, version, about, long_about = None)] #[command(author, version, about, long_about = None)]
pub struct Cli { pub struct Cli {
/// Directory where repository metadata & SQLite database is stored
#[arg(env = "RIETER_DATA_DIR")]
pub data_dir: PathBuf,
/// API key to authenticate private routes with
#[arg(env = "RIETER_API_KEY")]
pub api_key: String,
#[arg( #[arg(
short, short,
long, long,
@ -19,54 +26,89 @@ pub struct Cli {
default_value = "./rieterd.toml" default_value = "./rieterd.toml"
)] )]
pub config_file: PathBuf, pub config_file: PathBuf,
/// Database connection URL; either sqlite:// or postgres://. Defaults to rieter.sqlite in the
/// data directory
#[arg(short, long, env = "RIETER_DATABASE_URL")]
pub database_url: Option<String>,
/// Port the server will listen on
#[arg(
short,
long,
value_name = "PORT",
default_value_t = 8000,
env = "RIETER_PORT"
)]
pub port: u16,
/// Log levels for the tracing
#[arg(
long,
value_name = "LOG_LEVEL",
default_value = "tower_http=debug,rieterd=debug,sea_orm=debug",
env = "RIETER_LOG"
)]
pub log: String,
} }
impl Cli { impl Cli {
pub async fn run(&self) -> crate::Result<()> { pub fn init_tracing(&self) {
let config: Config = Config::figment(&self.config_file)
.extract()
.inspect_err(|e| tracing::error!("{}", e))?;
tracing_subscriber::registry() tracing_subscriber::registry()
.with(tracing_subscriber::EnvFilter::new(config.log_level.clone())) .with(tracing_subscriber::EnvFilter::new(self.log.clone()))
.with(tracing_subscriber::fmt::layer()) .with(tracing_subscriber::fmt::layer())
.init(); .init();
}
tracing::info!("Connecting to database"); pub async fn run(&self) -> crate::Result<()> {
let db = crate::db::connect(&config.db).await?; self.init_tracing();
crate::db::Migrator::up(&db, None).await?; //tracing::debug!("{:?}", &self.config_file);
//let new_config: crate::config::Config = crate::config::Config::figment(&self.config_file).extract().inspect_err(
// |e| tracing::error!("{}", e)
//)?;
//tracing::debug!("{:?}", new_config);
let mgr = match &config.fs { let db_url = if let Some(url) = &self.database_url {
FsConfig::Local { data_dir } => { url.clone()
crate::repo::RepoMgr::new(data_dir.join("repos"), db.clone()).await? } else {
} format!(
"sqlite://{}?mode=rwc",
self.data_dir.join("rieter.sqlite").to_string_lossy()
)
}; };
let mgr = Arc::new(mgr); debug!("Connecting to database with URL {}", db_url);
for _ in 0..config.pkg_workers { let mut options = sea_orm::ConnectOptions::new(db_url);
options.max_connections(16);
let db = sea_orm::Database::connect(options).await?;
crate::db::Migrator::up(&db, None).await?;
debug!("Successfully applied migrations");
let config = Config {
data_dir: self.data_dir.clone(),
};
let mgr =
Arc::new(crate::repo::RepoMgr::new(&self.data_dir.join("repos"), db.clone()).await?);
for _ in 0..1 {
let clone = Arc::clone(&mgr); let clone = Arc::clone(&mgr);
tokio::spawn(async move { clone.pkg_parse_task().await }); tokio::spawn(async move { clone.pkg_parse_task().await });
} }
let global = Global { let global = Global { config, mgr, db };
config: config.clone(),
mgr,
db,
};
// build our application with a single route // build our application with a single route
let app = Router::new() let app = Router::new()
.nest("/api", crate::api::router()) .nest("/api", crate::api::router())
.merge(crate::repo::router(&config.api_key)) .merge(crate::repo::router(&self.api_key))
.with_state(global) .with_state(global)
.layer(TraceLayer::new_for_http()); .layer(TraceLayer::new_for_http());
let domain: String = format!("{}:{}", config.domain, config.port) let domain: String = format!("0.0.0.0:{}", self.port).parse().unwrap();
.parse()
.unwrap();
let listener = tokio::net::TcpListener::bind(domain).await?; let listener = tokio::net::TcpListener::bind(domain).await?;
// run it with hyper on localhost:3000 // run it with hyper on localhost:3000
Ok(axum::serve(listener, app.into_make_service()) Ok(axum::serve(listener, app.into_make_service())

View File

@ -6,49 +6,34 @@ use figment::{
}; };
use serde::Deserialize; use serde::Deserialize;
#[derive(Deserialize, Debug, Clone)] #[derive(Deserialize, Debug)]
#[serde(rename_all = "lowercase")] #[serde(rename_all = "lowercase")]
#[serde(tag = "type")] #[serde(tag = "type")]
pub enum FsConfig { pub enum FsConfig {
Local { data_dir: PathBuf }, Local { data_dir: PathBuf },
} }
#[derive(Deserialize, Debug, Clone)] #[derive(Deserialize, Debug)]
#[serde(rename_all = "lowercase")] #[serde(rename_all = "lowercase")]
#[serde(tag = "type")] #[serde(tag = "type")]
pub enum DbConfig { pub enum DbConfig {
Sqlite { Sqlite {
db_dir: PathBuf, db_dir: PathBuf,
#[serde(default = "default_db_sqlite_max_connections")]
max_connections: u32,
}, },
Postgres { Postgres {
host: String, host: String,
#[serde(default = "default_db_postgres_port")]
port: u16,
user: String, user: String,
password: String, password: String,
db: String,
#[serde(default)]
schema: String,
#[serde(default = "default_db_postgres_max_connections")]
max_connections: u32,
}, },
} }
#[derive(Deserialize, Debug, Clone)] #[derive(Deserialize, Debug)]
pub struct Config { pub struct Config {
pub api_key: String, api_key: String,
#[serde(default = "default_domain")] port: u16,
pub domain: String, log_level: String,
#[serde(default = "default_port")] fs: FsConfig,
pub port: u16, db: DbConfig,
#[serde(default = "default_log_level")]
pub log_level: String,
pub fs: FsConfig,
pub db: DbConfig,
#[serde(default = "default_pkg_workers")]
pub pkg_workers: u32,
} }
impl Config { impl Config {
@ -58,31 +43,3 @@ impl Config {
.merge(Env::prefixed("RIETER_")) .merge(Env::prefixed("RIETER_"))
} }
} }
fn default_domain() -> String {
String::from("0.0.0.0")
}
fn default_port() -> u16 {
8000
}
fn default_log_level() -> String {
String::from("tower_http=debug,rieterd=debug,sea_orm=debug")
}
fn default_db_sqlite_max_connections() -> u32 {
16
}
fn default_db_postgres_port() -> u16 {
5432
}
fn default_db_postgres_max_connections() -> u32 {
16
}
fn default_pkg_workers() -> u32 {
1
}

View File

@ -2,12 +2,10 @@ pub mod entities;
mod migrator; mod migrator;
pub mod query; pub mod query;
use crate::config::DbConfig;
pub use entities::{prelude::*, *}; pub use entities::{prelude::*, *};
pub use migrator::Migrator; pub use migrator::Migrator;
use sea_orm::{ConnectionTrait, Database, DbConn, DeriveActiveEnum, EnumIter}; use sea_orm::{DeriveActiveEnum, EnumIter};
use serde::{Deserialize, Serialize}; use serde::{Deserialize, Serialize};
type Result<T> = std::result::Result<T, sea_orm::DbErr>; type Result<T> = std::result::Result<T, sea_orm::DbErr>;
@ -52,50 +50,3 @@ pub struct FullPackage {
related: Vec<(PackageRelatedEnum, String)>, related: Vec<(PackageRelatedEnum, String)>,
files: Vec<String>, files: Vec<String>,
} }
pub async fn connect(conn: &DbConfig) -> crate::Result<DbConn> {
match conn {
DbConfig::Sqlite {
db_dir,
max_connections,
} => {
let url = format!(
"sqlite://{}?mode=rwc",
db_dir.join("rieter.sqlite").to_string_lossy()
);
let options = sea_orm::ConnectOptions::new(url)
.max_connections(*max_connections)
.to_owned();
let conn = Database::connect(options).await?;
// synchronous=NORMAL still ensures database consistency with WAL mode, as per the docs
// https://www.sqlite.org/pragma.html#pragma_synchronous
conn.execute_unprepared("PRAGMA journal_mode=WAL;").await?;
conn.execute_unprepared("PRAGMA synchronous=NORMAL;")
.await?;
Ok(conn)
}
DbConfig::Postgres {
host,
port,
db,
user,
password,
schema,
max_connections,
} => {
let mut url = format!("postgres://{}:{}@{}:{}/{}", user, password, host, port, db);
if schema != "" {
url = format!("{url}?currentSchema={schema}");
}
let options = sea_orm::ConnectOptions::new(url)
.max_connections(*max_connections)
.to_owned();
Ok(Database::connect(options).await?)
}
}
}

View File

@ -2,7 +2,7 @@ use crate::db::{self, *};
use futures::Stream; use futures::Stream;
use sea_orm::{sea_query::IntoCondition, *}; use sea_orm::{sea_query::IntoCondition, *};
use sea_query::{Alias, Asterisk, Expr, IntoColumnRef, Query, SelectStatement}; use sea_query::{Alias, Asterisk, Expr, Query, SelectStatement};
use serde::Deserialize; use serde::Deserialize;
#[derive(Deserialize)] #[derive(Deserialize)]
@ -222,8 +222,8 @@ pub struct PkgToRemove {
pub id: i32, pub id: i32,
} }
fn max_pkg_ids_query(committed: bool) -> SelectStatement { fn max_pkg_ids_query() -> SelectStatement {
let mut query = Query::select() Query::select()
.from(db::package::Entity) .from(db::package::Entity)
.columns([ .columns([
db::package::Column::RepoId, db::package::Column::RepoId,
@ -236,29 +236,39 @@ fn max_pkg_ids_query(committed: bool) -> SelectStatement {
db::package::Column::Arch, db::package::Column::Arch,
db::package::Column::Name, db::package::Column::Name,
]) ])
.to_owned(); .cond_where(
Condition::all().add(db::package::Column::State.eq(db::PackageState::Committed)),
if committed { )
query.cond_where(db::package::Column::State.eq(db::PackageState::Committed)); .to_owned()
}
query
} }
/// Query that returns all packages that should be included in a sync for the given repository and
/// arch.
pub fn pkgs_to_sync( pub fn pkgs_to_sync(
conn: &DbConn, conn: &DbConn,
repo: i32, repo: i32,
arch: &str, arch: &str,
) -> SelectorRaw<SelectModel<package::Model>> { ) -> SelectorRaw<SelectModel<package::Model>> {
let max_id_query = Query::select()
.columns([
db::package::Column::RepoId,
db::package::Column::Arch,
db::package::Column::Name,
])
.expr_as(db::package::Column::Id.max(), Alias::new("max_id"))
.from(db::package::Entity)
.group_by_columns([
db::package::Column::RepoId,
db::package::Column::Arch,
db::package::Column::Name,
])
.to_owned();
let (p1, p2) = (Alias::new("p1"), Alias::new("p2")); let (p1, p2) = (Alias::new("p1"), Alias::new("p2"));
let query = Query::select() let query = Query::select()
.columns(db::package::Column::iter().map(|c| (p1.clone(), c))) .column((p1.clone(), Asterisk))
.from_as(db::package::Entity, p1.clone()) .from_as(db::package::Entity, p1.clone())
.join_subquery( .join_subquery(
JoinType::InnerJoin, JoinType::InnerJoin,
max_pkg_ids_query(false), max_id_query,
p2.clone(), p2.clone(),
Expr::col((p1.clone(), db::package::Column::Id)) Expr::col((p1.clone(), db::package::Column::Id))
.eq(Expr::col((p2.clone(), Alias::new("max_id")))), .eq(Expr::col((p2.clone(), Alias::new("max_id")))),
@ -266,13 +276,13 @@ pub fn pkgs_to_sync(
.cond_where( .cond_where(
Condition::all() Condition::all()
.add(Expr::col((p1.clone(), db::package::Column::RepoId)).eq(repo)) .add(Expr::col((p1.clone(), db::package::Column::RepoId)).eq(repo))
.add(
Expr::col((p1.clone(), db::package::Column::Arch))
.is_in([arch, crate::ANY_ARCH]),
)
.add( .add(
Expr::col((p1.clone(), db::package::Column::State)) Expr::col((p1.clone(), db::package::Column::State))
.ne(db::PackageState::PendingDeletion), .ne(db::PackageState::PendingDeletion),
)
.add(
Expr::col((p1.clone(), db::package::Column::Arch))
.is_in([arch, crate::ANY_ARCH]),
), ),
) )
.to_owned(); .to_owned();
@ -283,10 +293,36 @@ pub fn pkgs_to_sync(
} }
fn stale_pkgs_query(include_repo: bool) -> SelectStatement { fn stale_pkgs_query(include_repo: bool) -> SelectStatement {
// In each repository, only one version of a package can exist for any given arch. Because ids
// are monotonically increasing, we know that the row that represents the actual package
// currently in the repository is the row with the largest id whose state is "committed". This
// query finds this id for each (repo, arch, name) tuple.
let mut max_id_query = Query::select();
max_id_query
.from(db::package::Entity)
.columns([
db::package::Column::RepoId,
db::package::Column::Arch,
db::package::Column::Name,
])
.expr_as(db::package::Column::Id.max(), Alias::new("max_id"))
.group_by_columns([
db::package::Column::RepoId,
db::package::Column::Arch,
db::package::Column::Name,
])
.cond_where(
Condition::all().add(db::package::Column::State.eq(db::PackageState::Committed)),
);
let (p1, p2) = (Alias::new("p1"), Alias::new("p2")); let (p1, p2) = (Alias::new("p1"), Alias::new("p2"));
let mut query = Query::select() let mut query = Query::select();
.from_as(db::package::Entity, p1.clone())
.to_owned(); // We then perform an inner join between the max id query above and the package table, where we
// filter on rows whose id is less than their respective package's max id or whose state is set
// to "pending deletion". This gives us all rows in the database that correspond to packages
// that are no longer needed, and can thus be removed.
query.from_as(db::package::Entity, p1.clone());
if include_repo { if include_repo {
query.columns([ query.columns([
@ -297,13 +333,10 @@ fn stale_pkgs_query(include_repo: bool) -> SelectStatement {
query.column((p1.clone(), db::package::Column::Id)); query.column((p1.clone(), db::package::Column::Id));
} }
// We left join on the max pkgs query because a repository that has all its packages set to
// "pending deletion" doesn't show up in the query. These are also included with a where clause
// on the joined rows.
query query
.join_subquery( .join_subquery(
JoinType::LeftJoin, JoinType::InnerJoin,
max_pkg_ids_query(true), max_id_query,
p2.clone(), p2.clone(),
Condition::all() Condition::all()
.add( .add(
@ -326,12 +359,11 @@ fn stale_pkgs_query(include_repo: bool) -> SelectStatement {
.lt(Expr::col((p2.clone(), Alias::new("max_id")))), .lt(Expr::col((p2.clone(), Alias::new("max_id")))),
) )
.add( .add(
Expr::col((p1.clone(), db::package::Column::State)) Expr::col((p1.clone(), db::package::Column::Id))
.eq(db::PackageState::PendingDeletion), .eq(db::PackageState::PendingDeletion),
), ),
); )
.to_owned()
query
} }
pub fn stale_pkgs(conn: &DbConn) -> SelectorRaw<SelectModel<PkgToRemove>> { pub fn stale_pkgs(conn: &DbConn) -> SelectorRaw<SelectModel<PkgToRemove>> {

View File

@ -6,7 +6,6 @@ mod distro;
mod error; mod error;
mod repo; mod repo;
pub use config::{Config, DbConfig, FsConfig};
pub use error::{Result, ServerError}; pub use error::{Result, ServerError};
use repo::DistroMgr; use repo::DistroMgr;
@ -15,9 +14,14 @@ use std::{path::PathBuf, sync::Arc};
pub const ANY_ARCH: &'static str = "any"; pub const ANY_ARCH: &'static str = "any";
#[derive(Clone)]
pub struct Config {
data_dir: PathBuf,
}
#[derive(Clone)] #[derive(Clone)]
pub struct Global { pub struct Global {
config: crate::config::Config, config: Config,
mgr: Arc<repo::RepoMgr>, mgr: Arc<repo::RepoMgr>,
db: sea_orm::DbConn, db: sea_orm::DbConn,
} }

View File

@ -100,6 +100,42 @@ impl RepoMgr {
Ok(()) Ok(())
} }
/// Clean any remaining old package files from the database and file system
pub async fn remove_stale_pkgs(&self) -> crate::Result<()> {
let mut pkgs = db::query::package::stale_pkgs(&self.conn)
.stream(&self.conn)
.await?;
// Ids are monotonically increasing, so the max id suffices to know which packages to
// remove later
let mut max_id = -1;
let mut removed_pkgs = 0;
while let Some(pkg) = pkgs.next().await.transpose()? {
// Failing to remove the package file isn't the biggest problem
let _ = tokio::fs::remove_file(
self.repos_dir
.join(pkg.repo_id.to_string())
.join(pkg.id.to_string()),
)
.await;
if pkg.id > max_id {
max_id = pkg.id;
}
removed_pkgs += 1;
}
if removed_pkgs > 0 {
db::query::package::delete_stale_pkgs(&self.conn, max_id).await?;
}
tracing::info!("Removed {removed_pkgs} stale package(s)");
Ok(())
}
/// Generate the archive databases for the given repository and architecture. /// Generate the archive databases for the given repository and architecture.
async fn generate_archives(&self, repo: i32, arch: &str) -> crate::Result<()> { async fn generate_archives(&self, repo: i32, arch: &str) -> crate::Result<()> {
let [tmp_ar_db_path, tmp_ar_files_path, files_tmp_file_path, desc_tmp_file_path] = let [tmp_ar_db_path, tmp_ar_files_path, files_tmp_file_path, desc_tmp_file_path] =
@ -173,42 +209,6 @@ impl RepoMgr {
Ok(()) Ok(())
} }
/// Clean any remaining old package files from the database and file system
pub async fn remove_stale_pkgs(&self) -> crate::Result<()> {
let mut pkgs = db::query::package::stale_pkgs(&self.conn)
.stream(&self.conn)
.await?;
// Ids are monotonically increasing, so the max id suffices to know which packages to
// remove later
let mut max_id = -1;
let mut removed_pkgs = 0;
while let Some(pkg) = pkgs.next().await.transpose()? {
// Failing to remove the package file isn't the biggest problem
let _ = tokio::fs::remove_file(
self.repos_dir
.join(pkg.repo_id.to_string())
.join(pkg.id.to_string()),
)
.await;
if pkg.id > max_id {
max_id = pkg.id;
}
removed_pkgs += 1;
}
if removed_pkgs > 0 {
db::query::package::delete_stale_pkgs(&self.conn, max_id).await?;
}
tracing::info!("Removed {removed_pkgs} stale package(s)");
Ok(())
}
pub async fn pkg_parse_task(&self) { pub async fn pkg_parse_task(&self) {
loop { loop {
// Receive the next message and immediately drop the mutex afterwards. As long as the // Receive the next message and immediately drop the mutex afterwards. As long as the
@ -248,7 +248,7 @@ impl RepoMgr {
} }
pub async fn queue_pkg(&self, repo: i32, path: PathBuf) { pub async fn queue_pkg(&self, repo: i32, path: PathBuf) {
self.pkg_queue.0.send(PkgQueueMsg { path, repo }).unwrap(); let _ = self.pkg_queue.0.send(PkgQueueMsg { path, repo });
self.repos.read().await.get(&repo).inspect(|n| { self.repos.read().await.get(&repo).inspect(|n| {
n.0.fetch_add(1, Ordering::SeqCst); n.0.fetch_add(1, Ordering::SeqCst);
}); });
@ -291,7 +291,6 @@ impl RepoMgr {
}; };
let repo_id: Option<i32> = db::Repo::find() let repo_id: Option<i32> = db::Repo::find()
.filter(db::repo::Column::DistroId.eq(distro_id))
.filter(db::repo::Column::Name.eq(repo)) .filter(db::repo::Column::Name.eq(repo))
.select_only() .select_only()
.column(db::repo::Column::Id) .column(db::repo::Column::Id)

View File

@ -6,8 +6,6 @@ pub mod package;
pub use manager::DistroMgr; pub use manager::DistroMgr;
pub use manager2::RepoMgr; pub use manager2::RepoMgr;
use crate::FsConfig;
use axum::{ use axum::{
body::Body, body::Body,
extract::{Path, State}, extract::{Path, State},
@ -52,26 +50,25 @@ async fn get_file(
req: Request<Body>, req: Request<Body>,
) -> crate::Result<impl IntoResponse> { ) -> crate::Result<impl IntoResponse> {
if let Some(repo_id) = global.mgr.get_repo(&distro, &repo).await? { if let Some(repo_id) = global.mgr.get_repo(&distro, &repo).await? {
match global.config.fs { let repo_dir = global
FsConfig::Local { data_dir } => { .config
let repo_dir = data_dir.join("repos").join(repo_id.to_string()); .data_dir
.join("repos")
.join(repo_id.to_string());
let file_name = if file_name == format!("{}.db", repo) let file_name =
|| file_name == format!("{}.db.tar.gz", repo) if file_name == format!("{}.db", repo) || file_name == format!("{}.db.tar.gz", repo) {
{ format!("{}.db.tar.gz", arch)
format!("{}.db.tar.gz", arch) } else if file_name == format!("{}.files", repo)
} else if file_name == format!("{}.files", repo) || file_name == format!("{}.files.tar.gz", repo)
|| file_name == format!("{}.files.tar.gz", repo) {
{ format!("{}.files.tar.gz", arch)
format!("{}.files.tar.gz", arch) } else {
} else { file_name
file_name };
};
let path = repo_dir.join(file_name); let path = repo_dir.join(file_name);
Ok(ServeFile::new(path).oneshot(req).await) Ok(ServeFile::new(path).oneshot(req).await)
}
}
} else { } else {
Err(StatusCode::NOT_FOUND.into()) Err(StatusCode::NOT_FOUND.into())
} }
@ -81,7 +78,7 @@ async fn post_package_archive(
State(global): State<crate::Global>, State(global): State<crate::Global>,
Path((distro, repo)): Path<(String, String)>, Path((distro, repo)): Path<(String, String)>,
body: Body, body: Body,
) -> crate::Result<StatusCode> { ) -> crate::Result<()> {
let mut body = StreamReader::new(body.into_data_stream().map_err(std::io::Error::other)); let mut body = StreamReader::new(body.into_data_stream().map_err(std::io::Error::other));
let repo = global.mgr.get_or_create_repo(&distro, &repo).await?; let repo = global.mgr.get_or_create_repo(&distro, &repo).await?;
let [tmp_path] = global.mgr.random_file_paths(); let [tmp_path] = global.mgr.random_file_paths();
@ -91,7 +88,7 @@ async fn post_package_archive(
global.mgr.queue_pkg(repo, tmp_path).await; global.mgr.queue_pkg(repo, tmp_path).await;
Ok(StatusCode::ACCEPTED) Ok(())
} }
async fn delete_repo( async fn delete_repo(
@ -113,15 +110,7 @@ async fn delete_arch_repo(
State(global): State<crate::Global>, State(global): State<crate::Global>,
Path((distro, repo, arch)): Path<(String, String, String)>, Path((distro, repo, arch)): Path<(String, String, String)>,
) -> crate::Result<StatusCode> { ) -> crate::Result<StatusCode> {
if let Some(repo) = global.mgr.get_repo(&distro, &repo).await? { Ok(StatusCode::NOT_FOUND)
global.mgr.remove_repo_arch(repo, &arch).await?;
tracing::info!("Removed architecture '{arch}' from repository {repo}");
Ok(StatusCode::OK)
} else {
Ok(StatusCode::NOT_FOUND)
}
//if let Some(mgr) = global.mgr.get_mgr(&distro).await { //if let Some(mgr) = global.mgr.get_mgr(&distro).await {
// let repo_removed = mgr.remove_repo_arch(&repo, &arch).await?; // let repo_removed = mgr.remove_repo_arch(&repo, &arch).await?;
// //