Compare commits
45 Commits
incrementa
...
main
Author | SHA1 | Date |
---|---|---|
Jef Roosens | d23227dd0b | |
Jef Roosens | d3cb29b52e | |
Jef Roosens | 3cddea19c3 | |
Jef Roosens | 9ce2417528 | |
Jef Roosens | f2e781dd5a | |
Jef Roosens | 5bdd4e21b0 | |
Jef Roosens | 8f190c489b | |
Jef Roosens | 5f6366078c | |
Jef Roosens | b3d1cec078 | |
Jef Roosens | a51ff3937d | |
Jef Roosens | db3bba5a42 | |
Jef Roosens | 34d016fd3f | |
Jef Roosens | bf83357464 | |
Jef Roosens | bfb264e823 | |
Jef Roosens | 241bb4d68e | |
Jef Roosens | 6cdc18742e | |
Jef Roosens | b924a054a6 | |
Jef Roosens | 32d923e64b | |
Jef Roosens | 1acfc9c422 | |
Jef Roosens | fc8e8d37d3 | |
Jef Roosens | 5567323473 | |
Jef Roosens | 80b814bcff | |
Jef Roosens | 4ec336eb86 | |
Jef Roosens | 6e216aa88f | |
Jef Roosens | 75e9d7a9d2 | |
Jef Roosens | e6fa8a0eeb | |
Jef Roosens | 55c5f24937 | |
Jef Roosens | 36c441b8c2 | |
Jef Roosens | f71db90922 | |
Jef Roosens | 2c256cf904 | |
Jef Roosens | bfd278abbe | |
Jef Roosens | c5193f0f3c | |
Jef Roosens | a4a03ca4c5 | |
Jef Roosens | 5159bfdddd | |
Jef Roosens | 0eda768c03 | |
Jef Roosens | a4e2a1276f | |
Jef Roosens | e373fc85f1 | |
Jef Roosens | 1cfe13674d | |
Jef Roosens | d5cea49c8b | |
Jef Roosens | 03e21fda87 | |
Jef Roosens | 29636ffcdb | |
Jef Roosens | a236c36a4f | |
Jef Roosens | 0a459ee30b | |
Jef Roosens | 4e8d0a8d25 | |
Jef Roosens | 188fb30343 |
|
@ -1,3 +1,3 @@
|
||||||
[alias]
|
[alias]
|
||||||
runs = "run -- run paper 1.19.4-550 --config data/config --backup data/backups --world data/worlds --jar paper-1.19.4-550.jar"
|
runs = "run -- --config data/config --backup data/backups --world data/worlds --layers 2min,2,4,4;3min,3,2,2"
|
||||||
runrs = "run --release -- paper 1.19.4-545 --config data/config --backup data/backups --world data/worlds --jar data/paper-1.19.4-525.jar"
|
runrs = "run --release -- --config data/config --backup data/backups --world data/worlds --layers 2min,2,4,4;3min,3,2,2"
|
||||||
|
|
|
@ -5,17 +5,17 @@ matrix:
|
||||||
|
|
||||||
platform: "linux/${ARCH}"
|
platform: "linux/${ARCH}"
|
||||||
|
|
||||||
branches:
|
when:
|
||||||
|
branch:
|
||||||
exclude: [main]
|
exclude: [main]
|
||||||
|
event: push
|
||||||
|
|
||||||
pipeline:
|
steps:
|
||||||
build:
|
build:
|
||||||
image: 'rust:1.70-alpine3.18'
|
image: 'rust:1.71-alpine3.18'
|
||||||
commands:
|
commands:
|
||||||
- apk add --no-cache build-base
|
- apk add --no-cache build-base
|
||||||
- cargo build --verbose
|
- cargo build --verbose
|
||||||
- cargo test --verbose
|
- cargo test --verbose
|
||||||
# Binaries, even debug ones, should be statically compiled
|
# Binaries, even debug ones, should be statically compiled
|
||||||
- '[ "$(readelf -d target/debug/alex | grep NEEDED | wc -l)" = 0 ]'
|
- '[ "$(readelf -d target/debug/alex | grep NEEDED | wc -l)" = 0 ]'
|
||||||
when:
|
|
||||||
event: [push]
|
|
||||||
|
|
|
@ -1,13 +1,13 @@
|
||||||
platform: 'linux/amd64'
|
platform: 'linux/amd64'
|
||||||
|
|
||||||
branches:
|
when:
|
||||||
|
branch:
|
||||||
exclude: [ main ]
|
exclude: [ main ]
|
||||||
|
event: push
|
||||||
|
|
||||||
pipeline:
|
steps:
|
||||||
clippy:
|
clippy:
|
||||||
image: 'rust:1.70'
|
image: 'rust:1.71'
|
||||||
commands:
|
commands:
|
||||||
- rustup component add clippy
|
- rustup component add clippy
|
||||||
- cargo clippy -- --no-deps -Dwarnings
|
- cargo clippy -- --no-deps -Dwarnings
|
||||||
when:
|
|
||||||
event: [push]
|
|
||||||
|
|
|
@ -1,13 +1,13 @@
|
||||||
platform: 'linux/amd64'
|
platform: 'linux/amd64'
|
||||||
|
|
||||||
branches:
|
when:
|
||||||
|
branch:
|
||||||
exclude: [ main ]
|
exclude: [ main ]
|
||||||
|
event: push
|
||||||
|
|
||||||
pipeline:
|
steps:
|
||||||
lint:
|
lint:
|
||||||
image: 'rust:1.70'
|
image: 'rust:1.71'
|
||||||
commands:
|
commands:
|
||||||
- rustup component add rustfmt
|
- rustup component add rustfmt
|
||||||
- cargo fmt -- --check
|
- cargo fmt -- --check
|
||||||
when:
|
|
||||||
event: [push]
|
|
||||||
|
|
|
@ -4,19 +4,19 @@ matrix:
|
||||||
- 'linux/arm64'
|
- 'linux/arm64'
|
||||||
|
|
||||||
platform: ${PLATFORM}
|
platform: ${PLATFORM}
|
||||||
branches: [ main ]
|
|
||||||
|
|
||||||
pipeline:
|
when:
|
||||||
|
event: tag
|
||||||
|
|
||||||
|
steps:
|
||||||
build:
|
build:
|
||||||
image: 'rust:1.70-alpine3.18'
|
image: 'rust:1.71-alpine3.18'
|
||||||
commands:
|
commands:
|
||||||
- apk add --no-cache build-base
|
- apk add --no-cache build-base
|
||||||
- cargo build --release --verbose
|
- cargo build --release --verbose
|
||||||
# Ensure the release binary is also statically compiled
|
# Ensure the release binary is also statically compiled
|
||||||
- '[ "$(readelf -d target/release/alex | grep NEEDED | wc -l)" = 0 ]'
|
- '[ "$(readelf -d target/release/alex | grep NEEDED | wc -l)" = 0 ]'
|
||||||
- du -h target/release/alex
|
- du -h target/release/alex
|
||||||
when:
|
|
||||||
event: tag
|
|
||||||
|
|
||||||
publish:
|
publish:
|
||||||
image: 'curlimages/curl'
|
image: 'curlimages/curl'
|
||||||
|
@ -28,5 +28,3 @@ pipeline:
|
||||||
--user "Chewing_Bever:$GITEA_PASSWORD"
|
--user "Chewing_Bever:$GITEA_PASSWORD"
|
||||||
--upload-file target/release/alex
|
--upload-file target/release/alex
|
||||||
https://git.rustybever.be/api/packages/Chewing_Bever/generic/alex/"${CI_COMMIT_TAG}"/alex-"$(echo '${PLATFORM}' | sed 's:/:-:g')"
|
https://git.rustybever.be/api/packages/Chewing_Bever/generic/alex/"${CI_COMMIT_TAG}"/alex-"$(echo '${PLATFORM}' | sed 's:/:-:g')"
|
||||||
when:
|
|
||||||
event: tag
|
|
||||||
|
|
41
CHANGELOG.md
41
CHANGELOG.md
|
@ -7,14 +7,49 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
|
||||||
|
|
||||||
## [Unreleased](https://git.rustybever.be/Chewing_Bever/alex/src/branch/dev)
|
## [Unreleased](https://git.rustybever.be/Chewing_Bever/alex/src/branch/dev)
|
||||||
|
|
||||||
|
## [0.4.1](https://git.rustybever.be/Chewing_Bever/alex/src/tag/0.4.1)
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
|
||||||
|
* Moved PKGBUILD to separate repo
|
||||||
|
* Properly update lock file
|
||||||
|
|
||||||
|
## [0.4.0](https://git.rustybever.be/Chewing_Bever/alex/src/tag/0.4.0)
|
||||||
|
|
||||||
|
### Added
|
||||||
|
|
||||||
|
* Extract command for working with the output of export
|
||||||
|
* Arch packages are now published to my bur repo
|
||||||
|
* Allow passing configuration variables from TOML file
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
|
||||||
|
* Export command no longer reads backups that do not contribute to the final
|
||||||
|
state
|
||||||
|
* Running backups no longer block stdin input or shutdown
|
||||||
|
* Env vars `ALEX_CONFIG_DIR`, `ALEX_WORLD_DIR` and `ALEX_BACKUP_DIR` renamed to
|
||||||
|
`ALEX_CONFIG`, `ALEX_WORLD` and `ALEX_BACKUP` respectively
|
||||||
|
|
||||||
|
## [0.3.1](https://git.rustybever.be/Chewing_Bever/alex/src/tag/0.3.1)
|
||||||
|
|
||||||
|
### Added
|
||||||
|
|
||||||
|
* Export command to export any backup as a new full backup
|
||||||
|
|
||||||
|
## [0.3.0](https://git.rustybever.be/Chewing_Bever/alex/src/tag/0.3.0)
|
||||||
|
|
||||||
### Added
|
### Added
|
||||||
|
|
||||||
* `backup` CLI command
|
|
||||||
* Incremental backups
|
* Incremental backups
|
||||||
* Chain length descibres how many incremental backups to create from the
|
* Chain length describes how many incremental backups to create from the
|
||||||
same full backup
|
same full backup
|
||||||
* "backups to keep" has been replaced by "chains to keep"
|
* "backups to keep" has been replaced by "chains to keep"
|
||||||
* Server type & version is now stored as metadata in the metadata file
|
* Server type & version and backup size are now stored as metadata in the
|
||||||
|
metadata file
|
||||||
|
* Backup layers
|
||||||
|
* Store multiple chains of backups in parallel, configuring each with
|
||||||
|
different parameters (son-father-grandfather principle)
|
||||||
|
* CLI commands for creating, restoring & listing backups
|
||||||
|
|
||||||
### Changed
|
### Changed
|
||||||
|
|
||||||
|
|
|
@ -10,10 +10,11 @@ checksum = "f26201604c87b1e01bd3d98f8d5d9a8fcbb815e8cedb41ffccbeb4bf593a35fe"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "alex"
|
name = "alex"
|
||||||
version = "0.2.2"
|
version = "0.4.1"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"chrono",
|
"chrono",
|
||||||
"clap",
|
"clap",
|
||||||
|
"figment",
|
||||||
"flate2",
|
"flate2",
|
||||||
"serde",
|
"serde",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
|
@ -53,15 +54,15 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "anstyle"
|
name = "anstyle"
|
||||||
version = "1.0.0"
|
version = "1.0.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "41ed9a86bf92ae6580e0a31281f65a1b1d867c0cc68d5346e2ae128dddfa6a7d"
|
checksum = "3a30da5c5f2d5e72842e00bcb57657162cdabef0931f40e2deb9b4140440cecd"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "anstyle-parse"
|
name = "anstyle-parse"
|
||||||
version = "0.2.0"
|
version = "0.2.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "e765fd216e48e067936442276d1d57399e37bce53c264d6fefbe298080cb57ee"
|
checksum = "938874ff5980b03a87c5524b3ae5b59cf99b1d6bc836848df7bc5ada9643c333"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"utf8parse",
|
"utf8parse",
|
||||||
]
|
]
|
||||||
|
@ -77,14 +78,20 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "anstyle-wincon"
|
name = "anstyle-wincon"
|
||||||
version = "1.0.1"
|
version = "1.0.2"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "180abfa45703aebe0093f79badacc01b8fd4ea2e35118747e5811127f926e188"
|
checksum = "c677ab05e09154296dd37acecd46420c17b9713e8366facafa8fc0885167cf4c"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"anstyle",
|
"anstyle",
|
||||||
"windows-sys",
|
"windows-sys",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "atomic"
|
||||||
|
version = "0.5.3"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "c59bdb34bc650a32731b31bd8f0829cc15d24a708ee31559e0bb34f2bc320cba"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "autocfg"
|
name = "autocfg"
|
||||||
version = "1.1.0"
|
version = "1.1.0"
|
||||||
|
@ -97,6 +104,12 @@ version = "1.3.2"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "bef38d45163c2f1dde094a7dfd33ccf595c92905c8f8f4fdc18d06fb1037718a"
|
checksum = "bef38d45163c2f1dde094a7dfd33ccf595c92905c8f8f4fdc18d06fb1037718a"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "bitflags"
|
||||||
|
version = "2.4.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "b4682ae6287fcf752ecaabbfcc7b6f9b72aa33933dc23a554d853aea8eea8635"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "bumpalo"
|
name = "bumpalo"
|
||||||
version = "3.13.0"
|
version = "3.13.0"
|
||||||
|
@ -105,9 +118,12 @@ checksum = "a3e2c3daef883ecc1b5d58c15adae93470a91d425f3532ba1695849656af3fc1"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "cc"
|
name = "cc"
|
||||||
version = "1.0.79"
|
version = "1.0.82"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "50d30906286121d95be3d479533b458f87493b30a4b5f79a607db8f5d11aa91f"
|
checksum = "305fe645edc1442a0fa8b6726ba61d422798d37a52e12eaecf4b022ebbb88f01"
|
||||||
|
dependencies = [
|
||||||
|
"libc",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "cfg-if"
|
name = "cfg-if"
|
||||||
|
@ -133,9 +149,9 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "clap"
|
name = "clap"
|
||||||
version = "4.3.1"
|
version = "4.3.21"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "b4ed2379f8603fa2b7509891660e802b88c70a79a6427a70abb5968054de2c28"
|
checksum = "c27cdf28c0f604ba3f512b0c9a409f8de8513e4816705deb0498b627e7c3a3fd"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"clap_builder",
|
"clap_builder",
|
||||||
"clap_derive",
|
"clap_derive",
|
||||||
|
@ -144,22 +160,21 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "clap_builder"
|
name = "clap_builder"
|
||||||
version = "4.3.1"
|
version = "4.3.21"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "72394f3339a76daf211e57d4bcb374410f3965dcc606dd0e03738c7888766980"
|
checksum = "08a9f1ab5e9f01a9b81f202e8562eb9a10de70abf9eaeac1be465c28b75aa4aa"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"anstream",
|
"anstream",
|
||||||
"anstyle",
|
"anstyle",
|
||||||
"bitflags",
|
|
||||||
"clap_lex",
|
"clap_lex",
|
||||||
"strsim",
|
"strsim",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "clap_derive"
|
name = "clap_derive"
|
||||||
version = "4.3.1"
|
version = "4.3.12"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "59e9ef9a08ee1c0e1f2e162121665ac45ac3783b0f897db7244ae75ad9a8f65b"
|
checksum = "54a9bb5758fc5dfe728d1019941681eccaf0cf8a4189b692a0ee2f2ecf90a050"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"heck",
|
"heck",
|
||||||
"proc-macro2",
|
"proc-macro2",
|
||||||
|
@ -195,10 +210,16 @@ dependencies = [
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "errno"
|
name = "equivalent"
|
||||||
version = "0.3.1"
|
version = "1.0.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "4bcfec3a70f97c962c307b2d2c56e358cf1d00b558d74262b5f929ee8cc7e73a"
|
checksum = "5443807d6dff69373d433ab9ef5378ad8df50ca6298caf15de6e52e24aaf54d5"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "errno"
|
||||||
|
version = "0.3.2"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "6b30f669a7961ef1631673d2766cc92f52d64f7ef354d4fe0ddfd30ed52f0f4f"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"errno-dragonfly",
|
"errno-dragonfly",
|
||||||
"libc",
|
"libc",
|
||||||
|
@ -216,10 +237,24 @@ dependencies = [
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "filetime"
|
name = "figment"
|
||||||
version = "0.2.21"
|
version = "0.10.10"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "5cbc844cecaee9d4443931972e1289c8ff485cb4cc2767cb03ca139ed6885153"
|
checksum = "4547e226f4c9ab860571e070a9034192b3175580ecea38da34fcdb53a018c9a5"
|
||||||
|
dependencies = [
|
||||||
|
"atomic",
|
||||||
|
"pear",
|
||||||
|
"serde",
|
||||||
|
"toml",
|
||||||
|
"uncased",
|
||||||
|
"version_check",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "filetime"
|
||||||
|
version = "0.2.22"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "d4029edd3e734da6fe05b6cd7bd2960760a616bd2ddd0d59a0124746d6272af0"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"cfg-if",
|
"cfg-if",
|
||||||
"libc",
|
"libc",
|
||||||
|
@ -237,6 +272,12 @@ dependencies = [
|
||||||
"miniz_oxide",
|
"miniz_oxide",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "hashbrown"
|
||||||
|
version = "0.14.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "2c6201b9ff9fd90a5a3bac2e56a830d0caa509576f0e503818ee82c181b3437a"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "heck"
|
name = "heck"
|
||||||
version = "0.4.1"
|
version = "0.4.1"
|
||||||
|
@ -245,15 +286,15 @@ checksum = "95505c38b4572b2d910cecb0281560f54b440a19336cbbcb27bf6ce6adc6f5a8"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "hermit-abi"
|
name = "hermit-abi"
|
||||||
version = "0.3.1"
|
version = "0.3.2"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "fed44880c466736ef9a5c5b5facefb5ed0785676d0c02d612db14e54f0d84286"
|
checksum = "443144c8cdadd93ebf52ddb4056d257f5b52c04d3c804e657d19eb73fc33668b"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "iana-time-zone"
|
name = "iana-time-zone"
|
||||||
version = "0.1.56"
|
version = "0.1.57"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "0722cd7114b7de04316e7ea5456a0bbb20e4adb46fd27a3697adb812cff0f37c"
|
checksum = "2fad5b825842d2b38bd206f3e81d6957625fd7f0a361e345c30e01a0ae2dd613"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"android_system_properties",
|
"android_system_properties",
|
||||||
"core-foundation-sys",
|
"core-foundation-sys",
|
||||||
|
@ -273,60 +314,70 @@ dependencies = [
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "io-lifetimes"
|
name = "indexmap"
|
||||||
version = "1.0.11"
|
version = "2.0.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "eae7b9aee968036d54dce06cebaefd919e4472e753296daccd6d344e3e2df0c2"
|
checksum = "d5477fe2230a79769d8dc68e0eabf5437907c0457a5614a9e8dddb67f65eb65d"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"hermit-abi",
|
"equivalent",
|
||||||
"libc",
|
"hashbrown",
|
||||||
"windows-sys",
|
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "is-terminal"
|
name = "inlinable_string"
|
||||||
version = "0.4.7"
|
version = "0.1.15"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "adcf93614601c8129ddf72e2d5633df827ba6551541c6d8c59520a371475be1f"
|
checksum = "c8fae54786f62fb2918dcfae3d568594e50eb9b5c25bf04371af6fe7516452fb"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "is-terminal"
|
||||||
|
version = "0.4.9"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "cb0889898416213fab133e1d33a0e5858a48177452750691bde3666d0fdbaf8b"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"hermit-abi",
|
"hermit-abi",
|
||||||
"io-lifetimes",
|
|
||||||
"rustix",
|
"rustix",
|
||||||
"windows-sys",
|
"windows-sys",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "itoa"
|
name = "itoa"
|
||||||
version = "1.0.6"
|
version = "1.0.9"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "453ad9f582a441959e5f0d088b02ce04cfe8d51a8eaf077f12ac6d3e94164ca6"
|
checksum = "af150ab688ff2122fcef229be89cb50dd66af9e01a4ff320cc137eecc9bacc38"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "js-sys"
|
name = "js-sys"
|
||||||
version = "0.3.63"
|
version = "0.3.64"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "2f37a4a5928311ac501dee68b3c7613a1037d0edb30c8e5427bd832d55d1b790"
|
checksum = "c5f195fe497f702db0f318b07fdd68edb16955aed830df8363d837542f8f935a"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"wasm-bindgen",
|
"wasm-bindgen",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "libc"
|
name = "libc"
|
||||||
version = "0.2.144"
|
version = "0.2.147"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "2b00cc1c228a6782d0f076e7b232802e0c5689d41bb5df366f2a6b6621cfdfe1"
|
checksum = "b4668fb0ea861c1df094127ac5f1da3409a82116a4ba74fca2e58ef927159bb3"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "linux-raw-sys"
|
name = "linux-raw-sys"
|
||||||
version = "0.3.8"
|
version = "0.4.5"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "ef53942eb7bf7ff43a617b3e2c1c4a5ecf5944a7c1bc12d7ee39bbb15e5c1519"
|
checksum = "57bcfdad1b858c2db7c38303a6d2ad4dfaf5eb53dfeb0910128b2c26d6158503"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "log"
|
name = "log"
|
||||||
version = "0.4.18"
|
version = "0.4.20"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "518ef76f2f87365916b142844c16d8fefd85039bc5699050210a7778ee1cd1de"
|
checksum = "b5e6163cb8c49088c2c36f57875e58ccd8c87c7427f7fbd50ea6710b2f3f2e8f"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "memchr"
|
||||||
|
version = "2.5.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "2dffe52ecf27772e601905b7522cb4ef790d2cc203488bbd0e2fe85fcb74566d"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "miniz_oxide"
|
name = "miniz_oxide"
|
||||||
|
@ -339,55 +390,90 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "num-traits"
|
name = "num-traits"
|
||||||
version = "0.2.15"
|
version = "0.2.16"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "578ede34cf02f8924ab9447f50c28075b4d3e5b269972345e7e0372b38c6cdcd"
|
checksum = "f30b0abd723be7e2ffca1272140fac1a2f084c77ec3e123c192b66af1ee9e6c2"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"autocfg",
|
"autocfg",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "once_cell"
|
name = "once_cell"
|
||||||
version = "1.17.2"
|
version = "1.18.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "9670a07f94779e00908f3e686eab508878ebb390ba6e604d3a284c00e8d0487b"
|
checksum = "dd8b5dd2ae5ed71462c540258bedcb51965123ad7e7ccf4b9a8cafaa4a63576d"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "pear"
|
||||||
|
version = "0.2.7"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "61a386cd715229d399604b50d1361683fe687066f42d56f54be995bc6868f71c"
|
||||||
|
dependencies = [
|
||||||
|
"inlinable_string",
|
||||||
|
"pear_codegen",
|
||||||
|
"yansi",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "pear_codegen"
|
||||||
|
version = "0.2.7"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "da9f0f13dac8069c139e8300a6510e3f4143ecf5259c60b116a9b271b4ca0d54"
|
||||||
|
dependencies = [
|
||||||
|
"proc-macro2",
|
||||||
|
"proc-macro2-diagnostics",
|
||||||
|
"quote",
|
||||||
|
"syn",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "proc-macro2"
|
name = "proc-macro2"
|
||||||
version = "1.0.59"
|
version = "1.0.66"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "6aeca18b86b413c660b781aa319e4e2648a3e6f9eadc9b47e9038e6fe9f3451b"
|
checksum = "18fb31db3f9bddb2ea821cde30a9f70117e3f119938b5ee630b7403aa6e2ead9"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"unicode-ident",
|
"unicode-ident",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "quote"
|
name = "proc-macro2-diagnostics"
|
||||||
version = "1.0.28"
|
version = "0.10.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "1b9ab9c7eadfd8df19006f1cf1a4aed13540ed5cbc047010ece5826e10825488"
|
checksum = "af066a9c399a26e020ada66a034357a868728e72cd426f3adcd35f80d88d88c8"
|
||||||
|
dependencies = [
|
||||||
|
"proc-macro2",
|
||||||
|
"quote",
|
||||||
|
"syn",
|
||||||
|
"version_check",
|
||||||
|
"yansi",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "quote"
|
||||||
|
version = "1.0.32"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "50f3b39ccfb720540debaa0164757101c08ecb8d326b15358ce76a62c7e85965"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"proc-macro2",
|
"proc-macro2",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "redox_syscall"
|
name = "redox_syscall"
|
||||||
version = "0.2.16"
|
version = "0.3.5"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "fb5a58c1855b4b6819d59012155603f0b22ad30cad752600aadfcb695265519a"
|
checksum = "567664f262709473930a4bf9e51bf2ebf3348f2e748ccc50dea20646858f8f29"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"bitflags",
|
"bitflags 1.3.2",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "rustix"
|
name = "rustix"
|
||||||
version = "0.37.19"
|
version = "0.38.8"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "acf8729d8542766f1b2cf77eb034d52f40d375bb8b615d0b147089946e16613d"
|
checksum = "19ed4fa021d81c8392ce04db050a3da9a60299050b7ae1cf482d862b54a7218f"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"bitflags",
|
"bitflags 2.4.0",
|
||||||
"errno",
|
"errno",
|
||||||
"io-lifetimes",
|
|
||||||
"libc",
|
"libc",
|
||||||
"linux-raw-sys",
|
"linux-raw-sys",
|
||||||
"windows-sys",
|
"windows-sys",
|
||||||
|
@ -395,24 +481,24 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "ryu"
|
name = "ryu"
|
||||||
version = "1.0.13"
|
version = "1.0.15"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "f91339c0467de62360649f8d3e185ca8de4224ff281f66000de5eb2a77a79041"
|
checksum = "1ad4cc8da4ef723ed60bced201181d83791ad433213d8c24efffda1eec85d741"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "serde"
|
name = "serde"
|
||||||
version = "1.0.164"
|
version = "1.0.183"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "9e8c8cf938e98f769bc164923b06dce91cea1751522f46f8466461af04c9027d"
|
checksum = "32ac8da02677876d532745a130fc9d8e6edfa81a269b107c5b00829b91d8eb3c"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"serde_derive",
|
"serde_derive",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "serde_derive"
|
name = "serde_derive"
|
||||||
version = "1.0.164"
|
version = "1.0.183"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "d9735b638ccc51c28bf6914d90a2e9725b377144fc612c49a611fddd1b631d68"
|
checksum = "aafe972d60b0b9bee71a91b92fee2d4fb3c9d7e8f6b179aa99f27203d99a4816"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"proc-macro2",
|
"proc-macro2",
|
||||||
"quote",
|
"quote",
|
||||||
|
@ -421,9 +507,9 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "serde_json"
|
name = "serde_json"
|
||||||
version = "1.0.96"
|
version = "1.0.104"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "057d394a50403bcac12672b2b18fb387ab6d289d957dab67dd201875391e52f1"
|
checksum = "076066c5f1078eac5b722a31827a8832fe108bed65dfa75e233c89f8206e976c"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"itoa",
|
"itoa",
|
||||||
"ryu",
|
"ryu",
|
||||||
|
@ -431,10 +517,19 @@ dependencies = [
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "signal-hook"
|
name = "serde_spanned"
|
||||||
version = "0.3.15"
|
version = "0.6.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "732768f1176d21d09e076c23a93123d40bba92d50c4058da34d45c8de8e682b9"
|
checksum = "96426c9936fd7a0124915f9185ea1d20aa9445cc9821142f0a73bc9207a2e186"
|
||||||
|
dependencies = [
|
||||||
|
"serde",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "signal-hook"
|
||||||
|
version = "0.3.17"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "8621587d4798caf8eb44879d42e56b9a93ea5dcd315a6487c357130095b62801"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"libc",
|
"libc",
|
||||||
"signal-hook-registry",
|
"signal-hook-registry",
|
||||||
|
@ -457,9 +552,9 @@ checksum = "73473c0e59e6d5812c5dfe2a064a6444949f089e20eec9a2e5506596494e4623"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "syn"
|
name = "syn"
|
||||||
version = "2.0.18"
|
version = "2.0.28"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "32d41677bcbe24c20c52e7c70b0d8db04134c5d1066bf98662e2871ad200ea3e"
|
checksum = "04361975b3f5e348b2189d8dc55bc942f278b2d482a6a0365de5bdd62d351567"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"proc-macro2",
|
"proc-macro2",
|
||||||
"quote",
|
"quote",
|
||||||
|
@ -468,9 +563,9 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "tar"
|
name = "tar"
|
||||||
version = "0.4.38"
|
version = "0.4.40"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "4b55807c0344e1e6c04d7c965f5289c39a8d94ae23ed5c0b57aabac549f871c6"
|
checksum = "b16afcea1f22891c49a00c751c7b63b2233284064f11a200fc624137c51e2ddb"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"filetime",
|
"filetime",
|
||||||
"libc",
|
"libc",
|
||||||
|
@ -489,10 +584,53 @@ dependencies = [
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "unicode-ident"
|
name = "toml"
|
||||||
version = "1.0.9"
|
version = "0.7.6"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "b15811caf2415fb889178633e7724bad2509101cde276048e013b9def5e51fa0"
|
checksum = "c17e963a819c331dcacd7ab957d80bc2b9a9c1e71c804826d2f283dd65306542"
|
||||||
|
dependencies = [
|
||||||
|
"serde",
|
||||||
|
"serde_spanned",
|
||||||
|
"toml_datetime",
|
||||||
|
"toml_edit",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "toml_datetime"
|
||||||
|
version = "0.6.3"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "7cda73e2f1397b1262d6dfdcef8aafae14d1de7748d66822d3bfeeb6d03e5e4b"
|
||||||
|
dependencies = [
|
||||||
|
"serde",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "toml_edit"
|
||||||
|
version = "0.19.14"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "f8123f27e969974a3dfba720fdb560be359f57b44302d280ba72e76a74480e8a"
|
||||||
|
dependencies = [
|
||||||
|
"indexmap",
|
||||||
|
"serde",
|
||||||
|
"serde_spanned",
|
||||||
|
"toml_datetime",
|
||||||
|
"winnow",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "uncased"
|
||||||
|
version = "0.9.9"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "9b9bc53168a4be7402ab86c3aad243a84dd7381d09be0eddc81280c1da95ca68"
|
||||||
|
dependencies = [
|
||||||
|
"version_check",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "unicode-ident"
|
||||||
|
version = "1.0.11"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "301abaae475aa91687eb82514b328ab47a211a533026cb25fc3e519b86adfc3c"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "utf8parse"
|
name = "utf8parse"
|
||||||
|
@ -500,6 +638,12 @@ version = "0.2.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "711b9620af191e0cdc7468a8d14e709c3dcdb115b36f838e601583af800a370a"
|
checksum = "711b9620af191e0cdc7468a8d14e709c3dcdb115b36f838e601583af800a370a"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "version_check"
|
||||||
|
version = "0.9.4"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "49874b5167b65d7193b8aba1567f5c7d93d001cafc34600cee003eda787e483f"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "wasi"
|
name = "wasi"
|
||||||
version = "0.10.0+wasi-snapshot-preview1"
|
version = "0.10.0+wasi-snapshot-preview1"
|
||||||
|
@ -508,9 +652,9 @@ checksum = "1a143597ca7c7793eff794def352d41792a93c481eb1042423ff7ff72ba2c31f"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "wasm-bindgen"
|
name = "wasm-bindgen"
|
||||||
version = "0.2.86"
|
version = "0.2.87"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "5bba0e8cb82ba49ff4e229459ff22a191bbe9a1cb3a341610c9c33efc27ddf73"
|
checksum = "7706a72ab36d8cb1f80ffbf0e071533974a60d0a308d01a5d0375bf60499a342"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"cfg-if",
|
"cfg-if",
|
||||||
"wasm-bindgen-macro",
|
"wasm-bindgen-macro",
|
||||||
|
@ -518,9 +662,9 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "wasm-bindgen-backend"
|
name = "wasm-bindgen-backend"
|
||||||
version = "0.2.86"
|
version = "0.2.87"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "19b04bc93f9d6bdee709f6bd2118f57dd6679cf1176a1af464fca3ab0d66d8fb"
|
checksum = "5ef2b6d3c510e9625e5fe6f509ab07d66a760f0885d858736483c32ed7809abd"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"bumpalo",
|
"bumpalo",
|
||||||
"log",
|
"log",
|
||||||
|
@ -533,9 +677,9 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "wasm-bindgen-macro"
|
name = "wasm-bindgen-macro"
|
||||||
version = "0.2.86"
|
version = "0.2.87"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "14d6b024f1a526bb0234f52840389927257beb670610081360e5a03c5df9c258"
|
checksum = "dee495e55982a3bd48105a7b947fd2a9b4a8ae3010041b9e0faab3f9cd028f1d"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"quote",
|
"quote",
|
||||||
"wasm-bindgen-macro-support",
|
"wasm-bindgen-macro-support",
|
||||||
|
@ -543,9 +687,9 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "wasm-bindgen-macro-support"
|
name = "wasm-bindgen-macro-support"
|
||||||
version = "0.2.86"
|
version = "0.2.87"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "e128beba882dd1eb6200e1dc92ae6c5dbaa4311aa7bb211ca035779e5efc39f8"
|
checksum = "54681b18a46765f095758388f2d0cf16eb8d4169b639ab575a8f5693af210c7b"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"proc-macro2",
|
"proc-macro2",
|
||||||
"quote",
|
"quote",
|
||||||
|
@ -556,9 +700,9 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "wasm-bindgen-shared"
|
name = "wasm-bindgen-shared"
|
||||||
version = "0.2.86"
|
version = "0.2.87"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "ed9d5b4305409d1fc9482fee2d7f9bcbf24b3972bf59817ef757e23982242a93"
|
checksum = "ca6ad05a4870b2bf5fe995117d3728437bd27d7cd5f06f13c17443ef369775a1"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "winapi"
|
name = "winapi"
|
||||||
|
@ -602,9 +746,9 @@ dependencies = [
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "windows-targets"
|
name = "windows-targets"
|
||||||
version = "0.48.0"
|
version = "0.48.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "7b1eb6f0cd7c80c79759c929114ef071b87354ce476d9d94271031c0497adfd5"
|
checksum = "05d4b17490f70499f20b9e791dcf6a299785ce8af4d709018206dc5b4953e95f"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"windows_aarch64_gnullvm",
|
"windows_aarch64_gnullvm",
|
||||||
"windows_aarch64_msvc",
|
"windows_aarch64_msvc",
|
||||||
|
@ -658,10 +802,25 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "1a515f5799fe4961cb532f983ce2b23082366b898e52ffbce459c86f67c8378a"
|
checksum = "1a515f5799fe4961cb532f983ce2b23082366b898e52ffbce459c86f67c8378a"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "xattr"
|
name = "winnow"
|
||||||
version = "0.2.3"
|
version = "0.5.10"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "6d1526bbe5aaeb5eb06885f4d987bcdfa5e23187055de9b83fe00156a821fabc"
|
checksum = "5504cc7644f4b593cbc05c4a55bf9bd4e94b867c3c0bd440934174d50482427d"
|
||||||
|
dependencies = [
|
||||||
|
"memchr",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "xattr"
|
||||||
|
version = "1.0.1"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "f4686009f71ff3e5c4dbcf1a282d0a44db3f021ba69350cd42086b3e5f1c6985"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"libc",
|
"libc",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "yansi"
|
||||||
|
version = "1.0.0-rc.1"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "1367295b8f788d371ce2dbc842c7b709c73ee1364d30351dd300ec2203b12377"
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
[package]
|
[package]
|
||||||
name = "alex"
|
name = "alex"
|
||||||
version = "0.2.2"
|
version = "0.4.1"
|
||||||
description = "Wrapper around Minecraft server processes, designed to complement Docker image installations."
|
description = "Wrapper around Minecraft server processes, designed to complement Docker image installations."
|
||||||
authors = ["Jef Roosens"]
|
authors = ["Jef Roosens"]
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
|
@ -15,8 +15,9 @@ flate2 = "1.0.26"
|
||||||
chrono = { version = "0.4.26", features = ["serde"] }
|
chrono = { version = "0.4.26", features = ["serde"] }
|
||||||
clap = { version = "4.3.1", features = ["derive", "env"] }
|
clap = { version = "4.3.1", features = ["derive", "env"] }
|
||||||
signal-hook = "0.3.15"
|
signal-hook = "0.3.15"
|
||||||
serde = { version = "1.0.164", features = ["derive", "rc"] }
|
serde = { version = "1.0.164", features = ["derive"] }
|
||||||
serde_json = "1.0.96"
|
serde_json = "1.0.96"
|
||||||
|
figment = { version = "0.10.10", features = ["env", "toml"] }
|
||||||
|
|
||||||
[profile.release]
|
[profile.release]
|
||||||
lto = "fat"
|
lto = "fat"
|
||||||
|
|
14
Dockerfile
14
Dockerfile
|
@ -47,13 +47,15 @@ COPY --from=builder /app/target/debug/alex /bin/alex
|
||||||
RUN chmod +x /bin/alex
|
RUN chmod +x /bin/alex
|
||||||
|
|
||||||
# Default value to keep users from eating up all ram accidentally
|
# Default value to keep users from eating up all ram accidentally
|
||||||
ENV ALEX_XMS=1024 \
|
ENV ALEX_CONFIG=/app/config \
|
||||||
|
ALEX_WORLD=/app/worlds \
|
||||||
|
ALEX_BACKUP=/app/backups \
|
||||||
|
ALEX_SERVER=paper \
|
||||||
|
ALEX_XMS=1024 \
|
||||||
ALEX_XMX=2048 \
|
ALEX_XMX=2048 \
|
||||||
ALEX_JAR=/app/server.jar \
|
ALEX_JAR=/app/server.jar \
|
||||||
ALEX_CONFIG_DIR=/app/config \
|
ALEX_SERVER_VERSION="${MC_VERSION}-${PAPERMC_VERSION}" \
|
||||||
ALEX_WORLD_DIR=/app/worlds \
|
ALEX_LAYERS="2min,2,4,4;3min,3,2,2"
|
||||||
ALEX_BACKUPS_DIR=/app/backups \
|
|
||||||
ALEX_SERVER_VERSION="${MC_VERSION}-${PAPERMC_VERSION}"
|
|
||||||
|
|
||||||
# Document exposed ports
|
# Document exposed ports
|
||||||
EXPOSE 25565
|
EXPOSE 25565
|
||||||
|
@ -62,4 +64,4 @@ EXPOSE 25565
|
||||||
USER paper:paper
|
USER paper:paper
|
||||||
|
|
||||||
ENTRYPOINT ["/bin/dumb-init", "--"]
|
ENTRYPOINT ["/bin/dumb-init", "--"]
|
||||||
CMD ["/bin/alex", "paper"]
|
CMD ["/bin/alex", "run"]
|
||||||
|
|
119
README.md
119
README.md
|
@ -2,7 +2,38 @@
|
||||||
|
|
||||||
Alex is a wrapper around a typical Minecraft server process. It acts as the
|
Alex is a wrapper around a typical Minecraft server process. It acts as the
|
||||||
parent process, and sits in between the user's input and the server's stdin.
|
parent process, and sits in between the user's input and the server's stdin.
|
||||||
This allows Alex to support additional commands that execute Rust code.
|
This allows Alex to support additional commands that execute Rust code, notably
|
||||||
|
creating periodic backups.
|
||||||
|
|
||||||
|
## Installation
|
||||||
|
|
||||||
|
Alex is distributed as statically compiled binaries for Linux amd64 and arm64.
|
||||||
|
These can be found
|
||||||
|
[here](https://git.rustybever.be/Chewing_Bever/alex/packages).
|
||||||
|
|
||||||
|
### Arch
|
||||||
|
|
||||||
|
Arch users can install prebuilt `x86_64` & `aarch64` packages from my `bur`
|
||||||
|
repository. Add the following at the bottom of your `pacman.conf`:
|
||||||
|
|
||||||
|
```toml
|
||||||
|
[bur]
|
||||||
|
Server = https://arch.r8r.be/$repo/$arch
|
||||||
|
SigLevel = Optional
|
||||||
|
```
|
||||||
|
|
||||||
|
If you prefer building the package yourself, the PKGBUILD can be found
|
||||||
|
[here](https://git.rustybever.be/bur/alex-mc).
|
||||||
|
|
||||||
|
### Dockerfiles
|
||||||
|
|
||||||
|
You can easily install alex in your Docker images by letting Docker download it
|
||||||
|
for you. Add the following to your Dockerfile (replace with your required
|
||||||
|
version & architecture):
|
||||||
|
|
||||||
|
```dockerfile
|
||||||
|
ADD "https://git.rustybever.be/api/packages/Chewing_Bever/generic/alex/0.2.2/alex-linux-amd64" /bin/alex
|
||||||
|
```
|
||||||
|
|
||||||
## Why
|
## Why
|
||||||
|
|
||||||
|
@ -19,8 +50,86 @@ Afterwards, saving is enabled again with `save-on`.
|
||||||
* Properly configures the process (working directory, optimisation flags)
|
* Properly configures the process (working directory, optimisation flags)
|
||||||
* Configure everything as CLI arguments or environment variables
|
* Configure everything as CLI arguments or environment variables
|
||||||
|
|
||||||
## Installation
|
## Configuration
|
||||||
|
|
||||||
Alex is distributed as statically compiled binaries for Linux amd64 and arm64.
|
Most information can be retrieved easily by looking at the help command:
|
||||||
These can be found
|
|
||||||
[here](https://git.rustybever.be/Chewing_Bever/alex/packages).
|
```
|
||||||
|
Wrapper around Minecraft server processes, designed to complement Docker image installations.
|
||||||
|
|
||||||
|
Usage: alex [OPTIONS] <COMMAND>
|
||||||
|
|
||||||
|
Commands:
|
||||||
|
run Run the server
|
||||||
|
backup Interact with the backup system without starting a server
|
||||||
|
help Print this message or the help of the given subcommand(s)
|
||||||
|
|
||||||
|
Options:
|
||||||
|
--config <CONFIG_DIR>
|
||||||
|
Directory where configs are stored, and where the server will run [env: ALEX_CONFIG_DIR=] [default: .]
|
||||||
|
--world <WORLD_DIR>
|
||||||
|
Directory where world files will be saved [env: ALEX_WORLD_DIR=] [default: ../worlds]
|
||||||
|
--backup <BACKUP_DIR>
|
||||||
|
Directory where backups will be stored [env: ALEX_BACKUP_DIR=] [default: ../backups]
|
||||||
|
--layers <LAYERS>
|
||||||
|
What backup layers to employ, provided as a list of tuples name,frequency,chains,chain_len delimited by semicolons (;) [env: ALEX_LAYERS=]
|
||||||
|
--server <SERVER>
|
||||||
|
Type of server [env: ALEX_SERVER=] [default: unknown] [possible values: unknown, paper, forge, vanilla]
|
||||||
|
--server-version <SERVER_VERSION>
|
||||||
|
Version string for the server, e.g. 1.19.4-545 [env: ALEX_SERVER_VERSION=] [default: ]
|
||||||
|
-h, --help
|
||||||
|
Print help
|
||||||
|
-V, --version
|
||||||
|
Print version
|
||||||
|
|
||||||
|
```
|
||||||
|
|
||||||
|
### Choosing layer parameters
|
||||||
|
|
||||||
|
One part of the configuration that does require some clarification is the layer
|
||||||
|
system. Alex can manage an arbitrary number of backup layers, each having its
|
||||||
|
own configuration. These layers can either use incremental or full backups,
|
||||||
|
depending on how they're configured.
|
||||||
|
|
||||||
|
These layers mostly correspond to the grandfather-father-son backup rotation
|
||||||
|
scheme. For example, one could have a layer that creates incremental backups
|
||||||
|
every 30 minutes, which are stored for 24 hours. This gives you 24 hours of
|
||||||
|
granular rollback in case your server suffers a crash. A second layer might
|
||||||
|
create a full backup every 24 hours, with backups being stored for 7 days. This
|
||||||
|
gives you 7 days worth of backups with the granularity of 24 hours. This
|
||||||
|
approach allows for greater versatility, while not having to store a large
|
||||||
|
amount of data. Thanks to incremental backups, frequent backups don't have to
|
||||||
|
take long at all.
|
||||||
|
|
||||||
|
A layer consists of 4 pieces of metadata:
|
||||||
|
|
||||||
|
* A name, which will be used in the file system and the in-game notifications
|
||||||
|
* The frequency, which describes in minutes how frequently a backup should be
|
||||||
|
created
|
||||||
|
* How many chains should be kept at all times
|
||||||
|
* How long each chain should be
|
||||||
|
|
||||||
|
These last two require some clarification. In Alex, a "chain" describes an
|
||||||
|
initial full backup and zero or more incremental backups that are created from
|
||||||
|
that initial full backup. This concept exists because an incremental backup has
|
||||||
|
no real meaning if its ancestors are not known. To restore one of these chains,
|
||||||
|
all backups in the chain need to be restored in-order. Note that a chain length
|
||||||
|
of 1 disables incremental backups entirely.
|
||||||
|
|
||||||
|
How many backups to keep is defined by how many chains should be stored.
|
||||||
|
Because an incremental backup needs to have its ancestors in order to be
|
||||||
|
restored, we can't simply "keep the last n backups", as this would break these
|
||||||
|
chains. Therefore, you configure how many backups to store using these chains.
|
||||||
|
|
||||||
|
For example, if you configure a layer to store 5 chains of length 4, you will
|
||||||
|
have 20 archive files on disk, namely 5 full backups and 15 incremental
|
||||||
|
backups. Note that Alex applies these rules to *full* chains. An in-progress
|
||||||
|
chain does not count towards this total. Therefore, you can have up to `n-1`
|
||||||
|
additional archive files, with `n` being the chain length, on disk.
|
||||||
|
|
||||||
|
To look at it from another perspective, say we wish to have a granularity of 30
|
||||||
|
minutes for a timespan of 24 hours. Then we could configure the layer to only
|
||||||
|
save a single chain, with a chain length of 48. If we prefer to have a few full
|
||||||
|
backups instead of a long chain of incremental backups, we could instead use a
|
||||||
|
chain length of 12 and store 4 chains. Either way, the total comes out to 48,
|
||||||
|
which spans 24 hours if we make a backup every 30 minutes.
|
||||||
|
|
|
@ -0,0 +1,16 @@
|
||||||
|
config = "data/config"
|
||||||
|
world = "data/worlds"
|
||||||
|
backup = "data/backups"
|
||||||
|
server = "Paper"
|
||||||
|
|
||||||
|
# [[layers]]
|
||||||
|
# name = "2min"
|
||||||
|
# frequency = 2
|
||||||
|
# chains = 4
|
||||||
|
# chain_len = 4
|
||||||
|
|
||||||
|
# [[layers]]
|
||||||
|
# name = "3min"
|
||||||
|
# frequency = 3
|
||||||
|
# chains = 2
|
||||||
|
# chain_len = 2
|
|
@ -1,76 +1,162 @@
|
||||||
|
use std::{borrow::Borrow, fmt};
|
||||||
|
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use std::collections::{HashMap, HashSet};
|
|
||||||
use std::path::PathBuf;
|
use super::State;
|
||||||
|
|
||||||
/// Represents the changes relative to the previous backup
|
/// Represents the changes relative to the previous backup
|
||||||
#[derive(Debug, Serialize, Deserialize)]
|
#[derive(Debug, Serialize, Deserialize, Clone)]
|
||||||
pub struct Delta {
|
pub struct Delta {
|
||||||
/// What files were added/modified in each part of the tarball.
|
/// What files were added/modified in each part of the tarball.
|
||||||
pub added: HashMap<PathBuf, HashSet<PathBuf>>,
|
pub added: State,
|
||||||
/// What files were removed in this backup, in comparison to the previous backup. For full
|
/// What files were removed in this backup, in comparison to the previous backup. For full
|
||||||
/// backups, this will always be empty, as they do not consider previous backups.
|
/// backups, this will always be empty, as they do not consider previous backups.
|
||||||
/// The map stores a separate list for each top-level directory, as the contents of these
|
/// The map stores a separate list for each top-level directory, as the contents of these
|
||||||
/// directories can come for different source directories.
|
/// directories can come for different source directories.
|
||||||
pub removed: HashMap<PathBuf, HashSet<PathBuf>>,
|
pub removed: State,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Delta {
|
impl Delta {
|
||||||
pub fn new() -> Self {
|
pub fn new() -> Self {
|
||||||
Self {
|
Self {
|
||||||
added: HashMap::new(),
|
added: State::new(),
|
||||||
removed: HashMap::new(),
|
removed: State::new(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Update the current state so that its result becomes the merge of itself and the other
|
/// Returns whether the delta is empty by checking whether both its added and removed state
|
||||||
/// state.
|
/// return true for their `is_empty`.
|
||||||
#[allow(dead_code)]
|
#[allow(dead_code)]
|
||||||
pub fn merge(&mut self, delta: &Self) {
|
pub fn is_empty(&self) -> bool {
|
||||||
|
self.added.is_empty() && self.removed.is_empty()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Calculate the union of this delta with another delta.
|
||||||
|
///
|
||||||
|
/// The union of two deltas is a delta that produces the same state as if you were to apply
|
||||||
|
/// both deltas in-order. Note that this operation is not commutative.
|
||||||
|
pub fn union(&self, delta: &Self) -> Self {
|
||||||
|
let mut out = self.clone();
|
||||||
|
|
||||||
for (dir, added) in delta.added.iter() {
|
for (dir, added) in delta.added.iter() {
|
||||||
// Files that were removed in the current state, but added in the new state, are no
|
// Files that were removed in the current state, but added in the new state, are no
|
||||||
// longer removed
|
// longer removed
|
||||||
if let Some(orig_removed) = self.removed.get_mut(dir) {
|
if let Some(orig_removed) = out.removed.get_mut(dir) {
|
||||||
orig_removed.retain(|k| !added.contains(k));
|
orig_removed.retain(|k| !added.contains(k));
|
||||||
}
|
}
|
||||||
|
|
||||||
// Newly added files are added to the state as well
|
// Newly added files are added to the state as well
|
||||||
if let Some(orig_added) = self.added.get_mut(dir) {
|
if let Some(orig_added) = out.added.get_mut(dir) {
|
||||||
orig_added.extend(added.iter().cloned());
|
orig_added.extend(added.iter().cloned());
|
||||||
} else {
|
} else {
|
||||||
self.added.insert(dir.clone(), added.clone());
|
out.added.insert(dir.clone(), added.clone());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
for (dir, removed) in delta.removed.iter() {
|
for (dir, removed) in delta.removed.iter() {
|
||||||
// Files that were originally added, but now deleted are removed from the added list
|
// Files that were originally added, but now deleted are removed from the added list
|
||||||
if let Some(orig_added) = self.added.get_mut(dir) {
|
if let Some(orig_added) = out.added.get_mut(dir) {
|
||||||
orig_added.retain(|k| !removed.contains(k));
|
orig_added.retain(|k| !removed.contains(k));
|
||||||
}
|
}
|
||||||
|
|
||||||
// Newly removed files are added to the state as well
|
// Newly removed files are added to the state as well
|
||||||
if let Some(orig_removed) = self.removed.get_mut(dir) {
|
if let Some(orig_removed) = out.removed.get_mut(dir) {
|
||||||
orig_removed.extend(removed.iter().cloned());
|
orig_removed.extend(removed.iter().cloned());
|
||||||
} else {
|
} else {
|
||||||
self.removed.insert(dir.clone(), removed.clone());
|
out.removed.insert(dir.clone(), removed.clone());
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Modify the given state by applying this delta's changes to it
|
|
||||||
pub fn apply(&self, state: &mut HashMap<PathBuf, HashSet<PathBuf>>) {
|
|
||||||
// First we add new files, then we remove the old ones
|
|
||||||
for (dir, added) in self.added.iter() {
|
|
||||||
if let Some(current) = state.get_mut(dir) {
|
|
||||||
current.extend(added.iter().cloned());
|
|
||||||
} else {
|
|
||||||
state.insert(dir.clone(), added.clone());
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
for (dir, removed) in self.removed.iter() {
|
out
|
||||||
if let Some(current) = state.get_mut(dir) {
|
}
|
||||||
current.retain(|k| !removed.contains(k));
|
|
||||||
|
// Calculate the difference between this delta and the other delta.
|
||||||
|
//
|
||||||
|
// The difference simply means removing all adds and removes that are also performed in the
|
||||||
|
// other delta.
|
||||||
|
pub fn difference(&self, other: &Self) -> Self {
|
||||||
|
let mut out = self.clone();
|
||||||
|
|
||||||
|
for (dir, added) in out.added.iter_mut() {
|
||||||
|
// If files are added in the other delta, we don't add them in this delta
|
||||||
|
if let Some(other_added) = other.added.get(dir) {
|
||||||
|
added.retain(|k| !other_added.contains(k));
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
for (dir, removed) in out.removed.iter_mut() {
|
||||||
|
// If files are removed in the other delta, we don't remove them in this delta either
|
||||||
|
if let Some(other_removed) = other.removed.get(dir) {
|
||||||
|
removed.retain(|k| !other_removed.contains(k));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
out
|
||||||
|
}
|
||||||
|
|
||||||
|
// Calculate the strict difference between this delta and the other delta.
|
||||||
|
//
|
||||||
|
// The strict difference is a difference where all operations that would be overwritten by the
|
||||||
|
// other delta are also removed (a.k.a. adding a file after removing it, or vice versa)
|
||||||
|
pub fn strict_difference(&self, other: &Self) -> Self {
|
||||||
|
let mut out = self.difference(other);
|
||||||
|
|
||||||
|
for (dir, added) in out.added.iter_mut() {
|
||||||
|
// Remove additions that are removed in the other delta
|
||||||
|
if let Some(other_removed) = other.removed.get(dir) {
|
||||||
|
added.retain(|k| !other_removed.contains(k));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for (dir, removed) in out.removed.iter_mut() {
|
||||||
|
// Remove removals that are re-added in the other delta
|
||||||
|
if let Some(other_added) = other.added.get(dir) {
|
||||||
|
removed.retain(|k| !other_added.contains(k));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
out
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Given a chain of deltas, ordered from last to first, calculate the "contribution" for each
|
||||||
|
/// state.
|
||||||
|
///
|
||||||
|
/// The contribution of a delta in a given chain is defined as the parts of the state produced
|
||||||
|
/// by this chain that are actually provided by this delta. This comes down to calculating the
|
||||||
|
/// strict difference of this delta and all of its successive deltas.
|
||||||
|
pub fn contributions<I>(deltas: I) -> Vec<State>
|
||||||
|
where
|
||||||
|
I: IntoIterator,
|
||||||
|
I::Item: Borrow<Delta>,
|
||||||
|
{
|
||||||
|
let mut contributions: Vec<State> = Vec::new();
|
||||||
|
|
||||||
|
let mut deltas = deltas.into_iter();
|
||||||
|
|
||||||
|
if let Some(first_delta) = deltas.next() {
|
||||||
|
// From last to first, we calculate the strict difference of the delta with the union of all its
|
||||||
|
// following deltas. The list of added files of this difference is the contribution for
|
||||||
|
// that delta.
|
||||||
|
contributions.push(first_delta.borrow().added.clone());
|
||||||
|
let mut union_future = first_delta.borrow().clone();
|
||||||
|
|
||||||
|
for delta in deltas {
|
||||||
|
contributions.push(delta.borrow().strict_difference(&union_future).added);
|
||||||
|
union_future = union_future.union(delta.borrow());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// contributions.reverse();
|
||||||
|
|
||||||
|
contributions
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl fmt::Display for Delta {
|
||||||
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||||
|
let added_count: usize = self.added.values().map(|s| s.len()).sum();
|
||||||
|
let removed_count: usize = self.removed.values().map(|s| s.len()).sum();
|
||||||
|
|
||||||
|
write!(f, "+{}-{}", added_count, removed_count)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,43 @@
|
||||||
|
use std::io::{self, Write};
|
||||||
|
|
||||||
|
/// Wrapper around the Write trait that counts how many bytes have been written in total.
|
||||||
|
/// Heavily inspired by https://stackoverflow.com/a/42189386
|
||||||
|
pub struct CountingWrite<W> {
|
||||||
|
inner: W,
|
||||||
|
count: usize,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<W> CountingWrite<W>
|
||||||
|
where
|
||||||
|
W: Write,
|
||||||
|
{
|
||||||
|
pub fn new(writer: W) -> Self {
|
||||||
|
Self {
|
||||||
|
inner: writer,
|
||||||
|
count: 0,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn bytes_written(&self) -> usize {
|
||||||
|
self.count
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<W> Write for CountingWrite<W>
|
||||||
|
where
|
||||||
|
W: Write,
|
||||||
|
{
|
||||||
|
fn write(&mut self, buf: &[u8]) -> io::Result<usize> {
|
||||||
|
let res = self.inner.write(buf);
|
||||||
|
|
||||||
|
if let Ok(count) = res {
|
||||||
|
self.count += count;
|
||||||
|
}
|
||||||
|
|
||||||
|
res
|
||||||
|
}
|
||||||
|
|
||||||
|
fn flush(&mut self) -> io::Result<()> {
|
||||||
|
self.inner.flush()
|
||||||
|
}
|
||||||
|
}
|
|
@ -1,140 +0,0 @@
|
||||||
use super::Backup;
|
|
||||||
use serde::de::DeserializeOwned;
|
|
||||||
use serde::Serialize;
|
|
||||||
use std::fs::File;
|
|
||||||
use std::io;
|
|
||||||
use std::path::PathBuf;
|
|
||||||
|
|
||||||
pub struct Manager<T>
|
|
||||||
where
|
|
||||||
T: Clone + Serialize + DeserializeOwned,
|
|
||||||
{
|
|
||||||
backup_dir: PathBuf,
|
|
||||||
config_dir: PathBuf,
|
|
||||||
world_dir: PathBuf,
|
|
||||||
default_metadata: T,
|
|
||||||
chain_len: u64,
|
|
||||||
chains_to_keep: u64,
|
|
||||||
chains: Vec<Vec<Backup<T>>>,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<T> Manager<T>
|
|
||||||
where
|
|
||||||
T: Clone + Serialize + DeserializeOwned,
|
|
||||||
{
|
|
||||||
const METADATA_FILE: &str = "alex.json";
|
|
||||||
|
|
||||||
/// Initialize a new instance of a `BackupManager`.
|
|
||||||
pub fn new(
|
|
||||||
backup_dir: PathBuf,
|
|
||||||
config_dir: PathBuf,
|
|
||||||
world_dir: PathBuf,
|
|
||||||
metadata: T,
|
|
||||||
chain_len: u64,
|
|
||||||
chains_to_keep: u64,
|
|
||||||
) -> Self {
|
|
||||||
Self {
|
|
||||||
backup_dir,
|
|
||||||
config_dir,
|
|
||||||
world_dir,
|
|
||||||
default_metadata: metadata,
|
|
||||||
chain_len,
|
|
||||||
chains_to_keep,
|
|
||||||
chains: Vec::new(),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Create a new backup with the expected type.
|
|
||||||
pub fn create_backup(&mut self) -> io::Result<()> {
|
|
||||||
let dirs = vec![
|
|
||||||
(PathBuf::from("config"), self.config_dir.clone()),
|
|
||||||
(PathBuf::from("worlds"), self.world_dir.clone()),
|
|
||||||
];
|
|
||||||
|
|
||||||
// We start a new chain if the current chain is complete, or if there isn't a first chain
|
|
||||||
// yet
|
|
||||||
if let Some(current_chain) = self.chains.last() {
|
|
||||||
let current_chain_len: u64 = current_chain.len().try_into().unwrap();
|
|
||||||
|
|
||||||
if current_chain_len >= self.chain_len {
|
|
||||||
self.chains.push(Vec::new());
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
self.chains.push(Vec::new());
|
|
||||||
}
|
|
||||||
|
|
||||||
let current_chain = self.chains.last_mut().unwrap();
|
|
||||||
|
|
||||||
let mut backup = if !current_chain.is_empty() {
|
|
||||||
let previous_backup = current_chain.last().unwrap();
|
|
||||||
let state = Backup::state(current_chain);
|
|
||||||
|
|
||||||
Backup::create_from(state, previous_backup.start_time, &self.backup_dir, dirs)?
|
|
||||||
} else {
|
|
||||||
Backup::create(&self.backup_dir, dirs)?
|
|
||||||
};
|
|
||||||
|
|
||||||
backup.set_metadata(self.default_metadata.clone());
|
|
||||||
|
|
||||||
current_chain.push(backup);
|
|
||||||
|
|
||||||
self.save()?;
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Delete all backups associated with outdated chains, and forget those chains.
|
|
||||||
pub fn remove_old_backups(&mut self) -> io::Result<()> {
|
|
||||||
let chains_to_store: usize = self.chains_to_keep.try_into().unwrap();
|
|
||||||
|
|
||||||
if chains_to_store < self.chains.len() {
|
|
||||||
let mut remove_count: usize = self.chains.len() - chains_to_store;
|
|
||||||
|
|
||||||
// We only count finished chains towards the list of stored chains
|
|
||||||
let chain_len: usize = self.chain_len.try_into().unwrap();
|
|
||||||
if self.chains.last().unwrap().len() < chain_len {
|
|
||||||
remove_count -= 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
for chain in self.chains.drain(..remove_count) {
|
|
||||||
for backup in chain {
|
|
||||||
let path = Backup::path(&self.backup_dir, backup.start_time);
|
|
||||||
std::fs::remove_file(path)?;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
self.save()?;
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Write the in-memory state to disk.
|
|
||||||
pub fn save(&self) -> io::Result<()> {
|
|
||||||
let json_file = File::create(self.backup_dir.join(Self::METADATA_FILE))?;
|
|
||||||
serde_json::to_writer(json_file, &self.chains)?;
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Overwrite the in-memory state with the on-disk state.
|
|
||||||
pub fn load(&mut self) -> io::Result<()> {
|
|
||||||
let json_file = match File::open(self.backup_dir.join(Self::METADATA_FILE)) {
|
|
||||||
Ok(f) => f,
|
|
||||||
Err(e) => {
|
|
||||||
// Don't error out if the file isn't there, it will be created when necessary
|
|
||||||
if e.kind() == io::ErrorKind::NotFound {
|
|
||||||
self.chains = Vec::new();
|
|
||||||
|
|
||||||
return Ok(());
|
|
||||||
} else {
|
|
||||||
return Err(e);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
self.chains = serde_json::from_reader(json_file)?;
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -0,0 +1,46 @@
|
||||||
|
use std::{error::Error, fmt, str::FromStr};
|
||||||
|
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
#[derive(Clone, Debug, Serialize, Deserialize)]
|
||||||
|
pub struct ManagerConfig {
|
||||||
|
pub name: String,
|
||||||
|
pub frequency: u32,
|
||||||
|
pub chains: u64,
|
||||||
|
pub chain_len: u64,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct ParseManagerConfigErr;
|
||||||
|
|
||||||
|
impl Error for ParseManagerConfigErr {}
|
||||||
|
|
||||||
|
impl fmt::Display for ParseManagerConfigErr {
|
||||||
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||||
|
write!(f, "parse manager config err")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl FromStr for ManagerConfig {
|
||||||
|
type Err = ParseManagerConfigErr;
|
||||||
|
|
||||||
|
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
||||||
|
let splits: Vec<&str> = s.split(',').collect();
|
||||||
|
|
||||||
|
if let [name, frequency, chains, chain_len] = splits[..] {
|
||||||
|
let name: String = name.parse().map_err(|_| ParseManagerConfigErr)?;
|
||||||
|
let frequency: u32 = frequency.parse().map_err(|_| ParseManagerConfigErr)?;
|
||||||
|
let chains: u64 = chains.parse().map_err(|_| ParseManagerConfigErr)?;
|
||||||
|
let chain_len: u64 = chain_len.parse().map_err(|_| ParseManagerConfigErr)?;
|
||||||
|
|
||||||
|
Ok(ManagerConfig {
|
||||||
|
name,
|
||||||
|
chains,
|
||||||
|
chain_len,
|
||||||
|
frequency,
|
||||||
|
})
|
||||||
|
} else {
|
||||||
|
Err(ParseManagerConfigErr)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,149 @@
|
||||||
|
use std::{
|
||||||
|
collections::HashMap,
|
||||||
|
io,
|
||||||
|
path::{Path, PathBuf},
|
||||||
|
};
|
||||||
|
|
||||||
|
use chrono::Utc;
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use super::{Manager, ManagerConfig};
|
||||||
|
|
||||||
|
/// Manages a collection of backup layers, allowing them to be utilized as a single object.
|
||||||
|
pub struct MetaManager<T>
|
||||||
|
where
|
||||||
|
T: Clone + Serialize + for<'de> Deserialize<'de> + std::fmt::Debug,
|
||||||
|
{
|
||||||
|
backup_dir: PathBuf,
|
||||||
|
dirs: Vec<(PathBuf, PathBuf)>,
|
||||||
|
default_metadata: T,
|
||||||
|
managers: HashMap<String, Manager<T>>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T> MetaManager<T>
|
||||||
|
where
|
||||||
|
T: Clone + Serialize + for<'de> Deserialize<'de> + std::fmt::Debug,
|
||||||
|
{
|
||||||
|
pub fn new<P: Into<PathBuf>>(
|
||||||
|
backup_dir: P,
|
||||||
|
dirs: Vec<(PathBuf, PathBuf)>,
|
||||||
|
default_metadata: T,
|
||||||
|
) -> Self {
|
||||||
|
MetaManager {
|
||||||
|
backup_dir: backup_dir.into(),
|
||||||
|
dirs,
|
||||||
|
default_metadata,
|
||||||
|
managers: HashMap::new(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Add a new manager to track, initializing it first.
|
||||||
|
pub fn add(&mut self, config: &ManagerConfig) -> io::Result<()> {
|
||||||
|
// Backup dir itself should exist, but we control its contents, so we can create
|
||||||
|
// separate directories for each layer
|
||||||
|
let path = self.backup_dir.join(&config.name);
|
||||||
|
|
||||||
|
// If the directory already exists, that's okay
|
||||||
|
match std::fs::create_dir(&path) {
|
||||||
|
Ok(()) => (),
|
||||||
|
Err(e) => match e.kind() {
|
||||||
|
io::ErrorKind::AlreadyExists => (),
|
||||||
|
_ => return Err(e),
|
||||||
|
},
|
||||||
|
};
|
||||||
|
|
||||||
|
let mut manager = Manager::new(
|
||||||
|
path,
|
||||||
|
self.dirs.clone(),
|
||||||
|
self.default_metadata.clone(),
|
||||||
|
config.chain_len,
|
||||||
|
config.chains,
|
||||||
|
chrono::Duration::minutes(config.frequency.into()),
|
||||||
|
);
|
||||||
|
manager.load()?;
|
||||||
|
self.managers.insert(config.name.clone(), manager);
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Convenient wrapper for `add`.
|
||||||
|
pub fn add_all(&mut self, configs: &Vec<ManagerConfig>) -> io::Result<()> {
|
||||||
|
for config in configs {
|
||||||
|
self.add(config)?;
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Return the name of the next scheduled layer, if one or more managers are present.
|
||||||
|
pub fn next_scheduled_layer(&self) -> Option<&str> {
|
||||||
|
self.managers
|
||||||
|
.iter()
|
||||||
|
.min_by_key(|(_, m)| m.next_scheduled_time())
|
||||||
|
.map(|(k, _)| k.as_str())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Return the earliest scheduled time for the underlying managers.
|
||||||
|
pub fn next_scheduled_time(&self) -> Option<chrono::DateTime<Utc>> {
|
||||||
|
self.managers
|
||||||
|
.values()
|
||||||
|
.map(|m| m.next_scheduled_time())
|
||||||
|
.min()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Perform a backup cycle for the earliest scheduled manager.
|
||||||
|
pub fn perform_backup_cycle(&mut self) -> io::Result<()> {
|
||||||
|
if let Some(manager) = self
|
||||||
|
.managers
|
||||||
|
.values_mut()
|
||||||
|
.min_by_key(|m| m.next_scheduled_time())
|
||||||
|
{
|
||||||
|
manager.create_backup()?;
|
||||||
|
manager.remove_old_backups()
|
||||||
|
} else {
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a manual backup for a specific layer
|
||||||
|
pub fn create_backup(&mut self, layer: &str) -> Option<io::Result<()>> {
|
||||||
|
if let Some(manager) = self.managers.get_mut(layer) {
|
||||||
|
let mut res = manager.create_backup();
|
||||||
|
|
||||||
|
if res.is_ok() {
|
||||||
|
res = manager.remove_old_backups();
|
||||||
|
}
|
||||||
|
|
||||||
|
Some(res)
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Restore a backup for a specific layer
|
||||||
|
pub fn restore_backup(
|
||||||
|
&self,
|
||||||
|
layer: &str,
|
||||||
|
start_time: chrono::DateTime<Utc>,
|
||||||
|
dirs: &Vec<(PathBuf, PathBuf)>,
|
||||||
|
) -> Option<io::Result<()>> {
|
||||||
|
self.managers
|
||||||
|
.get(layer)
|
||||||
|
.map(|manager| manager.restore_backup(start_time, dirs))
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn export_backup<P: AsRef<Path>>(
|
||||||
|
&self,
|
||||||
|
layer: &str,
|
||||||
|
start_time: chrono::DateTime<Utc>,
|
||||||
|
output_path: P,
|
||||||
|
) -> Option<io::Result<()>> {
|
||||||
|
self.managers
|
||||||
|
.get(layer)
|
||||||
|
.map(|manager| manager.export_backup(start_time, output_path))
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn managers(&self) -> &HashMap<String, Manager<T>> {
|
||||||
|
&self.managers
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,253 @@
|
||||||
|
mod config;
|
||||||
|
mod meta;
|
||||||
|
|
||||||
|
use std::{
|
||||||
|
fs::{File, OpenOptions},
|
||||||
|
io,
|
||||||
|
path::{Path, PathBuf},
|
||||||
|
};
|
||||||
|
|
||||||
|
use chrono::{SubsecRound, Utc};
|
||||||
|
use flate2::{write::GzEncoder, Compression};
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use super::{Backup, BackupType, Delta, State};
|
||||||
|
use crate::other;
|
||||||
|
pub use config::ManagerConfig;
|
||||||
|
pub use meta::MetaManager;
|
||||||
|
|
||||||
|
/// Manages a single backup layer consisting of one or more chains of backups.
|
||||||
|
pub struct Manager<T>
|
||||||
|
where
|
||||||
|
T: Clone + Serialize + for<'de> Deserialize<'de> + std::fmt::Debug,
|
||||||
|
{
|
||||||
|
backup_dir: PathBuf,
|
||||||
|
dirs: Vec<(PathBuf, PathBuf)>,
|
||||||
|
default_metadata: T,
|
||||||
|
chain_len: u64,
|
||||||
|
chains_to_keep: u64,
|
||||||
|
frequency: chrono::Duration,
|
||||||
|
chains: Vec<Vec<Backup<T>>>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T> Manager<T>
|
||||||
|
where
|
||||||
|
T: Clone + Serialize + for<'de> Deserialize<'de> + std::fmt::Debug,
|
||||||
|
{
|
||||||
|
const METADATA_FILE: &str = "alex.json";
|
||||||
|
|
||||||
|
pub fn new<P: Into<PathBuf>>(
|
||||||
|
backup_dir: P,
|
||||||
|
dirs: Vec<(PathBuf, PathBuf)>,
|
||||||
|
metadata: T,
|
||||||
|
chain_len: u64,
|
||||||
|
chains_to_keep: u64,
|
||||||
|
frequency: chrono::Duration,
|
||||||
|
) -> Self {
|
||||||
|
Self {
|
||||||
|
backup_dir: backup_dir.into(),
|
||||||
|
dirs,
|
||||||
|
default_metadata: metadata,
|
||||||
|
chain_len,
|
||||||
|
chains_to_keep,
|
||||||
|
frequency,
|
||||||
|
chains: Vec::new(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a new backup, either full or incremental, depending on the state of the current
|
||||||
|
/// chain.
|
||||||
|
pub fn create_backup(&mut self) -> io::Result<()> {
|
||||||
|
// We start a new chain if the current chain is complete, or if there isn't a first chain
|
||||||
|
// yet
|
||||||
|
if let Some(current_chain) = self.chains.last() {
|
||||||
|
let current_chain_len: u64 = current_chain.len().try_into().unwrap();
|
||||||
|
|
||||||
|
if current_chain_len >= self.chain_len {
|
||||||
|
self.chains.push(Vec::new());
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
self.chains.push(Vec::new());
|
||||||
|
}
|
||||||
|
|
||||||
|
let current_chain = self.chains.last_mut().unwrap();
|
||||||
|
|
||||||
|
let mut backup = if !current_chain.is_empty() {
|
||||||
|
let previous_backup = current_chain.last().unwrap();
|
||||||
|
let previous_state = State::from(current_chain.iter().map(|b| &b.delta));
|
||||||
|
|
||||||
|
Backup::create_from(
|
||||||
|
previous_state,
|
||||||
|
previous_backup.start_time,
|
||||||
|
&self.backup_dir,
|
||||||
|
&self.dirs,
|
||||||
|
)?
|
||||||
|
} else {
|
||||||
|
Backup::create(&self.backup_dir, &self.dirs)?
|
||||||
|
};
|
||||||
|
|
||||||
|
backup.set_metadata(self.default_metadata.clone());
|
||||||
|
|
||||||
|
current_chain.push(backup);
|
||||||
|
|
||||||
|
self.save()?;
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Delete all backups associated with outdated chains, and forget those chains.
|
||||||
|
pub fn remove_old_backups(&mut self) -> io::Result<()> {
|
||||||
|
let chains_to_store: usize = self.chains_to_keep.try_into().unwrap();
|
||||||
|
|
||||||
|
if chains_to_store < self.chains.len() {
|
||||||
|
let mut remove_count: usize = self.chains.len() - chains_to_store;
|
||||||
|
|
||||||
|
// We only count finished chains towards the list of stored chains
|
||||||
|
let chain_len: usize = self.chain_len.try_into().unwrap();
|
||||||
|
if self.chains.last().unwrap().len() < chain_len {
|
||||||
|
remove_count -= 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
for chain in self.chains.drain(..remove_count) {
|
||||||
|
for backup in chain {
|
||||||
|
let path = Backup::path(&self.backup_dir, backup.start_time);
|
||||||
|
std::fs::remove_file(path)?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
self.save()?;
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Write the in-memory state to disk.
|
||||||
|
pub fn save(&self) -> io::Result<()> {
|
||||||
|
let json_file = File::create(self.backup_dir.join(Self::METADATA_FILE))?;
|
||||||
|
serde_json::to_writer(json_file, &self.chains)?;
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Overwrite the in-memory state with the on-disk state.
|
||||||
|
pub fn load(&mut self) -> io::Result<()> {
|
||||||
|
let json_file = match File::open(self.backup_dir.join(Self::METADATA_FILE)) {
|
||||||
|
Ok(f) => f,
|
||||||
|
Err(e) => {
|
||||||
|
// Don't error out if the file isn't there, it will be created when necessary
|
||||||
|
if e.kind() == io::ErrorKind::NotFound {
|
||||||
|
self.chains = Vec::new();
|
||||||
|
|
||||||
|
return Ok(());
|
||||||
|
} else {
|
||||||
|
return Err(e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
self.chains = serde_json::from_reader(json_file)?;
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Calculate the next time a backup should be created. If no backup has been created yet, it
|
||||||
|
/// will return now.
|
||||||
|
pub fn next_scheduled_time(&self) -> chrono::DateTime<Utc> {
|
||||||
|
self.chains
|
||||||
|
.last()
|
||||||
|
.and_then(|last_chain| last_chain.last())
|
||||||
|
.map(|last_backup| last_backup.start_time + self.frequency)
|
||||||
|
.unwrap_or_else(chrono::offset::Utc::now)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Search for a chain containing a backup with the specified start time.
|
||||||
|
///
|
||||||
|
/// # Returns
|
||||||
|
///
|
||||||
|
/// A tuple (chain, index) with index being the index of the found backup in the returned
|
||||||
|
/// chain.
|
||||||
|
fn find(&self, start_time: chrono::DateTime<Utc>) -> Option<(&Vec<Backup<T>>, usize)> {
|
||||||
|
for chain in &self.chains {
|
||||||
|
if let Some(index) = chain
|
||||||
|
.iter()
|
||||||
|
.position(|b| b.start_time.trunc_subsecs(0) == start_time)
|
||||||
|
{
|
||||||
|
return Some((chain, index));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
None
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Restore the backup with the given start time by restoring its chain up to and including the
|
||||||
|
/// backup, in order.
|
||||||
|
pub fn restore_backup(
|
||||||
|
&self,
|
||||||
|
start_time: chrono::DateTime<Utc>,
|
||||||
|
dirs: &Vec<(PathBuf, PathBuf)>,
|
||||||
|
) -> io::Result<()> {
|
||||||
|
self.find(start_time)
|
||||||
|
.ok_or_else(|| other("Unknown layer."))
|
||||||
|
.and_then(|(chain, index)| {
|
||||||
|
for backup in chain.iter().take(index + 1) {
|
||||||
|
backup.restore(&self.backup_dir, dirs)?;
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Export the backup with the given start time as a new full archive.
|
||||||
|
pub fn export_backup<P: AsRef<Path>>(
|
||||||
|
&self,
|
||||||
|
start_time: chrono::DateTime<Utc>,
|
||||||
|
output_path: P,
|
||||||
|
) -> io::Result<()> {
|
||||||
|
self.find(start_time)
|
||||||
|
.ok_or_else(|| other("Unknown layer."))
|
||||||
|
.and_then(|(chain, index)| {
|
||||||
|
match chain[index].type_ {
|
||||||
|
// A full backup is simply copied to the output path
|
||||||
|
BackupType::Full => std::fs::copy(
|
||||||
|
Backup::path(&self.backup_dir, chain[index].start_time),
|
||||||
|
output_path,
|
||||||
|
)
|
||||||
|
.map(|_| ()),
|
||||||
|
// Incremental backups are exported one by one according to their contribution
|
||||||
|
BackupType::Incremental => {
|
||||||
|
let contributions = Delta::contributions(
|
||||||
|
chain.iter().take(index + 1).map(|b| &b.delta).rev(),
|
||||||
|
);
|
||||||
|
|
||||||
|
let tar_gz = OpenOptions::new()
|
||||||
|
.write(true)
|
||||||
|
.create(true)
|
||||||
|
.open(output_path.as_ref())?;
|
||||||
|
let enc = GzEncoder::new(tar_gz, Compression::default());
|
||||||
|
let mut ar = tar::Builder::new(enc);
|
||||||
|
|
||||||
|
// We only need to consider backups that have a non-empty contribution.
|
||||||
|
// This allows us to skip reading backups that have been completely
|
||||||
|
// overwritten by their successors anyways.
|
||||||
|
for (contribution, backup) in contributions
|
||||||
|
.iter()
|
||||||
|
.rev()
|
||||||
|
.zip(chain.iter().take(index + 1))
|
||||||
|
.filter(|(contribution, _)| !contribution.is_empty())
|
||||||
|
{
|
||||||
|
println!("{}", &backup);
|
||||||
|
backup.append(&self.backup_dir, contribution, &mut ar)?;
|
||||||
|
}
|
||||||
|
|
||||||
|
let mut enc = ar.into_inner()?;
|
||||||
|
enc.try_finish()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Get a reference to the underlying chains
|
||||||
|
pub fn chains(&self) -> &Vec<Vec<Backup<T>>> {
|
||||||
|
&self.chains
|
||||||
|
}
|
||||||
|
}
|
|
@ -1,19 +1,29 @@
|
||||||
mod delta;
|
mod delta;
|
||||||
mod manager;
|
mod io_ext;
|
||||||
|
pub mod manager;
|
||||||
mod path;
|
mod path;
|
||||||
|
mod state;
|
||||||
|
|
||||||
|
use std::{
|
||||||
|
collections::HashSet,
|
||||||
|
fmt,
|
||||||
|
fs::File,
|
||||||
|
io,
|
||||||
|
path::{Path, PathBuf},
|
||||||
|
};
|
||||||
|
|
||||||
|
use chrono::Utc;
|
||||||
|
use flate2::{read::GzDecoder, write::GzEncoder, Compression};
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
use delta::Delta;
|
use delta::Delta;
|
||||||
pub use manager::Manager;
|
pub use manager::Manager;
|
||||||
|
pub use manager::ManagerConfig;
|
||||||
use chrono::Utc;
|
pub use manager::MetaManager;
|
||||||
use flate2::write::GzEncoder;
|
|
||||||
use flate2::Compression;
|
|
||||||
use path::PathExt;
|
use path::PathExt;
|
||||||
use serde::{Deserialize, Serialize};
|
pub use state::State;
|
||||||
use std::collections::{HashMap, HashSet};
|
|
||||||
use std::fs::File;
|
const BYTE_SUFFIXES: [&str; 5] = ["B", "KiB", "MiB", "GiB", "TiB"];
|
||||||
use std::io;
|
|
||||||
use std::path::{Path, PathBuf};
|
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Serialize, Deserialize)]
|
#[derive(Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub enum BackupType {
|
pub enum BackupType {
|
||||||
|
@ -22,10 +32,13 @@ pub enum BackupType {
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Represents a successful backup
|
/// Represents a successful backup
|
||||||
#[derive(Serialize, Deserialize)]
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
pub struct Backup<T: Clone> {
|
pub struct Backup<T: Clone> {
|
||||||
/// When the backup was started (also corresponds to the name)
|
/// When the backup was started (also corresponds to the name)
|
||||||
pub start_time: chrono::DateTime<Utc>,
|
pub start_time: chrono::DateTime<Utc>,
|
||||||
|
/// When the backup finished
|
||||||
|
pub end_time: chrono::DateTime<Utc>,
|
||||||
|
pub size: usize,
|
||||||
/// Type of the backup
|
/// Type of the backup
|
||||||
pub type_: BackupType,
|
pub type_: BackupType,
|
||||||
pub delta: Delta,
|
pub delta: Delta,
|
||||||
|
@ -34,6 +47,8 @@ pub struct Backup<T: Clone> {
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Backup<()> {
|
impl Backup<()> {
|
||||||
|
pub const FILENAME_FORMAT: &str = "%Y-%m-%d_%H-%M-%S.tar.gz";
|
||||||
|
|
||||||
/// Return the path to a backup file by properly formatting the data.
|
/// Return the path to a backup file by properly formatting the data.
|
||||||
pub fn path<P: AsRef<Path>>(backup_dir: P, start_time: chrono::DateTime<Utc>) -> PathBuf {
|
pub fn path<P: AsRef<Path>>(backup_dir: P, start_time: chrono::DateTime<Utc>) -> PathBuf {
|
||||||
let backup_dir = backup_dir.as_ref();
|
let backup_dir = backup_dir.as_ref();
|
||||||
|
@ -41,27 +56,53 @@ impl Backup<()> {
|
||||||
let filename = format!("{}", start_time.format(Self::FILENAME_FORMAT));
|
let filename = format!("{}", start_time.format(Self::FILENAME_FORMAT));
|
||||||
backup_dir.join(filename)
|
backup_dir.join(filename)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Extract an archive.
|
||||||
|
///
|
||||||
|
/// # Arguments
|
||||||
|
///
|
||||||
|
/// * `backup_path` - Path to the archive to extract
|
||||||
|
/// * `dirs` - list of tuples `(path_in_tar, dst_dir)` with `dst_dir` the directory on-disk
|
||||||
|
/// where the files stored under `path_in_tar` inside the tarball should be extracted to.
|
||||||
|
pub fn extract_archive<P: AsRef<Path>>(
|
||||||
|
archive_path: P,
|
||||||
|
dirs: &Vec<(PathBuf, PathBuf)>,
|
||||||
|
) -> io::Result<()> {
|
||||||
|
let tar_gz = File::open(archive_path)?;
|
||||||
|
let enc = GzDecoder::new(tar_gz);
|
||||||
|
let mut ar = tar::Archive::new(enc);
|
||||||
|
|
||||||
|
// Unpack each file by matching it with one of the destination directories and extracting
|
||||||
|
// it to the right path
|
||||||
|
for entry in ar.entries()? {
|
||||||
|
let mut entry = entry?;
|
||||||
|
let entry_path_in_tar = entry.path()?.to_path_buf();
|
||||||
|
|
||||||
|
for (path_in_tar, dst_dir) in dirs {
|
||||||
|
if entry_path_in_tar.starts_with(path_in_tar) {
|
||||||
|
let dst_path =
|
||||||
|
dst_dir.join(entry_path_in_tar.strip_prefix(path_in_tar).unwrap());
|
||||||
|
|
||||||
|
// Ensure all parent directories are present
|
||||||
|
std::fs::create_dir_all(dst_path.parent().unwrap())?;
|
||||||
|
|
||||||
|
entry.unpack(dst_path)?;
|
||||||
|
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<T: Clone> Backup<T> {
|
impl<T: Clone> Backup<T> {
|
||||||
const FILENAME_FORMAT: &str = "%Y-%m-%d_%H-%M-%S.tar.gz";
|
/// Set the backup's metadata.
|
||||||
|
|
||||||
pub fn set_metadata(&mut self, metadata: T) {
|
pub fn set_metadata(&mut self, metadata: T) {
|
||||||
self.metadata = Some(metadata);
|
self.metadata = Some(metadata);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Resolve the state of the list of backups by applying their deltas in-order to an initially
|
|
||||||
/// empty state.
|
|
||||||
pub fn state(backups: &Vec<Self>) -> HashMap<PathBuf, HashSet<PathBuf>> {
|
|
||||||
let mut state: HashMap<PathBuf, HashSet<PathBuf>> = HashMap::new();
|
|
||||||
|
|
||||||
for backup in backups {
|
|
||||||
backup.delta.apply(&mut state);
|
|
||||||
}
|
|
||||||
|
|
||||||
state
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Create a new Full backup, populated with the given directories.
|
/// Create a new Full backup, populated with the given directories.
|
||||||
///
|
///
|
||||||
/// # Arguments
|
/// # Arguments
|
||||||
|
@ -75,12 +116,12 @@ impl<T: Clone> Backup<T> {
|
||||||
/// The `Backup` instance describing this new backup.
|
/// The `Backup` instance describing this new backup.
|
||||||
pub fn create<P: AsRef<Path>>(
|
pub fn create<P: AsRef<Path>>(
|
||||||
backup_dir: P,
|
backup_dir: P,
|
||||||
dirs: Vec<(PathBuf, PathBuf)>,
|
dirs: &Vec<(PathBuf, PathBuf)>,
|
||||||
) -> io::Result<Self> {
|
) -> io::Result<Self> {
|
||||||
let start_time = chrono::offset::Utc::now();
|
let start_time = chrono::offset::Utc::now();
|
||||||
|
|
||||||
let path = Backup::path(backup_dir, start_time);
|
let path = Backup::path(backup_dir, start_time);
|
||||||
let tar_gz = File::create(path)?;
|
let tar_gz = io_ext::CountingWrite::new(File::create(path)?);
|
||||||
let enc = GzEncoder::new(tar_gz, Compression::default());
|
let enc = GzEncoder::new(tar_gz, Compression::default());
|
||||||
let mut ar = tar::Builder::new(enc);
|
let mut ar = tar::Builder::new(enc);
|
||||||
|
|
||||||
|
@ -91,34 +132,54 @@ impl<T: Clone> Backup<T> {
|
||||||
|
|
||||||
for entry in src_dir.read_dir_recursive()?.ignored("cache").files() {
|
for entry in src_dir.read_dir_recursive()?.ignored("cache").files() {
|
||||||
let path = entry?.path();
|
let path = entry?.path();
|
||||||
let stripped = path.strip_prefix(&src_dir).unwrap();
|
let stripped = path.strip_prefix(src_dir).unwrap();
|
||||||
|
|
||||||
ar.append_path_with_name(&path, dir_in_tar.join(stripped))?;
|
ar.append_path_with_name(&path, dir_in_tar.join(stripped))?;
|
||||||
added_files.insert(stripped.to_path_buf());
|
added_files.insert(stripped.to_path_buf());
|
||||||
}
|
}
|
||||||
|
|
||||||
delta.added.insert(dir_in_tar, added_files);
|
delta.added.insert(dir_in_tar.to_path_buf(), added_files);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
let mut enc = ar.into_inner()?;
|
||||||
|
|
||||||
|
// The docs recommend running try_finish before unwrapping using finish
|
||||||
|
enc.try_finish()?;
|
||||||
|
let tar_gz = enc.finish()?;
|
||||||
|
|
||||||
Ok(Backup {
|
Ok(Backup {
|
||||||
type_: BackupType::Full,
|
type_: BackupType::Full,
|
||||||
start_time,
|
start_time,
|
||||||
|
end_time: chrono::Utc::now(),
|
||||||
|
size: tar_gz.bytes_written(),
|
||||||
delta,
|
delta,
|
||||||
metadata: None,
|
metadata: None,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Create a new incremental backup from a given previous backup
|
/// Create a new Incremental backup from the given state, populated with the given directories.
|
||||||
|
///
|
||||||
|
/// # Arguments
|
||||||
|
///
|
||||||
|
/// * `previous_state` - State the file system was in during the previous backup in the chain
|
||||||
|
/// * `previous_start_time` - Start time of the previous backup; used to filter files
|
||||||
|
/// * `backup_dir` - Directory to store archive in
|
||||||
|
/// * `dirs` - list of tuples `(path_in_tar, src_dir)` with `path_in_tar` the directory name
|
||||||
|
/// under which `src_dir`'s contents should be stored in the archive
|
||||||
|
///
|
||||||
|
/// # Returns
|
||||||
|
///
|
||||||
|
/// The `Backup` instance describing this new backup.
|
||||||
pub fn create_from<P: AsRef<Path>>(
|
pub fn create_from<P: AsRef<Path>>(
|
||||||
previous_state: HashMap<PathBuf, HashSet<PathBuf>>,
|
previous_state: State,
|
||||||
previous_start_time: chrono::DateTime<Utc>,
|
previous_start_time: chrono::DateTime<Utc>,
|
||||||
backup_dir: P,
|
backup_dir: P,
|
||||||
dirs: Vec<(PathBuf, PathBuf)>,
|
dirs: &Vec<(PathBuf, PathBuf)>,
|
||||||
) -> io::Result<Self> {
|
) -> io::Result<Self> {
|
||||||
let start_time = chrono::offset::Utc::now();
|
let start_time = chrono::offset::Utc::now();
|
||||||
|
|
||||||
let path = Backup::path(backup_dir, start_time);
|
let path = Backup::path(backup_dir, start_time);
|
||||||
let tar_gz = File::create(path)?;
|
let tar_gz = io_ext::CountingWrite::new(File::create(path)?);
|
||||||
let enc = GzEncoder::new(tar_gz, Compression::default());
|
let enc = GzEncoder::new(tar_gz, Compression::default());
|
||||||
let mut ar = tar::Builder::new(enc);
|
let mut ar = tar::Builder::new(enc);
|
||||||
|
|
||||||
|
@ -130,7 +191,7 @@ impl<T: Clone> Backup<T> {
|
||||||
|
|
||||||
for entry in src_dir.read_dir_recursive()?.ignored("cache").files() {
|
for entry in src_dir.read_dir_recursive()?.ignored("cache").files() {
|
||||||
let path = entry?.path();
|
let path = entry?.path();
|
||||||
let stripped = path.strip_prefix(&src_dir).unwrap();
|
let stripped = path.strip_prefix(src_dir).unwrap();
|
||||||
|
|
||||||
if !path.not_modified_since(previous_start_time) {
|
if !path.not_modified_since(previous_start_time) {
|
||||||
ar.append_path_with_name(&path, dir_in_tar.join(stripped))?;
|
ar.append_path_with_name(&path, dir_in_tar.join(stripped))?;
|
||||||
|
@ -142,19 +203,113 @@ impl<T: Clone> Backup<T> {
|
||||||
|
|
||||||
delta.added.insert(dir_in_tar.clone(), added_files);
|
delta.added.insert(dir_in_tar.clone(), added_files);
|
||||||
|
|
||||||
if let Some(previous_files) = previous_state.get(&dir_in_tar) {
|
if let Some(previous_files) = previous_state.get(dir_in_tar) {
|
||||||
delta.removed.insert(
|
delta.removed.insert(
|
||||||
dir_in_tar,
|
dir_in_tar.to_path_buf(),
|
||||||
previous_files.difference(&all_files).cloned().collect(),
|
previous_files.difference(&all_files).cloned().collect(),
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
let mut enc = ar.into_inner()?;
|
||||||
|
|
||||||
|
// The docs recommend running try_finish before unwrapping using finish
|
||||||
|
enc.try_finish()?;
|
||||||
|
let tar_gz = enc.finish()?;
|
||||||
|
|
||||||
Ok(Backup {
|
Ok(Backup {
|
||||||
type_: BackupType::Incremental,
|
type_: BackupType::Incremental,
|
||||||
start_time,
|
start_time,
|
||||||
|
end_time: chrono::Utc::now(),
|
||||||
|
size: tar_gz.bytes_written(),
|
||||||
delta,
|
delta,
|
||||||
metadata: None,
|
metadata: None,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Restore the backup by extracting its contents to the respective directories.
|
||||||
|
///
|
||||||
|
/// # Arguments
|
||||||
|
///
|
||||||
|
/// * `backup_dir` - Backup directory where the file is stored
|
||||||
|
/// * `dirs` - list of tuples `(path_in_tar, dst_dir)` with `dst_dir` the directory on-disk
|
||||||
|
/// where the files stored under `path_in_tar` inside the tarball should be extracted to.
|
||||||
|
pub fn restore<P: AsRef<Path>>(
|
||||||
|
&self,
|
||||||
|
backup_dir: P,
|
||||||
|
dirs: &Vec<(PathBuf, PathBuf)>,
|
||||||
|
) -> io::Result<()> {
|
||||||
|
let backup_path = Backup::path(backup_dir, self.start_time);
|
||||||
|
Backup::extract_archive(backup_path, dirs)?;
|
||||||
|
|
||||||
|
// Remove any files
|
||||||
|
for (path_in_tar, dst_dir) in dirs {
|
||||||
|
if let Some(removed) = self.delta.removed.get(path_in_tar) {
|
||||||
|
for path in removed {
|
||||||
|
let dst_path = dst_dir.join(path);
|
||||||
|
std::fs::remove_file(dst_path)?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn open<P: AsRef<Path>>(&self, backup_dir: P) -> io::Result<tar::Archive<GzDecoder<File>>> {
|
||||||
|
let path = Backup::path(backup_dir, self.start_time);
|
||||||
|
let tar_gz = File::open(path)?;
|
||||||
|
let enc = GzDecoder::new(tar_gz);
|
||||||
|
Ok(tar::Archive::new(enc))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Open this backup's archive and append all its files that are part of the provided state to
|
||||||
|
/// the archive file.
|
||||||
|
pub fn append<P: AsRef<Path>>(
|
||||||
|
&self,
|
||||||
|
backup_dir: P,
|
||||||
|
state: &State,
|
||||||
|
ar: &mut tar::Builder<GzEncoder<File>>,
|
||||||
|
) -> io::Result<()> {
|
||||||
|
let mut own_ar = self.open(backup_dir)?;
|
||||||
|
|
||||||
|
for entry in own_ar.entries()? {
|
||||||
|
let entry = entry?;
|
||||||
|
let entry_path_in_tar = entry.path()?.to_path_buf();
|
||||||
|
|
||||||
|
if state.contains(&entry_path_in_tar) {
|
||||||
|
let header = entry.header().clone();
|
||||||
|
ar.append(&header, entry)?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T: Clone> fmt::Display for Backup<T> {
|
||||||
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||||
|
let letter = match self.type_ {
|
||||||
|
BackupType::Full => 'F',
|
||||||
|
BackupType::Incremental => 'I',
|
||||||
|
};
|
||||||
|
|
||||||
|
// Pretty-print size
|
||||||
|
// If your backup is a petabyte or larger, this will crash and you need to re-evaluate your
|
||||||
|
// life choices
|
||||||
|
let index = self.size.ilog(1024) as usize;
|
||||||
|
let size = self.size as f64 / (1024.0_f64.powi(index as i32));
|
||||||
|
let duration = self.end_time - self.start_time;
|
||||||
|
|
||||||
|
write!(
|
||||||
|
f,
|
||||||
|
"{} ({}, {}m{}s, {:.2}{}, {})",
|
||||||
|
self.start_time.format(Backup::FILENAME_FORMAT),
|
||||||
|
letter,
|
||||||
|
duration.num_seconds() / 60,
|
||||||
|
duration.num_seconds() % 60,
|
||||||
|
size,
|
||||||
|
BYTE_SUFFIXES[index],
|
||||||
|
self.delta
|
||||||
|
)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,9 +1,12 @@
|
||||||
|
use std::{
|
||||||
|
collections::HashSet,
|
||||||
|
ffi::OsString,
|
||||||
|
fs::{self, DirEntry},
|
||||||
|
io,
|
||||||
|
path::{Path, PathBuf},
|
||||||
|
};
|
||||||
|
|
||||||
use chrono::{Local, Utc};
|
use chrono::{Local, Utc};
|
||||||
use std::collections::HashSet;
|
|
||||||
use std::ffi::OsString;
|
|
||||||
use std::fs::DirEntry;
|
|
||||||
use std::path::{Path, PathBuf};
|
|
||||||
use std::{fs, io};
|
|
||||||
|
|
||||||
pub struct ReadDirRecursive {
|
pub struct ReadDirRecursive {
|
||||||
ignored: HashSet<OsString>,
|
ignored: HashSet<OsString>,
|
||||||
|
@ -129,16 +132,15 @@ pub trait PathExt {
|
||||||
|
|
||||||
impl PathExt for Path {
|
impl PathExt for Path {
|
||||||
fn not_modified_since(&self, timestamp: chrono::DateTime<Utc>) -> bool {
|
fn not_modified_since(&self, timestamp: chrono::DateTime<Utc>) -> bool {
|
||||||
if let Ok(metadata) = self.metadata() {
|
self.metadata()
|
||||||
if let Ok(last_modified) = metadata.modified() {
|
.and_then(|m| m.modified())
|
||||||
|
.map(|last_modified| {
|
||||||
let t: chrono::DateTime<Utc> = last_modified.into();
|
let t: chrono::DateTime<Utc> = last_modified.into();
|
||||||
let t = t.with_timezone(&Local);
|
let t = t.with_timezone(&Local);
|
||||||
|
|
||||||
return t < timestamp;
|
t < timestamp
|
||||||
}
|
})
|
||||||
}
|
.unwrap_or(false)
|
||||||
|
|
||||||
false
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn read_dir_recursive(&self) -> io::Result<ReadDirRecursive> {
|
fn read_dir_recursive(&self) -> io::Result<ReadDirRecursive> {
|
||||||
|
|
|
@ -0,0 +1,98 @@
|
||||||
|
use std::{
|
||||||
|
borrow::Borrow,
|
||||||
|
collections::{HashMap, HashSet},
|
||||||
|
ops::{Deref, DerefMut},
|
||||||
|
path::{Path, PathBuf},
|
||||||
|
};
|
||||||
|
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use crate::backup::Delta;
|
||||||
|
|
||||||
|
/// Struct that represents a current state for a backup. This struct acts as a smart pointer around
|
||||||
|
/// a HashMap.
|
||||||
|
#[derive(Debug, Clone, PartialEq, Eq, Serialize, Deserialize)]
|
||||||
|
pub struct State(HashMap<PathBuf, HashSet<PathBuf>>);
|
||||||
|
|
||||||
|
impl State {
|
||||||
|
pub fn new() -> Self {
|
||||||
|
State(HashMap::new())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Apply the delta to the current state.
|
||||||
|
pub fn apply(&mut self, delta: &Delta) {
|
||||||
|
// First we add new files, then we remove the old ones
|
||||||
|
for (dir, added) in delta.added.iter() {
|
||||||
|
if let Some(current) = self.0.get_mut(dir) {
|
||||||
|
current.extend(added.iter().cloned());
|
||||||
|
} else {
|
||||||
|
self.0.insert(dir.clone(), added.clone());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for (dir, removed) in delta.removed.iter() {
|
||||||
|
if let Some(current) = self.0.get_mut(dir) {
|
||||||
|
current.retain(|k| !removed.contains(k));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Returns whether the provided relative path is part of the given state.
|
||||||
|
pub fn contains<P: AsRef<Path>>(&self, path: P) -> bool {
|
||||||
|
let path = path.as_ref();
|
||||||
|
|
||||||
|
self.0.iter().any(|(dir, files)| {
|
||||||
|
path.starts_with(dir) && files.contains(path.strip_prefix(dir).unwrap())
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Returns whether the state is empty.
|
||||||
|
///
|
||||||
|
/// Note that this does not necessarily mean that the state does not contain any sets, but
|
||||||
|
/// rather that any sets that it does contain are also empty.
|
||||||
|
pub fn is_empty(&self) -> bool {
|
||||||
|
self.0.values().all(|s| s.is_empty())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T> From<T> for State
|
||||||
|
where
|
||||||
|
T: IntoIterator,
|
||||||
|
T::Item: Borrow<Delta>,
|
||||||
|
{
|
||||||
|
fn from(deltas: T) -> Self {
|
||||||
|
let mut state = State::new();
|
||||||
|
|
||||||
|
for delta in deltas {
|
||||||
|
state.apply(delta.borrow());
|
||||||
|
}
|
||||||
|
|
||||||
|
state
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl AsRef<HashMap<PathBuf, HashSet<PathBuf>>> for State {
|
||||||
|
fn as_ref(&self) -> &HashMap<PathBuf, HashSet<PathBuf>> {
|
||||||
|
&self.0
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Deref for State {
|
||||||
|
type Target = HashMap<PathBuf, HashSet<PathBuf>>;
|
||||||
|
|
||||||
|
fn deref(&self) -> &Self::Target {
|
||||||
|
&self.0
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl DerefMut for State {
|
||||||
|
fn deref_mut(&mut self) -> &mut Self::Target {
|
||||||
|
&mut self.0
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Default for State {
|
||||||
|
fn default() -> Self {
|
||||||
|
Self::new()
|
||||||
|
}
|
||||||
|
}
|
106
src/cli.rs
106
src/cli.rs
|
@ -1,106 +0,0 @@
|
||||||
use crate::server::ServerType;
|
|
||||||
use clap::{Args, Parser, Subcommand};
|
|
||||||
use std::path::PathBuf;
|
|
||||||
|
|
||||||
#[derive(Parser)]
|
|
||||||
#[command(author, version, about, long_about = None)]
|
|
||||||
pub struct Cli {
|
|
||||||
#[command(subcommand)]
|
|
||||||
pub command: Commands,
|
|
||||||
/// Directory where configs are stored, and where the server will run
|
|
||||||
#[arg(
|
|
||||||
long,
|
|
||||||
value_name = "CONFIG_DIR",
|
|
||||||
default_value = ".",
|
|
||||||
env = "ALEX_CONFIG_DIR",
|
|
||||||
global = true
|
|
||||||
)]
|
|
||||||
pub config: PathBuf,
|
|
||||||
/// Directory where world files will be saved
|
|
||||||
#[arg(
|
|
||||||
long,
|
|
||||||
value_name = "WORLD_DIR",
|
|
||||||
default_value = "../worlds",
|
|
||||||
env = "ALEX_WORLD_DIR",
|
|
||||||
global = true
|
|
||||||
)]
|
|
||||||
pub world: PathBuf,
|
|
||||||
/// Directory where backups will be stored
|
|
||||||
#[arg(
|
|
||||||
long,
|
|
||||||
value_name = "BACKUP_DIR",
|
|
||||||
default_value = "../backups",
|
|
||||||
env = "ALEX_BACKUP_DIR",
|
|
||||||
global = true
|
|
||||||
)]
|
|
||||||
pub backup: PathBuf,
|
|
||||||
|
|
||||||
/// Length of a backup chain
|
|
||||||
#[arg(
|
|
||||||
short = 'l',
|
|
||||||
long,
|
|
||||||
default_value_t = 4,
|
|
||||||
env = "ALEX_CHAIN_LEN",
|
|
||||||
global = true
|
|
||||||
)]
|
|
||||||
pub chain_len: u64,
|
|
||||||
/// How many backup chains to keep
|
|
||||||
#[arg(
|
|
||||||
short = 'n',
|
|
||||||
long,
|
|
||||||
default_value_t = 7,
|
|
||||||
env = "ALEX_CHAINS",
|
|
||||||
global = true
|
|
||||||
)]
|
|
||||||
pub chains: u64,
|
|
||||||
/// Type of server
|
|
||||||
#[arg(long, default_value = "unknown", env = "ALEX_SERVER")]
|
|
||||||
pub server: ServerType,
|
|
||||||
/// Version string for the server, e.g. 1.19.4-545
|
|
||||||
#[arg(long, default_value = "", env = "ALEX_SERVER_VERSION")]
|
|
||||||
pub server_version: String,
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Subcommand)]
|
|
||||||
pub enum Commands {
|
|
||||||
/// Run the server
|
|
||||||
Run(RunArgs),
|
|
||||||
/// Create a new backup of the server. This command should only be used when the server is not
|
|
||||||
/// running.
|
|
||||||
Backup(BackupArgs),
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Args)]
|
|
||||||
pub struct RunArgs {
|
|
||||||
/// Server jar to execute
|
|
||||||
#[arg(
|
|
||||||
long,
|
|
||||||
value_name = "JAR_PATH",
|
|
||||||
default_value = "server.jar",
|
|
||||||
env = "ALEX_JAR"
|
|
||||||
)]
|
|
||||||
pub jar: PathBuf,
|
|
||||||
|
|
||||||
/// Java command to run the server jar with
|
|
||||||
#[arg(long, value_name = "JAVA_CMD", default_value_t = String::from("java"), env = "ALEX_JAVA")]
|
|
||||||
pub java: String,
|
|
||||||
|
|
||||||
/// XMS value in megabytes for the server instance
|
|
||||||
#[arg(long, default_value_t = 1024, env = "ALEX_XMS")]
|
|
||||||
pub xms: u64,
|
|
||||||
/// XMX value in megabytes for the server instance
|
|
||||||
#[arg(long, default_value_t = 2048, env = "ALEX_XMX")]
|
|
||||||
pub xmx: u64,
|
|
||||||
|
|
||||||
/// How frequently to perform a backup, in minutes; 0 to disable.
|
|
||||||
#[arg(short = 't', long, default_value_t = 0, env = "ALEX_FREQUENCY")]
|
|
||||||
pub frequency: u64,
|
|
||||||
|
|
||||||
/// Don't actually run the server, but simply output the server configuration that would have
|
|
||||||
/// been ran
|
|
||||||
#[arg(short, long, default_value_t = false)]
|
|
||||||
pub dry: bool,
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Args)]
|
|
||||||
pub struct BackupArgs {}
|
|
|
@ -0,0 +1,302 @@
|
||||||
|
use std::io;
|
||||||
|
use std::path::{Path, PathBuf};
|
||||||
|
|
||||||
|
use chrono::{TimeZone, Utc};
|
||||||
|
use clap::{Args, Subcommand};
|
||||||
|
|
||||||
|
use crate::{backup::Backup, other};
|
||||||
|
|
||||||
|
#[derive(Subcommand)]
|
||||||
|
pub enum BackupCommands {
|
||||||
|
/// List all tracked backups
|
||||||
|
///
|
||||||
|
/// Note that this will only list backups for the layers currently configured, and will ignore
|
||||||
|
/// any other layers also present in the backup directory.
|
||||||
|
List(BackupListArgs),
|
||||||
|
/// Manually create a new backup
|
||||||
|
///
|
||||||
|
/// Note that backups created using this command will count towards the length of a chain, and
|
||||||
|
/// can therefore shorten how far back in time your backups will be stored.
|
||||||
|
Create(BackupCreateArgs),
|
||||||
|
/// Restore a backup
|
||||||
|
///
|
||||||
|
/// This command will restore the selected backup by extracting its entire chain up to and
|
||||||
|
/// including the requested backup in-order.
|
||||||
|
Restore(BackupRestoreArgs),
|
||||||
|
/// Export a backup into a full archive
|
||||||
|
///
|
||||||
|
/// Just like the restore command, this will extract each backup from the chain up to and
|
||||||
|
/// including the requested backup, but instead of writing the files to disk, they will be
|
||||||
|
/// recompressed into a new tarball, resulting in a new tarball containing a full backup.
|
||||||
|
Export(BackupExportArgs),
|
||||||
|
/// Extract an archive file, which is assumed to be a full backup.
|
||||||
|
///
|
||||||
|
/// This command mostly exists as a convenience method for working with the output of `export`.
|
||||||
|
Extract(BackupExtractArgs),
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Args)]
|
||||||
|
pub struct BackupArgs {
|
||||||
|
#[command(subcommand)]
|
||||||
|
pub command: BackupCommands,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Args)]
|
||||||
|
pub struct BackupCreateArgs {
|
||||||
|
/// What layer to create a backup in
|
||||||
|
layer: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Args)]
|
||||||
|
pub struct BackupListArgs {
|
||||||
|
/// What layer to list
|
||||||
|
layer: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Args)]
|
||||||
|
pub struct BackupRestoreArgs {
|
||||||
|
/// Path to the backup inside the backup directory to restore
|
||||||
|
path: PathBuf,
|
||||||
|
/// Directory to store config in
|
||||||
|
output_config: PathBuf,
|
||||||
|
/// Directory to store worlds in
|
||||||
|
output_worlds: PathBuf,
|
||||||
|
/// Whether to overwrite the contents of the output directories
|
||||||
|
///
|
||||||
|
/// If set, the output directories will be completely cleared before trying to restore the
|
||||||
|
/// backup.
|
||||||
|
#[arg(short, long, default_value_t = false)]
|
||||||
|
force: bool,
|
||||||
|
/// Create output directories if they don't exist
|
||||||
|
#[arg(short, long, default_value_t = false)]
|
||||||
|
make: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Args)]
|
||||||
|
pub struct BackupExportArgs {
|
||||||
|
/// Path to the backup inside the backup directory to export
|
||||||
|
path: PathBuf,
|
||||||
|
/// Path to store the exported archive
|
||||||
|
output: PathBuf,
|
||||||
|
/// Create output directories if they don't exist
|
||||||
|
#[arg(short, long, default_value_t = false)]
|
||||||
|
make: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Args)]
|
||||||
|
pub struct BackupExtractArgs {
|
||||||
|
/// Path to the backup to extract
|
||||||
|
path: PathBuf,
|
||||||
|
/// Directory to store config in
|
||||||
|
output_config: PathBuf,
|
||||||
|
/// Directory to store worlds in
|
||||||
|
output_worlds: PathBuf,
|
||||||
|
/// Whether to overwrite the contents of the output directories
|
||||||
|
///
|
||||||
|
/// If set, the output directories will be completely cleared before trying to restore the
|
||||||
|
/// backup.
|
||||||
|
#[arg(short, long, default_value_t = false)]
|
||||||
|
force: bool,
|
||||||
|
/// Create output directories if they don't exist
|
||||||
|
#[arg(short, long, default_value_t = false)]
|
||||||
|
make: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl BackupArgs {
|
||||||
|
pub fn run(&self, cli: &super::Config) -> io::Result<()> {
|
||||||
|
match &self.command {
|
||||||
|
BackupCommands::Create(args) => args.run(cli),
|
||||||
|
BackupCommands::List(args) => args.run(cli),
|
||||||
|
BackupCommands::Restore(args) => args.run(cli),
|
||||||
|
BackupCommands::Export(args) => args.run(cli),
|
||||||
|
BackupCommands::Extract(args) => args.run(cli),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl BackupCreateArgs {
|
||||||
|
pub fn run(&self, cli: &super::Config) -> io::Result<()> {
|
||||||
|
let mut meta = cli.meta()?;
|
||||||
|
|
||||||
|
if let Some(res) = meta.create_backup(&self.layer) {
|
||||||
|
res
|
||||||
|
} else {
|
||||||
|
Err(io::Error::new(io::ErrorKind::Other, "Unknown layer"))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl BackupListArgs {
|
||||||
|
pub fn run(&self, cli: &super::Config) -> io::Result<()> {
|
||||||
|
let meta = cli.meta()?;
|
||||||
|
|
||||||
|
// A bit scuffed? Sure
|
||||||
|
for (name, manager) in meta
|
||||||
|
.managers()
|
||||||
|
.iter()
|
||||||
|
.filter(|(name, _)| self.layer.is_none() || &self.layer.as_ref().unwrap() == name)
|
||||||
|
{
|
||||||
|
println!("{}", name);
|
||||||
|
|
||||||
|
for chain in manager.chains().iter().filter(|c| !c.is_empty()) {
|
||||||
|
let mut iter = chain.iter();
|
||||||
|
println!(" {}", iter.next().unwrap());
|
||||||
|
|
||||||
|
for backup in iter {
|
||||||
|
println!(" {}", backup);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Tries to parse the given path as the path to a backup inside the backup directory with a
|
||||||
|
/// formatted timestamp.
|
||||||
|
fn parse_backup_path(
|
||||||
|
backup_dir: &Path,
|
||||||
|
backup_path: &Path,
|
||||||
|
) -> io::Result<(String, chrono::DateTime<Utc>)> {
|
||||||
|
if !backup_path.starts_with(backup_dir) {
|
||||||
|
return Err(other("Provided file is not inside the backup directory."));
|
||||||
|
}
|
||||||
|
|
||||||
|
let layer = if let Some(parent) = backup_path.parent() {
|
||||||
|
// Backup files should be stored nested inside a layer's folder
|
||||||
|
if parent != backup_dir {
|
||||||
|
parent.file_name().unwrap().to_string_lossy()
|
||||||
|
} else {
|
||||||
|
return Err(other("Invalid path."));
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
return Err(other("Invalid path."));
|
||||||
|
};
|
||||||
|
|
||||||
|
let timestamp = if let Some(filename) = backup_path.file_name() {
|
||||||
|
Utc.datetime_from_str(&filename.to_string_lossy(), Backup::FILENAME_FORMAT)
|
||||||
|
.map_err(|_| other("Invalid filename."))?
|
||||||
|
} else {
|
||||||
|
return Err(other("Invalid filename."));
|
||||||
|
};
|
||||||
|
|
||||||
|
Ok((layer.to_string(), timestamp))
|
||||||
|
}
|
||||||
|
|
||||||
|
impl BackupRestoreArgs {
|
||||||
|
pub fn run(&self, cli: &super::Config) -> io::Result<()> {
|
||||||
|
let backup_dir = cli.backup.canonicalize()?;
|
||||||
|
|
||||||
|
// Create directories if needed
|
||||||
|
if self.make {
|
||||||
|
std::fs::create_dir_all(&self.output_config)?;
|
||||||
|
std::fs::create_dir_all(&self.output_worlds)?;
|
||||||
|
}
|
||||||
|
|
||||||
|
let output_config = self.output_config.canonicalize()?;
|
||||||
|
let output_worlds = self.output_worlds.canonicalize()?;
|
||||||
|
|
||||||
|
// Parse input path
|
||||||
|
let backup_path = self.path.canonicalize()?;
|
||||||
|
let (layer, timestamp) = parse_backup_path(&backup_dir, &backup_path)?;
|
||||||
|
|
||||||
|
let meta = cli.meta()?;
|
||||||
|
|
||||||
|
// Clear previous contents of directories
|
||||||
|
let mut entries = output_config
|
||||||
|
.read_dir()?
|
||||||
|
.chain(output_worlds.read_dir()?)
|
||||||
|
.peekable();
|
||||||
|
|
||||||
|
if entries.peek().is_some() && !self.force {
|
||||||
|
return Err(other("Output directories are not empty. If you wish to overwrite these contents, use the force flag."));
|
||||||
|
}
|
||||||
|
|
||||||
|
for entry in entries {
|
||||||
|
let path = entry?.path();
|
||||||
|
|
||||||
|
if path.is_dir() {
|
||||||
|
std::fs::remove_dir_all(path)?;
|
||||||
|
} else {
|
||||||
|
std::fs::remove_file(path)?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
let dirs = vec![
|
||||||
|
(PathBuf::from("config"), output_config),
|
||||||
|
(PathBuf::from("worlds"), output_worlds),
|
||||||
|
];
|
||||||
|
|
||||||
|
// Restore the backup
|
||||||
|
if let Some(res) = meta.restore_backup(&layer, timestamp, &dirs) {
|
||||||
|
res
|
||||||
|
} else {
|
||||||
|
Err(other("Unknown layer"))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl BackupExportArgs {
|
||||||
|
pub fn run(&self, cli: &super::Config) -> io::Result<()> {
|
||||||
|
let backup_dir = cli.backup.canonicalize()?;
|
||||||
|
|
||||||
|
if self.make {
|
||||||
|
if let Some(parent) = &self.output.parent() {
|
||||||
|
std::fs::create_dir_all(parent)?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parse input path
|
||||||
|
let backup_path = self.path.canonicalize()?;
|
||||||
|
let (layer, timestamp) = parse_backup_path(&backup_dir, &backup_path)?;
|
||||||
|
|
||||||
|
let meta = cli.meta()?;
|
||||||
|
|
||||||
|
if let Some(res) = meta.export_backup(&layer, timestamp, &self.output) {
|
||||||
|
res
|
||||||
|
} else {
|
||||||
|
Err(other("Unknown layer"))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl BackupExtractArgs {
|
||||||
|
pub fn run(&self, _cli: &super::Config) -> io::Result<()> {
|
||||||
|
// Create directories if needed
|
||||||
|
if self.make {
|
||||||
|
std::fs::create_dir_all(&self.output_config)?;
|
||||||
|
std::fs::create_dir_all(&self.output_worlds)?;
|
||||||
|
}
|
||||||
|
|
||||||
|
let output_config = self.output_config.canonicalize()?;
|
||||||
|
let output_worlds = self.output_worlds.canonicalize()?;
|
||||||
|
let backup_path = self.path.canonicalize()?;
|
||||||
|
|
||||||
|
// Clear previous contents of directories
|
||||||
|
let mut entries = output_config
|
||||||
|
.read_dir()?
|
||||||
|
.chain(output_worlds.read_dir()?)
|
||||||
|
.peekable();
|
||||||
|
|
||||||
|
if entries.peek().is_some() && !self.force {
|
||||||
|
return Err(other("Output directories are not empty. If you wish to overwrite these contents, use the force flag."));
|
||||||
|
}
|
||||||
|
|
||||||
|
for entry in entries {
|
||||||
|
let path = entry?.path();
|
||||||
|
|
||||||
|
if path.is_dir() {
|
||||||
|
std::fs::remove_dir_all(path)?;
|
||||||
|
} else {
|
||||||
|
std::fs::remove_file(path)?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
let dirs = vec![
|
||||||
|
(PathBuf::from("config"), output_config),
|
||||||
|
(PathBuf::from("worlds"), output_worlds),
|
||||||
|
];
|
||||||
|
|
||||||
|
Backup::extract_archive(backup_path, &dirs)
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,49 @@
|
||||||
|
use std::{io, path::PathBuf};
|
||||||
|
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
backup::{ManagerConfig, MetaManager},
|
||||||
|
server::{Metadata, ServerType},
|
||||||
|
};
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct Config {
|
||||||
|
pub config: PathBuf,
|
||||||
|
pub world: PathBuf,
|
||||||
|
pub backup: PathBuf,
|
||||||
|
pub layers: Vec<ManagerConfig>,
|
||||||
|
pub server: ServerType,
|
||||||
|
pub server_version: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Default for Config {
|
||||||
|
fn default() -> Self {
|
||||||
|
Self {
|
||||||
|
config: PathBuf::from("."),
|
||||||
|
world: PathBuf::from("../worlds"),
|
||||||
|
backup: PathBuf::from("../backups"),
|
||||||
|
layers: Vec::new(),
|
||||||
|
server: ServerType::Unknown,
|
||||||
|
server_version: String::from(""),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Config {
|
||||||
|
/// Convenience method to initialize backup manager from the cli arguments
|
||||||
|
pub fn meta(&self) -> io::Result<MetaManager<Metadata>> {
|
||||||
|
let metadata = Metadata {
|
||||||
|
server_type: self.server,
|
||||||
|
server_version: self.server_version.clone(),
|
||||||
|
};
|
||||||
|
let dirs = vec![
|
||||||
|
(PathBuf::from("config"), self.config.canonicalize()?),
|
||||||
|
(PathBuf::from("worlds"), self.world.canonicalize()?),
|
||||||
|
];
|
||||||
|
let mut meta = MetaManager::new(self.backup.canonicalize()?, dirs, metadata);
|
||||||
|
meta.add_all(&self.layers)?;
|
||||||
|
|
||||||
|
Ok(meta)
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,119 @@
|
||||||
|
mod backup;
|
||||||
|
mod config;
|
||||||
|
mod run;
|
||||||
|
|
||||||
|
use std::{path::PathBuf, str::FromStr};
|
||||||
|
|
||||||
|
use clap::{Args, Parser, Subcommand};
|
||||||
|
use figment::{
|
||||||
|
providers::{Env, Format, Serialized, Toml},
|
||||||
|
Figment,
|
||||||
|
};
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use crate::{backup::ManagerConfig, server::ServerType};
|
||||||
|
use backup::BackupArgs;
|
||||||
|
use config::Config;
|
||||||
|
use run::RunCli;
|
||||||
|
|
||||||
|
#[derive(Parser, Serialize)]
|
||||||
|
#[command(author, version, about, long_about = None)]
|
||||||
|
pub struct Cli {
|
||||||
|
#[command(subcommand)]
|
||||||
|
#[serde(skip)]
|
||||||
|
pub command: Commands,
|
||||||
|
|
||||||
|
/// Path to a TOML configuration file
|
||||||
|
#[arg(long = "config-file", global = true)]
|
||||||
|
pub config_file: Option<PathBuf>,
|
||||||
|
|
||||||
|
#[command(flatten)]
|
||||||
|
pub args: CliArgs,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Args, Serialize, Deserialize, Clone)]
|
||||||
|
pub struct CliArgs {
|
||||||
|
/// Directory where configs are stored, and where the server will run
|
||||||
|
#[arg(long, value_name = "CONFIG_DIR", global = true)]
|
||||||
|
#[serde(skip_serializing_if = "::std::option::Option::is_none")]
|
||||||
|
pub config: Option<PathBuf>,
|
||||||
|
|
||||||
|
/// Directory where world files will be saved
|
||||||
|
#[arg(long, value_name = "WORLD_DIR", global = true)]
|
||||||
|
#[serde(skip_serializing_if = "::std::option::Option::is_none")]
|
||||||
|
pub world: Option<PathBuf>,
|
||||||
|
|
||||||
|
/// Directory where backups will be stored
|
||||||
|
#[arg(long, value_name = "BACKUP_DIR", global = true)]
|
||||||
|
#[serde(skip_serializing_if = "::std::option::Option::is_none")]
|
||||||
|
pub backup: Option<PathBuf>,
|
||||||
|
|
||||||
|
/// What backup layers to employ, provided as a list of tuples name,frequency,chains,chain_len
|
||||||
|
/// delimited by semicolons (;).
|
||||||
|
#[arg(long, global = true, value_delimiter = ';')]
|
||||||
|
#[serde(skip_serializing_if = "::std::option::Option::is_none")]
|
||||||
|
pub layers: Option<Vec<ManagerConfig>>,
|
||||||
|
|
||||||
|
/// Type of server
|
||||||
|
#[arg(long, global = true)]
|
||||||
|
#[serde(skip_serializing_if = "::std::option::Option::is_none")]
|
||||||
|
pub server: Option<ServerType>,
|
||||||
|
|
||||||
|
/// Version string for the server, e.g. 1.19.4-545
|
||||||
|
#[arg(long, global = true)]
|
||||||
|
#[serde(skip_serializing_if = "::std::option::Option::is_none")]
|
||||||
|
pub server_version: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Subcommand)]
|
||||||
|
pub enum Commands {
|
||||||
|
/// Run the server
|
||||||
|
Run(RunCli),
|
||||||
|
/// Interact with the backup system without starting a server
|
||||||
|
Backup(BackupArgs),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Cli {
|
||||||
|
pub fn run(&self) -> crate::Result<()> {
|
||||||
|
let config = self.config(&self.args)?;
|
||||||
|
|
||||||
|
match &self.command {
|
||||||
|
Commands::Run(args) => args.run(self, &config),
|
||||||
|
Commands::Backup(args) => Ok(args.run(&config)?),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn config<T, U>(&self, args: &U) -> crate::Result<T>
|
||||||
|
where
|
||||||
|
T: Default + Serialize + for<'de> Deserialize<'de>,
|
||||||
|
U: Serialize,
|
||||||
|
{
|
||||||
|
let toml_file = self
|
||||||
|
.config_file
|
||||||
|
.clone()
|
||||||
|
.unwrap_or(PathBuf::from(Env::var_or("ALEX_CONFIG_FILE", "")));
|
||||||
|
|
||||||
|
let mut figment = Figment::new()
|
||||||
|
.merge(Serialized::defaults(T::default()))
|
||||||
|
.merge(Toml::file(toml_file))
|
||||||
|
.merge(Env::prefixed("ALEX_").ignore(&["ALEX_LAYERS"]));
|
||||||
|
|
||||||
|
// Layers need to be parsed separately, as the env var format is different than the one
|
||||||
|
// serde expects
|
||||||
|
if let Some(layers_env) = Env::var("ALEX_LAYERS") {
|
||||||
|
let res = layers_env
|
||||||
|
.split(';')
|
||||||
|
.map(ManagerConfig::from_str)
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
|
||||||
|
if res.iter().any(|e| e.is_err()) {
|
||||||
|
return Err(crate::other("Invalid layer configuration").into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let layers: Vec<_> = res.iter().flatten().collect();
|
||||||
|
figment = figment.merge(Serialized::default("layers", layers));
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(figment.merge(Serialized::defaults(args)).extract()?)
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,22 @@
|
||||||
|
use std::path::PathBuf;
|
||||||
|
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct Config {
|
||||||
|
pub jar: PathBuf,
|
||||||
|
pub java: String,
|
||||||
|
pub xms: u64,
|
||||||
|
pub xmx: u64,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Default for Config {
|
||||||
|
fn default() -> Self {
|
||||||
|
Self {
|
||||||
|
jar: PathBuf::from("server.jar"),
|
||||||
|
java: String::from("java"),
|
||||||
|
xms: 1024,
|
||||||
|
xmx: 2048,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,103 @@
|
||||||
|
mod config;
|
||||||
|
|
||||||
|
use std::{path::PathBuf, sync::Arc};
|
||||||
|
|
||||||
|
use clap::Args;
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use crate::{server, signals, stdin};
|
||||||
|
use config::Config;
|
||||||
|
|
||||||
|
#[derive(Args)]
|
||||||
|
pub struct RunCli {
|
||||||
|
#[command(flatten)]
|
||||||
|
pub args: RunArgs,
|
||||||
|
|
||||||
|
/// Don't actually run the server, but simply output the server configuration that would have
|
||||||
|
/// been ran
|
||||||
|
#[arg(short, long, default_value_t = false)]
|
||||||
|
pub dry: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Args, Serialize, Deserialize, Clone)]
|
||||||
|
pub struct RunArgs {
|
||||||
|
/// Server jar to execute
|
||||||
|
#[arg(long, value_name = "JAR_PATH")]
|
||||||
|
#[serde(skip_serializing_if = "::std::option::Option::is_none")]
|
||||||
|
pub jar: Option<PathBuf>,
|
||||||
|
|
||||||
|
/// Java command to run the server jar with
|
||||||
|
#[arg(long, value_name = "JAVA_CMD")]
|
||||||
|
#[serde(skip_serializing_if = "::std::option::Option::is_none")]
|
||||||
|
pub java: Option<String>,
|
||||||
|
|
||||||
|
/// XMS value in megabytes for the server instance
|
||||||
|
#[arg(long)]
|
||||||
|
#[serde(skip_serializing_if = "::std::option::Option::is_none")]
|
||||||
|
pub xms: Option<u64>,
|
||||||
|
|
||||||
|
/// XMX value in megabytes for the server instance
|
||||||
|
#[arg(long)]
|
||||||
|
#[serde(skip_serializing_if = "::std::option::Option::is_none")]
|
||||||
|
pub xmx: Option<u64>,
|
||||||
|
}
|
||||||
|
|
||||||
|
fn backups_thread(server: Arc<server::ServerProcess>) {
|
||||||
|
loop {
|
||||||
|
let next_scheduled_time = {
|
||||||
|
server
|
||||||
|
.backups
|
||||||
|
.read()
|
||||||
|
.unwrap()
|
||||||
|
.next_scheduled_time()
|
||||||
|
.unwrap()
|
||||||
|
};
|
||||||
|
|
||||||
|
let now = chrono::offset::Utc::now();
|
||||||
|
if next_scheduled_time > now {
|
||||||
|
std::thread::sleep((next_scheduled_time - now).to_std().unwrap());
|
||||||
|
}
|
||||||
|
|
||||||
|
// We explicitely ignore the error here, as we don't want the thread to fail
|
||||||
|
let _ = server.backup();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl RunCli {
|
||||||
|
pub fn run(&self, cli: &super::Cli, global: &super::Config) -> crate::Result<()> {
|
||||||
|
let config: Config = cli.config(&self.args)?;
|
||||||
|
|
||||||
|
let (_, mut signals) = signals::install_signal_handlers()?;
|
||||||
|
|
||||||
|
let mut cmd = server::ServerCommand::new(global.server, &global.server_version)
|
||||||
|
.java(&config.java)
|
||||||
|
.jar(config.jar.clone())
|
||||||
|
.config(global.config.clone())
|
||||||
|
.world(global.world.clone())
|
||||||
|
.backup(global.backup.clone())
|
||||||
|
.managers(global.layers.clone())
|
||||||
|
.xms(config.xms)
|
||||||
|
.xmx(config.xmx);
|
||||||
|
cmd.canonicalize()?;
|
||||||
|
|
||||||
|
if self.dry {
|
||||||
|
print!("{}", cmd);
|
||||||
|
|
||||||
|
return Ok(());
|
||||||
|
}
|
||||||
|
|
||||||
|
let counter = Arc::new(cmd.spawn()?);
|
||||||
|
|
||||||
|
if !global.layers.is_empty() {
|
||||||
|
let clone = Arc::clone(&counter);
|
||||||
|
std::thread::spawn(move || backups_thread(clone));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Spawn thread that handles the main stdin loop
|
||||||
|
let clone = Arc::clone(&counter);
|
||||||
|
std::thread::spawn(move || stdin::handle_stdin(clone));
|
||||||
|
|
||||||
|
// Signal handler loop exits the process when necessary
|
||||||
|
Ok(signals::handle_signals(&mut signals, counter)?)
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,32 @@
|
||||||
|
use std::{fmt, io};
|
||||||
|
|
||||||
|
pub type Result<T> = std::result::Result<T, Error>;
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Error {
|
||||||
|
IO(io::Error),
|
||||||
|
Figment(figment::Error),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl fmt::Display for Error {
|
||||||
|
fn fmt(&self, fmt: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||||
|
match self {
|
||||||
|
Error::IO(err) => write!(fmt, "{}", err),
|
||||||
|
Error::Figment(err) => write!(fmt, "{}", err),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl std::error::Error for Error {}
|
||||||
|
|
||||||
|
impl From<io::Error> for Error {
|
||||||
|
fn from(err: io::Error) -> Self {
|
||||||
|
Error::IO(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<figment::Error> for Error {
|
||||||
|
fn from(err: figment::Error) -> Self {
|
||||||
|
Error::Figment(err)
|
||||||
|
}
|
||||||
|
}
|
96
src/main.rs
96
src/main.rs
|
@ -1,89 +1,43 @@
|
||||||
mod backup;
|
mod backup;
|
||||||
mod cli;
|
mod cli;
|
||||||
|
mod error;
|
||||||
mod server;
|
mod server;
|
||||||
mod signals;
|
mod signals;
|
||||||
mod stdin;
|
mod stdin;
|
||||||
|
|
||||||
use clap::Parser;
|
|
||||||
use cli::{BackupArgs, Cli, Commands, RunArgs};
|
|
||||||
use std::io;
|
use std::io;
|
||||||
use std::sync::{Arc, Mutex};
|
|
||||||
|
|
||||||
fn backups_thread(counter: Arc<Mutex<server::ServerProcess>>, frequency: u64) {
|
use clap::Parser;
|
||||||
loop {
|
|
||||||
std::thread::sleep(std::time::Duration::from_secs(frequency * 60));
|
|
||||||
|
|
||||||
{
|
use crate::cli::Cli;
|
||||||
let mut server = counter.lock().unwrap();
|
pub use error::{Error, Result};
|
||||||
|
|
||||||
// We explicitely ignore the error here, as we don't want the thread to fail
|
pub fn other(msg: &str) -> io::Error {
|
||||||
let _ = server.backup();
|
io::Error::new(io::ErrorKind::Other, msg)
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn command_run(cli: &Cli, args: &RunArgs) -> io::Result<()> {
|
// fn commands_backup(cli: &Cli, args: &BackupArgs) -> io::Result<()> {
|
||||||
let (_, mut signals) = signals::install_signal_handlers()?;
|
// let metadata = server::Metadata {
|
||||||
|
// server_type: cli.server,
|
||||||
|
// server_version: cli.server_version.clone(),
|
||||||
|
// };
|
||||||
|
// let dirs = vec![
|
||||||
|
// (PathBuf::from("config"), cli.config.clone()),
|
||||||
|
// (PathBuf::from("worlds"), cli.world.clone()),
|
||||||
|
// ];
|
||||||
|
// let mut meta = MetaManager::new(cli.backup.clone(), dirs, metadata);
|
||||||
|
// meta.add_all(&cli.layers)?;
|
||||||
|
|
||||||
let mut cmd = server::ServerCommand::new(cli.server, &cli.server_version)
|
// match &args.command {
|
||||||
.java(&args.java)
|
// BackupCommands::List => ()
|
||||||
.jar(args.jar.clone())
|
// }
|
||||||
.config(cli.config.clone())
|
|
||||||
.world(cli.world.clone())
|
|
||||||
.backup(cli.backup.clone())
|
|
||||||
.xms(args.xms)
|
|
||||||
.xmx(args.xmx)
|
|
||||||
.chain_len(cli.chain_len)
|
|
||||||
.chains_to_keep(cli.chains);
|
|
||||||
cmd.canonicalize()?;
|
|
||||||
|
|
||||||
if args.dry {
|
// // manager.create_backup()?;
|
||||||
print!("{}", cmd);
|
// // manager.remove_old_backups()
|
||||||
|
// }
|
||||||
|
|
||||||
return Ok(());
|
fn main() -> crate::Result<()> {
|
||||||
}
|
|
||||||
|
|
||||||
let counter = Arc::new(Mutex::new(cmd.spawn()?));
|
|
||||||
|
|
||||||
if args.frequency > 0 {
|
|
||||||
let clone = Arc::clone(&counter);
|
|
||||||
let frequency = args.frequency;
|
|
||||||
std::thread::spawn(move || backups_thread(clone, frequency));
|
|
||||||
}
|
|
||||||
|
|
||||||
// Spawn thread that handles the main stdin loop
|
|
||||||
let clone = Arc::clone(&counter);
|
|
||||||
std::thread::spawn(move || stdin::handle_stdin(clone));
|
|
||||||
|
|
||||||
// Signal handler loop exits the process when necessary
|
|
||||||
signals::handle_signals(&mut signals, counter)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn commands_backup(cli: &Cli, _args: &BackupArgs) -> io::Result<()> {
|
|
||||||
let metadata = server::Metadata {
|
|
||||||
server_type: cli.server,
|
|
||||||
server_version: cli.server_version.clone(),
|
|
||||||
};
|
|
||||||
|
|
||||||
let mut manager = backup::Manager::new(
|
|
||||||
cli.backup.clone(),
|
|
||||||
cli.config.clone(),
|
|
||||||
cli.world.clone(),
|
|
||||||
metadata,
|
|
||||||
cli.chain_len,
|
|
||||||
cli.chains,
|
|
||||||
);
|
|
||||||
manager.load()?;
|
|
||||||
|
|
||||||
manager.create_backup()?;
|
|
||||||
manager.remove_old_backups()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn main() -> io::Result<()> {
|
|
||||||
let cli = Cli::parse();
|
let cli = Cli::parse();
|
||||||
|
|
||||||
match &cli.command {
|
cli.run()
|
||||||
Commands::Run(args) => command_run(&cli, args),
|
|
||||||
Commands::Backup(args) => commands_backup(&cli, args),
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,40 +1,12 @@
|
||||||
use crate::backup::Manager as BackupManager;
|
use crate::backup::ManagerConfig;
|
||||||
use crate::server::ServerProcess;
|
use crate::backup::MetaManager;
|
||||||
use clap::ValueEnum;
|
use crate::server::{Metadata, ServerProcess, ServerType};
|
||||||
use serde::{Deserialize, Serialize};
|
|
||||||
use std::fmt;
|
use std::fmt;
|
||||||
use std::fs::File;
|
use std::fs::File;
|
||||||
use std::io::Write;
|
use std::io::Write;
|
||||||
use std::path::{Path, PathBuf};
|
use std::path::{Path, PathBuf};
|
||||||
use std::process::{Command, Stdio};
|
use std::process::{Command, Stdio};
|
||||||
|
|
||||||
#[derive(Copy, Clone, PartialEq, Eq, PartialOrd, Ord, ValueEnum, Serialize, Deserialize)]
|
|
||||||
pub enum ServerType {
|
|
||||||
Unknown,
|
|
||||||
Paper,
|
|
||||||
Forge,
|
|
||||||
Vanilla,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl fmt::Display for ServerType {
|
|
||||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
|
||||||
let s = match self {
|
|
||||||
ServerType::Unknown => "Unknown",
|
|
||||||
ServerType::Paper => "PaperMC",
|
|
||||||
ServerType::Forge => "Forge",
|
|
||||||
ServerType::Vanilla => "Vanilla",
|
|
||||||
};
|
|
||||||
|
|
||||||
write!(f, "{}", s)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Clone, Serialize, Deserialize)]
|
|
||||||
pub struct Metadata {
|
|
||||||
pub server_type: ServerType,
|
|
||||||
pub server_version: String,
|
|
||||||
}
|
|
||||||
|
|
||||||
pub struct ServerCommand {
|
pub struct ServerCommand {
|
||||||
type_: ServerType,
|
type_: ServerType,
|
||||||
version: String,
|
version: String,
|
||||||
|
@ -45,8 +17,7 @@ pub struct ServerCommand {
|
||||||
backup_dir: PathBuf,
|
backup_dir: PathBuf,
|
||||||
xms: u64,
|
xms: u64,
|
||||||
xmx: u64,
|
xmx: u64,
|
||||||
chain_len: u64,
|
managers: Vec<ManagerConfig>,
|
||||||
chains_to_keep: u64,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl ServerCommand {
|
impl ServerCommand {
|
||||||
|
@ -61,8 +32,7 @@ impl ServerCommand {
|
||||||
backup_dir: PathBuf::from("backups"),
|
backup_dir: PathBuf::from("backups"),
|
||||||
xms: 1024,
|
xms: 1024,
|
||||||
xmx: 2048,
|
xmx: 2048,
|
||||||
chain_len: 4,
|
managers: Vec::new(),
|
||||||
chains_to_keep: 7,
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -103,13 +73,9 @@ impl ServerCommand {
|
||||||
self
|
self
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn chain_len(mut self, v: u64) -> Self {
|
pub fn managers(mut self, configs: Vec<ManagerConfig>) -> Self {
|
||||||
self.chain_len = v;
|
self.managers = configs;
|
||||||
self
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn chains_to_keep(mut self, v: u64) -> Self {
|
|
||||||
self.chains_to_keep = v;
|
|
||||||
self
|
self
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -200,21 +166,20 @@ impl ServerCommand {
|
||||||
server_type: self.type_,
|
server_type: self.type_,
|
||||||
server_version: self.version.clone(),
|
server_version: self.version.clone(),
|
||||||
};
|
};
|
||||||
let mut manager = BackupManager::new(
|
|
||||||
self.backup_dir.clone(),
|
let dirs = vec![
|
||||||
self.config_dir.clone(),
|
(PathBuf::from("config"), self.config_dir.clone()),
|
||||||
self.world_dir.clone(),
|
(PathBuf::from("worlds"), self.world_dir.clone()),
|
||||||
metadata,
|
];
|
||||||
self.chain_len,
|
|
||||||
self.chains_to_keep,
|
let mut meta = MetaManager::new(self.backup_dir.clone(), dirs, metadata);
|
||||||
);
|
meta.add_all(&self.managers)?;
|
||||||
manager.load()?;
|
|
||||||
|
|
||||||
let mut cmd = self.create_cmd();
|
let mut cmd = self.create_cmd();
|
||||||
self.accept_eula()?;
|
self.accept_eula()?;
|
||||||
let child = cmd.spawn()?;
|
let child = cmd.spawn()?;
|
||||||
|
|
||||||
Ok(ServerProcess::new(manager, child))
|
Ok(ServerProcess::new(meta, child))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,5 +1,37 @@
|
||||||
mod command;
|
mod command;
|
||||||
mod process;
|
mod process;
|
||||||
|
|
||||||
pub use command::{Metadata, ServerCommand, ServerType};
|
pub use command::ServerCommand;
|
||||||
pub use process::ServerProcess;
|
pub use process::ServerProcess;
|
||||||
|
|
||||||
|
use clap::ValueEnum;
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
use std::fmt;
|
||||||
|
|
||||||
|
#[derive(Copy, Clone, PartialEq, Eq, PartialOrd, Ord, ValueEnum, Serialize, Deserialize, Debug)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
pub enum ServerType {
|
||||||
|
Unknown,
|
||||||
|
Paper,
|
||||||
|
Forge,
|
||||||
|
Vanilla,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl fmt::Display for ServerType {
|
||||||
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||||
|
let s = match self {
|
||||||
|
ServerType::Unknown => "Unknown",
|
||||||
|
ServerType::Paper => "PaperMC",
|
||||||
|
ServerType::Forge => "Forge",
|
||||||
|
ServerType::Vanilla => "Vanilla",
|
||||||
|
};
|
||||||
|
|
||||||
|
write!(f, "{}", s)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Clone, Serialize, Deserialize, Debug)]
|
||||||
|
pub struct Metadata {
|
||||||
|
pub server_type: ServerType,
|
||||||
|
pub server_version: String,
|
||||||
|
}
|
||||||
|
|
|
@ -1,22 +1,21 @@
|
||||||
use crate::backup::Manager as BackupManager;
|
use std::{io::Write, process::Child, sync::RwLock};
|
||||||
use crate::server::Metadata;
|
|
||||||
use std::io::Write;
|
use crate::{backup::MetaManager, server::Metadata};
|
||||||
use std::process::Child;
|
|
||||||
|
|
||||||
pub struct ServerProcess {
|
pub struct ServerProcess {
|
||||||
child: Child,
|
child: RwLock<Child>,
|
||||||
backups: BackupManager<Metadata>,
|
pub backups: RwLock<MetaManager<Metadata>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl ServerProcess {
|
impl ServerProcess {
|
||||||
pub fn new(manager: BackupManager<Metadata>, child: Child) -> ServerProcess {
|
pub fn new(manager: MetaManager<Metadata>, child: Child) -> ServerProcess {
|
||||||
ServerProcess {
|
ServerProcess {
|
||||||
child,
|
child: RwLock::new(child),
|
||||||
backups: manager,
|
backups: RwLock::new(manager),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn send_command(&mut self, cmd: &str) -> std::io::Result<()> {
|
pub fn send_command(&self, cmd: &str) -> std::io::Result<()> {
|
||||||
match cmd.trim() {
|
match cmd.trim() {
|
||||||
"stop" | "exit" => self.stop()?,
|
"stop" | "exit" => self.stop()?,
|
||||||
"backup" => self.backup()?,
|
"backup" => self.backup()?,
|
||||||
|
@ -26,29 +25,35 @@ impl ServerProcess {
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn custom(&mut self, cmd: &str) -> std::io::Result<()> {
|
fn custom(&self, cmd: &str) -> std::io::Result<()> {
|
||||||
let mut stdin = self.child.stdin.as_ref().unwrap();
|
let child = self.child.write().unwrap();
|
||||||
|
let mut stdin = child.stdin.as_ref().unwrap();
|
||||||
stdin.write_all(format!("{}\n", cmd.trim()).as_bytes())?;
|
stdin.write_all(format!("{}\n", cmd.trim()).as_bytes())?;
|
||||||
stdin.flush()?;
|
stdin.flush()?;
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn stop(&mut self) -> std::io::Result<()> {
|
pub fn stop(&self) -> std::io::Result<()> {
|
||||||
self.custom("stop")?;
|
self.custom("stop")?;
|
||||||
self.child.wait()?;
|
|
||||||
|
self.child.write().unwrap().wait()?;
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn kill(&mut self) -> std::io::Result<()> {
|
pub fn kill(&self) -> std::io::Result<()> {
|
||||||
self.child.kill()
|
self.child.write().unwrap().kill()
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Perform a backup by disabling the server's save feature and flushing its data, before
|
/// Perform a backup by disabling the server's save feature and flushing its data, before
|
||||||
/// creating an archive file.
|
/// creating an archive file.
|
||||||
pub fn backup(&mut self) -> std::io::Result<()> {
|
pub fn backup(&self) -> std::io::Result<()> {
|
||||||
self.custom("say backing up server")?;
|
// We explicitely lock this entire function to prevent parallel backups
|
||||||
|
let mut backups = self.backups.write().unwrap();
|
||||||
|
|
||||||
|
let layer_name = String::from(backups.next_scheduled_layer().unwrap());
|
||||||
|
self.custom(&format!("say starting backup for layer '{}'", layer_name))?;
|
||||||
|
|
||||||
// Make sure the server isn't modifying the files during the backup
|
// Make sure the server isn't modifying the files during the backup
|
||||||
self.custom("save-off")?;
|
self.custom("save-off")?;
|
||||||
|
@ -59,16 +64,12 @@ impl ServerProcess {
|
||||||
std::thread::sleep(std::time::Duration::from_secs(10));
|
std::thread::sleep(std::time::Duration::from_secs(10));
|
||||||
|
|
||||||
let start_time = chrono::offset::Utc::now();
|
let start_time = chrono::offset::Utc::now();
|
||||||
let res = self.backups.create_backup();
|
let res = backups.perform_backup_cycle();
|
||||||
|
|
||||||
// The server's save feature needs to be enabled again even if the archive failed to create
|
// The server's save feature needs to be enabled again even if the archive failed to create
|
||||||
self.custom("save-on")?;
|
self.custom("save-on")?;
|
||||||
self.custom("save-all")?;
|
self.custom("save-all")?;
|
||||||
|
|
||||||
if res.is_ok() {
|
|
||||||
self.backups.remove_old_backups()?;
|
|
||||||
}
|
|
||||||
|
|
||||||
let duration = chrono::offset::Utc::now() - start_time;
|
let duration = chrono::offset::Utc::now() - start_time;
|
||||||
let duration_str = format!(
|
let duration_str = format!(
|
||||||
"{}m{}s",
|
"{}m{}s",
|
||||||
|
@ -77,11 +78,14 @@ impl ServerProcess {
|
||||||
);
|
);
|
||||||
|
|
||||||
if res.is_ok() {
|
if res.is_ok() {
|
||||||
self.custom(&format!("say server backed up in {}", duration_str))?;
|
self.custom(&format!(
|
||||||
|
"say backup created for layer '{}' in {}",
|
||||||
|
layer_name, duration_str
|
||||||
|
))?;
|
||||||
} else {
|
} else {
|
||||||
self.custom(&format!(
|
self.custom(&format!(
|
||||||
"an error occured after {} while backing up the server",
|
"an error occured after {} while creating backup for layer '{}'",
|
||||||
duration_str
|
duration_str, layer_name
|
||||||
))?;
|
))?;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,10 +1,13 @@
|
||||||
use std::io;
|
use std::{
|
||||||
use std::sync::atomic::AtomicBool;
|
io,
|
||||||
use std::sync::{Arc, Mutex};
|
sync::{atomic::AtomicBool, Arc},
|
||||||
|
};
|
||||||
|
|
||||||
use signal_hook::consts::TERM_SIGNALS;
|
use signal_hook::{
|
||||||
use signal_hook::flag;
|
consts::TERM_SIGNALS,
|
||||||
use signal_hook::iterator::{Signals, SignalsInfo};
|
flag,
|
||||||
|
iterator::{Signals, SignalsInfo},
|
||||||
|
};
|
||||||
|
|
||||||
use crate::server;
|
use crate::server;
|
||||||
|
|
||||||
|
@ -34,7 +37,7 @@ pub fn install_signal_handlers() -> io::Result<(Arc<AtomicBool>, SignalsInfo)> {
|
||||||
/// Loop that handles terminating signals as they come in.
|
/// Loop that handles terminating signals as they come in.
|
||||||
pub fn handle_signals(
|
pub fn handle_signals(
|
||||||
signals: &mut SignalsInfo,
|
signals: &mut SignalsInfo,
|
||||||
counter: Arc<Mutex<server::ServerProcess>>,
|
server: Arc<server::ServerProcess>,
|
||||||
) -> io::Result<()> {
|
) -> io::Result<()> {
|
||||||
let mut force = false;
|
let mut force = false;
|
||||||
|
|
||||||
|
@ -46,17 +49,15 @@ pub fn handle_signals(
|
||||||
// This will currently not work, as the initial stop command will block the kill from
|
// This will currently not work, as the initial stop command will block the kill from
|
||||||
// happening.
|
// happening.
|
||||||
if force {
|
if force {
|
||||||
let mut server = counter.lock().unwrap();
|
|
||||||
return server.kill();
|
return server.kill();
|
||||||
}
|
}
|
||||||
// The stop command runs in a separate thread to avoid blocking the signal handling loop.
|
// The stop command runs in a separate thread to avoid blocking the signal handling loop.
|
||||||
// After stopping the server, the thread terminates the process.
|
// After stopping the server, the thread terminates the process.
|
||||||
else {
|
else {
|
||||||
let clone = Arc::clone(&counter);
|
let clone = Arc::clone(&server);
|
||||||
|
|
||||||
std::thread::spawn(move || {
|
std::thread::spawn(move || {
|
||||||
let mut server = clone.lock().unwrap();
|
let _ = clone.stop();
|
||||||
let _ = server.stop();
|
|
||||||
std::process::exit(0);
|
std::process::exit(0);
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,9 +1,8 @@
|
||||||
use std::io;
|
use std::{io, sync::Arc};
|
||||||
use std::sync::{Arc, Mutex};
|
|
||||||
|
|
||||||
use crate::server;
|
use crate::server;
|
||||||
|
|
||||||
pub fn handle_stdin(counter: Arc<Mutex<server::ServerProcess>>) {
|
pub fn handle_stdin(server: Arc<server::ServerProcess>) {
|
||||||
let stdin = io::stdin();
|
let stdin = io::stdin();
|
||||||
let input = &mut String::new();
|
let input = &mut String::new();
|
||||||
|
|
||||||
|
@ -14,13 +13,9 @@ pub fn handle_stdin(counter: Arc<Mutex<server::ServerProcess>>) {
|
||||||
continue;
|
continue;
|
||||||
};
|
};
|
||||||
|
|
||||||
{
|
|
||||||
let mut server = counter.lock().unwrap();
|
|
||||||
|
|
||||||
if let Err(e) = server.send_command(input) {
|
if let Err(e) = server.send_command(input) {
|
||||||
println!("{}", e);
|
println!("{}", e);
|
||||||
};
|
};
|
||||||
}
|
|
||||||
|
|
||||||
if input.trim() == "stop" {
|
if input.trim() == "stop" {
|
||||||
std::process::exit(0);
|
std::process::exit(0);
|
||||||
|
|
Loading…
Reference in New Issue