forked from vieter-v/vieter
Compare commits
26 Commits
2c4b948a94
...
48e2ae7645
| Author | SHA1 | Date |
|---|---|---|
|
|
48e2ae7645 | |
|
|
9f753f9c93 | |
|
|
ec92b16a73 | |
|
|
aded6d438a | |
|
|
edd71b41c2 | |
|
|
f0565c4168 | |
|
|
a4ffc2c0e3 | |
|
|
401e0291e3 | |
|
|
a39c1aa5eb | |
|
|
4f32dec5b5 | |
|
|
31e903ebeb | |
|
|
596da100b6 | |
|
|
0d5704ba15 | |
|
|
cdb88e1620 | |
|
|
768da5b790 | |
|
|
bd4bb9a9fb | |
|
|
c0b739035b | |
|
|
7f6e9e636c | |
|
|
96d0c2f1eb | |
|
|
0233b8559d | |
|
|
2fc25f1afe | |
|
|
6bd5b7cb48 | |
|
|
5e81dadce3 | |
|
|
1e079143cd | |
|
|
7627b28bcf | |
|
|
0de5ffb45d |
|
|
@ -45,7 +45,7 @@ pipeline:
|
||||||
|
|
||||||
- export OBJ_PATH="/vieter/commits/$CI_COMMIT_SHA/vieter-$(echo '${PLATFORM}' | sed 's:/:-:g')"
|
- export OBJ_PATH="/vieter/commits/$CI_COMMIT_SHA/vieter-$(echo '${PLATFORM}' | sed 's:/:-:g')"
|
||||||
- export SIG_STRING="PUT\n\n$CONTENT_TYPE\n$DATE\n$OBJ_PATH"
|
- export SIG_STRING="PUT\n\n$CONTENT_TYPE\n$DATE\n$OBJ_PATH"
|
||||||
- export SIGNATURE=`echo -en $SIG_STRING | openssl sha1 -hmac $S3_PASSWORD -binary | base64`
|
- export SIGNATURE="$(echo -en $SIG_STRING | openssl sha1 -hmac $S3_PASSWORD -binary | base64)"
|
||||||
- >
|
- >
|
||||||
curl
|
curl
|
||||||
--silent
|
--silent
|
||||||
|
|
|
||||||
20
CHANGELOG.md
20
CHANGELOG.md
|
|
@ -9,6 +9,26 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
|
||||||
|
|
||||||
### Added
|
### Added
|
||||||
|
|
||||||
|
* Database migrations
|
||||||
|
* Improved GitRepo & BuildLog API
|
||||||
|
* Pagination using `limit` & `offset` query params
|
||||||
|
* GitRepo: filter by repo
|
||||||
|
* BuildLog: filter by start & end date, repo, exit code & arch
|
||||||
|
* CLI flags to take advantage of above API improvements
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
|
||||||
|
* Packages from target repo are available during builds
|
||||||
|
* This can be used as a basic way to support AUR dependencies, by adding
|
||||||
|
the dependencies to the same repository
|
||||||
|
* Every build now updates its packages first instead of solely relying on the
|
||||||
|
updated builder image
|
||||||
|
* Build logs now show commands being executed
|
||||||
|
|
||||||
|
## [0.3.0-alpha.2](https://git.rustybever.be/vieter/vieter/src/tag/0.3.0-alpha.2)
|
||||||
|
|
||||||
|
### Added
|
||||||
|
|
||||||
* Web API for adding & querying build logs
|
* Web API for adding & querying build logs
|
||||||
* CLI commands to access build logs API
|
* CLI commands to access build logs API
|
||||||
* Cron build logs are uploaded to above API
|
* Cron build logs are uploaded to above API
|
||||||
|
|
|
||||||
2
PKGBUILD
2
PKGBUILD
|
|
@ -3,7 +3,7 @@
|
||||||
|
|
||||||
pkgbase='vieter'
|
pkgbase='vieter'
|
||||||
pkgname='vieter'
|
pkgname='vieter'
|
||||||
pkgver='0.3.0_alpha.1'
|
pkgver='0.3.0_alpha.2'
|
||||||
pkgrel=1
|
pkgrel=1
|
||||||
depends=('glibc' 'openssl' 'libarchive' 'sqlite')
|
depends=('glibc' 'openssl' 'libarchive' 'sqlite')
|
||||||
makedepends=('git' 'vieter-v')
|
makedepends=('git' 'vieter-v')
|
||||||
|
|
|
||||||
|
|
@ -4,9 +4,9 @@ import docker
|
||||||
import encoding.base64
|
import encoding.base64
|
||||||
import time
|
import time
|
||||||
import os
|
import os
|
||||||
import db
|
|
||||||
import strings
|
import strings
|
||||||
import util
|
import util
|
||||||
|
import models { GitRepo }
|
||||||
|
|
||||||
const (
|
const (
|
||||||
container_build_dir = '/build'
|
container_build_dir = '/build'
|
||||||
|
|
@ -93,7 +93,7 @@ pub:
|
||||||
// build_repo builds, packages & publishes a given Arch package based on the
|
// build_repo builds, packages & publishes a given Arch package based on the
|
||||||
// provided GitRepo. The base image ID should be of an image previously created
|
// provided GitRepo. The base image ID should be of an image previously created
|
||||||
// by create_build_image. It returns the logs of the container.
|
// by create_build_image. It returns the logs of the container.
|
||||||
pub fn build_repo(address string, api_key string, base_image_id string, repo &db.GitRepo) ?BuildResult {
|
pub fn build_repo(address string, api_key string, base_image_id string, repo &GitRepo) ?BuildResult {
|
||||||
mut dd := docker.new_conn()?
|
mut dd := docker.new_conn()?
|
||||||
|
|
||||||
defer {
|
defer {
|
||||||
|
|
@ -101,30 +101,19 @@ pub fn build_repo(address string, api_key string, base_image_id string, repo &db
|
||||||
}
|
}
|
||||||
|
|
||||||
build_arch := os.uname().machine
|
build_arch := os.uname().machine
|
||||||
|
build_script := create_build_script(address, repo, build_arch)
|
||||||
|
|
||||||
// TODO what to do with PKGBUILDs that build multiple packages?
|
// We convert the build script into a base64 string, which then gets passed
|
||||||
commands := [
|
// to the container as an env var
|
||||||
'git clone --single-branch --depth 1 --branch $repo.branch $repo.url repo',
|
base64_script := base64.encode_str(build_script)
|
||||||
'cd repo',
|
|
||||||
'makepkg --nobuild --syncdeps --needed --noconfirm',
|
|
||||||
'source PKGBUILD',
|
|
||||||
// The build container checks whether the package is already
|
|
||||||
// present on the server
|
|
||||||
'curl -s --head --fail $address/$repo.repo/$build_arch/\$pkgname-\$pkgver-\$pkgrel && exit 0',
|
|
||||||
'MAKEFLAGS="-j\$(nproc)" makepkg -s --noconfirm --needed && for pkg in \$(ls -1 *.pkg*); do curl -XPOST -T "\$pkg" -H "X-API-KEY: \$API_KEY" $address/$repo.repo/publish; done',
|
|
||||||
]
|
|
||||||
|
|
||||||
// We convert the list of commands into a base64 string, which then gets
|
|
||||||
// passed to the container as an env var
|
|
||||||
cmds_str := base64.encode_str(commands.join('\n'))
|
|
||||||
|
|
||||||
c := docker.NewContainer{
|
c := docker.NewContainer{
|
||||||
image: '$base_image_id'
|
image: '$base_image_id'
|
||||||
env: ['BUILD_SCRIPT=$cmds_str', 'API_KEY=$api_key']
|
env: ['BUILD_SCRIPT=$base64_script', 'API_KEY=$api_key']
|
||||||
entrypoint: ['/bin/sh', '-c']
|
entrypoint: ['/bin/sh', '-c']
|
||||||
cmd: ['echo \$BUILD_SCRIPT | base64 -d | /bin/bash -e']
|
cmd: ['echo \$BUILD_SCRIPT | base64 -d | /bin/bash -e']
|
||||||
work_dir: '/build'
|
work_dir: '/build'
|
||||||
user: 'builder:builder'
|
user: '0:0'
|
||||||
}
|
}
|
||||||
|
|
||||||
id := dd.create_container(c)?.id
|
id := dd.create_container(c)?.id
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,20 @@
|
||||||
|
echo -e '+ echo -e '\''[vieter]\\nServer = https://example.com/$repo/$arch\\nSigLevel = Optional'\'' >> /etc/pacman.conf'
|
||||||
|
echo -e '[vieter]\nServer = https://example.com/$repo/$arch\nSigLevel = Optional' >> /etc/pacman.conf
|
||||||
|
echo -e '+ pacman -Syu --needed --noconfirm'
|
||||||
|
pacman -Syu --needed --noconfirm
|
||||||
|
echo -e '+ su builder'
|
||||||
|
su builder
|
||||||
|
echo -e '+ git clone --single-branch --depth 1 --branch main https://examplerepo.com repo'
|
||||||
|
git clone --single-branch --depth 1 --branch main https://examplerepo.com repo
|
||||||
|
echo -e '+ cd repo'
|
||||||
|
cd repo
|
||||||
|
echo -e '+ makepkg --nobuild --syncdeps --needed --noconfirm'
|
||||||
|
makepkg --nobuild --syncdeps --needed --noconfirm
|
||||||
|
echo -e '+ source PKGBUILD'
|
||||||
|
source PKGBUILD
|
||||||
|
echo -e '+ curl -s --head --fail https://example.com/vieter/x86_64/$pkgname-$pkgver-$pkgrel && exit 0'
|
||||||
|
curl -s --head --fail https://example.com/vieter/x86_64/$pkgname-$pkgver-$pkgrel && exit 0
|
||||||
|
echo -e '+ [ "$(id -u)" == 0 ] && exit 0'
|
||||||
|
[ "$(id -u)" == 0 ] && exit 0
|
||||||
|
echo -e '+ MAKEFLAGS="-j$(nproc)" makepkg -s --noconfirm --needed && for pkg in $(ls -1 *.pkg*); do curl -XPOST -T "$pkg" -H "X-API-KEY: $API_KEY" https://example.com/vieter/publish; done'
|
||||||
|
MAKEFLAGS="-j$(nproc)" makepkg -s --noconfirm --needed && for pkg in $(ls -1 *.pkg*); do curl -XPOST -T "$pkg" -H "X-API-KEY: $API_KEY" https://example.com/vieter/publish; done
|
||||||
|
|
@ -0,0 +1,55 @@
|
||||||
|
module build
|
||||||
|
|
||||||
|
import models { GitRepo }
|
||||||
|
|
||||||
|
// escape_shell_string escapes any characters that could be interpreted
|
||||||
|
// incorrectly by a shell. The resulting value should be safe to use inside an
|
||||||
|
// echo statement.
|
||||||
|
fn escape_shell_string(s string) string {
|
||||||
|
return s.replace(r'\', r'\\').replace("'", r"'\''")
|
||||||
|
}
|
||||||
|
|
||||||
|
// echo_commands takes a list of shell commands & prepends each one with
|
||||||
|
// an echo call displaying said command.
|
||||||
|
pub fn echo_commands(cmds []string) []string {
|
||||||
|
mut out := []string{cap: 2 * cmds.len}
|
||||||
|
|
||||||
|
for cmd in cmds {
|
||||||
|
out << "echo -e '+ ${escape_shell_string(cmd)}'"
|
||||||
|
out << cmd
|
||||||
|
}
|
||||||
|
|
||||||
|
return out
|
||||||
|
}
|
||||||
|
|
||||||
|
// create_build_script generates a shell script that builds a given GitRepo.
|
||||||
|
fn create_build_script(address string, repo &GitRepo, build_arch string) string {
|
||||||
|
repo_url := '$address/$repo.repo'
|
||||||
|
|
||||||
|
commands := echo_commands([
|
||||||
|
// This will later be replaced by a proper setting for changing the
|
||||||
|
// mirrorlist
|
||||||
|
"echo -e '[$repo.repo]\\nServer = $address/\$repo/\$arch\\nSigLevel = Optional' >> /etc/pacman.conf"
|
||||||
|
// We need to update the package list of the repo we just added above.
|
||||||
|
// This should however not pull in a lot of packages as long as the
|
||||||
|
// builder image is rebuilt frequently.
|
||||||
|
'pacman -Syu --needed --noconfirm',
|
||||||
|
// makepkg can't run as root
|
||||||
|
'su builder',
|
||||||
|
'git clone --single-branch --depth 1 --branch $repo.branch $repo.url repo',
|
||||||
|
'cd repo',
|
||||||
|
'makepkg --nobuild --syncdeps --needed --noconfirm',
|
||||||
|
'source PKGBUILD',
|
||||||
|
// The build container checks whether the package is already present on
|
||||||
|
// the server.
|
||||||
|
'curl -s --head --fail $repo_url/$build_arch/\$pkgname-\$pkgver-\$pkgrel && exit 0',
|
||||||
|
// If the above curl command succeeds, we don't need to rebuild the
|
||||||
|
// package. However, because we're in a su shell, the exit command will
|
||||||
|
// drop us back into the root shell. Therefore, we must check whether
|
||||||
|
// we're in root so we don't proceed.
|
||||||
|
'[ "\$(id -u)" == 0 ] && exit 0',
|
||||||
|
'MAKEFLAGS="-j\$(nproc)" makepkg -s --noconfirm --needed && for pkg in \$(ls -1 *.pkg*); do curl -XPOST -T "\$pkg" -H "X-API-KEY: \$API_KEY" $repo_url/publish; done',
|
||||||
|
])
|
||||||
|
|
||||||
|
return commands.join('\n')
|
||||||
|
}
|
||||||
|
|
@ -0,0 +1,16 @@
|
||||||
|
module build
|
||||||
|
|
||||||
|
import models { GitRepo }
|
||||||
|
|
||||||
|
fn test_create_build_script() {
|
||||||
|
repo := GitRepo{
|
||||||
|
id: 1
|
||||||
|
url: 'https://examplerepo.com'
|
||||||
|
branch: 'main'
|
||||||
|
repo: 'vieter'
|
||||||
|
}
|
||||||
|
build_script := create_build_script('https://example.com', repo, 'x86_64')
|
||||||
|
expected := $embed_file('build_script.sh')
|
||||||
|
|
||||||
|
assert build_script == expected.to_string().trim_space()
|
||||||
|
}
|
||||||
|
|
@ -29,7 +29,10 @@ fn (c &Client) send_request_raw(method Method, url string, params map[string]str
|
||||||
|
|
||||||
// Escape each query param
|
// Escape each query param
|
||||||
for k, v in params {
|
for k, v in params {
|
||||||
params_escaped[k] = urllib.query_escape(v)
|
// An empty parameter should be the same as not providing it at all
|
||||||
|
if v != '' {
|
||||||
|
params_escaped[k] = urllib.query_escape(v)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
params_str := params_escaped.keys().map('$it=${params[it]}').join('&')
|
params_str := params_escaped.keys().map('$it=${params[it]}').join('&')
|
||||||
|
|
|
||||||
|
|
@ -1,16 +1,38 @@
|
||||||
module client
|
module client
|
||||||
|
|
||||||
import db { GitRepo }
|
import models { GitRepo, GitRepoFilter }
|
||||||
import net.http { Method }
|
import net.http { Method }
|
||||||
import response { Response }
|
import response { Response }
|
||||||
|
|
||||||
// get_git_repos returns the current list of repos.
|
// get_git_repos returns a list of GitRepo's, given a filter object.
|
||||||
pub fn (c &Client) get_git_repos() ?[]GitRepo {
|
pub fn (c &Client) get_git_repos(filter GitRepoFilter) ?[]GitRepo {
|
||||||
data := c.send_request<[]GitRepo>(Method.get, '/api/repos', {})?
|
params := models.params_from(filter)
|
||||||
|
data := c.send_request<[]GitRepo>(Method.get, '/api/repos', params)?
|
||||||
|
|
||||||
return data.data
|
return data.data
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// get_all_git_repos retrieves *all* GitRepo's from the API using the default
|
||||||
|
// limit.
|
||||||
|
pub fn (c &Client) get_all_git_repos() ?[]GitRepo {
|
||||||
|
mut repos := []GitRepo{}
|
||||||
|
mut offset := u64(0)
|
||||||
|
|
||||||
|
for {
|
||||||
|
sub_repos := c.get_git_repos(offset: offset)?
|
||||||
|
|
||||||
|
if sub_repos.len == 0 {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
repos << sub_repos
|
||||||
|
|
||||||
|
offset += u64(sub_repos.len)
|
||||||
|
}
|
||||||
|
|
||||||
|
return repos
|
||||||
|
}
|
||||||
|
|
||||||
// get_git_repo returns the repo for a specific ID.
|
// get_git_repo returns the repo for a specific ID.
|
||||||
pub fn (c &Client) get_git_repo(id int) ?GitRepo {
|
pub fn (c &Client) get_git_repo(id int) ?GitRepo {
|
||||||
data := c.send_request<GitRepo>(Method.get, '/api/repos/$id', {})?
|
data := c.send_request<GitRepo>(Method.get, '/api/repos/$id', {})?
|
||||||
|
|
|
||||||
|
|
@ -1,13 +1,14 @@
|
||||||
module client
|
module client
|
||||||
|
|
||||||
import db { BuildLog }
|
import models { BuildLog, BuildLogFilter }
|
||||||
import net.http { Method }
|
import net.http { Method }
|
||||||
import response { Response }
|
import response { Response }
|
||||||
import time
|
import time
|
||||||
|
|
||||||
// get_build_logs returns all build logs.
|
// get_build_logs returns all build logs.
|
||||||
pub fn (c &Client) get_build_logs() ?Response<[]BuildLog> {
|
pub fn (c &Client) get_build_logs(filter BuildLogFilter) ?Response<[]BuildLog> {
|
||||||
data := c.send_request<[]BuildLog>(Method.get, '/api/logs', {})?
|
params := models.params_from(filter)
|
||||||
|
data := c.send_request<[]BuildLog>(Method.get, '/api/logs', params)?
|
||||||
|
|
||||||
return data
|
return data
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -5,6 +5,7 @@ import env
|
||||||
import cron.expression { parse_expression }
|
import cron.expression { parse_expression }
|
||||||
import client
|
import client
|
||||||
import console
|
import console
|
||||||
|
import models { GitRepoFilter }
|
||||||
|
|
||||||
struct Config {
|
struct Config {
|
||||||
address string [required]
|
address string [required]
|
||||||
|
|
@ -21,11 +22,50 @@ pub fn cmd() cli.Command {
|
||||||
cli.Command{
|
cli.Command{
|
||||||
name: 'list'
|
name: 'list'
|
||||||
description: 'List the current repos.'
|
description: 'List the current repos.'
|
||||||
|
flags: [
|
||||||
|
cli.Flag{
|
||||||
|
name: 'limit'
|
||||||
|
description: 'How many results to return.'
|
||||||
|
flag: cli.FlagType.int
|
||||||
|
},
|
||||||
|
cli.Flag{
|
||||||
|
name: 'offset'
|
||||||
|
description: 'Minimum index to return.'
|
||||||
|
flag: cli.FlagType.int
|
||||||
|
},
|
||||||
|
cli.Flag{
|
||||||
|
name: 'repo'
|
||||||
|
description: 'Only return Git repos that publish to this repo.'
|
||||||
|
flag: cli.FlagType.string
|
||||||
|
},
|
||||||
|
cli.Flag{
|
||||||
|
name: 'arch'
|
||||||
|
description: 'Only return repos enabled for this architecture.'
|
||||||
|
flag: cli.FlagType.string
|
||||||
|
},
|
||||||
|
]
|
||||||
execute: fn (cmd cli.Command) ? {
|
execute: fn (cmd cli.Command) ? {
|
||||||
config_file := cmd.flags.get_string('config-file')?
|
config_file := cmd.flags.get_string('config-file')?
|
||||||
conf := env.load<Config>(config_file)?
|
conf := env.load<Config>(config_file)?
|
||||||
|
|
||||||
list(conf)?
|
mut filter := GitRepoFilter{}
|
||||||
|
|
||||||
|
limit := cmd.flags.get_int('limit')?
|
||||||
|
if limit != 0 {
|
||||||
|
filter.limit = u64(limit)
|
||||||
|
}
|
||||||
|
|
||||||
|
offset := cmd.flags.get_int('offset')?
|
||||||
|
if offset != 0 {
|
||||||
|
filter.offset = u64(offset)
|
||||||
|
}
|
||||||
|
|
||||||
|
repo := cmd.flags.get_string('repo')?
|
||||||
|
if repo != '' {
|
||||||
|
filter.repo = repo
|
||||||
|
}
|
||||||
|
|
||||||
|
list(conf, filter)?
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
cli.Command{
|
cli.Command{
|
||||||
|
|
@ -133,9 +173,9 @@ pub fn cmd() cli.Command {
|
||||||
// ID. If multiple or none are found, an error is raised.
|
// ID. If multiple or none are found, an error is raised.
|
||||||
|
|
||||||
// list prints out a list of all repositories.
|
// list prints out a list of all repositories.
|
||||||
fn list(conf Config) ? {
|
fn list(conf Config, filter GitRepoFilter) ? {
|
||||||
c := client.new(conf.address, conf.api_key)
|
c := client.new(conf.address, conf.api_key)
|
||||||
repos := c.get_git_repos()?
|
repos := c.get_git_repos(filter)?
|
||||||
data := repos.map([it.id.str(), it.url, it.branch, it.repo])
|
data := repos.map([it.id.str(), it.url, it.branch, it.repo])
|
||||||
|
|
||||||
println(console.pretty_table(['id', 'url', 'branch', 'repo'], data)?)
|
println(console.pretty_table(['id', 'url', 'branch', 'repo'], data)?)
|
||||||
|
|
|
||||||
|
|
@ -3,8 +3,9 @@ module logs
|
||||||
import cli
|
import cli
|
||||||
import env
|
import env
|
||||||
import client
|
import client
|
||||||
import db
|
|
||||||
import console
|
import console
|
||||||
|
import time
|
||||||
|
import models { BuildLog, BuildLogFilter }
|
||||||
|
|
||||||
struct Config {
|
struct Config {
|
||||||
address string [required]
|
address string [required]
|
||||||
|
|
@ -19,21 +20,120 @@ pub fn cmd() cli.Command {
|
||||||
commands: [
|
commands: [
|
||||||
cli.Command{
|
cli.Command{
|
||||||
name: 'list'
|
name: 'list'
|
||||||
description: 'List the build logs. If a repo ID is provided, only list the build logs for that repo.'
|
description: 'List build logs. All date strings in the output are converted to the local timezone. Any time strings provided as input should be in the local timezone as well.'
|
||||||
flags: [
|
flags: [
|
||||||
cli.Flag{
|
cli.Flag{
|
||||||
name: 'repo'
|
name: 'limit'
|
||||||
description: 'ID of the Git repo to restrict list to.'
|
description: 'How many results to return.'
|
||||||
flag: cli.FlagType.int
|
flag: cli.FlagType.int
|
||||||
},
|
},
|
||||||
|
cli.Flag{
|
||||||
|
name: 'offset'
|
||||||
|
description: 'Minimum index to return.'
|
||||||
|
flag: cli.FlagType.int
|
||||||
|
},
|
||||||
|
cli.Flag{
|
||||||
|
name: 'repo'
|
||||||
|
description: 'Only return logs for this repo id.'
|
||||||
|
flag: cli.FlagType.int
|
||||||
|
},
|
||||||
|
cli.Flag{
|
||||||
|
name: 'today'
|
||||||
|
description: 'Only list logs started today.'
|
||||||
|
flag: cli.FlagType.bool
|
||||||
|
},
|
||||||
|
cli.Flag{
|
||||||
|
name: 'failed'
|
||||||
|
description: 'Only list logs with non-zero exit codes.'
|
||||||
|
flag: cli.FlagType.bool
|
||||||
|
},
|
||||||
|
cli.Flag{
|
||||||
|
name: 'day'
|
||||||
|
description: 'Only list logs started on this day. (format: YYYY-MM-DD)'
|
||||||
|
flag: cli.FlagType.string
|
||||||
|
},
|
||||||
|
cli.Flag{
|
||||||
|
name: 'before'
|
||||||
|
description: 'Only list logs started before this timestamp. (format: YYYY-MM-DD HH:mm:ss)'
|
||||||
|
flag: cli.FlagType.string
|
||||||
|
},
|
||||||
|
cli.Flag{
|
||||||
|
name: 'after'
|
||||||
|
description: 'Only list logs started after this timestamp. (format: YYYY-MM-DD HH:mm:ss)'
|
||||||
|
flag: cli.FlagType.string
|
||||||
|
},
|
||||||
]
|
]
|
||||||
execute: fn (cmd cli.Command) ? {
|
execute: fn (cmd cli.Command) ? {
|
||||||
config_file := cmd.flags.get_string('config-file')?
|
config_file := cmd.flags.get_string('config-file')?
|
||||||
conf := env.load<Config>(config_file)?
|
conf := env.load<Config>(config_file)?
|
||||||
|
|
||||||
repo_id := cmd.flags.get_int('repo')?
|
mut filter := BuildLogFilter{}
|
||||||
|
|
||||||
if repo_id == 0 { list(conf)? } else { list_for_repo(conf, repo_id)? }
|
limit := cmd.flags.get_int('limit')?
|
||||||
|
if limit != 0 {
|
||||||
|
filter.limit = u64(limit)
|
||||||
|
}
|
||||||
|
|
||||||
|
offset := cmd.flags.get_int('offset')?
|
||||||
|
if offset != 0 {
|
||||||
|
filter.offset = u64(offset)
|
||||||
|
}
|
||||||
|
|
||||||
|
repo_id := cmd.flags.get_int('repo')?
|
||||||
|
if repo_id != 0 {
|
||||||
|
filter.repo = repo_id
|
||||||
|
}
|
||||||
|
|
||||||
|
tz_offset := time.offset()
|
||||||
|
|
||||||
|
if cmd.flags.get_bool('today')? {
|
||||||
|
today := time.now()
|
||||||
|
|
||||||
|
filter.after = time.new_time(time.Time{
|
||||||
|
year: today.year
|
||||||
|
month: today.month
|
||||||
|
day: today.day
|
||||||
|
}).add_seconds(-tz_offset)
|
||||||
|
filter.before = filter.after.add_days(1)
|
||||||
|
}
|
||||||
|
// The -today flag overwrites any of the other date flags.
|
||||||
|
else {
|
||||||
|
day_str := cmd.flags.get_string('day')?
|
||||||
|
before_str := cmd.flags.get_string('before')?
|
||||||
|
after_str := cmd.flags.get_string('after')?
|
||||||
|
|
||||||
|
if day_str != '' {
|
||||||
|
day := time.parse_rfc3339(day_str)?
|
||||||
|
day_utc := time.new_time(time.Time{
|
||||||
|
year: day.year
|
||||||
|
month: day.month
|
||||||
|
day: day.day
|
||||||
|
}).add_seconds(-tz_offset)
|
||||||
|
|
||||||
|
// The extra -1 is so we also return logs that
|
||||||
|
// started at exactly midnight (filter bounds are
|
||||||
|
// exclusive). therefore, we have to request logs
|
||||||
|
// started after 23:59:59 the previous day.
|
||||||
|
filter.after = day_utc.add_seconds(-1)
|
||||||
|
filter.before = day_utc.add_days(1)
|
||||||
|
} else {
|
||||||
|
if before_str != '' {
|
||||||
|
filter.before = time.parse(before_str)?.add_seconds(-tz_offset)
|
||||||
|
}
|
||||||
|
|
||||||
|
if after_str != '' {
|
||||||
|
filter.after = time.parse(after_str)?.add_seconds(-tz_offset)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if cmd.flags.get_bool('failed')? {
|
||||||
|
filter.exit_codes = [
|
||||||
|
'!0',
|
||||||
|
]
|
||||||
|
}
|
||||||
|
|
||||||
|
list(conf, filter)?
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
cli.Command{
|
cli.Command{
|
||||||
|
|
@ -67,17 +167,17 @@ pub fn cmd() cli.Command {
|
||||||
}
|
}
|
||||||
|
|
||||||
// print_log_list prints a list of logs.
|
// print_log_list prints a list of logs.
|
||||||
fn print_log_list(logs []db.BuildLog) ? {
|
fn print_log_list(logs []BuildLog) ? {
|
||||||
data := logs.map([it.id.str(), it.repo_id.str(), it.start_time.str(),
|
data := logs.map([it.id.str(), it.repo_id.str(), it.start_time.local().str(),
|
||||||
it.exit_code.str()])
|
it.exit_code.str()])
|
||||||
|
|
||||||
println(console.pretty_table(['id', 'repo', 'start time', 'exit code'], data)?)
|
println(console.pretty_table(['id', 'repo', 'start time', 'exit code'], data)?)
|
||||||
}
|
}
|
||||||
|
|
||||||
// list prints a list of all build logs.
|
// list prints a list of all build logs.
|
||||||
fn list(conf Config) ? {
|
fn list(conf Config, filter BuildLogFilter) ? {
|
||||||
c := client.new(conf.address, conf.api_key)
|
c := client.new(conf.address, conf.api_key)
|
||||||
logs := c.get_build_logs()?.data
|
logs := c.get_build_logs(filter)?.data
|
||||||
|
|
||||||
print_log_list(logs)?
|
print_log_list(logs)?
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,30 @@
|
||||||
|
module schedule
|
||||||
|
|
||||||
|
import cli
|
||||||
|
import cron.expression { parse_expression }
|
||||||
|
import time
|
||||||
|
|
||||||
|
// cmd returns the cli submodule for previewing a cron schedule.
|
||||||
|
pub fn cmd() cli.Command {
|
||||||
|
return cli.Command{
|
||||||
|
name: 'schedule'
|
||||||
|
usage: 'schedule'
|
||||||
|
description: 'Preview the behavior of a cron schedule.'
|
||||||
|
flags: [
|
||||||
|
cli.Flag{
|
||||||
|
name: 'count'
|
||||||
|
description: 'How many scheduled times to show.'
|
||||||
|
flag: cli.FlagType.int
|
||||||
|
default_value: ['5']
|
||||||
|
},
|
||||||
|
]
|
||||||
|
execute: fn (cmd cli.Command) ? {
|
||||||
|
ce := parse_expression(cmd.args.join(' '))?
|
||||||
|
count := cmd.flags.get_int('count')?
|
||||||
|
|
||||||
|
for t in ce.next_n(time.now(), count)? {
|
||||||
|
println(t)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
@ -7,9 +7,9 @@ import cron.expression { CronExpression, parse_expression }
|
||||||
import math
|
import math
|
||||||
import build
|
import build
|
||||||
import docker
|
import docker
|
||||||
import db
|
|
||||||
import os
|
import os
|
||||||
import client
|
import client
|
||||||
|
import models { GitRepo }
|
||||||
|
|
||||||
const (
|
const (
|
||||||
// How many seconds to wait before retrying to update API if failed
|
// How many seconds to wait before retrying to update API if failed
|
||||||
|
|
@ -20,7 +20,7 @@ const (
|
||||||
|
|
||||||
struct ScheduledBuild {
|
struct ScheduledBuild {
|
||||||
pub:
|
pub:
|
||||||
repo db.GitRepo
|
repo GitRepo
|
||||||
timestamp time.Time
|
timestamp time.Time
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -38,7 +38,7 @@ mut:
|
||||||
api_update_frequency int
|
api_update_frequency int
|
||||||
image_rebuild_frequency int
|
image_rebuild_frequency int
|
||||||
// Repos currently loaded from API.
|
// Repos currently loaded from API.
|
||||||
repos []db.GitRepo
|
repos []GitRepo
|
||||||
// At what point to update the list of repositories.
|
// At what point to update the list of repositories.
|
||||||
api_update_timestamp time.Time
|
api_update_timestamp time.Time
|
||||||
image_build_timestamp time.Time
|
image_build_timestamp time.Time
|
||||||
|
|
@ -149,7 +149,7 @@ pub fn (mut d Daemon) run() {
|
||||||
}
|
}
|
||||||
|
|
||||||
// schedule_build adds the next occurence of the given repo build to the queue.
|
// schedule_build adds the next occurence of the given repo build to the queue.
|
||||||
fn (mut d Daemon) schedule_build(repo db.GitRepo) {
|
fn (mut d Daemon) schedule_build(repo GitRepo) {
|
||||||
ce := if repo.schedule != '' {
|
ce := if repo.schedule != '' {
|
||||||
parse_expression(repo.schedule) or {
|
parse_expression(repo.schedule) or {
|
||||||
// TODO This shouldn't return an error if the expression is empty.
|
// TODO This shouldn't return an error if the expression is empty.
|
||||||
|
|
@ -178,7 +178,7 @@ fn (mut d Daemon) schedule_build(repo db.GitRepo) {
|
||||||
fn (mut d Daemon) renew_repos() {
|
fn (mut d Daemon) renew_repos() {
|
||||||
d.linfo('Renewing repos...')
|
d.linfo('Renewing repos...')
|
||||||
|
|
||||||
mut new_repos := d.client.get_git_repos() or {
|
mut new_repos := d.client.get_all_git_repos() or {
|
||||||
d.lerror('Failed to renew repos. Retrying in ${daemon.api_update_retry_timeout}s...')
|
d.lerror('Failed to renew repos. Retrying in ${daemon.api_update_retry_timeout}s...')
|
||||||
d.api_update_timestamp = time.now().add_seconds(daemon.api_update_retry_timeout)
|
d.api_update_timestamp = time.now().add_seconds(daemon.api_update_retry_timeout)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -121,6 +121,20 @@ pub fn (ce &CronExpression) next_from_now() ?time.Time {
|
||||||
return ce.next(time.now())
|
return ce.next(time.now())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// next_n returns the n next occurences of the expression, given a starting
|
||||||
|
// time.
|
||||||
|
pub fn (ce &CronExpression) next_n(ref time.Time, n int) ?[]time.Time {
|
||||||
|
mut times := []time.Time{cap: n}
|
||||||
|
|
||||||
|
times << ce.next(ref)?
|
||||||
|
|
||||||
|
for i in 1 .. n {
|
||||||
|
times << ce.next(times[i - 1])?
|
||||||
|
}
|
||||||
|
|
||||||
|
return times
|
||||||
|
}
|
||||||
|
|
||||||
// parse_range parses a given string into a range of sorted integers, if
|
// parse_range parses a given string into a range of sorted integers, if
|
||||||
// possible.
|
// possible.
|
||||||
fn parse_range(s string, min int, max int, mut bitv []bool) ? {
|
fn parse_range(s string, min int, max int, mut bitv []bool) ? {
|
||||||
|
|
|
||||||
72
src/db/db.v
72
src/db/db.v
|
|
@ -1,21 +1,89 @@
|
||||||
module db
|
module db
|
||||||
|
|
||||||
import sqlite
|
import sqlite
|
||||||
|
import time
|
||||||
|
|
||||||
struct VieterDb {
|
struct VieterDb {
|
||||||
conn sqlite.DB
|
conn sqlite.DB
|
||||||
}
|
}
|
||||||
|
|
||||||
|
struct MigrationVersion {
|
||||||
|
id int [primary]
|
||||||
|
version int
|
||||||
|
}
|
||||||
|
|
||||||
|
const (
|
||||||
|
migrations_up = [$embed_file('migrations/001-initial/up.sql')]
|
||||||
|
migrations_down = [$embed_file('migrations/001-initial/down.sql')]
|
||||||
|
)
|
||||||
|
|
||||||
// init initializes a database & adds the correct tables.
|
// init initializes a database & adds the correct tables.
|
||||||
pub fn init(db_path string) ?VieterDb {
|
pub fn init(db_path string) ?VieterDb {
|
||||||
conn := sqlite.connect(db_path)?
|
conn := sqlite.connect(db_path)?
|
||||||
|
|
||||||
sql conn {
|
sql conn {
|
||||||
create table GitRepo
|
create table MigrationVersion
|
||||||
create table BuildLog
|
}
|
||||||
|
|
||||||
|
cur_version := sql conn {
|
||||||
|
select from MigrationVersion limit 1
|
||||||
|
}
|
||||||
|
|
||||||
|
// If there's no row yet, we add it here
|
||||||
|
if cur_version == MigrationVersion{} {
|
||||||
|
sql conn {
|
||||||
|
insert cur_version into MigrationVersion
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Apply each migration in order
|
||||||
|
for i in cur_version.version .. db.migrations_up.len {
|
||||||
|
migration := db.migrations_up[i].to_string()
|
||||||
|
|
||||||
|
version_num := i + 1
|
||||||
|
|
||||||
|
// vfmt does not like these dots
|
||||||
|
println('Applying migration $version_num' + '...')
|
||||||
|
|
||||||
|
// The sqlite library seems to not like it when multiple statements are
|
||||||
|
// passed in a single exec. Therefore, we split them & run them all
|
||||||
|
// separately.
|
||||||
|
for part in migration.split(';').map(it.trim_space()).filter(it != '') {
|
||||||
|
res := conn.exec_none(part)
|
||||||
|
|
||||||
|
if res != sqlite.sqlite_done {
|
||||||
|
return error('An error occurred while applying migration $version_num')
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// The where clause doesn't really matter, as there will always only be
|
||||||
|
// one entry anyways.
|
||||||
|
sql conn {
|
||||||
|
update MigrationVersion set version = version_num where id > 0
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return VieterDb{
|
return VieterDb{
|
||||||
conn: conn
|
conn: conn
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// row_into<T> converts an sqlite.Row into a given type T by parsing each field
|
||||||
|
// from a string according to its type.
|
||||||
|
pub fn row_into<T>(row sqlite.Row) T {
|
||||||
|
mut i := 0
|
||||||
|
mut out := T{}
|
||||||
|
|
||||||
|
$for field in T.fields {
|
||||||
|
$if field.typ is string {
|
||||||
|
out.$(field.name) = row.vals[i]
|
||||||
|
} $else $if field.typ is int {
|
||||||
|
out.$(field.name) = row.vals[i].int()
|
||||||
|
} $else $if field.typ is time.Time {
|
||||||
|
out.$(field.name) = time.unix(row.vals[i].int())
|
||||||
|
}
|
||||||
|
|
||||||
|
i += 1
|
||||||
|
}
|
||||||
|
return out
|
||||||
|
}
|
||||||
|
|
|
||||||
91
src/db/git.v
91
src/db/git.v
|
|
@ -1,85 +1,21 @@
|
||||||
module db
|
module db
|
||||||
|
|
||||||
pub struct GitRepoArch {
|
import models { GitRepo, GitRepoArch, GitRepoFilter }
|
||||||
pub:
|
|
||||||
id int [primary; sql: serial]
|
|
||||||
repo_id int [nonull]
|
|
||||||
value string [nonull]
|
|
||||||
}
|
|
||||||
|
|
||||||
// str returns a string representation.
|
|
||||||
pub fn (gra &GitRepoArch) str() string {
|
|
||||||
return gra.value
|
|
||||||
}
|
|
||||||
|
|
||||||
pub struct GitRepo {
|
|
||||||
pub mut:
|
|
||||||
id int [optional; primary; sql: serial]
|
|
||||||
// URL of the Git repository
|
|
||||||
url string [nonull]
|
|
||||||
// Branch of the Git repository to use
|
|
||||||
branch string [nonull]
|
|
||||||
// Which repo the builder should publish packages to
|
|
||||||
repo string [nonull]
|
|
||||||
// Cron schedule describing how frequently to build the repo.
|
|
||||||
schedule string [optional]
|
|
||||||
// On which architectures the package is allowed to be built. In reality,
|
|
||||||
// this controls which builders will periodically build the image.
|
|
||||||
arch []GitRepoArch [fkey: 'repo_id']
|
|
||||||
}
|
|
||||||
|
|
||||||
// str returns a string representation.
|
|
||||||
pub fn (gr &GitRepo) str() string {
|
|
||||||
mut parts := [
|
|
||||||
'id: $gr.id',
|
|
||||||
'url: $gr.url',
|
|
||||||
'branch: $gr.branch',
|
|
||||||
'repo: $gr.repo',
|
|
||||||
'schedule: $gr.schedule',
|
|
||||||
'arch: ${gr.arch.map(it.value).join(', ')}',
|
|
||||||
]
|
|
||||||
str := parts.join('\n')
|
|
||||||
|
|
||||||
return str
|
|
||||||
}
|
|
||||||
|
|
||||||
// patch_from_params patches a GitRepo from a map[string]string, usually
|
|
||||||
// provided from a web.App's params
|
|
||||||
pub fn (mut r GitRepo) patch_from_params(params map[string]string) {
|
|
||||||
$for field in GitRepo.fields {
|
|
||||||
if field.name in params {
|
|
||||||
$if field.typ is string {
|
|
||||||
r.$(field.name) = params[field.name]
|
|
||||||
// This specific type check is needed for the compiler to ensure
|
|
||||||
// our types are correct
|
|
||||||
} $else $if field.typ is []GitRepoArch {
|
|
||||||
r.$(field.name) = params[field.name].split(',').map(GitRepoArch{ value: it })
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// git_repo_from_params creates a GitRepo from a map[string]string, usually
|
|
||||||
// provided from a web.App's params
|
|
||||||
pub fn git_repo_from_params(params map[string]string) ?GitRepo {
|
|
||||||
mut repo := GitRepo{}
|
|
||||||
|
|
||||||
// If we're creating a new GitRepo, we want all fields to be present before
|
|
||||||
// "patching".
|
|
||||||
$for field in GitRepo.fields {
|
|
||||||
if field.name !in params && !field.attrs.contains('optional') {
|
|
||||||
return error('Missing parameter: ${field.name}.')
|
|
||||||
}
|
|
||||||
}
|
|
||||||
repo.patch_from_params(params)
|
|
||||||
|
|
||||||
return repo
|
|
||||||
}
|
|
||||||
|
|
||||||
// get_git_repos returns all GitRepo's in the database.
|
// get_git_repos returns all GitRepo's in the database.
|
||||||
pub fn (db &VieterDb) get_git_repos() []GitRepo {
|
pub fn (db &VieterDb) get_git_repos(filter GitRepoFilter) []GitRepo {
|
||||||
|
// This seems to currently be blocked by a bug in the ORM, I'll have to ask
|
||||||
|
// around.
|
||||||
|
if filter.repo != '' {
|
||||||
|
res := sql db.conn {
|
||||||
|
select from GitRepo where repo == filter.repo order by id limit filter.limit offset filter.offset
|
||||||
|
}
|
||||||
|
|
||||||
|
return res
|
||||||
|
}
|
||||||
|
|
||||||
res := sql db.conn {
|
res := sql db.conn {
|
||||||
select from GitRepo order by id
|
select from GitRepo order by id limit filter.limit offset filter.offset
|
||||||
}
|
}
|
||||||
|
|
||||||
return res
|
return res
|
||||||
|
|
@ -130,6 +66,7 @@ pub fn (db &VieterDb) update_git_repo(repo_id int, params map[string]string) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
values_str := values.join(', ')
|
values_str := values.join(', ')
|
||||||
|
// I think this is actual SQL & not the ORM language
|
||||||
query := 'update GitRepo set $values_str where id == $repo_id'
|
query := 'update GitRepo set $values_str where id == $repo_id'
|
||||||
|
|
||||||
db.conn.exec_none(query)
|
db.conn.exec_none(query)
|
||||||
|
|
|
||||||
|
|
@ -1,38 +1,57 @@
|
||||||
module db
|
module db
|
||||||
|
|
||||||
|
import models { BuildLog, BuildLogFilter }
|
||||||
import time
|
import time
|
||||||
|
|
||||||
pub struct BuildLog {
|
|
||||||
pub:
|
|
||||||
id int [primary; sql: serial]
|
|
||||||
repo_id int [nonull]
|
|
||||||
start_time time.Time [nonull]
|
|
||||||
end_time time.Time [nonull]
|
|
||||||
arch string [nonull]
|
|
||||||
exit_code int [nonull]
|
|
||||||
}
|
|
||||||
|
|
||||||
// str returns a string representation.
|
|
||||||
pub fn (bl &BuildLog) str() string {
|
|
||||||
mut parts := [
|
|
||||||
'id: $bl.id',
|
|
||||||
'repo id: $bl.repo_id',
|
|
||||||
'start time: $bl.start_time',
|
|
||||||
'end time: $bl.end_time',
|
|
||||||
'arch: $bl.arch',
|
|
||||||
'exit code: $bl.exit_code',
|
|
||||||
]
|
|
||||||
str := parts.join('\n')
|
|
||||||
|
|
||||||
return str
|
|
||||||
}
|
|
||||||
|
|
||||||
// get_build_logs returns all BuildLog's in the database.
|
// get_build_logs returns all BuildLog's in the database.
|
||||||
pub fn (db &VieterDb) get_build_logs() []BuildLog {
|
pub fn (db &VieterDb) get_build_logs(filter BuildLogFilter) []BuildLog {
|
||||||
res := sql db.conn {
|
mut where_parts := []string{}
|
||||||
select from BuildLog order by id
|
|
||||||
|
if filter.repo != 0 {
|
||||||
|
where_parts << 'repo_id == $filter.repo'
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if filter.before != time.Time{} {
|
||||||
|
where_parts << 'start_time < $filter.before.unix_time()'
|
||||||
|
}
|
||||||
|
|
||||||
|
if filter.after != time.Time{} {
|
||||||
|
where_parts << 'start_time > $filter.after.unix_time()'
|
||||||
|
}
|
||||||
|
|
||||||
|
// NOTE: possible SQL injection
|
||||||
|
if filter.arch != '' {
|
||||||
|
where_parts << "arch == '$filter.arch'"
|
||||||
|
}
|
||||||
|
|
||||||
|
mut parts := []string{}
|
||||||
|
|
||||||
|
for exp in filter.exit_codes {
|
||||||
|
if exp[0] == `!` {
|
||||||
|
code := exp[1..].int()
|
||||||
|
|
||||||
|
parts << 'exit_code != $code'
|
||||||
|
} else {
|
||||||
|
code := exp.int()
|
||||||
|
|
||||||
|
parts << 'exit_code == $code'
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if parts.len > 0 {
|
||||||
|
where_parts << parts.map('($it)').join(' or ')
|
||||||
|
}
|
||||||
|
|
||||||
|
mut where_str := ''
|
||||||
|
|
||||||
|
if where_parts.len > 0 {
|
||||||
|
where_str = 'where ' + where_parts.map('($it)').join(' and ')
|
||||||
|
}
|
||||||
|
|
||||||
|
query := 'select * from BuildLog $where_str limit $filter.limit offset $filter.offset'
|
||||||
|
rows, _ := db.conn.exec(query)
|
||||||
|
res := rows.map(row_into<BuildLog>(it))
|
||||||
|
|
||||||
return res
|
return res
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,3 @@
|
||||||
|
DROP TABLE IF EXISTS BuildLog;
|
||||||
|
DROP TABLE IF EXISTS GitRepoArch;
|
||||||
|
DROP TABLE IF EXISTS GitRepo;
|
||||||
|
|
@ -0,0 +1,22 @@
|
||||||
|
CREATE TABLE IF NOT EXISTS GitRepo (
|
||||||
|
id INTEGER PRIMARY KEY,
|
||||||
|
url TEXT NOT NULL,
|
||||||
|
branch TEXT NOT NULL,
|
||||||
|
repo TEXT NOT NULL,
|
||||||
|
schedule TEXT
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS GitRepoArch (
|
||||||
|
id INTEGER PRIMARY KEY,
|
||||||
|
repo_id INTEGER NOT NULL,
|
||||||
|
value TEXT NOT NULL
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS BuildLog (
|
||||||
|
id INTEGER PRIMARY KEY,
|
||||||
|
repo_id INTEGER NOT NULL,
|
||||||
|
start_time INTEGER NOT NULL,
|
||||||
|
end_time iNTEGER NOT NULL,
|
||||||
|
arch TEXT NOT NULL,
|
||||||
|
exit_code INTEGER NOT NULL
|
||||||
|
);
|
||||||
|
|
@ -83,6 +83,7 @@ pub fn (mut d DockerConn) inspect_container(id string) ?ContainerInspect {
|
||||||
|
|
||||||
mut data := json.decode(ContainerInspect, body)?
|
mut data := json.decode(ContainerInspect, body)?
|
||||||
|
|
||||||
|
// The Docker engine API *should* always return UTC time.
|
||||||
data.state.start_time = time.parse_rfc3339(data.state.start_time_str)?
|
data.state.start_time = time.parse_rfc3339(data.state.start_time_str)?
|
||||||
|
|
||||||
if data.state.status == 'exited' {
|
if data.state.status == 'exited' {
|
||||||
|
|
|
||||||
|
|
@ -5,13 +5,14 @@ import server
|
||||||
import cli
|
import cli
|
||||||
import console.git
|
import console.git
|
||||||
import console.logs
|
import console.logs
|
||||||
|
import console.schedule
|
||||||
import cron
|
import cron
|
||||||
|
|
||||||
fn main() {
|
fn main() {
|
||||||
mut app := cli.Command{
|
mut app := cli.Command{
|
||||||
name: 'vieter'
|
name: 'vieter'
|
||||||
description: 'Vieter is a lightweight implementation of an Arch repository server.'
|
description: 'Vieter is a lightweight implementation of an Arch repository server.'
|
||||||
version: '0.3.0-alpha.1'
|
version: '0.3.0-alpha.2'
|
||||||
flags: [
|
flags: [
|
||||||
cli.Flag{
|
cli.Flag{
|
||||||
flag: cli.FlagType.string
|
flag: cli.FlagType.string
|
||||||
|
|
@ -27,6 +28,7 @@ fn main() {
|
||||||
git.cmd(),
|
git.cmd(),
|
||||||
cron.cmd(),
|
cron.cmd(),
|
||||||
logs.cmd(),
|
logs.cmd(),
|
||||||
|
schedule.cmd(),
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
app.setup()
|
app.setup()
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,52 @@
|
||||||
|
module models
|
||||||
|
|
||||||
|
pub struct GitRepoArch {
|
||||||
|
pub:
|
||||||
|
id int [primary; sql: serial]
|
||||||
|
repo_id int [nonull]
|
||||||
|
value string [nonull]
|
||||||
|
}
|
||||||
|
|
||||||
|
// str returns a string representation.
|
||||||
|
pub fn (gra &GitRepoArch) str() string {
|
||||||
|
return gra.value
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct GitRepo {
|
||||||
|
pub mut:
|
||||||
|
id int [primary; sql: serial]
|
||||||
|
// URL of the Git repository
|
||||||
|
url string [nonull]
|
||||||
|
// Branch of the Git repository to use
|
||||||
|
branch string [nonull]
|
||||||
|
// Which repo the builder should publish packages to
|
||||||
|
repo string [nonull]
|
||||||
|
// Cron schedule describing how frequently to build the repo.
|
||||||
|
schedule string
|
||||||
|
// On which architectures the package is allowed to be built. In reality,
|
||||||
|
// this controls which builders will periodically build the image.
|
||||||
|
arch []GitRepoArch [fkey: 'repo_id']
|
||||||
|
}
|
||||||
|
|
||||||
|
// str returns a string representation.
|
||||||
|
pub fn (gr &GitRepo) str() string {
|
||||||
|
mut parts := [
|
||||||
|
'id: $gr.id',
|
||||||
|
'url: $gr.url',
|
||||||
|
'branch: $gr.branch',
|
||||||
|
'repo: $gr.repo',
|
||||||
|
'schedule: $gr.schedule',
|
||||||
|
'arch: ${gr.arch.map(it.value).join(', ')}',
|
||||||
|
]
|
||||||
|
str := parts.join('\n')
|
||||||
|
|
||||||
|
return str
|
||||||
|
}
|
||||||
|
|
||||||
|
[params]
|
||||||
|
pub struct GitRepoFilter {
|
||||||
|
pub mut:
|
||||||
|
limit u64 = 25
|
||||||
|
offset u64
|
||||||
|
repo string
|
||||||
|
}
|
||||||
|
|
@ -0,0 +1,41 @@
|
||||||
|
module models
|
||||||
|
|
||||||
|
import time
|
||||||
|
|
||||||
|
pub struct BuildLog {
|
||||||
|
pub mut:
|
||||||
|
id int [primary; sql: serial]
|
||||||
|
repo_id int [nonull]
|
||||||
|
start_time time.Time [nonull]
|
||||||
|
end_time time.Time [nonull]
|
||||||
|
arch string [nonull]
|
||||||
|
exit_code int [nonull]
|
||||||
|
}
|
||||||
|
|
||||||
|
// str returns a string representation.
|
||||||
|
pub fn (bl &BuildLog) str() string {
|
||||||
|
mut parts := [
|
||||||
|
'id: $bl.id',
|
||||||
|
'repo id: $bl.repo_id',
|
||||||
|
'start time: $bl.start_time.local()',
|
||||||
|
'end time: $bl.end_time.local()',
|
||||||
|
'duration: ${bl.end_time - bl.start_time}',
|
||||||
|
'arch: $bl.arch',
|
||||||
|
'exit code: $bl.exit_code',
|
||||||
|
]
|
||||||
|
str := parts.join('\n')
|
||||||
|
|
||||||
|
return str
|
||||||
|
}
|
||||||
|
|
||||||
|
[params]
|
||||||
|
pub struct BuildLogFilter {
|
||||||
|
pub mut:
|
||||||
|
limit u64 = 25
|
||||||
|
offset u64
|
||||||
|
repo int
|
||||||
|
before time.Time
|
||||||
|
after time.Time
|
||||||
|
arch string
|
||||||
|
exit_codes []string
|
||||||
|
}
|
||||||
|
|
@ -0,0 +1,53 @@
|
||||||
|
module models
|
||||||
|
|
||||||
|
import time
|
||||||
|
|
||||||
|
// from_params<T> creates a new instance of T from the given map by parsing all
|
||||||
|
// of its fields from the map.
|
||||||
|
pub fn from_params<T>(params map[string]string) ?T {
|
||||||
|
mut o := T{}
|
||||||
|
|
||||||
|
patch_from_params<T>(mut o, params)?
|
||||||
|
|
||||||
|
return o
|
||||||
|
}
|
||||||
|
|
||||||
|
// patch_from_params<T> updates the given T object with the params defined in
|
||||||
|
// the map.
|
||||||
|
pub fn patch_from_params<T>(mut o T, params map[string]string) ? {
|
||||||
|
$for field in T.fields {
|
||||||
|
if field.name in params && params[field.name] != '' {
|
||||||
|
$if field.typ is string {
|
||||||
|
o.$(field.name) = params[field.name]
|
||||||
|
} $else $if field.typ is int {
|
||||||
|
o.$(field.name) = params[field.name].int()
|
||||||
|
} $else $if field.typ is u64 {
|
||||||
|
o.$(field.name) = params[field.name].u64()
|
||||||
|
} $else $if field.typ is []GitRepoArch {
|
||||||
|
o.$(field.name) = params[field.name].split(',').map(GitRepoArch{ value: it })
|
||||||
|
} $else $if field.typ is time.Time {
|
||||||
|
o.$(field.name) = time.unix(params[field.name].int())
|
||||||
|
} $else $if field.typ is []string {
|
||||||
|
o.$(field.name) = params[field.name].split(',')
|
||||||
|
}
|
||||||
|
} else if field.attrs.contains('nonull') {
|
||||||
|
return error('Missing parameter: ${field.name}.')
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// params_from<T> converts a given T struct into a map of strings.
|
||||||
|
pub fn params_from<T>(o &T) map[string]string {
|
||||||
|
mut out := map[string]string{}
|
||||||
|
|
||||||
|
$for field in T.fields {
|
||||||
|
$if field.typ is time.Time {
|
||||||
|
out[field.name] = o.$(field.name).unix_time().str()
|
||||||
|
} $else $if field.typ is []string {
|
||||||
|
out[field.name] = o.$(field.name).join(',')
|
||||||
|
} $else {
|
||||||
|
out[field.name] = o.$(field.name).str()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return out
|
||||||
|
}
|
||||||
|
|
@ -4,6 +4,7 @@ import web
|
||||||
import net.http
|
import net.http
|
||||||
import response { new_data_response, new_response }
|
import response { new_data_response, new_response }
|
||||||
import db
|
import db
|
||||||
|
import models { GitRepo, GitRepoArch, GitRepoFilter }
|
||||||
|
|
||||||
// get_repos returns the current list of repos.
|
// get_repos returns the current list of repos.
|
||||||
['/api/repos'; get]
|
['/api/repos'; get]
|
||||||
|
|
@ -12,7 +13,10 @@ fn (mut app App) get_repos() web.Result {
|
||||||
return app.json(http.Status.unauthorized, new_response('Unauthorized.'))
|
return app.json(http.Status.unauthorized, new_response('Unauthorized.'))
|
||||||
}
|
}
|
||||||
|
|
||||||
repos := app.db.get_git_repos()
|
filter := models.from_params<GitRepoFilter>(app.query) or {
|
||||||
|
return app.json(http.Status.bad_request, new_response('Invalid query parameters.'))
|
||||||
|
}
|
||||||
|
repos := app.db.get_git_repos(filter)
|
||||||
|
|
||||||
return app.json(http.Status.ok, new_data_response(repos))
|
return app.json(http.Status.ok, new_data_response(repos))
|
||||||
}
|
}
|
||||||
|
|
@ -44,7 +48,7 @@ fn (mut app App) post_repo() web.Result {
|
||||||
params['arch'] = app.conf.default_arch
|
params['arch'] = app.conf.default_arch
|
||||||
}
|
}
|
||||||
|
|
||||||
new_repo := db.git_repo_from_params(params) or {
|
new_repo := models.from_params<GitRepo>(params) or {
|
||||||
return app.json(http.Status.bad_request, new_response(err.msg()))
|
return app.json(http.Status.bad_request, new_response(err.msg()))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -75,7 +79,7 @@ fn (mut app App) patch_repo(id int) web.Result {
|
||||||
app.db.update_git_repo(id, app.query)
|
app.db.update_git_repo(id, app.query)
|
||||||
|
|
||||||
if 'arch' in app.query {
|
if 'arch' in app.query {
|
||||||
arch_objs := app.query['arch'].split(',').map(db.GitRepoArch{ value: it })
|
arch_objs := app.query['arch'].split(',').map(GitRepoArch{ value: it })
|
||||||
|
|
||||||
app.db.update_git_repo_archs(id, arch_objs)
|
app.db.update_git_repo_archs(id, arch_objs)
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -8,6 +8,7 @@ import db
|
||||||
import time
|
import time
|
||||||
import os
|
import os
|
||||||
import util
|
import util
|
||||||
|
import models { BuildLog, BuildLogFilter }
|
||||||
|
|
||||||
// get_logs returns all build logs in the database. A 'repo' query param can
|
// get_logs returns all build logs in the database. A 'repo' query param can
|
||||||
// optionally be added to limit the list of build logs to that repository.
|
// optionally be added to limit the list of build logs to that repository.
|
||||||
|
|
@ -17,11 +18,10 @@ fn (mut app App) get_logs() web.Result {
|
||||||
return app.json(http.Status.unauthorized, new_response('Unauthorized.'))
|
return app.json(http.Status.unauthorized, new_response('Unauthorized.'))
|
||||||
}
|
}
|
||||||
|
|
||||||
logs := if 'repo' in app.query {
|
filter := models.from_params<BuildLogFilter>(app.query) or {
|
||||||
app.db.get_build_logs_for_repo(app.query['repo'].int())
|
return app.json(http.Status.bad_request, new_response('Invalid query parameters.'))
|
||||||
} else {
|
|
||||||
app.db.get_build_logs()
|
|
||||||
}
|
}
|
||||||
|
logs := app.db.get_build_logs(filter)
|
||||||
|
|
||||||
return app.json(http.Status.ok, new_data_response(logs))
|
return app.json(http.Status.ok, new_data_response(logs))
|
||||||
}
|
}
|
||||||
|
|
@ -97,7 +97,7 @@ fn (mut app App) post_log() web.Result {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Store log in db
|
// Store log in db
|
||||||
log := db.BuildLog{
|
log := BuildLog{
|
||||||
repo_id: repo_id
|
repo_id: repo_id
|
||||||
start_time: start_time
|
start_time: start_time
|
||||||
end_time: end_time
|
end_time: end_time
|
||||||
|
|
|
||||||
|
|
@ -68,7 +68,9 @@ pub fn server(conf Config) ? {
|
||||||
}
|
}
|
||||||
|
|
||||||
db_file := os.join_path_single(conf.data_dir, server.db_file_name)
|
db_file := os.join_path_single(conf.data_dir, server.db_file_name)
|
||||||
db := db.init(db_file) or { util.exit_with_message(1, 'Failed to initialize database.') }
|
db := db.init(db_file) or {
|
||||||
|
util.exit_with_message(1, 'Failed to initialize database: $err.msg()')
|
||||||
|
}
|
||||||
|
|
||||||
web.run(&App{
|
web.run(&App{
|
||||||
logger: logger
|
logger: logger
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue