forked from vieter-v/vieter
Merge pull request 'Release 0.3.0-alpha.2' (#185) from release-0.3.0-alpha.2 into main
Reviewed-on: vieter/vieter#185main 0.3.0-alpha.2
commit
7627b28bcf
|
@ -0,0 +1,39 @@
|
|||
matrix:
|
||||
PLATFORM:
|
||||
- linux/amd64
|
||||
- linux/arm64
|
||||
|
||||
platform: ${PLATFORM}
|
||||
branches: [main]
|
||||
skip_clone: true
|
||||
|
||||
pipeline:
|
||||
build:
|
||||
image: 'menci/archlinuxarm:base-devel'
|
||||
commands:
|
||||
# Add the vieter repository so we can use the compiler
|
||||
- echo -e '[vieter]\nServer = https://arch.r8r.be/$repo/$arch\nSigLevel = Optional' >> /etc/pacman.conf
|
||||
# Update packages
|
||||
- pacman -Syu --noconfirm
|
||||
# Create non-root user to perform build & switch to their home
|
||||
- groupadd -g 1000 builder
|
||||
- useradd -mg builder builder
|
||||
- chown -R builder:builder "$PWD"
|
||||
- "echo 'builder ALL=(ALL) NOPASSWD: ALL' >> /etc/sudoers"
|
||||
- su builder
|
||||
# Due to a bug with the V compiler, we can't just use the PKGBUILD from
|
||||
# inside the repo
|
||||
- curl -OL "https://git.rustybever.be/vieter/vieter/raw/tag/$CI_COMMIT_TAG/PKGBUILD"
|
||||
- makepkg -s --noconfirm --needed
|
||||
when:
|
||||
event: tag
|
||||
|
||||
publish:
|
||||
image: 'curlimages/curl'
|
||||
commands:
|
||||
# Publish the package
|
||||
- 'for pkg in $(ls -1 *.pkg*); do curl -XPOST -T "$pkg" -H "X-API-KEY: $VIETER_API_KEY" https://arch.r8r.be/vieter/publish; done'
|
||||
secrets:
|
||||
- vieter_api_key
|
||||
when:
|
||||
event: tag
|
|
@ -23,7 +23,7 @@ pipeline:
|
|||
- su builder
|
||||
# Due to a bug with the V compiler, we can't just use the PKGBUILD from
|
||||
# inside the repo
|
||||
- curl -OL https://git.rustybever.be/vieter/vieter/raw/branch/dev/PKGBUILD
|
||||
- curl -o PKGBUILD -L https://git.rustybever.be/vieter/vieter/raw/branch/dev/PKGBUILD.dev
|
||||
- makepkg -s --noconfirm --needed
|
||||
when:
|
||||
event: push
|
||||
|
|
|
@ -11,7 +11,9 @@ pipeline:
|
|||
- 'docker_password'
|
||||
settings:
|
||||
repo: 'chewingbever/vieter'
|
||||
tag: 'dev'
|
||||
tags:
|
||||
- 'dev'
|
||||
- ${CI_COMMIT_SHA}
|
||||
platforms: [ 'linux/arm64/v8', 'linux/amd64' ]
|
||||
build_args_from_env:
|
||||
- 'CI_COMMIT_SHA'
|
||||
|
|
27
CHANGELOG.md
27
CHANGELOG.md
|
@ -5,6 +5,33 @@ All notable changes to this project will be documented in this file.
|
|||
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/),
|
||||
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
|
||||
|
||||
## [Unreleased](https://git.rustybever.be/vieter/vieter/src/branch/dev)
|
||||
|
||||
## [0.3.0-alpha.2](https://git.rustybever.be/vieter/vieter/src/tag/0.3.0-alpha.2)
|
||||
|
||||
### Added
|
||||
|
||||
* Web API for adding & querying build logs
|
||||
* CLI commands to access build logs API
|
||||
* Cron build logs are uploaded to above API
|
||||
* Proper ASCII table output in CLI
|
||||
* `vieter repos build id` command to run builds locally
|
||||
|
||||
### Removed
|
||||
|
||||
* `vieter build` command
|
||||
* This command was used alongside cron for periodic builds, but this has
|
||||
been replaced by `vieter cron`
|
||||
|
||||
### Changed
|
||||
|
||||
* `vieter build` command now only builds a single repository & uploads the
|
||||
build logs
|
||||
* Official Arch packages are now split between `vieter` & `vieter-git`
|
||||
* `vieter` is the latest release
|
||||
* `vieter-git` is the latest commit on the dev branch
|
||||
* Full refactor of Docker socket code
|
||||
|
||||
## [0.3.0-alpha.1](https://git.rustybever.be/vieter/vieter/src/tag/0.3.0-alpha.1)
|
||||
|
||||
### Changed
|
||||
|
|
|
@ -36,15 +36,8 @@ ENV PATH=/bin \
|
|||
|
||||
COPY --from=builder /app/dumb-init /app/vieter /bin/
|
||||
|
||||
HEALTHCHECK --interval=30s \
|
||||
--timeout=3s \
|
||||
--start-period=5s \
|
||||
CMD /bin/wget --spider http://localhost:8000/health || exit 1
|
||||
|
||||
RUN mkdir /data && \
|
||||
chown -R www-data:www-data /data && \
|
||||
mkdir -p '/var/spool/cron/crontabs' && \
|
||||
echo '0 3 * * * /bin/vieter build' | crontab -
|
||||
chown -R www-data:www-data /data
|
||||
|
||||
WORKDIR /data
|
||||
|
||||
|
|
17
PKGBUILD
17
PKGBUILD
|
@ -1,23 +1,18 @@
|
|||
# vim: ft=bash
|
||||
# Maintainer: Jef Roosens
|
||||
|
||||
pkgbase='vieter'
|
||||
pkgname='vieter'
|
||||
pkgver=0.2.0.r25.g20112b8
|
||||
pkgver='0.3.0_alpha.2'
|
||||
pkgrel=1
|
||||
depends=('glibc' 'openssl' 'libarchive' 'gc' 'sqlite')
|
||||
makedepends=('git' 'gcc' 'vieter-v')
|
||||
depends=('glibc' 'openssl' 'libarchive' 'sqlite')
|
||||
makedepends=('git' 'vieter-v')
|
||||
arch=('x86_64' 'aarch64')
|
||||
url='https://git.rustybever.be/vieter/vieter'
|
||||
license=('AGPL3')
|
||||
source=($pkgname::git+https://git.rustybever.be/vieter/vieter#branch=dev)
|
||||
source=("$pkgname::git+https://git.rustybever.be/vieter/vieter#tag=${pkgver//_/-}")
|
||||
md5sums=('SKIP')
|
||||
|
||||
pkgver() {
|
||||
cd "$pkgname"
|
||||
|
||||
git describe --long --tags | sed 's/^v//;s/\([^-]*-g\)/r\1/;s/-/./g'
|
||||
}
|
||||
|
||||
build() {
|
||||
cd "$pkgname"
|
||||
|
||||
|
@ -28,5 +23,5 @@ package() {
|
|||
pkgdesc="Vieter is a lightweight implementation of an Arch repository server."
|
||||
|
||||
install -dm755 "$pkgdir/usr/bin"
|
||||
install -Dm755 "$pkgbase/pvieter" "$pkgdir/usr/bin/vieter"
|
||||
install -Dm755 "$pkgname/pvieter" "$pkgdir/usr/bin/vieter"
|
||||
}
|
||||
|
|
|
@ -0,0 +1,35 @@
|
|||
# vim: ft=bash
|
||||
# Maintainer: Jef Roosens
|
||||
|
||||
pkgbase='vieter-git'
|
||||
pkgname='vieter-git'
|
||||
pkgver=0.2.0.r25.g20112b8
|
||||
pkgrel=1
|
||||
depends=('glibc' 'openssl' 'libarchive' 'sqlite')
|
||||
makedepends=('git' 'vieter-v')
|
||||
arch=('x86_64' 'aarch64')
|
||||
url='https://git.rustybever.be/vieter/vieter'
|
||||
license=('AGPL3')
|
||||
source=("$pkgname::git+https://git.rustybever.be/vieter/vieter#branch=dev")
|
||||
md5sums=('SKIP')
|
||||
provides=('vieter')
|
||||
conflicts=('vieter')
|
||||
|
||||
pkgver() {
|
||||
cd "$pkgname"
|
||||
|
||||
git describe --long --tags | sed 's/^v//;s/\([^-]*-g\)/r\1/;s/-/./g'
|
||||
}
|
||||
|
||||
build() {
|
||||
cd "$pkgname"
|
||||
|
||||
make prod
|
||||
}
|
||||
|
||||
package() {
|
||||
pkgdesc="Vieter is a lightweight implementation of an Arch repository server."
|
||||
|
||||
install -dm755 "$pkgdir/usr/bin"
|
||||
install -Dm755 "$pkgname/pvieter" "$pkgdir/usr/bin/vieter"
|
||||
}
|
10
README.md
10
README.md
|
@ -55,3 +55,13 @@ clone my compiler in the `v` directory & build it. Afterwards, you can use this
|
|||
compiler with make by prepending all make commands with `V_PATH=v/v`. If you do
|
||||
encounter this issue, please let me know so I can update my mirror & the
|
||||
codebase to fix it!
|
||||
|
||||
## Contributing
|
||||
|
||||
If you wish to contribute to the project, please take note of the following:
|
||||
|
||||
* Rebase instead of merging whenever possible, e.g. when updating your branch
|
||||
with the dev branch.
|
||||
* Please follow the
|
||||
[Conventional Commits](https://www.conventionalcommits.org/) style for your
|
||||
commit messages.
|
||||
|
|
|
@ -56,6 +56,11 @@ Vieter only supports uploading archives compressed using either gzip, zstd or
|
|||
xz at the moment.
|
||||
{{< /hint >}}
|
||||
|
||||
### `GET /health`
|
||||
|
||||
This endpoint's only use is to be used with healthchecks. It returns a JSON
|
||||
response with the message "Healthy.".
|
||||
|
||||
## API
|
||||
|
||||
All API routes require the API key to provided using the `X-Api-Key` header.
|
||||
|
|
|
@ -15,7 +15,7 @@ repositories. After the image has been created, each repository returned by
|
|||
previously created image as a base. Each container goes through the following steps:
|
||||
|
||||
1. The repository is cloned
|
||||
2. `makepkg --nobuild --nodeps` is ran to update the `pkgver` variable inside
|
||||
2. `makepkg --nobuild --syncdeps --needed --noconfirm` is ran to update the `pkgver` variable inside
|
||||
the `PKGBUILD` file
|
||||
3. A HEAD request is sent to the Vieter server to check whether the specific
|
||||
version of the package is already present. If it is, the container exits.
|
||||
|
|
|
@ -3,13 +3,15 @@ module build
|
|||
import docker
|
||||
import encoding.base64
|
||||
import time
|
||||
import git
|
||||
import os
|
||||
import db
|
||||
import strings
|
||||
import util
|
||||
|
||||
const container_build_dir = '/build'
|
||||
|
||||
const build_image_repo = 'vieter-build'
|
||||
const (
|
||||
container_build_dir = '/build'
|
||||
build_image_repo = 'vieter-build'
|
||||
)
|
||||
|
||||
// create_build_image creates a builder image given some base image which can
|
||||
// then be used to build & package Arch images. It mostly just updates the
|
||||
|
@ -17,6 +19,12 @@ const build_image_repo = 'vieter-build'
|
|||
// makepkg with. The base image should be some Linux distribution that uses
|
||||
// Pacman as its package manager.
|
||||
pub fn create_build_image(base_image string) ?string {
|
||||
mut dd := docker.new_conn()?
|
||||
|
||||
defer {
|
||||
dd.close() or {}
|
||||
}
|
||||
|
||||
commands := [
|
||||
// Update repos & install required packages
|
||||
'pacman -Syu --needed --noconfirm base-devel git'
|
||||
|
@ -46,14 +54,15 @@ pub fn create_build_image(base_image string) ?string {
|
|||
image_tag := if image_parts.len > 1 { image_parts[1] } else { 'latest' }
|
||||
|
||||
// We pull the provided image
|
||||
docker.pull_image(image_name, image_tag) ?
|
||||
dd.pull_image(image_name, image_tag)?
|
||||
|
||||
id := docker.create_container(c) ?
|
||||
docker.start_container(id) ?
|
||||
id := dd.create_container(c)?.id
|
||||
// id := docker.create_container(c)?
|
||||
dd.start_container(id)?
|
||||
|
||||
// This loop waits until the container has stopped, so we can remove it after
|
||||
for {
|
||||
data := docker.inspect_container(id) ?
|
||||
data := dd.inspect_container(id)?
|
||||
|
||||
if !data.state.running {
|
||||
break
|
||||
|
@ -67,27 +76,41 @@ pub fn create_build_image(base_image string) ?string {
|
|||
// TODO also add the base image's name into the image name to prevent
|
||||
// conflicts.
|
||||
tag := time.sys_mono_now().str()
|
||||
image := docker.create_image_from_container(id, 'vieter-build', tag) ?
|
||||
docker.remove_container(id) ?
|
||||
image := dd.create_image_from_container(id, 'vieter-build', tag)?
|
||||
dd.remove_container(id)?
|
||||
|
||||
return image.id
|
||||
}
|
||||
|
||||
pub struct BuildResult {
|
||||
pub:
|
||||
start_time time.Time
|
||||
end_time time.Time
|
||||
exit_code int
|
||||
logs string
|
||||
}
|
||||
|
||||
// build_repo builds, packages & publishes a given Arch package based on the
|
||||
// provided GitRepo. The base image ID should be of an image previously created
|
||||
// by create_build_image.
|
||||
pub fn build_repo(address string, api_key string, base_image_id string, repo &db.GitRepo) ? {
|
||||
// by create_build_image. It returns the logs of the container.
|
||||
pub fn build_repo(address string, api_key string, base_image_id string, repo &db.GitRepo) ?BuildResult {
|
||||
mut dd := docker.new_conn()?
|
||||
|
||||
defer {
|
||||
dd.close() or {}
|
||||
}
|
||||
|
||||
build_arch := os.uname().machine
|
||||
|
||||
// TODO what to do with PKGBUILDs that build multiple packages?
|
||||
commands := [
|
||||
'git clone --single-branch --depth 1 --branch $repo.branch $repo.url repo',
|
||||
'cd repo',
|
||||
'makepkg --nobuild --nodeps',
|
||||
'makepkg --nobuild --syncdeps --needed --noconfirm',
|
||||
'source PKGBUILD',
|
||||
// The build container checks whether the package is already
|
||||
// present on the server
|
||||
'curl --head --fail $address/$repo.repo/$build_arch/\$pkgname-\$pkgver-\$pkgrel && exit 0',
|
||||
'curl -s --head --fail $address/$repo.repo/$build_arch/\$pkgname-\$pkgver-\$pkgrel && exit 0',
|
||||
'MAKEFLAGS="-j\$(nproc)" makepkg -s --noconfirm --needed && for pkg in \$(ls -1 *.pkg*); do curl -XPOST -T "\$pkg" -H "X-API-KEY: \$API_KEY" $address/$repo.repo/publish; done',
|
||||
]
|
||||
|
||||
|
@ -104,46 +127,30 @@ pub fn build_repo(address string, api_key string, base_image_id string, repo &db
|
|||
user: 'builder:builder'
|
||||
}
|
||||
|
||||
id := docker.create_container(c) ?
|
||||
docker.start_container(id) ?
|
||||
id := dd.create_container(c)?.id
|
||||
dd.start_container(id)?
|
||||
|
||||
mut data := dd.inspect_container(id)?
|
||||
|
||||
// This loop waits until the container has stopped, so we can remove it after
|
||||
for {
|
||||
data := docker.inspect_container(id) ?
|
||||
|
||||
if !data.state.running {
|
||||
break
|
||||
}
|
||||
|
||||
for data.state.running {
|
||||
time.sleep(1 * time.second)
|
||||
|
||||
data = dd.inspect_container(id)?
|
||||
}
|
||||
|
||||
docker.remove_container(id) ?
|
||||
}
|
||||
|
||||
// build builds every Git repo in the server's list.
|
||||
fn build(conf Config) ? {
|
||||
build_arch := os.uname().machine
|
||||
|
||||
// We get the repos map from the Vieter instance
|
||||
repos := git.get_repos(conf.address, conf.api_key) ?
|
||||
|
||||
// We filter out any repos that aren't allowed to be built on this
|
||||
// architecture
|
||||
filtered_repos := repos.filter(it.arch.map(it.value).contains(build_arch))
|
||||
|
||||
// No point in doing work if there's no repos present
|
||||
if filtered_repos.len == 0 {
|
||||
return
|
||||
}
|
||||
|
||||
// First, we create a base image which has updated repos n stuff
|
||||
image_id := create_build_image(conf.base_image) ?
|
||||
|
||||
for repo in filtered_repos {
|
||||
build_repo(conf.address, conf.api_key, image_id, repo) ?
|
||||
}
|
||||
|
||||
// Finally, we remove the builder image
|
||||
docker.remove_image(image_id) ?
|
||||
mut logs_stream := dd.get_container_logs(id)?
|
||||
|
||||
// Read in the entire stream
|
||||
mut logs_builder := strings.new_builder(10 * 1024)
|
||||
util.reader_to_writer(mut logs_stream, mut logs_builder)?
|
||||
|
||||
dd.remove_container(id)?
|
||||
|
||||
return BuildResult{
|
||||
start_time: data.state.start_time
|
||||
end_time: data.state.end_time
|
||||
exit_code: data.state.exit_code
|
||||
logs: logs_builder.str()
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1,25 +0,0 @@
|
|||
module build
|
||||
|
||||
import cli
|
||||
import env
|
||||
|
||||
pub struct Config {
|
||||
pub:
|
||||
api_key string
|
||||
address string
|
||||
base_image string = 'archlinux:base-devel'
|
||||
}
|
||||
|
||||
// cmd returns the cli submodule that handles the build process
|
||||
pub fn cmd() cli.Command {
|
||||
return cli.Command{
|
||||
name: 'build'
|
||||
description: 'Run the build process.'
|
||||
execute: fn (cmd cli.Command) ? {
|
||||
config_file := cmd.flags.get_string('config-file') ?
|
||||
conf := env.load<Config>(config_file) ?
|
||||
|
||||
build(conf) ?
|
||||
}
|
||||
}
|
||||
}
|
|
@ -0,0 +1,67 @@
|
|||
module client
|
||||
|
||||
import net.http { Method }
|
||||
import net.urllib
|
||||
import response { Response }
|
||||
import json
|
||||
|
||||
pub struct Client {
|
||||
pub:
|
||||
address string
|
||||
api_key string
|
||||
}
|
||||
|
||||
// new creates a new Client instance.
|
||||
pub fn new(address string, api_key string) Client {
|
||||
return Client{
|
||||
address: address
|
||||
api_key: api_key
|
||||
}
|
||||
}
|
||||
|
||||
// send_request_raw sends an HTTP request, returning the http.Response object.
|
||||
// It encodes the params so that they're safe to pass as HTTP query parameters.
|
||||
fn (c &Client) send_request_raw(method Method, url string, params map[string]string, body string) ?http.Response {
|
||||
mut full_url := '$c.address$url'
|
||||
|
||||
if params.len > 0 {
|
||||
mut params_escaped := map[string]string{}
|
||||
|
||||
// Escape each query param
|
||||
for k, v in params {
|
||||
params_escaped[k] = urllib.query_escape(v)
|
||||
}
|
||||
|
||||
params_str := params_escaped.keys().map('$it=${params[it]}').join('&')
|
||||
|
||||
full_url = '$full_url?$params_str'
|
||||
}
|
||||
|
||||
mut req := http.new_request(method, full_url, body)?
|
||||
req.add_custom_header('X-Api-Key', c.api_key)?
|
||||
|
||||
res := req.do()?
|
||||
|
||||
return res
|
||||
}
|
||||
|
||||
// send_request<T> just calls send_request_with_body<T> with an empty body.
|
||||
fn (c &Client) send_request<T>(method Method, url string, params map[string]string) ?Response<T> {
|
||||
return c.send_request_with_body<T>(method, url, params, '')
|
||||
}
|
||||
|
||||
// send_request_with_body<T> calls send_request_raw_response & parses its
|
||||
// output as a Response<T> object.
|
||||
fn (c &Client) send_request_with_body<T>(method Method, url string, params map[string]string, body string) ?Response<T> {
|
||||
res_text := c.send_request_raw_response(method, url, params, body)?
|
||||
data := json.decode(Response<T>, res_text)?
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// send_request_raw_response returns the raw text response for an HTTP request.
|
||||
fn (c &Client) send_request_raw_response(method Method, url string, params map[string]string, body string) ?string {
|
||||
res := c.send_request_raw(method, url, params, body)?
|
||||
|
||||
return res.text
|
||||
}
|
|
@ -0,0 +1,51 @@
|
|||
module client
|
||||
|
||||
import db { GitRepo }
|
||||
import net.http { Method }
|
||||
import response { Response }
|
||||
|
||||
// get_git_repos returns the current list of repos.
|
||||
pub fn (c &Client) get_git_repos() ?[]GitRepo {
|
||||
data := c.send_request<[]GitRepo>(Method.get, '/api/repos', {})?
|
||||
|
||||
return data.data
|
||||
}
|
||||
|
||||
// get_git_repo returns the repo for a specific ID.
|
||||
pub fn (c &Client) get_git_repo(id int) ?GitRepo {
|
||||
data := c.send_request<GitRepo>(Method.get, '/api/repos/$id', {})?
|
||||
|
||||
return data.data
|
||||
}
|
||||
|
||||
// add_git_repo adds a new repo to the server.
|
||||
pub fn (c &Client) add_git_repo(url string, branch string, repo string, arch []string) ?Response<string> {
|
||||
mut params := {
|
||||
'url': url
|
||||
'branch': branch
|
||||
'repo': repo
|
||||
}
|
||||
|
||||
if arch.len > 0 {
|
||||
params['arch'] = arch.join(',')
|
||||
}
|
||||
|
||||
data := c.send_request<string>(Method.post, '/api/repos', params)?
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// remove_git_repo removes the repo with the given ID from the server.
|
||||
pub fn (c &Client) remove_git_repo(id int) ?Response<string> {
|
||||
data := c.send_request<string>(Method.delete, '/api/repos/$id', {})?
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// patch_git_repo sends a PATCH request to the given repo with the params as
|
||||
// payload.
|
||||
pub fn (c &Client) patch_git_repo(id int, params map[string]string) ?Response<string> {
|
||||
data := c.send_request<string>(Method.patch, '/api/repos/$id', params)?
|
||||
|
||||
return data
|
||||
}
|
|
@ -0,0 +1,53 @@
|
|||
module client
|
||||
|
||||
import db { BuildLog }
|
||||
import net.http { Method }
|
||||
import response { Response }
|
||||
import time
|
||||
|
||||
// get_build_logs returns all build logs.
|
||||
pub fn (c &Client) get_build_logs() ?Response<[]BuildLog> {
|
||||
data := c.send_request<[]BuildLog>(Method.get, '/api/logs', {})?
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// get_build_logs_for_repo returns all build logs for a given repo.
|
||||
pub fn (c &Client) get_build_logs_for_repo(repo_id int) ?Response<[]BuildLog> {
|
||||
params := {
|
||||
'repo': repo_id.str()
|
||||
}
|
||||
|
||||
data := c.send_request<[]BuildLog>(Method.get, '/api/logs', params)?
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// get_build_log returns a specific build log.
|
||||
pub fn (c &Client) get_build_log(id int) ?Response<BuildLog> {
|
||||
data := c.send_request<BuildLog>(Method.get, '/api/logs/$id', {})?
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// get_build_log_content returns the contents of the build log file.
|
||||
pub fn (c &Client) get_build_log_content(id int) ?string {
|
||||
data := c.send_request_raw_response(Method.get, '/api/logs/$id/content', {}, '')?
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// add_build_log adds a new build log to the server.
|
||||
pub fn (c &Client) add_build_log(repo_id int, start_time time.Time, end_time time.Time, arch string, exit_code int, content string) ?Response<string> {
|
||||
params := {
|
||||
'repo': repo_id.str()
|
||||
'startTime': start_time.str()
|
||||
'endTime': end_time.str()
|
||||
'arch': arch
|
||||
'exitCode': exit_code.str()
|
||||
}
|
||||
|
||||
data := c.send_request_with_body<string>(Method.post, '/api/logs', params, content)?
|
||||
|
||||
return data
|
||||
}
|
|
@ -0,0 +1,56 @@
|
|||
module console
|
||||
|
||||
import arrays
|
||||
import strings
|
||||
|
||||
// pretty_table converts a list of string data into a pretty table. Many thanks
|
||||
// to @hungrybluedev in the Vlang Discord for providing this code!
|
||||
// https://ptb.discord.com/channels/592103645835821068/592106336838352923/970278787143045192
|
||||
pub fn pretty_table(header []string, data [][]string) ?string {
|
||||
column_count := header.len
|
||||
|
||||
mut column_widths := []int{len: column_count, init: header[it].len}
|
||||
|
||||
for values in data {
|
||||
for col, value in values {
|
||||
if column_widths[col] < value.len {
|
||||
column_widths[col] = value.len
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
single_line_length := arrays.sum(column_widths)? + (column_count + 1) * 3 - 4
|
||||
|
||||
horizontal_line := '+' + strings.repeat(`-`, single_line_length) + '+'
|
||||
mut buffer := strings.new_builder(data.len * single_line_length)
|
||||
|
||||
buffer.writeln(horizontal_line)
|
||||
|
||||
buffer.write_string('| ')
|
||||
for col, head in header {
|
||||
if col != 0 {
|
||||
buffer.write_string(' | ')
|
||||
}
|
||||
buffer.write_string(head)
|
||||
buffer.write_string(strings.repeat(` `, column_widths[col] - head.len))
|
||||
}
|
||||
buffer.writeln(' |')
|
||||
|
||||
buffer.writeln(horizontal_line)
|
||||
|
||||
for values in data {
|
||||
buffer.write_string('| ')
|
||||
for col, value in values {
|
||||
if col != 0 {
|
||||
buffer.write_string(' | ')
|
||||
}
|
||||
buffer.write_string(value)
|
||||
buffer.write_string(strings.repeat(` `, column_widths[col] - value.len))
|
||||
}
|
||||
buffer.writeln(' |')
|
||||
}
|
||||
|
||||
buffer.writeln(horizontal_line)
|
||||
|
||||
return buffer.str()
|
||||
}
|
|
@ -0,0 +1,34 @@
|
|||
module git
|
||||
|
||||
import client
|
||||
import docker
|
||||
import os
|
||||
import build
|
||||
|
||||
// build builds every Git repo in the server's list.
|
||||
fn build(conf Config, repo_id int) ? {
|
||||
c := client.new(conf.address, conf.api_key)
|
||||
repo := c.get_git_repo(repo_id)?
|
||||
|
||||
build_arch := os.uname().machine
|
||||
|
||||
println('Creating base image...')
|
||||
image_id := build.create_build_image(conf.base_image)?
|
||||
|
||||
println('Running build...')
|
||||
res := build.build_repo(conf.address, conf.api_key, image_id, repo)?
|
||||
|
||||
println('Removing build image...')
|
||||
|
||||
mut dd := docker.new_conn()?
|
||||
|
||||
defer {
|
||||
dd.close() or {}
|
||||
}
|
||||
|
||||
dd.remove_image(image_id)?
|
||||
|
||||
println('Uploading logs to Vieter...')
|
||||
c.add_build_log(repo.id, res.start_time, res.end_time, build_arch, res.exit_code,
|
||||
res.logs)?
|
||||
}
|
|
@ -3,10 +3,13 @@ module git
|
|||
import cli
|
||||
import env
|
||||
import cron.expression { parse_expression }
|
||||
import client
|
||||
import console
|
||||
|
||||
struct Config {
|
||||
address string [required]
|
||||
api_key string [required]
|
||||
address string [required]
|
||||
api_key string [required]
|
||||
base_image string = 'archlinux:base-devel'
|
||||
}
|
||||
|
||||
// cmd returns the cli submodule that handles the repos API interaction
|
||||
|
@ -19,10 +22,10 @@ pub fn cmd() cli.Command {
|
|||
name: 'list'
|
||||
description: 'List the current repos.'
|
||||
execute: fn (cmd cli.Command) ? {
|
||||
config_file := cmd.flags.get_string('config-file') ?
|
||||
conf := env.load<Config>(config_file) ?
|
||||
config_file := cmd.flags.get_string('config-file')?
|
||||
conf := env.load<Config>(config_file)?
|
||||
|
||||
list(conf) ?
|
||||
list(conf)?
|
||||
}
|
||||
},
|
||||
cli.Command{
|
||||
|
@ -31,10 +34,10 @@ pub fn cmd() cli.Command {
|
|||
usage: 'url branch repo'
|
||||
description: 'Add a new repository.'
|
||||
execute: fn (cmd cli.Command) ? {
|
||||
config_file := cmd.flags.get_string('config-file') ?
|
||||
conf := env.load<Config>(config_file) ?
|
||||
config_file := cmd.flags.get_string('config-file')?
|
||||
conf := env.load<Config>(config_file)?
|
||||
|
||||
add(conf, cmd.args[0], cmd.args[1], cmd.args[2]) ?
|
||||
add(conf, cmd.args[0], cmd.args[1], cmd.args[2])?
|
||||
}
|
||||
},
|
||||
cli.Command{
|
||||
|
@ -43,10 +46,10 @@ pub fn cmd() cli.Command {
|
|||
usage: 'id'
|
||||
description: 'Remove a repository that matches the given ID prefix.'
|
||||
execute: fn (cmd cli.Command) ? {
|
||||
config_file := cmd.flags.get_string('config-file') ?
|
||||
conf := env.load<Config>(config_file) ?
|
||||
config_file := cmd.flags.get_string('config-file')?
|
||||
conf := env.load<Config>(config_file)?
|
||||
|
||||
remove(conf, cmd.args[0]) ?
|
||||
remove(conf, cmd.args[0])?
|
||||
}
|
||||
},
|
||||
cli.Command{
|
||||
|
@ -55,10 +58,10 @@ pub fn cmd() cli.Command {
|
|||
usage: 'id'
|
||||
description: 'Show detailed information for the repo matching the ID prefix.'
|
||||
execute: fn (cmd cli.Command) ? {
|
||||
config_file := cmd.flags.get_string('config-file') ?
|
||||
conf := env.load<Config>(config_file) ?
|
||||
config_file := cmd.flags.get_string('config-file')?
|
||||
conf := env.load<Config>(config_file)?
|
||||
|
||||
info(conf, cmd.args[0]) ?
|
||||
info(conf, cmd.args[0])?
|
||||
}
|
||||
},
|
||||
cli.Command{
|
||||
|
@ -94,8 +97,8 @@ pub fn cmd() cli.Command {
|
|||
},
|
||||
]
|
||||
execute: fn (cmd cli.Command) ? {
|
||||
config_file := cmd.flags.get_string('config-file') ?
|
||||
conf := env.load<Config>(config_file) ?
|
||||
config_file := cmd.flags.get_string('config-file')?
|
||||
conf := env.load<Config>(config_file)?
|
||||
|
||||
found := cmd.flags.get_all_found()
|
||||
|
||||
|
@ -103,11 +106,23 @@ pub fn cmd() cli.Command {
|
|||
|
||||
for f in found {
|
||||
if f.name != 'config-file' {
|
||||
params[f.name] = f.get_string() ?
|
||||
params[f.name] = f.get_string()?
|
||||
}
|
||||
}
|
||||
|
||||
patch(conf, cmd.args[0], params) ?
|
||||
patch(conf, cmd.args[0], params)?
|
||||
}
|
||||
},
|
||||
cli.Command{
|
||||
name: 'build'
|
||||
required_args: 1
|
||||
usage: 'id'
|
||||
description: 'Build the repo with the given id & publish it.'
|
||||
execute: fn (cmd cli.Command) ? {
|
||||
config_file := cmd.flags.get_string('config-file')?
|
||||
conf := env.load<Config>(config_file)?
|
||||
|
||||
build(conf, cmd.args[0].int())?
|
||||
}
|
||||
},
|
||||
]
|
||||
|
@ -119,16 +134,17 @@ pub fn cmd() cli.Command {
|
|||
|
||||
// list prints out a list of all repositories.
|
||||
fn list(conf Config) ? {
|
||||
repos := get_repos(conf.address, conf.api_key) ?
|
||||
c := client.new(conf.address, conf.api_key)
|
||||
repos := c.get_git_repos()?
|
||||
data := repos.map([it.id.str(), it.url, it.branch, it.repo])
|
||||
|
||||
for repo in repos {
|
||||
println('$repo.id\t$repo.url\t$repo.branch\t$repo.repo')
|
||||
}
|
||||
println(console.pretty_table(['id', 'url', 'branch', 'repo'], data)?)
|
||||
}
|
||||
|
||||
// add adds a new repository to the server's list.
|
||||
fn add(conf Config, url string, branch string, repo string) ? {
|
||||
res := add_repo(conf.address, conf.api_key, url, branch, repo, []) ?
|
||||
c := client.new(conf.address, conf.api_key)
|
||||
res := c.add_git_repo(url, branch, repo, [])?
|
||||
|
||||
println(res.message)
|
||||
}
|
||||
|
@ -139,7 +155,8 @@ fn remove(conf Config, id string) ? {
|
|||
id_int := id.int()
|
||||
|
||||
if id_int != 0 {
|
||||
res := remove_repo(conf.address, conf.api_key, id_int) ?
|
||||
c := client.new(conf.address, conf.api_key)
|
||||
res := c.remove_git_repo(id_int)?
|
||||
println(res.message)
|
||||
}
|
||||
}
|
||||
|
@ -156,7 +173,8 @@ fn patch(conf Config, id string, params map[string]string) ? {
|
|||
|
||||
id_int := id.int()
|
||||
if id_int != 0 {
|
||||
res := patch_repo(conf.address, conf.api_key, id_int, params) ?
|
||||
c := client.new(conf.address, conf.api_key)
|
||||
res := c.patch_git_repo(id_int, params)?
|
||||
|
||||
println(res.message)
|
||||
}
|
||||
|
@ -170,6 +188,7 @@ fn info(conf Config, id string) ? {
|
|||
return
|
||||
}
|
||||
|
||||
repo := get_repo(conf.address, conf.api_key, id_int) ?
|
||||
c := client.new(conf.address, conf.api_key)
|
||||
repo := c.get_git_repo(id_int)?
|
||||
println(repo)
|
||||
}
|
|
@ -0,0 +1,108 @@
|
|||
module logs
|
||||
|
||||
import cli
|
||||
import env
|
||||
import client
|
||||
import db
|
||||
import console
|
||||
|
||||
struct Config {
|
||||
address string [required]
|
||||
api_key string [required]
|
||||
}
|
||||
|
||||
// cmd returns the cli module that handles the build repos API.
|
||||
pub fn cmd() cli.Command {
|
||||
return cli.Command{
|
||||
name: 'logs'
|
||||
description: 'Interact with the build logs API.'
|
||||
commands: [
|
||||
cli.Command{
|
||||
name: 'list'
|
||||
description: 'List the build logs. If a repo ID is provided, only list the build logs for that repo.'
|
||||
flags: [
|
||||
cli.Flag{
|
||||
name: 'repo'
|
||||
description: 'ID of the Git repo to restrict list to.'
|
||||
flag: cli.FlagType.int
|
||||
},
|
||||
]
|
||||
execute: fn (cmd cli.Command) ? {
|
||||
config_file := cmd.flags.get_string('config-file')?
|
||||
conf := env.load<Config>(config_file)?
|
||||
|
||||
repo_id := cmd.flags.get_int('repo')?
|
||||
|
||||
if repo_id == 0 { list(conf)? } else { list_for_repo(conf, repo_id)? }
|
||||
}
|
||||
},
|
||||
cli.Command{
|
||||
name: 'info'
|
||||
required_args: 1
|
||||
usage: 'id'
|
||||
description: 'Show all info for a specific build log.'
|
||||
execute: fn (cmd cli.Command) ? {
|
||||
config_file := cmd.flags.get_string('config-file')?
|
||||
conf := env.load<Config>(config_file)?
|
||||
|
||||
id := cmd.args[0].int()
|
||||
info(conf, id)?
|
||||
}
|
||||
},
|
||||
cli.Command{
|
||||
name: 'content'
|
||||
required_args: 1
|
||||
usage: 'id'
|
||||
description: 'Output the content of a build log to stdout.'
|
||||
execute: fn (cmd cli.Command) ? {
|
||||
config_file := cmd.flags.get_string('config-file')?
|
||||
conf := env.load<Config>(config_file)?
|
||||
|
||||
id := cmd.args[0].int()
|
||||
content(conf, id)?
|
||||
}
|
||||
},
|
||||
]
|
||||
}
|
||||
}
|
||||
|
||||
// print_log_list prints a list of logs.
|
||||
fn print_log_list(logs []db.BuildLog) ? {
|
||||
data := logs.map([it.id.str(), it.repo_id.str(), it.start_time.str(),
|
||||
it.exit_code.str()])
|
||||
|
||||
println(console.pretty_table(['id', 'repo', 'start time', 'exit code'], data)?)
|
||||
}
|
||||
|
||||
// list prints a list of all build logs.
|
||||
fn list(conf Config) ? {
|
||||
c := client.new(conf.address, conf.api_key)
|
||||
logs := c.get_build_logs()?.data
|
||||
|
||||
print_log_list(logs)?
|
||||
}
|
||||
|
||||
// list prints a list of all build logs for a given repo.
|
||||
fn list_for_repo(conf Config, repo_id int) ? {
|
||||
c := client.new(conf.address, conf.api_key)
|
||||
logs := c.get_build_logs_for_repo(repo_id)?.data
|
||||
|
||||
print_log_list(logs)?
|
||||
}
|
||||
|
||||
// info print the detailed info for a given build log.
|
||||
fn info(conf Config, id int) ? {
|
||||
c := client.new(conf.address, conf.api_key)
|
||||
log := c.get_build_log(id)?.data
|
||||
|
||||
print(log)
|
||||
}
|
||||
|
||||
// content outputs the contents of the log file for a given build log to
|
||||
// stdout.
|
||||
fn content(conf Config, id int) ? {
|
||||
c := client.new(conf.address, conf.api_key)
|
||||
content := c.get_build_log_content(id)?
|
||||
|
||||
println(content)
|
||||
}
|
|
@ -23,10 +23,10 @@ pub fn cmd() cli.Command {
|
|||
name: 'cron'
|
||||
description: 'Start the cron service that periodically runs builds.'
|
||||
execute: fn (cmd cli.Command) ? {
|
||||
config_file := cmd.flags.get_string('config-file') ?
|
||||
conf := env.load<Config>(config_file) ?
|
||||
config_file := cmd.flags.get_string('config-file')?
|
||||
conf := env.load<Config>(config_file)?
|
||||
|
||||
cron(conf) ?
|
||||
cron(conf)?
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -27,7 +27,7 @@ pub fn cron(conf Config) ? {
|
|||
}
|
||||
|
||||
mut d := daemon.init_daemon(logger, conf.address, conf.api_key, conf.base_image, ce,
|
||||
conf.max_concurrent_builds, conf.api_update_frequency, conf.image_rebuild_frequency) ?
|
||||
conf.max_concurrent_builds, conf.api_update_frequency, conf.image_rebuild_frequency)?
|
||||
|
||||
d.run()
|
||||
}
|
||||
|
|
|
@ -3,6 +3,7 @@ module daemon
|
|||
import time
|
||||
import sync.stdatomic
|
||||
import build
|
||||
import os
|
||||
|
||||
const (
|
||||
build_empty = 0
|
||||
|
@ -77,13 +78,20 @@ fn (mut d Daemon) run_build(build_index int, sb ScheduledBuild) {
|
|||
// 0 means success, 1 means failure
|
||||
mut status := 0
|
||||
|
||||
build.build_repo(d.address, d.api_key, d.builder_images.last(), &sb.repo) or {
|
||||
res := build.build_repo(d.client.address, d.client.api_key, d.builder_images.last(),
|
||||
&sb.repo) or {
|
||||
d.ldebug('build_repo error: $err.msg()')
|
||||
status = 1
|
||||
|
||||
build.BuildResult{}
|
||||
}
|
||||
|
||||
if status == 0 {
|
||||
d.linfo('finished build: $sb.repo.url $sb.repo.branch')
|
||||
d.linfo('finished build: $sb.repo.url $sb.repo.branch; uploading logs...')
|
||||
|
||||
build_arch := os.uname().machine
|
||||
d.client.add_build_log(sb.repo.id, res.start_time, res.end_time, build_arch, res.exit_code,
|
||||
res.logs) or { d.lerror('Failed to upload logs for $sb.repo.url $sb.repo.arch') }
|
||||
} else {
|
||||
d.linfo('failed build: $sb.repo.url $sb.repo.branch')
|
||||
}
|
||||
|
|
|
@ -1,6 +1,5 @@
|
|||
module daemon
|
||||
|
||||
import git
|
||||
import time
|
||||
import log
|
||||
import datatypes { MinHeap }
|
||||
|
@ -10,6 +9,7 @@ import build
|
|||
import docker
|
||||
import db
|
||||
import os
|
||||
import client
|
||||
|
||||
const (
|
||||
// How many seconds to wait before retrying to update API if failed
|
||||
|
@ -31,8 +31,7 @@ fn (r1 ScheduledBuild) < (r2 ScheduledBuild) bool {
|
|||
|
||||
pub struct Daemon {
|
||||
mut:
|
||||
address string
|
||||
api_key string
|
||||
client client.Client
|
||||
base_image string
|
||||
builder_images []string
|
||||
global_schedule CronExpression
|
||||
|
@ -56,8 +55,7 @@ mut:
|
|||
// populates the build queue for the first time.
|
||||
pub fn init_daemon(logger log.Log, address string, api_key string, base_image string, global_schedule CronExpression, max_concurrent_builds int, api_update_frequency int, image_rebuild_frequency int) ?Daemon {
|
||||
mut d := Daemon{
|
||||
address: address
|
||||
api_key: api_key
|
||||
client: client.new(address, api_key)
|
||||
base_image: base_image
|
||||
global_schedule: global_schedule
|
||||
api_update_frequency: api_update_frequency
|
||||
|
@ -180,7 +178,7 @@ fn (mut d Daemon) schedule_build(repo db.GitRepo) {
|
|||
fn (mut d Daemon) renew_repos() {
|
||||
d.linfo('Renewing repos...')
|
||||
|
||||
mut new_repos := git.get_repos(d.address, d.api_key) or {
|
||||
mut new_repos := d.client.get_git_repos() or {
|
||||
d.lerror('Failed to renew repos. Retrying in ${daemon.api_update_retry_timeout}s...')
|
||||
d.api_update_timestamp = time.now().add_seconds(daemon.api_update_retry_timeout)
|
||||
|
||||
|
@ -255,14 +253,21 @@ fn (mut d Daemon) rebuild_base_image() bool {
|
|||
fn (mut d Daemon) clean_old_base_images() {
|
||||
mut i := 0
|
||||
|
||||
mut dd := docker.new_conn() or {
|
||||
d.lerror('Failed to connect to Docker socket.')
|
||||
return
|
||||
}
|
||||
|
||||
defer {
|
||||
dd.close() or {}
|
||||
}
|
||||
|
||||
for i < d.builder_images.len - 1 {
|
||||
// For each builder image, we try to remove it by calling the Docker
|
||||
// API. If the function returns an error or false, that means the image
|
||||
// wasn't deleted. Therefore, we move the index over. If the function
|
||||
// returns true, the array's length has decreased by one so we don't
|
||||
// move the index.
|
||||
if !docker.remove_image(d.builder_images[i]) or { false } {
|
||||
i += 1
|
||||
}
|
||||
dd.remove_image(d.builder_images[i]) or { i += 1 }
|
||||
}
|
||||
}
|
||||
|
|
|
@ -218,7 +218,7 @@ fn parse_part(s string, min int, max int) ?[]int {
|
|||
mut bitv := []bool{len: max - min + 1, init: false}
|
||||
|
||||
for range in s.split(',') {
|
||||
parse_range(range, min, max, mut bitv) ?
|
||||
parse_range(range, min, max, mut bitv)?
|
||||
}
|
||||
|
||||
return bitv_to_ints(bitv, min)
|
||||
|
|
|
@ -13,14 +13,14 @@ fn parse_range_error(s string, min int, max int) string {
|
|||
// =====parse_range=====
|
||||
fn test_range_star_range() ? {
|
||||
mut bitv := []bool{len: 6, init: false}
|
||||
parse_range('*', 0, 5, mut bitv) ?
|
||||
parse_range('*', 0, 5, mut bitv)?
|
||||
|
||||
assert bitv == [true, true, true, true, true, true]
|
||||
}
|
||||
|
||||
fn test_range_number() ? {
|
||||
mut bitv := []bool{len: 6, init: false}
|
||||
parse_range('4', 0, 5, mut bitv) ?
|
||||
parse_range('4', 0, 5, mut bitv)?
|
||||
|
||||
assert bitv_to_ints(bitv, 0) == [4]
|
||||
}
|
||||
|
@ -39,14 +39,14 @@ fn test_range_number_invalid() ? {
|
|||
|
||||
fn test_range_step_star_1() ? {
|
||||
mut bitv := []bool{len: 21, init: false}
|
||||
parse_range('*/4', 0, 20, mut bitv) ?
|
||||
parse_range('*/4', 0, 20, mut bitv)?
|
||||
|
||||
assert bitv_to_ints(bitv, 0) == [0, 4, 8, 12, 16, 20]
|
||||
}
|
||||
|
||||
fn test_range_step_star_2() ? {
|
||||
mut bitv := []bool{len: 8, init: false}
|
||||
parse_range('*/3', 1, 8, mut bitv) ?
|
||||
parse_range('*/3', 1, 8, mut bitv)?
|
||||
|
||||
assert bitv_to_ints(bitv, 1) == [1, 4, 7]
|
||||
}
|
||||
|
@ -61,7 +61,7 @@ fn test_range_step_zero() ? {
|
|||
|
||||
fn test_range_step_number() ? {
|
||||
mut bitv := []bool{len: 21, init: false}
|
||||
parse_range('5/4', 2, 22, mut bitv) ?
|
||||
parse_range('5/4', 2, 22, mut bitv)?
|
||||
|
||||
assert bitv_to_ints(bitv, 2) == [5, 9, 13, 17, 21]
|
||||
}
|
||||
|
@ -76,23 +76,23 @@ fn test_range_step_number_too_small() ? {
|
|||
|
||||
fn test_range_dash() ? {
|
||||
mut bitv := []bool{len: 10, init: false}
|
||||
parse_range('4-8', 0, 9, mut bitv) ?
|
||||
parse_range('4-8', 0, 9, mut bitv)?
|
||||
|
||||
assert bitv_to_ints(bitv, 0) == [4, 5, 6, 7, 8]
|
||||
}
|
||||
|
||||
fn test_range_dash_step() ? {
|
||||
mut bitv := []bool{len: 10, init: false}
|
||||
parse_range('4-8/2', 0, 9, mut bitv) ?
|
||||
parse_range('4-8/2', 0, 9, mut bitv)?
|
||||
|
||||
assert bitv_to_ints(bitv, 0) == [4, 6, 8]
|
||||
}
|
||||
|
||||
// =====parse_part=====
|
||||
fn test_part_single() ? {
|
||||
assert parse_part('*', 0, 5) ? == [0, 1, 2, 3, 4, 5]
|
||||
assert parse_part('*', 0, 5)? == [0, 1, 2, 3, 4, 5]
|
||||
}
|
||||
|
||||
fn test_part_multiple() ? {
|
||||
assert parse_part('*/2,2/3', 1, 8) ? == [1, 2, 3, 5, 7, 8]
|
||||
assert parse_part('*/2,2/3', 1, 8)? == [1, 2, 3, 5, 7, 8]
|
||||
}
|
||||
|
|
|
@ -3,11 +3,11 @@ module expression
|
|||
import time { parse }
|
||||
|
||||
fn util_test_time(exp string, t1_str string, t2_str string) ? {
|
||||
ce := parse_expression(exp) ?
|
||||
t1 := parse(t1_str) ?
|
||||
t2 := parse(t2_str) ?
|
||||
ce := parse_expression(exp)?
|
||||
t1 := parse(t1_str)?
|
||||
t2 := parse(t2_str)?
|
||||
|
||||
t3 := ce.next(t1) ?
|
||||
t3 := ce.next(t1)?
|
||||
|
||||
assert t2.year == t3.year
|
||||
assert t2.month == t3.month
|
||||
|
@ -18,17 +18,17 @@ fn util_test_time(exp string, t1_str string, t2_str string) ? {
|
|||
|
||||
fn test_next_simple() ? {
|
||||
// Very simple
|
||||
util_test_time('0 3', '2002-01-01 00:00:00', '2002-01-01 03:00:00') ?
|
||||
util_test_time('0 3', '2002-01-01 00:00:00', '2002-01-01 03:00:00')?
|
||||
|
||||
// Overlap to next day
|
||||
util_test_time('0 3', '2002-01-01 03:00:00', '2002-01-02 03:00:00') ?
|
||||
util_test_time('0 3', '2002-01-01 04:00:00', '2002-01-02 03:00:00') ?
|
||||
util_test_time('0 3', '2002-01-01 03:00:00', '2002-01-02 03:00:00')?
|
||||
util_test_time('0 3', '2002-01-01 04:00:00', '2002-01-02 03:00:00')?
|
||||
|
||||
util_test_time('0 3/4', '2002-01-01 04:00:00', '2002-01-01 07:00:00') ?
|
||||
util_test_time('0 3/4', '2002-01-01 04:00:00', '2002-01-01 07:00:00')?
|
||||
|
||||
// Overlap to next month
|
||||
util_test_time('0 3', '2002-11-31 04:00:00', '2002-12-01 03:00:00') ?
|
||||
util_test_time('0 3', '2002-11-31 04:00:00', '2002-12-01 03:00:00')?
|
||||
|
||||
// Overlap to next year
|
||||
util_test_time('0 3', '2002-12-31 04:00:00', '2003-01-01 03:00:00') ?
|
||||
util_test_time('0 3', '2002-12-31 04:00:00', '2003-01-01 03:00:00')?
|
||||
}
|
||||
|
|
|
@ -8,10 +8,11 @@ struct VieterDb {
|
|||
|
||||
// init initializes a database & adds the correct tables.
|
||||
pub fn init(db_path string) ?VieterDb {
|
||||
conn := sqlite.connect(db_path) ?
|
||||
conn := sqlite.connect(db_path)?
|
||||
|
||||
sql conn {
|
||||
create table GitRepo
|
||||
create table BuildLog
|
||||
}
|
||||
|
||||
return VieterDb{
|
||||
|
|
10
src/db/git.v
10
src/db/git.v
|
@ -94,7 +94,7 @@ pub fn (db &VieterDb) get_git_repo(repo_id int) ?GitRepo {
|
|||
// If a select statement fails, it returns a zeroed object. By
|
||||
// checking one of the required fields, we can see whether the query
|
||||
// returned a result or not.
|
||||
if res.url == '' {
|
||||
if res.id == 0 {
|
||||
return none
|
||||
}
|
||||
|
||||
|
@ -152,3 +152,11 @@ pub fn (db &VieterDb) update_git_repo_archs(repo_id int, archs []GitRepoArch) {
|
|||
}
|
||||
}
|
||||
}
|
||||
|
||||
// git_repo_exists is a utility function that checks whether a repo with the
|
||||
// given id exists.
|
||||
pub fn (db &VieterDb) git_repo_exists(repo_id int) bool {
|
||||
db.get_git_repo(repo_id) or { return false }
|
||||
|
||||
return true
|
||||
}
|
||||
|
|
|
@ -0,0 +1,74 @@
|
|||
module db
|
||||
|
||||
import time
|
||||
|
||||
pub struct BuildLog {
|
||||
pub:
|
||||
id int [primary; sql: serial]
|
||||
repo_id int [nonull]
|
||||
start_time time.Time [nonull]
|
||||
end_time time.Time [nonull]
|
||||
arch string [nonull]
|
||||
exit_code int [nonull]
|
||||
}
|
||||
|
||||
// str returns a string representation.
|
||||
pub fn (bl &BuildLog) str() string {
|
||||
mut parts := [
|
||||
'id: $bl.id',
|
||||
'repo id: $bl.repo_id',
|
||||
'start time: $bl.start_time',
|
||||
'end time: $bl.end_time',
|
||||
'arch: $bl.arch',
|
||||
'exit code: $bl.exit_code',
|
||||
]
|
||||
str := parts.join('\n')
|
||||
|
||||
return str
|
||||
}
|
||||
|
||||
// get_build_logs returns all BuildLog's in the database.
|
||||
pub fn (db &VieterDb) get_build_logs() []BuildLog {
|
||||
res := sql db.conn {
|
||||
select from BuildLog order by id
|
||||
}
|
||||
|
||||
return res
|
||||
}
|
||||
|
||||
// get_build_logs_for_repo returns all BuildLog's in the database for a given
|
||||
// repo.
|
||||
pub fn (db &VieterDb) get_build_logs_for_repo(repo_id int) []BuildLog {
|
||||
res := sql db.conn {
|
||||
select from BuildLog where repo_id == repo_id order by id
|
||||
}
|
||||
|
||||
return res
|
||||
}
|
||||
|
||||
// get_build_log tries to return a specific BuildLog.
|
||||
pub fn (db &VieterDb) get_build_log(id int) ?BuildLog {
|
||||
res := sql db.conn {
|
||||
select from BuildLog where id == id
|
||||
}
|
||||
|
||||
if res.id == 0 {
|
||||
return none
|
||||
}
|
||||
|
||||
return res
|
||||
}
|
||||
|
||||
// add_build_log inserts the given BuildLog into the database.
|
||||
pub fn (db &VieterDb) add_build_log(log BuildLog) {
|
||||
sql db.conn {
|
||||
insert log into BuildLog
|
||||
}
|
||||
}
|
||||
|
||||
// delete_build_log delete the BuildLog with the given ID from the database.
|
||||
pub fn (db &VieterDb) delete_build_log(id int) {
|
||||
sql db.conn {
|
||||
delete from BuildLog where id == id
|
||||
}
|
||||
}
|
|
@ -0,0 +1,3 @@
|
|||
This module implements part of the Docker Engine API v1.41
|
||||
([documentation](https://docs.docker.com/engine/api/v1.41/)) using socket-based
|
||||
HTTP communication.
|
|
@ -2,17 +2,11 @@ module docker
|
|||
|
||||
import json
|
||||
import net.urllib
|
||||
import time
|
||||
import net.http { Method }
|
||||
|
||||
struct Container {
|
||||
id string [json: Id]
|
||||
names []string [json: Names]
|
||||
}
|
||||
|
||||
// containers returns a list of all currently running containers
|
||||
pub fn containers() ?[]Container {
|
||||
res := request('GET', urllib.parse('/v1.41/containers/json') ?) ?
|
||||
|
||||
return json.decode([]Container, res.text) or {}
|
||||
struct DockerError {
|
||||
message string
|
||||
}
|
||||
|
||||
pub struct NewContainer {
|
||||
|
@ -25,54 +19,104 @@ pub struct NewContainer {
|
|||
}
|
||||
|
||||
struct CreatedContainer {
|
||||
id string [json: Id]
|
||||
pub:
|
||||
id string [json: Id]
|
||||
warnings []string [json: Warnings]
|
||||
}
|
||||
|
||||
// create_container creates a container defined by the given configuration. If
|
||||
// successful, it returns the ID of the newly created container.
|
||||
pub fn create_container(c &NewContainer) ?string {
|
||||
res := request_with_json('POST', urllib.parse('/v1.41/containers/create') ?, c) ?
|
||||
// create_container creates a new container with the given config.
|
||||
pub fn (mut d DockerConn) create_container(c NewContainer) ?CreatedContainer {
|
||||
d.send_request_with_json(Method.post, urllib.parse('/v1.41/containers/create')?, c)?
|
||||
head, res := d.read_response()?
|
||||
|
||||
if res.status_code != 201 {
|
||||
return error('Failed to create container.')
|
||||
if head.status_code != 201 {
|
||||
data := json.decode(DockerError, res)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
|
||||
return json.decode(CreatedContainer, res.text) ?.id
|
||||
data := json.decode(CreatedContainer, res)?
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// start_container starts a container with a given ID. It returns whether the
|
||||
// container was started or not.
|
||||
pub fn start_container(id string) ?bool {
|
||||
res := request('POST', urllib.parse('/v1.41/containers/$id/start') ?) ?
|
||||
// start_container starts the container with the given id.
|
||||
pub fn (mut d DockerConn) start_container(id string) ? {
|
||||
d.send_request(Method.post, urllib.parse('/v1.41/containers/$id/start')?)?
|
||||
head, body := d.read_response()?
|
||||
|
||||
return res.status_code == 204
|
||||
if head.status_code != 204 {
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
}
|
||||
|
||||
struct ContainerInspect {
|
||||
pub:
|
||||
pub mut:
|
||||
state ContainerState [json: State]
|
||||
}
|
||||
|
||||
struct ContainerState {
|
||||
pub:
|
||||
running bool [json: Running]
|
||||
running bool [json: Running]
|
||||
status string [json: Status]
|
||||
exit_code int [json: ExitCode]
|
||||
// These use a rather specific format so they have to be parsed later
|
||||
start_time_str string [json: StartedAt]
|
||||
end_time_str string [json: FinishedAt]
|
||||
pub mut:
|
||||
start_time time.Time [skip]
|
||||
end_time time.Time [skip]
|
||||
}
|
||||
|
||||
// inspect_container returns the result of inspecting a container with a given
|
||||
// ID.
|
||||
pub fn inspect_container(id string) ?ContainerInspect {
|
||||
res := request('GET', urllib.parse('/v1.41/containers/$id/json') ?) ?
|
||||
// inspect_container returns detailed information for a given container.
|
||||
pub fn (mut d DockerConn) inspect_container(id string) ?ContainerInspect {
|
||||
d.send_request(Method.get, urllib.parse('/v1.41/containers/$id/json')?)?
|
||||
head, body := d.read_response()?
|
||||
|
||||
if res.status_code != 200 {
|
||||
return error('Failed to inspect container.')
|
||||
if head.status_code != 200 {
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
|
||||
return json.decode(ContainerInspect, res.text) or {}
|
||||
mut data := json.decode(ContainerInspect, body)?
|
||||
|
||||
data.state.start_time = time.parse_rfc3339(data.state.start_time_str)?
|
||||
|
||||
if data.state.status == 'exited' {
|
||||
data.state.end_time = time.parse_rfc3339(data.state.end_time_str)?
|
||||
}
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// remove_container removes a container with a given ID.
|
||||
pub fn remove_container(id string) ?bool {
|
||||
res := request('DELETE', urllib.parse('/v1.41/containers/$id') ?) ?
|
||||
// remove_container removes the container with the given id.
|
||||
pub fn (mut d DockerConn) remove_container(id string) ? {
|
||||
d.send_request(Method.delete, urllib.parse('/v1.41/containers/$id')?)?
|
||||
head, body := d.read_response()?
|
||||
|
||||
return res.status_code == 204
|
||||
if head.status_code != 204 {
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
}
|
||||
|
||||
// get_container_logs returns a reader object allowing access to the
|
||||
// container's logs.
|
||||
pub fn (mut d DockerConn) get_container_logs(id string) ?&StreamFormatReader {
|
||||
d.send_request(Method.get, urllib.parse('/v1.41/containers/$id/logs?stdout=true&stderr=true')?)?
|
||||
head := d.read_response_head()?
|
||||
|
||||
if head.status_code != 200 {
|
||||
content_length := head.header.get(http.CommonHeader.content_length)?.int()
|
||||
body := d.read_response_body(content_length)?
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
|
||||
return d.get_stream_format_reader()
|
||||
}
|
||||
|
|
|
@ -1,97 +1,137 @@
|
|||
module docker
|
||||
|
||||
import net.unix
|
||||
import net.urllib
|
||||
import io
|
||||
import net.http
|
||||
import strings
|
||||
import net.urllib
|
||||
import json
|
||||
import util
|
||||
|
||||
const socket = '/var/run/docker.sock'
|
||||
const (
|
||||
socket = '/var/run/docker.sock'
|
||||
buf_len = 10 * 1024
|
||||
http_separator = [u8(`\r`), `\n`, `\r`, `\n`]
|
||||
http_chunk_separator = [u8(`\r`), `\n`]
|
||||
)
|
||||
|
||||
const buf_len = 1024
|
||||
pub struct DockerConn {
|
||||
mut:
|
||||
socket &unix.StreamConn
|
||||
reader &io.BufferedReader
|
||||
}
|
||||
|
||||
// send writes a request to the Docker socket, waits for a response & returns
|
||||
// it.
|
||||
fn send(req &string) ?http.Response {
|
||||
// Open a connection to the socket
|
||||
mut s := unix.connect_stream(docker.socket) or {
|
||||
return error('Failed to connect to socket ${docker.socket}.')
|
||||
// new_conn creates a new connection to the Docker daemon.
|
||||
pub fn new_conn() ?&DockerConn {
|
||||
s := unix.connect_stream(docker.socket)?
|
||||
|
||||
d := &DockerConn{
|
||||
socket: s
|
||||
reader: io.new_buffered_reader(reader: s)
|
||||
}
|
||||
|
||||
defer {
|
||||
// This or is required because otherwise, the V compiler segfaults for
|
||||
// some reason
|
||||
// https://github.com/vlang/v/issues/13534
|
||||
s.close() or {}
|
||||
}
|
||||
return d
|
||||
}
|
||||
|
||||
// Write the request to the socket
|
||||
s.write_string(req) or { return error('Failed to write request to socket ${docker.socket}.') }
|
||||
// close closes the underlying socket connection.
|
||||
pub fn (mut d DockerConn) close() ? {
|
||||
d.socket.close()?
|
||||
}
|
||||
|
||||
s.wait_for_write() ?
|
||||
// send_request sends an HTTP request without body.
|
||||
pub fn (mut d DockerConn) send_request(method http.Method, url urllib.URL) ? {
|
||||
req := '$method $url.request_uri() HTTP/1.1\nHost: localhost\n\n'
|
||||
|
||||
mut c := 0
|
||||
mut buf := []u8{len: docker.buf_len}
|
||||
d.socket.write_string(req)?
|
||||
|
||||
// When starting a new request, the reader needs to be reset.
|
||||
d.reader = io.new_buffered_reader(reader: d.socket)
|
||||
}
|
||||
|
||||
// send_request_with_body sends an HTTP request with the given body.
|
||||
pub fn (mut d DockerConn) send_request_with_body(method http.Method, url urllib.URL, content_type string, body string) ? {
|
||||
req := '$method $url.request_uri() HTTP/1.1\nHost: localhost\nContent-Type: $content_type\nContent-Length: $body.len\n\n$body\n\n'
|
||||
|
||||
d.socket.write_string(req)?
|
||||
|
||||
// When starting a new request, the reader needs to be reset.
|
||||
d.reader = io.new_buffered_reader(reader: d.socket)
|
||||
}
|
||||
|
||||
// send_request_with_json<T> is a convenience wrapper around
|
||||
// send_request_with_body that encodes the input as JSON.
|
||||
pub fn (mut d DockerConn) send_request_with_json<T>(method http.Method, url urllib.URL, data &T) ? {
|
||||
body := json.encode(data)
|
||||
|
||||
return d.send_request_with_body(method, url, 'application/json', body)
|
||||
}
|
||||
|
||||
// read_response_head consumes the socket's contents until it encounters
|
||||
// '\r\n\r\n', after which it parses the response as an HTTP response.
|
||||
// Importantly, this function never consumes the reader past the HTTP
|
||||
// separator, so the body can be read fully later on.
|
||||
pub fn (mut d DockerConn) read_response_head() ?http.Response {
|
||||
mut res := []u8{}
|
||||
|
||||
for {
|
||||
c = s.read(mut buf) or { return error('Failed to read data from socket ${docker.socket}.') }
|
||||
res << buf[..c]
|
||||
util.read_until_separator(mut d.reader, mut res, docker.http_separator)?
|
||||
|
||||
if c < docker.buf_len {
|
||||
break
|
||||
}
|
||||
}
|
||||
|
||||
// After reading the first part of the response, we parse it into an HTTP
|
||||
// response. If it isn't chunked, we return early with the data.
|
||||
parsed := http.parse_response(res.bytestr()) or {
|
||||
return error('Failed to parse HTTP response from socket ${docker.socket}.')
|
||||
}
|
||||
|
||||
if parsed.header.get(http.CommonHeader.transfer_encoding) or { '' } != 'chunked' {
|
||||
return parsed
|
||||
}
|
||||
|
||||
// We loop until we've encountered the end of the chunked response
|
||||
// A chunked HTTP response always ends with '0\r\n\r\n'.
|
||||
for res.len < 5 || res#[-5..] != [u8(`0`), `\r`, `\n`, `\r`, `\n`] {
|
||||
// Wait for the server to respond
|
||||
s.wait_for_write() ?
|
||||
|
||||
for {
|
||||
c = s.read(mut buf) or {
|
||||
return error('Failed to read data from socket ${docker.socket}.')
|
||||
}
|
||||
res << buf[..c]
|
||||
|
||||
if c < docker.buf_len {
|
||||
break
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Decode chunked response
|
||||
return http.parse_response(res.bytestr())
|
||||
}
|
||||
|
||||
// request_with_body sends a request to the Docker socket with the given body.
|
||||
fn request_with_body(method string, url urllib.URL, content_type string, body string) ?http.Response {
|
||||
req := '$method $url.request_uri() HTTP/1.1\nHost: localhost\nContent-Type: $content_type\nContent-Length: $body.len\n\n$body\n\n'
|
||||
// read_response_body reads `length` bytes from the stream. It can be used when
|
||||
// the response encoding isn't chunked to fully read it.
|
||||
pub fn (mut d DockerConn) read_response_body(length int) ?string {
|
||||
if length == 0 {
|
||||
return ''
|
||||
}
|
||||
|
||||
return send(req)
|
||||
mut buf := []u8{len: docker.buf_len}
|
||||
mut c := 0
|
||||
mut builder := strings.new_builder(docker.buf_len)
|
||||
|
||||
for builder.len < length {
|
||||
c = d.reader.read(mut buf) or { break }
|
||||
|
||||
builder.write(buf[..c])?
|
||||
}
|
||||
|
||||
return builder.str()
|
||||
}
|
||||
|
||||
// request sends a request to the Docker socket with an empty body.
|
||||
fn request(method string, url urllib.URL) ?http.Response {
|
||||
req := '$method $url.request_uri() HTTP/1.1\nHost: localhost\n\n'
|
||||
// read_response is a convenience function which always consumes the entire
|
||||
// response & returns it. It should only be used when we're certain that the
|
||||
// result isn't too large.
|
||||
pub fn (mut d DockerConn) read_response() ?(http.Response, string) {
|
||||
head := d.read_response_head()?
|
||||
|
||||
return send(req)
|
||||
if head.header.get(http.CommonHeader.transfer_encoding) or { '' } == 'chunked' {
|
||||
mut builder := strings.new_builder(1024)
|
||||
mut body := d.get_chunked_response_reader()
|
||||
|
||||
util.reader_to_writer(mut body, mut builder)?
|
||||
|
||||
return head, builder.str()
|
||||
}
|
||||
|
||||
content_length := head.header.get(http.CommonHeader.content_length)?.int()
|
||||
res := d.read_response_body(content_length)?
|
||||
|
||||
return head, res
|
||||
}
|
||||
|
||||
// request_with_json<T> sends a request to the Docker socket with a given JSON
|
||||
// payload
|
||||
pub fn request_with_json<T>(method string, url urllib.URL, data &T) ?http.Response {
|
||||
body := json.encode(data)
|
||||
// get_chunked_response_reader returns a ChunkedResponseReader using the socket
|
||||
// as reader.
|
||||
pub fn (mut d DockerConn) get_chunked_response_reader() &ChunkedResponseReader {
|
||||
r := new_chunked_response_reader(d.reader)
|
||||
|
||||
return request_with_body(method, url, 'application/json', body)
|
||||
return r
|
||||
}
|
||||
|
||||
// get_stream_format_reader returns a StreamFormatReader using the socket as
|
||||
// reader.
|
||||
pub fn (mut d DockerConn) get_stream_format_reader() &StreamFormatReader {
|
||||
r := new_chunked_response_reader(d.reader)
|
||||
r2 := new_stream_format_reader(r)
|
||||
|
||||
return r2
|
||||
}
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
module docker
|
||||
|
||||
import net.http
|
||||
import net.http { Method }
|
||||
import net.urllib
|
||||
import json
|
||||
|
||||
|
@ -9,26 +9,53 @@ pub:
|
|||
id string [json: Id]
|
||||
}
|
||||
|
||||
// pull_image pulls tries to pull the image for the given image & tag
|
||||
pub fn pull_image(image string, tag string) ?http.Response {
|
||||
return request('POST', urllib.parse('/v1.41/images/create?fromImage=$image&tag=$tag') ?)
|
||||
}
|
||||
// pull_image pulls the given image:tag.
|
||||
pub fn (mut d DockerConn) pull_image(image string, tag string) ? {
|
||||
d.send_request(Method.post, urllib.parse('/v1.41/images/create?fromImage=$image&tag=$tag')?)?
|
||||
head := d.read_response_head()?
|
||||
|
||||
// create_image_from_container creates a new image from a container with the
|
||||
// given repo & tag, given the container's ID.
|
||||
pub fn create_image_from_container(id string, repo string, tag string) ?Image {
|
||||
res := request('POST', urllib.parse('/v1.41/commit?container=$id&repo=$repo&tag=$tag') ?) ?
|
||||
if head.status_code != 200 {
|
||||
content_length := head.header.get(http.CommonHeader.content_length)?.int()
|
||||
body := d.read_response_body(content_length)?
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
if res.status_code != 201 {
|
||||
return error('Failed to create image from container.')
|
||||
return error(data.message)
|
||||
}
|
||||
|
||||
return json.decode(Image, res.text) or {}
|
||||
// Keep reading the body until the pull has completed
|
||||
mut body := d.get_chunked_response_reader()
|
||||
|
||||
mut buf := []u8{len: 1024}
|
||||
|
||||
for {
|
||||
body.read(mut buf) or { break }
|
||||
}
|
||||
}
|
||||
|
||||
// remove_image removes the image with the given ID.
|
||||
pub fn remove_image(id string) ?bool {
|
||||
res := request('DELETE', urllib.parse('/v1.41/images/$id') ?) ?
|
||||
// create_image_from_container creates a new image from a container.
|
||||
pub fn (mut d DockerConn) create_image_from_container(id string, repo string, tag string) ?Image {
|
||||
d.send_request(Method.post, urllib.parse('/v1.41/commit?container=$id&repo=$repo&tag=$tag')?)?
|
||||
head, body := d.read_response()?
|
||||
|
||||
return res.status_code == 200
|
||||
if head.status_code != 201 {
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
|
||||
data := json.decode(Image, body)?
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// remove_image removes the image with the given id.
|
||||
pub fn (mut d DockerConn) remove_image(id string) ? {
|
||||
d.send_request(Method.delete, urllib.parse('/v1.41/images/$id')?)?
|
||||
head, body := d.read_response()?
|
||||
|
||||
if head.status_code != 200 {
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
}
|
||||
|
|
|
@ -0,0 +1,135 @@
|
|||
module docker
|
||||
|
||||
import io
|
||||
import util
|
||||
import encoding.binary
|
||||
import encoding.hex
|
||||
|
||||
// ChunkedResponseReader parses an underlying HTTP chunked response, exposing
|
||||
// it as if it was a continuous stream of data.
|
||||
struct ChunkedResponseReader {
|
||||
mut:
|
||||
reader io.BufferedReader
|
||||
bytes_left_in_chunk u64
|
||||
started bool
|
||||
}
|
||||
|
||||
// new_chunked_response_reader creates a new ChunkedResponseReader on the heap
|
||||
// with the provided reader.
|
||||
pub fn new_chunked_response_reader(reader io.BufferedReader) &ChunkedResponseReader {
|
||||
r := &ChunkedResponseReader{
|
||||
reader: reader
|
||||
}
|
||||
|
||||
return r
|
||||
}
|
||||
|
||||
// read satisfies the io.Reader interface.
|
||||
pub fn (mut r ChunkedResponseReader) read(mut buf []u8) ?int {
|
||||
if r.bytes_left_in_chunk == 0 {
|
||||
// An io.BufferedReader always returns none if its stream has
|
||||
// ended.
|
||||
r.bytes_left_in_chunk = r.read_chunk_size()?
|
||||
}
|
||||
|
||||
mut c := 0
|
||||
|
||||
// Make sure we don't read more than we can safely read. This is to avoid
|
||||
// the underlying reader from becoming out of sync with our parsing:
|
||||
if buf.len > r.bytes_left_in_chunk {
|
||||
c = r.reader.read(mut buf[..r.bytes_left_in_chunk])?
|
||||
} else {
|
||||
c = r.reader.read(mut buf)?
|
||||
}
|
||||
|
||||
r.bytes_left_in_chunk -= u64(c)
|
||||
|
||||
return c
|
||||
}
|
||||
|
||||
// read_chunk_size advances the reader & reads the size of the next HTTP chunk.
|
||||
// This function should only be called if the previous chunk has been
|
||||
// completely consumed.
|
||||
fn (mut r ChunkedResponseReader) read_chunk_size() ?u64 {
|
||||
if r.started {
|
||||
mut buf := []u8{len: 2}
|
||||
|
||||
// Each chunk ends with a `\r\n` which we want to skip first
|
||||
r.reader.read(mut buf)?
|
||||
}
|
||||
|
||||
r.started = true
|
||||
|
||||
mut res := []u8{}
|
||||
util.read_until_separator(mut r.reader, mut res, http_chunk_separator)?
|
||||
|
||||
// The length of the next chunk is provided as a hexadecimal
|
||||
mut num_data := hex.decode(res#[..-2].bytestr())?
|
||||
|
||||
for num_data.len < 8 {
|
||||
num_data.insert(0, 0)
|
||||
}
|
||||
|
||||
num := binary.big_endian_u64(num_data)
|
||||
|
||||
// This only occurs for the very last chunk, which always reports a size of
|
||||
// 0.
|
||||
if num == 0 {
|
||||
return none
|
||||
}
|
||||
|
||||
return num
|
||||
}
|
||||
|
||||
// StreamFormatReader parses an underlying stream of Docker logs, removing the
|
||||
// header bytes.
|
||||
struct StreamFormatReader {
|
||||
mut:
|
||||
reader ChunkedResponseReader
|
||||
bytes_left_in_chunk u32
|
||||
}
|
||||
|
||||
// new_stream_format_reader creates a new StreamFormatReader using the given
|
||||
// reader.
|
||||
pub fn new_stream_format_reader(reader ChunkedResponseReader) &StreamFormatReader {
|
||||
r := &StreamFormatReader{
|
||||
reader: reader
|
||||
}
|
||||
|
||||
return r
|
||||
}
|
||||
|
||||
// read satisfies the io.Reader interface.
|
||||
pub fn (mut r StreamFormatReader) read(mut buf []u8) ?int {
|
||||
if r.bytes_left_in_chunk == 0 {
|
||||
r.bytes_left_in_chunk = r.read_chunk_size()?
|
||||
}
|
||||
|
||||
mut c := 0
|
||||
|
||||
if buf.len > r.bytes_left_in_chunk {
|
||||
c = r.reader.read(mut buf[..r.bytes_left_in_chunk])?
|
||||
} else {
|
||||
c = r.reader.read(mut buf)?
|
||||
}
|
||||
|
||||
r.bytes_left_in_chunk -= u32(c)
|
||||
|
||||
return c
|
||||
}
|
||||
|
||||
// read_chunk_size advances the reader & reads the header bytes for the length
|
||||
// of the next chunk.
|
||||
fn (mut r StreamFormatReader) read_chunk_size() ?u32 {
|
||||
mut buf := []u8{len: 8}
|
||||
|
||||
r.reader.read(mut buf)?
|
||||
|
||||
num := binary.big_endian_u32(buf[4..])
|
||||
|
||||
if num == 0 {
|
||||
return none
|
||||
}
|
||||
|
||||
return num
|
||||
}
|
|
@ -0,0 +1,7 @@
|
|||
This module provides a framework for parsing a configuration, defined as a
|
||||
struct, from both a TOML configuration file & environment variables. Some
|
||||
notable features are:
|
||||
|
||||
* Overwrite values in config file using environment variables
|
||||
* Allow default values in config struct
|
||||
* Read environment variable value from file
|
|
@ -3,13 +3,19 @@ module env
|
|||
import os
|
||||
import toml
|
||||
|
||||
// The prefix that every environment variable should have
|
||||
const prefix = 'VIETER_'
|
||||
|
||||
// The suffix an environment variable in order for it to be loaded from a file
|
||||
// instead
|
||||
const file_suffix = '_FILE'
|
||||
const (
|
||||
// The prefix that every environment variable should have
|
||||
prefix = 'VIETER_'
|
||||
// The suffix an environment variable in order for it to be loaded from a file
|
||||
// instead
|
||||
file_suffix = '_FILE'
|
||||
)
|
||||
|
||||
// get_env_var tries to read the contents of the given environment variable. It
|
||||
// looks for either `${env.prefix}${field_name.to_upper()}` or
|
||||
// `${env.prefix}${field_name.to_upper()}${env.file_suffix}`, returning the
|
||||
// contents of the file instead if the latter. If both or neither exist, the
|
||||
// function returns an error.
|
||||
fn get_env_var(field_name string) ?string {
|
||||
env_var_name := '$env.prefix$field_name.to_upper()'
|
||||
env_file_name := '$env.prefix$field_name.to_upper()$env.file_suffix'
|
||||
|
@ -50,7 +56,7 @@ pub fn load<T>(path string) ?T {
|
|||
if os.exists(path) {
|
||||
// We don't use reflect here because reflect also sets any fields not
|
||||
// in the toml back to their zero value, which we don't want
|
||||
doc := toml.parse_file(path) ?
|
||||
doc := toml.parse_file(path)?
|
||||
|
||||
$for field in T.fields {
|
||||
s := doc.value(field.name)
|
||||
|
@ -66,7 +72,7 @@ pub fn load<T>(path string) ?T {
|
|||
}
|
||||
|
||||
$for field in T.fields {
|
||||
env_value := get_env_var(field.name) ?
|
||||
env_value := get_env_var(field.name)?
|
||||
|
||||
// The value of an env var will always take precedence over the toml
|
||||
// file.
|
||||
|
|
|
@ -1,77 +0,0 @@
|
|||
module git
|
||||
|
||||
import json
|
||||
import response { Response }
|
||||
import net.http
|
||||
import db
|
||||
|
||||
// send_request<T> is a convenience method for sending requests to the repos
|
||||
// API. It mostly does string manipulation to create a query string containing
|
||||
// the provided params.
|
||||
fn send_request<T>(method http.Method, address string, url string, api_key string, params map[string]string) ?Response<T> {
|
||||
mut full_url := '$address$url'
|
||||
|
||||
if params.len > 0 {
|
||||
params_str := params.keys().map('$it=${params[it]}').join('&')
|
||||
|
||||
full_url = '$full_url?$params_str'
|
||||
}
|
||||
|
||||
mut req := http.new_request(method, full_url, '') ?
|
||||
req.add_custom_header('X-API-Key', api_key) ?
|
||||
|
||||
res := req.do() ?
|
||||
data := json.decode(Response<T>, res.text) ?
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// get_repos returns the current list of repos.
|
||||
pub fn get_repos(address string, api_key string) ?[]db.GitRepo {
|
||||
data := send_request<[]db.GitRepo>(http.Method.get, address, '/api/repos', api_key,
|
||||
{}) ?
|
||||
|
||||
return data.data
|
||||
}
|
||||
|
||||
// get_repo returns the repo for a specific ID.
|
||||
pub fn get_repo(address string, api_key string, id int) ?db.GitRepo {
|
||||
data := send_request<db.GitRepo>(http.Method.get, address, '/api/repos/$id', api_key,
|
||||
{}) ?
|
||||
|
||||
return data.data
|
||||
}
|
||||
|
||||
// add_repo adds a new repo to the server.
|
||||
pub fn add_repo(address string, api_key string, url string, branch string, repo string, arch []string) ?Response<string> {
|
||||
mut params := {
|
||||
'url': url
|
||||
'branch': branch
|
||||
'repo': repo
|
||||
}
|
||||
|
||||
if arch.len > 0 {
|
||||
params['arch'] = arch.join(',')
|
||||
}
|
||||
|
||||
data := send_request<string>(http.Method.post, address, '/api/repos', api_key, params) ?
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// remove_repo removes the repo with the given ID from the server.
|
||||
pub fn remove_repo(address string, api_key string, id int) ?Response<string> {
|
||||
data := send_request<string>(http.Method.delete, address, '/api/repos/$id', api_key,
|
||||
{}) ?
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// patch_repo sends a PATCH request to the given repo with the params as
|
||||
// payload.
|
||||
pub fn patch_repo(address string, api_key string, id int, params map[string]string) ?Response<string> {
|
||||
data := send_request<string>(http.Method.patch, address, '/api/repos/$id', api_key,
|
||||
params) ?
|
||||
|
||||
return data
|
||||
}
|
10
src/main.v
10
src/main.v
|
@ -3,15 +3,15 @@ module main
|
|||
import os
|
||||
import server
|
||||
import cli
|
||||
import build
|
||||
import git
|
||||
import console.git
|
||||
import console.logs
|
||||
import cron
|
||||
|
||||
fn main() {
|
||||
mut app := cli.Command{
|
||||
name: 'vieter'
|
||||
description: 'Vieter is a lightweight implementation of an Arch repository server.'
|
||||
version: '0.3.0-alpha.1'
|
||||
version: '0.3.0-alpha.2'
|
||||
flags: [
|
||||
cli.Flag{
|
||||
flag: cli.FlagType.string
|
||||
|
@ -24,12 +24,12 @@ fn main() {
|
|||
]
|
||||
commands: [
|
||||
server.cmd(),
|
||||
build.cmd(),
|
||||
git.cmd(),
|
||||
cron.cmd(),
|
||||
logs.cmd(),
|
||||
]
|
||||
}
|
||||
|
||||
app.setup()
|
||||
app.parse(os.args)
|
||||
return
|
||||
}
|
||||
|
|
|
@ -159,7 +159,7 @@ pub fn read_pkg_archive(pkg_path string) ?Pkg {
|
|||
|
||||
pkg_text := unsafe { buf.vstring_with_len(size).clone() }
|
||||
|
||||
pkg_info = parse_pkg_info_string(pkg_text) ?
|
||||
pkg_info = parse_pkg_info_string(pkg_text)?
|
||||
} else {
|
||||
C.archive_read_data_skip(a)
|
||||
}
|
||||
|
|
|
@ -53,22 +53,22 @@ pub fn (r &RepoGroupManager) add_pkg_from_path(repo string, pkg_path string) ?Re
|
|||
return error('Failed to read package file: $err.msg()')
|
||||
}
|
||||
|
||||
added := r.add_pkg_in_repo(repo, pkg) ?
|
||||
added := r.add_pkg_in_repo(repo, pkg)?
|
||||
|
||||
// If the add was successful, we move the file to the packages directory
|
||||
for arch in added {
|
||||
repo_pkg_path := os.real_path(os.join_path(r.pkg_dir, repo, arch))
|
||||
dest_path := os.join_path_single(repo_pkg_path, pkg.filename())
|
||||
|
||||
os.mkdir_all(repo_pkg_path) ?
|
||||
os.mkdir_all(repo_pkg_path)?
|
||||
|
||||
// We create hard links so that "any" arch packages aren't stored
|
||||
// multiple times
|
||||
os.link(pkg_path, dest_path) ?
|
||||
os.link(pkg_path, dest_path)?
|
||||
}
|
||||
|
||||
// After linking, we can remove the original file
|
||||
os.rm(pkg_path) ?
|
||||
os.rm(pkg_path)?
|
||||
|
||||
return RepoAddResult{
|
||||
added: added.len > 0
|
||||
|
@ -87,7 +87,7 @@ fn (r &RepoGroupManager) add_pkg_in_repo(repo string, pkg &package.Pkg) ?[]strin
|
|||
// A package not of arch 'any' can be handled easily by adding it to the
|
||||
// respective repo
|
||||
if pkg.info.arch != 'any' {
|
||||
if r.add_pkg_in_arch_repo(repo, pkg.info.arch, pkg) ? {
|
||||
if r.add_pkg_in_arch_repo(repo, pkg.info.arch, pkg)? {
|
||||
return [pkg.info.arch]
|
||||
} else {
|
||||
return []
|
||||
|
@ -104,7 +104,7 @@ fn (r &RepoGroupManager) add_pkg_in_repo(repo string, pkg &package.Pkg) ?[]strin
|
|||
// If this is the first package that's added to the repo, the directory
|
||||
// won't exist yet
|
||||
if os.exists(repo_dir) {
|
||||
arch_repos = os.ls(repo_dir) ?
|
||||
arch_repos = os.ls(repo_dir)?
|
||||
}
|
||||
|
||||
// The default_arch should always be updated when a package with arch 'any'
|
||||
|
@ -118,7 +118,7 @@ fn (r &RepoGroupManager) add_pkg_in_repo(repo string, pkg &package.Pkg) ?[]strin
|
|||
// We add the package to each repository. If any of the repositories
|
||||
// return true, the result of the function is also true.
|
||||
for arch in arch_repos {
|
||||
if r.add_pkg_in_arch_repo(repo, arch, pkg) ? {
|
||||
if r.add_pkg_in_arch_repo(repo, arch, pkg)? {
|
||||
added << arch
|
||||
}
|
||||
}
|
||||
|
@ -135,22 +135,22 @@ fn (r &RepoGroupManager) add_pkg_in_arch_repo(repo string, arch string, pkg &pac
|
|||
pkg_dir := os.join_path(r.repos_dir, repo, arch, '$pkg.info.name-$pkg.info.version')
|
||||
|
||||
// Remove the previous version of the package, if present
|
||||
r.remove_pkg_from_arch_repo(repo, arch, pkg.info.name, false) ?
|
||||
r.remove_pkg_from_arch_repo(repo, arch, pkg.info.name, false)?
|
||||
|
||||
os.mkdir_all(pkg_dir) or { return error('Failed to create package directory.') }
|
||||
|
||||
os.write_file(os.join_path_single(pkg_dir, 'desc'), pkg.to_desc()) or {
|
||||
os.rmdir_all(pkg_dir) ?
|
||||
os.rmdir_all(pkg_dir)?
|
||||
|
||||
return error('Failed to write desc file.')
|
||||
}
|
||||
os.write_file(os.join_path_single(pkg_dir, 'files'), pkg.to_files()) or {
|
||||
os.rmdir_all(pkg_dir) ?
|
||||
os.rmdir_all(pkg_dir)?
|
||||
|
||||
return error('Failed to write files file.')
|
||||
}
|
||||
|
||||
r.sync(repo, arch) ?
|
||||
r.sync(repo, arch)?
|
||||
|
||||
return true
|
||||
}
|
||||
|
@ -168,7 +168,7 @@ fn (r &RepoGroupManager) remove_pkg_from_arch_repo(repo string, arch string, pkg
|
|||
|
||||
// We iterate over every directory in the repo dir
|
||||
// TODO filter so we only check directories
|
||||
for d in os.ls(repo_dir) ? {
|
||||
for d in os.ls(repo_dir)? {
|
||||
// Because a repository only allows a single version of each package,
|
||||
// we need only compare whether the name of the package is the same,
|
||||
// not the version.
|
||||
|
@ -178,22 +178,22 @@ fn (r &RepoGroupManager) remove_pkg_from_arch_repo(repo string, arch string, pkg
|
|||
// We lock the mutex here to prevent other routines from creating a
|
||||
// new archive while we remove an entry
|
||||
lock r.mutex {
|
||||
os.rmdir_all(os.join_path_single(repo_dir, d)) ?
|
||||
os.rmdir_all(os.join_path_single(repo_dir, d))?
|
||||
}
|
||||
|
||||
// Also remove the package archive
|
||||
repo_pkg_dir := os.join_path(r.pkg_dir, repo, arch)
|
||||
|
||||
archives := os.ls(repo_pkg_dir) ?.filter(it.split('-')#[..-3].join('-') == name)
|
||||
archives := os.ls(repo_pkg_dir)?.filter(it.split('-')#[..-3].join('-') == name)
|
||||
|
||||
for archive_name in archives {
|
||||
full_path := os.join_path_single(repo_pkg_dir, archive_name)
|
||||
os.rm(full_path) ?
|
||||
os.rm(full_path)?
|
||||
}
|
||||
|
||||
// Sync the db archives if requested
|
||||
if sync {
|
||||
r.sync(repo, arch) ?
|
||||
r.sync(repo, arch)?
|
||||
}
|
||||
|
||||
return true
|
||||
|
|
|
@ -54,7 +54,7 @@ fn (r &RepoGroupManager) sync(repo string, arch string) ? {
|
|||
C.archive_write_open_filename(a_files, &char(files_path.str))
|
||||
|
||||
// Iterate over each directory
|
||||
for d in os.ls(subrepo_path) ?.filter(os.is_dir(os.join_path_single(subrepo_path,
|
||||
for d in os.ls(subrepo_path)?.filter(os.is_dir(os.join_path_single(subrepo_path,
|
||||
it))) {
|
||||
// desc
|
||||
mut inner_path := os.join_path_single(d, 'desc')
|
||||
|
|
|
@ -18,10 +18,10 @@ pub fn cmd() cli.Command {
|
|||
name: 'server'
|
||||
description: 'Start the Vieter server.'
|
||||
execute: fn (cmd cli.Command) ? {
|
||||
config_file := cmd.flags.get_string('config-file') ?
|
||||
conf := env.load<Config>(config_file) ?
|
||||
config_file := cmd.flags.get_string('config-file')?
|
||||
conf := env.load<Config>(config_file)?
|
||||
|
||||
server(conf) ?
|
||||
server(conf)?
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -0,0 +1,136 @@
|
|||
module server
|
||||
|
||||
import web
|
||||
import net.http
|
||||
import net.urllib
|
||||
import response { new_data_response, new_response }
|
||||
import db
|
||||
import time
|
||||
import os
|
||||
import util
|
||||
|
||||
// get_logs returns all build logs in the database. A 'repo' query param can
|
||||
// optionally be added to limit the list of build logs to that repository.
|
||||
['/api/logs'; get]
|
||||
fn (mut app App) get_logs() web.Result {
|
||||
if !app.is_authorized() {
|
||||
return app.json(http.Status.unauthorized, new_response('Unauthorized.'))
|
||||
}
|
||||
|
||||
logs := if 'repo' in app.query {
|
||||
app.db.get_build_logs_for_repo(app.query['repo'].int())
|
||||
} else {
|
||||
app.db.get_build_logs()
|
||||
}
|
||||
|
||||
return app.json(http.Status.ok, new_data_response(logs))
|
||||
}
|
||||
|
||||
// get_single_log returns the build log with the given id.
|
||||
['/api/logs/:id'; get]
|
||||
fn (mut app App) get_single_log(id int) web.Result {
|
||||
if !app.is_authorized() {
|
||||
return app.json(http.Status.unauthorized, new_response('Unauthorized.'))
|
||||
}
|
||||
|
||||
log := app.db.get_build_log(id) or { return app.not_found() }
|
||||
|
||||
return app.json(http.Status.ok, new_data_response(log))
|
||||
}
|
||||
|
||||
// get_log_content returns the actual build log file for the given id.
|
||||
['/api/logs/:id/content'; get]
|
||||
fn (mut app App) get_log_content(id int) web.Result {
|
||||
if !app.is_authorized() {
|
||||
return app.json(http.Status.unauthorized, new_response('Unauthorized.'))
|
||||
}
|
||||
|
||||
log := app.db.get_build_log(id) or { return app.not_found() }
|
||||
file_name := log.start_time.custom_format('YYYY-MM-DD_HH-mm-ss')
|
||||
full_path := os.join_path(app.conf.data_dir, logs_dir_name, log.repo_id.str(), log.arch,
|
||||
file_name)
|
||||
|
||||
return app.file(full_path)
|
||||
}
|
||||
|
||||
// parse_query_time unescapes an HTTP query parameter & tries to parse it as a
|
||||
// time.Time struct.
|
||||
fn parse_query_time(query string) ?time.Time {
|
||||
unescaped := urllib.query_unescape(query)?
|
||||
t := time.parse(unescaped)?
|
||||
|
||||
return t
|
||||
}
|
||||
|
||||
// post_log adds a new log to the database.
|
||||
['/api/logs'; post]
|
||||
fn (mut app App) post_log() web.Result {
|
||||
if !app.is_authorized() {
|
||||
return app.json(http.Status.unauthorized, new_response('Unauthorized.'))
|
||||
}
|
||||
|
||||
// Parse query params
|
||||
start_time := parse_query_time(app.query['startTime']) or {
|
||||
return app.json(http.Status.bad_request, new_response('Invalid or missing start time.'))
|
||||
}
|
||||
|
||||
end_time := parse_query_time(app.query['endTime']) or {
|
||||
return app.json(http.Status.bad_request, new_response('Invalid or missing end time.'))
|
||||
}
|
||||
|
||||
if 'exitCode' !in app.query {
|
||||
return app.json(http.Status.bad_request, new_response('Missing exit code.'))
|
||||
}
|
||||
|
||||
exit_code := app.query['exitCode'].int()
|
||||
|
||||
if 'arch' !in app.query {
|
||||
return app.json(http.Status.bad_request, new_response("Missing parameter 'arch'."))
|
||||
}
|
||||
|
||||
arch := app.query['arch']
|
||||
|
||||
repo_id := app.query['repo'].int()
|
||||
|
||||
if !app.db.git_repo_exists(repo_id) {
|
||||
return app.json(http.Status.bad_request, new_response('Unknown Git repo.'))
|
||||
}
|
||||
|
||||
// Store log in db
|
||||
log := db.BuildLog{
|
||||
repo_id: repo_id
|
||||
start_time: start_time
|
||||
end_time: end_time
|
||||
arch: arch
|
||||
exit_code: exit_code
|
||||
}
|
||||
|
||||
app.db.add_build_log(log)
|
||||
|
||||
repo_logs_dir := os.join_path(app.conf.data_dir, logs_dir_name, repo_id.str(), arch)
|
||||
|
||||
// Create the logs directory of it doesn't exist
|
||||
if !os.exists(repo_logs_dir) {
|
||||
os.mkdir_all(repo_logs_dir) or {
|
||||
app.lerror("Couldn't create dir '$repo_logs_dir'.")
|
||||
|
||||
return app.json(http.Status.internal_server_error, new_response('An error occured while processing the request.'))
|
||||
}
|
||||
}
|
||||
|
||||
// Stream log contents to correct file
|
||||
file_name := start_time.custom_format('YYYY-MM-DD_HH-mm-ss')
|
||||
full_path := os.join_path_single(repo_logs_dir, file_name)
|
||||
|
||||
if length := app.req.header.get(.content_length) {
|
||||
util.reader_to_file(mut app.reader, length.int(), full_path) or {
|
||||
app.lerror('An error occured while receiving logs: $err.msg()')
|
||||
|
||||
return app.json(http.Status.internal_server_error, new_response('Failed to upload logs.'))
|
||||
}
|
||||
} else {
|
||||
return app.status(http.Status.length_required)
|
||||
}
|
||||
|
||||
return app.json(http.Status.ok, new_response('Logs added successfully.'))
|
||||
}
|
|
@ -12,6 +12,7 @@ const (
|
|||
log_file_name = 'vieter.log'
|
||||
repo_dir_name = 'repos'
|
||||
db_file_name = 'vieter.sqlite'
|
||||
logs_dir_name = 'logs'
|
||||
)
|
||||
|
||||
struct App {
|
||||
|
@ -37,6 +38,14 @@ pub fn server(conf Config) ? {
|
|||
|
||||
os.mkdir_all(conf.data_dir) or { util.exit_with_message(1, 'Failed to create data directory.') }
|
||||
|
||||
logs_dir := os.join_path_single(conf.data_dir, server.logs_dir_name)
|
||||
|
||||
if !os.exists(logs_dir) {
|
||||
os.mkdir(os.join_path_single(conf.data_dir, server.logs_dir_name)) or {
|
||||
util.exit_with_message(1, 'Failed to create logs directory.')
|
||||
}
|
||||
}
|
||||
|
||||
mut logger := log.Log{
|
||||
level: log_level
|
||||
}
|
||||
|
|
|
@ -0,0 +1,2 @@
|
|||
This module defines a few useful functions used throughout the codebase that
|
||||
don't specifically fit inside a module.
|
|
@ -0,0 +1,95 @@
|
|||
// Functions for interacting with `io.Reader` & `io.Writer` objects.
|
||||
module util
|
||||
|
||||
import io
|
||||
import os
|
||||
|
||||
// reader_to_writer tries to consume the entire reader & write it to the writer.
|
||||
pub fn reader_to_writer(mut reader io.Reader, mut writer io.Writer) ? {
|
||||
mut buf := []u8{len: 10 * 1024}
|
||||
|
||||
for {
|
||||
bytes_read := reader.read(mut buf) or { break }
|
||||
mut bytes_written := 0
|
||||
|
||||
for bytes_written < bytes_read {
|
||||
c := writer.write(buf[bytes_written..bytes_read]) or { break }
|
||||
|
||||
bytes_written += c
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// reader_to_file writes the contents of a BufferedReader to a file
|
||||
pub fn reader_to_file(mut reader io.BufferedReader, length int, path string) ? {
|
||||
mut file := os.create(path)?
|
||||
defer {
|
||||
file.close()
|
||||
}
|
||||
|
||||
mut buf := []u8{len: reader_buf_size}
|
||||
mut bytes_left := length
|
||||
|
||||
// Repeat as long as the stream still has data
|
||||
for bytes_left > 0 {
|
||||
// TODO check if just breaking here is safe
|
||||
bytes_read := reader.read(mut buf) or { break }
|
||||
bytes_left -= bytes_read
|
||||
|
||||
mut to_write := bytes_read
|
||||
|
||||
for to_write > 0 {
|
||||
// TODO don't just loop infinitely here
|
||||
bytes_written := file.write(buf[bytes_read - to_write..bytes_read]) or { continue }
|
||||
// file.flush()
|
||||
|
||||
to_write = to_write - bytes_written
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// match_array_in_array<T> returns how many elements of a2 overlap with a1. For
|
||||
// example, if a1 = "abcd" & a2 = "cd", the result will be 2. If the match is
|
||||
// not at the end of a1, the result is 0.
|
||||
pub fn match_array_in_array<T>(a1 []T, a2 []T) int {
|
||||
mut i := 0
|
||||
mut match_len := 0
|
||||
|
||||
for i + match_len < a1.len {
|
||||
if a1[i + match_len] == a2[match_len] {
|
||||
match_len += 1
|
||||
} else {
|
||||
i += match_len + 1
|
||||
match_len = 0
|
||||
}
|
||||
}
|
||||
|
||||
return match_len
|
||||
}
|
||||
|
||||
// read_until_separator consumes an io.Reader until it encounters some
|
||||
// separator array. The data read is stored inside the provided res array.
|
||||
pub fn read_until_separator(mut reader io.Reader, mut res []u8, sep []u8) ? {
|
||||
mut buf := []u8{len: sep.len}
|
||||
|
||||
for {
|
||||
c := reader.read(mut buf)?
|
||||
res << buf[..c]
|
||||
|
||||
match_len := match_array_in_array(buf[..c], sep)
|
||||
|
||||
if match_len == sep.len {
|
||||
break
|
||||
}
|
||||
|
||||
if match_len > 0 {
|
||||
match_left := sep.len - match_len
|
||||
c2 := reader.read(mut buf[..match_left])?
|
||||
res << buf[..c2]
|
||||
|
||||
if buf[..c2] == sep[match_len..] {
|
||||
break
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
|
@ -1,13 +1,13 @@
|
|||
module util
|
||||
|
||||
import os
|
||||
import io
|
||||
import crypto.md5
|
||||
import crypto.sha256
|
||||
|
||||
const reader_buf_size = 1_000_000
|
||||
|
||||
const prefixes = ['B', 'KB', 'MB', 'GB']
|
||||
const (
|
||||
reader_buf_size = 1_000_000
|
||||
prefixes = ['B', 'KB', 'MB', 'GB']
|
||||
)
|
||||
|
||||
// Dummy struct to work around the fact that you can only share structs, maps &
|
||||
// arrays
|
||||
|
@ -23,34 +23,6 @@ pub fn exit_with_message(code int, msg string) {
|
|||
exit(code)
|
||||
}
|
||||
|
||||
// reader_to_file writes the contents of a BufferedReader to a file
|
||||
pub fn reader_to_file(mut reader io.BufferedReader, length int, path string) ? {
|
||||
mut file := os.create(path) ?
|
||||
defer {
|
||||
file.close()
|
||||
}
|
||||
|
||||
mut buf := []u8{len: util.reader_buf_size}
|
||||
mut bytes_left := length
|
||||
|
||||
// Repeat as long as the stream still has data
|
||||
for bytes_left > 0 {
|
||||
// TODO check if just breaking here is safe
|
||||
bytes_read := reader.read(mut buf) or { break }
|
||||
bytes_left -= bytes_read
|
||||
|
||||
mut to_write := bytes_read
|
||||
|
||||
for to_write > 0 {
|
||||
// TODO don't just loop infinitely here
|
||||
bytes_written := file.write(buf[bytes_read - to_write..bytes_read]) or { continue }
|
||||
// file.flush()
|
||||
|
||||
to_write = to_write - bytes_written
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// hash_file returns the md5 & sha256 hash of a given file
|
||||
// TODO actually implement sha256
|
||||
pub fn hash_file(path &string) ?(string, string) {
|
||||
|
|
|
@ -190,7 +190,7 @@ pub fn (ctx Context) before_request() {}
|
|||
|
||||
// send_string
|
||||
fn send_string(mut conn net.TcpConn, s string) ? {
|
||||
conn.write(s.bytes()) ?
|
||||
conn.write(s.bytes())?
|
||||
}
|
||||
|
||||
// send_response_to_client sends a response to the client
|
||||
|
|
Loading…
Reference in New Issue