forked from vieter-v/vieter
chore: ran v fmt for v 0.3.3 changes
parent
e10b450abd
commit
b3a119f221
|
@ -23,7 +23,7 @@ pub fn cmd() cli.Command {
|
|||
description: 'Start an agent daemon.'
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
agent(conf)!
|
||||
}
|
||||
|
|
|
@ -51,7 +51,7 @@ pub fn (mut d AgentDaemon) run() {
|
|||
|
||||
for {
|
||||
if sleep_time > 0 {
|
||||
d.ldebug('Sleeping for $sleep_time')
|
||||
d.ldebug('Sleeping for ${sleep_time}')
|
||||
time.sleep(sleep_time)
|
||||
}
|
||||
|
||||
|
@ -80,14 +80,14 @@ pub fn (mut d AgentDaemon) run() {
|
|||
d.ldebug('Polling for new jobs')
|
||||
|
||||
new_configs := d.client.poll_jobs(d.conf.arch, finished + empty) or {
|
||||
d.lerror('Failed to poll jobs: $err.msg()')
|
||||
d.lerror('Failed to poll jobs: ${err.msg()}')
|
||||
|
||||
// TODO pick a better delay here
|
||||
sleep_time = 5 * time.second
|
||||
continue
|
||||
}
|
||||
|
||||
d.ldebug('Received $new_configs.len jobs')
|
||||
d.ldebug('Received ${new_configs.len} jobs')
|
||||
|
||||
last_poll_time = time.now()
|
||||
|
||||
|
@ -95,7 +95,7 @@ pub fn (mut d AgentDaemon) run() {
|
|||
// Make sure a recent build base image is available for
|
||||
// building the config
|
||||
if !d.images.up_to_date(config.base_image) {
|
||||
d.linfo('Building builder image from base image $config.base_image')
|
||||
d.linfo('Building builder image from base image ${config.base_image}')
|
||||
|
||||
// TODO handle this better than to just skip the config
|
||||
d.images.refresh_image(config.base_image) or {
|
||||
|
@ -154,7 +154,7 @@ fn (mut d AgentDaemon) start_build(config BuildConfig) bool {
|
|||
stdatomic.store_u64(&d.atomics[i], agent.build_running)
|
||||
d.builds[i] = config
|
||||
|
||||
go d.run_build(i, config)
|
||||
spawn d.run_build(i, config)
|
||||
|
||||
return true
|
||||
}
|
||||
|
@ -165,7 +165,7 @@ fn (mut d AgentDaemon) start_build(config BuildConfig) bool {
|
|||
|
||||
// run_build actually starts the build process for a given target.
|
||||
fn (mut d AgentDaemon) run_build(build_index int, config BuildConfig) {
|
||||
d.linfo('started build: $config')
|
||||
d.linfo('started build: ${config}')
|
||||
|
||||
// 0 means success, 1 means failure
|
||||
mut status := 0
|
||||
|
@ -176,21 +176,21 @@ fn (mut d AgentDaemon) run_build(build_index int, config BuildConfig) {
|
|||
}
|
||||
|
||||
res := build.build_config(d.client.address, d.client.api_key, new_config) or {
|
||||
d.ldebug('build_config error: $err.msg()')
|
||||
d.ldebug('build_config error: ${err.msg()}')
|
||||
status = 1
|
||||
|
||||
build.BuildResult{}
|
||||
}
|
||||
|
||||
if status == 0 {
|
||||
d.linfo('Uploading build logs for $config')
|
||||
d.linfo('Uploading build logs for ${config}')
|
||||
|
||||
// TODO use the arch value here
|
||||
build_arch := os.uname().machine
|
||||
d.client.add_build_log(config.target_id, res.start_time, res.end_time, build_arch,
|
||||
res.exit_code, res.logs) or { d.lerror('Failed to upload logs for $config') }
|
||||
res.exit_code, res.logs) or { d.lerror('Failed to upload logs for ${config}') }
|
||||
} else {
|
||||
d.lwarn('an error occurred during build: $config')
|
||||
d.lwarn('an error occurred during build: ${config}')
|
||||
}
|
||||
|
||||
stdatomic.store_u64(&d.atomics[build_index], agent.build_done)
|
||||
|
|
|
@ -71,7 +71,7 @@ pub fn (mut m ImageManager) up_to_date(base_image string) bool {
|
|||
fn (mut m ImageManager) refresh_image(base_image string) ! {
|
||||
// TODO use better image tags for built images
|
||||
new_image := build.create_build_image(base_image) or {
|
||||
return error('Failed to build builder image from base image $base_image')
|
||||
return error('Failed to build builder image from base image ${base_image}')
|
||||
}
|
||||
|
||||
m.images[base_image] << new_image
|
||||
|
|
|
@ -45,7 +45,7 @@ pub fn create_build_image(base_image string) !string {
|
|||
|
||||
c := docker.NewContainer{
|
||||
image: base_image
|
||||
env: ['BUILD_SCRIPT=$cmds_str']
|
||||
env: ['BUILD_SCRIPT=${cmds_str}']
|
||||
entrypoint: ['/bin/sh', '-c']
|
||||
cmd: ['echo \$BUILD_SCRIPT | base64 -d | /bin/sh -e']
|
||||
}
|
||||
|
@ -118,10 +118,10 @@ pub fn build_config(address string, api_key string, config BuildConfig) !BuildRe
|
|||
base64_script := base64.encode_str(build_script)
|
||||
|
||||
c := docker.NewContainer{
|
||||
image: '$config.base_image'
|
||||
image: '${config.base_image}'
|
||||
env: [
|
||||
'BUILD_SCRIPT=$base64_script',
|
||||
'API_KEY=$api_key',
|
||||
'BUILD_SCRIPT=${base64_script}',
|
||||
'API_KEY=${api_key}',
|
||||
// `archlinux:base-devel` does not correctly set the path variable,
|
||||
// causing certain builds to fail. This fixes it.
|
||||
'PATH=${build.path_dirs.join(':')}',
|
||||
|
|
|
@ -36,7 +36,7 @@ pub struct BuildJobQueue {
|
|||
mut:
|
||||
mutex shared util.Dummy
|
||||
// For each architecture, a priority queue is tracked
|
||||
queues map[string]MinHeap<BuildJob>
|
||||
queues map[string]MinHeap[BuildJob]
|
||||
// When a target is removed from the server or edited, its previous build
|
||||
// configs will be invalid. This map allows for those to be simply skipped
|
||||
// by ignoring any build configs created before this timestamp.
|
||||
|
@ -74,7 +74,7 @@ pub struct InsertConfig {
|
|||
pub fn (mut q BuildJobQueue) insert(input InsertConfig) ! {
|
||||
lock q.mutex {
|
||||
if input.arch !in q.queues {
|
||||
q.queues[input.arch] = MinHeap<BuildJob>{}
|
||||
q.queues[input.arch] = MinHeap[BuildJob]{}
|
||||
}
|
||||
|
||||
mut job := BuildJob{
|
||||
|
@ -86,7 +86,7 @@ pub fn (mut q BuildJobQueue) insert(input InsertConfig) ! {
|
|||
if !input.now {
|
||||
ce := if input.target.schedule != '' {
|
||||
cron.parse_expression(input.target.schedule) or {
|
||||
return error("Error while parsing cron expression '$input.target.schedule' (id $input.target.id): $err.msg()")
|
||||
return error("Error while parsing cron expression '${input.target.schedule}' (id ${input.target.id}): ${err.msg()}")
|
||||
}
|
||||
} else {
|
||||
q.default_schedule
|
||||
|
|
|
@ -24,12 +24,12 @@ pub fn echo_commands(cmds []string) []string {
|
|||
|
||||
// create_build_script generates a shell script that builds a given Target.
|
||||
fn create_build_script(address string, config BuildConfig, build_arch string) string {
|
||||
repo_url := '$address/$config.repo'
|
||||
repo_url := '${address}/${config.repo}'
|
||||
|
||||
mut commands := [
|
||||
// This will later be replaced by a proper setting for changing the
|
||||
// mirrorlist
|
||||
"echo -e '[$config.repo]\\nServer = $address/\$repo/\$arch\\nSigLevel = Optional' >> /etc/pacman.conf"
|
||||
"echo -e '[${config.repo}]\\nServer = ${address}/\$repo/\$arch\\nSigLevel = Optional' >> /etc/pacman.conf"
|
||||
// We need to update the package list of the repo we just added above.
|
||||
// This should however not pull in a lot of packages as long as the
|
||||
// builder image is rebuilt frequently.
|
||||
|
@ -42,18 +42,18 @@ fn create_build_script(address string, config BuildConfig, build_arch string) st
|
|||
'git' {
|
||||
if config.branch == '' {
|
||||
[
|
||||
"git clone --single-branch --depth 1 '$config.url' repo",
|
||||
"git clone --single-branch --depth 1 '${config.url}' repo",
|
||||
]
|
||||
} else {
|
||||
[
|
||||
"git clone --single-branch --depth 1 --branch $config.branch '$config.url' repo",
|
||||
"git clone --single-branch --depth 1 --branch ${config.branch} '${config.url}' repo",
|
||||
]
|
||||
}
|
||||
}
|
||||
'url' {
|
||||
[
|
||||
'mkdir repo',
|
||||
"curl -o repo/PKGBUILD -L '$config.url'",
|
||||
"curl -o repo/PKGBUILD -L '${config.url}'",
|
||||
]
|
||||
}
|
||||
else {
|
||||
|
@ -62,7 +62,7 @@ fn create_build_script(address string, config BuildConfig, build_arch string) st
|
|||
}
|
||||
|
||||
commands << if config.path != '' {
|
||||
"cd 'repo/$config.path'"
|
||||
"cd 'repo/${config.path}'"
|
||||
} else {
|
||||
'cd repo'
|
||||
}
|
||||
|
@ -76,7 +76,7 @@ fn create_build_script(address string, config BuildConfig, build_arch string) st
|
|||
// The build container checks whether the package is already present on
|
||||
// the server.
|
||||
commands << [
|
||||
'curl -s --head --fail $repo_url/$build_arch/\$pkgname-\$pkgver-\$pkgrel && exit 0',
|
||||
'curl -s --head --fail ${repo_url}/${build_arch}/\$pkgname-\$pkgver-\$pkgrel && exit 0',
|
||||
// If the above curl command succeeds, we don't need to rebuild the
|
||||
// package. However, because we're in a su shell, the exit command will
|
||||
// drop us back into the root shell. Therefore, we must check whether
|
||||
|
@ -86,7 +86,7 @@ fn create_build_script(address string, config BuildConfig, build_arch string) st
|
|||
}
|
||||
|
||||
commands << [
|
||||
'MAKEFLAGS="-j\$(nproc)" makepkg -s --noconfirm --needed --noextract && for pkg in \$(ls -1 *.pkg*); do curl -XPOST -T "\$pkg" -H "X-API-KEY: \$API_KEY" $repo_url/publish; done',
|
||||
'MAKEFLAGS="-j\$(nproc)" makepkg -s --noconfirm --needed --noextract && for pkg in \$(ls -1 *.pkg*); do curl -XPOST -T "\$pkg" -H "X-API-KEY: \$API_KEY" ${repo_url}/publish; done',
|
||||
]
|
||||
|
||||
return echo_commands(commands).join('\n')
|
||||
|
|
|
@ -22,7 +22,7 @@ pub fn new(address string, api_key string) Client {
|
|||
// send_request_raw sends an HTTP request, returning the http.Response object.
|
||||
// It encodes the params so that they're safe to pass as HTTP query parameters.
|
||||
fn (c &Client) send_request_raw(method Method, url string, params map[string]string, body string) !http.Response {
|
||||
mut full_url := '$c.address$url'
|
||||
mut full_url := '${c.address}${url}'
|
||||
|
||||
if params.len > 0 {
|
||||
mut params_escaped := map[string]string{}
|
||||
|
@ -33,9 +33,9 @@ fn (c &Client) send_request_raw(method Method, url string, params map[string]str
|
|||
params_escaped[k] = urllib.query_escape(v)
|
||||
}
|
||||
|
||||
params_str := params_escaped.keys().map('$it=${params_escaped[it]}').join('&')
|
||||
params_str := params_escaped.keys().map('${it}=${params_escaped[it]}').join('&')
|
||||
|
||||
full_url = '$full_url?$params_str'
|
||||
full_url = '${full_url}?${params_str}'
|
||||
}
|
||||
|
||||
// Looking at the source code, this function doesn't actually fail, so I'm
|
||||
|
@ -49,13 +49,13 @@ fn (c &Client) send_request_raw(method Method, url string, params map[string]str
|
|||
}
|
||||
|
||||
// send_request<T> just calls send_request_with_body<T> with an empty body.
|
||||
fn (c &Client) send_request<T>(method Method, url string, params map[string]string) !Response<T> {
|
||||
return c.send_request_with_body<T>(method, url, params, '')
|
||||
fn (c &Client) send_request[T](method Method, url string, params map[string]string) !Response[T] {
|
||||
return c.send_request_with_body[T](method, url, params, '')
|
||||
}
|
||||
|
||||
// send_request_with_body<T> calls send_request_raw_response & parses its
|
||||
// output as a Response<T> object.
|
||||
fn (c &Client) send_request_with_body<T>(method Method, url string, params map[string]string, body string) !Response<T> {
|
||||
fn (c &Client) send_request_with_body[T](method Method, url string, params map[string]string, body string) !Response[T] {
|
||||
res := c.send_request_raw(method, url, params, body)!
|
||||
status := res.status()
|
||||
|
||||
|
@ -64,12 +64,12 @@ fn (c &Client) send_request_with_body<T>(method Method, url string, params map[s
|
|||
if status.is_error() {
|
||||
// A non-successful status call will have an empty body
|
||||
if res.body == '' {
|
||||
return error('Error $res.status_code ($status.str()): (empty response)')
|
||||
return error('Error ${res.status_code} (${status.str()}): (empty response)')
|
||||
}
|
||||
|
||||
data := json.decode(Response<string>, res.body)!
|
||||
data := json.decode(Response[string], res.body)!
|
||||
|
||||
return error('Status $res.status_code ($status.str()): $data.message')
|
||||
return error('Status ${res.status_code} (${status.str()}): ${data.message}')
|
||||
}
|
||||
|
||||
// Just return an empty successful response
|
||||
|
@ -77,7 +77,7 @@ fn (c &Client) send_request_with_body<T>(method Method, url string, params map[s
|
|||
return new_data_response(T{})
|
||||
}
|
||||
|
||||
data := json.decode(Response<T>, res.body)!
|
||||
data := json.decode(Response[T], res.body)!
|
||||
|
||||
return data
|
||||
}
|
||||
|
|
|
@ -4,7 +4,7 @@ import models { BuildConfig }
|
|||
|
||||
// poll_jobs requests a list of new build jobs from the server.
|
||||
pub fn (c &Client) poll_jobs(arch string, max int) ![]BuildConfig {
|
||||
data := c.send_request<[]BuildConfig>(.get, '/api/v1/jobs/poll', {
|
||||
data := c.send_request[[]BuildConfig](.get, '/api/v1/jobs/poll', {
|
||||
'arch': arch
|
||||
'max': max.str()
|
||||
})!
|
||||
|
@ -15,7 +15,7 @@ pub fn (c &Client) poll_jobs(arch string, max int) ![]BuildConfig {
|
|||
// queue_job adds a new one-time build job for the given target to the job
|
||||
// queue.
|
||||
pub fn (c &Client) queue_job(target_id int, arch string, force bool) ! {
|
||||
c.send_request<string>(.post, '/api/v1/jobs/queue', {
|
||||
c.send_request[string](.post, '/api/v1/jobs/queue', {
|
||||
'target': target_id.str()
|
||||
'arch': arch
|
||||
'force': force.str()
|
||||
|
|
|
@ -7,27 +7,27 @@ import time
|
|||
// get_build_logs returns all build logs.
|
||||
pub fn (c &Client) get_build_logs(filter BuildLogFilter) ![]BuildLog {
|
||||
params := models.params_from(filter)
|
||||
data := c.send_request<[]BuildLog>(.get, '/api/v1/logs', params)!
|
||||
data := c.send_request[[]BuildLog](.get, '/api/v1/logs', params)!
|
||||
|
||||
return data.data
|
||||
}
|
||||
|
||||
// get_build_log returns a specific build log.
|
||||
pub fn (c &Client) get_build_log(id int) !BuildLog {
|
||||
data := c.send_request<BuildLog>(.get, '/api/v1/logs/$id', {})!
|
||||
data := c.send_request[BuildLog](.get, '/api/v1/logs/${id}', {})!
|
||||
|
||||
return data.data
|
||||
}
|
||||
|
||||
// get_build_log_content returns the contents of the build log file.
|
||||
pub fn (c &Client) get_build_log_content(id int) !string {
|
||||
data := c.send_request_raw_response(.get, '/api/v1/logs/$id/content', {}, '')!
|
||||
data := c.send_request_raw_response(.get, '/api/v1/logs/${id}/content', {}, '')!
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// add_build_log adds a new build log to the server.
|
||||
pub fn (c &Client) add_build_log(target_id int, start_time time.Time, end_time time.Time, arch string, exit_code int, content string) !Response<int> {
|
||||
pub fn (c &Client) add_build_log(target_id int, start_time time.Time, end_time time.Time, arch string, exit_code int, content string) !Response[int] {
|
||||
params := {
|
||||
'target': target_id.str()
|
||||
'startTime': start_time.unix_time().str()
|
||||
|
@ -36,12 +36,12 @@ pub fn (c &Client) add_build_log(target_id int, start_time time.Time, end_time t
|
|||
'exitCode': exit_code.str()
|
||||
}
|
||||
|
||||
data := c.send_request_with_body<int>(.post, '/api/v1/logs', params, content)!
|
||||
data := c.send_request_with_body[int](.post, '/api/v1/logs', params, content)!
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// remove_build_log removes the build log with the given id from the server.
|
||||
pub fn (c &Client) remove_build_log(id int) ! {
|
||||
c.send_request<string>(.delete, '/api/v1/logs/$id', {})!
|
||||
c.send_request[string](.delete, '/api/v1/logs/${id}', {})!
|
||||
}
|
||||
|
|
|
@ -2,15 +2,15 @@ module client
|
|||
|
||||
// remove_repo removes an entire repository.
|
||||
pub fn (c &Client) remove_repo(repo string) ! {
|
||||
c.send_request<string>(.delete, '/$repo', {})!
|
||||
c.send_request[string](.delete, '/${repo}', {})!
|
||||
}
|
||||
|
||||
// remove_arch_repo removes an entire arch-repo.
|
||||
pub fn (c &Client) remove_arch_repo(repo string, arch string) ! {
|
||||
c.send_request<string>(.delete, '/$repo/$arch', {})!
|
||||
c.send_request[string](.delete, '/${repo}/${arch}', {})!
|
||||
}
|
||||
|
||||
// remove_package removes a single package from the given arch-repo.
|
||||
pub fn (c &Client) remove_package(repo string, arch string, pkgname string) ! {
|
||||
c.send_request<string>(.delete, '/$repo/$arch/$pkgname', {})!
|
||||
c.send_request[string](.delete, '/${repo}/${arch}/${pkgname}', {})!
|
||||
}
|
||||
|
|
|
@ -5,7 +5,7 @@ import models { Target, TargetFilter }
|
|||
// get_targets returns a list of targets, given a filter object.
|
||||
pub fn (c &Client) get_targets(filter TargetFilter) ![]Target {
|
||||
params := models.params_from(filter)
|
||||
data := c.send_request<[]Target>(.get, '/api/v1/targets', params)!
|
||||
data := c.send_request[[]Target](.get, '/api/v1/targets', params)!
|
||||
|
||||
return data.data
|
||||
}
|
||||
|
@ -33,7 +33,7 @@ pub fn (c &Client) get_all_targets() ![]Target {
|
|||
|
||||
// get_target returns the target for a specific id.
|
||||
pub fn (c &Client) get_target(id int) !Target {
|
||||
data := c.send_request<Target>(.get, '/api/v1/targets/$id', {})!
|
||||
data := c.send_request[Target](.get, '/api/v1/targets/${id}', {})!
|
||||
|
||||
return data.data
|
||||
}
|
||||
|
@ -49,15 +49,15 @@ pub struct NewTarget {
|
|||
|
||||
// add_target adds a new target to the server.
|
||||
pub fn (c &Client) add_target(t NewTarget) !int {
|
||||
params := models.params_from<NewTarget>(t)
|
||||
data := c.send_request<int>(.post, '/api/v1/targets', params)!
|
||||
params := models.params_from[NewTarget](t)
|
||||
data := c.send_request[int](.post, '/api/v1/targets', params)!
|
||||
|
||||
return data.data
|
||||
}
|
||||
|
||||
// remove_target removes the target with the given id from the server.
|
||||
pub fn (c &Client) remove_target(id int) !string {
|
||||
data := c.send_request<string>(.delete, '/api/v1/targets/$id', {})!
|
||||
data := c.send_request[string](.delete, '/api/v1/targets/${id}', {})!
|
||||
|
||||
return data.data
|
||||
}
|
||||
|
@ -65,7 +65,7 @@ pub fn (c &Client) remove_target(id int) !string {
|
|||
// patch_target sends a PATCH request to the given target with the params as
|
||||
// payload.
|
||||
pub fn (c &Client) patch_target(id int, params map[string]string) !string {
|
||||
data := c.send_request<string>(.patch, '/api/v1/targets/$id', params)!
|
||||
data := c.send_request[string](.patch, '/api/v1/targets/${id}', params)!
|
||||
|
||||
return data.data
|
||||
}
|
||||
|
|
|
@ -36,7 +36,7 @@ pub fn cmd() cli.Command {
|
|||
required_args: 2
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
c := aur.new()
|
||||
pkgs := c.info(cmd.args[1..])!
|
||||
|
@ -46,14 +46,14 @@ pub fn cmd() cli.Command {
|
|||
for pkg in pkgs {
|
||||
vc.add_target(
|
||||
kind: 'git'
|
||||
url: 'https://aur.archlinux.org/$pkg.package_base' + '.git'
|
||||
url: 'https://aur.archlinux.org/${pkg.package_base}' + '.git'
|
||||
repo: cmd.args[0]
|
||||
) or {
|
||||
println('Failed to add $pkg.name: $err.msg()')
|
||||
println('Failed to add ${pkg.name}: ${err.msg()}')
|
||||
continue
|
||||
}
|
||||
|
||||
println('Added $pkg.name' + '.')
|
||||
println('Added ${pkg.name}' + '.')
|
||||
}
|
||||
}
|
||||
},
|
||||
|
|
|
@ -74,7 +74,7 @@ pub fn cmd() cli.Command {
|
|||
]
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
mut filter := BuildLogFilter{}
|
||||
|
||||
|
@ -156,7 +156,7 @@ pub fn cmd() cli.Command {
|
|||
description: 'Remove a build log that matches the given id.'
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
remove(conf, cmd.args[0])!
|
||||
}
|
||||
|
@ -168,7 +168,7 @@ pub fn cmd() cli.Command {
|
|||
description: 'Show all info for a specific build log.'
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
id := cmd.args[0].int()
|
||||
info(conf, id)!
|
||||
|
@ -181,7 +181,7 @@ pub fn cmd() cli.Command {
|
|||
description: 'Output the content of a build log to stdout.'
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
id := cmd.args[0].int()
|
||||
content(conf, id)!
|
||||
|
|
|
@ -28,7 +28,7 @@ pub fn cmd() cli.Command {
|
|||
]
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
if cmd.args.len < 3 {
|
||||
if !cmd.flags.get_bool('force')! {
|
||||
|
|
|
@ -54,7 +54,7 @@ pub fn cmd() cli.Command {
|
|||
]
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
mut filter := TargetFilter{}
|
||||
|
||||
|
@ -113,7 +113,7 @@ pub fn cmd() cli.Command {
|
|||
]
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
t := NewTarget{
|
||||
kind: cmd.flags.get_string('kind')!
|
||||
|
@ -135,7 +135,7 @@ pub fn cmd() cli.Command {
|
|||
description: 'Remove a target that matches the given id.'
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
remove(conf, cmd.args[0])!
|
||||
}
|
||||
|
@ -147,7 +147,7 @@ pub fn cmd() cli.Command {
|
|||
description: 'Show detailed information for the target matching the id.'
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
info(conf, cmd.args[0])!
|
||||
}
|
||||
|
@ -196,7 +196,7 @@ pub fn cmd() cli.Command {
|
|||
]
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
found := cmd.flags.get_all_found()
|
||||
|
||||
|
@ -235,7 +235,7 @@ pub fn cmd() cli.Command {
|
|||
]
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
remote := cmd.flags.get_bool('remote')!
|
||||
force := cmd.flags.get_bool('force')!
|
||||
|
@ -280,7 +280,7 @@ fn add(conf Config, t &NewTarget, raw bool) ! {
|
|||
if raw {
|
||||
println(target_id)
|
||||
} else {
|
||||
println('Target added with id $target_id')
|
||||
println('Target added with id ${target_id}')
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -296,7 +296,7 @@ fn patch(conf Config, id string, params map[string]string) ! {
|
|||
// invalid one to the server.
|
||||
if 'schedule' in params && params['schedule'] != '' {
|
||||
cron.parse_expression(params['schedule']) or {
|
||||
return error('Invalid cron expression: $err.msg()')
|
||||
return error('Invalid cron expression: ${err.msg()}')
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -25,7 +25,7 @@ fn test_not_allowed() {
|
|||
for exp in illegal_expressions {
|
||||
res = false
|
||||
parse_expression(exp) or { res = true }
|
||||
assert res, "'$exp' should produce an error"
|
||||
assert res, "'${exp}' should produce an error"
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -55,7 +55,7 @@ pub fn init(db_path string) !VieterDb {
|
|||
version_num := i + 1
|
||||
|
||||
// vfmt does not like these dots
|
||||
println('Applying migration $version_num' + '...')
|
||||
println('Applying migration ${version_num}' + '...')
|
||||
|
||||
// The sqlite library seems to not like it when multiple statements are
|
||||
// passed in a single exec. Therefore, we split them & run them all
|
||||
|
@ -64,7 +64,7 @@ pub fn init(db_path string) !VieterDb {
|
|||
res := conn.exec_none(part)
|
||||
|
||||
if res != sqlite.sqlite_done {
|
||||
return error('An error occurred while applying migration $version_num: SQLite error code $res')
|
||||
return error('An error occurred while applying migration ${version_num}: SQLite error code ${res}')
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -82,7 +82,7 @@ pub fn init(db_path string) !VieterDb {
|
|||
|
||||
// row_into<T> converts an sqlite.Row into a given type T by parsing each field
|
||||
// from a string according to its type.
|
||||
pub fn row_into<T>(row sqlite.Row) T {
|
||||
pub fn row_into[T](row sqlite.Row) T {
|
||||
mut i := 0
|
||||
mut out := T{}
|
||||
|
||||
|
|
|
@ -8,20 +8,20 @@ pub fn (db &VieterDb) get_build_logs(filter BuildLogFilter) []BuildLog {
|
|||
mut where_parts := []string{}
|
||||
|
||||
if filter.target != 0 {
|
||||
where_parts << 'target_id == $filter.target'
|
||||
where_parts << 'target_id == ${filter.target}'
|
||||
}
|
||||
|
||||
if filter.before != time.Time{} {
|
||||
where_parts << 'start_time < $filter.before.unix_time()'
|
||||
where_parts << 'start_time < ${filter.before.unix_time()}'
|
||||
}
|
||||
|
||||
if filter.after != time.Time{} {
|
||||
where_parts << 'start_time > $filter.after.unix_time()'
|
||||
where_parts << 'start_time > ${filter.after.unix_time()}'
|
||||
}
|
||||
|
||||
// NOTE: possible SQL injection
|
||||
if filter.arch != '' {
|
||||
where_parts << "arch == '$filter.arch'"
|
||||
where_parts << "arch == '${filter.arch}'"
|
||||
}
|
||||
|
||||
mut parts := []string{}
|
||||
|
@ -30,27 +30,27 @@ pub fn (db &VieterDb) get_build_logs(filter BuildLogFilter) []BuildLog {
|
|||
if exp[0] == `!` {
|
||||
code := exp[1..].int()
|
||||
|
||||
parts << 'exit_code != $code'
|
||||
parts << 'exit_code != ${code}'
|
||||
} else {
|
||||
code := exp.int()
|
||||
|
||||
parts << 'exit_code == $code'
|
||||
parts << 'exit_code == ${code}'
|
||||
}
|
||||
}
|
||||
|
||||
if parts.len > 0 {
|
||||
where_parts << parts.map('($it)').join(' or ')
|
||||
where_parts << parts.map('(${it})').join(' or ')
|
||||
}
|
||||
|
||||
mut where_str := ''
|
||||
|
||||
if where_parts.len > 0 {
|
||||
where_str = 'where ' + where_parts.map('($it)').join(' and ')
|
||||
where_str = 'where ' + where_parts.map('(${it})').join(' and ')
|
||||
}
|
||||
|
||||
query := 'select * from BuildLog $where_str limit $filter.limit offset $filter.offset'
|
||||
query := 'select * from BuildLog ${where_str} limit ${filter.limit} offset ${filter.offset}'
|
||||
rows, _ := db.conn.exec(query)
|
||||
res := rows.map(row_into<BuildLog>(it))
|
||||
res := rows.map(row_into[BuildLog](it))
|
||||
|
||||
return res
|
||||
}
|
||||
|
|
|
@ -49,13 +49,13 @@ pub fn (db &VieterDb) update_target(target_id int, params map[string]string) {
|
|||
if field.name in params {
|
||||
// Any fields that are array types require their own update method
|
||||
$if field.typ is string {
|
||||
values << "$field.name = '${params[field.name]}'"
|
||||
values << "${field.name} = '${params[field.name]}'"
|
||||
}
|
||||
}
|
||||
}
|
||||
values_str := values.join(', ')
|
||||
// I think this is actual SQL & not the ORM language
|
||||
query := 'update Target set $values_str where id == $target_id'
|
||||
query := 'update Target set ${values_str} where id == ${target_id}'
|
||||
|
||||
db.conn.exec_none(query)
|
||||
}
|
||||
|
|
|
@ -14,5 +14,5 @@ pub:
|
|||
|
||||
// str return a single-line string representation of a build log
|
||||
pub fn (c BuildConfig) str() string {
|
||||
return '{ target: $c.target_id, kind: $c.kind, url: $c.url, branch: $c.branch, path: $c.path, repo: $c.repo, base_image: $c.base_image, force: $c.force }'
|
||||
return '{ target: ${c.target_id}, kind: ${c.kind}, url: ${c.url}, branch: ${c.branch}, path: ${c.path}, repo: ${c.repo}, base_image: ${c.base_image}, force: ${c.force} }'
|
||||
}
|
||||
|
|
|
@ -16,13 +16,13 @@ pub mut:
|
|||
// str returns a string representation.
|
||||
pub fn (bl &BuildLog) str() string {
|
||||
mut parts := [
|
||||
'id: $bl.id',
|
||||
'target id: $bl.target_id',
|
||||
'start time: $bl.start_time.local()',
|
||||
'end time: $bl.end_time.local()',
|
||||
'id: ${bl.id}',
|
||||
'target id: ${bl.target_id}',
|
||||
'start time: ${bl.start_time.local()}',
|
||||
'end time: ${bl.end_time.local()}',
|
||||
'duration: ${bl.end_time - bl.start_time}',
|
||||
'arch: $bl.arch',
|
||||
'exit code: $bl.exit_code',
|
||||
'arch: ${bl.arch}',
|
||||
'exit code: ${bl.exit_code}',
|
||||
]
|
||||
str := parts.join('\n')
|
||||
|
||||
|
|
|
@ -4,17 +4,17 @@ import time
|
|||
|
||||
// from_params<T> creates a new instance of T from the given map by parsing all
|
||||
// of its fields from the map.
|
||||
pub fn from_params<T>(params map[string]string) ?T {
|
||||
pub fn from_params[T](params map[string]string) ?T {
|
||||
mut o := T{}
|
||||
|
||||
patch_from_params<T>(mut o, params)?
|
||||
patch_from_params[T](mut o, params)?
|
||||
|
||||
return o
|
||||
}
|
||||
|
||||
// patch_from_params<T> updates the given T object with the params defined in
|
||||
// the map.
|
||||
pub fn patch_from_params<T>(mut o T, params map[string]string) ? {
|
||||
pub fn patch_from_params[T](mut o T, params map[string]string) ? {
|
||||
$for field in T.fields {
|
||||
if field.name in params && params[field.name] != '' {
|
||||
$if field.typ is string {
|
||||
|
@ -37,7 +37,7 @@ pub fn patch_from_params<T>(mut o T, params map[string]string) ? {
|
|||
}
|
||||
|
||||
// params_from<T> converts a given T struct into a map of strings.
|
||||
pub fn params_from<T>(o &T) map[string]string {
|
||||
pub fn params_from[T](o &T) map[string]string {
|
||||
mut out := map[string]string{}
|
||||
|
||||
$for field in T.fields {
|
||||
|
|
|
@ -38,13 +38,13 @@ pub mut:
|
|||
// str returns a string representation.
|
||||
pub fn (t &Target) str() string {
|
||||
mut parts := [
|
||||
'id: $t.id',
|
||||
'kind: $t.kind',
|
||||
'url: $t.url',
|
||||
'branch: $t.branch',
|
||||
'path: $t.path',
|
||||
'repo: $t.repo',
|
||||
'schedule: $t.schedule',
|
||||
'id: ${t.id}',
|
||||
'kind: ${t.kind}',
|
||||
'url: ${t.url}',
|
||||
'branch: ${t.branch}',
|
||||
'path: ${t.path}',
|
||||
'repo: ${t.repo}',
|
||||
'schedule: ${t.schedule}',
|
||||
'arch: ${t.arch.map(it.value).join(', ')}',
|
||||
]
|
||||
str := parts.join('\n')
|
||||
|
|
|
@ -3,14 +3,14 @@ module package
|
|||
// format_entry returns a string properly formatted to be added to a desc file.
|
||||
[inline]
|
||||
fn format_entry(key string, value string) string {
|
||||
return '\n%$key%\n$value\n'
|
||||
return '\n%${key}%\n${value}\n'
|
||||
}
|
||||
|
||||
// full_name returns the properly formatted name for the package, including
|
||||
// version & architecture
|
||||
pub fn (pkg &Pkg) full_name() string {
|
||||
p := pkg.info
|
||||
return '$p.name-$p.version-$p.arch'
|
||||
return '${p.name}-${p.version}-${p.arch}'
|
||||
}
|
||||
|
||||
// filename returns the correct filename of the package file
|
||||
|
@ -20,10 +20,10 @@ pub fn (pkg &Pkg) filename() string {
|
|||
1 { '.tar.gz' }
|
||||
6 { '.tar.xz' }
|
||||
14 { '.tar.zst' }
|
||||
else { panic("Another compression code shouldn't be possible. Faulty code: $pkg.compression") }
|
||||
else { panic("Another compression code shouldn't be possible. Faulty code: ${pkg.compression}") }
|
||||
}
|
||||
|
||||
return '${pkg.full_name()}.pkg$ext'
|
||||
return '${pkg.full_name()}.pkg${ext}'
|
||||
}
|
||||
|
||||
// to_desc returns a desc file valid string representation
|
||||
|
@ -31,7 +31,7 @@ pub fn (pkg &Pkg) to_desc() !string {
|
|||
p := pkg.info
|
||||
|
||||
// filename
|
||||
mut desc := '%FILENAME%\n$pkg.filename()\n'
|
||||
mut desc := '%FILENAME%\n${pkg.filename()}\n'
|
||||
|
||||
desc += format_entry('NAME', p.name)
|
||||
desc += format_entry('BASE', p.base)
|
||||
|
@ -94,10 +94,10 @@ pub fn (pkg &Pkg) to_desc() !string {
|
|||
desc += format_entry('CHECKDEPENDS', p.checkdepends.join_lines())
|
||||
}
|
||||
|
||||
return '$desc\n'
|
||||
return '${desc}\n'
|
||||
}
|
||||
|
||||
// to_files returns a files file valid string representation
|
||||
pub fn (pkg &Pkg) to_files() string {
|
||||
return '%FILES%\n$pkg.files.join_lines()\n'
|
||||
return '%FILES%\n${pkg.files.join_lines()}\n'
|
||||
}
|
||||
|
|
|
@ -103,7 +103,7 @@ fn parse_pkg_info_string(pkg_info_str &string) !PkgInfo {
|
|||
// NOTE: this command only supports zstd-, xz- & gzip-compressed tarballs.
|
||||
pub fn read_pkg_archive(pkg_path string) !Pkg {
|
||||
if !os.is_file(pkg_path) {
|
||||
return error("'$pkg_path' doesn't exist or isn't a file.")
|
||||
return error("'${pkg_path}' doesn't exist or isn't a file.")
|
||||
}
|
||||
|
||||
a := C.archive_read_new()
|
||||
|
|
|
@ -31,11 +31,15 @@ pub:
|
|||
// new creates a new RepoGroupManager & creates the directories as needed
|
||||
pub fn new(repos_dir string, pkg_dir string, default_arch string) !RepoGroupManager {
|
||||
if !os.is_dir(repos_dir) {
|
||||
os.mkdir_all(repos_dir) or { return error('Failed to create repos directory: $err.msg()') }
|
||||
os.mkdir_all(repos_dir) or {
|
||||
return error('Failed to create repos directory: ${err.msg()}')
|
||||
}
|
||||
}
|
||||
|
||||
if !os.is_dir(pkg_dir) {
|
||||
os.mkdir_all(pkg_dir) or { return error('Failed to create package directory: $err.msg()') }
|
||||
os.mkdir_all(pkg_dir) or {
|
||||
return error('Failed to create package directory: ${err.msg()}')
|
||||
}
|
||||
}
|
||||
|
||||
return RepoGroupManager{
|
||||
|
@ -51,7 +55,7 @@ pub fn new(repos_dir string, pkg_dir string, default_arch string) !RepoGroupMana
|
|||
// the right subdirectories in r.pkg_dir if it was successfully added.
|
||||
pub fn (r &RepoGroupManager) add_pkg_from_path(repo string, pkg_path string) !RepoAddResult {
|
||||
pkg := package.read_pkg_archive(pkg_path) or {
|
||||
return error('Failed to read package file: $err.msg()')
|
||||
return error('Failed to read package file: ${err.msg()}')
|
||||
}
|
||||
|
||||
archs := r.add_pkg_in_repo(repo, pkg)!
|
||||
|
@ -129,7 +133,7 @@ fn (r &RepoGroupManager) add_pkg_in_repo(repo string, pkg &package.Pkg) ![]strin
|
|||
// files, and afterwards updates the db & files archives to reflect these
|
||||
// changes.
|
||||
fn (r &RepoGroupManager) add_pkg_in_arch_repo(repo string, arch string, pkg &package.Pkg) ! {
|
||||
pkg_dir := os.join_path(r.repos_dir, repo, arch, '$pkg.info.name-$pkg.info.version')
|
||||
pkg_dir := os.join_path(r.repos_dir, repo, arch, '${pkg.info.name}-${pkg.info.version}')
|
||||
|
||||
// Remove the previous version of the package, if present
|
||||
r.remove_pkg_from_arch_repo(repo, arch, pkg.info.name, false)!
|
||||
|
|
|
@ -13,7 +13,7 @@ import models { BuildLog, BuildLogFilter }
|
|||
// optionally be added to limit the list of build logs to that repository.
|
||||
['/api/v1/logs'; auth; get; markused]
|
||||
fn (mut app App) v1_get_logs() web.Result {
|
||||
filter := models.from_params<BuildLogFilter>(app.query) or {
|
||||
filter := models.from_params[BuildLogFilter](app.query) or {
|
||||
return app.json(.bad_request, new_response('Invalid query parameters.'))
|
||||
}
|
||||
logs := app.db.get_build_logs(filter)
|
||||
|
@ -101,7 +101,7 @@ fn (mut app App) v1_post_log() web.Result {
|
|||
// Create the logs directory of it doesn't exist
|
||||
if !os.exists(os.dir(log_file_path)) {
|
||||
os.mkdir_all(os.dir(log_file_path)) or {
|
||||
app.lerror('Error while creating log file: $err.msg()')
|
||||
app.lerror('Error while creating log file: ${err.msg()}')
|
||||
|
||||
return app.status(.internal_server_error)
|
||||
}
|
||||
|
@ -109,7 +109,7 @@ fn (mut app App) v1_post_log() web.Result {
|
|||
|
||||
if length := app.req.header.get(.content_length) {
|
||||
util.reader_to_file(mut app.reader, length.int(), log_file_path) or {
|
||||
app.lerror('An error occured while receiving logs: $err.msg()')
|
||||
app.lerror('An error occured while receiving logs: ${err.msg()}')
|
||||
|
||||
return app.status(.internal_server_error)
|
||||
}
|
||||
|
@ -127,7 +127,7 @@ fn (mut app App) v1_delete_log(id int) web.Result {
|
|||
full_path := os.join_path(app.conf.data_dir, logs_dir_name, log.path())
|
||||
|
||||
os.rm(full_path) or {
|
||||
app.lerror('Failed to remove log file $full_path: $err.msg()')
|
||||
app.lerror('Failed to remove log file ${full_path}: ${err.msg()}')
|
||||
|
||||
return app.status(.internal_server_error)
|
||||
}
|
||||
|
|
|
@ -8,7 +8,7 @@ import models { Target, TargetArch, TargetFilter }
|
|||
// v1_get_targets returns the current list of targets.
|
||||
['/api/v1/targets'; auth; get; markused]
|
||||
fn (mut app App) v1_get_targets() web.Result {
|
||||
filter := models.from_params<TargetFilter>(app.query) or {
|
||||
filter := models.from_params[TargetFilter](app.query) or {
|
||||
return app.json(.bad_request, new_response('Invalid query parameters.'))
|
||||
}
|
||||
mut iter := app.db.targets(filter)
|
||||
|
@ -35,7 +35,7 @@ fn (mut app App) v1_post_target() web.Result {
|
|||
params['arch'] = app.conf.default_arch
|
||||
}
|
||||
|
||||
mut new_target := models.from_params<Target>(params) or {
|
||||
mut new_target := models.from_params[Target](params) or {
|
||||
return app.json(.bad_request, new_response(err.msg()))
|
||||
}
|
||||
|
||||
|
|
|
@ -25,7 +25,7 @@ pub fn cmd() cli.Command {
|
|||
description: 'Start the Vieter server.'
|
||||
execute: fn (cmd cli.Command) ! {
|
||||
config_file := cmd.flags.get_string('config-file')!
|
||||
conf := vconf.load<Config>(prefix: 'VIETER_', default_path: config_file)!
|
||||
conf := vconf.load[Config](prefix: 'VIETER_', default_path: config_file)!
|
||||
|
||||
server(conf)!
|
||||
}
|
||||
|
|
|
@ -12,7 +12,7 @@ fn (mut app App) log_removal_daemon(schedule &cron.Expression) {
|
|||
for {
|
||||
mut too_old_timestamp := time.now().add_days(-app.conf.max_log_age)
|
||||
|
||||
app.linfo('Cleaning logs before $too_old_timestamp')
|
||||
app.linfo('Cleaning logs before ${too_old_timestamp}')
|
||||
|
||||
mut logs := []BuildLog{}
|
||||
mut counter := 0
|
||||
|
@ -29,7 +29,7 @@ fn (mut app App) log_removal_daemon(schedule &cron.Expression) {
|
|||
log_file_path := os.join_path(app.conf.data_dir, logs_dir_name, log.path())
|
||||
|
||||
os.rm(log_file_path) or {
|
||||
app.lerror('Failed to remove log file $log_file_path: $err.msg()')
|
||||
app.lerror('Failed to remove log file ${log_file_path}: ${err.msg()}')
|
||||
failed += 1
|
||||
|
||||
continue
|
||||
|
@ -44,7 +44,7 @@ fn (mut app App) log_removal_daemon(schedule &cron.Expression) {
|
|||
}
|
||||
}
|
||||
|
||||
app.linfo('Cleaned $counter logs ($failed failed)')
|
||||
app.linfo('Cleaned ${counter} logs (${failed} failed)')
|
||||
|
||||
// Sleep until the next cycle
|
||||
next_time := schedule.next_from_now()
|
||||
|
|
|
@ -26,7 +26,7 @@ fn (mut app App) get_repo_file(repo string, arch string, filename string) web.Re
|
|||
|
||||
// There's no point in having the ability to serve db archives with wrong
|
||||
// filenames
|
||||
if db_exts.any(filename == '$repo$it') {
|
||||
if db_exts.any(filename == '${repo}${it}') {
|
||||
full_path = os.join_path(app.repo.repos_dir, repo, arch, filename)
|
||||
|
||||
// repo-add does this using symlinks, but we just change the requested
|
||||
|
@ -62,19 +62,19 @@ fn (mut app App) put_package(repo string) web.Result {
|
|||
// Generate a random filename for the temp file
|
||||
pkg_path = os.join_path_single(app.repo.pkg_dir, rand.uuid_v4())
|
||||
|
||||
app.ldebug("Uploading $length bytes (${util.pretty_bytes(length.int())}) to '$pkg_path'.")
|
||||
app.ldebug("Uploading ${length} bytes (${util.pretty_bytes(length.int())}) to '${pkg_path}'.")
|
||||
|
||||
// This is used to time how long it takes to upload a file
|
||||
mut sw := time.new_stopwatch(time.StopWatchOptions{ auto_start: true })
|
||||
|
||||
util.reader_to_file(mut app.reader, length.int(), pkg_path) or {
|
||||
app.lwarn("Failed to upload '$pkg_path'")
|
||||
app.lwarn("Failed to upload '${pkg_path}'")
|
||||
|
||||
return app.status(.internal_server_error)
|
||||
}
|
||||
|
||||
sw.stop()
|
||||
app.ldebug("Upload of '$pkg_path' completed in ${sw.elapsed().seconds():.3}s.")
|
||||
app.ldebug("Upload of '${pkg_path}' completed in ${sw.elapsed().seconds():.3}s.")
|
||||
} else {
|
||||
app.lwarn('Tried to upload package without specifying a Content-Length.')
|
||||
|
||||
|
@ -83,14 +83,14 @@ fn (mut app App) put_package(repo string) web.Result {
|
|||
}
|
||||
|
||||
res := app.repo.add_pkg_from_path(repo, pkg_path) or {
|
||||
app.lerror('Error while adding package: $err.msg()')
|
||||
app.lerror('Error while adding package: ${err.msg()}')
|
||||
|
||||
os.rm(pkg_path) or { app.lerror("Failed to remove download '$pkg_path': $err.msg()") }
|
||||
os.rm(pkg_path) or { app.lerror("Failed to remove download '${pkg_path}': ${err.msg()}") }
|
||||
|
||||
return app.status(.internal_server_error)
|
||||
}
|
||||
|
||||
app.linfo("Added '$res.name-$res.version' to '$repo (${res.archs.join(',')})'.")
|
||||
app.linfo("Added '${res.name}-${res.version}' to '${repo} (${res.archs.join(',')})'.")
|
||||
|
||||
return app.json(.ok, new_data_response(res))
|
||||
}
|
||||
|
|
|
@ -6,17 +6,17 @@ import web
|
|||
['/:repo/:arch/:pkg'; auth; delete; markused]
|
||||
fn (mut app App) delete_package(repo string, arch string, pkg string) web.Result {
|
||||
res := app.repo.remove_pkg_from_arch_repo(repo, arch, pkg, true) or {
|
||||
app.lerror('Error while deleting package: $err.msg()')
|
||||
app.lerror('Error while deleting package: ${err.msg()}')
|
||||
|
||||
return app.status(.internal_server_error)
|
||||
}
|
||||
|
||||
if res {
|
||||
app.linfo("Removed package '$pkg' from '$repo/$arch'")
|
||||
app.linfo("Removed package '${pkg}' from '${repo}/${arch}'")
|
||||
|
||||
return app.status(.ok)
|
||||
} else {
|
||||
app.linfo("Tried removing package '$pkg' from '$repo/$arch', but it doesn't exist.")
|
||||
app.linfo("Tried removing package '${pkg}' from '${repo}/${arch}', but it doesn't exist.")
|
||||
|
||||
return app.status(.not_found)
|
||||
}
|
||||
|
@ -26,17 +26,17 @@ fn (mut app App) delete_package(repo string, arch string, pkg string) web.Result
|
|||
['/:repo/:arch'; auth; delete; markused]
|
||||
fn (mut app App) delete_arch_repo(repo string, arch string) web.Result {
|
||||
res := app.repo.remove_arch_repo(repo, arch) or {
|
||||
app.lerror('Error while deleting arch-repo: $err.msg()')
|
||||
app.lerror('Error while deleting arch-repo: ${err.msg()}')
|
||||
|
||||
return app.status(.internal_server_error)
|
||||
}
|
||||
|
||||
if res {
|
||||
app.linfo("Removed arch-repo '$repo/$arch'")
|
||||
app.linfo("Removed arch-repo '${repo}/${arch}'")
|
||||
|
||||
return app.status(.ok)
|
||||
} else {
|
||||
app.linfo("Tried removing '$repo/$arch', but it doesn't exist.")
|
||||
app.linfo("Tried removing '${repo}/${arch}', but it doesn't exist.")
|
||||
|
||||
return app.status(.not_found)
|
||||
}
|
||||
|
@ -46,17 +46,17 @@ fn (mut app App) delete_arch_repo(repo string, arch string) web.Result {
|
|||
['/:repo'; auth; delete; markused]
|
||||
fn (mut app App) delete_repo(repo string) web.Result {
|
||||
res := app.repo.remove_repo(repo) or {
|
||||
app.lerror('Error while deleting repo: $err.msg()')
|
||||
app.lerror('Error while deleting repo: ${err.msg()}')
|
||||
|
||||
return app.status(.internal_server_error)
|
||||
}
|
||||
|
||||
if res {
|
||||
app.linfo("Removed repo '$repo'")
|
||||
app.linfo("Removed repo '${repo}'")
|
||||
|
||||
return app.status(.ok)
|
||||
} else {
|
||||
app.linfo("Tried removing '$repo', but it doesn't exist.")
|
||||
app.linfo("Tried removing '${repo}', but it doesn't exist.")
|
||||
|
||||
return app.status(.not_found)
|
||||
}
|
||||
|
|
|
@ -44,11 +44,11 @@ pub fn server(conf Config) ! {
|
|||
}
|
||||
|
||||
global_ce := cron.parse_expression(conf.global_schedule) or {
|
||||
util.exit_with_message(1, 'Invalid global cron expression: $err.msg()')
|
||||
util.exit_with_message(1, 'Invalid global cron expression: ${err.msg()}')
|
||||
}
|
||||
|
||||
log_removal_ce := cron.parse_expression(conf.log_removal_schedule) or {
|
||||
util.exit_with_message(1, 'Invalid log removal cron expression: $err.msg()')
|
||||
util.exit_with_message(1, 'Invalid log removal cron expression: ${err.msg()}')
|
||||
}
|
||||
|
||||
// Configure logger
|
||||
|
@ -89,7 +89,7 @@ pub fn server(conf Config) ! {
|
|||
|
||||
db_file := os.join_path_single(conf.data_dir, server.db_file_name)
|
||||
db := db.init(db_file) or {
|
||||
util.exit_with_message(1, 'Failed to initialize database: $err.msg()')
|
||||
util.exit_with_message(1, 'Failed to initialize database: ${err.msg()}')
|
||||
}
|
||||
|
||||
mut collector := if conf.collect_metrics {
|
||||
|
@ -98,8 +98,8 @@ pub fn server(conf Config) ! {
|
|||
&metrics.MetricsCollector(metrics.new_null_collector())
|
||||
}
|
||||
|
||||
collector.histogram_buckets_set('http_requests_duration_seconds', [0.001, 0.005, 0.01, 0.05, 0.1, 0.5, 1, 5,
|
||||
10] )
|
||||
collector.histogram_buckets_set('http_requests_duration_seconds', [0.001, 0.005, 0.01, 0.05,
|
||||
0.1, 0.5, 1, 5, 10])
|
||||
|
||||
mut app := &App{
|
||||
logger: logger
|
||||
|
@ -111,11 +111,11 @@ pub fn server(conf Config) ! {
|
|||
job_queue: build.new_job_queue(global_ce, conf.base_image)
|
||||
}
|
||||
app.init_job_queue() or {
|
||||
util.exit_with_message(1, 'Failed to inialize job queue: $err.msg()')
|
||||
util.exit_with_message(1, 'Failed to inialize job queue: ${err.msg()}')
|
||||
}
|
||||
|
||||
if conf.max_log_age > 0 {
|
||||
go app.log_removal_daemon(log_removal_ce)
|
||||
spawn app.log_removal_daemon(log_removal_ce)
|
||||
}
|
||||
|
||||
web.run(app, conf.port)
|
||||
|
|
|
@ -51,7 +51,7 @@ pub fn reader_to_file(mut reader io.BufferedReader, length int, path string) ! {
|
|||
// match_array_in_array<T> returns how many elements of a2 overlap with a1. For
|
||||
// example, if a1 = "abcd" & a2 = "cd", the result will be 2. If the match is
|
||||
// not at the end of a1, the result is 0.
|
||||
pub fn match_array_in_array<T>(a1 []T, a2 []T) int {
|
||||
pub fn match_array_in_array[T](a1 []T, a2 []T) int {
|
||||
mut i := 0
|
||||
mut match_len := 0
|
||||
|
||||
|
|
|
@ -10,7 +10,7 @@ const attrs_to_ignore = ['auth', 'markused']
|
|||
// Parsing function attributes for methods and path.
|
||||
fn parse_attrs(name string, attrs []string) !([]http.Method, string) {
|
||||
if attrs.len == 0 {
|
||||
return [http.Method.get], '/$name'
|
||||
return [http.Method.get], '/${name}'
|
||||
}
|
||||
|
||||
mut x := attrs.clone()
|
||||
|
@ -45,7 +45,7 @@ fn parse_attrs(name string, attrs []string) !([]http.Method, string) {
|
|||
methods = [http.Method.get]
|
||||
}
|
||||
if path == '' {
|
||||
path = '/$name'
|
||||
path = '/${name}'
|
||||
}
|
||||
// Make path lowercase for case-insensitive comparisons
|
||||
return methods, path.to_lower()
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
module response
|
||||
|
||||
pub struct Response<T> {
|
||||
pub struct Response[T] {
|
||||
pub:
|
||||
message string
|
||||
data T
|
||||
|
@ -8,8 +8,8 @@ pub:
|
|||
|
||||
// new_response constructs a new Response<String> object with the given message
|
||||
// & an empty data field.
|
||||
pub fn new_response(message string) Response<string> {
|
||||
return Response<string>{
|
||||
pub fn new_response(message string) Response[string] {
|
||||
return Response[string]{
|
||||
message: message
|
||||
data: ''
|
||||
}
|
||||
|
@ -17,8 +17,8 @@ pub fn new_response(message string) Response<string> {
|
|||
|
||||
// new_data_response<T> constructs a new Response<T> object with the given data
|
||||
// & an empty message field.
|
||||
pub fn new_data_response<T>(data T) Response<T> {
|
||||
return Response<T>{
|
||||
pub fn new_data_response[T](data T) Response[T] {
|
||||
return Response[T]{
|
||||
message: ''
|
||||
data: data
|
||||
}
|
||||
|
@ -26,8 +26,8 @@ pub fn new_data_response<T>(data T) Response<T> {
|
|||
|
||||
// new_full_response<T> constructs a new Response<T> object with the given
|
||||
// message & data.
|
||||
pub fn new_full_response<T>(message string, data T) Response<T> {
|
||||
return Response<T>{
|
||||
pub fn new_full_response[T](message string, data T) Response[T] {
|
||||
return Response[T]{
|
||||
message: message
|
||||
data: data
|
||||
}
|
||||
|
|
|
@ -158,7 +158,7 @@ pub fn (mut ctx Context) body(status http.Status, content_type string, body stri
|
|||
}
|
||||
|
||||
// json<T> HTTP_OK with json_s as payload with content-type `application/json`
|
||||
pub fn (mut ctx Context) json<T>(status http.Status, j T) Result {
|
||||
pub fn (mut ctx Context) json[T](status http.Status, j T) Result {
|
||||
ctx.status = status
|
||||
ctx.content_type = 'application/json'
|
||||
|
||||
|
@ -278,14 +278,14 @@ interface DbInterface {
|
|||
|
||||
// run runs the app
|
||||
[manualfree]
|
||||
pub fn run<T>(global_app &T, port int) {
|
||||
mut l := net.listen_tcp(.ip6, ':$port') or { panic('failed to listen $err.code() $err') }
|
||||
pub fn run[T](global_app &T, port int) {
|
||||
mut l := net.listen_tcp(.ip6, ':${port}') or { panic('failed to listen ${err.code()} ${err}') }
|
||||
|
||||
// Parsing methods attributes
|
||||
mut routes := map[string]Route{}
|
||||
$for method in T.methods {
|
||||
http_methods, route_path := parse_attrs(method.name, method.attrs) or {
|
||||
eprintln('error parsing method attributes: $err')
|
||||
eprintln('error parsing method attributes: ${err}')
|
||||
return
|
||||
}
|
||||
|
||||
|
@ -294,7 +294,7 @@ pub fn run<T>(global_app &T, port int) {
|
|||
path: route_path
|
||||
}
|
||||
}
|
||||
println('[Vweb] Running app on http://localhost:$port')
|
||||
println('[Vweb] Running app on http://localhost:${port}')
|
||||
for {
|
||||
// Create a new app object for each connection, copy global data like db connections
|
||||
mut request_app := &T{}
|
||||
|
@ -311,16 +311,16 @@ pub fn run<T>(global_app &T, port int) {
|
|||
request_app.Context = global_app.Context // copy the context ref that contains static files map etc
|
||||
mut conn := l.accept() or {
|
||||
// failures should not panic
|
||||
eprintln('accept() failed with error: $err.msg()')
|
||||
eprintln('accept() failed with error: ${err.msg()}')
|
||||
continue
|
||||
}
|
||||
go handle_conn<T>(mut conn, mut request_app, routes)
|
||||
spawn handle_conn[T](mut conn, mut request_app, routes)
|
||||
}
|
||||
}
|
||||
|
||||
// handle_conn handles a connection
|
||||
[manualfree]
|
||||
fn handle_conn<T>(mut conn net.TcpConn, mut app T, routes map[string]Route) {
|
||||
fn handle_conn[T](mut conn net.TcpConn, mut app T, routes map[string]Route) {
|
||||
conn.set_read_timeout(30 * time.second)
|
||||
conn.set_write_timeout(30 * time.second)
|
||||
|
||||
|
@ -362,8 +362,8 @@ fn handle_conn<T>(mut conn net.TcpConn, mut app T, routes map[string]Route) {
|
|||
// Request parse
|
||||
head := http.parse_request_head(mut reader) or {
|
||||
// Prevents errors from being thrown when BufferedReader is empty
|
||||
if '$err' != 'none' {
|
||||
eprintln('error parsing request head: $err')
|
||||
if '${err}' != 'none' {
|
||||
eprintln('error parsing request head: ${err}')
|
||||
}
|
||||
return
|
||||
}
|
||||
|
@ -371,7 +371,7 @@ fn handle_conn<T>(mut conn net.TcpConn, mut app T, routes map[string]Route) {
|
|||
// The healthcheck spams the logs, which isn't very useful
|
||||
if head.url != '/health' {
|
||||
lock app.logger {
|
||||
app.logger.debug('$head.method $head.url $head.version')
|
||||
app.logger.debug('${head.method} ${head.url} ${head.version}')
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -385,7 +385,7 @@ fn handle_conn<T>(mut conn net.TcpConn, mut app T, routes map[string]Route) {
|
|||
|
||||
// URL Parse
|
||||
url := urllib.parse(head.url) or {
|
||||
eprintln('error parsing path: $err')
|
||||
eprintln('error parsing path: ${err}')
|
||||
return
|
||||
}
|
||||
|
||||
|
@ -423,7 +423,7 @@ fn handle_conn<T>(mut conn net.TcpConn, mut app T, routes map[string]Route) {
|
|||
$for method in T.methods {
|
||||
$if method.return_type is Result {
|
||||
route := routes[method.name] or {
|
||||
eprintln('parsed attributes for the `$method.name` are not found, skipping...')
|
||||
eprintln('parsed attributes for the `${method.name}` are not found, skipping...')
|
||||
Route{}
|
||||
}
|
||||
|
||||
|
@ -455,7 +455,7 @@ fn handle_conn<T>(mut conn net.TcpConn, mut app T, routes map[string]Route) {
|
|||
|
||||
method_args := params.clone()
|
||||
if method_args.len != method.args.len {
|
||||
eprintln('warning: uneven parameters count ($method.args.len) in `$method.name`, compared to the web route `$method.attrs` ($method_args.len)')
|
||||
eprintln('warning: uneven parameters count (${method.args.len}) in `${method.name}`, compared to the web route `${method.attrs}` (${method_args.len})')
|
||||
}
|
||||
app.$method(method_args)
|
||||
return
|
||||
|
|
Loading…
Reference in New Issue