Compare commits
24 Commits
Author | SHA1 | Date |
---|---|---|
Jef Roosens | 46fcbd8be0 | |
Jef Roosens | af2512034c | |
Jef Roosens | d47a3c71a1 | |
Jef Roosens | 12d8e5c04d | |
Jef Roosens | 4db13cdd6e | |
Jef Roosens | 7913b271dd | |
Jef Roosens | 0743085bac | |
Jef Roosens | ea8e49d55f | |
Jef Roosens | d31681843c | |
Jef Roosens | 5bc54c37f3 | |
Jef Roosens | 9258bb8146 | |
Jef Roosens | 6b578a80cd | |
Jef Roosens | 2db4afc226 | |
Jef Roosens | 7149c93b6f | |
Jef Roosens | c8191a19e1 | |
Jef Roosens | 9aebb3bedc | |
Jef Roosens | 598597f726 | |
Jef Roosens | 2dcee45406 | |
Jef Roosens | 20874bfb0b | |
Jef Roosens | f881122b48 | |
Jef Roosens | db0818024e | |
Jef Roosens | d4d567240d | |
Jef Roosens | b653f076d4 | |
Jef Roosens | 80946dd295 |
|
@ -1,13 +0,0 @@
|
|||
branches:
|
||||
exclude: [ main ]
|
||||
|
||||
platform: 'linux/amd64'
|
||||
|
||||
pipeline:
|
||||
lint:
|
||||
image: 'chewingbever/vlang:latest'
|
||||
pull: true
|
||||
commands:
|
||||
- make lint
|
||||
when:
|
||||
event: [ push ]
|
|
@ -0,0 +1,22 @@
|
|||
variables:
|
||||
- &vlang_image 'git.rustybever.be/chewing_bever/vlang:0.3.3-alpine3.17'
|
||||
|
||||
branches:
|
||||
exclude: [ main ]
|
||||
|
||||
platform: 'linux/amd64'
|
||||
|
||||
pipeline:
|
||||
build:
|
||||
image: *vlang_image
|
||||
commands:
|
||||
- make
|
||||
when:
|
||||
event: [ push ]
|
||||
|
||||
lint:
|
||||
image: *vlang_image
|
||||
commands:
|
||||
- make lint
|
||||
when:
|
||||
event: [ push ]
|
6
Makefile
6
Makefile
|
@ -1,6 +1,6 @@
|
|||
# =====CONFIG=====
|
||||
V_PATH ?= v
|
||||
V := $(V_PATH) -showcc
|
||||
V := $(V_PATH) -showcc -d use_openssl
|
||||
|
||||
all: vdocker
|
||||
|
||||
|
@ -10,6 +10,10 @@ all: vdocker
|
|||
vdocker:
|
||||
$(V) -g -shared .
|
||||
|
||||
.PHONY: c
|
||||
c:
|
||||
$(V) -o docker.c .
|
||||
|
||||
|
||||
# =====DOCS=====
|
||||
.PHONY: api-docs
|
||||
|
|
|
@ -20,7 +20,7 @@ reference](https://docs.docker.com/engine/api/v1.41/).
|
|||
- [ ] Start a container
|
||||
- [ ] Stop a container
|
||||
- [ ] Restart a container
|
||||
- [ ] Kill a container
|
||||
- [x] Kill a container
|
||||
- [ ] Update a container
|
||||
- [ ] Rename a container
|
||||
- [ ] Pause a container
|
||||
|
@ -42,7 +42,7 @@ reference](https://docs.docker.com/engine/api/v1.41/).
|
|||
- [ ] Inspect an image
|
||||
- [ ] Get the history of an image
|
||||
- [ ] Push an image
|
||||
- [ ] Tag an image
|
||||
- [x] Tag an image
|
||||
- [ ] Remove an image
|
||||
- [ ] Search images
|
||||
- [ ] Delete unused images
|
||||
|
|
95
containers.v
95
containers.v
|
@ -1,8 +1,6 @@
|
|||
module docker
|
||||
|
||||
import json
|
||||
import time
|
||||
import net.http { Method }
|
||||
import types { ContainerListItem }
|
||||
|
||||
[params]
|
||||
|
@ -13,12 +11,12 @@ pub struct ContainerListConfig {
|
|||
filters map[string][]string
|
||||
}
|
||||
|
||||
pub fn (mut d DockerConn) container_list(c ContainerListConfig) ?[]ContainerListItem {
|
||||
d.get('/containers/json')
|
||||
pub fn (mut d DockerConn) container_list(c ContainerListConfig) ![]ContainerListItem {
|
||||
d.request(.get, '/containers/json')
|
||||
d.params(c)
|
||||
d.send()?
|
||||
d.send()!
|
||||
|
||||
return d.read_json_response<[]ContainerListItem>()
|
||||
return d.read_json_response[[]ContainerListItem]()
|
||||
}
|
||||
|
||||
pub struct NewContainer {
|
||||
|
@ -36,31 +34,19 @@ pub:
|
|||
warnings []string [json: Warnings]
|
||||
}
|
||||
|
||||
pub fn (mut d DockerConn) container_create(c NewContainer) ?CreatedContainer {
|
||||
d.send_request_with_json(Method.post, '/containers/create', c)?
|
||||
head, res := d.read_response()?
|
||||
pub fn (mut d DockerConn) container_create(c NewContainer) !CreatedContainer {
|
||||
d.request(.post, '/containers/create')
|
||||
d.body_json(c)
|
||||
d.send()!
|
||||
|
||||
if head.status_code != 201 {
|
||||
data := json.decode(DockerError, res)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
|
||||
data := json.decode(CreatedContainer, res)?
|
||||
|
||||
return data
|
||||
return d.read_json_response[CreatedContainer]()
|
||||
}
|
||||
|
||||
// start_container starts the container with the given id.
|
||||
pub fn (mut d DockerConn) container_start(id string) ? {
|
||||
d.send_request(Method.post, '/containers/$id/start')?
|
||||
head, body := d.read_response()?
|
||||
|
||||
if head.status_code != 204 {
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
pub fn (mut d DockerConn) container_start(id string) ! {
|
||||
d.request(.post, '/containers/${id}/start')
|
||||
d.send()!
|
||||
d.read_response()!
|
||||
}
|
||||
|
||||
struct ContainerInspect {
|
||||
|
@ -81,50 +67,43 @@ pub mut:
|
|||
end_time time.Time [skip]
|
||||
}
|
||||
|
||||
pub fn (mut d DockerConn) container_inspect(id string) ?ContainerInspect {
|
||||
d.send_request(Method.get, '/containers/$id/json')?
|
||||
head, body := d.read_response()?
|
||||
pub fn (mut d DockerConn) container_inspect(id string) !ContainerInspect {
|
||||
d.request(.get, '/containers/${id}/json')
|
||||
d.send()!
|
||||
|
||||
if head.status_code != 200 {
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
|
||||
mut data := json.decode(ContainerInspect, body)?
|
||||
mut data := d.read_json_response[ContainerInspect]()!
|
||||
|
||||
// The Docker engine API *should* always return UTC time.
|
||||
data.state.start_time = time.parse_rfc3339(data.state.start_time_str)?
|
||||
data.state.start_time = time.parse_rfc3339(data.state.start_time_str)!
|
||||
|
||||
if data.state.status == 'exited' {
|
||||
data.state.end_time = time.parse_rfc3339(data.state.end_time_str)?
|
||||
data.state.end_time = time.parse_rfc3339(data.state.end_time_str)!
|
||||
}
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
pub fn (mut d DockerConn) container_remove(id string) ? {
|
||||
d.send_request(Method.delete, '/containers/$id')?
|
||||
head, body := d.read_response()?
|
||||
|
||||
if head.status_code != 204 {
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
pub fn (mut d DockerConn) container_remove(id string) ! {
|
||||
d.request(.delete, '/containers/${id}')
|
||||
d.send()!
|
||||
d.read_response()!
|
||||
}
|
||||
|
||||
pub fn (mut d DockerConn) container_get_logs(id string) ?&StreamFormatReader {
|
||||
d.send_request(Method.get, '/containers/$id/logs?stdout=true&stderr=true')?
|
||||
head := d.read_response_head()?
|
||||
pub fn (mut d DockerConn) container_kill(id string) ! {
|
||||
d.request(.post, '/containers/${id}/kill')
|
||||
d.send()!
|
||||
d.read_response()!
|
||||
}
|
||||
|
||||
if head.status_code != 200 {
|
||||
content_length := head.header.get(http.CommonHeader.content_length)?.int()
|
||||
body := d.read_response_body(content_length)?
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
pub fn (mut d DockerConn) container_get_logs(id string) !&StreamFormatReader {
|
||||
d.request(.get, '/containers/${id}/logs')
|
||||
d.params({
|
||||
'stdout': 'true'
|
||||
'stderr': 'true'
|
||||
})
|
||||
d.send()!
|
||||
d.read_response_head()!
|
||||
d.check_error()!
|
||||
|
||||
return d.get_stream_format_reader()
|
||||
}
|
||||
|
|
153
docker.v
153
docker.v
|
@ -4,7 +4,6 @@ import net.unix
|
|||
import io
|
||||
import net.http
|
||||
import strings
|
||||
import net.urllib
|
||||
import json
|
||||
import util
|
||||
|
||||
|
@ -27,12 +26,16 @@ mut:
|
|||
url string
|
||||
params map[string]string
|
||||
content_type string
|
||||
body string
|
||||
// Before send: body of the request
|
||||
// After send: body of response
|
||||
body string
|
||||
// HTTP head of the response
|
||||
head http.Response
|
||||
}
|
||||
|
||||
// new_conn creates a new connection to the Docker daemon.
|
||||
pub fn new_conn() ?&DockerConn {
|
||||
s := unix.connect_stream(docker.socket)?
|
||||
pub fn new_conn() !&DockerConn {
|
||||
s := unix.connect_stream(docker.socket)!
|
||||
|
||||
d := &DockerConn{
|
||||
socket: s
|
||||
|
@ -43,103 +46,75 @@ pub fn new_conn() ?&DockerConn {
|
|||
}
|
||||
|
||||
// close closes the underlying socket connection.
|
||||
pub fn (mut d DockerConn) close() ? {
|
||||
d.socket.close()?
|
||||
}
|
||||
|
||||
// send_request sends an HTTP request without body.
|
||||
fn (mut d DockerConn) send_request(method http.Method, url_str string) ? {
|
||||
url := urllib.parse('/$docker.api_version$url_str')?
|
||||
req := '$method $url.request_uri() HTTP/1.1\nHost: localhost\n\n'
|
||||
|
||||
d.socket.write_string(req)?
|
||||
|
||||
// When starting a new request, the reader needs to be reset.
|
||||
d.reader = io.new_buffered_reader(reader: d.socket)
|
||||
}
|
||||
|
||||
// send_request_with_body sends an HTTP request with the given body.
|
||||
fn (mut d DockerConn) send_request_with_body(method http.Method, url_str string, content_type string, body string) ? {
|
||||
url := urllib.parse('/$docker.api_version$url_str')?
|
||||
req := '$method $url.request_uri() HTTP/1.1\nHost: localhost\nContent-Type: $content_type\nContent-Length: $body.len\n\n$body\n\n'
|
||||
|
||||
d.socket.write_string(req)?
|
||||
|
||||
// When starting a new request, the reader needs to be reset.
|
||||
d.reader = io.new_buffered_reader(reader: d.socket)
|
||||
}
|
||||
|
||||
// send_request_with_json<T> is a convenience wrapper around
|
||||
// send_request_with_body that encodes the input as JSON.
|
||||
fn (mut d DockerConn) send_request_with_json<T>(method http.Method, url_str string, data &T) ? {
|
||||
body := json.encode(data)
|
||||
|
||||
return d.send_request_with_body(method, url_str, 'application/json', body)
|
||||
pub fn (mut d DockerConn) close() ! {
|
||||
d.socket.close()!
|
||||
}
|
||||
|
||||
// read_response_head consumes the socket's contents until it encounters
|
||||
// '\r\n\r\n', after which it parses the response as an HTTP response.
|
||||
// Importantly, this function never consumes the reader past the HTTP
|
||||
// separator, so the body can be read fully later on.
|
||||
fn (mut d DockerConn) read_response_head() ?http.Response {
|
||||
fn (mut d DockerConn) read_response_head() ! {
|
||||
mut res := []u8{}
|
||||
|
||||
util.read_until_separator(mut d.reader, mut res, docker.http_separator)?
|
||||
util.read_until_separator(mut d.reader, mut res, docker.http_separator)!
|
||||
|
||||
return http.parse_response(res.bytestr())
|
||||
d.head = http.parse_response(res.bytestr())!
|
||||
}
|
||||
|
||||
// read_response_body reads `length` bytes from the stream. It can be used when
|
||||
// the response encoding isn't chunked to fully read it.
|
||||
fn (mut d DockerConn) read_response_body(length int) ?string {
|
||||
if length == 0 {
|
||||
return ''
|
||||
// read_response_body consumes the rest of the HTTP response and stores it as
|
||||
// the response body. This function should only be called after
|
||||
// read_response_head. This function always reads the entire response into
|
||||
// memory, even if it's chunked.
|
||||
fn (mut d DockerConn) read_response_body() ! {
|
||||
if d.head.status() == .no_content {
|
||||
return
|
||||
}
|
||||
|
||||
mut buf := []u8{len: docker.buf_len}
|
||||
mut c := 0
|
||||
mut builder := strings.new_builder(docker.buf_len)
|
||||
|
||||
for builder.len < length {
|
||||
c = d.reader.read(mut buf) or { break }
|
||||
if d.head.header.get(.transfer_encoding) or { '' } == 'chunked' {
|
||||
mut body_stream := d.get_chunked_response_reader()
|
||||
|
||||
builder.write(buf[..c])?
|
||||
util.reader_to_writer(mut body_stream, mut builder)!
|
||||
} else {
|
||||
content_length := d.head.header.get(.content_length)!.int()
|
||||
|
||||
if content_length == 0 {
|
||||
d.body = ''
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
mut buf := []u8{len: docker.buf_len}
|
||||
mut c := 0
|
||||
|
||||
for builder.len < content_length {
|
||||
c = d.reader.read(mut buf)!
|
||||
|
||||
builder.write(buf[..c])!
|
||||
}
|
||||
}
|
||||
|
||||
return builder.str()
|
||||
d.body = builder.str()
|
||||
}
|
||||
|
||||
// read_response is a convenience function which always consumes the entire
|
||||
// response & returns it. It should only be used when we're certain that the
|
||||
// result isn't too large.
|
||||
fn (mut d DockerConn) read_response() ?(http.Response, string) {
|
||||
head := d.read_response_head()?
|
||||
|
||||
if head.header.get(http.CommonHeader.transfer_encoding) or { '' } == 'chunked' {
|
||||
mut builder := strings.new_builder(1024)
|
||||
mut body := d.get_chunked_response_reader()
|
||||
|
||||
util.reader_to_writer(mut body, mut builder)?
|
||||
|
||||
return head, builder.str()
|
||||
}
|
||||
|
||||
content_length := head.header.get(http.CommonHeader.content_length)?.int()
|
||||
res := d.read_response_body(content_length)?
|
||||
|
||||
return head, res
|
||||
// read_response is a convenience function that always consumes the entire
|
||||
// response and loads it into memory. It should only be used when we're certain
|
||||
// that the result isn't too large, as even chunked responses will get fully
|
||||
// loaded into memory.
|
||||
fn (mut d DockerConn) read_response() ! {
|
||||
d.read_response_head()!
|
||||
d.check_error()!
|
||||
d.read_response_body()!
|
||||
}
|
||||
|
||||
fn (mut d DockerConn) read_json_response<T>() ?T {
|
||||
head, body := d.read_response()?
|
||||
// read_json_response<T> is a convenience function that runs read_response
|
||||
// before parsing its contents, which is assumed to be JSON, into a struct.
|
||||
fn (mut d DockerConn) read_json_response[T]() !T {
|
||||
d.read_response()!
|
||||
|
||||
if head.status_code < 200 || head.status_code > 300 {
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return docker_error(head.status_code, data.message)
|
||||
}
|
||||
|
||||
mut data := json.decode(T, body)?
|
||||
data := json.decode(T, d.body)!
|
||||
|
||||
//$for field in T.fields {
|
||||
//$if field.typ is time.Time {
|
||||
|
@ -151,7 +126,7 @@ fn (mut d DockerConn) read_json_response<T>() ?T {
|
|||
}
|
||||
|
||||
// get_chunked_response_reader returns a ChunkedResponseReader using the socket
|
||||
// as reader.
|
||||
// as reader. This function should only be called after check_error.
|
||||
fn (mut d DockerConn) get_chunked_response_reader() &ChunkedResponseReader {
|
||||
r := new_chunked_response_reader(d.reader)
|
||||
|
||||
|
@ -159,10 +134,28 @@ fn (mut d DockerConn) get_chunked_response_reader() &ChunkedResponseReader {
|
|||
}
|
||||
|
||||
// get_stream_format_reader returns a StreamFormatReader using the socket as
|
||||
// reader.
|
||||
// reader. This function should only be called after check_error.
|
||||
fn (mut d DockerConn) get_stream_format_reader() &StreamFormatReader {
|
||||
r := new_chunked_response_reader(d.reader)
|
||||
r2 := new_stream_format_reader(r)
|
||||
|
||||
return r2
|
||||
}
|
||||
|
||||
struct DockerError {
|
||||
pub:
|
||||
message string
|
||||
}
|
||||
|
||||
// check_error should be called after read_response_head. If the status code of
|
||||
// the response is an error, the body is consumed and the Docker HTTP error is
|
||||
// returned as a V error. If the status isn't the error, this function is a
|
||||
// no-op, and the body can be read.
|
||||
fn (mut d DockerConn) check_error() ! {
|
||||
if d.head.status().is_error() {
|
||||
d.read_response_body()!
|
||||
d_err := json.decode(DockerError, d.body)!
|
||||
|
||||
return error_with_code('${d.head.status()}: ${d_err.message}', d.head.status_code)
|
||||
}
|
||||
}
|
||||
|
|
21
errors.v
21
errors.v
|
@ -1,21 +0,0 @@
|
|||
module docker
|
||||
|
||||
struct DockerError {
|
||||
status int [skip]
|
||||
message string
|
||||
}
|
||||
|
||||
fn (err DockerError) code() int {
|
||||
return err.status
|
||||
}
|
||||
|
||||
fn (err DockerError) msg() string {
|
||||
return err.message
|
||||
}
|
||||
|
||||
fn docker_error(status int, message string) IError {
|
||||
return IError(DockerError{
|
||||
status: status
|
||||
message: message
|
||||
})
|
||||
}
|
80
images.v
80
images.v
|
@ -1,25 +1,26 @@
|
|||
module docker
|
||||
|
||||
import net.http { Method }
|
||||
import json
|
||||
import types { Image }
|
||||
|
||||
struct Image {
|
||||
pub:
|
||||
id string [json: Id]
|
||||
pub fn (mut d DockerConn) image_inspect(image string) !Image {
|
||||
d.request(.get, '/images/${image}/json')
|
||||
d.send()!
|
||||
|
||||
data := d.read_json_response[Image]()!
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
// pull_image pulls the given image:tag.
|
||||
pub fn (mut d DockerConn) pull_image(image string, tag string) ? {
|
||||
d.send_request(Method.post, '/images/create?fromImage=$image&tag=$tag')?
|
||||
head := d.read_response_head()?
|
||||
|
||||
if head.status_code != 200 {
|
||||
content_length := head.header.get(http.CommonHeader.content_length)?.int()
|
||||
body := d.read_response_body(content_length)?
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
// image_pull pulls the given image:tag.
|
||||
pub fn (mut d DockerConn) image_pull(image string, tag string) ! {
|
||||
d.request(.post, '/images/create')
|
||||
d.params({
|
||||
'fromImage': image
|
||||
'tag': tag
|
||||
})
|
||||
d.send()!
|
||||
d.read_response_head()!
|
||||
d.check_error()!
|
||||
|
||||
// Keep reading the body until the pull has completed
|
||||
mut body := d.get_chunked_response_reader()
|
||||
|
@ -32,29 +33,32 @@ pub fn (mut d DockerConn) pull_image(image string, tag string) ? {
|
|||
}
|
||||
|
||||
// create_image_from_container creates a new image from a container.
|
||||
pub fn (mut d DockerConn) create_image_from_container(id string, repo string, tag string) ?Image {
|
||||
d.send_request(Method.post, '/commit?container=$id&repo=$repo&tag=$tag')?
|
||||
head, body := d.read_response()?
|
||||
pub fn (mut d DockerConn) image_from_container(id string, repo string, tag string) !Image {
|
||||
d.request(.post, '/commit')
|
||||
d.params({
|
||||
'container': id
|
||||
'repo': repo
|
||||
'tag': tag
|
||||
})
|
||||
d.body('application/json', '{}')
|
||||
d.send()!
|
||||
|
||||
if head.status_code != 201 {
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
|
||||
data := json.decode(Image, body)?
|
||||
|
||||
return data
|
||||
return d.read_json_response[Image]()!
|
||||
}
|
||||
|
||||
// remove_image removes the image with the given id.
|
||||
pub fn (mut d DockerConn) remove_image(id string) ? {
|
||||
d.send_request(Method.delete, '/images/$id')?
|
||||
head, body := d.read_response()?
|
||||
|
||||
if head.status_code != 200 {
|
||||
data := json.decode(DockerError, body)?
|
||||
|
||||
return error(data.message)
|
||||
}
|
||||
pub fn (mut d DockerConn) image_remove(id string) ! {
|
||||
d.request(.delete, '/images/${id}')
|
||||
d.send()!
|
||||
d.read_response()!
|
||||
}
|
||||
|
||||
pub fn (mut d DockerConn) image_tag(name string, repo string, tag string) ! {
|
||||
d.request(.post, '/images/${name}/tag')
|
||||
d.params({
|
||||
'repo': repo
|
||||
'tag': tag
|
||||
})
|
||||
d.send()!
|
||||
d.read_response()!
|
||||
}
|
||||
|
|
56
request.v
56
request.v
|
@ -3,48 +3,68 @@ module docker
|
|||
import net.http
|
||||
import net.urllib
|
||||
import io
|
||||
import json
|
||||
|
||||
fn (mut d DockerConn) request(method http.Method, url_str string) {
|
||||
fn (mut d DockerConn) request(method http.Method, url string) {
|
||||
d.method = method
|
||||
d.url = url_str
|
||||
d.params.clear()
|
||||
d.url = url
|
||||
d.content_type = ''
|
||||
d.body = ''
|
||||
|
||||
d.params.clear()
|
||||
}
|
||||
|
||||
fn (mut d DockerConn) get(url_str string) {
|
||||
d.request(http.Method.get, url_str)
|
||||
fn (mut d DockerConn) body(content_type string, body string) {
|
||||
d.content_type = content_type
|
||||
d.body = body
|
||||
}
|
||||
|
||||
fn (mut d DockerConn) params<T>(o T) {
|
||||
$for field in T.fields {
|
||||
v := o.$(field.name)
|
||||
fn (mut d DockerConn) body_json[T](data T) {
|
||||
d.content_type = 'application/json'
|
||||
d.body = json.encode(data)
|
||||
}
|
||||
|
||||
if !isnil(v) {
|
||||
d.params[field.name] = urllib.query_escape(v.str().replace("'", '"'))
|
||||
fn (mut d DockerConn) params[T](o T) {
|
||||
$if T is map[string]string {
|
||||
for key, value in o {
|
||||
d.params[key] = urllib.query_escape(value.replace("'", '"'))
|
||||
}
|
||||
} $else {
|
||||
$for field in T.fields {
|
||||
v := o.$(field.name)
|
||||
|
||||
if !isnil(v) {
|
||||
d.params[field.name] = urllib.query_escape(v.str().replace("'", '"'))
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn (mut d DockerConn) send() ? {
|
||||
mut full_url := d.url
|
||||
fn (mut d DockerConn) send() ! {
|
||||
mut full_url := '/${docker.api_version}${d.url}'
|
||||
|
||||
if d.params.len > 0 {
|
||||
params_str := d.params.keys().map('$it=${d.params[it]}').join('&')
|
||||
full_url += '?$params_str'
|
||||
mut fields := []string{cap: d.params.len}
|
||||
|
||||
for key, value in d.params {
|
||||
fields << '${key}=${value}'
|
||||
}
|
||||
params_str := fields.join('&')
|
||||
// params_str := d.params.keys().map('${it}=${d.params[it]}').join('&')
|
||||
full_url += '?${params_str}'
|
||||
}
|
||||
|
||||
// This is to make sure we actually created a valid URL
|
||||
parsed_url := urllib.parse(full_url)?
|
||||
parsed_url := urllib.parse(full_url)!
|
||||
final_url := parsed_url.request_uri()
|
||||
|
||||
req := if d.body == '' {
|
||||
'$d.method $final_url HTTP/1.1\nHost: localhost\n\n'
|
||||
'${d.method} ${final_url} HTTP/1.1\nHost: localhost\n\n'
|
||||
} else {
|
||||
'$d.method $final_url HTTP/1.1\nHost: localhost\nContent-Type: $d.content_type\nContent-Length: $d.body.len\n\n$d.body\n\n'
|
||||
'${d.method} ${final_url} HTTP/1.1\nHost: localhost\nContent-Type: ${d.content_type}\nContent-Length: ${d.body.len}\n\n${d.body}\n\n'
|
||||
}
|
||||
|
||||
d.socket.write_string(req)?
|
||||
d.socket.write_string(req)!
|
||||
|
||||
// When starting a new request, the reader needs to be reset.
|
||||
d.reader = io.new_buffered_reader(reader: d.socket)
|
||||
|
|
32
stream.v
32
stream.v
|
@ -25,11 +25,11 @@ pub fn new_chunked_response_reader(reader io.BufferedReader) &ChunkedResponseRea
|
|||
}
|
||||
|
||||
// read satisfies the io.Reader interface.
|
||||
pub fn (mut r ChunkedResponseReader) read(mut buf []u8) ?int {
|
||||
pub fn (mut r ChunkedResponseReader) read(mut buf []u8) !int {
|
||||
if r.bytes_left_in_chunk == 0 {
|
||||
// An io.BufferedReader always returns none if its stream has
|
||||
// ended.
|
||||
r.bytes_left_in_chunk = r.read_chunk_size()?
|
||||
r.bytes_left_in_chunk = r.read_chunk_size()!
|
||||
}
|
||||
|
||||
mut c := 0
|
||||
|
@ -37,9 +37,9 @@ pub fn (mut r ChunkedResponseReader) read(mut buf []u8) ?int {
|
|||
// Make sure we don't read more than we can safely read. This is to avoid
|
||||
// the underlying reader from becoming out of sync with our parsing:
|
||||
if buf.len > r.bytes_left_in_chunk {
|
||||
c = r.reader.read(mut buf[..r.bytes_left_in_chunk])?
|
||||
c = r.reader.read(mut buf[..r.bytes_left_in_chunk])!
|
||||
} else {
|
||||
c = r.reader.read(mut buf)?
|
||||
c = r.reader.read(mut buf)!
|
||||
}
|
||||
|
||||
r.bytes_left_in_chunk -= u64(c)
|
||||
|
@ -50,21 +50,21 @@ pub fn (mut r ChunkedResponseReader) read(mut buf []u8) ?int {
|
|||
// read_chunk_size advances the reader & reads the size of the next HTTP chunk.
|
||||
// This function should only be called if the previous chunk has been
|
||||
// completely consumed.
|
||||
fn (mut r ChunkedResponseReader) read_chunk_size() ?u64 {
|
||||
fn (mut r ChunkedResponseReader) read_chunk_size() !u64 {
|
||||
if r.started {
|
||||
mut buf := []u8{len: 2}
|
||||
|
||||
// Each chunk ends with a `\r\n` which we want to skip first
|
||||
r.reader.read(mut buf)?
|
||||
r.reader.read(mut buf)!
|
||||
}
|
||||
|
||||
r.started = true
|
||||
|
||||
mut res := []u8{}
|
||||
util.read_until_separator(mut r.reader, mut res, http_chunk_separator)?
|
||||
util.read_until_separator(mut r.reader, mut res, http_chunk_separator)!
|
||||
|
||||
// The length of the next chunk is provided as a hexadecimal
|
||||
mut num_data := hex.decode(res#[..-2].bytestr())?
|
||||
mut num_data := hex.decode(res#[..-2].bytestr())!
|
||||
|
||||
for num_data.len < 8 {
|
||||
num_data.insert(0, 0)
|
||||
|
@ -75,7 +75,7 @@ fn (mut r ChunkedResponseReader) read_chunk_size() ?u64 {
|
|||
// This only occurs for the very last chunk, which always reports a size of
|
||||
// 0.
|
||||
if num == 0 {
|
||||
return none
|
||||
return error('end of stream')
|
||||
}
|
||||
|
||||
return num
|
||||
|
@ -100,17 +100,17 @@ pub fn new_stream_format_reader(reader ChunkedResponseReader) &StreamFormatReade
|
|||
}
|
||||
|
||||
// read satisfies the io.Reader interface.
|
||||
pub fn (mut r StreamFormatReader) read(mut buf []u8) ?int {
|
||||
pub fn (mut r StreamFormatReader) read(mut buf []u8) !int {
|
||||
if r.bytes_left_in_chunk == 0 {
|
||||
r.bytes_left_in_chunk = r.read_chunk_size()?
|
||||
r.bytes_left_in_chunk = r.read_chunk_size()!
|
||||
}
|
||||
|
||||
mut c := 0
|
||||
|
||||
if buf.len > r.bytes_left_in_chunk {
|
||||
c = r.reader.read(mut buf[..r.bytes_left_in_chunk])?
|
||||
c = r.reader.read(mut buf[..r.bytes_left_in_chunk])!
|
||||
} else {
|
||||
c = r.reader.read(mut buf)?
|
||||
c = r.reader.read(mut buf)!
|
||||
}
|
||||
|
||||
r.bytes_left_in_chunk -= u32(c)
|
||||
|
@ -120,15 +120,15 @@ pub fn (mut r StreamFormatReader) read(mut buf []u8) ?int {
|
|||
|
||||
// read_chunk_size advances the reader & reads the header bytes for the length
|
||||
// of the next chunk.
|
||||
fn (mut r StreamFormatReader) read_chunk_size() ?u32 {
|
||||
fn (mut r StreamFormatReader) read_chunk_size() !u32 {
|
||||
mut buf := []u8{len: 8}
|
||||
|
||||
r.reader.read(mut buf)?
|
||||
r.reader.read(mut buf)!
|
||||
|
||||
num := binary.big_endian_u32(buf[4..])
|
||||
|
||||
if num == 0 {
|
||||
return none
|
||||
return error('end of stream')
|
||||
}
|
||||
|
||||
return num
|
||||
|
|
|
@ -75,37 +75,37 @@ pub:
|
|||
|
||||
pub struct HealthConfig {
|
||||
pub:
|
||||
test []string [json: Test]
|
||||
interval int [json: Interval]
|
||||
timeout int [json: Timeout]
|
||||
retries int [json: Retries]
|
||||
start_period int [json: StartPeriod]
|
||||
test []string [json: Test]
|
||||
interval int [json: Interval]
|
||||
timeout int [json: Timeout]
|
||||
retries int [json: Retries]
|
||||
start_period int [json: StartPeriod]
|
||||
}
|
||||
|
||||
pub struct ContainerCreate {
|
||||
pub:
|
||||
hostname string [json: Hostname]
|
||||
domain_name string [json: Domainname]
|
||||
user string [json: User]
|
||||
attach_stdin bool [json: AttachStdin]
|
||||
attach_stdout bool [json: AttachStderr] = true
|
||||
// ExposedPorts
|
||||
tty bool [json: Tty]
|
||||
open_stdin bool [json: OpenStdin]
|
||||
stdin_once bool [json: StdinOnce]
|
||||
env []string [json: Env]
|
||||
cmd []string [json: Cmd]
|
||||
healthcheck HealthConfig [json: Healthcheck]
|
||||
args_escaped bool [json: ArgsEscaped]
|
||||
image string [json: Image]
|
||||
// Volumes
|
||||
working_dir string [json: WorkingDir]
|
||||
entrypoint []string [json: Entrypoint]
|
||||
network_disabled bool [json: NetworkDisabled]
|
||||
mac_address string [json: MacAddress]
|
||||
on_build []string [json: OnBuild]
|
||||
labels map[string]string [json: Labels]
|
||||
stop_signal string [json: StopSignal]
|
||||
stop_timeout int [json: StopTimeout]
|
||||
shell []string [json: Shell]
|
||||
hostname string [json: Hostname]
|
||||
domain_name string [json: Domainname]
|
||||
user string [json: User]
|
||||
attach_stdin bool [json: AttachStdin]
|
||||
attach_stdout bool [json: AttachStderr] = true
|
||||
// ExposedPorts
|
||||
tty bool [json: Tty]
|
||||
open_stdin bool [json: OpenStdin]
|
||||
stdin_once bool [json: StdinOnce]
|
||||
env []string [json: Env]
|
||||
cmd []string [json: Cmd]
|
||||
healthcheck HealthConfig [json: Healthcheck]
|
||||
args_escaped bool [json: ArgsEscaped]
|
||||
image string [json: Image]
|
||||
// Volumes
|
||||
working_dir string [json: WorkingDir]
|
||||
entrypoint []string [json: Entrypoint]
|
||||
network_disabled bool [json: NetworkDisabled]
|
||||
mac_address string [json: MacAddress]
|
||||
on_build []string [json: OnBuild]
|
||||
labels map[string]string [json: Labels]
|
||||
stop_signal string [json: StopSignal]
|
||||
stop_timeout int [json: StopTimeout]
|
||||
shell []string [json: Shell]
|
||||
}
|
||||
|
|
|
@ -0,0 +1,6 @@
|
|||
module types
|
||||
|
||||
pub struct Image {
|
||||
pub:
|
||||
id string [json: Id]
|
||||
}
|
|
@ -0,0 +1,22 @@
|
|||
module types
|
||||
|
||||
import time
|
||||
|
||||
pub struct UsageData {
|
||||
size int [json: Size]
|
||||
ref_count int [json: RefCount]
|
||||
}
|
||||
|
||||
pub struct Volume {
|
||||
pub mut:
|
||||
created_at_str string [json: CreatedAt]
|
||||
created_at time.Time [skip]
|
||||
name string [json: Name]
|
||||
driver string [json: Driver]
|
||||
mountpoint string [json: Mountpoint]
|
||||
status map[string]string [json: Status]
|
||||
labels map[string]string [json: Labels]
|
||||
scope string [json: Scope]
|
||||
options map[string]string [json: Options]
|
||||
usage_data UsageData [json: UsageData]
|
||||
}
|
10
util/util.v
10
util/util.v
|
@ -3,7 +3,7 @@ module util
|
|||
import io
|
||||
|
||||
// reader_to_writer tries to consume the entire reader & write it to the writer.
|
||||
pub fn reader_to_writer(mut reader io.Reader, mut writer io.Writer) ? {
|
||||
pub fn reader_to_writer(mut reader io.Reader, mut writer io.Writer) ! {
|
||||
mut buf := []u8{len: 10 * 1024}
|
||||
|
||||
for {
|
||||
|
@ -21,7 +21,7 @@ pub fn reader_to_writer(mut reader io.Reader, mut writer io.Writer) ? {
|
|||
// match_array_in_array<T> returns how many elements of a2 overlap with a1. For
|
||||
// example, if a1 = "abcd" & a2 = "cd", the result will be 2. If the match is
|
||||
// not at the end of a1, the result is 0.
|
||||
pub fn match_array_in_array<T>(a1 []T, a2 []T) int {
|
||||
pub fn match_array_in_array[T](a1 []T, a2 []T) int {
|
||||
mut i := 0
|
||||
mut match_len := 0
|
||||
|
||||
|
@ -39,11 +39,11 @@ pub fn match_array_in_array<T>(a1 []T, a2 []T) int {
|
|||
|
||||
// read_until_separator consumes an io.Reader until it encounters some
|
||||
// separator array. The data read is stored inside the provided res array.
|
||||
pub fn read_until_separator(mut reader io.Reader, mut res []u8, sep []u8) ? {
|
||||
pub fn read_until_separator(mut reader io.Reader, mut res []u8, sep []u8) ! {
|
||||
mut buf := []u8{len: sep.len}
|
||||
|
||||
for {
|
||||
c := reader.read(mut buf)?
|
||||
c := reader.read(mut buf)!
|
||||
res << buf[..c]
|
||||
|
||||
match_len := match_array_in_array(buf[..c], sep)
|
||||
|
@ -54,7 +54,7 @@ pub fn read_until_separator(mut reader io.Reader, mut res []u8, sep []u8) ? {
|
|||
|
||||
if match_len > 0 {
|
||||
match_left := sep.len - match_len
|
||||
c2 := reader.read(mut buf[..match_left])?
|
||||
c2 := reader.read(mut buf[..match_left])!
|
||||
res << buf[..c2]
|
||||
|
||||
if buf[..c2] == sep[match_len..] {
|
||||
|
|
30
volumes.v
30
volumes.v
|
@ -1,26 +1,7 @@
|
|||
module docker
|
||||
|
||||
import net.http { Method }
|
||||
import time
|
||||
|
||||
struct UsageData {
|
||||
size int [json: Size]
|
||||
ref_count int [json: RefCount]
|
||||
}
|
||||
|
||||
struct Volume {
|
||||
created_at_str string [json: CreatedAt]
|
||||
pub mut:
|
||||
created_at time.Time [skip]
|
||||
name string [json: Name]
|
||||
driver string [json: Driver]
|
||||
mountpoint string [json: Mountpoint]
|
||||
status map[string]string [json: Status]
|
||||
labels map[string]string [json: Labels]
|
||||
scope string [json: Scope]
|
||||
options map[string]string [json: Options]
|
||||
usage_data UsageData [json: UsageData]
|
||||
}
|
||||
import types { Volume }
|
||||
|
||||
[params]
|
||||
pub struct VolumeListFilter {
|
||||
|
@ -35,13 +16,14 @@ struct VolumeListResponse {
|
|||
warnings []string [json: Warnings]
|
||||
}
|
||||
|
||||
pub fn (mut d DockerConn) volume_list() ?VolumeListResponse {
|
||||
d.send_request(Method.get, '/volumes')?
|
||||
pub fn (mut d DockerConn) volume_list() !VolumeListResponse {
|
||||
d.request(.get, '/volumes')
|
||||
d.send()!
|
||||
|
||||
mut data := d.read_json_response<VolumeListResponse>()?
|
||||
mut data := d.read_json_response[VolumeListResponse]()!
|
||||
|
||||
for mut vol in data.volumes {
|
||||
vol.created_at = time.parse_rfc3339(vol.created_at_str)?
|
||||
vol.created_at = time.parse_rfc3339(vol.created_at_str)!
|
||||
}
|
||||
|
||||
return data
|
||||
|
|
Loading…
Reference in New Issue