v/vlib/v/doc/doc.v

517 lines
13 KiB
V
Raw Normal View History

2020-02-19 16:12:39 +01:00
module doc
import os
2020-04-25 20:58:00 +02:00
import strings
import time
import v.ast
import v.checker
import v.fmt
import v.parser
2020-04-25 20:58:00 +02:00
import v.pref
import v.scanner
2020-04-25 20:58:00 +02:00
import v.table
import v.token
import v.util
2020-02-19 16:12:39 +01:00
pub struct Doc {
pub mut:
2020-09-09 14:24:42 +02:00
input_path string
2020-06-02 16:19:55 +02:00
prefs &pref.Preferences = &pref.Preferences{}
table &table.Table = &table.Table{}
checker checker.Checker = checker.Checker{
table: 0
cur_fn: 0
pref: 0
}
2020-06-02 16:19:55 +02:00
pub_only bool = true
head DocNode
with_comments bool = true
contents []DocNode
fmt fmt.Fmt
time_generated time.Time
with_pos bool
filename string
pos int
is_vlib bool
2020-02-19 16:12:39 +01:00
}
pub struct DocPos {
pub:
line int
2020-06-02 16:19:55 +02:00
col int
len int
}
2020-02-26 22:44:29 +01:00
pub struct DocNode {
pub mut:
name string
2020-09-09 14:24:42 +02:00
content string
comment string
pos DocPos = DocPos{-1, -1, 0}
2020-09-09 14:24:42 +02:00
file_path string
attrs map[string]string
}
pub fn merge_comments(comments []ast.Comment) string {
mut res := []string{}
for comment in comments {
res << comment.text.trim_left('|')
}
return res.join('\n')
}
pub fn get_comment_block_right_before(comments []ast.Comment) string {
if comments.len == 0 {
return ''
}
mut comment := ''
mut last_comment_line_nr := 0
for i := comments.len - 1; i >= 0; i-- {
cmt := comments[i]
if last_comment_line_nr != 0 && cmt.pos.line_nr < last_comment_line_nr - 1 {
// skip comments that are not part of a continuous block,
// located right above the top level statement.
2020-09-09 14:24:42 +02:00
// break
}
2020-06-07 12:27:42 +02:00
mut cmt_content := cmt.text.trim_left('|')
if cmt_content.len == cmt.text.len || cmt.is_multi {
// ignore /* */ style comments for now
continue
2020-06-07 12:27:42 +02:00
// if cmt_content.len == 0 {
2020-09-09 14:24:42 +02:00
// continue
2020-06-07 12:27:42 +02:00
// }
// mut new_cmt_content := ''
// mut is_codeblock := false
// // println(cmt_content)
// lines := cmt_content.split_into_lines()
// for j, line in lines {
2020-09-09 14:24:42 +02:00
// trimmed := line.trim_space().trim_left(cmt_prefix)
// if trimmed.starts_with('- ') || (trimmed.len >= 2 && trimmed[0].is_digit() && trimmed[1] == `.`) || is_codeblock {
// new_cmt_content += line + '\n'
// } else if line.starts_with('```') {
// is_codeblock = !is_codeblock
// new_cmt_content += line + '\n'
// } else {
// new_cmt_content += trimmed + '\n'
// }
2020-06-07 12:27:42 +02:00
// }
// return new_cmt_content
2020-05-15 23:09:38 +02:00
}
// eprintln('cmt: $cmt')
cseparator := if cmt_content.starts_with('```') { '\n' } else { ' ' }
comment = cmt_content + cseparator + comment
last_comment_line_nr = cmt.pos.line_nr
2020-02-19 16:12:39 +01:00
}
return comment
2020-02-19 16:12:39 +01:00
}
fn convert_pos(file_path string, pos token.Position) DocPos {
2020-06-02 16:19:55 +02:00
source := util.read_file(file_path) or {
''
}
mut p := util.imax(0, util.imin(source.len - 1, pos.pos))
column := util.imax(0, pos.pos - p - 1)
return DocPos{
line: pos.line_nr + 1
col: util.imax(1, column + 1)
len: pos.len
2020-06-02 16:19:55 +02:00
}
2020-02-26 22:44:29 +01:00
}
pub fn (mut d Doc) get_signature(stmt ast.Stmt, file &ast.File) string {
match stmt {
ast.Module {
return 'module $stmt.name'
}
ast.FnDecl {
return stmt.stringify(d.table, d.fmt.cur_mod)
}
else {
d.fmt.out = strings.new_builder(1000)
d.fmt.stmt(stmt)
return d.fmt.out.str().trim_space()
}
}
2020-02-26 22:44:29 +01:00
}
pub fn (d Doc) get_pos(stmt ast.Stmt) token.Position {
match stmt {
ast.FnDecl { return stmt.pos }
ast.StructDecl { return stmt.pos }
ast.EnumDecl { return stmt.pos }
ast.InterfaceDecl { return stmt.pos }
ast.ConstDecl { return stmt.pos }
else { return token.Position{} }
2020-02-19 16:12:39 +01:00
}
}
2020-06-07 12:27:42 +02:00
pub fn (d Doc) get_type_name(decl ast.TypeDecl) string {
match decl {
ast.SumTypeDecl { return decl.name }
ast.FnTypeDecl { return decl.name }
ast.AliasTypeDecl { return decl.name }
2020-06-07 12:27:42 +02:00
}
}
pub fn (d Doc) get_name(stmt ast.Stmt) string {
match stmt {
ast.FnDecl { return stmt.name }
ast.StructDecl { return stmt.name }
ast.EnumDecl { return stmt.name }
ast.InterfaceDecl { return stmt.name }
ast.TypeDecl { return d.get_type_name(stmt) }
ast.ConstDecl { return 'Constants' }
else { return '' }
2020-02-19 16:12:39 +01:00
}
2020-02-26 22:44:29 +01:00
}
pub fn new_vdoc_preferences() &pref.Preferences {
// vdoc should be able to parse as much user code as possible
// so its preferences should be permissive:
2020-09-09 14:24:42 +02:00
return &pref.Preferences{
enable_globals: true
}
}
pub fn new(input_path string) Doc {
mut d := Doc{
2020-06-02 16:19:55 +02:00
input_path: os.real_path(input_path)
prefs: new_vdoc_preferences()
2020-06-02 16:19:55 +02:00
table: table.new_table()
head: DocNode{}
contents: []DocNode{}
time_generated: time.now()
}
d.fmt = fmt.Fmt{
indent: 0
is_debug: false
table: d.table
}
d.checker = checker.new_checker(d.table, d.prefs)
return d
}
pub fn (mut nodes []DocNode) sort_by_name() {
nodes.sort_with_compare(compare_nodes_by_name)
}
pub fn (mut nodes []DocNode) sort_by_category() {
nodes.sort_with_compare(compare_nodes_by_category)
}
fn compare_nodes_by_name(a, b &DocNode) int {
al := a.name.to_lower()
bl := b.name.to_lower()
return compare_strings(al, bl)
}
fn compare_nodes_by_category(a, b &DocNode) int {
al := a.attrs['category']
bl := b.attrs['category']
return compare_strings(al, bl)
2020-02-26 22:44:29 +01:00
}
pub fn (nodes []DocNode) index_by_name(node_name string) int {
2020-06-08 10:28:46 +02:00
for i, node in nodes {
if node.name != node_name {
continue
}
2020-06-08 10:28:46 +02:00
return i
}
return -1
2020-06-08 10:28:46 +02:00
}
pub fn (nodes []DocNode) find_children_of(parent string) []DocNode {
return nodes.find_nodes_with_attr('parent', parent)
}
pub fn (nodes []DocNode) find_nodes_with_attr(attr_name, value string) []DocNode {
mut subgroup := []DocNode{}
if attr_name.len == 0 {
return subgroup
2020-06-02 16:19:55 +02:00
}
for node in nodes {
if !node.attrs.exists(attr_name) || node.attrs[attr_name] != value {
2020-06-02 16:19:55 +02:00
continue
}
subgroup << node
}
subgroup.sort_by_name()
return subgroup
2020-02-19 16:12:39 +01:00
}
2020-03-10 19:49:04 +01:00
// get_parent_mod - return the parent mod name, in dot format.
// It works by climbing up the folder hierarchy, until a folder,
// that either contains main .v files, or a v.mod file is reached.
// For example, given something like /languages/v/vlib/x/websocket/tests/autobahn
// it returns `x.websocket.tests`, because /languages/v/ has v.mod file in it.
// NB: calling this is expensive, so keep the result, instead of recomputing it.
fn get_parent_mod(dir string) ?string {
2020-06-04 23:50:59 +02:00
$if windows {
// windows root path is C: or D:
if dir.len <= 2 {
return error('root folder reached')
}
2020-06-04 23:50:59 +02:00
} $else {
if dir.len == 0 {
return error('root folder reached')
}
2020-06-04 23:50:59 +02:00
}
2020-10-01 01:30:22 +02:00
base_dir := os.dir(dir)
fname_base_dir := os.file_name(base_dir)
prefs := new_vdoc_preferences()
fentries := os.ls(base_dir) or {
2020-06-02 16:19:55 +02:00
[]string{}
}
files := fentries.filter(!os.is_dir(it))
if 'v.mod' in files {
// the top level is reached, no point in climbing up further
return ''
}
v_files := prefs.should_compile_filtered_files(base_dir, files)
if v_files.len == 0 {
2020-06-02 16:19:55 +02:00
parent_mod := get_parent_mod(base_dir) or {
return fname_base_dir
2020-06-02 16:19:55 +02:00
}
if parent_mod.len > 0 {
return parent_mod + '.' + fname_base_dir
2020-03-10 19:49:04 +01:00
}
return error('No V files found.')
2020-03-10 19:49:04 +01:00
}
tbl := table.new_table()
scope := &ast.Scope{
parent: 0
}
file_ast := parser.parse_file(v_files[0], tbl, .skip_comments, prefs, scope)
if file_ast.mod.name == 'main' {
return ''
}
2020-06-02 16:19:55 +02:00
parent_mod := get_parent_mod(base_dir) or {
return fname_base_dir
2020-06-02 16:19:55 +02:00
}
if parent_mod.len > 0 {
return parent_mod + '.' + file_ast.mod.name
}
return file_ast.mod.name
2020-03-10 19:49:04 +01:00
}
fn (mut d Doc) generate_from_ast(file_ast ast.File, orig_mod_name string) {
mut const_idx := -1
stmts := file_ast.stmts
d.fmt.file = file_ast
d.fmt.set_current_module_name(orig_mod_name)
d.fmt.process_file_imports(file_ast)
mut last_import_stmt_idx := 0
for sidx, stmt in stmts {
if stmt is ast.Import {
last_import_stmt_idx = sidx
}
}
mut prev_comments := []ast.Comment{}
mut imports_section := true
for sidx, stmt in stmts {
// eprintln('stmt typeof: ' + typeof(stmt))
if stmt is ast.ExprStmt {
if stmt.expr is ast.Comment as cmt {
prev_comments << cmt
continue
}
}
// TODO: Fetch head comment once
if stmt is ast.Module {
// the previous comments were probably a copyright/license one
module_comment := get_comment_block_right_before(prev_comments)
prev_comments = []
if !d.is_vlib && !module_comment.starts_with('Copyright (c)') {
if module_comment in ['', d.head.comment] {
continue
}
if d.head.comment != '' {
d.head.comment += '\n'
}
d.head.comment += module_comment
}
continue
}
if last_import_stmt_idx > 0 && sidx == last_import_stmt_idx {
// the accumulated comments were interspersed before/between the imports;
// just add them all to the module comment:
import_comments := merge_comments(prev_comments)
if d.head.comment != '' {
d.head.comment += '\n'
}
d.head.comment += import_comments
prev_comments = []
imports_section = false
}
if stmt is ast.Import {
continue
}
signature := d.get_signature(stmt, file_ast)
pos := d.get_pos(stmt)
mut name := d.get_name(stmt)
if (!signature.starts_with('pub') && d.pub_only) || stmt is ast.GlobalDecl {
prev_comments = []
continue
}
if name.starts_with(orig_mod_name + '.') {
name = name.all_after(orig_mod_name + '.')
}
mut node := DocNode{
name: name
content: signature
comment: ''
pos: convert_pos(file_ast.path, pos)
file_path: file_ast.path
}
if node.name.len == 0 && node.comment.len == 0 && node.content.len == 0 {
continue
}
if stmt is ast.FnDecl {
if stmt.is_deprecated {
continue
}
if stmt.receiver.typ != 0 {
node.attrs['parent'] = d.fmt.type_to_str(stmt.receiver.typ).trim_left('&')
p_idx := d.contents.index_by_name(node.attrs['parent'])
if p_idx == -1 && node.attrs['parent'] != 'void' {
d.contents << DocNode{
name: node.attrs['parent']
content: ''
comment: ''
attrs: {
'category': 'Structs'
}
}
}
}
}
if stmt is ast.ConstDecl {
if const_idx == -1 {
const_idx = sidx
} else {
node.attrs['parent'] = 'Constants'
}
}
match stmt {
ast.ConstDecl { node.attrs['category'] = 'Constants' }
ast.EnumDecl { node.attrs['category'] = 'Enums' }
ast.InterfaceDecl { node.attrs['category'] = 'Interfaces' }
ast.StructDecl { node.attrs['category'] = 'Structs' }
ast.TypeDecl { node.attrs['category'] = 'Typedefs' }
ast.FnDecl { node.attrs['category'] = if node.attrs['parent'] in ['void', ''] ||
!node.attrs.exists('parent') { 'Functions' } else { 'Methods' } }
else {}
}
d.contents << node
if d.with_comments && (prev_comments.len > 0) {
last_comment := d.contents[d.contents.len - 1].comment
cmt := last_comment + '\n' + get_comment_block_right_before(prev_comments)
d.contents[d.contents.len - 1].comment = cmt
}
prev_comments = []
}
}
fn (mut d Doc) expr_typ_to_string(ex ast.Expr) string {
expr_typ := d.checker.expr(ex)
return d.fmt.type_to_str(expr_typ)
}
fn (mut d Doc) generate() ?Doc {
// get all files
2020-10-01 01:30:22 +02:00
base_path := if os.is_dir(d.input_path) { d.input_path } else { os.real_path(os.dir(d.input_path)) }
d.is_vlib = 'vlib' !in base_path
2020-06-02 16:19:55 +02:00
project_files := os.ls(base_path) or {
return error_with_code(err, 0)
2020-06-02 16:19:55 +02:00
}
v_files := d.prefs.should_compile_filtered_files(base_path, project_files)
if v_files.len == 0 {
return error_with_code('vdoc: No valid V files were found.', 1)
}
// parse files
mut file_asts := []ast.File{}
// TODO: remove later for vlib
comments_mode := if d.with_comments { scanner.CommentsMode.toplevel_comments } else { scanner.CommentsMode.skip_comments }
mut fname_has_set := false
for file in v_files {
2020-06-02 16:19:55 +02:00
file_ast := parser.parse_file(file, d.table, comments_mode, d.prefs, &ast.Scope{
parent: 0
})
if d.filename.len > 0 && d.filename in file && !fname_has_set {
d.filename = file
fname_has_set = true
}
file_asts << file_ast
}
mut module_name := ''
mut parent_mod_name := ''
mut orig_mod_name := ''
for i, file_ast in file_asts {
d.checker.check(file_ast)
if i == 0 {
2020-06-02 16:19:55 +02:00
parent_mod_name = get_parent_mod(base_path) or {
''
}
module_name = file_ast.mod.name
orig_mod_name = module_name
if module_name != 'main' && parent_mod_name.len > 0 {
module_name = parent_mod_name + '.' + module_name
}
d.head = DocNode{
2020-06-02 16:19:55 +02:00
name: module_name
content: 'module $module_name'
comment: ''
}
} else if file_ast.mod.name != orig_mod_name {
2020-03-10 19:49:04 +01:00
continue
}
d.generate_from_ast(file_ast, orig_mod_name)
if file_ast.path == d.filename {
lscope := file_ast.scope.innermost(d.pos)
for name, val in lscope.objects {
if val !is ast.Var {
continue
}
vr_data := val as ast.Var
vr_expr := vr_data.expr
l_node := DocNode{
name: name
content: ''
comment: ''
pos: convert_pos(file_ast.path, vr_data.pos)
file_path: file_ast.path
attrs: {
'category': 'Variable'
'return_type': d.expr_typ_to_string(vr_expr)
'local': 'true'
2020-06-10 11:14:03 +02:00
}
}
d.contents << l_node
}
2020-03-10 19:49:04 +01:00
}
d.fmt.mod2alias = map[string]string{}
2020-03-10 19:49:04 +01:00
}
d.time_generated = time.now()
d.contents.sort_by_name()
d.contents.sort_by_category()
return *d
2020-03-10 19:49:04 +01:00
}
pub fn generate_from_pos(input_path, filename string, pos int) ?Doc {
mut doc := new(input_path)
doc.pub_only = false
doc.with_comments = true
doc.with_pos = true
doc.filename = filename
doc.pos = pos
return doc.generate()
}
2020-06-02 16:19:55 +02:00
pub fn generate(input_path string, pub_only, with_comments bool) ?Doc {
mut doc := new(input_path)
doc.pub_only = pub_only
doc.with_comments = with_comments
return doc.generate()
2020-06-02 16:18:12 +02:00
}