| // Copyright 2023 The Dawn & Tint Authors. |
| // |
| // Redistribution and use in source and binary forms, with or without |
| // modification, are permitted provided that the following conditions are met: |
| // |
| // 1. Redistributions of source code must retain the above copyright notice, this |
| // list of conditions and the following disclaimer. |
| // |
| // 2. Redistributions in binary form must reproduce the above copyright notice, |
| // this list of conditions and the following disclaimer in the documentation |
| // and/or other materials provided with the distribution. |
| // |
| // 3. Neither the name of the copyright holder nor the names of its |
| // contributors may be used to endorse or promote products derived from |
| // this software without specific prior written permission. |
| // |
| // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
| // AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| // IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE |
| // DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE |
| // FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL |
| // DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR |
| // SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER |
| // CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, |
| // OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| |
| // Package build implements a extensible build file list and module dependency |
| // generator for Tint. |
| // See: docs/tint/gen.md |
| package build |
| |
| import ( |
| "bytes" |
| "context" |
| "encoding/json" |
| "flag" |
| "fmt" |
| "log" |
| "os" |
| "os/exec" |
| "path" |
| "path/filepath" |
| "regexp" |
| "strings" |
| |
| "dawn.googlesource.com/dawn/tools/src/cmd/gen/common" |
| "dawn.googlesource.com/dawn/tools/src/cnf" |
| "dawn.googlesource.com/dawn/tools/src/container" |
| "dawn.googlesource.com/dawn/tools/src/fileutils" |
| "dawn.googlesource.com/dawn/tools/src/glob" |
| "dawn.googlesource.com/dawn/tools/src/match" |
| "dawn.googlesource.com/dawn/tools/src/template" |
| "dawn.googlesource.com/dawn/tools/src/transform" |
| "github.com/mzohreva/gographviz/graphviz" |
| "github.com/tidwall/jsonc" |
| ) |
| |
| const srcTint = "src/tint" |
| |
| func init() { |
| common.Register(&Cmd{}) |
| } |
| |
| type Cmd struct { |
| flags struct { |
| dot bool |
| } |
| } |
| |
| func (Cmd) Name() string { |
| return "build" |
| } |
| |
| func (Cmd) Desc() string { |
| return `build generates BUILD.* files in each of Tint's source directories` |
| } |
| |
| func (c *Cmd) RegisterFlags(ctx context.Context, cfg *common.Config) ([]string, error) { |
| flag.BoolVar(&c.flags.dot, "dot", false, "emit GraphViz DOT files for each target kind") |
| return nil, nil |
| } |
| |
| func (c Cmd) Run(ctx context.Context, cfg *common.Config) error { |
| p := NewProject(CanonicalizePath(path.Join(fileutils.DawnRoot(), srcTint)), cfg) |
| |
| for _, stage := range []struct { |
| desc string |
| fn func(p *Project) error |
| }{ |
| {"loading 'externals.json'", loadExternals}, |
| {"populating source files", populateSourceFiles}, |
| {"scanning source files", scanSourceFiles}, |
| {"loading directory configs", applyDirectoryConfigs}, |
| {"building dependencies", buildDependencies}, |
| {"checking for cycles", checkForCycles}, |
| {"emitting build files", emitBuildFiles}, |
| } { |
| if cfg.Flags.Verbose { |
| log.Printf("%v...\n", stage.desc) |
| } |
| if err := stage.fn(p); err != nil { |
| return err |
| } |
| } |
| |
| if c.flags.dot { |
| for _, kind := range AllTargetKinds { |
| if err := emitDotFile(p, kind); err != nil { |
| return err |
| } |
| } |
| } |
| |
| if cfg.Flags.Verbose { |
| log.Println("done") |
| } |
| return nil |
| } |
| |
| // loadExternals loads the 'externals.json' file in this directory. |
| func loadExternals(p *Project) error { |
| content, err := os.ReadFile(p.externalsJsonPath) |
| if err != nil { |
| return err |
| } |
| |
| externals := container.NewMap[string, struct { |
| IncludePatterns []string |
| Condition string |
| }]() |
| if err := json.Unmarshal(jsonc.ToJSON(content), &externals); err != nil { |
| return fmt.Errorf("failed to parse '%v': %w", p.externalsJsonPath, err) |
| } |
| |
| for _, name := range externals.Keys() { |
| external := externals[name] |
| |
| includePatternMatch := func(s string) bool { return false } |
| if len(external.IncludePatterns) > 0 { |
| matchers := []match.Test{} |
| for _, pattern := range external.IncludePatterns { |
| matcher, err := match.New(pattern) |
| if err != nil { |
| return fmt.Errorf("%v: matcher error: %w", p.externalsJsonPath, err) |
| } |
| matchers = append(matchers, matcher) |
| } |
| includePatternMatch = func(s string) bool { |
| for _, matcher := range matchers { |
| if matcher(s) { |
| return true |
| } |
| } |
| return false |
| } |
| } |
| |
| cond, err := cnf.Parse(external.Condition) |
| if err != nil { |
| return fmt.Errorf("%v: could not parse condition: %w", |
| p.externalsJsonPath, err) |
| } |
| |
| name := ExternalDependencyName(name) |
| p.externals.Add(name, ExternalDependency{ |
| Name: name, |
| Condition: cond, |
| includePatternMatch: includePatternMatch, |
| }) |
| } |
| |
| return nil |
| } |
| |
| // Globs all the source files, and creates populates the Project with Directory, Target and File. |
| // File name patterns are used to bin each file into a target for the directory. |
| func populateSourceFiles(p *Project) error { |
| paths, err := glob.Scan(p.Root, glob.MustParseConfig(`{ |
| "paths": [ |
| { |
| "include": [ |
| "*/**.cc", |
| "*/**.h", |
| "*/**.inl", |
| "*/**.mm" |
| ] |
| }, |
| { |
| "exclude": [ |
| "fuzzers/**" |
| ] |
| }] |
| }`)) |
| if err != nil { |
| return err |
| } |
| |
| for _, filepath := range paths { |
| filepath = CanonicalizePath(filepath) |
| dir, name := path.Split(filepath) |
| if kind := targetKindFromFilename(name); kind != targetInvalid { |
| directory := p.AddDirectory(dir) |
| p.AddTarget(directory, kind).AddSourceFile(p.AddFile(filepath)) |
| } |
| } |
| |
| return nil |
| } |
| |
| // scanSourceFiles scans all the source files for: |
| // * #includes to build a dependencies between targets |
| // * 'GEN_BUILD:' directives |
| func scanSourceFiles(p *Project) error { |
| // ParsedFile describes all the includes and conditions found in a source file |
| type ParsedFile struct { |
| removeFromProject bool |
| conditions []string |
| includes []Include |
| } |
| |
| // parseFile parses the source file at 'path' represented by 'file' |
| // As this is run concurrently, it must not modify any shared state (including file) |
| parseFile := func(path string, file *File) (string, *ParsedFile, error) { |
| conditions := []Condition{} |
| |
| body, err := os.ReadFile(file.AbsPath()) |
| if err != nil { |
| return path, nil, err |
| } |
| out := &ParsedFile{} |
| for i, line := range strings.Split(string(body), "\n") { |
| wrapErr := func(err error) error { |
| return fmt.Errorf("%v:%v %w", file.Path(), i+1, err) |
| } |
| if match := reIgnoreFile.FindStringSubmatch(line); len(match) > 0 { |
| out.removeFromProject = true |
| continue |
| } |
| if match := reIf.FindStringSubmatch(line); len(match) > 0 { |
| condition, err := cnf.Parse(strings.ToLower(match[1])) |
| if err != nil { |
| condition = Condition{{cnf.Unary{Var: "FAILED_TO_PARSE_CONDITION"}}} |
| } |
| if len(conditions) > 0 { |
| condition = cnf.And(condition, conditions[len(conditions)-1]) |
| } |
| conditions = append(conditions, condition) |
| } |
| if match := reIfdef.FindStringSubmatch(line); len(match) > 0 { |
| conditions = append(conditions, Condition{}) |
| } |
| if match := reIfndef.FindStringSubmatch(line); len(match) > 0 { |
| conditions = append(conditions, Condition{}) |
| } |
| if match := reElse.FindStringSubmatch(line); len(match) > 0 { |
| if len(conditions) == 0 { |
| return path, nil, wrapErr(fmt.Errorf("#else without #if")) |
| } |
| conditions[len(conditions)-1] = cnf.Not(conditions[len(conditions)-1]) |
| } |
| if match := reElif.FindStringSubmatch(line); len(match) > 0 { |
| condition, err := cnf.Parse(strings.ToLower(match[1])) |
| if err != nil { |
| condition = Condition{{cnf.Unary{Var: "FAILED_TO_PARSE_CONDITION"}}} |
| } |
| if len(conditions) == 0 { |
| return path, nil, wrapErr(fmt.Errorf("#elif without #if")) |
| } |
| conditions[len(conditions)-1] = cnf.And(cnf.Not(conditions[len(conditions)-1]), condition) |
| } |
| if match := reEndif.FindStringSubmatch(line); len(match) > 0 { |
| if len(conditions) == 0 { |
| return path, nil, wrapErr(fmt.Errorf("#endif without #if")) |
| } |
| conditions = conditions[:len(conditions)-1] |
| } |
| if match := reCondition.FindStringSubmatch(line); len(match) > 0 { |
| out.conditions = append(out.conditions, match[1]) |
| } |
| if !reIgnoreInclude.MatchString(line) { |
| for _, re := range []*regexp.Regexp{reInclude, reHashImport, reAtImport} { |
| if match := re.FindStringSubmatch(line); len(match) > 0 { |
| include := Include{ |
| Path: match[1], |
| Line: i + 1, |
| } |
| if len(conditions) > 0 { |
| include.Condition = conditions[len(conditions)-1] |
| } |
| out.includes = append(out.includes, include) |
| } |
| } |
| } |
| } |
| return path, out, nil |
| } |
| |
| // Create a new map by calling parseFile() on each entry of p.Files |
| // This is performed over multiple concurrent goroutines. |
| parsedFiles, err := transform.GoMap(p.Files, parseFile) |
| if err != nil { |
| return err |
| } |
| |
| // For each file, of each target, of each directory... |
| for _, dir := range p.Directories { |
| for _, target := range dir.Targets() { |
| for _, file := range target.SourceFiles() { |
| // Retrieve the parsed file information |
| parsed := parsedFiles[file.Path()] |
| |
| if parsed.removeFromProject { |
| file.RemoveFromProject() |
| continue |
| } |
| |
| // Apply any conditions |
| for _, condition := range parsed.conditions { |
| cond, err := cnf.Parse(condition) |
| if err != nil { |
| return fmt.Errorf("%v: could not parse condition: %w", file, err) |
| } |
| if file.Condition != nil { |
| cond = cnf.Optimize(cnf.And(file.Condition, cond)) |
| } |
| file.Condition = cond |
| } |
| |
| file.Includes = append(file.Includes, parsed.includes...) |
| } |
| } |
| } |
| return nil |
| } |
| |
| // applyDirectoryConfigs loads a 'BUILD.cfg' file in each source directory (if found), and |
| // applies the config to the Directory and/or Targets. |
| func applyDirectoryConfigs(p *Project) error { |
| // For each directory in the project... |
| for _, dir := range p.Directories.Values() { |
| path := path.Join(dir.AbsPath(), "BUILD.cfg") |
| content, err := os.ReadFile(path) |
| if err != nil { |
| continue |
| } |
| |
| // Parse the config |
| cfg := DirectoryConfig{} |
| if err := json.Unmarshal(jsonc.ToJSON(content), &cfg); err != nil { |
| return fmt.Errorf("error while parsing '%v': %w", path, err) |
| } |
| |
| // Apply any directory-level condition |
| for _, target := range dir.Targets() { |
| cond, err := cnf.Parse(cfg.Condition) |
| if err != nil { |
| return fmt.Errorf("%v: could not parse condition: %w", path, err) |
| } |
| target.Condition = cond |
| } |
| |
| // For each target config... |
| for _, tc := range []struct { |
| cfg *TargetConfig |
| kind TargetKind |
| }{ |
| {cfg.Lib, targetLib}, |
| {cfg.Test, targetTest}, |
| {cfg.TestCmd, targetTestCmd}, |
| {cfg.Bench, targetBench}, |
| {cfg.BenchCmd, targetBenchCmd}, |
| {cfg.Fuzz, targetFuzz}, |
| {cfg.FuzzCmd, targetFuzzCmd}, |
| {cfg.Cmd, targetCmd}, |
| } { |
| if tc.cfg == nil { |
| continue |
| } |
| target := p.Target(dir, tc.kind) |
| if target == nil { |
| return fmt.Errorf("%v: no files for target %v", path, tc.kind) |
| } |
| |
| // Apply any custom output name |
| target.OutputName = tc.cfg.OutputName |
| |
| if tc.cfg.Condition != "" { |
| condition, err := cnf.Parse(tc.cfg.Condition) |
| if err != nil { |
| return fmt.Errorf("%v: %v", path, err) |
| } |
| target.Condition = cnf.And(target.Condition, condition) |
| } |
| |
| // Add any additional internal dependencies |
| for _, depPattern := range tc.cfg.AdditionalDependencies.Internal { |
| match, err := match.New(depPattern) |
| if err != nil { |
| return fmt.Errorf("%v: invalid pattern for '%v'.AdditionalDependencies.Internal.'%v': %w", path, tc.kind, depPattern, err) |
| } |
| additionalDeps := []*Target{} |
| for _, target := range p.Targets.Keys() { |
| if match(string(target)) { |
| additionalDeps = append(additionalDeps, p.Targets[target]) |
| } |
| } |
| if len(additionalDeps) == 0 { |
| return fmt.Errorf("%v: '%v'.AdditionalDependencies.Internal.'%v' did not match any targets", path, tc.kind, depPattern) |
| } |
| for _, dep := range additionalDeps { |
| if dep != target { |
| target.Dependencies.AddInternal(dep) |
| } |
| } |
| } |
| // Add any additional internal dependencies |
| for _, name := range tc.cfg.AdditionalDependencies.External { |
| dep, ok := p.externals[name] |
| if !ok { |
| return fmt.Errorf("%v: external dependency '%v'.AdditionalDependencies.External.'%v' not declared in '%v'", |
| path, tc.kind, name, p.externalsJsonPath) |
| } |
| target.Dependencies.AddExternal(dep) |
| } |
| } |
| } |
| |
| return nil |
| } |
| |
| // buildDependencies walks all the #includes in all files, building the dependency information for |
| // all targets and files in the project. Errors if any cyclic includes are found. |
| func buildDependencies(p *Project) error { |
| type state int |
| const ( |
| unvisited state = iota |
| visiting |
| checked |
| ) |
| |
| cache := container.NewMap[string, state]() |
| |
| type FileInclude struct { |
| file string |
| inc Include |
| } |
| |
| var walk func(file *File, route []FileInclude) error |
| walk = func(file *File, route []FileInclude) error { |
| // Adds the dependency to the file and target's list of internal dependencies |
| addInternalDependency := func(dep *Target) { |
| file.TransitiveDependencies.AddInternal(dep) |
| if file.Target != dep { |
| file.Target.Dependencies.AddInternal(dep) |
| } |
| } |
| // Adds the dependency to the file and target's list of external dependencies |
| addExternalDependency := func(dep ExternalDependency) { |
| file.TransitiveDependencies.AddExternal(dep) |
| file.Target.Dependencies.AddExternal(dep) |
| } |
| |
| filePath := file.Path() |
| switch cache[filePath] { |
| case unvisited: |
| cache[filePath] = visiting |
| |
| for _, include := range file.Includes { |
| if strings.HasPrefix(include.Path, srcTint) { |
| // #include "src/tint/..." |
| path := include.Path[len(srcTint)+1:] // Strip 'src/tint/' |
| |
| includeFile := p.File(path) |
| if includeFile == nil { |
| return fmt.Errorf(`%v:%v includes non-existent file '%v'`, file.Path(), include.Line, path) |
| } |
| |
| if !isValidDependency(file.Target.Kind, includeFile.Target.Kind) { |
| return fmt.Errorf(`%v:%v %v target must not include %v target`, |
| file.Path(), include.Line, file.Target.Kind, includeFile.Target.Kind) |
| } |
| |
| addInternalDependency(includeFile.Target) |
| |
| // Gather the dependencies for the included file |
| if err := walk(includeFile, append(route, FileInclude{file: file.Path(), inc: include})); err != nil { |
| return err |
| } |
| |
| for _, dependency := range includeFile.TransitiveDependencies.Internal() { |
| addInternalDependency(dependency) |
| } |
| for _, dependency := range includeFile.TransitiveDependencies.External() { |
| addExternalDependency(dependency) |
| } |
| |
| noneIfEmpty := func(cond Condition) string { |
| if len(cond) == 0 { |
| return "<none>" |
| } |
| return cond.String() |
| } |
| sourceConditions := cnf.And(cnf.And(include.Condition, file.Condition), file.Target.Condition) |
| targetConditions := cnf.And(includeFile.Condition, includeFile.Target.Condition) |
| if missing := targetConditions.Remove(sourceConditions); len(missing) > 0 { |
| return fmt.Errorf(`%v:%v #include "%v" requires guard: #if %v |
| |
| %v build conditions: %v |
| %v build conditions: %v`, |
| file.Path(), include.Line, include.Path, strings.ToUpper(missing.String()), |
| file.Path(), noneIfEmpty(sourceConditions), |
| include.Path, targetConditions, |
| ) |
| } |
| |
| } else { |
| // Check for external includes |
| for _, external := range p.externals.Values() { |
| if external.includePatternMatch(include.Path) { |
| addExternalDependency(external) |
| } |
| } |
| } |
| |
| } |
| |
| cache[filePath] = checked |
| |
| case visiting: |
| err := strings.Builder{} |
| fmt.Fprintln(&err, "cyclic include found:") |
| for _, include := range route { |
| fmt.Fprintf(&err, " %v:%v includes '%v'\n", include.file, include.inc.Line, include.inc.Path) |
| } |
| return fmt.Errorf(err.String()) |
| } |
| return nil |
| } |
| |
| for _, file := range p.Files.Values() { |
| if err := walk(file, []FileInclude{}); err != nil { |
| return err |
| } |
| } |
| return nil |
| |
| } |
| |
| // checkForCycles ensures that the graph of target dependencies are acyclic (a DAG) |
| func checkForCycles(p *Project) error { |
| type state int |
| const ( |
| unvisited state = iota |
| visiting |
| checked |
| ) |
| |
| cache := container.NewMap[TargetName, state]() |
| |
| var walk func(t *Target, path []TargetName) error |
| walk = func(t *Target, path []TargetName) error { |
| switch cache[t.Name] { |
| case unvisited: |
| cache[t.Name] = visiting |
| for _, dep := range t.Dependencies.Internal() { |
| if err := walk(dep, append(path, dep.Name)); err != nil { |
| return err |
| } |
| } |
| cache[t.Name] = checked |
| case visiting: |
| err := strings.Builder{} |
| fmt.Fprintln(&err, "cyclic target dependency found:") |
| for _, t := range path { |
| fmt.Fprintln(&err, " ", string(t)) |
| } |
| fmt.Fprintln(&err, " ", string(t.Name)) |
| return fmt.Errorf(err.String()) |
| } |
| return nil |
| } |
| |
| for _, target := range p.Targets.Values() { |
| if err := walk(target, []TargetName{target.Name}); err != nil { |
| return err |
| } |
| } |
| return nil |
| } |
| |
| // emitBuildFiles emits a 'BUILD.*' file in each source directory for each |
| // 'BUILD.*.tmpl' found in this directory. |
| func emitBuildFiles(p *Project) error { |
| // Glob all the template files |
| templatePaths, err := glob.Glob(path.Join(fileutils.ThisDir(), "*.tmpl")) |
| if err != nil { |
| return err |
| } |
| if len(templatePaths) == 0 { |
| return fmt.Errorf("no template files found") |
| } |
| |
| // Load the templates |
| templates := container.NewMap[string, *template.Template]() |
| for _, path := range templatePaths { |
| tmpl, err := template.FromFile(path) |
| if err != nil { |
| return err |
| } |
| templates[path] = tmpl |
| } |
| |
| // process executes all the templates for the directory dir |
| // This is run concurrently, so must not modify shared state |
| process := func(dir *Directory) (common.StaleFiles, error) { |
| stale := common.StaleFiles{} |
| |
| // For each template... |
| for _, tmplPath := range templatePaths { |
| _, tmplName := filepath.Split(tmplPath) |
| outputName := strings.TrimSuffix(tmplName, ".tmpl") |
| outputPath := path.Join(dir.AbsPath(), outputName) |
| |
| // Attempt to read the existing output file |
| existing, err := os.ReadFile(outputPath) |
| if err != nil { |
| existing = nil |
| } |
| // If the file is annotated with a GEN_BUILD:DO_NOT_GENERATE directive, leave it alone |
| if reDoNotGenerate.Match(existing) { |
| continue |
| } |
| // Buffer for output |
| w := &bytes.Buffer{} |
| |
| // Write the header |
| relTmplPath, err := filepath.Rel(fileutils.DawnRoot(), tmplPath) |
| if err != nil { |
| return nil, err |
| } |
| w.WriteString(common.Header(string(existing), CanonicalizePath(relTmplPath), "#")) |
| |
| // Write the template output |
| err = templates[tmplPath].Run(w, dir, map[string]any{}) |
| if err != nil { |
| return nil, err |
| } |
| |
| // Format the output if it's a GN file. |
| if path.Ext(outputName) == ".gn" { |
| unformatted := w.String() |
| gn := exec.Command("gn", "format", "--stdin") |
| gn.Stdin = bytes.NewReader([]byte(unformatted)) |
| w.Reset() |
| gn.Stdout = w |
| gn.Stderr = w |
| if err := gn.Run(); err != nil { |
| return nil, fmt.Errorf("%v\ngn format failed: %w\n%v", unformatted, err, w.String()) |
| } |
| } |
| |
| if string(existing) != w.String() { |
| if !p.cfg.Flags.CheckStale { |
| if err := os.WriteFile(outputPath, w.Bytes(), 0666); err != nil { |
| return nil, err |
| } |
| } |
| |
| stale = append(stale, outputPath) |
| } |
| |
| } |
| |
| return stale, nil |
| } |
| |
| // Concurrently run process() on all the directories. |
| staleLists, err := transform.GoSlice(p.Directories.Values(), process) |
| if err != nil { |
| return err |
| } |
| |
| if p.cfg.Flags.Verbose || p.cfg.Flags.CheckStale { |
| // Collect all stale files into a flat list |
| stale := transform.Flatten(staleLists) |
| if p.cfg.Flags.CheckStale && len(stale) > 0 { |
| return stale |
| } |
| if p.cfg.Flags.Verbose { |
| log.Printf("generated %v files\n", len(stale)) |
| } |
| } |
| |
| return nil |
| } |
| |
| // emitDotFile writes a GraphViz DOT file visualizing the target dependency graph |
| func emitDotFile(p *Project, kind TargetKind) error { |
| g := graphviz.Graph{} |
| nodes := container.NewMap[TargetName, int]() |
| targets := []*Target{} |
| for _, target := range p.Targets.Values() { |
| if target.Kind == kind { |
| targets = append(targets, target) |
| } |
| } |
| for _, target := range targets { |
| nodes.Add(target.Name, g.AddNode(string(target.Name))) |
| } |
| for _, target := range targets { |
| for _, dep := range target.Dependencies.Internal() { |
| g.AddEdge(nodes[target.Name], nodes[dep.Name], "") |
| } |
| } |
| |
| g.MakeDirected() |
| |
| g.DefaultNodeAttribute(graphviz.Shape, graphviz.ShapeBox) |
| g.DefaultNodeAttribute(graphviz.FontName, "Courier") |
| g.DefaultNodeAttribute(graphviz.FontSize, "14") |
| g.DefaultNodeAttribute(graphviz.Style, graphviz.StyleFilled+","+graphviz.StyleRounded) |
| g.DefaultNodeAttribute(graphviz.FillColor, "yellow") |
| |
| g.DefaultEdgeAttribute(graphviz.FontName, "Courier") |
| g.DefaultEdgeAttribute(graphviz.FontSize, "12") |
| |
| file, err := os.Create(path.Join(p.Root, fmt.Sprintf("%v.dot", kind))) |
| if err != nil { |
| return err |
| } |
| defer file.Close() |
| |
| g.GenerateDOT(file) |
| return nil |
| } |
| |
| var ( |
| // Regular expressions used by this file |
| reIf = regexp.MustCompile(`\s*#\s*if\s+(.*)`) |
| reIfdef = regexp.MustCompile(`\s*#\s*ifdef\s+(.*)`) |
| reIfndef = regexp.MustCompile(`\s*#\s*ifndef\s+(.*)`) |
| reElse = regexp.MustCompile(`\s*#\s*else\s+(.*)`) |
| reElif = regexp.MustCompile(`\s*#\s*elif\s+(.*)`) |
| reEndif = regexp.MustCompile(`\s*#\s*endif`) |
| reInclude = regexp.MustCompile(`\s*#\s*include\s*(?:\"|<)([^(\"|>)]+)(?:\"|>)`) |
| reHashImport = regexp.MustCompile(`\s*#\s*import\s*\<([\w\/\.]+)\>`) |
| reAtImport = regexp.MustCompile(`\s*@\s*import\s*(\w+)\s*;`) |
| reIgnoreFile = regexp.MustCompile(`//\s*GEN_BUILD:IGNORE_FILE`) |
| reIgnoreInclude = regexp.MustCompile(`//\s*GEN_BUILD:IGNORE_INCLUDE`) |
| reCondition = regexp.MustCompile(`//\s*GEN_BUILD:CONDITION\((.*)\)\s*$`) |
| reDoNotGenerate = regexp.MustCompile(`#\s*GEN_BUILD:DO_NOT_GENERATE`) |
| ) |