| // Copyright 2011 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| // Package load loads packages. |
| package load |
| |
| import ( |
| "bytes" |
| "crypto/sha1" |
| "fmt" |
| "go/build" |
| "go/token" |
| "io/ioutil" |
| "os" |
| pathpkg "path" |
| "path/filepath" |
| "runtime" |
| "sort" |
| "strconv" |
| "strings" |
| "unicode" |
| "unicode/utf8" |
| |
| "cmd/go/internal/base" |
| "cmd/go/internal/buildid" |
| "cmd/go/internal/cfg" |
| "cmd/go/internal/str" |
| ) |
| |
| var IgnoreImports bool // control whether we ignore imports in packages |
| |
| // A Package describes a single package found in a directory. |
| type Package struct { |
| PackagePublic // visible in 'go list' |
| Internal PackageInternal // for use inside go command only |
| } |
| |
| type PackagePublic struct { |
| // Note: These fields are part of the go command's public API. |
| // See list.go. It is okay to add fields, but not to change or |
| // remove existing ones. Keep in sync with list.go |
| Dir string `json:",omitempty"` // directory containing package sources |
| ImportPath string `json:",omitempty"` // import path of package in dir |
| ImportComment string `json:",omitempty"` // path in import comment on package statement |
| Name string `json:",omitempty"` // package name |
| Doc string `json:",omitempty"` // package documentation string |
| Target string `json:",omitempty"` // install path |
| Shlib string `json:",omitempty"` // the shared library that contains this package (only set when -linkshared) |
| Goroot bool `json:",omitempty"` // is this package found in the Go root? |
| Standard bool `json:",omitempty"` // is this package part of the standard Go library? |
| Stale bool `json:",omitempty"` // would 'go install' do anything for this package? |
| StaleReason string `json:",omitempty"` // why is Stale true? |
| Root string `json:",omitempty"` // Go root or Go path dir containing this package |
| ConflictDir string `json:",omitempty"` // Dir is hidden by this other directory |
| BinaryOnly bool `json:",omitempty"` // package cannot be recompiled |
| |
| // Source files |
| // If you add to this list you MUST add to p.AllFiles (below) too. |
| // Otherwise file name security lists will not apply to any new additions. |
| GoFiles []string `json:",omitempty"` // .go source files (excluding CgoFiles, TestGoFiles, XTestGoFiles) |
| CgoFiles []string `json:",omitempty"` // .go sources files that import "C" |
| IgnoredGoFiles []string `json:",omitempty"` // .go sources ignored due to build constraints |
| CFiles []string `json:",omitempty"` // .c source files |
| CXXFiles []string `json:",omitempty"` // .cc, .cpp and .cxx source files |
| MFiles []string `json:",omitempty"` // .m source files |
| HFiles []string `json:",omitempty"` // .h, .hh, .hpp and .hxx source files |
| FFiles []string `json:",omitempty"` // .f, .F, .for and .f90 Fortran source files |
| SFiles []string `json:",omitempty"` // .s source files |
| SwigFiles []string `json:",omitempty"` // .swig files |
| SwigCXXFiles []string `json:",omitempty"` // .swigcxx files |
| SysoFiles []string `json:",omitempty"` // .syso system object files added to package |
| |
| // Cgo directives |
| CgoCFLAGS []string `json:",omitempty"` // cgo: flags for C compiler |
| CgoCPPFLAGS []string `json:",omitempty"` // cgo: flags for C preprocessor |
| CgoCXXFLAGS []string `json:",omitempty"` // cgo: flags for C++ compiler |
| CgoFFLAGS []string `json:",omitempty"` // cgo: flags for Fortran compiler |
| CgoLDFLAGS []string `json:",omitempty"` // cgo: flags for linker |
| CgoPkgConfig []string `json:",omitempty"` // cgo: pkg-config names |
| |
| // Dependency information |
| Imports []string `json:",omitempty"` // import paths used by this package |
| Deps []string `json:",omitempty"` // all (recursively) imported dependencies |
| |
| // Error information |
| Incomplete bool `json:",omitempty"` // was there an error loading this package or dependencies? |
| Error *PackageError `json:",omitempty"` // error loading this package (not dependencies) |
| DepsErrors []*PackageError `json:",omitempty"` // errors loading dependencies |
| |
| // Test information |
| // If you add to this list you MUST add to p.AllFiles (below) too. |
| // Otherwise file name security lists will not apply to any new additions. |
| TestGoFiles []string `json:",omitempty"` // _test.go files in package |
| TestImports []string `json:",omitempty"` // imports from TestGoFiles |
| XTestGoFiles []string `json:",omitempty"` // _test.go files outside package |
| XTestImports []string `json:",omitempty"` // imports from XTestGoFiles |
| } |
| |
| // AllFiles returns the names of all the files considered for the package. |
| // This is used for sanity and security checks, so we include all files, |
| // even IgnoredGoFiles, because some subcommands consider them. |
| // The go/build package filtered others out (like foo_wrongGOARCH.s) |
| // and that's OK. |
| func (p *Package) AllFiles() []string { |
| return str.StringList( |
| p.GoFiles, |
| p.CgoFiles, |
| p.IgnoredGoFiles, |
| p.CFiles, |
| p.CXXFiles, |
| p.MFiles, |
| p.HFiles, |
| p.FFiles, |
| p.SFiles, |
| p.SwigFiles, |
| p.SwigCXXFiles, |
| p.SysoFiles, |
| p.TestGoFiles, |
| p.XTestGoFiles, |
| ) |
| } |
| |
| type PackageInternal struct { |
| // Unexported fields are not part of the public API. |
| Build *build.Package |
| Pkgdir string // overrides build.PkgDir |
| Imports []*Package // this package's direct imports |
| RawImports []string // this package's original imports as they appear in the text of the program |
| Deps []*Package |
| GoFiles []string // GoFiles+CgoFiles+TestGoFiles+XTestGoFiles files, absolute paths |
| SFiles []string |
| AllGoFiles []string // gofiles + IgnoredGoFiles, absolute paths |
| Target string // installed file for this package (may be executable) |
| Fake bool // synthesized package |
| External bool // synthesized external test package |
| ForceLibrary bool // this package is a library (even if named "main") |
| Cmdline bool // defined by files listed on command line |
| Local bool // imported via local path (./ or ../) |
| LocalPrefix string // interpret ./ and ../ imports relative to this prefix |
| ExeName string // desired name for temporary executable |
| CoverMode string // preprocess Go source files with the coverage tool in this mode |
| CoverVars map[string]*CoverVar // variables created by coverage analysis |
| OmitDebug bool // tell linker not to write debug information |
| BuildID string // expected build ID for generated package |
| GobinSubdir bool // install target would be subdir of GOBIN |
| } |
| |
| type NoGoError struct { |
| Package *Package |
| } |
| |
| func (e *NoGoError) Error() string { |
| // Count files beginning with _ and ., which we will pretend don't exist at all. |
| dummy := 0 |
| for _, name := range e.Package.IgnoredGoFiles { |
| if strings.HasPrefix(name, "_") || strings.HasPrefix(name, ".") { |
| dummy++ |
| } |
| } |
| |
| if len(e.Package.IgnoredGoFiles) > dummy { |
| // Go files exist, but they were ignored due to build constraints. |
| return "build constraints exclude all Go files in " + e.Package.Dir |
| } |
| if len(e.Package.TestGoFiles)+len(e.Package.XTestGoFiles) > 0 { |
| // Test Go files exist, but we're not interested in them. |
| // The double-negative is unfortunate but we want e.Package.Dir |
| // to appear at the end of error message. |
| return "no non-test Go files in " + e.Package.Dir |
| } |
| return "no Go files in " + e.Package.Dir |
| } |
| |
| // Resolve returns the resolved version of imports, |
| // which should be p.TestImports or p.XTestImports, NOT p.Imports. |
| // The imports in p.TestImports and p.XTestImports are not recursively |
| // loaded during the initial load of p, so they list the imports found in |
| // the source file, but most processing should be over the vendor-resolved |
| // import paths. We do this resolution lazily both to avoid file system work |
| // and because the eventual real load of the test imports (during 'go test') |
| // can produce better error messages if it starts with the original paths. |
| // The initial load of p loads all the non-test imports and rewrites |
| // the vendored paths, so nothing should ever call p.vendored(p.Imports). |
| func (p *Package) Resolve(imports []string) []string { |
| if len(imports) > 0 && len(p.Imports) > 0 && &imports[0] == &p.Imports[0] { |
| panic("internal error: p.Resolve(p.Imports) called") |
| } |
| seen := make(map[string]bool) |
| var all []string |
| for _, path := range imports { |
| path = ResolveImportPath(p, path) |
| if !seen[path] { |
| seen[path] = true |
| all = append(all, path) |
| } |
| } |
| sort.Strings(all) |
| return all |
| } |
| |
| // CoverVar holds the name of the generated coverage variables targeting the named file. |
| type CoverVar struct { |
| File string // local file name |
| Var string // name of count struct |
| } |
| |
| func (p *Package) copyBuild(pp *build.Package) { |
| p.Internal.Build = pp |
| |
| if pp.PkgTargetRoot != "" && cfg.BuildPkgdir != "" { |
| old := pp.PkgTargetRoot |
| pp.PkgRoot = cfg.BuildPkgdir |
| pp.PkgTargetRoot = cfg.BuildPkgdir |
| pp.PkgObj = filepath.Join(cfg.BuildPkgdir, strings.TrimPrefix(pp.PkgObj, old)) |
| } |
| |
| p.Dir = pp.Dir |
| p.ImportPath = pp.ImportPath |
| p.ImportComment = pp.ImportComment |
| p.Name = pp.Name |
| p.Doc = pp.Doc |
| p.Root = pp.Root |
| p.ConflictDir = pp.ConflictDir |
| p.BinaryOnly = pp.BinaryOnly |
| |
| // TODO? Target |
| p.Goroot = pp.Goroot |
| p.Standard = p.Goroot && p.ImportPath != "" && isStandardImportPath(p.ImportPath) |
| p.GoFiles = pp.GoFiles |
| p.CgoFiles = pp.CgoFiles |
| p.IgnoredGoFiles = pp.IgnoredGoFiles |
| p.CFiles = pp.CFiles |
| p.CXXFiles = pp.CXXFiles |
| p.MFiles = pp.MFiles |
| p.HFiles = pp.HFiles |
| p.FFiles = pp.FFiles |
| p.SFiles = pp.SFiles |
| p.SwigFiles = pp.SwigFiles |
| p.SwigCXXFiles = pp.SwigCXXFiles |
| p.SysoFiles = pp.SysoFiles |
| if cfg.BuildMSan { |
| // There's no way for .syso files to be built both with and without |
| // support for memory sanitizer. Assume they are built without, |
| // and drop them. |
| p.SysoFiles = nil |
| } |
| p.CgoCFLAGS = pp.CgoCFLAGS |
| p.CgoCPPFLAGS = pp.CgoCPPFLAGS |
| p.CgoCXXFLAGS = pp.CgoCXXFLAGS |
| p.CgoFFLAGS = pp.CgoFFLAGS |
| p.CgoLDFLAGS = pp.CgoLDFLAGS |
| p.CgoPkgConfig = pp.CgoPkgConfig |
| // We modify p.Imports in place, so make copy now. |
| p.Imports = make([]string, len(pp.Imports)) |
| copy(p.Imports, pp.Imports) |
| p.Internal.RawImports = pp.Imports |
| p.TestGoFiles = pp.TestGoFiles |
| p.TestImports = pp.TestImports |
| p.XTestGoFiles = pp.XTestGoFiles |
| p.XTestImports = pp.XTestImports |
| if IgnoreImports { |
| p.Imports = nil |
| p.TestImports = nil |
| p.XTestImports = nil |
| } |
| } |
| |
| // isStandardImportPath reports whether $GOROOT/src/path should be considered |
| // part of the standard distribution. For historical reasons we allow people to add |
| // their own code to $GOROOT instead of using $GOPATH, but we assume that |
| // code will start with a domain name (dot in the first element). |
| func isStandardImportPath(path string) bool { |
| i := strings.Index(path, "/") |
| if i < 0 { |
| i = len(path) |
| } |
| elem := path[:i] |
| return !strings.Contains(elem, ".") |
| } |
| |
| // A PackageError describes an error loading information about a package. |
| type PackageError struct { |
| ImportStack []string // shortest path from package named on command line to this one |
| Pos string // position of error |
| Err string // the error itself |
| IsImportCycle bool `json:"-"` // the error is an import cycle |
| Hard bool `json:"-"` // whether the error is soft or hard; soft errors are ignored in some places |
| } |
| |
| func (p *PackageError) Error() string { |
| // Import cycles deserve special treatment. |
| if p.IsImportCycle { |
| return fmt.Sprintf("%s\npackage %s\n", p.Err, strings.Join(p.ImportStack, "\n\timports ")) |
| } |
| if p.Pos != "" { |
| // Omit import stack. The full path to the file where the error |
| // is the most important thing. |
| return p.Pos + ": " + p.Err |
| } |
| if len(p.ImportStack) == 0 { |
| return p.Err |
| } |
| return "package " + strings.Join(p.ImportStack, "\n\timports ") + ": " + p.Err |
| } |
| |
| // An ImportStack is a stack of import paths. |
| type ImportStack []string |
| |
| func (s *ImportStack) Push(p string) { |
| *s = append(*s, p) |
| } |
| |
| func (s *ImportStack) Pop() { |
| *s = (*s)[0 : len(*s)-1] |
| } |
| |
| func (s *ImportStack) Copy() []string { |
| return append([]string{}, *s...) |
| } |
| |
| // shorterThan reports whether sp is shorter than t. |
| // We use this to record the shortest import sequence |
| // that leads to a particular package. |
| func (sp *ImportStack) shorterThan(t []string) bool { |
| s := *sp |
| if len(s) != len(t) { |
| return len(s) < len(t) |
| } |
| // If they are the same length, settle ties using string ordering. |
| for i := range s { |
| if s[i] != t[i] { |
| return s[i] < t[i] |
| } |
| } |
| return false // they are equal |
| } |
| |
| // packageCache is a lookup cache for loadPackage, |
| // so that if we look up a package multiple times |
| // we return the same pointer each time. |
| var packageCache = map[string]*Package{} |
| |
| func ClearPackageCache() { |
| for name := range packageCache { |
| delete(packageCache, name) |
| } |
| } |
| |
| func ClearPackageCachePartial(args []string) { |
| for _, arg := range args { |
| p := packageCache[arg] |
| if p != nil { |
| delete(packageCache, p.Dir) |
| delete(packageCache, p.ImportPath) |
| } |
| } |
| } |
| |
| // reloadPackage is like loadPackage but makes sure |
| // not to use the package cache. |
| func ReloadPackage(arg string, stk *ImportStack) *Package { |
| p := packageCache[arg] |
| if p != nil { |
| delete(packageCache, p.Dir) |
| delete(packageCache, p.ImportPath) |
| } |
| return LoadPackage(arg, stk) |
| } |
| |
| // dirToImportPath returns the pseudo-import path we use for a package |
| // outside the Go path. It begins with _/ and then contains the full path |
| // to the directory. If the package lives in c:\home\gopher\my\pkg then |
| // the pseudo-import path is _/c_/home/gopher/my/pkg. |
| // Using a pseudo-import path like this makes the ./ imports no longer |
| // a special case, so that all the code to deal with ordinary imports works |
| // automatically. |
| func dirToImportPath(dir string) string { |
| return pathpkg.Join("_", strings.Map(makeImportValid, filepath.ToSlash(dir))) |
| } |
| |
| func makeImportValid(r rune) rune { |
| // Should match Go spec, compilers, and ../../go/parser/parser.go:/isValidImport. |
| const illegalChars = `!"#$%&'()*,:;<=>?[\]^{|}` + "`\uFFFD" |
| if !unicode.IsGraphic(r) || unicode.IsSpace(r) || strings.ContainsRune(illegalChars, r) { |
| return '_' |
| } |
| return r |
| } |
| |
| // Mode flags for loadImport and download (in get.go). |
| const ( |
| // ResolveImport means that loadImport should do import path expansion. |
| // That is, ResolveImport means that the import path came from |
| // a source file and has not been expanded yet to account for |
| // vendoring or possible module adjustment. |
| // Every import path should be loaded initially with ResolveImport, |
| // and then the expanded version (for example with the /vendor/ in it) |
| // gets recorded as the canonical import path. At that point, future loads |
| // of that package must not pass ResolveImport, because |
| // disallowVendor will reject direct use of paths containing /vendor/. |
| ResolveImport = 1 << iota |
| |
| // ResolveModule is for download (part of "go get") and indicates |
| // that the module adjustment should be done, but not vendor adjustment. |
| ResolveModule |
| |
| // GetTestDeps is for download (part of "go get") and indicates |
| // that test dependencies should be fetched too. |
| GetTestDeps |
| ) |
| |
| // loadImport scans the directory named by path, which must be an import path, |
| // but possibly a local import path (an absolute file system path or one beginning |
| // with ./ or ../). A local relative path is interpreted relative to srcDir. |
| // It returns a *Package describing the package found in that directory. |
| func LoadImport(path, srcDir string, parent *Package, stk *ImportStack, importPos []token.Position, mode int) *Package { |
| stk.Push(path) |
| defer stk.Pop() |
| |
| // Determine canonical identifier for this package. |
| // For a local import the identifier is the pseudo-import path |
| // we create from the full directory to the package. |
| // Otherwise it is the usual import path. |
| // For vendored imports, it is the expanded form. |
| importPath := path |
| origPath := path |
| isLocal := build.IsLocalImport(path) |
| if isLocal { |
| importPath = dirToImportPath(filepath.Join(srcDir, path)) |
| } else if mode&ResolveImport != 0 { |
| // We do our own path resolution, because we want to |
| // find out the key to use in packageCache without the |
| // overhead of repeated calls to buildContext.Import. |
| // The code is also needed in a few other places anyway. |
| path = ResolveImportPath(parent, path) |
| importPath = path |
| } else if mode&ResolveModule != 0 { |
| path = ModuleImportPath(parent, path) |
| importPath = path |
| } |
| |
| p := packageCache[importPath] |
| if p != nil { |
| p = reusePackage(p, stk) |
| } else { |
| p = new(Package) |
| p.Internal.Local = isLocal |
| p.ImportPath = importPath |
| packageCache[importPath] = p |
| |
| // Load package. |
| // Import always returns bp != nil, even if an error occurs, |
| // in order to return partial information. |
| // |
| // TODO: After Go 1, decide when to pass build.AllowBinary here. |
| // See issue 3268 for mistakes to avoid. |
| buildMode := build.ImportComment |
| if mode&ResolveImport == 0 || path != origPath { |
| // Not vendoring, or we already found the vendored path. |
| buildMode |= build.IgnoreVendor |
| } |
| bp, err := cfg.BuildContext.Import(path, srcDir, buildMode) |
| bp.ImportPath = importPath |
| if cfg.GOBIN != "" { |
| bp.BinDir = cfg.GOBIN |
| } |
| if err == nil && !isLocal && bp.ImportComment != "" && bp.ImportComment != path && |
| !strings.Contains(path, "/vendor/") && !strings.HasPrefix(path, "vendor/") { |
| err = fmt.Errorf("code in directory %s expects import %q", bp.Dir, bp.ImportComment) |
| } |
| p.load(stk, bp, err) |
| if p.Error != nil && p.Error.Pos == "" { |
| p = setErrorPos(p, importPos) |
| } |
| |
| if origPath != cleanImport(origPath) { |
| p.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: fmt.Sprintf("non-canonical import path: %q should be %q", origPath, pathpkg.Clean(origPath)), |
| } |
| p.Incomplete = true |
| } |
| } |
| |
| // Checked on every import because the rules depend on the code doing the importing. |
| if perr := disallowInternal(srcDir, p, stk); perr != p { |
| return setErrorPos(perr, importPos) |
| } |
| if mode&ResolveImport != 0 { |
| if perr := disallowVendor(srcDir, origPath, p, stk); perr != p { |
| return setErrorPos(perr, importPos) |
| } |
| } |
| |
| if p.Name == "main" && parent != nil && parent.Dir != p.Dir { |
| perr := *p |
| perr.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: fmt.Sprintf("import %q is a program, not an importable package", path), |
| } |
| return setErrorPos(&perr, importPos) |
| } |
| |
| if p.Internal.Local && parent != nil && !parent.Internal.Local { |
| perr := *p |
| perr.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: fmt.Sprintf("local import %q in non-local package", path), |
| } |
| return setErrorPos(&perr, importPos) |
| } |
| |
| return p |
| } |
| |
| func setErrorPos(p *Package, importPos []token.Position) *Package { |
| if len(importPos) > 0 { |
| pos := importPos[0] |
| pos.Filename = base.ShortPath(pos.Filename) |
| p.Error.Pos = pos.String() |
| } |
| return p |
| } |
| |
| func cleanImport(path string) string { |
| orig := path |
| path = pathpkg.Clean(path) |
| if strings.HasPrefix(orig, "./") && path != ".." && !strings.HasPrefix(path, "../") { |
| path = "./" + path |
| } |
| return path |
| } |
| |
| var isDirCache = map[string]bool{} |
| |
| func isDir(path string) bool { |
| result, ok := isDirCache[path] |
| if ok { |
| return result |
| } |
| |
| fi, err := os.Stat(path) |
| result = err == nil && fi.IsDir() |
| isDirCache[path] = result |
| return result |
| } |
| |
| // ResolveImportPath returns the true meaning of path when it appears in parent. |
| // There are two different resolutions applied. |
| // First, there is Go 1.5 vendoring (golang.org/s/go15vendor). |
| // If vendor expansion doesn't trigger, then the path is also subject to |
| // Go 1.11 vgo legacy conversion (golang.org/issue/25069). |
| func ResolveImportPath(parent *Package, path string) (found string) { |
| found = VendoredImportPath(parent, path) |
| if found != path { |
| return found |
| } |
| return ModuleImportPath(parent, path) |
| } |
| |
| // dirAndRoot returns the source directory and workspace root |
| // for the package p, guaranteeing that root is a path prefix of dir. |
| func dirAndRoot(p *Package) (dir, root string) { |
| dir = filepath.Clean(p.Dir) |
| root = filepath.Join(p.Root, "src") |
| if !hasFilePathPrefix(dir, root) || p.ImportPath != "command-line-arguments" && filepath.Join(root, p.ImportPath) != dir { |
| // Look for symlinks before reporting error. |
| dir = expandPath(dir) |
| root = expandPath(root) |
| } |
| |
| if !hasFilePathPrefix(dir, root) || len(dir) <= len(root) || dir[len(root)] != filepath.Separator || p.ImportPath != "command-line-arguments" && !p.Internal.Local && filepath.Join(root, p.ImportPath) != dir { |
| base.Fatalf("unexpected directory layout:\n"+ |
| " import path: %s\n"+ |
| " root: %s\n"+ |
| " dir: %s\n"+ |
| " expand root: %s\n"+ |
| " expand dir: %s\n"+ |
| " separator: %s", |
| p.ImportPath, |
| filepath.Join(p.Root, "src"), |
| filepath.Clean(p.Dir), |
| root, |
| dir, |
| string(filepath.Separator)) |
| } |
| |
| return dir, root |
| } |
| |
| // VendoredImportPath returns the vendor-expansion of path when it appears in parent. |
| // If parent is x/y/z, then path might expand to x/y/z/vendor/path, x/y/vendor/path, |
| // x/vendor/path, vendor/path, or else stay path if none of those exist. |
| // VendoredImportPath returns the expanded path or, if no expansion is found, the original. |
| func VendoredImportPath(parent *Package, path string) (found string) { |
| if parent == nil || parent.Root == "" { |
| return path |
| } |
| |
| dir, root := dirAndRoot(parent) |
| |
| vpath := "vendor/" + path |
| for i := len(dir); i >= len(root); i-- { |
| if i < len(dir) && dir[i] != filepath.Separator { |
| continue |
| } |
| // Note: checking for the vendor directory before checking |
| // for the vendor/path directory helps us hit the |
| // isDir cache more often. It also helps us prepare a more useful |
| // list of places we looked, to report when an import is not found. |
| if !isDir(filepath.Join(dir[:i], "vendor")) { |
| continue |
| } |
| targ := filepath.Join(dir[:i], vpath) |
| if isDir(targ) && hasGoFiles(targ) { |
| importPath := parent.ImportPath |
| if importPath == "command-line-arguments" { |
| // If parent.ImportPath is 'command-line-arguments'. |
| // set to relative directory to root (also chopped root directory) |
| importPath = dir[len(root)+1:] |
| } |
| // We started with parent's dir c:\gopath\src\foo\bar\baz\quux\xyzzy. |
| // We know the import path for parent's dir. |
| // We chopped off some number of path elements and |
| // added vendor\path to produce c:\gopath\src\foo\bar\baz\vendor\path. |
| // Now we want to know the import path for that directory. |
| // Construct it by chopping the same number of path elements |
| // (actually the same number of bytes) from parent's import path |
| // and then append /vendor/path. |
| chopped := len(dir) - i |
| if chopped == len(importPath)+1 { |
| // We walked up from c:\gopath\src\foo\bar |
| // and found c:\gopath\src\vendor\path. |
| // We chopped \foo\bar (length 8) but the import path is "foo/bar" (length 7). |
| // Use "vendor/path" without any prefix. |
| return vpath |
| } |
| return importPath[:len(importPath)-chopped] + "/" + vpath |
| } |
| } |
| return path |
| } |
| |
| var ( |
| modulePrefix = []byte("\nmodule ") |
| goModPathCache = make(map[string]string) |
| ) |
| |
| // goModPath returns the module path in the go.mod in dir, if any. |
| func goModPath(dir string) (path string) { |
| path, ok := goModPathCache[dir] |
| if ok { |
| return path |
| } |
| defer func() { |
| goModPathCache[dir] = path |
| }() |
| |
| data, err := ioutil.ReadFile(filepath.Join(dir, "go.mod")) |
| if err != nil { |
| return "" |
| } |
| var i int |
| if bytes.HasPrefix(data, modulePrefix[1:]) { |
| i = 0 |
| } else { |
| i = bytes.Index(data, modulePrefix) |
| if i < 0 { |
| return "" |
| } |
| i++ |
| } |
| line := data[i:] |
| |
| // Cut line at \n, drop trailing \r if present. |
| if j := bytes.IndexByte(line, '\n'); j >= 0 { |
| line = line[:j] |
| } |
| if line[len(line)-1] == '\r' { |
| line = line[:len(line)-1] |
| } |
| line = line[len("module "):] |
| |
| // If quoted, unquote. |
| path = strings.TrimSpace(string(line)) |
| if path != "" && path[0] == '"' { |
| s, err := strconv.Unquote(path) |
| if err != nil { |
| return "" |
| } |
| path = s |
| } |
| return path |
| } |
| |
| // findVersionElement returns the slice indices of the final version element /vN in path. |
| // If there is no such element, it returns -1, -1. |
| func findVersionElement(path string) (i, j int) { |
| j = len(path) |
| for i = len(path) - 1; i >= 0; i-- { |
| if path[i] == '/' { |
| if isVersionElement(path[i:j]) { |
| return i, j |
| } |
| j = i |
| } |
| } |
| return -1, -1 |
| } |
| |
| // isVersionElement reports whether s is a well-formed path version element: |
| // v2, v3, v10, etc, but not v0, v05, v1. |
| func isVersionElement(s string) bool { |
| if len(s) < 3 || s[0] != '/' || s[1] != 'v' || s[2] == '0' || s[2] == '1' && len(s) == 3 { |
| return false |
| } |
| for i := 2; i < len(s); i++ { |
| if s[i] < '0' || '9' < s[i] { |
| return false |
| } |
| } |
| return true |
| } |
| |
| // ModuleImportPath translates import paths found in go modules |
| // back down to paths that can be resolved in ordinary builds. |
| // |
| // Define “new” code as code with a go.mod file in the same directory |
| // or a parent directory. If an import in new code says x/y/v2/z but |
| // x/y/v2/z does not exist and x/y/go.mod says “module x/y/v2”, |
| // then go build will read the import as x/y/z instead. |
| // See golang.org/issue/25069. |
| func ModuleImportPath(parent *Package, path string) (found string) { |
| if parent == nil || parent.Root == "" { |
| return path |
| } |
| |
| // If there are no vN elements in path, leave it alone. |
| // (The code below would do the same, but only after |
| // some other file system accesses that we can avoid |
| // here by returning early.) |
| if i, _ := findVersionElement(path); i < 0 { |
| return path |
| } |
| |
| dir, root := dirAndRoot(parent) |
| |
| // Consider dir and parents, up to and including root. |
| for i := len(dir); i >= len(root); i-- { |
| if i < len(dir) && dir[i] != filepath.Separator { |
| continue |
| } |
| if goModPath(dir[:i]) != "" { |
| goto HaveGoMod |
| } |
| } |
| // This code is not in a tree with a go.mod, |
| // so apply no changes to the path. |
| return path |
| |
| HaveGoMod: |
| // This import is in a tree with a go.mod. |
| // Allow it to refer to code in GOPATH/src/x/y/z as x/y/v2/z |
| // if GOPATH/src/x/y/go.mod says module "x/y/v2", |
| |
| // If x/y/v2/z exists, use it unmodified. |
| if bp, _ := cfg.BuildContext.Import(path, "", build.IgnoreVendor); bp.Dir != "" { |
| return path |
| } |
| |
| // Otherwise look for a go.mod supplying a version element. |
| // Some version-like elements may appear in paths but not |
| // be module versions; we skip over those to look for module |
| // versions. For example the module m/v2 might have a |
| // package m/v2/api/v1/foo. |
| limit := len(path) |
| for limit > 0 { |
| i, j := findVersionElement(path[:limit]) |
| if i < 0 { |
| return path |
| } |
| if bp, _ := cfg.BuildContext.Import(path[:i], "", build.IgnoreVendor); bp.Dir != "" { |
| if mpath := goModPath(bp.Dir); mpath != "" { |
| // Found a valid go.mod file, so we're stopping the search. |
| // If the path is m/v2/p and we found m/go.mod that says |
| // "module m/v2", then we return "m/p". |
| if mpath == path[:j] { |
| return path[:i] + path[j:] |
| } |
| // Otherwise just return the original path. |
| // We didn't find anything worth rewriting, |
| // and the go.mod indicates that we should |
| // not consider parent directories. |
| return path |
| } |
| } |
| limit = i |
| } |
| return path |
| } |
| |
| // hasGoFiles reports whether dir contains any files with names ending in .go. |
| // For a vendor check we must exclude directories that contain no .go files. |
| // Otherwise it is not possible to vendor just a/b/c and still import the |
| // non-vendored a/b. See golang.org/issue/13832. |
| func hasGoFiles(dir string) bool { |
| fis, _ := ioutil.ReadDir(dir) |
| for _, fi := range fis { |
| if !fi.IsDir() && strings.HasSuffix(fi.Name(), ".go") { |
| return true |
| } |
| } |
| return false |
| } |
| |
| // reusePackage reuses package p to satisfy the import at the top |
| // of the import stack stk. If this use causes an import loop, |
| // reusePackage updates p's error information to record the loop. |
| func reusePackage(p *Package, stk *ImportStack) *Package { |
| // We use p.Internal.Imports==nil to detect a package that |
| // is in the midst of its own loadPackage call |
| // (all the recursion below happens before p.Internal.Imports gets set). |
| if p.Internal.Imports == nil { |
| if p.Error == nil { |
| p.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: "import cycle not allowed", |
| IsImportCycle: true, |
| } |
| } |
| p.Incomplete = true |
| } |
| // Don't rewrite the import stack in the error if we have an import cycle. |
| // If we do, we'll lose the path that describes the cycle. |
| if p.Error != nil && !p.Error.IsImportCycle && stk.shorterThan(p.Error.ImportStack) { |
| p.Error.ImportStack = stk.Copy() |
| } |
| return p |
| } |
| |
| // disallowInternal checks that srcDir is allowed to import p. |
| // If the import is allowed, disallowInternal returns the original package p. |
| // If not, it returns a new package containing just an appropriate error. |
| func disallowInternal(srcDir string, p *Package, stk *ImportStack) *Package { |
| // golang.org/s/go14internal: |
| // An import of a path containing the element “internal” |
| // is disallowed if the importing code is outside the tree |
| // rooted at the parent of the “internal” directory. |
| |
| // There was an error loading the package; stop here. |
| if p.Error != nil { |
| return p |
| } |
| |
| // The generated 'testmain' package is allowed to access testing/internal/..., |
| // as if it were generated into the testing directory tree |
| // (it's actually in a temporary directory outside any Go tree). |
| // This cleans up a former kludge in passing functionality to the testing package. |
| if strings.HasPrefix(p.ImportPath, "testing/internal") && len(*stk) >= 2 && (*stk)[len(*stk)-2] == "testmain" { |
| return p |
| } |
| |
| // We can't check standard packages with gccgo. |
| if cfg.BuildContext.Compiler == "gccgo" && p.Standard { |
| return p |
| } |
| |
| // The stack includes p.ImportPath. |
| // If that's the only thing on the stack, we started |
| // with a name given on the command line, not an |
| // import. Anything listed on the command line is fine. |
| if len(*stk) == 1 { |
| return p |
| } |
| |
| // Check for "internal" element: three cases depending on begin of string and/or end of string. |
| i, ok := findInternal(p.ImportPath) |
| if !ok { |
| return p |
| } |
| |
| // Internal is present. |
| // Map import path back to directory corresponding to parent of internal. |
| if i > 0 { |
| i-- // rewind over slash in ".../internal" |
| } |
| parent := p.Dir[:i+len(p.Dir)-len(p.ImportPath)] |
| if hasFilePathPrefix(filepath.Clean(srcDir), filepath.Clean(parent)) { |
| return p |
| } |
| |
| // Look for symlinks before reporting error. |
| srcDir = expandPath(srcDir) |
| parent = expandPath(parent) |
| if hasFilePathPrefix(filepath.Clean(srcDir), filepath.Clean(parent)) { |
| return p |
| } |
| |
| // Internal is present, and srcDir is outside parent's tree. Not allowed. |
| perr := *p |
| perr.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: "use of internal package not allowed", |
| } |
| perr.Incomplete = true |
| return &perr |
| } |
| |
| // findInternal looks for the final "internal" path element in the given import path. |
| // If there isn't one, findInternal returns ok=false. |
| // Otherwise, findInternal returns ok=true and the index of the "internal". |
| func findInternal(path string) (index int, ok bool) { |
| // Three cases, depending on internal at start/end of string or not. |
| // The order matters: we must return the index of the final element, |
| // because the final one produces the most restrictive requirement |
| // on the importer. |
| switch { |
| case strings.HasSuffix(path, "/internal"): |
| return len(path) - len("internal"), true |
| case strings.Contains(path, "/internal/"): |
| return strings.LastIndex(path, "/internal/") + 1, true |
| case path == "internal", strings.HasPrefix(path, "internal/"): |
| return 0, true |
| } |
| return 0, false |
| } |
| |
| // disallowVendor checks that srcDir is allowed to import p as path. |
| // If the import is allowed, disallowVendor returns the original package p. |
| // If not, it returns a new package containing just an appropriate error. |
| func disallowVendor(srcDir, path string, p *Package, stk *ImportStack) *Package { |
| // The stack includes p.ImportPath. |
| // If that's the only thing on the stack, we started |
| // with a name given on the command line, not an |
| // import. Anything listed on the command line is fine. |
| if len(*stk) == 1 { |
| return p |
| } |
| |
| if perr := disallowVendorVisibility(srcDir, p, stk); perr != p { |
| return perr |
| } |
| |
| // Paths like x/vendor/y must be imported as y, never as x/vendor/y. |
| if i, ok := FindVendor(path); ok { |
| perr := *p |
| perr.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: "must be imported as " + path[i+len("vendor/"):], |
| } |
| perr.Incomplete = true |
| return &perr |
| } |
| |
| return p |
| } |
| |
| // disallowVendorVisibility checks that srcDir is allowed to import p. |
| // The rules are the same as for /internal/ except that a path ending in /vendor |
| // is not subject to the rules, only subdirectories of vendor. |
| // This allows people to have packages and commands named vendor, |
| // for maximal compatibility with existing source trees. |
| func disallowVendorVisibility(srcDir string, p *Package, stk *ImportStack) *Package { |
| // The stack includes p.ImportPath. |
| // If that's the only thing on the stack, we started |
| // with a name given on the command line, not an |
| // import. Anything listed on the command line is fine. |
| if len(*stk) == 1 { |
| return p |
| } |
| |
| // Check for "vendor" element. |
| i, ok := FindVendor(p.ImportPath) |
| if !ok { |
| return p |
| } |
| |
| // Vendor is present. |
| // Map import path back to directory corresponding to parent of vendor. |
| if i > 0 { |
| i-- // rewind over slash in ".../vendor" |
| } |
| truncateTo := i + len(p.Dir) - len(p.ImportPath) |
| if truncateTo < 0 || len(p.Dir) < truncateTo { |
| return p |
| } |
| parent := p.Dir[:truncateTo] |
| if hasFilePathPrefix(filepath.Clean(srcDir), filepath.Clean(parent)) { |
| return p |
| } |
| |
| // Look for symlinks before reporting error. |
| srcDir = expandPath(srcDir) |
| parent = expandPath(parent) |
| if hasFilePathPrefix(filepath.Clean(srcDir), filepath.Clean(parent)) { |
| return p |
| } |
| |
| // Vendor is present, and srcDir is outside parent's tree. Not allowed. |
| perr := *p |
| perr.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: "use of vendored package not allowed", |
| } |
| perr.Incomplete = true |
| return &perr |
| } |
| |
| // FindVendor looks for the last non-terminating "vendor" path element in the given import path. |
| // If there isn't one, FindVendor returns ok=false. |
| // Otherwise, FindVendor returns ok=true and the index of the "vendor". |
| // |
| // Note that terminating "vendor" elements don't count: "x/vendor" is its own package, |
| // not the vendored copy of an import "" (the empty import path). |
| // This will allow people to have packages or commands named vendor. |
| // This may help reduce breakage, or it may just be confusing. We'll see. |
| func FindVendor(path string) (index int, ok bool) { |
| // Two cases, depending on internal at start of string or not. |
| // The order matters: we must return the index of the final element, |
| // because the final one is where the effective import path starts. |
| switch { |
| case strings.Contains(path, "/vendor/"): |
| return strings.LastIndex(path, "/vendor/") + 1, true |
| case strings.HasPrefix(path, "vendor/"): |
| return 0, true |
| } |
| return 0, false |
| } |
| |
| type targetDir int |
| |
| const ( |
| ToRoot targetDir = iota // to bin dir inside package root (default) |
| ToTool // GOROOT/pkg/tool |
| StalePath // the old import path; fail to build |
| ) |
| |
| // goTools is a map of Go program import path to install target directory. |
| var GoTools = map[string]targetDir{ |
| "cmd/addr2line": ToTool, |
| "cmd/api": ToTool, |
| "cmd/asm": ToTool, |
| "cmd/compile": ToTool, |
| "cmd/cgo": ToTool, |
| "cmd/cover": ToTool, |
| "cmd/dist": ToTool, |
| "cmd/doc": ToTool, |
| "cmd/fix": ToTool, |
| "cmd/link": ToTool, |
| "cmd/newlink": ToTool, |
| "cmd/nm": ToTool, |
| "cmd/objdump": ToTool, |
| "cmd/pack": ToTool, |
| "cmd/pprof": ToTool, |
| "cmd/trace": ToTool, |
| "cmd/vet": ToTool, |
| "code.google.com/p/go.tools/cmd/cover": StalePath, |
| "code.google.com/p/go.tools/cmd/godoc": StalePath, |
| "code.google.com/p/go.tools/cmd/vet": StalePath, |
| } |
| |
| var raceExclude = map[string]bool{ |
| "runtime/race": true, |
| "runtime/msan": true, |
| "runtime/cgo": true, |
| "cmd/cgo": true, |
| "syscall": true, |
| "errors": true, |
| } |
| |
| var cgoExclude = map[string]bool{ |
| "runtime/cgo": true, |
| } |
| |
| var cgoSyscallExclude = map[string]bool{ |
| "runtime/cgo": true, |
| "runtime/race": true, |
| "runtime/msan": true, |
| } |
| |
| var foldPath = make(map[string]string) |
| |
| // load populates p using information from bp, err, which should |
| // be the result of calling build.Context.Import. |
| func (p *Package) load(stk *ImportStack, bp *build.Package, err error) *Package { |
| p.copyBuild(bp) |
| |
| // The localPrefix is the path we interpret ./ imports relative to. |
| // Synthesized main packages sometimes override this. |
| p.Internal.LocalPrefix = dirToImportPath(p.Dir) |
| |
| if err != nil { |
| if _, ok := err.(*build.NoGoError); ok { |
| err = &NoGoError{Package: p} |
| } |
| p.Incomplete = true |
| err = base.ExpandScanner(err) |
| p.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: err.Error(), |
| } |
| return p |
| } |
| |
| useBindir := p.Name == "main" |
| if !p.Standard { |
| switch cfg.BuildBuildmode { |
| case "c-archive", "c-shared", "plugin": |
| useBindir = false |
| } |
| } |
| |
| if useBindir { |
| // Report an error when the old code.google.com/p/go.tools paths are used. |
| if GoTools[p.ImportPath] == StalePath { |
| newPath := strings.Replace(p.ImportPath, "code.google.com/p/go.", "golang.org/x/", 1) |
| e := fmt.Sprintf("the %v command has moved; use %v instead.", p.ImportPath, newPath) |
| p.Error = &PackageError{Err: e} |
| return p |
| } |
| _, elem := filepath.Split(p.Dir) |
| full := cfg.BuildContext.GOOS + "_" + cfg.BuildContext.GOARCH + "/" + elem |
| if cfg.BuildContext.GOOS != base.ToolGOOS || cfg.BuildContext.GOARCH != base.ToolGOARCH { |
| // Install cross-compiled binaries to subdirectories of bin. |
| elem = full |
| } |
| if p.Internal.Build.BinDir != "" { |
| // Install to GOBIN or bin of GOPATH entry. |
| p.Internal.Target = filepath.Join(p.Internal.Build.BinDir, elem) |
| if !p.Goroot && strings.Contains(elem, "/") && cfg.GOBIN != "" { |
| // Do not create $GOBIN/goos_goarch/elem. |
| p.Internal.Target = "" |
| p.Internal.GobinSubdir = true |
| } |
| } |
| if GoTools[p.ImportPath] == ToTool { |
| // This is for 'go tool'. |
| // Override all the usual logic and force it into the tool directory. |
| p.Internal.Target = filepath.Join(cfg.GOROOTpkg, "tool", full) |
| } |
| if p.Internal.Target != "" && cfg.BuildContext.GOOS == "windows" { |
| p.Internal.Target += ".exe" |
| } |
| } else if p.Internal.Local { |
| // Local import turned into absolute path. |
| // No permanent install target. |
| p.Internal.Target = "" |
| } else { |
| p.Internal.Target = p.Internal.Build.PkgObj |
| if cfg.BuildLinkshared { |
| shlibnamefile := p.Internal.Target[:len(p.Internal.Target)-2] + ".shlibname" |
| shlib, err := ioutil.ReadFile(shlibnamefile) |
| if err == nil { |
| libname := strings.TrimSpace(string(shlib)) |
| if cfg.BuildContext.Compiler == "gccgo" { |
| p.Shlib = filepath.Join(p.Internal.Build.PkgTargetRoot, "shlibs", libname) |
| } else { |
| p.Shlib = filepath.Join(p.Internal.Build.PkgTargetRoot, libname) |
| |
| } |
| } else if !os.IsNotExist(err) { |
| base.Fatalf("unexpected error reading %s: %v", shlibnamefile, err) |
| } |
| } |
| } |
| |
| ImportPaths := p.Imports |
| // Packages that use cgo import runtime/cgo implicitly. |
| // Packages that use cgo also import syscall implicitly, |
| // to wrap errno. |
| // Exclude certain packages to avoid circular dependencies. |
| if len(p.CgoFiles) > 0 && (!p.Standard || !cgoExclude[p.ImportPath]) { |
| ImportPaths = append(ImportPaths, "runtime/cgo") |
| } |
| if len(p.CgoFiles) > 0 && (!p.Standard || !cgoSyscallExclude[p.ImportPath]) { |
| ImportPaths = append(ImportPaths, "syscall") |
| } |
| |
| if cfg.BuildContext.CgoEnabled && p.Name == "main" && !p.Goroot { |
| // Currently build modes c-shared, pie (on systems that do not |
| // support PIE with internal linking mode), plugin, and |
| // -linkshared force external linking mode, as of course does |
| // -ldflags=-linkmode=external. External linking mode forces |
| // an import of runtime/cgo. |
| pieCgo := cfg.BuildBuildmode == "pie" && (cfg.BuildContext.GOOS != "linux" || cfg.BuildContext.GOARCH != "amd64") |
| linkmodeExternal := false |
| for i, a := range cfg.BuildLdflags { |
| if a == "-linkmode=external" { |
| linkmodeExternal = true |
| } |
| if a == "-linkmode" && i+1 < len(cfg.BuildLdflags) && cfg.BuildLdflags[i+1] == "external" { |
| linkmodeExternal = true |
| } |
| } |
| if cfg.BuildBuildmode == "c-shared" || cfg.BuildBuildmode == "plugin" || pieCgo || cfg.BuildLinkshared || linkmodeExternal { |
| ImportPaths = append(ImportPaths, "runtime/cgo") |
| } |
| } |
| |
| // Everything depends on runtime, except runtime, its internal |
| // subpackages, and unsafe. |
| if !p.Standard || (p.ImportPath != "runtime" && !strings.HasPrefix(p.ImportPath, "runtime/internal/") && p.ImportPath != "unsafe") { |
| ImportPaths = append(ImportPaths, "runtime") |
| // When race detection enabled everything depends on runtime/race. |
| // Exclude certain packages to avoid circular dependencies. |
| if cfg.BuildRace && (!p.Standard || !raceExclude[p.ImportPath]) { |
| ImportPaths = append(ImportPaths, "runtime/race") |
| } |
| // MSan uses runtime/msan. |
| if cfg.BuildMSan && (!p.Standard || !raceExclude[p.ImportPath]) { |
| ImportPaths = append(ImportPaths, "runtime/msan") |
| } |
| // On ARM with GOARM=5, everything depends on math for the link. |
| if p.Name == "main" && cfg.Goarch == "arm" { |
| ImportPaths = append(ImportPaths, "math") |
| } |
| } |
| |
| // Runtime and its internal packages depend on runtime/internal/sys, |
| // so that they pick up the generated zversion.go file. |
| // This can be an issue particularly for runtime/internal/atomic; |
| // see issue 13655. |
| if p.Standard && (p.ImportPath == "runtime" || strings.HasPrefix(p.ImportPath, "runtime/internal/")) && p.ImportPath != "runtime/internal/sys" { |
| ImportPaths = append(ImportPaths, "runtime/internal/sys") |
| } |
| |
| // Build list of full paths to all Go files in the package, |
| // for use by commands like go fmt. |
| p.Internal.GoFiles = str.StringList(p.GoFiles, p.CgoFiles, p.TestGoFiles, p.XTestGoFiles) |
| for i := range p.Internal.GoFiles { |
| p.Internal.GoFiles[i] = filepath.Join(p.Dir, p.Internal.GoFiles[i]) |
| } |
| sort.Strings(p.Internal.GoFiles) |
| |
| p.Internal.SFiles = str.StringList(p.SFiles) |
| for i := range p.Internal.SFiles { |
| p.Internal.SFiles[i] = filepath.Join(p.Dir, p.Internal.SFiles[i]) |
| } |
| sort.Strings(p.Internal.SFiles) |
| |
| p.Internal.AllGoFiles = str.StringList(p.IgnoredGoFiles) |
| for i := range p.Internal.AllGoFiles { |
| p.Internal.AllGoFiles[i] = filepath.Join(p.Dir, p.Internal.AllGoFiles[i]) |
| } |
| p.Internal.AllGoFiles = append(p.Internal.AllGoFiles, p.Internal.GoFiles...) |
| sort.Strings(p.Internal.AllGoFiles) |
| |
| // Check for case-insensitive collision of input files. |
| // To avoid problems on case-insensitive files, we reject any package |
| // where two different input files have equal names under a case-insensitive |
| // comparison. |
| inputs := p.AllFiles() |
| f1, f2 := str.FoldDup(inputs) |
| if f1 != "" { |
| p.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: fmt.Sprintf("case-insensitive file name collision: %q and %q", f1, f2), |
| } |
| return p |
| } |
| |
| // If first letter of input file is ASCII, it must be alphanumeric. |
| // This avoids files turning into flags when invoking commands, |
| // and other problems we haven't thought of yet. |
| // Also, _cgo_ files must be generated by us, not supplied. |
| // They are allowed to have //go:cgo_ldflag directives. |
| // The directory scan ignores files beginning with _, |
| // so we shouldn't see any _cgo_ files anyway, but just be safe. |
| for _, file := range inputs { |
| if !SafeArg(file) || strings.HasPrefix(file, "_cgo_") { |
| p.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: fmt.Sprintf("invalid input file name %q", file), |
| } |
| return p |
| } |
| } |
| if name := pathpkg.Base(p.ImportPath); !SafeArg(name) { |
| p.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: fmt.Sprintf("invalid input directory name %q", name), |
| } |
| return p |
| } |
| if !SafeArg(p.ImportPath) { |
| p.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: fmt.Sprintf("invalid import path %q", p.ImportPath), |
| } |
| return p |
| } |
| |
| // Build list of imported packages and full dependency list. |
| imports := make([]*Package, 0, len(p.Imports)) |
| deps := make(map[string]*Package) |
| save := func(path string, p1 *Package) { |
| // The same import path could produce an error or not, |
| // depending on what tries to import it. |
| // Prefer to record entries with errors, so we can report them. |
| p0 := deps[path] |
| if p0 == nil || p1.Error != nil && (p0.Error == nil || len(p0.Error.ImportStack) > len(p1.Error.ImportStack)) { |
| deps[path] = p1 |
| } |
| } |
| |
| for i, path := range ImportPaths { |
| if path == "C" { |
| continue |
| } |
| p1 := LoadImport(path, p.Dir, p, stk, p.Internal.Build.ImportPos[path], ResolveImport) |
| if p.Standard && p.Error == nil && !p1.Standard && p1.Error == nil { |
| p.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: fmt.Sprintf("non-standard import %q in standard package %q", path, p.ImportPath), |
| } |
| pos := p.Internal.Build.ImportPos[path] |
| if len(pos) > 0 { |
| p.Error.Pos = pos[0].String() |
| } |
| } |
| |
| path = p1.ImportPath |
| ImportPaths[i] = path |
| if i < len(p.Imports) { |
| p.Imports[i] = path |
| } |
| |
| save(path, p1) |
| imports = append(imports, p1) |
| for _, dep := range p1.Internal.Deps { |
| save(dep.ImportPath, dep) |
| } |
| if p1.Incomplete { |
| p.Incomplete = true |
| } |
| } |
| p.Internal.Imports = imports |
| |
| p.Deps = make([]string, 0, len(deps)) |
| for dep := range deps { |
| p.Deps = append(p.Deps, dep) |
| } |
| sort.Strings(p.Deps) |
| for _, dep := range p.Deps { |
| p1 := deps[dep] |
| if p1 == nil { |
| panic("impossible: missing entry in package cache for " + dep + " imported by " + p.ImportPath) |
| } |
| p.Internal.Deps = append(p.Internal.Deps, p1) |
| if p1.Error != nil { |
| p.DepsErrors = append(p.DepsErrors, p1.Error) |
| } |
| } |
| |
| // unsafe is a fake package. |
| if p.Standard && (p.ImportPath == "unsafe" || cfg.BuildContext.Compiler == "gccgo") { |
| p.Internal.Target = "" |
| } |
| p.Target = p.Internal.Target |
| |
| // If cgo is not enabled, ignore cgo supporting sources |
| // just as we ignore go files containing import "C". |
| if !cfg.BuildContext.CgoEnabled { |
| p.CFiles = nil |
| p.CXXFiles = nil |
| p.MFiles = nil |
| p.SwigFiles = nil |
| p.SwigCXXFiles = nil |
| // Note that SFiles are okay (they go to the Go assembler) |
| // and HFiles are okay (they might be used by the SFiles). |
| // Also Sysofiles are okay (they might not contain object |
| // code; see issue #16050). |
| } |
| |
| // The gc toolchain only permits C source files with cgo. |
| if len(p.CFiles) > 0 && !p.UsesCgo() && !p.UsesSwig() && cfg.BuildContext.Compiler == "gc" { |
| p.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: fmt.Sprintf("C source files not allowed when not using cgo or SWIG: %s", strings.Join(p.CFiles, " ")), |
| } |
| return p |
| } |
| |
| // Check for case-insensitive collisions of import paths. |
| fold := str.ToFold(p.ImportPath) |
| if other := foldPath[fold]; other == "" { |
| foldPath[fold] = p.ImportPath |
| } else if other != p.ImportPath { |
| p.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: fmt.Sprintf("case-insensitive import collision: %q and %q", p.ImportPath, other), |
| } |
| return p |
| } |
| |
| if p.BinaryOnly { |
| // For binary-only package, use build ID from supplied package binary. |
| buildID, err := buildid.ReadBuildID(p.Name, p.Target) |
| if err == nil { |
| p.Internal.BuildID = buildID |
| } |
| } else { |
| computeBuildID(p) |
| } |
| return p |
| } |
| |
| // SafeArg reports whether arg is a "safe" command-line argument, |
| // meaning that when it appears in a command-line, it probably |
| // doesn't have some special meaning other than its own name. |
| // Obviously args beginning with - are not safe (they look like flags). |
| // Less obviously, args beginning with @ are not safe (they look like |
| // GNU binutils flagfile specifiers, sometimes called "response files"). |
| // To be conservative, we reject almost any arg beginning with non-alphanumeric ASCII. |
| // We accept leading . _ and / as likely in file system paths. |
| // There is a copy of this function in cmd/compile/internal/gc/noder.go. |
| func SafeArg(name string) bool { |
| if name == "" { |
| return false |
| } |
| c := name[0] |
| return '0' <= c && c <= '9' || 'A' <= c && c <= 'Z' || 'a' <= c && c <= 'z' || c == '.' || c == '_' || c == '/' || c >= utf8.RuneSelf |
| } |
| |
| // usesSwig reports whether the package needs to run SWIG. |
| func (p *Package) UsesSwig() bool { |
| return len(p.SwigFiles) > 0 || len(p.SwigCXXFiles) > 0 |
| } |
| |
| // usesCgo reports whether the package needs to run cgo |
| func (p *Package) UsesCgo() bool { |
| return len(p.CgoFiles) > 0 |
| } |
| |
| // packageList returns the list of packages in the dag rooted at roots |
| // as visited in a depth-first post-order traversal. |
| func PackageList(roots []*Package) []*Package { |
| seen := map[*Package]bool{} |
| all := []*Package{} |
| var walk func(*Package) |
| walk = func(p *Package) { |
| if seen[p] { |
| return |
| } |
| seen[p] = true |
| for _, p1 := range p.Internal.Imports { |
| walk(p1) |
| } |
| all = append(all, p) |
| } |
| for _, root := range roots { |
| walk(root) |
| } |
| return all |
| } |
| |
| // computeStale computes the Stale flag in the package dag that starts |
| // at the named pkgs (command-line arguments). |
| func ComputeStale(pkgs ...*Package) { |
| for _, p := range PackageList(pkgs) { |
| p.Stale, p.StaleReason = isStale(p) |
| } |
| } |
| |
| // The runtime version string takes one of two forms: |
| // "go1.X[.Y]" for Go releases, and "devel +hash" at tip. |
| // Determine whether we are in a released copy by |
| // inspecting the version. |
| var isGoRelease = strings.HasPrefix(runtime.Version(), "go1") |
| |
| // isStale and computeBuildID |
| // |
| // Theory of Operation |
| // |
| // There is an installed copy of the package (or binary). |
| // Can we reuse the installed copy, or do we need to build a new one? |
| // |
| // We can use the installed copy if it matches what we'd get |
| // by building a new one. The hard part is predicting that without |
| // actually running a build. |
| // |
| // To start, we must know the set of inputs to the build process that can |
| // affect the generated output. At a minimum, that includes the source |
| // files for the package and also any compiled packages imported by those |
| // source files. The *Package has these, and we use them. One might also |
| // argue for including in the input set: the build tags, whether the race |
| // detector is in use, the target operating system and architecture, the |
| // compiler and linker binaries being used, the additional flags being |
| // passed to those, the cgo binary being used, the additional flags cgo |
| // passes to the host C compiler, the host C compiler being used, the set |
| // of host C include files and installed C libraries, and so on. |
| // We include some but not all of this information. |
| // |
| // Once we have decided on a set of inputs, we must next decide how to |
| // tell whether the content of that set has changed since the last build |
| // of p. If there have been no changes, then we assume a new build would |
| // produce the same result and reuse the installed package or binary. |
| // But if there have been changes, then we assume a new build might not |
| // produce the same result, so we rebuild. |
| // |
| // There are two common ways to decide whether the content of the set has |
| // changed: modification times and content hashes. We use a mixture of both. |
| // |
| // The use of modification times (mtimes) was pioneered by make: |
| // assuming that a file's mtime is an accurate record of when that file was last written, |
| // and assuming that the modification time of an installed package or |
| // binary is the time that it was built, if the mtimes of the inputs |
| // predate the mtime of the installed object, then the build of that |
| // object saw those versions of the files, and therefore a rebuild using |
| // those same versions would produce the same object. In contrast, if any |
| // mtime of an input is newer than the mtime of the installed object, a |
| // change has occurred since the build, and the build should be redone. |
| // |
| // Modification times are attractive because the logic is easy to |
| // understand and the file system maintains the mtimes automatically |
| // (less work for us). Unfortunately, there are a variety of ways in |
| // which the mtime approach fails to detect a change and reuses a stale |
| // object file incorrectly. (Making the opposite mistake, rebuilding |
| // unnecessarily, is only a performance problem and not a correctness |
| // problem, so we ignore that one.) |
| // |
| // As a warmup, one problem is that to be perfectly precise, we need to |
| // compare the input mtimes against the time at the beginning of the |
| // build, but the object file time is the time at the end of the build. |
| // If an input file changes after being read but before the object is |
| // written, the next build will see an object newer than the input and |
| // will incorrectly decide that the object is up to date. We make no |
| // attempt to detect or solve this problem. |
| // |
| // Another problem is that due to file system imprecision, an input and |
| // output that are actually ordered in time have the same mtime. |
| // This typically happens on file systems with 1-second (or, worse, |
| // 2-second) mtime granularity and with automated scripts that write an |
| // input and then immediately run a build, or vice versa. If an input and |
| // an output have the same mtime, the conservative behavior is to treat |
| // the output as out-of-date and rebuild. This can cause one or more |
| // spurious rebuilds, but only for 1 second, until the object finally has |
| // an mtime later than the input. |
| // |
| // Another problem is that binary distributions often set the mtime on |
| // all files to the same time. If the distribution includes both inputs |
| // and cached build outputs, the conservative solution to the previous |
| // problem will cause unnecessary rebuilds. Worse, in such a binary |
| // distribution, those rebuilds might not even have permission to update |
| // the cached build output. To avoid these write errors, if an input and |
| // output have the same mtime, we assume the output is up-to-date. |
| // This is the opposite of what the previous problem would have us do, |
| // but binary distributions are more common than instances of the |
| // previous problem. |
| // |
| // A variant of the last problem is that some binary distributions do not |
| // set the mtime on all files to the same time. Instead they let the file |
| // system record mtimes as the distribution is unpacked. If the outputs |
| // are unpacked before the inputs, they'll be older and a build will try |
| // to rebuild them. That rebuild might hit the same write errors as in |
| // the last scenario. We don't make any attempt to solve this, and we |
| // haven't had many reports of it. Perhaps the only time this happens is |
| // when people manually unpack the distribution, and most of the time |
| // that's done as the same user who will be using it, so an initial |
| // rebuild on first use succeeds quietly. |
| // |
| // More generally, people and programs change mtimes on files. The last |
| // few problems were specific examples of this, but it's a general problem. |
| // For example, instead of a binary distribution, copying a home |
| // directory from one directory or machine to another might copy files |
| // but not preserve mtimes. If the inputs are new than the outputs on the |
| // first machine but copied first, they end up older than the outputs on |
| // the second machine. |
| // |
| // Because many other build systems have the same sensitivity to mtimes, |
| // most programs manipulating source code take pains not to break the |
| // mtime assumptions. For example, Git does not set the mtime of files |
| // during a checkout operation, even when checking out an old version of |
| // the code. This decision was made specifically to work well with |
| // mtime-based build systems. |
| // |
| // The killer problem, though, for mtime-based build systems is that the |
| // build only has access to the mtimes of the inputs that still exist. |
| // If it is possible to remove an input without changing any other inputs, |
| // a later build will think the object is up-to-date when it is not. |
| // This happens for Go because a package is made up of all source |
| // files in a directory. If a source file is removed, there is no newer |
| // mtime available recording that fact. The mtime on the directory could |
| // be used, but it also changes when unrelated files are added to or |
| // removed from the directory, so including the directory mtime would |
| // cause unnecessary rebuilds, possibly many. It would also exacerbate |
| // the problems mentioned earlier, since even programs that are careful |
| // to maintain mtimes on files rarely maintain mtimes on directories. |
| // |
| // A variant of the last problem is when the inputs change for other |
| // reasons. For example, Go 1.4 and Go 1.5 both install $GOPATH/src/mypkg |
| // into the same target, $GOPATH/pkg/$GOOS_$GOARCH/mypkg.a. |
| // If Go 1.4 has built mypkg into mypkg.a, a build using Go 1.5 must |
| // rebuild mypkg.a, but from mtimes alone mypkg.a looks up-to-date. |
| // If Go 1.5 has just been installed, perhaps the compiler will have a |
| // newer mtime; since the compiler is considered an input, that would |
| // trigger a rebuild. But only once, and only the last Go 1.4 build of |
| // mypkg.a happened before Go 1.5 was installed. If a user has the two |
| // versions installed in different locations and flips back and forth, |
| // mtimes alone cannot tell what to do. Changing the toolchain is |
| // changing the set of inputs, without affecting any mtimes. |
| // |
| // To detect the set of inputs changing, we turn away from mtimes and to |
| // an explicit data comparison. Specifically, we build a list of the |
| // inputs to the build, compute its SHA1 hash, and record that as the |
| // ``build ID'' in the generated object. At the next build, we can |
| // recompute the build ID and compare it to the one in the generated |
| // object. If they differ, the list of inputs has changed, so the object |
| // is out of date and must be rebuilt. |
| // |
| // Because this build ID is computed before the build begins, the |
| // comparison does not have the race that mtime comparison does. |
| // |
| // Making the build sensitive to changes in other state is |
| // straightforward: include the state in the build ID hash, and if it |
| // changes, so does the build ID, triggering a rebuild. |
| // |
| // To detect changes in toolchain, we include the toolchain version in |
| // the build ID hash for package runtime, and then we include the build |
| // IDs of all imported packages in the build ID for p. |
| // |
| // It is natural to think about including build tags in the build ID, but |
| // the naive approach of just dumping the tags into the hash would cause |
| // spurious rebuilds. For example, 'go install' and 'go install -tags neverusedtag' |
| // produce the same binaries (assuming neverusedtag is never used). |
| // A more precise approach would be to include only tags that have an |
| // effect on the build. But the effect of a tag on the build is to |
| // include or exclude a file from the compilation, and that file list is |
| // already in the build ID hash. So the build ID is already tag-sensitive |
| // in a perfectly precise way. So we do NOT explicitly add build tags to |
| // the build ID hash. |
| // |
| // We do not include as part of the build ID the operating system, |
| // architecture, or whether the race detector is enabled, even though all |
| // three have an effect on the output, because that information is used |
| // to decide the install location. Binaries for linux and binaries for |
| // darwin are written to different directory trees; including that |
| // information in the build ID is unnecessary (although it would be |
| // harmless). |
| // |
| // TODO(rsc): Investigate the cost of putting source file content into |
| // the build ID hash as a replacement for the use of mtimes. Using the |
| // file content would avoid all the mtime problems, but it does require |
| // reading all the source files, something we avoid today (we read the |
| // beginning to find the build tags and the imports, but we stop as soon |
| // as we see the import block is over). If the package is stale, the compiler |
| // is going to read the files anyway. But if the package is up-to-date, the |
| // read is overhead. |
| // |
| // TODO(rsc): Investigate the complexity of making the build more |
| // precise about when individual results are needed. To be fully precise, |
| // there are two results of a compilation: the entire .a file used by the link |
| // and the subpiece used by later compilations (__.PKGDEF only). |
| // If a rebuild is needed but produces the previous __.PKGDEF, then |
| // no more recompilation due to the rebuilt package is needed, only |
| // relinking. To date, there is nothing in the Go command to express this. |
| // |
| // Special Cases |
| // |
| // When the go command makes the wrong build decision and does not |
| // rebuild something it should, users fall back to adding the -a flag. |
| // Any common use of the -a flag should be considered prima facie evidence |
| // that isStale is returning an incorrect false result in some important case. |
| // Bugs reported in the behavior of -a itself should prompt the question |
| // ``Why is -a being used at all? What bug does that indicate?'' |
| // |
| // There is a long history of changes to isStale to try to make -a into a |
| // suitable workaround for bugs in the mtime-based decisions. |
| // It is worth recording that history to inform (and, as much as possible, deter) future changes. |
| // |
| // (1) Before the build IDs were introduced, building with alternate tags |
| // would happily reuse installed objects built without those tags. |
| // For example, "go build -tags netgo myprog.go" would use the installed |
| // copy of package net, even if that copy had been built without netgo. |
| // (The netgo tag controls whether package net uses cgo or pure Go for |
| // functionality such as name resolution.) |
| // Using the installed non-netgo package defeats the purpose. |
| // |
| // Users worked around this with "go build -tags netgo -a myprog.go". |
| // |
| // Build IDs have made that workaround unnecessary: |
| // "go build -tags netgo myprog.go" |
| // cannot use a non-netgo copy of package net. |
| // |
| // (2) Before the build IDs were introduced, building with different toolchains, |
| // especially changing between toolchains, tried to reuse objects stored in |
| // $GOPATH/pkg, resulting in link-time errors about object file mismatches. |
| // |
| // Users worked around this with "go install -a ./...". |
| // |
| // Build IDs have made that workaround unnecessary: |
| // "go install ./..." will rebuild any objects it finds that were built against |
| // a different toolchain. |
| // |
| // (3) The common use of "go install -a ./..." led to reports of problems |
| // when the -a forced the rebuild of the standard library, which for some |
| // users was not writable. Because we didn't understand that the real |
| // problem was the bug -a was working around, we changed -a not to |
| // apply to the standard library. |
| // |
| // (4) The common use of "go build -tags netgo -a myprog.go" broke |
| // when we changed -a not to apply to the standard library, because |
| // if go build doesn't rebuild package net, it uses the non-netgo version. |
| // |
| // Users worked around this with "go build -tags netgo -installsuffix barf myprog.go". |
| // The -installsuffix here is making the go command look for packages |
| // in pkg/$GOOS_$GOARCH_barf instead of pkg/$GOOS_$GOARCH. |
| // Since the former presumably doesn't exist, go build decides to rebuild |
| // everything, including the standard library. Since go build doesn't |
| // install anything it builds, nothing is ever written to pkg/$GOOS_$GOARCH_barf, |
| // so repeated invocations continue to work. |
| // |
| // If the use of -a wasn't a red flag, the use of -installsuffix to point to |
| // a non-existent directory in a command that installs nothing should |
| // have been. |
| // |
| // (5) Now that (1) and (2) no longer need -a, we have removed the kludge |
| // introduced in (3): once again, -a means ``rebuild everything,'' not |
| // ``rebuild everything except the standard library.'' Only Go 1.4 had |
| // the restricted meaning. |
| // |
| // In addition to these cases trying to trigger rebuilds, there are |
| // special cases trying NOT to trigger rebuilds. The main one is that for |
| // a variety of reasons (see above), the install process for a Go release |
| // cannot be relied upon to set the mtimes such that the go command will |
| // think the standard library is up to date. So the mtime evidence is |
| // ignored for the standard library if we find ourselves in a release |
| // version of Go. Build ID-based staleness checks still apply to the |
| // standard library, even in release versions. This makes |
| // 'go build -tags netgo' work, among other things. |
| |
| // isStale reports whether package p needs to be rebuilt, |
| // along with the reason why. |
| func isStale(p *Package) (bool, string) { |
| if p.Standard && (p.ImportPath == "unsafe" || cfg.BuildContext.Compiler == "gccgo") { |
| // fake, builtin package |
| return false, "builtin package" |
| } |
| if p.Error != nil { |
| return true, "errors loading package" |
| } |
| if p.Stale { |
| return true, p.StaleReason |
| } |
| |
| // If this is a package with no source code, it cannot be rebuilt. |
| // If the binary is missing, we mark the package stale so that |
| // if a rebuild is needed, that rebuild attempt will produce a useful error. |
| // (Some commands, such as 'go list', do not attempt to rebuild.) |
| if p.BinaryOnly { |
| if p.Internal.Target == "" { |
| // Fail if a build is attempted. |
| return true, "no source code for package, but no install target" |
| } |
| if _, err := os.Stat(p.Internal.Target); err != nil { |
| // Fail if a build is attempted. |
| return true, "no source code for package, but cannot access install target: " + err.Error() |
| } |
| return false, "no source code for package" |
| } |
| |
| // If the -a flag is given, rebuild everything. |
| if cfg.BuildA { |
| return true, "build -a flag in use" |
| } |
| |
| // If there's no install target, we have to rebuild. |
| if p.Internal.Target == "" { |
| return true, "no install target" |
| } |
| |
| // Package is stale if completely unbuilt. |
| fi, err := os.Stat(p.Internal.Target) |
| if err != nil { |
| return true, "cannot stat install target" |
| } |
| |
| // Package is stale if the expected build ID differs from the |
| // recorded build ID. This catches changes like a source file |
| // being removed from a package directory. See issue 3895. |
| // It also catches changes in build tags that affect the set of |
| // files being compiled. See issue 9369. |
| // It also catches changes in toolchain, like when flipping between |
| // two versions of Go compiling a single GOPATH. |
| // See issue 8290 and issue 10702. |
| targetBuildID, err := buildid.ReadBuildID(p.Name, p.Target) |
| if err == nil && targetBuildID != p.Internal.BuildID { |
| return true, "build ID mismatch" |
| } |
| |
| // Package is stale if a dependency is. |
| for _, p1 := range p.Internal.Deps { |
| if p1.Stale { |
| return true, "stale dependency" |
| } |
| } |
| |
| // The checks above are content-based staleness. |
| // We assume they are always accurate. |
| // |
| // The checks below are mtime-based staleness. |
| // We hope they are accurate, but we know that they fail in the case of |
| // prebuilt Go installations that don't preserve the build mtimes |
| // (for example, if the pkg/ mtimes are before the src/ mtimes). |
| // See the large comment above isStale for details. |
| |
| // If we are running a release copy of Go and didn't find a content-based |
| // reason to rebuild the standard packages, do not rebuild them. |
| // They may not be writable anyway, but they are certainly not changing. |
| // This makes 'go build' skip the standard packages when |
| // using an official release, even when the mtimes have been changed. |
| // See issue 3036, issue 3149, issue 4106, issue 8290. |
| // (If a change to a release tree must be made by hand, the way to force the |
| // install is to run make.bash, which will remove the old package archives |
| // before rebuilding.) |
| if p.Standard && isGoRelease { |
| return false, "standard package in Go release distribution" |
| } |
| |
| // Time-based staleness. |
| |
| built := fi.ModTime() |
| |
| olderThan := func(file string) bool { |
| fi, err := os.Stat(file) |
| return err != nil || fi.ModTime().After(built) |
| } |
| |
| // Package is stale if a dependency is, or if a dependency is newer. |
| for _, p1 := range p.Internal.Deps { |
| if p1.Internal.Target != "" && olderThan(p1.Internal.Target) { |
| return true, "newer dependency" |
| } |
| } |
| |
| // As a courtesy to developers installing new versions of the compiler |
| // frequently, define that packages are stale if they are |
| // older than the compiler, and commands if they are older than |
| // the linker. This heuristic will not work if the binaries are |
| // back-dated, as some binary distributions may do, but it does handle |
| // a very common case. |
| // See issue 3036. |
| // Exclude $GOROOT, under the assumption that people working on |
| // the compiler may want to control when everything gets rebuilt, |
| // and people updating the Go repository will run make.bash or all.bash |
| // and get a full rebuild anyway. |
| // Excluding $GOROOT used to also fix issue 4106, but that's now |
| // taken care of above (at least when the installed Go is a released version). |
| if p.Root != cfg.GOROOT { |
| if olderThan(cfg.BuildToolchainCompiler()) { |
| return true, "newer compiler" |
| } |
| if p.Internal.Build.IsCommand() && olderThan(cfg.BuildToolchainLinker()) { |
| return true, "newer linker" |
| } |
| } |
| |
| // Note: Until Go 1.5, we had an additional shortcut here. |
| // We built a list of the workspace roots ($GOROOT, each $GOPATH) |
| // containing targets directly named on the command line, |
| // and if p were not in any of those, it would be treated as up-to-date |
| // as long as it is built. The goal was to avoid rebuilding a system-installed |
| // $GOROOT, unless something from $GOROOT were explicitly named |
| // on the command line (like go install math). |
| // That's now handled by the isGoRelease clause above. |
| // The other effect of the shortcut was to isolate different entries in |
| // $GOPATH from each other. This had the unfortunate effect that |
| // if you had (say), GOPATH listing two entries, one for commands |
| // and one for libraries, and you did a 'git pull' in the library one |
| // and then tried 'go install commands/...', it would build the new libraries |
| // during the first build (because they wouldn't have been installed at all) |
| // but then subsequent builds would not rebuild the libraries, even if the |
| // mtimes indicate they are stale, because the different GOPATH entries |
| // were treated differently. This behavior was confusing when using |
| // non-trivial GOPATHs, which were particularly common with some |
| // code management conventions, like the original godep. |
| // Since the $GOROOT case (the original motivation) is handled separately, |
| // we no longer put a barrier between the different $GOPATH entries. |
| // |
| // One implication of this is that if there is a system directory for |
| // non-standard Go packages that is included in $GOPATH, the mtimes |
| // on those compiled packages must be no earlier than the mtimes |
| // on the source files. Since most distributions use the same mtime |
| // for all files in a tree, they will be unaffected. People using plain |
| // tar x to extract system-installed packages will need to adjust mtimes, |
| // but it's better to force them to get the mtimes right than to ignore |
| // the mtimes and thereby do the wrong thing in common use cases. |
| // |
| // So there is no GOPATH vs GOPATH shortcut here anymore. |
| // |
| // If something needs to come back here, we could try writing a dummy |
| // file with a random name to the $GOPATH/pkg directory (and removing it) |
| // to test for write access, and then skip GOPATH roots we don't have write |
| // access to. But hopefully we can just use the mtimes always. |
| |
| srcs := str.StringList(p.GoFiles, p.CFiles, p.CXXFiles, p.MFiles, p.HFiles, p.FFiles, p.SFiles, p.CgoFiles, p.SysoFiles, p.SwigFiles, p.SwigCXXFiles) |
| for _, src := range srcs { |
| if olderThan(filepath.Join(p.Dir, src)) { |
| return true, "newer source file" |
| } |
| } |
| |
| return false, "" |
| } |
| |
| // computeBuildID computes the build ID for p, leaving it in p.Internal.BuildID. |
| // Build ID is a hash of the information we want to detect changes in. |
| // See the long comment in isStale for details. |
| func computeBuildID(p *Package) { |
| h := sha1.New() |
| |
| // Include the list of files compiled as part of the package. |
| // This lets us detect removed files. See issue 3895. |
| inputFiles := str.StringList( |
| p.GoFiles, |
| p.CgoFiles, |
| p.CFiles, |
| p.CXXFiles, |
| p.FFiles, |
| p.MFiles, |
| p.HFiles, |
| p.SFiles, |
| p.SysoFiles, |
| p.SwigFiles, |
| p.SwigCXXFiles, |
| ) |
| for _, file := range inputFiles { |
| fmt.Fprintf(h, "file %s\n", file) |
| } |
| |
| // Include the content of runtime/internal/sys/zversion.go in the hash |
| // for package runtime. This will give package runtime a |
| // different build ID in each Go release. |
| if p.Standard && p.ImportPath == "runtime/internal/sys" && cfg.BuildContext.Compiler != "gccgo" { |
| data, err := ioutil.ReadFile(filepath.Join(p.Dir, "zversion.go")) |
| if os.IsNotExist(err) { |
| p.Stale = true |
| p.StaleReason = fmt.Sprintf("missing zversion.go") |
| } else if err != nil { |
| base.Fatalf("go: %s", err) |
| } |
| fmt.Fprintf(h, "zversion %q\n", string(data)) |
| |
| // Add environment variables that affect code generation. |
| switch cfg.BuildContext.GOARCH { |
| case "arm": |
| fmt.Fprintf(h, "GOARM=%s\n", cfg.GOARM) |
| case "386": |
| fmt.Fprintf(h, "GO386=%s\n", cfg.GO386) |
| } |
| } |
| |
| // Include the build IDs of any dependencies in the hash. |
| // This, combined with the runtime/zversion content, |
| // will cause packages to have different build IDs when |
| // compiled with different Go releases. |
| // This helps the go command know to recompile when |
| // people use the same GOPATH but switch between |
| // different Go releases. See issue 10702. |
| // This is also a better fix for issue 8290. |
| for _, p1 := range p.Internal.Deps { |
| fmt.Fprintf(h, "dep %s %s\n", p1.ImportPath, p1.Internal.BuildID) |
| } |
| |
| p.Internal.BuildID = fmt.Sprintf("%x", h.Sum(nil)) |
| } |
| |
| var cmdCache = map[string]*Package{} |
| |
| func ClearCmdCache() { |
| for name := range cmdCache { |
| delete(cmdCache, name) |
| } |
| } |
| |
| // loadPackage is like loadImport but is used for command-line arguments, |
| // not for paths found in import statements. In addition to ordinary import paths, |
| // loadPackage accepts pseudo-paths beginning with cmd/ to denote commands |
| // in the Go command directory, as well as paths to those directories. |
| func LoadPackage(arg string, stk *ImportStack) *Package { |
| if build.IsLocalImport(arg) { |
| dir := arg |
| if !filepath.IsAbs(dir) { |
| if abs, err := filepath.Abs(dir); err == nil { |
| // interpret relative to current directory |
| dir = abs |
| } |
| } |
| if sub, ok := hasSubdir(cfg.GOROOTsrc, dir); ok && strings.HasPrefix(sub, "cmd/") && !strings.Contains(sub[4:], "/") { |
| arg = sub |
| } |
| } |
| if strings.HasPrefix(arg, "cmd/") && !strings.Contains(arg[4:], "/") { |
| if p := cmdCache[arg]; p != nil { |
| return p |
| } |
| stk.Push(arg) |
| defer stk.Pop() |
| |
| bp, err := cfg.BuildContext.ImportDir(filepath.Join(cfg.GOROOTsrc, arg), 0) |
| bp.ImportPath = arg |
| bp.Goroot = true |
| bp.BinDir = cfg.GOROOTbin |
| if cfg.GOROOTbin != "" { |
| bp.BinDir = cfg.GOROOTbin |
| } |
| bp.Root = cfg.GOROOT |
| bp.SrcRoot = cfg.GOROOTsrc |
| p := new(Package) |
| cmdCache[arg] = p |
| p.load(stk, bp, err) |
| if p.Error == nil && p.Name != "main" { |
| p.Error = &PackageError{ |
| ImportStack: stk.Copy(), |
| Err: fmt.Sprintf("expected package main but found package %s in %s", p.Name, p.Dir), |
| } |
| } |
| return p |
| } |
| |
| // Wasn't a command; must be a package. |
| // If it is a local import path but names a standard package, |
| // we treat it as if the user specified the standard package. |
| // This lets you run go test ./ioutil in package io and be |
| // referring to io/ioutil rather than a hypothetical import of |
| // "./ioutil". |
| if build.IsLocalImport(arg) { |
| bp, _ := cfg.BuildContext.ImportDir(filepath.Join(base.Cwd, arg), build.FindOnly) |
| if bp.ImportPath != "" && bp.ImportPath != "." { |
| arg = bp.ImportPath |
| } |
| } |
| |
| return LoadImport(arg, base.Cwd, nil, stk, nil, 0) |
| } |
| |
| // packages returns the packages named by the |
| // command line arguments 'args'. If a named package |
| // cannot be loaded at all (for example, if the directory does not exist), |
| // then packages prints an error and does not include that |
| // package in the results. However, if errors occur trying |
| // to load dependencies of a named package, the named |
| // package is still returned, with p.Incomplete = true |
| // and details in p.DepsErrors. |
| func Packages(args []string) []*Package { |
| var pkgs []*Package |
| for _, pkg := range PackagesAndErrors(args) { |
| if pkg.Error != nil { |
| base.Errorf("can't load package: %s", pkg.Error) |
| continue |
| } |
| pkgs = append(pkgs, pkg) |
| } |
| return pkgs |
| } |
| |
| // packagesAndErrors is like 'packages' but returns a |
| // *Package for every argument, even the ones that |
| // cannot be loaded at all. |
| // The packages that fail to load will have p.Error != nil. |
| func PackagesAndErrors(args []string) []*Package { |
| if len(args) > 0 && strings.HasSuffix(args[0], ".go") { |
| return []*Package{GoFilesPackage(args)} |
| } |
| |
| args = ImportPaths(args) |
| var ( |
| pkgs []*Package |
| stk ImportStack |
| seenArg = make(map[string]bool) |
| seenPkg = make(map[*Package]bool) |
| ) |
| |
| for _, arg := range args { |
| if seenArg[arg] { |
| continue |
| } |
| seenArg[arg] = true |
| pkg := LoadPackage(arg, &stk) |
| if seenPkg[pkg] { |
| continue |
| } |
| seenPkg[pkg] = true |
| pkgs = append(pkgs, pkg) |
| } |
| ComputeStale(pkgs...) |
| |
| return pkgs |
| } |
| |
| // packagesForBuild is like 'packages' but fails if any of |
| // the packages or their dependencies have errors |
| // (cannot be built). |
| func PackagesForBuild(args []string) []*Package { |
| pkgs := PackagesAndErrors(args) |
| printed := map[*PackageError]bool{} |
| for _, pkg := range pkgs { |
| if pkg.Error != nil { |
| base.Errorf("can't load package: %s", pkg.Error) |
| } |
| for _, err := range pkg.DepsErrors { |
| // Since these are errors in dependencies, |
| // the same error might show up multiple times, |
| // once in each package that depends on it. |
| // Only print each once. |
| if !printed[err] { |
| printed[err] = true |
| base.Errorf("%s", err) |
| } |
| } |
| } |
| base.ExitIfErrors() |
| |
| // Check for duplicate loads of the same package. |
| // That should be impossible, but if it does happen then |
| // we end up trying to build the same package twice, |
| // usually in parallel overwriting the same files, |
| // which doesn't work very well. |
| seen := map[string]bool{} |
| reported := map[string]bool{} |
| for _, pkg := range PackageList(pkgs) { |
| if seen[pkg.ImportPath] && !reported[pkg.ImportPath] { |
| reported[pkg.ImportPath] = true |
| base.Errorf("internal error: duplicate loads of %s", pkg.ImportPath) |
| } |
| seen[pkg.ImportPath] = true |
| } |
| base.ExitIfErrors() |
| |
| return pkgs |
| } |
| |
| // GoFilesPackage creates a package for building a collection of Go files |
| // (typically named on the command line). The target is named p.a for |
| // package p or named after the first Go file for package main. |
| func GoFilesPackage(gofiles []string) *Package { |
| // TODO: Remove this restriction. |
| for _, f := range gofiles { |
| if !strings.HasSuffix(f, ".go") { |
| base.Fatalf("named files must be .go files") |
| } |
| } |
| |
| var stk ImportStack |
| ctxt := cfg.BuildContext |
| ctxt.UseAllFiles = true |
| |
| // Synthesize fake "directory" that only shows the named files, |
| // to make it look like this is a standard package or |
| // command directory. So that local imports resolve |
| // consistently, the files must all be in the same directory. |
| var dirent []os.FileInfo |
| var dir string |
| for _, file := range gofiles { |
| fi, err := os.Stat(file) |
| if err != nil { |
| base.Fatalf("%s", err) |
| } |
| if fi.IsDir() { |
| base.Fatalf("%s is a directory, should be a Go file", file) |
| } |
| dir1, _ := filepath.Split(file) |
| if dir1 == "" { |
| dir1 = "./" |
| } |
| if dir == "" { |
| dir = dir1 |
| } else if dir != dir1 { |
| base.Fatalf("named files must all be in one directory; have %s and %s", dir, dir1) |
| } |
| dirent = append(dirent, fi) |
| } |
| ctxt.ReadDir = func(string) ([]os.FileInfo, error) { return dirent, nil } |
| |
| var err error |
| if dir == "" { |
| dir = base.Cwd |
| } |
| dir, err = filepath.Abs(dir) |
| if err != nil { |
| base.Fatalf("%s", err) |
| } |
| |
| bp, err := ctxt.ImportDir(dir, 0) |
| pkg := new(Package) |
| pkg.Internal.Local = true |
| pkg.Internal.Cmdline = true |
| stk.Push("main") |
| pkg.load(&stk, bp, err) |
| stk.Pop() |
| pkg.Internal.LocalPrefix = dirToImportPath(dir) |
| pkg.ImportPath = "command-line-arguments" |
| pkg.Internal.Target = "" |
| |
| if pkg.Name == "main" { |
| _, elem := filepath.Split(gofiles[0]) |
| exe := elem[:len(elem)-len(".go")] + cfg.ExeSuffix |
| if cfg.BuildO == "" { |
| cfg.BuildO = exe |
| } |
| if cfg.GOBIN != "" { |
| pkg.Internal.Target = filepath.Join(cfg.GOBIN, exe) |
| } |
| } |
| |
| pkg.Target = pkg.Internal.Target |
| pkg.Stale = true |
| pkg.StaleReason = "files named on command line" |
| |
| ComputeStale(pkg) |
| return pkg |
| } |