mirror of https://github.com/golang/go.git
add files to fix typos
This commit is contained in:
parent
45928e38c0
commit
8d289b73a3
|
|
@ -48,7 +48,7 @@ func test9400(t *testing.T) {
|
|||
}
|
||||
|
||||
// Disable GC for the duration of the test.
|
||||
// This avoids a potential GC deadlock when spinning in uninterruptable ASM below #49695.
|
||||
// This avoids a potential GC deadlock when spinning in uninterruptible ASM below #49695.
|
||||
defer debug.SetGCPercent(debug.SetGCPercent(-1))
|
||||
// SetGCPercent waits until the mark phase is over, but the runtime
|
||||
// also preempts at the start of the sweep phase, so make sure that's
|
||||
|
|
|
|||
|
|
@ -344,7 +344,7 @@ func runGet(ctx context.Context, cmd *base.Command, args []string) {
|
|||
// The result of any version query for a given module — even "upgrade" or
|
||||
// "patch" — is always relative to the build list at the start of
|
||||
// the 'go get' command, not an intermediate state, and is therefore
|
||||
// deterministic and therefore cachable, and the constraints on the
|
||||
// deterministic and therefore cacheable, and the constraints on the
|
||||
// selected version of each module can only narrow as we iterate.
|
||||
//
|
||||
// "all" is functionally very similar to a wildcard pattern. The set of
|
||||
|
|
|
|||
|
|
@ -96,7 +96,7 @@ Examples:
|
|||
7. Move large constants to vector registers.
|
||||
|
||||
Go asm uses VMOVQ/VMOVD/VMOVS to move 128-bit, 64-bit and 32-bit constants into vector registers, respectively.
|
||||
And for a 128-bit interger, it take two 64-bit operands, for the low and high parts separately.
|
||||
And for a 128-bit integer, it take two 64-bit operands, for the low and high parts separately.
|
||||
|
||||
Examples:
|
||||
|
||||
|
|
|
|||
|
|
@ -423,7 +423,7 @@ func (w *writer) Hash(s *LSym) {
|
|||
// contentHashSection returns a mnemonic for s's section.
|
||||
// The goal is to prevent content-addressability from moving symbols between sections.
|
||||
// contentHashSection only distinguishes between sets of sections for which this matters.
|
||||
// Allowing flexibility increases the effectiveness of content-addressibility.
|
||||
// Allowing flexibility increases the effectiveness of content-addressability.
|
||||
// But in some cases, such as doing addressing based on a base symbol,
|
||||
// we need to ensure that a symbol is always in a particular section.
|
||||
// Some of these conditions are duplicated in cmd/link/internal/ld.(*Link).symtab.
|
||||
|
|
|
|||
|
|
@ -1978,7 +1978,7 @@ func fusedJump(p *obj.Prog) (bool, uint8) {
|
|||
type padJumpsCtx int32
|
||||
|
||||
func makePjcCtx(ctxt *obj.Link) padJumpsCtx {
|
||||
// Disable jump padding on 32 bit builds by settting
|
||||
// Disable jump padding on 32 bit builds by setting
|
||||
// padJumps to 0.
|
||||
if ctxt.Arch.Family == sys.I386 {
|
||||
return padJumpsCtx(0)
|
||||
|
|
|
|||
|
|
@ -168,7 +168,7 @@ func cgocall(fn, arg unsafe.Pointer) int32 {
|
|||
// any C on the call stack, which there will be after this point. If
|
||||
// there isn't, we can use frame pointer unwinding to collect call
|
||||
// stacks efficiently. This will be the case for the first Go-to-C call
|
||||
// on a stack, so it's prefereable to update it here, after we emit a
|
||||
// on a stack, so it's preferable to update it here, after we emit a
|
||||
// trace event in entersyscall above.
|
||||
mp.ncgo++
|
||||
|
||||
|
|
|
|||
|
|
@ -588,7 +588,7 @@ func (s *emitState) emitCounterDataFile(finalHash [16]byte, w io.Writer) error {
|
|||
}
|
||||
|
||||
// markProfileEmitted signals the runtime/coverage machinery that
|
||||
// coverate data output files have already been written out, and there
|
||||
// coverage data output files have already been written out, and there
|
||||
// is no need to take any additional action at exit time. This
|
||||
// function is called (via linknamed reference) from the
|
||||
// coverage-related boilerplate code in _testmain.go emitted for go
|
||||
|
|
|
|||
|
|
@ -537,7 +537,7 @@ func gcAssistAlloc1(gp *g, scanWork int64) {
|
|||
// The gcBlackenEnabled check in malloc races with the
|
||||
// store that clears it but an atomic check in every malloc
|
||||
// would be a performance hit.
|
||||
// Instead we recheck it here on the non-preemptable system
|
||||
// Instead we recheck it here on the non-preemptible system
|
||||
// stack to determine if we should perform an assist.
|
||||
|
||||
// GC is done, so ignore any remaining debt.
|
||||
|
|
|
|||
Loading…
Reference in New Issue