import (
"internal/abi"
"internal/goarch"
+ "internal/goexperiment"
"runtime/internal/atomic"
"runtime/internal/sys"
"unsafe"
// If the CPU limiter is enabled, intentionally don't
// assist to reduce the amount of CPU time spent in the GC.
if traced {
- traceGCMarkAssistDone()
+ trace := traceAcquire()
+ if trace.ok() {
+ trace.GCMarkAssistDone()
+ traceRelease(trace)
+ }
}
return
}
// We were able to steal all of the credit we
// needed.
if traced {
- traceGCMarkAssistDone()
+ trace := traceAcquire()
+ if trace.ok() {
+ trace.GCMarkAssistDone()
+ traceRelease(trace)
+ }
}
return
}
}
-
if traceEnabled() && !traced {
- traced = true
- traceGCMarkAssistStart()
+ trace := traceAcquire()
+ if trace.ok() {
+ traced = true
+ trace.GCMarkAssistStart()
+ traceRelease(trace)
+ }
}
// Perform assist work
// this G's assist debt, or the GC cycle is over.
}
if traced {
- traceGCMarkAssistDone()
+ trace := traceAcquire()
+ if trace.ok() {
+ trace.GCMarkAssistDone()
+ traceRelease(trace)
+ }
}
}
throw("scanobject of a noscan object")
}
+ var tp typePointers
if n > maxObletBytes {
// Large object. Break into oblets for better
// parallelism and lower latency.
// of the object.
n = s.base() + s.elemsize - b
n = min(n, maxObletBytes)
+ if goexperiment.AllocHeaders {
+ tp = s.typePointersOfUnchecked(s.base())
+ tp = tp.fastForward(b-tp.addr, b+n)
+ }
+ } else {
+ if goexperiment.AllocHeaders {
+ tp = s.typePointersOfUnchecked(b)
+ }
}
- hbits := heapBitsForAddr(b, n)
+ var hbits heapBits
+ if !goexperiment.AllocHeaders {
+ hbits = heapBitsForAddr(b, n)
+ }
var scanSize uintptr
for {
var addr uintptr
- if hbits, addr = hbits.nextFast(); addr == 0 {
- if hbits, addr = hbits.next(); addr == 0 {
- break
+ if goexperiment.AllocHeaders {
+ if tp, addr = tp.nextFast(); addr == 0 {
+ if tp, addr = tp.next(b + n); addr == 0 {
+ break
+ }
+ }
+ } else {
+ if hbits, addr = hbits.nextFast(); addr == 0 {
+ if hbits, addr = hbits.next(); addr == 0 {
+ break
+ }
}
}