internal/abi.Type.Pointers (method)

58 uses

	internal/abi (current package)
		type.go#L196: func (t *Type) Pointers() bool { return t.PtrBytes != 0 }

	internal/reflectlite
		swapper.go#L36: 	hasPtr := typ.Pointers()
		value.go#L103: 	if v.typ().Size() != goarch.PtrSize || !v.typ().Pointers() {

	reflect
		abi.go#L169: 	if rcvr.IfaceIndir() || rcvr.Pointers() {
		deepequal.go#L42: 			if !v1.typ().Pointers() {
		swapper.go#L37: 	hasPtr := typ.Pointers()
		type.go#L2130: 	if ktyp.Pointers() || etyp.Pointers() {
		type.go#L2139: 		if ktyp.Pointers() {
		type.go#L2144: 		if etyp.Pointers() {
		type.go#L2654: 			if ft.Typ.Pointers() {
		type.go#L2666: 			if !ft.Typ.Pointers() {
		type.go#L2762: 			if ft.Pointers() {
		type.go#L2825: 	if length > 0 && typ.Pointers() {
		type.go#L2834: 	case !typ.Pointers() || array.Size_ == 0:
		type.go#L3042: 	if !t.Pointers() {
		value.go#L112: 	if v.typ().Size() != goarch.PtrSize || !v.typ().Pointers() {
		value.go#L2206: 		if !v.typ().Pointers() {
		value.go#L2802: 		if !v.typ().Pointers() {

	runtime
		arena.go#L483: 	if !typ.Pointers() {
		arena.go#L505: 	if typ.Pointers() {
		cgocall.go#L580: 	if !t.Pointers() || p == nil {
		cgocall.go#L643: 		if !st.Elem.Pointers() {
		cgocall.go#L668: 			if !f.Typ.Pointers() {
		cgocheck.go#L93: 	if !typ.Pointers() {
		cgocheck.go#L114: 	if !typ.Pointers() {
		cgocheck.go#L236: 	if !typ.Pointers() {
		chan.go#L100: 	case !elem.Pointers():
		checkptr.go#L19: 	if elem.Pointers() && uintptr(p)&(uintptr(elem.Align_)-1) != 0 {
		malloc.go#L1058: 	noscan := typ == nil || !typ.Pointers()
		map.go#L259: 	if !t.Bucket.Pointers() {
		map.go#L370: 		up := roundupsize(sz, !t.Bucket.Pointers())
		map.go#L384: 		if t.Bucket.Pointers() {
		map.go#L797: 			} else if t.Key.Pointers() {
		map.go#L803: 			} else if t.Elem.Pointers() {
		map.go#L896: 	if !t.Bucket.Pointers() {
		map.go#L1354: 		if h.flags&oldIterator == 0 && t.Bucket.Pointers() {
		map_fast32.go#L334: 			if goarch.PtrSize == 4 && t.Key.Pointers() {
		map_fast32.go#L340: 			if t.Elem.Pointers() {
		map_fast32.go#L460: 				if goarch.PtrSize == 4 && t.Key.Pointers() && writeBarrier.enabled {
		map_fast32.go#L478: 		if h.flags&oldIterator == 0 && t.Bucket.Pointers() {
		map_fast64.go#L334: 			if t.Key.Pointers() {
		map_fast64.go#L344: 			if t.Elem.Pointers() {
		map_fast64.go#L464: 				if t.Key.Pointers() && writeBarrier.enabled {
		map_fast64.go#L488: 		if h.flags&oldIterator == 0 && t.Bucket.Pointers() {
		map_faststr.go#L359: 			if t.Elem.Pointers() {
		map_faststr.go#L492: 		if h.flags&oldIterator == 0 && t.Bucket.Pointers() {
		mbarrier.go#L166: 	if writeBarrier.enabled && typ.Pointers() {
		mbarrier.go#L257: 	if writeBarrier.enabled && typ != nil && typ.Pointers() && size >= goarch.PtrSize {
		mbarrier.go#L351: 	if !elemType.Pointers() {
		mbarrier.go#L369: 	if writeBarrier.enabled && typ.Pointers() {
		mbarrier.go#L393: 	if writeBarrier.enabled && typ.Pointers() {
		mbarrier.go#L406: 	if writeBarrier.enabled && typ.Pointers() {
		mfinal.go#L453: 		if ot.Elem == nil || ot.Elem.Pointers() || ot.Elem.Size_ >= maxTinySize {
		slice.go#L56: 	if !et.Pointers() {
		slice.go#L208: 	noscan := !et.Pointers()
		slice.go#L263: 	if !et.Pointers() {
		slice.go#L341: 	if !et.Pointers() {