aboutsummaryrefslogtreecommitdiffstats
path: root/community/go
diff options
context:
space:
mode:
authorNatanael Copa <ncopa@alpinelinux.org>2016-06-27 11:36:13 +0000
committerNatanael Copa <ncopa@alpinelinux.org>2016-06-27 11:36:13 +0000
commit8ea73bd1fa44bde5ade8119015b498aa289daf67 (patch)
tree2d894ad30b967a12f5844a484610b479c53ae7d0 /community/go
parentc21409608a22c46e07b51bf4caea1fe96dbe90f8 (diff)
downloadaports-8ea73bd1fa44bde5ade8119015b498aa289daf67.tar.bz2
aports-8ea73bd1fa44bde5ade8119015b498aa289daf67.tar.xz
community/go: remove unused patches
Diffstat (limited to 'community/go')
-rw-r--r--community/go/0001-runtime-use-entire-address-space-on-32-bit.patch303
-rw-r--r--community/go/stacksize.patch120
2 files changed, 0 insertions, 423 deletions
diff --git a/community/go/0001-runtime-use-entire-address-space-on-32-bit.patch b/community/go/0001-runtime-use-entire-address-space-on-32-bit.patch
deleted file mode 100644
index 20ccd45d41..0000000000
--- a/community/go/0001-runtime-use-entire-address-space-on-32-bit.patch
+++ /dev/null
@@ -1,303 +0,0 @@
-From 2017cb5f865afa822b2d2355eeb4d5555395143c Mon Sep 17 00:00:00 2001
-From: Elias Naur <elias.naur@gmail.com>
-Date: Wed, 9 Mar 2016 10:00:12 +0100
-Subject: [PATCH] runtime: use entire address space on 32 bit
-
-In issue #13992, Russ mentioned that the heap bitmap footprint was
-halved but that the bitmap size calculation hadn't been updated. This
-presents the opportunity to either halve the bitmap size or double
-the addressable virtual space. This CL doubles the addressable virtual
-space. On 32 bit this can be tweaked further to allow the bitmap to
-cover the entire 4GB virtual address space, removing a failure mode
-if the kernel hands out memory with a too low address.
-
-First, fix the calculation and double _MaxArena32 to cover 4GB virtual
-memory space with the same bitmap size (256 MB).
-
-Then, allow the fallback mode for the initial memory reservation
-on 32 bit (or 64 bit with too little available virtual memory) to not
-include space for the arena. mheap.sysAlloc will automatically reserve
-additional space when the existing arena is full.
-
-Finally, set arena_start to 0 in 32 bit mode, so that any address is
-acceptable for subsequent (additional) reservations.
-
-Before, the bitmap was always located just before arena_start, so
-fix the two places relying on that assumption: Point the otherwise unused
-mheap.bitmap to one byte after the end of the bitmap, and use it for
-bitmap addressing instead of arena_start.
-
-With arena_start set to 0 on 32 bit, the cgoInRange check is no longer a
-sufficient check for Go pointers. Introduce and call inHeapOrStack to
-check whether a pointer is to the Go heap or stack.
-
-While we're here, remove sysReserveHigh which seems to be unused.
-
-Fixes #13992
-
-Change-Id: I592b513148a50b9d3967b5c5d94b86b3ec39acc2
-Reviewed-on: https://go-review.googlesource.com/20471
-Reviewed-by: Austin Clements <austin@google.com>
-Run-TryBot: Austin Clements <austin@google.com>
-TryBot-Result: Gobot Gobot <gobot@golang.org>
-(cherry picked from commit e6ec82067a9068c93db6e7041017060a1c863452)
----
- src/runtime/cgocall.go | 2 +-
- src/runtime/malloc.go | 82 +++++++++++++++++++-------------------------------
- src/runtime/mbitmap.go | 4 +--
- src/runtime/mheap.go | 24 ++++++++++++++-
- 4 files changed, 57 insertions(+), 55 deletions(-)
-
-diff --git a/src/runtime/cgocall.go b/src/runtime/cgocall.go
-index fef8add..30c3a67 100644
---- a/src/runtime/cgocall.go
-+++ b/src/runtime/cgocall.go
-@@ -568,7 +568,7 @@ func cgoIsGoPointer(p unsafe.Pointer) bool {
- return false
- }
-
-- if cgoInRange(p, mheap_.arena_start, mheap_.arena_used) {
-+ if inHeapOrStack(uintptr(p)) {
- return true
- }
-
-diff --git a/src/runtime/malloc.go b/src/runtime/malloc.go
-index b520c68..f3c6e88 100644
---- a/src/runtime/malloc.go
-+++ b/src/runtime/malloc.go
-@@ -173,7 +173,7 @@ const (
- // Page number (address>>pageShift)
- type pageID uintptr
-
--const _MaxArena32 = 2 << 30
-+const _MaxArena32 = 1<<32 - 1
-
- // OS-defined helpers:
- //
-@@ -229,8 +229,8 @@ func mallocinit() {
- limit = 0
-
- // Set up the allocation arena, a contiguous area of memory where
-- // allocated data will be found. The arena begins with a bitmap large
-- // enough to hold 4 bits per allocated word.
-+ // allocated data will be found. The arena begins with a bitmap large
-+ // enough to hold 2 bits per allocated word.
- if sys.PtrSize == 8 && (limit == 0 || limit > 1<<30) {
- // On a 64-bit machine, allocate from a single contiguous reservation.
- // 512 GB (MaxMem) should be big enough for now.
-@@ -262,7 +262,7 @@ func mallocinit() {
- // translation buffers, the user address space is limited to 39 bits
- // On darwin/arm64, the address space is even smaller.
- arenaSize := round(_MaxMem, _PageSize)
-- bitmapSize = arenaSize / (sys.PtrSize * 8 / 4)
-+ bitmapSize = arenaSize / (sys.PtrSize * 8 / 2)
- spansSize = arenaSize / _PageSize * sys.PtrSize
- spansSize = round(spansSize, _PageSize)
- for i := 0; i <= 0x7f; i++ {
-@@ -287,32 +287,26 @@ func mallocinit() {
- // with a giant virtual address space reservation.
- // Instead we map the memory information bitmap
- // immediately after the data segment, large enough
-- // to handle another 2GB of mappings (256 MB),
-+ // to handle the entire 4GB address space (256 MB),
- // along with a reservation for an initial arena.
- // When that gets used up, we'll start asking the kernel
-- // for any memory anywhere and hope it's in the 2GB
-- // following the bitmap (presumably the executable begins
-- // near the bottom of memory, so we'll have to use up
-- // most of memory before the kernel resorts to giving out
-- // memory before the beginning of the text segment).
-- //
-- // Alternatively we could reserve 512 MB bitmap, enough
-- // for 4GB of mappings, and then accept any memory the
-- // kernel threw at us, but normally that's a waste of 512 MB
-- // of address space, which is probably too much in a 32-bit world.
-+ // for any memory anywhere.
-
- // If we fail to allocate, try again with a smaller arena.
- // This is necessary on Android L where we share a process
- // with ART, which reserves virtual memory aggressively.
-+ // In the worst case, fall back to a 0-sized initial arena,
-+ // in the hope that subsequent reservations will succeed.
- arenaSizes := []uintptr{
- 512 << 20,
- 256 << 20,
- 128 << 20,
-+ 0,
- }
-
- for _, arenaSize := range arenaSizes {
-- bitmapSize = _MaxArena32 / (sys.PtrSize * 8 / 4)
-- spansSize = _MaxArena32 / _PageSize * sys.PtrSize
-+ bitmapSize = (_MaxArena32 + 1) / (sys.PtrSize * 8 / 2)
-+ spansSize = (_MaxArena32 + 1) / _PageSize * sys.PtrSize
- if limit > 0 && arenaSize+bitmapSize+spansSize > limit {
- bitmapSize = (limit / 9) &^ ((1 << _PageShift) - 1)
- arenaSize = bitmapSize * 8
-@@ -347,10 +341,16 @@ func mallocinit() {
- p1 := round(p, _PageSize)
-
- mheap_.spans = (**mspan)(unsafe.Pointer(p1))
-- mheap_.bitmap = p1 + spansSize
-- mheap_.arena_start = p1 + (spansSize + bitmapSize)
-- mheap_.arena_used = mheap_.arena_start
-+ mheap_.bitmap = p1 + spansSize + bitmapSize
-+ if sys.PtrSize == 4 {
-+ // Set arena_start such that we can accept memory
-+ // reservations located anywhere in the 4GB virtual space.
-+ mheap_.arena_start = 0
-+ } else {
-+ mheap_.arena_start = p1 + (spansSize + bitmapSize)
-+ }
- mheap_.arena_end = p + pSize
-+ mheap_.arena_used = p1 + (spansSize + bitmapSize)
- mheap_.arena_reserved = reserved
-
- if mheap_.arena_start&(_PageSize-1) != 0 {
-@@ -364,36 +364,17 @@ func mallocinit() {
- _g_.m.mcache = allocmcache()
- }
-
--// sysReserveHigh reserves space somewhere high in the address space.
--// sysReserve doesn't actually reserve the full amount requested on
--// 64-bit systems, because of problems with ulimit. Instead it checks
--// that it can get the first 64 kB and assumes it can grab the rest as
--// needed. This doesn't work well with the "let the kernel pick an address"
--// mode, so don't do that. Pick a high address instead.
--func sysReserveHigh(n uintptr, reserved *bool) unsafe.Pointer {
-- if sys.PtrSize == 4 {
-- return sysReserve(nil, n, reserved)
-- }
--
-- for i := 0; i <= 0x7f; i++ {
-- p := uintptr(i)<<40 | uintptrMask&(0x00c0<<32)
-- *reserved = false
-- p = uintptr(sysReserve(unsafe.Pointer(p), n, reserved))
-- if p != 0 {
-- return unsafe.Pointer(p)
-- }
-- }
--
-- return sysReserve(nil, n, reserved)
--}
--
-+// sysAlloc allocates the next n bytes from the heap arena. The
-+// returned pointer is always _PageSize aligned and between
-+// h.arena_start and h.arena_end. sysAlloc returns nil on failure.
-+// There is no corresponding free function.
- func (h *mheap) sysAlloc(n uintptr) unsafe.Pointer {
- if n > h.arena_end-h.arena_used {
- // We are in 32-bit mode, maybe we didn't use all possible address space yet.
- // Reserve some more space.
- p_size := round(n+_PageSize, 256<<20)
- new_end := h.arena_end + p_size // Careful: can overflow
-- if h.arena_end <= new_end && new_end <= h.arena_start+_MaxArena32 {
-+ if h.arena_end <= new_end && new_end-h.arena_start-1 <= _MaxArena32 {
- // TODO: It would be bad if part of the arena
- // is reserved and part is not.
- var reserved bool
-@@ -404,7 +385,7 @@ func (h *mheap) sysAlloc(n uintptr) unsafe.Pointer {
- if p == h.arena_end {
- h.arena_end = new_end
- h.arena_reserved = reserved
-- } else if h.arena_start <= p && p+p_size <= h.arena_start+_MaxArena32 {
-+ } else if h.arena_start <= p && p+p_size-h.arena_start-1 <= _MaxArena32 {
- // Keep everything page-aligned.
- // Our pages are bigger than hardware pages.
- h.arena_end = p + p_size
-@@ -441,23 +422,22 @@ func (h *mheap) sysAlloc(n uintptr) unsafe.Pointer {
- }
-
- // If using 64-bit, our reservation is all we have.
-- if h.arena_end-h.arena_start >= _MaxArena32 {
-+ if h.arena_end-h.arena_start > _MaxArena32 {
- return nil
- }
-
- // On 32-bit, once the reservation is gone we can
-- // try to get memory at a location chosen by the OS
-- // and hope that it is in the range we allocated bitmap for.
-+ // try to get memory at a location chosen by the OS.
- p_size := round(n, _PageSize) + _PageSize
- p := uintptr(sysAlloc(p_size, &memstats.heap_sys))
- if p == 0 {
- return nil
- }
-
-- if p < h.arena_start || uintptr(p)+p_size-h.arena_start >= _MaxArena32 {
-+ if p < h.arena_start || uintptr(p)+p_size-h.arena_start > _MaxArena32 {
- top := ^uintptr(0)
-- if top-h.arena_start > _MaxArena32 {
-- top = h.arena_start + _MaxArena32
-+ if top-h.arena_start-1 > _MaxArena32 {
-+ top = h.arena_start + _MaxArena32 + 1
- }
- print("runtime: memory allocated by OS (", hex(p), ") not in usable range [", hex(h.arena_start), ",", hex(top), ")\n")
- sysFree(unsafe.Pointer(p), p_size, &memstats.heap_sys)
-diff --git a/src/runtime/mbitmap.go b/src/runtime/mbitmap.go
-index 336d4d8..9a5d83b 100644
---- a/src/runtime/mbitmap.go
-+++ b/src/runtime/mbitmap.go
-@@ -145,7 +145,7 @@ func (h *mheap) mapBits(arena_used uintptr) {
- return
- }
-
-- sysMap(unsafe.Pointer(h.arena_start-n), n-h.bitmap_mapped, h.arena_reserved, &memstats.gc_sys)
-+ sysMap(unsafe.Pointer(h.bitmap-n), n-h.bitmap_mapped, h.arena_reserved, &memstats.gc_sys)
- h.bitmap_mapped = n
- }
-
-@@ -166,7 +166,7 @@ type heapBits struct {
- func heapBitsForAddr(addr uintptr) heapBits {
- // 2 bits per work, 4 pairs per byte, and a mask is hard coded.
- off := (addr - mheap_.arena_start) / sys.PtrSize
-- return heapBits{(*uint8)(unsafe.Pointer(mheap_.arena_start - off/4 - 1)), uint32(off & 3)}
-+ return heapBits{(*uint8)(unsafe.Pointer(mheap_.bitmap - off/4 - 1)), uint32(off & 3)}
- }
-
- // heapBitsForSpan returns the heapBits for the span base address base.
-diff --git a/src/runtime/mheap.go b/src/runtime/mheap.go
-index a153df0..1db6a49 100644
---- a/src/runtime/mheap.go
-+++ b/src/runtime/mheap.go
-@@ -46,7 +46,7 @@ type mheap struct {
- nsmallfree [_NumSizeClasses]uint64 // number of frees for small objects (<=maxsmallsize)
-
- // range of addresses we might see in the heap
-- bitmap uintptr
-+ bitmap uintptr // Points to one byte past the end of the bitmap
- bitmap_mapped uintptr
- arena_start uintptr
- arena_used uintptr // always mHeap_Map{Bits,Spans} before updating
-@@ -217,6 +217,28 @@ func inheap(b uintptr) bool {
- return true
- }
-
-+// inHeapOrStack is a variant of inheap that returns true for pointers into stack spans.
-+//go:nowritebarrier
-+//go:nosplit
-+func inHeapOrStack(b uintptr) bool {
-+ if b == 0 || b < mheap_.arena_start || b >= mheap_.arena_used {
-+ return false
-+ }
-+ // Not a beginning of a block, consult span table to find the block beginning.
-+ s := h_spans[(b-mheap_.arena_start)>>_PageShift]
-+ if s == nil || b < s.base() {
-+ return false
-+ }
-+ switch s.state {
-+ case mSpanInUse:
-+ return b < s.limit
-+ case _MSpanStack:
-+ return b < s.base()+s.npages<<_PageShift
-+ default:
-+ return false
-+ }
-+}
-+
- // TODO: spanOf and spanOfUnchecked are open-coded in a lot of places.
- // Use the functions instead.
-
---
-2.8.2
-
diff --git a/community/go/stacksize.patch b/community/go/stacksize.patch
deleted file mode 100644
index 33119ad670..0000000000
--- a/community/go/stacksize.patch
+++ /dev/null
@@ -1,120 +0,0 @@
-diff --git a/src/runtime/cgo/gcc_linux_386.c b/src/runtime/cgo/gcc_linux_386.c
-index 30fe92b..7a6b846 100644
---- a/src/runtime/cgo/gcc_linux_386.c
-+++ b/src/runtime/cgo/gcc_linux_386.c
-@@ -14,6 +14,8 @@ static void (*setg_gcc)(void*);
- void (*x_cgo_inittls)(void);
- void* (*x_cgo_threadentry)(void*);
-
-+#define MIN_STACKSIZE (256 * 1024)
-+
- void
- x_cgo_init(G *g, void (*setg)(void*))
- {
-@@ -51,6 +53,10 @@ _cgo_sys_thread_start(ThreadStart *ts)
- pthread_attr_init(&attr);
- size = 0;
- pthread_attr_getstacksize(&attr, &size);
-+ // musl libc set default stack size to 80k which is a bit small
-+ if (size < MIN_STACKSIZE)
-+ if (pthread_attr_setstacksize(&attr, MIN_STACKSIZE) == 0)
-+ size = MIN_STACKSIZE;
- // Leave stacklo=0 and set stackhi=size; mstack will do the rest.
- ts->g->stackhi = size;
- err = pthread_create(&p, &attr, threadentry, ts);
-diff --git a/src/runtime/cgo/gcc_linux_amd64.c b/src/runtime/cgo/gcc_linux_amd64.c
-index 0c34c66..6c090b7 100644
---- a/src/runtime/cgo/gcc_linux_amd64.c
-+++ b/src/runtime/cgo/gcc_linux_amd64.c
-@@ -16,6 +16,8 @@ static void (*setg_gcc)(void*);
- void (*x_cgo_inittls)(void);
- void* (*x_cgo_threadentry)(void*);
-
-+#define MIN_STACKSIZE (512 * 1024)
-+
- void
- x_cgo_init(G* g, void (*setg)(void*))
- {
-@@ -68,6 +70,10 @@ _cgo_sys_thread_start(ThreadStart *ts)
-
- pthread_attr_init(&attr);
- pthread_attr_getstacksize(&attr, &size);
-+ // musl libc set default stack size to 80k which is a bit small
-+ if (size < MIN_STACKSIZE)
-+ if (pthread_attr_setstacksize(&attr, MIN_STACKSIZE) == 0)
-+ size = MIN_STACKSIZE;
- // Leave stacklo=0 and set stackhi=size; mstack will do the rest.
- ts->g->stackhi = size;
- err = pthread_create(&p, &attr, threadentry, ts);
-diff --git a/src/runtime/cgo/gcc_linux_arm.c b/src/runtime/cgo/gcc_linux_arm.c
-index 945c3f1..7abe781 100644
---- a/src/runtime/cgo/gcc_linux_arm.c
-+++ b/src/runtime/cgo/gcc_linux_arm.c
-@@ -12,6 +12,8 @@ static void *threadentry(void*);
- void (*x_cgo_inittls)(void **tlsg, void **tlsbase);
- void (*setg_gcc)(void*);
-
-+#define MIN_STACKSIZE (256*1024)
-+
- void
- _cgo_sys_thread_start(ThreadStart *ts)
- {
-@@ -31,6 +33,10 @@ _cgo_sys_thread_start(ThreadStart *ts)
- pthread_attr_init(&attr);
- size = 0;
- pthread_attr_getstacksize(&attr, &size);
-+ // musl libc set default stack size to 80k which is a bit small
-+ if (size < MIN_STACKSIZE)
-+ if (pthread_attr_setstacksize(&attr, MIN_STACKSIZE) == 0)
-+ size = MIN_STACKSIZE;
- // Leave stacklo=0 and set stackhi=size; mstack will do the rest.
- ts->g->stackhi = size;
- err = pthread_create(&p, &attr, threadentry, ts);
-diff --git a/src/runtime/cgo/gcc_linux_arm64.c b/src/runtime/cgo/gcc_linux_arm64.c
-index ca9ba0b..8e0ff14 100644
---- a/src/runtime/cgo/gcc_linux_arm64.c
-+++ b/src/runtime/cgo/gcc_linux_arm64.c
-@@ -12,6 +12,8 @@ static void *threadentry(void*);
- void (*x_cgo_inittls)(void **tlsg, void **tlsbase);
- void (*setg_gcc)(void*);
-
-+#define MIN_STACKSIZE (512*1024)
-+
- void
- _cgo_sys_thread_start(ThreadStart *ts)
- {
-@@ -31,6 +33,10 @@ _cgo_sys_thread_start(ThreadStart *ts)
- pthread_attr_init(&attr);
- size = 0;
- pthread_attr_getstacksize(&attr, &size);
-+ // musl libc set default stack size to 80k which is a bit small
-+ if (size < MIN_STACKSIZE)
-+ if (pthread_attr_setstacksize(&attr, MIN_STACKSIZE) == 0)
-+ size = MIN_STACKSIZE;
- // Leave stacklo=0 and set stackhi=size; mstack will do the rest.
- ts->g->stackhi = size;
- err = pthread_create(&p, &attr, threadentry, ts);
-diff --git a/src/runtime/cgo/gcc_linux_ppc64x.c b/src/runtime/cgo/gcc_linux_ppc64x.c
-index fb19805..88911c6 100644
---- a/src/runtime/cgo/gcc_linux_ppc64x.c
-+++ b/src/runtime/cgo/gcc_linux_ppc64x.c
-@@ -14,6 +14,8 @@ static void *threadentry(void*);
- void (*x_cgo_inittls)(void **tlsg, void **tlsbase);
- static void (*setg_gcc)(void*);
-
-+#define MIN_STACKSIZE (512*1024)
-+
- void
- x_cgo_init(G *g, void (*setg)(void*), void **tlsbase)
- {
-@@ -41,6 +43,10 @@ _cgo_sys_thread_start(ThreadStart *ts)
-
- pthread_attr_init(&attr);
- pthread_attr_getstacksize(&attr, &size);
-+ // musl libc set default stack size to 80k which is a bit small
-+ if (size < MIN_STACKSIZE)
-+ if (pthread_attr_setstacksize(&attr, MIN_STACKSIZE) == 0)
-+ size = MIN_STACKSIZE;
- // Leave stacklo=0 and set stackhi=size; mstack will do the rest.
- ts->g->stackhi = size;
- err = pthread_create(&p, &attr, threadentry, ts);