Merge pull request #47743 from thaJeztah/bump_go_winio
vendor: github.com/Microsoft/go-winio v0.6.2 (for go1.22 compatibility)
diff --git a/Dockerfile b/Dockerfile
index 6e8d61c..97a0ac9 100644
--- a/Dockerfile
+++ b/Dockerfile
@@ -8,7 +8,7 @@
ARG VPNKIT_VERSION=0.5.0
ARG DOCKERCLI_REPOSITORY="https://github.com/docker/cli.git"
-ARG DOCKERCLI_VERSION=v26.0.0
+ARG DOCKERCLI_VERSION=v26.1.0
# cli version used for integration-cli tests
ARG DOCKERCLI_INTEGRATION_REPOSITORY="https://github.com/docker/cli.git"
ARG DOCKERCLI_INTEGRATION_VERSION=v17.06.2-ce
diff --git a/vendor.mod b/vendor.mod
index d3ed9cb..3d92406 100644
--- a/vendor.mod
+++ b/vendor.mod
@@ -136,7 +136,7 @@
github.com/bmizerany/assert v0.0.0-20160611221934-b7ed37b82869 // indirect
github.com/cenkalti/backoff/v4 v4.2.1 // indirect
github.com/cespare/xxhash/v2 v2.2.0 // indirect
- github.com/cilium/ebpf v0.11.0 // indirect
+ github.com/cilium/ebpf v0.12.3 // indirect
github.com/container-storage-interface/spec v1.5.0 // indirect
github.com/containerd/cgroups v1.1.0 // indirect
github.com/containerd/console v1.0.4 // indirect
diff --git a/vendor.sum b/vendor.sum
index cc95147..3652f07 100644
--- a/vendor.sum
+++ b/vendor.sum
@@ -124,8 +124,8 @@
github.com/chzyer/logex v1.1.10/go.mod h1:+Ywpsq7O8HXn0nuIou7OrIPyXbp3wmkHB+jjWRnGsAI=
github.com/chzyer/readline v0.0.0-20180603132655-2972be24d48e/go.mod h1:nSuG5e5PlCu98SY8svDHJxuZscDgtXS6KTTbou5AhLI=
github.com/chzyer/test v0.0.0-20180213035817-a1ea475d72b1/go.mod h1:Q3SI9o4m/ZMnBNeIyt5eFwwo7qiLfzFZmjNmxjkiQlU=
-github.com/cilium/ebpf v0.11.0 h1:V8gS/bTCCjX9uUnkUFUpPsksM8n1lXBAvHcpiFk1X2Y=
-github.com/cilium/ebpf v0.11.0/go.mod h1:WE7CZAnqOL2RouJ4f1uyNhqr2P4CCvXFIqdRDUgWsVs=
+github.com/cilium/ebpf v0.12.3 h1:8ht6F9MquybnY97at+VDZb3eQQr8ev79RueWeVaEcG4=
+github.com/cilium/ebpf v0.12.3/go.mod h1:TctK1ivibvI3znr66ljgi4hqOT8EYQjz1KWBfb1UVgM=
github.com/circonus-labs/circonus-gometrics v2.3.1+incompatible/go.mod h1:nmEj6Dob7S7YxXgwXpfOuvO54S+tGdZdw9fuRZt25Ag=
github.com/circonus-labs/circonusllhist v0.1.3/go.mod h1:kMXHVDlOchFAehlya5ePtbp5jckzBHf4XRpQvBOLI+I=
github.com/client9/misspell v0.3.4/go.mod h1:qj6jICC3Q7zFZvVWo7KLAzC3yx5G7kyvSDkc90ppPyw=
diff --git a/vendor/github.com/cilium/ebpf/.clang-format b/vendor/github.com/cilium/ebpf/.clang-format
index 3f74dc0..0ff4257 100644
--- a/vendor/github.com/cilium/ebpf/.clang-format
+++ b/vendor/github.com/cilium/ebpf/.clang-format
@@ -4,6 +4,9 @@
AlignAfterOpenBracket: DontAlign
AlignConsecutiveAssignments: true
AlignEscapedNewlines: DontAlign
+# mkdocs annotations in source code are written as trailing comments
+# and alignment pushes these really far away from the content.
+AlignTrailingComments: false
AlwaysBreakBeforeMultilineStrings: true
AlwaysBreakTemplateDeclarations: false
AllowAllParametersOfDeclarationOnNextLine: false
@@ -16,4 +19,7 @@
ColumnLimit: 1000
# Go compiler comments need to stay unindented.
CommentPragmas: '^go:.*'
+# linux/bpf.h needs to be included before bpf/bpf_helpers.h for types like __u64
+# and sorting makes this impossible.
+SortIncludes: false
...
diff --git a/vendor/github.com/cilium/ebpf/.golangci.yaml b/vendor/github.com/cilium/ebpf/.golangci.yaml
index 06743df..65f91b9 100644
--- a/vendor/github.com/cilium/ebpf/.golangci.yaml
+++ b/vendor/github.com/cilium/ebpf/.golangci.yaml
@@ -1,15 +1,7 @@
---
-issues:
- exclude-rules:
- # syscall param structs will have unused fields in Go code.
- - path: syscall.*.go
- linters:
- - structcheck
-
linters:
disable-all: true
enable:
- - errcheck
- goimports
- gosimple
- govet
@@ -19,8 +11,3 @@
- typecheck
- unused
- gofmt
-
- # Could be enabled later:
- # - gocyclo
- # - maligned
- # - gosec
diff --git a/vendor/github.com/cilium/ebpf/Makefile b/vendor/github.com/cilium/ebpf/Makefile
index abcd6c1..0fa8cdc 100644
--- a/vendor/github.com/cilium/ebpf/Makefile
+++ b/vendor/github.com/cilium/ebpf/Makefile
@@ -1,9 +1,9 @@
# The development version of clang is distributed as the 'clang' binary,
# while stable/released versions have a version number attached.
# Pin the default clang to a stable version.
-CLANG ?= clang-14
-STRIP ?= llvm-strip-14
-OBJCOPY ?= llvm-objcopy-14
+CLANG ?= clang-17
+STRIP ?= llvm-strip-17
+OBJCOPY ?= llvm-objcopy-17
CFLAGS := -O2 -g -Wall -Werror $(CFLAGS)
CI_KERNEL_URL ?= https://github.com/cilium/ci-kernels/raw/master/
@@ -21,12 +21,9 @@
IMAGE := $(shell cat ${REPODIR}/testdata/docker/IMAGE)
VERSION := $(shell cat ${REPODIR}/testdata/docker/VERSION)
-
-# clang <8 doesn't tag relocs properly (STT_NOTYPE)
-# clang 9 is the first version emitting BTF
TARGETS := \
- testdata/loader-clang-7 \
- testdata/loader-clang-9 \
+ testdata/loader-clang-11 \
+ testdata/loader-clang-14 \
testdata/loader-$(CLANG) \
testdata/manyprogs \
testdata/btf_map_init \
@@ -36,6 +33,7 @@
testdata/invalid_btf_map_init \
testdata/strings \
testdata/freplace \
+ testdata/fentry_fexit \
testdata/iproute2_map_compat \
testdata/map_spin_lock \
testdata/subprog_reloc \
@@ -45,6 +43,7 @@
testdata/kfunc \
testdata/invalid-kfunc \
testdata/kfunc-kmod \
+ testdata/constants \
btf/testdata/relocs \
btf/testdata/relocs_read \
btf/testdata/relocs_read_tgt \
@@ -56,22 +55,26 @@
# Build all ELF binaries using a containerized LLVM toolchain.
container-all:
- +${CONTAINER_ENGINE} run --rm -ti ${CONTAINER_RUN_ARGS} \
+ +${CONTAINER_ENGINE} run --rm -t ${CONTAINER_RUN_ARGS} \
-v "${REPODIR}":/ebpf -w /ebpf --env MAKEFLAGS \
- --env CFLAGS="-fdebug-prefix-map=/ebpf=." \
--env HOME="/tmp" \
+ --env BPF2GO_CC="$(CLANG)" \
+ --env BPF2GO_FLAGS="-fdebug-prefix-map=/ebpf=. $(CFLAGS)" \
"${IMAGE}:${VERSION}" \
make all
# (debug) Drop the user into a shell inside the container as root.
+# Set BPF2GO_ envs to make 'make generate' just work.
container-shell:
${CONTAINER_ENGINE} run --rm -ti \
-v "${REPODIR}":/ebpf -w /ebpf \
+ --env BPF2GO_CC="$(CLANG)" \
+ --env BPF2GO_FLAGS="-fdebug-prefix-map=/ebpf=. $(CFLAGS)" \
"${IMAGE}:${VERSION}"
clean:
- -$(RM) testdata/*.elf
- -$(RM) btf/testdata/*.elf
+ find "$(CURDIR)" -name "*.elf" -delete
+ find "$(CURDIR)" -name "*.o" -delete
format:
find . -type f -name "*.c" | xargs clang-format -i
@@ -80,9 +83,6 @@
ln -srf testdata/loader-$(CLANG)-el.elf testdata/loader-el.elf
ln -srf testdata/loader-$(CLANG)-eb.elf testdata/loader-eb.elf
-# $BPF_CLANG is used in go:generate invocations.
-generate: export BPF_CLANG := $(CLANG)
-generate: export BPF_CFLAGS := $(CFLAGS)
generate:
go generate ./...
@@ -103,13 +103,12 @@
$(STRIP) -g $@
.PHONY: generate-btf
-generate-btf: KERNEL_VERSION?=5.19
+generate-btf: KERNEL_VERSION?=6.1.29
generate-btf:
$(eval TMP := $(shell mktemp -d))
- curl -fL "$(CI_KERNEL_URL)/linux-$(KERNEL_VERSION).bz" -o "$(TMP)/bzImage"
- /lib/modules/$(uname -r)/build/scripts/extract-vmlinux "$(TMP)/bzImage" > "$(TMP)/vmlinux"
+ curl -fL "$(CI_KERNEL_URL)/linux-$(KERNEL_VERSION)-amd64.tgz" -o "$(TMP)/linux.tgz"
+ tar xvf "$(TMP)/linux.tgz" -C "$(TMP)" --strip-components=2 ./boot/vmlinuz ./lib/modules
+ /lib/modules/$(shell uname -r)/build/scripts/extract-vmlinux "$(TMP)/vmlinuz" > "$(TMP)/vmlinux"
$(OBJCOPY) --dump-section .BTF=/dev/stdout "$(TMP)/vmlinux" /dev/null | gzip > "btf/testdata/vmlinux.btf.gz"
- curl -fL "$(CI_KERNEL_URL)/linux-$(KERNEL_VERSION)-selftests-bpf.tgz" -o "$(TMP)/selftests.tgz"
- tar -xf "$(TMP)/selftests.tgz" --to-stdout tools/testing/selftests/bpf/bpf_testmod/bpf_testmod.ko | \
- $(OBJCOPY) --dump-section .BTF="btf/testdata/btf_testmod.btf" - /dev/null
+ find "$(TMP)/modules" -type f -name bpf_testmod.ko -exec $(OBJCOPY) --dump-section .BTF="btf/testdata/btf_testmod.btf" {} /dev/null \;
$(RM) -r "$(TMP)"
diff --git a/vendor/github.com/cilium/ebpf/README.md b/vendor/github.com/cilium/ebpf/README.md
index eff08d8..81235a6 100644
--- a/vendor/github.com/cilium/ebpf/README.md
+++ b/vendor/github.com/cilium/ebpf/README.md
@@ -2,7 +2,7 @@
[![PkgGoDev](https://pkg.go.dev/badge/github.com/cilium/ebpf)](https://pkg.go.dev/github.com/cilium/ebpf)
-![HoneyGopher](.github/images/cilium-ebpf.png)
+![HoneyGopher](docs/ebpf/ebpf-go.png)
ebpf-go is a pure Go library that provides utilities for loading, compiling, and
debugging eBPF programs. It has minimal external dependencies and is intended to
diff --git a/vendor/github.com/cilium/ebpf/asm/alu.go b/vendor/github.com/cilium/ebpf/asm/alu.go
index 3f60245..282233d 100644
--- a/vendor/github.com/cilium/ebpf/asm/alu.go
+++ b/vendor/github.com/cilium/ebpf/asm/alu.go
@@ -1,26 +1,26 @@
package asm
-//go:generate stringer -output alu_string.go -type=Source,Endianness,ALUOp
+//go:generate go run golang.org/x/tools/cmd/stringer@latest -output alu_string.go -type=Source,Endianness,ALUOp
// Source of ALU / ALU64 / Branch operations
//
-// msb lsb
-// +----+-+---+
-// |op |S|cls|
-// +----+-+---+
-type Source uint8
+// msb lsb
+// +------------+-+---+
+// | op |S|cls|
+// +------------+-+---+
+type Source uint16
-const sourceMask OpCode = 0x08
+const sourceMask OpCode = 0x0008
// Source bitmask
const (
// InvalidSource is returned by getters when invoked
// on non ALU / branch OpCodes.
- InvalidSource Source = 0xff
+ InvalidSource Source = 0xffff
// ImmSource src is from constant
- ImmSource Source = 0x00
+ ImmSource Source = 0x0000
// RegSource src is from register
- RegSource Source = 0x08
+ RegSource Source = 0x0008
)
// The Endianness of a byte swap instruction.
@@ -39,46 +39,56 @@
// ALUOp are ALU / ALU64 operations
//
-// msb lsb
-// +----+-+---+
-// |OP |s|cls|
-// +----+-+---+
-type ALUOp uint8
+// msb lsb
+// +-------+----+-+---+
+// | EXT | OP |s|cls|
+// +-------+----+-+---+
+type ALUOp uint16
-const aluMask OpCode = 0xf0
+const aluMask OpCode = 0x3ff0
const (
// InvalidALUOp is returned by getters when invoked
// on non ALU OpCodes
- InvalidALUOp ALUOp = 0xff
+ InvalidALUOp ALUOp = 0xffff
// Add - addition
- Add ALUOp = 0x00
+ Add ALUOp = 0x0000
// Sub - subtraction
- Sub ALUOp = 0x10
+ Sub ALUOp = 0x0010
// Mul - multiplication
- Mul ALUOp = 0x20
+ Mul ALUOp = 0x0020
// Div - division
- Div ALUOp = 0x30
+ Div ALUOp = 0x0030
+ // SDiv - signed division
+ SDiv ALUOp = Div + 0x0100
// Or - bitwise or
- Or ALUOp = 0x40
+ Or ALUOp = 0x0040
// And - bitwise and
- And ALUOp = 0x50
+ And ALUOp = 0x0050
// LSh - bitwise shift left
- LSh ALUOp = 0x60
+ LSh ALUOp = 0x0060
// RSh - bitwise shift right
- RSh ALUOp = 0x70
+ RSh ALUOp = 0x0070
// Neg - sign/unsign signing bit
- Neg ALUOp = 0x80
+ Neg ALUOp = 0x0080
// Mod - modulo
- Mod ALUOp = 0x90
+ Mod ALUOp = 0x0090
+ // SMod - signed modulo
+ SMod ALUOp = Mod + 0x0100
// Xor - bitwise xor
- Xor ALUOp = 0xa0
+ Xor ALUOp = 0x00a0
// Mov - move value from one place to another
- Mov ALUOp = 0xb0
- // ArSh - arithmatic shift
- ArSh ALUOp = 0xc0
+ Mov ALUOp = 0x00b0
+ // MovSX8 - move lower 8 bits, sign extended upper bits of target
+ MovSX8 ALUOp = Mov + 0x0100
+ // MovSX16 - move lower 16 bits, sign extended upper bits of target
+ MovSX16 ALUOp = Mov + 0x0200
+ // MovSX32 - move lower 32 bits, sign extended upper bits of target
+ MovSX32 ALUOp = Mov + 0x0300
+ // ArSh - arithmetic shift
+ ArSh ALUOp = 0x00c0
// Swap - endian conversions
- Swap ALUOp = 0xd0
+ Swap ALUOp = 0x00d0
)
// HostTo converts from host to another endianness.
@@ -102,6 +112,27 @@
}
}
+// BSwap unconditionally reverses the order of bytes in a register.
+func BSwap(dst Register, size Size) Instruction {
+ var imm int64
+ switch size {
+ case Half:
+ imm = 16
+ case Word:
+ imm = 32
+ case DWord:
+ imm = 64
+ default:
+ return Instruction{OpCode: InvalidOpCode}
+ }
+
+ return Instruction{
+ OpCode: OpCode(ALU64Class).SetALUOp(Swap),
+ Dst: dst,
+ Constant: imm,
+ }
+}
+
// Op returns the OpCode for an ALU operation with a given source.
func (op ALUOp) Op(source Source) OpCode {
return OpCode(ALU64Class).SetALUOp(op).SetSource(source)
diff --git a/vendor/github.com/cilium/ebpf/asm/alu_string.go b/vendor/github.com/cilium/ebpf/asm/alu_string.go
index 72d3fe6..35b406b 100644
--- a/vendor/github.com/cilium/ebpf/asm/alu_string.go
+++ b/vendor/github.com/cilium/ebpf/asm/alu_string.go
@@ -8,7 +8,7 @@
// An "invalid array index" compiler error signifies that the constant values have changed.
// Re-run the stringer command to generate them again.
var x [1]struct{}
- _ = x[InvalidSource-255]
+ _ = x[InvalidSource-65535]
_ = x[ImmSource-0]
_ = x[RegSource-8]
}
@@ -25,7 +25,7 @@
return _Source_name_0
case i == 8:
return _Source_name_1
- case i == 255:
+ case i == 65535:
return _Source_name_2
default:
return "Source(" + strconv.FormatInt(int64(i), 10) + ")"
@@ -62,41 +62,51 @@
// An "invalid array index" compiler error signifies that the constant values have changed.
// Re-run the stringer command to generate them again.
var x [1]struct{}
- _ = x[InvalidALUOp-255]
+ _ = x[InvalidALUOp-65535]
_ = x[Add-0]
_ = x[Sub-16]
_ = x[Mul-32]
_ = x[Div-48]
+ _ = x[SDiv-304]
_ = x[Or-64]
_ = x[And-80]
_ = x[LSh-96]
_ = x[RSh-112]
_ = x[Neg-128]
_ = x[Mod-144]
+ _ = x[SMod-400]
_ = x[Xor-160]
_ = x[Mov-176]
+ _ = x[MovSX8-432]
+ _ = x[MovSX16-688]
+ _ = x[MovSX32-944]
_ = x[ArSh-192]
_ = x[Swap-208]
}
-const _ALUOp_name = "AddSubMulDivOrAndLShRShNegModXorMovArShSwapInvalidALUOp"
+const _ALUOp_name = "AddSubMulDivOrAndLShRShNegModXorMovArShSwapSDivSModMovSX8MovSX16MovSX32InvalidALUOp"
var _ALUOp_map = map[ALUOp]string{
- 0: _ALUOp_name[0:3],
- 16: _ALUOp_name[3:6],
- 32: _ALUOp_name[6:9],
- 48: _ALUOp_name[9:12],
- 64: _ALUOp_name[12:14],
- 80: _ALUOp_name[14:17],
- 96: _ALUOp_name[17:20],
- 112: _ALUOp_name[20:23],
- 128: _ALUOp_name[23:26],
- 144: _ALUOp_name[26:29],
- 160: _ALUOp_name[29:32],
- 176: _ALUOp_name[32:35],
- 192: _ALUOp_name[35:39],
- 208: _ALUOp_name[39:43],
- 255: _ALUOp_name[43:55],
+ 0: _ALUOp_name[0:3],
+ 16: _ALUOp_name[3:6],
+ 32: _ALUOp_name[6:9],
+ 48: _ALUOp_name[9:12],
+ 64: _ALUOp_name[12:14],
+ 80: _ALUOp_name[14:17],
+ 96: _ALUOp_name[17:20],
+ 112: _ALUOp_name[20:23],
+ 128: _ALUOp_name[23:26],
+ 144: _ALUOp_name[26:29],
+ 160: _ALUOp_name[29:32],
+ 176: _ALUOp_name[32:35],
+ 192: _ALUOp_name[35:39],
+ 208: _ALUOp_name[39:43],
+ 304: _ALUOp_name[43:47],
+ 400: _ALUOp_name[47:51],
+ 432: _ALUOp_name[51:57],
+ 688: _ALUOp_name[57:64],
+ 944: _ALUOp_name[64:71],
+ 65535: _ALUOp_name[71:83],
}
func (i ALUOp) String() string {
diff --git a/vendor/github.com/cilium/ebpf/asm/func.go b/vendor/github.com/cilium/ebpf/asm/func.go
index 18f6a75..84a40b2 100644
--- a/vendor/github.com/cilium/ebpf/asm/func.go
+++ b/vendor/github.com/cilium/ebpf/asm/func.go
@@ -1,6 +1,6 @@
package asm
-//go:generate stringer -output func_string.go -type=BuiltinFunc
+//go:generate go run golang.org/x/tools/cmd/stringer@latest -output func_string.go -type=BuiltinFunc
// BuiltinFunc is a built-in eBPF function.
type BuiltinFunc int32
diff --git a/vendor/github.com/cilium/ebpf/asm/instruction.go b/vendor/github.com/cilium/ebpf/asm/instruction.go
index ef01eaa..67cd39d 100644
--- a/vendor/github.com/cilium/ebpf/asm/instruction.go
+++ b/vendor/github.com/cilium/ebpf/asm/instruction.go
@@ -60,6 +60,34 @@
}
ins.Offset = int16(bo.Uint16(data[2:4]))
+
+ if ins.OpCode.Class().IsALU() {
+ switch ins.OpCode.ALUOp() {
+ case Div:
+ if ins.Offset == 1 {
+ ins.OpCode = ins.OpCode.SetALUOp(SDiv)
+ ins.Offset = 0
+ }
+ case Mod:
+ if ins.Offset == 1 {
+ ins.OpCode = ins.OpCode.SetALUOp(SMod)
+ ins.Offset = 0
+ }
+ case Mov:
+ switch ins.Offset {
+ case 8:
+ ins.OpCode = ins.OpCode.SetALUOp(MovSX8)
+ ins.Offset = 0
+ case 16:
+ ins.OpCode = ins.OpCode.SetALUOp(MovSX16)
+ ins.Offset = 0
+ case 32:
+ ins.OpCode = ins.OpCode.SetALUOp(MovSX32)
+ ins.Offset = 0
+ }
+ }
+ }
+
// Convert to int32 before widening to int64
// to ensure the signed bit is carried over.
ins.Constant = int64(int32(bo.Uint32(data[4:8])))
@@ -106,8 +134,38 @@
return 0, fmt.Errorf("can't marshal registers: %s", err)
}
+ if ins.OpCode.Class().IsALU() {
+ newOffset := int16(0)
+ switch ins.OpCode.ALUOp() {
+ case SDiv:
+ ins.OpCode = ins.OpCode.SetALUOp(Div)
+ newOffset = 1
+ case SMod:
+ ins.OpCode = ins.OpCode.SetALUOp(Mod)
+ newOffset = 1
+ case MovSX8:
+ ins.OpCode = ins.OpCode.SetALUOp(Mov)
+ newOffset = 8
+ case MovSX16:
+ ins.OpCode = ins.OpCode.SetALUOp(Mov)
+ newOffset = 16
+ case MovSX32:
+ ins.OpCode = ins.OpCode.SetALUOp(Mov)
+ newOffset = 32
+ }
+ if newOffset != 0 && ins.Offset != 0 {
+ return 0, fmt.Errorf("extended ALU opcodes should have an .Offset of 0: %s", ins)
+ }
+ ins.Offset = newOffset
+ }
+
+ op, err := ins.OpCode.bpfOpCode()
+ if err != nil {
+ return 0, err
+ }
+
data := make([]byte, InstructionSize)
- data[0] = byte(ins.OpCode)
+ data[0] = op
data[1] = byte(regs)
bo.PutUint16(data[2:4], uint16(ins.Offset))
bo.PutUint32(data[4:8], uint32(cons))
@@ -298,9 +356,9 @@
goto ref
}
- fmt.Fprintf(f, "%v ", op)
switch cls := op.Class(); {
case cls.isLoadOrStore():
+ fmt.Fprintf(f, "%v ", op)
switch op.Mode() {
case ImmMode:
fmt.Fprintf(f, "dst: %s imm: %d", ins.Dst, ins.Constant)
@@ -308,21 +366,30 @@
fmt.Fprintf(f, "imm: %d", ins.Constant)
case IndMode:
fmt.Fprintf(f, "dst: %s src: %s imm: %d", ins.Dst, ins.Src, ins.Constant)
- case MemMode:
+ case MemMode, MemSXMode:
fmt.Fprintf(f, "dst: %s src: %s off: %d imm: %d", ins.Dst, ins.Src, ins.Offset, ins.Constant)
case XAddMode:
fmt.Fprintf(f, "dst: %s src: %s", ins.Dst, ins.Src)
}
case cls.IsALU():
- fmt.Fprintf(f, "dst: %s ", ins.Dst)
- if op.ALUOp() == Swap || op.Source() == ImmSource {
+ fmt.Fprintf(f, "%v", op)
+ if op == Swap.Op(ImmSource) {
+ fmt.Fprintf(f, "%d", ins.Constant)
+ }
+
+ fmt.Fprintf(f, " dst: %s ", ins.Dst)
+ switch {
+ case op.ALUOp() == Swap:
+ break
+ case op.Source() == ImmSource:
fmt.Fprintf(f, "imm: %d", ins.Constant)
- } else {
+ default:
fmt.Fprintf(f, "src: %s", ins.Src)
}
case cls.IsJump():
+ fmt.Fprintf(f, "%v ", op)
switch jop := op.JumpOp(); jop {
case Call:
switch ins.Src {
@@ -336,6 +403,13 @@
fmt.Fprint(f, BuiltinFunc(ins.Constant))
}
+ case Ja:
+ if ins.OpCode.Class() == Jump32Class {
+ fmt.Fprintf(f, "imm: %d", ins.Constant)
+ } else {
+ fmt.Fprintf(f, "off: %d", ins.Offset)
+ }
+
default:
fmt.Fprintf(f, "dst: %s off: %d ", ins.Dst, ins.Offset)
if op.Source() == ImmSource {
@@ -344,6 +418,8 @@
fmt.Fprintf(f, "src: %s", ins.Src)
}
}
+ default:
+ fmt.Fprintf(f, "%v ", op)
}
ref:
@@ -772,7 +848,8 @@
}
switch {
- case ins.IsFunctionReference() && ins.Constant == -1:
+ case ins.IsFunctionReference() && ins.Constant == -1,
+ ins.OpCode == Ja.opCode(Jump32Class, ImmSource) && ins.Constant == -1:
symOffset, ok := symbolOffsets[ins.Reference()]
if !ok {
return fmt.Errorf("%s at insn %d: symbol %q: %w", ins.OpCode, i, ins.Reference(), ErrUnsatisfiedProgramReference)
diff --git a/vendor/github.com/cilium/ebpf/asm/jump.go b/vendor/github.com/cilium/ebpf/asm/jump.go
index 2c8a3db..2738d73 100644
--- a/vendor/github.com/cilium/ebpf/asm/jump.go
+++ b/vendor/github.com/cilium/ebpf/asm/jump.go
@@ -1,6 +1,6 @@
package asm
-//go:generate stringer -output jump_string.go -type=JumpOp
+//go:generate go run golang.org/x/tools/cmd/stringer@latest -output jump_string.go -type=JumpOp
// JumpOp affect control flow.
//
@@ -10,7 +10,7 @@
// +----+-+---+
type JumpOp uint8
-const jumpMask OpCode = aluMask
+const jumpMask OpCode = 0xf0
const (
// InvalidJumpOp is returned by getters when invoked
@@ -103,13 +103,21 @@
}
func (op JumpOp) opCode(class Class, source Source) OpCode {
- if op == Exit || op == Call || op == Ja {
+ if op == Exit || op == Call {
return InvalidOpCode
}
return OpCode(class).SetJumpOp(op).SetSource(source)
}
+// LongJump returns a jump always instruction with a range of [-2^31, 2^31 - 1].
+func LongJump(label string) Instruction {
+ return Instruction{
+ OpCode: Ja.opCode(Jump32Class, ImmSource),
+ Constant: -1,
+ }.WithReference(label)
+}
+
// Label adjusts PC to the address of the label.
func (op JumpOp) Label(label string) Instruction {
if op == Call {
diff --git a/vendor/github.com/cilium/ebpf/asm/load_store.go b/vendor/github.com/cilium/ebpf/asm/load_store.go
index f109497..cdb5c5c 100644
--- a/vendor/github.com/cilium/ebpf/asm/load_store.go
+++ b/vendor/github.com/cilium/ebpf/asm/load_store.go
@@ -1,6 +1,6 @@
package asm
-//go:generate stringer -output load_store_string.go -type=Mode,Size
+//go:generate go run golang.org/x/tools/cmd/stringer@latest -output load_store_string.go -type=Mode,Size
// Mode for load and store operations
//
@@ -24,6 +24,8 @@
IndMode Mode = 0x40
// MemMode - load from memory
MemMode Mode = 0x60
+ // MemSXMode - load from memory, sign extension
+ MemSXMode Mode = 0x80
// XAddMode - add atomically across processors.
XAddMode Mode = 0xc0
)
@@ -73,6 +75,11 @@
return OpCode(LdXClass).SetMode(MemMode).SetSize(size)
}
+// LoadMemSXOp returns the OpCode to load a value of given size from memory sign extended.
+func LoadMemSXOp(size Size) OpCode {
+ return OpCode(LdXClass).SetMode(MemSXMode).SetSize(size)
+}
+
// LoadMem emits `dst = *(size *)(src + offset)`.
func LoadMem(dst, src Register, offset int16, size Size) Instruction {
return Instruction{
@@ -83,6 +90,20 @@
}
}
+// LoadMemSX emits `dst = *(size *)(src + offset)` but sign extends dst.
+func LoadMemSX(dst, src Register, offset int16, size Size) Instruction {
+ if size == DWord {
+ return Instruction{OpCode: InvalidOpCode}
+ }
+
+ return Instruction{
+ OpCode: LoadMemSXOp(size),
+ Dst: dst,
+ Src: src,
+ Offset: offset,
+ }
+}
+
// LoadImmOp returns the OpCode to load an immediate of given size.
//
// As of kernel 4.20, only DWord size is accepted.
diff --git a/vendor/github.com/cilium/ebpf/asm/load_store_string.go b/vendor/github.com/cilium/ebpf/asm/load_store_string.go
index 76d29a0..c480803 100644
--- a/vendor/github.com/cilium/ebpf/asm/load_store_string.go
+++ b/vendor/github.com/cilium/ebpf/asm/load_store_string.go
@@ -13,6 +13,7 @@
_ = x[AbsMode-32]
_ = x[IndMode-64]
_ = x[MemMode-96]
+ _ = x[MemSXMode-128]
_ = x[XAddMode-192]
}
@@ -21,8 +22,9 @@
_Mode_name_1 = "AbsMode"
_Mode_name_2 = "IndMode"
_Mode_name_3 = "MemMode"
- _Mode_name_4 = "XAddMode"
- _Mode_name_5 = "InvalidMode"
+ _Mode_name_4 = "MemSXMode"
+ _Mode_name_5 = "XAddMode"
+ _Mode_name_6 = "InvalidMode"
)
func (i Mode) String() string {
@@ -35,10 +37,12 @@
return _Mode_name_2
case i == 96:
return _Mode_name_3
- case i == 192:
+ case i == 128:
return _Mode_name_4
- case i == 255:
+ case i == 192:
return _Mode_name_5
+ case i == 255:
+ return _Mode_name_6
default:
return "Mode(" + strconv.FormatInt(int64(i), 10) + ")"
}
diff --git a/vendor/github.com/cilium/ebpf/asm/opcode.go b/vendor/github.com/cilium/ebpf/asm/opcode.go
index 9e3c30b..1dfd0b1 100644
--- a/vendor/github.com/cilium/ebpf/asm/opcode.go
+++ b/vendor/github.com/cilium/ebpf/asm/opcode.go
@@ -5,7 +5,7 @@
"strings"
)
-//go:generate stringer -output opcode_string.go -type=Class
+//go:generate go run golang.org/x/tools/cmd/stringer@latest -output opcode_string.go -type=Class
// Class of operations
//
@@ -66,18 +66,43 @@
return cls.IsJump() || cls.IsALU()
}
-// OpCode is a packed eBPF opcode.
+// OpCode represents a single operation.
+// It is not a 1:1 mapping to real eBPF opcodes.
//
-// Its encoding is defined by a Class value:
+// The encoding varies based on a 3-bit Class:
//
-// msb lsb
-// +----+-+---+
-// | ???? |CLS|
-// +----+-+---+
-type OpCode uint8
+// 7 6 5 4 3 2 1 0 7 6 5 4 3 2 1 0
+// ??? | CLS
+//
+// For ALUClass and ALUCLass32:
+//
+// 7 6 5 4 3 2 1 0 7 6 5 4 3 2 1 0
+// OPC |S| CLS
+//
+// For LdClass, LdXclass, StClass and StXClass:
+//
+// 7 6 5 4 3 2 1 0 7 6 5 4 3 2 1 0
+// 0 | MDE |SIZ| CLS
+//
+// For JumpClass, Jump32Class:
+//
+// 7 6 5 4 3 2 1 0 7 6 5 4 3 2 1 0
+// 0 | OPC |S| CLS
+type OpCode uint16
// InvalidOpCode is returned by setters on OpCode
-const InvalidOpCode OpCode = 0xff
+const InvalidOpCode OpCode = 0xffff
+
+// bpfOpCode returns the actual BPF opcode.
+func (op OpCode) bpfOpCode() (byte, error) {
+ const opCodeMask = 0xff
+
+ if !valid(op, opCodeMask) {
+ return 0, fmt.Errorf("invalid opcode %x", op)
+ }
+
+ return byte(op & opCodeMask), nil
+}
// rawInstructions returns the number of BPF instructions required
// to encode this opcode.
@@ -147,7 +172,7 @@
jumpOp := JumpOp(op & jumpMask)
// Some JumpOps are only supported by JumpClass, not Jump32Class.
- if op.Class() == Jump32Class && (jumpOp == Exit || jumpOp == Call || jumpOp == Ja) {
+ if op.Class() == Jump32Class && (jumpOp == Exit || jumpOp == Call) {
return InvalidJumpOp
}
@@ -234,17 +259,24 @@
}
case class.IsALU():
+ if op.ALUOp() == Swap && op.Class() == ALU64Class {
+ // B to make BSwap, uncontitional byte swap
+ f.WriteString("B")
+ }
+
f.WriteString(op.ALUOp().String())
if op.ALUOp() == Swap {
- // Width for Endian is controlled by Constant
- f.WriteString(op.Endianness().String())
+ if op.Class() == ALUClass {
+ // Width for Endian is controlled by Constant
+ f.WriteString(op.Endianness().String())
+ }
} else {
+ f.WriteString(strings.TrimSuffix(op.Source().String(), "Source"))
+
if class == ALUClass {
f.WriteString("32")
}
-
- f.WriteString(strings.TrimSuffix(op.Source().String(), "Source"))
}
case class.IsJump():
@@ -254,7 +286,7 @@
f.WriteString("32")
}
- if jop := op.JumpOp(); jop != Exit && jop != Call {
+ if jop := op.JumpOp(); jop != Exit && jop != Call && jop != Ja {
f.WriteString(strings.TrimSuffix(op.Source().String(), "Source"))
}
diff --git a/vendor/github.com/cilium/ebpf/btf/btf.go b/vendor/github.com/cilium/ebpf/btf/btf.go
index 86eb7d6..80f64d7 100644
--- a/vendor/github.com/cilium/ebpf/btf/btf.go
+++ b/vendor/github.com/cilium/ebpf/btf/btf.go
@@ -14,7 +14,6 @@
"github.com/cilium/ebpf/internal"
"github.com/cilium/ebpf/internal/sys"
- "github.com/cilium/ebpf/internal/unix"
)
const btfMagic = 0xeB9F
@@ -47,51 +46,13 @@
// Includes all struct flavors and types with the same name.
namedTypes map[essentialName][]Type
- // String table from ELF, may be nil.
+ // String table from ELF.
strings *stringTable
// Byte order of the ELF we decoded the spec from, may be nil.
byteOrder binary.ByteOrder
}
-var btfHeaderLen = binary.Size(&btfHeader{})
-
-type btfHeader struct {
- Magic uint16
- Version uint8
- Flags uint8
- HdrLen uint32
-
- TypeOff uint32
- TypeLen uint32
- StringOff uint32
- StringLen uint32
-}
-
-// typeStart returns the offset from the beginning of the .BTF section
-// to the start of its type entries.
-func (h *btfHeader) typeStart() int64 {
- return int64(h.HdrLen + h.TypeOff)
-}
-
-// stringStart returns the offset from the beginning of the .BTF section
-// to the start of its string table.
-func (h *btfHeader) stringStart() int64 {
- return int64(h.HdrLen + h.StringOff)
-}
-
-// newSpec creates a Spec containing only Void.
-func newSpec() *Spec {
- return &Spec{
- []Type{(*Void)(nil)},
- map[Type]TypeID{(*Void)(nil): 0},
- 0,
- make(map[essentialName][]Type),
- nil,
- nil,
- }
-}
-
// LoadSpec opens file and calls LoadSpecFromReader on it.
func LoadSpec(file string) (*Spec, error) {
fh, err := os.Open(file)
@@ -240,10 +201,6 @@
return nil, fmt.Errorf("can't use split BTF as base")
}
- if base.strings == nil {
- return nil, fmt.Errorf("parse split BTF: base must be loaded from an ELF")
- }
-
baseStrings = base.strings
firstTypeID, err = base.nextTypeID()
@@ -252,12 +209,7 @@
}
}
- rawTypes, rawStrings, err := parseBTF(btf, bo, baseStrings)
- if err != nil {
- return nil, err
- }
-
- types, err := inflateRawTypes(rawTypes, rawStrings, base)
+ types, rawStrings, err := parseBTF(btf, bo, baseStrings, base)
if err != nil {
return nil, err
}
@@ -365,12 +317,12 @@
}
defer file.Close()
- spec, err := loadSpecFromELF(file)
+ spec, err := LoadSpecFromReader(file)
return spec, true, err
}
// findVMLinux scans multiple well-known paths for vmlinux kernel images.
-func findVMLinux() (*internal.SafeELFFile, error) {
+func findVMLinux() (*os.File, error) {
release, err := internal.KernelRelease()
if err != nil {
return nil, err
@@ -389,7 +341,7 @@
}
for _, loc := range locations {
- file, err := internal.OpenSafeELFFile(fmt.Sprintf(loc, release))
+ file, err := os.Open(fmt.Sprintf(loc, release))
if errors.Is(err, os.ErrNotExist) {
continue
}
@@ -399,37 +351,6 @@
return nil, fmt.Errorf("no BTF found for kernel version %s: %w", release, internal.ErrNotSupported)
}
-// parseBTFHeader parses the header of the .BTF section.
-func parseBTFHeader(r io.Reader, bo binary.ByteOrder) (*btfHeader, error) {
- var header btfHeader
- if err := binary.Read(r, bo, &header); err != nil {
- return nil, fmt.Errorf("can't read header: %v", err)
- }
-
- if header.Magic != btfMagic {
- return nil, fmt.Errorf("incorrect magic value %v", header.Magic)
- }
-
- if header.Version != 1 {
- return nil, fmt.Errorf("unexpected version %v", header.Version)
- }
-
- if header.Flags != 0 {
- return nil, fmt.Errorf("unsupported flags %v", header.Flags)
- }
-
- remainder := int64(header.HdrLen) - int64(binary.Size(&header))
- if remainder < 0 {
- return nil, errors.New("header length shorter than btfHeader size")
- }
-
- if _, err := io.CopyN(internal.DiscardZeroes{}, r, remainder); err != nil {
- return nil, fmt.Errorf("header padding: %v", err)
- }
-
- return &header, nil
-}
-
func guessRawBTFByteOrder(r io.ReaderAt) binary.ByteOrder {
buf := new(bufio.Reader)
for _, bo := range []binary.ByteOrder{
@@ -447,7 +368,7 @@
// parseBTF reads a .BTF section into memory and parses it into a list of
// raw types and a string table.
-func parseBTF(btf io.ReaderAt, bo binary.ByteOrder, baseStrings *stringTable) ([]rawType, *stringTable, error) {
+func parseBTF(btf io.ReaderAt, bo binary.ByteOrder, baseStrings *stringTable, base *Spec) ([]Type, *stringTable, error) {
buf := internal.NewBufferedSectionReader(btf, 0, math.MaxInt64)
header, err := parseBTFHeader(buf, bo)
if err != nil {
@@ -461,12 +382,12 @@
}
buf.Reset(io.NewSectionReader(btf, header.typeStart(), int64(header.TypeLen)))
- rawTypes, err := readTypes(buf, bo, header.TypeLen)
+ types, err := readAndInflateTypes(buf, bo, header.TypeLen, rawStrings, base)
if err != nil {
- return nil, nil, fmt.Errorf("can't read types: %w", err)
+ return nil, nil, err
}
- return rawTypes, rawStrings, nil
+ return types, rawStrings, nil
}
type symbol struct {
@@ -773,97 +694,3 @@
iter.index++
return true
}
-
-// haveBTF attempts to load a BTF blob containing an Int. It should pass on any
-// kernel that supports BPF_BTF_LOAD.
-var haveBTF = internal.NewFeatureTest("BTF", "4.18", func() error {
- // 0-length anonymous integer
- err := probeBTF(&Int{})
- if errors.Is(err, unix.EINVAL) || errors.Is(err, unix.EPERM) {
- return internal.ErrNotSupported
- }
- return err
-})
-
-// haveMapBTF attempts to load a minimal BTF blob containing a Var. It is
-// used as a proxy for .bss, .data and .rodata map support, which generally
-// come with a Var and Datasec. These were introduced in Linux 5.2.
-var haveMapBTF = internal.NewFeatureTest("Map BTF (Var/Datasec)", "5.2", func() error {
- if err := haveBTF(); err != nil {
- return err
- }
-
- v := &Var{
- Name: "a",
- Type: &Pointer{(*Void)(nil)},
- }
-
- err := probeBTF(v)
- if errors.Is(err, unix.EINVAL) || errors.Is(err, unix.EPERM) {
- // Treat both EINVAL and EPERM as not supported: creating the map may still
- // succeed without Btf* attrs.
- return internal.ErrNotSupported
- }
- return err
-})
-
-// haveProgBTF attempts to load a BTF blob containing a Func and FuncProto. It
-// is used as a proxy for ext_info (func_info) support, which depends on
-// Func(Proto) by definition.
-var haveProgBTF = internal.NewFeatureTest("Program BTF (func/line_info)", "5.0", func() error {
- if err := haveBTF(); err != nil {
- return err
- }
-
- fn := &Func{
- Name: "a",
- Type: &FuncProto{Return: (*Void)(nil)},
- }
-
- err := probeBTF(fn)
- if errors.Is(err, unix.EINVAL) || errors.Is(err, unix.EPERM) {
- return internal.ErrNotSupported
- }
- return err
-})
-
-var haveFuncLinkage = internal.NewFeatureTest("BTF func linkage", "5.6", func() error {
- if err := haveProgBTF(); err != nil {
- return err
- }
-
- fn := &Func{
- Name: "a",
- Type: &FuncProto{Return: (*Void)(nil)},
- Linkage: GlobalFunc,
- }
-
- err := probeBTF(fn)
- if errors.Is(err, unix.EINVAL) {
- return internal.ErrNotSupported
- }
- return err
-})
-
-func probeBTF(typ Type) error {
- b, err := NewBuilder([]Type{typ})
- if err != nil {
- return err
- }
-
- buf, err := b.Marshal(nil, nil)
- if err != nil {
- return err
- }
-
- fd, err := sys.BtfLoad(&sys.BtfLoadAttr{
- Btf: sys.NewSlicePointer(buf),
- BtfSize: uint32(len(buf)),
- })
-
- if err == nil {
- fd.Close()
- }
-
- return err
-}
diff --git a/vendor/github.com/cilium/ebpf/btf/btf_types.go b/vendor/github.com/cilium/ebpf/btf/btf_types.go
index a253b7c..f0e327a 100644
--- a/vendor/github.com/cilium/ebpf/btf/btf_types.go
+++ b/vendor/github.com/cilium/ebpf/btf/btf_types.go
@@ -2,12 +2,15 @@
import (
"encoding/binary"
+ "errors"
"fmt"
"io"
"unsafe"
+
+ "github.com/cilium/ebpf/internal"
)
-//go:generate stringer -linecomment -output=btf_types_string.go -type=FuncLinkage,VarLinkage,btfKind
+//go:generate go run golang.org/x/tools/cmd/stringer@latest -linecomment -output=btf_types_string.go -type=FuncLinkage,VarLinkage,btfKind
// btfKind describes a Type.
type btfKind uint8
@@ -69,6 +72,63 @@
btfTypeKindFlagMask = 1
)
+var btfHeaderLen = binary.Size(&btfHeader{})
+
+type btfHeader struct {
+ Magic uint16
+ Version uint8
+ Flags uint8
+ HdrLen uint32
+
+ TypeOff uint32
+ TypeLen uint32
+ StringOff uint32
+ StringLen uint32
+}
+
+// typeStart returns the offset from the beginning of the .BTF section
+// to the start of its type entries.
+func (h *btfHeader) typeStart() int64 {
+ return int64(h.HdrLen + h.TypeOff)
+}
+
+// stringStart returns the offset from the beginning of the .BTF section
+// to the start of its string table.
+func (h *btfHeader) stringStart() int64 {
+ return int64(h.HdrLen + h.StringOff)
+}
+
+// parseBTFHeader parses the header of the .BTF section.
+func parseBTFHeader(r io.Reader, bo binary.ByteOrder) (*btfHeader, error) {
+ var header btfHeader
+ if err := binary.Read(r, bo, &header); err != nil {
+ return nil, fmt.Errorf("can't read header: %v", err)
+ }
+
+ if header.Magic != btfMagic {
+ return nil, fmt.Errorf("incorrect magic value %v", header.Magic)
+ }
+
+ if header.Version != 1 {
+ return nil, fmt.Errorf("unexpected version %v", header.Version)
+ }
+
+ if header.Flags != 0 {
+ return nil, fmt.Errorf("unsupported flags %v", header.Flags)
+ }
+
+ remainder := int64(header.HdrLen) - int64(binary.Size(&header))
+ if remainder < 0 {
+ return nil, errors.New("header length shorter than btfHeader size")
+ }
+
+ if _, err := io.CopyN(internal.DiscardZeroes{}, r, remainder); err != nil {
+ return nil, fmt.Errorf("header padding: %v", err)
+ }
+
+ return &header, nil
+}
+
var btfTypeLen = binary.Size(btfType{})
// btfType is equivalent to struct btf_type in Documentation/bpf/btf.rst.
@@ -93,6 +153,19 @@
SizeType uint32
}
+var btfTypeSize = int(unsafe.Sizeof(btfType{}))
+
+func unmarshalBtfType(bt *btfType, b []byte, bo binary.ByteOrder) (int, error) {
+ if len(b) < btfTypeSize {
+ return 0, fmt.Errorf("not enough bytes to unmarshal btfType")
+ }
+
+ bt.NameOff = bo.Uint32(b[0:])
+ bt.Info = bo.Uint32(b[4:])
+ bt.SizeType = bo.Uint32(b[8:])
+ return btfTypeSize, nil
+}
+
func mask(len uint32) uint32 {
return (1 << len) - 1
}
@@ -240,6 +313,17 @@
btfIntBitsShift = 0
)
+var btfIntLen = int(unsafe.Sizeof(btfInt{}))
+
+func unmarshalBtfInt(bi *btfInt, b []byte, bo binary.ByteOrder) (int, error) {
+ if len(b) < btfIntLen {
+ return 0, fmt.Errorf("not enough bytes to unmarshal btfInt")
+ }
+
+ bi.Raw = bo.Uint32(b[0:])
+ return btfIntLen, nil
+}
+
func (bi btfInt) Encoding() IntEncoding {
return IntEncoding(readBits(bi.Raw, btfIntEncodingLen, btfIntEncodingShift))
}
@@ -270,102 +354,166 @@
Nelems uint32
}
+var btfArrayLen = int(unsafe.Sizeof(btfArray{}))
+
+func unmarshalBtfArray(ba *btfArray, b []byte, bo binary.ByteOrder) (int, error) {
+ if len(b) < btfArrayLen {
+ return 0, fmt.Errorf("not enough bytes to unmarshal btfArray")
+ }
+
+ ba.Type = TypeID(bo.Uint32(b[0:]))
+ ba.IndexType = TypeID(bo.Uint32(b[4:]))
+ ba.Nelems = bo.Uint32(b[8:])
+ return btfArrayLen, nil
+}
+
type btfMember struct {
NameOff uint32
Type TypeID
Offset uint32
}
+var btfMemberLen = int(unsafe.Sizeof(btfMember{}))
+
+func unmarshalBtfMembers(members []btfMember, b []byte, bo binary.ByteOrder) (int, error) {
+ off := 0
+ for i := range members {
+ if off+btfMemberLen > len(b) {
+ return 0, fmt.Errorf("not enough bytes to unmarshal btfMember %d", i)
+ }
+
+ members[i].NameOff = bo.Uint32(b[off+0:])
+ members[i].Type = TypeID(bo.Uint32(b[off+4:]))
+ members[i].Offset = bo.Uint32(b[off+8:])
+
+ off += btfMemberLen
+ }
+
+ return off, nil
+}
+
type btfVarSecinfo struct {
Type TypeID
Offset uint32
Size uint32
}
+var btfVarSecinfoLen = int(unsafe.Sizeof(btfVarSecinfo{}))
+
+func unmarshalBtfVarSecInfos(secinfos []btfVarSecinfo, b []byte, bo binary.ByteOrder) (int, error) {
+ off := 0
+ for i := range secinfos {
+ if off+btfVarSecinfoLen > len(b) {
+ return 0, fmt.Errorf("not enough bytes to unmarshal btfVarSecinfo %d", i)
+ }
+
+ secinfos[i].Type = TypeID(bo.Uint32(b[off+0:]))
+ secinfos[i].Offset = bo.Uint32(b[off+4:])
+ secinfos[i].Size = bo.Uint32(b[off+8:])
+
+ off += btfVarSecinfoLen
+ }
+
+ return off, nil
+}
+
type btfVariable struct {
Linkage uint32
}
+var btfVariableLen = int(unsafe.Sizeof(btfVariable{}))
+
+func unmarshalBtfVariable(bv *btfVariable, b []byte, bo binary.ByteOrder) (int, error) {
+ if len(b) < btfVariableLen {
+ return 0, fmt.Errorf("not enough bytes to unmarshal btfVariable")
+ }
+
+ bv.Linkage = bo.Uint32(b[0:])
+ return btfVariableLen, nil
+}
+
type btfEnum struct {
NameOff uint32
Val uint32
}
+var btfEnumLen = int(unsafe.Sizeof(btfEnum{}))
+
+func unmarshalBtfEnums(enums []btfEnum, b []byte, bo binary.ByteOrder) (int, error) {
+ off := 0
+ for i := range enums {
+ if off+btfEnumLen > len(b) {
+ return 0, fmt.Errorf("not enough bytes to unmarshal btfEnum %d", i)
+ }
+
+ enums[i].NameOff = bo.Uint32(b[off+0:])
+ enums[i].Val = bo.Uint32(b[off+4:])
+
+ off += btfEnumLen
+ }
+
+ return off, nil
+}
+
type btfEnum64 struct {
NameOff uint32
ValLo32 uint32
ValHi32 uint32
}
+var btfEnum64Len = int(unsafe.Sizeof(btfEnum64{}))
+
+func unmarshalBtfEnums64(enums []btfEnum64, b []byte, bo binary.ByteOrder) (int, error) {
+ off := 0
+ for i := range enums {
+ if off+btfEnum64Len > len(b) {
+ return 0, fmt.Errorf("not enough bytes to unmarshal btfEnum64 %d", i)
+ }
+
+ enums[i].NameOff = bo.Uint32(b[off+0:])
+ enums[i].ValLo32 = bo.Uint32(b[off+4:])
+ enums[i].ValHi32 = bo.Uint32(b[off+8:])
+
+ off += btfEnum64Len
+ }
+
+ return off, nil
+}
+
type btfParam struct {
NameOff uint32
Type TypeID
}
+var btfParamLen = int(unsafe.Sizeof(btfParam{}))
+
+func unmarshalBtfParams(params []btfParam, b []byte, bo binary.ByteOrder) (int, error) {
+ off := 0
+ for i := range params {
+ if off+btfParamLen > len(b) {
+ return 0, fmt.Errorf("not enough bytes to unmarshal btfParam %d", i)
+ }
+
+ params[i].NameOff = bo.Uint32(b[off+0:])
+ params[i].Type = TypeID(bo.Uint32(b[off+4:]))
+
+ off += btfParamLen
+ }
+
+ return off, nil
+}
+
type btfDeclTag struct {
ComponentIdx uint32
}
-func readTypes(r io.Reader, bo binary.ByteOrder, typeLen uint32) ([]rawType, error) {
- var header btfType
- // because of the interleaving between types and struct members it is difficult to
- // precompute the numbers of raw types this will parse
- // this "guess" is a good first estimation
- sizeOfbtfType := uintptr(btfTypeLen)
- tyMaxCount := uintptr(typeLen) / sizeOfbtfType / 2
- types := make([]rawType, 0, tyMaxCount)
+var btfDeclTagLen = int(unsafe.Sizeof(btfDeclTag{}))
- for id := TypeID(1); ; id++ {
- if err := binary.Read(r, bo, &header); err == io.EOF {
- return types, nil
- } else if err != nil {
- return nil, fmt.Errorf("can't read type info for id %v: %v", id, err)
- }
-
- var data interface{}
- switch header.Kind() {
- case kindInt:
- data = new(btfInt)
- case kindPointer:
- case kindArray:
- data = new(btfArray)
- case kindStruct:
- fallthrough
- case kindUnion:
- data = make([]btfMember, header.Vlen())
- case kindEnum:
- data = make([]btfEnum, header.Vlen())
- case kindForward:
- case kindTypedef:
- case kindVolatile:
- case kindConst:
- case kindRestrict:
- case kindFunc:
- case kindFuncProto:
- data = make([]btfParam, header.Vlen())
- case kindVar:
- data = new(btfVariable)
- case kindDatasec:
- data = make([]btfVarSecinfo, header.Vlen())
- case kindFloat:
- case kindDeclTag:
- data = new(btfDeclTag)
- case kindTypeTag:
- case kindEnum64:
- data = make([]btfEnum64, header.Vlen())
- default:
- return nil, fmt.Errorf("type id %v: unknown kind: %v", id, header.Kind())
- }
-
- if data == nil {
- types = append(types, rawType{header, nil})
- continue
- }
-
- if err := binary.Read(r, bo, data); err != nil {
- return nil, fmt.Errorf("type id %d: kind %v: can't read %T: %v", id, header.Kind(), data, err)
- }
-
- types = append(types, rawType{header, data})
+func unmarshalBtfDeclTag(bdt *btfDeclTag, b []byte, bo binary.ByteOrder) (int, error) {
+ if len(b) < btfDeclTagLen {
+ return 0, fmt.Errorf("not enough bytes to unmarshal btfDeclTag")
}
+
+ bdt.ComponentIdx = bo.Uint32(b[0:])
+ return btfDeclTagLen, nil
}
diff --git a/vendor/github.com/cilium/ebpf/btf/core.go b/vendor/github.com/cilium/ebpf/btf/core.go
index a5c40d3..ded7d43 100644
--- a/vendor/github.com/cilium/ebpf/btf/core.go
+++ b/vendor/github.com/cilium/ebpf/btf/core.go
@@ -18,8 +18,8 @@
// COREFixup is the result of computing a CO-RE relocation for a target.
type COREFixup struct {
kind coreKind
- local uint32
- target uint32
+ local uint64
+ target uint64
// True if there is no valid fixup. The instruction is replaced with an
// invalid dummy.
poison bool
@@ -196,12 +196,12 @@
result[i] = COREFixup{
kind: relo.kind,
- local: uint32(relo.id),
+ local: uint64(relo.id),
// NB: Using relo.id as the target here is incorrect, since
// it doesn't match the BTF we generate on the fly. This isn't
// too bad for now since there are no uses of the local type ID
// in the kernel, yet.
- target: uint32(relo.id),
+ target: uint64(relo.id),
}
continue
}
@@ -311,10 +311,10 @@
// coreCalculateFixup calculates the fixup for a single local type, target type
// and relocation.
func coreCalculateFixup(relo *CORERelocation, target Type, targetID TypeID, bo binary.ByteOrder) (COREFixup, error) {
- fixup := func(local, target uint32) (COREFixup, error) {
+ fixup := func(local, target uint64) (COREFixup, error) {
return COREFixup{kind: relo.kind, local: local, target: target}, nil
}
- fixupWithoutValidation := func(local, target uint32) (COREFixup, error) {
+ fixupWithoutValidation := func(local, target uint64) (COREFixup, error) {
return COREFixup{kind: relo.kind, local: local, target: target, skipLocalValidation: true}, nil
}
poison := func() (COREFixup, error) {
@@ -346,7 +346,7 @@
return fixup(1, 1)
case reloTypeIDTarget:
- return fixup(uint32(relo.id), uint32(targetID))
+ return fixup(uint64(relo.id), uint64(targetID))
case reloTypeSize:
localSize, err := Sizeof(local)
@@ -359,7 +359,7 @@
return zero, err
}
- return fixup(uint32(localSize), uint32(targetSize))
+ return fixup(uint64(localSize), uint64(targetSize))
}
case reloEnumvalValue, reloEnumvalExists:
@@ -376,7 +376,7 @@
return fixup(1, 1)
case reloEnumvalValue:
- return fixup(uint32(localValue.Value), uint32(targetValue.Value))
+ return fixup(localValue.Value, targetValue.Value)
}
case reloFieldByteOffset, reloFieldByteSize, reloFieldExists, reloFieldLShiftU64, reloFieldRShiftU64, reloFieldSigned:
@@ -405,7 +405,7 @@
return fixup(1, 1)
case reloFieldByteOffset:
- return maybeSkipValidation(fixup(localField.offset, targetField.offset))
+ return maybeSkipValidation(fixup(uint64(localField.offset), uint64(targetField.offset)))
case reloFieldByteSize:
localSize, err := Sizeof(localField.Type)
@@ -417,24 +417,24 @@
if err != nil {
return zero, err
}
- return maybeSkipValidation(fixup(uint32(localSize), uint32(targetSize)))
+ return maybeSkipValidation(fixup(uint64(localSize), uint64(targetSize)))
case reloFieldLShiftU64:
- var target uint32
+ var target uint64
if bo == binary.LittleEndian {
targetSize, err := targetField.sizeBits()
if err != nil {
return zero, err
}
- target = uint32(64 - targetField.bitfieldOffset - targetSize)
+ target = uint64(64 - targetField.bitfieldOffset - targetSize)
} else {
loadWidth, err := Sizeof(targetField.Type)
if err != nil {
return zero, err
}
- target = uint32(64 - Bits(loadWidth*8) + targetField.bitfieldOffset)
+ target = uint64(64 - Bits(loadWidth*8) + targetField.bitfieldOffset)
}
return fixupWithoutValidation(0, target)
@@ -444,7 +444,7 @@
return zero, err
}
- return fixupWithoutValidation(0, uint32(64-targetSize))
+ return fixupWithoutValidation(0, uint64(64-targetSize))
case reloFieldSigned:
switch local := UnderlyingType(localField.Type).(type) {
@@ -454,7 +454,7 @@
return zero, fmt.Errorf("target isn't *Enum but %T", targetField.Type)
}
- return fixup(boolToUint32(local.Signed), boolToUint32(target.Signed))
+ return fixup(boolToUint64(local.Signed), boolToUint64(target.Signed))
case *Int:
target, ok := as[*Int](targetField.Type)
if !ok {
@@ -462,8 +462,8 @@
}
return fixup(
- uint32(local.Encoding&Signed),
- uint32(target.Encoding&Signed),
+ uint64(local.Encoding&Signed),
+ uint64(target.Encoding&Signed),
)
default:
return zero, fmt.Errorf("type %T: %w", local, errNoSignedness)
@@ -474,7 +474,7 @@
return zero, ErrNotSupported
}
-func boolToUint32(val bool) uint32 {
+func boolToUint64(val bool) uint64 {
if val {
return 1
}
@@ -799,7 +799,7 @@
if visited[target] {
continue
}
- if len(visited) >= maxTypeDepth {
+ if len(visited) >= maxResolveDepth {
// This check is different than libbpf, which restricts the entire
// path to BPF_CORE_SPEC_MAX_LEN items.
return Member{}, false, fmt.Errorf("type is nested too deep")
@@ -895,7 +895,7 @@
)
for ; l != nil && t != nil; l, t = localTs.Shift(), targetTs.Shift() {
- if depth >= maxTypeDepth {
+ if depth >= maxResolveDepth {
return errors.New("types are nested too deep")
}
diff --git a/vendor/github.com/cilium/ebpf/btf/ext_info.go b/vendor/github.com/cilium/ebpf/btf/ext_info.go
index b764fb7..d85f45a 100644
--- a/vendor/github.com/cilium/ebpf/btf/ext_info.go
+++ b/vendor/github.com/cilium/ebpf/btf/ext_info.go
@@ -16,9 +16,9 @@
// ExtInfos contains ELF section metadata.
type ExtInfos struct {
// The slices are sorted by offset in ascending order.
- funcInfos map[string][]funcInfo
- lineInfos map[string][]lineInfo
- relocationInfos map[string][]coreRelocationInfo
+ funcInfos map[string]FuncInfos
+ lineInfos map[string]LineInfos
+ relocationInfos map[string]CORERelocationInfos
}
// loadExtInfosFromELF parses ext infos from the .BTF.ext section in an ELF.
@@ -34,11 +34,11 @@
return nil, fmt.Errorf("compressed ext_info is not supported")
}
- return loadExtInfos(section.ReaderAt, file.ByteOrder, spec, spec.strings)
+ return loadExtInfos(section.ReaderAt, file.ByteOrder, spec)
}
// loadExtInfos parses bare ext infos.
-func loadExtInfos(r io.ReaderAt, bo binary.ByteOrder, spec *Spec, strings *stringTable) (*ExtInfos, error) {
+func loadExtInfos(r io.ReaderAt, bo binary.ByteOrder, spec *Spec) (*ExtInfos, error) {
// Open unbuffered section reader. binary.Read() calls io.ReadFull on
// the header structs, resulting in one syscall per header.
headerRd := io.NewSectionReader(r, 0, math.MaxInt64)
@@ -53,12 +53,12 @@
}
buf := internal.NewBufferedSectionReader(r, extHeader.funcInfoStart(), int64(extHeader.FuncInfoLen))
- btfFuncInfos, err := parseFuncInfos(buf, bo, strings)
+ btfFuncInfos, err := parseFuncInfos(buf, bo, spec.strings)
if err != nil {
return nil, fmt.Errorf("parsing BTF function info: %w", err)
}
- funcInfos := make(map[string][]funcInfo, len(btfFuncInfos))
+ funcInfos := make(map[string]FuncInfos, len(btfFuncInfos))
for section, bfis := range btfFuncInfos {
funcInfos[section], err = newFuncInfos(bfis, spec)
if err != nil {
@@ -67,14 +67,14 @@
}
buf = internal.NewBufferedSectionReader(r, extHeader.lineInfoStart(), int64(extHeader.LineInfoLen))
- btfLineInfos, err := parseLineInfos(buf, bo, strings)
+ btfLineInfos, err := parseLineInfos(buf, bo, spec.strings)
if err != nil {
return nil, fmt.Errorf("parsing BTF line info: %w", err)
}
- lineInfos := make(map[string][]lineInfo, len(btfLineInfos))
+ lineInfos := make(map[string]LineInfos, len(btfLineInfos))
for section, blis := range btfLineInfos {
- lineInfos[section], err = newLineInfos(blis, strings)
+ lineInfos[section], err = newLineInfos(blis, spec.strings)
if err != nil {
return nil, fmt.Errorf("section %s: line infos: %w", section, err)
}
@@ -86,14 +86,14 @@
var btfCORERelos map[string][]bpfCORERelo
buf = internal.NewBufferedSectionReader(r, extHeader.coreReloStart(coreHeader), int64(coreHeader.COREReloLen))
- btfCORERelos, err = parseCORERelos(buf, bo, strings)
+ btfCORERelos, err = parseCORERelos(buf, bo, spec.strings)
if err != nil {
return nil, fmt.Errorf("parsing CO-RE relocation info: %w", err)
}
- coreRelos := make(map[string][]coreRelocationInfo, len(btfCORERelos))
+ coreRelos := make(map[string]CORERelocationInfos, len(btfCORERelos))
for section, brs := range btfCORERelos {
- coreRelos[section], err = newRelocationInfos(brs, spec, strings)
+ coreRelos[section], err = newRelocationInfos(brs, spec, spec.strings)
if err != nil {
return nil, fmt.Errorf("section %s: CO-RE relocations: %w", section, err)
}
@@ -111,21 +111,31 @@
lineInfos := ei.lineInfos[section]
reloInfos := ei.relocationInfos[section]
+ AssignMetadataToInstructions(insns, funcInfos, lineInfos, reloInfos)
+}
+
+// Assign per-instruction metadata to the instructions in insns.
+func AssignMetadataToInstructions(
+ insns asm.Instructions,
+ funcInfos FuncInfos,
+ lineInfos LineInfos,
+ reloInfos CORERelocationInfos,
+) {
iter := insns.Iterate()
for iter.Next() {
- if len(funcInfos) > 0 && funcInfos[0].offset == iter.Offset {
- *iter.Ins = WithFuncMetadata(*iter.Ins, funcInfos[0].fn)
- funcInfos = funcInfos[1:]
+ if len(funcInfos.infos) > 0 && funcInfos.infos[0].offset == iter.Offset {
+ *iter.Ins = WithFuncMetadata(*iter.Ins, funcInfos.infos[0].fn)
+ funcInfos.infos = funcInfos.infos[1:]
}
- if len(lineInfos) > 0 && lineInfos[0].offset == iter.Offset {
- *iter.Ins = iter.Ins.WithSource(lineInfos[0].line)
- lineInfos = lineInfos[1:]
+ if len(lineInfos.infos) > 0 && lineInfos.infos[0].offset == iter.Offset {
+ *iter.Ins = iter.Ins.WithSource(lineInfos.infos[0].line)
+ lineInfos.infos = lineInfos.infos[1:]
}
- if len(reloInfos) > 0 && reloInfos[0].offset == iter.Offset {
- iter.Ins.Metadata.Set(coreRelocationMeta{}, reloInfos[0].relo)
- reloInfos = reloInfos[1:]
+ if len(reloInfos.infos) > 0 && reloInfos.infos[0].offset == iter.Offset {
+ iter.Ins.Metadata.Set(coreRelocationMeta{}, reloInfos.infos[0].relo)
+ reloInfos.infos = reloInfos.infos[1:]
}
}
}
@@ -323,6 +333,11 @@
return recordSize, nil
}
+// FuncInfos contains a sorted list of func infos.
+type FuncInfos struct {
+ infos []funcInfo
+}
+
// The size of a FuncInfo in BTF wire format.
var FuncInfoSize = uint32(binary.Size(bpfFuncInfo{}))
@@ -359,21 +374,39 @@
}, nil
}
-func newFuncInfos(bfis []bpfFuncInfo, spec *Spec) ([]funcInfo, error) {
- fis := make([]funcInfo, 0, len(bfis))
+func newFuncInfos(bfis []bpfFuncInfo, spec *Spec) (FuncInfos, error) {
+ fis := FuncInfos{
+ infos: make([]funcInfo, 0, len(bfis)),
+ }
for _, bfi := range bfis {
fi, err := newFuncInfo(bfi, spec)
if err != nil {
- return nil, fmt.Errorf("offset %d: %w", bfi.InsnOff, err)
+ return FuncInfos{}, fmt.Errorf("offset %d: %w", bfi.InsnOff, err)
}
- fis = append(fis, *fi)
+ fis.infos = append(fis.infos, *fi)
}
- sort.Slice(fis, func(i, j int) bool {
- return fis[i].offset <= fis[j].offset
+ sort.Slice(fis.infos, func(i, j int) bool {
+ return fis.infos[i].offset <= fis.infos[j].offset
})
return fis, nil
}
+// LoadFuncInfos parses BTF func info in kernel wire format.
+func LoadFuncInfos(reader io.Reader, bo binary.ByteOrder, recordNum uint32, spec *Spec) (FuncInfos, error) {
+ fis, err := parseFuncInfoRecords(
+ reader,
+ bo,
+ FuncInfoSize,
+ recordNum,
+ false,
+ )
+ if err != nil {
+ return FuncInfos{}, fmt.Errorf("parsing BTF func info: %w", err)
+ }
+
+ return newFuncInfos(fis, spec)
+}
+
// marshal into the BTF wire format.
func (fi *funcInfo) marshal(w *bytes.Buffer, b *Builder) error {
id, err := b.Add(fi.fn)
@@ -409,7 +442,7 @@
return nil, err
}
- records, err := parseFuncInfoRecords(r, bo, recordSize, infoHeader.NumInfo)
+ records, err := parseFuncInfoRecords(r, bo, recordSize, infoHeader.NumInfo, true)
if err != nil {
return nil, fmt.Errorf("section %v: %w", secName, err)
}
@@ -421,7 +454,7 @@
// parseFuncInfoRecords parses a stream of func_infos into a funcInfos.
// These records appear after a btf_ext_info_sec header in the func_info
// sub-section of .BTF.ext.
-func parseFuncInfoRecords(r io.Reader, bo binary.ByteOrder, recordSize uint32, recordNum uint32) ([]bpfFuncInfo, error) {
+func parseFuncInfoRecords(r io.Reader, bo binary.ByteOrder, recordSize uint32, recordNum uint32, offsetInBytes bool) ([]bpfFuncInfo, error) {
var out []bpfFuncInfo
var fi bpfFuncInfo
@@ -435,13 +468,15 @@
return nil, fmt.Errorf("can't read function info: %v", err)
}
- if fi.InsnOff%asm.InstructionSize != 0 {
- return nil, fmt.Errorf("offset %v is not aligned with instruction size", fi.InsnOff)
- }
+ if offsetInBytes {
+ if fi.InsnOff%asm.InstructionSize != 0 {
+ return nil, fmt.Errorf("offset %v is not aligned with instruction size", fi.InsnOff)
+ }
- // ELF tracks offset in bytes, the kernel expects raw BPF instructions.
- // Convert as early as possible.
- fi.InsnOff /= asm.InstructionSize
+ // ELF tracks offset in bytes, the kernel expects raw BPF instructions.
+ // Convert as early as possible.
+ fi.InsnOff /= asm.InstructionSize
+ }
out = append(out, fi)
}
@@ -480,6 +515,11 @@
return li.line
}
+// LineInfos contains a sorted list of line infos.
+type LineInfos struct {
+ infos []lineInfo
+}
+
type lineInfo struct {
line *Line
offset asm.RawInstructionOffset
@@ -500,21 +540,37 @@
LineCol uint32
}
-func newLineInfo(li bpfLineInfo, strings *stringTable) (*lineInfo, error) {
+// LoadLineInfos parses BTF line info in kernel wire format.
+func LoadLineInfos(reader io.Reader, bo binary.ByteOrder, recordNum uint32, spec *Spec) (LineInfos, error) {
+ lis, err := parseLineInfoRecords(
+ reader,
+ bo,
+ LineInfoSize,
+ recordNum,
+ false,
+ )
+ if err != nil {
+ return LineInfos{}, fmt.Errorf("parsing BTF line info: %w", err)
+ }
+
+ return newLineInfos(lis, spec.strings)
+}
+
+func newLineInfo(li bpfLineInfo, strings *stringTable) (lineInfo, error) {
line, err := strings.Lookup(li.LineOff)
if err != nil {
- return nil, fmt.Errorf("lookup of line: %w", err)
+ return lineInfo{}, fmt.Errorf("lookup of line: %w", err)
}
fileName, err := strings.Lookup(li.FileNameOff)
if err != nil {
- return nil, fmt.Errorf("lookup of filename: %w", err)
+ return lineInfo{}, fmt.Errorf("lookup of filename: %w", err)
}
lineNumber := li.LineCol >> bpfLineShift
lineColumn := li.LineCol & bpfColumnMax
- return &lineInfo{
+ return lineInfo{
&Line{
fileName,
line,
@@ -525,17 +581,19 @@
}, nil
}
-func newLineInfos(blis []bpfLineInfo, strings *stringTable) ([]lineInfo, error) {
- lis := make([]lineInfo, 0, len(blis))
+func newLineInfos(blis []bpfLineInfo, strings *stringTable) (LineInfos, error) {
+ lis := LineInfos{
+ infos: make([]lineInfo, 0, len(blis)),
+ }
for _, bli := range blis {
li, err := newLineInfo(bli, strings)
if err != nil {
- return nil, fmt.Errorf("offset %d: %w", bli.InsnOff, err)
+ return LineInfos{}, fmt.Errorf("offset %d: %w", bli.InsnOff, err)
}
- lis = append(lis, *li)
+ lis.infos = append(lis.infos, li)
}
- sort.Slice(lis, func(i, j int) bool {
- return lis[i].offset <= lis[j].offset
+ sort.Slice(lis.infos, func(i, j int) bool {
+ return lis.infos[i].offset <= lis.infos[j].offset
})
return lis, nil
}
@@ -595,7 +653,7 @@
return nil, err
}
- records, err := parseLineInfoRecords(r, bo, recordSize, infoHeader.NumInfo)
+ records, err := parseLineInfoRecords(r, bo, recordSize, infoHeader.NumInfo, true)
if err != nil {
return nil, fmt.Errorf("section %v: %w", secName, err)
}
@@ -607,8 +665,7 @@
// parseLineInfoRecords parses a stream of line_infos into a lineInfos.
// These records appear after a btf_ext_info_sec header in the line_info
// sub-section of .BTF.ext.
-func parseLineInfoRecords(r io.Reader, bo binary.ByteOrder, recordSize uint32, recordNum uint32) ([]bpfLineInfo, error) {
- var out []bpfLineInfo
+func parseLineInfoRecords(r io.Reader, bo binary.ByteOrder, recordSize uint32, recordNum uint32, offsetInBytes bool) ([]bpfLineInfo, error) {
var li bpfLineInfo
if exp, got := uint32(binary.Size(li)), recordSize; exp != got {
@@ -616,18 +673,21 @@
return nil, fmt.Errorf("expected LineInfo record size %d, but BTF blob contains %d", exp, got)
}
+ out := make([]bpfLineInfo, 0, recordNum)
for i := uint32(0); i < recordNum; i++ {
if err := binary.Read(r, bo, &li); err != nil {
return nil, fmt.Errorf("can't read line info: %v", err)
}
- if li.InsnOff%asm.InstructionSize != 0 {
- return nil, fmt.Errorf("offset %v is not aligned with instruction size", li.InsnOff)
- }
+ if offsetInBytes {
+ if li.InsnOff%asm.InstructionSize != 0 {
+ return nil, fmt.Errorf("offset %v is not aligned with instruction size", li.InsnOff)
+ }
- // ELF tracks offset in bytes, the kernel expects raw BPF instructions.
- // Convert as early as possible.
- li.InsnOff /= asm.InstructionSize
+ // ELF tracks offset in bytes, the kernel expects raw BPF instructions.
+ // Convert as early as possible.
+ li.InsnOff /= asm.InstructionSize
+ }
out = append(out, li)
}
@@ -661,6 +721,11 @@
return relo
}
+// CORERelocationInfos contains a sorted list of co:re relocation infos.
+type CORERelocationInfos struct {
+ infos []coreRelocationInfo
+}
+
type coreRelocationInfo struct {
relo *CORERelocation
offset asm.RawInstructionOffset
@@ -693,17 +758,19 @@
}, nil
}
-func newRelocationInfos(brs []bpfCORERelo, spec *Spec, strings *stringTable) ([]coreRelocationInfo, error) {
- rs := make([]coreRelocationInfo, 0, len(brs))
+func newRelocationInfos(brs []bpfCORERelo, spec *Spec, strings *stringTable) (CORERelocationInfos, error) {
+ rs := CORERelocationInfos{
+ infos: make([]coreRelocationInfo, 0, len(brs)),
+ }
for _, br := range brs {
relo, err := newRelocationInfo(br, spec, strings)
if err != nil {
- return nil, fmt.Errorf("offset %d: %w", br.InsnOff, err)
+ return CORERelocationInfos{}, fmt.Errorf("offset %d: %w", br.InsnOff, err)
}
- rs = append(rs, *relo)
+ rs.infos = append(rs.infos, *relo)
}
- sort.Slice(rs, func(i, j int) bool {
- return rs[i].offset < rs[j].offset
+ sort.Slice(rs.infos, func(i, j int) bool {
+ return rs.infos[i].offset < rs.infos[j].offset
})
return rs, nil
}
diff --git a/vendor/github.com/cilium/ebpf/btf/feature.go b/vendor/github.com/cilium/ebpf/btf/feature.go
new file mode 100644
index 0000000..6feb08d
--- /dev/null
+++ b/vendor/github.com/cilium/ebpf/btf/feature.go
@@ -0,0 +1,123 @@
+package btf
+
+import (
+ "errors"
+ "math"
+
+ "github.com/cilium/ebpf/internal"
+ "github.com/cilium/ebpf/internal/sys"
+ "github.com/cilium/ebpf/internal/unix"
+)
+
+// haveBTF attempts to load a BTF blob containing an Int. It should pass on any
+// kernel that supports BPF_BTF_LOAD.
+var haveBTF = internal.NewFeatureTest("BTF", "4.18", func() error {
+ // 0-length anonymous integer
+ err := probeBTF(&Int{})
+ if errors.Is(err, unix.EINVAL) || errors.Is(err, unix.EPERM) {
+ return internal.ErrNotSupported
+ }
+ return err
+})
+
+// haveMapBTF attempts to load a minimal BTF blob containing a Var. It is
+// used as a proxy for .bss, .data and .rodata map support, which generally
+// come with a Var and Datasec. These were introduced in Linux 5.2.
+var haveMapBTF = internal.NewFeatureTest("Map BTF (Var/Datasec)", "5.2", func() error {
+ if err := haveBTF(); err != nil {
+ return err
+ }
+
+ v := &Var{
+ Name: "a",
+ Type: &Pointer{(*Void)(nil)},
+ }
+
+ err := probeBTF(v)
+ if errors.Is(err, unix.EINVAL) || errors.Is(err, unix.EPERM) {
+ // Treat both EINVAL and EPERM as not supported: creating the map may still
+ // succeed without Btf* attrs.
+ return internal.ErrNotSupported
+ }
+ return err
+})
+
+// haveProgBTF attempts to load a BTF blob containing a Func and FuncProto. It
+// is used as a proxy for ext_info (func_info) support, which depends on
+// Func(Proto) by definition.
+var haveProgBTF = internal.NewFeatureTest("Program BTF (func/line_info)", "5.0", func() error {
+ if err := haveBTF(); err != nil {
+ return err
+ }
+
+ fn := &Func{
+ Name: "a",
+ Type: &FuncProto{Return: (*Void)(nil)},
+ }
+
+ err := probeBTF(fn)
+ if errors.Is(err, unix.EINVAL) || errors.Is(err, unix.EPERM) {
+ return internal.ErrNotSupported
+ }
+ return err
+})
+
+var haveFuncLinkage = internal.NewFeatureTest("BTF func linkage", "5.6", func() error {
+ if err := haveProgBTF(); err != nil {
+ return err
+ }
+
+ fn := &Func{
+ Name: "a",
+ Type: &FuncProto{Return: (*Void)(nil)},
+ Linkage: GlobalFunc,
+ }
+
+ err := probeBTF(fn)
+ if errors.Is(err, unix.EINVAL) {
+ return internal.ErrNotSupported
+ }
+ return err
+})
+
+var haveEnum64 = internal.NewFeatureTest("ENUM64", "6.0", func() error {
+ if err := haveBTF(); err != nil {
+ return err
+ }
+
+ enum := &Enum{
+ Size: 8,
+ Values: []EnumValue{
+ {"TEST", math.MaxUint32 + 1},
+ },
+ }
+
+ err := probeBTF(enum)
+ if errors.Is(err, unix.EINVAL) {
+ return internal.ErrNotSupported
+ }
+ return err
+})
+
+func probeBTF(typ Type) error {
+ b, err := NewBuilder([]Type{typ})
+ if err != nil {
+ return err
+ }
+
+ buf, err := b.Marshal(nil, nil)
+ if err != nil {
+ return err
+ }
+
+ fd, err := sys.BtfLoad(&sys.BtfLoadAttr{
+ Btf: sys.NewSlicePointer(buf),
+ BtfSize: uint32(len(buf)),
+ })
+
+ if err == nil {
+ fd.Close()
+ }
+
+ return err
+}
diff --git a/vendor/github.com/cilium/ebpf/btf/format.go b/vendor/github.com/cilium/ebpf/btf/format.go
index e852202..5e581b4 100644
--- a/vendor/github.com/cilium/ebpf/btf/format.go
+++ b/vendor/github.com/cilium/ebpf/btf/format.go
@@ -77,7 +77,13 @@
gf.w.WriteString("; const ( ")
for _, ev := range e.Values {
id := gf.enumIdentifier(name, ev.Name)
- fmt.Fprintf(&gf.w, "%s %s = %d; ", id, name, ev.Value)
+ var value any
+ if e.Signed {
+ value = int64(ev.Value)
+ } else {
+ value = ev.Value
+ }
+ fmt.Fprintf(&gf.w, "%s %s = %d; ", id, name, value)
}
gf.w.WriteString(")")
@@ -112,7 +118,7 @@
// uint32
func (gf *GoFormatter) writeTypeLit(typ Type, depth int) error {
depth++
- if depth > maxTypeDepth {
+ if depth > maxResolveDepth {
return errNestedTooDeep
}
@@ -259,7 +265,7 @@
}
depth++
- if depth > maxTypeDepth {
+ if depth > maxResolveDepth {
return errNestedTooDeep
}
@@ -332,7 +338,7 @@
func skipQualifiers(typ Type) Type {
result := typ
- for depth := 0; depth <= maxTypeDepth; depth++ {
+ for depth := 0; depth <= maxResolveDepth; depth++ {
switch v := (result).(type) {
case qualifier:
result = v.qualify()
diff --git a/vendor/github.com/cilium/ebpf/btf/marshal.go b/vendor/github.com/cilium/ebpf/btf/marshal.go
index bfe53b4..0d093c6 100644
--- a/vendor/github.com/cilium/ebpf/btf/marshal.go
+++ b/vendor/github.com/cilium/ebpf/btf/marshal.go
@@ -18,6 +18,8 @@
Order binary.ByteOrder
// Remove function linkage information for compatibility with <5.6 kernels.
StripFuncLinkage bool
+ // Replace Enum64 with a placeholder for compatibility with <6.0 kernels.
+ ReplaceEnum64 bool
}
// KernelMarshalOptions will generate BTF suitable for the current kernel.
@@ -25,6 +27,7 @@
return &MarshalOptions{
Order: internal.NativeEndian,
StripFuncLinkage: haveFuncLinkage() != nil,
+ ReplaceEnum64: haveEnum64() != nil,
}
}
@@ -328,21 +331,13 @@
raw.data, err = e.convertMembers(&raw.btfType, v.Members)
case *Union:
- raw.SetKind(kindUnion)
- raw.SetSize(v.Size)
- raw.data, err = e.convertMembers(&raw.btfType, v.Members)
+ err = e.deflateUnion(&raw, v)
case *Enum:
- raw.SetSize(v.size())
- raw.SetVlen(len(v.Values))
- raw.SetSigned(v.Signed)
-
- if v.has64BitValues() {
- raw.SetKind(kindEnum64)
- raw.data, err = e.deflateEnum64Values(v.Values)
+ if v.Size == 8 {
+ err = e.deflateEnum64(&raw, v)
} else {
- raw.SetKind(kindEnum)
- raw.data, err = e.deflateEnumValues(v.Values)
+ err = e.deflateEnum(&raw, v)
}
case *Fwd:
@@ -415,6 +410,13 @@
return raw.Marshal(e.buf, e.Order)
}
+func (e *encoder) deflateUnion(raw *rawType, union *Union) (err error) {
+ raw.SetKind(kindUnion)
+ raw.SetSize(union.Size)
+ raw.data, err = e.convertMembers(&raw.btfType, union.Members)
+ return
+}
+
func (e *encoder) convertMembers(header *btfType, members []Member) ([]btfMember, error) {
bms := make([]btfMember, 0, len(members))
isBitfield := false
@@ -443,16 +445,32 @@
return bms, nil
}
-func (e *encoder) deflateEnumValues(values []EnumValue) ([]btfEnum, error) {
- bes := make([]btfEnum, 0, len(values))
- for _, value := range values {
+func (e *encoder) deflateEnum(raw *rawType, enum *Enum) (err error) {
+ raw.SetKind(kindEnum)
+ raw.SetSize(enum.Size)
+ raw.SetVlen(len(enum.Values))
+ // Signedness appeared together with ENUM64 support.
+ raw.SetSigned(enum.Signed && !e.ReplaceEnum64)
+ raw.data, err = e.deflateEnumValues(enum)
+ return
+}
+
+func (e *encoder) deflateEnumValues(enum *Enum) ([]btfEnum, error) {
+ bes := make([]btfEnum, 0, len(enum.Values))
+ for _, value := range enum.Values {
nameOff, err := e.strings.Add(value.Name)
if err != nil {
return nil, err
}
- if value.Value > math.MaxUint32 {
- return nil, fmt.Errorf("value of enum %q exceeds 32 bits", value.Name)
+ if enum.Signed {
+ if signedValue := int64(value.Value); signedValue < math.MinInt32 || signedValue > math.MaxInt32 {
+ return nil, fmt.Errorf("value %d of enum %q exceeds 32 bits", signedValue, value.Name)
+ }
+ } else {
+ if value.Value > math.MaxUint32 {
+ return nil, fmt.Errorf("value %d of enum %q exceeds 32 bits", value.Value, value.Name)
+ }
}
bes = append(bes, btfEnum{
@@ -464,6 +482,41 @@
return bes, nil
}
+func (e *encoder) deflateEnum64(raw *rawType, enum *Enum) (err error) {
+ if e.ReplaceEnum64 {
+ // Replace the ENUM64 with a union of fields with the correct size.
+ // This matches libbpf behaviour on purpose.
+ placeholder := &Int{
+ "enum64_placeholder",
+ enum.Size,
+ Unsigned,
+ }
+ if enum.Signed {
+ placeholder.Encoding = Signed
+ }
+ if err := e.allocateID(placeholder); err != nil {
+ return fmt.Errorf("add enum64 placeholder: %w", err)
+ }
+
+ members := make([]Member, 0, len(enum.Values))
+ for _, v := range enum.Values {
+ members = append(members, Member{
+ Name: v.Name,
+ Type: placeholder,
+ })
+ }
+
+ return e.deflateUnion(raw, &Union{enum.Name, enum.Size, members})
+ }
+
+ raw.SetKind(kindEnum64)
+ raw.SetSize(enum.Size)
+ raw.SetVlen(len(enum.Values))
+ raw.SetSigned(enum.Signed)
+ raw.data, err = e.deflateEnum64Values(enum.Values)
+ return
+}
+
func (e *encoder) deflateEnum64Values(values []EnumValue) ([]btfEnum64, error) {
bes := make([]btfEnum64, 0, len(values))
for _, value := range values {
diff --git a/vendor/github.com/cilium/ebpf/btf/strings.go b/vendor/github.com/cilium/ebpf/btf/strings.go
index bc6aff2..114d250 100644
--- a/vendor/github.com/cilium/ebpf/btf/strings.go
+++ b/vendor/github.com/cilium/ebpf/btf/strings.go
@@ -9,11 +9,13 @@
"strings"
"golang.org/x/exp/maps"
+ "golang.org/x/exp/slices"
)
type stringTable struct {
base *stringTable
offsets []uint32
+ prevIdx int
strings []string
}
@@ -60,7 +62,7 @@
return nil, errors.New("first item in string table is non-empty")
}
- return &stringTable{base, offsets, strings}, nil
+ return &stringTable{base, offsets, 0, strings}, nil
}
func splitNull(data []byte, atEOF bool) (advance int, token []byte, err error) {
@@ -83,53 +85,36 @@
}
func (st *stringTable) lookup(offset uint32) (string, error) {
- i := search(st.offsets, offset)
- if i == len(st.offsets) || st.offsets[i] != offset {
+ // Fast path: zero offset is the empty string, looked up frequently.
+ if offset == 0 && st.base == nil {
+ return "", nil
+ }
+
+ // Accesses tend to be globally increasing, so check if the next string is
+ // the one we want. This skips the binary search in about 50% of cases.
+ if st.prevIdx+1 < len(st.offsets) && st.offsets[st.prevIdx+1] == offset {
+ st.prevIdx++
+ return st.strings[st.prevIdx], nil
+ }
+
+ i, found := slices.BinarySearch(st.offsets, offset)
+ if !found {
return "", fmt.Errorf("offset %d isn't start of a string", offset)
}
+ // Set the new increment index, but only if its greater than the current.
+ if i > st.prevIdx+1 {
+ st.prevIdx = i
+ }
+
return st.strings[i], nil
}
-func (st *stringTable) Marshal(w io.Writer) error {
- for _, str := range st.strings {
- _, err := io.WriteString(w, str)
- if err != nil {
- return err
- }
- _, err = w.Write([]byte{0})
- if err != nil {
- return err
- }
- }
- return nil
-}
-
// Num returns the number of strings in the table.
func (st *stringTable) Num() int {
return len(st.strings)
}
-// search is a copy of sort.Search specialised for uint32.
-//
-// Licensed under https://go.dev/LICENSE
-func search(ints []uint32, needle uint32) int {
- // Define f(-1) == false and f(n) == true.
- // Invariant: f(i-1) == false, f(j) == true.
- i, j := 0, len(ints)
- for i < j {
- h := int(uint(i+j) >> 1) // avoid overflow when computing h
- // i ≤ h < j
- if !(ints[h] >= needle) {
- i = h + 1 // preserves f(i-1) == false
- } else {
- j = h // preserves f(j) == true
- }
- }
- // i == j, f(i-1) == false, and f(j) (= f(i)) == true => answer is i.
- return i
-}
-
// stringTableBuilder builds BTF string tables.
type stringTableBuilder struct {
length uint32
diff --git a/vendor/github.com/cilium/ebpf/btf/types.go b/vendor/github.com/cilium/ebpf/btf/types.go
index 68d4a17..8bd7fc7 100644
--- a/vendor/github.com/cilium/ebpf/btf/types.go
+++ b/vendor/github.com/cilium/ebpf/btf/types.go
@@ -1,19 +1,22 @@
package btf
import (
+ "encoding/binary"
"errors"
"fmt"
"io"
"math"
- "reflect"
"strings"
"github.com/cilium/ebpf/asm"
"github.com/cilium/ebpf/internal"
"github.com/cilium/ebpf/internal/sys"
+ "golang.org/x/exp/slices"
)
-const maxTypeDepth = 32
+// Mirrors MAX_RESOLVE_DEPTH in libbpf.
+// https://github.com/libbpf/libbpf/blob/e26b84dc330c9644c07428c271ab491b0f01f4e1/src/btf.c#L761
+const maxResolveDepth = 32
// TypeID identifies a type in a BTF section.
type TypeID = sys.TypeID
@@ -116,7 +119,7 @@
}
func (i *Int) Format(fs fmt.State, verb rune) {
- formatType(fs, verb, i, i.Encoding, "size=", i.Size*8)
+ formatType(fs, verb, i, i.Encoding, "size=", i.Size)
}
func (i *Int) TypeName() string { return i.Name }
@@ -278,21 +281,6 @@
return &cpy
}
-// has64BitValues returns true if the Enum contains a value larger than 32 bits.
-// Kernels before 6.0 have enum values that overrun u32 replaced with zeroes.
-//
-// 64-bit enums have their Enum.Size attributes correctly set to 8, but if we
-// use the size attribute as a heuristic during BTF marshaling, we'll emit
-// ENUM64s to kernels that don't support them.
-func (e *Enum) has64BitValues() bool {
- for _, v := range e.Values {
- if v.Value > math.MaxUint32 {
- return true
- }
- }
- return false
-}
-
// FwdKind is the type of forward declaration.
type FwdKind int
@@ -605,7 +593,7 @@
elem int64
)
- for i := 0; i < maxTypeDepth; i++ {
+ for i := 0; i < maxResolveDepth; i++ {
switch v := typ.(type) {
case *Array:
if n > 0 && int64(v.Nelems) > math.MaxInt64/n {
@@ -740,17 +728,22 @@
type typeDeque = internal.Deque[*Type]
-// inflateRawTypes takes a list of raw btf types linked via type IDs, and turns
-// it into a graph of Types connected via pointers.
+// readAndInflateTypes reads the raw btf type info and turns it into a graph
+// of Types connected via pointers.
//
-// If base is provided, then the raw types are considered to be of a split BTF
+// If base is provided, then the types are considered to be of a split BTF
// (e.g., a kernel module).
//
// Returns a slice of types indexed by TypeID. Since BTF ignores compilation
// units, multiple types may share the same name. A Type may form a cyclic graph
// by pointing at itself.
-func inflateRawTypes(rawTypes []rawType, rawStrings *stringTable, base *Spec) ([]Type, error) {
- types := make([]Type, 0, len(rawTypes)+1) // +1 for Void added to base types
+func readAndInflateTypes(r io.Reader, bo binary.ByteOrder, typeLen uint32, rawStrings *stringTable, base *Spec) ([]Type, error) {
+ // because of the interleaving between types and struct members it is difficult to
+ // precompute the numbers of raw types this will parse
+ // this "guess" is a good first estimation
+ sizeOfbtfType := uintptr(btfTypeLen)
+ tyMaxCount := uintptr(typeLen) / sizeOfbtfType / 2
+ types := make([]Type, 0, tyMaxCount)
// Void is defined to always be type ID 0, and is thus omitted from BTF.
types = append(types, (*Void)(nil))
@@ -773,11 +766,11 @@
}
var fixups []fixupDef
- fixup := func(id TypeID, typ *Type) bool {
+ fixup := func(id TypeID, typ *Type) {
if id < firstTypeID {
if baseType, err := base.TypeByID(id); err == nil {
*typ = baseType
- return true
+ return
}
}
@@ -785,31 +778,10 @@
if idx < len(types) {
// We've already inflated this type, fix it up immediately.
*typ = types[idx]
- return true
+ return
}
fixups = append(fixups, fixupDef{id, typ})
- return false
- }
-
- type assertion struct {
- id TypeID
- typ *Type
- want reflect.Type
- }
-
- var assertions []assertion
- fixupAndAssert := func(id TypeID, typ *Type, want reflect.Type) error {
- if !fixup(id, typ) {
- assertions = append(assertions, assertion{id, typ, want})
- return nil
- }
-
- // The type has already been fixed up, check the type immediately.
- if reflect.TypeOf(*typ) != want {
- return fmt.Errorf("type ID %d: expected %s, got %T", id, want, *typ)
- }
- return nil
}
type bitfieldFixupDef struct {
@@ -876,62 +848,128 @@
return members, nil
}
+ var (
+ buf = make([]byte, 1024)
+ header btfType
+ bInt btfInt
+ bArr btfArray
+ bMembers []btfMember
+ bEnums []btfEnum
+ bParams []btfParam
+ bVariable btfVariable
+ bSecInfos []btfVarSecinfo
+ bDeclTag btfDeclTag
+ bEnums64 []btfEnum64
+ )
+
var declTags []*declTag
- for _, raw := range rawTypes {
+ for {
var (
id = firstTypeID + TypeID(len(types))
typ Type
)
+ if _, err := io.ReadFull(r, buf[:btfTypeLen]); err == io.EOF {
+ break
+ } else if err != nil {
+ return nil, fmt.Errorf("can't read type info for id %v: %v", id, err)
+ }
+
+ if _, err := unmarshalBtfType(&header, buf[:btfTypeLen], bo); err != nil {
+ return nil, fmt.Errorf("can't unmarshal type info for id %v: %v", id, err)
+ }
+
if id < firstTypeID {
return nil, fmt.Errorf("no more type IDs")
}
- name, err := rawStrings.Lookup(raw.NameOff)
+ name, err := rawStrings.Lookup(header.NameOff)
if err != nil {
return nil, fmt.Errorf("get name for type id %d: %w", id, err)
}
- switch raw.Kind() {
+ switch header.Kind() {
case kindInt:
- size := raw.Size()
- bi := raw.data.(*btfInt)
- if bi.Offset() > 0 || bi.Bits().Bytes() != size {
- legacyBitfields[id] = [2]Bits{bi.Offset(), bi.Bits()}
+ size := header.Size()
+ buf = buf[:btfIntLen]
+ if _, err := io.ReadFull(r, buf); err != nil {
+ return nil, fmt.Errorf("can't read btfInt, id: %d: %w", id, err)
}
- typ = &Int{name, raw.Size(), bi.Encoding()}
+ if _, err := unmarshalBtfInt(&bInt, buf, bo); err != nil {
+ return nil, fmt.Errorf("can't unmarshal btfInt, id: %d: %w", id, err)
+ }
+ if bInt.Offset() > 0 || bInt.Bits().Bytes() != size {
+ legacyBitfields[id] = [2]Bits{bInt.Offset(), bInt.Bits()}
+ }
+ typ = &Int{name, header.Size(), bInt.Encoding()}
case kindPointer:
ptr := &Pointer{nil}
- fixup(raw.Type(), &ptr.Target)
+ fixup(header.Type(), &ptr.Target)
typ = ptr
case kindArray:
- btfArr := raw.data.(*btfArray)
- arr := &Array{nil, nil, btfArr.Nelems}
- fixup(btfArr.IndexType, &arr.Index)
- fixup(btfArr.Type, &arr.Type)
+ buf = buf[:btfArrayLen]
+ if _, err := io.ReadFull(r, buf); err != nil {
+ return nil, fmt.Errorf("can't read btfArray, id: %d: %w", id, err)
+ }
+ if _, err := unmarshalBtfArray(&bArr, buf, bo); err != nil {
+ return nil, fmt.Errorf("can't unmarshal btfArray, id: %d: %w", id, err)
+ }
+
+ arr := &Array{nil, nil, bArr.Nelems}
+ fixup(bArr.IndexType, &arr.Index)
+ fixup(bArr.Type, &arr.Type)
typ = arr
case kindStruct:
- members, err := convertMembers(raw.data.([]btfMember), raw.Bitfield())
+ vlen := header.Vlen()
+ bMembers = slices.Grow(bMembers[:0], vlen)[:vlen]
+ buf = slices.Grow(buf[:0], vlen*btfMemberLen)[:vlen*btfMemberLen]
+ if _, err := io.ReadFull(r, buf); err != nil {
+ return nil, fmt.Errorf("can't read btfMembers, id: %d: %w", id, err)
+ }
+ if _, err := unmarshalBtfMembers(bMembers, buf, bo); err != nil {
+ return nil, fmt.Errorf("can't unmarshal btfMembers, id: %d: %w", id, err)
+ }
+
+ members, err := convertMembers(bMembers, header.Bitfield())
if err != nil {
return nil, fmt.Errorf("struct %s (id %d): %w", name, id, err)
}
- typ = &Struct{name, raw.Size(), members}
+ typ = &Struct{name, header.Size(), members}
case kindUnion:
- members, err := convertMembers(raw.data.([]btfMember), raw.Bitfield())
+ vlen := header.Vlen()
+ bMembers = slices.Grow(bMembers[:0], vlen)[:vlen]
+ buf = slices.Grow(buf[:0], vlen*btfMemberLen)[:vlen*btfMemberLen]
+ if _, err := io.ReadFull(r, buf); err != nil {
+ return nil, fmt.Errorf("can't read btfMembers, id: %d: %w", id, err)
+ }
+ if _, err := unmarshalBtfMembers(bMembers, buf, bo); err != nil {
+ return nil, fmt.Errorf("can't unmarshal btfMembers, id: %d: %w", id, err)
+ }
+
+ members, err := convertMembers(bMembers, header.Bitfield())
if err != nil {
return nil, fmt.Errorf("union %s (id %d): %w", name, id, err)
}
- typ = &Union{name, raw.Size(), members}
+ typ = &Union{name, header.Size(), members}
case kindEnum:
- rawvals := raw.data.([]btfEnum)
- vals := make([]EnumValue, 0, len(rawvals))
- signed := raw.Signed()
- for i, btfVal := range rawvals {
+ vlen := header.Vlen()
+ bEnums = slices.Grow(bEnums[:0], vlen)[:vlen]
+ buf = slices.Grow(buf[:0], vlen*btfEnumLen)[:vlen*btfEnumLen]
+ if _, err := io.ReadFull(r, buf); err != nil {
+ return nil, fmt.Errorf("can't read btfEnums, id: %d: %w", id, err)
+ }
+ if _, err := unmarshalBtfEnums(bEnums, buf, bo); err != nil {
+ return nil, fmt.Errorf("can't unmarshal btfEnums, id: %d: %w", id, err)
+ }
+
+ vals := make([]EnumValue, 0, vlen)
+ signed := header.Signed()
+ for i, btfVal := range bEnums {
name, err := rawStrings.Lookup(btfVal.NameOff)
if err != nil {
return nil, fmt.Errorf("get name for enum value %d: %s", i, err)
@@ -943,42 +981,49 @@
}
vals = append(vals, EnumValue{name, value})
}
- typ = &Enum{name, raw.Size(), signed, vals}
+ typ = &Enum{name, header.Size(), signed, vals}
case kindForward:
- typ = &Fwd{name, raw.FwdKind()}
+ typ = &Fwd{name, header.FwdKind()}
case kindTypedef:
typedef := &Typedef{name, nil}
- fixup(raw.Type(), &typedef.Type)
+ fixup(header.Type(), &typedef.Type)
typ = typedef
case kindVolatile:
volatile := &Volatile{nil}
- fixup(raw.Type(), &volatile.Type)
+ fixup(header.Type(), &volatile.Type)
typ = volatile
case kindConst:
cnst := &Const{nil}
- fixup(raw.Type(), &cnst.Type)
+ fixup(header.Type(), &cnst.Type)
typ = cnst
case kindRestrict:
restrict := &Restrict{nil}
- fixup(raw.Type(), &restrict.Type)
+ fixup(header.Type(), &restrict.Type)
typ = restrict
case kindFunc:
- fn := &Func{name, nil, raw.Linkage()}
- if err := fixupAndAssert(raw.Type(), &fn.Type, reflect.TypeOf((*FuncProto)(nil))); err != nil {
- return nil, err
- }
+ fn := &Func{name, nil, header.Linkage()}
+ fixup(header.Type(), &fn.Type)
typ = fn
case kindFuncProto:
- rawparams := raw.data.([]btfParam)
- params := make([]FuncParam, 0, len(rawparams))
- for i, param := range rawparams {
+ vlen := header.Vlen()
+ bParams = slices.Grow(bParams[:0], vlen)[:vlen]
+ buf = slices.Grow(buf[:0], vlen*btfParamLen)[:vlen*btfParamLen]
+ if _, err := io.ReadFull(r, buf); err != nil {
+ return nil, fmt.Errorf("can't read btfParams, id: %d: %w", id, err)
+ }
+ if _, err := unmarshalBtfParams(bParams, buf, bo); err != nil {
+ return nil, fmt.Errorf("can't unmarshal btfParams, id: %d: %w", id, err)
+ }
+
+ params := make([]FuncParam, 0, vlen)
+ for i, param := range bParams {
name, err := rawStrings.Lookup(param.NameOff)
if err != nil {
return nil, fmt.Errorf("get name for func proto parameter %d: %s", i, err)
@@ -988,57 +1033,90 @@
})
}
for i := range params {
- fixup(rawparams[i].Type, ¶ms[i].Type)
+ fixup(bParams[i].Type, ¶ms[i].Type)
}
fp := &FuncProto{nil, params}
- fixup(raw.Type(), &fp.Return)
+ fixup(header.Type(), &fp.Return)
typ = fp
case kindVar:
- variable := raw.data.(*btfVariable)
- v := &Var{name, nil, VarLinkage(variable.Linkage)}
- fixup(raw.Type(), &v.Type)
+ buf = buf[:btfVariableLen]
+ if _, err := io.ReadFull(r, buf); err != nil {
+ return nil, fmt.Errorf("can't read btfVariable, id: %d: %w", id, err)
+ }
+ if _, err := unmarshalBtfVariable(&bVariable, buf, bo); err != nil {
+ return nil, fmt.Errorf("can't read btfVariable, id: %d: %w", id, err)
+ }
+
+ v := &Var{name, nil, VarLinkage(bVariable.Linkage)}
+ fixup(header.Type(), &v.Type)
typ = v
case kindDatasec:
- btfVars := raw.data.([]btfVarSecinfo)
- vars := make([]VarSecinfo, 0, len(btfVars))
- for _, btfVar := range btfVars {
+ vlen := header.Vlen()
+ bSecInfos = slices.Grow(bSecInfos[:0], vlen)[:vlen]
+ buf = slices.Grow(buf[:0], vlen*btfVarSecinfoLen)[:vlen*btfVarSecinfoLen]
+ if _, err := io.ReadFull(r, buf); err != nil {
+ return nil, fmt.Errorf("can't read btfVarSecInfos, id: %d: %w", id, err)
+ }
+ if _, err := unmarshalBtfVarSecInfos(bSecInfos, buf, bo); err != nil {
+ return nil, fmt.Errorf("can't unmarshal btfVarSecInfos, id: %d: %w", id, err)
+ }
+
+ vars := make([]VarSecinfo, 0, vlen)
+ for _, btfVar := range bSecInfos {
vars = append(vars, VarSecinfo{
Offset: btfVar.Offset,
Size: btfVar.Size,
})
}
for i := range vars {
- fixup(btfVars[i].Type, &vars[i].Type)
+ fixup(bSecInfos[i].Type, &vars[i].Type)
}
- typ = &Datasec{name, raw.Size(), vars}
+ typ = &Datasec{name, header.Size(), vars}
case kindFloat:
- typ = &Float{name, raw.Size()}
+ typ = &Float{name, header.Size()}
case kindDeclTag:
- btfIndex := raw.data.(*btfDeclTag).ComponentIdx
+ buf = buf[:btfDeclTagLen]
+ if _, err := io.ReadFull(r, buf); err != nil {
+ return nil, fmt.Errorf("can't read btfDeclTag, id: %d: %w", id, err)
+ }
+ if _, err := unmarshalBtfDeclTag(&bDeclTag, buf, bo); err != nil {
+ return nil, fmt.Errorf("can't read btfDeclTag, id: %d: %w", id, err)
+ }
+
+ btfIndex := bDeclTag.ComponentIdx
if uint64(btfIndex) > math.MaxInt {
return nil, fmt.Errorf("type id %d: index exceeds int", id)
}
dt := &declTag{nil, name, int(int32(btfIndex))}
- fixup(raw.Type(), &dt.Type)
+ fixup(header.Type(), &dt.Type)
typ = dt
declTags = append(declTags, dt)
case kindTypeTag:
tt := &typeTag{nil, name}
- fixup(raw.Type(), &tt.Type)
+ fixup(header.Type(), &tt.Type)
typ = tt
case kindEnum64:
- rawvals := raw.data.([]btfEnum64)
- vals := make([]EnumValue, 0, len(rawvals))
- for i, btfVal := range rawvals {
+ vlen := header.Vlen()
+ bEnums64 = slices.Grow(bEnums64[:0], vlen)[:vlen]
+ buf = slices.Grow(buf[:0], vlen*btfEnum64Len)[:vlen*btfEnum64Len]
+ if _, err := io.ReadFull(r, buf); err != nil {
+ return nil, fmt.Errorf("can't read btfEnum64s, id: %d: %w", id, err)
+ }
+ if _, err := unmarshalBtfEnums64(bEnums64, buf, bo); err != nil {
+ return nil, fmt.Errorf("can't unmarshal btfEnum64s, id: %d: %w", id, err)
+ }
+
+ vals := make([]EnumValue, 0, vlen)
+ for i, btfVal := range bEnums64 {
name, err := rawStrings.Lookup(btfVal.NameOff)
if err != nil {
return nil, fmt.Errorf("get name for enum64 value %d: %s", i, err)
@@ -1046,10 +1124,10 @@
value := (uint64(btfVal.ValHi32) << 32) | uint64(btfVal.ValLo32)
vals = append(vals, EnumValue{name, value})
}
- typ = &Enum{name, raw.Size(), raw.Signed(), vals}
+ typ = &Enum{name, header.Size(), header.Signed(), vals}
default:
- return nil, fmt.Errorf("type id %d: unknown kind: %v", id, raw.Kind())
+ return nil, fmt.Errorf("type id %d: unknown kind: %v", id, header.Kind())
}
types = append(types, typ)
@@ -1081,12 +1159,6 @@
}
}
- for _, assertion := range assertions {
- if reflect.TypeOf(*assertion.typ) != assertion.want {
- return nil, fmt.Errorf("type ID %d: expected %s, got %T", assertion.id, assertion.want, *assertion.typ)
- }
- }
-
for _, dt := range declTags {
switch t := dt.Type.(type) {
case *Var, *Typedef:
@@ -1100,7 +1172,12 @@
}
case *Func:
- if dt.Index >= len(t.Type.(*FuncProto).Params) {
+ fp, ok := t.Type.(*FuncProto)
+ if !ok {
+ return nil, fmt.Errorf("type %s: %s is not a FuncProto", dt, t.Type)
+ }
+
+ if dt.Index >= len(fp.Params) {
return nil, fmt.Errorf("type %s: index %d exceeds params of %s", dt, dt.Index, t)
}
@@ -1136,7 +1213,7 @@
// UnderlyingType skips qualifiers and Typedefs.
func UnderlyingType(typ Type) Type {
result := typ
- for depth := 0; depth <= maxTypeDepth; depth++ {
+ for depth := 0; depth <= maxResolveDepth; depth++ {
switch v := (result).(type) {
case qualifier:
result = v.qualify()
@@ -1155,7 +1232,7 @@
// Returns the zero value and false if there is no T or if the type is nested
// too deeply.
func as[T Type](typ Type) (T, bool) {
- for depth := 0; depth <= maxTypeDepth; depth++ {
+ for depth := 0; depth <= maxResolveDepth; depth++ {
switch v := (typ).(type) {
case T:
return v, true
diff --git a/vendor/github.com/cilium/ebpf/collection.go b/vendor/github.com/cilium/ebpf/collection.go
index fb720be..a553222 100644
--- a/vendor/github.com/cilium/ebpf/collection.go
+++ b/vendor/github.com/cilium/ebpf/collection.go
@@ -11,6 +11,7 @@
"github.com/cilium/ebpf/btf"
"github.com/cilium/ebpf/internal"
"github.com/cilium/ebpf/internal/kconfig"
+ "github.com/cilium/ebpf/internal/sysenc"
)
// CollectionOptions control loading a collection into the kernel.
@@ -175,12 +176,12 @@
return fmt.Errorf("section %s: offset %d(+%d) for variable %s is out of bounds", name, v.Offset, v.Size, vname)
}
- b, err := marshalBytes(replacement, int(v.Size))
+ b, err := sysenc.Marshal(replacement, int(v.Size))
if err != nil {
return fmt.Errorf("marshaling constant replacement %s: %w", vname, err)
}
- copy(cpy[v.Offset:v.Offset+v.Size], b)
+ b.CopyTo(cpy[v.Offset : v.Offset+v.Size])
replaced[vname] = true
}
@@ -308,7 +309,7 @@
}
// Populate the requested maps. Has a chance of lazy-loading other dependent maps.
- if err := loader.populateMaps(); err != nil {
+ if err := loader.populateDeferredMaps(); err != nil {
return err
}
@@ -388,7 +389,7 @@
// Maps can contain Program and Map stubs, so populate them after
// all Maps and Programs have been successfully loaded.
- if err := loader.populateMaps(); err != nil {
+ if err := loader.populateDeferredMaps(); err != nil {
return nil, err
}
@@ -470,6 +471,15 @@
return nil, fmt.Errorf("map %s: %w", mapName, err)
}
+ // Finalize 'scalar' maps that don't refer to any other eBPF resources
+ // potentially pending creation. This is needed for frozen maps like .rodata
+ // that need to be finalized before invoking the verifier.
+ if !mapSpec.Type.canStoreMapOrProgram() {
+ if err := m.finalize(mapSpec); err != nil {
+ return nil, fmt.Errorf("finalizing map %s: %w", mapName, err)
+ }
+ }
+
cl.maps[mapName] = m
return m, nil
}
@@ -527,44 +537,50 @@
return prog, nil
}
-func (cl *collectionLoader) populateMaps() error {
+// populateDeferredMaps iterates maps holding programs or other maps and loads
+// any dependencies. Populates all maps in cl and freezes them if specified.
+func (cl *collectionLoader) populateDeferredMaps() error {
for mapName, m := range cl.maps {
mapSpec, ok := cl.coll.Maps[mapName]
if !ok {
return fmt.Errorf("missing map spec %s", mapName)
}
+ // Scalar maps without Map or Program references are finalized during
+ // creation. Don't finalize them again.
+ if !mapSpec.Type.canStoreMapOrProgram() {
+ continue
+ }
+
+ mapSpec = mapSpec.Copy()
+
// MapSpecs that refer to inner maps or programs within the same
// CollectionSpec do so using strings. These strings are used as the key
// to look up the respective object in the Maps or Programs fields.
// Resolve those references to actual Map or Program resources that
// have been loaded into the kernel.
- if mapSpec.Type.canStoreMap() || mapSpec.Type.canStoreProgram() {
- mapSpec = mapSpec.Copy()
+ for i, kv := range mapSpec.Contents {
+ objName, ok := kv.Value.(string)
+ if !ok {
+ continue
+ }
- for i, kv := range mapSpec.Contents {
- objName, ok := kv.Value.(string)
- if !ok {
- continue
+ switch t := mapSpec.Type; {
+ case t.canStoreProgram():
+ // loadProgram is idempotent and could return an existing Program.
+ prog, err := cl.loadProgram(objName)
+ if err != nil {
+ return fmt.Errorf("loading program %s, for map %s: %w", objName, mapName, err)
}
+ mapSpec.Contents[i] = MapKV{kv.Key, prog}
- switch t := mapSpec.Type; {
- case t.canStoreProgram():
- // loadProgram is idempotent and could return an existing Program.
- prog, err := cl.loadProgram(objName)
- if err != nil {
- return fmt.Errorf("loading program %s, for map %s: %w", objName, mapName, err)
- }
- mapSpec.Contents[i] = MapKV{kv.Key, prog}
-
- case t.canStoreMap():
- // loadMap is idempotent and could return an existing Map.
- innerMap, err := cl.loadMap(objName)
- if err != nil {
- return fmt.Errorf("loading inner map %s, for map %s: %w", objName, mapName, err)
- }
- mapSpec.Contents[i] = MapKV{kv.Key, innerMap}
+ case t.canStoreMap():
+ // loadMap is idempotent and could return an existing Map.
+ innerMap, err := cl.loadMap(objName)
+ if err != nil {
+ return fmt.Errorf("loading inner map %s, for map %s: %w", objName, mapName, err)
}
+ mapSpec.Contents[i] = MapKV{kv.Key, innerMap}
}
}
@@ -610,17 +626,20 @@
internal.NativeEndian.PutUint32(data[vsi.Offset:], kv.Kernel())
case "LINUX_HAS_SYSCALL_WRAPPER":
- if integer, ok := v.Type.(*btf.Int); !ok || integer.Size != 4 {
- return fmt.Errorf("variable %s must be a 32 bits integer, got %s", n, v.Type)
+ integer, ok := v.Type.(*btf.Int)
+ if !ok {
+ return fmt.Errorf("variable %s must be an integer, got %s", n, v.Type)
}
- var value uint32 = 1
+ var value uint64 = 1
if err := haveSyscallWrapper(); errors.Is(err, ErrNotSupported) {
value = 0
} else if err != nil {
return fmt.Errorf("unable to derive a value for LINUX_HAS_SYSCALL_WRAPPER: %w", err)
}
- internal.NativeEndian.PutUint32(data[vsi.Offset:], value)
+ if err := kconfig.PutInteger(data[vsi.Offset:], integer, value); err != nil {
+ return fmt.Errorf("set LINUX_HAS_SYSCALL_WRAPPER: %w", err)
+ }
default: // Catch CONFIG_*.
configs[n] = configInfo{
@@ -679,6 +698,71 @@
return NewCollection(spec)
}
+// Assign the contents of a Collection to a struct.
+//
+// This function bridges functionality between bpf2go generated
+// code and any functionality better implemented in Collection.
+//
+// 'to' must be a pointer to a struct. A field of the
+// struct is updated with values from Programs or Maps if it
+// has an `ebpf` tag and its type is *Program or *Map.
+// The tag's value specifies the name of the program or map as
+// found in the CollectionSpec.
+//
+// struct {
+// Foo *ebpf.Program `ebpf:"xdp_foo"`
+// Bar *ebpf.Map `ebpf:"bar_map"`
+// Ignored int
+// }
+//
+// Returns an error if any of the eBPF objects can't be found, or
+// if the same Map or Program is assigned multiple times.
+//
+// Ownership and Close()ing responsibility is transferred to `to`
+// for any successful assigns. On error `to` is left in an undefined state.
+func (coll *Collection) Assign(to interface{}) error {
+ assignedMaps := make(map[string]bool)
+ assignedProgs := make(map[string]bool)
+
+ // Assign() only transfers already-loaded Maps and Programs. No extra
+ // loading is done.
+ getValue := func(typ reflect.Type, name string) (interface{}, error) {
+ switch typ {
+
+ case reflect.TypeOf((*Program)(nil)):
+ if p := coll.Programs[name]; p != nil {
+ assignedProgs[name] = true
+ return p, nil
+ }
+ return nil, fmt.Errorf("missing program %q", name)
+
+ case reflect.TypeOf((*Map)(nil)):
+ if m := coll.Maps[name]; m != nil {
+ assignedMaps[name] = true
+ return m, nil
+ }
+ return nil, fmt.Errorf("missing map %q", name)
+
+ default:
+ return nil, fmt.Errorf("unsupported type %s", typ)
+ }
+ }
+
+ if err := assignValues(to, getValue); err != nil {
+ return err
+ }
+
+ // Finalize ownership transfer
+ for p := range assignedProgs {
+ delete(coll.Programs, p)
+ }
+ for m := range assignedMaps {
+ delete(coll.Maps, m)
+ }
+
+ return nil
+}
+
// Close frees all maps and programs associated with the collection.
//
// The collection mustn't be used afterwards.
diff --git a/vendor/github.com/cilium/ebpf/elf_reader.go b/vendor/github.com/cilium/ebpf/elf_reader.go
index 8d92672..5a85cbc 100644
--- a/vendor/github.com/cilium/ebpf/elf_reader.go
+++ b/vendor/github.com/cilium/ebpf/elf_reader.go
@@ -81,6 +81,8 @@
// Collect all the sections we're interested in. This includes relocations
// which we parse later.
+ //
+ // Keep the documentation at docs/ebpf/loading/elf-sections.md up-to-date.
for i, sec := range f.Sections {
idx := elf.SectionIndex(i)
@@ -371,7 +373,7 @@
r := bufio.NewReader(section.Open())
// Decode the section's instruction stream.
- var insns asm.Instructions
+ insns := make(asm.Instructions, 0, section.Size/asm.InstructionSize)
if err := insns.Unmarshal(r, ec.ByteOrder); err != nil {
return nil, fmt.Errorf("decoding instructions for section %s: %w", section.Name, err)
}
@@ -465,10 +467,14 @@
switch target.kind {
case mapSection, btfMapSection:
- if bind != elf.STB_GLOBAL {
+ if bind == elf.STB_LOCAL {
return fmt.Errorf("possible erroneous static qualifier on map definition: found reference to %q", name)
}
+ if bind != elf.STB_GLOBAL {
+ return fmt.Errorf("map %q: unsupported relocation %s", name, bind)
+ }
+
if typ != elf.STT_OBJECT && typ != elf.STT_NOTYPE {
// STT_NOTYPE is generated on clang < 8 which doesn't tag
// relocations appropriately.
@@ -694,10 +700,6 @@
spec.Extra = bytes.NewReader(extra)
}
- if err := spec.clampPerfEventArraySize(); err != nil {
- return fmt.Errorf("map %s: %w", mapName, err)
- }
-
ec.maps[mapName] = &spec
}
}
@@ -752,7 +754,7 @@
}
// Each Var representing a BTF map definition contains a Struct.
- mapStruct, ok := v.Type.(*btf.Struct)
+ mapStruct, ok := btf.UnderlyingType(v.Type).(*btf.Struct)
if !ok {
return fmt.Errorf("expected struct, got %s", v.Type)
}
@@ -762,10 +764,6 @@
return fmt.Errorf("map %v: %w", name, err)
}
- if err := mapSpec.clampPerfEventArraySize(); err != nil {
- return fmt.Errorf("map %v: %w", name, err)
- }
-
ec.maps[name] = mapSpec
}
@@ -785,7 +783,7 @@
// mapSpecFromBTF produces a MapSpec based on a btf.Struct def representing
// a BTF map definition. The name and spec arguments will be copied to the
-// resulting MapSpec, and inner must be true on any resursive invocations.
+// resulting MapSpec, and inner must be true on any recursive invocations.
func mapSpecFromBTF(es *elfSection, vs *btf.VarSecinfo, def *btf.Struct, spec *btf.Spec, name string, inner bool) (*MapSpec, error) {
var (
key, value btf.Type
@@ -1150,7 +1148,7 @@
KeySize: uint32(4),
ValueSize: ds.Size,
MaxEntries: 1,
- Flags: unix.BPF_F_RDONLY_PROG | unix.BPF_F_MMAPABLE,
+ Flags: unix.BPF_F_RDONLY_PROG,
Freeze: true,
Key: &btf.Int{Size: 4},
Value: ds,
@@ -1268,6 +1266,7 @@
{"seccomp", SocketFilter, AttachNone, 0},
{"kprobe.multi", Kprobe, AttachTraceKprobeMulti, 0},
{"kretprobe.multi", Kprobe, AttachTraceKprobeMulti, 0},
+ // Document all prefixes in docs/ebpf/concepts/elf-sections.md.
}
for _, t := range types {
diff --git a/vendor/github.com/cilium/ebpf/info.go b/vendor/github.com/cilium/ebpf/info.go
index a02e8a4..79b11c9 100644
--- a/vendor/github.com/cilium/ebpf/info.go
+++ b/vendor/github.com/cilium/ebpf/info.go
@@ -101,6 +101,11 @@
maps []MapID
insns []byte
+
+ lineInfos []byte
+ numLineInfos uint32
+ funcInfos []byte
+ numFuncInfos uint32
}
func newProgramInfoFromFd(fd *sys.FD) (*ProgramInfo, error) {
@@ -128,10 +133,13 @@
// Start with a clean struct for the second call, otherwise we may get EFAULT.
var info2 sys.ProgInfo
+ makeSecondCall := false
+
if info.NrMapIds > 0 {
pi.maps = make([]MapID, info.NrMapIds)
info2.NrMapIds = info.NrMapIds
info2.MapIds = sys.NewPointer(unsafe.Pointer(&pi.maps[0]))
+ makeSecondCall = true
} else if haveProgramInfoMapIDs() == nil {
// This program really has no associated maps.
pi.maps = make([]MapID, 0)
@@ -150,9 +158,28 @@
pi.insns = make([]byte, info.XlatedProgLen)
info2.XlatedProgLen = info.XlatedProgLen
info2.XlatedProgInsns = sys.NewSlicePointer(pi.insns)
+ makeSecondCall = true
}
- if info.NrMapIds > 0 || info.XlatedProgLen > 0 {
+ if info.NrLineInfo > 0 {
+ pi.lineInfos = make([]byte, btf.LineInfoSize*info.NrLineInfo)
+ info2.LineInfo = sys.NewSlicePointer(pi.lineInfos)
+ info2.LineInfoRecSize = btf.LineInfoSize
+ info2.NrLineInfo = info.NrLineInfo
+ pi.numLineInfos = info.NrLineInfo
+ makeSecondCall = true
+ }
+
+ if info.NrFuncInfo > 0 {
+ pi.funcInfos = make([]byte, btf.FuncInfoSize*info.NrFuncInfo)
+ info2.FuncInfo = sys.NewSlicePointer(pi.funcInfos)
+ info2.FuncInfoRecSize = btf.FuncInfoSize
+ info2.NrFuncInfo = info.NrFuncInfo
+ pi.numFuncInfos = info.NrFuncInfo
+ makeSecondCall = true
+ }
+
+ if makeSecondCall {
if err := sys.ObjInfo(fd, &info2); err != nil {
return nil, err
}
@@ -245,7 +272,13 @@
//
// The first instruction is marked as a symbol using the Program's name.
//
-// Available from 4.13. Requires CAP_BPF or equivalent.
+// If available, the instructions will be annotated with metadata from the
+// BTF. This includes line information and function information. Reading
+// this metadata requires CAP_SYS_ADMIN or equivalent. If capability is
+// unavailable, the instructions will be returned without metadata.
+//
+// Available from 4.13. Requires CAP_BPF or equivalent for plain instructions.
+// Requires CAP_SYS_ADMIN for instructions with metadata.
func (pi *ProgramInfo) Instructions() (asm.Instructions, error) {
// If the calling process is not BPF-capable or if the kernel doesn't
// support getting xlated instructions, the field will be zero.
@@ -259,8 +292,55 @@
return nil, fmt.Errorf("unmarshaling instructions: %w", err)
}
- // Tag the first instruction with the name of the program, if available.
- insns[0] = insns[0].WithSymbol(pi.Name)
+ if pi.btf != 0 {
+ btfh, err := btf.NewHandleFromID(pi.btf)
+ if err != nil {
+ // Getting a BTF handle requires CAP_SYS_ADMIN, if not available we get an -EPERM.
+ // Ignore it and fall back to instructions without metadata.
+ if !errors.Is(err, unix.EPERM) {
+ return nil, fmt.Errorf("unable to get BTF handle: %w", err)
+ }
+ }
+
+ // If we have a BTF handle, we can use it to assign metadata to the instructions.
+ if btfh != nil {
+ defer btfh.Close()
+
+ spec, err := btfh.Spec(nil)
+ if err != nil {
+ return nil, fmt.Errorf("unable to get BTF spec: %w", err)
+ }
+
+ lineInfos, err := btf.LoadLineInfos(
+ bytes.NewReader(pi.lineInfos),
+ internal.NativeEndian,
+ pi.numLineInfos,
+ spec,
+ )
+ if err != nil {
+ return nil, fmt.Errorf("parse line info: %w", err)
+ }
+
+ funcInfos, err := btf.LoadFuncInfos(
+ bytes.NewReader(pi.funcInfos),
+ internal.NativeEndian,
+ pi.numFuncInfos,
+ spec,
+ )
+ if err != nil {
+ return nil, fmt.Errorf("parse func info: %w", err)
+ }
+
+ btf.AssignMetadataToInstructions(insns, funcInfos, lineInfos, btf.CORERelocationInfos{})
+ }
+ }
+
+ fn := btf.FuncMetadata(&insns[0])
+ name := pi.Name
+ if fn != nil {
+ name = fn.Name
+ }
+ insns[0] = insns[0].WithSymbol(name)
return insns, nil
}
diff --git a/vendor/github.com/cilium/ebpf/internal/endian_be.go b/vendor/github.com/cilium/ebpf/internal/endian_be.go
index 96a2ac0..39f49ba 100644
--- a/vendor/github.com/cilium/ebpf/internal/endian_be.go
+++ b/vendor/github.com/cilium/ebpf/internal/endian_be.go
@@ -6,7 +6,7 @@
// NativeEndian is set to either binary.BigEndian or binary.LittleEndian,
// depending on the host's endianness.
-var NativeEndian binary.ByteOrder = binary.BigEndian
+var NativeEndian = binary.BigEndian
// ClangEndian is set to either "el" or "eb" depending on the host's endianness.
const ClangEndian = "eb"
diff --git a/vendor/github.com/cilium/ebpf/internal/endian_le.go b/vendor/github.com/cilium/ebpf/internal/endian_le.go
index fde4c55..9488e30 100644
--- a/vendor/github.com/cilium/ebpf/internal/endian_le.go
+++ b/vendor/github.com/cilium/ebpf/internal/endian_le.go
@@ -6,7 +6,7 @@
// NativeEndian is set to either binary.BigEndian or binary.LittleEndian,
// depending on the host's endianness.
-var NativeEndian binary.ByteOrder = binary.LittleEndian
+var NativeEndian = binary.LittleEndian
// ClangEndian is set to either "el" or "eb" depending on the host's endianness.
const ClangEndian = "el"
diff --git a/vendor/github.com/cilium/ebpf/internal/kconfig/kconfig.go b/vendor/github.com/cilium/ebpf/internal/kconfig/kconfig.go
index d95e7eb..fa53085 100644
--- a/vendor/github.com/cilium/ebpf/internal/kconfig/kconfig.go
+++ b/vendor/github.com/cilium/ebpf/internal/kconfig/kconfig.go
@@ -250,7 +250,20 @@
return fmt.Errorf("cannot parse value: %w", err)
}
- switch size {
+ return PutInteger(data, integer, n)
+}
+
+// PutInteger writes n into data.
+//
+// integer determines how much is written into data and what the valid values
+// are.
+func PutInteger(data []byte, integer *btf.Int, n uint64) error {
+ // This function should match set_kcfg_value_num in libbpf.
+ if integer.Encoding == btf.Bool && n > 1 {
+ return fmt.Errorf("invalid boolean value: %d", n)
+ }
+
+ switch integer.Size {
case 1:
data[0] = byte(n)
case 2:
@@ -260,7 +273,7 @@
case 8:
internal.NativeEndian.PutUint64(data, uint64(n))
default:
- return fmt.Errorf("size (%d) is not valid, expected: 1, 2, 4 or 8", size)
+ return fmt.Errorf("size (%d) is not valid, expected: 1, 2, 4 or 8", integer.Size)
}
return nil
diff --git a/vendor/github.com/cilium/ebpf/internal/sys/syscall.go b/vendor/github.com/cilium/ebpf/internal/sys/syscall.go
index 4fae04d..088e82e 100644
--- a/vendor/github.com/cilium/ebpf/internal/sys/syscall.go
+++ b/vendor/github.com/cilium/ebpf/internal/sys/syscall.go
@@ -11,7 +11,7 @@
// ENOTSUPP is a Linux internal error code that has leaked into UAPI.
//
// It is not the same as ENOTSUP or EOPNOTSUPP.
-var ENOTSUPP = syscall.Errno(524)
+const ENOTSUPP = syscall.Errno(524)
// BPF wraps SYS_BPF.
//
@@ -123,7 +123,7 @@
// MapFlags control map behaviour.
type MapFlags uint32
-//go:generate stringer -type MapFlags
+//go:generate go run golang.org/x/tools/cmd/stringer@latest -type MapFlags
const (
BPF_F_NO_PREALLOC MapFlags = 1 << iota
diff --git a/vendor/github.com/cilium/ebpf/internal/sys/types.go b/vendor/github.com/cilium/ebpf/internal/sys/types.go
index 2af7759..51698e0 100644
--- a/vendor/github.com/cilium/ebpf/internal/sys/types.go
+++ b/vendor/github.com/cilium/ebpf/internal/sys/types.go
@@ -59,7 +59,13 @@
BPF_SK_REUSEPORT_SELECT_OR_MIGRATE AttachType = 40
BPF_PERF_EVENT AttachType = 41
BPF_TRACE_KPROBE_MULTI AttachType = 42
- __MAX_BPF_ATTACH_TYPE AttachType = 43
+ BPF_LSM_CGROUP AttachType = 43
+ BPF_STRUCT_OPS AttachType = 44
+ BPF_NETFILTER AttachType = 45
+ BPF_TCX_INGRESS AttachType = 46
+ BPF_TCX_EGRESS AttachType = 47
+ BPF_TRACE_UPROBE_MULTI AttachType = 48
+ __MAX_BPF_ATTACH_TYPE AttachType = 49
)
type Cmd uint32
@@ -311,7 +317,15 @@
BPF_FUNC_dynptr_read FunctionId = 201
BPF_FUNC_dynptr_write FunctionId = 202
BPF_FUNC_dynptr_data FunctionId = 203
- __BPF_FUNC_MAX_ID FunctionId = 204
+ BPF_FUNC_tcp_raw_gen_syncookie_ipv4 FunctionId = 204
+ BPF_FUNC_tcp_raw_gen_syncookie_ipv6 FunctionId = 205
+ BPF_FUNC_tcp_raw_check_syncookie_ipv4 FunctionId = 206
+ BPF_FUNC_tcp_raw_check_syncookie_ipv6 FunctionId = 207
+ BPF_FUNC_ktime_get_tai_ns FunctionId = 208
+ BPF_FUNC_user_ringbuf_drain FunctionId = 209
+ BPF_FUNC_cgrp_storage_get FunctionId = 210
+ BPF_FUNC_cgrp_storage_delete FunctionId = 211
+ __BPF_FUNC_MAX_ID FunctionId = 212
)
type HdrStartOff uint32
@@ -334,43 +348,49 @@
BPF_LINK_TYPE_PERF_EVENT LinkType = 7
BPF_LINK_TYPE_KPROBE_MULTI LinkType = 8
BPF_LINK_TYPE_STRUCT_OPS LinkType = 9
- MAX_BPF_LINK_TYPE LinkType = 10
+ BPF_LINK_TYPE_NETFILTER LinkType = 10
+ BPF_LINK_TYPE_TCX LinkType = 11
+ BPF_LINK_TYPE_UPROBE_MULTI LinkType = 12
+ MAX_BPF_LINK_TYPE LinkType = 13
)
type MapType uint32
const (
- BPF_MAP_TYPE_UNSPEC MapType = 0
- BPF_MAP_TYPE_HASH MapType = 1
- BPF_MAP_TYPE_ARRAY MapType = 2
- BPF_MAP_TYPE_PROG_ARRAY MapType = 3
- BPF_MAP_TYPE_PERF_EVENT_ARRAY MapType = 4
- BPF_MAP_TYPE_PERCPU_HASH MapType = 5
- BPF_MAP_TYPE_PERCPU_ARRAY MapType = 6
- BPF_MAP_TYPE_STACK_TRACE MapType = 7
- BPF_MAP_TYPE_CGROUP_ARRAY MapType = 8
- BPF_MAP_TYPE_LRU_HASH MapType = 9
- BPF_MAP_TYPE_LRU_PERCPU_HASH MapType = 10
- BPF_MAP_TYPE_LPM_TRIE MapType = 11
- BPF_MAP_TYPE_ARRAY_OF_MAPS MapType = 12
- BPF_MAP_TYPE_HASH_OF_MAPS MapType = 13
- BPF_MAP_TYPE_DEVMAP MapType = 14
- BPF_MAP_TYPE_SOCKMAP MapType = 15
- BPF_MAP_TYPE_CPUMAP MapType = 16
- BPF_MAP_TYPE_XSKMAP MapType = 17
- BPF_MAP_TYPE_SOCKHASH MapType = 18
- BPF_MAP_TYPE_CGROUP_STORAGE MapType = 19
- BPF_MAP_TYPE_REUSEPORT_SOCKARRAY MapType = 20
- BPF_MAP_TYPE_PERCPU_CGROUP_STORAGE MapType = 21
- BPF_MAP_TYPE_QUEUE MapType = 22
- BPF_MAP_TYPE_STACK MapType = 23
- BPF_MAP_TYPE_SK_STORAGE MapType = 24
- BPF_MAP_TYPE_DEVMAP_HASH MapType = 25
- BPF_MAP_TYPE_STRUCT_OPS MapType = 26
- BPF_MAP_TYPE_RINGBUF MapType = 27
- BPF_MAP_TYPE_INODE_STORAGE MapType = 28
- BPF_MAP_TYPE_TASK_STORAGE MapType = 29
- BPF_MAP_TYPE_BLOOM_FILTER MapType = 30
+ BPF_MAP_TYPE_UNSPEC MapType = 0
+ BPF_MAP_TYPE_HASH MapType = 1
+ BPF_MAP_TYPE_ARRAY MapType = 2
+ BPF_MAP_TYPE_PROG_ARRAY MapType = 3
+ BPF_MAP_TYPE_PERF_EVENT_ARRAY MapType = 4
+ BPF_MAP_TYPE_PERCPU_HASH MapType = 5
+ BPF_MAP_TYPE_PERCPU_ARRAY MapType = 6
+ BPF_MAP_TYPE_STACK_TRACE MapType = 7
+ BPF_MAP_TYPE_CGROUP_ARRAY MapType = 8
+ BPF_MAP_TYPE_LRU_HASH MapType = 9
+ BPF_MAP_TYPE_LRU_PERCPU_HASH MapType = 10
+ BPF_MAP_TYPE_LPM_TRIE MapType = 11
+ BPF_MAP_TYPE_ARRAY_OF_MAPS MapType = 12
+ BPF_MAP_TYPE_HASH_OF_MAPS MapType = 13
+ BPF_MAP_TYPE_DEVMAP MapType = 14
+ BPF_MAP_TYPE_SOCKMAP MapType = 15
+ BPF_MAP_TYPE_CPUMAP MapType = 16
+ BPF_MAP_TYPE_XSKMAP MapType = 17
+ BPF_MAP_TYPE_SOCKHASH MapType = 18
+ BPF_MAP_TYPE_CGROUP_STORAGE_DEPRECATED MapType = 19
+ BPF_MAP_TYPE_CGROUP_STORAGE MapType = 19
+ BPF_MAP_TYPE_REUSEPORT_SOCKARRAY MapType = 20
+ BPF_MAP_TYPE_PERCPU_CGROUP_STORAGE MapType = 21
+ BPF_MAP_TYPE_QUEUE MapType = 22
+ BPF_MAP_TYPE_STACK MapType = 23
+ BPF_MAP_TYPE_SK_STORAGE MapType = 24
+ BPF_MAP_TYPE_DEVMAP_HASH MapType = 25
+ BPF_MAP_TYPE_STRUCT_OPS MapType = 26
+ BPF_MAP_TYPE_RINGBUF MapType = 27
+ BPF_MAP_TYPE_INODE_STORAGE MapType = 28
+ BPF_MAP_TYPE_TASK_STORAGE MapType = 29
+ BPF_MAP_TYPE_BLOOM_FILTER MapType = 30
+ BPF_MAP_TYPE_USER_RINGBUF MapType = 31
+ BPF_MAP_TYPE_CGRP_STORAGE MapType = 32
)
type ProgType uint32
@@ -408,15 +428,17 @@
BPF_PROG_TYPE_LSM ProgType = 29
BPF_PROG_TYPE_SK_LOOKUP ProgType = 30
BPF_PROG_TYPE_SYSCALL ProgType = 31
+ BPF_PROG_TYPE_NETFILTER ProgType = 32
)
type RetCode uint32
const (
- BPF_OK RetCode = 0
- BPF_DROP RetCode = 2
- BPF_REDIRECT RetCode = 7
- BPF_LWT_REROUTE RetCode = 128
+ BPF_OK RetCode = 0
+ BPF_DROP RetCode = 2
+ BPF_REDIRECT RetCode = 7
+ BPF_LWT_REROUTE RetCode = 128
+ BPF_FLOW_DISSECTOR_CONTINUE RetCode = 129
)
type SkAction uint32
@@ -476,7 +498,7 @@
Id LinkID
ProgId uint32
_ [4]byte
- Extra [16]uint8
+ Extra [32]uint8
}
type MapInfo struct {
@@ -521,10 +543,10 @@
JitedFuncLens uint64
BtfId BTFID
FuncInfoRecSize uint32
- FuncInfo uint64
+ FuncInfo Pointer
NrFuncInfo uint32
NrLineInfo uint32
- LineInfo uint64
+ LineInfo Pointer
JitedLineInfo uint64
NrJitedLineInfo uint32
LineInfoRecSize uint32
@@ -535,6 +557,8 @@
RunCnt uint64
RecursionMisses uint64
VerifiedInsns uint32
+ AttachBtfObjId BTFID
+ AttachBtfId TypeID
_ [4]byte
}
@@ -583,12 +607,12 @@
}
type BtfLoadAttr struct {
- Btf Pointer
- BtfLogBuf Pointer
- BtfSize uint32
- BtfLogSize uint32
- BtfLogLevel uint32
- _ [4]byte
+ Btf Pointer
+ BtfLogBuf Pointer
+ BtfSize uint32
+ BtfLogSize uint32
+ BtfLogLevel uint32
+ BtfLogTrueSize uint32
}
func BtfLoad(attr *BtfLoadAttr) (*FD, error) {
@@ -628,7 +652,7 @@
AttachType AttachType
Flags uint32
TargetBtfId TypeID
- _ [28]byte
+ _ [44]byte
}
func LinkCreate(attr *LinkCreateAttr) (*FD, error) {
@@ -646,7 +670,7 @@
Flags uint32
IterInfo Pointer
IterInfoLen uint32
- _ [20]byte
+ _ [36]byte
}
func LinkCreateIter(attr *LinkCreateIterAttr) (*FD, error) {
@@ -667,6 +691,7 @@
Syms Pointer
Addrs Pointer
Cookies Pointer
+ _ [16]byte
}
func LinkCreateKprobeMulti(attr *LinkCreateKprobeMultiAttr) (*FD, error) {
@@ -683,7 +708,7 @@
AttachType AttachType
Flags uint32
BpfCookie uint64
- _ [24]byte
+ _ [40]byte
}
func LinkCreatePerfEvent(attr *LinkCreatePerfEventAttr) (*FD, error) {
@@ -702,7 +727,7 @@
TargetBtfId BTFID
_ [4]byte
Cookie uint64
- _ [16]byte
+ _ [32]byte
}
func LinkCreateTracing(attr *LinkCreateTracingAttr) (*FD, error) {
@@ -909,6 +934,8 @@
Pathname Pointer
BpfFd uint32
FileFlags uint32
+ PathFd int32
+ _ [4]byte
}
func ObjGet(attr *ObjGetAttr) (*FD, error) {
@@ -934,6 +961,8 @@
Pathname Pointer
BpfFd uint32
FileFlags uint32
+ PathFd int32
+ _ [4]byte
}
func ObjPin(attr *ObjPinAttr) error {
@@ -942,11 +971,13 @@
}
type ProgAttachAttr struct {
- TargetFd uint32
- AttachBpfFd uint32
- AttachType uint32
- AttachFlags uint32
- ReplaceBpfFd uint32
+ TargetFd uint32
+ AttachBpfFd uint32
+ AttachType uint32
+ AttachFlags uint32
+ ReplaceBpfFd uint32
+ RelativeFd uint32
+ ExpectedRevision uint64
}
func ProgAttach(attr *ProgAttachAttr) error {
@@ -1022,7 +1053,7 @@
FdArray Pointer
CoreRelos Pointer
CoreReloRecSize uint32
- _ [4]byte
+ LogTrueSize uint32
}
func ProgLoad(attr *ProgLoadAttr) (*FD, error) {
@@ -1034,13 +1065,17 @@
}
type ProgQueryAttr struct {
- TargetFd uint32
- AttachType AttachType
- QueryFlags uint32
- AttachFlags uint32
- ProgIds Pointer
- ProgCount uint32
- _ [4]byte
+ TargetFd uint32
+ AttachType AttachType
+ QueryFlags uint32
+ AttachFlags uint32
+ ProgIds Pointer
+ ProgCount uint32
+ _ [4]byte
+ ProgAttachFlags Pointer
+ LinkIds Pointer
+ LinkAttachFlags Pointer
+ Revision uint64
}
func ProgQuery(attr *ProgQueryAttr) error {
diff --git a/vendor/github.com/cilium/ebpf/internal/sysenc/buffer.go b/vendor/github.com/cilium/ebpf/internal/sysenc/buffer.go
new file mode 100644
index 0000000..c6959d9
--- /dev/null
+++ b/vendor/github.com/cilium/ebpf/internal/sysenc/buffer.go
@@ -0,0 +1,77 @@
+package sysenc
+
+import (
+ "unsafe"
+
+ "github.com/cilium/ebpf/internal/sys"
+)
+
+type Buffer struct {
+ ptr unsafe.Pointer
+ // Size of the buffer. syscallPointerOnly if created from UnsafeBuffer or when using
+ // zero-copy unmarshaling.
+ size int
+}
+
+const syscallPointerOnly = -1
+
+func newBuffer(buf []byte) Buffer {
+ if len(buf) == 0 {
+ return Buffer{}
+ }
+ return Buffer{unsafe.Pointer(&buf[0]), len(buf)}
+}
+
+// UnsafeBuffer constructs a Buffer for zero-copy unmarshaling.
+//
+// [Pointer] is the only valid method to call on such a Buffer.
+// Use [SyscallBuffer] instead if possible.
+func UnsafeBuffer(ptr unsafe.Pointer) Buffer {
+ return Buffer{ptr, syscallPointerOnly}
+}
+
+// SyscallOutput prepares a Buffer for a syscall to write into.
+//
+// The buffer may point at the underlying memory of dst, in which case [Unmarshal]
+// becomes a no-op.
+//
+// The contents of the buffer are undefined and may be non-zero.
+func SyscallOutput(dst any, size int) Buffer {
+ if dstBuf := unsafeBackingMemory(dst); len(dstBuf) == size {
+ buf := newBuffer(dstBuf)
+ buf.size = syscallPointerOnly
+ return buf
+ }
+
+ return newBuffer(make([]byte, size))
+}
+
+// CopyTo copies the buffer into dst.
+//
+// Returns the number of copied bytes.
+func (b Buffer) CopyTo(dst []byte) int {
+ return copy(dst, b.unsafeBytes())
+}
+
+// Pointer returns the location where a syscall should write.
+func (b Buffer) Pointer() sys.Pointer {
+ // NB: This deliberately ignores b.length to support zero-copy
+ // marshaling / unmarshaling using unsafe.Pointer.
+ return sys.NewPointer(b.ptr)
+}
+
+// Unmarshal the buffer into the provided value.
+func (b Buffer) Unmarshal(data any) error {
+ if b.size == syscallPointerOnly {
+ return nil
+ }
+
+ return Unmarshal(data, b.unsafeBytes())
+}
+
+func (b Buffer) unsafeBytes() []byte {
+ if b.size == syscallPointerOnly {
+ return nil
+ }
+ return unsafe.Slice((*byte)(b.ptr), b.size)
+}
diff --git a/vendor/github.com/cilium/ebpf/internal/sysenc/doc.go b/vendor/github.com/cilium/ebpf/internal/sysenc/doc.go
new file mode 100644
index 0000000..676ad98
--- /dev/null
+++ b/vendor/github.com/cilium/ebpf/internal/sysenc/doc.go
@@ -0,0 +1,3 @@
+// Package sysenc provides efficient conversion of Go values to system
+// call interfaces.
+package sysenc
diff --git a/vendor/github.com/cilium/ebpf/internal/sysenc/layout.go b/vendor/github.com/cilium/ebpf/internal/sysenc/layout.go
new file mode 100644
index 0000000..52d111e
--- /dev/null
+++ b/vendor/github.com/cilium/ebpf/internal/sysenc/layout.go
@@ -0,0 +1,41 @@
+// Copyright 2009 The Go Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style
+// license that can be found at https://go.dev/LICENSE.
+
+package sysenc
+
+import (
+ "reflect"
+ "sync"
+)
+
+var hasUnexportedFieldsCache sync.Map // map[reflect.Type]bool
+
+func hasUnexportedFields(typ reflect.Type) bool {
+ switch typ.Kind() {
+ case reflect.Slice, reflect.Array, reflect.Pointer:
+ return hasUnexportedFields(typ.Elem())
+
+ case reflect.Struct:
+ if unexported, ok := hasUnexportedFieldsCache.Load(typ); ok {
+ return unexported.(bool)
+ }
+
+ unexported := false
+ for i, n := 0, typ.NumField(); i < n; i++ {
+ field := typ.Field(i)
+ // Package binary allows _ fields but always writes zeroes into them.
+ if (!field.IsExported() && field.Name != "_") || hasUnexportedFields(field.Type) {
+ unexported = true
+ break
+ }
+ }
+
+ hasUnexportedFieldsCache.Store(typ, unexported)
+ return unexported
+
+ default:
+ // NB: It's not clear what this means for Chan and so on.
+ return false
+ }
+}
diff --git a/vendor/github.com/cilium/ebpf/internal/sysenc/marshal.go b/vendor/github.com/cilium/ebpf/internal/sysenc/marshal.go
new file mode 100644
index 0000000..5233945
--- /dev/null
+++ b/vendor/github.com/cilium/ebpf/internal/sysenc/marshal.go
@@ -0,0 +1,178 @@
+package sysenc
+
+import (
+ "bytes"
+ "encoding"
+ "encoding/binary"
+ "errors"
+ "fmt"
+ "reflect"
+ "sync"
+ "unsafe"
+
+ "github.com/cilium/ebpf/internal"
+
+ "golang.org/x/exp/slices"
+)
+
+// Marshal turns data into a byte slice using the system's native endianness.
+//
+// If possible, avoids allocations by directly using the backing memory
+// of data. This means that the variable must not be modified for the lifetime
+// of the returned [Buffer].
+//
+// Returns an error if the data can't be turned into a byte slice according to
+// the behaviour of [binary.Write].
+func Marshal(data any, size int) (Buffer, error) {
+ if data == nil {
+ return Buffer{}, errors.New("can't marshal a nil value")
+ }
+
+ var buf []byte
+ var err error
+ switch value := data.(type) {
+ case encoding.BinaryMarshaler:
+ buf, err = value.MarshalBinary()
+ case string:
+ buf = unsafe.Slice(unsafe.StringData(value), len(value))
+ case []byte:
+ buf = value
+ case int16:
+ buf = internal.NativeEndian.AppendUint16(make([]byte, 0, 2), uint16(value))
+ case uint16:
+ buf = internal.NativeEndian.AppendUint16(make([]byte, 0, 2), value)
+ case int32:
+ buf = internal.NativeEndian.AppendUint32(make([]byte, 0, 4), uint32(value))
+ case uint32:
+ buf = internal.NativeEndian.AppendUint32(make([]byte, 0, 4), value)
+ case int64:
+ buf = internal.NativeEndian.AppendUint64(make([]byte, 0, 8), uint64(value))
+ case uint64:
+ buf = internal.NativeEndian.AppendUint64(make([]byte, 0, 8), value)
+ default:
+ if buf := unsafeBackingMemory(data); len(buf) == size {
+ return newBuffer(buf), nil
+ }
+
+ wr := internal.NewBuffer(make([]byte, 0, size))
+ defer internal.PutBuffer(wr)
+
+ err = binary.Write(wr, internal.NativeEndian, value)
+ buf = wr.Bytes()
+ }
+ if err != nil {
+ return Buffer{}, err
+ }
+
+ if len(buf) != size {
+ return Buffer{}, fmt.Errorf("%T doesn't marshal to %d bytes", data, size)
+ }
+
+ return newBuffer(buf), nil
+}
+
+var bytesReaderPool = sync.Pool{
+ New: func() interface{} {
+ return new(bytes.Reader)
+ },
+}
+
+// Unmarshal a byte slice in the system's native endianness into data.
+//
+// Returns an error if buf can't be unmarshalled according to the behaviour
+// of [binary.Read].
+func Unmarshal(data interface{}, buf []byte) error {
+ switch value := data.(type) {
+ case encoding.BinaryUnmarshaler:
+ return value.UnmarshalBinary(buf)
+
+ case *string:
+ *value = string(buf)
+ return nil
+
+ case *[]byte:
+ // Backwards compat: unmarshaling into a slice replaces the whole slice.
+ *value = slices.Clone(buf)
+ return nil
+
+ default:
+ if dataBuf := unsafeBackingMemory(data); len(dataBuf) == len(buf) {
+ copy(dataBuf, buf)
+ return nil
+ }
+
+ rd := bytesReaderPool.Get().(*bytes.Reader)
+ defer bytesReaderPool.Put(rd)
+
+ rd.Reset(buf)
+
+ if err := binary.Read(rd, internal.NativeEndian, value); err != nil {
+ return err
+ }
+
+ if rd.Len() != 0 {
+ return fmt.Errorf("unmarshaling %T doesn't consume all data", data)
+ }
+
+ return nil
+ }
+}
+
+// unsafeBackingMemory returns the backing memory of data if it can be used
+// instead of calling into package binary.
+//
+// Returns nil if the value is not a pointer or a slice, or if it contains
+// padding or unexported fields.
+func unsafeBackingMemory(data any) []byte {
+ if data == nil {
+ return nil
+ }
+
+ value := reflect.ValueOf(data)
+ var valueSize int
+ switch value.Kind() {
+ case reflect.Pointer:
+ if value.IsNil() {
+ return nil
+ }
+
+ if elemType := value.Type().Elem(); elemType.Kind() != reflect.Slice {
+ valueSize = int(elemType.Size())
+ break
+ }
+
+ // We're dealing with a pointer to a slice. Dereference and
+ // handle it like a regular slice.
+ value = value.Elem()
+ fallthrough
+
+ case reflect.Slice:
+ valueSize = int(value.Type().Elem().Size()) * value.Len()
+
+ default:
+ // Prevent Value.UnsafePointer from panicking.
+ return nil
+ }
+
+ // Some nil pointer types currently crash binary.Size. Call it after our own
+ // code so that the panic isn't reachable.
+ // See https://github.com/golang/go/issues/60892
+ if size := binary.Size(data); size == -1 || size != valueSize {
+ // The type contains padding or unsupported types.
+ return nil
+ }
+
+ if hasUnexportedFields(reflect.TypeOf(data)) {
+ return nil
+ }
+
+ // Reinterpret the pointer as a byte slice. This violates the unsafe.Pointer
+ // rules because it's very unlikely that the source data has "an equivalent
+ // memory layout". However, we can make it safe-ish because of the
+ // following reasons:
+ // - There is no alignment mismatch since we cast to a type with an
+ // alignment of 1.
+ // - There are no pointers in the source type so we don't upset the GC.
+ // - The length is verified at runtime.
+ return unsafe.Slice((*byte)(value.UnsafePointer()), valueSize)
+}
diff --git a/vendor/github.com/cilium/ebpf/internal/tracefs/kprobe.go b/vendor/github.com/cilium/ebpf/internal/tracefs/kprobe.go
index 4059a09..1b45a9a 100644
--- a/vendor/github.com/cilium/ebpf/internal/tracefs/kprobe.go
+++ b/vendor/github.com/cilium/ebpf/internal/tracefs/kprobe.go
@@ -20,7 +20,7 @@
ErrInvalidMaxActive = errors.New("can only set maxactive on kretprobes")
)
-//go:generate stringer -type=ProbeType -linecomment
+//go:generate go run golang.org/x/tools/cmd/stringer@latest -type=ProbeType -linecomment
type ProbeType uint8
diff --git a/vendor/github.com/cilium/ebpf/internal/unix/types_linux.go b/vendor/github.com/cilium/ebpf/internal/unix/types_linux.go
index 7c97059..51ed7d0 100644
--- a/vendor/github.com/cilium/ebpf/internal/unix/types_linux.go
+++ b/vendor/github.com/cilium/ebpf/internal/unix/types_linux.go
@@ -85,6 +85,8 @@
BPF_FS_MAGIC = linux.BPF_FS_MAGIC
TRACEFS_MAGIC = linux.TRACEFS_MAGIC
DEBUGFS_MAGIC = linux.DEBUGFS_MAGIC
+ BPF_RB_NO_WAKEUP = linux.BPF_RB_NO_WAKEUP
+ BPF_RB_FORCE_WAKEUP = linux.BPF_RB_FORCE_WAKEUP
)
type Statfs_t = linux.Statfs_t
diff --git a/vendor/github.com/cilium/ebpf/internal/unix/types_other.go b/vendor/github.com/cilium/ebpf/internal/unix/types_other.go
index 5e86b50..1760e9e 100644
--- a/vendor/github.com/cilium/ebpf/internal/unix/types_other.go
+++ b/vendor/github.com/cilium/ebpf/internal/unix/types_other.go
@@ -89,6 +89,8 @@
BPF_FS_MAGIC
TRACEFS_MAGIC
DEBUGFS_MAGIC
+ BPF_RB_NO_WAKEUP
+ BPF_RB_FORCE_WAKEUP
)
type Statfs_t struct {
diff --git a/vendor/github.com/cilium/ebpf/internal/vdso.go b/vendor/github.com/cilium/ebpf/internal/vdso.go
index 10e639b..c444a41 100644
--- a/vendor/github.com/cilium/ebpf/internal/vdso.go
+++ b/vendor/github.com/cilium/ebpf/internal/vdso.go
@@ -8,6 +8,7 @@
"io"
"math"
"os"
+ "unsafe"
"github.com/cilium/ebpf/internal/unix"
)
@@ -19,6 +20,8 @@
// vdsoVersion returns the LINUX_VERSION_CODE embedded in the vDSO library
// linked into the current process image.
func vdsoVersion() (uint32, error) {
+ const uintptrIs32bits = unsafe.Sizeof((uintptr)(0)) == 4
+
// Read data from the auxiliary vector, which is normally passed directly
// to the process. Go does not expose that data, so we must read it from procfs.
// https://man7.org/linux/man-pages/man3/getauxval.3.html
@@ -31,7 +34,7 @@
}
defer av.Close()
- vdsoAddr, err := vdsoMemoryAddress(av)
+ vdsoAddr, err := vdsoMemoryAddress(av, NativeEndian, uintptrIs32bits)
if err != nil {
return 0, fmt.Errorf("finding vDSO memory address: %w", err)
}
@@ -52,9 +55,34 @@
return c, nil
}
+type auxvPair32 struct {
+ Tag, Value uint32
+}
+
+type auxvPair64 struct {
+ Tag, Value uint64
+}
+
+func readAuxvPair(r io.Reader, order binary.ByteOrder, uintptrIs32bits bool) (tag, value uint64, _ error) {
+ if uintptrIs32bits {
+ var aux auxvPair32
+ if err := binary.Read(r, order, &aux); err != nil {
+ return 0, 0, fmt.Errorf("reading auxv entry: %w", err)
+ }
+ return uint64(aux.Tag), uint64(aux.Value), nil
+ }
+
+ var aux auxvPair64
+ if err := binary.Read(r, order, &aux); err != nil {
+ return 0, 0, fmt.Errorf("reading auxv entry: %w", err)
+ }
+ return aux.Tag, aux.Value, nil
+}
+
// vdsoMemoryAddress returns the memory address of the vDSO library
// linked into the current process image. r is an io.Reader into an auxv blob.
-func vdsoMemoryAddress(r io.Reader) (uint64, error) {
+func vdsoMemoryAddress(r io.Reader, order binary.ByteOrder, uintptrIs32bits bool) (uintptr, error) {
+ // See https://elixir.bootlin.com/linux/v6.5.5/source/include/uapi/linux/auxvec.h
const (
_AT_NULL = 0 // End of vector
_AT_SYSINFO_EHDR = 33 // Offset to vDSO blob in process image
@@ -62,16 +90,16 @@
// Loop through all tag/value pairs in auxv until we find `AT_SYSINFO_EHDR`,
// the address of a page containing the virtual Dynamic Shared Object (vDSO).
- aux := struct{ Tag, Val uint64 }{}
for {
- if err := binary.Read(r, NativeEndian, &aux); err != nil {
- return 0, fmt.Errorf("reading auxv entry: %w", err)
+ tag, value, err := readAuxvPair(r, order, uintptrIs32bits)
+ if err != nil {
+ return 0, err
}
- switch aux.Tag {
+ switch tag {
case _AT_SYSINFO_EHDR:
- if aux.Val != 0 {
- return aux.Val, nil
+ if value != 0 {
+ return uintptr(value), nil
}
return 0, fmt.Errorf("invalid vDSO address in auxv")
// _AT_NULL is always the last tag/val pair in the aux vector
diff --git a/vendor/github.com/cilium/ebpf/link/iter.go b/vendor/github.com/cilium/ebpf/link/iter.go
index d2b32ef..0a39fae 100644
--- a/vendor/github.com/cilium/ebpf/link/iter.go
+++ b/vendor/github.com/cilium/ebpf/link/iter.go
@@ -25,10 +25,6 @@
// AttachIter attaches a BPF seq_file iterator.
func AttachIter(opts IterOptions) (*Iter, error) {
- if err := haveBPFLink(); err != nil {
- return nil, err
- }
-
progFd := opts.Program.FD()
if progFd < 0 {
return nil, fmt.Errorf("invalid program: %s", sys.ErrClosedFd)
@@ -52,6 +48,9 @@
fd, err := sys.LinkCreateIter(&attr)
if err != nil {
+ if haveFeatErr := haveBPFLink(); haveFeatErr != nil {
+ return nil, haveFeatErr
+ }
return nil, fmt.Errorf("can't link iterator: %w", err)
}
diff --git a/vendor/github.com/cilium/ebpf/link/kprobe_multi.go b/vendor/github.com/cilium/ebpf/link/kprobe_multi.go
index 697c6d7..4d364d8 100644
--- a/vendor/github.com/cilium/ebpf/link/kprobe_multi.go
+++ b/vendor/github.com/cilium/ebpf/link/kprobe_multi.go
@@ -82,10 +82,6 @@
return nil, fmt.Errorf("Cookies must be exactly Symbols or Addresses in length: %w", errInvalidInput)
}
- if err := haveBPFLinkKprobeMulti(); err != nil {
- return nil, err
- }
-
attr := &sys.LinkCreateKprobeMultiAttr{
ProgFd: uint32(prog.FD()),
AttachType: sys.BPF_TRACE_KPROBE_MULTI,
@@ -113,7 +109,11 @@
if errors.Is(err, unix.EINVAL) {
return nil, fmt.Errorf("%w (missing kernel symbol or prog's AttachType not AttachTraceKprobeMulti?)", err)
}
+
if err != nil {
+ if haveFeatErr := haveBPFLinkKprobeMulti(); haveFeatErr != nil {
+ return nil, haveFeatErr
+ }
return nil, err
}
diff --git a/vendor/github.com/cilium/ebpf/link/program.go b/vendor/github.com/cilium/ebpf/link/program.go
index ea31817..053735a 100644
--- a/vendor/github.com/cilium/ebpf/link/program.go
+++ b/vendor/github.com/cilium/ebpf/link/program.go
@@ -25,10 +25,6 @@
// You should use one of the higher level abstractions available in this
// package if possible.
func RawAttachProgram(opts RawAttachProgramOptions) error {
- if err := haveProgAttach(); err != nil {
- return err
- }
-
var replaceFd uint32
if opts.Replace != nil {
replaceFd = uint32(opts.Replace.FD())
@@ -43,8 +39,12 @@
}
if err := sys.ProgAttach(&attr); err != nil {
+ if haveFeatErr := haveProgAttach(); haveFeatErr != nil {
+ return haveFeatErr
+ }
return fmt.Errorf("can't attach program: %w", err)
}
+
return nil
}
@@ -59,16 +59,15 @@
// You should use one of the higher level abstractions available in this
// package if possible.
func RawDetachProgram(opts RawDetachProgramOptions) error {
- if err := haveProgAttach(); err != nil {
- return err
- }
-
attr := sys.ProgDetachAttr{
TargetFd: uint32(opts.Target),
AttachBpfFd: uint32(opts.Program.FD()),
AttachType: uint32(opts.Attach),
}
if err := sys.ProgDetach(&attr); err != nil {
+ if haveFeatErr := haveProgAttach(); haveFeatErr != nil {
+ return haveFeatErr
+ }
return fmt.Errorf("can't detach program: %w", err)
}
diff --git a/vendor/github.com/cilium/ebpf/link/syscalls.go b/vendor/github.com/cilium/ebpf/link/syscalls.go
index c9c998c..012970e 100644
--- a/vendor/github.com/cilium/ebpf/link/syscalls.go
+++ b/vendor/github.com/cilium/ebpf/link/syscalls.go
@@ -60,9 +60,11 @@
asm.Return(),
},
})
+
if err != nil {
return internal.ErrNotSupported
}
+
defer prog.Close()
// We know that we have BPF_PROG_ATTACH since we can load CGroupSKB programs.
@@ -113,11 +115,12 @@
}
err := sys.ProgQuery(&attr)
- if errors.Is(err, unix.EINVAL) {
- return internal.ErrNotSupported
- }
+
if errors.Is(err, unix.EBADF) {
return nil
}
- return err
+ if err != nil {
+ return ErrNotSupported
+ }
+ return errors.New("syscall succeeded unexpectedly")
})
diff --git a/vendor/github.com/cilium/ebpf/link/uprobe.go b/vendor/github.com/cilium/ebpf/link/uprobe.go
index 272bac4..83977e0 100644
--- a/vendor/github.com/cilium/ebpf/link/uprobe.go
+++ b/vendor/github.com/cilium/ebpf/link/uprobe.go
@@ -18,9 +18,12 @@
uprobeRefCtrOffsetShift = 32
haveRefCtrOffsetPMU = internal.NewFeatureTest("RefCtrOffsetPMU", "4.20", func() error {
_, err := os.Stat(uprobeRefCtrOffsetPMUPath)
- if err != nil {
+ if errors.Is(err, os.ErrNotExist) {
return internal.ErrNotSupported
}
+ if err != nil {
+ return err
+ }
return nil
})
diff --git a/vendor/github.com/cilium/ebpf/linker.go b/vendor/github.com/cilium/ebpf/linker.go
index e0dbfcf..cf5b02d 100644
--- a/vendor/github.com/cilium/ebpf/linker.go
+++ b/vendor/github.com/cilium/ebpf/linker.go
@@ -7,6 +7,8 @@
"io"
"math"
+ "golang.org/x/exp/slices"
+
"github.com/cilium/ebpf/asm"
"github.com/cilium/ebpf/btf"
"github.com/cilium/ebpf/internal"
@@ -40,10 +42,12 @@
return fda
}
-func (hs handles) close() {
- for _, h := range hs {
- h.Close()
+func (hs *handles) Close() error {
+ var errs []error
+ for _, h := range *hs {
+ errs = append(errs, h.Close())
}
+ return errors.Join(errs...)
}
// splitSymbols splits insns into subsections delimited by Symbol Instructions.
@@ -55,21 +59,33 @@
return nil, errors.New("insns is empty")
}
- if insns[0].Symbol() == "" {
+ currentSym := insns[0].Symbol()
+ if currentSym == "" {
return nil, errors.New("insns must start with a Symbol")
}
- var name string
+ start := 0
progs := make(map[string]asm.Instructions)
- for _, ins := range insns {
- if sym := ins.Symbol(); sym != "" {
- if progs[sym] != nil {
- return nil, fmt.Errorf("insns contains duplicate Symbol %s", sym)
- }
- name = sym
+ for i, ins := range insns[1:] {
+ i := i + 1
+
+ sym := ins.Symbol()
+ if sym == "" {
+ continue
}
- progs[name] = append(progs[name], ins)
+ // New symbol, flush the old one out.
+ progs[currentSym] = slices.Clone(insns[start:i])
+
+ if progs[sym] != nil {
+ return nil, fmt.Errorf("insns contains duplicate Symbol %s", sym)
+ }
+ currentSym = sym
+ start = i
+ }
+
+ if tail := insns[start:]; len(tail) > 0 {
+ progs[currentSym] = slices.Clone(tail)
}
return progs, nil
@@ -231,7 +247,13 @@
// fixupKfuncs loops over all instructions in search for kfunc calls.
// If at least one is found, the current kernels BTF and module BTFis are searched to set Instruction.Constant
// and Instruction.Offset to the correct values.
-func fixupKfuncs(insns asm.Instructions) (handles, error) {
+func fixupKfuncs(insns asm.Instructions) (_ handles, err error) {
+ closeOnError := func(c io.Closer) {
+ if err != nil {
+ c.Close()
+ }
+ }
+
iter := insns.Iterate()
for iter.Next() {
ins := iter.Ins
@@ -250,6 +272,8 @@
}
fdArray := make(handles, 0)
+ defer closeOnError(&fdArray)
+
for {
ins := iter.Ins
@@ -276,6 +300,11 @@
return nil, err
}
+ idx, err := fdArray.add(module)
+ if err != nil {
+ return nil, err
+ }
+
if err := btf.CheckTypeCompatibility(kfm.Type, target.(*btf.Func).Type); err != nil {
return nil, &incompatibleKfuncError{kfm.Name, err}
}
@@ -285,11 +314,6 @@
return nil, err
}
- idx, err := fdArray.add(module)
- if err != nil {
- return nil, err
- }
-
ins.Constant = int64(id)
ins.Offset = int16(idx)
diff --git a/vendor/github.com/cilium/ebpf/map.go b/vendor/github.com/cilium/ebpf/map.go
index a11664c..ce945ac 100644
--- a/vendor/github.com/cilium/ebpf/map.go
+++ b/vendor/github.com/cilium/ebpf/map.go
@@ -9,12 +9,14 @@
"os"
"path/filepath"
"reflect"
+ "strings"
"time"
"unsafe"
"github.com/cilium/ebpf/btf"
"github.com/cilium/ebpf/internal"
"github.com/cilium/ebpf/internal/sys"
+ "github.com/cilium/ebpf/internal/sysenc"
"github.com/cilium/ebpf/internal/unix"
)
@@ -102,26 +104,55 @@
return &cpy
}
-func (ms *MapSpec) clampPerfEventArraySize() error {
- if ms.Type != PerfEventArray {
- return nil
+// fixupMagicFields fills fields of MapSpec which are usually
+// left empty in ELF or which depend on runtime information.
+//
+// The method doesn't modify Spec, instead returning a copy.
+// The copy is only performed if fixups are necessary, so callers mustn't mutate
+// the returned spec.
+func (spec *MapSpec) fixupMagicFields() (*MapSpec, error) {
+ switch spec.Type {
+ case ArrayOfMaps, HashOfMaps:
+ if spec.ValueSize != 0 && spec.ValueSize != 4 {
+ return nil, errors.New("ValueSize must be zero or four for map of map")
+ }
+
+ spec = spec.Copy()
+ spec.ValueSize = 4
+
+ case PerfEventArray:
+ if spec.KeySize != 0 && spec.KeySize != 4 {
+ return nil, errors.New("KeySize must be zero or four for perf event array")
+ }
+
+ if spec.ValueSize != 0 && spec.ValueSize != 4 {
+ return nil, errors.New("ValueSize must be zero or four for perf event array")
+ }
+
+ spec = spec.Copy()
+ spec.KeySize = 4
+ spec.ValueSize = 4
+
+ n, err := internal.PossibleCPUs()
+ if err != nil {
+ return nil, fmt.Errorf("fixup perf event array: %w", err)
+ }
+
+ if n := uint32(n); spec.MaxEntries == 0 || spec.MaxEntries > n {
+ // MaxEntries should be zero most of the time, but there is code
+ // out there which hardcodes large constants. Clamp the number
+ // of entries to the number of CPUs at most. Allow creating maps with
+ // less than n items since some kernel selftests relied on this
+ // behaviour in the past.
+ spec.MaxEntries = n
+ }
}
- n, err := internal.PossibleCPUs()
- if err != nil {
- return fmt.Errorf("perf event array: %w", err)
- }
-
- if n := uint32(n); ms.MaxEntries > n {
- ms.MaxEntries = n
- }
-
- return nil
+ return spec, nil
}
// dataSection returns the contents and BTF Datasec descriptor of the spec.
func (ms *MapSpec) dataSection() ([]byte, *btf.Datasec, error) {
-
if ms.Value == nil {
return nil, nil, errMapNoBTFValue
}
@@ -155,27 +186,37 @@
//
// Returns an error wrapping [ErrMapIncompatible] otherwise.
func (ms *MapSpec) Compatible(m *Map) error {
- switch {
- case m.typ != ms.Type:
- return fmt.Errorf("expected type %v, got %v: %w", ms.Type, m.typ, ErrMapIncompatible)
-
- case m.keySize != ms.KeySize:
- return fmt.Errorf("expected key size %v, got %v: %w", ms.KeySize, m.keySize, ErrMapIncompatible)
-
- case m.valueSize != ms.ValueSize:
- return fmt.Errorf("expected value size %v, got %v: %w", ms.ValueSize, m.valueSize, ErrMapIncompatible)
-
- case !(ms.Type == PerfEventArray && ms.MaxEntries == 0) &&
- m.maxEntries != ms.MaxEntries:
- return fmt.Errorf("expected max entries %v, got %v: %w", ms.MaxEntries, m.maxEntries, ErrMapIncompatible)
-
- // BPF_F_RDONLY_PROG is set unconditionally for devmaps. Explicitly allow
- // this mismatch.
- case !((ms.Type == DevMap || ms.Type == DevMapHash) && m.flags^ms.Flags == unix.BPF_F_RDONLY_PROG) &&
- m.flags != ms.Flags:
- return fmt.Errorf("expected flags %v, got %v: %w", ms.Flags, m.flags, ErrMapIncompatible)
+ ms, err := ms.fixupMagicFields()
+ if err != nil {
+ return err
}
- return nil
+
+ diffs := []string{}
+ if m.typ != ms.Type {
+ diffs = append(diffs, fmt.Sprintf("Type: %s changed to %s", m.typ, ms.Type))
+ }
+ if m.keySize != ms.KeySize {
+ diffs = append(diffs, fmt.Sprintf("KeySize: %d changed to %d", m.keySize, ms.KeySize))
+ }
+ if m.valueSize != ms.ValueSize {
+ diffs = append(diffs, fmt.Sprintf("ValueSize: %d changed to %d", m.valueSize, ms.ValueSize))
+ }
+ if m.maxEntries != ms.MaxEntries {
+ diffs = append(diffs, fmt.Sprintf("MaxEntries: %d changed to %d", m.maxEntries, ms.MaxEntries))
+ }
+
+ // BPF_F_RDONLY_PROG is set unconditionally for devmaps. Explicitly allow this
+ // mismatch.
+ if !((ms.Type == DevMap || ms.Type == DevMapHash) && m.flags^ms.Flags == unix.BPF_F_RDONLY_PROG) &&
+ m.flags != ms.Flags {
+ diffs = append(diffs, fmt.Sprintf("Flags: %d changed to %d", m.flags, ms.Flags))
+ }
+
+ if len(diffs) == 0 {
+ return nil
+ }
+
+ return fmt.Errorf("%s: %w", strings.Join(diffs, ", "), ErrMapIncompatible)
}
// Map represents a Map file descriptor.
@@ -350,60 +391,9 @@
}
}
- switch spec.Type {
- case ArrayOfMaps, HashOfMaps:
- if err := haveNestedMaps(); err != nil {
- return nil, err
- }
-
- if spec.ValueSize != 0 && spec.ValueSize != 4 {
- return nil, errors.New("ValueSize must be zero or four for map of map")
- }
-
- spec = spec.Copy()
- spec.ValueSize = 4
-
- case PerfEventArray:
- if spec.KeySize != 0 && spec.KeySize != 4 {
- return nil, errors.New("KeySize must be zero or four for perf event array")
- }
-
- if spec.ValueSize != 0 && spec.ValueSize != 4 {
- return nil, errors.New("ValueSize must be zero or four for perf event array")
- }
-
- spec = spec.Copy()
- spec.KeySize = 4
- spec.ValueSize = 4
-
- if spec.MaxEntries == 0 {
- n, err := internal.PossibleCPUs()
- if err != nil {
- return nil, fmt.Errorf("perf event array: %w", err)
- }
- spec.MaxEntries = uint32(n)
- }
- }
-
- if spec.Flags&(unix.BPF_F_RDONLY_PROG|unix.BPF_F_WRONLY_PROG) > 0 || spec.Freeze {
- if err := haveMapMutabilityModifiers(); err != nil {
- return nil, fmt.Errorf("map create: %w", err)
- }
- }
- if spec.Flags&unix.BPF_F_MMAPABLE > 0 {
- if err := haveMmapableMaps(); err != nil {
- return nil, fmt.Errorf("map create: %w", err)
- }
- }
- if spec.Flags&unix.BPF_F_INNER_MAP > 0 {
- if err := haveInnerMaps(); err != nil {
- return nil, fmt.Errorf("map create: %w", err)
- }
- }
- if spec.Flags&unix.BPF_F_NO_PREALLOC > 0 {
- if err := haveNoPreallocMaps(); err != nil {
- return nil, fmt.Errorf("map create: %w", err)
- }
+ spec, err = spec.fixupMagicFields()
+ if err != nil {
+ return nil, err
}
attr := sys.MapCreateAttr{
@@ -440,38 +430,72 @@
}
fd, err := sys.MapCreate(&attr)
+
// Some map types don't support BTF k/v in earlier kernel versions.
// Remove BTF metadata and retry map creation.
if (errors.Is(err, sys.ENOTSUPP) || errors.Is(err, unix.EINVAL)) && attr.BtfFd != 0 {
attr.BtfFd, attr.BtfKeyTypeId, attr.BtfValueTypeId = 0, 0, 0
fd, err = sys.MapCreate(&attr)
}
-
if err != nil {
- if errors.Is(err, unix.EPERM) {
- return nil, fmt.Errorf("map create: %w (MEMLOCK may be too low, consider rlimit.RemoveMemlock)", err)
- }
- if errors.Is(err, unix.EINVAL) && attr.MaxEntries == 0 {
- return nil, fmt.Errorf("map create: %w (MaxEntries may be incorrectly set to zero)", err)
- }
- if errors.Is(err, unix.EINVAL) && spec.Type == UnspecifiedMap {
- return nil, fmt.Errorf("map create: cannot use type %s", UnspecifiedMap)
- }
- if attr.BtfFd == 0 {
- return nil, fmt.Errorf("map create: %w (without BTF k/v)", err)
- }
- return nil, fmt.Errorf("map create: %w", err)
+ return nil, handleMapCreateError(attr, spec, err)
}
- defer closeOnError(fd)
+ defer closeOnError(fd)
m, err := newMap(fd, spec.Name, spec.Type, spec.KeySize, spec.ValueSize, spec.MaxEntries, spec.Flags)
if err != nil {
return nil, fmt.Errorf("map create: %w", err)
}
-
return m, nil
}
+func handleMapCreateError(attr sys.MapCreateAttr, spec *MapSpec, err error) error {
+ if errors.Is(err, unix.EPERM) {
+ return fmt.Errorf("map create: %w (MEMLOCK may be too low, consider rlimit.RemoveMemlock)", err)
+ }
+ if errors.Is(err, unix.EINVAL) && spec.MaxEntries == 0 {
+ return fmt.Errorf("map create: %w (MaxEntries may be incorrectly set to zero)", err)
+ }
+ if errors.Is(err, unix.EINVAL) && spec.Type == UnspecifiedMap {
+ return fmt.Errorf("map create: cannot use type %s", UnspecifiedMap)
+ }
+ if errors.Is(err, unix.EINVAL) && spec.Flags&unix.BPF_F_NO_PREALLOC > 0 {
+ return fmt.Errorf("map create: %w (noPrealloc flag may be incompatible with map type %s)", err, spec.Type)
+ }
+
+ switch spec.Type {
+ case ArrayOfMaps, HashOfMaps:
+ if haveFeatErr := haveNestedMaps(); haveFeatErr != nil {
+ return fmt.Errorf("map create: %w", haveFeatErr)
+ }
+ }
+ if spec.Flags&(unix.BPF_F_RDONLY_PROG|unix.BPF_F_WRONLY_PROG) > 0 || spec.Freeze {
+ if haveFeatErr := haveMapMutabilityModifiers(); haveFeatErr != nil {
+ return fmt.Errorf("map create: %w", haveFeatErr)
+ }
+ }
+ if spec.Flags&unix.BPF_F_MMAPABLE > 0 {
+ if haveFeatErr := haveMmapableMaps(); haveFeatErr != nil {
+ return fmt.Errorf("map create: %w", haveFeatErr)
+ }
+ }
+ if spec.Flags&unix.BPF_F_INNER_MAP > 0 {
+ if haveFeatErr := haveInnerMaps(); haveFeatErr != nil {
+ return fmt.Errorf("map create: %w", haveFeatErr)
+ }
+ }
+ if spec.Flags&unix.BPF_F_NO_PREALLOC > 0 {
+ if haveFeatErr := haveNoPreallocMaps(); haveFeatErr != nil {
+ return fmt.Errorf("map create: %w", haveFeatErr)
+ }
+ }
+ if attr.BtfFd == 0 {
+ return fmt.Errorf("map create: %w (without BTF k/v)", err)
+ }
+
+ return fmt.Errorf("map create: %w", err)
+}
+
// newMap allocates and returns a new Map structure.
// Sets the fullValueSize on per-CPU maps.
func newMap(fd *sys.FD, name string, typ MapType, keySize, valueSize, maxEntries, flags uint32) (*Map, error) {
@@ -568,8 +592,8 @@
return m.lookupPerCPU(key, valueOut, flags)
}
- valuePtr, valueBytes := makeBuffer(valueOut, m.fullValueSize)
- if err := m.lookup(key, valuePtr, flags); err != nil {
+ valueBytes := makeMapSyscallOutput(valueOut, m.fullValueSize)
+ if err := m.lookup(key, valueBytes.Pointer(), flags); err != nil {
return err
}
@@ -595,8 +619,8 @@
return m.lookupAndDeletePerCPU(key, valueOut, flags)
}
- valuePtr, valueBytes := makeBuffer(valueOut, m.fullValueSize)
- if err := m.lookupAndDelete(key, valuePtr, flags); err != nil {
+ valueBytes := makeMapSyscallOutput(valueOut, m.fullValueSize)
+ if err := m.lookupAndDelete(key, valueBytes.Pointer(), flags); err != nil {
return err
}
return m.unmarshalValue(valueOut, valueBytes)
@@ -764,13 +788,13 @@
//
// Returns ErrKeyNotExist if there is no next key.
func (m *Map) NextKey(key, nextKeyOut interface{}) error {
- nextKeyPtr, nextKeyBytes := makeBuffer(nextKeyOut, int(m.keySize))
+ nextKeyBytes := makeMapSyscallOutput(nextKeyOut, int(m.keySize))
- if err := m.nextKey(key, nextKeyPtr); err != nil {
+ if err := m.nextKey(key, nextKeyBytes.Pointer()); err != nil {
return err
}
- if err := m.unmarshalKey(nextKeyOut, nextKeyBytes); err != nil {
+ if err := nextKeyBytes.Unmarshal(nextKeyOut); err != nil {
return fmt.Errorf("can't unmarshal next key: %w", err)
}
return nil
@@ -941,14 +965,14 @@
keyPtr := sys.NewSlicePointer(keyBuf)
valueBuf := make([]byte, count*int(m.fullValueSize))
valuePtr := sys.NewSlicePointer(valueBuf)
- nextPtr, nextBuf := makeBuffer(nextKeyOut, int(m.keySize))
+ nextBuf := makeMapSyscallOutput(nextKeyOut, int(m.keySize))
attr := sys.MapLookupBatchAttr{
MapFd: m.fd.Uint(),
Keys: keyPtr,
Values: valuePtr,
Count: uint32(count),
- OutBatch: nextPtr,
+ OutBatch: nextBuf.Pointer(),
}
if opts != nil {
@@ -958,7 +982,7 @@
var err error
if startKey != nil {
- attr.InBatch, err = marshalPtr(startKey, int(m.keySize))
+ attr.InBatch, err = marshalMapSyscallInput(startKey, int(m.keySize))
if err != nil {
return 0, err
}
@@ -970,15 +994,15 @@
return 0, sysErr
}
- err = m.unmarshalKey(nextKeyOut, nextBuf)
+ err = nextBuf.Unmarshal(nextKeyOut)
if err != nil {
return 0, err
}
- err = unmarshalBytes(keysOut, keyBuf)
+ err = sysenc.Unmarshal(keysOut, keyBuf)
if err != nil {
return 0, err
}
- err = unmarshalBytes(valuesOut, valueBuf)
+ err = sysenc.Unmarshal(valuesOut, valueBuf)
if err != nil {
return 0, err
}
@@ -991,9 +1015,6 @@
// "keys" and "values" must be of type slice, a pointer
// to a slice or buffer will not work.
func (m *Map) BatchUpdate(keys, values interface{}, opts *BatchOptions) (int, error) {
- if err := haveBatchAPI(); err != nil {
- return 0, err
- }
if m.typ.hasPerCPUValue() {
return 0, ErrNotSupported
}
@@ -1013,11 +1034,11 @@
if count != valuesValue.Len() {
return 0, fmt.Errorf("keys and values must be the same length")
}
- keyPtr, err := marshalPtr(keys, count*int(m.keySize))
+ keyPtr, err := marshalMapSyscallInput(keys, count*int(m.keySize))
if err != nil {
return 0, err
}
- valuePtr, err = marshalPtr(values, count*int(m.valueSize))
+ valuePtr, err = marshalMapSyscallInput(values, count*int(m.valueSize))
if err != nil {
return 0, err
}
@@ -1035,6 +1056,9 @@
err = sys.MapUpdateBatch(&attr)
if err != nil {
+ if haveFeatErr := haveBatchAPI(); haveFeatErr != nil {
+ return 0, haveFeatErr
+ }
return int(attr.Count), fmt.Errorf("batch update: %w", wrapMapError(err))
}
@@ -1044,9 +1068,6 @@
// BatchDelete batch deletes entries in the map by keys.
// "keys" must be of type slice, a pointer to a slice or buffer will not work.
func (m *Map) BatchDelete(keys interface{}, opts *BatchOptions) (int, error) {
- if err := haveBatchAPI(); err != nil {
- return 0, err
- }
if m.typ.hasPerCPUValue() {
return 0, ErrNotSupported
}
@@ -1055,7 +1076,7 @@
return 0, fmt.Errorf("keys must be a slice")
}
count := keysValue.Len()
- keyPtr, err := marshalPtr(keys, count*int(m.keySize))
+ keyPtr, err := marshalMapSyscallInput(keys, count*int(m.keySize))
if err != nil {
return 0, fmt.Errorf("cannot marshal keys: %v", err)
}
@@ -1072,6 +1093,9 @@
}
if err = sys.MapDeleteBatch(&attr); err != nil {
+ if haveFeatErr := haveBatchAPI(); haveFeatErr != nil {
+ return 0, haveFeatErr
+ }
return int(attr.Count), fmt.Errorf("batch delete: %w", wrapMapError(err))
}
@@ -1176,15 +1200,14 @@
//
// It makes no changes to kernel-side restrictions.
func (m *Map) Freeze() error {
- if err := haveMapMutabilityModifiers(); err != nil {
- return fmt.Errorf("can't freeze map: %w", err)
- }
-
attr := sys.MapFreezeAttr{
MapFd: m.fd.Uint(),
}
if err := sys.MapFreeze(&attr); err != nil {
+ if haveFeatErr := haveMapMutabilityModifiers(); haveFeatErr != nil {
+ return fmt.Errorf("can't freeze map: %w", haveFeatErr)
+ }
return fmt.Errorf("can't freeze map: %w", err)
}
return nil
@@ -1217,16 +1240,7 @@
return sys.Pointer{}, errors.New("can't use nil as key of map")
}
- return marshalPtr(data, int(m.keySize))
-}
-
-func (m *Map) unmarshalKey(data interface{}, buf []byte) error {
- if buf == nil {
- // This is from a makeBuffer call, nothing do do here.
- return nil
- }
-
- return unmarshalBytes(data, buf)
+ return marshalMapSyscallInput(data, int(m.keySize))
}
func (m *Map) marshalValue(data interface{}) (sys.Pointer, error) {
@@ -1249,7 +1263,7 @@
buf, err = marshalProgram(value, int(m.valueSize))
default:
- return marshalPtr(data, int(m.valueSize))
+ return marshalMapSyscallInput(data, int(m.valueSize))
}
if err != nil {
@@ -1259,16 +1273,7 @@
return sys.NewSlicePointer(buf), nil
}
-func (m *Map) unmarshalValue(value interface{}, buf []byte) error {
- if buf == nil {
- // This is from a makeBuffer call, nothing do do here.
- return nil
- }
-
- if m.typ.hasPerCPUValue() {
- return unmarshalPerCPUValue(value, int(m.valueSize), buf)
- }
-
+func (m *Map) unmarshalValue(value any, buf sysenc.Buffer) error {
switch value := value.(type) {
case **Map:
if !m.typ.canStoreMap() {
@@ -1315,7 +1320,7 @@
return errors.New("require pointer to *Program")
}
- return unmarshalBytes(value, buf)
+ return buf.Unmarshal(value)
}
// LoadPinnedMap loads a Map from a BPF file.
@@ -1337,12 +1342,11 @@
}
// unmarshalMap creates a map from a map ID encoded in host endianness.
-func unmarshalMap(buf []byte) (*Map, error) {
- if len(buf) != 4 {
- return nil, errors.New("map id requires 4 byte value")
+func unmarshalMap(buf sysenc.Buffer) (*Map, error) {
+ var id uint32
+ if err := buf.Unmarshal(&id); err != nil {
+ return nil, err
}
-
- id := internal.NativeEndian.Uint32(buf)
return NewMapFromID(MapID(id))
}
@@ -1414,11 +1418,7 @@
return false
}
- // The user can get access to nextKey since unmarshalBytes
- // does not copy when unmarshaling into a []byte.
- // Make a copy to prevent accidental corruption of
- // iterator state.
- copy(mi.curKey, nextKey)
+ mi.curKey = nextKey
mi.count++
mi.err = mi.target.Lookup(nextKey, valueOut)
@@ -1438,7 +1438,12 @@
return false
}
- mi.err = mi.target.unmarshalKey(keyOut, nextKey)
+ if ptr, ok := keyOut.(unsafe.Pointer); ok {
+ copy(unsafe.Slice((*byte)(ptr), len(nextKey)), nextKey)
+ } else {
+ mi.err = sysenc.Unmarshal(keyOut, nextKey)
+ }
+
return mi.err == nil
}
diff --git a/vendor/github.com/cilium/ebpf/marshalers.go b/vendor/github.com/cilium/ebpf/marshalers.go
index a568bff..e89a12f 100644
--- a/vendor/github.com/cilium/ebpf/marshalers.go
+++ b/vendor/github.com/cilium/ebpf/marshalers.go
@@ -1,166 +1,53 @@
package ebpf
import (
- "bytes"
"encoding"
- "encoding/binary"
"errors"
"fmt"
"reflect"
- "runtime"
- "sync"
"unsafe"
"github.com/cilium/ebpf/internal"
"github.com/cilium/ebpf/internal/sys"
+ "github.com/cilium/ebpf/internal/sysenc"
)
-// marshalPtr converts an arbitrary value into a pointer suitable
+// marshalMapSyscallInput converts an arbitrary value into a pointer suitable
// to be passed to the kernel.
//
// As an optimization, it returns the original value if it is an
// unsafe.Pointer.
-func marshalPtr(data interface{}, length int) (sys.Pointer, error) {
+func marshalMapSyscallInput(data any, length int) (sys.Pointer, error) {
if ptr, ok := data.(unsafe.Pointer); ok {
return sys.NewPointer(ptr), nil
}
- buf, err := marshalBytes(data, length)
+ buf, err := sysenc.Marshal(data, length)
if err != nil {
return sys.Pointer{}, err
}
- return sys.NewSlicePointer(buf), nil
+ return buf.Pointer(), nil
}
-// marshalBytes converts an arbitrary value into a byte buffer.
-//
-// Prefer using Map.marshalKey and Map.marshalValue if possible, since
-// those have special cases that allow more types to be encoded.
-//
-// Returns an error if the given value isn't representable in exactly
-// length bytes.
-func marshalBytes(data interface{}, length int) (buf []byte, err error) {
- if data == nil {
- return nil, errors.New("can't marshal a nil value")
- }
-
- switch value := data.(type) {
- case encoding.BinaryMarshaler:
- buf, err = value.MarshalBinary()
- case string:
- buf = []byte(value)
- case []byte:
- buf = value
- case unsafe.Pointer:
- err = errors.New("can't marshal from unsafe.Pointer")
- case Map, *Map, Program, *Program:
- err = fmt.Errorf("can't marshal %T", value)
- default:
- wr := internal.NewBuffer(make([]byte, 0, length))
- defer internal.PutBuffer(wr)
-
- err = binary.Write(wr, internal.NativeEndian, value)
- if err != nil {
- err = fmt.Errorf("encoding %T: %v", value, err)
- }
- buf = wr.Bytes()
- }
- if err != nil {
- return nil, err
- }
-
- if len(buf) != length {
- return nil, fmt.Errorf("%T doesn't marshal to %d bytes", data, length)
- }
- return buf, nil
-}
-
-func makeBuffer(dst interface{}, length int) (sys.Pointer, []byte) {
+func makeMapSyscallOutput(dst any, length int) sysenc.Buffer {
if ptr, ok := dst.(unsafe.Pointer); ok {
- return sys.NewPointer(ptr), nil
+ return sysenc.UnsafeBuffer(ptr)
}
- buf := make([]byte, length)
- return sys.NewSlicePointer(buf), buf
-}
-
-var bytesReaderPool = sync.Pool{
- New: func() interface{} {
- return new(bytes.Reader)
- },
-}
-
-// unmarshalBytes converts a byte buffer into an arbitrary value.
-//
-// Prefer using Map.unmarshalKey and Map.unmarshalValue if possible, since
-// those have special cases that allow more types to be encoded.
-//
-// The common int32 and int64 types are directly handled to avoid
-// unnecessary heap allocations as happening in the default case.
-func unmarshalBytes(data interface{}, buf []byte) error {
- switch value := data.(type) {
- case unsafe.Pointer:
- dst := unsafe.Slice((*byte)(value), len(buf))
- copy(dst, buf)
- runtime.KeepAlive(value)
- return nil
- case Map, *Map, Program, *Program:
- return fmt.Errorf("can't unmarshal into %T", value)
- case encoding.BinaryUnmarshaler:
- return value.UnmarshalBinary(buf)
- case *string:
- *value = string(buf)
- return nil
- case *[]byte:
- *value = buf
- return nil
- case *int32:
- if len(buf) < 4 {
- return errors.New("int32 requires 4 bytes")
- }
- *value = int32(internal.NativeEndian.Uint32(buf))
- return nil
- case *uint32:
- if len(buf) < 4 {
- return errors.New("uint32 requires 4 bytes")
- }
- *value = internal.NativeEndian.Uint32(buf)
- return nil
- case *int64:
- if len(buf) < 8 {
- return errors.New("int64 requires 8 bytes")
- }
- *value = int64(internal.NativeEndian.Uint64(buf))
- return nil
- case *uint64:
- if len(buf) < 8 {
- return errors.New("uint64 requires 8 bytes")
- }
- *value = internal.NativeEndian.Uint64(buf)
- return nil
- case string:
- return errors.New("require pointer to string")
- case []byte:
- return errors.New("require pointer to []byte")
- default:
- rd := bytesReaderPool.Get().(*bytes.Reader)
- rd.Reset(buf)
- defer bytesReaderPool.Put(rd)
- if err := binary.Read(rd, internal.NativeEndian, value); err != nil {
- return fmt.Errorf("decoding %T: %v", value, err)
- }
- return nil
+ _, ok := dst.(encoding.BinaryUnmarshaler)
+ if ok {
+ return sysenc.SyscallOutput(nil, length)
}
+
+ return sysenc.SyscallOutput(dst, length)
}
// marshalPerCPUValue encodes a slice containing one value per
// possible CPU into a buffer of bytes.
//
// Values are initialized to zero if the slice has less elements than CPUs.
-//
-// slice must have a type like []elementType.
-func marshalPerCPUValue(slice interface{}, elemLength int) (sys.Pointer, error) {
+func marshalPerCPUValue(slice any, elemLength int) (sys.Pointer, error) {
sliceType := reflect.TypeOf(slice)
if sliceType.Kind() != reflect.Slice {
return sys.Pointer{}, errors.New("per-CPU value requires slice")
@@ -182,13 +69,13 @@
for i := 0; i < sliceLen; i++ {
elem := sliceValue.Index(i).Interface()
- elemBytes, err := marshalBytes(elem, elemLength)
+ elemBytes, err := sysenc.Marshal(elem, elemLength)
if err != nil {
return sys.Pointer{}, err
}
offset := i * alignedElemLength
- copy(buf[offset:offset+elemLength], elemBytes)
+ elemBytes.CopyTo(buf[offset : offset+elemLength])
}
return sys.NewSlicePointer(buf), nil
@@ -197,8 +84,8 @@
// unmarshalPerCPUValue decodes a buffer into a slice containing one value per
// possible CPU.
//
-// valueOut must have a type like *[]elementType
-func unmarshalPerCPUValue(slicePtr interface{}, elemLength int, buf []byte) error {
+// slicePtr must be a pointer to a slice.
+func unmarshalPerCPUValue(slicePtr any, elemLength int, buf []byte) error {
slicePtrType := reflect.TypeOf(slicePtr)
if slicePtrType.Kind() != reflect.Ptr || slicePtrType.Elem().Kind() != reflect.Slice {
return fmt.Errorf("per-cpu value requires pointer to slice")
@@ -218,12 +105,9 @@
sliceElemType = sliceElemType.Elem()
}
- step := len(buf) / possibleCPUs
- if step < elemLength {
- return fmt.Errorf("per-cpu element length is larger than available data")
- }
+ stride := internal.Align(elemLength, 8)
for i := 0; i < possibleCPUs; i++ {
- var elem interface{}
+ var elem any
if sliceElemIsPointer {
newElem := reflect.New(sliceElemType)
slice.Index(i).Set(newElem)
@@ -232,16 +116,12 @@
elem = slice.Index(i).Addr().Interface()
}
- // Make a copy, since unmarshal can hold on to itemBytes
- elemBytes := make([]byte, elemLength)
- copy(elemBytes, buf[:elemLength])
-
- err := unmarshalBytes(elem, elemBytes)
+ err := sysenc.Unmarshal(elem, buf[:elemLength])
if err != nil {
return fmt.Errorf("cpu %d: %w", i, err)
}
- buf = buf[step:]
+ buf = buf[stride:]
}
reflect.ValueOf(slicePtr).Elem().Set(slice)
diff --git a/vendor/github.com/cilium/ebpf/netlify.toml b/vendor/github.com/cilium/ebpf/netlify.toml
new file mode 100644
index 0000000..67c83f3
--- /dev/null
+++ b/vendor/github.com/cilium/ebpf/netlify.toml
@@ -0,0 +1,4 @@
+[build]
+ base = "docs/"
+ publish = "site/"
+ command = "mkdocs build"
diff --git a/vendor/github.com/cilium/ebpf/prog.go b/vendor/github.com/cilium/ebpf/prog.go
index 70aaef5..6d46a04 100644
--- a/vendor/github.com/cilium/ebpf/prog.go
+++ b/vendor/github.com/cilium/ebpf/prog.go
@@ -16,6 +16,7 @@
"github.com/cilium/ebpf/btf"
"github.com/cilium/ebpf/internal"
"github.com/cilium/ebpf/internal/sys"
+ "github.com/cilium/ebpf/internal/sysenc"
"github.com/cilium/ebpf/internal/unix"
)
@@ -277,7 +278,7 @@
if err != nil {
return nil, fmt.Errorf("fixing up kfuncs: %w", err)
}
- defer handles.close()
+ defer handles.Close()
if len(handles) > 0 {
fdArray := handles.fdArray()
@@ -763,14 +764,14 @@
return attr.Retval, total, nil
}
-func unmarshalProgram(buf []byte) (*Program, error) {
- if len(buf) != 4 {
- return nil, errors.New("program id requires 4 byte value")
+func unmarshalProgram(buf sysenc.Buffer) (*Program, error) {
+ var id uint32
+ if err := buf.Unmarshal(&id); err != nil {
+ return nil, err
}
// Looking up an entry in a nested map or prog array returns an id,
// not an fd.
- id := internal.NativeEndian.Uint32(buf)
return NewProgramFromID(ProgramID(id))
}
@@ -921,7 +922,12 @@
}
id, err := spec.TypeID(target)
- return module, id, err
+ if err != nil {
+ module.Close()
+ return nil, 0, err
+ }
+
+ return module, id, nil
}
// findTargetInKernel attempts to find a named type in the current kernel.
@@ -999,7 +1005,9 @@
var typeName string
switch (match{progType, attachType}) {
- case match{Extension, AttachNone}:
+ case match{Extension, AttachNone},
+ match{Tracing, AttachTraceFEntry},
+ match{Tracing, AttachTraceFExit}:
typeName = name
default:
return 0, errUnrecognizedAttachType
diff --git a/vendor/github.com/cilium/ebpf/run-tests.sh b/vendor/github.com/cilium/ebpf/run-tests.sh
index 1d1490a..629a069 100644
--- a/vendor/github.com/cilium/ebpf/run-tests.sh
+++ b/vendor/github.com/cilium/ebpf/run-tests.sh
@@ -14,6 +14,21 @@
script="$(realpath "$0")"
readonly script
+quote_env() {
+ for var in "$@"; do
+ if [ -v "$var" ]; then
+ printf "%s=%q " "$var" "${!var}"
+ fi
+ done
+}
+
+declare -a preserved_env=(
+ PATH
+ CI_MAX_KERNEL_VERSION
+ TEST_SEED
+ KERNEL_VERSION
+)
+
# This script is a bit like a Matryoshka doll since it keeps re-executing itself
# in various different contexts:
#
@@ -51,11 +66,11 @@
fi
for ((i = 0; i < 3; i++)); do
- if ! $sudo virtme-run --kimg "${input}/bzImage" --memory 768M --pwd \
+ if ! $sudo virtme-run --kimg "${input}/boot/vmlinuz" --memory 768M --pwd \
--rwdir="${testdir}=${testdir}" \
--rodir=/run/input="${input}" \
--rwdir=/run/output="${output}" \
- --script-sh "PATH=\"$PATH\" CI_MAX_KERNEL_VERSION="${CI_MAX_KERNEL_VERSION:-}" \"$script\" --exec-test $cmd" \
+ --script-sh "$(quote_env "${preserved_env[@]}") \"$script\" --exec-test $cmd" \
--kopt possible_cpus=2; then # need at least two CPUs for some tests
exit 23
fi
@@ -85,8 +100,8 @@
export KERNEL_SELFTESTS="/run/input/bpf"
fi
- if [[ -f "/run/input/bpf/bpf_testmod/bpf_testmod.ko" ]]; then
- insmod "/run/input/bpf/bpf_testmod/bpf_testmod.ko"
+ if [[ -d "/run/input/lib/modules" ]]; then
+ find /run/input/lib/modules -type f -name bpf_testmod.ko -exec insmod {} \;
fi
dmesg --clear
@@ -114,6 +129,9 @@
return $ret
}
+machine="$(uname -m)"
+readonly machine
+
if [[ -f "${1}" ]]; then
readonly kernel="${1}"
cp "${1}" "${input}/bzImage"
@@ -121,16 +139,24 @@
# LINUX_VERSION_CODE test compares this to discovered value.
export KERNEL_VERSION="${1}"
- readonly kernel="linux-${1}.bz"
- readonly selftests="linux-${1}-selftests-bpf.tgz"
+ if [ "${machine}" = "x86_64" ]; then
+ readonly kernel="linux-${1}-amd64.tgz"
+ readonly selftests="linux-${1}-amd64-selftests-bpf.tgz"
+ elif [ "${machine}" = "aarch64" ]; then
+ readonly kernel="linux-${1}-arm64.tgz"
+ readonly selftests=""
+ else
+ echo "Arch ${machine} is not supported"
+ exit 1
+ fi
fetch "${kernel}"
- cp "${tmp_dir}/${kernel}" "${input}/bzImage"
+ tar xf "${tmp_dir}/${kernel}" -C "${input}"
- if fetch "${selftests}"; then
+ if [ -n "${selftests}" ] && fetch "${selftests}"; then
echo "Decompressing selftests"
mkdir "${input}/bpf"
- tar --strip-components=4 -xf "${tmp_dir}/${selftests}" -C "${input}/bpf"
+ tar --strip-components=5 -xf "${tmp_dir}/${selftests}" -C "${input}/bpf"
else
echo "No selftests found, disabling"
fi
diff --git a/vendor/github.com/cilium/ebpf/syscalls.go b/vendor/github.com/cilium/ebpf/syscalls.go
index fd21dea..cdf1fcf 100644
--- a/vendor/github.com/cilium/ebpf/syscalls.go
+++ b/vendor/github.com/cilium/ebpf/syscalls.go
@@ -119,6 +119,7 @@
MaxEntries: 1,
MapFlags: unix.BPF_F_INNER_MAP,
})
+
if err != nil {
return internal.ErrNotSupported
}
@@ -135,6 +136,7 @@
MaxEntries: 1,
MapFlags: unix.BPF_F_NO_PREALLOC,
})
+
if err != nil {
return internal.ErrNotSupported
}
@@ -223,8 +225,8 @@
keys := []uint32{1, 2}
values := []uint32{3, 4}
- kp, _ := marshalPtr(keys, 8)
- vp, _ := marshalPtr(values, 8)
+ kp, _ := marshalMapSyscallInput(keys, 8)
+ vp, _ := marshalMapSyscallInput(values, 8)
err = sys.MapUpdateBatch(&sys.MapUpdateBatchAttr{
MapFd: fd.Uint(),
@@ -265,11 +267,8 @@
}
fd, err := progLoad(insns, SocketFilter, "MIT")
- if errors.Is(err, unix.EINVAL) {
- return internal.ErrNotSupported
- }
if err != nil {
- return err
+ return internal.ErrNotSupported
}
_ = fd.Close()
return nil
diff --git a/vendor/github.com/cilium/ebpf/types.go b/vendor/github.com/cilium/ebpf/types.go
index 35927e2..e921551 100644
--- a/vendor/github.com/cilium/ebpf/types.go
+++ b/vendor/github.com/cilium/ebpf/types.go
@@ -5,7 +5,7 @@
"github.com/cilium/ebpf/internal/unix"
)
-//go:generate stringer -output types_string.go -type=MapType,ProgramType,PinType
+//go:generate go run golang.org/x/tools/cmd/stringer@latest -output types_string.go -type=MapType,ProgramType,PinType
// MapType indicates the type map structure
// that will be initialized in the kernel.
@@ -44,7 +44,7 @@
// if an skb is from a socket belonging to a specific cgroup
CGroupArray
// LRUHash - This allows you to create a small hash structure that will purge the
- // least recently used items rather than thow an error when you run out of memory
+ // least recently used items rather than throw an error when you run out of memory
LRUHash
// LRUCPUHash - This is NOT like PerCPUHash, this structure is shared among the CPUs,
// it has more to do with including the CPU id with the LRU calculation so that if a
@@ -102,6 +102,12 @@
return mt == PerCPUHash || mt == PerCPUArray || mt == LRUCPUHash || mt == PerCPUCGroupStorage
}
+// canStoreMapOrProgram returns true if the Map stores references to another Map
+// or Program.
+func (mt MapType) canStoreMapOrProgram() bool {
+ return mt.canStoreMap() || mt.canStoreProgram()
+}
+
// canStoreMap returns true if the map type accepts a map fd
// for update and returns a map id for lookup.
func (mt MapType) canStoreMap() bool {
@@ -158,7 +164,7 @@
// Will cause invalid argument (EINVAL) at program load time if set incorrectly.
type AttachType uint32
-//go:generate stringer -type AttachType -trimprefix Attach
+//go:generate go run golang.org/x/tools/cmd/stringer@latest -type AttachType -trimprefix Attach
// AttachNone is an alias for AttachCGroupInetIngress for readability reasons.
const AttachNone AttachType = 0
@@ -213,7 +219,7 @@
type AttachFlags uint32
// PinType determines whether a map is pinned into a BPFFS.
-type PinType int
+type PinType uint32
// Valid pin types.
//
diff --git a/vendor/github.com/cilium/ebpf/types_string.go b/vendor/github.com/cilium/ebpf/types_string.go
index 5679f22..e20c37a 100644
--- a/vendor/github.com/cilium/ebpf/types_string.go
+++ b/vendor/github.com/cilium/ebpf/types_string.go
@@ -111,7 +111,7 @@
var _PinType_index = [...]uint8{0, 7, 16}
func (i PinType) String() string {
- if i < 0 || i >= PinType(len(_PinType_index)-1) {
+ if i >= PinType(len(_PinType_index)-1) {
return "PinType(" + strconv.FormatInt(int64(i), 10) + ")"
}
return _PinType_name[_PinType_index[i]:_PinType_index[i+1]]
diff --git a/vendor/modules.txt b/vendor/modules.txt
index 79d2a5a..d7cf00b 100644
--- a/vendor/modules.txt
+++ b/vendor/modules.txt
@@ -216,14 +216,15 @@
# github.com/cespare/xxhash/v2 v2.2.0
## explicit; go 1.11
github.com/cespare/xxhash/v2
-# github.com/cilium/ebpf v0.11.0
-## explicit; go 1.19
+# github.com/cilium/ebpf v0.12.3
+## explicit; go 1.20
github.com/cilium/ebpf
github.com/cilium/ebpf/asm
github.com/cilium/ebpf/btf
github.com/cilium/ebpf/internal
github.com/cilium/ebpf/internal/kconfig
github.com/cilium/ebpf/internal/sys
+github.com/cilium/ebpf/internal/sysenc
github.com/cilium/ebpf/internal/tracefs
github.com/cilium/ebpf/internal/unix
github.com/cilium/ebpf/link