mirror of
https://github.com/varun-r-mallya/Python-BPF.git
synced 2025-12-31 21:06:25 +00:00
Compare commits
119 Commits
798f07986a
...
32int_supp
| Author | SHA1 | Date | |
|---|---|---|---|
| 5ad33b011e | |||
| 2f4785b796 | |||
| c5fdd3bce2 | |||
| b0d35693b9 | |||
| 44c6ceda27 | |||
| 2685d0a0ee | |||
| 3489f45b63 | |||
| 204ec26154 | |||
| a9d82d40d3 | |||
| 85a62d6cd8 | |||
| c3fc790c71 | |||
| 22e30f04b4 | |||
| 620b8cb1e7 | |||
| 1207fe9f92 | |||
| b138405931 | |||
| 262f00f635 | |||
| 07580dabf2 | |||
| ac74b03b14 | |||
| 3bf85e733e | |||
| 73f7c80eca | |||
| 238697469a | |||
| 8bd210cede | |||
| 7bf6f9c48c | |||
| a1fe2ed4bc | |||
| 93285dbdd8 | |||
| 1ea44dd8e1 | |||
| 96216d4411 | |||
| 028d9c2c08 | |||
| c6b5ecb47e | |||
| 30bcfcbbd0 | |||
| f18a4399ea | |||
| 4e01df735f | |||
| 64674cf646 | |||
| 5c1e7103a6 | |||
| 576fa2f106 | |||
| 76a873cb0d | |||
| e86c6082c9 | |||
| cb1ad15f43 | |||
| b24b3ed250 | |||
| beaad996db | |||
| 99b92e44e3 | |||
| ce7adaadb6 | |||
| 5ac316a1ac | |||
| 36a1a0903e | |||
| f2bc7f1434 | |||
| b3921c424d | |||
| 7a99d21b24 | |||
| cf05e4959d | |||
| a7394ccafa | |||
| 63f378c34b | |||
| 37af7d2e20 | |||
| 77c0d131be | |||
| 84fdf52658 | |||
| f4d903d4b5 | |||
| f9494c870b | |||
| 0d4ebf72b6 | |||
| adf32560a0 | |||
| c65900b733 | |||
| 711e34cae1 | |||
| cf3f4a0002 | |||
| d50157fa09 | |||
| ba860b5039 | |||
| 21cea97d78 | |||
| d8729342dc | |||
| 4179fbfc88 | |||
| ba397036b4 | |||
| caecb8c9b0 | |||
| 1a0e21eaa8 | |||
| 190baf2674 | |||
| c3f3d1e564 | |||
| e7734629a5 | |||
| 5955db88cf | |||
| 66caa3cf1d | |||
| e499c29d42 | |||
| 76d0dbfbf4 | |||
| 56a2fbaf5b | |||
| 3b323132f0 | |||
| c9363e62a9 | |||
| a20643f3a7 | |||
| eee212795f | |||
| 8da50b7068 | |||
| e636fcaea7 | |||
| 5512bf52e4 | |||
| 079ceaa0d6 | |||
| 328b792e4e | |||
| 5dafa5bd0d | |||
| 33aa794718 | |||
| d855e9ef2e | |||
| de19c8fc90 | |||
| dc1b243e82 | |||
| 1b4272b408 | |||
| 101183c315 | |||
| 3a3116253f | |||
| 9b7aa6d8be | |||
| 51a1be0b0b | |||
| 7ae629e8f7 | |||
| dd734ea2aa | |||
| 71d005b6b1 | |||
| 5d9a29ee8e | |||
| 041e538b53 | |||
| 5413cc793b | |||
| f21837aefe | |||
| 0f5c1fa752 | |||
| de02731ea1 | |||
| c22d85ceb8 | |||
| 2b3c81affa | |||
| 8372111616 | |||
| eb4ee64ee5 | |||
| ce7b170fea | |||
| 9a60dd87e3 | |||
| c499fe7421 | |||
| 8239097fbb | |||
| a4cfc2b7aa | |||
| 69b73003ca | |||
| 11e8e72188 | |||
| d3f0e3b2ef | |||
| 09ba749b46 | |||
| a03d3e5d4c | |||
| e1f9ac6ba0 |
2
.gitattributes
vendored
2
.gitattributes
vendored
@ -1 +1,3 @@
|
||||
tests/c-form/vmlinux.h linguist-vendored
|
||||
examples/ linguist-vendored
|
||||
BCC-Examples/ linguist-vendored
|
||||
|
||||
4
.github/workflows/python-publish.yml
vendored
4
.github/workflows/python-publish.yml
vendored
@ -33,7 +33,7 @@ jobs:
|
||||
python -m build
|
||||
|
||||
- name: Upload distributions
|
||||
uses: actions/upload-artifact@v4
|
||||
uses: actions/upload-artifact@v5
|
||||
with:
|
||||
name: release-dists
|
||||
path: dist/
|
||||
@ -59,7 +59,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: Retrieve release distributions
|
||||
uses: actions/download-artifact@v5
|
||||
uses: actions/download-artifact@v6
|
||||
with:
|
||||
name: release-dists
|
||||
path: dist/
|
||||
|
||||
20
BCC-Examples/README.md
Normal file
20
BCC-Examples/README.md
Normal file
@ -0,0 +1,20 @@
|
||||
## BCC examples ported to PythonBPF
|
||||
|
||||
This folder contains examples of BCC tutorial examples that have been ported to use **PythonBPF**.
|
||||
|
||||
## Requirements
|
||||
- install `pythonbpf` and `pylibbpf` using pip.
|
||||
- You will also need `matplotlib` for vfsreadlat.py example.
|
||||
- You will also need `rich` for vfsreadlat_rich.py example.
|
||||
- You will also need `plotly` and `dash` for vfsreadlat_plotly.py example.
|
||||
|
||||
## Usage
|
||||
- You'll need root privileges to run these examples. If you are using a virtualenv, use the following command to run the scripts:
|
||||
```bash
|
||||
sudo <path_to_virtualenv>/bin/python3 <script_name>.py
|
||||
```
|
||||
- For vfsreadlat_plotly.py, run the following command to start the Dash server:
|
||||
```bash
|
||||
sudo <path_to_virtualenv>/bin/python3 vfsreadlat_plotly/bpf_program.py
|
||||
```
|
||||
Then open your web browser and navigate to the given URL.
|
||||
83
BCC-Examples/hello_fields.ipynb
Normal file
83
BCC-Examples/hello_fields.ipynb
Normal file
@ -0,0 +1,83 @@
|
||||
{
|
||||
"cells": [
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "28cf2e27-41e2-461c-a39c-147417141a4e",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from pythonbpf import bpf, section, bpfglobal, BPF, trace_fields\n",
|
||||
"from ctypes import c_void_p, c_int64"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "133190e5-5a99-4585-b6e1-91224ed973c2",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"@bpf\n",
|
||||
"@section(\"tracepoint/syscalls/sys_enter_clone\")\n",
|
||||
"def hello_world(ctx: c_void_p) -> c_int64:\n",
|
||||
" print(\"Hello, World!\")\n",
|
||||
" return 0\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@bpfglobal\n",
|
||||
"def LICENSE() -> str:\n",
|
||||
" return \"GPL\"\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"# Compile and load\n",
|
||||
"b = BPF()\n",
|
||||
"b.load()\n",
|
||||
"b.attach_all()"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "d3934efb-4043-4545-ae4c-c50ec40a24fd",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"# header\n",
|
||||
"print(f\"{'TIME(s)':<18} {'COMM':<16} {'PID':<6} {'MESSAGE'}\")\n",
|
||||
"\n",
|
||||
"# format output\n",
|
||||
"while True:\n",
|
||||
" try:\n",
|
||||
" (task, pid, cpu, flags, ts, msg) = trace_fields()\n",
|
||||
" except ValueError:\n",
|
||||
" continue\n",
|
||||
" except KeyboardInterrupt:\n",
|
||||
" exit()\n",
|
||||
" print(f\"{ts:<18} {task:<16} {pid:<6} {msg}\")"
|
||||
]
|
||||
}
|
||||
],
|
||||
"metadata": {
|
||||
"kernelspec": {
|
||||
"display_name": "Python 3 (ipykernel)",
|
||||
"language": "python",
|
||||
"name": "python3"
|
||||
},
|
||||
"language_info": {
|
||||
"codemirror_mode": {
|
||||
"name": "ipython",
|
||||
"version": 3
|
||||
},
|
||||
"file_extension": ".py",
|
||||
"mimetype": "text/x-python",
|
||||
"name": "python",
|
||||
"nbconvert_exporter": "python",
|
||||
"pygments_lexer": "ipython3",
|
||||
"version": "3.13.3"
|
||||
}
|
||||
},
|
||||
"nbformat": 4,
|
||||
"nbformat_minor": 5
|
||||
}
|
||||
110
BCC-Examples/hello_perf_output.ipynb
Normal file
110
BCC-Examples/hello_perf_output.ipynb
Normal file
@ -0,0 +1,110 @@
|
||||
{
|
||||
"cells": [
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "79b74928-f4b4-4320-96e3-d973997de2f4",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from pythonbpf import bpf, map, struct, section, bpfglobal, BPF\n",
|
||||
"from pythonbpf.helper import ktime, pid, comm\n",
|
||||
"from pythonbpf.maps import PerfEventArray\n",
|
||||
"from ctypes import c_void_p, c_int64"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "5bdb0329-ae2d-45e8-808e-5ed5b1374204",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"@bpf\n",
|
||||
"@struct\n",
|
||||
"class data_t:\n",
|
||||
" pid: c_int64\n",
|
||||
" ts: c_int64\n",
|
||||
" comm: str(16)\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@map\n",
|
||||
"def events() -> PerfEventArray:\n",
|
||||
" return PerfEventArray(key_size=c_int64, value_size=c_int64)\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@section(\"tracepoint/syscalls/sys_enter_clone\")\n",
|
||||
"def hello(ctx: c_void_p) -> c_int64:\n",
|
||||
" dataobj = data_t()\n",
|
||||
" dataobj.pid, dataobj.ts = pid(), ktime()\n",
|
||||
" comm(dataobj.comm)\n",
|
||||
" events.output(dataobj)\n",
|
||||
" return 0\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@bpfglobal\n",
|
||||
"def LICENSE() -> str:\n",
|
||||
" return \"GPL\"\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"# Compile and load\n",
|
||||
"b = BPF()\n",
|
||||
"b.load()\n",
|
||||
"b.attach_all()"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "4bcc7d57-6cc4-48a3-bbd2-42ad6263afdf",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"start = 0\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"def callback(cpu, event):\n",
|
||||
" global start\n",
|
||||
" if start == 0:\n",
|
||||
" start = event.ts\n",
|
||||
" ts = (event.ts - start) / 1e9\n",
|
||||
" print(f\"[CPU {cpu}] PID: {event.pid}, TS: {ts}, COMM: {event.comm.decode()}\")\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"perf = b[\"events\"].open_perf_buffer(callback, struct_name=\"data_t\")\n",
|
||||
"print(\"Starting to poll... (Ctrl+C to stop)\")\n",
|
||||
"print(\"Try running: fork() or clone() system calls to trigger events\")\n",
|
||||
"\n",
|
||||
"try:\n",
|
||||
" while True:\n",
|
||||
" b[\"events\"].poll(1000)\n",
|
||||
"except KeyboardInterrupt:\n",
|
||||
" print(\"Stopping...\")"
|
||||
]
|
||||
}
|
||||
],
|
||||
"metadata": {
|
||||
"kernelspec": {
|
||||
"display_name": "Python 3 (ipykernel)",
|
||||
"language": "python",
|
||||
"name": "python3"
|
||||
},
|
||||
"language_info": {
|
||||
"codemirror_mode": {
|
||||
"name": "ipython",
|
||||
"version": 3
|
||||
},
|
||||
"file_extension": ".py",
|
||||
"mimetype": "text/x-python",
|
||||
"name": "python",
|
||||
"nbconvert_exporter": "python",
|
||||
"pygments_lexer": "ipython3",
|
||||
"version": "3.13.3"
|
||||
}
|
||||
},
|
||||
"nbformat": 4,
|
||||
"nbformat_minor": 5
|
||||
}
|
||||
116
BCC-Examples/hello_world.ipynb
Normal file
116
BCC-Examples/hello_world.ipynb
Normal file
@ -0,0 +1,116 @@
|
||||
{
|
||||
"cells": [
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": 9,
|
||||
"id": "7d5d3cfb-39ba-4516-9856-b3bed47a0cef",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from pythonbpf import bpf, section, bpfglobal, BPF, trace_pipe\n",
|
||||
"from ctypes import c_void_p, c_int64"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": 10,
|
||||
"id": "cf1c87aa-e173-4156-8f2d-762225bc6d19",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"@bpf\n",
|
||||
"@section(\"tracepoint/syscalls/sys_enter_clone\")\n",
|
||||
"def hello_world(ctx: c_void_p) -> c_int64:\n",
|
||||
" print(\"Hello, World!\")\n",
|
||||
" return 0\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@bpfglobal\n",
|
||||
"def LICENSE() -> str:\n",
|
||||
" return \"GPL\"\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"b = BPF()"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "bd81383d-f75a-4269-8451-3d985d85b124",
|
||||
"metadata": {
|
||||
"scrolled": true
|
||||
},
|
||||
"outputs": [
|
||||
{
|
||||
"name": "stdout",
|
||||
"output_type": "stream",
|
||||
"text": [
|
||||
" Cache2 I/O-4716 [003] ...21 8218.000492: bpf_trace_printk: count: 11 with 4716\n",
|
||||
"\n",
|
||||
" Cache2 I/O-4716 [003] ...21 8218.000499: bpf_trace_printk: Hello, World!\n",
|
||||
"\n",
|
||||
" WebExtensions-5168 [002] ...21 8219.320392: bpf_trace_printk: count: 13 with 5168\n",
|
||||
"\n",
|
||||
" WebExtensions-5168 [002] ...21 8219.320399: bpf_trace_printk: Hello, World!\n",
|
||||
"\n",
|
||||
" python-21155 [001] ...21 8220.933716: bpf_trace_printk: count: 5 with 21155\n",
|
||||
"\n",
|
||||
" python-21155 [001] ...21 8220.933721: bpf_trace_printk: Hello, World!\n",
|
||||
"\n",
|
||||
" python-21155 [002] ...21 8221.341290: bpf_trace_printk: count: 6 with 21155\n",
|
||||
"\n",
|
||||
" python-21155 [002] ...21 8221.341295: bpf_trace_printk: Hello, World!\n",
|
||||
"\n",
|
||||
" Isolated Web Co-5462 [000] ...21 8223.095033: bpf_trace_printk: count: 7 with 5462\n",
|
||||
"\n",
|
||||
" Isolated Web Co-5462 [000] ...21 8223.095043: bpf_trace_printk: Hello, World!\n",
|
||||
"\n",
|
||||
" firefox-4542 [000] ...21 8227.760067: bpf_trace_printk: count: 8 with 4542\n",
|
||||
"\n",
|
||||
" firefox-4542 [000] ...21 8227.760080: bpf_trace_printk: Hello, World!\n",
|
||||
"\n",
|
||||
" Isolated Web Co-12404 [003] ...21 8227.917086: bpf_trace_printk: count: 7 with 12404\n",
|
||||
"\n",
|
||||
" Isolated Web Co-12404 [003] ...21 8227.917095: bpf_trace_printk: Hello, World!\n",
|
||||
"\n"
|
||||
]
|
||||
}
|
||||
],
|
||||
"source": [
|
||||
"b.load()\n",
|
||||
"b.attach_all()\n",
|
||||
"trace_pipe()"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "01e1f25b-decc-425b-a1aa-a5e701082574",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": []
|
||||
}
|
||||
],
|
||||
"metadata": {
|
||||
"kernelspec": {
|
||||
"display_name": "Python 3 (ipykernel)",
|
||||
"language": "python",
|
||||
"name": "python3"
|
||||
},
|
||||
"language_info": {
|
||||
"codemirror_mode": {
|
||||
"name": "ipython",
|
||||
"version": 3
|
||||
},
|
||||
"file_extension": ".py",
|
||||
"mimetype": "text/x-python",
|
||||
"name": "python",
|
||||
"nbconvert_exporter": "python",
|
||||
"pygments_lexer": "ipython3",
|
||||
"version": "3.13.3"
|
||||
}
|
||||
},
|
||||
"nbformat": 4,
|
||||
"nbformat_minor": 5
|
||||
}
|
||||
107
BCC-Examples/sync_count.ipynb
Normal file
107
BCC-Examples/sync_count.ipynb
Normal file
@ -0,0 +1,107 @@
|
||||
{
|
||||
"cells": [
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "dcab010c-f5e9-446f-9f9f-056cc794ad14",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from pythonbpf import bpf, map, section, bpfglobal, BPF, trace_fields\n",
|
||||
"from pythonbpf.helper import ktime\n",
|
||||
"from pythonbpf.maps import HashMap\n",
|
||||
"\n",
|
||||
"from ctypes import c_void_p, c_int64"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "720797e8-9c81-4af6-a385-80f1ec4c0f15",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"@bpf\n",
|
||||
"@map\n",
|
||||
"def last() -> HashMap:\n",
|
||||
" return HashMap(key=c_int64, value=c_int64, max_entries=2)\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@section(\"tracepoint/syscalls/sys_enter_sync\")\n",
|
||||
"def do_trace(ctx: c_void_p) -> c_int64:\n",
|
||||
" ts_key, cnt_key = 0, 1\n",
|
||||
" tsp, cntp = last.lookup(ts_key), last.lookup(cnt_key)\n",
|
||||
" if not cntp:\n",
|
||||
" last.update(cnt_key, 0)\n",
|
||||
" cntp = last.lookup(cnt_key)\n",
|
||||
" if tsp:\n",
|
||||
" delta = ktime() - tsp\n",
|
||||
" if delta < 1000000000:\n",
|
||||
" time_ms = delta // 1000000\n",
|
||||
" print(f\"{time_ms} {cntp}\")\n",
|
||||
" last.delete(ts_key)\n",
|
||||
" else:\n",
|
||||
" last.update(ts_key, ktime())\n",
|
||||
" last.update(cnt_key, cntp + 1)\n",
|
||||
" return 0\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@bpfglobal\n",
|
||||
"def LICENSE() -> str:\n",
|
||||
" return \"GPL\"\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"# Compile and load\n",
|
||||
"b = BPF()\n",
|
||||
"b.load()\n",
|
||||
"b.attach_all()"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "78a8b82c-7c5f-43c1-9de1-cd982a0f345b",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"print(\"Tracing for quick sync's... Ctrl-C to end\")\n",
|
||||
"\n",
|
||||
"# format output\n",
|
||||
"start = 0\n",
|
||||
"while True:\n",
|
||||
" try:\n",
|
||||
" task, pid, cpu, flags, ts, msg = trace_fields()\n",
|
||||
" if start == 0:\n",
|
||||
" start = ts\n",
|
||||
" ts -= start\n",
|
||||
" ms, cnt = msg.split()\n",
|
||||
" print(f\"At time {ts} s: Multiple syncs detected, last {ms} ms ago. Count {cnt}\")\n",
|
||||
" except KeyboardInterrupt:\n",
|
||||
" exit()"
|
||||
]
|
||||
}
|
||||
],
|
||||
"metadata": {
|
||||
"kernelspec": {
|
||||
"display_name": "Python 3 (ipykernel)",
|
||||
"language": "python",
|
||||
"name": "python3"
|
||||
},
|
||||
"language_info": {
|
||||
"codemirror_mode": {
|
||||
"name": "ipython",
|
||||
"version": 3
|
||||
},
|
||||
"file_extension": ".py",
|
||||
"mimetype": "text/x-python",
|
||||
"name": "python",
|
||||
"nbconvert_exporter": "python",
|
||||
"pygments_lexer": "ipython3",
|
||||
"version": "3.13.3"
|
||||
}
|
||||
},
|
||||
"nbformat": 4,
|
||||
"nbformat_minor": 5
|
||||
}
|
||||
134
BCC-Examples/sync_perf_output.ipynb
Normal file
134
BCC-Examples/sync_perf_output.ipynb
Normal file
@ -0,0 +1,134 @@
|
||||
{
|
||||
"cells": [
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "b0d1ab05-0c1f-4578-9c1b-568202b95a5c",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from pythonbpf import bpf, map, struct, section, bpfglobal, BPF\n",
|
||||
"from pythonbpf.helper import ktime\n",
|
||||
"from pythonbpf.maps import HashMap, PerfEventArray\n",
|
||||
"from ctypes import c_void_p, c_int64"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "85e50d0a-f9d8-468f-8e03-f5f7128f05d8",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"@bpf\n",
|
||||
"@struct\n",
|
||||
"class data_t:\n",
|
||||
" ts: c_int64\n",
|
||||
" ms: c_int64\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@map\n",
|
||||
"def events() -> PerfEventArray:\n",
|
||||
" return PerfEventArray(key_size=c_int64, value_size=c_int64)\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@map\n",
|
||||
"def last() -> HashMap:\n",
|
||||
" return HashMap(key=c_int64, value=c_int64, max_entries=1)\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@section(\"tracepoint/syscalls/sys_enter_sync\")\n",
|
||||
"def do_trace(ctx: c_void_p) -> c_int64:\n",
|
||||
" dat, dat.ts, key = data_t(), ktime(), 0\n",
|
||||
" tsp = last.lookup(key)\n",
|
||||
" if tsp:\n",
|
||||
" delta = ktime() - tsp\n",
|
||||
" if delta < 1000000000:\n",
|
||||
" dat.ms = delta // 1000000\n",
|
||||
" events.output(dat)\n",
|
||||
" last.delete(key)\n",
|
||||
" else:\n",
|
||||
" last.update(key, ktime())\n",
|
||||
" return 0\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@bpfglobal\n",
|
||||
"def LICENSE() -> str:\n",
|
||||
" return \"GPL\"\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"# Compile and load\n",
|
||||
"b = BPF()\n",
|
||||
"b.load()\n",
|
||||
"b.attach_all()"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "40bb1107-369f-4be7-9f10-37201900c16b",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"print(\"Tracing for quick sync's... Ctrl-C to end\")\n",
|
||||
"\n",
|
||||
"# format output\n",
|
||||
"start = 0\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"def callback(cpu, event):\n",
|
||||
" global start\n",
|
||||
" if start == 0:\n",
|
||||
" start = event.ts\n",
|
||||
" event.ts -= start\n",
|
||||
" print(\n",
|
||||
" f\"At time {event.ts / 1e9} s: Multiple sync detected, Last sync: {event.ms} ms ago\"\n",
|
||||
" )\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"perf = b[\"events\"].open_perf_buffer(callback, struct_name=\"data_t\")\n",
|
||||
"print(\"Starting to poll... (Ctrl+C to stop)\")\n",
|
||||
"print(\"Try running: fork() or clone() system calls to trigger events\")\n",
|
||||
"\n",
|
||||
"try:\n",
|
||||
" while True:\n",
|
||||
" b[\"events\"].poll(1000)\n",
|
||||
"except KeyboardInterrupt:\n",
|
||||
" print(\"Stopping...\")"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "94a588d9-3a40-437c-a35b-fc40410f3eb7",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": []
|
||||
}
|
||||
],
|
||||
"metadata": {
|
||||
"kernelspec": {
|
||||
"display_name": "Python 3 (ipykernel)",
|
||||
"language": "python",
|
||||
"name": "python3"
|
||||
},
|
||||
"language_info": {
|
||||
"codemirror_mode": {
|
||||
"name": "ipython",
|
||||
"version": 3
|
||||
},
|
||||
"file_extension": ".py",
|
||||
"mimetype": "text/x-python",
|
||||
"name": "python",
|
||||
"nbconvert_exporter": "python",
|
||||
"pygments_lexer": "ipython3",
|
||||
"version": "3.13.3"
|
||||
}
|
||||
},
|
||||
"nbformat": 4,
|
||||
"nbformat_minor": 5
|
||||
}
|
||||
@ -1,7 +1,6 @@
|
||||
from pythonbpf import bpf, map, struct, section, bpfglobal, BPF
|
||||
from pythonbpf.helper import ktime
|
||||
from pythonbpf.maps import HashMap
|
||||
from pythonbpf.maps import PerfEventArray
|
||||
from pythonbpf.maps import HashMap, PerfEventArray
|
||||
from ctypes import c_void_p, c_int64
|
||||
|
||||
|
||||
|
||||
102
BCC-Examples/sync_timing.ipynb
Normal file
102
BCC-Examples/sync_timing.ipynb
Normal file
@ -0,0 +1,102 @@
|
||||
{
|
||||
"cells": [
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "bfe01ceb-2f27-41b3-b3ba-50ec65cfddda",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from pythonbpf import bpf, map, section, bpfglobal, BPF, trace_fields\n",
|
||||
"from pythonbpf.helper import ktime\n",
|
||||
"from pythonbpf.maps import HashMap\n",
|
||||
"\n",
|
||||
"from ctypes import c_void_p, c_int64"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "ddb115f4-20a7-43bc-bb5b-ccbfd6031fc2",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"@bpf\n",
|
||||
"@map\n",
|
||||
"def last() -> HashMap:\n",
|
||||
" return HashMap(key=c_int64, value=c_int64, max_entries=1)\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@section(\"tracepoint/syscalls/sys_enter_sync\")\n",
|
||||
"def do_trace(ctx: c_void_p) -> c_int64:\n",
|
||||
" key = 0\n",
|
||||
" tsp = last.lookup(key)\n",
|
||||
" if tsp:\n",
|
||||
" delta = ktime() - tsp\n",
|
||||
" if delta < 1000000000:\n",
|
||||
" time_ms = delta // 1000000\n",
|
||||
" print(f\"{time_ms}\")\n",
|
||||
" last.delete(key)\n",
|
||||
" else:\n",
|
||||
" last.update(key, ktime())\n",
|
||||
" return 0\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@bpfglobal\n",
|
||||
"def LICENSE() -> str:\n",
|
||||
" return \"GPL\"\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"# Compile and load\n",
|
||||
"b = BPF()\n",
|
||||
"b.load()\n",
|
||||
"b.attach_all()"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "e4f46574-9fd8-46e7-9c7b-27a36d07f200",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"print(\"Tracing for quick sync's... Ctrl-C to end\")\n",
|
||||
"\n",
|
||||
"# format output\n",
|
||||
"start = 0\n",
|
||||
"while True:\n",
|
||||
" try:\n",
|
||||
" task, pid, cpu, flags, ts, ms = trace_fields()\n",
|
||||
" if start == 0:\n",
|
||||
" start = ts\n",
|
||||
" ts -= start\n",
|
||||
" print(f\"At time {ts} s: Multiple syncs detected, last {ms} ms ago\")\n",
|
||||
" except KeyboardInterrupt:\n",
|
||||
" exit()"
|
||||
]
|
||||
}
|
||||
],
|
||||
"metadata": {
|
||||
"kernelspec": {
|
||||
"display_name": "Python 3 (ipykernel)",
|
||||
"language": "python",
|
||||
"name": "python3"
|
||||
},
|
||||
"language_info": {
|
||||
"codemirror_mode": {
|
||||
"name": "ipython",
|
||||
"version": 3
|
||||
},
|
||||
"file_extension": ".py",
|
||||
"mimetype": "text/x-python",
|
||||
"name": "python",
|
||||
"nbconvert_exporter": "python",
|
||||
"pygments_lexer": "ipython3",
|
||||
"version": "3.13.3"
|
||||
}
|
||||
},
|
||||
"nbformat": 4,
|
||||
"nbformat_minor": 5
|
||||
}
|
||||
73
BCC-Examples/sys_sync.ipynb
Normal file
73
BCC-Examples/sys_sync.ipynb
Normal file
@ -0,0 +1,73 @@
|
||||
{
|
||||
"cells": [
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "bb49598f-b9cc-4ea8-8391-923cad513711",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from pythonbpf import bpf, section, bpfglobal, BPF, trace_pipe\n",
|
||||
"from ctypes import c_void_p, c_int64"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "5da237b0-1c7d-4ec5-8c24-696b1c1d97fa",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"@bpf\n",
|
||||
"@section(\"tracepoint/syscalls/sys_enter_sync\")\n",
|
||||
"def hello_world(ctx: c_void_p) -> c_int64:\n",
|
||||
" print(\"sys_sync() called\")\n",
|
||||
" return 0\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"@bpf\n",
|
||||
"@bpfglobal\n",
|
||||
"def LICENSE() -> str:\n",
|
||||
" return \"GPL\"\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"# Compile and load\n",
|
||||
"b = BPF()\n",
|
||||
"b.load()\n",
|
||||
"b.attach_all()"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "e4c218ac-fe47-4fd1-a27b-c07e02f3cd05",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"print(\"Tracing sys_sync()... Ctrl-C to end.\")\n",
|
||||
"trace_pipe()"
|
||||
]
|
||||
}
|
||||
],
|
||||
"metadata": {
|
||||
"kernelspec": {
|
||||
"display_name": "Python 3 (ipykernel)",
|
||||
"language": "python",
|
||||
"name": "python3"
|
||||
},
|
||||
"language_info": {
|
||||
"codemirror_mode": {
|
||||
"name": "ipython",
|
||||
"version": 3
|
||||
},
|
||||
"file_extension": ".py",
|
||||
"mimetype": "text/x-python",
|
||||
"name": "python",
|
||||
"nbconvert_exporter": "python",
|
||||
"pygments_lexer": "ipython3",
|
||||
"version": "3.13.3"
|
||||
}
|
||||
},
|
||||
"nbformat": 4,
|
||||
"nbformat_minor": 5
|
||||
}
|
||||
252
BCC-Examples/vfsreadlat.ipynb
Normal file
252
BCC-Examples/vfsreadlat.ipynb
Normal file
File diff suppressed because one or more lines are too long
@ -40,6 +40,12 @@ Python-BPF is an LLVM IR generator for eBPF programs written in Python. It uses
|
||||
|
||||
---
|
||||
|
||||
## Try It Out!
|
||||
Run
|
||||
```bash
|
||||
curl -s https://raw.githubusercontent.com/pythonbpf/Python-BPF/refs/heads/master/tools/setup.sh | sudo bash
|
||||
```
|
||||
|
||||
## Installation
|
||||
|
||||
Dependencies:
|
||||
|
||||
@ -21,17 +21,17 @@ def last() -> HashMap:
|
||||
@section("tracepoint/syscalls/sys_enter_execve")
|
||||
def do_trace(ctx: c_void_p) -> c_int64:
|
||||
key = 0
|
||||
tsp = last().lookup(key)
|
||||
tsp = last.lookup(key)
|
||||
if tsp:
|
||||
kt = ktime()
|
||||
delta = kt - tsp
|
||||
if delta < 1000000000:
|
||||
time_ms = delta // 1000000
|
||||
print(f"Execve syscall entered within last second, last {time_ms} ms ago")
|
||||
last().delete(key)
|
||||
last.delete(key)
|
||||
else:
|
||||
kt = ktime()
|
||||
last().update(key, kt)
|
||||
last.update(key, kt)
|
||||
return c_int64(0)
|
||||
|
||||
|
||||
|
||||
File diff suppressed because one or more lines are too long
@ -3,7 +3,6 @@ import time
|
||||
from pythonbpf import bpf, map, section, bpfglobal, BPF
|
||||
from pythonbpf.helper import pid
|
||||
from pythonbpf.maps import HashMap
|
||||
from pylibbpf import BpfMap
|
||||
from ctypes import c_void_p, c_int64, c_uint64, c_int32
|
||||
import matplotlib.pyplot as plt
|
||||
|
||||
@ -26,14 +25,14 @@ def hist() -> HashMap:
|
||||
def hello(ctx: c_void_p) -> c_int64:
|
||||
process_id = pid()
|
||||
one = 1
|
||||
prev = hist().lookup(process_id)
|
||||
prev = hist.lookup(process_id)
|
||||
if prev:
|
||||
previous_value = prev + 1
|
||||
print(f"count: {previous_value} with {process_id}")
|
||||
hist().update(process_id, previous_value)
|
||||
hist.update(process_id, previous_value)
|
||||
return c_int64(0)
|
||||
else:
|
||||
hist().update(process_id, one)
|
||||
hist.update(process_id, one)
|
||||
return c_int64(0)
|
||||
|
||||
|
||||
@ -44,12 +43,12 @@ def LICENSE() -> str:
|
||||
|
||||
|
||||
b = BPF()
|
||||
b.load_and_attach()
|
||||
hist = BpfMap(b, hist)
|
||||
b.load()
|
||||
b.attach_all()
|
||||
print("Recording")
|
||||
time.sleep(10)
|
||||
|
||||
counts = list(hist.values())
|
||||
counts = list(b["hist"].values())
|
||||
|
||||
plt.hist(counts, bins=20)
|
||||
plt.xlabel("Clone calls per PID")
|
||||
|
||||
@ -1,4 +1,4 @@
|
||||
from pythonbpf import bpf, section, bpfglobal, BPF
|
||||
from pythonbpf import bpf, section, bpfglobal, BPF, trace_pipe
|
||||
from ctypes import c_void_p, c_int64
|
||||
|
||||
# Instructions to how to run this program
|
||||
@ -21,10 +21,6 @@ def LICENSE() -> str:
|
||||
|
||||
|
||||
b = BPF()
|
||||
b.load_and_attach()
|
||||
if b.is_loaded() and b.is_attached():
|
||||
print("Successfully loaded and attached")
|
||||
else:
|
||||
print("Could not load successfully")
|
||||
|
||||
# Now cat /sys/kernel/debug/tracing/trace_pipe to see results of the execve syscall.
|
||||
b.load()
|
||||
b.attach_all()
|
||||
trace_pipe()
|
||||
|
||||
@ -1,4 +1,4 @@
|
||||
from pythonbpf import bpf, section, bpfglobal, BPF
|
||||
from pythonbpf import bpf, section, bpfglobal, BPF, trace_pipe
|
||||
from ctypes import c_void_p, c_int64
|
||||
|
||||
|
||||
@ -23,7 +23,7 @@ def LICENSE() -> str:
|
||||
|
||||
|
||||
b = BPF()
|
||||
b.load_and_attach()
|
||||
while True:
|
||||
print("running")
|
||||
# Now cat /sys/kernel/debug/tracing/trace_pipe to see results of unlink kprobe.
|
||||
b.load()
|
||||
b.attach_all()
|
||||
print("running")
|
||||
trace_pipe()
|
||||
|
||||
@ -23,14 +23,14 @@ def count() -> HashMap:
|
||||
def hello_world(ctx: c_void_p) -> c_int64:
|
||||
key = 0
|
||||
one = 1
|
||||
prev = count().lookup(key)
|
||||
prev = count.lookup(key)
|
||||
if prev:
|
||||
prevval = prev + 1
|
||||
print(f"count: {prevval}")
|
||||
count().update(key, prevval)
|
||||
count.update(key, prevval)
|
||||
return XDP_PASS
|
||||
else:
|
||||
count().update(key, one)
|
||||
count.update(key, one)
|
||||
|
||||
return XDP_PASS
|
||||
|
||||
|
||||
@ -4,12 +4,26 @@ build-backend = "setuptools.build_meta"
|
||||
|
||||
[project]
|
||||
name = "pythonbpf"
|
||||
version = "0.1.4"
|
||||
version = "0.1.6"
|
||||
description = "Reduced Python frontend for eBPF"
|
||||
authors = [
|
||||
{ name = "r41k0u", email="pragyanshchaturvedi18@gmail.com" },
|
||||
{ name = "varun-r-mallya", email="varunrmallya@gmail.com" }
|
||||
]
|
||||
classifiers = [
|
||||
"Development Status :: 3 - Alpha",
|
||||
"Intended Audience :: Developers",
|
||||
"Operating System :: POSIX :: Linux",
|
||||
"Programming Language :: Python :: 3",
|
||||
"Programming Language :: Python :: 3.8",
|
||||
"Programming Language :: Python :: 3.9",
|
||||
"Programming Language :: Python :: 3.10",
|
||||
"Programming Language :: Python :: 3.11",
|
||||
"Programming Language :: Python :: 3.12",
|
||||
"Programming Language :: Python",
|
||||
"Topic :: Software Development :: Libraries :: Python Modules",
|
||||
"Topic :: System :: Operating System Kernels :: Linux",
|
||||
]
|
||||
readme = "README.md"
|
||||
license = {text = "Apache-2.0"}
|
||||
requires-python = ">=3.8"
|
||||
|
||||
@ -1,37 +1,26 @@
|
||||
import ast
|
||||
import logging
|
||||
|
||||
import ctypes
|
||||
from llvmlite import ir
|
||||
from dataclasses import dataclass
|
||||
from typing import Any
|
||||
from .local_symbol import LocalSymbol
|
||||
from pythonbpf.helper import HelperHandlerRegistry
|
||||
from pythonbpf.vmlinux_parser.dependency_node import Field
|
||||
from .expr import VmlinuxHandlerRegistry
|
||||
from pythonbpf.type_deducer import ctypes_to_ir
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
@dataclass
|
||||
class LocalSymbol:
|
||||
var: ir.AllocaInstr
|
||||
ir_type: ir.Type
|
||||
metadata: Any = None
|
||||
|
||||
def __iter__(self):
|
||||
yield self.var
|
||||
yield self.ir_type
|
||||
yield self.metadata
|
||||
|
||||
|
||||
def create_targets_and_rvals(stmt):
|
||||
"""Create lists of targets and right-hand values from an assignment statement."""
|
||||
if isinstance(stmt.targets[0], ast.Tuple):
|
||||
if not isinstance(stmt.value, ast.Tuple):
|
||||
logger.warning("Mismatched multi-target assignment, skipping allocation")
|
||||
return
|
||||
return [], []
|
||||
targets, rvals = stmt.targets[0].elts, stmt.value.elts
|
||||
if len(targets) != len(rvals):
|
||||
logger.warning("length of LHS != length of RHS, skipping allocation")
|
||||
return
|
||||
return [], []
|
||||
return targets, rvals
|
||||
return stmt.targets, [stmt.value]
|
||||
|
||||
@ -59,7 +48,6 @@ def handle_assign_allocation(builder, stmt, local_sym_tab, structs_sym_tab):
|
||||
continue
|
||||
|
||||
var_name = target.id
|
||||
|
||||
# Skip if already allocated
|
||||
if var_name in local_sym_tab:
|
||||
logger.debug(f"Variable {var_name} already allocated, skipping")
|
||||
@ -93,7 +81,7 @@ def _allocate_for_call(builder, var_name, rval, local_sym_tab, structs_sym_tab):
|
||||
call_type = rval.func.id
|
||||
|
||||
# C type constructors
|
||||
if call_type in ("c_int32", "c_int64", "c_uint32", "c_uint64"):
|
||||
if call_type in ("c_int32", "c_int64", "c_uint32", "c_uint64", "c_void_p"):
|
||||
ir_type = ctypes_to_ir(call_type)
|
||||
var = builder.alloca(ir_type, name=var_name)
|
||||
var.align = ir_type.width // 8
|
||||
@ -238,9 +226,92 @@ def _allocate_for_attribute(builder, var_name, rval, local_sym_tab, structs_sym_
|
||||
logger.error(f"Struct variable '{struct_var}' not found")
|
||||
return
|
||||
|
||||
struct_type = local_sym_tab[struct_var].metadata
|
||||
struct_type: type = local_sym_tab[struct_var].metadata
|
||||
if not struct_type or struct_type not in structs_sym_tab:
|
||||
logger.error(f"Struct type '{struct_type}' not found")
|
||||
if VmlinuxHandlerRegistry.is_vmlinux_struct(struct_type.__name__):
|
||||
# Handle vmlinux struct field access
|
||||
vmlinux_struct_name = struct_type.__name__
|
||||
if not VmlinuxHandlerRegistry.has_field(vmlinux_struct_name, field_name):
|
||||
logger.error(
|
||||
f"Field '{field_name}' not found in vmlinux struct '{vmlinux_struct_name}'"
|
||||
)
|
||||
return
|
||||
|
||||
field_type: tuple[ir.GlobalVariable, Field] = (
|
||||
VmlinuxHandlerRegistry.get_field_type(vmlinux_struct_name, field_name)
|
||||
)
|
||||
field_ir, field = field_type
|
||||
# TODO: For now, we only support integer type allocations.
|
||||
# This always assumes first argument of function to be the context struct
|
||||
base_ptr = builder.function.args[0]
|
||||
local_sym_tab[
|
||||
struct_var
|
||||
].var = base_ptr # This is repurposing of var to store the pointer of the base type
|
||||
local_sym_tab[struct_var].ir_type = field_ir
|
||||
|
||||
# Determine the actual IR type based on the field's type
|
||||
actual_ir_type = None
|
||||
|
||||
# Check if it's a ctypes primitive
|
||||
if field.type.__module__ == ctypes.__name__:
|
||||
try:
|
||||
field_size_bytes = ctypes.sizeof(field.type)
|
||||
field_size_bits = field_size_bytes * 8
|
||||
|
||||
if field_size_bits in [8, 16, 32, 64]:
|
||||
# Special case: struct_xdp_md i32 fields should allocate as i64
|
||||
# because load_ctx_field will zero-extend them to i64
|
||||
if (
|
||||
vmlinux_struct_name == "struct_xdp_md"
|
||||
and field_size_bits == 32
|
||||
):
|
||||
actual_ir_type = ir.IntType(64)
|
||||
logger.info(
|
||||
f"Allocating {var_name} as i64 for i32 field from struct_xdp_md.{field_name} "
|
||||
"(will be zero-extended during load)"
|
||||
)
|
||||
else:
|
||||
actual_ir_type = ir.IntType(field_size_bits)
|
||||
else:
|
||||
logger.warning(
|
||||
f"Unusual field size {field_size_bits} bits for {field_name}"
|
||||
)
|
||||
actual_ir_type = ir.IntType(64)
|
||||
except Exception as e:
|
||||
logger.warning(
|
||||
f"Could not determine size for ctypes field {field_name}: {e}"
|
||||
)
|
||||
actual_ir_type = ir.IntType(64)
|
||||
|
||||
# Check if it's a nested vmlinux struct or complex type
|
||||
elif field.type.__module__ == "vmlinux":
|
||||
# For pointers to structs, use pointer type (64-bit)
|
||||
if field.ctype_complex_type is not None and issubclass(
|
||||
field.ctype_complex_type, ctypes._Pointer
|
||||
):
|
||||
actual_ir_type = ir.IntType(64) # Pointer is always 64-bit
|
||||
# For embedded structs, this is more complex - might need different handling
|
||||
else:
|
||||
logger.warning(
|
||||
f"Field {field_name} is a nested vmlinux struct, using i64 for now"
|
||||
)
|
||||
actual_ir_type = ir.IntType(64)
|
||||
else:
|
||||
logger.warning(
|
||||
f"Unknown field type module {field.type.__module__} for {field_name}"
|
||||
)
|
||||
actual_ir_type = ir.IntType(64)
|
||||
|
||||
# Allocate with the actual IR type, not the GlobalVariable
|
||||
var = _allocate_with_type(builder, var_name, actual_ir_type)
|
||||
local_sym_tab[var_name] = LocalSymbol(var, actual_ir_type, field)
|
||||
|
||||
logger.info(
|
||||
f"Pre-allocated {var_name} from vmlinux struct {vmlinux_struct_name}.{field_name}"
|
||||
)
|
||||
return
|
||||
else:
|
||||
logger.error(f"Struct type '{struct_type}' not found")
|
||||
return
|
||||
|
||||
struct_info = structs_sym_tab[struct_type]
|
||||
|
||||
@ -3,6 +3,8 @@ import logging
|
||||
from llvmlite import ir
|
||||
from pythonbpf.expr import eval_expr
|
||||
from pythonbpf.helper import emit_probe_read_kernel_str_call
|
||||
from pythonbpf.type_deducer import ctypes_to_ir
|
||||
from pythonbpf.vmlinux_parser.dependency_node import Field
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
@ -148,7 +150,33 @@ def handle_variable_assignment(
|
||||
val, val_type = val_result
|
||||
logger.info(f"Evaluated value for {var_name}: {val} of type {val_type}, {var_type}")
|
||||
if val_type != var_type:
|
||||
if isinstance(val_type, ir.IntType) and isinstance(var_type, ir.IntType):
|
||||
if isinstance(val_type, Field):
|
||||
logger.info("Handling assignment to struct field")
|
||||
# Special handling for struct_xdp_md i32 fields that are zero-extended to i64
|
||||
# The load_ctx_field already extended them, so val is i64 but val_type.type shows c_uint
|
||||
if (
|
||||
hasattr(val_type, "type")
|
||||
and val_type.type.__name__ == "c_uint"
|
||||
and isinstance(var_type, ir.IntType)
|
||||
and var_type.width == 64
|
||||
):
|
||||
# This is the struct_xdp_md case - value is already i64
|
||||
builder.store(val, var_ptr)
|
||||
logger.info(
|
||||
f"Assigned zero-extended struct_xdp_md i32 field to {var_name} (i64)"
|
||||
)
|
||||
return True
|
||||
# TODO: handling only ctype struct fields for now. Handle other stuff too later.
|
||||
elif var_type == ctypes_to_ir(val_type.type.__name__):
|
||||
builder.store(val, var_ptr)
|
||||
logger.info(f"Assigned ctype struct field to {var_name}")
|
||||
return True
|
||||
else:
|
||||
logger.error(
|
||||
f"Failed to assign ctype struct field to {var_name}: {val_type} != {var_type}"
|
||||
)
|
||||
return False
|
||||
elif isinstance(val_type, ir.IntType) and isinstance(var_type, ir.IntType):
|
||||
# Allow implicit int widening
|
||||
if val_type.width < var_type.width:
|
||||
val = builder.sext(val, var_type)
|
||||
|
||||
@ -5,6 +5,8 @@ from .functions import func_proc
|
||||
from .maps import maps_proc
|
||||
from .structs import structs_proc
|
||||
from .vmlinux_parser import vmlinux_proc
|
||||
from pythonbpf.vmlinux_parser.vmlinux_exports_handler import VmlinuxHandler
|
||||
from .expr import VmlinuxHandlerRegistry
|
||||
from .globals_pass import (
|
||||
globals_list_creation,
|
||||
globals_processing,
|
||||
@ -19,10 +21,38 @@ from pylibbpf import BpfObject
|
||||
import tempfile
|
||||
from logging import Logger
|
||||
import logging
|
||||
import re
|
||||
|
||||
logger: Logger = logging.getLogger(__name__)
|
||||
|
||||
VERSION = "v0.1.4"
|
||||
VERSION = "v0.1.6"
|
||||
|
||||
|
||||
def finalize_module(original_str):
|
||||
"""After all IR generation is complete, we monkey patch btf_ama attribute"""
|
||||
|
||||
# Create a string with applied transformation of btf_ama attribute addition to BTF struct field accesses.
|
||||
pattern = r'(@"llvm\.[^"]+:[^"]*" = external global i64, !llvm\.preserve\.access\.index ![0-9]+)'
|
||||
replacement = r'\1 "btf_ama"'
|
||||
return re.sub(pattern, replacement, original_str)
|
||||
|
||||
|
||||
def bpf_passthrough_gen(module):
|
||||
i32_ty = ir.IntType(32)
|
||||
ptr_ty = ir.PointerType(ir.IntType(8))
|
||||
fnty = ir.FunctionType(ptr_ty, [i32_ty, ptr_ty])
|
||||
|
||||
# Declare the intrinsic
|
||||
passthrough = ir.Function(module, fnty, "llvm.bpf.passthrough.p0.p0")
|
||||
|
||||
# Set function attributes
|
||||
# TODO: the ones commented are supposed to be there but cannot be added due to llvmlite limitations at the moment
|
||||
# passthrough.attributes.add("nofree")
|
||||
# passthrough.attributes.add("nosync")
|
||||
passthrough.attributes.add("nounwind")
|
||||
# passthrough.attributes.add("memory(none)")
|
||||
|
||||
return passthrough
|
||||
|
||||
|
||||
def find_bpf_chunks(tree):
|
||||
@ -45,11 +75,16 @@ def processor(source_code, filename, module):
|
||||
for func_node in bpf_chunks:
|
||||
logger.info(f"Found BPF function/struct: {func_node.name}")
|
||||
|
||||
vmlinux_proc(tree, module)
|
||||
bpf_passthrough_gen(module)
|
||||
|
||||
vmlinux_symtab = vmlinux_proc(tree, module)
|
||||
if vmlinux_symtab:
|
||||
handler = VmlinuxHandler.initialize(vmlinux_symtab)
|
||||
VmlinuxHandlerRegistry.set_handler(handler)
|
||||
|
||||
populate_global_symbol_table(tree, module)
|
||||
license_processing(tree, module)
|
||||
globals_processing(tree, module)
|
||||
|
||||
structs_sym_tab = structs_proc(tree, module, bpf_chunks)
|
||||
map_sym_tab = maps_proc(tree, module, bpf_chunks)
|
||||
func_proc(tree, module, bpf_chunks, map_sym_tab, structs_sym_tab)
|
||||
@ -122,10 +157,12 @@ def compile_to_ir(filename: str, output: str, loglevel=logging.INFO):
|
||||
|
||||
module.add_named_metadata("llvm.ident", [f"PythonBPF {VERSION}"])
|
||||
|
||||
module_string: str = finalize_module(str(module))
|
||||
|
||||
logger.info(f"IR written to {output}")
|
||||
with open(output, "w") as f:
|
||||
f.write(f'source_filename = "{filename}"\n')
|
||||
f.write(str(module))
|
||||
f.write(module_string)
|
||||
f.write("\n")
|
||||
|
||||
return output, structs_sym_tab, maps_sym_tab
|
||||
|
||||
@ -81,6 +81,20 @@ class DebugInfoGenerator:
|
||||
},
|
||||
)
|
||||
|
||||
def create_array_type_vmlinux(self, type_info: Any, count: int) -> Any:
|
||||
"""Create an array type of the given base type with specified count"""
|
||||
base_type, type_sizing = type_info
|
||||
subrange = self.module.add_debug_info("DISubrange", {"count": count})
|
||||
return self.module.add_debug_info(
|
||||
"DICompositeType",
|
||||
{
|
||||
"tag": dc.DW_TAG_array_type,
|
||||
"baseType": base_type,
|
||||
"size": type_sizing,
|
||||
"elements": [subrange],
|
||||
},
|
||||
)
|
||||
|
||||
@staticmethod
|
||||
def _compute_array_size(base_type: Any, count: int) -> int:
|
||||
# Extract size from base_type if possible
|
||||
@ -101,6 +115,23 @@ class DebugInfoGenerator:
|
||||
},
|
||||
)
|
||||
|
||||
def create_struct_member_vmlinux(
|
||||
self, name: str, base_type_with_size: Any, offset: int
|
||||
) -> Any:
|
||||
"""Create a struct member with the given name, type, and offset"""
|
||||
base_type, type_size = base_type_with_size
|
||||
return self.module.add_debug_info(
|
||||
"DIDerivedType",
|
||||
{
|
||||
"tag": dc.DW_TAG_member,
|
||||
"name": name,
|
||||
"file": self.module._file_metadata,
|
||||
"baseType": base_type,
|
||||
"size": type_size,
|
||||
"offset": offset,
|
||||
},
|
||||
)
|
||||
|
||||
def create_struct_type(
|
||||
self, members: List[Any], size: int, is_distinct: bool
|
||||
) -> Any:
|
||||
@ -116,6 +147,22 @@ class DebugInfoGenerator:
|
||||
is_distinct=is_distinct,
|
||||
)
|
||||
|
||||
def create_struct_type_with_name(
|
||||
self, name: str, members: List[Any], size: int, is_distinct: bool
|
||||
) -> Any:
|
||||
"""Create a struct type with the given members and size"""
|
||||
return self.module.add_debug_info(
|
||||
"DICompositeType",
|
||||
{
|
||||
"name": name,
|
||||
"tag": dc.DW_TAG_structure_type,
|
||||
"file": self.module._file_metadata,
|
||||
"size": size,
|
||||
"elements": members,
|
||||
},
|
||||
is_distinct=is_distinct,
|
||||
)
|
||||
|
||||
def create_global_var_debug_info(
|
||||
self, name: str, var_type: Any, is_local: bool = False
|
||||
) -> Any:
|
||||
@ -137,3 +184,83 @@ class DebugInfoGenerator:
|
||||
"DIGlobalVariableExpression",
|
||||
{"var": global_var, "expr": self.module.add_debug_info("DIExpression", {})},
|
||||
)
|
||||
|
||||
def get_int64_type(self):
|
||||
return self.get_basic_type("long", 64, dc.DW_ATE_signed)
|
||||
|
||||
def create_subroutine_type(self, return_type, param_types):
|
||||
"""
|
||||
Create a DISubroutineType given return type and list of parameter types.
|
||||
Equivalent to: !DISubroutineType(types: !{ret, args...})
|
||||
"""
|
||||
type_array = [return_type]
|
||||
if isinstance(param_types, (list, tuple)):
|
||||
type_array.extend(param_types)
|
||||
else:
|
||||
type_array.append(param_types)
|
||||
return self.module.add_debug_info("DISubroutineType", {"types": type_array})
|
||||
|
||||
def create_local_variable_debug_info(
|
||||
self, name: str, arg: int, var_type: Any
|
||||
) -> Any:
|
||||
"""
|
||||
Create debug info for a local variable (DILocalVariable) without scope.
|
||||
Example:
|
||||
!DILocalVariable(name: "ctx", arg: 1, file: !3, line: 20, type: !7)
|
||||
"""
|
||||
return self.module.add_debug_info(
|
||||
"DILocalVariable",
|
||||
{
|
||||
"name": name,
|
||||
"arg": arg,
|
||||
"file": self.module._file_metadata,
|
||||
"type": var_type,
|
||||
},
|
||||
)
|
||||
|
||||
def add_scope_to_local_variable(self, local_variable_debug_info, scope_value):
|
||||
"""
|
||||
Add scope information to an existing local variable debug info object.
|
||||
"""
|
||||
# TODO: this is a workaround a flaw in the debug info generation. Fix this if possible in the future.
|
||||
# We should not be touching llvmlite's internals like this.
|
||||
if hasattr(local_variable_debug_info, "operands"):
|
||||
# LLVM metadata operands is a tuple, so we need to rebuild it
|
||||
existing_operands = local_variable_debug_info.operands
|
||||
|
||||
# Convert tuple to list, add scope, convert back to tuple
|
||||
operands_list = list(existing_operands)
|
||||
operands_list.append(("scope", scope_value))
|
||||
|
||||
# Reassign the new tuple
|
||||
local_variable_debug_info.operands = tuple(operands_list)
|
||||
|
||||
def create_subprogram(
|
||||
self, name: str, subroutine_type: Any, retained_nodes: List[Any]
|
||||
) -> Any:
|
||||
"""
|
||||
Create a DISubprogram for a function.
|
||||
|
||||
Args:
|
||||
name: Function name
|
||||
subroutine_type: DISubroutineType for the function signature
|
||||
retained_nodes: List of DILocalVariable nodes for function parameters/variables
|
||||
|
||||
Returns:
|
||||
DISubprogram metadata
|
||||
"""
|
||||
return self.module.add_debug_info(
|
||||
"DISubprogram",
|
||||
{
|
||||
"name": name,
|
||||
"scope": self.module._file_metadata,
|
||||
"file": self.module._file_metadata,
|
||||
"type": subroutine_type,
|
||||
# TODO: the following flags do not exist at the moment in our dwarf constants file. We need to add them.
|
||||
# "flags": dc.DW_FLAG_Prototyped | dc.DW_FLAG_AllCallsDescribed,
|
||||
# "spFlags": dc.DW_SPFLAG_Definition | dc.DW_SPFLAG_Optimized,
|
||||
"unit": self.module._debug_compile_unit,
|
||||
"retainedNodes": retained_nodes,
|
||||
},
|
||||
is_distinct=True,
|
||||
)
|
||||
|
||||
@ -2,6 +2,7 @@ from .expr_pass import eval_expr, handle_expr, get_operand_value
|
||||
from .type_normalization import convert_to_bool, get_base_type_and_depth
|
||||
from .ir_ops import deref_to_depth
|
||||
from .call_registry import CallHandlerRegistry
|
||||
from .vmlinux_registry import VmlinuxHandlerRegistry
|
||||
|
||||
__all__ = [
|
||||
"eval_expr",
|
||||
@ -11,4 +12,5 @@ __all__ = [
|
||||
"deref_to_depth",
|
||||
"get_operand_value",
|
||||
"CallHandlerRegistry",
|
||||
"VmlinuxHandlerRegistry",
|
||||
]
|
||||
|
||||
@ -12,6 +12,8 @@ from .type_normalization import (
|
||||
get_base_type_and_depth,
|
||||
deref_to_depth,
|
||||
)
|
||||
from pythonbpf.vmlinux_parser.assignment_info import Field
|
||||
from .vmlinux_registry import VmlinuxHandlerRegistry
|
||||
|
||||
logger: Logger = logging.getLogger(__name__)
|
||||
|
||||
@ -27,8 +29,12 @@ def _handle_name_expr(expr: ast.Name, local_sym_tab: Dict, builder: ir.IRBuilder
|
||||
val = builder.load(var)
|
||||
return val, local_sym_tab[expr.id].ir_type
|
||||
else:
|
||||
logger.info(f"Undefined variable {expr.id}")
|
||||
return None
|
||||
# Check if it's a vmlinux enum/constant
|
||||
vmlinux_result = VmlinuxHandlerRegistry.handle_name(expr.id)
|
||||
if vmlinux_result is not None:
|
||||
return vmlinux_result
|
||||
|
||||
raise SyntaxError(f"Undefined variable {expr.id}")
|
||||
|
||||
|
||||
def _handle_constant_expr(module, builder, expr: ast.Constant):
|
||||
@ -67,7 +73,22 @@ def _handle_attribute_expr(
|
||||
if var_name in local_sym_tab:
|
||||
var_ptr, var_type, var_metadata = local_sym_tab[var_name]
|
||||
logger.info(f"Loading attribute {attr_name} from variable {var_name}")
|
||||
logger.info(f"Variable type: {var_type}, Variable ptr: {var_ptr}")
|
||||
logger.info(
|
||||
f"Variable type: {var_type}, Variable ptr: {var_ptr}, Variable Metadata: {var_metadata}"
|
||||
)
|
||||
if (
|
||||
hasattr(var_metadata, "__module__")
|
||||
and var_metadata.__module__ == "vmlinux"
|
||||
):
|
||||
# Try vmlinux handler when var_metadata is not a string, but has a module attribute.
|
||||
# This has been done to keep everything separate in vmlinux struct handling.
|
||||
vmlinux_result = VmlinuxHandlerRegistry.handle_attribute(
|
||||
expr, local_sym_tab, None, builder
|
||||
)
|
||||
if vmlinux_result is not None:
|
||||
return vmlinux_result
|
||||
else:
|
||||
raise RuntimeError("Vmlinux struct did not process successfully")
|
||||
metadata = structs_sym_tab[var_metadata]
|
||||
if attr_name in metadata.fields:
|
||||
gep = metadata.gep(builder, var_ptr, attr_name)
|
||||
@ -130,7 +151,12 @@ def get_operand_value(
|
||||
logger.info(f"var is {var}, base_type is {base_type}, depth is {depth}")
|
||||
val = deref_to_depth(func, builder, var, depth)
|
||||
return val
|
||||
raise ValueError(f"Undefined variable: {operand.id}")
|
||||
else:
|
||||
# Check if it's a vmlinux enum/constant
|
||||
vmlinux_result = VmlinuxHandlerRegistry.handle_name(operand.id)
|
||||
if vmlinux_result is not None:
|
||||
val, _ = vmlinux_result
|
||||
return val
|
||||
elif isinstance(operand, ast.Constant):
|
||||
if isinstance(operand.value, int):
|
||||
cst = ir.Constant(ir.IntType(64), int(operand.value))
|
||||
@ -254,16 +280,45 @@ def _handle_ctypes_call(
|
||||
call_type = expr.func.id
|
||||
expected_type = ctypes_to_ir(call_type)
|
||||
|
||||
if val[1] != expected_type:
|
||||
# Extract the actual IR value and type
|
||||
# val could be (value, ir_type) or (value, Field)
|
||||
value, val_type = val
|
||||
|
||||
# If val_type is a Field object (from vmlinux struct), get the actual IR type of the value
|
||||
if isinstance(val_type, Field):
|
||||
# The value is already the correct IR value (potentially zero-extended)
|
||||
# Get the IR type from the value itself
|
||||
actual_ir_type = value.type
|
||||
logger.info(
|
||||
f"Converting vmlinux field {val_type.name} (IR type: {actual_ir_type}) to {call_type}"
|
||||
)
|
||||
else:
|
||||
actual_ir_type = val_type
|
||||
|
||||
if actual_ir_type != expected_type:
|
||||
# NOTE: We are only considering casting to and from int types for now
|
||||
if isinstance(val[1], ir.IntType) and isinstance(expected_type, ir.IntType):
|
||||
if val[1].width < expected_type.width:
|
||||
val = (builder.sext(val[0], expected_type), expected_type)
|
||||
if isinstance(actual_ir_type, ir.IntType) and isinstance(
|
||||
expected_type, ir.IntType
|
||||
):
|
||||
if actual_ir_type.width < expected_type.width:
|
||||
value = builder.sext(value, expected_type)
|
||||
logger.info(
|
||||
f"Sign-extended from i{actual_ir_type.width} to i{expected_type.width}"
|
||||
)
|
||||
elif actual_ir_type.width > expected_type.width:
|
||||
value = builder.trunc(value, expected_type)
|
||||
logger.info(
|
||||
f"Truncated from i{actual_ir_type.width} to i{expected_type.width}"
|
||||
)
|
||||
else:
|
||||
val = (builder.trunc(val[0], expected_type), expected_type)
|
||||
# Same width, just use as-is (e.g., both i64)
|
||||
pass
|
||||
else:
|
||||
raise ValueError(f"Type mismatch: expected {expected_type}, got {val[1]}")
|
||||
return val
|
||||
raise ValueError(
|
||||
f"Type mismatch: expected {expected_type}, got {actual_ir_type} (original type: {val_type})"
|
||||
)
|
||||
|
||||
return value, expected_type
|
||||
|
||||
|
||||
def _handle_compare(
|
||||
@ -332,6 +387,7 @@ def _handle_unary_op(
|
||||
neg_one = ir.Constant(ir.IntType(64), -1)
|
||||
result = builder.mul(operand, neg_one)
|
||||
return result, ir.IntType(64)
|
||||
return None
|
||||
|
||||
|
||||
# ============================================================================
|
||||
|
||||
75
pythonbpf/expr/vmlinux_registry.py
Normal file
75
pythonbpf/expr/vmlinux_registry.py
Normal file
@ -0,0 +1,75 @@
|
||||
import ast
|
||||
|
||||
from pythonbpf.vmlinux_parser.vmlinux_exports_handler import VmlinuxHandler
|
||||
|
||||
|
||||
class VmlinuxHandlerRegistry:
|
||||
"""Registry for vmlinux handler operations"""
|
||||
|
||||
_handler = None
|
||||
|
||||
@classmethod
|
||||
def set_handler(cls, handler: VmlinuxHandler):
|
||||
"""Set the vmlinux handler"""
|
||||
cls._handler = handler
|
||||
|
||||
@classmethod
|
||||
def get_handler(cls):
|
||||
"""Get the vmlinux handler"""
|
||||
return cls._handler
|
||||
|
||||
@classmethod
|
||||
def handle_name(cls, name):
|
||||
"""Try to handle a name as vmlinux enum/constant"""
|
||||
if cls._handler is None:
|
||||
return None
|
||||
return cls._handler.handle_vmlinux_enum(name)
|
||||
|
||||
@classmethod
|
||||
def handle_attribute(cls, expr, local_sym_tab, module, builder):
|
||||
"""Try to handle an attribute access as vmlinux struct field"""
|
||||
if cls._handler is None:
|
||||
return None
|
||||
|
||||
if isinstance(expr.value, ast.Name):
|
||||
var_name = expr.value.id
|
||||
field_name = expr.attr
|
||||
return cls._handler.handle_vmlinux_struct_field(
|
||||
var_name, field_name, module, builder, local_sym_tab
|
||||
)
|
||||
return None
|
||||
|
||||
@classmethod
|
||||
def get_struct_debug_info(cls, name):
|
||||
if cls._handler is None:
|
||||
return False
|
||||
return cls._handler.get_struct_debug_info(name)
|
||||
|
||||
@classmethod
|
||||
def is_vmlinux_struct(cls, name):
|
||||
"""Check if a name refers to a vmlinux struct"""
|
||||
if cls._handler is None:
|
||||
return False
|
||||
return cls._handler.is_vmlinux_struct(name)
|
||||
|
||||
@classmethod
|
||||
def get_struct_type(cls, name):
|
||||
"""Try to handle a struct name as vmlinux struct"""
|
||||
if cls._handler is None:
|
||||
return None
|
||||
return cls._handler.get_vmlinux_struct_type(name)
|
||||
|
||||
@classmethod
|
||||
def has_field(cls, vmlinux_struct_name, field_name):
|
||||
"""Check if a vmlinux struct has a specific field"""
|
||||
if cls._handler is None:
|
||||
return False
|
||||
return cls._handler.has_field(vmlinux_struct_name, field_name)
|
||||
|
||||
@classmethod
|
||||
def get_field_type(cls, vmlinux_struct_name, field_name):
|
||||
"""Get the type of a field in a vmlinux struct"""
|
||||
if cls._handler is None:
|
||||
return None
|
||||
assert isinstance(cls._handler, VmlinuxHandler)
|
||||
return cls._handler.get_field_type(vmlinux_struct_name, field_name)
|
||||
82
pythonbpf/functions/function_debug_info.py
Normal file
82
pythonbpf/functions/function_debug_info.py
Normal file
@ -0,0 +1,82 @@
|
||||
import ast
|
||||
|
||||
import llvmlite.ir as ir
|
||||
import logging
|
||||
from pythonbpf.debuginfo import DebugInfoGenerator
|
||||
from pythonbpf.expr import VmlinuxHandlerRegistry
|
||||
import ctypes
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def generate_function_debug_info(
|
||||
func_node: ast.FunctionDef, module: ir.Module, func: ir.Function
|
||||
):
|
||||
generator = DebugInfoGenerator(module)
|
||||
leading_argument = func_node.args.args[0]
|
||||
leading_argument_name = leading_argument.arg
|
||||
annotation = leading_argument.annotation
|
||||
if func_node.returns is None:
|
||||
# TODO: should check if this logic is consistent with function return type handling elsewhere
|
||||
return_type = ctypes.c_int64()
|
||||
elif hasattr(func_node.returns, "id"):
|
||||
return_type = func_node.returns.id
|
||||
if return_type == "c_int32":
|
||||
return_type = generator.get_int32_type()
|
||||
elif return_type == "c_int64":
|
||||
return_type = generator.get_int64_type()
|
||||
elif return_type == "c_uint32":
|
||||
return_type = generator.get_uint32_type()
|
||||
elif return_type == "c_uint64":
|
||||
return_type = generator.get_uint64_type()
|
||||
else:
|
||||
logger.warning(
|
||||
"Return type should be int32, int64, uint32 or uint64 only. Falling back to int64"
|
||||
)
|
||||
return_type = generator.get_int64_type()
|
||||
else:
|
||||
return_type = ctypes.c_int64()
|
||||
# context processing
|
||||
if annotation is None:
|
||||
logger.warning("Type of context of function not found.")
|
||||
return
|
||||
if hasattr(annotation, "id"):
|
||||
ctype_name = annotation.id
|
||||
if ctype_name == "c_void_p":
|
||||
return
|
||||
elif ctype_name.startswith("ctypes"):
|
||||
raise SyntaxError(
|
||||
"The first argument should always be a pointer to a struct or a void pointer"
|
||||
)
|
||||
context_debug_info = VmlinuxHandlerRegistry.get_struct_debug_info(annotation.id)
|
||||
|
||||
# Create pointer to context this must be created fresh for each function
|
||||
# to avoid circular reference issues when the same struct is used in multiple functions
|
||||
pointer_to_context_debug_info = generator.create_pointer_type(
|
||||
context_debug_info, 64
|
||||
)
|
||||
|
||||
# Create subroutine type - also fresh for each function
|
||||
subroutine_type = generator.create_subroutine_type(
|
||||
return_type, pointer_to_context_debug_info
|
||||
)
|
||||
|
||||
# Create local variable - fresh for each function with unique name
|
||||
context_local_variable = generator.create_local_variable_debug_info(
|
||||
leading_argument_name, 1, pointer_to_context_debug_info
|
||||
)
|
||||
|
||||
retained_nodes = [context_local_variable]
|
||||
logger.info(f"Generating debug info for function {func_node.name}")
|
||||
|
||||
# Create subprogram with is_distinct=True to ensure each function gets unique debug info
|
||||
subprogram_debug_info = generator.create_subprogram(
|
||||
func_node.name, subroutine_type, retained_nodes
|
||||
)
|
||||
generator.add_scope_to_local_variable(
|
||||
context_local_variable, subprogram_debug_info
|
||||
)
|
||||
func.set_metadata("dbg", subprogram_debug_info)
|
||||
|
||||
else:
|
||||
logger.error(f"Invalid annotation type for argument '{leading_argument_name}'")
|
||||
@ -7,7 +7,12 @@ from pythonbpf.helper import (
|
||||
reset_scratch_pool,
|
||||
)
|
||||
from pythonbpf.type_deducer import ctypes_to_ir
|
||||
from pythonbpf.expr import eval_expr, handle_expr, convert_to_bool
|
||||
from pythonbpf.expr import (
|
||||
eval_expr,
|
||||
handle_expr,
|
||||
convert_to_bool,
|
||||
VmlinuxHandlerRegistry,
|
||||
)
|
||||
from pythonbpf.assign_pass import (
|
||||
handle_variable_assignment,
|
||||
handle_struct_field_assignment,
|
||||
@ -16,8 +21,9 @@ from pythonbpf.allocation_pass import (
|
||||
handle_assign_allocation,
|
||||
allocate_temp_pool,
|
||||
create_targets_and_rvals,
|
||||
LocalSymbol,
|
||||
)
|
||||
|
||||
from .function_debug_info import generate_function_debug_info
|
||||
from .return_utils import handle_none_return, handle_xdp_return, is_xdp_name
|
||||
from .function_metadata import get_probe_string, is_global_function, infer_return_type
|
||||
|
||||
@ -310,7 +316,13 @@ def process_stmt(
|
||||
|
||||
|
||||
def process_func_body(
|
||||
module, builder, func_node, func, ret_type, map_sym_tab, structs_sym_tab
|
||||
module,
|
||||
builder,
|
||||
func_node,
|
||||
func,
|
||||
ret_type,
|
||||
map_sym_tab,
|
||||
structs_sym_tab,
|
||||
):
|
||||
"""Process the body of a bpf function"""
|
||||
# TODO: A lot. We just have print -> bpf_trace_printk for now
|
||||
@ -318,6 +330,28 @@ def process_func_body(
|
||||
|
||||
local_sym_tab = {}
|
||||
|
||||
# Add the context parameter (first function argument) to the local symbol table
|
||||
if func_node.args.args and len(func_node.args.args) > 0:
|
||||
context_arg = func_node.args.args[0]
|
||||
context_name = context_arg.arg
|
||||
|
||||
if hasattr(context_arg, "annotation") and context_arg.annotation:
|
||||
if isinstance(context_arg.annotation, ast.Name):
|
||||
context_type_name = context_arg.annotation.id
|
||||
elif isinstance(context_arg.annotation, ast.Attribute):
|
||||
context_type_name = context_arg.annotation.attr
|
||||
else:
|
||||
raise TypeError(
|
||||
f"Unsupported annotation type: {ast.dump(context_arg.annotation)}"
|
||||
)
|
||||
if VmlinuxHandlerRegistry.is_vmlinux_struct(context_type_name):
|
||||
resolved_type = VmlinuxHandlerRegistry.get_struct_type(
|
||||
context_type_name
|
||||
)
|
||||
context_type = LocalSymbol(None, None, resolved_type)
|
||||
local_sym_tab[context_name] = context_type
|
||||
logger.info(f"Added argument '{context_name}' to local symbol table")
|
||||
|
||||
# pre-allocate dynamic variables
|
||||
local_sym_tab = allocate_mem(
|
||||
module,
|
||||
@ -368,7 +402,7 @@ def process_bpf_chunk(func_node, module, return_type, map_sym_tab, structs_sym_t
|
||||
func.linkage = "dso_local"
|
||||
func.attributes.add("nounwind")
|
||||
func.attributes.add("noinline")
|
||||
func.attributes.add("optnone")
|
||||
# func.attributes.add("optnone")
|
||||
|
||||
if func_node.args.args:
|
||||
# Only look at the first argument for now
|
||||
@ -383,7 +417,13 @@ def process_bpf_chunk(func_node, module, return_type, map_sym_tab, structs_sym_t
|
||||
builder = ir.IRBuilder(block)
|
||||
|
||||
process_func_body(
|
||||
module, builder, func_node, func, ret_type, map_sym_tab, structs_sym_tab
|
||||
module,
|
||||
builder,
|
||||
func_node,
|
||||
func,
|
||||
ret_type,
|
||||
map_sym_tab,
|
||||
structs_sym_tab,
|
||||
)
|
||||
return func
|
||||
|
||||
@ -400,7 +440,7 @@ def func_proc(tree, module, chunks, map_sym_tab, structs_sym_tab):
|
||||
func_type = get_probe_string(func_node)
|
||||
logger.info(f"Found probe_string of {func_node.name}: {func_type}")
|
||||
|
||||
process_bpf_chunk(
|
||||
func = process_bpf_chunk(
|
||||
func_node,
|
||||
module,
|
||||
ctypes_to_ir(infer_return_type(func_node)),
|
||||
@ -408,6 +448,9 @@ def func_proc(tree, module, chunks, map_sym_tab, structs_sym_tab):
|
||||
structs_sym_tab,
|
||||
)
|
||||
|
||||
logger.info(f"Generating Debug Info for Function {func_node.name}")
|
||||
generate_function_debug_info(func_node, module, func)
|
||||
|
||||
|
||||
# TODO: WIP, for string assignment to fixed-size arrays
|
||||
def assign_string_to_array(builder, target_array_ptr, source_string_ptr, array_length):
|
||||
|
||||
@ -78,9 +78,9 @@ def bpf_map_lookup_elem_emitter(
|
||||
map_void_ptr = builder.bitcast(map_ptr, ir.PointerType())
|
||||
|
||||
# TODO: I have changed the return type to i64*, as we are
|
||||
# allocating space for that type in allocate_mem. This is
|
||||
# temporary, and we will honour other widths later. But this
|
||||
# allows us to have cool binary ops on the returned value.
|
||||
# allocating space for that type in allocate_mem. This is
|
||||
# temporary, and we will honour other widths later. But this
|
||||
# allows us to have cool binary ops on the returned value.
|
||||
fn_type = ir.FunctionType(
|
||||
ir.PointerType(ir.IntType(64)), # Return type: void*
|
||||
[ir.PointerType(), ir.PointerType()], # Args: (void*, void*)
|
||||
|
||||
@ -236,7 +236,7 @@ def get_char_array_ptr_and_size(buf_arg, builder, local_sym_tab, struct_sym_tab)
|
||||
var_ptr = local_sym_tab[var_name].var
|
||||
var_type = local_sym_tab[var_name].ir_type
|
||||
|
||||
if not isinstance(var_type, ir.PointerType) and not isinstance(
|
||||
if not isinstance(var_type, ir.PointerType) or not isinstance(
|
||||
var_type.pointee, ir.IntType(8)
|
||||
):
|
||||
raise ValueError("Expected str ptr variable")
|
||||
|
||||
@ -3,6 +3,7 @@ import logging
|
||||
|
||||
from llvmlite import ir
|
||||
from pythonbpf.expr import eval_expr, get_base_type_and_depth, deref_to_depth
|
||||
from pythonbpf.expr.vmlinux_registry import VmlinuxHandlerRegistry
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
@ -108,6 +109,16 @@ def _process_name_in_fval(name_node, fmt_parts, exprs, local_sym_tab):
|
||||
if local_sym_tab and name_node.id in local_sym_tab:
|
||||
_, var_type, tmp = local_sym_tab[name_node.id]
|
||||
_populate_fval(var_type, name_node, fmt_parts, exprs)
|
||||
else:
|
||||
# Try to resolve through vmlinux registry if not in local symbol table
|
||||
result = VmlinuxHandlerRegistry.handle_name(name_node.id)
|
||||
if result:
|
||||
val, var_type = result
|
||||
_populate_fval(var_type, name_node, fmt_parts, exprs)
|
||||
else:
|
||||
raise ValueError(
|
||||
f"Variable '{name_node.id}' not found in symbol table or vmlinux"
|
||||
)
|
||||
|
||||
|
||||
def _process_attr_in_fval(attr_node, fmt_parts, exprs, local_sym_tab, struct_sym_tab):
|
||||
|
||||
15
pythonbpf/local_symbol.py
Normal file
15
pythonbpf/local_symbol.py
Normal file
@ -0,0 +1,15 @@
|
||||
import llvmlite.ir as ir
|
||||
from dataclasses import dataclass
|
||||
from typing import Any
|
||||
|
||||
|
||||
@dataclass
|
||||
class LocalSymbol:
|
||||
var: ir.AllocaInstr
|
||||
ir_type: ir.Type
|
||||
metadata: Any = None
|
||||
|
||||
def __iter__(self):
|
||||
yield self.var
|
||||
yield self.ir_type
|
||||
yield self.metadata
|
||||
@ -6,6 +6,8 @@ from llvmlite import ir
|
||||
from .maps_utils import MapProcessorRegistry
|
||||
from .map_types import BPFMapType
|
||||
from .map_debug_info import create_map_debug_info, create_ringbuf_debug_info
|
||||
from pythonbpf.expr.vmlinux_registry import VmlinuxHandlerRegistry
|
||||
|
||||
|
||||
logger: Logger = logging.getLogger(__name__)
|
||||
|
||||
@ -51,7 +53,7 @@ def _parse_map_params(rval, expected_args=None):
|
||||
"""Parse map parameters from call arguments and keywords."""
|
||||
|
||||
params = {}
|
||||
|
||||
handler = VmlinuxHandlerRegistry.get_handler()
|
||||
# Parse positional arguments
|
||||
if expected_args:
|
||||
for i, arg_name in enumerate(expected_args):
|
||||
@ -65,7 +67,12 @@ def _parse_map_params(rval, expected_args=None):
|
||||
# Parse keyword arguments (override positional)
|
||||
for keyword in rval.keywords:
|
||||
if isinstance(keyword.value, ast.Name):
|
||||
params[keyword.arg] = keyword.value.id
|
||||
name = keyword.value.id
|
||||
if handler and handler.is_vmlinux_enum(name):
|
||||
result = handler.get_vmlinux_enum_value(name)
|
||||
params[keyword.arg] = result if result is not None else name
|
||||
else:
|
||||
params[keyword.arg] = name
|
||||
elif isinstance(keyword.value, ast.Constant):
|
||||
params[keyword.arg] = keyword.value.value
|
||||
|
||||
|
||||
@ -13,6 +13,11 @@ mapping = {
|
||||
"c_float": ir.FloatType(),
|
||||
"c_double": ir.DoubleType(),
|
||||
"c_void_p": ir.IntType(64),
|
||||
"c_long": ir.IntType(64),
|
||||
"c_ulong": ir.IntType(64),
|
||||
"c_longlong": ir.IntType(64),
|
||||
"c_uint": ir.IntType(32),
|
||||
"c_int": ir.IntType(32),
|
||||
# Not so sure about this one
|
||||
"str": ir.PointerType(ir.IntType(8)),
|
||||
}
|
||||
|
||||
@ -7,6 +7,8 @@ def trace_pipe():
|
||||
subprocess.run(["cat", "/sys/kernel/tracing/trace_pipe"])
|
||||
except KeyboardInterrupt:
|
||||
print("Tracing stopped.")
|
||||
except (FileNotFoundError, PermissionError) as e:
|
||||
print(f"Error accessing trace_pipe: {e}. Try running as root.")
|
||||
|
||||
|
||||
def trace_fields():
|
||||
|
||||
36
pythonbpf/vmlinux_parser/assignment_info.py
Normal file
36
pythonbpf/vmlinux_parser/assignment_info.py
Normal file
@ -0,0 +1,36 @@
|
||||
from enum import Enum, auto
|
||||
from typing import Any, Dict, List, Optional
|
||||
from dataclasses import dataclass
|
||||
import llvmlite.ir as ir
|
||||
|
||||
from pythonbpf.vmlinux_parser.dependency_node import Field
|
||||
|
||||
|
||||
class AssignmentType(Enum):
|
||||
CONSTANT = auto()
|
||||
STRUCT = auto()
|
||||
ARRAY = auto() # probably won't be used
|
||||
FUNCTION_POINTER = auto()
|
||||
POINTER = auto() # again, probably won't be used
|
||||
|
||||
|
||||
@dataclass
|
||||
class FunctionSignature:
|
||||
return_type: str
|
||||
param_types: List[str]
|
||||
varargs: bool
|
||||
|
||||
|
||||
# Thew name of the assignment will be in the dict that uses this class
|
||||
@dataclass
|
||||
class AssignmentInfo:
|
||||
value_type: AssignmentType
|
||||
python_type: type
|
||||
value: Optional[Any]
|
||||
pointer_level: Optional[int]
|
||||
signature: Optional[FunctionSignature] # For function pointers
|
||||
# The key of the dict is the name of the field.
|
||||
# Value is a tuple that contains the global variable representing that field
|
||||
# along with all the information about that field as a Field type.
|
||||
members: Optional[Dict[str, tuple[ir.GlobalVariable, Field]]] # For structs.
|
||||
debug_info: Any
|
||||
@ -1,6 +1,7 @@
|
||||
import logging
|
||||
from functools import lru_cache
|
||||
import importlib
|
||||
|
||||
from .dependency_handler import DependencyHandler
|
||||
from .dependency_node import DependencyNode
|
||||
import ctypes
|
||||
@ -15,7 +16,11 @@ def get_module_symbols(module_name: str):
|
||||
return [name for name in dir(imported_module)], imported_module
|
||||
|
||||
|
||||
def process_vmlinux_class(node, llvm_module, handler: DependencyHandler):
|
||||
def process_vmlinux_class(
|
||||
node,
|
||||
llvm_module,
|
||||
handler: DependencyHandler,
|
||||
):
|
||||
symbols_in_module, imported_module = get_module_symbols("vmlinux")
|
||||
if node.name in symbols_in_module:
|
||||
vmlinux_type = getattr(imported_module, node.name)
|
||||
@ -25,7 +30,10 @@ def process_vmlinux_class(node, llvm_module, handler: DependencyHandler):
|
||||
|
||||
|
||||
def process_vmlinux_post_ast(
|
||||
elem_type_class, llvm_handler, handler: DependencyHandler, processing_stack=None
|
||||
elem_type_class,
|
||||
llvm_handler,
|
||||
handler: DependencyHandler,
|
||||
processing_stack=None,
|
||||
):
|
||||
# Initialize processing stack on first call
|
||||
if processing_stack is None:
|
||||
@ -46,7 +54,7 @@ def process_vmlinux_post_ast(
|
||||
logger.debug(f"Node {current_symbol_name} already processed and ready")
|
||||
return True
|
||||
|
||||
# XXX:Check it's use. It's probably not being used.
|
||||
# XXX:Check its use. It's probably not being used.
|
||||
if current_symbol_name in processing_stack:
|
||||
logger.debug(
|
||||
f"Dependency already in processing stack for {current_symbol_name}, skipping"
|
||||
@ -60,6 +68,10 @@ def process_vmlinux_post_ast(
|
||||
pass
|
||||
else:
|
||||
new_dep_node = DependencyNode(name=current_symbol_name)
|
||||
|
||||
# elem_type_class is the actual vmlinux struct/class
|
||||
new_dep_node.set_ctype_struct(elem_type_class)
|
||||
|
||||
handler.add_node(new_dep_node)
|
||||
class_obj = getattr(imported_module, current_symbol_name)
|
||||
# Inspect the class fields
|
||||
@ -94,12 +106,47 @@ def process_vmlinux_post_ast(
|
||||
[elem_type, elem_bitfield_size] = elem_temp_list
|
||||
local_module_name = getattr(elem_type, "__module__", None)
|
||||
new_dep_node.add_field(elem_name, elem_type, ready=False)
|
||||
|
||||
if local_module_name == ctypes.__name__:
|
||||
# TODO: need to process pointer to ctype and also CFUNCTYPES here recursively. Current processing is a single dereference
|
||||
new_dep_node.set_field_bitfield_size(elem_name, elem_bitfield_size)
|
||||
new_dep_node.set_field_ready(elem_name, is_ready=True)
|
||||
logger.debug(
|
||||
f"Field {elem_name} is direct ctypes type: {elem_type}"
|
||||
)
|
||||
|
||||
# Process pointer to ctype
|
||||
if isinstance(elem_type, type) and issubclass(
|
||||
elem_type, ctypes._Pointer
|
||||
):
|
||||
# Get the pointed-to type
|
||||
pointed_type = elem_type._type_
|
||||
logger.debug(f"Found pointer to type: {pointed_type}")
|
||||
new_dep_node.set_field_containing_type(elem_name, pointed_type)
|
||||
new_dep_node.set_field_ctype_complex_type(
|
||||
elem_name, ctypes._Pointer
|
||||
)
|
||||
new_dep_node.set_field_ready(elem_name, is_ready=True)
|
||||
|
||||
# Process function pointers (CFUNCTYPE)
|
||||
elif hasattr(elem_type, "_restype_") and hasattr(
|
||||
elem_type, "_argtypes_"
|
||||
):
|
||||
# This is a CFUNCTYPE or similar
|
||||
logger.info(
|
||||
f"Function pointer detected for {elem_name} with return type {elem_type._restype_} and arguments {elem_type._argtypes_}"
|
||||
)
|
||||
# Set the field as ready but mark it with special handling
|
||||
new_dep_node.set_field_ctype_complex_type(
|
||||
elem_name, ctypes.CFUNCTYPE
|
||||
)
|
||||
new_dep_node.set_field_ready(elem_name, is_ready=True)
|
||||
logger.warning(
|
||||
"Blindly processing CFUNCTYPE ctypes to ensure compilation. Unsupported"
|
||||
)
|
||||
|
||||
else:
|
||||
# Regular ctype
|
||||
new_dep_node.set_field_ready(elem_name, is_ready=True)
|
||||
logger.debug(
|
||||
f"Field {elem_name} is direct ctypes type: {elem_type}"
|
||||
)
|
||||
elif local_module_name == "vmlinux":
|
||||
new_dep_node.set_field_bitfield_size(elem_name, elem_bitfield_size)
|
||||
logger.debug(
|
||||
@ -112,13 +159,21 @@ def process_vmlinux_post_ast(
|
||||
type_length = elem_type._length_
|
||||
|
||||
if containing_type.__module__ == "vmlinux":
|
||||
pass
|
||||
new_dep_node.add_dependent(
|
||||
elem_type._type_.__name__
|
||||
if hasattr(elem_type._type_, "__name__")
|
||||
else str(elem_type._type_)
|
||||
)
|
||||
elif containing_type.__module__ == ctypes.__name__:
|
||||
if isinstance(elem_type, type):
|
||||
if issubclass(elem_type, ctypes.Array):
|
||||
ctype_complex_type = ctypes.Array
|
||||
elif issubclass(elem_type, ctypes._Pointer):
|
||||
ctype_complex_type = ctypes._Pointer
|
||||
else:
|
||||
raise ImportError(
|
||||
"Non Array and Pointer type ctype imports not supported in current version"
|
||||
)
|
||||
else:
|
||||
raise TypeError("Unsupported ctypes subclass")
|
||||
else:
|
||||
@ -137,10 +192,35 @@ def process_vmlinux_post_ast(
|
||||
)
|
||||
new_dep_node.set_field_type(elem_name, elem_type)
|
||||
if containing_type.__module__ == "vmlinux":
|
||||
process_vmlinux_post_ast(
|
||||
containing_type, llvm_handler, handler, processing_stack
|
||||
containing_type_name = (
|
||||
containing_type.__name__
|
||||
if hasattr(containing_type, "__name__")
|
||||
else str(containing_type)
|
||||
)
|
||||
new_dep_node.set_field_ready(elem_name, True)
|
||||
|
||||
# Check for self-reference or already processed
|
||||
if containing_type_name == current_symbol_name:
|
||||
# Self-referential pointer
|
||||
logger.debug(
|
||||
f"Self-referential pointer in {current_symbol_name}.{elem_name}"
|
||||
)
|
||||
new_dep_node.set_field_ready(elem_name, True)
|
||||
elif handler.has_node(containing_type_name):
|
||||
# Already processed
|
||||
logger.debug(
|
||||
f"Reusing already processed {containing_type_name}"
|
||||
)
|
||||
new_dep_node.set_field_ready(elem_name, True)
|
||||
else:
|
||||
# Process recursively - THIS WAS MISSING
|
||||
new_dep_node.add_dependent(containing_type_name)
|
||||
process_vmlinux_post_ast(
|
||||
containing_type,
|
||||
llvm_handler,
|
||||
handler,
|
||||
processing_stack,
|
||||
)
|
||||
new_dep_node.set_field_ready(elem_name, True)
|
||||
elif containing_type.__module__ == ctypes.__name__:
|
||||
logger.debug(f"Processing ctype internal{containing_type}")
|
||||
new_dep_node.set_field_ready(elem_name, True)
|
||||
@ -149,8 +229,16 @@ def process_vmlinux_post_ast(
|
||||
"Module not supported in recursive resolution"
|
||||
)
|
||||
else:
|
||||
new_dep_node.add_dependent(
|
||||
elem_type.__name__
|
||||
if hasattr(elem_type, "__name__")
|
||||
else str(elem_type)
|
||||
)
|
||||
process_vmlinux_post_ast(
|
||||
elem_type, llvm_handler, handler, processing_stack
|
||||
elem_type,
|
||||
llvm_handler,
|
||||
handler,
|
||||
processing_stack,
|
||||
)
|
||||
new_dep_node.set_field_ready(elem_name, True)
|
||||
else:
|
||||
@ -161,7 +249,7 @@ def process_vmlinux_post_ast(
|
||||
else:
|
||||
raise ImportError("UNSUPPORTED Module")
|
||||
|
||||
logging.info(
|
||||
logger.info(
|
||||
f"{current_symbol_name} processed and handler readiness {handler.is_ready}"
|
||||
)
|
||||
return True
|
||||
|
||||
@ -147,3 +147,27 @@ class DependencyHandler:
|
||||
int: The number of nodes
|
||||
"""
|
||||
return len(self._nodes)
|
||||
|
||||
def __getitem__(self, name: str) -> DependencyNode:
|
||||
"""
|
||||
Get a node by name using dictionary-style access.
|
||||
|
||||
Args:
|
||||
name: The name of the node to retrieve
|
||||
|
||||
Returns:
|
||||
DependencyNode: The node with the given name
|
||||
|
||||
Raises:
|
||||
KeyError: If no node with the given name exists
|
||||
|
||||
Example:
|
||||
node = handler["some-dep_node_name"]
|
||||
"""
|
||||
if name not in self._nodes:
|
||||
raise KeyError(f"No node with name '{name}' found")
|
||||
return self._nodes[name]
|
||||
|
||||
@property
|
||||
def nodes(self):
|
||||
return self._nodes
|
||||
|
||||
@ -1,5 +1,6 @@
|
||||
from dataclasses import dataclass, field
|
||||
from typing import Dict, Any, Optional
|
||||
import ctypes
|
||||
|
||||
|
||||
# TODO: FIX THE FUCKING TYPE NAME CONVENTION.
|
||||
@ -13,9 +14,35 @@ class Field:
|
||||
containing_type: Optional[Any]
|
||||
type_size: Optional[int]
|
||||
bitfield_size: Optional[int]
|
||||
offset: int
|
||||
value: Any = None
|
||||
ready: bool = False
|
||||
|
||||
def __hash__(self):
|
||||
"""
|
||||
Create a hash based on the immutable attributes that define this field's identity.
|
||||
This allows Field objects to be used as dictionary keys.
|
||||
"""
|
||||
# Use a tuple of the fields that uniquely identify this field
|
||||
identity = (
|
||||
self.name,
|
||||
id(self.type), # Use id for non-hashable types
|
||||
id(self.ctype_complex_type) if self.ctype_complex_type else None,
|
||||
id(self.containing_type) if self.containing_type else None,
|
||||
self.type_size,
|
||||
self.bitfield_size,
|
||||
self.offset,
|
||||
self.value if self.value else None,
|
||||
)
|
||||
return hash(identity)
|
||||
|
||||
def __eq__(self, other):
|
||||
"""
|
||||
Define equality consistent with the hash function.
|
||||
Two fields are equal if they have they are the same
|
||||
"""
|
||||
return self is other
|
||||
|
||||
def set_ready(self, is_ready: bool = True) -> None:
|
||||
"""Set the readiness state of this field."""
|
||||
self.ready = is_ready
|
||||
@ -60,6 +87,10 @@ class Field:
|
||||
if mark_ready:
|
||||
self.ready = True
|
||||
|
||||
def set_offset(self, offset: int) -> None:
|
||||
"""Set the offset of this field"""
|
||||
self.offset = offset
|
||||
|
||||
|
||||
@dataclass
|
||||
class DependencyNode:
|
||||
@ -106,8 +137,11 @@ class DependencyNode:
|
||||
"""
|
||||
|
||||
name: str
|
||||
depends_on: Optional[list[str]] = None
|
||||
fields: Dict[str, Field] = field(default_factory=dict)
|
||||
_ready_cache: Optional[bool] = field(default=None, repr=False)
|
||||
current_offset: int = 0
|
||||
ctype_struct: Optional[Any] = field(default=None, repr=False)
|
||||
|
||||
def add_field(
|
||||
self,
|
||||
@ -119,8 +153,11 @@ class DependencyNode:
|
||||
ctype_complex_type: Optional[int] = None,
|
||||
bitfield_size: Optional[int] = None,
|
||||
ready: bool = False,
|
||||
offset: int = 0,
|
||||
) -> None:
|
||||
"""Add a field to the node with an optional initial value and readiness state."""
|
||||
if self.depends_on is None:
|
||||
self.depends_on = []
|
||||
self.fields[name] = Field(
|
||||
name=name,
|
||||
type=field_type,
|
||||
@ -130,10 +167,21 @@ class DependencyNode:
|
||||
type_size=type_size,
|
||||
ctype_complex_type=ctype_complex_type,
|
||||
bitfield_size=bitfield_size,
|
||||
offset=offset,
|
||||
)
|
||||
# Invalidate readiness cache
|
||||
self._ready_cache = None
|
||||
|
||||
def set_ctype_struct(self, ctype_struct: Any) -> None:
|
||||
"""Set the ctypes structure for automatic offset calculation."""
|
||||
self.ctype_struct = ctype_struct
|
||||
|
||||
def __sizeof__(self):
|
||||
# If we have a ctype_struct, use its size
|
||||
if self.ctype_struct is not None:
|
||||
return ctypes.sizeof(self.ctype_struct)
|
||||
return self.current_offset
|
||||
|
||||
def get_field(self, name: str) -> Field:
|
||||
"""Get a field by name."""
|
||||
return self.fields[name]
|
||||
@ -200,15 +248,112 @@ class DependencyNode:
|
||||
# Invalidate readiness cache
|
||||
self._ready_cache = None
|
||||
|
||||
def set_field_ready(self, name: str, is_ready: bool = False) -> None:
|
||||
def set_field_ready(
|
||||
self,
|
||||
name: str,
|
||||
is_ready: bool = False,
|
||||
size_of_containing_type: Optional[int] = None,
|
||||
) -> None:
|
||||
"""Mark a field as ready or not ready."""
|
||||
if name not in self.fields:
|
||||
raise KeyError(f"Field '{name}' does not exist in node '{self.name}'")
|
||||
|
||||
self.fields[name].set_ready(is_ready)
|
||||
|
||||
# Use ctypes built-in offset if available
|
||||
if self.ctype_struct is not None:
|
||||
try:
|
||||
self.fields[name].set_offset(getattr(self.ctype_struct, name).offset)
|
||||
except AttributeError:
|
||||
# Fallback to manual calculation if field not found in ctype_struct
|
||||
self.fields[name].set_offset(self.current_offset)
|
||||
self.current_offset += self._calculate_size(
|
||||
name, size_of_containing_type
|
||||
)
|
||||
else:
|
||||
# Manual offset calculation when no ctype_struct is available
|
||||
self.fields[name].set_offset(self.current_offset)
|
||||
self.current_offset += self._calculate_size(name, size_of_containing_type)
|
||||
|
||||
# Invalidate readiness cache
|
||||
self._ready_cache = None
|
||||
|
||||
def _calculate_size(
|
||||
self, name: str, size_of_containing_type: Optional[int] = None
|
||||
) -> int:
|
||||
processing_field = self.fields[name]
|
||||
# size_of_field will be in bytes
|
||||
if processing_field.type.__module__ == ctypes.__name__:
|
||||
size_of_field = ctypes.sizeof(processing_field.type)
|
||||
return size_of_field
|
||||
elif processing_field.type.__module__ == "vmlinux":
|
||||
if processing_field.ctype_complex_type is not None:
|
||||
if issubclass(processing_field.ctype_complex_type, ctypes.Array):
|
||||
if processing_field.containing_type.__module__ == ctypes.__name__:
|
||||
if (
|
||||
processing_field.containing_type is not None
|
||||
and processing_field.type_size is not None
|
||||
):
|
||||
size_of_field = (
|
||||
ctypes.sizeof(processing_field.containing_type)
|
||||
* processing_field.type_size
|
||||
)
|
||||
else:
|
||||
raise RuntimeError(
|
||||
f"{processing_field} has no containing_type or type_size"
|
||||
)
|
||||
return size_of_field
|
||||
elif processing_field.containing_type.__module__ == "vmlinux":
|
||||
if (
|
||||
size_of_containing_type is not None
|
||||
and processing_field.type_size is not None
|
||||
):
|
||||
size_of_field = (
|
||||
size_of_containing_type * processing_field.type_size
|
||||
)
|
||||
else:
|
||||
raise RuntimeError(
|
||||
f"{processing_field} has no containing_type or type_size"
|
||||
)
|
||||
return size_of_field
|
||||
elif issubclass(processing_field.ctype_complex_type, ctypes._Pointer):
|
||||
return ctypes.sizeof(ctypes.c_void_p)
|
||||
else:
|
||||
raise NotImplementedError(
|
||||
"This subclass of ctype not supported yet"
|
||||
)
|
||||
elif processing_field.type_size is not None:
|
||||
# Handle vmlinux types with type_size but no ctype_complex_type
|
||||
# This means it's a direct vmlinux struct field (not array/pointer wrapped)
|
||||
# The type_size should already contain the full size of the struct
|
||||
# But if there's a containing_type from vmlinux, we need that size
|
||||
if processing_field.containing_type is not None:
|
||||
if processing_field.containing_type.__module__ == "vmlinux":
|
||||
# For vmlinux containing types, we need the pre-calculated size
|
||||
if size_of_containing_type is not None:
|
||||
return size_of_containing_type * processing_field.type_size
|
||||
else:
|
||||
raise RuntimeError(
|
||||
f"Field {name}: vmlinux containing_type requires size_of_containing_type"
|
||||
)
|
||||
else:
|
||||
raise ModuleNotFoundError(
|
||||
f"Containing type module {processing_field.containing_type.__module__} not supported"
|
||||
)
|
||||
else:
|
||||
raise RuntimeError("Wrong type found with no containing type")
|
||||
else:
|
||||
# No ctype_complex_type and no type_size, must rely on size_of_containing_type
|
||||
if size_of_containing_type is None:
|
||||
raise RuntimeError(
|
||||
f"Size of containing type {size_of_containing_type} is None"
|
||||
)
|
||||
return size_of_containing_type
|
||||
|
||||
else:
|
||||
raise ModuleNotFoundError("Module is not supported for the operation")
|
||||
raise RuntimeError("control should not reach here")
|
||||
|
||||
@property
|
||||
def is_ready(self) -> bool:
|
||||
"""Check if the node is ready (all fields are ready)."""
|
||||
@ -235,3 +380,9 @@ class DependencyNode:
|
||||
def get_not_ready_fields(self) -> Dict[str, Field]:
|
||||
"""Get all fields that are marked as not ready."""
|
||||
return {name: elem for name, elem in self.fields.items() if not elem.ready}
|
||||
|
||||
def add_dependent(self, dep_type):
|
||||
if dep_type in self.depends_on:
|
||||
return
|
||||
else:
|
||||
self.depends_on.append(dep_type)
|
||||
|
||||
@ -1,17 +1,17 @@
|
||||
import ast
|
||||
import logging
|
||||
from typing import List, Tuple, Dict
|
||||
import importlib
|
||||
import inspect
|
||||
|
||||
from .assignment_info import AssignmentInfo, AssignmentType
|
||||
from .dependency_handler import DependencyHandler
|
||||
from .ir_generation import IRGenerator
|
||||
from .ir_gen import IRGenerator
|
||||
from .class_handler import process_vmlinux_class
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def detect_import_statement(tree: ast.AST) -> List[Tuple[str, ast.ImportFrom]]:
|
||||
def detect_import_statement(tree: ast.AST) -> list[tuple[str, ast.ImportFrom]]:
|
||||
"""
|
||||
Parse AST and detect import statements from vmlinux.
|
||||
|
||||
@ -82,23 +82,23 @@ def vmlinux_proc(tree: ast.AST, module):
|
||||
# initialise dependency handler
|
||||
handler = DependencyHandler()
|
||||
# initialise assignment dictionary of name to type
|
||||
assignments: Dict[str, type] = {}
|
||||
assignments: dict[str, AssignmentInfo] = {}
|
||||
|
||||
if not import_statements:
|
||||
logger.info("No vmlinux imports found")
|
||||
return
|
||||
return None
|
||||
|
||||
# Import vmlinux module directly
|
||||
try:
|
||||
vmlinux_mod = importlib.import_module("vmlinux")
|
||||
except ImportError:
|
||||
logger.warning("Could not import vmlinux module")
|
||||
return
|
||||
return None
|
||||
|
||||
source_file = inspect.getsourcefile(vmlinux_mod)
|
||||
if source_file is None:
|
||||
logger.warning("Cannot find source for vmlinux module")
|
||||
return
|
||||
return None
|
||||
|
||||
with open(source_file, "r") as f:
|
||||
mod_ast = ast.parse(f.read(), filename=source_file)
|
||||
@ -128,8 +128,36 @@ def vmlinux_proc(tree: ast.AST, module):
|
||||
f"{imported_name} not found as ClassDef or Assign in vmlinux"
|
||||
)
|
||||
|
||||
IRGenerator(module, handler)
|
||||
IRGenerator(module, handler, assignments)
|
||||
return assignments
|
||||
|
||||
|
||||
def process_vmlinux_assign(node, module, assignments: Dict[str, type]):
|
||||
raise NotImplementedError("Assignment handling has not been implemented yet")
|
||||
def process_vmlinux_assign(node, module, assignments: dict[str, AssignmentInfo]):
|
||||
"""Process assignments from vmlinux module."""
|
||||
# Only handle single-target assignments
|
||||
if len(node.targets) == 1 and isinstance(node.targets[0], ast.Name):
|
||||
target_name = node.targets[0].id
|
||||
|
||||
# Handle constant value assignments
|
||||
if isinstance(node.value, ast.Constant):
|
||||
# Fixed: using proper TypedDict creation syntax with named arguments
|
||||
assignments[target_name] = AssignmentInfo(
|
||||
value_type=AssignmentType.CONSTANT,
|
||||
python_type=type(node.value.value),
|
||||
value=node.value.value,
|
||||
pointer_level=None,
|
||||
signature=None,
|
||||
members=None,
|
||||
debug_info=None,
|
||||
)
|
||||
logger.info(
|
||||
f"Added assignment: {target_name} = {node.value.value!r} of type {type(node.value.value)}"
|
||||
)
|
||||
|
||||
# Handle other assignment types that we may need to support
|
||||
else:
|
||||
logger.warning(
|
||||
f"Unsupported assignment type for {target_name}: {ast.dump(node.value)}"
|
||||
)
|
||||
else:
|
||||
raise ValueError("Not a simple assignment")
|
||||
|
||||
3
pythonbpf/vmlinux_parser/ir_gen/__init__.py
Normal file
3
pythonbpf/vmlinux_parser/ir_gen/__init__.py
Normal file
@ -0,0 +1,3 @@
|
||||
from .ir_generation import IRGenerator
|
||||
|
||||
__all__ = ["IRGenerator"]
|
||||
161
pythonbpf/vmlinux_parser/ir_gen/debug_info_gen.py
Normal file
161
pythonbpf/vmlinux_parser/ir_gen/debug_info_gen.py
Normal file
@ -0,0 +1,161 @@
|
||||
from pythonbpf.debuginfo import DebugInfoGenerator, dwarf_constants as dc
|
||||
from ..dependency_node import DependencyNode
|
||||
import ctypes
|
||||
import logging
|
||||
from typing import List, Any, Tuple
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def debug_info_generation(
|
||||
struct: DependencyNode,
|
||||
llvm_module,
|
||||
generated_debug_info: List[Tuple[DependencyNode, Any]],
|
||||
) -> Any:
|
||||
"""
|
||||
Generate DWARF debug information for a struct defined in a DependencyNode.
|
||||
|
||||
Args:
|
||||
struct: The dependency node containing struct information
|
||||
llvm_module: The LLVM module to add debug info to
|
||||
generated_debug_info: List of tuples (struct, debug_info) to track generated debug info
|
||||
|
||||
Returns:
|
||||
The generated global variable debug info
|
||||
"""
|
||||
# Set up debug info generator
|
||||
generator = DebugInfoGenerator(llvm_module)
|
||||
|
||||
# Check if debug info for this struct has already been generated
|
||||
for existing_struct, debug_info in generated_debug_info:
|
||||
if existing_struct.name == struct.name:
|
||||
return debug_info
|
||||
|
||||
# Process all fields and create members for the struct
|
||||
members = []
|
||||
for field_name, field in struct.fields.items():
|
||||
# Get appropriate debug type for this field
|
||||
field_type = _get_field_debug_type(
|
||||
field_name, field, generator, struct, generated_debug_info
|
||||
)
|
||||
# Create struct member with proper offset
|
||||
member = generator.create_struct_member_vmlinux(
|
||||
field_name, field_type, field.offset * 8
|
||||
)
|
||||
members.append(member)
|
||||
|
||||
if struct.name.startswith("struct_"):
|
||||
struct_name = struct.name.removeprefix("struct_")
|
||||
else:
|
||||
raise ValueError("Unions are not supported in the current version")
|
||||
# Create struct type with all members
|
||||
struct_type = generator.create_struct_type_with_name(
|
||||
struct_name, members, struct.__sizeof__() * 8, is_distinct=True
|
||||
)
|
||||
|
||||
return struct_type
|
||||
|
||||
|
||||
def _get_field_debug_type(
|
||||
field_name: str,
|
||||
field,
|
||||
generator: DebugInfoGenerator,
|
||||
parent_struct: DependencyNode,
|
||||
generated_debug_info: List[Tuple[DependencyNode, Any]],
|
||||
) -> tuple[Any, int]:
|
||||
"""
|
||||
Determine the appropriate debug type for a field based on its Python/ctypes type.
|
||||
|
||||
Args:
|
||||
field_name: Name of the field
|
||||
field: Field object containing type information
|
||||
generator: DebugInfoGenerator instance
|
||||
parent_struct: The parent struct containing this field
|
||||
generated_debug_info: List of already generated debug info
|
||||
|
||||
Returns:
|
||||
The debug info type for this field
|
||||
"""
|
||||
# Handle complex types (arrays, pointers)
|
||||
if field.ctype_complex_type is not None:
|
||||
if issubclass(field.ctype_complex_type, ctypes.Array):
|
||||
# Handle array types
|
||||
element_type, base_type_size = _get_basic_debug_type(
|
||||
field.containing_type, generator
|
||||
)
|
||||
return generator.create_array_type_vmlinux(
|
||||
(element_type, base_type_size * field.type_size), field.type_size
|
||||
), field.type_size * base_type_size
|
||||
elif issubclass(field.ctype_complex_type, ctypes._Pointer):
|
||||
# Handle pointer types
|
||||
pointee_type, _ = _get_basic_debug_type(field.containing_type, generator)
|
||||
return generator.create_pointer_type(pointee_type), 64
|
||||
|
||||
# Handle other vmlinux types (nested structs)
|
||||
if field.type.__module__ == "vmlinux":
|
||||
# If it's a struct from vmlinux, check if we've already generated debug info for it
|
||||
struct_name = field.type.__name__
|
||||
|
||||
# Look for existing debug info in the list
|
||||
for existing_struct, debug_info in generated_debug_info:
|
||||
if existing_struct.name == struct_name:
|
||||
# Use existing debug info
|
||||
return debug_info, existing_struct.__sizeof__()
|
||||
|
||||
# If not found, create a forward declaration
|
||||
# This will be completed when the actual struct is processed
|
||||
logger.warning("Forward declaration in struct created")
|
||||
forward_type = generator.create_struct_type([], 0, is_distinct=True)
|
||||
return forward_type, 0
|
||||
|
||||
# Handle basic C types
|
||||
return _get_basic_debug_type(field.type, generator)
|
||||
|
||||
|
||||
def _get_basic_debug_type(ctype, generator: DebugInfoGenerator) -> Any:
|
||||
"""
|
||||
Map a ctypes type to a DWARF debug type.
|
||||
|
||||
Args:
|
||||
ctype: A ctypes type or Python type
|
||||
generator: DebugInfoGenerator instance
|
||||
|
||||
Returns:
|
||||
The corresponding debug type
|
||||
"""
|
||||
# Map ctypes to debug info types
|
||||
if ctype == ctypes.c_char or ctype == ctypes.c_byte:
|
||||
return generator.get_basic_type("char", 8, dc.DW_ATE_signed_char), 8
|
||||
elif ctype == ctypes.c_ubyte or ctype == ctypes.c_uint8:
|
||||
return generator.get_basic_type("unsigned char", 8, dc.DW_ATE_unsigned_char), 8
|
||||
elif ctype == ctypes.c_short or ctype == ctypes.c_int16:
|
||||
return generator.get_basic_type("short", 16, dc.DW_ATE_signed), 16
|
||||
elif ctype == ctypes.c_ushort or ctype == ctypes.c_uint16:
|
||||
return generator.get_basic_type("unsigned short", 16, dc.DW_ATE_unsigned), 16
|
||||
elif ctype == ctypes.c_int or ctype == ctypes.c_int32:
|
||||
return generator.get_basic_type("int", 32, dc.DW_ATE_signed), 32
|
||||
elif ctype == ctypes.c_uint or ctype == ctypes.c_uint32:
|
||||
return generator.get_basic_type("unsigned int", 32, dc.DW_ATE_unsigned), 32
|
||||
elif ctype == ctypes.c_long:
|
||||
return generator.get_basic_type("long", 64, dc.DW_ATE_signed), 64
|
||||
elif ctype == ctypes.c_ulong:
|
||||
return generator.get_basic_type("unsigned long", 64, dc.DW_ATE_unsigned), 64
|
||||
elif ctype == ctypes.c_longlong or ctype == ctypes.c_int64:
|
||||
return generator.get_basic_type("long long", 64, dc.DW_ATE_signed), 64
|
||||
elif ctype == ctypes.c_ulonglong or ctype == ctypes.c_uint64:
|
||||
return generator.get_basic_type(
|
||||
"unsigned long long", 64, dc.DW_ATE_unsigned
|
||||
), 64
|
||||
elif ctype == ctypes.c_float:
|
||||
return generator.get_basic_type("float", 32, dc.DW_ATE_float), 32
|
||||
elif ctype == ctypes.c_double:
|
||||
return generator.get_basic_type("double", 64, dc.DW_ATE_float), 64
|
||||
elif ctype == ctypes.c_bool:
|
||||
return generator.get_basic_type("bool", 8, dc.DW_ATE_boolean), 8
|
||||
elif ctype == ctypes.c_char_p:
|
||||
char_type = generator.get_basic_type("char", 8, dc.DW_ATE_signed_char), 8
|
||||
return generator.create_pointer_type(char_type)
|
||||
elif ctype == ctypes.c_void_p:
|
||||
return generator.create_pointer_type(None), 64
|
||||
else:
|
||||
return generator.get_uint64_type(), 64
|
||||
225
pythonbpf/vmlinux_parser/ir_gen/ir_generation.py
Normal file
225
pythonbpf/vmlinux_parser/ir_gen/ir_generation.py
Normal file
@ -0,0 +1,225 @@
|
||||
import ctypes
|
||||
import logging
|
||||
|
||||
from ..assignment_info import AssignmentInfo, AssignmentType
|
||||
from ..dependency_handler import DependencyHandler
|
||||
from .debug_info_gen import debug_info_generation
|
||||
from ..dependency_node import DependencyNode
|
||||
import llvmlite.ir as ir
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class IRGenerator:
|
||||
# get the assignments dict and add this stuff to it.
|
||||
def __init__(self, llvm_module, handler: DependencyHandler, assignments):
|
||||
self.llvm_module = llvm_module
|
||||
self.handler: DependencyHandler = handler
|
||||
self.generated: list[str] = []
|
||||
self.generated_debug_info: list = []
|
||||
# Use struct_name and field_name as key instead of Field object
|
||||
self.generated_field_names: dict[str, dict[str, ir.GlobalVariable]] = {}
|
||||
self.assignments: dict[str, AssignmentInfo] = assignments
|
||||
if not handler.is_ready:
|
||||
raise ImportError(
|
||||
"Semantic analysis of vmlinux imports failed. Cannot generate IR"
|
||||
)
|
||||
for struct in handler:
|
||||
self.struct_processor(struct)
|
||||
|
||||
def struct_processor(self, struct, processing_stack=None):
|
||||
# Initialize processing stack on first call
|
||||
if processing_stack is None:
|
||||
processing_stack = set()
|
||||
|
||||
# If already generated, skip
|
||||
if struct.name in self.generated:
|
||||
return
|
||||
|
||||
# Detect circular dependency
|
||||
if struct.name in processing_stack:
|
||||
logger.info(
|
||||
f"Circular dependency detected for {struct.name}, skipping recursive processing"
|
||||
)
|
||||
# For circular dependencies, we can either:
|
||||
# 1. Use forward declarations (opaque pointers)
|
||||
# 2. Mark as incomplete and process later
|
||||
# 3. Generate a placeholder type
|
||||
# Here we'll just skip and let it be processed in its own call
|
||||
return
|
||||
|
||||
logger.info(f"IR generating for {struct.name}")
|
||||
|
||||
# Add to processing stack before processing dependencies
|
||||
processing_stack.add(struct.name)
|
||||
|
||||
try:
|
||||
# Process all dependencies first
|
||||
if struct.depends_on is None:
|
||||
pass
|
||||
else:
|
||||
for dependency in struct.depends_on:
|
||||
if dependency not in self.generated:
|
||||
# Check if dependency exists in handler
|
||||
if dependency in self.handler.nodes:
|
||||
dep_node_from_dependency = self.handler[dependency]
|
||||
# Pass the processing_stack down to track circular refs
|
||||
self.struct_processor(
|
||||
dep_node_from_dependency, processing_stack
|
||||
)
|
||||
else:
|
||||
raise RuntimeError(
|
||||
f"Warning: Dependency {dependency} not found in handler"
|
||||
)
|
||||
|
||||
# Generate IR first to populate field names
|
||||
struct_debug_info = self.gen_ir(struct, self.generated_debug_info)
|
||||
self.generated_debug_info.append((struct, struct_debug_info))
|
||||
|
||||
# Fill the assignments dictionary with struct information
|
||||
if struct.name not in self.assignments:
|
||||
# Create a members dictionary for AssignmentInfo
|
||||
members_dict = {}
|
||||
for field_name, field in struct.fields.items():
|
||||
# Get the generated field name from our dictionary, or use field_name if not found
|
||||
if (
|
||||
struct.name in self.generated_field_names
|
||||
and field_name in self.generated_field_names[struct.name]
|
||||
):
|
||||
field_global_variable = self.generated_field_names[struct.name][
|
||||
field_name
|
||||
]
|
||||
members_dict[field_name] = (field_global_variable, field)
|
||||
else:
|
||||
raise ValueError(
|
||||
f"llvm global name not found for struct field {field_name}"
|
||||
)
|
||||
# members_dict[field_name] = (field_name, field)
|
||||
|
||||
# Add struct to assignments dictionary
|
||||
self.assignments[struct.name] = AssignmentInfo(
|
||||
value_type=AssignmentType.STRUCT,
|
||||
python_type=struct.ctype_struct,
|
||||
value=None,
|
||||
pointer_level=None,
|
||||
signature=None,
|
||||
members=members_dict,
|
||||
debug_info=struct_debug_info,
|
||||
)
|
||||
logger.info(f"Added struct assignment info for {struct.name}")
|
||||
|
||||
self.generated.append(struct.name)
|
||||
|
||||
finally:
|
||||
# Remove from processing stack after we're done
|
||||
processing_stack.discard(struct.name)
|
||||
|
||||
def gen_ir(self, struct, generated_debug_info):
|
||||
# TODO: we add the btf_ama attribute by monkey patching in the end of compilation, but once llvmlite
|
||||
# accepts our issue, we will resort to normal accessed attribute based attribute addition
|
||||
# currently we generate all possible field accesses for CO-RE and put into the assignment table
|
||||
debug_info = debug_info_generation(
|
||||
struct, self.llvm_module, generated_debug_info
|
||||
)
|
||||
field_index = 0
|
||||
|
||||
# Make sure the struct has an entry in our field names dictionary
|
||||
if struct.name not in self.generated_field_names:
|
||||
self.generated_field_names[struct.name] = {}
|
||||
|
||||
for field_name, field in struct.fields.items():
|
||||
# does not take arrays and similar types into consideration yet.
|
||||
if field.ctype_complex_type is not None and issubclass(
|
||||
field.ctype_complex_type, ctypes.Array
|
||||
):
|
||||
array_size = field.type_size
|
||||
containing_type = field.containing_type
|
||||
if containing_type.__module__ == ctypes.__name__:
|
||||
containing_type_size = ctypes.sizeof(containing_type)
|
||||
if array_size == 0:
|
||||
field_co_re_name = self._struct_name_generator(
|
||||
struct, field, field_index, True, 0, containing_type_size
|
||||
)
|
||||
globvar = ir.GlobalVariable(
|
||||
self.llvm_module, ir.IntType(64), name=field_co_re_name
|
||||
)
|
||||
globvar.linkage = "external"
|
||||
globvar.set_metadata("llvm.preserve.access.index", debug_info)
|
||||
self.generated_field_names[struct.name][field_name] = globvar
|
||||
field_index += 1
|
||||
continue
|
||||
for i in range(0, array_size):
|
||||
field_co_re_name = self._struct_name_generator(
|
||||
struct, field, field_index, True, i, containing_type_size
|
||||
)
|
||||
globvar = ir.GlobalVariable(
|
||||
self.llvm_module, ir.IntType(64), name=field_co_re_name
|
||||
)
|
||||
globvar.linkage = "external"
|
||||
globvar.set_metadata("llvm.preserve.access.index", debug_info)
|
||||
self.generated_field_names[struct.name][field_name] = globvar
|
||||
field_index += 1
|
||||
elif field.type_size is not None:
|
||||
array_size = field.type_size
|
||||
containing_type = field.containing_type
|
||||
if containing_type.__module__ == "vmlinux":
|
||||
containing_type_size = self.handler[
|
||||
containing_type.__name__
|
||||
].current_offset
|
||||
for i in range(0, array_size):
|
||||
field_co_re_name = self._struct_name_generator(
|
||||
struct, field, field_index, True, i, containing_type_size
|
||||
)
|
||||
globvar = ir.GlobalVariable(
|
||||
self.llvm_module, ir.IntType(64), name=field_co_re_name
|
||||
)
|
||||
globvar.linkage = "external"
|
||||
globvar.set_metadata("llvm.preserve.access.index", debug_info)
|
||||
self.generated_field_names[struct.name][field_name] = globvar
|
||||
field_index += 1
|
||||
else:
|
||||
field_co_re_name = self._struct_name_generator(
|
||||
struct, field, field_index
|
||||
)
|
||||
field_index += 1
|
||||
globvar = ir.GlobalVariable(
|
||||
self.llvm_module, ir.IntType(64), name=field_co_re_name
|
||||
)
|
||||
globvar.linkage = "external"
|
||||
globvar.set_metadata("llvm.preserve.access.index", debug_info)
|
||||
self.generated_field_names[struct.name][field_name] = globvar
|
||||
return debug_info
|
||||
|
||||
def _struct_name_generator(
|
||||
self,
|
||||
struct: DependencyNode,
|
||||
field,
|
||||
field_index: int,
|
||||
is_indexed: bool = False,
|
||||
index: int = 0,
|
||||
containing_type_size: int = 0,
|
||||
) -> str:
|
||||
# TODO: Does not support Unions as well as recursive pointer and array type naming
|
||||
if is_indexed:
|
||||
name = (
|
||||
"llvm."
|
||||
+ struct.name.removeprefix("struct_")
|
||||
+ f":0:{field.offset + index * containing_type_size}"
|
||||
+ "$"
|
||||
+ f"0:{field_index}:{index}"
|
||||
)
|
||||
return name
|
||||
elif struct.name.startswith("struct_"):
|
||||
name = (
|
||||
"llvm."
|
||||
+ struct.name.removeprefix("struct_")
|
||||
+ f":0:{field.offset}"
|
||||
+ "$"
|
||||
+ f"0:{field_index}"
|
||||
)
|
||||
return name
|
||||
else:
|
||||
print(self.handler[struct.name])
|
||||
raise TypeError(
|
||||
"Name generation cannot occur due to type name not starting with struct"
|
||||
)
|
||||
@ -1,14 +0,0 @@
|
||||
import logging
|
||||
from .dependency_handler import DependencyHandler
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class IRGenerator:
|
||||
def __init__(self, module, handler: DependencyHandler):
|
||||
self.module = module
|
||||
self.handler: DependencyHandler = handler
|
||||
if not handler.is_ready:
|
||||
raise ImportError(
|
||||
"Semantic analysis of vmlinux imports failed. Cannot generate IR"
|
||||
)
|
||||
255
pythonbpf/vmlinux_parser/vmlinux_exports_handler.py
Normal file
255
pythonbpf/vmlinux_parser/vmlinux_exports_handler.py
Normal file
@ -0,0 +1,255 @@
|
||||
import logging
|
||||
from typing import Any
|
||||
import ctypes
|
||||
from llvmlite import ir
|
||||
|
||||
from pythonbpf.local_symbol import LocalSymbol
|
||||
from pythonbpf.vmlinux_parser.assignment_info import AssignmentType
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class VmlinuxHandler:
|
||||
"""Handler for vmlinux-related operations"""
|
||||
|
||||
_instance = None
|
||||
|
||||
@classmethod
|
||||
def get_instance(cls):
|
||||
"""Get the singleton instance"""
|
||||
if cls._instance is None:
|
||||
logger.warning("VmlinuxHandler used before initialization")
|
||||
return None
|
||||
return cls._instance
|
||||
|
||||
@classmethod
|
||||
def initialize(cls, vmlinux_symtab):
|
||||
"""Initialize the handler with vmlinux symbol table"""
|
||||
cls._instance = cls(vmlinux_symtab)
|
||||
return cls._instance
|
||||
|
||||
def __init__(self, vmlinux_symtab):
|
||||
"""Initialize with vmlinux symbol table"""
|
||||
self.vmlinux_symtab = vmlinux_symtab
|
||||
logger.info(
|
||||
f"VmlinuxHandler initialized with {len(vmlinux_symtab) if vmlinux_symtab else 0} symbols"
|
||||
)
|
||||
|
||||
def is_vmlinux_enum(self, name):
|
||||
"""Check if name is a vmlinux enum constant"""
|
||||
return (
|
||||
name in self.vmlinux_symtab
|
||||
and self.vmlinux_symtab[name].value_type == AssignmentType.CONSTANT
|
||||
)
|
||||
|
||||
def get_struct_debug_info(self, name: str) -> Any:
|
||||
if (
|
||||
name in self.vmlinux_symtab
|
||||
and self.vmlinux_symtab[name].value_type == AssignmentType.STRUCT
|
||||
):
|
||||
return self.vmlinux_symtab[name].debug_info
|
||||
else:
|
||||
raise ValueError(f"{name} is not a vmlinux struct type")
|
||||
|
||||
def get_vmlinux_struct_type(self, name):
|
||||
"""Check if name is a vmlinux struct type"""
|
||||
if (
|
||||
name in self.vmlinux_symtab
|
||||
and self.vmlinux_symtab[name].value_type == AssignmentType.STRUCT
|
||||
):
|
||||
return self.vmlinux_symtab[name].python_type
|
||||
else:
|
||||
raise ValueError(f"{name} is not a vmlinux struct type")
|
||||
|
||||
def is_vmlinux_struct(self, name):
|
||||
"""Check if name is a vmlinux struct"""
|
||||
return (
|
||||
name in self.vmlinux_symtab
|
||||
and self.vmlinux_symtab[name].value_type == AssignmentType.STRUCT
|
||||
)
|
||||
|
||||
def handle_vmlinux_enum(self, name):
|
||||
"""Handle vmlinux enum constants by returning LLVM IR constants"""
|
||||
if self.is_vmlinux_enum(name):
|
||||
value = self.vmlinux_symtab[name].value
|
||||
logger.info(f"Resolving vmlinux enum {name} = {value}")
|
||||
return ir.Constant(ir.IntType(64), value), ir.IntType(64)
|
||||
return None
|
||||
|
||||
def get_vmlinux_enum_value(self, name):
|
||||
"""Handle vmlinux enum constants by returning LLVM IR constants"""
|
||||
if self.is_vmlinux_enum(name):
|
||||
value = self.vmlinux_symtab[name].value
|
||||
logger.info(f"The value of vmlinux enum {name} = {value}")
|
||||
return value
|
||||
return None
|
||||
|
||||
def handle_vmlinux_struct_field(
|
||||
self, struct_var_name, field_name, module, builder, local_sym_tab
|
||||
):
|
||||
"""Handle access to vmlinux struct fields"""
|
||||
if struct_var_name in local_sym_tab:
|
||||
var_info: LocalSymbol = local_sym_tab[struct_var_name]
|
||||
logger.info(
|
||||
f"Attempting to access field {field_name} of possible vmlinux struct {struct_var_name}"
|
||||
)
|
||||
python_type: type = var_info.metadata
|
||||
struct_name = python_type.__name__
|
||||
globvar_ir, field_data = self.get_field_type(struct_name, field_name)
|
||||
builder.function.args[0].type = ir.PointerType(ir.IntType(8))
|
||||
field_ptr = self.load_ctx_field(
|
||||
builder, builder.function.args[0], globvar_ir, field_data, struct_name
|
||||
)
|
||||
# Return pointer to field and field type
|
||||
return field_ptr, field_data
|
||||
else:
|
||||
raise RuntimeError("Variable accessed not found in symbol table")
|
||||
|
||||
@staticmethod
|
||||
def load_ctx_field(builder, ctx_arg, offset_global, field_data, struct_name=None):
|
||||
"""
|
||||
Generate LLVM IR to load a field from BPF context using offset.
|
||||
|
||||
Args:
|
||||
builder: llvmlite IRBuilder instance
|
||||
ctx_arg: The context pointer argument (ptr/i8*)
|
||||
offset_global: Global variable containing the field offset (i64)
|
||||
field_data: contains data about the field
|
||||
struct_name: Name of the struct being accessed (optional)
|
||||
Returns:
|
||||
The loaded value (i64 register or appropriately sized)
|
||||
"""
|
||||
|
||||
# Load the offset value
|
||||
offset = builder.load(offset_global)
|
||||
|
||||
# Ensure ctx_arg is treated as i8* (byte pointer)
|
||||
i8_ptr_type = ir.PointerType()
|
||||
|
||||
# Cast ctx_arg to i8* if it isn't already
|
||||
if str(ctx_arg.type) != str(i8_ptr_type):
|
||||
ctx_i8_ptr = builder.bitcast(ctx_arg, i8_ptr_type)
|
||||
else:
|
||||
ctx_i8_ptr = ctx_arg
|
||||
|
||||
# GEP with explicit type - this is the key fix
|
||||
field_ptr = builder.gep(
|
||||
ctx_i8_ptr,
|
||||
[offset],
|
||||
inbounds=False,
|
||||
)
|
||||
|
||||
# Get or declare the BPF passthrough intrinsic
|
||||
module = builder.function.module
|
||||
|
||||
try:
|
||||
passthrough_fn = module.globals.get("llvm.bpf.passthrough.p0.p0")
|
||||
if passthrough_fn is None:
|
||||
raise KeyError
|
||||
except (KeyError, AttributeError):
|
||||
passthrough_type = ir.FunctionType(
|
||||
i8_ptr_type,
|
||||
[ir.IntType(32), i8_ptr_type],
|
||||
)
|
||||
passthrough_fn = ir.Function(
|
||||
module,
|
||||
passthrough_type,
|
||||
name="llvm.bpf.passthrough.p0.p0",
|
||||
)
|
||||
|
||||
# Call passthrough to satisfy BPF verifier
|
||||
verified_ptr = builder.call(
|
||||
passthrough_fn, [ir.Constant(ir.IntType(32), 0), field_ptr], tail=True
|
||||
)
|
||||
|
||||
# Determine the appropriate IR type based on field information
|
||||
int_width = 64 # Default to 64-bit
|
||||
needs_zext = False # Track if we need zero-extension for xdp_md
|
||||
|
||||
if field_data is not None:
|
||||
# Try to determine the size from field metadata
|
||||
if field_data.type.__module__ == ctypes.__name__:
|
||||
try:
|
||||
field_size_bytes = ctypes.sizeof(field_data.type)
|
||||
field_size_bits = field_size_bytes * 8
|
||||
|
||||
if field_size_bits in [8, 16, 32, 64]:
|
||||
int_width = field_size_bits
|
||||
logger.info(f"Determined field size: {int_width} bits")
|
||||
|
||||
# Special handling for struct_xdp_md i32 fields
|
||||
# Load as i32 but extend to i64 before storing
|
||||
if struct_name == "struct_xdp_md" and int_width == 32:
|
||||
needs_zext = True
|
||||
logger.info(
|
||||
"struct_xdp_md i32 field detected, will zero-extend to i64"
|
||||
)
|
||||
else:
|
||||
logger.warning(
|
||||
f"Unusual field size {field_size_bits} bits, using default 64"
|
||||
)
|
||||
except Exception as e:
|
||||
logger.warning(
|
||||
f"Could not determine field size: {e}, using default 64"
|
||||
)
|
||||
|
||||
elif field_data.type.__module__ == "vmlinux":
|
||||
# For pointers to structs or complex vmlinux types
|
||||
if field_data.ctype_complex_type is not None and issubclass(
|
||||
field_data.ctype_complex_type, ctypes._Pointer
|
||||
):
|
||||
int_width = 64 # Pointers are always 64-bit
|
||||
logger.info("Field is a pointer type, using 64 bits")
|
||||
# TODO: Add handling for other complex types (arrays, embedded structs, etc.)
|
||||
else:
|
||||
logger.warning("Complex vmlinux field type, using default 64 bits")
|
||||
|
||||
# Bitcast to appropriate pointer type based on determined width
|
||||
ptr_type = ir.PointerType(ir.IntType(int_width))
|
||||
|
||||
typed_ptr = builder.bitcast(verified_ptr, ptr_type)
|
||||
|
||||
# Load and return the value
|
||||
value = builder.load(typed_ptr)
|
||||
|
||||
# Zero-extend i32 to i64 for struct_xdp_md fields
|
||||
if needs_zext:
|
||||
value = builder.zext(value, ir.IntType(64))
|
||||
logger.info("Zero-extended i32 value to i64 for struct_xdp_md field")
|
||||
|
||||
return value
|
||||
|
||||
def has_field(self, struct_name, field_name):
|
||||
"""Check if a vmlinux struct has a specific field"""
|
||||
if self.is_vmlinux_struct(struct_name):
|
||||
python_type = self.vmlinux_symtab[struct_name].python_type
|
||||
return hasattr(python_type, field_name)
|
||||
return False
|
||||
|
||||
def get_field_type(self, vmlinux_struct_name, field_name):
|
||||
"""Get the type of a field in a vmlinux struct"""
|
||||
if self.is_vmlinux_struct(vmlinux_struct_name):
|
||||
python_type = self.vmlinux_symtab[vmlinux_struct_name].python_type
|
||||
if hasattr(python_type, field_name):
|
||||
return self.vmlinux_symtab[vmlinux_struct_name].members[field_name]
|
||||
else:
|
||||
raise ValueError(
|
||||
f"Field {field_name} not found in vmlinux struct {vmlinux_struct_name}"
|
||||
)
|
||||
else:
|
||||
raise ValueError(f"{vmlinux_struct_name} is not a vmlinux struct")
|
||||
|
||||
def get_field_index(self, vmlinux_struct_name, field_name):
|
||||
"""Get the type of a field in a vmlinux struct"""
|
||||
if self.is_vmlinux_struct(vmlinux_struct_name):
|
||||
python_type = self.vmlinux_symtab[vmlinux_struct_name].python_type
|
||||
if hasattr(python_type, field_name):
|
||||
return list(
|
||||
self.vmlinux_symtab[vmlinux_struct_name].members.keys()
|
||||
).index(field_name)
|
||||
else:
|
||||
raise ValueError(
|
||||
f"Field {field_name} not found in vmlinux struct {vmlinux_struct_name}"
|
||||
)
|
||||
else:
|
||||
raise ValueError(f"{vmlinux_struct_name} is not a vmlinux struct")
|
||||
@ -1,19 +1,23 @@
|
||||
BPF_CLANG := clang
|
||||
CFLAGS := -O2 -emit-llvm -target bpf -c
|
||||
CFLAGS := -emit-llvm -target bpf -c
|
||||
|
||||
SRC := $(wildcard *.bpf.c)
|
||||
LL := $(SRC:.bpf.c=.bpf.ll)
|
||||
LL2 := $(SRC:.bpf.c=.bpf.o2.ll)
|
||||
OBJ := $(SRC:.bpf.c=.bpf.o)
|
||||
|
||||
.PHONY: all clean
|
||||
|
||||
all: $(LL) $(OBJ)
|
||||
all: $(LL) $(OBJ) $(LL2)
|
||||
|
||||
%.bpf.o: %.bpf.c
|
||||
$(BPF_CLANG) -O2 -g -target bpf -c $< -o $@
|
||||
|
||||
%.bpf.ll: %.bpf.c
|
||||
$(BPF_CLANG) $(CFLAGS) -g -S $< -o $@
|
||||
$(BPF_CLANG) -O0 $(CFLAGS) -g -S $< -o $@
|
||||
|
||||
%.bpf.o2.ll: %.bpf.c
|
||||
$(BPF_CLANG) -O2 $(CFLAGS) -g -S $< -o $@
|
||||
|
||||
clean:
|
||||
rm -f $(LL) $(OBJ)
|
||||
rm -f $(LL) $(OBJ) $(LL2)
|
||||
|
||||
@ -1,25 +0,0 @@
|
||||
#define __TARGET_ARCH_arm64
|
||||
|
||||
#include "vmlinux.h"
|
||||
#include <bpf/bpf_helpers.h>
|
||||
#include <bpf/bpf_tracing.h>
|
||||
#include <bpf/bpf_core_read.h>
|
||||
|
||||
// Map: key = struct request*, value = u64 timestamp
|
||||
struct {
|
||||
__uint(type, BPF_MAP_TYPE_HASH);
|
||||
__type(key, struct request *);
|
||||
__type(value, u64);
|
||||
__uint(max_entries, 1024);
|
||||
} start SEC(".maps");
|
||||
|
||||
// Attach to kprobe for blk_start_request
|
||||
SEC("kprobe/blk_start_request")
|
||||
int BPF_KPROBE(trace_start, struct request *req)
|
||||
{
|
||||
u64 ts = bpf_ktime_get_ns();
|
||||
bpf_map_update_elem(&start, &req, &ts, BPF_ANY);
|
||||
return 0;
|
||||
}
|
||||
|
||||
char LICENSE[] SEC("license") = "GPL";
|
||||
@ -1,23 +1,9 @@
|
||||
// SPDX-License-Identifier: GPL-2.0
|
||||
|
||||
#include <linux/bpf.h>
|
||||
#include "vmlinux.h"
|
||||
#include <bpf/bpf_helpers.h>
|
||||
#include <bpf/bpf_tracing.h>
|
||||
|
||||
struct trace_entry {
|
||||
short unsigned int type;
|
||||
unsigned char flags;
|
||||
unsigned char preempt_count;
|
||||
int pid;
|
||||
};
|
||||
|
||||
struct trace_event_raw_sys_enter {
|
||||
struct trace_entry ent;
|
||||
long int id;
|
||||
long unsigned int args[6];
|
||||
char __data[0];
|
||||
};
|
||||
|
||||
struct event {
|
||||
__u32 pid;
|
||||
__u32 uid;
|
||||
@ -33,7 +19,7 @@ struct {
|
||||
SEC("tp/syscalls/sys_enter_setuid")
|
||||
int handle_setuid_entry(struct trace_event_raw_sys_enter *ctx) {
|
||||
struct event data = {};
|
||||
|
||||
struct blk_integrity_iter it = {};
|
||||
// Extract UID from the syscall arguments
|
||||
data.uid = (unsigned int)ctx->args[0];
|
||||
data.ts = bpf_ktime_get_ns();
|
||||
|
||||
15
tests/c-form/i32test.bpf.c
Normal file
15
tests/c-form/i32test.bpf.c
Normal file
@ -0,0 +1,15 @@
|
||||
#include <linux/bpf.h>
|
||||
#include <bpf/bpf_helpers.h>
|
||||
|
||||
SEC("xdp")
|
||||
int print_xdp_data(struct xdp_md *ctx)
|
||||
{
|
||||
// 'data' is a pointer to the start of packet data
|
||||
long data = (long)ctx->data;
|
||||
|
||||
bpf_printk("ctx->data = %lld\n", data);
|
||||
|
||||
return XDP_PASS;
|
||||
}
|
||||
|
||||
char LICENSE[] SEC("license") = "GPL";
|
||||
@ -2,18 +2,75 @@
|
||||
#include <bpf/bpf_helpers.h>
|
||||
#include <bpf/bpf_tracing.h>
|
||||
|
||||
char LICENSE[] SEC("license") = "Dual BSD/GPL";
|
||||
char LICENSE[] SEC("license") = "GPL";
|
||||
|
||||
SEC("kprobe/do_unlinkat")
|
||||
int kprobe_execve(struct pt_regs *ctx)
|
||||
{
|
||||
bpf_printk("unlinkat created");
|
||||
return 0;
|
||||
}
|
||||
|
||||
SEC("kretprobe/do_unlinkat")
|
||||
int kretprobe_execve(struct pt_regs *ctx)
|
||||
{
|
||||
bpf_printk("unlinkat returned\n");
|
||||
unsigned long r15 = ctx->r15;
|
||||
bpf_printk("r15: %lld", r15);
|
||||
|
||||
unsigned long r14 = ctx->r14;
|
||||
bpf_printk("r14: %lld", r14);
|
||||
|
||||
unsigned long r13 = ctx->r13;
|
||||
bpf_printk("r13: %lld", r13);
|
||||
|
||||
unsigned long r12 = ctx->r12;
|
||||
bpf_printk("r12: %lld", r12);
|
||||
|
||||
unsigned long bp = ctx->bp;
|
||||
bpf_printk("rbp: %lld", bp);
|
||||
|
||||
unsigned long bx = ctx->bx;
|
||||
bpf_printk("rbx: %lld", bx);
|
||||
|
||||
unsigned long r11 = ctx->r11;
|
||||
bpf_printk("r11: %lld", r11);
|
||||
|
||||
unsigned long r10 = ctx->r10;
|
||||
bpf_printk("r10: %lld", r10);
|
||||
|
||||
unsigned long r9 = ctx->r9;
|
||||
bpf_printk("r9: %lld", r9);
|
||||
|
||||
unsigned long r8 = ctx->r8;
|
||||
bpf_printk("r8: %lld", r8);
|
||||
|
||||
unsigned long ax = ctx->ax;
|
||||
bpf_printk("rax: %lld", ax);
|
||||
|
||||
unsigned long cx = ctx->cx;
|
||||
bpf_printk("rcx: %lld", cx);
|
||||
|
||||
unsigned long dx = ctx->dx;
|
||||
bpf_printk("rdx: %lld", dx);
|
||||
|
||||
unsigned long si = ctx->si;
|
||||
bpf_printk("rsi: %lld", si);
|
||||
|
||||
unsigned long di = ctx->di;
|
||||
bpf_printk("rdi: %lld", di);
|
||||
|
||||
unsigned long orig_ax = ctx->orig_ax;
|
||||
bpf_printk("orig_rax: %lld", orig_ax);
|
||||
|
||||
unsigned long ip = ctx->ip;
|
||||
bpf_printk("rip: %lld", ip);
|
||||
|
||||
unsigned long cs = ctx->cs;
|
||||
bpf_printk("cs: %lld", cs);
|
||||
|
||||
unsigned long flags = ctx->flags;
|
||||
bpf_printk("eflags: %lld", flags);
|
||||
|
||||
unsigned long sp = ctx->sp;
|
||||
bpf_printk("rsp: %lld", sp);
|
||||
|
||||
unsigned long ss = ctx->ss;
|
||||
bpf_printk("ss: %lld", ss);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
42
tests/c-form/struct_field_tests.bpf.c
Normal file
42
tests/c-form/struct_field_tests.bpf.c
Normal file
@ -0,0 +1,42 @@
|
||||
// SPDX-License-Identifier: GPL-2.0
|
||||
|
||||
#include "vmlinux.h"
|
||||
#include <bpf/bpf_helpers.h>
|
||||
#include <bpf/bpf_tracing.h>
|
||||
|
||||
/*
|
||||
Information gained from reversing this (multiple kernel versions):
|
||||
There is no point of
|
||||
```llvm
|
||||
tail call void @llvm.dbg.value(metadata ptr %0, metadata !60, metadata !DIExpression()), !dbg !70
|
||||
```
|
||||
and the first argument of passthrough is fucking useless. It just needs to be a distinct integer:
|
||||
```llvm
|
||||
%9 = tail call ptr @llvm.bpf.passthrough.p0.p0(i32 3, ptr %8)
|
||||
```
|
||||
*/
|
||||
|
||||
SEC("tp/syscalls/sys_enter_execve")
|
||||
int handle_setuid_entry(struct trace_event_raw_sys_enter *ctx) {
|
||||
// Access each argument separately with clear variable assignments
|
||||
long int id = ctx->id;
|
||||
bpf_printk("This is context field %d", id);
|
||||
/*
|
||||
* the IR to aim for is
|
||||
* %2 = alloca ptr, align 8
|
||||
* store ptr %0, ptr %2, align 8
|
||||
* Above, %0 is the arg pointer
|
||||
* %5 = load ptr, ptr %2, align 8
|
||||
* %6 = getelementptr inbounds %struct.trace_event_raw_sys_enter, ptr %5, i32 0, i32 2
|
||||
* %7 = load i64, ptr @"llvm.trace_event_raw_sys_enter:0:16$0:2:0", align 8
|
||||
* %8 = bitcast ptr %5 to ptr
|
||||
* %9 = getelementptr i8, ptr %8, i64 %7
|
||||
* %10 = bitcast ptr %9 to ptr
|
||||
* %11 = call ptr @llvm.bpf.passthrough.p0.p0(i32 0, ptr %10)
|
||||
* %12 = load i64, ptr %11, align 8, !dbg !101
|
||||
*
|
||||
*/
|
||||
return 0;
|
||||
}
|
||||
|
||||
char LICENSE[] SEC("license") = "GPL";
|
||||
121617
tests/c-form/vmlinux.h
vendored
121617
tests/c-form/vmlinux.h
vendored
File diff suppressed because it is too large
Load Diff
21
tests/c-form/xdp_modify.bpf.c
Normal file
21
tests/c-form/xdp_modify.bpf.c
Normal file
@ -0,0 +1,21 @@
|
||||
// xdp_rewrite.c
|
||||
#include <linux/bpf.h>
|
||||
#include <bpf/bpf_helpers.h>
|
||||
#include <linux/if_ether.h>
|
||||
|
||||
SEC("xdp")
|
||||
int xdp_rewrite_mac(struct xdp_md *ctx)
|
||||
{
|
||||
void *data_end = (void *)(long)ctx->data_end;
|
||||
void *data = (void *)(long)ctx->data;
|
||||
|
||||
struct ethhdr *eth = data;
|
||||
if ((void*)(eth + 1) > data_end)
|
||||
return XDP_PASS;
|
||||
__u8 new_src[ETH_ALEN] = {0x02,0x00,0x00,0x00,0x00,0x02};
|
||||
for (int i = 0; i < ETH_ALEN; i++) eth->h_source[i] = new_src[i];
|
||||
|
||||
return XDP_PASS;
|
||||
}
|
||||
|
||||
char _license[] SEC("license") = "GPL";
|
||||
30
tests/failing_tests/vmlinux/args_test.py
Normal file
30
tests/failing_tests/vmlinux/args_test.py
Normal file
@ -0,0 +1,30 @@
|
||||
import logging
|
||||
|
||||
from pythonbpf import bpf, section, bpfglobal, compile_to_ir
|
||||
from pythonbpf import compile # noqa: F401
|
||||
from vmlinux import TASK_COMM_LEN # noqa: F401
|
||||
from vmlinux import struct_trace_event_raw_sys_enter # noqa: F401
|
||||
from ctypes import c_int64, c_int32, c_void_p # noqa: F401
|
||||
|
||||
|
||||
# from vmlinux import struct_uinput_device
|
||||
# from vmlinux import struct_blk_integrity_iter
|
||||
|
||||
|
||||
@bpf
|
||||
@section("tracepoint/syscalls/sys_enter_execve")
|
||||
def hello_world(ctx: struct_trace_event_raw_sys_enter) -> c_int64:
|
||||
b = ctx.args
|
||||
c = b[0]
|
||||
print(f"This is context args field {c}")
|
||||
return c_int64(0)
|
||||
|
||||
|
||||
@bpf
|
||||
@bpfglobal
|
||||
def LICENSE() -> str:
|
||||
return "GPL"
|
||||
|
||||
|
||||
compile_to_ir("args_test.py", "args_test.ll", loglevel=logging.INFO)
|
||||
compile()
|
||||
@ -1,10 +1,17 @@
|
||||
from pythonbpf import bpf, map, section, bpfglobal, compile_to_ir
|
||||
from pythonbpf.maps import HashMap
|
||||
from pythonbpf.helper import XDP_PASS
|
||||
from vmlinux import struct_xdp_md
|
||||
from vmlinux import struct_xdp_buff # noqa: F401
|
||||
from vmlinux import struct_ring_buffer_per_cpu # noqa: F401
|
||||
from vmlinux import TASK_COMM_LEN # noqa: F401
|
||||
|
||||
from vmlinux import struct_qspinlock # noqa: F401
|
||||
|
||||
# from vmlinux import struct_trace_event_raw_sys_enter # noqa: F401
|
||||
# from vmlinux import struct_posix_cputimers # noqa: F401
|
||||
from vmlinux import struct_xdp_md
|
||||
|
||||
# from vmlinux import struct_trace_event_raw_sys_enter # noqa: F401
|
||||
# from vmlinux import struct_ring_buffer_per_cpu # noqa: F401
|
||||
# from vmlinux import struct_request # noqa: F401
|
||||
from ctypes import c_int64
|
||||
|
||||
# Instructions to how to run this program
|
||||
|
||||
@ -1,4 +1,4 @@
|
||||
from pythonbpf import bpf, map, section, bpfglobal, compile, struct
|
||||
from pythonbpf import bpf, map, section, bpfglobal, compile, struct, compile_to_ir
|
||||
from ctypes import c_void_p, c_int64, c_int32, c_uint64
|
||||
from pythonbpf.maps import HashMap
|
||||
from pythonbpf.helper import ktime
|
||||
@ -71,4 +71,5 @@ def LICENSE() -> str:
|
||||
return "GPL"
|
||||
|
||||
|
||||
compile_to_ir("comprehensive.py", "comprehensive.ll")
|
||||
compile()
|
||||
|
||||
31
tests/passing_tests/vmlinux/i32_test.py
Normal file
31
tests/passing_tests/vmlinux/i32_test.py
Normal file
@ -0,0 +1,31 @@
|
||||
from ctypes import c_int64, c_void_p
|
||||
from pythonbpf import bpf, section, bpfglobal, compile_to_ir, compile
|
||||
from vmlinux import struct_xdp_md
|
||||
from vmlinux import XDP_PASS
|
||||
|
||||
|
||||
@bpf
|
||||
@section("xdp")
|
||||
def print_xdp_dat2a(ct2x: struct_xdp_md) -> c_int64:
|
||||
data = ct2x.data # 32-bit field: packet start pointer
|
||||
print(f"ct2x->data = {data}")
|
||||
return c_int64(XDP_PASS)
|
||||
|
||||
|
||||
@bpf
|
||||
@section("xdp")
|
||||
def print_xdp_data(ctx: struct_xdp_md) -> c_int64:
|
||||
data = ctx.data # 32-bit field: packet start pointer
|
||||
something = c_void_p(data)
|
||||
print(f"ctx->data = {something}")
|
||||
return c_int64(XDP_PASS)
|
||||
|
||||
|
||||
@bpf
|
||||
@bpfglobal
|
||||
def LICENSE() -> str:
|
||||
return "GPL"
|
||||
|
||||
|
||||
compile_to_ir("i32_test.py", "i32_test.ll")
|
||||
compile()
|
||||
24
tests/passing_tests/vmlinux/i32_test_fail_1.py
Normal file
24
tests/passing_tests/vmlinux/i32_test_fail_1.py
Normal file
@ -0,0 +1,24 @@
|
||||
from ctypes import c_int64
|
||||
from pythonbpf import bpf, section, bpfglobal, compile
|
||||
from vmlinux import struct_xdp_md
|
||||
from vmlinux import XDP_PASS
|
||||
import logging
|
||||
|
||||
|
||||
@bpf
|
||||
@section("xdp")
|
||||
def print_xdp_data(ctx: struct_xdp_md) -> c_int64:
|
||||
data = 0
|
||||
data = ctx.data # 32-bit field: packet start pointer
|
||||
something = 2 + data
|
||||
print(f"ctx->data = {something}")
|
||||
return c_int64(XDP_PASS)
|
||||
|
||||
|
||||
@bpf
|
||||
@bpfglobal
|
||||
def LICENSE() -> str:
|
||||
return "GPL"
|
||||
|
||||
|
||||
compile(logging.INFO)
|
||||
24
tests/passing_tests/vmlinux/i32_test_fail_2.py
Normal file
24
tests/passing_tests/vmlinux/i32_test_fail_2.py
Normal file
@ -0,0 +1,24 @@
|
||||
from ctypes import c_int64
|
||||
from pythonbpf import bpf, section, bpfglobal, compile, compile_to_ir
|
||||
from vmlinux import struct_xdp_md
|
||||
from vmlinux import XDP_PASS
|
||||
import logging
|
||||
|
||||
|
||||
@bpf
|
||||
@section("xdp")
|
||||
def print_xdp_data(ctx: struct_xdp_md) -> c_int64:
|
||||
data = c_int64(ctx.data) # 32-bit field: packet start pointer
|
||||
something = 2 + data
|
||||
print(f"ctx->data = {something}")
|
||||
return c_int64(XDP_PASS)
|
||||
|
||||
|
||||
@bpf
|
||||
@bpfglobal
|
||||
def LICENSE() -> str:
|
||||
return "GPL"
|
||||
|
||||
|
||||
compile_to_ir("i32_test_fail_2.py", "i32_test_fail_2.ll")
|
||||
compile(logging.INFO)
|
||||
53
tests/passing_tests/vmlinux/register_state_dump.py
Normal file
53
tests/passing_tests/vmlinux/register_state_dump.py
Normal file
@ -0,0 +1,53 @@
|
||||
from pythonbpf import bpf, section, bpfglobal, BPF, trace_pipe
|
||||
from pythonbpf import compile # noqa: F401
|
||||
from vmlinux import struct_pt_regs
|
||||
from ctypes import c_int64, c_int32, c_void_p # noqa: F401
|
||||
|
||||
|
||||
@bpf
|
||||
@section("kprobe/do_unlinkat")
|
||||
def kprobe_execve(ctx: struct_pt_regs) -> c_int64:
|
||||
r15 = ctx.r15
|
||||
r14 = ctx.r14
|
||||
r13 = ctx.r13
|
||||
r12 = ctx.r12
|
||||
bp = ctx.bp
|
||||
bx = ctx.bx
|
||||
r11 = ctx.r11
|
||||
r10 = ctx.r10
|
||||
r9 = ctx.r9
|
||||
r8 = ctx.r8
|
||||
ax = ctx.ax
|
||||
cx = ctx.cx
|
||||
dx = ctx.dx
|
||||
si = ctx.si
|
||||
di = ctx.di
|
||||
orig_ax = ctx.orig_ax
|
||||
ip = ctx.ip
|
||||
cs = ctx.cs
|
||||
flags = ctx.flags
|
||||
sp = ctx.sp
|
||||
ss = ctx.ss
|
||||
|
||||
print(f"r15={r15} r14={r14} r13={r13}")
|
||||
print(f"r12={r12} rbp={bp} rbx={bx}")
|
||||
print(f"r11={r11} r10={r10} r9={r9}")
|
||||
print(f"r8={r8} rax={ax} rcx={cx}")
|
||||
print(f"rdx={dx} rsi={si} rdi={di}")
|
||||
print(f"orig_rax={orig_ax} rip={ip} cs={cs}")
|
||||
print(f"eflags={flags} rsp={sp} ss={ss}")
|
||||
|
||||
return c_int64(0)
|
||||
|
||||
|
||||
@bpf
|
||||
@bpfglobal
|
||||
def LICENSE() -> str:
|
||||
return "GPL"
|
||||
|
||||
|
||||
b = BPF()
|
||||
b.load()
|
||||
b.attach_all()
|
||||
|
||||
trace_pipe()
|
||||
47
tests/passing_tests/vmlinux/simple_struct_test.py
Normal file
47
tests/passing_tests/vmlinux/simple_struct_test.py
Normal file
@ -0,0 +1,47 @@
|
||||
import logging
|
||||
|
||||
from pythonbpf import bpf, section, bpfglobal, compile_to_ir, map
|
||||
from pythonbpf import compile # noqa: F401
|
||||
from vmlinux import TASK_COMM_LEN # noqa: F401
|
||||
from vmlinux import struct_trace_event_raw_sys_enter # noqa: F401
|
||||
from ctypes import c_uint64, c_int32, c_int64
|
||||
from pythonbpf.maps import HashMap
|
||||
|
||||
# from vmlinux import struct_uinput_device
|
||||
# from vmlinux import struct_blk_integrity_iter
|
||||
|
||||
|
||||
@bpf
|
||||
@map
|
||||
def mymap() -> HashMap:
|
||||
return HashMap(key=c_int32, value=c_uint64, max_entries=TASK_COMM_LEN)
|
||||
|
||||
|
||||
@bpf
|
||||
@map
|
||||
def mymap2() -> HashMap:
|
||||
return HashMap(key=c_int32, value=c_uint64, max_entries=18)
|
||||
|
||||
|
||||
# Instructions to how to run this program
|
||||
# 1. Install PythonBPF: pip install pythonbpf
|
||||
# 2. Run the program: python examples/simple_struct_test.py
|
||||
# 3. Run the program with sudo: sudo tools/check.sh run examples/simple_struct_test.o
|
||||
# 4. Attach object file to any network device with something like ./check.sh run examples/simple_struct_test.o tailscale0
|
||||
# 5. send traffic through the device and observe effects
|
||||
@bpf
|
||||
@section("tracepoint/syscalls/sys_enter_execve")
|
||||
def hello_world(ctx: struct_trace_event_raw_sys_enter) -> c_int64:
|
||||
a = 2 + TASK_COMM_LEN + TASK_COMM_LEN
|
||||
print(f"Hello, World{TASK_COMM_LEN} and {a}")
|
||||
return c_int64(TASK_COMM_LEN + 2)
|
||||
|
||||
|
||||
@bpf
|
||||
@bpfglobal
|
||||
def LICENSE() -> str:
|
||||
return "GPL"
|
||||
|
||||
|
||||
compile_to_ir("simple_struct_test.py", "simple_struct_test.ll", loglevel=logging.DEBUG)
|
||||
compile()
|
||||
29
tests/passing_tests/vmlinux/struct_field_access.py
Normal file
29
tests/passing_tests/vmlinux/struct_field_access.py
Normal file
@ -0,0 +1,29 @@
|
||||
import logging
|
||||
|
||||
from pythonbpf import bpf, section, bpfglobal, compile_to_ir
|
||||
from pythonbpf import compile # noqa: F401
|
||||
from vmlinux import TASK_COMM_LEN # noqa: F401
|
||||
from vmlinux import struct_trace_event_raw_sys_enter # noqa: F401
|
||||
from ctypes import c_int64, c_int32, c_void_p # noqa: F401
|
||||
|
||||
|
||||
# from vmlinux import struct_uinput_device
|
||||
# from vmlinux import struct_blk_integrity_iter
|
||||
|
||||
|
||||
@bpf
|
||||
@section("tracepoint/syscalls/sys_enter_execve")
|
||||
def hello_world(ctx: struct_trace_event_raw_sys_enter) -> c_int64:
|
||||
b = ctx.id
|
||||
print(f"This is context field {b}")
|
||||
return c_int64(0)
|
||||
|
||||
|
||||
@bpf
|
||||
@bpfglobal
|
||||
def LICENSE() -> str:
|
||||
return "GPL"
|
||||
|
||||
|
||||
compile_to_ir("struct_field_access.py", "struct_field_access.ll", loglevel=logging.INFO)
|
||||
compile()
|
||||
209
tools/setup.sh
Executable file
209
tools/setup.sh
Executable file
@ -0,0 +1,209 @@
|
||||
#!/bin/bash
|
||||
|
||||
print_warning() {
|
||||
echo -e "\033[1;33m$1\033[0m"
|
||||
}
|
||||
print_info() {
|
||||
echo -e "\033[1;32m$1\033[0m"
|
||||
}
|
||||
|
||||
if [ "$EUID" -ne 0 ]; then
|
||||
echo "Please run this script with sudo."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
print_warning "===================================================================="
|
||||
print_warning " WARNING "
|
||||
print_warning " This script will run kernel-level BPF programs. "
|
||||
print_warning " BPF programs run with kernel privileges and could potentially "
|
||||
print_warning " affect system stability if not used properly. "
|
||||
print_warning " "
|
||||
print_warning " This is a non-interactive version for curl piping. "
|
||||
print_warning " The script will proceed automatically with installation. "
|
||||
print_warning "===================================================================="
|
||||
echo
|
||||
|
||||
print_info "This script will:"
|
||||
echo "1. Check and install required dependencies (libelf, clang, python, bpftool)"
|
||||
echo "2. Download example programs from the Python-BPF GitHub repository"
|
||||
echo "3. Create a Python virtual environment with necessary packages"
|
||||
echo "4. Set up a Jupyter notebook server"
|
||||
echo "Starting in 5 seconds. Press Ctrl+C to cancel..."
|
||||
sleep 5
|
||||
|
||||
WORK_DIR="/tmp/python_bpf_setup"
|
||||
REAL_USER=$(logname || echo "$SUDO_USER")
|
||||
|
||||
echo "Creating temporary directory: $WORK_DIR"
|
||||
mkdir -p "$WORK_DIR"
|
||||
cd "$WORK_DIR" || exit 1
|
||||
|
||||
if [ -f /etc/os-release ]; then
|
||||
. /etc/os-release
|
||||
DISTRO=$ID
|
||||
else
|
||||
echo "Cannot determine Linux distribution. Exiting."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
install_dependencies() {
|
||||
case $DISTRO in
|
||||
ubuntu|debian|pop|mint|elementary|zorin)
|
||||
echo "Detected Ubuntu/Debian-based system"
|
||||
apt update
|
||||
|
||||
# Check and install libelf
|
||||
if ! dpkg -l libelf-dev >/dev/null 2>&1; then
|
||||
echo "Installing libelf-dev..."
|
||||
apt install -y libelf-dev
|
||||
else
|
||||
echo "libelf-dev is already installed."
|
||||
fi
|
||||
|
||||
# Check and install clang
|
||||
if ! command -v clang >/dev/null 2>&1; then
|
||||
echo "Installing clang..."
|
||||
apt install -y clang
|
||||
else
|
||||
echo "clang is already installed."
|
||||
fi
|
||||
|
||||
# Check and install python
|
||||
if ! command -v python3 >/dev/null 2>&1; then
|
||||
echo "Installing python3..."
|
||||
apt install -y python3 python3-pip python3-venv
|
||||
else
|
||||
echo "python3 is already installed."
|
||||
fi
|
||||
|
||||
# Check and install bpftool
|
||||
if ! command -v bpftool >/dev/null 2>&1; then
|
||||
echo "Installing bpftool..."
|
||||
apt install -y linux-tools-common linux-tools-generic
|
||||
|
||||
# If bpftool still not found, try installing linux-tools-$(uname -r)
|
||||
if ! command -v bpftool >/dev/null 2>&1; then
|
||||
KERNEL_VERSION=$(uname -r)
|
||||
apt install -y linux-tools-$KERNEL_VERSION
|
||||
fi
|
||||
else
|
||||
echo "bpftool is already installed."
|
||||
fi
|
||||
;;
|
||||
|
||||
arch|manjaro|endeavouros)
|
||||
echo "Detected Arch-based Linux system"
|
||||
|
||||
# Check and install libelf
|
||||
if ! pacman -Q libelf >/dev/null 2>&1; then
|
||||
echo "Installing libelf..."
|
||||
pacman -S --noconfirm libelf
|
||||
else
|
||||
echo "libelf is already installed."
|
||||
fi
|
||||
|
||||
# Check and install clang
|
||||
if ! command -v clang >/dev/null 2>&1; then
|
||||
echo "Installing clang..."
|
||||
pacman -S --noconfirm clang
|
||||
else
|
||||
echo "clang is already installed."
|
||||
fi
|
||||
|
||||
# Check and install python
|
||||
if ! command -v python3 >/dev/null 2>&1; then
|
||||
echo "Installing python3..."
|
||||
pacman -S --noconfirm python python-pip
|
||||
else
|
||||
echo "python3 is already installed."
|
||||
fi
|
||||
|
||||
# Check and install bpftool
|
||||
if ! command -v bpftool >/dev/null 2>&1; then
|
||||
echo "Installing bpftool..."
|
||||
pacman -S --noconfirm bpf linux-headers
|
||||
else
|
||||
echo "bpftool is already installed."
|
||||
fi
|
||||
;;
|
||||
|
||||
*)
|
||||
echo "Unsupported distribution: $DISTRO"
|
||||
echo "This script only supports Ubuntu/Debian and Arch Linux derivatives."
|
||||
exit 1
|
||||
;;
|
||||
esac
|
||||
}
|
||||
|
||||
echo "Checking and installing dependencies..."
|
||||
install_dependencies
|
||||
|
||||
# Download example programs
|
||||
echo "Downloading example programs from Python-BPF GitHub repository..."
|
||||
mkdir -p examples
|
||||
cd examples || exit 1
|
||||
|
||||
echo "Fetching example files list..."
|
||||
FILES=$(curl -s "https://api.github.com/repos/pythonbpf/Python-BPF/contents/examples" | grep -E -o '"path": "examples/[^"]*(\.md|\.ipynb|\.py)"' | awk -F'"' '{print $4}')
|
||||
BCC_EXAMPLES=$(curl -s "https://api.github.com/repos/pythonbpf/Python-BPF/contents/BCC-Examples" | grep -E -o '"path": "BCC-Examples/[^"]*(\.md|\.ipynb|\.py)"' | awk -F'"' '{print $4}')
|
||||
|
||||
if [ -z "$FILES" ]; then
|
||||
echo "Failed to fetch file list from repository. Using fallback method..."
|
||||
# Fallback to downloading common example files
|
||||
EXAMPLES=(
|
||||
"binops_demo.py"
|
||||
"blk_request.py"
|
||||
"clone-matplotlib.ipynb"
|
||||
"clone_plot.py"
|
||||
"hello_world.py"
|
||||
"kprobes.py"
|
||||
"struct_and_perf.py"
|
||||
"sys_sync.py"
|
||||
"xdp_pass.py"
|
||||
)
|
||||
|
||||
for example in "${EXAMPLES[@]}"; do
|
||||
echo "Downloading: $example"
|
||||
curl -s -O "https://raw.githubusercontent.com/pythonbpf/Python-BPF/master/examples/$example"
|
||||
done
|
||||
else
|
||||
mkdir examples && cd examples
|
||||
for file in $FILES; do
|
||||
filename=$(basename "$file")
|
||||
echo "Downloading: $filename"
|
||||
curl -s -o "$filename" "https://raw.githubusercontent.com/pythonbpf/Python-BPF/master/$file"
|
||||
done
|
||||
cd ..
|
||||
mkdir BCC-Examples && cd BCC-Examples
|
||||
for file in $BCC_EXAMPLES; do
|
||||
filename=$(basename "$file")
|
||||
echo "Downloading: $filename"
|
||||
curl -s -o "$filename" "https://raw.githubusercontent.com/pythonbpf/Python-BPF/master/$file"
|
||||
done
|
||||
cd ..
|
||||
fi
|
||||
|
||||
cd "$WORK_DIR" || exit 1
|
||||
chown -R "$REAL_USER:$(id -gn "$REAL_USER")" .
|
||||
|
||||
echo "Creating Python virtual environment..."
|
||||
su - "$REAL_USER" -c "cd \"$WORK_DIR\" && python3 -m venv venv"
|
||||
|
||||
echo "Installing Python packages..."
|
||||
su - "$REAL_USER" -c "cd \"$WORK_DIR\" && source venv/bin/activate && pip install --upgrade pip && pip install jupyter pythonbpf pylibbpf matplotlib"
|
||||
|
||||
cat > "$WORK_DIR/start_jupyter.sh" << EOF
|
||||
#!/bin/bash
|
||||
cd "$WORK_DIR"
|
||||
source venv/bin/activate
|
||||
cd examples
|
||||
sudo ../venv/bin/python -m notebook --ip=0.0.0.0 --allow-root
|
||||
EOF
|
||||
|
||||
chmod +x "$WORK_DIR/start_jupyter.sh"
|
||||
chown "$REAL_USER:$(id -gn "$REAL_USER")" "$WORK_DIR/start_jupyter.sh"
|
||||
|
||||
print_info "========================================================"
|
||||
print_info "Setup complete! To start Jupyter Notebook, run:"
|
||||
print_info "$ sudo $WORK_DIR/start_jupyter.sh"
|
||||
print_info "========================================================"
|
||||
Reference in New Issue
Block a user