~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/tools/testing/selftests/bpf/progs/profiler.inc.h

Version: ~ [ linux-6.11.5 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.58 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.114 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.169 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.228 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.284 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.322 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 // SPDX-License-Identifier: GPL-2.0
  2 /* Copyright (c) 2020 Facebook */
  3 #include <vmlinux.h>
  4 #include <bpf/bpf_core_read.h>
  5 #include <bpf/bpf_helpers.h>
  6 #include <bpf/bpf_tracing.h>
  7 
  8 #include "profiler.h"
  9 #include "err.h"
 10 #include "bpf_experimental.h"
 11 #include "bpf_compiler.h"
 12 #include "bpf_misc.h"
 13 
 14 #ifndef NULL
 15 #define NULL 0
 16 #endif
 17 
 18 #define O_WRONLY 00000001
 19 #define O_RDWR 00000002
 20 #define O_DIRECTORY 00200000
 21 #define __O_TMPFILE 020000000
 22 #define O_TMPFILE (__O_TMPFILE | O_DIRECTORY)
 23 #define S_IFMT 00170000
 24 #define S_IFSOCK 0140000
 25 #define S_IFLNK 0120000
 26 #define S_IFREG 0100000
 27 #define S_IFBLK 0060000
 28 #define S_IFDIR 0040000
 29 #define S_IFCHR 0020000
 30 #define S_IFIFO 0010000
 31 #define S_ISUID 0004000
 32 #define S_ISGID 0002000
 33 #define S_ISVTX 0001000
 34 #define S_ISLNK(m) (((m)&S_IFMT) == S_IFLNK)
 35 #define S_ISDIR(m) (((m)&S_IFMT) == S_IFDIR)
 36 #define S_ISCHR(m) (((m)&S_IFMT) == S_IFCHR)
 37 #define S_ISBLK(m) (((m)&S_IFMT) == S_IFBLK)
 38 #define S_ISFIFO(m) (((m)&S_IFMT) == S_IFIFO)
 39 #define S_ISSOCK(m) (((m)&S_IFMT) == S_IFSOCK)
 40 
 41 #define KILL_DATA_ARRAY_SIZE 8
 42 
 43 struct var_kill_data_arr_t {
 44         struct var_kill_data_t array[KILL_DATA_ARRAY_SIZE];
 45 };
 46 
 47 union any_profiler_data_t {
 48         struct var_exec_data_t var_exec;
 49         struct var_kill_data_t var_kill;
 50         struct var_sysctl_data_t var_sysctl;
 51         struct var_filemod_data_t var_filemod;
 52         struct var_fork_data_t var_fork;
 53         struct var_kill_data_arr_t var_kill_data_arr;
 54 };
 55 
 56 volatile struct profiler_config_struct bpf_config = {};
 57 
 58 #define FETCH_CGROUPS_FROM_BPF (bpf_config.fetch_cgroups_from_bpf)
 59 #define CGROUP_FS_INODE (bpf_config.cgroup_fs_inode)
 60 #define CGROUP_LOGIN_SESSION_INODE \
 61         (bpf_config.cgroup_login_session_inode)
 62 #define KILL_SIGNALS (bpf_config.kill_signals_mask)
 63 #define STALE_INFO (bpf_config.stale_info_secs)
 64 #define INODE_FILTER (bpf_config.inode_filter)
 65 #define READ_ENVIRON_FROM_EXEC (bpf_config.read_environ_from_exec)
 66 #define ENABLE_CGROUP_V1_RESOLVER (bpf_config.enable_cgroup_v1_resolver)
 67 
 68 struct kernfs_iattrs___52 {
 69         struct iattr ia_iattr;
 70 };
 71 
 72 struct kernfs_node___52 {
 73         union /* kernfs_node_id */ {
 74                 struct {
 75                         u32 ino;
 76                         u32 generation;
 77                 };
 78                 u64 id;
 79         } id;
 80 };
 81 
 82 struct {
 83         __uint(type, BPF_MAP_TYPE_PERCPU_ARRAY);
 84         __uint(max_entries, 1);
 85         __type(key, u32);
 86         __type(value, union any_profiler_data_t);
 87 } data_heap SEC(".maps");
 88 
 89 struct {
 90         __uint(type, BPF_MAP_TYPE_PERF_EVENT_ARRAY);
 91         __uint(key_size, sizeof(int));
 92         __uint(value_size, sizeof(int));
 93 } events SEC(".maps");
 94 
 95 struct {
 96         __uint(type, BPF_MAP_TYPE_HASH);
 97         __uint(max_entries, KILL_DATA_ARRAY_SIZE);
 98         __type(key, u32);
 99         __type(value, struct var_kill_data_arr_t);
100 } var_tpid_to_data SEC(".maps");
101 
102 struct {
103         __uint(type, BPF_MAP_TYPE_PERCPU_ARRAY);
104         __uint(max_entries, profiler_bpf_max_function_id);
105         __type(key, u32);
106         __type(value, struct bpf_func_stats_data);
107 } bpf_func_stats SEC(".maps");
108 
109 struct {
110         __uint(type, BPF_MAP_TYPE_HASH);
111         __type(key, u32);
112         __type(value, bool);
113         __uint(max_entries, 16);
114 } allowed_devices SEC(".maps");
115 
116 struct {
117         __uint(type, BPF_MAP_TYPE_HASH);
118         __type(key, u64);
119         __type(value, bool);
120         __uint(max_entries, 1024);
121 } allowed_file_inodes SEC(".maps");
122 
123 struct {
124         __uint(type, BPF_MAP_TYPE_HASH);
125         __type(key, u64);
126         __type(value, bool);
127         __uint(max_entries, 1024);
128 } allowed_directory_inodes SEC(".maps");
129 
130 struct {
131         __uint(type, BPF_MAP_TYPE_HASH);
132         __type(key, u32);
133         __type(value, bool);
134         __uint(max_entries, 16);
135 } disallowed_exec_inodes SEC(".maps");
136 
137 static INLINE bool IS_ERR(const void* ptr)
138 {
139         return IS_ERR_VALUE((unsigned long)ptr);
140 }
141 
142 static INLINE u32 get_userspace_pid()
143 {
144         return bpf_get_current_pid_tgid() >> 32;
145 }
146 
147 static INLINE bool is_init_process(u32 tgid)
148 {
149         return tgid == 1 || tgid == 0;
150 }
151 
152 static INLINE unsigned long
153 probe_read_lim(void* dst, void* src, unsigned long len, unsigned long max)
154 {
155         len = len < max ? len : max;
156         if (len > 1) {
157                 if (bpf_probe_read_kernel(dst, len, src))
158                         return 0;
159         } else if (len == 1) {
160                 if (bpf_probe_read_kernel(dst, 1, src))
161                         return 0;
162         }
163         return len;
164 }
165 
166 static INLINE int get_var_spid_index(struct var_kill_data_arr_t* arr_struct,
167                                      int spid)
168 {
169 #ifdef UNROLL
170         __pragma_loop_unroll
171 #endif
172         for (int i = 0; i < ARRAY_SIZE(arr_struct->array); i++)
173                 if (arr_struct->array[i].meta.pid == spid)
174                         return i;
175         return -1;
176 }
177 
178 static INLINE void populate_ancestors(struct task_struct* task,
179                                       struct ancestors_data_t* ancestors_data)
180 {
181         struct task_struct* parent = task;
182         u32 num_ancestors, ppid;
183 
184         ancestors_data->num_ancestors = 0;
185 #ifdef UNROLL
186         __pragma_loop_unroll
187 #endif
188         for (num_ancestors = 0; num_ancestors < MAX_ANCESTORS; num_ancestors++) {
189                 parent = BPF_CORE_READ(parent, real_parent);
190                 if (parent == NULL)
191                         break;
192                 ppid = BPF_CORE_READ(parent, tgid);
193                 if (is_init_process(ppid))
194                         break;
195                 ancestors_data->ancestor_pids[num_ancestors] = ppid;
196                 ancestors_data->ancestor_exec_ids[num_ancestors] =
197                         BPF_CORE_READ(parent, self_exec_id);
198                 ancestors_data->ancestor_start_times[num_ancestors] =
199                         BPF_CORE_READ(parent, start_time);
200                 ancestors_data->num_ancestors = num_ancestors;
201         }
202 }
203 
204 static INLINE void* read_full_cgroup_path(struct kernfs_node* cgroup_node,
205                                           struct kernfs_node* cgroup_root_node,
206                                           void* payload,
207                                           int* root_pos)
208 {
209         void* payload_start = payload;
210         size_t filepart_length;
211 
212 #ifdef UNROLL
213         __pragma_loop_unroll
214 #endif
215         for (int i = 0; i < MAX_CGROUPS_PATH_DEPTH; i++) {
216                 filepart_length =
217                         bpf_probe_read_kernel_str(payload, MAX_PATH,
218                                                   BPF_CORE_READ(cgroup_node, name));
219                 if (!cgroup_node)
220                         return payload;
221                 if (cgroup_node == cgroup_root_node)
222                         *root_pos = payload - payload_start;
223                 if (bpf_cmp_likely(filepart_length, <=, MAX_PATH)) {
224                         payload += filepart_length;
225                 }
226                 cgroup_node = BPF_CORE_READ(cgroup_node, parent);
227         }
228         return payload;
229 }
230 
231 static ino_t get_inode_from_kernfs(struct kernfs_node* node)
232 {
233         struct kernfs_node___52* node52 = (void*)node;
234 
235         if (bpf_core_field_exists(node52->id.ino)) {
236                 barrier_var(node52);
237                 return BPF_CORE_READ(node52, id.ino);
238         } else {
239                 barrier_var(node);
240                 return (u64)BPF_CORE_READ(node, id);
241         }
242 }
243 
244 extern bool CONFIG_CGROUP_PIDS __kconfig __weak;
245 enum cgroup_subsys_id___local {
246         pids_cgrp_id___local = 123, /* value doesn't matter */
247 };
248 
249 static INLINE void* populate_cgroup_info(struct cgroup_data_t* cgroup_data,
250                                          struct task_struct* task,
251                                          void* payload)
252 {
253         struct kernfs_node* root_kernfs =
254                 BPF_CORE_READ(task, nsproxy, cgroup_ns, root_cset, dfl_cgrp, kn);
255         struct kernfs_node* proc_kernfs = BPF_CORE_READ(task, cgroups, dfl_cgrp, kn);
256 
257 #if __has_builtin(__builtin_preserve_enum_value)
258         if (ENABLE_CGROUP_V1_RESOLVER && CONFIG_CGROUP_PIDS) {
259                 int cgrp_id = bpf_core_enum_value(enum cgroup_subsys_id___local,
260                                                   pids_cgrp_id___local);
261 #ifdef UNROLL
262                 __pragma_loop_unroll
263 #endif
264                 for (int i = 0; i < CGROUP_SUBSYS_COUNT; i++) {
265                         struct cgroup_subsys_state* subsys =
266                                 BPF_CORE_READ(task, cgroups, subsys[i]);
267                         if (subsys != NULL) {
268                                 int subsys_id = BPF_CORE_READ(subsys, ss, id);
269                                 if (subsys_id == cgrp_id) {
270                                         proc_kernfs = BPF_CORE_READ(subsys, cgroup, kn);
271                                         root_kernfs = BPF_CORE_READ(subsys, ss, root, kf_root, kn);
272                                         break;
273                                 }
274                         }
275                 }
276         }
277 #endif
278 
279         cgroup_data->cgroup_root_inode = get_inode_from_kernfs(root_kernfs);
280         cgroup_data->cgroup_proc_inode = get_inode_from_kernfs(proc_kernfs);
281 
282         if (bpf_core_field_exists(root_kernfs->iattr->ia_mtime)) {
283                 cgroup_data->cgroup_root_mtime =
284                         BPF_CORE_READ(root_kernfs, iattr, ia_mtime.tv_nsec);
285                 cgroup_data->cgroup_proc_mtime =
286                         BPF_CORE_READ(proc_kernfs, iattr, ia_mtime.tv_nsec);
287         } else {
288                 struct kernfs_iattrs___52* root_iattr =
289                         (struct kernfs_iattrs___52*)BPF_CORE_READ(root_kernfs, iattr);
290                 cgroup_data->cgroup_root_mtime =
291                         BPF_CORE_READ(root_iattr, ia_iattr.ia_mtime.tv_nsec);
292 
293                 struct kernfs_iattrs___52* proc_iattr =
294                         (struct kernfs_iattrs___52*)BPF_CORE_READ(proc_kernfs, iattr);
295                 cgroup_data->cgroup_proc_mtime =
296                         BPF_CORE_READ(proc_iattr, ia_iattr.ia_mtime.tv_nsec);
297         }
298 
299         cgroup_data->cgroup_root_length = 0;
300         cgroup_data->cgroup_proc_length = 0;
301         cgroup_data->cgroup_full_length = 0;
302 
303         size_t cgroup_root_length =
304                 bpf_probe_read_kernel_str(payload, MAX_PATH,
305                                           BPF_CORE_READ(root_kernfs, name));
306         if (bpf_cmp_likely(cgroup_root_length, <=, MAX_PATH)) {
307                 cgroup_data->cgroup_root_length = cgroup_root_length;
308                 payload += cgroup_root_length;
309         }
310 
311         size_t cgroup_proc_length =
312                 bpf_probe_read_kernel_str(payload, MAX_PATH,
313                                           BPF_CORE_READ(proc_kernfs, name));
314         if (bpf_cmp_likely(cgroup_proc_length, <=, MAX_PATH)) {
315                 cgroup_data->cgroup_proc_length = cgroup_proc_length;
316                 payload += cgroup_proc_length;
317         }
318 
319         if (FETCH_CGROUPS_FROM_BPF) {
320                 cgroup_data->cgroup_full_path_root_pos = -1;
321                 void* payload_end_pos = read_full_cgroup_path(proc_kernfs, root_kernfs, payload,
322                                                               &cgroup_data->cgroup_full_path_root_pos);
323                 cgroup_data->cgroup_full_length = payload_end_pos - payload;
324                 payload = payload_end_pos;
325         }
326 
327         return (void*)payload;
328 }
329 
330 static INLINE void* populate_var_metadata(struct var_metadata_t* metadata,
331                                           struct task_struct* task,
332                                           u32 pid, void* payload)
333 {
334         u64 uid_gid = bpf_get_current_uid_gid();
335 
336         metadata->uid = (u32)uid_gid;
337         metadata->gid = uid_gid >> 32;
338         metadata->pid = pid;
339         metadata->exec_id = BPF_CORE_READ(task, self_exec_id);
340         metadata->start_time = BPF_CORE_READ(task, start_time);
341         metadata->comm_length = 0;
342 
343         size_t comm_length = bpf_core_read_str(payload, TASK_COMM_LEN, &task->comm);
344         if (bpf_cmp_likely(comm_length, <=, TASK_COMM_LEN)) {
345                 metadata->comm_length = comm_length;
346                 payload += comm_length;
347         }
348 
349         return (void*)payload;
350 }
351 
352 static INLINE struct var_kill_data_t*
353 get_var_kill_data(struct pt_regs* ctx, int spid, int tpid, int sig)
354 {
355         int zero = 0;
356         struct var_kill_data_t* kill_data = bpf_map_lookup_elem(&data_heap, &zero);
357 
358         if (kill_data == NULL)
359                 return NULL;
360         struct task_struct* task = (struct task_struct*)bpf_get_current_task();
361 
362         void* payload = populate_var_metadata(&kill_data->meta, task, spid, kill_data->payload);
363         payload = populate_cgroup_info(&kill_data->cgroup_data, task, payload);
364         size_t payload_length = payload - (void*)kill_data->payload;
365         kill_data->payload_length = payload_length;
366         populate_ancestors(task, &kill_data->ancestors_info);
367         kill_data->meta.type = KILL_EVENT;
368         kill_data->kill_target_pid = tpid;
369         kill_data->kill_sig = sig;
370         kill_data->kill_count = 1;
371         kill_data->last_kill_time = bpf_ktime_get_ns();
372         return kill_data;
373 }
374 
375 static INLINE int trace_var_sys_kill(void* ctx, int tpid, int sig)
376 {
377         if ((KILL_SIGNALS & (1ULL << sig)) == 0)
378                 return 0;
379 
380         u32 spid = get_userspace_pid();
381         struct var_kill_data_arr_t* arr_struct = bpf_map_lookup_elem(&var_tpid_to_data, &tpid);
382 
383         if (arr_struct == NULL) {
384                 struct var_kill_data_t* kill_data = get_var_kill_data(ctx, spid, tpid, sig);
385                 int zero = 0;
386 
387                 if (kill_data == NULL)
388                         return 0;
389                 arr_struct = bpf_map_lookup_elem(&data_heap, &zero);
390                 if (arr_struct == NULL)
391                         return 0;
392                 bpf_probe_read_kernel(&arr_struct->array[0],
393                                       sizeof(arr_struct->array[0]), kill_data);
394         } else {
395                 int index = get_var_spid_index(arr_struct, spid);
396 
397                 if (index == -1) {
398                         struct var_kill_data_t* kill_data =
399                                 get_var_kill_data(ctx, spid, tpid, sig);
400                         if (kill_data == NULL)
401                                 return 0;
402 #ifdef UNROLL
403                         __pragma_loop_unroll
404 #endif
405                         for (int i = 0; i < ARRAY_SIZE(arr_struct->array); i++)
406                                 if (arr_struct->array[i].meta.pid == 0) {
407                                         bpf_probe_read_kernel(&arr_struct->array[i],
408                                                               sizeof(arr_struct->array[i]),
409                                                               kill_data);
410                                         bpf_map_update_elem(&var_tpid_to_data, &tpid,
411                                                             arr_struct, 0);
412 
413                                         return 0;
414                                 }
415                         return 0;
416                 }
417 
418                 struct var_kill_data_t* kill_data = &arr_struct->array[index];
419 
420                 u64 delta_sec =
421                         (bpf_ktime_get_ns() - kill_data->last_kill_time) / 1000000000;
422 
423                 if (delta_sec < STALE_INFO) {
424                         kill_data->kill_count++;
425                         kill_data->last_kill_time = bpf_ktime_get_ns();
426                         bpf_probe_read_kernel(&arr_struct->array[index],
427                                               sizeof(arr_struct->array[index]),
428                                               kill_data);
429                 } else {
430                         struct var_kill_data_t* kill_data =
431                                 get_var_kill_data(ctx, spid, tpid, sig);
432                         if (kill_data == NULL)
433                                 return 0;
434                         bpf_probe_read_kernel(&arr_struct->array[index],
435                                               sizeof(arr_struct->array[index]),
436                                               kill_data);
437                 }
438         }
439         bpf_map_update_elem(&var_tpid_to_data, &tpid, arr_struct, 0);
440         return 0;
441 }
442 
443 static INLINE void bpf_stats_enter(struct bpf_func_stats_ctx* bpf_stat_ctx,
444                                    enum bpf_function_id func_id)
445 {
446         int func_id_key = func_id;
447 
448         bpf_stat_ctx->start_time_ns = bpf_ktime_get_ns();
449         bpf_stat_ctx->bpf_func_stats_data_val =
450                 bpf_map_lookup_elem(&bpf_func_stats, &func_id_key);
451         if (bpf_stat_ctx->bpf_func_stats_data_val)
452                 bpf_stat_ctx->bpf_func_stats_data_val->num_executions++;
453 }
454 
455 static INLINE void bpf_stats_exit(struct bpf_func_stats_ctx* bpf_stat_ctx)
456 {
457         if (bpf_stat_ctx->bpf_func_stats_data_val)
458                 bpf_stat_ctx->bpf_func_stats_data_val->time_elapsed_ns +=
459                         bpf_ktime_get_ns() - bpf_stat_ctx->start_time_ns;
460 }
461 
462 static INLINE void
463 bpf_stats_pre_submit_var_perf_event(struct bpf_func_stats_ctx* bpf_stat_ctx,
464                                     struct var_metadata_t* meta)
465 {
466         if (bpf_stat_ctx->bpf_func_stats_data_val) {
467                 bpf_stat_ctx->bpf_func_stats_data_val->num_perf_events++;
468                 meta->bpf_stats_num_perf_events =
469                         bpf_stat_ctx->bpf_func_stats_data_val->num_perf_events;
470         }
471         meta->bpf_stats_start_ktime_ns = bpf_stat_ctx->start_time_ns;
472         meta->cpu_id = bpf_get_smp_processor_id();
473 }
474 
475 static INLINE size_t
476 read_absolute_file_path_from_dentry(struct dentry* filp_dentry, void* payload)
477 {
478         size_t length = 0;
479         size_t filepart_length;
480         struct dentry* parent_dentry;
481 
482 #ifdef UNROLL
483         __pragma_loop_unroll
484 #endif
485         for (int i = 0; i < MAX_PATH_DEPTH; i++) {
486                 filepart_length =
487                         bpf_probe_read_kernel_str(payload, MAX_PATH,
488                                                   BPF_CORE_READ(filp_dentry, d_name.name));
489                 bpf_nop_mov(filepart_length);
490                 if (bpf_cmp_unlikely(filepart_length, >, MAX_PATH))
491                         break;
492                 payload += filepart_length;
493                 length += filepart_length;
494 
495                 parent_dentry = BPF_CORE_READ(filp_dentry, d_parent);
496                 if (filp_dentry == parent_dentry)
497                         break;
498                 filp_dentry = parent_dentry;
499         }
500 
501         return length;
502 }
503 
504 static INLINE bool
505 is_ancestor_in_allowed_inodes(struct dentry* filp_dentry)
506 {
507         struct dentry* parent_dentry;
508 #ifdef UNROLL
509         __pragma_loop_unroll
510 #endif
511         for (int i = 0; i < MAX_PATH_DEPTH; i++) {
512                 u64 dir_ino = BPF_CORE_READ(filp_dentry, d_inode, i_ino);
513                 bool* allowed_dir = bpf_map_lookup_elem(&allowed_directory_inodes, &dir_ino);
514 
515                 if (allowed_dir != NULL)
516                         return true;
517                 parent_dentry = BPF_CORE_READ(filp_dentry, d_parent);
518                 if (filp_dentry == parent_dentry)
519                         break;
520                 filp_dentry = parent_dentry;
521         }
522         return false;
523 }
524 
525 static INLINE bool is_dentry_allowed_for_filemod(struct dentry* file_dentry,
526                                                  u32* device_id,
527                                                  u64* file_ino)
528 {
529         u32 dev_id = BPF_CORE_READ(file_dentry, d_sb, s_dev);
530         *device_id = dev_id;
531         bool* allowed_device = bpf_map_lookup_elem(&allowed_devices, &dev_id);
532 
533         if (allowed_device == NULL)
534                 return false;
535 
536         u64 ino = BPF_CORE_READ(file_dentry, d_inode, i_ino);
537         *file_ino = ino;
538         bool* allowed_file = bpf_map_lookup_elem(&allowed_file_inodes, &ino);
539 
540         if (allowed_file == NULL)
541                 if (!is_ancestor_in_allowed_inodes(BPF_CORE_READ(file_dentry, d_parent)))
542                         return false;
543         return true;
544 }
545 
546 SEC("kprobe/proc_sys_write")
547 ssize_t BPF_KPROBE(kprobe__proc_sys_write,
548                    struct file* filp, const char* buf,
549                    size_t count, loff_t* ppos)
550 {
551         struct bpf_func_stats_ctx stats_ctx;
552         bpf_stats_enter(&stats_ctx, profiler_bpf_proc_sys_write);
553 
554         u32 pid = get_userspace_pid();
555         int zero = 0;
556         struct var_sysctl_data_t* sysctl_data =
557                 bpf_map_lookup_elem(&data_heap, &zero);
558         if (!sysctl_data)
559                 goto out;
560 
561         struct task_struct* task = (struct task_struct*)bpf_get_current_task();
562         sysctl_data->meta.type = SYSCTL_EVENT;
563         void* payload = populate_var_metadata(&sysctl_data->meta, task, pid, sysctl_data->payload);
564         payload = populate_cgroup_info(&sysctl_data->cgroup_data, task, payload);
565 
566         populate_ancestors(task, &sysctl_data->ancestors_info);
567 
568         sysctl_data->sysctl_val_length = 0;
569         sysctl_data->sysctl_path_length = 0;
570 
571         size_t sysctl_val_length = bpf_probe_read_kernel_str(payload,
572                                                              CTL_MAXNAME, buf);
573         if (bpf_cmp_likely(sysctl_val_length, <=, CTL_MAXNAME)) {
574                 sysctl_data->sysctl_val_length = sysctl_val_length;
575                 payload += sysctl_val_length;
576         }
577 
578         size_t sysctl_path_length =
579                 bpf_probe_read_kernel_str(payload, MAX_PATH,
580                                           BPF_CORE_READ(filp, f_path.dentry,
581                                                         d_name.name));
582         if (bpf_cmp_likely(sysctl_path_length, <=, MAX_PATH)) {
583                 sysctl_data->sysctl_path_length = sysctl_path_length;
584                 payload += sysctl_path_length;
585         }
586 
587         bpf_stats_pre_submit_var_perf_event(&stats_ctx, &sysctl_data->meta);
588         unsigned long data_len = payload - (void*)sysctl_data;
589         data_len = data_len > sizeof(struct var_sysctl_data_t)
590                 ? sizeof(struct var_sysctl_data_t)
591                 : data_len;
592         bpf_perf_event_output(ctx, &events, BPF_F_CURRENT_CPU, sysctl_data, data_len);
593 out:
594         bpf_stats_exit(&stats_ctx);
595         return 0;
596 }
597 
598 SEC("tracepoint/syscalls/sys_enter_kill")
599 int tracepoint__syscalls__sys_enter_kill(struct syscall_trace_enter* ctx)
600 {
601         struct bpf_func_stats_ctx stats_ctx;
602 
603         bpf_stats_enter(&stats_ctx, profiler_bpf_sys_enter_kill);
604         int pid = ctx->args[0];
605         int sig = ctx->args[1];
606         int ret = trace_var_sys_kill(ctx, pid, sig);
607         bpf_stats_exit(&stats_ctx);
608         return ret;
609 };
610 
611 SEC("raw_tracepoint/sched_process_exit")
612 int raw_tracepoint__sched_process_exit(void* ctx)
613 {
614         int zero = 0;
615         struct bpf_func_stats_ctx stats_ctx;
616         bpf_stats_enter(&stats_ctx, profiler_bpf_sched_process_exit);
617 
618         u32 tpid = get_userspace_pid();
619 
620         struct var_kill_data_arr_t* arr_struct = bpf_map_lookup_elem(&var_tpid_to_data, &tpid);
621         struct var_kill_data_t* kill_data = bpf_map_lookup_elem(&data_heap, &zero);
622 
623         if (arr_struct == NULL || kill_data == NULL)
624                 goto out;
625 
626         struct task_struct* task = (struct task_struct*)bpf_get_current_task();
627         struct kernfs_node* proc_kernfs = BPF_CORE_READ(task, cgroups, dfl_cgrp, kn);
628 
629 #ifdef UNROLL
630         __pragma_loop_unroll
631 #endif
632         for (int i = 0; i < ARRAY_SIZE(arr_struct->array); i++) {
633                 struct var_kill_data_t* past_kill_data = &arr_struct->array[i];
634 
635                 if (past_kill_data != NULL && past_kill_data->kill_target_pid == (pid_t)tpid) {
636                         bpf_probe_read_kernel(kill_data, sizeof(*past_kill_data),
637                                               past_kill_data);
638                         void* payload = kill_data->payload;
639                         size_t offset = kill_data->payload_length;
640                         if (offset >= MAX_METADATA_PAYLOAD_LEN + MAX_CGROUP_PAYLOAD_LEN)
641                                 return 0;
642                         payload += offset;
643 
644                         kill_data->kill_target_name_length = 0;
645                         kill_data->kill_target_cgroup_proc_length = 0;
646 
647                         size_t comm_length = bpf_core_read_str(payload, TASK_COMM_LEN, &task->comm);
648                         if (bpf_cmp_likely(comm_length, <=, TASK_COMM_LEN)) {
649                                 kill_data->kill_target_name_length = comm_length;
650                                 payload += comm_length;
651                         }
652 
653                         size_t cgroup_proc_length =
654                                 bpf_probe_read_kernel_str(payload,
655                                                           KILL_TARGET_LEN,
656                                                           BPF_CORE_READ(proc_kernfs, name));
657                         if (bpf_cmp_likely(cgroup_proc_length, <=, KILL_TARGET_LEN)) {
658                                 kill_data->kill_target_cgroup_proc_length = cgroup_proc_length;
659                                 payload += cgroup_proc_length;
660                         }
661 
662                         bpf_stats_pre_submit_var_perf_event(&stats_ctx, &kill_data->meta);
663                         unsigned long data_len = (void*)payload - (void*)kill_data;
664                         data_len = data_len > sizeof(struct var_kill_data_t)
665                                 ? sizeof(struct var_kill_data_t)
666                                 : data_len;
667                         bpf_perf_event_output(ctx, &events, BPF_F_CURRENT_CPU, kill_data, data_len);
668                 }
669         }
670         bpf_map_delete_elem(&var_tpid_to_data, &tpid);
671 out:
672         bpf_stats_exit(&stats_ctx);
673         return 0;
674 }
675 
676 SEC("raw_tracepoint/sched_process_exec")
677 int raw_tracepoint__sched_process_exec(struct bpf_raw_tracepoint_args* ctx)
678 {
679         struct bpf_func_stats_ctx stats_ctx;
680         bpf_stats_enter(&stats_ctx, profiler_bpf_sched_process_exec);
681 
682         struct linux_binprm* bprm = (struct linux_binprm*)ctx->args[2];
683         u64 inode = BPF_CORE_READ(bprm, file, f_inode, i_ino);
684 
685         bool* should_filter_binprm = bpf_map_lookup_elem(&disallowed_exec_inodes, &inode);
686         if (should_filter_binprm != NULL)
687                 goto out;
688 
689         int zero = 0;
690         struct var_exec_data_t* proc_exec_data = bpf_map_lookup_elem(&data_heap, &zero);
691         if (!proc_exec_data)
692                 goto out;
693 
694         if (INODE_FILTER && inode != INODE_FILTER)
695                 return 0;
696 
697         u32 pid = get_userspace_pid();
698         struct task_struct* task = (struct task_struct*)bpf_get_current_task();
699 
700         proc_exec_data->meta.type = EXEC_EVENT;
701         proc_exec_data->bin_path_length = 0;
702         proc_exec_data->cmdline_length = 0;
703         proc_exec_data->environment_length = 0;
704         void* payload = populate_var_metadata(&proc_exec_data->meta, task, pid,
705                                               proc_exec_data->payload);
706         payload = populate_cgroup_info(&proc_exec_data->cgroup_data, task, payload);
707 
708         struct task_struct* parent_task = BPF_CORE_READ(task, real_parent);
709         proc_exec_data->parent_pid = BPF_CORE_READ(parent_task, tgid);
710         proc_exec_data->parent_uid = BPF_CORE_READ(parent_task, real_cred, uid.val);
711         proc_exec_data->parent_exec_id = BPF_CORE_READ(parent_task, self_exec_id);
712         proc_exec_data->parent_start_time = BPF_CORE_READ(parent_task, start_time);
713 
714         const char* filename = BPF_CORE_READ(bprm, filename);
715         size_t bin_path_length =
716                 bpf_probe_read_kernel_str(payload, MAX_FILENAME_LEN, filename);
717         if (bpf_cmp_likely(bin_path_length, <=, MAX_FILENAME_LEN)) {
718                 proc_exec_data->bin_path_length = bin_path_length;
719                 payload += bin_path_length;
720         }
721 
722         void* arg_start = (void*)BPF_CORE_READ(task, mm, arg_start);
723         void* arg_end = (void*)BPF_CORE_READ(task, mm, arg_end);
724         unsigned int cmdline_length = probe_read_lim(payload, arg_start,
725                                                      arg_end - arg_start, MAX_ARGS_LEN);
726 
727         if (bpf_cmp_likely(cmdline_length, <=, MAX_ARGS_LEN)) {
728                 proc_exec_data->cmdline_length = cmdline_length;
729                 payload += cmdline_length;
730         }
731 
732         if (READ_ENVIRON_FROM_EXEC) {
733                 void* env_start = (void*)BPF_CORE_READ(task, mm, env_start);
734                 void* env_end = (void*)BPF_CORE_READ(task, mm, env_end);
735                 unsigned long env_len = probe_read_lim(payload, env_start,
736                                                        env_end - env_start, MAX_ENVIRON_LEN);
737                 if (cmdline_length <= MAX_ENVIRON_LEN) {
738                         proc_exec_data->environment_length = env_len;
739                         payload += env_len;
740                 }
741         }
742 
743         bpf_stats_pre_submit_var_perf_event(&stats_ctx, &proc_exec_data->meta);
744         unsigned long data_len = payload - (void*)proc_exec_data;
745         data_len = data_len > sizeof(struct var_exec_data_t)
746                 ? sizeof(struct var_exec_data_t)
747                 : data_len;
748         bpf_perf_event_output(ctx, &events, BPF_F_CURRENT_CPU, proc_exec_data, data_len);
749 out:
750         bpf_stats_exit(&stats_ctx);
751         return 0;
752 }
753 
754 SEC("kretprobe/do_filp_open")
755 int kprobe_ret__do_filp_open(struct pt_regs* ctx)
756 {
757         struct bpf_func_stats_ctx stats_ctx;
758         bpf_stats_enter(&stats_ctx, profiler_bpf_do_filp_open_ret);
759 
760         struct file* filp = (struct file*)PT_REGS_RC_CORE(ctx);
761 
762         if (filp == NULL || IS_ERR(filp))
763                 goto out;
764         unsigned int flags = BPF_CORE_READ(filp, f_flags);
765         if ((flags & (O_RDWR | O_WRONLY)) == 0)
766                 goto out;
767         if ((flags & O_TMPFILE) > 0)
768                 goto out;
769         struct inode* file_inode = BPF_CORE_READ(filp, f_inode);
770         umode_t mode = BPF_CORE_READ(file_inode, i_mode);
771         if (S_ISDIR(mode) || S_ISCHR(mode) || S_ISBLK(mode) || S_ISFIFO(mode) ||
772             S_ISSOCK(mode))
773                 goto out;
774 
775         struct dentry* filp_dentry = BPF_CORE_READ(filp, f_path.dentry);
776         u32 device_id = 0;
777         u64 file_ino = 0;
778         if (!is_dentry_allowed_for_filemod(filp_dentry, &device_id, &file_ino))
779                 goto out;
780 
781         int zero = 0;
782         struct var_filemod_data_t* filemod_data = bpf_map_lookup_elem(&data_heap, &zero);
783         if (!filemod_data)
784                 goto out;
785 
786         u32 pid = get_userspace_pid();
787         struct task_struct* task = (struct task_struct*)bpf_get_current_task();
788 
789         filemod_data->meta.type = FILEMOD_EVENT;
790         filemod_data->fmod_type = FMOD_OPEN;
791         filemod_data->dst_flags = flags;
792         filemod_data->src_inode = 0;
793         filemod_data->dst_inode = file_ino;
794         filemod_data->src_device_id = 0;
795         filemod_data->dst_device_id = device_id;
796         filemod_data->src_filepath_length = 0;
797         filemod_data->dst_filepath_length = 0;
798 
799         void* payload = populate_var_metadata(&filemod_data->meta, task, pid,
800                                               filemod_data->payload);
801         payload = populate_cgroup_info(&filemod_data->cgroup_data, task, payload);
802 
803         size_t len = read_absolute_file_path_from_dentry(filp_dentry, payload);
804         if (bpf_cmp_likely(len, <=, MAX_FILEPATH_LENGTH)) {
805                 payload += len;
806                 filemod_data->dst_filepath_length = len;
807         }
808         bpf_stats_pre_submit_var_perf_event(&stats_ctx, &filemod_data->meta);
809         unsigned long data_len = payload - (void*)filemod_data;
810         data_len = data_len > sizeof(*filemod_data) ? sizeof(*filemod_data) : data_len;
811         bpf_perf_event_output(ctx, &events, BPF_F_CURRENT_CPU, filemod_data, data_len);
812 out:
813         bpf_stats_exit(&stats_ctx);
814         return 0;
815 }
816 
817 SEC("kprobe/vfs_link")
818 int BPF_KPROBE(kprobe__vfs_link,
819                struct dentry* old_dentry, struct mnt_idmap *idmap,
820                struct inode* dir, struct dentry* new_dentry,
821                struct inode** delegated_inode)
822 {
823         struct bpf_func_stats_ctx stats_ctx;
824         bpf_stats_enter(&stats_ctx, profiler_bpf_vfs_link);
825 
826         u32 src_device_id = 0;
827         u64 src_file_ino = 0;
828         u32 dst_device_id = 0;
829         u64 dst_file_ino = 0;
830         if (!is_dentry_allowed_for_filemod(old_dentry, &src_device_id, &src_file_ino) &&
831             !is_dentry_allowed_for_filemod(new_dentry, &dst_device_id, &dst_file_ino))
832                 goto out;
833 
834         int zero = 0;
835         struct var_filemod_data_t* filemod_data = bpf_map_lookup_elem(&data_heap, &zero);
836         if (!filemod_data)
837                 goto out;
838 
839         u32 pid = get_userspace_pid();
840         struct task_struct* task = (struct task_struct*)bpf_get_current_task();
841 
842         filemod_data->meta.type = FILEMOD_EVENT;
843         filemod_data->fmod_type = FMOD_LINK;
844         filemod_data->dst_flags = 0;
845         filemod_data->src_inode = src_file_ino;
846         filemod_data->dst_inode = dst_file_ino;
847         filemod_data->src_device_id = src_device_id;
848         filemod_data->dst_device_id = dst_device_id;
849         filemod_data->src_filepath_length = 0;
850         filemod_data->dst_filepath_length = 0;
851 
852         void* payload = populate_var_metadata(&filemod_data->meta, task, pid,
853                                               filemod_data->payload);
854         payload = populate_cgroup_info(&filemod_data->cgroup_data, task, payload);
855 
856         size_t len = read_absolute_file_path_from_dentry(old_dentry, payload);
857         if (bpf_cmp_likely(len, <=, MAX_FILEPATH_LENGTH)) {
858                 payload += len;
859                 filemod_data->src_filepath_length = len;
860         }
861 
862         len = read_absolute_file_path_from_dentry(new_dentry, payload);
863         if (bpf_cmp_likely(len, <=, MAX_FILEPATH_LENGTH)) {
864                 payload += len;
865                 filemod_data->dst_filepath_length = len;
866         }
867 
868         bpf_stats_pre_submit_var_perf_event(&stats_ctx, &filemod_data->meta);
869         unsigned long data_len = payload - (void*)filemod_data;
870         data_len = data_len > sizeof(*filemod_data) ? sizeof(*filemod_data) : data_len;
871         bpf_perf_event_output(ctx, &events, BPF_F_CURRENT_CPU, filemod_data, data_len);
872 out:
873         bpf_stats_exit(&stats_ctx);
874         return 0;
875 }
876 
877 SEC("kprobe/vfs_symlink")
878 int BPF_KPROBE(kprobe__vfs_symlink, struct inode* dir, struct dentry* dentry,
879                const char* oldname)
880 {
881         struct bpf_func_stats_ctx stats_ctx;
882         bpf_stats_enter(&stats_ctx, profiler_bpf_vfs_symlink);
883 
884         u32 dst_device_id = 0;
885         u64 dst_file_ino = 0;
886         if (!is_dentry_allowed_for_filemod(dentry, &dst_device_id, &dst_file_ino))
887                 goto out;
888 
889         int zero = 0;
890         struct var_filemod_data_t* filemod_data = bpf_map_lookup_elem(&data_heap, &zero);
891         if (!filemod_data)
892                 goto out;
893 
894         u32 pid = get_userspace_pid();
895         struct task_struct* task = (struct task_struct*)bpf_get_current_task();
896 
897         filemod_data->meta.type = FILEMOD_EVENT;
898         filemod_data->fmod_type = FMOD_SYMLINK;
899         filemod_data->dst_flags = 0;
900         filemod_data->src_inode = 0;
901         filemod_data->dst_inode = dst_file_ino;
902         filemod_data->src_device_id = 0;
903         filemod_data->dst_device_id = dst_device_id;
904         filemod_data->src_filepath_length = 0;
905         filemod_data->dst_filepath_length = 0;
906 
907         void* payload = populate_var_metadata(&filemod_data->meta, task, pid,
908                                               filemod_data->payload);
909         payload = populate_cgroup_info(&filemod_data->cgroup_data, task, payload);
910 
911         size_t len = bpf_probe_read_kernel_str(payload, MAX_FILEPATH_LENGTH,
912                                                oldname);
913         if (bpf_cmp_likely(len, <=, MAX_FILEPATH_LENGTH)) {
914                 payload += len;
915                 filemod_data->src_filepath_length = len;
916         }
917         len = read_absolute_file_path_from_dentry(dentry, payload);
918         if (bpf_cmp_likely(len, <=, MAX_FILEPATH_LENGTH)) {
919                 payload += len;
920                 filemod_data->dst_filepath_length = len;
921         }
922         bpf_stats_pre_submit_var_perf_event(&stats_ctx, &filemod_data->meta);
923         unsigned long data_len = payload - (void*)filemod_data;
924         data_len = data_len > sizeof(*filemod_data) ? sizeof(*filemod_data) : data_len;
925         bpf_perf_event_output(ctx, &events, BPF_F_CURRENT_CPU, filemod_data, data_len);
926 out:
927         bpf_stats_exit(&stats_ctx);
928         return 0;
929 }
930 
931 SEC("raw_tracepoint/sched_process_fork")
932 int raw_tracepoint__sched_process_fork(struct bpf_raw_tracepoint_args* ctx)
933 {
934         struct bpf_func_stats_ctx stats_ctx;
935         bpf_stats_enter(&stats_ctx, profiler_bpf_sched_process_fork);
936 
937         int zero = 0;
938         struct var_fork_data_t* fork_data = bpf_map_lookup_elem(&data_heap, &zero);
939         if (!fork_data)
940                 goto out;
941 
942         struct task_struct* parent = (struct task_struct*)ctx->args[0];
943         struct task_struct* child = (struct task_struct*)ctx->args[1];
944         fork_data->meta.type = FORK_EVENT;
945 
946         void* payload = populate_var_metadata(&fork_data->meta, child,
947                                               BPF_CORE_READ(child, pid), fork_data->payload);
948         fork_data->parent_pid = BPF_CORE_READ(parent, pid);
949         fork_data->parent_exec_id = BPF_CORE_READ(parent, self_exec_id);
950         fork_data->parent_start_time = BPF_CORE_READ(parent, start_time);
951         bpf_stats_pre_submit_var_perf_event(&stats_ctx, &fork_data->meta);
952 
953         unsigned long data_len = payload - (void*)fork_data;
954         data_len = data_len > sizeof(*fork_data) ? sizeof(*fork_data) : data_len;
955         bpf_perf_event_output(ctx, &events, BPF_F_CURRENT_CPU, fork_data, data_len);
956 out:
957         bpf_stats_exit(&stats_ctx);
958         return 0;
959 }
960 char _license[] SEC("license") = "GPL";
961 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php