Profile.cpp 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511
  1. /*
  2. * Copyright (c) 2018-2021, Andreas Kling <kling@serenityos.org>
  3. *
  4. * SPDX-License-Identifier: BSD-2-Clause
  5. */
  6. #include "Profile.h"
  7. #include "DisassemblyModel.h"
  8. #include "ProfileModel.h"
  9. #include "SamplesModel.h"
  10. #include <AK/HashTable.h>
  11. #include <AK/LexicalPath.h>
  12. #include <AK/MappedFile.h>
  13. #include <AK/NonnullOwnPtrVector.h>
  14. #include <AK/QuickSort.h>
  15. #include <AK/RefPtr.h>
  16. #include <LibCore/File.h>
  17. #include <LibELF/Image.h>
  18. #include <sys/stat.h>
  19. namespace Profiler {
  20. static void sort_profile_nodes(Vector<NonnullRefPtr<ProfileNode>>& nodes)
  21. {
  22. quick_sort(nodes.begin(), nodes.end(), [](auto& a, auto& b) {
  23. return a->event_count() >= b->event_count();
  24. });
  25. for (auto& child : nodes)
  26. child->sort_children();
  27. }
  28. Profile::Profile(Vector<Process> processes, Vector<Event> events)
  29. : m_processes(move(processes))
  30. , m_events(move(events))
  31. {
  32. m_first_timestamp = m_events.first().timestamp;
  33. m_last_timestamp = m_events.last().timestamp;
  34. m_model = ProfileModel::create(*this);
  35. m_samples_model = SamplesModel::create(*this);
  36. rebuild_tree();
  37. }
  38. GUI::Model& Profile::model()
  39. {
  40. return *m_model;
  41. }
  42. GUI::Model& Profile::samples_model()
  43. {
  44. return *m_samples_model;
  45. }
  46. void Profile::rebuild_tree()
  47. {
  48. Vector<NonnullRefPtr<ProfileNode>> roots;
  49. auto find_or_create_process_node = [this, &roots](pid_t pid, EventSerialNumber serial) -> ProfileNode& {
  50. auto* process = find_process(pid, serial);
  51. if (!process) {
  52. dbgln("Profile contains event for unknown process with pid={}, serial={}", pid, serial.to_number());
  53. VERIFY_NOT_REACHED();
  54. }
  55. for (auto root : roots) {
  56. if (&root->process() == process)
  57. return root;
  58. }
  59. auto new_root = ProfileNode::create_process_node(*process);
  60. roots.append(new_root);
  61. return new_root;
  62. };
  63. HashTable<FlatPtr> live_allocations;
  64. for_each_event_in_filter_range([&](auto& event) {
  65. if (event.type == "malloc"sv)
  66. live_allocations.set(event.ptr);
  67. else if (event.type == "free"sv)
  68. live_allocations.remove(event.ptr);
  69. });
  70. m_filtered_event_indices.clear();
  71. for (size_t event_index = 0; event_index < m_events.size(); ++event_index) {
  72. auto& event = m_events.at(event_index);
  73. if (has_timestamp_filter_range()) {
  74. auto timestamp = event.timestamp;
  75. if (timestamp < m_timestamp_filter_range_start || timestamp > m_timestamp_filter_range_end)
  76. continue;
  77. }
  78. if (!process_filter_contains(event.pid, event.serial))
  79. continue;
  80. m_filtered_event_indices.append(event_index);
  81. if (event.type == "malloc"sv && !live_allocations.contains(event.ptr))
  82. continue;
  83. if (event.type == "free"sv)
  84. continue;
  85. auto for_each_frame = [&]<typename Callback>(Callback callback) {
  86. if (!m_inverted) {
  87. for (size_t i = 0; i < event.frames.size(); ++i) {
  88. if (callback(event.frames.at(i), i == event.frames.size() - 1) == IterationDecision::Break)
  89. break;
  90. }
  91. } else {
  92. for (ssize_t i = event.frames.size() - 1; i >= 0; --i) {
  93. if (callback(event.frames.at(i), static_cast<size_t>(i) == event.frames.size() - 1) == IterationDecision::Break)
  94. break;
  95. }
  96. }
  97. };
  98. if (!m_show_top_functions) {
  99. ProfileNode* node = nullptr;
  100. auto& process_node = find_or_create_process_node(event.pid, event.serial);
  101. process_node.increment_event_count();
  102. for_each_frame([&](const Frame& frame, bool is_innermost_frame) {
  103. auto& object_name = frame.object_name;
  104. auto& symbol = frame.symbol;
  105. auto& address = frame.address;
  106. auto& offset = frame.offset;
  107. if (symbol.is_empty())
  108. return IterationDecision::Break;
  109. // FIXME: More cheating with intentional mixing of TID/PID here:
  110. if (!node)
  111. node = &process_node;
  112. node = &node->find_or_create_child(object_name, symbol, address, offset, event.timestamp, event.pid);
  113. node->increment_event_count();
  114. if (is_innermost_frame) {
  115. node->add_event_address(address);
  116. node->increment_self_count();
  117. }
  118. return IterationDecision::Continue;
  119. });
  120. } else {
  121. auto& process_node = find_or_create_process_node(event.pid, event.serial);
  122. process_node.increment_event_count();
  123. for (size_t i = 0; i < event.frames.size(); ++i) {
  124. ProfileNode* node = nullptr;
  125. ProfileNode* root = nullptr;
  126. for (size_t j = i; j < event.frames.size(); ++j) {
  127. auto& frame = event.frames.at(j);
  128. auto& object_name = frame.object_name;
  129. auto& symbol = frame.symbol;
  130. auto& address = frame.address;
  131. auto& offset = frame.offset;
  132. if (symbol.is_empty())
  133. break;
  134. // FIXME: More PID/TID mixing cheats here:
  135. if (!node) {
  136. node = &find_or_create_process_node(event.pid, event.serial);
  137. node = &node->find_or_create_child(object_name, symbol, address, offset, event.timestamp, event.pid);
  138. root = node;
  139. root->will_track_seen_events(m_events.size());
  140. } else {
  141. node = &node->find_or_create_child(object_name, symbol, address, offset, event.timestamp, event.pid);
  142. }
  143. if (!root->has_seen_event(event_index)) {
  144. root->did_see_event(event_index);
  145. root->increment_event_count();
  146. } else if (node != root) {
  147. node->increment_event_count();
  148. }
  149. if (j == event.frames.size() - 1) {
  150. node->add_event_address(address);
  151. node->increment_self_count();
  152. }
  153. }
  154. }
  155. }
  156. }
  157. sort_profile_nodes(roots);
  158. m_roots = move(roots);
  159. m_model->update();
  160. }
  161. Result<NonnullOwnPtr<Profile>, String> Profile::load_from_perfcore_file(const StringView& path)
  162. {
  163. auto file = Core::File::construct(path);
  164. if (!file->open(Core::OpenMode::ReadOnly))
  165. return String::formatted("Unable to open {}, error: {}", path, file->error_string());
  166. auto json = JsonValue::from_string(file->read_all());
  167. if (!json.has_value() || !json.value().is_object())
  168. return String { "Invalid perfcore format (not a JSON object)" };
  169. auto& object = json.value().as_object();
  170. auto file_or_error = MappedFile::map("/boot/Kernel.debug");
  171. OwnPtr<ELF::Image> kernel_elf;
  172. if (!file_or_error.is_error())
  173. kernel_elf = make<ELF::Image>(file_or_error.value()->bytes());
  174. auto events_value = object.get_ptr("events");
  175. if (!events_value || !events_value->is_array())
  176. return String { "Malformed profile (events is not an array)" };
  177. auto& perf_events = events_value->as_array();
  178. NonnullOwnPtrVector<Process> all_processes;
  179. HashMap<pid_t, Process*> current_processes;
  180. Vector<Event> events;
  181. EventSerialNumber next_serial;
  182. for (auto& perf_event_value : perf_events.values()) {
  183. auto& perf_event = perf_event_value.as_object();
  184. Event event;
  185. event.serial = next_serial;
  186. next_serial.increment();
  187. event.timestamp = perf_event.get("timestamp").to_number<u64>();
  188. event.lost_samples = perf_event.get("lost_samples").to_number<u32>();
  189. event.type = perf_event.get("type").to_string();
  190. event.pid = perf_event.get("pid").to_i32();
  191. event.tid = perf_event.get("tid").to_i32();
  192. if (event.type == "malloc"sv) {
  193. event.ptr = perf_event.get("ptr").to_number<FlatPtr>();
  194. event.size = perf_event.get("size").to_number<size_t>();
  195. } else if (event.type == "free"sv) {
  196. event.ptr = perf_event.get("ptr").to_number<FlatPtr>();
  197. } else if (event.type == "mmap"sv) {
  198. event.ptr = perf_event.get("ptr").to_number<FlatPtr>();
  199. event.size = perf_event.get("size").to_number<size_t>();
  200. event.name = perf_event.get("name").to_string();
  201. auto it = current_processes.find(event.pid);
  202. if (it != current_processes.end())
  203. it->value->library_metadata.handle_mmap(event.ptr, event.size, event.name);
  204. continue;
  205. } else if (event.type == "munmap"sv) {
  206. event.ptr = perf_event.get("ptr").to_number<FlatPtr>();
  207. event.size = perf_event.get("size").to_number<size_t>();
  208. continue;
  209. } else if (event.type == "process_create"sv) {
  210. event.parent_pid = perf_event.get("parent_pid").to_number<FlatPtr>();
  211. event.executable = perf_event.get("executable").to_string();
  212. auto sampled_process = adopt_own(*new Process {
  213. .pid = event.pid,
  214. .executable = event.executable,
  215. .basename = LexicalPath::basename(event.executable),
  216. .start_valid = event.serial,
  217. .end_valid = {},
  218. });
  219. current_processes.set(sampled_process->pid, sampled_process);
  220. all_processes.append(move(sampled_process));
  221. continue;
  222. } else if (event.type == "process_exec"sv) {
  223. event.executable = perf_event.get("executable").to_string();
  224. auto old_process = current_processes.get(event.pid).value();
  225. old_process->end_valid = event.serial;
  226. current_processes.remove(event.pid);
  227. auto sampled_process = adopt_own(*new Process {
  228. .pid = event.pid,
  229. .executable = event.executable,
  230. .basename = LexicalPath::basename(event.executable),
  231. .start_valid = event.serial,
  232. .end_valid = {},
  233. });
  234. current_processes.set(sampled_process->pid, sampled_process);
  235. all_processes.append(move(sampled_process));
  236. continue;
  237. } else if (event.type == "process_exit"sv) {
  238. auto old_process = current_processes.get(event.pid).value();
  239. old_process->end_valid = event.serial;
  240. current_processes.remove(event.pid);
  241. continue;
  242. } else if (event.type == "thread_create"sv) {
  243. event.parent_tid = perf_event.get("parent_tid").to_i32();
  244. auto it = current_processes.find(event.pid);
  245. if (it != current_processes.end())
  246. it->value->handle_thread_create(event.tid, event.serial);
  247. continue;
  248. } else if (event.type == "thread_exit"sv) {
  249. auto it = current_processes.find(event.pid);
  250. if (it != current_processes.end())
  251. it->value->handle_thread_exit(event.tid, event.serial);
  252. continue;
  253. }
  254. // FIXME: Use /proc for this
  255. #if ARCH(I386)
  256. FlatPtr kernel_base = 0xc0000000;
  257. #else
  258. FlatPtr kernel_base = 0x2000000000;
  259. #endif
  260. auto* stack = perf_event.get_ptr("stack");
  261. VERIFY(stack);
  262. auto& stack_array = stack->as_array();
  263. for (ssize_t i = stack_array.values().size() - 1; i >= 0; --i) {
  264. auto& frame = stack_array.at(i);
  265. auto ptr = frame.to_number<u64>();
  266. u32 offset = 0;
  267. FlyString object_name;
  268. String symbol;
  269. if (ptr >= kernel_base) {
  270. if (kernel_elf) {
  271. symbol = kernel_elf->symbolicate(ptr, &offset);
  272. } else {
  273. symbol = String::formatted("?? <{:p}>", ptr);
  274. }
  275. } else {
  276. auto it = current_processes.find(event.pid);
  277. // FIXME: This logic is kinda gnarly, find a way to clean it up.
  278. LibraryMetadata* library_metadata {};
  279. if (it != current_processes.end())
  280. library_metadata = &it->value->library_metadata;
  281. if (auto* library = library_metadata ? library_metadata->library_containing(ptr) : nullptr) {
  282. object_name = library->name;
  283. symbol = library->symbolicate(ptr, &offset);
  284. } else {
  285. symbol = String::formatted("?? <{:p}>", ptr);
  286. }
  287. }
  288. event.frames.append({ object_name, symbol, (FlatPtr)ptr, offset });
  289. }
  290. if (event.frames.size() < 2)
  291. continue;
  292. FlatPtr innermost_frame_address = event.frames.at(1).address;
  293. event.in_kernel = innermost_frame_address >= kernel_base;
  294. events.append(move(event));
  295. }
  296. if (events.is_empty())
  297. return String { "No events captured (targeted process was never on CPU)" };
  298. quick_sort(all_processes, [](auto& a, auto& b) {
  299. if (a.pid == b.pid)
  300. return a.start_valid < b.start_valid;
  301. else
  302. return a.pid < b.pid;
  303. });
  304. Vector<Process> processes;
  305. for (auto& it : all_processes)
  306. processes.append(move(it));
  307. return adopt_own(*new Profile(move(processes), move(events)));
  308. }
  309. void ProfileNode::sort_children()
  310. {
  311. sort_profile_nodes(m_children);
  312. }
  313. void Profile::set_timestamp_filter_range(u64 start, u64 end)
  314. {
  315. if (m_has_timestamp_filter_range && m_timestamp_filter_range_start == start && m_timestamp_filter_range_end == end)
  316. return;
  317. m_has_timestamp_filter_range = true;
  318. m_timestamp_filter_range_start = min(start, end);
  319. m_timestamp_filter_range_end = max(start, end);
  320. rebuild_tree();
  321. m_samples_model->update();
  322. }
  323. void Profile::clear_timestamp_filter_range()
  324. {
  325. if (!m_has_timestamp_filter_range)
  326. return;
  327. m_has_timestamp_filter_range = false;
  328. rebuild_tree();
  329. m_samples_model->update();
  330. }
  331. void Profile::add_process_filter(pid_t pid, EventSerialNumber start_valid, EventSerialNumber end_valid)
  332. {
  333. auto filter = ProcessFilter { pid, start_valid, end_valid };
  334. if (m_process_filters.contains_slow(filter))
  335. return;
  336. m_process_filters.append(move(filter));
  337. rebuild_tree();
  338. if (m_disassembly_model)
  339. m_disassembly_model->update();
  340. m_samples_model->update();
  341. }
  342. void Profile::remove_process_filter(pid_t pid, EventSerialNumber start_valid, EventSerialNumber end_valid)
  343. {
  344. auto filter = ProcessFilter { pid, start_valid, end_valid };
  345. if (!m_process_filters.contains_slow(filter))
  346. return;
  347. m_process_filters.remove_first_matching([&filter](ProcessFilter const& other_filter) {
  348. return other_filter == filter;
  349. });
  350. rebuild_tree();
  351. if (m_disassembly_model)
  352. m_disassembly_model->update();
  353. m_samples_model->update();
  354. }
  355. void Profile::clear_process_filter()
  356. {
  357. if (m_process_filters.is_empty())
  358. return;
  359. m_process_filters.clear();
  360. rebuild_tree();
  361. if (m_disassembly_model)
  362. m_disassembly_model->update();
  363. m_samples_model->update();
  364. }
  365. bool Profile::process_filter_contains(pid_t pid, EventSerialNumber serial)
  366. {
  367. if (!has_process_filter())
  368. return true;
  369. for (auto const& process_filter : m_process_filters)
  370. if (pid == process_filter.pid && serial >= process_filter.start_valid && serial <= process_filter.end_valid)
  371. return true;
  372. return false;
  373. }
  374. void Profile::set_inverted(bool inverted)
  375. {
  376. if (m_inverted == inverted)
  377. return;
  378. m_inverted = inverted;
  379. rebuild_tree();
  380. }
  381. void Profile::set_show_top_functions(bool show)
  382. {
  383. if (m_show_top_functions == show)
  384. return;
  385. m_show_top_functions = show;
  386. rebuild_tree();
  387. }
  388. void Profile::set_show_percentages(bool show_percentages)
  389. {
  390. if (m_show_percentages == show_percentages)
  391. return;
  392. m_show_percentages = show_percentages;
  393. }
  394. void Profile::set_disassembly_index(const GUI::ModelIndex& index)
  395. {
  396. if (m_disassembly_index == index)
  397. return;
  398. m_disassembly_index = index;
  399. auto* node = static_cast<ProfileNode*>(index.internal_data());
  400. if (!node)
  401. m_disassembly_model = nullptr;
  402. else
  403. m_disassembly_model = DisassemblyModel::create(*this, *node);
  404. }
  405. GUI::Model* Profile::disassembly_model()
  406. {
  407. return m_disassembly_model;
  408. }
  409. ProfileNode::ProfileNode(Process const& process)
  410. : m_root(true)
  411. , m_process(process)
  412. {
  413. }
  414. ProfileNode::ProfileNode(Process const& process, const String& object_name, String symbol, FlatPtr address, u32 offset, u64 timestamp, pid_t pid)
  415. : m_process(process)
  416. , m_symbol(move(symbol))
  417. , m_pid(pid)
  418. , m_address(address)
  419. , m_offset(offset)
  420. , m_timestamp(timestamp)
  421. {
  422. String object;
  423. if (object_name.ends_with(": .text"sv)) {
  424. object = object_name.view().substring_view(0, object_name.length() - 7);
  425. } else {
  426. object = object_name;
  427. }
  428. m_object_name = LexicalPath::basename(object);
  429. }
  430. }