| /* |
| * Copyright (C) 2012 Fusion-io |
| * |
| * This program is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU General Public |
| * License v2 as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| * |
| * You should have received a copy of the GNU General Public License |
| * along with this program; if not, write to the Free Software |
| * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA |
| * |
| * Parts of this file were imported from Jens Axboe's blktrace sources (also GPL) |
| */ |
| #include <sys/types.h> |
| #include <sys/stat.h> |
| #include <fcntl.h> |
| #include <unistd.h> |
| #include <stdlib.h> |
| #include <stdio.h> |
| #include <math.h> |
| #include <inttypes.h> |
| #include <string.h> |
| #include <asm/types.h> |
| #include <errno.h> |
| #include <sys/mman.h> |
| #include <time.h> |
| #include <math.h> |
| #include <dirent.h> |
| |
| #include "plot.h" |
| #include "blkparse.h" |
| #include "list.h" |
| #include "tracers.h" |
| |
| #define IO_HASH_TABLE_BITS 11 |
| #define IO_HASH_TABLE_SIZE (1 << IO_HASH_TABLE_BITS) |
| static struct list_head io_hash_table[IO_HASH_TABLE_SIZE]; |
| static u64 ios_in_flight = 0; |
| |
| #define PROCESS_HASH_TABLE_BITS 7 |
| #define PROCESS_HASH_TABLE_SIZE (1 << PROCESS_HASH_TABLE_BITS) |
| static struct list_head process_hash_table[PROCESS_HASH_TABLE_SIZE]; |
| |
| extern int plot_io_action; |
| extern int io_per_process; |
| |
| static const int line_len = 1024; |
| static char line[1024]; |
| |
| /* |
| * Trace categories |
| */ |
| enum { |
| BLK_TC_READ = 1 << 0, /* reads */ |
| BLK_TC_WRITE = 1 << 1, /* writes */ |
| BLK_TC_FLUSH = 1 << 2, /* flush */ |
| BLK_TC_SYNC = 1 << 3, /* sync */ |
| BLK_TC_QUEUE = 1 << 4, /* queueing/merging */ |
| BLK_TC_REQUEUE = 1 << 5, /* requeueing */ |
| BLK_TC_ISSUE = 1 << 6, /* issue */ |
| BLK_TC_COMPLETE = 1 << 7, /* completions */ |
| BLK_TC_FS = 1 << 8, /* fs requests */ |
| BLK_TC_PC = 1 << 9, /* pc requests */ |
| BLK_TC_NOTIFY = 1 << 10, /* special message */ |
| BLK_TC_AHEAD = 1 << 11, /* readahead */ |
| BLK_TC_META = 1 << 12, /* metadata */ |
| BLK_TC_DISCARD = 1 << 13, /* discard requests */ |
| BLK_TC_DRV_DATA = 1 << 14, /* binary driver data */ |
| BLK_TC_FUA = 1 << 15, /* fua requests */ |
| |
| BLK_TC_END = 1 << 15, /* we've run out of bits! */ |
| }; |
| |
| #define BLK_TC_SHIFT (16) |
| #define BLK_TC_ACT(act) ((act) << BLK_TC_SHIFT) |
| #define BLK_DATADIR(a) (((a) >> BLK_TC_SHIFT) & (BLK_TC_READ | BLK_TC_WRITE)) |
| |
| /* |
| * Basic trace actions |
| */ |
| enum { |
| __BLK_TA_QUEUE = 1, /* queued */ |
| __BLK_TA_BACKMERGE, /* back merged to existing rq */ |
| __BLK_TA_FRONTMERGE, /* front merge to existing rq */ |
| __BLK_TA_GETRQ, /* allocated new request */ |
| __BLK_TA_SLEEPRQ, /* sleeping on rq allocation */ |
| __BLK_TA_REQUEUE, /* request requeued */ |
| __BLK_TA_ISSUE, /* sent to driver */ |
| __BLK_TA_COMPLETE, /* completed by driver */ |
| __BLK_TA_PLUG, /* queue was plugged */ |
| __BLK_TA_UNPLUG_IO, /* queue was unplugged by io */ |
| __BLK_TA_UNPLUG_TIMER, /* queue was unplugged by timer */ |
| __BLK_TA_INSERT, /* insert request */ |
| __BLK_TA_SPLIT, /* bio was split */ |
| __BLK_TA_BOUNCE, /* bio was bounced */ |
| __BLK_TA_REMAP, /* bio was remapped */ |
| __BLK_TA_ABORT, /* request aborted */ |
| __BLK_TA_DRV_DATA, /* binary driver data */ |
| }; |
| |
| #define BLK_TA_MASK ((1 << BLK_TC_SHIFT) - 1) |
| |
| /* |
| * Notify events. |
| */ |
| enum blktrace_notify { |
| __BLK_TN_PROCESS = 0, /* establish pid/name mapping */ |
| __BLK_TN_TIMESTAMP, /* include system clock */ |
| __BLK_TN_MESSAGE, /* Character string message */ |
| }; |
| |
| /* |
| * Trace actions in full. Additionally, read or write is masked |
| */ |
| #define BLK_TA_QUEUE (__BLK_TA_QUEUE | BLK_TC_ACT(BLK_TC_QUEUE)) |
| #define BLK_TA_BACKMERGE (__BLK_TA_BACKMERGE | BLK_TC_ACT(BLK_TC_QUEUE)) |
| #define BLK_TA_FRONTMERGE (__BLK_TA_FRONTMERGE | BLK_TC_ACT(BLK_TC_QUEUE)) |
| #define BLK_TA_GETRQ (__BLK_TA_GETRQ | BLK_TC_ACT(BLK_TC_QUEUE)) |
| #define BLK_TA_SLEEPRQ (__BLK_TA_SLEEPRQ | BLK_TC_ACT(BLK_TC_QUEUE)) |
| #define BLK_TA_REQUEUE (__BLK_TA_REQUEUE | BLK_TC_ACT(BLK_TC_REQUEUE)) |
| #define BLK_TA_ISSUE (__BLK_TA_ISSUE | BLK_TC_ACT(BLK_TC_ISSUE)) |
| #define BLK_TA_COMPLETE (__BLK_TA_COMPLETE| BLK_TC_ACT(BLK_TC_COMPLETE)) |
| #define BLK_TA_PLUG (__BLK_TA_PLUG | BLK_TC_ACT(BLK_TC_QUEUE)) |
| #define BLK_TA_UNPLUG_IO (__BLK_TA_UNPLUG_IO | BLK_TC_ACT(BLK_TC_QUEUE)) |
| #define BLK_TA_UNPLUG_TIMER (__BLK_TA_UNPLUG_TIMER | BLK_TC_ACT(BLK_TC_QUEUE)) |
| #define BLK_TA_INSERT (__BLK_TA_INSERT | BLK_TC_ACT(BLK_TC_QUEUE)) |
| #define BLK_TA_SPLIT (__BLK_TA_SPLIT) |
| #define BLK_TA_BOUNCE (__BLK_TA_BOUNCE) |
| #define BLK_TA_REMAP (__BLK_TA_REMAP | BLK_TC_ACT(BLK_TC_QUEUE)) |
| #define BLK_TA_ABORT (__BLK_TA_ABORT | BLK_TC_ACT(BLK_TC_QUEUE)) |
| #define BLK_TA_DRV_DATA (__BLK_TA_DRV_DATA | BLK_TC_ACT(BLK_TC_DRV_DATA)) |
| |
| #define BLK_TN_PROCESS (__BLK_TN_PROCESS | BLK_TC_ACT(BLK_TC_NOTIFY)) |
| #define BLK_TN_TIMESTAMP (__BLK_TN_TIMESTAMP | BLK_TC_ACT(BLK_TC_NOTIFY)) |
| #define BLK_TN_MESSAGE (__BLK_TN_MESSAGE | BLK_TC_ACT(BLK_TC_NOTIFY)) |
| |
| #define BLK_IO_TRACE_MAGIC 0x65617400 |
| #define BLK_IO_TRACE_VERSION 0x07 |
| /* |
| * The trace itself |
| */ |
| struct blk_io_trace { |
| __u32 magic; /* MAGIC << 8 | version */ |
| __u32 sequence; /* event number */ |
| __u64 time; /* in nanoseconds */ |
| __u64 sector; /* disk offset */ |
| __u32 bytes; /* transfer length */ |
| __u32 action; /* what happened */ |
| __u32 pid; /* who did it */ |
| __u32 device; /* device identifier (dev_t) */ |
| __u32 cpu; /* on what cpu did it happen */ |
| __u16 error; /* completion error */ |
| __u16 pdu_len; /* length of data after this trace */ |
| }; |
| |
| struct pending_io { |
| /* sector offset of this IO */ |
| u64 sector; |
| |
| /* dev_t for this IO */ |
| u32 device; |
| |
| /* time this IO was dispatched */ |
| u64 dispatch_time; |
| /* time this IO was finished */ |
| u64 completion_time; |
| struct list_head hash_list; |
| /* process which queued this IO */ |
| u32 pid; |
| }; |
| |
| struct pid_map { |
| struct list_head hash_list; |
| u32 pid; |
| int index; |
| char name[0]; |
| }; |
| |
| #define MINORBITS 20 |
| #define MINORMASK ((1 << MINORBITS) - 1) |
| #define SECONDS(x) ((unsigned long long)(x) / 1000000000) |
| #define NANO_SECONDS(x) ((unsigned long long)(x) % 1000000000) |
| #define DOUBLE_TO_NANO_ULL(d) ((unsigned long long)((d) * 1000000000)) |
| #define CHECK_MAGIC(t) (((t)->magic & 0xffffff00) == BLK_IO_TRACE_MAGIC) |
| |
| void init_io_hash_table(void) |
| { |
| int i; |
| struct list_head *head; |
| |
| for (i = 0; i < IO_HASH_TABLE_SIZE; i++) { |
| head = io_hash_table + i; |
| INIT_LIST_HEAD(head); |
| } |
| } |
| |
| /* taken from the kernel hash.h */ |
| static inline u64 hash_sector(u64 val) |
| { |
| u64 hash = val; |
| |
| /* Sigh, gcc can't optimise this alone like it does for 32 bits. */ |
| u64 n = hash; |
| n <<= 18; |
| hash -= n; |
| n <<= 33; |
| hash -= n; |
| n <<= 3; |
| hash += n; |
| n <<= 3; |
| hash -= n; |
| n <<= 4; |
| hash += n; |
| n <<= 2; |
| hash += n; |
| |
| /* High bits are more random, so use them. */ |
| return hash >> (64 - IO_HASH_TABLE_BITS); |
| } |
| |
| static int io_hash_table_insert(struct pending_io *ins_pio) |
| { |
| u64 sector = ins_pio->sector; |
| u32 dev = ins_pio->device; |
| int slot = hash_sector(sector); |
| struct list_head *head; |
| struct pending_io *pio; |
| |
| head = io_hash_table + slot; |
| list_for_each_entry(pio, head, hash_list) { |
| if (pio->sector == sector && pio->device == dev) |
| return -EEXIST; |
| } |
| list_add_tail(&ins_pio->hash_list, head); |
| return 0; |
| } |
| |
| static struct pending_io *io_hash_table_search(u64 sector, u32 dev) |
| { |
| int slot = hash_sector(sector); |
| struct list_head *head; |
| struct pending_io *pio; |
| |
| head = io_hash_table + slot; |
| list_for_each_entry(pio, head, hash_list) { |
| if (pio->sector == sector && pio->device == dev) |
| return pio; |
| } |
| return NULL; |
| } |
| |
| static struct pending_io *hash_queued_io(struct blk_io_trace *io) |
| { |
| struct pending_io *pio; |
| int ret; |
| |
| pio = calloc(1, sizeof(*pio)); |
| pio->sector = io->sector; |
| pio->device = io->device; |
| pio->pid = io->pid; |
| |
| ret = io_hash_table_insert(pio); |
| if (ret < 0) { |
| /* crud, the IO is there already */ |
| free(pio); |
| return NULL; |
| } |
| return pio; |
| } |
| |
| static struct pending_io *hash_dispatched_io(struct blk_io_trace *io) |
| { |
| struct pending_io *pio; |
| |
| pio = io_hash_table_search(io->sector, io->device); |
| if (!pio) { |
| pio = hash_queued_io(io); |
| if (!pio) |
| return NULL; |
| } |
| pio->dispatch_time = io->time; |
| return pio; |
| } |
| |
| static struct pending_io *hash_completed_io(struct blk_io_trace *io) |
| { |
| struct pending_io *pio; |
| |
| pio = io_hash_table_search(io->sector, io->device); |
| |
| if (!pio) |
| return NULL; |
| return pio; |
| } |
| |
| void init_process_hash_table(void) |
| { |
| int i; |
| struct list_head *head; |
| |
| for (i = 0; i < PROCESS_HASH_TABLE_SIZE; i++) { |
| head = process_hash_table + i; |
| INIT_LIST_HEAD(head); |
| } |
| } |
| |
| static u32 hash_pid(u32 pid) |
| { |
| u32 hash = pid; |
| |
| hash ^= pid >> 3; |
| hash ^= pid >> 3; |
| hash ^= pid >> 4; |
| hash ^= pid >> 6; |
| return (hash & (PROCESS_HASH_TABLE_SIZE - 1)); |
| } |
| |
| static struct pid_map *process_hash_search(u32 pid) |
| { |
| int slot = hash_pid(pid); |
| struct list_head *head; |
| struct pid_map *pm; |
| |
| head = process_hash_table + slot; |
| list_for_each_entry(pm, head, hash_list) { |
| if (pm->pid == pid) |
| return pm; |
| } |
| return NULL; |
| } |
| |
| static struct pid_map *process_hash_insert(u32 pid, char *name) |
| { |
| int slot = hash_pid(pid); |
| struct pid_map *pm; |
| int old_index = 0; |
| char buf[16]; |
| |
| pm = process_hash_search(pid); |
| if (pm) { |
| /* Entry exists and name shouldn't be changed? */ |
| if (!name || !strcmp(name, pm->name)) |
| return pm; |
| list_del(&pm->hash_list); |
| old_index = pm->index; |
| free(pm); |
| } |
| if (!name) { |
| sprintf(buf, "[%u]", pid); |
| name = buf; |
| } |
| pm = malloc(sizeof(struct pid_map) + strlen(name) + 1); |
| pm->pid = pid; |
| pm->index = old_index; |
| strcpy(pm->name, name); |
| list_add_tail(&pm->hash_list, process_hash_table + slot); |
| |
| return pm; |
| } |
| |
| static void handle_notify(struct trace *trace) |
| { |
| struct blk_io_trace *io = trace->io; |
| void *payload = (char *)io + sizeof(*io); |
| u32 two32[2]; |
| |
| if (io->action == BLK_TN_PROCESS) { |
| if (io_per_process) |
| process_hash_insert(io->pid, payload); |
| return; |
| } |
| |
| if (io->action != BLK_TN_TIMESTAMP) |
| return; |
| |
| if (io->pdu_len != sizeof(two32)) |
| return; |
| |
| memcpy(two32, payload, sizeof(two32)); |
| trace->start_timestamp = io->time; |
| trace->abs_start_time.tv_sec = two32[0]; |
| trace->abs_start_time.tv_nsec = two32[1]; |
| if (trace->abs_start_time.tv_nsec < 0) { |
| trace->abs_start_time.tv_sec--; |
| trace->abs_start_time.tv_nsec += 1000000000; |
| } |
| } |
| |
| int next_record(struct trace *trace) |
| { |
| int skip = trace->io->pdu_len; |
| u64 offset; |
| |
| trace->cur += sizeof(*trace->io) + skip; |
| offset = trace->cur - trace->start; |
| if (offset >= trace->len) |
| return 1; |
| |
| trace->io = (struct blk_io_trace *)trace->cur; |
| return 0; |
| } |
| |
| void first_record(struct trace *trace) |
| { |
| trace->cur = trace->start; |
| trace->io = (struct blk_io_trace *)trace->cur; |
| } |
| |
| int is_io_event(struct blk_io_trace *test) |
| { |
| char *message; |
| if (!(test->action & BLK_TC_ACT(BLK_TC_NOTIFY))) |
| return 1; |
| if (test->action == BLK_TN_MESSAGE) { |
| int len = test->pdu_len; |
| if (len < 3) |
| return 0; |
| message = (char *)(test + 1); |
| if (strncmp(message, "fio ", 4) == 0) { |
| return 1; |
| } |
| } |
| return 0; |
| } |
| |
| u64 find_last_time(struct trace *trace) |
| { |
| char *p = trace->start + trace->len; |
| struct blk_io_trace *test; |
| int search_len = 0; |
| u64 found = 0; |
| |
| if (trace->len < sizeof(*trace->io)) |
| return 0; |
| p -= sizeof(*trace->io); |
| while (p >= trace->start) { |
| test = (struct blk_io_trace *)p; |
| if (CHECK_MAGIC(test) && is_io_event(test)) { |
| u64 offset = p - trace->start; |
| if (offset + sizeof(*test) + test->pdu_len == trace->len) { |
| return test->time; |
| } |
| } |
| p--; |
| search_len++; |
| if (search_len > 8192) { |
| break; |
| } |
| } |
| |
| /* searching backwards didn't work out, we'll have to scan the file */ |
| first_record(trace); |
| while (1) { |
| if (is_io_event(trace->io)) |
| found = trace->io->time; |
| if (next_record(trace)) |
| break; |
| } |
| first_record(trace); |
| return found; |
| } |
| |
| int parse_fio_bank_message(struct trace *trace, u64 *bank_ret, u64 *offset_ret, |
| u64 *num_banks_ret) |
| { |
| char *s; |
| char *next; |
| char *message; |
| struct blk_io_trace *test = trace->io; |
| int len = test->pdu_len; |
| u64 bank; |
| u64 offset; |
| u64 num_banks; |
| |
| if (!(test->action & BLK_TC_ACT(BLK_TC_NOTIFY))) |
| return -1; |
| if (test->action != BLK_TN_MESSAGE) |
| return -1; |
| |
| /* the message is fio rw bank offset num_banks */ |
| if (len < 3) |
| return -1; |
| message = (char *)(test + 1); |
| if (strncmp(message, "fio r ", 6) != 0) |
| return -1; |
| |
| message = strndup(message, len); |
| s = strchr(message, ' '); |
| if (!s) |
| goto out; |
| s++; |
| s = strchr(s, ' '); |
| if (!s) |
| goto out; |
| |
| bank = strtoll(s, &next, 10); |
| if (s == next) |
| goto out; |
| s = next; |
| |
| offset = strtoll(s, &next, 10); |
| if (s == next) |
| goto out; |
| s = next; |
| |
| num_banks = strtoll(s, &next, 10); |
| if (s == next) |
| goto out; |
| |
| *bank_ret = bank; |
| *offset_ret = offset; |
| *num_banks_ret = num_banks; |
| |
| return 0; |
| out: |
| free(message); |
| return -1; |
| } |
| |
| static struct dev_info *lookup_dev(struct trace *trace, struct blk_io_trace *io) |
| { |
| u32 dev = io->device; |
| int i; |
| struct dev_info *di = NULL; |
| |
| for (i = 0; i < trace->num_devices; i++) { |
| if (trace->devices[i].device == dev) { |
| di = trace->devices + i; |
| goto found; |
| } |
| } |
| i = trace->num_devices++; |
| if (i >= MAX_DEVICES_PER_TRACE) { |
| fprintf(stderr, "Trace contains too many devices (%d)\n", i); |
| exit(1); |
| } |
| di = trace->devices + i; |
| di->device = dev; |
| found: |
| return di; |
| } |
| |
| static void map_devices(struct trace *trace) |
| { |
| struct dev_info *di; |
| u64 found; |
| u64 map_start = 0; |
| int i; |
| |
| first_record(trace); |
| while (1) { |
| if (!(trace->io->action & BLK_TC_ACT(BLK_TC_NOTIFY))) { |
| di = lookup_dev(trace, trace->io); |
| found = trace->io->sector << 9; |
| if (found < di->min) |
| di->min = found; |
| |
| found += trace->io->bytes; |
| if (di->max < found) |
| di->max = found; |
| } |
| if (next_record(trace)) |
| break; |
| } |
| first_record(trace); |
| for (i = 0; i < trace->num_devices; i++) { |
| di = trace->devices + i; |
| di->map = map_start; |
| map_start += di->max - di->min; |
| } |
| } |
| |
| u64 map_io(struct trace *trace, struct blk_io_trace *io) |
| { |
| struct dev_info *di = lookup_dev(trace, io); |
| u64 val = trace->io->sector << 9; |
| return di->map + val - di->min; |
| } |
| |
| void find_extreme_offsets(struct trace *trace, u64 *min_ret, u64 *max_ret, u64 *max_bank_ret, |
| u64 *max_offset_ret) |
| { |
| u64 found = 0; |
| u64 max = 0, min = ~(u64)0; |
| u64 max_bank = 0; |
| u64 max_bank_offset = 0; |
| u64 num_banks = 0; |
| |
| map_devices(trace); |
| |
| first_record(trace); |
| while (1) { |
| if (!(trace->io->action & BLK_TC_ACT(BLK_TC_NOTIFY))) { |
| found = map_io(trace, trace->io); |
| if (found < min) |
| min = found; |
| |
| found += trace->io->bytes; |
| if (max < found) |
| max = found; |
| } else { |
| u64 bank; |
| u64 offset; |
| if (!parse_fio_bank_message(trace, &bank, |
| &offset, &num_banks)) { |
| if (bank > max_bank) |
| max_bank = bank; |
| if (offset > max_bank_offset) |
| max_bank_offset = offset; |
| } |
| } |
| if (next_record(trace)) |
| break; |
| } |
| first_record(trace); |
| *min_ret = min; |
| *max_ret = max; |
| *max_bank_ret = max_bank; |
| *max_offset_ret = max_bank_offset; |
| } |
| |
| static void check_io_types(struct trace *trace) |
| { |
| struct blk_io_trace *io = trace->io; |
| int action = io->action & BLK_TA_MASK; |
| |
| if (!(io->action & BLK_TC_ACT(BLK_TC_NOTIFY))) { |
| switch (action) { |
| case __BLK_TA_COMPLETE: |
| trace->found_completion = 1; |
| break; |
| case __BLK_TA_ISSUE: |
| trace->found_issue = 1; |
| break; |
| case __BLK_TA_QUEUE: |
| trace->found_queue = 1; |
| break; |
| }; |
| } |
| } |
| |
| |
| int filter_outliers(struct trace *trace, u64 min_offset, u64 max_offset, |
| u64 *yzoom_min, u64 *yzoom_max) |
| { |
| int hits[11]; |
| u64 max_per_bucket[11]; |
| u64 min_per_bucket[11]; |
| u64 bytes_per_bucket = (max_offset - min_offset + 1) / 10; |
| int slot; |
| int fat_count = 0; |
| |
| memset(hits, 0, sizeof(int) * 11); |
| memset(max_per_bucket, 0, sizeof(u64) * 11); |
| memset(min_per_bucket, 0xff, sizeof(u64) * 11); |
| first_record(trace); |
| while (1) { |
| check_io_types(trace); |
| if (!(trace->io->action & BLK_TC_ACT(BLK_TC_NOTIFY)) && |
| (trace->io->action & BLK_TA_MASK) == __BLK_TA_QUEUE) { |
| u64 off = map_io(trace, trace->io) - min_offset; |
| |
| slot = (int)(off / bytes_per_bucket); |
| hits[slot]++; |
| if (off < min_per_bucket[slot]) |
| min_per_bucket[slot] = off; |
| |
| off += trace->io->bytes; |
| slot = (int)(off / bytes_per_bucket); |
| hits[slot]++; |
| if (off > max_per_bucket[slot]) |
| max_per_bucket[slot] = off; |
| } |
| if (next_record(trace)) |
| break; |
| } |
| first_record(trace); |
| for (slot = 0; slot < 11; slot++) { |
| if (hits[slot] > fat_count) { |
| fat_count = hits[slot]; |
| } |
| } |
| |
| *yzoom_max = max_offset; |
| for (slot = 10; slot >= 0; slot--) { |
| double d = hits[slot]; |
| |
| if (d >= (double)fat_count * .05) { |
| *yzoom_max = max_per_bucket[slot] + min_offset; |
| break; |
| } |
| } |
| |
| *yzoom_min = min_offset; |
| for (slot = 0; slot < 10; slot++) { |
| double d = hits[slot]; |
| |
| if (d >= (double)fat_count * .05) { |
| *yzoom_min = min_per_bucket[slot] + min_offset; |
| break; |
| } |
| } |
| return 0; |
| } |
| |
| static char footer[] = ".blktrace.0"; |
| static int footer_len = sizeof(footer); |
| |
| static void match_trace(char *name, char **traces) |
| { |
| int match_len; |
| char *match; |
| int footer_start; |
| |
| match_len = strlen(name); |
| if (match_len <= footer_len) |
| return; |
| |
| footer_start = match_len - footer_len; |
| if (strcmp(name + footer_start + 1, footer) != 0) |
| return; |
| |
| match = strdup(name); |
| if (!match) |
| goto enomem; |
| |
| match[footer_start + 1] = '\0'; |
| snprintf(line, line_len, "%s -i '%s'", *traces ? *traces : "", match); |
| free(match); |
| |
| match = strdup(line); |
| if (!match) |
| goto enomem; |
| |
| free(*traces); |
| *traces = match; |
| return; |
| |
| enomem: |
| perror("memory allocation failed"); |
| exit(1); |
| return; |
| } |
| |
| static char *combine_blktrace_devs(char *dir_name) |
| { |
| DIR *dir; |
| char *traces = NULL; |
| struct dirent *d; |
| int len; |
| int ret; |
| |
| dir = opendir(dir_name); |
| if (!dir) |
| return NULL; |
| |
| while (1) { |
| d = readdir(dir); |
| if (!d) |
| break; |
| |
| len = strlen(d->d_name); |
| if (len > footer_len) |
| match_trace(d->d_name, &traces); |
| } |
| |
| closedir(dir); |
| |
| if (!traces) |
| return NULL; |
| |
| snprintf(line, line_len, "blkparse -O %s -D %s -d '%s.%s'", |
| traces, dir_name, dir_name, "dump"); |
| |
| ret = system(line); |
| if (ret) { |
| fprintf(stderr, "blkparse failure %s\n", line); |
| exit(1); |
| } |
| snprintf(line, line_len, "%s.%s", dir_name, "dump"); |
| return strdup(line); |
| } |
| |
| static char *find_trace_file(char *filename) |
| { |
| int ret; |
| struct stat st; |
| char *dot; |
| char *try; |
| int found_dir = 0; |
| |
| /* look for an exact match of whatever they pass in. |
| * If it is a file, assume it is the dump file. |
| * If a directory, remember that it existed so we |
| * can combine traces in that directory later |
| */ |
| ret = stat(filename, &st); |
| if (ret == 0) { |
| if (S_ISREG(st.st_mode)) |
| return strdup(filename); |
| |
| if (S_ISDIR(st.st_mode)) |
| found_dir = 1; |
| } |
| |
| /* |
| * try tacking .dump onto the end and see if that already |
| * has been generated |
| */ |
| snprintf(line, line_len, "%s.%s", filename, "dump"); |
| ret = stat(line, &st); |
| if (ret == 0) |
| return strdup(line); |
| |
| /* |
| * try to generate the .dump from all the traces in |
| * a single dir. |
| */ |
| if (found_dir) { |
| try = combine_blktrace_devs(filename); |
| if (try) |
| return try; |
| } |
| |
| /* |
| * try to generate the .dump from all the blktrace |
| * files for a named trace |
| */ |
| try = strdup(filename); |
| dot = strrchr(try, '.'); |
| if (!dot || strcmp(".dump", dot) != 0) { |
| if (dot && dot != try) |
| *dot = '\0'; |
| snprintf(line, line_len, "%s%s", try, ".blktrace.0"); |
| ret = stat(line, &st); |
| if (ret == 0) { |
| blktrace_to_dump(try); |
| snprintf(line, line_len, "%s.%s", try, "dump"); |
| ret = stat(line, &st); |
| if (ret == 0) { |
| free(try); |
| return strdup(line); |
| } |
| } |
| } |
| free(try); |
| return NULL; |
| } |
| struct trace *open_trace(char *filename) |
| { |
| int fd; |
| char *p; |
| struct stat st; |
| int ret; |
| struct trace *trace; |
| char *found_filename; |
| |
| trace = calloc(1, sizeof(*trace)); |
| if (!trace) { |
| fprintf(stderr, "unable to allocate memory for trace\n"); |
| return NULL; |
| } |
| |
| found_filename = find_trace_file(filename); |
| if (!found_filename) { |
| fprintf(stderr, "Unable to find trace file %s\n", filename); |
| goto fail; |
| } |
| filename = found_filename; |
| |
| fd = open(filename, O_RDONLY); |
| if (fd < 0) { |
| fprintf(stderr, "Unable to open trace file %s err %s\n", filename, strerror(errno)); |
| goto fail; |
| } |
| ret = fstat(fd, &st); |
| if (ret < 0) { |
| fprintf(stderr, "stat failed on %s err %s\n", filename, strerror(errno)); |
| goto fail_fd; |
| } |
| p = mmap(NULL, st.st_size, PROT_READ, MAP_PRIVATE, fd, 0); |
| if (p == MAP_FAILED) { |
| fprintf(stderr, "Unable to mmap trace file %s, err %s\n", filename, strerror(errno)); |
| goto fail_fd; |
| } |
| trace->fd = fd; |
| trace->len = st.st_size; |
| trace->start = p; |
| trace->cur = p; |
| trace->io = (struct blk_io_trace *)p; |
| return trace; |
| |
| fail_fd: |
| close(fd); |
| fail: |
| free(trace); |
| return NULL; |
| } |
| static inline int tput_event(struct trace *trace) |
| { |
| if (trace->found_completion) |
| return __BLK_TA_COMPLETE; |
| if (trace->found_issue) |
| return __BLK_TA_ISSUE; |
| if (trace->found_queue) |
| return __BLK_TA_QUEUE; |
| |
| return __BLK_TA_COMPLETE; |
| } |
| |
| int action_char_to_num(char action) |
| { |
| switch (action) { |
| case 'Q': |
| return __BLK_TA_QUEUE; |
| case 'D': |
| return __BLK_TA_ISSUE; |
| case 'C': |
| return __BLK_TA_COMPLETE; |
| } |
| return -1; |
| } |
| |
| static inline int io_event(struct trace *trace) |
| { |
| if (plot_io_action) |
| return plot_io_action; |
| if (trace->found_queue) |
| return __BLK_TA_QUEUE; |
| if (trace->found_issue) |
| return __BLK_TA_ISSUE; |
| if (trace->found_completion) |
| return __BLK_TA_COMPLETE; |
| |
| return __BLK_TA_COMPLETE; |
| } |
| |
| void add_tput(struct trace *trace, struct graph_line_data *writes_gld, |
| struct graph_line_data *reads_gld) |
| { |
| struct blk_io_trace *io = trace->io; |
| struct graph_line_data *gld; |
| int action = io->action & BLK_TA_MASK; |
| int seconds; |
| |
| if (io->action & BLK_TC_ACT(BLK_TC_NOTIFY)) |
| return; |
| |
| if (action != tput_event(trace)) |
| return; |
| |
| if (BLK_DATADIR(io->action) & BLK_TC_READ) |
| gld = reads_gld; |
| else |
| gld = writes_gld; |
| |
| seconds = SECONDS(io->time); |
| if (seconds > gld->max_seconds) |
| return; |
| |
| gld->data[seconds].sum += io->bytes; |
| |
| gld->data[seconds].count = 1; |
| if (gld->data[seconds].sum > gld->max) |
| gld->max = gld->data[seconds].sum; |
| } |
| |
| #define GDD_PTR_ALLOC_STEP 16 |
| |
| static struct pid_map *get_pid_map(struct trace_file *tf, u32 pid) |
| { |
| struct pid_map *pm; |
| |
| if (!io_per_process) { |
| if (!tf->io_plots) |
| tf->io_plots = 1; |
| return NULL; |
| } |
| |
| pm = process_hash_insert(pid, NULL); |
| /* New entry? */ |
| if (!pm->index) { |
| if (tf->io_plots == tf->io_plots_allocated) { |
| tf->io_plots_allocated += GDD_PTR_ALLOC_STEP; |
| tf->gdd_reads = realloc(tf->gdd_reads, tf->io_plots_allocated * sizeof(struct graph_dot_data *)); |
| if (!tf->gdd_reads) |
| abort(); |
| tf->gdd_writes = realloc(tf->gdd_writes, tf->io_plots_allocated * sizeof(struct graph_dot_data *)); |
| if (!tf->gdd_writes) |
| abort(); |
| memset(tf->gdd_reads + tf->io_plots_allocated - GDD_PTR_ALLOC_STEP, |
| 0, GDD_PTR_ALLOC_STEP * sizeof(struct graph_dot_data *)); |
| memset(tf->gdd_writes + tf->io_plots_allocated - GDD_PTR_ALLOC_STEP, |
| 0, GDD_PTR_ALLOC_STEP * sizeof(struct graph_dot_data *)); |
| } |
| pm->index = tf->io_plots++; |
| |
| return pm; |
| } |
| return pm; |
| } |
| |
| void add_io(struct trace *trace, struct trace_file *tf) |
| { |
| struct blk_io_trace *io = trace->io; |
| int action = io->action & BLK_TA_MASK; |
| u64 offset; |
| int index; |
| char *label; |
| struct pid_map *pm; |
| |
| if (io->action & BLK_TC_ACT(BLK_TC_NOTIFY)) |
| return; |
| |
| if (action != io_event(trace)) |
| return; |
| |
| offset = map_io(trace, io); |
| |
| pm = get_pid_map(tf, io->pid); |
| if (!pm) { |
| index = 0; |
| label = ""; |
| } else { |
| index = pm->index; |
| label = pm->name; |
| } |
| if (BLK_DATADIR(io->action) & BLK_TC_READ) { |
| if (!tf->gdd_reads[index]) |
| tf->gdd_reads[index] = alloc_dot_data(tf->min_seconds, tf->max_seconds, tf->min_offset, tf->max_offset, tf->stop_seconds, pick_color(), strdup(label)); |
| set_gdd_bit(tf->gdd_reads[index], offset, io->bytes, io->time); |
| } else if (BLK_DATADIR(io->action) & BLK_TC_WRITE) { |
| if (!tf->gdd_writes[index]) |
| tf->gdd_writes[index] = alloc_dot_data(tf->min_seconds, tf->max_seconds, tf->min_offset, tf->max_offset, tf->stop_seconds, pick_color(), strdup(label)); |
| set_gdd_bit(tf->gdd_writes[index], offset, io->bytes, io->time); |
| } |
| } |
| |
| void add_pending_io(struct trace *trace, struct graph_line_data *gld) |
| { |
| int seconds; |
| struct blk_io_trace *io = trace->io; |
| int action = io->action & BLK_TA_MASK; |
| double avg; |
| struct pending_io *pio; |
| |
| if (io->action & BLK_TC_ACT(BLK_TC_NOTIFY)) |
| return; |
| |
| if (action == __BLK_TA_QUEUE) { |
| if (trace->found_issue || trace->found_completion) |
| hash_queued_io(trace->io); |
| return; |
| } |
| if (action != __BLK_TA_ISSUE) |
| return; |
| |
| seconds = SECONDS(io->time); |
| if (seconds > gld->max_seconds) |
| return; |
| |
| pio = hash_dispatched_io(trace->io); |
| if (!pio) |
| return; |
| |
| if (!trace->found_completion) { |
| list_del(&pio->hash_list); |
| free(pio); |
| } |
| |
| ios_in_flight++; |
| |
| gld->data[seconds].sum += ios_in_flight; |
| gld->data[seconds].count++; |
| |
| avg = (double)gld->data[seconds].sum / gld->data[seconds].count; |
| if (gld->max < (u64)avg) { |
| gld->max = avg; |
| } |
| } |
| |
| void add_completed_io(struct trace *trace, |
| struct graph_line_data *latency_gld) |
| { |
| struct blk_io_trace *io = trace->io; |
| int seconds; |
| int action = io->action & BLK_TA_MASK; |
| struct pending_io *pio; |
| double avg; |
| u64 latency; |
| |
| if (io->action & BLK_TC_ACT(BLK_TC_NOTIFY)) |
| return; |
| |
| if (action != __BLK_TA_COMPLETE) |
| return; |
| |
| seconds = SECONDS(io->time); |
| |
| pio = hash_completed_io(trace->io); |
| if (!pio) |
| return; |
| |
| if (ios_in_flight > 0) |
| ios_in_flight--; |
| if (io->time >= pio->dispatch_time) { |
| latency = io->time - pio->dispatch_time; |
| latency_gld->data[seconds].sum += latency; |
| latency_gld->data[seconds].count++; |
| } |
| |
| list_del(&pio->hash_list); |
| free(pio); |
| |
| avg = (double)latency_gld->data[seconds].sum / |
| latency_gld->data[seconds].count; |
| if (latency_gld->max < (u64)avg) { |
| latency_gld->max = avg; |
| } |
| } |
| |
| void add_iop(struct trace *trace, struct graph_line_data *gld) |
| { |
| struct blk_io_trace *io = trace->io; |
| int action = io->action & BLK_TA_MASK; |
| int seconds; |
| |
| if (io->action & BLK_TC_ACT(BLK_TC_NOTIFY)) |
| return; |
| |
| /* iops and tput use the same events */ |
| if (action != tput_event(trace)) |
| return; |
| |
| seconds = SECONDS(io->time); |
| if (seconds > gld->max_seconds) |
| return; |
| |
| gld->data[seconds].sum += 1; |
| gld->data[seconds].count = 1; |
| if (gld->data[seconds].sum > gld->max) |
| gld->max = gld->data[seconds].sum; |
| } |
| |
| void check_record(struct trace *trace) |
| { |
| handle_notify(trace); |
| } |