The Linux Kernel Tracepoint API

Author:

Jason Baron

Author:

William Cohen

Introduction

Tracepoints are static probe points that are located in strategic points throughout the kernel. 'Probes' register/unregister with tracepoints via a callback mechanism. The 'probes' are strictly typed functions that are passed a unique set of parameters defined by each tracepoint.

From this simple callback mechanism, 'probes' can be used to profile, debug, and understand kernel behavior. There are a number of tools that provide a framework for using 'probes'. These tools include Systemtap, ftrace, and LTTng.

Tracepoints are defined in a number of header files via various macros. Thus, the purpose of this document is to provide a clear accounting of the available tracepoints. The intention is to understand not only what tracepoints are available but also to understand where future tracepoints might be added.

The API presented has functions of the form: trace_tracepointname(function parameters). These are the tracepoints callbacks that are found throughout the code. Registering and unregistering probes with these callback sites is covered in the Documentation/trace/* directory.

IRQ

void trace_irq_handler_entry(int irq, struct irqaction *action)

called immediately before the irq action handler

Parameters

int irq

irq number

struct irqaction *action

pointer to struct irqaction

Description

The struct irqaction pointed to by action contains various information about the handler, including the device name, action->name, and the device id, action->dev_id. When used in conjunction with the irq_handler_exit tracepoint, we can figure out irq handler latencies.

void trace_irq_handler_exit(int irq, struct irqaction *action, int ret)

called immediately after the irq action handler returns

Parameters

int irq

irq number

struct irqaction *action

pointer to struct irqaction

int ret

return value

Description

If the ret value is set to IRQ_HANDLED, then we know that the corresponding action->handler successfully handled this irq. Otherwise, the irq might be a shared irq line, or the irq was not handled successfully. Can be used in conjunction with the irq_handler_entry to understand irq handler latencies.

void trace_softirq_entry(unsigned int vec_nr)

called immediately before the softirq handler

Parameters

unsigned int vec_nr

softirq vector number

Description

When used in combination with the softirq_exit tracepoint we can determine the softirq handler routine.

void trace_softirq_exit(unsigned int vec_nr)

called immediately after the softirq handler returns

Parameters

unsigned int vec_nr

softirq vector number

Description

When used in combination with the softirq_entry tracepoint we can determine the softirq handler routine.

void trace_softirq_raise(unsigned int vec_nr)

called immediately when a softirq is raised

Parameters

unsigned int vec_nr

softirq vector number

Description

When used in combination with the softirq_entry tracepoint we can determine the softirq raise to run latency.

void trace_tasklet_entry(struct tasklet_struct *t, void *func)

called immediately before the tasklet is run

Parameters

struct tasklet_struct *t

tasklet pointer

void *func

tasklet callback or function being run

Description

Used to find individual tasklet execution time

void trace_tasklet_exit(struct tasklet_struct *t, void *func)

called immediately after the tasklet is run

Parameters

struct tasklet_struct *t

tasklet pointer

void *func

tasklet callback or function being run

Description

Used to find individual tasklet execution time

SIGNAL

void trace_signal_generate(int sig, struct kernel_siginfo *info, struct task_struct *task, int group, int result)

called when a signal is generated

Parameters

int sig

signal number

struct kernel_siginfo *info

pointer to struct siginfo

struct task_struct *task

pointer to struct task_struct

int group

shared or private

int result

TRACE_SIGNAL_*

Description

Current process sends a 'sig' signal to 'task' process with 'info' siginfo. If 'info' is SEND_SIG_NOINFO or SEND_SIG_PRIV, 'info' is not a pointer and you can't access its field. Instead, SEND_SIG_NOINFO means that si_code is SI_USER, and SEND_SIG_PRIV means that si_code is SI_KERNEL.

void trace_signal_deliver(int sig, struct kernel_siginfo *info, struct k_sigaction *ka)

called when a signal is delivered

Parameters

int sig

signal number

struct kernel_siginfo *info

pointer to struct siginfo

struct k_sigaction *ka

pointer to struct k_sigaction

Description

A 'sig' signal is delivered to current process with 'info' siginfo, and it will be handled by 'ka'. ka->sa.sa_handler can be SIG_IGN or SIG_DFL. Note that some signals reported by signal_generate tracepoint can be lost, ignored or modified (by debugger) before hitting this tracepoint. This means, this can show which signals are actually delivered, but matching generated signals and delivered signals may not be correct.

Block IO

void trace_block_touch_buffer(struct buffer_head *bh)

mark a buffer accessed

Parameters

struct buffer_head *bh

buffer_head being touched

Description

Called from touch_buffer().

void trace_block_dirty_buffer(struct buffer_head *bh)

mark a buffer dirty

Parameters

struct buffer_head *bh

buffer_head being dirtied

Description

Called from mark_buffer_dirty().

void trace_block_rq_requeue(struct request *rq)

place block IO request back on a queue

Parameters

struct request *rq

block IO operation request

Description

The block operation request rq is being placed back into queue q. For some reason the request was not completed and needs to be put back in the queue.

void trace_block_rq_complete(struct request *rq, blk_status_t error, unsigned int nr_bytes)

block IO operation completed by device driver

Parameters

struct request *rq

block operations request

blk_status_t error

status code

unsigned int nr_bytes

number of completed bytes

Description

The block_rq_complete tracepoint event indicates that some portion of operation request has been completed by the device driver. If the rq->bio is NULL, then there is absolutely no additional work to do for the request. If rq->bio is non-NULL then there is additional work required to complete the request.

void trace_block_rq_error(struct request *rq, blk_status_t error, unsigned int nr_bytes)

block IO operation error reported by device driver

Parameters

struct request *rq

block operations request

blk_status_t error

status code

unsigned int nr_bytes

number of completed bytes

Description

The block_rq_error tracepoint event indicates that some portion of operation request has failed as reported by the device driver.

void trace_block_rq_insert(struct request *rq)

insert block operation request into queue

Parameters

struct request *rq

block IO operation request

Description

Called immediately before block operation request rq is inserted into queue q. The fields in the operation request rq struct can be examined to determine which device and sectors the pending operation would access.

void trace_block_rq_issue(struct request *rq)

issue pending block IO request operation to device driver

Parameters

struct request *rq

block IO operation request

Description

Called when block operation request rq from queue q is sent to a device driver for processing.

void trace_block_rq_merge(struct request *rq)

merge request with another one in the elevator

Parameters

struct request *rq

block IO operation request

Description

Called when block operation request rq from queue q is merged to another request queued in the elevator.

void trace_block_io_start(struct request *rq)

insert a request for execution

Parameters

struct request *rq

block IO operation request

Description

Called when block operation request rq is queued for execution

void trace_block_io_done(struct request *rq)

block IO operation request completed

Parameters

struct request *rq

block IO operation request

Description

Called when block operation request rq is completed

void trace_block_bio_complete(struct request_queue *q, struct bio *bio)

completed all work on the block operation

Parameters

struct request_queue *q

queue holding the block operation

struct bio *bio

block operation completed

Description

This tracepoint indicates there is no further work to do on this block IO operation bio.

void trace_block_bio_bounce(struct bio *bio)

used bounce buffer when processing block operation

Parameters

struct bio *bio

block operation

Description

A bounce buffer was used to handle the block operation bio in q. This occurs when hardware limitations prevent a direct transfer of data between the bio data memory area and the IO device. Use of a bounce buffer requires extra copying of data and decreases performance.

void trace_block_bio_backmerge(struct bio *bio)

merging block operation to the end of an existing operation

Parameters

struct bio *bio

new block operation to merge

Description

Merging block request bio to the end of an existing block request.

void trace_block_bio_frontmerge(struct bio *bio)

merging block operation to the beginning of an existing operation

Parameters

struct bio *bio

new block operation to merge

Description

Merging block IO operation bio to the beginning of an existing block request.

void trace_block_bio_queue(struct bio *bio)

putting new block IO operation in queue

Parameters

struct bio *bio

new block operation

Description

About to place the block IO operation bio into queue q.

void trace_block_getrq(struct bio *bio)

get a free request entry in queue for block IO operations

Parameters

struct bio *bio

pending block IO operation (can be NULL)

Description

A request struct has been allocated to handle the block IO operation bio.

void trace_block_plug(struct request_queue *q)

keep operations requests in request queue

Parameters

struct request_queue *q

request queue to plug

Description

Plug the request queue q. Do not allow block operation requests to be sent to the device driver. Instead, accumulate requests in the queue to improve throughput performance of the block device.

void trace_block_unplug(struct request_queue *q, unsigned int depth, bool explicit)

release of operations requests in request queue

Parameters

struct request_queue *q

request queue to unplug

unsigned int depth

number of requests just added to the queue

bool explicit

whether this was an explicit unplug, or one from schedule()

Description

Unplug request queue q because device driver is scheduled to work on elements in the request queue.

void trace_block_split(struct bio *bio, unsigned int new_sector)

split a single bio struct into two bio structs

Parameters

struct bio *bio

block operation being split

unsigned int new_sector

The starting sector for the new bio

Description

The bio request bio needs to be split into two bio requests. The newly created bio request starts at new_sector. This split may be required due to hardware limitations such as operation crossing device boundaries in a RAID system.

void trace_block_bio_remap(struct bio *bio, dev_t dev, sector_t from)

map request for a logical device to the raw device

Parameters

struct bio *bio

revised operation

dev_t dev

original device for the operation

sector_t from

original sector for the operation

Description

An operation for a logical device has been mapped to the raw block device.

void trace_block_rq_remap(struct request *rq, dev_t dev, sector_t from)

map request for a block operation request

Parameters

struct request *rq

block IO operation request

dev_t dev

device for the operation

sector_t from

original sector for the operation

Description

The block operation request rq in q has been remapped. The block operation request rq holds the current information and from hold the original sector.

Workqueue

void trace_workqueue_queue_work(int req_cpu, struct pool_workqueue *pwq, struct work_struct *work)

called when a work gets queued

Parameters

int req_cpu

the requested cpu

struct pool_workqueue *pwq

pointer to struct pool_workqueue

struct work_struct *work

pointer to struct work_struct

Description

This event occurs when a work is queued immediately or once a delayed work is actually queued on a workqueue (ie: once the delay has been reached).

void trace_workqueue_activate_work(struct work_struct *work)

called when a work gets activated

Parameters

struct work_struct *work

pointer to struct work_struct

Description

This event occurs when a queued work is put on the active queue, which happens immediately after queueing unless max_active limit is reached.

void trace_workqueue_execute_start(struct work_struct *work)

called immediately before the workqueue callback

Parameters

struct work_struct *work

pointer to struct work_struct

Description

Allows to track workqueue execution.

void trace_workqueue_execute_end(struct work_struct *work, work_func_t function)

called immediately after the workqueue callback

Parameters

struct work_struct *work

pointer to struct work_struct

work_func_t function

pointer to worker function

Description

Allows to track workqueue execution.