InfiniBand and Remote DMA (RDMA) Interfaces

Introduction and Overview

TBD

InfiniBand core interfaces

struct iwpm_nlmsg_request * iwpm_get_nlmsg_request(__u32 nlmsg_seq, u8 nl_client, gfp_t gfp)

Allocate and initialize netlink message request

Parameters

__u32 nlmsg_seq
Sequence number of the netlink message
u8 nl_client
The index of the netlink client
gfp_t gfp
Indicates how the memory for the request should be allocated

Description

Returns the newly allocated netlink request object if successful, otherwise returns NULL

void iwpm_free_nlmsg_request(struct kref * kref)

Deallocate netlink message request

Parameters

struct kref * kref
Holds reference of netlink message request
struct iwpm_nlmsg_request * iwpm_find_nlmsg_request(__u32 echo_seq)

Find netlink message request in the request list

Parameters

__u32 echo_seq
Sequence number of the netlink request to find

Description

Returns the found netlink message request, if not found, returns NULL

int iwpm_wait_complete_req(struct iwpm_nlmsg_request * nlmsg_request)

Block while servicing the netlink request

Parameters

struct iwpm_nlmsg_request * nlmsg_request
Netlink message request to service

Description

Wakes up, after the request is completed or expired Returns 0 if the request is complete without error

int iwpm_get_nlmsg_seq(void)

Get the sequence number for a netlink message to send to the port mapper

Parameters

void
no arguments

Description

Returns the sequence number for the netlink message.

void iwpm_add_remote_info(struct iwpm_remote_info * reminfo)

Add remote address info of the connecting peer to the remote info hash table

Parameters

struct iwpm_remote_info * reminfo
The remote info to be added
int iwpm_valid_client(u8 nl_client)

Check if the port mapper client is valid

Parameters

u8 nl_client
The index of the netlink client

Description

Valid clients need to call iwpm_init() before using the port mapper

void iwpm_set_valid(u8 nl_client, int valid)

Set the port mapper client to valid or not

Parameters

u8 nl_client
The index of the netlink client
int valid
1 if valid or 0 if invalid
u32 iwpm_check_registration(u8 nl_client, u32 reg)

Check if the client registration matches the given one

Parameters

u8 nl_client
The index of the netlink client
u32 reg
The given registration type to compare with

Description

Call iwpm_register_pid() to register a client Returns true if the client registration matches reg, otherwise returns false

void iwpm_set_registration(u8 nl_client, u32 reg)

Set the client registration

Parameters

u8 nl_client
The index of the netlink client
u32 reg
Registration type to set
u32 iwpm_get_registration(u8 nl_client)

Parameters

u8 nl_client
The index of the netlink client

Description

Returns the client registration type

int iwpm_send_mapinfo(u8 nl_client, int iwpm_pid)

Send local and mapped IPv4/IPv6 address info of a client to the user space port mapper

Parameters

u8 nl_client
The index of the netlink client
int iwpm_pid
The pid of the user space port mapper

Description

If successful, returns the number of sent mapping info records

int iwpm_mapinfo_available(void)

Check if any mapping info records is available in the hash table

Parameters

void
no arguments

Description

Returns 1 if mapping information is available, otherwise returns 0

int iwpm_compare_sockaddr(struct sockaddr_storage * a_sockaddr, struct sockaddr_storage * b_sockaddr)

Compare two sockaddr storage structs

Parameters

struct sockaddr_storage * a_sockaddr
first sockaddr to compare
struct sockaddr_storage * b_sockaddr
second sockaddr to compare

Return

0 if they are holding the same ip/tcp address info, otherwise returns 1

int iwpm_validate_nlmsg_attr(struct nlattr * nltb, int nla_count)

Check for NULL netlink attributes

Parameters

struct nlattr * nltb
Holds address of each netlink message attributes
int nla_count
Number of netlink message attributes

Description

Returns error if any of the nla_count attributes is NULL

struct sk_buff * iwpm_create_nlmsg(u32 nl_op, struct nlmsghdr ** nlh, int nl_client)

Allocate skb and form a netlink message

Parameters

u32 nl_op
Netlink message opcode
struct nlmsghdr ** nlh
Holds address of the netlink message header in skb
int nl_client
The index of the netlink client

Description

Returns the newly allcated skb, or NULL if the tailroom of the skb is insufficient to store the message header and payload

int iwpm_parse_nlmsg(struct netlink_callback * cb, int policy_max, const struct nla_policy * nlmsg_policy, struct nlattr * nltb, const char * msg_type)

Validate and parse the received netlink message

Parameters

struct netlink_callback * cb
Netlink callback structure
int policy_max
Maximum attribute type to be expected
const struct nla_policy * nlmsg_policy
Validation policy
struct nlattr * nltb
Array to store policy_max parsed elements
const char * msg_type
Type of netlink message

Description

Returns 0 on success or a negative error code

void iwpm_print_sockaddr(struct sockaddr_storage * sockaddr, char * msg)

Print IPv4/IPv6 address and TCP port

Parameters

struct sockaddr_storage * sockaddr
Socket address to print
char * msg
Message to print
int iwpm_send_hello(u8 nl_client, int iwpm_pid, u16 abi_version)

Send hello response to iwpmd

Parameters

u8 nl_client
The index of the netlink client
int iwpm_pid
The pid of the user space port mapper
u16 abi_version
The kernel's abi_version

Description

Returns 0 on success or a negative error code

int ib_process_cq_direct(struct ib_cq * cq, int budget)

process a CQ in caller context

Parameters

struct ib_cq * cq
CQ to process
int budget
number of CQEs to poll for

Description

This function is used to process all outstanding CQ entries. It does not offload CQ processing to a different context and does not ask for completion interrupts from the HCA. Using direct processing on CQ with non IB_POLL_DIRECT type may trigger concurrent processing.

Note

do not pass -1 as budget unless it is guaranteed that the number of completions that will be processed is small.

struct ib_cq * __ib_alloc_cq_user(struct ib_device * dev, void * private, int nr_cqe, int comp_vector, enum ib_poll_context poll_ctx, const char * caller, struct ib_udata * udata)

allocate a completion queue

Parameters

struct ib_device * dev
device to allocate the CQ for
void * private
driver private data, accessible from cq->cq_context
int nr_cqe
number of CQEs to allocate
int comp_vector
HCA completion vectors for this CQ
enum ib_poll_context poll_ctx
context to poll the CQ from.
const char * caller
module owner name.
struct ib_udata * udata
Valid user data or NULL for kernel object

Description

This is the proper interface to allocate a CQ for in-kernel users. A CQ allocated with this interface will automatically be polled from the specified context. The ULP must use wr->wr_cqe instead of wr->wr_id to use this CQ abstraction.

struct ib_cq * __ib_alloc_cq_any(struct ib_device * dev, void * private, int nr_cqe, enum ib_poll_context poll_ctx, const char * caller)

allocate a completion queue

Parameters

struct ib_device * dev
device to allocate the CQ for
void * private
driver private data, accessible from cq->cq_context
int nr_cqe
number of CQEs to allocate
enum ib_poll_context poll_ctx
context to poll the CQ from
const char * caller
module owner name

Description

Attempt to spread ULP Completion Queues over each device's interrupt vectors. A simple best-effort mechanism is used.

void ib_free_cq_user(struct ib_cq * cq, struct ib_udata * udata)

free a completion queue

Parameters

struct ib_cq * cq
completion queue to free.
struct ib_udata * udata
User data or NULL for kernel object
struct ib_cm_id * ib_cm_insert_listen(struct ib_device * device, ib_cm_handler cm_handler, __be64 service_id)

Parameters

struct ib_device * device
Device associated with the cm_id. All related communication will be associated with the specified device.
ib_cm_handler cm_handler
Callback invoked to notify the user of CM events.
__be64 service_id
Service identifier matched against incoming connection and service ID resolution requests. The service ID should be specified network-byte order. If set to IB_CM_ASSIGN_SERVICE_ID, the CM will assign a service ID to the caller.

Description

If there's an existing ID listening on that same device and service ID, return it.

Callers should call ib_destroy_cm_id when done with the listener ID.

int rdma_rw_ctx_init(struct rdma_rw_ctx * ctx, struct ib_qp * qp, u8 port_num, struct scatterlist * sg, u32 sg_cnt, u32 sg_offset, u64 remote_addr, u32 rkey, enum dma_data_direction dir)

initialize a RDMA READ/WRITE context

Parameters

struct rdma_rw_ctx * ctx
context to initialize
struct ib_qp * qp
queue pair to operate on
u8 port_num
port num to which the connection is bound
struct scatterlist * sg
scatterlist to READ/WRITE from/to
u32 sg_cnt
number of entries in sg
u32 sg_offset
current byte offset into sg
u64 remote_addr
remote address to read/write (relative to rkey)
u32 rkey
remote key to operate on
enum dma_data_direction dir
DMA_TO_DEVICE for RDMA WRITE, DMA_FROM_DEVICE for RDMA READ

Description

Returns the number of WQEs that will be needed on the workqueue if successful, or a negative error code.

int rdma_rw_ctx_signature_init(struct rdma_rw_ctx * ctx, struct ib_qp * qp, u8 port_num, struct scatterlist * sg, u32 sg_cnt, struct scatterlist * prot_sg, u32 prot_sg_cnt, struct ib_sig_attrs * sig_attrs, u64 remote_addr, u32 rkey, enum dma_data_direction dir)

initialize a RW context with signature offload

Parameters

struct rdma_rw_ctx * ctx
context to initialize
struct ib_qp * qp
queue pair to operate on
u8 port_num
port num to which the connection is bound
struct scatterlist * sg
scatterlist to READ/WRITE from/to
u32 sg_cnt
number of entries in sg
struct scatterlist * prot_sg
scatterlist to READ/WRITE protection information from/to
u32 prot_sg_cnt
number of entries in prot_sg
struct ib_sig_attrs * sig_attrs
signature offloading algorithms
u64 remote_addr
remote address to read/write (relative to rkey)
u32 rkey
remote key to operate on
enum dma_data_direction dir
DMA_TO_DEVICE for RDMA WRITE, DMA_FROM_DEVICE for RDMA READ

Description

Returns the number of WQEs that will be needed on the workqueue if successful, or a negative error code.

struct ib_send_wr * rdma_rw_ctx_wrs(struct rdma_rw_ctx * ctx, struct ib_qp * qp, u8 port_num, struct ib_cqe * cqe, struct ib_send_wr * chain_wr)

return chain of WRs for a RDMA READ or WRITE operation

Parameters

struct rdma_rw_ctx * ctx
context to operate on
struct ib_qp * qp
queue pair to operate on
u8 port_num
port num to which the connection is bound
struct ib_cqe * cqe
completion queue entry for the last WR
struct ib_send_wr * chain_wr
WR to append to the posted chain

Description

Return the WR chain for the set of RDMA READ/WRITE operations described by ctx, as well as any memory registration operations needed. If chain_wr is non-NULL the WR it points to will be appended to the chain of WRs posted. If chain_wr is not set cqe must be set so that the caller gets a completion notification.

int rdma_rw_ctx_post(struct rdma_rw_ctx * ctx, struct ib_qp * qp, u8 port_num, struct ib_cqe * cqe, struct ib_send_wr * chain_wr)

post a RDMA READ or RDMA WRITE operation

Parameters

struct rdma_rw_ctx * ctx
context to operate on
struct ib_qp * qp
queue pair to operate on
u8 port_num
port num to which the connection is bound
struct ib_cqe * cqe
completion queue entry for the last WR
struct ib_send_wr * chain_wr
WR to append to the posted chain

Description

Post the set of RDMA READ/WRITE operations described by ctx, as well as any memory registration operations needed. If chain_wr is non-NULL the WR it points to will be appended to the chain of WRs posted. If chain_wr is not set cqe must be set so that the caller gets a completion notification.

void rdma_rw_ctx_destroy(struct rdma_rw_ctx * ctx, struct ib_qp * qp, u8 port_num, struct scatterlist * sg, u32 sg_cnt, enum dma_data_direction dir)

release all resources allocated by rdma_rw_ctx_init

Parameters

struct rdma_rw_ctx * ctx
context to release
struct ib_qp * qp
queue pair to operate on
u8 port_num
port num to which the connection is bound
struct scatterlist * sg
scatterlist that was used for the READ/WRITE
u32 sg_cnt
number of entries in sg
enum dma_data_direction dir
DMA_TO_DEVICE for RDMA WRITE, DMA_FROM_DEVICE for RDMA READ
void rdma_rw_ctx_destroy_signature(struct rdma_rw_ctx * ctx, struct ib_qp * qp, u8 port_num, struct scatterlist * sg, u32 sg_cnt, struct scatterlist * prot_sg, u32 prot_sg_cnt, enum dma_data_direction dir)

release all resources allocated by rdma_rw_ctx_signature_init

Parameters

struct rdma_rw_ctx * ctx
context to release
struct ib_qp * qp
queue pair to operate on
u8 port_num
port num to which the connection is bound
struct scatterlist * sg
scatterlist that was used for the READ/WRITE
u32 sg_cnt
number of entries in sg
struct scatterlist * prot_sg
scatterlist that was used for the READ/WRITE of the PI
u32 prot_sg_cnt
number of entries in prot_sg
enum dma_data_direction dir
DMA_TO_DEVICE for RDMA WRITE, DMA_FROM_DEVICE for RDMA READ
unsigned int rdma_rw_mr_factor(struct ib_device * device, u8 port_num, unsigned int maxpages)

return number of MRs required for a payload

Parameters

struct ib_device * device
device handling the connection
u8 port_num
port num to which the connection is bound
unsigned int maxpages
maximum payload pages per rdma_rw_ctx

Description

Returns the number of MRs the device requires to move maxpayload bytes. The returned value is used during transport creation to compute max_rdma_ctxts and the size of the transport's Send and Send Completion Queues.

bool rdma_dev_access_netns(const struct ib_device * dev, const struct net * net)

Return whether an rdma device can be accessed from a specified net namespace or not.

Parameters

const struct ib_device * dev
Pointer to rdma device which needs to be checked
const struct net * net
Pointer to net namesapce for which access to be checked

Description

When the rdma device is in shared mode, it ignores the net namespace. When the rdma device is exclusive to a net namespace, rdma device net namespace is checked against the specified one.

void ib_device_put(struct ib_device * device)

Release IB device reference

Parameters

struct ib_device * device
device whose reference to be released

Description

ib_device_put() releases reference to the IB device to allow it to be unregistered and eventually free.

struct ib_device * ib_device_get_by_name(const char * name, enum rdma_driver_id driver_id)

Find an IB device by name

Parameters

const char * name
The name to look for
enum rdma_driver_id driver_id
The driver ID that must match (RDMA_DRIVER_UNKNOWN matches all)

Description

Find and hold an ib_device by its name. The caller must call ib_device_put() on the returned pointer.

struct ib_device * _ib_alloc_device(size_t size)

allocate an IB device struct

Parameters

size_t size
size of structure to allocate

Description

Low-level drivers should use ib_alloc_device() to allocate struct ib_device. size is the size of the structure to be allocated, including any private data used by the low-level driver. ib_dealloc_device() must be used to free structures allocated with ib_alloc_device().

void ib_dealloc_device(struct ib_device * device)

free an IB device struct

Parameters

struct ib_device * device
structure to free

Description

Free a structure allocated with ib_alloc_device().

int ib_register_device(struct ib_device * device, const char * name)

Register an IB device with IB core

Parameters

struct ib_device * device
Device to register
const char * name
unique string device name. This may include a '%' which will cause a unique index to be added to the passed device name.

Description

Low-level drivers use ib_register_device() to register their devices with the IB core. All registered clients will receive a callback for each device that is added. device must be allocated with ib_alloc_device().

If the driver uses ops.dealloc_driver and calls any ib_unregister_device() asynchronously then the device pointer may become freed as soon as this function returns.

void ib_unregister_device(struct ib_device * ib_dev)

Unregister an IB device

Parameters

struct ib_device * ib_dev
The device to unregister

Description

Unregister an IB device. All clients will receive a remove callback.

Callers should call this routine only once, and protect against races with registration. Typically it should only be called as part of a remove callback in an implementation of driver core's struct device_driver and related.

If ops.dealloc_driver is used then ib_dev will be freed upon return from this function.

void ib_unregister_device_and_put(struct ib_device * ib_dev)

Unregister a device while holding a 'get'

Parameters

struct ib_device * ib_dev
The device to unregister

Description

This is the same as ib_unregister_device(), except it includes an internal ib_device_put() that should match a 'get' obtained by the caller.

It is safe to call this routine concurrently from multiple threads while holding the 'get'. When the function returns the device is fully unregistered.

Drivers using this flow MUST use the driver_unregister callback to clean up their resources associated with the device and dealloc it.

void ib_unregister_driver(enum rdma_driver_id driver_id)

Unregister all IB devices for a driver

Parameters

enum rdma_driver_id driver_id
The driver to unregister

Description

This implements a fence for device unregistration. It only returns once all devices associated with the driver_id have fully completed their unregistration and returned from ib_unregister_device*().

If device's are not yet unregistered it goes ahead and starts unregistering them.

This does not block creation of new devices with the given driver_id, that is the responsibility of the caller.

void ib_unregister_device_queued(struct ib_device * ib_dev)

Unregister a device using a work queue

Parameters

struct ib_device * ib_dev
The device to unregister

Description

This schedules an asynchronous unregistration using a WQ for the device. A driver should use this to avoid holding locks while doing unregistration, such as holding the RTNL lock.

Drivers using this API must use ib_unregister_driver before module unload to ensure that all scheduled unregistrations have completed.

int ib_register_client(struct ib_client * client)

Register an IB client

Parameters

struct ib_client * client
Client to register

Description

Upper level users of the IB drivers can use ib_register_client() to register callbacks for IB device addition and removal. When an IB device is added, each registered client's add method will be called (in the order the clients were registered), and when a device is removed, each client's remove method will be called (in the reverse order that clients were registered). In addition, when ib_register_client() is called, the client will receive an add callback for all devices already registered.

void ib_unregister_client(struct ib_client * client)

Unregister an IB client

Parameters

struct ib_client * client
Client to unregister

Description

Upper level users use ib_unregister_client() to remove their client registration. When ib_unregister_client() is called, the client will receive a remove callback for each IB device still registered.

This is a full fence, once it returns no client callbacks will be called, or are running in another thread.

void ib_set_client_data(struct ib_device * device, struct ib_client * client, void * data)

Set IB client context

Parameters

struct ib_device * device
Device to set context for
struct ib_client * client
Client to set context for
void * data
Context to set

Description

ib_set_client_data() sets client context data that can be retrieved with ib_get_client_data(). This can only be called while the client is registered to the device, once the ib_client remove() callback returns this cannot be called.

void ib_register_event_handler(struct ib_event_handler * event_handler)

Register an IB event handler

Parameters

struct ib_event_handler * event_handler
Handler to register

Description

ib_register_event_handler() registers an event handler that will be called back when asynchronous IB events occur (as defined in chapter 11 of the InfiniBand Architecture Specification). This callback may occur in interrupt context.

void ib_unregister_event_handler(struct ib_event_handler * event_handler)

Unregister an event handler

Parameters

struct ib_event_handler * event_handler
Handler to unregister

Description

Unregister an event handler registered with ib_register_event_handler().

void ib_dispatch_event(struct ib_event * event)

Dispatch an asynchronous event

Parameters

struct ib_event * event
Event to dispatch

Description

Low-level drivers must call ib_dispatch_event() to dispatch the event to all registered event handlers when an asynchronous event occurs.

int ib_query_port(struct ib_device * device, u8 port_num, struct ib_port_attr * port_attr)

Query IB port attributes

Parameters

struct ib_device * device
Device to query
u8 port_num
Port number to query
struct ib_port_attr * port_attr
Port attributes

Description

ib_query_port() returns the attributes of a port through the port_attr pointer.

int ib_device_set_netdev(struct ib_device * ib_dev, struct net_device * ndev, unsigned int port)

Associate the ib_dev with an underlying net_device

Parameters

struct ib_device * ib_dev
Device to modify
struct net_device * ndev
net_device to affiliate, may be NULL
unsigned int port
IB port the net_device is connected to

Description

Drivers should use this to link the ib_device to a netdev so the netdev shows up in interfaces like ib_enum_roce_netdev. Only one netdev may be affiliated with any port.

The caller must ensure that the given ndev is not unregistered or unregistering, and that either the ib_device is unregistered or ib_device_set_netdev() is called with NULL when the ndev sends a NETDEV_UNREGISTER event.

struct ib_device * ib_device_get_by_netdev(struct net_device * ndev, enum rdma_driver_id driver_id)

Find an IB device associated with a netdev

Parameters

struct net_device * ndev
netdev to locate
enum rdma_driver_id driver_id
The driver ID that must match (RDMA_DRIVER_UNKNOWN matches all)

Description

Find and hold an ib_device that is associated with a netdev via ib_device_set_netdev(). The caller must call ib_device_put() on the returned pointer.

int ib_query_pkey(struct ib_device * device, u8 port_num, u16 index, u16 * pkey)

Get P_Key table entry

Parameters

struct ib_device * device
Device to query
u8 port_num
Port number to query
u16 index
P_Key table index to query
u16 * pkey
Returned P_Key

Description

ib_query_pkey() fetches the specified P_Key table entry.

int ib_modify_device(struct ib_device * device, int device_modify_mask, struct ib_device_modify * device_modify)

Change IB device attributes

Parameters

struct ib_device * device
Device to modify
int device_modify_mask
Mask of attributes to change
struct ib_device_modify * device_modify
New attribute values

Description

ib_modify_device() changes a device's attributes as specified by the device_modify_mask and device_modify structure.

int ib_modify_port(struct ib_device * device, u8 port_num, int port_modify_mask, struct ib_port_modify * port_modify)

Modifies the attributes for the specified port.

Parameters

struct ib_device * device
The device to modify.
u8 port_num
The number of the port to modify.
int port_modify_mask
Mask used to specify which attributes of the port to change.
struct ib_port_modify * port_modify
New attribute values for the port.

Description

ib_modify_port() changes a port's attributes as specified by the port_modify_mask and port_modify structure.

int ib_find_gid(struct ib_device * device, union ib_gid * gid, u8 * port_num, u16 * index)

Returns the port number and GID table index where a specified GID value occurs. Its searches only for IB link layer.

Parameters

struct ib_device * device
The device to query.
union ib_gid * gid
The GID value to search for.
u8 * port_num
The port number of the device where the GID value was found.
u16 * index
The index into the GID table where the GID was found. This parameter may be NULL.
int ib_find_pkey(struct ib_device * device, u8 port_num, u16 pkey, u16 * index)

Returns the PKey table index where a specified PKey value occurs.

Parameters

struct ib_device * device
The device to query.
u8 port_num
The port number of the device to search for the PKey.
u16 pkey
The PKey value to search for.
u16 * index
The index into the PKey table where the PKey was found.
struct net_device * ib_get_net_dev_by_params(struct ib_device * dev, u8 port, u16 pkey, const union ib_gid * gid, const struct sockaddr * addr)

Return the appropriate net_dev for a received CM request

Parameters

struct ib_device * dev
An RDMA device on which the request has been received.
u8 port
Port number on the RDMA device.
u16 pkey
The Pkey the request came on.
const union ib_gid * gid
A GID that the net_dev uses to communicate.
const struct sockaddr * addr
Contains the IP address that the request specified as its destination.
struct ib_pd * __ib_alloc_pd(struct ib_device * device, unsigned int flags, const char * caller)

Allocates an unused protection domain.

Parameters

struct ib_device * device
The device on which to allocate the protection domain.
unsigned int flags
protection domain flags
const char * caller
caller's build-time module name

Description

A protection domain object provides an association between QPs, shared receive queues, address handles, memory regions, and memory windows.

Every PD has a local_dma_lkey which can be used as the lkey value for local memory operations.

void ib_dealloc_pd_user(struct ib_pd * pd, struct ib_udata * udata)

Deallocates a protection domain.

Parameters

struct ib_pd * pd
The protection domain to deallocate.
struct ib_udata * udata
Valid user data or NULL for kernel object

Description

It is an error to call this function while any resources in the pd still exist. The caller is responsible to synchronously destroy them and guarantee no new allocations will happen.

void rdma_copy_ah_attr(struct rdma_ah_attr * dest, const struct rdma_ah_attr * src)

Copy rdma ah attribute from source to destination.

Parameters

struct rdma_ah_attr * dest
Pointer to destination ah_attr. Contents of the destination pointer is assumed to be invalid and attribute are overwritten.
const struct rdma_ah_attr * src
Pointer to source ah_attr.
void rdma_replace_ah_attr(struct rdma_ah_attr * old, const struct rdma_ah_attr * new)

Replace valid ah_attr with new new one.

Parameters

struct rdma_ah_attr * old
Pointer to existing ah_attr which needs to be replaced. old is assumed to be valid or zero'd
const struct rdma_ah_attr * new
Pointer to the new ah_attr.

Description

rdma_replace_ah_attr() first releases any reference in the old ah_attr if old the ah_attr is valid; after that it copies the new attribute and holds the reference to the replaced ah_attr.

void rdma_move_ah_attr(struct rdma_ah_attr * dest, struct rdma_ah_attr * src)

Move ah_attr pointed by source to destination.

Parameters

struct rdma_ah_attr * dest
Pointer to destination ah_attr to copy to. dest is assumed to be valid or zero'd
struct rdma_ah_attr * src
Pointer to the new ah_attr.

Description

rdma_move_ah_attr() first releases any reference in the destination ah_attr if it is valid. This also transfers ownership of internal references from src to dest, making src invalid in the process. No new reference of the src ah_attr is taken.

struct ib_ah * rdma_create_ah(struct ib_pd * pd, struct rdma_ah_attr * ah_attr, u32 flags)

Creates an address handle for the given address vector.

Parameters

struct ib_pd * pd
The protection domain associated with the address handle.
struct rdma_ah_attr * ah_attr
The attributes of the address vector.
u32 flags
Create address handle flags (see enum rdma_create_ah_flags).

Description

It returns 0 on success and returns appropriate error code on error. The address handle is used to reference a local or global destination in all UD QP post sends.

struct ib_ah * rdma_create_user_ah(struct ib_pd * pd, struct rdma_ah_attr * ah_attr, struct ib_udata * udata)

Creates an address handle for the given address vector. It resolves destination mac address for ah attribute of RoCE type.

Parameters

struct ib_pd * pd
The protection domain associated with the address handle.
struct rdma_ah_attr * ah_attr
The attributes of the address vector.
struct ib_udata * udata
pointer to user's input output buffer information need by provider driver.

Description

It returns 0 on success and returns appropriate error code on error. The address handle is used to reference a local or global destination in all UD QP post sends.

void rdma_move_grh_sgid_attr(struct rdma_ah_attr * attr, union ib_gid * dgid, u32 flow_label, u8 hop_limit, u8 traffic_class, const struct ib_gid_attr * sgid_attr)

Sets the sgid attribute of GRH, taking ownership of the reference

Parameters

struct rdma_ah_attr * attr
Pointer to AH attribute structure
union ib_gid * dgid
Destination GID
u32 flow_label
Flow label
u8 hop_limit
Hop limit
u8 traffic_class
traffic class
const struct ib_gid_attr * sgid_attr
Pointer to SGID attribute

Description

This takes ownership of the sgid_attr reference. The caller must ensure rdma_destroy_ah_attr() is called before destroying the rdma_ah_attr after calling this function.

void rdma_destroy_ah_attr(struct rdma_ah_attr * ah_attr)

Release reference to SGID attribute of ah attribute.

Parameters

struct rdma_ah_attr * ah_attr
Pointer to ah attribute

Description

Release reference to the SGID attribute of the ah attribute if it is non NULL. It is safe to call this multiple times, and safe to call it on a zero initialized ah_attr.

int ib_modify_qp_with_udata(struct ib_qp * ib_qp, struct ib_qp_attr * attr, int attr_mask, struct ib_udata * udata)

Modifies the attributes for the specified QP.

Parameters

struct ib_qp * ib_qp
The QP to modify.
struct ib_qp_attr * attr
On input, specifies the QP attributes to modify. On output, the current values of selected QP attributes are returned.
int attr_mask
A bit-mask used to specify which attributes of the QP are being modified.
struct ib_udata * udata
pointer to user's input output buffer information are being modified. It returns 0 on success and returns appropriate error code on error.
struct ib_mr * ib_alloc_mr_user(struct ib_pd * pd, enum ib_mr_type mr_type, u32 max_num_sg, struct ib_udata * udata)

Allocates a memory region

Parameters

struct ib_pd * pd
protection domain associated with the region
enum ib_mr_type mr_type
memory region type
u32 max_num_sg
maximum sg entries available for registration.
struct ib_udata * udata
user data or null for kernel objects

Notes

Memory registeration page/sg lists must not exceed max_num_sg. For mr_type IB_MR_TYPE_MEM_REG, the total length cannot exceed max_num_sg * used_page_size.

struct ib_mr * ib_alloc_mr_integrity(struct ib_pd * pd, u32 max_num_data_sg, u32 max_num_meta_sg)

Allocates an integrity memory region

Parameters

struct ib_pd * pd
protection domain associated with the region
u32 max_num_data_sg
maximum data sg entries available for registration
u32 max_num_meta_sg
maximum metadata sg entries available for registration

Notes

Memory registration page/sg lists must not exceed max_num_sg, also the integrity page/sg lists must not exceed max_num_meta_sg.

struct ib_wq * ib_create_wq(struct ib_pd * pd, struct ib_wq_init_attr * wq_attr)

Creates a WQ associated with the specified protection domain.

Parameters

struct ib_pd * pd
The protection domain associated with the WQ.
struct ib_wq_init_attr * wq_attr
A list of initial attributes required to create the WQ. If WQ creation succeeds, then the attributes are updated to the actual capabilities of the created WQ.

Description

wq_attr->max_wr and wq_attr->max_sge determine the requested size of the WQ, and set to the actual values allocated on return. If ib_create_wq() succeeds, then max_wr and max_sge will always be at least as large as the requested values.

int ib_destroy_wq(struct ib_wq * wq, struct ib_udata * udata)

Destroys the specified user WQ.

Parameters

struct ib_wq * wq
The WQ to destroy.
struct ib_udata * udata
Valid user data
int ib_modify_wq(struct ib_wq * wq, struct ib_wq_attr * wq_attr, u32 wq_attr_mask)

Modifies the specified WQ.

Parameters

struct ib_wq * wq
The WQ to modify.
struct ib_wq_attr * wq_attr
On input, specifies the WQ attributes to modify.
u32 wq_attr_mask
A bit-mask used to specify which attributes of the WQ are being modified. On output, the current values of selected WQ attributes are returned.
int ib_map_mr_sg_pi(struct ib_mr * mr, struct scatterlist * data_sg, int data_sg_nents, unsigned int * data_sg_offset, struct scatterlist * meta_sg, int meta_sg_nents, unsigned int * meta_sg_offset, unsigned int page_size)

Map the dma mapped SG lists for PI (protection information) and set an appropriate memory region for registration.

Parameters

struct ib_mr * mr
memory region
struct scatterlist * data_sg
dma mapped scatterlist for data
int data_sg_nents
number of entries in data_sg
unsigned int * data_sg_offset
offset in bytes into data_sg
struct scatterlist * meta_sg
dma mapped scatterlist for metadata
int meta_sg_nents
number of entries in meta_sg
unsigned int * meta_sg_offset
offset in bytes into meta_sg
unsigned int page_size
page vector desired page size

Description

Constraints: - The MR must be allocated with type IB_MR_TYPE_INTEGRITY.

Return

0 on success.

After this completes successfully, the memory region is ready for registration.

int ib_map_mr_sg(struct ib_mr * mr, struct scatterlist * sg, int sg_nents, unsigned int * sg_offset, unsigned int page_size)

Map the largest prefix of a dma mapped SG list and set it the memory region.

Parameters

struct ib_mr * mr
memory region
struct scatterlist * sg
dma mapped scatterlist
int sg_nents
number of entries in sg
unsigned int * sg_offset
offset in bytes into sg
unsigned int page_size
page vector desired page size

Description

Constraints: - The first sg element is allowed to have an offset. - Each sg element must either be aligned to page_size or virtually

contiguous to the previous element. In case an sg element has a non-contiguous offset, the mapping prefix will not include it.
  • The last sg element is allowed to have length less than page_size.
  • If sg_nents total byte length exceeds the mr max_num_sge * page_size then only max_num_sg entries will be mapped.
  • If the MR was allocated with type IB_MR_TYPE_SG_GAPS, none of these constraints holds and the page_size argument is ignored.

Returns the number of sg elements that were mapped to the memory region.

After this completes successfully, the memory region is ready for registration.

int ib_sg_to_pages(struct ib_mr * mr, struct scatterlist * sgl, int sg_nents, unsigned int * sg_offset_p, int (*set_page) (struct ib_mr *, u64)

Convert the largest prefix of a sg list to a page vector

Parameters

struct ib_mr * mr
memory region
struct scatterlist * sgl
dma mapped scatterlist
int sg_nents
number of entries in sg
unsigned int * sg_offset_p

IN: start offset in bytes into sg OUT: offset in bytes for element n of the sg of the first

byte that has not been processed where n is the return value of this function.
int (*)(struct ib_mr *, u64) set_page
driver page assignment function pointer

Description

Core service helper for drivers to convert the largest prefix of given sg list to a page vector. The sg list prefix converted is the prefix that meet the requirements of ib_map_mr_sg.

Returns the number of sg elements that were assigned to a page vector.

void ib_drain_sq(struct ib_qp * qp)

Block until all SQ CQEs have been consumed by the application.

Parameters

struct ib_qp * qp
queue pair to drain

Description

If the device has a provider-specific drain function, then call that. Otherwise call the generic drain function __ib_drain_sq().

The caller must:

ensure there is room in the CQ and SQ for the drain work request and completion.

allocate the CQ using ib_alloc_cq().

ensure that there are no other contexts that are posting WRs concurrently. Otherwise the drain is not guaranteed.

void ib_drain_rq(struct ib_qp * qp)

Block until all RQ CQEs have been consumed by the application.

Parameters

struct ib_qp * qp
queue pair to drain

Description

If the device has a provider-specific drain function, then call that. Otherwise call the generic drain function __ib_drain_rq().

The caller must:

ensure there is room in the CQ and RQ for the drain work request and completion.

allocate the CQ using ib_alloc_cq().

ensure that there are no other contexts that are posting WRs concurrently. Otherwise the drain is not guaranteed.

void ib_drain_qp(struct ib_qp * qp)

Block until all CQEs have been consumed by the application on both the RQ and SQ.

Parameters

struct ib_qp * qp
queue pair to drain

Description

The caller must:

ensure there is room in the CQ(s), SQ, and RQ for drain work requests and completions.

allocate the CQs using ib_alloc_cq().

ensure that there are no other contexts that are posting WRs concurrently. Otherwise the drain is not guaranteed.

void ib_pack(const struct ib_field * desc, int desc_len, void * structure, void * buf)

Pack a structure into a buffer

Parameters

const struct ib_field * desc
Array of structure field descriptions
int desc_len
Number of entries in desc
void * structure
Structure to pack from
void * buf
Buffer to pack into

Description

ib_pack() packs a list of structure fields into a buffer, controlled by the array of fields in desc.

void ib_unpack(const struct ib_field * desc, int desc_len, void * buf, void * structure)

Unpack a buffer into a structure

Parameters

const struct ib_field * desc
Array of structure field descriptions
int desc_len
Number of entries in desc
void * buf
Buffer to unpack from
void * structure
Structure to unpack into

Description

ib_pack() unpacks a list of structure fields from a buffer, controlled by the array of fields in desc.

void ib_sa_cancel_query(int id, struct ib_sa_query * query)

try to cancel an SA query

Parameters

int id
ID of query to cancel
struct ib_sa_query * query
query pointer to cancel

Description

Try to cancel an SA query. If the id and query don't match up or the query has already completed, nothing is done. Otherwise the query is canceled and will complete with a status of -EINTR.

int ib_init_ah_attr_from_path(struct ib_device * device, u8 port_num, struct sa_path_rec * rec, struct rdma_ah_attr * ah_attr, const struct ib_gid_attr * gid_attr)

Initialize address handle attributes based on an SA path record.

Parameters

struct ib_device * device
Device associated ah attributes initialization.
u8 port_num
Port on the specified device.
struct sa_path_rec * rec
path record entry to use for ah attributes initialization.
struct rdma_ah_attr * ah_attr
address handle attributes to initialization from path record.
const struct ib_gid_attr * gid_attr
SGID attribute to consider during initialization.

Description

When ib_init_ah_attr_from_path() returns success, (a) for IB link layer it optionally contains a reference to SGID attribute when GRH is present for IB link layer. (b) for RoCE link layer it contains a reference to SGID attribute. User must invoke rdma_destroy_ah_attr() to release reference to SGID attributes which are initialized using ib_init_ah_attr_from_path().

int ib_sa_path_rec_get(struct ib_sa_client * client, struct ib_device * device, u8 port_num, struct sa_path_rec * rec, ib_sa_comp_mask comp_mask, unsigned long timeout_ms, gfp_t gfp_mask, void (*callback) (int status, struct sa_path_rec *resp, void *context, void * context, struct ib_sa_query ** sa_query)

Start a Path get query

Parameters

struct ib_sa_client * client
SA client
struct ib_device * device
device to send query on
u8 port_num
port number to send query on
struct sa_path_rec * rec
Path Record to send in query
ib_sa_comp_mask comp_mask
component mask to send in query
unsigned long timeout_ms
time to wait for response
gfp_t gfp_mask
GFP mask to use for internal allocations
void (*)(int status, struct sa_path_rec *resp, void *context) callback
function called when query completes, times out or is canceled
void * context
opaque user context passed to callback
struct ib_sa_query ** sa_query
query context, used to cancel query

Description

Send a Path Record Get query to the SA to look up a path. The callback function will be called when the query completes (or fails); status is 0 for a successful response, -EINTR if the query is canceled, -ETIMEDOUT is the query timed out, or -EIO if an error occurred sending the query. The resp parameter of the callback is only valid if status is 0.

If the return value of ib_sa_path_rec_get() is negative, it is an error code. Otherwise it is a query ID that can be used to cancel the query.

int ib_sa_service_rec_query(struct ib_sa_client * client, struct ib_device * device, u8 port_num, u8 method, struct ib_sa_service_rec * rec, ib_sa_comp_mask comp_mask, unsigned long timeout_ms, gfp_t gfp_mask, void (*callback) (int status, struct ib_sa_service_rec *resp, void *context, void * context, struct ib_sa_query ** sa_query)

Start Service Record operation

Parameters

struct ib_sa_client * client
SA client
struct ib_device * device
device to send request on
u8 port_num
port number to send request on
u8 method
SA method - should be get, set, or delete
struct ib_sa_service_rec * rec
Service Record to send in request
ib_sa_comp_mask comp_mask
component mask to send in request
unsigned long timeout_ms
time to wait for response
gfp_t gfp_mask
GFP mask to use for internal allocations
void (*)(int status, struct ib_sa_service_rec *resp, void *context) callback
function called when request completes, times out or is canceled
void * context
opaque user context passed to callback
struct ib_sa_query ** sa_query
request context, used to cancel request

Description

Send a Service Record set/get/delete to the SA to register, unregister or query a service record. The callback function will be called when the request completes (or fails); status is 0 for a successful response, -EINTR if the query is canceled, -ETIMEDOUT is the query timed out, or -EIO if an error occurred sending the query. The resp parameter of the callback is only valid if status is 0.

If the return value of ib_sa_service_rec_query() is negative, it is an error code. Otherwise it is a request ID that can be used to cancel the query.

int ib_ud_header_init(int payload_bytes, int lrh_present, int eth_present, int vlan_present, int grh_present, int ip_version, int udp_present, int immediate_present, struct ib_ud_header * header)

Initialize UD header structure

Parameters

int payload_bytes
Length of packet payload
int lrh_present
specify if LRH is present
int eth_present
specify if Eth header is present
int vlan_present
packet is tagged vlan
int grh_present
GRH flag (if non-zero, GRH will be included)
int ip_version
if non-zero, IP header, V4 or V6, will be included
int udp_present
if non-zero, UDP header will be included
int immediate_present
specify if immediate data is present
struct ib_ud_header * header
Structure to initialize
int ib_ud_header_pack(struct ib_ud_header * header, void * buf)

Pack UD header struct into wire format

Parameters

struct ib_ud_header * header
UD header struct
void * buf
Buffer to pack into

Description

ib_ud_header_pack() packs the UD header structure header into wire format in the buffer buf.

int ib_ud_header_unpack(void * buf, struct ib_ud_header * header)

Unpack UD header struct from wire format

Parameters

void * buf
Buffer to pack into
struct ib_ud_header * header
UD header struct

Description

ib_ud_header_pack() unpacks the UD header structure header from wire format in the buffer buf.

struct ib_fmr_pool * ib_create_fmr_pool(struct ib_pd * pd, struct ib_fmr_pool_param * params)

Create an FMR pool

Parameters

struct ib_pd * pd
Protection domain for FMRs
struct ib_fmr_pool_param * params
FMR pool parameters

Description

Create a pool of FMRs. Return value is pointer to new pool or error code if creation failed.

void ib_destroy_fmr_pool(struct ib_fmr_pool * pool)

Free FMR pool

Parameters

struct ib_fmr_pool * pool
FMR pool to free

Description

Destroy an FMR pool and free all associated resources.

int ib_flush_fmr_pool(struct ib_fmr_pool * pool)

Invalidate all unmapped FMRs

Parameters

struct ib_fmr_pool * pool
FMR pool to flush

Description

Ensure that all unmapped FMRs are fully invalidated.

struct ib_pool_fmr * ib_fmr_pool_map_phys(struct ib_fmr_pool * pool_handle, u64 * page_list, int list_len, u64 io_virtual_address)

Map an FMR from an FMR pool.

Parameters

struct ib_fmr_pool * pool_handle
FMR pool to allocate FMR from
u64 * page_list
List of pages to map
int list_len
Number of pages in page_list
u64 io_virtual_address
I/O virtual address for new FMR
void ib_fmr_pool_unmap(struct ib_pool_fmr * fmr)

Unmap FMR

Parameters

struct ib_pool_fmr * fmr
FMR to unmap

Description

Unmap an FMR. The FMR mapping may remain valid until the FMR is reused (or until ib_flush_fmr_pool() is called).

unsigned long ib_umem_find_best_pgsz(struct ib_umem * umem, unsigned long pgsz_bitmap, unsigned long virt)

Find best HW page size to use for this MR

Parameters

struct ib_umem * umem
umem struct
unsigned long pgsz_bitmap
bitmap of HW supported page sizes
unsigned long virt
IOVA

Description

This helper is intended for HW that support multiple page sizes but can do only a single page size in an MR.

Returns 0 if the umem requires page sizes not supported by the driver to be mapped. Drivers always supporting PAGE_SIZE or smaller will never see a 0 result.

struct ib_umem * ib_umem_get(struct ib_udata * udata, unsigned long addr, size_t size, int access)

Pin and DMA map userspace memory.

Parameters

struct ib_udata * udata
userspace context to pin memory for
unsigned long addr
userspace virtual address to start at
size_t size
length of region to pin
int access
IB_ACCESS_xxx flags for memory being pinned
void ib_umem_release(struct ib_umem * umem)

release memory pinned with ib_umem_get

Parameters

struct ib_umem * umem
umem struct to release
struct ib_umem_odp * ib_umem_odp_alloc_implicit(struct ib_udata * udata, int access)

Allocate a parent implicit ODP umem

Parameters

struct ib_udata * udata
udata from the syscall being used to create the umem
int access
ib_reg_mr access flags

Description

Implicit ODP umems do not have a VA range and do not have any page lists. They exist only to hold the per_mm reference to help the driver create children umems.

struct ib_umem_odp * ib_umem_odp_alloc_child(struct ib_umem_odp * root, unsigned long addr, size_t size, const struct mmu_interval_notifier_ops * ops)

Allocate a child ODP umem under an implicit parent ODP umem

Parameters

struct ib_umem_odp * root
The parent umem enclosing the child. This must be allocated using ib_alloc_implicit_odp_umem()
unsigned long addr
The starting userspace VA
size_t size
The length of the userspace VA
const struct mmu_interval_notifier_ops * ops
undescribed
struct ib_umem_odp * ib_umem_odp_get(struct ib_udata * udata, unsigned long addr, size_t size, int access, const struct mmu_interval_notifier_ops * ops)

Create a umem_odp for a userspace va

Parameters

struct ib_udata * udata
userspace context to pin memory for
unsigned long addr
userspace virtual address to start at
size_t size
length of region to pin
int access
IB_ACCESS_xxx flags for memory being pinned
const struct mmu_interval_notifier_ops * ops
undescribed

Description

The driver should use when the access flags indicate ODP memory. It avoids pinning, instead, stores the mm for future page fault handling in conjunction with MMU notifiers.

int ib_umem_odp_map_dma_pages(struct ib_umem_odp * umem_odp, u64 user_virt, u64 bcnt, u64 access_mask, unsigned long current_seq)

Pin and DMA map userspace memory in an ODP MR.

Parameters

struct ib_umem_odp * umem_odp
the umem to map and pin
u64 user_virt
the address from which we need to map.
u64 bcnt
the minimal number of bytes to pin and map. The mapping might be bigger due to alignment, and may also be smaller in case of an error pinning or mapping a page. The actual pages mapped is returned in the return value.
u64 access_mask
bit mask of the requested access permissions for the given range.
unsigned long current_seq
the MMU notifiers sequance value for synchronization with invalidations. the sequance number is read from umem_odp->notifiers_seq before calling this function

Description

Pins the range of pages passed in the argument, and maps them to DMA addresses. The DMA addresses of the mapped pages is updated in umem_odp->dma_list.

Returns the number of pages mapped in success, negative error code for failure. An -EAGAIN error code is returned when a concurrent mmu notifier prevents the function from completing its task. An -ENOENT error code indicates that userspace process is being terminated and mm was already destroyed.

RDMA Verbs transport library

int rvt_fast_reg_mr(struct rvt_qp * qp, struct ib_mr * ibmr, u32 key, int access)

fast register physical MR

Parameters

struct rvt_qp * qp
the queue pair where the work request comes from
struct ib_mr * ibmr
the memory region to be registered
u32 key
updated key for this memory region
int access
access flags for this memory region

Description

Returns 0 on success.

int rvt_invalidate_rkey(struct rvt_qp * qp, u32 rkey)

invalidate an MR rkey

Parameters

struct rvt_qp * qp
queue pair associated with the invalidate op
u32 rkey
rkey to invalidate

Description

Returns 0 on success.

int rvt_lkey_ok(struct rvt_lkey_table * rkt, struct rvt_pd * pd, struct rvt_sge * isge, struct rvt_sge * last_sge, struct ib_sge * sge, int acc)

check IB SGE for validity and initialize

Parameters

struct rvt_lkey_table * rkt
table containing lkey to check SGE against
struct rvt_pd * pd
protection domain
struct rvt_sge * isge
outgoing internal SGE
struct rvt_sge * last_sge
last outgoing SGE written
struct ib_sge * sge
SGE to check
int acc
access flags

Description

Check the IB SGE for validity and initialize our internal version of it.

Increments the reference count when a new sge is stored.

Return

0 if compressed, 1 if added , otherwise returns -errno.

int rvt_rkey_ok(struct rvt_qp * qp, struct rvt_sge * sge, u32 len, u64 vaddr, u32 rkey, int acc)

check the IB virtual address, length, and RKEY

Parameters

struct rvt_qp * qp
qp for validation
struct rvt_sge * sge
SGE state
u32 len
length of data
u64 vaddr
virtual address to place data
u32 rkey
rkey to check
int acc
access flags

Return

1 if successful, otherwise 0.

increments the reference count upon success

__be32 rvt_compute_aeth(struct rvt_qp * qp)

compute the AETH (syndrome + MSN)

Parameters

struct rvt_qp * qp
the queue pair to compute the AETH for

Description

Returns the AETH.

void rvt_get_credit(struct rvt_qp * qp, u32 aeth)

flush the send work queue of a QP

Parameters

struct rvt_qp * qp
the qp who's send work queue to flush
u32 aeth
the Acknowledge Extended Transport Header

Description

The QP s_lock should be held.

int rvt_check_ah(struct ib_device * ibdev, struct rdma_ah_attr * ah_attr)

validate the attributes of AH

Parameters

struct ib_device * ibdev
the ib device
struct rdma_ah_attr * ah_attr
the attributes of the AH

Description

If driver supports a more detailed check_ah function call back to it otherwise just check the basics.

Return

0 on success

struct rvt_dev_info * rvt_alloc_device(size_t size, int nports)

allocate rdi

Parameters

size_t size
how big of a structure to allocate
int nports
number of ports to allocate array slots for

Description

Use IB core device alloc to allocate space for the rdi which is assumed to be inside of the ib_device. Any extra space that drivers require should be included in size.

We also allocate a port array based on the number of ports.

Return

pointer to allocated rdi

void rvt_dealloc_device(struct rvt_dev_info * rdi)

deallocate rdi

Parameters

struct rvt_dev_info * rdi
structure to free

Description

Free a structure allocated with rvt_alloc_device()

int rvt_register_device(struct rvt_dev_info * rdi)

register a driver

Parameters

struct rvt_dev_info * rdi
main dev structure for all of rdmavt operations

Description

It is up to drivers to allocate the rdi and fill in the appropriate information.

Return

0 on success otherwise an errno.

void rvt_unregister_device(struct rvt_dev_info * rdi)

remove a driver

Parameters

struct rvt_dev_info * rdi
rvt dev struct
int rvt_init_port(struct rvt_dev_info * rdi, struct rvt_ibport * port, int port_index, u16 * pkey_table)

init internal data for driver port

Parameters

struct rvt_dev_info * rdi
rvt_dev_info struct
struct rvt_ibport * port
rvt port
int port_index
0 based index of ports, different from IB core port num
u16 * pkey_table
pkey_table for port

Description

Keep track of a list of ports. No need to have a detach port. They persist until the driver goes away.

Return

always 0

bool rvt_cq_enter(struct rvt_cq * cq, struct ib_wc * entry, bool solicited)

add a new entry to the completion queue

Parameters

struct rvt_cq * cq
completion queue
struct ib_wc * entry
work completion entry to add
bool solicited
true if entry is solicited

Description

This may be called with qp->s_lock held.

Return

return true on success, else return false if cq is full.

int rvt_error_qp(struct rvt_qp * qp, enum ib_wc_status err)

put a QP into the error state

Parameters

struct rvt_qp * qp
the QP to put into the error state
enum ib_wc_status err
the receive completion error to signal if a RWQE is active

Description

Flushes both send and receive work queues.

Return

true if last WQE event should be generated. The QP r_lock and s_lock should be held and interrupts disabled. If we are already in error state, just return.

int rvt_get_rwqe(struct rvt_qp * qp, bool wr_id_only)

copy the next RWQE into the QP's RWQE

Parameters

struct rvt_qp * qp
the QP
bool wr_id_only
update qp->r_wr_id only, not qp->r_sge

Description

Return -1 if there is a local error, 0 if no RWQE is available, otherwise return 1.

Can be called from interrupt level.

void rvt_comm_est(struct rvt_qp * qp)

handle trap with QP established

Parameters

struct rvt_qp * qp
the QP
void rvt_add_rnr_timer(struct rvt_qp * qp, u32 aeth)

add/start an rnr timer on the QP

Parameters

struct rvt_qp * qp
the QP
u32 aeth
aeth of RNR timeout, simulated aeth for loopback
void rvt_stop_rc_timers(struct rvt_qp * qp)

stop all timers

Parameters

struct rvt_qp * qp
the QP stop any pending timers
void rvt_del_timers_sync(struct rvt_qp * qp)

wait for any timeout routines to exit

Parameters

struct rvt_qp * qp
the QP
struct rvt_qp_iter * rvt_qp_iter_init(struct rvt_dev_info * rdi, u64 v, void (*cb) (struct rvt_qp *qp, u64 v)

initial for QP iteration

Parameters

struct rvt_dev_info * rdi
rvt devinfo
u64 v
u64 value
void (*)(struct rvt_qp *qp, u64 v) cb
user-defined callback

Description

This returns an iterator suitable for iterating QPs in the system.

The cb is a user-defined callback and v is a 64-bit value passed to and relevant for processing in the cb. An example use case would be to alter QP processing based on criteria not part of the rvt_qp.

Use cases that require memory allocation to succeed must preallocate appropriately.

Return

a pointer to an rvt_qp_iter or NULL

int rvt_qp_iter_next(struct rvt_qp_iter * iter)

return the next QP in iter

Parameters

struct rvt_qp_iter * iter
the iterator

Description

Fine grained QP iterator suitable for use with debugfs seq_file mechanisms.

Updates iter->qp with the current QP when the return value is 0.

Return

0 - iter->qp is valid 1 - no more QPs

void rvt_qp_iter(struct rvt_dev_info * rdi, u64 v, void (*cb) (struct rvt_qp *qp, u64 v)

iterate all QPs

Parameters

struct rvt_dev_info * rdi
rvt devinfo
u64 v
a 64-bit value
void (*)(struct rvt_qp *qp, u64 v) cb
a callback

Description

This provides a way for iterating all QPs.

The cb is a user-defined callback and v is a 64-bit value passed to and relevant for processing in the cb. An example use case would be to alter QP processing based on criteria not part of the rvt_qp.

The code has an internal iterator to simplify non seq_file use cases.

void rvt_copy_sge(struct rvt_qp * qp, struct rvt_sge_state * ss, void * data, u32 length, bool release, bool copy_last)

copy data to SGE memory

Parameters

struct rvt_qp * qp
associated QP
struct rvt_sge_state * ss
the SGE state
void * data
the data to copy
u32 length
the length of the data
bool release
boolean to release MR
bool copy_last
do a separate copy of the last 8 bytes
void rvt_ruc_loopback(struct rvt_qp * sqp)

handle UC and RC loopback requests

Parameters

struct rvt_qp * sqp
the sending QP

Description

This is called from rvt_do_send() to forward a WQE addressed to the same HFI Note that although we are single threaded due to the send engine, we still have to protect against post_send(). We don't have to worry about receive interrupts since this is a connected protocol and all packets will pass through here.

struct rvt_mcast * rvt_mcast_find(struct rvt_ibport * ibp, union ib_gid * mgid, u16 lid)

search the global table for the given multicast GID/LID

Parameters

struct rvt_ibport * ibp
the IB port structure
union ib_gid * mgid
the multicast GID to search for
u16 lid
the multicast LID portion of the multicast address (host order)

NOTE

It is valid to have 1 MLID with multiple MGIDs. It is not valid to have 1 MGID with multiple MLIDs.

Description

The caller is responsible for decrementing the reference count if found.

Return

NULL if not found.

Upper Layer Protocols

iSCSI Extensions for RDMA (iSER)

struct iser_data_buf

iSER data buffer

Definition

struct iser_data_buf {
  struct scatterlist *sg;
  int size;
  unsigned long      data_len;
  int dma_nents;
};

Members

sg
pointer to the sg list
size
num entries of this sg
data_len
total beffer byte len
dma_nents
returned by dma_map_sg
struct iser_mem_reg

iSER memory registration info

Definition

struct iser_mem_reg {
  struct ib_sge    sge;
  u32 rkey;
  void *mem_h;
};

Members

sge
memory region sg element
rkey
memory region remote key
mem_h
pointer to registration context (FMR/Fastreg)
struct iser_tx_desc

iSER TX descriptor

Definition

struct iser_tx_desc {
  struct iser_ctrl             iser_header;
  struct iscsi_hdr             iscsi_header;
  enum iser_desc_type        type;
  u64 dma_addr;
  struct ib_sge                tx_sg[2];
  int num_sge;
  struct ib_cqe                cqe;
  bool mapped;
  struct ib_reg_wr             reg_wr;
  struct ib_send_wr            send_wr;
  struct ib_send_wr            inv_wr;
};

Members

iser_header
iser header
iscsi_header
iscsi header
type
command/control/dataout
dma_addr
header buffer dma_address
tx_sg
sg[0] points to iser/iscsi headers sg[1] optionally points to either of immediate data unsolicited data-out or control
num_sge
number sges used on this TX task
cqe
completion handler
mapped
Is the task header mapped
reg_wr
registration WR
send_wr
send WR
inv_wr
invalidate WR
struct iser_rx_desc

iSER RX descriptor

Definition

struct iser_rx_desc {
  struct iser_ctrl             iser_header;
  struct iscsi_hdr             iscsi_header;
  char data[ISER_RECV_DATA_SEG_LEN];
  u64 dma_addr;
  struct ib_sge                rx_sg;
  struct ib_cqe                cqe;
  char pad[ISER_RX_PAD_SIZE];
};

Members

iser_header
iser header
iscsi_header
iscsi header
data
received data segment
dma_addr
receive buffer dma address
rx_sg
ib_sge of receive buffer
cqe
completion handler
pad
for sense data TODO: Modify to maximum sense length supported
struct iser_login_desc

iSER login descriptor

Definition

struct iser_login_desc {
  void *req;
  void *rsp;
  u64 req_dma;
  u64 rsp_dma;
  struct ib_sge                sge;
  struct ib_cqe                cqe;
};

Members

req
pointer to login request buffer
rsp
pointer to login response buffer
req_dma
DMA address of login request buffer
rsp_dma
DMA address of login response buffer
sge
IB sge for login post recv
cqe
completion handler
struct iser_comp

iSER completion context

Definition

struct iser_comp {
  struct ib_cq            *cq;
  int active_qps;
};

Members

cq
completion queue
active_qps
Number of active QPs attached to completion context
struct iser_reg_ops

Memory registration operations per-device registration schemes

Definition

struct iser_reg_ops {
  int (*alloc_reg_res)(struct ib_conn *ib_conn,unsigned cmds_max, unsigned int size);
  void (*free_reg_res)(struct ib_conn *ib_conn);
  int (*reg_mem)(struct iscsi_iser_task *iser_task,struct iser_data_buf *mem,struct iser_reg_resources *rsc, struct iser_mem_reg *reg);
  void (*unreg_mem)(struct iscsi_iser_task *iser_task, enum iser_data_dir cmd_dir);
  struct iser_fr_desc * (*reg_desc_get)(struct ib_conn *ib_conn);
  void (*reg_desc_put)(struct ib_conn *ib_conn, struct iser_fr_desc *desc);
};

Members

alloc_reg_res
Allocate registration resources
free_reg_res
Free registration resources
reg_mem
Register memory buffers
unreg_mem
Un-register memory buffers
reg_desc_get
Get a registration descriptor for pool
reg_desc_put
Get a registration descriptor to pool
struct iser_device

iSER device handle

Definition

struct iser_device {
  struct ib_device             *ib_device;
  struct ib_pd                 *pd;
  struct ib_event_handler      event_handler;
  struct list_head             ig_list;
  int refcount;
  int comps_used;
  struct iser_comp             *comps;
  const struct iser_reg_ops    *reg_ops;
  bool remote_inv_sup;
};

Members

ib_device
RDMA device
pd
Protection Domain for this device
event_handler
IB events handle routine
ig_list
entry in devices list
refcount
Reference counter, dominated by open iser connections
comps_used
Number of completion contexts used, Min between online cpus and device max completion vectors
comps
Dinamically allocated array of completion handlers
reg_ops
Registration ops
remote_inv_sup
Remote invalidate is supported on this device
struct iser_reg_resources

Fast registration resources

Definition

struct iser_reg_resources {
  union {
    struct ib_mr             *mr;
    struct ib_fmr_pool       *fmr_pool;
  };
  struct ib_mr                     *sig_mr;
  struct iser_page_vec             *page_vec;
  u8 mr_valid:1;
};

Members

{unnamed_union}
anonymous
mr
memory region
fmr_pool
pool of fmrs
sig_mr
signature memory region
page_vec
fast reg page list used by fmr pool
mr_valid
is mr valid indicator
struct iser_fr_desc

Fast registration descriptor

Definition

struct iser_fr_desc {
  struct list_head                  list;
  struct iser_reg_resources         rsc;
  bool sig_protected;
  struct list_head                  all_list;
};

Members

list
entry in connection fastreg pool
rsc
data buffer registration resources
sig_protected
is region protected indicator
struct iser_fr_pool

connection fast registration pool

Definition

struct iser_fr_pool {
  struct list_head        list;
  spinlock_t lock;
  int size;
  struct list_head        all_list;
};

Members

list
list of fastreg descriptors
lock
protects fmr/fastreg pool
size
size of the pool
struct ib_conn

Infiniband related objects

Definition

struct ib_conn {
  struct rdma_cm_id           *cma_id;
  struct ib_qp                *qp;
  int post_recv_buf_count;
  u8 sig_count;
  struct ib_recv_wr            rx_wr[ISER_MIN_POSTED_RX];
  struct iser_device          *device;
  struct iser_comp            *comp;
  struct iser_fr_pool          fr_pool;
  bool pi_support;
  struct ib_cqe                reg_cqe;
};

Members

cma_id
rdma_cm connection maneger handle
qp
Connection Queue-pair
post_recv_buf_count
post receive counter
sig_count
send work request signal count
rx_wr
receive work request for batch posts
device
reference to iser device
comp
iser completion context
fr_pool
connection fast registration poool
pi_support
Indicate device T10-PI support
reg_cqe
completion handler
struct iser_conn

iSER connection context

Definition

struct iser_conn {
  struct ib_conn               ib_conn;
  struct iscsi_conn            *iscsi_conn;
  struct iscsi_endpoint        *ep;
  enum iser_conn_state         state;
  unsigned qp_max_recv_dtos;
  unsigned qp_max_recv_dtos_mask;
  unsigned min_posted_rx;
  u16 max_cmds;
  char name[ISER_OBJECT_NAME_SIZE];
  struct work_struct           release_work;
  struct mutex                 state_mutex;
  struct completion            stop_completion;
  struct completion            ib_completion;
  struct completion            up_completion;
  struct list_head             conn_list;
  struct iser_login_desc       login_desc;
  unsigned int                 rx_desc_head;
  struct iser_rx_desc          *rx_descs;
  u32 num_rx_descs;
  unsigned short               scsi_sg_tablesize;
  unsigned short               pages_per_mr;
  bool snd_w_inv;
};

Members

ib_conn
connection RDMA resources
iscsi_conn
link to matching iscsi connection
ep
transport handle
state
connection logical state
qp_max_recv_dtos
maximum number of data outs, corresponds to max number of post recvs
qp_max_recv_dtos_mask
(qp_max_recv_dtos - 1)
min_posted_rx
(qp_max_recv_dtos >> 2)
max_cmds
maximum cmds allowed for this connection
name
connection peer portal
release_work
deffered work for release job
state_mutex
protects iser onnection state
stop_completion
conn_stop completion
ib_completion
RDMA cleanup completion
up_completion
connection establishment completed (state is ISER_CONN_UP)
conn_list
entry in ig conn list
login_desc
login descriptor
rx_desc_head
head of rx_descs cyclic buffer
rx_descs
rx buffers array (cyclic buffer)
num_rx_descs
number of rx descriptors
scsi_sg_tablesize
scsi host sg_tablesize
pages_per_mr
maximum pages available for registration
snd_w_inv
connection uses remote invalidation
struct iscsi_iser_task

iser task context

Definition

struct iscsi_iser_task {
  struct iser_tx_desc          desc;
  struct iser_conn             *iser_conn;
  enum iser_task_status        status;
  struct scsi_cmnd             *sc;
  int command_sent;
  int dir[ISER_DIRS_NUM];
  struct iser_mem_reg          rdma_reg[ISER_DIRS_NUM];
  struct iser_data_buf         data[ISER_DIRS_NUM];
  struct iser_data_buf         prot[ISER_DIRS_NUM];
};

Members

desc
TX descriptor
iser_conn
link to iser connection
status
current task status
sc
link to scsi command
command_sent
indicate if command was sent
dir
iser data direction
rdma_reg
task rdma registration desc
data
iser data buffer desc
prot
iser protection buffer desc
struct iser_global

iSER global context

Definition

struct iser_global {
  struct mutex      device_list_mutex;
  struct list_head  device_list;
  struct mutex      connlist_mutex;
  struct list_head  connlist;
  struct kmem_cache *desc_cache;
};

Members

device_list_mutex
protects device_list
device_list
iser devices global list
connlist_mutex
protects connlist
connlist
iser connections global list
desc_cache
kmem cache for tx dataout
int iscsi_iser_pdu_alloc(struct iscsi_task * task, uint8_t opcode)

allocate an iscsi-iser PDU

Parameters

struct iscsi_task * task
iscsi task
uint8_t opcode
iscsi command opcode

Description

Netes: This routine can't fail, just assign iscsi task
hdr and max hdr size.
int iser_initialize_task_headers(struct iscsi_task * task, struct iser_tx_desc * tx_desc)

Initialize task headers

Parameters

struct iscsi_task * task
iscsi task
struct iser_tx_desc * tx_desc
iser tx descriptor

Notes

This routine may race with iser teardown flow for scsi error handling TMFs. So for TMF we should acquire the state mutex to avoid dereferencing the IB device which may have already been terminated.

int iscsi_iser_task_init(struct iscsi_task * task)

Initialize iscsi-iser task

Parameters

struct iscsi_task * task
iscsi task

Description

Initialize the task for the scsi command or mgmt command.

Return

Returns zero on success or -ENOMEM when failing
to init task headers (dma mapping error).
int iscsi_iser_mtask_xmit(struct iscsi_conn * conn, struct iscsi_task * task)

xmit management (immediate) task

Parameters

struct iscsi_conn * conn
iscsi connection
struct iscsi_task * task
task management task

Notes

The function can return -EAGAIN in which case caller must call it again later, or recover. '0' return code means successful xmit.
int iscsi_iser_task_xmit(struct iscsi_task * task)

xmit iscsi-iser task

Parameters

struct iscsi_task * task
iscsi task

Return

zero on success or escalates $error on failure.

void iscsi_iser_cleanup_task(struct iscsi_task * task)

cleanup an iscsi-iser task

Parameters

struct iscsi_task * task
iscsi task

Notes

In case the RDMA device is already NULL (might have
been removed in DEVICE_REMOVAL CM event it will bail-out without doing dma unmapping.
u8 iscsi_iser_check_protection(struct iscsi_task * task, sector_t * sector)

check protection information status of task.

Parameters

struct iscsi_task * task
iscsi task
sector_t * sector
error sector if exsists (output)

Return

zero if no data-integrity errors have occured

0x1: data-integrity error occured in the guard-block 0x2: data-integrity error occured in the reference tag 0x3: data-integrity error occured in the application tag

In addition the error sector is marked.

struct iscsi_cls_conn * iscsi_iser_conn_create(struct iscsi_cls_session * cls_session, uint32_t conn_idx)

create a new iscsi-iser connection

Parameters

struct iscsi_cls_session * cls_session
iscsi class connection
uint32_t conn_idx
connection index within the session (for MCS)

Return

iscsi_cls_conn when iscsi_conn_setup succeeds or NULL
otherwise.
int iscsi_iser_conn_bind(struct iscsi_cls_session * cls_session, struct iscsi_cls_conn * cls_conn, uint64_t transport_eph, int is_leading)

bind iscsi and iser connection structures

Parameters

struct iscsi_cls_session * cls_session
iscsi class session
struct iscsi_cls_conn * cls_conn
iscsi class connection
uint64_t transport_eph
transport end-point handle
int is_leading
indicate if this is the session leading connection (MCS)

Return

zero on success, $error if iscsi_conn_bind fails and
-EINVAL in case end-point doesn't exsits anymore or iser connection state is not UP (teardown already started).
int iscsi_iser_conn_start(struct iscsi_cls_conn * cls_conn)

start iscsi-iser connection

Parameters

struct iscsi_cls_conn * cls_conn
iscsi class connection

Notes

Here iser intialize (or re-initialize) stop_completion as
from this point iscsi must call conn_stop in session/connection teardown so iser transport must wait for it.
void iscsi_iser_conn_stop(struct iscsi_cls_conn * cls_conn, int flag)

stop iscsi-iser connection

Parameters

struct iscsi_cls_conn * cls_conn
iscsi class connection
int flag
indicate if recover or terminate (passed as is)

Notes

Calling iscsi_conn_stop might theoretically race with
DEVICE_REMOVAL event and dereference a previously freed RDMA device handle, so we call it under iser the state lock to protect against this kind of race.
void iscsi_iser_session_destroy(struct iscsi_cls_session * cls_session)

destroy iscsi-iser session

Parameters

struct iscsi_cls_session * cls_session
iscsi class session

Description

Removes and free iscsi host.

struct iscsi_cls_session * iscsi_iser_session_create(struct iscsi_endpoint * ep, uint16_t cmds_max, uint16_t qdepth, uint32_t initial_cmdsn)

create an iscsi-iser session

Parameters

struct iscsi_endpoint * ep
iscsi end-point handle
uint16_t cmds_max
maximum commands in this session
uint16_t qdepth
session command queue depth
uint32_t initial_cmdsn
initiator command sequnce number

Description

Allocates and adds a scsi host, expose DIF supprot if exists, and sets up an iscsi session.

struct iscsi_endpoint * iscsi_iser_ep_connect(struct Scsi_Host * shost, struct sockaddr * dst_addr, int non_blocking)

Initiate iSER connection establishment

Parameters

struct Scsi_Host * shost
scsi_host
struct sockaddr * dst_addr
destination address
int non_blocking
indicate if routine can block

Description

Allocate an iscsi endpoint, an iser_conn structure and bind them. After that start RDMA connection establishment via rdma_cm. We don't allocate iser_conn embedded in iscsi_endpoint since in teardown the endpoint will be destroyed at ep_disconnect while iser_conn will cleanup its resources asynchronuously.

Return

iscsi_endpoint created by iscsi layer or ERR_PTR(error)
if fails.
int iscsi_iser_ep_poll(struct iscsi_endpoint * ep, int timeout_ms)

poll for iser connection establishment to complete

Parameters

struct iscsi_endpoint * ep
iscsi endpoint (created at ep_connect)
int timeout_ms
polling timeout allowed in ms.

Description

This routine boils down to waiting for up_completion signaling that cma_id got CONNECTED event.

Return

1 if succeeded in connection establishment, 0 if timeout expired
(libiscsi will retry will kick in) or -1 if interrupted by signal or more likely iser connection state transitioned to TEMINATING or DOWN during the wait period.
void iscsi_iser_ep_disconnect(struct iscsi_endpoint * ep)

Initiate connection teardown process

Parameters

struct iscsi_endpoint * ep
iscsi endpoint handle

Description

This routine is not blocked by iser and RDMA termination process completion as we queue a deffered work for iser/RDMA destruction and cleanup or actually call it immediately in case we didn't pass iscsi conn bind/start stage, thus it is safe.

int iser_send_command(struct iscsi_conn * conn, struct iscsi_task * task)

send command PDU

Parameters

struct iscsi_conn * conn
link to matching iscsi connection
struct iscsi_task * task
SCSI command task
int iser_send_data_out(struct iscsi_conn * conn, struct iscsi_task * task, struct iscsi_data * hdr)

send data out PDU

Parameters

struct iscsi_conn * conn
link to matching iscsi connection
struct iscsi_task * task
SCSI command task
struct iscsi_data * hdr
pointer to the LLD's iSCSI message header
int iser_alloc_fmr_pool(struct ib_conn * ib_conn, unsigned cmds_max, unsigned int size)

Creates FMR pool and page_vector

Parameters

struct ib_conn * ib_conn
connection RDMA resources
unsigned cmds_max
max number of SCSI commands for this connection
unsigned int size
max number of pages per map request

Return

0 on success, or errno code on failure

void iser_free_fmr_pool(struct ib_conn * ib_conn)

releases the FMR pool and page vec

Parameters

struct ib_conn * ib_conn
connection RDMA resources
int iser_alloc_fastreg_pool(struct ib_conn * ib_conn, unsigned cmds_max, unsigned int size)

Creates pool of fast_reg descriptors for fast registration work requests.

Parameters

struct ib_conn * ib_conn
connection RDMA resources
unsigned cmds_max
max number of SCSI commands for this connection
unsigned int size
max number of pages per map request

Return

0 on success, or errno code on failure

void iser_free_fastreg_pool(struct ib_conn * ib_conn)

releases the pool of fast_reg descriptors

Parameters

struct ib_conn * ib_conn
connection RDMA resources
void iser_free_ib_conn_res(struct iser_conn * iser_conn, bool destroy)

release IB related resources

Parameters

struct iser_conn * iser_conn
iser connection struct
bool destroy
indicator if we need to try to release the iser device and memory regoins pool (only iscsi shutdown and DEVICE_REMOVAL will use this).

Description

This routine is called with the iser state mutex held so the cm_id removal is out of here. It is Safe to be invoked multiple times.

void iser_conn_release(struct iser_conn * iser_conn)

Frees all conn objects and deallocs conn descriptor

Parameters

struct iser_conn * iser_conn
iSER connection context
int iser_conn_terminate(struct iser_conn * iser_conn)

triggers start of the disconnect procedures and waits for them to be done

Parameters

struct iser_conn * iser_conn
iSER connection context

Description

Called with state mutex held

int iser_post_send(struct ib_conn * ib_conn, struct iser_tx_desc * tx_desc, bool signal)

Initiate a Send DTO operation

Parameters

struct ib_conn * ib_conn
connection RDMA resources
struct iser_tx_desc * tx_desc
iSER TX descriptor
bool signal
true to send work request as SIGNALED

Return

0 on success, -1 on failure

Omni-Path (OPA) Virtual NIC support

struct opa_vnic_ctrl_port

OPA virtual NIC control port

Definition

struct opa_vnic_ctrl_port {
  struct ib_device           *ibdev;
  struct opa_vnic_ctrl_ops   *ops;
  u8 num_ports;
};

Members

ibdev
pointer to ib device
ops
opa vnic control operations
num_ports
number of opa ports
struct opa_vnic_adapter

OPA VNIC netdev private data structure

Definition

struct opa_vnic_adapter {
  struct net_device             *netdev;
  struct ib_device              *ibdev;
  struct opa_vnic_ctrl_port     *cport;
  const struct net_device_ops   *rn_ops;
  u8 port_num;
  u8 vport_num;
  struct mutex lock;
  struct __opa_veswport_info  info;
  u8 vema_mac_addr[ETH_ALEN];
  u32 umac_hash;
  u32 mmac_hash;
  struct hlist_head  __rcu   *mactbl;
  struct mutex mactbl_lock;
  spinlock_t stats_lock;
  u8 flow_tbl[OPA_VNIC_FLOW_TBL_SIZE];
  unsigned long trap_timeout;
  u8 trap_count;
};

Members

netdev
pointer to associated netdev
ibdev
ib device
cport
pointer to opa vnic control port
rn_ops
rdma netdev's net_device_ops
port_num
OPA port number
vport_num
vesw port number
lock
adapter lock
info
virtual ethernet switch port information
vema_mac_addr
mac address configured by vema
umac_hash
unicast maclist hash
mmac_hash
multicast maclist hash
mactbl
hash table of MAC entries
mactbl_lock
mac table lock
stats_lock
statistics lock
flow_tbl
flow to default port redirection table
trap_timeout
trap timeout
trap_count
no. of traps allowed within timeout period
struct opa_vnic_mac_tbl_node

OPA VNIC mac table node

Definition

struct opa_vnic_mac_tbl_node {
  struct hlist_node                    hlist;
  u16 index;
  struct __opa_vnic_mactable_entry     entry;
};

Members

hlist
hash list handle
index
index of entry in the mac table
entry
entry in the table
struct opa_vesw_info

OPA vnic switch information

Definition

struct opa_vesw_info {
  __be16 fabric_id;
  __be16 vesw_id;
  u8 rsvd0[6];
  __be16 def_port_mask;
  u8 rsvd1[2];
  __be16 pkey;
  u8 rsvd2[4];
  __be32 u_mcast_dlid;
  __be32 u_ucast_dlid[OPA_VESW_MAX_NUM_DEF_PORT];
  __be32 rc;
  u8 rsvd3[56];
  __be16 eth_mtu;
  u8 rsvd4[2];
};

Members

fabric_id
10-bit fabric id
vesw_id
12-bit virtual ethernet switch id
def_port_mask
bitmask of default ports
pkey
partition key
u_mcast_dlid
unknown multicast dlid
u_ucast_dlid
array of unknown unicast dlids
rc
routing control
eth_mtu
Ethernet MTU
struct opa_per_veswport_info

OPA vnic per port information

Definition

struct opa_per_veswport_info {
  __be32 port_num;
  u8 eth_link_status;
  u8 rsvd0[3];
  u8 base_mac_addr[ETH_ALEN];
  u8 config_state;
  u8 oper_state;
  __be16 max_mac_tbl_ent;
  __be16 max_smac_ent;
  __be32 mac_tbl_digest;
  u8 rsvd1[4];
  __be32 encap_slid;
  u8 pcp_to_sc_uc[OPA_VNIC_MAX_NUM_PCP];
  u8 pcp_to_vl_uc[OPA_VNIC_MAX_NUM_PCP];
  u8 pcp_to_sc_mc[OPA_VNIC_MAX_NUM_PCP];
  u8 pcp_to_vl_mc[OPA_VNIC_MAX_NUM_PCP];
  u8 non_vlan_sc_uc;
  u8 non_vlan_vl_uc;
  u8 non_vlan_sc_mc;
  u8 non_vlan_vl_mc;
  u8 rsvd2[48];
  __be16 uc_macs_gen_count;
  __be16 mc_macs_gen_count;
  u8 rsvd3[8];
};

Members

port_num
port number
eth_link_status
current ethernet link state
base_mac_addr
base mac address
config_state
configured port state
oper_state
operational port state
max_mac_tbl_ent
max number of mac table entries
max_smac_ent
max smac entries in mac table
mac_tbl_digest
mac table digest
encap_slid
base slid for the port
pcp_to_sc_uc
sc by pcp index for unicast ethernet packets
pcp_to_vl_uc
vl by pcp index for unicast ethernet packets
pcp_to_sc_mc
sc by pcp index for multicast ethernet packets
pcp_to_vl_mc
vl by pcp index for multicast ethernet packets
non_vlan_sc_uc
sc for non-vlan unicast ethernet packets
non_vlan_vl_uc
vl for non-vlan unicast ethernet packets
non_vlan_sc_mc
sc for non-vlan multicast ethernet packets
non_vlan_vl_mc
vl for non-vlan multicast ethernet packets
uc_macs_gen_count
generation count for unicast macs list
mc_macs_gen_count
generation count for multicast macs list
struct opa_veswport_info

OPA vnic port information

Definition

struct opa_veswport_info {
  struct opa_vesw_info          vesw;
  struct opa_per_veswport_info  vport;
};

Members

vesw
OPA vnic switch information
vport
OPA vnic per port information

Description

On host, each of the virtual ethernet ports belongs to a different virtual ethernet switches.

struct opa_veswport_mactable_entry

single entry in the forwarding table

Definition

struct opa_veswport_mactable_entry {
  u8 mac_addr[ETH_ALEN];
  u8 mac_addr_mask[ETH_ALEN];
  __be32 dlid_sd;
};

Members

mac_addr
MAC address
mac_addr_mask
MAC address bit mask
dlid_sd
Matching DLID and side data

Description

On the host each virtual ethernet port will have a forwarding table. These tables are used to map a MAC to a LID and other data. For more details see struct opa_veswport_mactable_entries. This is the structure of a single mactable entry

struct opa_veswport_mactable

Forwarding table array

Definition

struct opa_veswport_mactable {
  __be16 offset;
  __be16 num_entries;
  __be32 mac_tbl_digest;
  struct opa_veswport_mactable_entry  tbl_entries[0];
};

Members

offset
mac table starting offset
num_entries
Number of entries to get or set
mac_tbl_digest
mac table digest **tbl_entries**[]: Array of table entries

Description

The EM sends down this structure in a MAD indicating the starting offset in the forwarding table that this entry is to be loaded into and the number of entries that that this MAD instance contains The mac_tbl_digest has been added to this MAD structure. It will be set by the EM and it will be used by the EM to check if there are any discrepancies with this value and the value maintained by the EM in the case of VNIC port being deleted or unloaded A new instantiation of a VNIC will always have a value of zero. This value is stored as part of the vnic adapter structure and will be accessed by the GET and SET routines for both the mactable entries and the veswport info.

struct opa_veswport_summary_counters

summary counters

Definition

struct opa_veswport_summary_counters {
  __be16 vp_instance;
  __be16 vesw_id;
  __be32 veswport_num;
  __be64 tx_errors;
  __be64 rx_errors;
  __be64 tx_packets;
  __be64 rx_packets;
  __be64 tx_bytes;
  __be64 rx_bytes;
  __be64 tx_unicast;
  __be64 tx_mcastbcast;
  __be64 tx_untagged;
  __be64 tx_vlan;
  __be64 tx_64_size;
  __be64 tx_65_127;
  __be64 tx_128_255;
  __be64 tx_256_511;
  __be64 tx_512_1023;
  __be64 tx_1024_1518;
  __be64 tx_1519_max;
  __be64 rx_unicast;
  __be64 rx_mcastbcast;
  __be64 rx_untagged;
  __be64 rx_vlan;
  __be64 rx_64_size;
  __be64 rx_65_127;
  __be64 rx_128_255;
  __be64 rx_256_511;
  __be64 rx_512_1023;
  __be64 rx_1024_1518;
  __be64 rx_1519_max;
  __be64 reserved[16];
};

Members

vp_instance
vport instance on the OPA port
vesw_id
virtual ethernet switch id
veswport_num
virtual ethernet switch port number
tx_errors
transmit errors
rx_errors
receive errors
tx_packets
transmit packets
rx_packets
receive packets
tx_bytes
transmit bytes
rx_bytes
receive bytes
tx_unicast
unicast packets transmitted
tx_mcastbcast
multicast/broadcast packets transmitted
tx_untagged
non-vlan packets transmitted
tx_vlan
vlan packets transmitted
tx_64_size
transmit packet length is 64 bytes
tx_65_127
transmit packet length is >=65 and < 127 bytes
tx_128_255
transmit packet length is >=128 and < 255 bytes
tx_256_511
transmit packet length is >=256 and < 511 bytes
tx_512_1023
transmit packet length is >=512 and < 1023 bytes
tx_1024_1518
transmit packet length is >=1024 and < 1518 bytes
tx_1519_max
transmit packet length >= 1519 bytes
rx_unicast
unicast packets received
rx_mcastbcast
multicast/broadcast packets received
rx_untagged
non-vlan packets received
rx_vlan
vlan packets received
rx_64_size
received packet length is 64 bytes
rx_65_127
received packet length is >=65 and < 127 bytes
rx_128_255
received packet length is >=128 and < 255 bytes
rx_256_511
received packet length is >=256 and < 511 bytes
rx_512_1023
received packet length is >=512 and < 1023 bytes
rx_1024_1518
received packet length is >=1024 and < 1518 bytes
rx_1519_max
received packet length >= 1519 bytes

Description

All the above are counters of corresponding conditions.

struct opa_veswport_error_counters

error counters

Definition

struct opa_veswport_error_counters {
  __be16 vp_instance;
  __be16 vesw_id;
  __be32 veswport_num;
  __be64 tx_errors;
  __be64 rx_errors;
  __be64 rsvd0;
  __be64 tx_smac_filt;
  __be64 rsvd1;
  __be64 rsvd2;
  __be64 rsvd3;
  __be64 tx_dlid_zero;
  __be64 rsvd4;
  __be64 tx_logic;
  __be64 rsvd5;
  __be64 tx_drop_state;
  __be64 rx_bad_veswid;
  __be64 rsvd6;
  __be64 rx_runt;
  __be64 rx_oversize;
  __be64 rsvd7;
  __be64 rx_eth_down;
  __be64 rx_drop_state;
  __be64 rx_logic;
  __be64 rsvd8;
  __be64 rsvd9[16];
};

Members

vp_instance
vport instance on the OPA port
vesw_id
virtual ethernet switch id
veswport_num
virtual ethernet switch port number
tx_errors
transmit errors
rx_errors
receive errors
tx_smac_filt
smac filter errors
tx_dlid_zero
transmit packets with invalid dlid
tx_logic
other transmit errors
tx_drop_state
packet tansmission in non-forward port state
rx_bad_veswid
received packet with invalid vesw id
rx_runt
received ethernet packet with length < 64 bytes
rx_oversize
received ethernet packet with length > MTU size
rx_eth_down
received packets when interface is down
rx_drop_state
received packets in non-forwarding port state
rx_logic
other receive errors

Description

All the above are counters of corresponding erorr conditions.

struct opa_veswport_trap

Trap message sent to EM by VNIC

Definition

struct opa_veswport_trap {
  __be16 fabric_id;
  __be16 veswid;
  __be32 veswportnum;
  __be16 opaportnum;
  u8 veswportindex;
  u8 opcode;
  __be32 reserved;
};

Members

fabric_id
10 bit fabric id
veswid
12 bit virtual ethernet switch id
veswportnum
logical port number on the Virtual switch
opaportnum
physical port num (redundant on host)
veswportindex
switch port index on opa port 0 based
opcode
operation
reserved
32 bit for alignment

Description

The VNIC will send trap messages to the Ethernet manager to inform it about changes to the VNIC config, behaviour etc. This is the format of the trap payload.

struct opa_vnic_iface_mac_entry

single entry in the mac list

Definition

struct opa_vnic_iface_mac_entry {
  u8 mac_addr[ETH_ALEN];
};

Members

mac_addr
MAC address
struct opa_veswport_iface_macs

Msg to set globally administered MAC

Definition

struct opa_veswport_iface_macs {
  __be16 start_idx;
  __be16 num_macs_in_msg;
  __be16 tot_macs_in_lst;
  __be16 gen_count;
  struct opa_vnic_iface_mac_entry entry[0];
};

Members

start_idx
position of first entry (0 based)
num_macs_in_msg
number of MACs in this message
tot_macs_in_lst
The total number of MACs the agent has
gen_count
gen_count to indicate change
entry
The mac list entry

Description

Same attribute IDS and attribute modifiers as in locally administered addresses used to set globally administered addresses

struct opa_vnic_vema_mad

Generic VEMA MAD

Definition

struct opa_vnic_vema_mad {
  struct ib_mad_hdr  mad_hdr;
  struct ib_rmpp_hdr rmpp_hdr;
  u8 reserved;
  u8 oui[3];
  u8 data[OPA_VNIC_EMA_DATA];
};

Members

mad_hdr
Generic MAD header
rmpp_hdr
RMPP header for vendor specific MADs
oui
Unique org identifier
data
MAD data
struct opa_vnic_notice_attr

Generic Notice MAD

Definition

struct opa_vnic_notice_attr {
  u8 gen_type;
  u8 oui_1;
  u8 oui_2;
  u8 oui_3;
  __be16 trap_num;
  __be16 toggle_count;
  __be32 issuer_lid;
  __be32 reserved;
  u8 issuer_gid[16];
  u8 raw_data[64];
};

Members

gen_type
Generic/Specific bit and type of notice
oui_1
Vendor ID byte 1
oui_2
Vendor ID byte 2
oui_3
Vendor ID byte 3
trap_num
Trap number
toggle_count
Notice toggle bit and count value
issuer_lid
Trap issuer's lid
issuer_gid
Issuer GID (only if Report method)
raw_data
Trap message body
struct opa_vnic_vema_mad_trap

Generic VEMA MAD Trap

Definition

struct opa_vnic_vema_mad_trap {
  struct ib_mad_hdr            mad_hdr;
  struct ib_rmpp_hdr           rmpp_hdr;
  u8 reserved;
  u8 oui[3];
  struct opa_vnic_notice_attr  notice;
};

Members

mad_hdr
Generic MAD header
rmpp_hdr
RMPP header for vendor specific MADs
oui
Unique org identifier
notice
Notice structure
void opa_vnic_vema_report_event(struct opa_vnic_adapter * adapter, u8 event)

sent trap to report the specified event

Parameters

struct opa_vnic_adapter * adapter
vnic port adapter
u8 event
event to be reported

Description

This function calls vema api to sent a trap for the given event.

void opa_vnic_get_summary_counters(struct opa_vnic_adapter * adapter, struct opa_veswport_summary_counters * cntrs)

get summary counters

Parameters

struct opa_vnic_adapter * adapter
vnic port adapter
struct opa_veswport_summary_counters * cntrs
pointer to destination summary counters structure

Description

This function populates the summary counters that is maintained by the given adapter to destination address provided.

void opa_vnic_get_error_counters(struct opa_vnic_adapter * adapter, struct opa_veswport_error_counters * cntrs)

get error counters

Parameters

struct opa_vnic_adapter * adapter
vnic port adapter
struct opa_veswport_error_counters * cntrs
pointer to destination error counters structure

Description

This function populates the error counters that is maintained by the given adapter to destination address provided.

void opa_vnic_get_vesw_info(struct opa_vnic_adapter * adapter, struct opa_vesw_info * info)
  • Get the vesw information

Parameters

struct opa_vnic_adapter * adapter
vnic port adapter
struct opa_vesw_info * info
pointer to destination vesw info structure

Description

This function copies the vesw info that is maintained by the given adapter to destination address provided.

void opa_vnic_set_vesw_info(struct opa_vnic_adapter * adapter, struct opa_vesw_info * info)
  • Set the vesw information

Parameters

struct opa_vnic_adapter * adapter
vnic port adapter
struct opa_vesw_info * info
pointer to vesw info structure

Description

This function updates the vesw info that is maintained by the given adapter with vesw info provided. Reserved fields are stored and returned back to EM as is.

void opa_vnic_get_per_veswport_info(struct opa_vnic_adapter * adapter, struct opa_per_veswport_info * info)
  • Get the vesw per port information

Parameters

struct opa_vnic_adapter * adapter
vnic port adapter
struct opa_per_veswport_info * info
pointer to destination vport info structure

Description

This function copies the vesw per port info that is maintained by the given adapter to destination address provided. Note that the read only fields are not copied.

void opa_vnic_set_per_veswport_info(struct opa_vnic_adapter * adapter, struct opa_per_veswport_info * info)
  • Set vesw per port information

Parameters

struct opa_vnic_adapter * adapter
vnic port adapter
struct opa_per_veswport_info * info
pointer to vport info structure

Description

This function updates the vesw per port info that is maintained by the given adapter with vesw per port info provided. Reserved fields are stored and returned back to EM as is.

void opa_vnic_query_mcast_macs(struct opa_vnic_adapter * adapter, struct opa_veswport_iface_macs * macs)

query multicast mac list

Parameters

struct opa_vnic_adapter * adapter
vnic port adapter
struct opa_veswport_iface_macs * macs
pointer mac list

Description

This function populates the provided mac list with the configured multicast addresses in the adapter.

void opa_vnic_query_ucast_macs(struct opa_vnic_adapter * adapter, struct opa_veswport_iface_macs * macs)

query unicast mac list

Parameters

struct opa_vnic_adapter * adapter
vnic port adapter
struct opa_veswport_iface_macs * macs
pointer mac list

Description

This function populates the provided mac list with the configured unicast addresses in the adapter.

struct opa_vnic_vema_port
  • VNIC VEMA port details

Definition

struct opa_vnic_vema_port {
  struct opa_vnic_ctrl_port      *cport;
  struct ib_mad_agent            *mad_agent;
  struct opa_class_port_info      class_port_info;
  u64 tid;
  u8 port_num;
  struct xarray                   vports;
  struct ib_event_handler         event_handler;
  struct mutex                    lock;
};

Members

cport
pointer to port
mad_agent
pointer to mad agent for port
class_port_info
Class port info information.
tid
Transaction id
port_num
OPA port number
vports
vnic ports
event_handler
ib event handler
lock
adapter interface lock
u8 vema_get_vport_num(struct opa_vnic_vema_mad * recvd_mad)
  • Get the vnic from the mad

Parameters

struct opa_vnic_vema_mad * recvd_mad
Received mad

Return

returns value of the vnic port number

struct opa_vnic_adapter * vema_get_vport_adapter(struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_port * port)
  • Get vnic port adapter from recvd mad

Parameters

struct opa_vnic_vema_mad * recvd_mad
received mad
struct opa_vnic_vema_port * port
ptr to port struct on which MAD was recvd

Return

vnic adapter

bool vema_mac_tbl_req_ok(struct opa_veswport_mactable * mac_tbl)
  • Check if mac request has correct values

Parameters

struct opa_veswport_mactable * mac_tbl
mac table

Description

This function checks for the validity of the offset and number of entries required.

Return

true if offset and num_entries are valid

struct opa_vnic_adapter * vema_add_vport(struct opa_vnic_vema_port * port, u8 vport_num)
  • Add a new vnic port

Parameters

struct opa_vnic_vema_port * port
ptr to opa_vnic_vema_port struct
u8 vport_num
vnic port number (to be added)

Description

Return a pointer to the vnic adapter structure

void vema_get_class_port_info(struct opa_vnic_vema_port * port, struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_mad * rsp_mad)
  • Get class info for port

Parameters

struct opa_vnic_vema_port * port
Port on whic MAD was received
struct opa_vnic_vema_mad * recvd_mad
pointer to the received mad
struct opa_vnic_vema_mad * rsp_mad
pointer to respose mad

Description

This function copies the latest class port info value set for the port and stores it for generating traps

void vema_set_class_port_info(struct opa_vnic_vema_port * port, struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_mad * rsp_mad)
  • Get class info for port

Parameters

struct opa_vnic_vema_port * port
Port on whic MAD was received
struct opa_vnic_vema_mad * recvd_mad
pointer to the received mad
struct opa_vnic_vema_mad * rsp_mad
pointer to respose mad

Description

This function updates the port class info for the specific vnic and sets up the response mad data

void vema_get_veswport_info(struct opa_vnic_vema_port * port, struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_mad * rsp_mad)
  • Get veswport info

Parameters

struct opa_vnic_vema_port * port
source port on which MAD was received
struct opa_vnic_vema_mad * recvd_mad
pointer to the received mad
struct opa_vnic_vema_mad * rsp_mad
pointer to respose mad
void vema_set_veswport_info(struct opa_vnic_vema_port * port, struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_mad * rsp_mad)
  • Set veswport info

Parameters

struct opa_vnic_vema_port * port
source port on which MAD was received
struct opa_vnic_vema_mad * recvd_mad
pointer to the received mad
struct opa_vnic_vema_mad * rsp_mad
pointer to respose mad

Description

This function gets the port class infor for vnic

void vema_get_mac_entries(struct opa_vnic_vema_port * port, struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_mad * rsp_mad)
  • Get MAC entries in VNIC MAC table

Parameters

struct opa_vnic_vema_port * port
source port on which MAD was received
struct opa_vnic_vema_mad * recvd_mad
pointer to the received mad
struct opa_vnic_vema_mad * rsp_mad
pointer to respose mad

Description

This function gets the MAC entries that are programmed into the VNIC MAC forwarding table. It checks for the validity of the index into the MAC table and the number of entries that are to be retrieved.

void vema_set_mac_entries(struct opa_vnic_vema_port * port, struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_mad * rsp_mad)
  • Set MAC entries in VNIC MAC table

Parameters

struct opa_vnic_vema_port * port
source port on which MAD was received
struct opa_vnic_vema_mad * recvd_mad
pointer to the received mad
struct opa_vnic_vema_mad * rsp_mad
pointer to respose mad

Description

This function sets the MAC entries in the VNIC forwarding table It checks for the validity of the index and the number of forwarding table entries to be programmed.

void vema_set_delete_vesw(struct opa_vnic_vema_port * port, struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_mad * rsp_mad)
  • Reset VESW info to POD values

Parameters

struct opa_vnic_vema_port * port
source port on which MAD was received
struct opa_vnic_vema_mad * recvd_mad
pointer to the received mad
struct opa_vnic_vema_mad * rsp_mad
pointer to respose mad

Description

This function clears all the fields of veswport info for the requested vesw and sets them back to the power-on default values. It does not delete the vesw.

void vema_get_mac_list(struct opa_vnic_vema_port * port, struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_mad * rsp_mad, u16 attr_id)
  • Get the unicast/multicast macs.

Parameters

struct opa_vnic_vema_port * port
source port on which MAD was received
struct opa_vnic_vema_mad * recvd_mad
Received mad contains fields to set vnic parameters
struct opa_vnic_vema_mad * rsp_mad
Response mad to be built
u16 attr_id
Attribute ID indicating multicast or unicast mac list
void vema_get_summary_counters(struct opa_vnic_vema_port * port, struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_mad * rsp_mad)
  • Gets summary counters.

Parameters

struct opa_vnic_vema_port * port
source port on which MAD was received
struct opa_vnic_vema_mad * recvd_mad
Received mad contains fields to set vnic parameters
struct opa_vnic_vema_mad * rsp_mad
Response mad to be built
void vema_get_error_counters(struct opa_vnic_vema_port * port, struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_mad * rsp_mad)
  • Gets summary counters.

Parameters

struct opa_vnic_vema_port * port
source port on which MAD was received
struct opa_vnic_vema_mad * recvd_mad
Received mad contains fields to set vnic parameters
struct opa_vnic_vema_mad * rsp_mad
Response mad to be built
void vema_get(struct opa_vnic_vema_port * port, struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_mad * rsp_mad)
  • Process received get MAD

Parameters

struct opa_vnic_vema_port * port
source port on which MAD was received
struct opa_vnic_vema_mad * recvd_mad
Received mad
struct opa_vnic_vema_mad * rsp_mad
Response mad to be built
void vema_set(struct opa_vnic_vema_port * port, struct opa_vnic_vema_mad * recvd_mad, struct opa_vnic_vema_mad * rsp_mad)
  • Process received set MAD

Parameters

struct opa_vnic_vema_port * port
source port on which MAD was received
struct opa_vnic_vema_mad * recvd_mad
Received mad contains fields to set vnic parameters
struct opa_vnic_vema_mad * rsp_mad
Response mad to be built
void vema_send(struct ib_mad_agent * mad_agent, struct ib_mad_send_wc * mad_wc)
  • Send handler for VEMA MAD agent

Parameters

struct ib_mad_agent * mad_agent
pointer to the mad agent
struct ib_mad_send_wc * mad_wc
pointer to mad send work completion information

Description

Free all the data structures associated with the sent MAD

void vema_recv(struct ib_mad_agent * mad_agent, struct ib_mad_send_buf * send_buf, struct ib_mad_recv_wc * mad_wc)
  • Recv handler for VEMA MAD agent

Parameters

struct ib_mad_agent * mad_agent
pointer to the mad agent
struct ib_mad_send_buf * send_buf
Send buffer if found, else NULL
struct ib_mad_recv_wc * mad_wc
pointer to mad send work completion information

Description

Handle only set and get methods and respond to other methods as unsupported. Allocate response buffer and address handle for the response MAD.

struct opa_vnic_vema_port * vema_get_port(struct opa_vnic_ctrl_port * cport, u8 port_num)
  • Gets the opa_vnic_vema_port

Parameters

struct opa_vnic_ctrl_port * cport
pointer to control dev
u8 port_num
Port number

Description

This function loops through the ports and returns the opa_vnic_vema port structure that is associated with the OPA port number

Return

ptr to requested opa_vnic_vema_port strucure
if success, NULL if not
void opa_vnic_vema_send_trap(struct opa_vnic_adapter * adapter, struct __opa_veswport_trap * data, u32 lid)
  • This function sends a trap to the EM

Parameters

struct opa_vnic_adapter * adapter
pointer to vnic adapter
struct __opa_veswport_trap * data
pointer to trap data filled by calling function
u32 lid
issuers lid (encap_slid from vesw_port_info)

Description

This function is called from the VNIC driver to send a trap if there is somethng the EM should be notified about. These events currently are 1) UNICAST INTERFACE MACADDRESS changes 2) MULTICAST INTERFACE MACADDRESS changes 3) ETHERNET LINK STATUS changes While allocating the send mad the remote site qpn used is 1 as this is the well known QP.

void vema_unregister(struct opa_vnic_ctrl_port * cport)
  • Unregisters agent

Parameters

struct opa_vnic_ctrl_port * cport
pointer to control port

Description

This deletes the registration by VEMA for MADs

int vema_register(struct opa_vnic_ctrl_port * cport)
  • Registers agent

Parameters

struct opa_vnic_ctrl_port * cport
pointer to control port

Description

This function registers the handlers for the VEMA MADs

Return

returns 0 on success. non zero otherwise

void opa_vnic_ctrl_config_dev(struct opa_vnic_ctrl_port * cport, bool en)
  • This function sends a trap to the EM by way of ib_modify_port to indicate support for ethernet on the fabric.

Parameters

struct opa_vnic_ctrl_port * cport
pointer to control port
bool en
enable or disable ethernet on fabric support
void opa_vnic_vema_add_one(struct ib_device * device)
  • Handle new ib device

Parameters

struct ib_device * device
ib device pointer

Description

Allocate the vnic control port and initialize it.

void opa_vnic_vema_rem_one(struct ib_device * device, void * client_data)
  • Handle ib device removal

Parameters

struct ib_device * device
ib device pointer
void * client_data
ib client data

Description

Uninitialize and free the vnic control port.

InfiniBand SCSI RDMA protocol target support

enum srpt_command_state

SCSI command state managed by SRPT

Constants

SRPT_STATE_NEW
New command arrived and is being processed.
SRPT_STATE_NEED_DATA
Processing a write or bidir command and waiting for data arrival.
SRPT_STATE_DATA_IN
Data for the write or bidir command arrived and is being processed.
SRPT_STATE_CMD_RSP_SENT
SRP_RSP for SRP_CMD has been sent.
SRPT_STATE_MGMT
Processing a SCSI task management command.
SRPT_STATE_MGMT_RSP_SENT
SRP_RSP for SRP_TSK_MGMT has been sent.
SRPT_STATE_DONE
Command processing finished successfully, command processing has been aborted or command processing failed.
struct srpt_ioctx

shared SRPT I/O context information

Definition

struct srpt_ioctx {
  struct ib_cqe           cqe;
  void *buf;
  dma_addr_t dma;
  uint32_t offset;
  uint32_t index;
};

Members

cqe
Completion queue element.
buf
Pointer to the buffer.
dma
DMA address of the buffer.
offset
Offset of the first byte in buf and dma that is actually used.
index
Index of the I/O context in its ioctx_ring array.
struct srpt_recv_ioctx

SRPT receive I/O context

Definition

struct srpt_recv_ioctx {
  struct srpt_ioctx       ioctx;
  struct list_head        wait_list;
  int byte_len;
};

Members

ioctx
See above.
wait_list
Node for insertion in srpt_rdma_ch.cmd_wait_list.
byte_len
Number of bytes in ioctx.buf.
struct srpt_send_ioctx

SRPT send I/O context

Definition

struct srpt_send_ioctx {
  struct srpt_ioctx       ioctx;
  struct srpt_rdma_ch     *ch;
  struct srpt_recv_ioctx  *recv_ioctx;
  struct srpt_rw_ctx      s_rw_ctx;
  struct srpt_rw_ctx      *rw_ctxs;
  struct scatterlist      imm_sg;
  struct ib_cqe           rdma_cqe;
  enum srpt_command_state state;
  struct se_cmd           cmd;
  u8 n_rdma;
  u8 n_rw_ctx;
  bool queue_status_only;
  u8 sense_data[TRANSPORT_SENSE_BUFFER];
};

Members

ioctx
See above.
ch
Channel pointer.
recv_ioctx
Receive I/O context associated with this send I/O context. Only used for processing immediate data.
s_rw_ctx
rw_ctxs points here if only a single rw_ctx is needed.
rw_ctxs
RDMA read/write contexts.
imm_sg
Scatterlist for immediate data.
rdma_cqe
RDMA completion queue element.
state
I/O context state.
cmd
Target core command data structure.
n_rdma
Number of work requests needed to transfer this ioctx.
n_rw_ctx
Size of rw_ctxs array.
queue_status_only
Send a SCSI status back to the initiator but no data.
sense_data
Sense data to be sent to the initiator.
enum rdma_ch_state

SRP channel state

Constants

CH_CONNECTING
QP is in RTR state; waiting for RTU.
CH_LIVE
QP is in RTS state.
CH_DISCONNECTING
DREQ has been sent and waiting for DREP or DREQ has been received.
CH_DRAINING
DREP has been received or waiting for DREP timed out and last work request has been queued.
CH_DISCONNECTED
Last completion has been received.
struct srpt_rdma_ch

RDMA channel

Definition

struct srpt_rdma_ch {
  struct srpt_nexus       *nexus;
  struct ib_qp            *qp;
  union {
    struct {
      struct ib_cm_id         *cm_id;
    } ib_cm;
    struct {
      struct rdma_cm_id       *cm_id;
    } rdma_cm;
  };
  struct ib_cq            *cq;
  struct ib_cqe           zw_cqe;
  struct rcu_head         rcu;
  struct kref             kref;
  struct completion       *closed;
  int rq_size;
  u32 max_rsp_size;
  atomic_t sq_wr_avail;
  struct srpt_port        *sport;
  int max_ti_iu_len;
  atomic_t req_lim;
  atomic_t req_lim_delta;
  u16 imm_data_offset;
  spinlock_t spinlock;
  enum rdma_ch_state      state;
  struct kmem_cache       *rsp_buf_cache;
  struct srpt_send_ioctx  **ioctx_ring;
  struct kmem_cache       *req_buf_cache;
  struct srpt_recv_ioctx  **ioctx_recv_ring;
  struct list_head        list;
  struct list_head        cmd_wait_list;
  uint16_t pkey;
  bool using_rdma_cm;
  bool processing_wait_list;
  struct se_session       *sess;
  u8 sess_name[40];
  struct work_struct      release_work;
};

Members

nexus
I_T nexus this channel is associated with.
qp
IB queue pair used for communicating over this channel.
{unnamed_union}
anonymous
ib_cm
See below.
ib_cm.cm_id
IB CM ID associated with the channel.
rdma_cm
See below.
rdma_cm.cm_id
RDMA CM ID associated with the channel.
cq
IB completion queue for this channel.
zw_cqe
Zero-length write CQE.
rcu
RCU head.
kref
kref for this channel.
closed
Completion object that will be signaled as soon as a new channel object with the same identity can be created.
rq_size
IB receive queue size.
max_rsp_size
Maximum size of an RSP response message in bytes.
sq_wr_avail
number of work requests available in the send queue.
sport
pointer to the information of the HCA port used by this channel.
max_ti_iu_len
maximum target-to-initiator information unit length.
req_lim
request limit: maximum number of requests that may be sent by the initiator without having received a response.
req_lim_delta
Number of credits not yet sent back to the initiator.
imm_data_offset
Offset from start of SRP_CMD for immediate data.
spinlock
Protects free_list and state.
state
channel state. See also enum rdma_ch_state.
rsp_buf_cache
kmem_cache for ioctx_ring.
ioctx_ring
Send ring.
req_buf_cache
kmem_cache for ioctx_recv_ring.
ioctx_recv_ring
Receive I/O context ring.
list
Node in srpt_nexus.ch_list.
cmd_wait_list
List of SCSI commands that arrived before the RTU event. This list contains struct srpt_ioctx elements and is protected against concurrent modification by the cm_id spinlock.
pkey
P_Key of the IB partition for this SRP channel.
using_rdma_cm
Whether the RDMA/CM or IB/CM is used for this channel.
processing_wait_list
Whether or not cmd_wait_list is being processed.
sess
Session information associated with this SRP channel.
sess_name
Session name.
release_work
Allows scheduling of srpt_release_channel().
struct srpt_nexus

I_T nexus

Definition

struct srpt_nexus {
  struct rcu_head         rcu;
  struct list_head        entry;
  struct list_head        ch_list;
  u8 i_port_id[16];
  u8 t_port_id[16];
};

Members

rcu
RCU head for this data structure.
entry
srpt_port.nexus_list list node.
ch_list
struct srpt_rdma_ch list. Protected by srpt_port.mutex.
i_port_id
128-bit initiator port identifier copied from SRP_LOGIN_REQ.
t_port_id
128-bit target port identifier copied from SRP_LOGIN_REQ.
struct srpt_port_attrib

attributes for SRPT port

Definition

struct srpt_port_attrib {
  u32 srp_max_rdma_size;
  u32 srp_max_rsp_size;
  u32 srp_sq_size;
  bool use_srq;
};

Members

srp_max_rdma_size
Maximum size of SRP RDMA transfers for new connections.
srp_max_rsp_size
Maximum size of SRP response messages in bytes.
srp_sq_size
Shared receive queue (SRQ) size.
use_srq
Whether or not to use SRQ.
struct srpt_tpg

information about a single "target portal group"

Definition

struct srpt_tpg {
  struct list_head        entry;
  struct srpt_port_id     *sport_id;
  struct se_portal_group  tpg;
};

Members

entry
Entry in sport_id->tpg_list.
sport_id
Port name this TPG is associated with.
tpg
LIO TPG data structure.

Description

Zero or more target portal groups are associated with each port name (srpt_port_id). With each TPG an ACL list is associated.

struct srpt_port_id

information about an RDMA port name

Definition

struct srpt_port_id {
  struct mutex            mutex;
  struct list_head        tpg_list;
  struct se_wwn           wwn;
  char name[64];
};

Members

mutex
Protects tpg_list changes.
tpg_list
TPGs associated with the RDMA port name.
wwn
WWN associated with the RDMA port name.
name
ASCII representation of the port name.

Description

Multiple sysfs directories can be associated with a single RDMA port. This data structure represents a single (port, name) pair.

struct srpt_port

information associated by SRPT with a single IB port

Definition

struct srpt_port {
  struct srpt_device      *sdev;
  struct ib_mad_agent     *mad_agent;
  bool enabled;
  u8 port;
  u32 sm_lid;
  u32 lid;
  union ib_gid            gid;
  struct work_struct      work;
  struct srpt_port_id     port_guid_id;
  struct srpt_port_id     port_gid_id;
  struct srpt_port_attrib port_attrib;
  atomic_t refcount;
  struct completion       *freed_channels;
  struct mutex            mutex;
  struct list_head        nexus_list;
};

Members

sdev
backpointer to the HCA information.
mad_agent
per-port management datagram processing information.
enabled
Whether or not this target port is enabled.
port
one-based port number.
sm_lid
cached value of the port's sm_lid.
lid
cached value of the port's lid.
gid
cached value of the port's gid.
work
work structure for refreshing the aforementioned cached values.
port_guid_id
target port GUID
port_gid_id
target port GID
port_attrib
Port attributes that can be accessed through configfs.
refcount
Number of objects associated with this port.
freed_channels
Completion that will be signaled once refcount becomes 0.
mutex
Protects nexus_list.
nexus_list
Nexus list. See also srpt_nexus.entry.
struct srpt_device

information associated by SRPT with a single HCA

Definition

struct srpt_device {
  struct ib_device        *device;
  struct ib_pd            *pd;
  u32 lkey;
  struct ib_srq           *srq;
  struct ib_cm_id         *cm_id;
  int srq_size;
  struct mutex            sdev_mutex;
  bool use_srq;
  struct kmem_cache       *req_buf_cache;
  struct srpt_recv_ioctx  **ioctx_ring;
  struct ib_event_handler event_handler;
  struct list_head        list;
  struct srpt_port        port[];
};

Members

device
Backpointer to the struct ib_device managed by the IB core.
pd
IB protection domain.
lkey
L_Key (local key) with write access to all local memory.
srq
Per-HCA SRQ (shared receive queue).
cm_id
Connection identifier.
srq_size
SRQ size.
sdev_mutex
Serializes use_srq changes.
use_srq
Whether or not to use SRQ.
req_buf_cache
kmem_cache for ioctx_ring buffers.
ioctx_ring
Per-HCA SRQ.
event_handler
Per-HCA asynchronous IB event handler.
list
Node in srpt_dev_list.
port
Information about the ports owned by this HCA.
void srpt_event_handler(struct ib_event_handler * handler, struct ib_event * event)

asynchronous IB event callback function

Parameters

struct ib_event_handler * handler
IB event handler registered by ib_register_event_handler().
struct ib_event * event
Description of the event that occurred.

Description

Callback function called by the InfiniBand core when an asynchronous IB event occurs. This callback may occur in interrupt context. See also section 11.5.2, Set Asynchronous Event Handler in the InfiniBand Architecture Specification.

void srpt_srq_event(struct ib_event * event, void * ctx)

SRQ event callback function

Parameters

struct ib_event * event
Description of the event that occurred.
void * ctx
Context pointer specified at SRQ creation time.
void srpt_qp_event(struct ib_event * event, struct srpt_rdma_ch * ch)

QP event callback function

Parameters

struct ib_event * event
Description of the event that occurred.
struct srpt_rdma_ch * ch
SRPT RDMA channel.
void srpt_set_ioc(u8 * c_list, u32 slot, u8 value)

initialize a IOUnitInfo structure

Parameters

u8 * c_list
controller list.
u32 slot
one-based slot number.
u8 value
four-bit value.

Description

Copies the lowest four bits of value in element slot of the array of four bit elements called c_list (controller list). The index slot is one-based.

void srpt_get_class_port_info(struct ib_dm_mad * mad)

copy ClassPortInfo to a management datagram

Parameters

struct ib_dm_mad * mad
Datagram that will be sent as response to DM_ATTR_CLASS_PORT_INFO.

Description

See also section 16.3.3.1 ClassPortInfo in the InfiniBand Architecture Specification.

void srpt_get_iou(struct ib_dm_mad * mad)

write IOUnitInfo to a management datagram

Parameters

struct ib_dm_mad * mad
Datagram that will be sent as response to DM_ATTR_IOU_INFO.

Description

See also section 16.3.3.3 IOUnitInfo in the InfiniBand Architecture Specification. See also section B.7, table B.6 in the SRP r16a document.

void srpt_get_ioc(struct srpt_port * sport, u32 slot, struct ib_dm_mad * mad)

write IOControllerprofile to a management datagram

Parameters

struct srpt_port * sport
HCA port through which the MAD has been received.
u32 slot
Slot number specified in DM_ATTR_IOC_PROFILE query.
struct ib_dm_mad * mad
Datagram that will be sent as response to DM_ATTR_IOC_PROFILE.

Description

See also section 16.3.3.4 IOControllerProfile in the InfiniBand Architecture Specification. See also section B.7, table B.7 in the SRP r16a document.

void srpt_get_svc_entries(u64 ioc_guid, u16 slot, u8 hi, u8 lo, struct ib_dm_mad * mad)

write ServiceEntries to a management datagram

Parameters

u64 ioc_guid
I/O controller GUID to use in reply.
u16 slot
I/O controller number.
u8 hi
End of the range of service entries to be specified in the reply.
u8 lo
Start of the range of service entries to be specified in the reply..
struct ib_dm_mad * mad
Datagram that will be sent as response to DM_ATTR_SVC_ENTRIES.

Description

See also section 16.3.3.5 ServiceEntries in the InfiniBand Architecture Specification. See also section B.7, table B.8 in the SRP r16a document.

void srpt_mgmt_method_get(struct srpt_port * sp, struct ib_mad * rq_mad, struct ib_dm_mad * rsp_mad)

process a received management datagram

Parameters

struct srpt_port * sp
HCA port through which the MAD has been received.
struct ib_mad * rq_mad
received MAD.
struct ib_dm_mad * rsp_mad
response MAD.
void srpt_mad_send_handler(struct ib_mad_agent * mad_agent, struct ib_mad_send_wc * mad_wc)

MAD send completion callback

Parameters

struct ib_mad_agent * mad_agent
Return value of ib_register_mad_agent().
struct ib_mad_send_wc * mad_wc
Work completion reporting that the MAD has been sent.
void srpt_mad_recv_handler(struct ib_mad_agent * mad_agent, struct ib_mad_send_buf * send_buf, struct ib_mad_recv_wc * mad_wc)

MAD reception callback function

Parameters

struct ib_mad_agent * mad_agent
Return value of ib_register_mad_agent().
struct ib_mad_send_buf * send_buf
Not used.
struct ib_mad_recv_wc * mad_wc
Work completion reporting that a MAD has been received.
int srpt_refresh_port(struct srpt_port * sport)

configure a HCA port

Parameters

struct srpt_port * sport
SRPT HCA port.

Description

Enable InfiniBand management datagram processing, update the cached sm_lid, lid and gid values, and register a callback function for processing MADs on the specified port.

Note

It is safe to call this function more than once for the same port.

void srpt_unregister_mad_agent(struct srpt_device * sdev)

unregister MAD callback functions

Parameters

struct srpt_device * sdev
SRPT HCA pointer.

Note

It is safe to call this function more than once for the same device.

struct srpt_ioctx * srpt_alloc_ioctx(struct srpt_device * sdev, int ioctx_size, struct kmem_cache * buf_cache, enum dma_data_direction dir)

allocate a SRPT I/O context structure

Parameters

struct srpt_device * sdev
SRPT HCA pointer.
int ioctx_size
I/O context size.
struct kmem_cache * buf_cache
I/O buffer cache.
enum dma_data_direction dir
DMA data direction.
void srpt_free_ioctx(struct srpt_device * sdev, struct srpt_ioctx * ioctx, struct kmem_cache * buf_cache, enum dma_data_direction dir)

free a SRPT I/O context structure

Parameters

struct srpt_device * sdev
SRPT HCA pointer.
struct srpt_ioctx * ioctx
I/O context pointer.
struct kmem_cache * buf_cache
I/O buffer cache.
enum dma_data_direction dir
DMA data direction.
struct srpt_ioctx ** srpt_alloc_ioctx_ring(struct srpt_device * sdev, int ring_size, int ioctx_size, struct kmem_cache * buf_cache, int alignment_offset, enum dma_data_direction dir)

allocate a ring of SRPT I/O context structures

Parameters

struct srpt_device * sdev
Device to allocate the I/O context ring for.
int ring_size
Number of elements in the I/O context ring.
int ioctx_size
I/O context size.
struct kmem_cache * buf_cache
I/O buffer cache.
int alignment_offset
Offset in each ring buffer at which the SRP information unit starts.
enum dma_data_direction dir
DMA data direction.
void srpt_free_ioctx_ring(struct srpt_ioctx ** ioctx_ring, struct srpt_device * sdev, int ring_size, struct kmem_cache * buf_cache, enum dma_data_direction dir)

free the ring of SRPT I/O context structures

Parameters

struct srpt_ioctx ** ioctx_ring
I/O context ring to be freed.
struct srpt_device * sdev
SRPT HCA pointer.
int ring_size
Number of ring elements.
struct kmem_cache * buf_cache
I/O buffer cache.
enum dma_data_direction dir
DMA data direction.
enum srpt_command_state srpt_set_cmd_state(struct srpt_send_ioctx * ioctx, enum srpt_command_state new)

set the state of a SCSI command

Parameters

struct srpt_send_ioctx * ioctx
Send I/O context.
enum srpt_command_state new
New I/O context state.

Description

Does not modify the state of aborted commands. Returns the previous command state.

bool srpt_test_and_set_cmd_state(struct srpt_send_ioctx * ioctx, enum srpt_command_state old, enum srpt_command_state new)

test and set the state of a command

Parameters

struct srpt_send_ioctx * ioctx
Send I/O context.
enum srpt_command_state old
Current I/O context state.
enum srpt_command_state new
New I/O context state.

Description

Returns true if and only if the previous command state was equal to 'old'.

int srpt_post_recv(struct srpt_device * sdev, struct srpt_rdma_ch * ch, struct srpt_recv_ioctx * ioctx)

post an IB receive request

Parameters

struct srpt_device * sdev
SRPT HCA pointer.
struct srpt_rdma_ch * ch
SRPT RDMA channel.
struct srpt_recv_ioctx * ioctx
Receive I/O context pointer.
int srpt_zerolength_write(struct srpt_rdma_ch * ch)

perform a zero-length RDMA write

Parameters

struct srpt_rdma_ch * ch
SRPT RDMA channel.

Description

A quote from the InfiniBand specification: C9-88: For an HCA responder using Reliable Connection service, for each zero-length RDMA READ or WRITE request, the R_Key shall not be validated, even if the request includes Immediate data.

int srpt_get_desc_tbl(struct srpt_recv_ioctx * recv_ioctx, struct srpt_send_ioctx * ioctx, struct srp_cmd * srp_cmd, enum dma_data_direction * dir, struct scatterlist ** sg, unsigned int * sg_cnt, u64 * data_len, u16 imm_data_offset)

parse the data descriptors of a SRP_CMD request

Parameters

struct srpt_recv_ioctx * recv_ioctx
I/O context associated with the received command srp_cmd.
struct srpt_send_ioctx * ioctx
I/O context that will be used for responding to the initiator.
struct srp_cmd * srp_cmd
Pointer to the SRP_CMD request data.
enum dma_data_direction * dir
Pointer to the variable to which the transfer direction will be written.
struct scatterlist ** sg
[out] scatterlist for the parsed SRP_CMD.
unsigned int * sg_cnt
[out] length of sg.
u64 * data_len
Pointer to the variable to which the total data length of all descriptors in the SRP_CMD request will be written.
u16 imm_data_offset
[in] Offset in SRP_CMD requests at which immediate data starts.

Description

This function initializes ioctx->nrbuf and ioctx->r_bufs.

Returns -EINVAL when the SRP_CMD request contains inconsistent descriptors; -ENOMEM when memory allocation fails and zero upon success.

int srpt_init_ch_qp(struct srpt_rdma_ch * ch, struct ib_qp * qp)

initialize queue pair attributes

Parameters

struct srpt_rdma_ch * ch
SRPT RDMA channel.
struct ib_qp * qp
Queue pair pointer.

Description

Initialized the attributes of queue pair 'qp' by allowing local write, remote read and remote write. Also transitions 'qp' to state IB_QPS_INIT.

int srpt_ch_qp_rtr(struct srpt_rdma_ch * ch, struct ib_qp * qp)

change the state of a channel to 'ready to receive' (RTR)

Parameters

struct srpt_rdma_ch * ch
channel of the queue pair.
struct ib_qp * qp
queue pair to change the state of.

Description

Returns zero upon success and a negative value upon failure.

Note

currently a struct ib_qp_attr takes 136 bytes on a 64-bit system. If this structure ever becomes larger, it might be necessary to allocate it dynamically instead of on the stack.

int srpt_ch_qp_rts(struct srpt_rdma_ch * ch, struct ib_qp * qp)

change the state of a channel to 'ready to send' (RTS)

Parameters

struct srpt_rdma_ch * ch
channel of the queue pair.
struct ib_qp * qp
queue pair to change the state of.

Description

Returns zero upon success and a negative value upon failure.

Note

currently a struct ib_qp_attr takes 136 bytes on a 64-bit system. If this structure ever becomes larger, it might be necessary to allocate it dynamically instead of on the stack.

int srpt_ch_qp_err(struct srpt_rdma_ch * ch)

set the channel queue pair state to 'error'

Parameters

struct srpt_rdma_ch * ch
SRPT RDMA channel.
struct srpt_send_ioctx * srpt_get_send_ioctx(struct srpt_rdma_ch * ch)

obtain an I/O context for sending to the initiator

Parameters

struct srpt_rdma_ch * ch
SRPT RDMA channel.
int srpt_abort_cmd(struct srpt_send_ioctx * ioctx)

abort a SCSI command

Parameters

struct srpt_send_ioctx * ioctx
I/O context associated with the SCSI command.
void srpt_rdma_read_done(struct ib_cq * cq, struct ib_wc * wc)

RDMA read completion callback

Parameters

struct ib_cq * cq
Completion queue.
struct ib_wc * wc
Work completion.

Description

XXX: what is now target_execute_cmd used to be asynchronous, and unmapping the data that has been transferred via IB RDMA had to be postponed until the check_stop_free() callback. None of this is necessary anymore and needs to be cleaned up.

int srpt_build_cmd_rsp(struct srpt_rdma_ch * ch, struct srpt_send_ioctx * ioctx, u64 tag, int status)

build a SRP_RSP response

Parameters

struct srpt_rdma_ch * ch
RDMA channel through which the request has been received.
struct srpt_send_ioctx * ioctx
I/O context associated with the SRP_CMD request. The response will be built in the buffer ioctx->buf points at and hence this function will overwrite the request data.
u64 tag
tag of the request for which this response is being generated.
int status
value for the STATUS field of the SRP_RSP information unit.

Description

Returns the size in bytes of the SRP_RSP response.

An SRP_RSP response contains a SCSI status or service response. See also section 6.9 in the SRP r16a document for the format of an SRP_RSP response. See also SPC-2 for more information about sense data.

int srpt_build_tskmgmt_rsp(struct srpt_rdma_ch * ch, struct srpt_send_ioctx * ioctx, u8 rsp_code, u64 tag)

build a task management response

Parameters

struct srpt_rdma_ch * ch
RDMA channel through which the request has been received.
struct srpt_send_ioctx * ioctx
I/O context in which the SRP_RSP response will be built.
u8 rsp_code
RSP_CODE that will be stored in the response.
u64 tag
Tag of the request for which this response is being generated.

Description

Returns the size in bytes of the SRP_RSP response.

An SRP_RSP response contains a SCSI status or service response. See also section 6.9 in the SRP r16a document for the format of an SRP_RSP response.

void srpt_handle_cmd(struct srpt_rdma_ch * ch, struct srpt_recv_ioctx * recv_ioctx, struct srpt_send_ioctx * send_ioctx)

process a SRP_CMD information unit

Parameters

struct srpt_rdma_ch * ch
SRPT RDMA channel.
struct srpt_recv_ioctx * recv_ioctx
Receive I/O context.
struct srpt_send_ioctx * send_ioctx
Send I/O context.
void srpt_handle_tsk_mgmt(struct srpt_rdma_ch * ch, struct srpt_recv_ioctx * recv_ioctx, struct srpt_send_ioctx * send_ioctx)

process a SRP_TSK_MGMT information unit

Parameters

struct srpt_rdma_ch * ch
SRPT RDMA channel.
struct srpt_recv_ioctx * recv_ioctx
Receive I/O context.
struct srpt_send_ioctx * send_ioctx
Send I/O context.

Description

Returns 0 if and only if the request will be processed by the target core.

For more information about SRP_TSK_MGMT information units, see also section 6.7 in the SRP r16a document.

bool srpt_handle_new_iu(struct srpt_rdma_ch * ch, struct srpt_recv_ioctx * recv_ioctx)

process a newly received information unit

Parameters

struct srpt_rdma_ch * ch
RDMA channel through which the information unit has been received.
struct srpt_recv_ioctx * recv_ioctx
Receive I/O context associated with the information unit.
void srpt_send_done(struct ib_cq * cq, struct ib_wc * wc)

send completion callback

Parameters

struct ib_cq * cq
Completion queue.
struct ib_wc * wc
Work completion.

Note

Although this has not yet been observed during tests, at least in theory it is possible that the srpt_get_send_ioctx() call invoked by srpt_handle_new_iu() fails. This is possible because the req_lim_delta value in each response is set to one, and it is possible that this response makes the initiator send a new request before the send completion for that response has been processed. This could e.g. happen if the call to srpt_put_send_iotcx() is delayed because of a higher priority interrupt or if IB retransmission causes generation of the send completion to be delayed. Incoming information units for which srpt_get_send_ioctx() fails are queued on cmd_wait_list. The code below processes these delayed requests one at a time.

int srpt_create_ch_ib(struct srpt_rdma_ch * ch)

create receive and send completion queues

Parameters

struct srpt_rdma_ch * ch
SRPT RDMA channel.
bool srpt_close_ch(struct srpt_rdma_ch * ch)

close a RDMA channel

Parameters

struct srpt_rdma_ch * ch
SRPT RDMA channel.

Description

Make sure all resources associated with the channel will be deallocated at an appropriate time.

Returns true if and only if the channel state has been modified into CH_DRAINING.

int srpt_cm_req_recv(struct srpt_device *const sdev, struct ib_cm_id * ib_cm_id, struct rdma_cm_id * rdma_cm_id, u8 port_num, __be16 pkey, const struct srp_login_req * req, const char * src_addr)

process the event IB_CM_REQ_RECEIVED

Parameters

struct srpt_device *const sdev
HCA through which the login request was received.
struct ib_cm_id * ib_cm_id
IB/CM connection identifier in case of IB/CM.
struct rdma_cm_id * rdma_cm_id
RDMA/CM connection identifier in case of RDMA/CM.
u8 port_num
Port through which the REQ message was received.
__be16 pkey
P_Key of the incoming connection.
const struct srp_login_req * req
SRP login request.
const char * src_addr
GID (IB/CM) or IP address (RDMA/CM) of the port that submitted the login request.

Description

Ownership of the cm_id is transferred to the target session if this function returns zero. Otherwise the caller remains the owner of cm_id.

void srpt_cm_rtu_recv(struct srpt_rdma_ch * ch)

process an IB_CM_RTU_RECEIVED or USER_ESTABLISHED event

Parameters

struct srpt_rdma_ch * ch
SRPT RDMA channel.

Description

An RTU (ready to use) message indicates that the connection has been established and that the recipient may begin transmitting.

int srpt_cm_handler(struct ib_cm_id * cm_id, const struct ib_cm_event * event)

IB connection manager callback function

Parameters

struct ib_cm_id * cm_id
IB/CM connection identifier.
const struct ib_cm_event * event
IB/CM event.

Description

A non-zero return value will cause the caller destroy the CM ID.

Note

srpt_cm_handler() must only return a non-zero value when transferring ownership of the cm_id to a channel by srpt_cm_req_recv() failed. Returning a non-zero value in any other case will trigger a race with the ib_destroy_cm_id() call in srpt_release_channel().

void srpt_queue_response(struct se_cmd * cmd)

transmit the response to a SCSI command

Parameters

struct se_cmd * cmd
SCSI target command.

Description

Callback function called by the TCM core. Must not block since it can be invoked on the context of the IB completion handler.

int srpt_release_sport(struct srpt_port * sport)

disable login and wait for associated channels

Parameters

struct srpt_port * sport
SRPT HCA port.
void srpt_add_one(struct ib_device * device)

InfiniBand device addition callback function

Parameters

struct ib_device * device
Describes a HCA.
void srpt_remove_one(struct ib_device * device, void * client_data)

InfiniBand device removal callback function

Parameters

struct ib_device * device
Describes a HCA.
void * client_data
The value passed as the third argument to ib_set_client_data().
void srpt_close_session(struct se_session * se_sess)

forcibly close a session

Parameters

struct se_session * se_sess
SCSI target session.

Description

Callback function invoked by the TCM core to clean up sessions associated with a node ACL when the user invokes rmdir /sys/kernel/config/target/$driver/$port/$tpg/acls/$i_port_id

u32 srpt_sess_get_index(struct se_session * se_sess)

return the value of scsiAttIntrPortIndex (SCSI-MIB)

Parameters

struct se_session * se_sess
SCSI target session.

Description

A quote from RFC 4455 (SCSI-MIB) about this MIB object: This object represents an arbitrary integer used to uniquely identify a particular attached remote initiator port to a particular SCSI target port within a particular SCSI target device within a particular SCSI instance.

int srpt_parse_i_port_id(u8 i_port_id, const char * name)

parse an initiator port ID

Parameters

u8 i_port_id
Binary 128-bit port ID.
const char * name
ASCII representation of a 128-bit initiator port ID.
struct se_portal_group * srpt_make_tpg(struct se_wwn * wwn, const char * name)

configfs callback invoked for mkdir /sys/kernel/config/target/$driver/$port/$tpg

Parameters

struct se_wwn * wwn
Corresponds to $driver/$port.
const char * name
$tpg.
void srpt_drop_tpg(struct se_portal_group * tpg)

configfs callback invoked for rmdir /sys/kernel/config/target/$driver/$port/$tpg

Parameters

struct se_portal_group * tpg
Target portal group to deregister.
struct se_wwn * srpt_make_tport(struct target_fabric_configfs * tf, struct config_group * group, const char * name)

configfs callback invoked for mkdir /sys/kernel/config/target/$driver/$port

Parameters

struct target_fabric_configfs * tf
Not used.
struct config_group * group
Not used.
const char * name
$port.
void srpt_drop_tport(struct se_wwn * wwn)

configfs callback invoked for rmdir /sys/kernel/config/target/$driver/$port

Parameters

struct se_wwn * wwn
$port.
int srpt_init_module(void)

kernel module initialization

Parameters

void
no arguments

Note

Since ib_register_client() registers callback functions, and since at least one of these callback functions (srpt_add_one()) calls target core functions, this driver must be registered with the target core before ib_register_client() is called.

iSCSI Extensions for RDMA (iSER) target support

void isert_conn_terminate(struct isert_conn * isert_conn)

Initiate connection termination

Parameters

struct isert_conn * isert_conn
isert connection struct

Notes

In case the connection state is BOUND, move state to TEMINATING and start teardown sequence (rdma_disconnect). In case the connection state is UP, complete flush as well.

This routine must be called with mutex held. Thus it is safe to call multiple times.

void isert_put_unsol_pending_cmds(struct iscsi_conn * conn)

Drop commands waiting for unsolicitate dataout

Parameters

struct iscsi_conn * conn
iscsi connection

Description

We might still have commands that are waiting for unsolicited dataouts messages. We must put the extra reference on those before blocking on the target_wait_for_session_cmds