File: shadow_queue.h

package info (click to toggle)
rdma-core 61.0-2
  • links: PTS, VCS
  • area: main
  • in suites: forky, sid
  • size: 13,124 kB
  • sloc: ansic: 176,798; python: 15,496; sh: 2,742; perl: 1,465; makefile: 73
file content (159 lines) | stat: -rw-r--r-- 4,222 bytes parent folder | download
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
/* SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB */
/*
 * Copyright (c) 2024, Microsoft Corporation. All rights reserved.
 */

#ifndef _SHADOW_QUEUE_H_
#define _SHADOW_QUEUE_H_

#include <stdio.h>
#include <linux/types.h>
#include <endian.h>
#include <infiniband/verbs.h>
#include <sys/mman.h>
#include <util/util.h>
#include <stdatomic.h>

typedef _Atomic(uint64_t) _atomic_t;

#define MANA_NO_SIGNAL_WC (0xff)

struct shadow_wqe_header {
	/* ibv_wc_opcode */
	uint64_t opcode : 8;
	/* ibv_wc_flags or MANA_NO_SIGNAL_WC */
	uint64_t flags : 8;
	uint64_t posted_wqe_size_in_bu : 8;
	/* ibv_wc_status */
	uint64_t vendor_error : 12;
	uint64_t unmasked_queue_offset : 28;
	uint64_t wr_id;
};

struct rc_sq_shadow_wqe {
	struct  shadow_wqe_header header;
	uint32_t end_psn;
	uint32_t read_posted_wqe_size_in_bu;
};

struct rc_rq_shadow_wqe {
	struct shadow_wqe_header header;
	uint32_t byte_len;
	uint32_t imm_or_rkey;
};

struct shadow_queue {
	uint64_t prod_idx;
	uint64_t cons_idx;
	uint64_t next_to_complete_idx;
	uint64_t next_to_signal_idx;
	uint32_t length;
	uint32_t stride;
	void *buffer;
};

static inline void reset_shadow_queue(struct shadow_queue *queue)
{
	queue->prod_idx = 0;
	queue->cons_idx = 0;
	queue->next_to_complete_idx = 0;
	queue->next_to_signal_idx = 0;
}

int create_shadow_queue(struct shadow_queue *queue, uint32_t length, uint32_t stride);

void destroy_shadow_queue(struct shadow_queue *queue);

static inline _atomic_t *producer(struct shadow_queue *queue)
{
	return (_atomic_t *)&queue->prod_idx;
}

static inline _atomic_t *consumer(struct shadow_queue *queue)
{
	return (_atomic_t *)&queue->cons_idx;
}

static inline struct shadow_wqe_header *
shadow_queue_get_element(struct shadow_queue *queue, uint64_t unmasked_index)
{
	uint32_t index = unmasked_index & (queue->length - 1);

	return (struct shadow_wqe_header *)((uint8_t *)queue->buffer + index * queue->stride);
}

static inline bool shadow_queue_full(struct shadow_queue *queue)
{
	uint64_t prod_idx = atomic_load_explicit(producer(queue), memory_order_relaxed);
	uint64_t cons_idx = atomic_load_explicit(consumer(queue), memory_order_acquire);

	return (prod_idx - cons_idx) >= queue->length;
}

static inline struct shadow_wqe_header *
shadow_queue_producer_entry(struct shadow_queue *queue)
{
	uint64_t prod_idx = atomic_load_explicit(producer(queue), memory_order_relaxed);

	return shadow_queue_get_element(queue, prod_idx);
}

static inline void shadow_queue_advance_producer(struct shadow_queue *queue)
{
	uint64_t prod_idx = atomic_load_explicit(producer(queue), memory_order_relaxed);

	atomic_store_explicit(producer(queue), prod_idx + 1, memory_order_release);
}

static inline void shadow_queue_advance_consumer(struct shadow_queue *queue)
{
	uint64_t cons_idx = atomic_load_explicit(consumer(queue), memory_order_relaxed);

	atomic_store_explicit(consumer(queue), cons_idx + 1, memory_order_release);
}

static inline struct shadow_wqe_header *
shadow_queue_get_next_to_consume(struct shadow_queue *queue)
{
	uint64_t cons_idx = atomic_load_explicit(consumer(queue), memory_order_relaxed);

	if (cons_idx == queue->next_to_complete_idx)
		return NULL;

	return shadow_queue_get_element(queue, cons_idx);
}

static inline struct shadow_wqe_header *
shadow_queue_get_next_to_complete(struct shadow_queue *queue)
{
	uint64_t prod_idx = atomic_load_explicit(producer(queue), memory_order_acquire);

	if (queue->next_to_complete_idx == prod_idx)
		return NULL;

	return shadow_queue_get_element(queue, queue->next_to_complete_idx);
}

static inline void shadow_queue_advance_next_to_complete(struct shadow_queue *queue)
{
	queue->next_to_complete_idx++;
}

static inline struct shadow_wqe_header *
shadow_queue_get_next_to_signal(struct shadow_queue *queue)
{
	uint64_t prod_idx = atomic_load_explicit(producer(queue), memory_order_acquire);
	struct shadow_wqe_header *wqe = NULL;

	queue->next_to_signal_idx = max(queue->next_to_signal_idx, queue->next_to_complete_idx);
	while (queue->next_to_signal_idx < prod_idx) {
		wqe = shadow_queue_get_element(queue, queue->next_to_signal_idx);
		queue->next_to_signal_idx++;
		if (wqe->flags != MANA_NO_SIGNAL_WC)
			return wqe;
	}

	return NULL;
}

#endif //_SHADOW_QUEUE_H_